var/home/core/zuul-output/0000755000175000017500000000000015070411706014526 5ustar corecorevar/home/core/zuul-output/logs/0000755000175000017500000000000015070440216015470 5ustar corecorevar/home/core/zuul-output/logs/kubelet.log0000644000000000000000006674677415070440207017727 0ustar rootrootOct 05 06:47:56 crc systemd[1]: Starting Kubernetes Kubelet... Oct 05 06:47:56 crc restorecon[4742]: Relabeled /var/lib/kubelet/config.json from system_u:object_r:unlabeled_t:s0 to system_u:object_r:container_var_lib_t:s0 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/device-plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/device-plugins/kubelet.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/volumes/kubernetes.io~configmap/nginx-conf/..2025_02_23_05_40_35.4114275528/nginx.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/22e96971 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/21c98286 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/0f1869e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/46889d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/5b6a5969 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/6c7921f5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4804f443 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/2a46b283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/a6b5573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4f88ee5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/5a4eee4b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/cd87c521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/38602af4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/1483b002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/0346718b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/d3ed4ada not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/3bb473a5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/8cd075a9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/00ab4760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/54a21c09 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/70478888 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/43802770 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/955a0edc not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/bca2d009 not reset as customized by admin to system_u:object_r:container_file_t:s0:c140,c1009 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/b295f9bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/bc46ea27 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5731fc1b not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5e1b2a3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/943f0936 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/3f764ee4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/8695e3f9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/aed7aa86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/c64d7448 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/0ba16bd2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/207a939f not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/54aa8cdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/1f5fa595 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/bf9c8153 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/47fba4ea not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/7ae55ce9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7906a268 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/ce43fa69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7fc7ea3a not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/d8c38b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/9ef015fb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/b9db6a41 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/b1733d79 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/afccd338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/9df0a185 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/18938cf8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/7ab4eb23 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/56930be6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_35.630010865 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/0d8e3722 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/d22b2e76 not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/e036759f not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/2734c483 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/57878fe7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/3f3c2e58 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/375bec3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/7bc41e08 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/48c7a72d not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/4b66701f not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/a5a1c202 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_40.1388695756 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/26f3df5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/6d8fb21d not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/50e94777 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208473b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/ec9e08ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3b787c39 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208eaed5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/93aa3a2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3c697968 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/ba950ec9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/cb5cdb37 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/f2df9827 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/fedaa673 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/9ca2df95 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/b2d7460e not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2207853c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/241c1c29 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2d910eaf not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/c6c0f2e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/399edc97 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8049f7cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/0cec5484 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/312446d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c406,c828 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8e56a35d not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/2d30ddb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/eca8053d not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/c3a25c9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c168,c522 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/b9609c22 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/e8b0eca9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/b36a9c3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/38af7b07 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/ae821620 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/baa23338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/2c534809 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/59b29eae not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/c91a8e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/4d87494a not reset as customized by admin to system_u:object_r:container_file_t:s0:c442,c857 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/1e33ca63 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/8dea7be2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d0b04a99 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d84f01e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/4109059b not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/a7258a3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/05bdf2b6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/f3261b51 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/315d045e not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/5fdcf278 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/d053f757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/c2850dc7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fcfb0b2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c7ac9b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fa0c0d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c609b6ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/2be6c296 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/89a32653 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/4eb9afeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/13af6efa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/b03f9724 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/e3d105cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/3aed4d83 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/0765fa6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/2cefc627 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/3dcc6345 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/365af391 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b1130c0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/236a5913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b9432e26 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/5ddb0e3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/986dc4fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/8a23ff9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/9728ae68 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/665f31d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/136c9b42 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/98a1575b not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/cac69136 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/5deb77a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/2ae53400 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/e46f2326 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/dc688d3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/3497c3cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/177eb008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/af5a2afa not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/d780cb1f not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/49b0f374 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/26fbb125 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/cf14125a not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/b7f86972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/e51d739c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/88ba6a69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/669a9acf not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/5cd51231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/75349ec7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/15c26839 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/45023dcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/2bb66a50 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/64d03bdd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/ab8e7ca0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/bb9be25f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/9a0b61d3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/d471b9d2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/8cb76b8e not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/11a00840 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/ec355a92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/992f735e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d59cdbbc not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/72133ff0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/c56c834c not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d13724c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/0a498258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa471982 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fc900d92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa7d68da not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/4bacf9b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/424021b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/fc2e31a3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/f51eefac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/c8997f2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/7481f599 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/fdafea19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/d0e1c571 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/ee398915 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/682bb6b8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a3e67855 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a989f289 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/915431bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/7796fdab not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/dcdb5f19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/a3aaa88c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/5508e3e6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/160585de not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/e99f8da3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/8bc85570 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/a5861c91 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/84db1135 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/9e1a6043 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/c1aba1c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/d55ccd6d not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/971cc9f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/8f2e3dcf not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/ceb35e9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/1c192745 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/5209e501 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/f83de4df not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/e7b978ac not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/c64304a1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/5384386b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/cce3e3ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/8fb75465 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/740f573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/32fd1134 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/0a861bd3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/80363026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/bfa952a8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..2025_02_23_05_33_31.333075221 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 05 06:47:56 crc restorecon[4742]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/793bf43d not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/7db1bb6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/4f6a0368 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/c12c7d86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/36c4a773 not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/4c1e98ae not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/a4c8115c not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/setup/7db1802e not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver/a008a7ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-syncer/2c836bac not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-regeneration-controller/0ce62299 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-insecure-readyz/945d2457 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-check-endpoints/7d5c1dd8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/index.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/bundle-v1.15.0.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/channel.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/package.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/bc8d0691 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/6b76097a not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/34d1af30 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/312ba61c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/645d5dd1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/16e825f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/4cf51fc9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/2a23d348 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/075dbd49 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/dd585ddd not reset as customized by admin to system_u:object_r:container_file_t:s0:c377,c642 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/17ebd0ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c343 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/005579f4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_23_11.1287037894 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/bf5f3b9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/af276eb7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/ea28e322 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/692e6683 not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/871746a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/4eb2e958 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/ca9b62da not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/0edd6fce not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/containers/controller-manager/89b4555f not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/655fcd71 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/0d43c002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/e68efd17 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/9acf9b65 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/5ae3ff11 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/1e59206a not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/27af16d1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c304,c1017 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/7918e729 not reset as customized by admin to system_u:object_r:container_file_t:s0:c853,c893 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/5d976d0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c585,c981 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/d7f55cbb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/f0812073 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/1a56cbeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/7fdd437e not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/cdfb5652 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/fix-audit-permissions/fb93119e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver/f1e8fc0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver-check-endpoints/218511f3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server/serving-certs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/ca8af7b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/72cc8a75 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/6e8a3760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4c3455c0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/2278acb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4b453e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/3ec09bda not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2/cacerts.bin not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java/cacerts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl/ca-bundle.trust.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/email-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/objsign-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2ae6433e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fde84897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75680d2e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/openshift-service-serving-signer_1740288168.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/facfc4fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f5a969c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CFCA_EV_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9ef4a08a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ingress-operator_1740288202.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2f332aed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/248c8271.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d10a21f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ACCVRAIZ1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a94d09e5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c9a4d3b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40193066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd8c0d63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b936d1c6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CA_Disig_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4fd49c6c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM_SERVIDORES_SEGUROS.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b81b93f0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f9a69fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b30d5fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ANF_Secure_Server_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b433981b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93851c9e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9282e51c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7dd1bc4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Actalis_Authentication_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/930ac5d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f47b495.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e113c810.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5931b5bc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Commercial.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2b349938.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e48193cf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/302904dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a716d4ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Networking.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93bc0acc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/86212b19.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b727005e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbc54cab.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f51bb24c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c28a8a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9c8dfbd4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ccc52f49.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cb1c3204.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ce5e74ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd08c599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6d41d539.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb5fa911.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e35234b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8cb5ee0f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a7c655d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f8fc53da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/de6d66f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d41b5e2a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/41a3f684.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1df5a75f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_2011.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e36a6752.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b872f2b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9576d26b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/228f89db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_ECC_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb717492.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d21b73c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b1b94ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/595e996b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_RSA_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b46e03d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/128f4b91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_3_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81f2d2b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Autoridad_de_Certificacion_Firmaprofesional_CIF_A62634068.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3bde41ac.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d16a5865.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_EC-384_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0179095f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ffa7f1eb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9482e63a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4dae3dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e359ba6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7e067d03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/95aff9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7746a63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Baltimore_CyberTrust_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/653b494a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3ad48a91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_2_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/54657681.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/82223c44.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8de2f56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d9dafe4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d96b65e2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee64a828.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40547a79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5a3f0ff8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a780d93.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/34d996fb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/eed8c118.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/89c02a45.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b1159c4c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d6325660.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4c339cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8312c4c1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_E1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8508e720.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5fdd185d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48bec511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/69105f4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b9bc432.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/32888f65.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b03dec0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/219d9499.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5acf816d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbf06781.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc99f41e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AAA_Certificate_Services.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/985c1f52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8794b4e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_BR_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7c037b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ef954a4e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_EV_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2add47b6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/90c5a3c8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0f3e76e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/53a1b57a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_EV_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5ad8a5d6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/68dd7389.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d04f354.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d6437c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/062cdee6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bd43e1dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7f3d5d1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c491639e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3513523f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/399e7759.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/feffd413.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d18e9066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/607986c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c90bc37d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1b0f7e5c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e08bfd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dd8e9d41.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed39abd0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a3418fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bc3f2570.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_High_Assurance_EV_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/244b5494.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81b9768f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4be590e0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_ECC_P384_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9846683b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/252252d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e8e7201.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_RSA4096_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d52c538d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c44cc0c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Trusted_Root_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75d1b2ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a2c66da8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ecccd8db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust.net_Certification_Authority__2048_.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/aee5f10d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e7271e8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0e59380.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4c3982f2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b99d060.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf64f35b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0a775a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/002c0b4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cc450945.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_EC1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/106f3e4d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b3fb433b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4042bcee.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/02265526.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/455f1b52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0d69c7e1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9f727ac7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5e98733a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0cd152c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc4d6a89.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6187b673.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/FIRMAPROFESIONAL_CA_ROOT-A_WEB.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ba8887ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/068570d1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f081611a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48a195d8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GDCA_TrustAUTH_R5_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f6fa695.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab59055e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b92fd57f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GLOBALTRUST_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fa5da96b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ec40989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7719f463.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1001acf7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f013ecaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/626dceaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c559d742.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1d3472b9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9479c8c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a81e292b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4bfab552.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e071171e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/57bcb2da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_ECC_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab5346f4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5046c355.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_RSA_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/865fbdf9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da0cfd1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/85cde254.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_ECC_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbb3f32b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureSign_RootCA11.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5860aaa6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/31188b5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HiPKI_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c7f1359b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f15c80c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hongkong_Post_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/09789157.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/18856ac4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e09d511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Commercial_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cf701eeb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d06393bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Public_Sector_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/10531352.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Izenpe.com.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureTrust_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0ed035a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsec_e-Szigno_Root_CA_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8160b96c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8651083.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2c63f966.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_ECC_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d89cda1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/01419da9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_RSA_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7a5b843.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_RSA_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf53fb88.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9591a472.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3afde786.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Gold_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NAVER_Global_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3fb36b73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d39b0a2c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a89d74c2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd58d51e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7db1890.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NetLock_Arany__Class_Gold__F__tan__s__tv__ny.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/988a38cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/60afe812.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f39fc864.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5443e9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GB_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e73d606e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dfc0fe80.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b66938e9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e1eab7c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GC_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/773e07ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c899c73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d59297b8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ddcda989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_1_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/749e9e03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/52b525c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7e8dc79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a819ef2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/08063a00.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b483515.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/064e0aa9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1f58a078.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6f7454b3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7fa05551.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76faf6c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9339512a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f387163d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee37c333.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e18bfb83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e442e424.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fe8a2cd8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/23f4c490.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5cd81ad7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0c70a8d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7892ad52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SZAFIR_ROOT_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4f316efb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_RSA_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/06dc52d5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/583d0756.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0bf05006.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/88950faa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9046744a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c860d51.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_RSA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6fa5da56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/33ee480d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Secure_Global_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/63a2c897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_ECC_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bdacca6f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ff34af3f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbff3a01.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_ECC_RootCA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_C1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/406c9bb1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_C3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Services_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Silver_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/99e1b953.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/14bc7599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TUBITAK_Kamu_SM_SSL_Kok_Sertifikasi_-_Surum_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a3adc42.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f459871d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_ECC_Root_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_RSA_Root_2023.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TeliaSonera_Root_CA_v1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telia_Root_CA_v2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f103249.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f058632f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-certificates.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9bf03295.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/98aaf404.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1cef98f5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/073bfcc5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2923b3f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f249de83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/edcbddb5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P256_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b5697b0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ae85e5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b74d2bd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P384_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d887a5bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9aef356c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TunTrust_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd64f3fc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e13665f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Extended_Validation_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f5dc4f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da7377f6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Global_G2_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c01eb047.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/304d27c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed858448.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f30dd6ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/04f60c28.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_ECC_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fc5a8f99.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/35105088.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee532fd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/XRamp_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/706f604c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76579174.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d86cdd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/882de061.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f618aec.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a9d40e02.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e-Szigno_Root_CA_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e868b802.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/83e9984f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ePKI_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca6e4ad9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d6523ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4b718d9b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/869fbf79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/containers/registry/f8d22bdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/6e8bbfac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/54dd7996 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/a4f1bb05 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/207129da not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/c1df39e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/15b8f1cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/77bd6913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/2382c1b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/704ce128 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/70d16fe0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/bfb95535 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/57a8e8e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/1b9d3e5e not reset as customized by admin to system_u:object_r:container_file_t:s0:c107,c917 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/fddb173c not reset as customized by admin to system_u:object_r:container_file_t:s0:c202,c983 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/95d3c6c4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/bfb5fff5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/2aef40aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/c0391cad not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/1119e69d not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/660608b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/8220bd53 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/85f99d5c not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/4b0225f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/9c2a3394 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/e820b243 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/1ca52ea0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/e6988e45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/6655f00b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/98bc3986 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/08e3458a not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/2a191cb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/6c4eeefb not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/f61a549c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/24891863 not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/fbdfd89c not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/9b63b3bc not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/8acde6d6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/node-driver-registrar/59ecbba3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/csi-provisioner/685d4be3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/containers/route-controller-manager/feaea55e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/63709497 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/d966b7fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/f5773757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/81c9edb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/57bf57ee not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/86f5e6aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/0aabe31d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/d2af85c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/09d157d9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c0fe7256 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c30319e4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/e6b1dd45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/2bb643f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/920de426 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/70fa1e87 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/a1c12a2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/9442e6c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/5b45ec72 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/3c9f3a59 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/1091c11b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/9a6821c6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/ec0c35e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/517f37e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/6214fe78 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/ba189c8b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/351e4f31 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/c0f219ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/8069f607 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/559c3d82 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/605ad488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/148df488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/3bf6dcb4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/022a2feb not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/938c3924 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/729fe23e not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/1fd5cbd4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/a96697e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/e155ddca not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/10dd0e0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/6f2c8392 not reset as customized by admin to system_u:object_r:container_file_t:s0:c267,c588 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/bd241ad9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/plugins/csi-hostpath not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/plugins/csi-hostpath/csi.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/plugins/kubernetes.io not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/plugins/kubernetes.io/csi not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983 not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/vol_data.json not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 05 06:47:57 crc restorecon[4742]: /var/lib/kubelet/plugins_registry not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 05 06:47:57 crc restorecon[4742]: Relabeled /var/usrlocal/bin/kubenswrapper from system_u:object_r:bin_t:s0 to system_u:object_r:kubelet_exec_t:s0 Oct 05 06:47:58 crc kubenswrapper[4846]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 05 06:47:58 crc kubenswrapper[4846]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Oct 05 06:47:58 crc kubenswrapper[4846]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 05 06:47:58 crc kubenswrapper[4846]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 05 06:47:58 crc kubenswrapper[4846]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Oct 05 06:47:58 crc kubenswrapper[4846]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.223645 4846 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.243019 4846 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.243084 4846 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.243099 4846 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.243109 4846 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.243118 4846 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.243127 4846 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.243136 4846 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.243146 4846 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.243157 4846 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.243167 4846 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.243205 4846 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.243213 4846 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.243221 4846 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.243229 4846 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.243237 4846 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.243248 4846 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.243258 4846 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.243268 4846 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.243278 4846 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.243288 4846 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.243298 4846 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.243309 4846 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.243320 4846 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.243330 4846 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.243341 4846 feature_gate.go:330] unrecognized feature gate: Example Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.243349 4846 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.243358 4846 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.243366 4846 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.243387 4846 feature_gate.go:330] unrecognized feature gate: SignatureStores Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.243395 4846 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.243403 4846 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.243411 4846 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.243420 4846 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.243430 4846 feature_gate.go:330] unrecognized feature gate: GatewayAPI Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.243437 4846 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.243445 4846 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.243453 4846 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.243461 4846 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.243469 4846 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.243477 4846 feature_gate.go:330] unrecognized feature gate: OVNObservability Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.243484 4846 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.243494 4846 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.243502 4846 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.243509 4846 feature_gate.go:330] unrecognized feature gate: InsightsConfig Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.243520 4846 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.243530 4846 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.243538 4846 feature_gate.go:330] unrecognized feature gate: PlatformOperators Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.243547 4846 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.243557 4846 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.243567 4846 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.243575 4846 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.243583 4846 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.243592 4846 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.243601 4846 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.243610 4846 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.243619 4846 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.243627 4846 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.243636 4846 feature_gate.go:330] unrecognized feature gate: NewOLM Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.243644 4846 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.243661 4846 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.243669 4846 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.243677 4846 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.243685 4846 feature_gate.go:330] unrecognized feature gate: PinnedImages Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.243693 4846 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.243701 4846 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.243708 4846 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.243716 4846 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.243724 4846 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.243734 4846 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.243742 4846 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.243749 4846 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.243951 4846 flags.go:64] FLAG: --address="0.0.0.0" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.243971 4846 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.243989 4846 flags.go:64] FLAG: --anonymous-auth="true" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.244014 4846 flags.go:64] FLAG: --application-metrics-count-limit="100" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.244027 4846 flags.go:64] FLAG: --authentication-token-webhook="false" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.244037 4846 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.244049 4846 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.244060 4846 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.244070 4846 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.244079 4846 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.244089 4846 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.244101 4846 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.244112 4846 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.244121 4846 flags.go:64] FLAG: --cgroup-root="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.244130 4846 flags.go:64] FLAG: --cgroups-per-qos="true" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.244139 4846 flags.go:64] FLAG: --client-ca-file="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.244148 4846 flags.go:64] FLAG: --cloud-config="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.244157 4846 flags.go:64] FLAG: --cloud-provider="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.244166 4846 flags.go:64] FLAG: --cluster-dns="[]" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.244211 4846 flags.go:64] FLAG: --cluster-domain="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.244225 4846 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.244234 4846 flags.go:64] FLAG: --config-dir="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.244243 4846 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.244254 4846 flags.go:64] FLAG: --container-log-max-files="5" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.244266 4846 flags.go:64] FLAG: --container-log-max-size="10Mi" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.244276 4846 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.244285 4846 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.244295 4846 flags.go:64] FLAG: --containerd-namespace="k8s.io" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.244305 4846 flags.go:64] FLAG: --contention-profiling="false" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.244314 4846 flags.go:64] FLAG: --cpu-cfs-quota="true" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.244323 4846 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.244333 4846 flags.go:64] FLAG: --cpu-manager-policy="none" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.244341 4846 flags.go:64] FLAG: --cpu-manager-policy-options="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.244354 4846 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.244364 4846 flags.go:64] FLAG: --enable-controller-attach-detach="true" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.244373 4846 flags.go:64] FLAG: --enable-debugging-handlers="true" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.244381 4846 flags.go:64] FLAG: --enable-load-reader="false" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.244391 4846 flags.go:64] FLAG: --enable-server="true" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.244401 4846 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.244414 4846 flags.go:64] FLAG: --event-burst="100" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.244423 4846 flags.go:64] FLAG: --event-qps="50" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.244432 4846 flags.go:64] FLAG: --event-storage-age-limit="default=0" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.244442 4846 flags.go:64] FLAG: --event-storage-event-limit="default=0" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.244450 4846 flags.go:64] FLAG: --eviction-hard="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.244463 4846 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.244472 4846 flags.go:64] FLAG: --eviction-minimum-reclaim="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.244484 4846 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.244494 4846 flags.go:64] FLAG: --eviction-soft="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.244503 4846 flags.go:64] FLAG: --eviction-soft-grace-period="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.244513 4846 flags.go:64] FLAG: --exit-on-lock-contention="false" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.244522 4846 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.244531 4846 flags.go:64] FLAG: --experimental-mounter-path="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.244543 4846 flags.go:64] FLAG: --fail-cgroupv1="false" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.244552 4846 flags.go:64] FLAG: --fail-swap-on="true" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.244561 4846 flags.go:64] FLAG: --feature-gates="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.244581 4846 flags.go:64] FLAG: --file-check-frequency="20s" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.244590 4846 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.244600 4846 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.244609 4846 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.244619 4846 flags.go:64] FLAG: --healthz-port="10248" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.244628 4846 flags.go:64] FLAG: --help="false" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.244638 4846 flags.go:64] FLAG: --hostname-override="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.244647 4846 flags.go:64] FLAG: --housekeeping-interval="10s" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.244657 4846 flags.go:64] FLAG: --http-check-frequency="20s" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.244666 4846 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.244676 4846 flags.go:64] FLAG: --image-credential-provider-config="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.244686 4846 flags.go:64] FLAG: --image-gc-high-threshold="85" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.244695 4846 flags.go:64] FLAG: --image-gc-low-threshold="80" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.244704 4846 flags.go:64] FLAG: --image-service-endpoint="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.244713 4846 flags.go:64] FLAG: --kernel-memcg-notification="false" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.244722 4846 flags.go:64] FLAG: --kube-api-burst="100" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.244731 4846 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.244741 4846 flags.go:64] FLAG: --kube-api-qps="50" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.244749 4846 flags.go:64] FLAG: --kube-reserved="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.244758 4846 flags.go:64] FLAG: --kube-reserved-cgroup="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.244767 4846 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.244777 4846 flags.go:64] FLAG: --kubelet-cgroups="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.244786 4846 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.244795 4846 flags.go:64] FLAG: --lock-file="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.244804 4846 flags.go:64] FLAG: --log-cadvisor-usage="false" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.244813 4846 flags.go:64] FLAG: --log-flush-frequency="5s" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.244822 4846 flags.go:64] FLAG: --log-json-info-buffer-size="0" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.244836 4846 flags.go:64] FLAG: --log-json-split-stream="false" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.244846 4846 flags.go:64] FLAG: --log-text-info-buffer-size="0" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.244858 4846 flags.go:64] FLAG: --log-text-split-stream="false" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.244867 4846 flags.go:64] FLAG: --logging-format="text" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.244877 4846 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.244886 4846 flags.go:64] FLAG: --make-iptables-util-chains="true" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.244895 4846 flags.go:64] FLAG: --manifest-url="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.244904 4846 flags.go:64] FLAG: --manifest-url-header="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.244917 4846 flags.go:64] FLAG: --max-housekeeping-interval="15s" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.244926 4846 flags.go:64] FLAG: --max-open-files="1000000" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.244938 4846 flags.go:64] FLAG: --max-pods="110" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.244947 4846 flags.go:64] FLAG: --maximum-dead-containers="-1" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.244956 4846 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.244965 4846 flags.go:64] FLAG: --memory-manager-policy="None" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.244974 4846 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.244983 4846 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.244993 4846 flags.go:64] FLAG: --node-ip="192.168.126.11" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.245002 4846 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.245038 4846 flags.go:64] FLAG: --node-status-max-images="50" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.245048 4846 flags.go:64] FLAG: --node-status-update-frequency="10s" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.245057 4846 flags.go:64] FLAG: --oom-score-adj="-999" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.245066 4846 flags.go:64] FLAG: --pod-cidr="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.245076 4846 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.245089 4846 flags.go:64] FLAG: --pod-manifest-path="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.245098 4846 flags.go:64] FLAG: --pod-max-pids="-1" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.245108 4846 flags.go:64] FLAG: --pods-per-core="0" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.245117 4846 flags.go:64] FLAG: --port="10250" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.245126 4846 flags.go:64] FLAG: --protect-kernel-defaults="false" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.245135 4846 flags.go:64] FLAG: --provider-id="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.245144 4846 flags.go:64] FLAG: --qos-reserved="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.245153 4846 flags.go:64] FLAG: --read-only-port="10255" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.245162 4846 flags.go:64] FLAG: --register-node="true" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.245170 4846 flags.go:64] FLAG: --register-schedulable="true" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.245209 4846 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.245230 4846 flags.go:64] FLAG: --registry-burst="10" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.245239 4846 flags.go:64] FLAG: --registry-qps="5" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.245247 4846 flags.go:64] FLAG: --reserved-cpus="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.245257 4846 flags.go:64] FLAG: --reserved-memory="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.245269 4846 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.245279 4846 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.245288 4846 flags.go:64] FLAG: --rotate-certificates="false" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.245297 4846 flags.go:64] FLAG: --rotate-server-certificates="false" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.245306 4846 flags.go:64] FLAG: --runonce="false" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.245315 4846 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.245325 4846 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.245335 4846 flags.go:64] FLAG: --seccomp-default="false" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.245344 4846 flags.go:64] FLAG: --serialize-image-pulls="true" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.245353 4846 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.245363 4846 flags.go:64] FLAG: --storage-driver-db="cadvisor" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.245376 4846 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.245386 4846 flags.go:64] FLAG: --storage-driver-password="root" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.245395 4846 flags.go:64] FLAG: --storage-driver-secure="false" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.245405 4846 flags.go:64] FLAG: --storage-driver-table="stats" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.245414 4846 flags.go:64] FLAG: --storage-driver-user="root" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.245422 4846 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.245433 4846 flags.go:64] FLAG: --sync-frequency="1m0s" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.245442 4846 flags.go:64] FLAG: --system-cgroups="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.245451 4846 flags.go:64] FLAG: --system-reserved="cpu=200m,ephemeral-storage=350Mi,memory=350Mi" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.245466 4846 flags.go:64] FLAG: --system-reserved-cgroup="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.245475 4846 flags.go:64] FLAG: --tls-cert-file="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.245484 4846 flags.go:64] FLAG: --tls-cipher-suites="[]" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.245494 4846 flags.go:64] FLAG: --tls-min-version="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.245503 4846 flags.go:64] FLAG: --tls-private-key-file="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.245513 4846 flags.go:64] FLAG: --topology-manager-policy="none" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.245521 4846 flags.go:64] FLAG: --topology-manager-policy-options="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.245531 4846 flags.go:64] FLAG: --topology-manager-scope="container" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.245543 4846 flags.go:64] FLAG: --v="2" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.245556 4846 flags.go:64] FLAG: --version="false" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.245568 4846 flags.go:64] FLAG: --vmodule="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.245580 4846 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.245590 4846 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.245814 4846 feature_gate.go:330] unrecognized feature gate: SignatureStores Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.245825 4846 feature_gate.go:330] unrecognized feature gate: NewOLM Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.245835 4846 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.245843 4846 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.245851 4846 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.245859 4846 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.245867 4846 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.245875 4846 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.245883 4846 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.245893 4846 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.245907 4846 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.245917 4846 feature_gate.go:330] unrecognized feature gate: Example Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.245926 4846 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.245935 4846 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.245943 4846 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.245951 4846 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.245959 4846 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.245966 4846 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.245974 4846 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.245982 4846 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.245990 4846 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.245998 4846 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.246005 4846 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.246014 4846 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.246022 4846 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.246030 4846 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.246039 4846 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.246050 4846 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.246058 4846 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.246066 4846 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.246074 4846 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.246085 4846 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.246095 4846 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.246104 4846 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.246113 4846 feature_gate.go:330] unrecognized feature gate: OVNObservability Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.246122 4846 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.246131 4846 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.246139 4846 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.246149 4846 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.246157 4846 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.246167 4846 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.246204 4846 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.246217 4846 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.246225 4846 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.246234 4846 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.246243 4846 feature_gate.go:330] unrecognized feature gate: InsightsConfig Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.246253 4846 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.246262 4846 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.246271 4846 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.246279 4846 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.246287 4846 feature_gate.go:330] unrecognized feature gate: GatewayAPI Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.246295 4846 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.246303 4846 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.246310 4846 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.246319 4846 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.246327 4846 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.246334 4846 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.246343 4846 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.246351 4846 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.246362 4846 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.246369 4846 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.246377 4846 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.246385 4846 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.246393 4846 feature_gate.go:330] unrecognized feature gate: PlatformOperators Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.246401 4846 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.246409 4846 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.246417 4846 feature_gate.go:330] unrecognized feature gate: PinnedImages Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.246424 4846 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.246432 4846 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.246442 4846 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.246453 4846 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.246467 4846 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.260858 4846 server.go:491] "Kubelet version" kubeletVersion="v1.31.5" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.260909 4846 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.261116 4846 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.261152 4846 feature_gate.go:330] unrecognized feature gate: PinnedImages Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.261166 4846 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.261215 4846 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.261226 4846 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.261237 4846 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.261247 4846 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.261258 4846 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.261269 4846 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.261280 4846 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.261289 4846 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.261300 4846 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.261311 4846 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.261321 4846 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.261331 4846 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.261341 4846 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.261351 4846 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.261362 4846 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.261372 4846 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.261382 4846 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.261391 4846 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.261402 4846 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.261412 4846 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.261423 4846 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.261432 4846 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.261443 4846 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.261453 4846 feature_gate.go:330] unrecognized feature gate: Example Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.261463 4846 feature_gate.go:330] unrecognized feature gate: OVNObservability Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.261473 4846 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.261482 4846 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.261492 4846 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.261502 4846 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.261512 4846 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.261527 4846 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.261545 4846 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.261556 4846 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.261567 4846 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.261582 4846 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.261595 4846 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.261608 4846 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.261618 4846 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.261629 4846 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.261639 4846 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.261650 4846 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.261660 4846 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.261671 4846 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.261681 4846 feature_gate.go:330] unrecognized feature gate: SignatureStores Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.261691 4846 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.261701 4846 feature_gate.go:330] unrecognized feature gate: NewOLM Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.261712 4846 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.261723 4846 feature_gate.go:330] unrecognized feature gate: PlatformOperators Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.261734 4846 feature_gate.go:330] unrecognized feature gate: InsightsConfig Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.261744 4846 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.261754 4846 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.261765 4846 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.261776 4846 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.261789 4846 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.261800 4846 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.261810 4846 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.261825 4846 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.261837 4846 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.261848 4846 feature_gate.go:330] unrecognized feature gate: GatewayAPI Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.261861 4846 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.261874 4846 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.261886 4846 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.261897 4846 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.261908 4846 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.261918 4846 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.261928 4846 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.261939 4846 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.261966 4846 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.261984 4846 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.262428 4846 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.262449 4846 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.262461 4846 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.262473 4846 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.262484 4846 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.262495 4846 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.262504 4846 feature_gate.go:330] unrecognized feature gate: InsightsConfig Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.262516 4846 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.262526 4846 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.262536 4846 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.262547 4846 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.262558 4846 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.262568 4846 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.262578 4846 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.262589 4846 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.262601 4846 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.262612 4846 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.262623 4846 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.262634 4846 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.262645 4846 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.262656 4846 feature_gate.go:330] unrecognized feature gate: PinnedImages Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.262666 4846 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.262677 4846 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.262708 4846 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.262720 4846 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.262730 4846 feature_gate.go:330] unrecognized feature gate: SignatureStores Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.262740 4846 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.262751 4846 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.262761 4846 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.262772 4846 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.262781 4846 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.262791 4846 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.262801 4846 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.262812 4846 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.262840 4846 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.262852 4846 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.262863 4846 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.262873 4846 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.262884 4846 feature_gate.go:330] unrecognized feature gate: GatewayAPI Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.262895 4846 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.262906 4846 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.262916 4846 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.262926 4846 feature_gate.go:330] unrecognized feature gate: Example Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.262936 4846 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.262947 4846 feature_gate.go:330] unrecognized feature gate: OVNObservability Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.262957 4846 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.262967 4846 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.262977 4846 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.262989 4846 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.263002 4846 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.263012 4846 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.263023 4846 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.263033 4846 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.263044 4846 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.263054 4846 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.263065 4846 feature_gate.go:330] unrecognized feature gate: PlatformOperators Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.263079 4846 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.263093 4846 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.263104 4846 feature_gate.go:330] unrecognized feature gate: NewOLM Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.263115 4846 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.263126 4846 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.263139 4846 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.263153 4846 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.263167 4846 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.263207 4846 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.263220 4846 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.263230 4846 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.263244 4846 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.263258 4846 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.263270 4846 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.263286 4846 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.263303 4846 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.263647 4846 server.go:940] "Client rotation is on, will bootstrap in background" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.271846 4846 bootstrap.go:85] "Current kubeconfig file contents are still valid, no bootstrap necessary" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.272057 4846 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.273880 4846 server.go:997] "Starting client certificate rotation" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.273938 4846 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.274776 4846 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-02-24 05:52:08 +0000 UTC, rotation deadline is 2026-01-10 13:54:21.079398098 +0000 UTC Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.274924 4846 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 2335h6m22.804479573s for next certificate rotation Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.300264 4846 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.303371 4846 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.323579 4846 log.go:25] "Validated CRI v1 runtime API" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.362721 4846 log.go:25] "Validated CRI v1 image API" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.365416 4846 server.go:1437] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.372613 4846 fs.go:133] Filesystem UUIDs: map[0b076daa-c26a-46d2-b3a6-72a8dbc6e257:/dev/vda4 2025-10-05-06-43-28-00:/dev/sr0 7B77-95E7:/dev/vda2 de0497b0-db1b-465a-b278-03db02455c71:/dev/vda3] Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.372661 4846 fs.go:134] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/user/1000:{mountpoint:/run/user/1000 major:0 minor:42 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:30 fsType:tmpfs blockSize:0} /var/lib/etcd:{mountpoint:/var/lib/etcd major:0 minor:43 fsType:tmpfs blockSize:0}] Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.395220 4846 manager.go:217] Machine: {Timestamp:2025-10-05 06:47:58.392984915 +0000 UTC m=+0.633837710 CPUVendorID:AuthenticAMD NumCores:12 NumPhysicalCores:1 NumSockets:12 CpuFrequency:2800000 MemoryCapacity:33654128640 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:21801e6708c44f15b81395eb736a7cec SystemUUID:a050a23b-e773-4ba9-989b-360aa53a0605 BootID:04ce628d-2f61-4931-aad6-615924ea53fb Filesystems:[{Device:/var/lib/etcd DeviceMajor:0 DeviceMinor:43 Capacity:1073741824 Type:vfs Inodes:4108170 HasInodes:true} {Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:16827064320 Type:vfs Inodes:4108170 HasInodes:true} {Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:6730825728 Type:vfs Inodes:819200 HasInodes:true} {Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true} {Device:/tmp DeviceMajor:0 DeviceMinor:30 Capacity:16827064320 Type:vfs Inodes:1048576 HasInodes:true} {Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true} {Device:/run/user/1000 DeviceMajor:0 DeviceMinor:42 Capacity:3365412864 Type:vfs Inodes:821634 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:214748364800 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:fa:16:3e:02:4a:f1 Speed:0 Mtu:1500} {Name:br-int MacAddress:d6:39:55:2e:22:71 Speed:0 Mtu:1400} {Name:ens3 MacAddress:fa:16:3e:02:4a:f1 Speed:-1 Mtu:1500} {Name:ens7 MacAddress:fa:16:3e:d0:fc:c9 Speed:-1 Mtu:1500} {Name:ens7.20 MacAddress:52:54:00:36:cb:a9 Speed:-1 Mtu:1496} {Name:ens7.21 MacAddress:52:54:00:a1:2f:c6 Speed:-1 Mtu:1496} {Name:ens7.22 MacAddress:52:54:00:53:11:cc Speed:-1 Mtu:1496} {Name:ens7.23 MacAddress:52:54:00:ae:e9:81 Speed:-1 Mtu:1496} {Name:eth10 MacAddress:3e:1a:79:fe:c9:1b Speed:0 Mtu:1500} {Name:ovn-k8s-mp0 MacAddress:0a:58:0a:d9:00:02 Speed:0 Mtu:1400} {Name:ovs-system MacAddress:86:e5:6a:d1:06:10 Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:33654128640 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[10] Caches:[{Id:10 Size:32768 Type:Data Level:1} {Id:10 Size:32768 Type:Instruction Level:1} {Id:10 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:10 Size:16777216 Type:Unified Level:3}] SocketID:10 BookID: DrawerID:} {Id:0 Threads:[11] Caches:[{Id:11 Size:32768 Type:Data Level:1} {Id:11 Size:32768 Type:Instruction Level:1} {Id:11 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:11 Size:16777216 Type:Unified Level:3}] SocketID:11 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:32768 Type:Data Level:1} {Id:4 Size:32768 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:32768 Type:Data Level:1} {Id:5 Size:32768 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:32768 Type:Data Level:1} {Id:6 Size:32768 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:32768 Type:Data Level:1} {Id:7 Size:32768 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:} {Id:0 Threads:[8] Caches:[{Id:8 Size:32768 Type:Data Level:1} {Id:8 Size:32768 Type:Instruction Level:1} {Id:8 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:8 Size:16777216 Type:Unified Level:3}] SocketID:8 BookID: DrawerID:} {Id:0 Threads:[9] Caches:[{Id:9 Size:32768 Type:Data Level:1} {Id:9 Size:32768 Type:Instruction Level:1} {Id:9 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:9 Size:16777216 Type:Unified Level:3}] SocketID:9 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.395459 4846 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.395652 4846 manager.go:233] Version: {KernelVersion:5.14.0-427.50.2.el9_4.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 418.94.202502100215-0 DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.397196 4846 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.397384 4846 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.397422 4846 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"crc","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"200m","ephemeral-storage":"350Mi","memory":"350Mi"},"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.397658 4846 topology_manager.go:138] "Creating topology manager with none policy" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.397670 4846 container_manager_linux.go:303] "Creating device plugin manager" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.398096 4846 manager.go:142] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.398132 4846 server.go:66] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.398552 4846 state_mem.go:36] "Initialized new in-memory state store" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.398655 4846 server.go:1245] "Using root directory" path="/var/lib/kubelet" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.401601 4846 kubelet.go:418] "Attempting to sync node with API server" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.401624 4846 kubelet.go:313] "Adding static pod path" path="/etc/kubernetes/manifests" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.401654 4846 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.401674 4846 kubelet.go:324] "Adding apiserver pod source" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.401691 4846 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.408241 4846 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="cri-o" version="1.31.5-4.rhaos4.18.gitdad78d5.el9" apiVersion="v1" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.410441 4846 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem". Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.411136 4846 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.111:6443: connect: connection refused Oct 05 06:47:58 crc kubenswrapper[4846]: E1005 06:47:58.411266 4846 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.111:6443: connect: connection refused" logger="UnhandledError" Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.411241 4846 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.111:6443: connect: connection refused Oct 05 06:47:58 crc kubenswrapper[4846]: E1005 06:47:58.411466 4846 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.111:6443: connect: connection refused" logger="UnhandledError" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.411884 4846 kubelet.go:854] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.413621 4846 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.413663 4846 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.413678 4846 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.413694 4846 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.413717 4846 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.413731 4846 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/secret" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.413745 4846 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.413767 4846 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.413784 4846 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/fc" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.413800 4846 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.413837 4846 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/projected" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.413856 4846 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.414707 4846 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/csi" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.415526 4846 server.go:1280] "Started kubelet" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.415863 4846 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.415866 4846 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.416711 4846 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.417659 4846 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.111:6443: connect: connection refused Oct 05 06:47:58 crc systemd[1]: Started Kubernetes Kubelet. Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.421710 4846 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate rotation is enabled Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.421758 4846 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.421986 4846 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-25 19:48:47.027129094 +0000 UTC Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.426716 4846 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 1957h0m48.600426692s for next certificate rotation Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.426905 4846 server.go:460] "Adding debug handlers to kubelet server" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.427260 4846 volume_manager.go:287] "The desired_state_of_world populator starts" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.427303 4846 volume_manager.go:289] "Starting Kubelet Volume Manager" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.427631 4846 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Oct 05 06:47:58 crc kubenswrapper[4846]: E1005 06:47:58.427660 4846 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Oct 05 06:47:58 crc kubenswrapper[4846]: E1005 06:47:58.428964 4846 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.111:6443: connect: connection refused" interval="200ms" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.429690 4846 factory.go:55] Registering systemd factory Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.429739 4846 factory.go:221] Registration of the systemd container factory successfully Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.430551 4846 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.111:6443: connect: connection refused Oct 05 06:47:58 crc kubenswrapper[4846]: E1005 06:47:58.430816 4846 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.111:6443: connect: connection refused" logger="UnhandledError" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.430959 4846 factory.go:153] Registering CRI-O factory Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.430999 4846 factory.go:221] Registration of the crio container factory successfully Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.431092 4846 factory.go:219] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.431123 4846 factory.go:103] Registering Raw factory Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.431149 4846 manager.go:1196] Started watching for new ooms in manager Oct 05 06:47:58 crc kubenswrapper[4846]: E1005 06:47:58.430325 4846 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.111:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.186b85834e1446ca default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-10-05 06:47:58.41547233 +0000 UTC m=+0.656325175,LastTimestamp:2025-10-05 06:47:58.41547233 +0000 UTC m=+0.656325175,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.435584 4846 manager.go:319] Starting recovery of all containers Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.449588 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" seLinuxMountContext="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.449702 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides" seLinuxMountContext="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.449738 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" seLinuxMountContext="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.449767 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf" seLinuxMountContext="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.449807 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" seLinuxMountContext="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.449835 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" seLinuxMountContext="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.449861 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" seLinuxMountContext="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.449894 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" seLinuxMountContext="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.449926 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" seLinuxMountContext="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.450050 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" seLinuxMountContext="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.450077 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" seLinuxMountContext="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.450111 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" seLinuxMountContext="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.450138 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" seLinuxMountContext="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.450171 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" seLinuxMountContext="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.450245 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" seLinuxMountContext="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.450342 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" seLinuxMountContext="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.450366 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" seLinuxMountContext="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.450388 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" volumeName="kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" seLinuxMountContext="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.450407 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" seLinuxMountContext="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.450427 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" seLinuxMountContext="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.450449 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" seLinuxMountContext="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.450469 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" seLinuxMountContext="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.450494 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" seLinuxMountContext="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.450514 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" seLinuxMountContext="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.450536 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" seLinuxMountContext="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.450558 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" seLinuxMountContext="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.450582 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d751cbb-f2e2-430d-9754-c882a5e924a5" volumeName="kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl" seLinuxMountContext="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.450618 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" seLinuxMountContext="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.450644 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" seLinuxMountContext="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.450667 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script" seLinuxMountContext="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.450686 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert" seLinuxMountContext="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.450707 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" seLinuxMountContext="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.450758 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" seLinuxMountContext="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.450779 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" seLinuxMountContext="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.450801 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" seLinuxMountContext="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.450821 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" seLinuxMountContext="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.450861 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" seLinuxMountContext="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.450891 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" seLinuxMountContext="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.450918 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" seLinuxMountContext="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.450948 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" seLinuxMountContext="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.450975 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" seLinuxMountContext="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.451006 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" seLinuxMountContext="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.451032 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" seLinuxMountContext="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.451059 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" seLinuxMountContext="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.451089 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" seLinuxMountContext="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.451118 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf" seLinuxMountContext="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.451147 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" seLinuxMountContext="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.451173 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49ef4625-1d3a-4a9f-b595-c2433d32326d" volumeName="kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" seLinuxMountContext="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.451237 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" seLinuxMountContext="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.451269 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" seLinuxMountContext="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.451298 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" seLinuxMountContext="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.451328 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" seLinuxMountContext="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.451376 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert" seLinuxMountContext="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.451442 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" seLinuxMountContext="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.451480 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" seLinuxMountContext="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.451511 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" seLinuxMountContext="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.451541 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" seLinuxMountContext="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.451570 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" seLinuxMountContext="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.451602 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" seLinuxMountContext="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.451631 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" seLinuxMountContext="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.451657 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" seLinuxMountContext="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.451682 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" seLinuxMountContext="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.451710 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" seLinuxMountContext="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.451740 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" seLinuxMountContext="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.451773 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3b6479f0-333b-4a96-9adf-2099afdc2447" volumeName="kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr" seLinuxMountContext="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.451803 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" seLinuxMountContext="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.451834 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" seLinuxMountContext="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.451862 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" seLinuxMountContext="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.451888 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" seLinuxMountContext="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.451914 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" seLinuxMountContext="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.451942 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" seLinuxMountContext="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.451969 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" seLinuxMountContext="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.451993 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" seLinuxMountContext="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.452020 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" seLinuxMountContext="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.452047 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls" seLinuxMountContext="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.452074 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" seLinuxMountContext="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.452100 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" seLinuxMountContext="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.452160 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" seLinuxMountContext="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.452248 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" seLinuxMountContext="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.452282 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" seLinuxMountContext="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.452312 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" seLinuxMountContext="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.452339 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" seLinuxMountContext="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.452368 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" seLinuxMountContext="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.452396 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" seLinuxMountContext="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.452424 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" seLinuxMountContext="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.452452 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" seLinuxMountContext="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.452482 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" seLinuxMountContext="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.452511 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" seLinuxMountContext="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.452538 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" seLinuxMountContext="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.452565 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" seLinuxMountContext="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.452594 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" seLinuxMountContext="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.452622 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" seLinuxMountContext="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.452652 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" seLinuxMountContext="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.452679 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" seLinuxMountContext="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.455906 4846 reconstruct.go:144] "Volume is marked device as uncertain and added into the actual state" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" deviceMountPath="/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.455981 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" seLinuxMountContext="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.456018 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb" seLinuxMountContext="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.456056 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" seLinuxMountContext="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.456096 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" seLinuxMountContext="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.456130 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" seLinuxMountContext="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.456161 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" seLinuxMountContext="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.456247 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" seLinuxMountContext="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.456281 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" seLinuxMountContext="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.456310 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" seLinuxMountContext="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.456363 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" seLinuxMountContext="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.456426 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" seLinuxMountContext="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.456462 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" seLinuxMountContext="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.456540 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" seLinuxMountContext="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.456595 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" seLinuxMountContext="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.456626 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" seLinuxMountContext="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.456655 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" seLinuxMountContext="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.456688 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5" seLinuxMountContext="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.456720 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" seLinuxMountContext="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.456753 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" seLinuxMountContext="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.456783 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" seLinuxMountContext="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.456812 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" seLinuxMountContext="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.456844 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" seLinuxMountContext="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.456908 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" seLinuxMountContext="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.456943 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" seLinuxMountContext="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.456972 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" seLinuxMountContext="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.457000 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" seLinuxMountContext="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.457035 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" seLinuxMountContext="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.457076 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" seLinuxMountContext="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.457129 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" seLinuxMountContext="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.457162 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" seLinuxMountContext="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.457227 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" seLinuxMountContext="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.457258 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="44663579-783b-4372-86d6-acf235a62d72" volumeName="kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" seLinuxMountContext="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.457289 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" seLinuxMountContext="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.457323 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" seLinuxMountContext="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.457353 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" seLinuxMountContext="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.457394 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" seLinuxMountContext="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.457423 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" seLinuxMountContext="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.457466 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" seLinuxMountContext="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.457494 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" seLinuxMountContext="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.457528 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" seLinuxMountContext="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.457557 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" seLinuxMountContext="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.457614 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" seLinuxMountContext="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.457641 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" seLinuxMountContext="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.457694 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" seLinuxMountContext="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.457721 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" seLinuxMountContext="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.457796 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" seLinuxMountContext="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.457824 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" seLinuxMountContext="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.457863 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" seLinuxMountContext="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.457889 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" seLinuxMountContext="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.457950 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" seLinuxMountContext="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.457988 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" seLinuxMountContext="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.458017 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" seLinuxMountContext="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.458049 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm" seLinuxMountContext="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.458078 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" seLinuxMountContext="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.458117 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" seLinuxMountContext="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.458242 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" seLinuxMountContext="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.458284 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" seLinuxMountContext="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.458520 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" seLinuxMountContext="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.458591 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" seLinuxMountContext="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.458612 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" volumeName="kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" seLinuxMountContext="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.458630 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" seLinuxMountContext="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.458676 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" seLinuxMountContext="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.458697 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" seLinuxMountContext="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.458716 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" seLinuxMountContext="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.458756 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" seLinuxMountContext="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.458775 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" seLinuxMountContext="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.458798 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" seLinuxMountContext="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.458839 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" seLinuxMountContext="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.458861 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" seLinuxMountContext="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.458879 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" seLinuxMountContext="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.458932 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" seLinuxMountContext="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.458954 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" seLinuxMountContext="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.459197 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" seLinuxMountContext="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.459217 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" seLinuxMountContext="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.459277 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" seLinuxMountContext="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.459308 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" seLinuxMountContext="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.459369 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" seLinuxMountContext="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.459388 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" seLinuxMountContext="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.459407 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" seLinuxMountContext="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.459561 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" seLinuxMountContext="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.459623 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" seLinuxMountContext="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.459651 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" seLinuxMountContext="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.459700 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" seLinuxMountContext="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.459886 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" seLinuxMountContext="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.460197 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" seLinuxMountContext="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.460218 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" seLinuxMountContext="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.460234 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" seLinuxMountContext="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.460277 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" seLinuxMountContext="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.460299 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" seLinuxMountContext="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.460314 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" seLinuxMountContext="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.460377 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" seLinuxMountContext="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.460390 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" seLinuxMountContext="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.460478 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" seLinuxMountContext="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.460493 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" seLinuxMountContext="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.460568 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" seLinuxMountContext="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.460609 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" seLinuxMountContext="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.460624 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" seLinuxMountContext="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.460639 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" seLinuxMountContext="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.460658 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" seLinuxMountContext="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.460704 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" seLinuxMountContext="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.460749 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" seLinuxMountContext="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.460813 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" seLinuxMountContext="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.460836 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" seLinuxMountContext="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.460993 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" seLinuxMountContext="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.461102 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" seLinuxMountContext="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.461149 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" seLinuxMountContext="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.461229 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" seLinuxMountContext="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.461247 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" seLinuxMountContext="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.461326 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" seLinuxMountContext="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.461401 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" seLinuxMountContext="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.461416 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" seLinuxMountContext="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.461431 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" seLinuxMountContext="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.461448 4846 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" seLinuxMountContext="" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.461482 4846 reconstruct.go:97] "Volume reconstruction finished" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.461493 4846 reconciler.go:26] "Reconciler: start to sync state" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.466101 4846 manager.go:324] Recovery completed Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.484204 4846 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.489301 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.489372 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.489492 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.490960 4846 cpu_manager.go:225] "Starting CPU manager" policy="none" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.490997 4846 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.491045 4846 state_mem.go:36] "Initialized new in-memory state store" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.493733 4846 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.496172 4846 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.496229 4846 status_manager.go:217] "Starting to sync pod status with apiserver" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.496249 4846 kubelet.go:2335] "Starting kubelet main sync loop" Oct 05 06:47:58 crc kubenswrapper[4846]: E1005 06:47:58.496292 4846 kubelet.go:2359] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.499358 4846 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.111:6443: connect: connection refused Oct 05 06:47:58 crc kubenswrapper[4846]: E1005 06:47:58.499441 4846 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.111:6443: connect: connection refused" logger="UnhandledError" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.504756 4846 policy_none.go:49] "None policy: Start" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.505728 4846 memory_manager.go:170] "Starting memorymanager" policy="None" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.505817 4846 state_mem.go:35] "Initializing new in-memory state store" Oct 05 06:47:58 crc kubenswrapper[4846]: E1005 06:47:58.529261 4846 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.580945 4846 manager.go:334] "Starting Device Plugin manager" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.581573 4846 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.581599 4846 server.go:79] "Starting device plugin registration server" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.582039 4846 eviction_manager.go:189] "Eviction manager: starting control loop" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.582058 4846 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.582517 4846 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.582592 4846 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.582605 4846 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.596524 4846 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc","openshift-machine-config-operator/kube-rbac-proxy-crio-crc","openshift-etcd/etcd-crc","openshift-kube-apiserver/kube-apiserver-crc","openshift-kube-controller-manager/kube-controller-manager-crc"] Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.596650 4846 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 05 06:47:58 crc kubenswrapper[4846]: E1005 06:47:58.597887 4846 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.598437 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.598494 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.598509 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.598718 4846 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.599908 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.600015 4846 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.600038 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.600097 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.600119 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.600401 4846 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.600490 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.600540 4846 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.601145 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.601200 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.601216 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.601878 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.601906 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.601931 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.601941 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.601951 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.601963 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.602262 4846 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.602325 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.602375 4846 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.603625 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.603631 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.603707 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.603678 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.603722 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.603741 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.603987 4846 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.604170 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.604244 4846 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.605397 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.605434 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.605448 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.605539 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.605565 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.605579 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.605766 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.605805 4846 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.606639 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.606693 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.606720 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:47:58 crc kubenswrapper[4846]: E1005 06:47:58.629748 4846 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.111:6443: connect: connection refused" interval="400ms" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.665064 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.665112 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.665161 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.665239 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.665377 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.665566 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.665648 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.665701 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.665750 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.665792 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.665833 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.665877 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.665916 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.665984 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.666034 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.684508 4846 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.686380 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.686440 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.686461 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.686568 4846 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 05 06:47:58 crc kubenswrapper[4846]: E1005 06:47:58.687158 4846 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.111:6443: connect: connection refused" node="crc" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.767422 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.767497 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.767538 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.767569 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.767615 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.767647 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.767675 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.767703 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.767743 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.767775 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.767806 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.767834 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.767862 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.767891 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.767930 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.768535 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.768587 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.769155 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.768588 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.769160 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.768483 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.769508 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.769566 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.769431 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.769294 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.769334 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.769664 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.769688 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.769773 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.769882 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.888258 4846 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.889867 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.889919 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.889939 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.889975 4846 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 05 06:47:58 crc kubenswrapper[4846]: E1005 06:47:58.890569 4846 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.111:6443: connect: connection refused" node="crc" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.929580 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.935048 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.958537 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.980847 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.984293 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1b160f5dda77d281dd8e69ec8d817f9.slice/crio-3990211ac56404042a0c0c8dee673e4f3b998073a64b903b4134601e8bd04328 WatchSource:0}: Error finding container 3990211ac56404042a0c0c8dee673e4f3b998073a64b903b4134601e8bd04328: Status 404 returned error can't find the container with id 3990211ac56404042a0c0c8dee673e4f3b998073a64b903b4134601e8bd04328 Oct 05 06:47:58 crc kubenswrapper[4846]: I1005 06:47:58.985377 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.986554 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3dcd261975c3d6b9a6ad6367fd4facd3.slice/crio-7b31906f90f3a70afcd50a957f9852a2ae3e6606284ca5c77d388ca85a6cd0e8 WatchSource:0}: Error finding container 7b31906f90f3a70afcd50a957f9852a2ae3e6606284ca5c77d388ca85a6cd0e8: Status 404 returned error can't find the container with id 7b31906f90f3a70afcd50a957f9852a2ae3e6606284ca5c77d388ca85a6cd0e8 Oct 05 06:47:58 crc kubenswrapper[4846]: W1005 06:47:58.995580 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2139d3e2895fc6797b9c76a1b4c9886d.slice/crio-3fb8d9627e9d6aff9a18c625f0e9fef9bbb3cccc77c0d1a1431eca09ffcaab26 WatchSource:0}: Error finding container 3fb8d9627e9d6aff9a18c625f0e9fef9bbb3cccc77c0d1a1431eca09ffcaab26: Status 404 returned error can't find the container with id 3fb8d9627e9d6aff9a18c625f0e9fef9bbb3cccc77c0d1a1431eca09ffcaab26 Oct 05 06:47:59 crc kubenswrapper[4846]: W1005 06:47:59.009854 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf4b27818a5e8e43d0dc095d08835c792.slice/crio-81753b215d75180c0fac90a64c55854ea94e2913a5cc15fb957a4c719835314b WatchSource:0}: Error finding container 81753b215d75180c0fac90a64c55854ea94e2913a5cc15fb957a4c719835314b: Status 404 returned error can't find the container with id 81753b215d75180c0fac90a64c55854ea94e2913a5cc15fb957a4c719835314b Oct 05 06:47:59 crc kubenswrapper[4846]: W1005 06:47:59.015732 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf614b9022728cf315e60c057852e563e.slice/crio-075725f6de009bb30d17f2efa74d2379c870db0e895a4d8681f5c1600cb218fa WatchSource:0}: Error finding container 075725f6de009bb30d17f2efa74d2379c870db0e895a4d8681f5c1600cb218fa: Status 404 returned error can't find the container with id 075725f6de009bb30d17f2efa74d2379c870db0e895a4d8681f5c1600cb218fa Oct 05 06:47:59 crc kubenswrapper[4846]: E1005 06:47:59.031892 4846 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.111:6443: connect: connection refused" interval="800ms" Oct 05 06:47:59 crc kubenswrapper[4846]: W1005 06:47:59.275071 4846 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.111:6443: connect: connection refused Oct 05 06:47:59 crc kubenswrapper[4846]: E1005 06:47:59.275210 4846 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.111:6443: connect: connection refused" logger="UnhandledError" Oct 05 06:47:59 crc kubenswrapper[4846]: I1005 06:47:59.290849 4846 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 05 06:47:59 crc kubenswrapper[4846]: I1005 06:47:59.292432 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:47:59 crc kubenswrapper[4846]: I1005 06:47:59.292482 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:47:59 crc kubenswrapper[4846]: I1005 06:47:59.292497 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:47:59 crc kubenswrapper[4846]: I1005 06:47:59.292539 4846 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 05 06:47:59 crc kubenswrapper[4846]: E1005 06:47:59.293041 4846 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.111:6443: connect: connection refused" node="crc" Oct 05 06:47:59 crc kubenswrapper[4846]: I1005 06:47:59.419092 4846 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.111:6443: connect: connection refused Oct 05 06:47:59 crc kubenswrapper[4846]: W1005 06:47:59.469373 4846 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.111:6443: connect: connection refused Oct 05 06:47:59 crc kubenswrapper[4846]: E1005 06:47:59.469498 4846 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.111:6443: connect: connection refused" logger="UnhandledError" Oct 05 06:47:59 crc kubenswrapper[4846]: I1005 06:47:59.500789 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"3fb8d9627e9d6aff9a18c625f0e9fef9bbb3cccc77c0d1a1431eca09ffcaab26"} Oct 05 06:47:59 crc kubenswrapper[4846]: I1005 06:47:59.502103 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"3990211ac56404042a0c0c8dee673e4f3b998073a64b903b4134601e8bd04328"} Oct 05 06:47:59 crc kubenswrapper[4846]: I1005 06:47:59.503278 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"7b31906f90f3a70afcd50a957f9852a2ae3e6606284ca5c77d388ca85a6cd0e8"} Oct 05 06:47:59 crc kubenswrapper[4846]: I1005 06:47:59.504440 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"075725f6de009bb30d17f2efa74d2379c870db0e895a4d8681f5c1600cb218fa"} Oct 05 06:47:59 crc kubenswrapper[4846]: I1005 06:47:59.505557 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"81753b215d75180c0fac90a64c55854ea94e2913a5cc15fb957a4c719835314b"} Oct 05 06:47:59 crc kubenswrapper[4846]: W1005 06:47:59.560463 4846 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.111:6443: connect: connection refused Oct 05 06:47:59 crc kubenswrapper[4846]: E1005 06:47:59.560613 4846 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.111:6443: connect: connection refused" logger="UnhandledError" Oct 05 06:47:59 crc kubenswrapper[4846]: W1005 06:47:59.648595 4846 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.111:6443: connect: connection refused Oct 05 06:47:59 crc kubenswrapper[4846]: E1005 06:47:59.649240 4846 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.111:6443: connect: connection refused" logger="UnhandledError" Oct 05 06:47:59 crc kubenswrapper[4846]: E1005 06:47:59.832832 4846 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.111:6443: connect: connection refused" interval="1.6s" Oct 05 06:48:00 crc kubenswrapper[4846]: I1005 06:48:00.093677 4846 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 05 06:48:00 crc kubenswrapper[4846]: I1005 06:48:00.096284 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:00 crc kubenswrapper[4846]: I1005 06:48:00.096360 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:00 crc kubenswrapper[4846]: I1005 06:48:00.096381 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:00 crc kubenswrapper[4846]: I1005 06:48:00.096422 4846 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 05 06:48:00 crc kubenswrapper[4846]: E1005 06:48:00.096986 4846 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.111:6443: connect: connection refused" node="crc" Oct 05 06:48:00 crc kubenswrapper[4846]: I1005 06:48:00.418894 4846 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.111:6443: connect: connection refused Oct 05 06:48:00 crc kubenswrapper[4846]: I1005 06:48:00.511582 4846 generic.go:334] "Generic (PLEG): container finished" podID="d1b160f5dda77d281dd8e69ec8d817f9" containerID="ef573b754a80e5148b8caa71b4f22c461f57dfa415394a5bccda9771b7684e4c" exitCode=0 Oct 05 06:48:00 crc kubenswrapper[4846]: I1005 06:48:00.511728 4846 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 05 06:48:00 crc kubenswrapper[4846]: I1005 06:48:00.511736 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerDied","Data":"ef573b754a80e5148b8caa71b4f22c461f57dfa415394a5bccda9771b7684e4c"} Oct 05 06:48:00 crc kubenswrapper[4846]: I1005 06:48:00.513131 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:00 crc kubenswrapper[4846]: I1005 06:48:00.513210 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:00 crc kubenswrapper[4846]: I1005 06:48:00.513230 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:00 crc kubenswrapper[4846]: I1005 06:48:00.514098 4846 generic.go:334] "Generic (PLEG): container finished" podID="3dcd261975c3d6b9a6ad6367fd4facd3" containerID="c236b4ed7f039a221baac043246353eecbb6812c1e6bf6836a5fa96ec8f5abdb" exitCode=0 Oct 05 06:48:00 crc kubenswrapper[4846]: I1005 06:48:00.514167 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerDied","Data":"c236b4ed7f039a221baac043246353eecbb6812c1e6bf6836a5fa96ec8f5abdb"} Oct 05 06:48:00 crc kubenswrapper[4846]: I1005 06:48:00.514329 4846 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 05 06:48:00 crc kubenswrapper[4846]: I1005 06:48:00.515790 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:00 crc kubenswrapper[4846]: I1005 06:48:00.515835 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:00 crc kubenswrapper[4846]: I1005 06:48:00.515853 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:00 crc kubenswrapper[4846]: I1005 06:48:00.520411 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"640ade2c733d7916ec213a201a1dd614d261c69eab654c80100efbc99a768995"} Oct 05 06:48:00 crc kubenswrapper[4846]: I1005 06:48:00.520458 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"a0b6a7afe3094f1c92093c05c28344897a615d590c2fbb507c87e18a44c6c831"} Oct 05 06:48:00 crc kubenswrapper[4846]: I1005 06:48:00.520485 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"f2b5ee9781c01d6aec3ac778efef1067b6c3bfe4ed37431bdebe188fcde0165f"} Oct 05 06:48:00 crc kubenswrapper[4846]: I1005 06:48:00.520503 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"f3fb4717692aa1fec0012b514ef8d18d1ca3707cfa4bb8248bb93841a0a40f11"} Oct 05 06:48:00 crc kubenswrapper[4846]: I1005 06:48:00.520523 4846 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 05 06:48:00 crc kubenswrapper[4846]: I1005 06:48:00.522013 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:00 crc kubenswrapper[4846]: I1005 06:48:00.522086 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:00 crc kubenswrapper[4846]: I1005 06:48:00.522105 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:00 crc kubenswrapper[4846]: I1005 06:48:00.524027 4846 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="5dc5411b642f99b13d514d2f21defad0c394aeeb82a966f72300076617e2586c" exitCode=0 Oct 05 06:48:00 crc kubenswrapper[4846]: I1005 06:48:00.524131 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"5dc5411b642f99b13d514d2f21defad0c394aeeb82a966f72300076617e2586c"} Oct 05 06:48:00 crc kubenswrapper[4846]: I1005 06:48:00.524171 4846 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 05 06:48:00 crc kubenswrapper[4846]: I1005 06:48:00.525281 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:00 crc kubenswrapper[4846]: I1005 06:48:00.525332 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:00 crc kubenswrapper[4846]: I1005 06:48:00.525350 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:00 crc kubenswrapper[4846]: I1005 06:48:00.526746 4846 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="44a9f740f0a06c8bf928e2a9d1d5979217558c31e73a333631baccebf00fd5d2" exitCode=0 Oct 05 06:48:00 crc kubenswrapper[4846]: I1005 06:48:00.526801 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"44a9f740f0a06c8bf928e2a9d1d5979217558c31e73a333631baccebf00fd5d2"} Oct 05 06:48:00 crc kubenswrapper[4846]: I1005 06:48:00.526948 4846 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 05 06:48:00 crc kubenswrapper[4846]: I1005 06:48:00.529398 4846 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 05 06:48:00 crc kubenswrapper[4846]: I1005 06:48:00.529463 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:00 crc kubenswrapper[4846]: I1005 06:48:00.529505 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:00 crc kubenswrapper[4846]: I1005 06:48:00.529522 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:00 crc kubenswrapper[4846]: I1005 06:48:00.530677 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:00 crc kubenswrapper[4846]: I1005 06:48:00.530718 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:00 crc kubenswrapper[4846]: I1005 06:48:00.530739 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:01 crc kubenswrapper[4846]: W1005 06:48:01.212752 4846 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.111:6443: connect: connection refused Oct 05 06:48:01 crc kubenswrapper[4846]: E1005 06:48:01.212854 4846 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.111:6443: connect: connection refused" logger="UnhandledError" Oct 05 06:48:01 crc kubenswrapper[4846]: I1005 06:48:01.418695 4846 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.111:6443: connect: connection refused Oct 05 06:48:01 crc kubenswrapper[4846]: E1005 06:48:01.433819 4846 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.111:6443: connect: connection refused" interval="3.2s" Oct 05 06:48:01 crc kubenswrapper[4846]: I1005 06:48:01.530283 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"7e13964315903099b3d7e3dc8c143fe58bb9f9654214702ba51581ac4044414d"} Oct 05 06:48:01 crc kubenswrapper[4846]: I1005 06:48:01.530416 4846 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 05 06:48:01 crc kubenswrapper[4846]: I1005 06:48:01.531300 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:01 crc kubenswrapper[4846]: I1005 06:48:01.531328 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:01 crc kubenswrapper[4846]: I1005 06:48:01.531338 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:01 crc kubenswrapper[4846]: I1005 06:48:01.535915 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"70f0d5b436d18a78a2d8e2767143ba1a3104332d86bd63f91d23cd8e77a952aa"} Oct 05 06:48:01 crc kubenswrapper[4846]: I1005 06:48:01.535990 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"b5e5cf9125d1d413c91666744aeab06980b44420dbf0341b165b05787527944b"} Oct 05 06:48:01 crc kubenswrapper[4846]: I1005 06:48:01.536008 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"2b94cc50fe15d58f375f8492067811be16278cdf13ee3d80cfed4682279157b6"} Oct 05 06:48:01 crc kubenswrapper[4846]: I1005 06:48:01.536150 4846 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 05 06:48:01 crc kubenswrapper[4846]: I1005 06:48:01.537300 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:01 crc kubenswrapper[4846]: I1005 06:48:01.537339 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:01 crc kubenswrapper[4846]: I1005 06:48:01.537352 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:01 crc kubenswrapper[4846]: I1005 06:48:01.541104 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"f4a091716ff935adb42ae2e88048619fd347bd174f64e5f9c97a01f81690e57f"} Oct 05 06:48:01 crc kubenswrapper[4846]: I1005 06:48:01.541143 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"15dbaf2445704da0987f256f81dff2ee95664988541fa48ce94898afeadab2e8"} Oct 05 06:48:01 crc kubenswrapper[4846]: I1005 06:48:01.541307 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"cd3b3ab323c6d5153054aa04af2d91f7b9e8436ff1e2f75f3594b57831b2a416"} Oct 05 06:48:01 crc kubenswrapper[4846]: I1005 06:48:01.541336 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"0839c1765df2c68d2554e9c66d289505e559a9672caf8069feeff2c4fdfd2171"} Oct 05 06:48:01 crc kubenswrapper[4846]: I1005 06:48:01.543353 4846 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="48ca0ba6ca6092bfcb5b153220c054f02cc36fb22707c42e021f06d329e18e2e" exitCode=0 Oct 05 06:48:01 crc kubenswrapper[4846]: I1005 06:48:01.543480 4846 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 05 06:48:01 crc kubenswrapper[4846]: I1005 06:48:01.544040 4846 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 05 06:48:01 crc kubenswrapper[4846]: I1005 06:48:01.544537 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"48ca0ba6ca6092bfcb5b153220c054f02cc36fb22707c42e021f06d329e18e2e"} Oct 05 06:48:01 crc kubenswrapper[4846]: I1005 06:48:01.545258 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:01 crc kubenswrapper[4846]: I1005 06:48:01.545294 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:01 crc kubenswrapper[4846]: I1005 06:48:01.545311 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:01 crc kubenswrapper[4846]: I1005 06:48:01.546031 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:01 crc kubenswrapper[4846]: I1005 06:48:01.546059 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:01 crc kubenswrapper[4846]: I1005 06:48:01.546069 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:01 crc kubenswrapper[4846]: W1005 06:48:01.683004 4846 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.111:6443: connect: connection refused Oct 05 06:48:01 crc kubenswrapper[4846]: E1005 06:48:01.683160 4846 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.111:6443: connect: connection refused" logger="UnhandledError" Oct 05 06:48:01 crc kubenswrapper[4846]: I1005 06:48:01.697619 4846 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 05 06:48:01 crc kubenswrapper[4846]: I1005 06:48:01.699267 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:01 crc kubenswrapper[4846]: I1005 06:48:01.699317 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:01 crc kubenswrapper[4846]: I1005 06:48:01.699332 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:01 crc kubenswrapper[4846]: I1005 06:48:01.699363 4846 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 05 06:48:01 crc kubenswrapper[4846]: E1005 06:48:01.699922 4846 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.111:6443: connect: connection refused" node="crc" Oct 05 06:48:02 crc kubenswrapper[4846]: W1005 06:48:02.147621 4846 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.111:6443: connect: connection refused Oct 05 06:48:02 crc kubenswrapper[4846]: E1005 06:48:02.147741 4846 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.111:6443: connect: connection refused" logger="UnhandledError" Oct 05 06:48:02 crc kubenswrapper[4846]: I1005 06:48:02.551320 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"25335315b00b3ac706ebfb936c850af574eae5ec0f572c1ad4594a16c37b8ab9"} Oct 05 06:48:02 crc kubenswrapper[4846]: I1005 06:48:02.551467 4846 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 05 06:48:02 crc kubenswrapper[4846]: I1005 06:48:02.556985 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:02 crc kubenswrapper[4846]: I1005 06:48:02.557113 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:02 crc kubenswrapper[4846]: I1005 06:48:02.557172 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:02 crc kubenswrapper[4846]: I1005 06:48:02.560687 4846 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="4006d0cea128675c8885d879171204abffc5e11ea0664750a12b4bc7344ccb2e" exitCode=0 Oct 05 06:48:02 crc kubenswrapper[4846]: I1005 06:48:02.560895 4846 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 05 06:48:02 crc kubenswrapper[4846]: I1005 06:48:02.561363 4846 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 05 06:48:02 crc kubenswrapper[4846]: I1005 06:48:02.561449 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"4006d0cea128675c8885d879171204abffc5e11ea0664750a12b4bc7344ccb2e"} Oct 05 06:48:02 crc kubenswrapper[4846]: I1005 06:48:02.561584 4846 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 05 06:48:02 crc kubenswrapper[4846]: I1005 06:48:02.561649 4846 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 05 06:48:02 crc kubenswrapper[4846]: I1005 06:48:02.563148 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:02 crc kubenswrapper[4846]: I1005 06:48:02.563225 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:02 crc kubenswrapper[4846]: I1005 06:48:02.563245 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:02 crc kubenswrapper[4846]: I1005 06:48:02.563349 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:02 crc kubenswrapper[4846]: I1005 06:48:02.563385 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:02 crc kubenswrapper[4846]: I1005 06:48:02.563405 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:02 crc kubenswrapper[4846]: I1005 06:48:02.563798 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:02 crc kubenswrapper[4846]: I1005 06:48:02.563827 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:02 crc kubenswrapper[4846]: I1005 06:48:02.563838 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:03 crc kubenswrapper[4846]: I1005 06:48:03.569122 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"496a4782faa30a9a20ff63038141600cd635f12902d1cf17989cd067fcdfe929"} Oct 05 06:48:03 crc kubenswrapper[4846]: I1005 06:48:03.569224 4846 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 05 06:48:03 crc kubenswrapper[4846]: I1005 06:48:03.569238 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"bfe45d30e4fa141541aca1d46bc5cc1da0a7b4a5eb1fc9bb2caa5b7fdcb0b264"} Oct 05 06:48:03 crc kubenswrapper[4846]: I1005 06:48:03.569272 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"c17a711c2c7473b076a3944ffe7c700808d05cba582fbd55c5c2ac8ff73838a4"} Oct 05 06:48:03 crc kubenswrapper[4846]: I1005 06:48:03.569287 4846 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 05 06:48:03 crc kubenswrapper[4846]: I1005 06:48:03.571082 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:03 crc kubenswrapper[4846]: I1005 06:48:03.571125 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:03 crc kubenswrapper[4846]: I1005 06:48:03.571142 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:03 crc kubenswrapper[4846]: I1005 06:48:03.944980 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 05 06:48:03 crc kubenswrapper[4846]: I1005 06:48:03.945207 4846 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 05 06:48:03 crc kubenswrapper[4846]: I1005 06:48:03.947154 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:03 crc kubenswrapper[4846]: I1005 06:48:03.947317 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:03 crc kubenswrapper[4846]: I1005 06:48:03.947341 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:04 crc kubenswrapper[4846]: I1005 06:48:04.299514 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 05 06:48:04 crc kubenswrapper[4846]: I1005 06:48:04.580057 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"6771c7215d2f5dd5838d17888375a4768b9d1755fc2e21bda1bbfbb2eb1dbb95"} Oct 05 06:48:04 crc kubenswrapper[4846]: I1005 06:48:04.580145 4846 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 05 06:48:04 crc kubenswrapper[4846]: I1005 06:48:04.580257 4846 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 05 06:48:04 crc kubenswrapper[4846]: I1005 06:48:04.580150 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"281be077869a6aa1ca815770871a14413d4a97e97aeb550fd5cd215cba88b21f"} Oct 05 06:48:04 crc kubenswrapper[4846]: I1005 06:48:04.580319 4846 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 05 06:48:04 crc kubenswrapper[4846]: I1005 06:48:04.581991 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:04 crc kubenswrapper[4846]: I1005 06:48:04.582143 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:04 crc kubenswrapper[4846]: I1005 06:48:04.582169 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:04 crc kubenswrapper[4846]: I1005 06:48:04.582172 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:04 crc kubenswrapper[4846]: I1005 06:48:04.582248 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:04 crc kubenswrapper[4846]: I1005 06:48:04.582267 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:04 crc kubenswrapper[4846]: I1005 06:48:04.677814 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 05 06:48:04 crc kubenswrapper[4846]: I1005 06:48:04.900512 4846 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 05 06:48:04 crc kubenswrapper[4846]: I1005 06:48:04.902117 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:04 crc kubenswrapper[4846]: I1005 06:48:04.902156 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:04 crc kubenswrapper[4846]: I1005 06:48:04.902169 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:04 crc kubenswrapper[4846]: I1005 06:48:04.902220 4846 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 05 06:48:04 crc kubenswrapper[4846]: I1005 06:48:04.982603 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 05 06:48:04 crc kubenswrapper[4846]: I1005 06:48:04.982811 4846 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 05 06:48:04 crc kubenswrapper[4846]: I1005 06:48:04.984328 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:04 crc kubenswrapper[4846]: I1005 06:48:04.984361 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:04 crc kubenswrapper[4846]: I1005 06:48:04.984374 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:05 crc kubenswrapper[4846]: I1005 06:48:05.424499 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 05 06:48:05 crc kubenswrapper[4846]: I1005 06:48:05.424842 4846 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 05 06:48:05 crc kubenswrapper[4846]: I1005 06:48:05.426692 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:05 crc kubenswrapper[4846]: I1005 06:48:05.426762 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:05 crc kubenswrapper[4846]: I1005 06:48:05.426784 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:05 crc kubenswrapper[4846]: I1005 06:48:05.437687 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 05 06:48:05 crc kubenswrapper[4846]: I1005 06:48:05.586924 4846 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 05 06:48:05 crc kubenswrapper[4846]: I1005 06:48:05.586946 4846 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 05 06:48:05 crc kubenswrapper[4846]: I1005 06:48:05.587144 4846 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 05 06:48:05 crc kubenswrapper[4846]: I1005 06:48:05.586955 4846 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 05 06:48:05 crc kubenswrapper[4846]: I1005 06:48:05.588895 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:05 crc kubenswrapper[4846]: I1005 06:48:05.588956 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:05 crc kubenswrapper[4846]: I1005 06:48:05.588985 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:05 crc kubenswrapper[4846]: I1005 06:48:05.589085 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:05 crc kubenswrapper[4846]: I1005 06:48:05.589141 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:05 crc kubenswrapper[4846]: I1005 06:48:05.589162 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:05 crc kubenswrapper[4846]: I1005 06:48:05.589349 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:05 crc kubenswrapper[4846]: I1005 06:48:05.589371 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:05 crc kubenswrapper[4846]: I1005 06:48:05.589389 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:06 crc kubenswrapper[4846]: I1005 06:48:06.945632 4846 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Oct 05 06:48:06 crc kubenswrapper[4846]: I1005 06:48:06.945761 4846 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 05 06:48:07 crc kubenswrapper[4846]: I1005 06:48:07.805654 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-crc" Oct 05 06:48:07 crc kubenswrapper[4846]: I1005 06:48:07.805925 4846 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 05 06:48:07 crc kubenswrapper[4846]: I1005 06:48:07.807653 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:07 crc kubenswrapper[4846]: I1005 06:48:07.807736 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:07 crc kubenswrapper[4846]: I1005 06:48:07.807752 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:07 crc kubenswrapper[4846]: I1005 06:48:07.855700 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 05 06:48:07 crc kubenswrapper[4846]: I1005 06:48:07.855899 4846 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 05 06:48:07 crc kubenswrapper[4846]: I1005 06:48:07.857369 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:07 crc kubenswrapper[4846]: I1005 06:48:07.857444 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:07 crc kubenswrapper[4846]: I1005 06:48:07.857468 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:07 crc kubenswrapper[4846]: I1005 06:48:07.880582 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 05 06:48:07 crc kubenswrapper[4846]: I1005 06:48:07.880798 4846 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 05 06:48:07 crc kubenswrapper[4846]: I1005 06:48:07.880889 4846 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 05 06:48:07 crc kubenswrapper[4846]: I1005 06:48:07.882400 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:07 crc kubenswrapper[4846]: I1005 06:48:07.882479 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:07 crc kubenswrapper[4846]: I1005 06:48:07.882505 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:07 crc kubenswrapper[4846]: I1005 06:48:07.896280 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 05 06:48:08 crc kubenswrapper[4846]: I1005 06:48:08.467517 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-etcd/etcd-crc" Oct 05 06:48:08 crc kubenswrapper[4846]: I1005 06:48:08.596596 4846 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 05 06:48:08 crc kubenswrapper[4846]: I1005 06:48:08.596656 4846 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 05 06:48:08 crc kubenswrapper[4846]: E1005 06:48:08.598094 4846 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Oct 05 06:48:08 crc kubenswrapper[4846]: I1005 06:48:08.598272 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:08 crc kubenswrapper[4846]: I1005 06:48:08.598316 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:08 crc kubenswrapper[4846]: I1005 06:48:08.598341 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:08 crc kubenswrapper[4846]: I1005 06:48:08.598354 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:08 crc kubenswrapper[4846]: I1005 06:48:08.598372 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:08 crc kubenswrapper[4846]: I1005 06:48:08.598376 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:12 crc kubenswrapper[4846]: W1005 06:48:12.352463 4846 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": net/http: TLS handshake timeout Oct 05 06:48:12 crc kubenswrapper[4846]: I1005 06:48:12.352584 4846 trace.go:236] Trace[708361166]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (05-Oct-2025 06:48:02.351) (total time: 10001ms): Oct 05 06:48:12 crc kubenswrapper[4846]: Trace[708361166]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": net/http: TLS handshake timeout 10001ms (06:48:12.352) Oct 05 06:48:12 crc kubenswrapper[4846]: Trace[708361166]: [10.001532892s] [10.001532892s] END Oct 05 06:48:12 crc kubenswrapper[4846]: E1005 06:48:12.352617 4846 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" Oct 05 06:48:12 crc kubenswrapper[4846]: I1005 06:48:12.419400 4846 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": net/http: TLS handshake timeout Oct 05 06:48:12 crc kubenswrapper[4846]: I1005 06:48:12.657711 4846 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Oct 05 06:48:12 crc kubenswrapper[4846]: I1005 06:48:12.657798 4846 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Oct 05 06:48:12 crc kubenswrapper[4846]: I1005 06:48:12.662760 4846 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Oct 05 06:48:12 crc kubenswrapper[4846]: I1005 06:48:12.662825 4846 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Oct 05 06:48:13 crc kubenswrapper[4846]: I1005 06:48:13.618391 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Oct 05 06:48:13 crc kubenswrapper[4846]: I1005 06:48:13.620481 4846 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="25335315b00b3ac706ebfb936c850af574eae5ec0f572c1ad4594a16c37b8ab9" exitCode=255 Oct 05 06:48:13 crc kubenswrapper[4846]: I1005 06:48:13.620535 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"25335315b00b3ac706ebfb936c850af574eae5ec0f572c1ad4594a16c37b8ab9"} Oct 05 06:48:13 crc kubenswrapper[4846]: I1005 06:48:13.620743 4846 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 05 06:48:13 crc kubenswrapper[4846]: I1005 06:48:13.621960 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:13 crc kubenswrapper[4846]: I1005 06:48:13.621998 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:13 crc kubenswrapper[4846]: I1005 06:48:13.622011 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:13 crc kubenswrapper[4846]: I1005 06:48:13.622596 4846 scope.go:117] "RemoveContainer" containerID="25335315b00b3ac706ebfb936c850af574eae5ec0f572c1ad4594a16c37b8ab9" Oct 05 06:48:14 crc kubenswrapper[4846]: I1005 06:48:14.305385 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 05 06:48:14 crc kubenswrapper[4846]: I1005 06:48:14.625441 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Oct 05 06:48:14 crc kubenswrapper[4846]: I1005 06:48:14.627979 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"c52e076eddfd307110b1f6de00a169b81af8f2785af7becef2a820ce363a5861"} Oct 05 06:48:14 crc kubenswrapper[4846]: I1005 06:48:14.628219 4846 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 05 06:48:14 crc kubenswrapper[4846]: I1005 06:48:14.629532 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:14 crc kubenswrapper[4846]: I1005 06:48:14.629620 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:14 crc kubenswrapper[4846]: I1005 06:48:14.629637 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:14 crc kubenswrapper[4846]: I1005 06:48:14.632147 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 05 06:48:15 crc kubenswrapper[4846]: I1005 06:48:15.634342 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Oct 05 06:48:15 crc kubenswrapper[4846]: I1005 06:48:15.635117 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Oct 05 06:48:15 crc kubenswrapper[4846]: I1005 06:48:15.637778 4846 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="c52e076eddfd307110b1f6de00a169b81af8f2785af7becef2a820ce363a5861" exitCode=255 Oct 05 06:48:15 crc kubenswrapper[4846]: I1005 06:48:15.637862 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"c52e076eddfd307110b1f6de00a169b81af8f2785af7becef2a820ce363a5861"} Oct 05 06:48:15 crc kubenswrapper[4846]: I1005 06:48:15.637950 4846 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 05 06:48:15 crc kubenswrapper[4846]: I1005 06:48:15.638066 4846 scope.go:117] "RemoveContainer" containerID="25335315b00b3ac706ebfb936c850af574eae5ec0f572c1ad4594a16c37b8ab9" Oct 05 06:48:15 crc kubenswrapper[4846]: I1005 06:48:15.639215 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:15 crc kubenswrapper[4846]: I1005 06:48:15.639280 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:15 crc kubenswrapper[4846]: I1005 06:48:15.639305 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:15 crc kubenswrapper[4846]: I1005 06:48:15.640208 4846 scope.go:117] "RemoveContainer" containerID="c52e076eddfd307110b1f6de00a169b81af8f2785af7becef2a820ce363a5861" Oct 05 06:48:15 crc kubenswrapper[4846]: E1005 06:48:15.640546 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Oct 05 06:48:15 crc kubenswrapper[4846]: I1005 06:48:15.965419 4846 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 05 06:48:16 crc kubenswrapper[4846]: I1005 06:48:16.641910 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Oct 05 06:48:16 crc kubenswrapper[4846]: I1005 06:48:16.644206 4846 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 05 06:48:16 crc kubenswrapper[4846]: I1005 06:48:16.645026 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:16 crc kubenswrapper[4846]: I1005 06:48:16.645063 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:16 crc kubenswrapper[4846]: I1005 06:48:16.645075 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:16 crc kubenswrapper[4846]: I1005 06:48:16.645687 4846 scope.go:117] "RemoveContainer" containerID="c52e076eddfd307110b1f6de00a169b81af8f2785af7becef2a820ce363a5861" Oct 05 06:48:16 crc kubenswrapper[4846]: E1005 06:48:16.645880 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Oct 05 06:48:16 crc kubenswrapper[4846]: I1005 06:48:16.946318 4846 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Oct 05 06:48:16 crc kubenswrapper[4846]: I1005 06:48:16.946424 4846 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.142665 4846 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.414700 4846 apiserver.go:52] "Watching apiserver" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.422882 4846 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.423337 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-network-console/networking-console-plugin-85b44fc459-gdk6g","openshift-network-diagnostics/network-check-source-55646444c4-trplf","openshift-network-diagnostics/network-check-target-xd92c","openshift-network-node-identity/network-node-identity-vrzqb","openshift-network-operator/iptables-alerter-4ln5h","openshift-network-operator/network-operator-58b4c7f79c-55gtf"] Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.423848 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.424023 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.424106 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 05 06:48:17 crc kubenswrapper[4846]: E1005 06:48:17.424512 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 05 06:48:17 crc kubenswrapper[4846]: E1005 06:48:17.424575 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.425047 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.425135 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 05 06:48:17 crc kubenswrapper[4846]: E1005 06:48:17.425546 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.425156 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.428645 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.429093 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.429275 4846 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.429338 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.429494 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.430976 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.431019 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.431024 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.431220 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.431367 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.472480 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.485597 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.515572 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.546635 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.562828 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.576033 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.591883 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 05 06:48:17 crc kubenswrapper[4846]: E1005 06:48:17.610704 4846 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": context deadline exceeded" interval="6.4s" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.615298 4846 trace.go:236] Trace[1862805429]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (05-Oct-2025 06:48:07.179) (total time: 10435ms): Oct 05 06:48:17 crc kubenswrapper[4846]: Trace[1862805429]: ---"Objects listed" error: 10435ms (06:48:17.615) Oct 05 06:48:17 crc kubenswrapper[4846]: Trace[1862805429]: [10.435477033s] [10.435477033s] END Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.615341 4846 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.617365 4846 trace.go:236] Trace[2054884154]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (05-Oct-2025 06:48:06.064) (total time: 11552ms): Oct 05 06:48:17 crc kubenswrapper[4846]: Trace[2054884154]: ---"Objects listed" error: 11552ms (06:48:17.617) Oct 05 06:48:17 crc kubenswrapper[4846]: Trace[2054884154]: [11.55235617s] [11.55235617s] END Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.617399 4846 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.618653 4846 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Oct 05 06:48:17 crc kubenswrapper[4846]: E1005 06:48:17.619530 4846 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes \"crc\" is forbidden: autoscaling.openshift.io/ManagedNode infra config cache not synchronized" node="crc" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.619704 4846 trace.go:236] Trace[1703806106]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (05-Oct-2025 06:48:06.707) (total time: 10912ms): Oct 05 06:48:17 crc kubenswrapper[4846]: Trace[1703806106]: ---"Objects listed" error: 10911ms (06:48:17.619) Oct 05 06:48:17 crc kubenswrapper[4846]: Trace[1703806106]: [10.912227339s] [10.912227339s] END Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.619759 4846 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.668274 4846 scope.go:117] "RemoveContainer" containerID="c52e076eddfd307110b1f6de00a169b81af8f2785af7becef2a820ce363a5861" Oct 05 06:48:17 crc kubenswrapper[4846]: E1005 06:48:17.668481 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.669052 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.719397 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.719455 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.719513 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.719547 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.719577 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.719603 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.719627 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.719653 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.719677 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.719701 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.719730 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.719776 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.719802 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.719824 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.719885 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.719896 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.719909 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.719960 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.719983 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.720051 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.720071 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.720090 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.720110 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.720128 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.720148 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.720169 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.720205 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.720225 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.720242 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.720246 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.720260 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.720279 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.720345 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.720362 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.720386 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.720402 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.720419 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.720435 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.720456 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.720463 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.720519 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.720558 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.720633 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.720684 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.720695 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.720703 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.720744 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.720782 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.720817 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.720845 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.720856 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.720891 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.720926 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.720969 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.721009 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.721049 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.721086 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.721122 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.721155 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.721228 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.721263 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.721297 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.721334 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.721370 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.721408 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.721447 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.721482 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.721515 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.721550 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.721585 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.721621 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.721657 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.721693 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.721731 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.721768 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.721803 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.721841 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.721877 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.722007 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.722044 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.722081 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.722127 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.722168 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.722238 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.722277 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.722314 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.722359 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.722395 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.722431 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.722469 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.722507 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.722542 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.722578 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.722616 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.722652 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.722688 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.720902 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.720924 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.720986 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.721092 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.721253 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.721301 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.723066 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.721634 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.721753 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.721720 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.721793 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.721780 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.721810 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.721889 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.722024 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.722024 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.723208 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.722231 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.722364 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.722264 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.722486 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.722566 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.724526 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.724947 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.725680 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.725828 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.726066 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.726133 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.726152 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.726270 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.726517 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.726584 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.726840 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.726873 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.727023 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.727237 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.727474 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.722724 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.727971 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.727818 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.728084 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.727845 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.722698 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.722000 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.721584 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.728083 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.728300 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.728343 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.728379 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.728409 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.728438 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.728472 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.728502 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.728538 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.728583 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.728612 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.728644 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.728776 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.728830 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.728864 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.728896 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.728923 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.728945 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.728966 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.728989 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.729017 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.729044 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.729066 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.729097 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.729130 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.729164 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.729238 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.729271 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.729305 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.729334 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.729365 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.729398 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.729430 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.729483 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.729517 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.729551 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.729586 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.729617 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.729650 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.729686 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.729717 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.729750 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.729784 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.729819 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.729852 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.729885 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.729917 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.729949 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.729981 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.730015 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.730047 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.730083 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.730170 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.730228 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.730253 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.730282 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.730305 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.730327 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.730352 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.730375 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.730398 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.730424 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.730455 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.730488 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.730522 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.730553 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.730583 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.730611 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.730635 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.730656 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.730706 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.730726 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.730747 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.730773 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.730793 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.730814 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.730837 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.730861 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.730883 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.730905 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.730928 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.730951 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.730975 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.730996 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.731020 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.731044 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.731070 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.731092 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.731114 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.731137 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.731160 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.727977 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.731243 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.731274 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.731298 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.731322 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.731347 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.731375 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.731398 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.731421 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.731469 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.731493 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.731515 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.731538 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.731559 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.731616 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.731674 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.731709 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.731735 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.731763 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.731789 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.731813 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.731834 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.731860 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.731887 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.731911 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.731935 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.731964 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.731987 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.732064 4846 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.732080 4846 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.732094 4846 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.732109 4846 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.732125 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.732138 4846 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.732151 4846 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.732648 4846 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.732724 4846 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.732746 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.732812 4846 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.732830 4846 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.732849 4846 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.732873 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.732892 4846 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.732910 4846 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.732930 4846 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.732949 4846 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.732965 4846 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.732984 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.733005 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.733022 4846 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.733039 4846 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.733056 4846 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.733073 4846 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.733094 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.733111 4846 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.733129 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.733145 4846 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.733162 4846 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.733199 4846 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.733216 4846 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.733241 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.733258 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.733277 4846 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.733294 4846 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.733315 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.733333 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.733352 4846 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.733370 4846 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.733388 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.733405 4846 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.733421 4846 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.733437 4846 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.733453 4846 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.733470 4846 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.733486 4846 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.733503 4846 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.733519 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.733536 4846 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.733555 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.734670 4846 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.731043 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.728040 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.722644 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.728688 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.728754 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.740237 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.729085 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.729153 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.729643 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.729657 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.730036 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.730924 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.730898 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.731023 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.731213 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.731656 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.731840 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.731918 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.731940 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.732062 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.732110 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.733296 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.733361 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.733553 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.734058 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.734135 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.734174 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.734407 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.734484 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.734494 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.734634 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.735221 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.735224 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.744678 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.742342 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.746459 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.735936 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.736203 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.736618 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.736923 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.736944 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.737067 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.737133 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.747140 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.737103 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.737827 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.737861 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.737866 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.737977 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.738266 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.738037 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.738921 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.738958 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.739029 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.739055 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.739293 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.739503 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.739527 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.739871 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.739984 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.739961 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.740359 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.740505 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.740547 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.740906 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.741787 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.741879 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.742156 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.742206 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.742308 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.742530 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.742778 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.742996 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.743048 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.743775 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.743950 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.744068 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.744106 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.744279 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 06:48:17 crc kubenswrapper[4846]: E1005 06:48:17.744421 4846 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.747802 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 06:48:17 crc kubenswrapper[4846]: E1005 06:48:17.744442 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-05 06:48:18.244394206 +0000 UTC m=+20.485247181 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.747842 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 06:48:17 crc kubenswrapper[4846]: E1005 06:48:17.744516 4846 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.744868 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.744869 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.745226 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.745283 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.746348 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.747131 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 06:48:17 crc kubenswrapper[4846]: E1005 06:48:17.748329 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-05 06:48:18.248285552 +0000 UTC m=+20.489138547 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.748331 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.748816 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.748908 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 06:48:17 crc kubenswrapper[4846]: E1005 06:48:17.748969 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-05 06:48:18.248908751 +0000 UTC m=+20.489761566 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.737450 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.749645 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.749803 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.749946 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.751523 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.752280 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.752604 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.754433 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.755749 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.756899 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.761481 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.761507 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.761591 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.761654 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.761839 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.762680 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.763630 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.763696 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.763957 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.764003 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.764716 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.764855 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.766584 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.766846 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.766848 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.766888 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.767400 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.767799 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.767891 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 06:48:17 crc kubenswrapper[4846]: E1005 06:48:17.768485 4846 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 05 06:48:17 crc kubenswrapper[4846]: E1005 06:48:17.771461 4846 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 05 06:48:17 crc kubenswrapper[4846]: E1005 06:48:17.771525 4846 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 05 06:48:17 crc kubenswrapper[4846]: E1005 06:48:17.771657 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-05 06:48:18.271595996 +0000 UTC m=+20.512448771 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.768507 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.768978 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.769395 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.769809 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.769849 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 06:48:17 crc kubenswrapper[4846]: E1005 06:48:17.771084 4846 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 05 06:48:17 crc kubenswrapper[4846]: E1005 06:48:17.771766 4846 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 05 06:48:17 crc kubenswrapper[4846]: E1005 06:48:17.771777 4846 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 05 06:48:17 crc kubenswrapper[4846]: E1005 06:48:17.771810 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-05 06:48:18.271802752 +0000 UTC m=+20.512655527 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.772366 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.772163 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.773728 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.773990 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.774071 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.774258 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.774452 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.775314 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.775912 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.776033 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.776038 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.776241 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.776192 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.776572 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.778155 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.778275 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.778625 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.779231 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.781227 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.782052 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.782905 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.783407 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.787624 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.787726 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.787847 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.803295 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.823281 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.832364 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.834223 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.834281 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.834334 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.834375 4846 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.834395 4846 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.834409 4846 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.834425 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.834440 4846 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.834454 4846 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.834468 4846 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.834482 4846 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.834497 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.834512 4846 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.834525 4846 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.834538 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.834549 4846 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.834563 4846 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.834576 4846 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.834591 4846 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.834605 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.834619 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.834632 4846 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.834644 4846 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.834656 4846 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.834668 4846 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.834683 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.834694 4846 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.834709 4846 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.834721 4846 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.834734 4846 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.834746 4846 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.834760 4846 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.834773 4846 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.834786 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.834799 4846 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.834812 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.834825 4846 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.834838 4846 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.834851 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.834867 4846 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.834881 4846 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.834893 4846 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.834915 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.834927 4846 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.834938 4846 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.834975 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.834989 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.835000 4846 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.835012 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.835026 4846 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.835037 4846 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.835048 4846 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.835060 4846 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.835074 4846 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.835086 4846 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.835098 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.835109 4846 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.835121 4846 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.835132 4846 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.835144 4846 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.835155 4846 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.835168 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.835198 4846 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.835210 4846 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.835223 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.835238 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.835254 4846 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.835267 4846 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.835279 4846 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.835290 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.835301 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.835314 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.835326 4846 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.835338 4846 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.835349 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.835363 4846 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.835375 4846 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.835387 4846 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.835400 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.835412 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.835425 4846 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.835436 4846 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.835451 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.835466 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.835480 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.835493 4846 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.835506 4846 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.835517 4846 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.835542 4846 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.835556 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.835569 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.835583 4846 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.835594 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.835632 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.835647 4846 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.835660 4846 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.835672 4846 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.835684 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.835697 4846 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.835710 4846 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.835723 4846 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.835735 4846 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.835747 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.835760 4846 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.835773 4846 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.835786 4846 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.835798 4846 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.835811 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.835793 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.835823 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.836353 4846 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.836367 4846 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.836380 4846 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.836396 4846 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.836409 4846 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.836428 4846 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.836441 4846 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.836453 4846 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.836466 4846 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.836480 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.836494 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.836507 4846 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.836520 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.836533 4846 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.836546 4846 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.836558 4846 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.836570 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.836583 4846 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.836596 4846 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.836607 4846 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.836619 4846 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.836633 4846 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.836647 4846 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.836661 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.836674 4846 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.836686 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.836700 4846 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.836712 4846 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.836724 4846 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.836737 4846 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.836753 4846 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.836767 4846 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.836780 4846 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.836793 4846 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.836812 4846 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.836826 4846 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.836837 4846 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.845800 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.848292 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.855860 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.859264 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.866012 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc"] Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.872534 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.886336 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.901699 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.903239 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"775b5702-a131-4d6c-852d-91bfa056cac5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0839c1765df2c68d2554e9c66d289505e559a9672caf8069feeff2c4fdfd2171\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15dbaf2445704da0987f256f81dff2ee95664988541fa48ce94898afeadab2e8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd3b3ab323c6d5153054aa04af2d91f7b9e8436ff1e2f75f3594b57831b2a416\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c52e076eddfd307110b1f6de00a169b81af8f2785af7becef2a820ce363a5861\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c52e076eddfd307110b1f6de00a169b81af8f2785af7becef2a820ce363a5861\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-05T06:48:14Z\\\",\\\"message\\\":\\\"le observer\\\\nW1005 06:48:14.337870 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1005 06:48:14.338086 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1005 06:48:14.339389 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1467752215/tls.crt::/tmp/serving-cert-1467752215/tls.key\\\\\\\"\\\\nI1005 06:48:14.864345 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1005 06:48:14.868087 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1005 06:48:14.868114 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1005 06:48:14.868140 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1005 06:48:14.868147 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1005 06:48:14.877354 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1005 06:48:14.877412 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1005 06:48:14.877456 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1005 06:48:14.877471 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1005 06:48:14.877490 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1005 06:48:14.877501 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1005 06:48:14.877507 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1005 06:48:14.877515 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1005 06:48:14.880326 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:13Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4a091716ff935adb42ae2e88048619fd347bd174f64e5f9c97a01f81690e57f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:01Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dc5411b642f99b13d514d2f21defad0c394aeeb82a966f72300076617e2586c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5dc5411b642f99b13d514d2f21defad0c394aeeb82a966f72300076617e2586c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:47:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:47:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:47:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.916085 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.929952 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.941484 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.945121 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/kube-controller-manager-crc"] Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.955170 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.966694 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.982774 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c482bb97-fd50-40cf-96ea-f806c3f01125\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bfe45d30e4fa141541aca1d46bc5cc1da0a7b4a5eb1fc9bb2caa5b7fdcb0b264\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://496a4782faa30a9a20ff63038141600cd635f12902d1cf17989cd067fcdfe929\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://281be077869a6aa1ca815770871a14413d4a97e97aeb550fd5cd215cba88b21f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6771c7215d2f5dd5838d17888375a4768b9d1755fc2e21bda1bbfbb2eb1dbb95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c17a711c2c7473b076a3944ffe7c700808d05cba582fbd55c5c2ac8ff73838a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44a9f740f0a06c8bf928e2a9d1d5979217558c31e73a333631baccebf00fd5d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://44a9f740f0a06c8bf928e2a9d1d5979217558c31e73a333631baccebf00fd5d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:47:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:47:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://48ca0ba6ca6092bfcb5b153220c054f02cc36fb22707c42e021f06d329e18e2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48ca0ba6ca6092bfcb5b153220c054f02cc36fb22707c42e021f06d329e18e2e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:00Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4006d0cea128675c8885d879171204abffc5e11ea0664750a12b4bc7344ccb2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4006d0cea128675c8885d879171204abffc5e11ea0664750a12b4bc7344ccb2e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:47:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 05 06:48:17 crc kubenswrapper[4846]: I1005 06:48:17.993551 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 05 06:48:18 crc kubenswrapper[4846]: I1005 06:48:18.004221 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 05 06:48:18 crc kubenswrapper[4846]: I1005 06:48:18.017690 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 05 06:48:18 crc kubenswrapper[4846]: I1005 06:48:18.035866 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"775b5702-a131-4d6c-852d-91bfa056cac5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0839c1765df2c68d2554e9c66d289505e559a9672caf8069feeff2c4fdfd2171\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15dbaf2445704da0987f256f81dff2ee95664988541fa48ce94898afeadab2e8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd3b3ab323c6d5153054aa04af2d91f7b9e8436ff1e2f75f3594b57831b2a416\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c52e076eddfd307110b1f6de00a169b81af8f2785af7becef2a820ce363a5861\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c52e076eddfd307110b1f6de00a169b81af8f2785af7becef2a820ce363a5861\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-05T06:48:14Z\\\",\\\"message\\\":\\\"le observer\\\\nW1005 06:48:14.337870 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1005 06:48:14.338086 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1005 06:48:14.339389 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1467752215/tls.crt::/tmp/serving-cert-1467752215/tls.key\\\\\\\"\\\\nI1005 06:48:14.864345 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1005 06:48:14.868087 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1005 06:48:14.868114 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1005 06:48:14.868140 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1005 06:48:14.868147 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1005 06:48:14.877354 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1005 06:48:14.877412 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1005 06:48:14.877456 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1005 06:48:14.877471 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1005 06:48:14.877490 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1005 06:48:14.877501 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1005 06:48:14.877507 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1005 06:48:14.877515 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1005 06:48:14.880326 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:13Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4a091716ff935adb42ae2e88048619fd347bd174f64e5f9c97a01f81690e57f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:01Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dc5411b642f99b13d514d2f21defad0c394aeeb82a966f72300076617e2586c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5dc5411b642f99b13d514d2f21defad0c394aeeb82a966f72300076617e2586c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:47:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:47:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:47:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 05 06:48:18 crc kubenswrapper[4846]: I1005 06:48:18.048116 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 05 06:48:18 crc kubenswrapper[4846]: I1005 06:48:18.065592 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 05 06:48:18 crc kubenswrapper[4846]: W1005 06:48:18.069764 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod37a5e44f_9a88_4405_be8a_b645485e7312.slice/crio-00c6777a5792703897f3ce24ef5886c91163bcea551ffb3bf75f1f5993c06e0a WatchSource:0}: Error finding container 00c6777a5792703897f3ce24ef5886c91163bcea551ffb3bf75f1f5993c06e0a: Status 404 returned error can't find the container with id 00c6777a5792703897f3ce24ef5886c91163bcea551ffb3bf75f1f5993c06e0a Oct 05 06:48:18 crc kubenswrapper[4846]: I1005 06:48:18.080399 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 05 06:48:18 crc kubenswrapper[4846]: W1005 06:48:18.101006 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd75a4c96_2883_4a0b_bab2_0fab2b6c0b49.slice/crio-2162023aaecb7adb26b7b5a58c1f9560331d07651d078882f6dc0b422c957a06 WatchSource:0}: Error finding container 2162023aaecb7adb26b7b5a58c1f9560331d07651d078882f6dc0b422c957a06: Status 404 returned error can't find the container with id 2162023aaecb7adb26b7b5a58c1f9560331d07651d078882f6dc0b422c957a06 Oct 05 06:48:18 crc kubenswrapper[4846]: I1005 06:48:18.340015 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 05 06:48:18 crc kubenswrapper[4846]: E1005 06:48:18.340210 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-05 06:48:19.340163656 +0000 UTC m=+21.581016431 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 06:48:18 crc kubenswrapper[4846]: I1005 06:48:18.340498 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 05 06:48:18 crc kubenswrapper[4846]: I1005 06:48:18.340523 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 05 06:48:18 crc kubenswrapper[4846]: I1005 06:48:18.340544 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 05 06:48:18 crc kubenswrapper[4846]: I1005 06:48:18.340585 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 05 06:48:18 crc kubenswrapper[4846]: E1005 06:48:18.340661 4846 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 05 06:48:18 crc kubenswrapper[4846]: E1005 06:48:18.340683 4846 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 05 06:48:18 crc kubenswrapper[4846]: E1005 06:48:18.340698 4846 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 05 06:48:18 crc kubenswrapper[4846]: E1005 06:48:18.340687 4846 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 05 06:48:18 crc kubenswrapper[4846]: E1005 06:48:18.340711 4846 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 05 06:48:18 crc kubenswrapper[4846]: E1005 06:48:18.340744 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-05 06:48:19.340735093 +0000 UTC m=+21.581587858 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 05 06:48:18 crc kubenswrapper[4846]: E1005 06:48:18.340766 4846 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 05 06:48:18 crc kubenswrapper[4846]: E1005 06:48:18.340784 4846 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 05 06:48:18 crc kubenswrapper[4846]: E1005 06:48:18.340812 4846 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 05 06:48:18 crc kubenswrapper[4846]: E1005 06:48:18.340793 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-05 06:48:19.340772694 +0000 UTC m=+21.581625469 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 05 06:48:18 crc kubenswrapper[4846]: E1005 06:48:18.340873 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-05 06:48:19.340858557 +0000 UTC m=+21.581711332 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 05 06:48:18 crc kubenswrapper[4846]: E1005 06:48:18.340886 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-05 06:48:19.340880428 +0000 UTC m=+21.581733203 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 05 06:48:18 crc kubenswrapper[4846]: I1005 06:48:18.496991 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 05 06:48:18 crc kubenswrapper[4846]: E1005 06:48:18.497297 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 05 06:48:18 crc kubenswrapper[4846]: I1005 06:48:18.503080 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Oct 05 06:48:18 crc kubenswrapper[4846]: I1005 06:48:18.503684 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Oct 05 06:48:18 crc kubenswrapper[4846]: I1005 06:48:18.516793 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c482bb97-fd50-40cf-96ea-f806c3f01125\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bfe45d30e4fa141541aca1d46bc5cc1da0a7b4a5eb1fc9bb2caa5b7fdcb0b264\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://496a4782faa30a9a20ff63038141600cd635f12902d1cf17989cd067fcdfe929\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://281be077869a6aa1ca815770871a14413d4a97e97aeb550fd5cd215cba88b21f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6771c7215d2f5dd5838d17888375a4768b9d1755fc2e21bda1bbfbb2eb1dbb95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c17a711c2c7473b076a3944ffe7c700808d05cba582fbd55c5c2ac8ff73838a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44a9f740f0a06c8bf928e2a9d1d5979217558c31e73a333631baccebf00fd5d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://44a9f740f0a06c8bf928e2a9d1d5979217558c31e73a333631baccebf00fd5d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:47:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:47:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://48ca0ba6ca6092bfcb5b153220c054f02cc36fb22707c42e021f06d329e18e2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48ca0ba6ca6092bfcb5b153220c054f02cc36fb22707c42e021f06d329e18e2e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:00Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4006d0cea128675c8885d879171204abffc5e11ea0664750a12b4bc7344ccb2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4006d0cea128675c8885d879171204abffc5e11ea0664750a12b4bc7344ccb2e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:47:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 05 06:48:18 crc kubenswrapper[4846]: I1005 06:48:18.529606 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Oct 05 06:48:18 crc kubenswrapper[4846]: I1005 06:48:18.530357 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Oct 05 06:48:18 crc kubenswrapper[4846]: I1005 06:48:18.530265 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 05 06:48:18 crc kubenswrapper[4846]: I1005 06:48:18.530915 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Oct 05 06:48:18 crc kubenswrapper[4846]: I1005 06:48:18.541617 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Oct 05 06:48:18 crc kubenswrapper[4846]: I1005 06:48:18.542258 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Oct 05 06:48:18 crc kubenswrapper[4846]: I1005 06:48:18.542794 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Oct 05 06:48:18 crc kubenswrapper[4846]: I1005 06:48:18.543338 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 05 06:48:18 crc kubenswrapper[4846]: I1005 06:48:18.543452 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Oct 05 06:48:18 crc kubenswrapper[4846]: I1005 06:48:18.544048 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Oct 05 06:48:18 crc kubenswrapper[4846]: I1005 06:48:18.544588 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Oct 05 06:48:18 crc kubenswrapper[4846]: I1005 06:48:18.546167 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Oct 05 06:48:18 crc kubenswrapper[4846]: I1005 06:48:18.547541 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Oct 05 06:48:18 crc kubenswrapper[4846]: I1005 06:48:18.548760 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Oct 05 06:48:18 crc kubenswrapper[4846]: I1005 06:48:18.560722 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 05 06:48:18 crc kubenswrapper[4846]: I1005 06:48:18.574070 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 05 06:48:18 crc kubenswrapper[4846]: I1005 06:48:18.595262 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Oct 05 06:48:18 crc kubenswrapper[4846]: I1005 06:48:18.595224 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"775b5702-a131-4d6c-852d-91bfa056cac5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0839c1765df2c68d2554e9c66d289505e559a9672caf8069feeff2c4fdfd2171\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15dbaf2445704da0987f256f81dff2ee95664988541fa48ce94898afeadab2e8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd3b3ab323c6d5153054aa04af2d91f7b9e8436ff1e2f75f3594b57831b2a416\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c52e076eddfd307110b1f6de00a169b81af8f2785af7becef2a820ce363a5861\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c52e076eddfd307110b1f6de00a169b81af8f2785af7becef2a820ce363a5861\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-05T06:48:14Z\\\",\\\"message\\\":\\\"le observer\\\\nW1005 06:48:14.337870 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1005 06:48:14.338086 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1005 06:48:14.339389 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1467752215/tls.crt::/tmp/serving-cert-1467752215/tls.key\\\\\\\"\\\\nI1005 06:48:14.864345 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1005 06:48:14.868087 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1005 06:48:14.868114 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1005 06:48:14.868140 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1005 06:48:14.868147 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1005 06:48:14.877354 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1005 06:48:14.877412 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1005 06:48:14.877456 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1005 06:48:14.877471 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1005 06:48:14.877490 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1005 06:48:14.877501 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1005 06:48:14.877507 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1005 06:48:14.877515 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1005 06:48:14.880326 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:13Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4a091716ff935adb42ae2e88048619fd347bd174f64e5f9c97a01f81690e57f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:01Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dc5411b642f99b13d514d2f21defad0c394aeeb82a966f72300076617e2586c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5dc5411b642f99b13d514d2f21defad0c394aeeb82a966f72300076617e2586c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:47:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:47:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:47:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 05 06:48:18 crc kubenswrapper[4846]: I1005 06:48:18.596762 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Oct 05 06:48:18 crc kubenswrapper[4846]: I1005 06:48:18.599342 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Oct 05 06:48:18 crc kubenswrapper[4846]: I1005 06:48:18.600319 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Oct 05 06:48:18 crc kubenswrapper[4846]: I1005 06:48:18.601907 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Oct 05 06:48:18 crc kubenswrapper[4846]: I1005 06:48:18.604466 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Oct 05 06:48:18 crc kubenswrapper[4846]: I1005 06:48:18.618573 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c9f812b-ecc8-4985-a670-53fa80fbb499\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"message\\\":\\\"containers with unready status: [cluster-policy-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"message\\\":\\\"containers with unready status: [cluster-policy-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2b5ee9781c01d6aec3ac778efef1067b6c3bfe4ed37431bdebe188fcde0165f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:47:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3fb4717692aa1fec0012b514ef8d18d1ca3707cfa4bb8248bb93841a0a40f11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:47:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0b6a7afe3094f1c92093c05c28344897a615d590c2fbb507c87e18a44c6c831\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://640ade2c733d7916ec213a201a1dd614d261c69eab654c80100efbc99a768995\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:47:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 05 06:48:18 crc kubenswrapper[4846]: I1005 06:48:18.636642 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 05 06:48:18 crc kubenswrapper[4846]: I1005 06:48:18.642966 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Oct 05 06:48:18 crc kubenswrapper[4846]: I1005 06:48:18.644908 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Oct 05 06:48:18 crc kubenswrapper[4846]: I1005 06:48:18.646224 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Oct 05 06:48:18 crc kubenswrapper[4846]: I1005 06:48:18.654615 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 05 06:48:18 crc kubenswrapper[4846]: I1005 06:48:18.656816 4846 scope.go:117] "RemoveContainer" containerID="c52e076eddfd307110b1f6de00a169b81af8f2785af7becef2a820ce363a5861" Oct 05 06:48:18 crc kubenswrapper[4846]: E1005 06:48:18.657017 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Oct 05 06:48:18 crc kubenswrapper[4846]: E1005 06:48:18.668576 4846 kubelet.go:1929] "Failed creating a mirror pod for" err="pods \"etcd-crc\" already exists" pod="openshift-etcd/etcd-crc" Oct 05 06:48:18 crc kubenswrapper[4846]: I1005 06:48:18.699405 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Oct 05 06:48:18 crc kubenswrapper[4846]: I1005 06:48:18.700170 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Oct 05 06:48:18 crc kubenswrapper[4846]: I1005 06:48:18.702288 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Oct 05 06:48:18 crc kubenswrapper[4846]: I1005 06:48:18.723744 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Oct 05 06:48:18 crc kubenswrapper[4846]: I1005 06:48:18.724853 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Oct 05 06:48:18 crc kubenswrapper[4846]: I1005 06:48:18.726622 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Oct 05 06:48:18 crc kubenswrapper[4846]: I1005 06:48:18.727395 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Oct 05 06:48:18 crc kubenswrapper[4846]: I1005 06:48:18.728159 4846 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Oct 05 06:48:18 crc kubenswrapper[4846]: I1005 06:48:18.728407 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Oct 05 06:48:18 crc kubenswrapper[4846]: I1005 06:48:18.730275 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Oct 05 06:48:18 crc kubenswrapper[4846]: I1005 06:48:18.731867 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Oct 05 06:48:18 crc kubenswrapper[4846]: I1005 06:48:18.732330 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Oct 05 06:48:18 crc kubenswrapper[4846]: I1005 06:48:18.733550 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Oct 05 06:48:18 crc kubenswrapper[4846]: I1005 06:48:18.734234 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Oct 05 06:48:18 crc kubenswrapper[4846]: I1005 06:48:18.734761 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Oct 05 06:48:18 crc kubenswrapper[4846]: I1005 06:48:18.735396 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Oct 05 06:48:18 crc kubenswrapper[4846]: I1005 06:48:18.736019 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Oct 05 06:48:18 crc kubenswrapper[4846]: I1005 06:48:18.736515 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Oct 05 06:48:18 crc kubenswrapper[4846]: I1005 06:48:18.737087 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Oct 05 06:48:18 crc kubenswrapper[4846]: I1005 06:48:18.737744 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Oct 05 06:48:18 crc kubenswrapper[4846]: I1005 06:48:18.738329 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Oct 05 06:48:18 crc kubenswrapper[4846]: I1005 06:48:18.738792 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Oct 05 06:48:18 crc kubenswrapper[4846]: I1005 06:48:18.739338 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Oct 05 06:48:18 crc kubenswrapper[4846]: I1005 06:48:18.739914 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Oct 05 06:48:18 crc kubenswrapper[4846]: I1005 06:48:18.743470 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Oct 05 06:48:18 crc kubenswrapper[4846]: I1005 06:48:18.743934 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Oct 05 06:48:18 crc kubenswrapper[4846]: I1005 06:48:18.744764 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Oct 05 06:48:18 crc kubenswrapper[4846]: I1005 06:48:18.745240 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Oct 05 06:48:18 crc kubenswrapper[4846]: I1005 06:48:18.745726 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Oct 05 06:48:18 crc kubenswrapper[4846]: I1005 06:48:18.746657 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Oct 05 06:48:18 crc kubenswrapper[4846]: I1005 06:48:18.747111 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Oct 05 06:48:18 crc kubenswrapper[4846]: I1005 06:48:18.747922 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"2162023aaecb7adb26b7b5a58c1f9560331d07651d078882f6dc0b422c957a06"} Oct 05 06:48:18 crc kubenswrapper[4846]: I1005 06:48:18.747952 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"56f1fe1f9f06f0249cf937a6963fda82467732dc056ca2b2189be6d8dd9d4a05"} Oct 05 06:48:18 crc kubenswrapper[4846]: I1005 06:48:18.747976 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"7212ce5f42c1fb55146abb190a557a504b43f16af6df84f90e1073567a329ec5"} Oct 05 06:48:18 crc kubenswrapper[4846]: I1005 06:48:18.747986 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"49f832893dee4f5d40984c5e1e77b2ef8c99c17045ba883c7ef6c936172cd8b5"} Oct 05 06:48:18 crc kubenswrapper[4846]: I1005 06:48:18.747995 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"7ff10c31459b3b3724ee7b769bd6f72f9ed80cce5419b6adcdf69a3239b70756"} Oct 05 06:48:18 crc kubenswrapper[4846]: I1005 06:48:18.748006 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"00c6777a5792703897f3ce24ef5886c91163bcea551ffb3bf75f1f5993c06e0a"} Oct 05 06:48:18 crc kubenswrapper[4846]: I1005 06:48:18.779646 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c482bb97-fd50-40cf-96ea-f806c3f01125\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bfe45d30e4fa141541aca1d46bc5cc1da0a7b4a5eb1fc9bb2caa5b7fdcb0b264\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://496a4782faa30a9a20ff63038141600cd635f12902d1cf17989cd067fcdfe929\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://281be077869a6aa1ca815770871a14413d4a97e97aeb550fd5cd215cba88b21f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6771c7215d2f5dd5838d17888375a4768b9d1755fc2e21bda1bbfbb2eb1dbb95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c17a711c2c7473b076a3944ffe7c700808d05cba582fbd55c5c2ac8ff73838a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44a9f740f0a06c8bf928e2a9d1d5979217558c31e73a333631baccebf00fd5d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://44a9f740f0a06c8bf928e2a9d1d5979217558c31e73a333631baccebf00fd5d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:47:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:47:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://48ca0ba6ca6092bfcb5b153220c054f02cc36fb22707c42e021f06d329e18e2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48ca0ba6ca6092bfcb5b153220c054f02cc36fb22707c42e021f06d329e18e2e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:00Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4006d0cea128675c8885d879171204abffc5e11ea0664750a12b4bc7344ccb2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4006d0cea128675c8885d879171204abffc5e11ea0664750a12b4bc7344ccb2e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:47:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 05 06:48:18 crc kubenswrapper[4846]: I1005 06:48:18.796991 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ff10c31459b3b3724ee7b769bd6f72f9ed80cce5419b6adcdf69a3239b70756\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 05 06:48:18 crc kubenswrapper[4846]: I1005 06:48:18.816441 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 05 06:48:18 crc kubenswrapper[4846]: I1005 06:48:18.834116 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 05 06:48:18 crc kubenswrapper[4846]: I1005 06:48:18.850503 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56f1fe1f9f06f0249cf937a6963fda82467732dc056ca2b2189be6d8dd9d4a05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7212ce5f42c1fb55146abb190a557a504b43f16af6df84f90e1073567a329ec5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 05 06:48:18 crc kubenswrapper[4846]: I1005 06:48:18.871691 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"775b5702-a131-4d6c-852d-91bfa056cac5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0839c1765df2c68d2554e9c66d289505e559a9672caf8069feeff2c4fdfd2171\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15dbaf2445704da0987f256f81dff2ee95664988541fa48ce94898afeadab2e8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd3b3ab323c6d5153054aa04af2d91f7b9e8436ff1e2f75f3594b57831b2a416\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c52e076eddfd307110b1f6de00a169b81af8f2785af7becef2a820ce363a5861\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c52e076eddfd307110b1f6de00a169b81af8f2785af7becef2a820ce363a5861\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-05T06:48:14Z\\\",\\\"message\\\":\\\"le observer\\\\nW1005 06:48:14.337870 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1005 06:48:14.338086 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1005 06:48:14.339389 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1467752215/tls.crt::/tmp/serving-cert-1467752215/tls.key\\\\\\\"\\\\nI1005 06:48:14.864345 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1005 06:48:14.868087 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1005 06:48:14.868114 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1005 06:48:14.868140 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1005 06:48:14.868147 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1005 06:48:14.877354 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1005 06:48:14.877412 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1005 06:48:14.877456 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1005 06:48:14.877471 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1005 06:48:14.877490 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1005 06:48:14.877501 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1005 06:48:14.877507 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1005 06:48:14.877515 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1005 06:48:14.880326 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:13Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4a091716ff935adb42ae2e88048619fd347bd174f64e5f9c97a01f81690e57f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:01Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dc5411b642f99b13d514d2f21defad0c394aeeb82a966f72300076617e2586c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5dc5411b642f99b13d514d2f21defad0c394aeeb82a966f72300076617e2586c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:47:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:47:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:47:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 05 06:48:18 crc kubenswrapper[4846]: I1005 06:48:18.888528 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c9f812b-ecc8-4985-a670-53fa80fbb499\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"message\\\":\\\"containers with unready status: [cluster-policy-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"message\\\":\\\"containers with unready status: [cluster-policy-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2b5ee9781c01d6aec3ac778efef1067b6c3bfe4ed37431bdebe188fcde0165f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:47:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3fb4717692aa1fec0012b514ef8d18d1ca3707cfa4bb8248bb93841a0a40f11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:47:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0b6a7afe3094f1c92093c05c28344897a615d590c2fbb507c87e18a44c6c831\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://640ade2c733d7916ec213a201a1dd614d261c69eab654c80100efbc99a768995\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:47:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 05 06:48:18 crc kubenswrapper[4846]: I1005 06:48:18.907128 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 05 06:48:18 crc kubenswrapper[4846]: I1005 06:48:18.924233 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 05 06:48:18 crc kubenswrapper[4846]: I1005 06:48:18.941282 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"775b5702-a131-4d6c-852d-91bfa056cac5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0839c1765df2c68d2554e9c66d289505e559a9672caf8069feeff2c4fdfd2171\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15dbaf2445704da0987f256f81dff2ee95664988541fa48ce94898afeadab2e8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd3b3ab323c6d5153054aa04af2d91f7b9e8436ff1e2f75f3594b57831b2a416\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c52e076eddfd307110b1f6de00a169b81af8f2785af7becef2a820ce363a5861\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c52e076eddfd307110b1f6de00a169b81af8f2785af7becef2a820ce363a5861\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-05T06:48:14Z\\\",\\\"message\\\":\\\"le observer\\\\nW1005 06:48:14.337870 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1005 06:48:14.338086 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1005 06:48:14.339389 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1467752215/tls.crt::/tmp/serving-cert-1467752215/tls.key\\\\\\\"\\\\nI1005 06:48:14.864345 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1005 06:48:14.868087 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1005 06:48:14.868114 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1005 06:48:14.868140 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1005 06:48:14.868147 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1005 06:48:14.877354 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1005 06:48:14.877412 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1005 06:48:14.877456 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1005 06:48:14.877471 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1005 06:48:14.877490 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1005 06:48:14.877501 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1005 06:48:14.877507 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1005 06:48:14.877515 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1005 06:48:14.880326 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:13Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4a091716ff935adb42ae2e88048619fd347bd174f64e5f9c97a01f81690e57f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:01Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dc5411b642f99b13d514d2f21defad0c394aeeb82a966f72300076617e2586c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5dc5411b642f99b13d514d2f21defad0c394aeeb82a966f72300076617e2586c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:47:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:47:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:47:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 05 06:48:18 crc kubenswrapper[4846]: I1005 06:48:18.959652 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c9f812b-ecc8-4985-a670-53fa80fbb499\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"message\\\":\\\"containers with unready status: [cluster-policy-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"message\\\":\\\"containers with unready status: [cluster-policy-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2b5ee9781c01d6aec3ac778efef1067b6c3bfe4ed37431bdebe188fcde0165f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:47:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3fb4717692aa1fec0012b514ef8d18d1ca3707cfa4bb8248bb93841a0a40f11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:47:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0b6a7afe3094f1c92093c05c28344897a615d590c2fbb507c87e18a44c6c831\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://640ade2c733d7916ec213a201a1dd614d261c69eab654c80100efbc99a768995\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:47:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 05 06:48:18 crc kubenswrapper[4846]: I1005 06:48:18.976985 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 05 06:48:18 crc kubenswrapper[4846]: I1005 06:48:18.989589 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 05 06:48:19 crc kubenswrapper[4846]: I1005 06:48:19.019975 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c482bb97-fd50-40cf-96ea-f806c3f01125\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bfe45d30e4fa141541aca1d46bc5cc1da0a7b4a5eb1fc9bb2caa5b7fdcb0b264\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://496a4782faa30a9a20ff63038141600cd635f12902d1cf17989cd067fcdfe929\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://281be077869a6aa1ca815770871a14413d4a97e97aeb550fd5cd215cba88b21f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6771c7215d2f5dd5838d17888375a4768b9d1755fc2e21bda1bbfbb2eb1dbb95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c17a711c2c7473b076a3944ffe7c700808d05cba582fbd55c5c2ac8ff73838a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44a9f740f0a06c8bf928e2a9d1d5979217558c31e73a333631baccebf00fd5d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://44a9f740f0a06c8bf928e2a9d1d5979217558c31e73a333631baccebf00fd5d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:47:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:47:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://48ca0ba6ca6092bfcb5b153220c054f02cc36fb22707c42e021f06d329e18e2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48ca0ba6ca6092bfcb5b153220c054f02cc36fb22707c42e021f06d329e18e2e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:00Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4006d0cea128675c8885d879171204abffc5e11ea0664750a12b4bc7344ccb2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4006d0cea128675c8885d879171204abffc5e11ea0664750a12b4bc7344ccb2e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:47:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 05 06:48:19 crc kubenswrapper[4846]: I1005 06:48:19.038590 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ff10c31459b3b3724ee7b769bd6f72f9ed80cce5419b6adcdf69a3239b70756\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 05 06:48:19 crc kubenswrapper[4846]: I1005 06:48:19.057121 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 05 06:48:19 crc kubenswrapper[4846]: I1005 06:48:19.074015 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 05 06:48:19 crc kubenswrapper[4846]: I1005 06:48:19.088251 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56f1fe1f9f06f0249cf937a6963fda82467732dc056ca2b2189be6d8dd9d4a05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7212ce5f42c1fb55146abb190a557a504b43f16af6df84f90e1073567a329ec5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 05 06:48:19 crc kubenswrapper[4846]: I1005 06:48:19.349239 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 05 06:48:19 crc kubenswrapper[4846]: I1005 06:48:19.349315 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 05 06:48:19 crc kubenswrapper[4846]: I1005 06:48:19.349345 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 05 06:48:19 crc kubenswrapper[4846]: I1005 06:48:19.349365 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 05 06:48:19 crc kubenswrapper[4846]: I1005 06:48:19.349388 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 05 06:48:19 crc kubenswrapper[4846]: E1005 06:48:19.349479 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-05 06:48:21.349434649 +0000 UTC m=+23.590287424 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 06:48:19 crc kubenswrapper[4846]: E1005 06:48:19.349503 4846 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 05 06:48:19 crc kubenswrapper[4846]: E1005 06:48:19.349522 4846 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 05 06:48:19 crc kubenswrapper[4846]: E1005 06:48:19.349534 4846 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 05 06:48:19 crc kubenswrapper[4846]: E1005 06:48:19.349588 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-05 06:48:21.349572383 +0000 UTC m=+23.590425158 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 05 06:48:19 crc kubenswrapper[4846]: E1005 06:48:19.349643 4846 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 05 06:48:19 crc kubenswrapper[4846]: E1005 06:48:19.349690 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-05 06:48:21.349679867 +0000 UTC m=+23.590532642 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 05 06:48:19 crc kubenswrapper[4846]: E1005 06:48:19.349794 4846 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 05 06:48:19 crc kubenswrapper[4846]: E1005 06:48:19.349806 4846 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 05 06:48:19 crc kubenswrapper[4846]: E1005 06:48:19.349820 4846 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 05 06:48:19 crc kubenswrapper[4846]: E1005 06:48:19.349848 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-05 06:48:21.349840871 +0000 UTC m=+23.590693646 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 05 06:48:19 crc kubenswrapper[4846]: E1005 06:48:19.349897 4846 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 05 06:48:19 crc kubenswrapper[4846]: E1005 06:48:19.349920 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-05 06:48:21.349913663 +0000 UTC m=+23.590766438 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 05 06:48:19 crc kubenswrapper[4846]: I1005 06:48:19.497433 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 05 06:48:19 crc kubenswrapper[4846]: I1005 06:48:19.497506 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 05 06:48:19 crc kubenswrapper[4846]: E1005 06:48:19.497621 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 05 06:48:19 crc kubenswrapper[4846]: E1005 06:48:19.497813 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 05 06:48:19 crc kubenswrapper[4846]: I1005 06:48:19.660242 4846 scope.go:117] "RemoveContainer" containerID="c52e076eddfd307110b1f6de00a169b81af8f2785af7becef2a820ce363a5861" Oct 05 06:48:19 crc kubenswrapper[4846]: E1005 06:48:19.660518 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Oct 05 06:48:20 crc kubenswrapper[4846]: I1005 06:48:20.497348 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 05 06:48:20 crc kubenswrapper[4846]: E1005 06:48:20.497568 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 05 06:48:21 crc kubenswrapper[4846]: I1005 06:48:21.364085 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 05 06:48:21 crc kubenswrapper[4846]: I1005 06:48:21.364286 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 05 06:48:21 crc kubenswrapper[4846]: I1005 06:48:21.364339 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 05 06:48:21 crc kubenswrapper[4846]: I1005 06:48:21.364394 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 05 06:48:21 crc kubenswrapper[4846]: I1005 06:48:21.364452 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 05 06:48:21 crc kubenswrapper[4846]: E1005 06:48:21.364711 4846 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 05 06:48:21 crc kubenswrapper[4846]: E1005 06:48:21.364748 4846 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 05 06:48:21 crc kubenswrapper[4846]: E1005 06:48:21.364771 4846 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 05 06:48:21 crc kubenswrapper[4846]: E1005 06:48:21.364854 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-05 06:48:25.36482614 +0000 UTC m=+27.605678955 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 05 06:48:21 crc kubenswrapper[4846]: E1005 06:48:21.365375 4846 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 05 06:48:21 crc kubenswrapper[4846]: E1005 06:48:21.365453 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-05 06:48:25.365400978 +0000 UTC m=+27.606253783 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 06:48:21 crc kubenswrapper[4846]: E1005 06:48:21.365458 4846 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 05 06:48:21 crc kubenswrapper[4846]: E1005 06:48:21.365761 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-05 06:48:25.365726877 +0000 UTC m=+27.606579832 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 05 06:48:21 crc kubenswrapper[4846]: E1005 06:48:21.365775 4846 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 05 06:48:21 crc kubenswrapper[4846]: E1005 06:48:21.365805 4846 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 05 06:48:21 crc kubenswrapper[4846]: E1005 06:48:21.365868 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-05 06:48:25.365857711 +0000 UTC m=+27.606710706 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 05 06:48:21 crc kubenswrapper[4846]: E1005 06:48:21.365477 4846 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 05 06:48:21 crc kubenswrapper[4846]: E1005 06:48:21.365905 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-05 06:48:25.365899092 +0000 UTC m=+27.606751867 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 05 06:48:21 crc kubenswrapper[4846]: I1005 06:48:21.497101 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 05 06:48:21 crc kubenswrapper[4846]: I1005 06:48:21.497143 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 05 06:48:21 crc kubenswrapper[4846]: E1005 06:48:21.497385 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 05 06:48:21 crc kubenswrapper[4846]: E1005 06:48:21.497550 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 05 06:48:21 crc kubenswrapper[4846]: I1005 06:48:21.667394 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"347f900743b7bba3daf92f21c7eba0625e8c774f50057bc9472f0c82119325e3"} Oct 05 06:48:21 crc kubenswrapper[4846]: I1005 06:48:21.710672 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:21Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:21 crc kubenswrapper[4846]: I1005 06:48:21.731269 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:21Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:21 crc kubenswrapper[4846]: I1005 06:48:21.754891 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"775b5702-a131-4d6c-852d-91bfa056cac5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0839c1765df2c68d2554e9c66d289505e559a9672caf8069feeff2c4fdfd2171\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15dbaf2445704da0987f256f81dff2ee95664988541fa48ce94898afeadab2e8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd3b3ab323c6d5153054aa04af2d91f7b9e8436ff1e2f75f3594b57831b2a416\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c52e076eddfd307110b1f6de00a169b81af8f2785af7becef2a820ce363a5861\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c52e076eddfd307110b1f6de00a169b81af8f2785af7becef2a820ce363a5861\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-05T06:48:14Z\\\",\\\"message\\\":\\\"le observer\\\\nW1005 06:48:14.337870 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1005 06:48:14.338086 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1005 06:48:14.339389 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1467752215/tls.crt::/tmp/serving-cert-1467752215/tls.key\\\\\\\"\\\\nI1005 06:48:14.864345 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1005 06:48:14.868087 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1005 06:48:14.868114 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1005 06:48:14.868140 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1005 06:48:14.868147 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1005 06:48:14.877354 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1005 06:48:14.877412 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1005 06:48:14.877456 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1005 06:48:14.877471 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1005 06:48:14.877490 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1005 06:48:14.877501 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1005 06:48:14.877507 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1005 06:48:14.877515 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1005 06:48:14.880326 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:13Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4a091716ff935adb42ae2e88048619fd347bd174f64e5f9c97a01f81690e57f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:01Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dc5411b642f99b13d514d2f21defad0c394aeeb82a966f72300076617e2586c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5dc5411b642f99b13d514d2f21defad0c394aeeb82a966f72300076617e2586c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:47:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:47:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:47:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:21Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:21 crc kubenswrapper[4846]: I1005 06:48:21.778554 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c9f812b-ecc8-4985-a670-53fa80fbb499\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"message\\\":\\\"containers with unready status: [cluster-policy-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"message\\\":\\\"containers with unready status: [cluster-policy-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2b5ee9781c01d6aec3ac778efef1067b6c3bfe4ed37431bdebe188fcde0165f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:47:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3fb4717692aa1fec0012b514ef8d18d1ca3707cfa4bb8248bb93841a0a40f11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:47:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0b6a7afe3094f1c92093c05c28344897a615d590c2fbb507c87e18a44c6c831\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://640ade2c733d7916ec213a201a1dd614d261c69eab654c80100efbc99a768995\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:47:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:21Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:21 crc kubenswrapper[4846]: I1005 06:48:21.797915 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:21Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:21 crc kubenswrapper[4846]: I1005 06:48:21.815375 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://347f900743b7bba3daf92f21c7eba0625e8c774f50057bc9472f0c82119325e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:21Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:21 crc kubenswrapper[4846]: I1005 06:48:21.832829 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56f1fe1f9f06f0249cf937a6963fda82467732dc056ca2b2189be6d8dd9d4a05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7212ce5f42c1fb55146abb190a557a504b43f16af6df84f90e1073567a329ec5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:21Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:21 crc kubenswrapper[4846]: I1005 06:48:21.865014 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c482bb97-fd50-40cf-96ea-f806c3f01125\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bfe45d30e4fa141541aca1d46bc5cc1da0a7b4a5eb1fc9bb2caa5b7fdcb0b264\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://496a4782faa30a9a20ff63038141600cd635f12902d1cf17989cd067fcdfe929\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://281be077869a6aa1ca815770871a14413d4a97e97aeb550fd5cd215cba88b21f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6771c7215d2f5dd5838d17888375a4768b9d1755fc2e21bda1bbfbb2eb1dbb95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c17a711c2c7473b076a3944ffe7c700808d05cba582fbd55c5c2ac8ff73838a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44a9f740f0a06c8bf928e2a9d1d5979217558c31e73a333631baccebf00fd5d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://44a9f740f0a06c8bf928e2a9d1d5979217558c31e73a333631baccebf00fd5d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:47:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:47:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://48ca0ba6ca6092bfcb5b153220c054f02cc36fb22707c42e021f06d329e18e2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48ca0ba6ca6092bfcb5b153220c054f02cc36fb22707c42e021f06d329e18e2e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:00Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4006d0cea128675c8885d879171204abffc5e11ea0664750a12b4bc7344ccb2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4006d0cea128675c8885d879171204abffc5e11ea0664750a12b4bc7344ccb2e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:47:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:21Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:21 crc kubenswrapper[4846]: I1005 06:48:21.888246 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ff10c31459b3b3724ee7b769bd6f72f9ed80cce5419b6adcdf69a3239b70756\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:21Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:22 crc kubenswrapper[4846]: I1005 06:48:22.496838 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 05 06:48:22 crc kubenswrapper[4846]: E1005 06:48:22.497053 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 05 06:48:23 crc kubenswrapper[4846]: I1005 06:48:23.496715 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 05 06:48:23 crc kubenswrapper[4846]: I1005 06:48:23.496792 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 05 06:48:23 crc kubenswrapper[4846]: E1005 06:48:23.497052 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 05 06:48:23 crc kubenswrapper[4846]: E1005 06:48:23.497231 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 05 06:48:23 crc kubenswrapper[4846]: I1005 06:48:23.950559 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 05 06:48:23 crc kubenswrapper[4846]: I1005 06:48:23.954720 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 05 06:48:23 crc kubenswrapper[4846]: I1005 06:48:23.969000 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c9f812b-ecc8-4985-a670-53fa80fbb499\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"message\\\":\\\"containers with unready status: [cluster-policy-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"message\\\":\\\"containers with unready status: [cluster-policy-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2b5ee9781c01d6aec3ac778efef1067b6c3bfe4ed37431bdebe188fcde0165f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:47:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3fb4717692aa1fec0012b514ef8d18d1ca3707cfa4bb8248bb93841a0a40f11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:47:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0b6a7afe3094f1c92093c05c28344897a615d590c2fbb507c87e18a44c6c831\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://640ade2c733d7916ec213a201a1dd614d261c69eab654c80100efbc99a768995\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:47:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:23Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:23 crc kubenswrapper[4846]: I1005 06:48:23.985199 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:23Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.004727 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:24Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.019830 4846 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.022638 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.022679 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.022691 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.022757 4846 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.024429 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"775b5702-a131-4d6c-852d-91bfa056cac5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0839c1765df2c68d2554e9c66d289505e559a9672caf8069feeff2c4fdfd2171\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15dbaf2445704da0987f256f81dff2ee95664988541fa48ce94898afeadab2e8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd3b3ab323c6d5153054aa04af2d91f7b9e8436ff1e2f75f3594b57831b2a416\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c52e076eddfd307110b1f6de00a169b81af8f2785af7becef2a820ce363a5861\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c52e076eddfd307110b1f6de00a169b81af8f2785af7becef2a820ce363a5861\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-05T06:48:14Z\\\",\\\"message\\\":\\\"le observer\\\\nW1005 06:48:14.337870 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1005 06:48:14.338086 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1005 06:48:14.339389 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1467752215/tls.crt::/tmp/serving-cert-1467752215/tls.key\\\\\\\"\\\\nI1005 06:48:14.864345 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1005 06:48:14.868087 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1005 06:48:14.868114 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1005 06:48:14.868140 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1005 06:48:14.868147 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1005 06:48:14.877354 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1005 06:48:14.877412 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1005 06:48:14.877456 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1005 06:48:14.877471 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1005 06:48:14.877490 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1005 06:48:14.877501 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1005 06:48:14.877507 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1005 06:48:14.877515 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1005 06:48:14.880326 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:13Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4a091716ff935adb42ae2e88048619fd347bd174f64e5f9c97a01f81690e57f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:01Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dc5411b642f99b13d514d2f21defad0c394aeeb82a966f72300076617e2586c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5dc5411b642f99b13d514d2f21defad0c394aeeb82a966f72300076617e2586c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:47:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:47:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:47:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:24Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.031355 4846 kubelet_node_status.go:115] "Node was previously registered" node="crc" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.031757 4846 kubelet_node_status.go:79] "Successfully registered node" node="crc" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.033068 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.033117 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.033129 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.033146 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.033155 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:24Z","lastTransitionTime":"2025-10-05T06:48:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.040843 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ff10c31459b3b3724ee7b769bd6f72f9ed80cce5419b6adcdf69a3239b70756\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:24Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.052905 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:24Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:24 crc kubenswrapper[4846]: E1005 06:48:24.060565 4846 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"04ce628d-2f61-4931-aad6-615924ea53fb\\\",\\\"systemUUID\\\":\\\"a050a23b-e773-4ba9-989b-360aa53a0605\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:24Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.067702 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.067760 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.067775 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.067798 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.067813 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:24Z","lastTransitionTime":"2025-10-05T06:48:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.072524 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://347f900743b7bba3daf92f21c7eba0625e8c774f50057bc9472f0c82119325e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:24Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:24 crc kubenswrapper[4846]: E1005 06:48:24.086722 4846 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"04ce628d-2f61-4931-aad6-615924ea53fb\\\",\\\"systemUUID\\\":\\\"a050a23b-e773-4ba9-989b-360aa53a0605\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:24Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.092275 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.092312 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.092325 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.092342 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.092351 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:24Z","lastTransitionTime":"2025-10-05T06:48:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.095008 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56f1fe1f9f06f0249cf937a6963fda82467732dc056ca2b2189be6d8dd9d4a05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7212ce5f42c1fb55146abb190a557a504b43f16af6df84f90e1073567a329ec5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:24Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:24 crc kubenswrapper[4846]: E1005 06:48:24.105129 4846 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"04ce628d-2f61-4931-aad6-615924ea53fb\\\",\\\"systemUUID\\\":\\\"a050a23b-e773-4ba9-989b-360aa53a0605\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:24Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.108822 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.108871 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.108880 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.108898 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.108909 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:24Z","lastTransitionTime":"2025-10-05T06:48:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.115514 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c482bb97-fd50-40cf-96ea-f806c3f01125\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bfe45d30e4fa141541aca1d46bc5cc1da0a7b4a5eb1fc9bb2caa5b7fdcb0b264\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://496a4782faa30a9a20ff63038141600cd635f12902d1cf17989cd067fcdfe929\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://281be077869a6aa1ca815770871a14413d4a97e97aeb550fd5cd215cba88b21f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6771c7215d2f5dd5838d17888375a4768b9d1755fc2e21bda1bbfbb2eb1dbb95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c17a711c2c7473b076a3944ffe7c700808d05cba582fbd55c5c2ac8ff73838a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44a9f740f0a06c8bf928e2a9d1d5979217558c31e73a333631baccebf00fd5d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://44a9f740f0a06c8bf928e2a9d1d5979217558c31e73a333631baccebf00fd5d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:47:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:47:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://48ca0ba6ca6092bfcb5b153220c054f02cc36fb22707c42e021f06d329e18e2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48ca0ba6ca6092bfcb5b153220c054f02cc36fb22707c42e021f06d329e18e2e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:00Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4006d0cea128675c8885d879171204abffc5e11ea0664750a12b4bc7344ccb2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4006d0cea128675c8885d879171204abffc5e11ea0664750a12b4bc7344ccb2e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:47:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:24Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:24 crc kubenswrapper[4846]: E1005 06:48:24.119165 4846 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"04ce628d-2f61-4931-aad6-615924ea53fb\\\",\\\"systemUUID\\\":\\\"a050a23b-e773-4ba9-989b-360aa53a0605\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:24Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.122655 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.122688 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.122699 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.122718 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.122731 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:24Z","lastTransitionTime":"2025-10-05T06:48:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.127277 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://347f900743b7bba3daf92f21c7eba0625e8c774f50057bc9472f0c82119325e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:24Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:24 crc kubenswrapper[4846]: E1005 06:48:24.135297 4846 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"04ce628d-2f61-4931-aad6-615924ea53fb\\\",\\\"systemUUID\\\":\\\"a050a23b-e773-4ba9-989b-360aa53a0605\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:24Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:24 crc kubenswrapper[4846]: E1005 06:48:24.135574 4846 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.137395 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.137482 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.137497 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.137521 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.137534 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:24Z","lastTransitionTime":"2025-10-05T06:48:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.140765 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56f1fe1f9f06f0249cf937a6963fda82467732dc056ca2b2189be6d8dd9d4a05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7212ce5f42c1fb55146abb190a557a504b43f16af6df84f90e1073567a329ec5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:24Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.161463 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c482bb97-fd50-40cf-96ea-f806c3f01125\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bfe45d30e4fa141541aca1d46bc5cc1da0a7b4a5eb1fc9bb2caa5b7fdcb0b264\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://496a4782faa30a9a20ff63038141600cd635f12902d1cf17989cd067fcdfe929\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://281be077869a6aa1ca815770871a14413d4a97e97aeb550fd5cd215cba88b21f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6771c7215d2f5dd5838d17888375a4768b9d1755fc2e21bda1bbfbb2eb1dbb95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c17a711c2c7473b076a3944ffe7c700808d05cba582fbd55c5c2ac8ff73838a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44a9f740f0a06c8bf928e2a9d1d5979217558c31e73a333631baccebf00fd5d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://44a9f740f0a06c8bf928e2a9d1d5979217558c31e73a333631baccebf00fd5d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:47:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:47:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://48ca0ba6ca6092bfcb5b153220c054f02cc36fb22707c42e021f06d329e18e2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48ca0ba6ca6092bfcb5b153220c054f02cc36fb22707c42e021f06d329e18e2e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:00Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4006d0cea128675c8885d879171204abffc5e11ea0664750a12b4bc7344ccb2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4006d0cea128675c8885d879171204abffc5e11ea0664750a12b4bc7344ccb2e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:47:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:24Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.176654 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ff10c31459b3b3724ee7b769bd6f72f9ed80cce5419b6adcdf69a3239b70756\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:24Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.191858 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:24Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.242807 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/node-resolver-r7ztb"] Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.243123 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-r7ztb" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.243744 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-daemon-fscvf"] Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.243930 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-additional-cni-plugins-sfmxd"] Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.244281 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.244319 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.244328 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.244332 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-7wm5q"] Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.244344 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.244354 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:24Z","lastTransitionTime":"2025-10-05T06:48:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.244468 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.244487 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-7wm5q" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.244878 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-sfmxd" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.245100 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.245442 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.248418 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.252092 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.252144 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.252207 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.252742 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.253472 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.253505 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.253539 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.253551 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.253491 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.254428 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.254786 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.257152 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.257677 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:24Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.274315 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"775b5702-a131-4d6c-852d-91bfa056cac5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0839c1765df2c68d2554e9c66d289505e559a9672caf8069feeff2c4fdfd2171\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15dbaf2445704da0987f256f81dff2ee95664988541fa48ce94898afeadab2e8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd3b3ab323c6d5153054aa04af2d91f7b9e8436ff1e2f75f3594b57831b2a416\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c52e076eddfd307110b1f6de00a169b81af8f2785af7becef2a820ce363a5861\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c52e076eddfd307110b1f6de00a169b81af8f2785af7becef2a820ce363a5861\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-05T06:48:14Z\\\",\\\"message\\\":\\\"le observer\\\\nW1005 06:48:14.337870 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1005 06:48:14.338086 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1005 06:48:14.339389 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1467752215/tls.crt::/tmp/serving-cert-1467752215/tls.key\\\\\\\"\\\\nI1005 06:48:14.864345 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1005 06:48:14.868087 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1005 06:48:14.868114 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1005 06:48:14.868140 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1005 06:48:14.868147 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1005 06:48:14.877354 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1005 06:48:14.877412 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1005 06:48:14.877456 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1005 06:48:14.877471 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1005 06:48:14.877490 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1005 06:48:14.877501 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1005 06:48:14.877507 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1005 06:48:14.877515 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1005 06:48:14.880326 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:13Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4a091716ff935adb42ae2e88048619fd347bd174f64e5f9c97a01f81690e57f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:01Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dc5411b642f99b13d514d2f21defad0c394aeeb82a966f72300076617e2586c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5dc5411b642f99b13d514d2f21defad0c394aeeb82a966f72300076617e2586c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:47:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:47:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:47:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:24Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.289425 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/9bc7744f-d700-450f-8b8d-98140877fee9-system-cni-dir\") pod \"multus-7wm5q\" (UID: \"9bc7744f-d700-450f-8b8d-98140877fee9\") " pod="openshift-multus/multus-7wm5q" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.289473 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/9bc7744f-d700-450f-8b8d-98140877fee9-host-var-lib-cni-bin\") pod \"multus-7wm5q\" (UID: \"9bc7744f-d700-450f-8b8d-98140877fee9\") " pod="openshift-multus/multus-7wm5q" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.289516 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9jfcj\" (UniqueName: \"kubernetes.io/projected/a5119cd9-af70-4324-a955-978306ab0b20-kube-api-access-9jfcj\") pod \"node-resolver-r7ztb\" (UID: \"a5119cd9-af70-4324-a955-978306ab0b20\") " pod="openshift-dns/node-resolver-r7ztb" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.289585 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/9bc7744f-d700-450f-8b8d-98140877fee9-host-var-lib-cni-multus\") pod \"multus-7wm5q\" (UID: \"9bc7744f-d700-450f-8b8d-98140877fee9\") " pod="openshift-multus/multus-7wm5q" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.289642 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/9bc7744f-d700-450f-8b8d-98140877fee9-host-var-lib-kubelet\") pod \"multus-7wm5q\" (UID: \"9bc7744f-d700-450f-8b8d-98140877fee9\") " pod="openshift-multus/multus-7wm5q" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.289701 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/9bc7744f-d700-450f-8b8d-98140877fee9-multus-conf-dir\") pod \"multus-7wm5q\" (UID: \"9bc7744f-d700-450f-8b8d-98140877fee9\") " pod="openshift-multus/multus-7wm5q" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.289754 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7-rootfs\") pod \"machine-config-daemon-fscvf\" (UID: \"ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7\") " pod="openshift-machine-config-operator/machine-config-daemon-fscvf" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.289783 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/9bc7744f-d700-450f-8b8d-98140877fee9-multus-cni-dir\") pod \"multus-7wm5q\" (UID: \"9bc7744f-d700-450f-8b8d-98140877fee9\") " pod="openshift-multus/multus-7wm5q" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.289836 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7-proxy-tls\") pod \"machine-config-daemon-fscvf\" (UID: \"ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7\") " pod="openshift-machine-config-operator/machine-config-daemon-fscvf" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.289859 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/9bc7744f-d700-450f-8b8d-98140877fee9-host-run-netns\") pod \"multus-7wm5q\" (UID: \"9bc7744f-d700-450f-8b8d-98140877fee9\") " pod="openshift-multus/multus-7wm5q" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.289879 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/1dbb13ac-136c-4a1c-9229-b4e8ac4f9c71-tuning-conf-dir\") pod \"multus-additional-cni-plugins-sfmxd\" (UID: \"1dbb13ac-136c-4a1c-9229-b4e8ac4f9c71\") " pod="openshift-multus/multus-additional-cni-plugins-sfmxd" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.289902 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/1dbb13ac-136c-4a1c-9229-b4e8ac4f9c71-cni-binary-copy\") pod \"multus-additional-cni-plugins-sfmxd\" (UID: \"1dbb13ac-136c-4a1c-9229-b4e8ac4f9c71\") " pod="openshift-multus/multus-additional-cni-plugins-sfmxd" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.289994 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h5p4q\" (UniqueName: \"kubernetes.io/projected/9bc7744f-d700-450f-8b8d-98140877fee9-kube-api-access-h5p4q\") pod \"multus-7wm5q\" (UID: \"9bc7744f-d700-450f-8b8d-98140877fee9\") " pod="openshift-multus/multus-7wm5q" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.290051 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/1dbb13ac-136c-4a1c-9229-b4e8ac4f9c71-system-cni-dir\") pod \"multus-additional-cni-plugins-sfmxd\" (UID: \"1dbb13ac-136c-4a1c-9229-b4e8ac4f9c71\") " pod="openshift-multus/multus-additional-cni-plugins-sfmxd" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.290074 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/9bc7744f-d700-450f-8b8d-98140877fee9-cni-binary-copy\") pod \"multus-7wm5q\" (UID: \"9bc7744f-d700-450f-8b8d-98140877fee9\") " pod="openshift-multus/multus-7wm5q" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.290093 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/9bc7744f-d700-450f-8b8d-98140877fee9-multus-daemon-config\") pod \"multus-7wm5q\" (UID: \"9bc7744f-d700-450f-8b8d-98140877fee9\") " pod="openshift-multus/multus-7wm5q" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.290124 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/9bc7744f-d700-450f-8b8d-98140877fee9-host-run-multus-certs\") pod \"multus-7wm5q\" (UID: \"9bc7744f-d700-450f-8b8d-98140877fee9\") " pod="openshift-multus/multus-7wm5q" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.290154 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/1dbb13ac-136c-4a1c-9229-b4e8ac4f9c71-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-sfmxd\" (UID: \"1dbb13ac-136c-4a1c-9229-b4e8ac4f9c71\") " pod="openshift-multus/multus-additional-cni-plugins-sfmxd" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.290196 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dp2wt\" (UniqueName: \"kubernetes.io/projected/1dbb13ac-136c-4a1c-9229-b4e8ac4f9c71-kube-api-access-dp2wt\") pod \"multus-additional-cni-plugins-sfmxd\" (UID: \"1dbb13ac-136c-4a1c-9229-b4e8ac4f9c71\") " pod="openshift-multus/multus-additional-cni-plugins-sfmxd" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.290217 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dkhl7\" (UniqueName: \"kubernetes.io/projected/ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7-kube-api-access-dkhl7\") pod \"machine-config-daemon-fscvf\" (UID: \"ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7\") " pod="openshift-machine-config-operator/machine-config-daemon-fscvf" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.290236 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/1dbb13ac-136c-4a1c-9229-b4e8ac4f9c71-cnibin\") pod \"multus-additional-cni-plugins-sfmxd\" (UID: \"1dbb13ac-136c-4a1c-9229-b4e8ac4f9c71\") " pod="openshift-multus/multus-additional-cni-plugins-sfmxd" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.290282 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/1dbb13ac-136c-4a1c-9229-b4e8ac4f9c71-os-release\") pod \"multus-additional-cni-plugins-sfmxd\" (UID: \"1dbb13ac-136c-4a1c-9229-b4e8ac4f9c71\") " pod="openshift-multus/multus-additional-cni-plugins-sfmxd" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.290302 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/9bc7744f-d700-450f-8b8d-98140877fee9-cnibin\") pod \"multus-7wm5q\" (UID: \"9bc7744f-d700-450f-8b8d-98140877fee9\") " pod="openshift-multus/multus-7wm5q" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.290318 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/9bc7744f-d700-450f-8b8d-98140877fee9-os-release\") pod \"multus-7wm5q\" (UID: \"9bc7744f-d700-450f-8b8d-98140877fee9\") " pod="openshift-multus/multus-7wm5q" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.290343 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/9bc7744f-d700-450f-8b8d-98140877fee9-multus-socket-dir-parent\") pod \"multus-7wm5q\" (UID: \"9bc7744f-d700-450f-8b8d-98140877fee9\") " pod="openshift-multus/multus-7wm5q" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.290362 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/9bc7744f-d700-450f-8b8d-98140877fee9-host-run-k8s-cni-cncf-io\") pod \"multus-7wm5q\" (UID: \"9bc7744f-d700-450f-8b8d-98140877fee9\") " pod="openshift-multus/multus-7wm5q" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.290378 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/9bc7744f-d700-450f-8b8d-98140877fee9-hostroot\") pod \"multus-7wm5q\" (UID: \"9bc7744f-d700-450f-8b8d-98140877fee9\") " pod="openshift-multus/multus-7wm5q" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.290424 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/a5119cd9-af70-4324-a955-978306ab0b20-hosts-file\") pod \"node-resolver-r7ztb\" (UID: \"a5119cd9-af70-4324-a955-978306ab0b20\") " pod="openshift-dns/node-resolver-r7ztb" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.290444 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/9bc7744f-d700-450f-8b8d-98140877fee9-etc-kubernetes\") pod \"multus-7wm5q\" (UID: \"9bc7744f-d700-450f-8b8d-98140877fee9\") " pod="openshift-multus/multus-7wm5q" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.290460 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7-mcd-auth-proxy-config\") pod \"machine-config-daemon-fscvf\" (UID: \"ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7\") " pod="openshift-machine-config-operator/machine-config-daemon-fscvf" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.295402 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c9f812b-ecc8-4985-a670-53fa80fbb499\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2b5ee9781c01d6aec3ac778efef1067b6c3bfe4ed37431bdebe188fcde0165f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:47:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3fb4717692aa1fec0012b514ef8d18d1ca3707cfa4bb8248bb93841a0a40f11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:47:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0b6a7afe3094f1c92093c05c28344897a615d590c2fbb507c87e18a44c6c831\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://640ade2c733d7916ec213a201a1dd614d261c69eab654c80100efbc99a768995\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:47:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:24Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.312913 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:24Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.336873 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c9f812b-ecc8-4985-a670-53fa80fbb499\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2b5ee9781c01d6aec3ac778efef1067b6c3bfe4ed37431bdebe188fcde0165f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:47:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3fb4717692aa1fec0012b514ef8d18d1ca3707cfa4bb8248bb93841a0a40f11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:47:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0b6a7afe3094f1c92093c05c28344897a615d590c2fbb507c87e18a44c6c831\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://640ade2c733d7916ec213a201a1dd614d261c69eab654c80100efbc99a768995\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:47:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:24Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.347151 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.347218 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.347232 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.347252 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.347264 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:24Z","lastTransitionTime":"2025-10-05T06:48:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.385620 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:24Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.391403 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9jfcj\" (UniqueName: \"kubernetes.io/projected/a5119cd9-af70-4324-a955-978306ab0b20-kube-api-access-9jfcj\") pod \"node-resolver-r7ztb\" (UID: \"a5119cd9-af70-4324-a955-978306ab0b20\") " pod="openshift-dns/node-resolver-r7ztb" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.391453 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/9bc7744f-d700-450f-8b8d-98140877fee9-system-cni-dir\") pod \"multus-7wm5q\" (UID: \"9bc7744f-d700-450f-8b8d-98140877fee9\") " pod="openshift-multus/multus-7wm5q" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.391478 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/9bc7744f-d700-450f-8b8d-98140877fee9-host-var-lib-cni-bin\") pod \"multus-7wm5q\" (UID: \"9bc7744f-d700-450f-8b8d-98140877fee9\") " pod="openshift-multus/multus-7wm5q" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.391521 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/9bc7744f-d700-450f-8b8d-98140877fee9-multus-conf-dir\") pod \"multus-7wm5q\" (UID: \"9bc7744f-d700-450f-8b8d-98140877fee9\") " pod="openshift-multus/multus-7wm5q" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.391544 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/9bc7744f-d700-450f-8b8d-98140877fee9-host-var-lib-cni-multus\") pod \"multus-7wm5q\" (UID: \"9bc7744f-d700-450f-8b8d-98140877fee9\") " pod="openshift-multus/multus-7wm5q" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.391566 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/9bc7744f-d700-450f-8b8d-98140877fee9-host-var-lib-kubelet\") pod \"multus-7wm5q\" (UID: \"9bc7744f-d700-450f-8b8d-98140877fee9\") " pod="openshift-multus/multus-7wm5q" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.391591 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7-rootfs\") pod \"machine-config-daemon-fscvf\" (UID: \"ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7\") " pod="openshift-machine-config-operator/machine-config-daemon-fscvf" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.391616 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/9bc7744f-d700-450f-8b8d-98140877fee9-multus-cni-dir\") pod \"multus-7wm5q\" (UID: \"9bc7744f-d700-450f-8b8d-98140877fee9\") " pod="openshift-multus/multus-7wm5q" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.391627 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/9bc7744f-d700-450f-8b8d-98140877fee9-host-var-lib-cni-bin\") pod \"multus-7wm5q\" (UID: \"9bc7744f-d700-450f-8b8d-98140877fee9\") " pod="openshift-multus/multus-7wm5q" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.391632 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/9bc7744f-d700-450f-8b8d-98140877fee9-system-cni-dir\") pod \"multus-7wm5q\" (UID: \"9bc7744f-d700-450f-8b8d-98140877fee9\") " pod="openshift-multus/multus-7wm5q" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.391639 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7-proxy-tls\") pod \"machine-config-daemon-fscvf\" (UID: \"ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7\") " pod="openshift-machine-config-operator/machine-config-daemon-fscvf" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.391698 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/9bc7744f-d700-450f-8b8d-98140877fee9-host-var-lib-kubelet\") pod \"multus-7wm5q\" (UID: \"9bc7744f-d700-450f-8b8d-98140877fee9\") " pod="openshift-multus/multus-7wm5q" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.391725 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/9bc7744f-d700-450f-8b8d-98140877fee9-host-run-netns\") pod \"multus-7wm5q\" (UID: \"9bc7744f-d700-450f-8b8d-98140877fee9\") " pod="openshift-multus/multus-7wm5q" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.391722 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/9bc7744f-d700-450f-8b8d-98140877fee9-host-var-lib-cni-multus\") pod \"multus-7wm5q\" (UID: \"9bc7744f-d700-450f-8b8d-98140877fee9\") " pod="openshift-multus/multus-7wm5q" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.391755 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/9bc7744f-d700-450f-8b8d-98140877fee9-host-run-netns\") pod \"multus-7wm5q\" (UID: \"9bc7744f-d700-450f-8b8d-98140877fee9\") " pod="openshift-multus/multus-7wm5q" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.391776 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/9bc7744f-d700-450f-8b8d-98140877fee9-multus-cni-dir\") pod \"multus-7wm5q\" (UID: \"9bc7744f-d700-450f-8b8d-98140877fee9\") " pod="openshift-multus/multus-7wm5q" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.391788 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/1dbb13ac-136c-4a1c-9229-b4e8ac4f9c71-tuning-conf-dir\") pod \"multus-additional-cni-plugins-sfmxd\" (UID: \"1dbb13ac-136c-4a1c-9229-b4e8ac4f9c71\") " pod="openshift-multus/multus-additional-cni-plugins-sfmxd" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.391820 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7-rootfs\") pod \"machine-config-daemon-fscvf\" (UID: \"ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7\") " pod="openshift-machine-config-operator/machine-config-daemon-fscvf" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.391829 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/9bc7744f-d700-450f-8b8d-98140877fee9-multus-conf-dir\") pod \"multus-7wm5q\" (UID: \"9bc7744f-d700-450f-8b8d-98140877fee9\") " pod="openshift-multus/multus-7wm5q" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.391831 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/1dbb13ac-136c-4a1c-9229-b4e8ac4f9c71-cni-binary-copy\") pod \"multus-additional-cni-plugins-sfmxd\" (UID: \"1dbb13ac-136c-4a1c-9229-b4e8ac4f9c71\") " pod="openshift-multus/multus-additional-cni-plugins-sfmxd" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.391998 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/9bc7744f-d700-450f-8b8d-98140877fee9-cni-binary-copy\") pod \"multus-7wm5q\" (UID: \"9bc7744f-d700-450f-8b8d-98140877fee9\") " pod="openshift-multus/multus-7wm5q" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.392019 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h5p4q\" (UniqueName: \"kubernetes.io/projected/9bc7744f-d700-450f-8b8d-98140877fee9-kube-api-access-h5p4q\") pod \"multus-7wm5q\" (UID: \"9bc7744f-d700-450f-8b8d-98140877fee9\") " pod="openshift-multus/multus-7wm5q" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.392054 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/1dbb13ac-136c-4a1c-9229-b4e8ac4f9c71-system-cni-dir\") pod \"multus-additional-cni-plugins-sfmxd\" (UID: \"1dbb13ac-136c-4a1c-9229-b4e8ac4f9c71\") " pod="openshift-multus/multus-additional-cni-plugins-sfmxd" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.392142 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dkhl7\" (UniqueName: \"kubernetes.io/projected/ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7-kube-api-access-dkhl7\") pod \"machine-config-daemon-fscvf\" (UID: \"ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7\") " pod="openshift-machine-config-operator/machine-config-daemon-fscvf" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.392159 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/9bc7744f-d700-450f-8b8d-98140877fee9-multus-daemon-config\") pod \"multus-7wm5q\" (UID: \"9bc7744f-d700-450f-8b8d-98140877fee9\") " pod="openshift-multus/multus-7wm5q" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.392112 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/1dbb13ac-136c-4a1c-9229-b4e8ac4f9c71-system-cni-dir\") pod \"multus-additional-cni-plugins-sfmxd\" (UID: \"1dbb13ac-136c-4a1c-9229-b4e8ac4f9c71\") " pod="openshift-multus/multus-additional-cni-plugins-sfmxd" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.392222 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/9bc7744f-d700-450f-8b8d-98140877fee9-host-run-multus-certs\") pod \"multus-7wm5q\" (UID: \"9bc7744f-d700-450f-8b8d-98140877fee9\") " pod="openshift-multus/multus-7wm5q" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.392246 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/1dbb13ac-136c-4a1c-9229-b4e8ac4f9c71-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-sfmxd\" (UID: \"1dbb13ac-136c-4a1c-9229-b4e8ac4f9c71\") " pod="openshift-multus/multus-additional-cni-plugins-sfmxd" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.392293 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/9bc7744f-d700-450f-8b8d-98140877fee9-host-run-multus-certs\") pod \"multus-7wm5q\" (UID: \"9bc7744f-d700-450f-8b8d-98140877fee9\") " pod="openshift-multus/multus-7wm5q" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.392517 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/1dbb13ac-136c-4a1c-9229-b4e8ac4f9c71-tuning-conf-dir\") pod \"multus-additional-cni-plugins-sfmxd\" (UID: \"1dbb13ac-136c-4a1c-9229-b4e8ac4f9c71\") " pod="openshift-multus/multus-additional-cni-plugins-sfmxd" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.392984 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/9bc7744f-d700-450f-8b8d-98140877fee9-cni-binary-copy\") pod \"multus-7wm5q\" (UID: \"9bc7744f-d700-450f-8b8d-98140877fee9\") " pod="openshift-multus/multus-7wm5q" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.393002 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/9bc7744f-d700-450f-8b8d-98140877fee9-multus-daemon-config\") pod \"multus-7wm5q\" (UID: \"9bc7744f-d700-450f-8b8d-98140877fee9\") " pod="openshift-multus/multus-7wm5q" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.393030 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/1dbb13ac-136c-4a1c-9229-b4e8ac4f9c71-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-sfmxd\" (UID: \"1dbb13ac-136c-4a1c-9229-b4e8ac4f9c71\") " pod="openshift-multus/multus-additional-cni-plugins-sfmxd" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.393077 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dp2wt\" (UniqueName: \"kubernetes.io/projected/1dbb13ac-136c-4a1c-9229-b4e8ac4f9c71-kube-api-access-dp2wt\") pod \"multus-additional-cni-plugins-sfmxd\" (UID: \"1dbb13ac-136c-4a1c-9229-b4e8ac4f9c71\") " pod="openshift-multus/multus-additional-cni-plugins-sfmxd" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.393113 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/1dbb13ac-136c-4a1c-9229-b4e8ac4f9c71-cnibin\") pod \"multus-additional-cni-plugins-sfmxd\" (UID: \"1dbb13ac-136c-4a1c-9229-b4e8ac4f9c71\") " pod="openshift-multus/multus-additional-cni-plugins-sfmxd" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.393130 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/1dbb13ac-136c-4a1c-9229-b4e8ac4f9c71-os-release\") pod \"multus-additional-cni-plugins-sfmxd\" (UID: \"1dbb13ac-136c-4a1c-9229-b4e8ac4f9c71\") " pod="openshift-multus/multus-additional-cni-plugins-sfmxd" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.393151 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/9bc7744f-d700-450f-8b8d-98140877fee9-cnibin\") pod \"multus-7wm5q\" (UID: \"9bc7744f-d700-450f-8b8d-98140877fee9\") " pod="openshift-multus/multus-7wm5q" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.393167 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/9bc7744f-d700-450f-8b8d-98140877fee9-os-release\") pod \"multus-7wm5q\" (UID: \"9bc7744f-d700-450f-8b8d-98140877fee9\") " pod="openshift-multus/multus-7wm5q" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.393200 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/a5119cd9-af70-4324-a955-978306ab0b20-hosts-file\") pod \"node-resolver-r7ztb\" (UID: \"a5119cd9-af70-4324-a955-978306ab0b20\") " pod="openshift-dns/node-resolver-r7ztb" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.393215 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/9bc7744f-d700-450f-8b8d-98140877fee9-multus-socket-dir-parent\") pod \"multus-7wm5q\" (UID: \"9bc7744f-d700-450f-8b8d-98140877fee9\") " pod="openshift-multus/multus-7wm5q" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.393252 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/1dbb13ac-136c-4a1c-9229-b4e8ac4f9c71-os-release\") pod \"multus-additional-cni-plugins-sfmxd\" (UID: \"1dbb13ac-136c-4a1c-9229-b4e8ac4f9c71\") " pod="openshift-multus/multus-additional-cni-plugins-sfmxd" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.393274 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/1dbb13ac-136c-4a1c-9229-b4e8ac4f9c71-cnibin\") pod \"multus-additional-cni-plugins-sfmxd\" (UID: \"1dbb13ac-136c-4a1c-9229-b4e8ac4f9c71\") " pod="openshift-multus/multus-additional-cni-plugins-sfmxd" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.393273 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/9bc7744f-d700-450f-8b8d-98140877fee9-cnibin\") pod \"multus-7wm5q\" (UID: \"9bc7744f-d700-450f-8b8d-98140877fee9\") " pod="openshift-multus/multus-7wm5q" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.393294 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/9bc7744f-d700-450f-8b8d-98140877fee9-os-release\") pod \"multus-7wm5q\" (UID: \"9bc7744f-d700-450f-8b8d-98140877fee9\") " pod="openshift-multus/multus-7wm5q" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.393306 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/a5119cd9-af70-4324-a955-978306ab0b20-hosts-file\") pod \"node-resolver-r7ztb\" (UID: \"a5119cd9-af70-4324-a955-978306ab0b20\") " pod="openshift-dns/node-resolver-r7ztb" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.393311 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/9bc7744f-d700-450f-8b8d-98140877fee9-multus-socket-dir-parent\") pod \"multus-7wm5q\" (UID: \"9bc7744f-d700-450f-8b8d-98140877fee9\") " pod="openshift-multus/multus-7wm5q" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.393313 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/9bc7744f-d700-450f-8b8d-98140877fee9-host-run-k8s-cni-cncf-io\") pod \"multus-7wm5q\" (UID: \"9bc7744f-d700-450f-8b8d-98140877fee9\") " pod="openshift-multus/multus-7wm5q" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.393337 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/9bc7744f-d700-450f-8b8d-98140877fee9-host-run-k8s-cni-cncf-io\") pod \"multus-7wm5q\" (UID: \"9bc7744f-d700-450f-8b8d-98140877fee9\") " pod="openshift-multus/multus-7wm5q" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.393355 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/9bc7744f-d700-450f-8b8d-98140877fee9-hostroot\") pod \"multus-7wm5q\" (UID: \"9bc7744f-d700-450f-8b8d-98140877fee9\") " pod="openshift-multus/multus-7wm5q" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.393383 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/9bc7744f-d700-450f-8b8d-98140877fee9-etc-kubernetes\") pod \"multus-7wm5q\" (UID: \"9bc7744f-d700-450f-8b8d-98140877fee9\") " pod="openshift-multus/multus-7wm5q" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.393403 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7-mcd-auth-proxy-config\") pod \"machine-config-daemon-fscvf\" (UID: \"ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7\") " pod="openshift-machine-config-operator/machine-config-daemon-fscvf" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.393417 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/9bc7744f-d700-450f-8b8d-98140877fee9-hostroot\") pod \"multus-7wm5q\" (UID: \"9bc7744f-d700-450f-8b8d-98140877fee9\") " pod="openshift-multus/multus-7wm5q" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.393455 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/9bc7744f-d700-450f-8b8d-98140877fee9-etc-kubernetes\") pod \"multus-7wm5q\" (UID: \"9bc7744f-d700-450f-8b8d-98140877fee9\") " pod="openshift-multus/multus-7wm5q" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.393897 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7-mcd-auth-proxy-config\") pod \"machine-config-daemon-fscvf\" (UID: \"ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7\") " pod="openshift-machine-config-operator/machine-config-daemon-fscvf" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.394133 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/1dbb13ac-136c-4a1c-9229-b4e8ac4f9c71-cni-binary-copy\") pod \"multus-additional-cni-plugins-sfmxd\" (UID: \"1dbb13ac-136c-4a1c-9229-b4e8ac4f9c71\") " pod="openshift-multus/multus-additional-cni-plugins-sfmxd" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.399722 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7-proxy-tls\") pod \"machine-config-daemon-fscvf\" (UID: \"ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7\") " pod="openshift-machine-config-operator/machine-config-daemon-fscvf" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.404817 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dkhl7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dkhl7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-fscvf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:24Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.413608 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9jfcj\" (UniqueName: \"kubernetes.io/projected/a5119cd9-af70-4324-a955-978306ab0b20-kube-api-access-9jfcj\") pod \"node-resolver-r7ztb\" (UID: \"a5119cd9-af70-4324-a955-978306ab0b20\") " pod="openshift-dns/node-resolver-r7ztb" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.419093 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dkhl7\" (UniqueName: \"kubernetes.io/projected/ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7-kube-api-access-dkhl7\") pod \"machine-config-daemon-fscvf\" (UID: \"ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7\") " pod="openshift-machine-config-operator/machine-config-daemon-fscvf" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.419587 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h5p4q\" (UniqueName: \"kubernetes.io/projected/9bc7744f-d700-450f-8b8d-98140877fee9-kube-api-access-h5p4q\") pod \"multus-7wm5q\" (UID: \"9bc7744f-d700-450f-8b8d-98140877fee9\") " pod="openshift-multus/multus-7wm5q" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.422493 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ff10c31459b3b3724ee7b769bd6f72f9ed80cce5419b6adcdf69a3239b70756\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:24Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.425880 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dp2wt\" (UniqueName: \"kubernetes.io/projected/1dbb13ac-136c-4a1c-9229-b4e8ac4f9c71-kube-api-access-dp2wt\") pod \"multus-additional-cni-plugins-sfmxd\" (UID: \"1dbb13ac-136c-4a1c-9229-b4e8ac4f9c71\") " pod="openshift-multus/multus-additional-cni-plugins-sfmxd" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.437934 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://347f900743b7bba3daf92f21c7eba0625e8c774f50057bc9472f0c82119325e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:24Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.450060 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.450117 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.450128 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.450146 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.450159 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:24Z","lastTransitionTime":"2025-10-05T06:48:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.453508 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56f1fe1f9f06f0249cf937a6963fda82467732dc056ca2b2189be6d8dd9d4a05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7212ce5f42c1fb55146abb190a557a504b43f16af6df84f90e1073567a329ec5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:24Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.466961 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-7wm5q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9bc7744f-d700-450f-8b8d-98140877fee9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h5p4q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-7wm5q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:24Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.480243 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"775b5702-a131-4d6c-852d-91bfa056cac5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0839c1765df2c68d2554e9c66d289505e559a9672caf8069feeff2c4fdfd2171\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15dbaf2445704da0987f256f81dff2ee95664988541fa48ce94898afeadab2e8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd3b3ab323c6d5153054aa04af2d91f7b9e8436ff1e2f75f3594b57831b2a416\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c52e076eddfd307110b1f6de00a169b81af8f2785af7becef2a820ce363a5861\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c52e076eddfd307110b1f6de00a169b81af8f2785af7becef2a820ce363a5861\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-05T06:48:14Z\\\",\\\"message\\\":\\\"le observer\\\\nW1005 06:48:14.337870 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1005 06:48:14.338086 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1005 06:48:14.339389 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1467752215/tls.crt::/tmp/serving-cert-1467752215/tls.key\\\\\\\"\\\\nI1005 06:48:14.864345 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1005 06:48:14.868087 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1005 06:48:14.868114 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1005 06:48:14.868140 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1005 06:48:14.868147 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1005 06:48:14.877354 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1005 06:48:14.877412 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1005 06:48:14.877456 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1005 06:48:14.877471 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1005 06:48:14.877490 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1005 06:48:14.877501 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1005 06:48:14.877507 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1005 06:48:14.877515 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1005 06:48:14.880326 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:13Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4a091716ff935adb42ae2e88048619fd347bd174f64e5f9c97a01f81690e57f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:01Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dc5411b642f99b13d514d2f21defad0c394aeeb82a966f72300076617e2586c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5dc5411b642f99b13d514d2f21defad0c394aeeb82a966f72300076617e2586c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:47:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:47:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:47:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:24Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.491598 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:24Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.496693 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 05 06:48:24 crc kubenswrapper[4846]: E1005 06:48:24.496848 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.504816 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-r7ztb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a5119cd9-af70-4324-a955-978306ab0b20\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jfcj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-r7ztb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:24Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.531577 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c482bb97-fd50-40cf-96ea-f806c3f01125\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bfe45d30e4fa141541aca1d46bc5cc1da0a7b4a5eb1fc9bb2caa5b7fdcb0b264\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://496a4782faa30a9a20ff63038141600cd635f12902d1cf17989cd067fcdfe929\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://281be077869a6aa1ca815770871a14413d4a97e97aeb550fd5cd215cba88b21f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6771c7215d2f5dd5838d17888375a4768b9d1755fc2e21bda1bbfbb2eb1dbb95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c17a711c2c7473b076a3944ffe7c700808d05cba582fbd55c5c2ac8ff73838a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44a9f740f0a06c8bf928e2a9d1d5979217558c31e73a333631baccebf00fd5d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://44a9f740f0a06c8bf928e2a9d1d5979217558c31e73a333631baccebf00fd5d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:47:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:47:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://48ca0ba6ca6092bfcb5b153220c054f02cc36fb22707c42e021f06d329e18e2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48ca0ba6ca6092bfcb5b153220c054f02cc36fb22707c42e021f06d329e18e2e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:00Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4006d0cea128675c8885d879171204abffc5e11ea0664750a12b4bc7344ccb2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4006d0cea128675c8885d879171204abffc5e11ea0664750a12b4bc7344ccb2e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:47:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:24Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.543975 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:24Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.555664 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.555778 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.555813 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.555834 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.555853 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:24Z","lastTransitionTime":"2025-10-05T06:48:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.557988 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-sfmxd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1dbb13ac-136c-4a1c-9229-b4e8ac4f9c71\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-sfmxd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:24Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.566420 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-r7ztb" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.575488 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-7wm5q" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.586322 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" Oct 05 06:48:24 crc kubenswrapper[4846]: W1005 06:48:24.588293 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9bc7744f_d700_450f_8b8d_98140877fee9.slice/crio-27caa176edf9306753199fc4a7b3cf5f7b1d278568c90dd7b4be344fa4183f9c WatchSource:0}: Error finding container 27caa176edf9306753199fc4a7b3cf5f7b1d278568c90dd7b4be344fa4183f9c: Status 404 returned error can't find the container with id 27caa176edf9306753199fc4a7b3cf5f7b1d278568c90dd7b4be344fa4183f9c Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.591918 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-sfmxd" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.634014 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-txzlk"] Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.634776 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-txzlk" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.639199 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.639351 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.639544 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.639683 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.639795 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.640216 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.640354 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.653453 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:24Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.661147 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.661213 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.661223 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.661242 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.661255 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:24Z","lastTransitionTime":"2025-10-05T06:48:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.673932 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dkhl7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dkhl7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-fscvf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:24Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.686448 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-sfmxd" event={"ID":"1dbb13ac-136c-4a1c-9229-b4e8ac4f9c71","Type":"ContainerStarted","Data":"5262ce958f3d3540d157aefcc1717f8d974a1466cb417da19d8bb019b276e55a"} Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.691312 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c9f812b-ecc8-4985-a670-53fa80fbb499\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2b5ee9781c01d6aec3ac778efef1067b6c3bfe4ed37431bdebe188fcde0165f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:47:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3fb4717692aa1fec0012b514ef8d18d1ca3707cfa4bb8248bb93841a0a40f11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:47:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0b6a7afe3094f1c92093c05c28344897a615d590c2fbb507c87e18a44c6c831\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://640ade2c733d7916ec213a201a1dd614d261c69eab654c80100efbc99a768995\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:47:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:24Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.693502 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" event={"ID":"ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7","Type":"ContainerStarted","Data":"17449f7fe85738483c250861b02e20663a576556461efbc9809345a462659bdc"} Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.694749 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-7wm5q" event={"ID":"9bc7744f-d700-450f-8b8d-98140877fee9","Type":"ContainerStarted","Data":"27caa176edf9306753199fc4a7b3cf5f7b1d278568c90dd7b4be344fa4183f9c"} Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.695553 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/93a6ace4-8d64-44d7-9637-457d3af6543c-run-systemd\") pod \"ovnkube-node-txzlk\" (UID: \"93a6ace4-8d64-44d7-9637-457d3af6543c\") " pod="openshift-ovn-kubernetes/ovnkube-node-txzlk" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.695580 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/93a6ace4-8d64-44d7-9637-457d3af6543c-run-openvswitch\") pod \"ovnkube-node-txzlk\" (UID: \"93a6ace4-8d64-44d7-9637-457d3af6543c\") " pod="openshift-ovn-kubernetes/ovnkube-node-txzlk" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.695595 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/93a6ace4-8d64-44d7-9637-457d3af6543c-host-run-netns\") pod \"ovnkube-node-txzlk\" (UID: \"93a6ace4-8d64-44d7-9637-457d3af6543c\") " pod="openshift-ovn-kubernetes/ovnkube-node-txzlk" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.695628 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/93a6ace4-8d64-44d7-9637-457d3af6543c-node-log\") pod \"ovnkube-node-txzlk\" (UID: \"93a6ace4-8d64-44d7-9637-457d3af6543c\") " pod="openshift-ovn-kubernetes/ovnkube-node-txzlk" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.695643 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/93a6ace4-8d64-44d7-9637-457d3af6543c-ovn-node-metrics-cert\") pod \"ovnkube-node-txzlk\" (UID: \"93a6ace4-8d64-44d7-9637-457d3af6543c\") " pod="openshift-ovn-kubernetes/ovnkube-node-txzlk" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.695661 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/93a6ace4-8d64-44d7-9637-457d3af6543c-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-txzlk\" (UID: \"93a6ace4-8d64-44d7-9637-457d3af6543c\") " pod="openshift-ovn-kubernetes/ovnkube-node-txzlk" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.695678 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/93a6ace4-8d64-44d7-9637-457d3af6543c-systemd-units\") pod \"ovnkube-node-txzlk\" (UID: \"93a6ace4-8d64-44d7-9637-457d3af6543c\") " pod="openshift-ovn-kubernetes/ovnkube-node-txzlk" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.695693 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/93a6ace4-8d64-44d7-9637-457d3af6543c-log-socket\") pod \"ovnkube-node-txzlk\" (UID: \"93a6ace4-8d64-44d7-9637-457d3af6543c\") " pod="openshift-ovn-kubernetes/ovnkube-node-txzlk" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.695728 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/93a6ace4-8d64-44d7-9637-457d3af6543c-ovnkube-config\") pod \"ovnkube-node-txzlk\" (UID: \"93a6ace4-8d64-44d7-9637-457d3af6543c\") " pod="openshift-ovn-kubernetes/ovnkube-node-txzlk" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.695743 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/93a6ace4-8d64-44d7-9637-457d3af6543c-host-cni-netd\") pod \"ovnkube-node-txzlk\" (UID: \"93a6ace4-8d64-44d7-9637-457d3af6543c\") " pod="openshift-ovn-kubernetes/ovnkube-node-txzlk" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.695761 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-96h7b\" (UniqueName: \"kubernetes.io/projected/93a6ace4-8d64-44d7-9637-457d3af6543c-kube-api-access-96h7b\") pod \"ovnkube-node-txzlk\" (UID: \"93a6ace4-8d64-44d7-9637-457d3af6543c\") " pod="openshift-ovn-kubernetes/ovnkube-node-txzlk" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.695813 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/93a6ace4-8d64-44d7-9637-457d3af6543c-host-kubelet\") pod \"ovnkube-node-txzlk\" (UID: \"93a6ace4-8d64-44d7-9637-457d3af6543c\") " pod="openshift-ovn-kubernetes/ovnkube-node-txzlk" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.695842 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/93a6ace4-8d64-44d7-9637-457d3af6543c-host-cni-bin\") pod \"ovnkube-node-txzlk\" (UID: \"93a6ace4-8d64-44d7-9637-457d3af6543c\") " pod="openshift-ovn-kubernetes/ovnkube-node-txzlk" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.695861 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/93a6ace4-8d64-44d7-9637-457d3af6543c-ovnkube-script-lib\") pod \"ovnkube-node-txzlk\" (UID: \"93a6ace4-8d64-44d7-9637-457d3af6543c\") " pod="openshift-ovn-kubernetes/ovnkube-node-txzlk" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.695921 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/93a6ace4-8d64-44d7-9637-457d3af6543c-var-lib-openvswitch\") pod \"ovnkube-node-txzlk\" (UID: \"93a6ace4-8d64-44d7-9637-457d3af6543c\") " pod="openshift-ovn-kubernetes/ovnkube-node-txzlk" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.695949 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/93a6ace4-8d64-44d7-9637-457d3af6543c-host-run-ovn-kubernetes\") pod \"ovnkube-node-txzlk\" (UID: \"93a6ace4-8d64-44d7-9637-457d3af6543c\") " pod="openshift-ovn-kubernetes/ovnkube-node-txzlk" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.695964 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/93a6ace4-8d64-44d7-9637-457d3af6543c-env-overrides\") pod \"ovnkube-node-txzlk\" (UID: \"93a6ace4-8d64-44d7-9637-457d3af6543c\") " pod="openshift-ovn-kubernetes/ovnkube-node-txzlk" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.696023 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/93a6ace4-8d64-44d7-9637-457d3af6543c-host-slash\") pod \"ovnkube-node-txzlk\" (UID: \"93a6ace4-8d64-44d7-9637-457d3af6543c\") " pod="openshift-ovn-kubernetes/ovnkube-node-txzlk" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.696044 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/93a6ace4-8d64-44d7-9637-457d3af6543c-etc-openvswitch\") pod \"ovnkube-node-txzlk\" (UID: \"93a6ace4-8d64-44d7-9637-457d3af6543c\") " pod="openshift-ovn-kubernetes/ovnkube-node-txzlk" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.696061 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/93a6ace4-8d64-44d7-9637-457d3af6543c-run-ovn\") pod \"ovnkube-node-txzlk\" (UID: \"93a6ace4-8d64-44d7-9637-457d3af6543c\") " pod="openshift-ovn-kubernetes/ovnkube-node-txzlk" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.696260 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-r7ztb" event={"ID":"a5119cd9-af70-4324-a955-978306ab0b20","Type":"ContainerStarted","Data":"1ceec26233d00d99633fba647e03b7a19f2180836955f4318d5abb1761849623"} Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.707612 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ff10c31459b3b3724ee7b769bd6f72f9ed80cce5419b6adcdf69a3239b70756\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:24Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.720758 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://347f900743b7bba3daf92f21c7eba0625e8c774f50057bc9472f0c82119325e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:24Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.755206 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56f1fe1f9f06f0249cf937a6963fda82467732dc056ca2b2189be6d8dd9d4a05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7212ce5f42c1fb55146abb190a557a504b43f16af6df84f90e1073567a329ec5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:24Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.766721 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.766773 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.766789 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.766808 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.766822 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:24Z","lastTransitionTime":"2025-10-05T06:48:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.781274 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-7wm5q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9bc7744f-d700-450f-8b8d-98140877fee9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h5p4q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-7wm5q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:24Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.797496 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/93a6ace4-8d64-44d7-9637-457d3af6543c-var-lib-openvswitch\") pod \"ovnkube-node-txzlk\" (UID: \"93a6ace4-8d64-44d7-9637-457d3af6543c\") " pod="openshift-ovn-kubernetes/ovnkube-node-txzlk" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.797545 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/93a6ace4-8d64-44d7-9637-457d3af6543c-host-run-ovn-kubernetes\") pod \"ovnkube-node-txzlk\" (UID: \"93a6ace4-8d64-44d7-9637-457d3af6543c\") " pod="openshift-ovn-kubernetes/ovnkube-node-txzlk" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.797569 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/93a6ace4-8d64-44d7-9637-457d3af6543c-run-ovn\") pod \"ovnkube-node-txzlk\" (UID: \"93a6ace4-8d64-44d7-9637-457d3af6543c\") " pod="openshift-ovn-kubernetes/ovnkube-node-txzlk" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.797597 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/93a6ace4-8d64-44d7-9637-457d3af6543c-env-overrides\") pod \"ovnkube-node-txzlk\" (UID: \"93a6ace4-8d64-44d7-9637-457d3af6543c\") " pod="openshift-ovn-kubernetes/ovnkube-node-txzlk" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.797621 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/93a6ace4-8d64-44d7-9637-457d3af6543c-host-slash\") pod \"ovnkube-node-txzlk\" (UID: \"93a6ace4-8d64-44d7-9637-457d3af6543c\") " pod="openshift-ovn-kubernetes/ovnkube-node-txzlk" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.797643 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/93a6ace4-8d64-44d7-9637-457d3af6543c-etc-openvswitch\") pod \"ovnkube-node-txzlk\" (UID: \"93a6ace4-8d64-44d7-9637-457d3af6543c\") " pod="openshift-ovn-kubernetes/ovnkube-node-txzlk" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.797669 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/93a6ace4-8d64-44d7-9637-457d3af6543c-run-systemd\") pod \"ovnkube-node-txzlk\" (UID: \"93a6ace4-8d64-44d7-9637-457d3af6543c\") " pod="openshift-ovn-kubernetes/ovnkube-node-txzlk" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.797690 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/93a6ace4-8d64-44d7-9637-457d3af6543c-run-openvswitch\") pod \"ovnkube-node-txzlk\" (UID: \"93a6ace4-8d64-44d7-9637-457d3af6543c\") " pod="openshift-ovn-kubernetes/ovnkube-node-txzlk" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.797710 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/93a6ace4-8d64-44d7-9637-457d3af6543c-host-run-netns\") pod \"ovnkube-node-txzlk\" (UID: \"93a6ace4-8d64-44d7-9637-457d3af6543c\") " pod="openshift-ovn-kubernetes/ovnkube-node-txzlk" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.797751 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/93a6ace4-8d64-44d7-9637-457d3af6543c-node-log\") pod \"ovnkube-node-txzlk\" (UID: \"93a6ace4-8d64-44d7-9637-457d3af6543c\") " pod="openshift-ovn-kubernetes/ovnkube-node-txzlk" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.797773 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/93a6ace4-8d64-44d7-9637-457d3af6543c-ovn-node-metrics-cert\") pod \"ovnkube-node-txzlk\" (UID: \"93a6ace4-8d64-44d7-9637-457d3af6543c\") " pod="openshift-ovn-kubernetes/ovnkube-node-txzlk" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.797796 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/93a6ace4-8d64-44d7-9637-457d3af6543c-log-socket\") pod \"ovnkube-node-txzlk\" (UID: \"93a6ace4-8d64-44d7-9637-457d3af6543c\") " pod="openshift-ovn-kubernetes/ovnkube-node-txzlk" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.797822 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/93a6ace4-8d64-44d7-9637-457d3af6543c-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-txzlk\" (UID: \"93a6ace4-8d64-44d7-9637-457d3af6543c\") " pod="openshift-ovn-kubernetes/ovnkube-node-txzlk" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.797851 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/93a6ace4-8d64-44d7-9637-457d3af6543c-systemd-units\") pod \"ovnkube-node-txzlk\" (UID: \"93a6ace4-8d64-44d7-9637-457d3af6543c\") " pod="openshift-ovn-kubernetes/ovnkube-node-txzlk" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.797876 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/93a6ace4-8d64-44d7-9637-457d3af6543c-ovnkube-config\") pod \"ovnkube-node-txzlk\" (UID: \"93a6ace4-8d64-44d7-9637-457d3af6543c\") " pod="openshift-ovn-kubernetes/ovnkube-node-txzlk" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.797902 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/93a6ace4-8d64-44d7-9637-457d3af6543c-host-cni-netd\") pod \"ovnkube-node-txzlk\" (UID: \"93a6ace4-8d64-44d7-9637-457d3af6543c\") " pod="openshift-ovn-kubernetes/ovnkube-node-txzlk" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.797929 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-96h7b\" (UniqueName: \"kubernetes.io/projected/93a6ace4-8d64-44d7-9637-457d3af6543c-kube-api-access-96h7b\") pod \"ovnkube-node-txzlk\" (UID: \"93a6ace4-8d64-44d7-9637-457d3af6543c\") " pod="openshift-ovn-kubernetes/ovnkube-node-txzlk" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.797954 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/93a6ace4-8d64-44d7-9637-457d3af6543c-host-kubelet\") pod \"ovnkube-node-txzlk\" (UID: \"93a6ace4-8d64-44d7-9637-457d3af6543c\") " pod="openshift-ovn-kubernetes/ovnkube-node-txzlk" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.797979 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/93a6ace4-8d64-44d7-9637-457d3af6543c-host-cni-bin\") pod \"ovnkube-node-txzlk\" (UID: \"93a6ace4-8d64-44d7-9637-457d3af6543c\") " pod="openshift-ovn-kubernetes/ovnkube-node-txzlk" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.798006 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/93a6ace4-8d64-44d7-9637-457d3af6543c-ovnkube-script-lib\") pod \"ovnkube-node-txzlk\" (UID: \"93a6ace4-8d64-44d7-9637-457d3af6543c\") " pod="openshift-ovn-kubernetes/ovnkube-node-txzlk" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.798816 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/93a6ace4-8d64-44d7-9637-457d3af6543c-ovnkube-script-lib\") pod \"ovnkube-node-txzlk\" (UID: \"93a6ace4-8d64-44d7-9637-457d3af6543c\") " pod="openshift-ovn-kubernetes/ovnkube-node-txzlk" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.798846 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/93a6ace4-8d64-44d7-9637-457d3af6543c-var-lib-openvswitch\") pod \"ovnkube-node-txzlk\" (UID: \"93a6ace4-8d64-44d7-9637-457d3af6543c\") " pod="openshift-ovn-kubernetes/ovnkube-node-txzlk" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.798906 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/93a6ace4-8d64-44d7-9637-457d3af6543c-host-run-ovn-kubernetes\") pod \"ovnkube-node-txzlk\" (UID: \"93a6ace4-8d64-44d7-9637-457d3af6543c\") " pod="openshift-ovn-kubernetes/ovnkube-node-txzlk" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.799271 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/93a6ace4-8d64-44d7-9637-457d3af6543c-env-overrides\") pod \"ovnkube-node-txzlk\" (UID: \"93a6ace4-8d64-44d7-9637-457d3af6543c\") " pod="openshift-ovn-kubernetes/ovnkube-node-txzlk" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.798886 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/93a6ace4-8d64-44d7-9637-457d3af6543c-node-log\") pod \"ovnkube-node-txzlk\" (UID: \"93a6ace4-8d64-44d7-9637-457d3af6543c\") " pod="openshift-ovn-kubernetes/ovnkube-node-txzlk" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.799409 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/93a6ace4-8d64-44d7-9637-457d3af6543c-host-slash\") pod \"ovnkube-node-txzlk\" (UID: \"93a6ace4-8d64-44d7-9637-457d3af6543c\") " pod="openshift-ovn-kubernetes/ovnkube-node-txzlk" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.799425 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/93a6ace4-8d64-44d7-9637-457d3af6543c-etc-openvswitch\") pod \"ovnkube-node-txzlk\" (UID: \"93a6ace4-8d64-44d7-9637-457d3af6543c\") " pod="openshift-ovn-kubernetes/ovnkube-node-txzlk" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.799438 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/93a6ace4-8d64-44d7-9637-457d3af6543c-run-systemd\") pod \"ovnkube-node-txzlk\" (UID: \"93a6ace4-8d64-44d7-9637-457d3af6543c\") " pod="openshift-ovn-kubernetes/ovnkube-node-txzlk" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.799452 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/93a6ace4-8d64-44d7-9637-457d3af6543c-run-openvswitch\") pod \"ovnkube-node-txzlk\" (UID: \"93a6ace4-8d64-44d7-9637-457d3af6543c\") " pod="openshift-ovn-kubernetes/ovnkube-node-txzlk" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.799468 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/93a6ace4-8d64-44d7-9637-457d3af6543c-host-run-netns\") pod \"ovnkube-node-txzlk\" (UID: \"93a6ace4-8d64-44d7-9637-457d3af6543c\") " pod="openshift-ovn-kubernetes/ovnkube-node-txzlk" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.799482 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/93a6ace4-8d64-44d7-9637-457d3af6543c-host-cni-netd\") pod \"ovnkube-node-txzlk\" (UID: \"93a6ace4-8d64-44d7-9637-457d3af6543c\") " pod="openshift-ovn-kubernetes/ovnkube-node-txzlk" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.799494 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/93a6ace4-8d64-44d7-9637-457d3af6543c-log-socket\") pod \"ovnkube-node-txzlk\" (UID: \"93a6ace4-8d64-44d7-9637-457d3af6543c\") " pod="openshift-ovn-kubernetes/ovnkube-node-txzlk" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.799508 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/93a6ace4-8d64-44d7-9637-457d3af6543c-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-txzlk\" (UID: \"93a6ace4-8d64-44d7-9637-457d3af6543c\") " pod="openshift-ovn-kubernetes/ovnkube-node-txzlk" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.799522 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/93a6ace4-8d64-44d7-9637-457d3af6543c-systemd-units\") pod \"ovnkube-node-txzlk\" (UID: \"93a6ace4-8d64-44d7-9637-457d3af6543c\") " pod="openshift-ovn-kubernetes/ovnkube-node-txzlk" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.800002 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/93a6ace4-8d64-44d7-9637-457d3af6543c-ovnkube-config\") pod \"ovnkube-node-txzlk\" (UID: \"93a6ace4-8d64-44d7-9637-457d3af6543c\") " pod="openshift-ovn-kubernetes/ovnkube-node-txzlk" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.800030 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/93a6ace4-8d64-44d7-9637-457d3af6543c-host-cni-bin\") pod \"ovnkube-node-txzlk\" (UID: \"93a6ace4-8d64-44d7-9637-457d3af6543c\") " pod="openshift-ovn-kubernetes/ovnkube-node-txzlk" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.800343 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/93a6ace4-8d64-44d7-9637-457d3af6543c-host-kubelet\") pod \"ovnkube-node-txzlk\" (UID: \"93a6ace4-8d64-44d7-9637-457d3af6543c\") " pod="openshift-ovn-kubernetes/ovnkube-node-txzlk" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.800393 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/93a6ace4-8d64-44d7-9637-457d3af6543c-run-ovn\") pod \"ovnkube-node-txzlk\" (UID: \"93a6ace4-8d64-44d7-9637-457d3af6543c\") " pod="openshift-ovn-kubernetes/ovnkube-node-txzlk" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.803874 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/93a6ace4-8d64-44d7-9637-457d3af6543c-ovn-node-metrics-cert\") pod \"ovnkube-node-txzlk\" (UID: \"93a6ace4-8d64-44d7-9637-457d3af6543c\") " pod="openshift-ovn-kubernetes/ovnkube-node-txzlk" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.836000 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"775b5702-a131-4d6c-852d-91bfa056cac5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0839c1765df2c68d2554e9c66d289505e559a9672caf8069feeff2c4fdfd2171\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15dbaf2445704da0987f256f81dff2ee95664988541fa48ce94898afeadab2e8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd3b3ab323c6d5153054aa04af2d91f7b9e8436ff1e2f75f3594b57831b2a416\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c52e076eddfd307110b1f6de00a169b81af8f2785af7becef2a820ce363a5861\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c52e076eddfd307110b1f6de00a169b81af8f2785af7becef2a820ce363a5861\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-05T06:48:14Z\\\",\\\"message\\\":\\\"le observer\\\\nW1005 06:48:14.337870 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1005 06:48:14.338086 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1005 06:48:14.339389 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1467752215/tls.crt::/tmp/serving-cert-1467752215/tls.key\\\\\\\"\\\\nI1005 06:48:14.864345 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1005 06:48:14.868087 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1005 06:48:14.868114 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1005 06:48:14.868140 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1005 06:48:14.868147 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1005 06:48:14.877354 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1005 06:48:14.877412 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1005 06:48:14.877456 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1005 06:48:14.877471 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1005 06:48:14.877490 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1005 06:48:14.877501 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1005 06:48:14.877507 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1005 06:48:14.877515 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1005 06:48:14.880326 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:13Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4a091716ff935adb42ae2e88048619fd347bd174f64e5f9c97a01f81690e57f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:01Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dc5411b642f99b13d514d2f21defad0c394aeeb82a966f72300076617e2586c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5dc5411b642f99b13d514d2f21defad0c394aeeb82a966f72300076617e2586c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:47:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:47:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:47:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:24Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.847867 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-96h7b\" (UniqueName: \"kubernetes.io/projected/93a6ace4-8d64-44d7-9637-457d3af6543c-kube-api-access-96h7b\") pod \"ovnkube-node-txzlk\" (UID: \"93a6ace4-8d64-44d7-9637-457d3af6543c\") " pod="openshift-ovn-kubernetes/ovnkube-node-txzlk" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.860026 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:24Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.877462 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.877519 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.877528 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.877548 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.877557 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:24Z","lastTransitionTime":"2025-10-05T06:48:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.890495 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-r7ztb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a5119cd9-af70-4324-a955-978306ab0b20\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jfcj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-r7ztb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:24Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.913447 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:24Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.946754 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-sfmxd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1dbb13ac-136c-4a1c-9229-b4e8ac4f9c71\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-sfmxd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:24Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.968655 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-txzlk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"93a6ace4-8d64-44d7-9637-457d3af6543c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:24Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-txzlk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:24Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.979797 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.979842 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.979852 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.979867 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.979878 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:24Z","lastTransitionTime":"2025-10-05T06:48:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:24 crc kubenswrapper[4846]: I1005 06:48:24.988794 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c482bb97-fd50-40cf-96ea-f806c3f01125\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bfe45d30e4fa141541aca1d46bc5cc1da0a7b4a5eb1fc9bb2caa5b7fdcb0b264\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://496a4782faa30a9a20ff63038141600cd635f12902d1cf17989cd067fcdfe929\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://281be077869a6aa1ca815770871a14413d4a97e97aeb550fd5cd215cba88b21f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6771c7215d2f5dd5838d17888375a4768b9d1755fc2e21bda1bbfbb2eb1dbb95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c17a711c2c7473b076a3944ffe7c700808d05cba582fbd55c5c2ac8ff73838a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44a9f740f0a06c8bf928e2a9d1d5979217558c31e73a333631baccebf00fd5d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://44a9f740f0a06c8bf928e2a9d1d5979217558c31e73a333631baccebf00fd5d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:47:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:47:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://48ca0ba6ca6092bfcb5b153220c054f02cc36fb22707c42e021f06d329e18e2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48ca0ba6ca6092bfcb5b153220c054f02cc36fb22707c42e021f06d329e18e2e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:00Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4006d0cea128675c8885d879171204abffc5e11ea0664750a12b4bc7344ccb2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4006d0cea128675c8885d879171204abffc5e11ea0664750a12b4bc7344ccb2e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:47:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:24Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:25 crc kubenswrapper[4846]: I1005 06:48:25.020548 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-txzlk" Oct 05 06:48:25 crc kubenswrapper[4846]: W1005 06:48:25.032016 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod93a6ace4_8d64_44d7_9637_457d3af6543c.slice/crio-96612529b2d28247403966c25096e92a88b39c2bef12601d90018451dbf8a6e7 WatchSource:0}: Error finding container 96612529b2d28247403966c25096e92a88b39c2bef12601d90018451dbf8a6e7: Status 404 returned error can't find the container with id 96612529b2d28247403966c25096e92a88b39c2bef12601d90018451dbf8a6e7 Oct 05 06:48:25 crc kubenswrapper[4846]: I1005 06:48:25.082013 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:25 crc kubenswrapper[4846]: I1005 06:48:25.082050 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:25 crc kubenswrapper[4846]: I1005 06:48:25.082072 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:25 crc kubenswrapper[4846]: I1005 06:48:25.082087 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:25 crc kubenswrapper[4846]: I1005 06:48:25.082097 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:25Z","lastTransitionTime":"2025-10-05T06:48:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:25 crc kubenswrapper[4846]: I1005 06:48:25.184577 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:25 crc kubenswrapper[4846]: I1005 06:48:25.184627 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:25 crc kubenswrapper[4846]: I1005 06:48:25.184637 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:25 crc kubenswrapper[4846]: I1005 06:48:25.184655 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:25 crc kubenswrapper[4846]: I1005 06:48:25.184668 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:25Z","lastTransitionTime":"2025-10-05T06:48:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:25 crc kubenswrapper[4846]: I1005 06:48:25.287451 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:25 crc kubenswrapper[4846]: I1005 06:48:25.287483 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:25 crc kubenswrapper[4846]: I1005 06:48:25.287494 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:25 crc kubenswrapper[4846]: I1005 06:48:25.287510 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:25 crc kubenswrapper[4846]: I1005 06:48:25.287520 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:25Z","lastTransitionTime":"2025-10-05T06:48:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:25 crc kubenswrapper[4846]: I1005 06:48:25.390229 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:25 crc kubenswrapper[4846]: I1005 06:48:25.390279 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:25 crc kubenswrapper[4846]: I1005 06:48:25.390288 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:25 crc kubenswrapper[4846]: I1005 06:48:25.390307 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:25 crc kubenswrapper[4846]: I1005 06:48:25.390318 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:25Z","lastTransitionTime":"2025-10-05T06:48:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:25 crc kubenswrapper[4846]: I1005 06:48:25.403809 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 05 06:48:25 crc kubenswrapper[4846]: I1005 06:48:25.403906 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 05 06:48:25 crc kubenswrapper[4846]: E1005 06:48:25.403943 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-05 06:48:33.403921912 +0000 UTC m=+35.644774687 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 06:48:25 crc kubenswrapper[4846]: I1005 06:48:25.403967 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 05 06:48:25 crc kubenswrapper[4846]: I1005 06:48:25.404006 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 05 06:48:25 crc kubenswrapper[4846]: I1005 06:48:25.404028 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 05 06:48:25 crc kubenswrapper[4846]: E1005 06:48:25.404012 4846 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 05 06:48:25 crc kubenswrapper[4846]: E1005 06:48:25.404085 4846 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 05 06:48:25 crc kubenswrapper[4846]: E1005 06:48:25.404072 4846 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 05 06:48:25 crc kubenswrapper[4846]: E1005 06:48:25.404132 4846 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 05 06:48:25 crc kubenswrapper[4846]: E1005 06:48:25.404144 4846 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 05 06:48:25 crc kubenswrapper[4846]: E1005 06:48:25.404119 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-05 06:48:33.404111038 +0000 UTC m=+35.644963813 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 05 06:48:25 crc kubenswrapper[4846]: E1005 06:48:25.404190 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-05 06:48:33.404166509 +0000 UTC m=+35.645019284 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 05 06:48:25 crc kubenswrapper[4846]: E1005 06:48:25.404200 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-05 06:48:33.40419511 +0000 UTC m=+35.645047885 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 05 06:48:25 crc kubenswrapper[4846]: E1005 06:48:25.404209 4846 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 05 06:48:25 crc kubenswrapper[4846]: E1005 06:48:25.404234 4846 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 05 06:48:25 crc kubenswrapper[4846]: E1005 06:48:25.404246 4846 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 05 06:48:25 crc kubenswrapper[4846]: E1005 06:48:25.404304 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-05 06:48:33.404287473 +0000 UTC m=+35.645140248 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 05 06:48:25 crc kubenswrapper[4846]: I1005 06:48:25.492812 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:25 crc kubenswrapper[4846]: I1005 06:48:25.492879 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:25 crc kubenswrapper[4846]: I1005 06:48:25.492897 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:25 crc kubenswrapper[4846]: I1005 06:48:25.492917 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:25 crc kubenswrapper[4846]: I1005 06:48:25.492933 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:25Z","lastTransitionTime":"2025-10-05T06:48:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:25 crc kubenswrapper[4846]: I1005 06:48:25.497374 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 05 06:48:25 crc kubenswrapper[4846]: I1005 06:48:25.497425 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 05 06:48:25 crc kubenswrapper[4846]: E1005 06:48:25.497518 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 05 06:48:25 crc kubenswrapper[4846]: E1005 06:48:25.497637 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 05 06:48:25 crc kubenswrapper[4846]: I1005 06:48:25.595693 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:25 crc kubenswrapper[4846]: I1005 06:48:25.595770 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:25 crc kubenswrapper[4846]: I1005 06:48:25.595790 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:25 crc kubenswrapper[4846]: I1005 06:48:25.595816 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:25 crc kubenswrapper[4846]: I1005 06:48:25.595835 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:25Z","lastTransitionTime":"2025-10-05T06:48:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:25 crc kubenswrapper[4846]: I1005 06:48:25.698859 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:25 crc kubenswrapper[4846]: I1005 06:48:25.698961 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:25 crc kubenswrapper[4846]: I1005 06:48:25.698982 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:25 crc kubenswrapper[4846]: I1005 06:48:25.699009 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:25 crc kubenswrapper[4846]: I1005 06:48:25.699031 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:25Z","lastTransitionTime":"2025-10-05T06:48:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:25 crc kubenswrapper[4846]: I1005 06:48:25.702439 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-7wm5q" event={"ID":"9bc7744f-d700-450f-8b8d-98140877fee9","Type":"ContainerStarted","Data":"90379e74c544fccca44f7acdc228ff378154390e211ac801b79f8d074012241e"} Oct 05 06:48:25 crc kubenswrapper[4846]: I1005 06:48:25.708263 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-r7ztb" event={"ID":"a5119cd9-af70-4324-a955-978306ab0b20","Type":"ContainerStarted","Data":"06231a87429ec292cb57ebd88cda7ad547bcb32e80343b776833621a11c28ad1"} Oct 05 06:48:25 crc kubenswrapper[4846]: I1005 06:48:25.710848 4846 generic.go:334] "Generic (PLEG): container finished" podID="1dbb13ac-136c-4a1c-9229-b4e8ac4f9c71" containerID="226d977b596d07aeda0486452e2a8605949ed4d2453b21f56174937c395d6f36" exitCode=0 Oct 05 06:48:25 crc kubenswrapper[4846]: I1005 06:48:25.710970 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-sfmxd" event={"ID":"1dbb13ac-136c-4a1c-9229-b4e8ac4f9c71","Type":"ContainerDied","Data":"226d977b596d07aeda0486452e2a8605949ed4d2453b21f56174937c395d6f36"} Oct 05 06:48:25 crc kubenswrapper[4846]: I1005 06:48:25.713215 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" event={"ID":"ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7","Type":"ContainerStarted","Data":"2e67d0e447a8a6a381259e25d25badc791194e1291f56b1cd77bd63989b20408"} Oct 05 06:48:25 crc kubenswrapper[4846]: I1005 06:48:25.713240 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" event={"ID":"ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7","Type":"ContainerStarted","Data":"516f88f4990f5962d8b694a3d889259cc5508a309919870191c2e1331640c973"} Oct 05 06:48:25 crc kubenswrapper[4846]: I1005 06:48:25.715283 4846 generic.go:334] "Generic (PLEG): container finished" podID="93a6ace4-8d64-44d7-9637-457d3af6543c" containerID="ff3c6545ff3d2436da27bbd223b950b30b0c69a3e7f3034cb0e61380cb38d338" exitCode=0 Oct 05 06:48:25 crc kubenswrapper[4846]: I1005 06:48:25.715314 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-txzlk" event={"ID":"93a6ace4-8d64-44d7-9637-457d3af6543c","Type":"ContainerDied","Data":"ff3c6545ff3d2436da27bbd223b950b30b0c69a3e7f3034cb0e61380cb38d338"} Oct 05 06:48:25 crc kubenswrapper[4846]: I1005 06:48:25.715332 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-txzlk" event={"ID":"93a6ace4-8d64-44d7-9637-457d3af6543c","Type":"ContainerStarted","Data":"96612529b2d28247403966c25096e92a88b39c2bef12601d90018451dbf8a6e7"} Oct 05 06:48:25 crc kubenswrapper[4846]: I1005 06:48:25.720147 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ff10c31459b3b3724ee7b769bd6f72f9ed80cce5419b6adcdf69a3239b70756\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:25Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:25 crc kubenswrapper[4846]: I1005 06:48:25.737240 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://347f900743b7bba3daf92f21c7eba0625e8c774f50057bc9472f0c82119325e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:25Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:25 crc kubenswrapper[4846]: I1005 06:48:25.759745 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56f1fe1f9f06f0249cf937a6963fda82467732dc056ca2b2189be6d8dd9d4a05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7212ce5f42c1fb55146abb190a557a504b43f16af6df84f90e1073567a329ec5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:25Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:25 crc kubenswrapper[4846]: I1005 06:48:25.777634 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-7wm5q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9bc7744f-d700-450f-8b8d-98140877fee9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90379e74c544fccca44f7acdc228ff378154390e211ac801b79f8d074012241e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h5p4q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-7wm5q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:25Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:25 crc kubenswrapper[4846]: I1005 06:48:25.798996 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"775b5702-a131-4d6c-852d-91bfa056cac5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0839c1765df2c68d2554e9c66d289505e559a9672caf8069feeff2c4fdfd2171\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15dbaf2445704da0987f256f81dff2ee95664988541fa48ce94898afeadab2e8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd3b3ab323c6d5153054aa04af2d91f7b9e8436ff1e2f75f3594b57831b2a416\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c52e076eddfd307110b1f6de00a169b81af8f2785af7becef2a820ce363a5861\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c52e076eddfd307110b1f6de00a169b81af8f2785af7becef2a820ce363a5861\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-05T06:48:14Z\\\",\\\"message\\\":\\\"le observer\\\\nW1005 06:48:14.337870 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1005 06:48:14.338086 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1005 06:48:14.339389 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1467752215/tls.crt::/tmp/serving-cert-1467752215/tls.key\\\\\\\"\\\\nI1005 06:48:14.864345 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1005 06:48:14.868087 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1005 06:48:14.868114 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1005 06:48:14.868140 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1005 06:48:14.868147 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1005 06:48:14.877354 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1005 06:48:14.877412 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1005 06:48:14.877456 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1005 06:48:14.877471 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1005 06:48:14.877490 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1005 06:48:14.877501 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1005 06:48:14.877507 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1005 06:48:14.877515 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1005 06:48:14.880326 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:13Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4a091716ff935adb42ae2e88048619fd347bd174f64e5f9c97a01f81690e57f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:01Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dc5411b642f99b13d514d2f21defad0c394aeeb82a966f72300076617e2586c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5dc5411b642f99b13d514d2f21defad0c394aeeb82a966f72300076617e2586c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:47:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:47:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:47:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:25Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:25 crc kubenswrapper[4846]: I1005 06:48:25.802576 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:25 crc kubenswrapper[4846]: I1005 06:48:25.802646 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:25 crc kubenswrapper[4846]: I1005 06:48:25.802662 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:25 crc kubenswrapper[4846]: I1005 06:48:25.802682 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:25 crc kubenswrapper[4846]: I1005 06:48:25.802719 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:25Z","lastTransitionTime":"2025-10-05T06:48:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:25 crc kubenswrapper[4846]: I1005 06:48:25.814428 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:25Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:25 crc kubenswrapper[4846]: I1005 06:48:25.827253 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-r7ztb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a5119cd9-af70-4324-a955-978306ab0b20\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jfcj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-r7ztb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:25Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:25 crc kubenswrapper[4846]: I1005 06:48:25.842667 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:25Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:25 crc kubenswrapper[4846]: I1005 06:48:25.859953 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-sfmxd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1dbb13ac-136c-4a1c-9229-b4e8ac4f9c71\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-sfmxd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:25Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:25 crc kubenswrapper[4846]: I1005 06:48:25.880438 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-txzlk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"93a6ace4-8d64-44d7-9637-457d3af6543c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:24Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-txzlk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:25Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:25 crc kubenswrapper[4846]: I1005 06:48:25.899944 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c482bb97-fd50-40cf-96ea-f806c3f01125\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bfe45d30e4fa141541aca1d46bc5cc1da0a7b4a5eb1fc9bb2caa5b7fdcb0b264\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://496a4782faa30a9a20ff63038141600cd635f12902d1cf17989cd067fcdfe929\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://281be077869a6aa1ca815770871a14413d4a97e97aeb550fd5cd215cba88b21f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6771c7215d2f5dd5838d17888375a4768b9d1755fc2e21bda1bbfbb2eb1dbb95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c17a711c2c7473b076a3944ffe7c700808d05cba582fbd55c5c2ac8ff73838a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44a9f740f0a06c8bf928e2a9d1d5979217558c31e73a333631baccebf00fd5d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://44a9f740f0a06c8bf928e2a9d1d5979217558c31e73a333631baccebf00fd5d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:47:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:47:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://48ca0ba6ca6092bfcb5b153220c054f02cc36fb22707c42e021f06d329e18e2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48ca0ba6ca6092bfcb5b153220c054f02cc36fb22707c42e021f06d329e18e2e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:00Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4006d0cea128675c8885d879171204abffc5e11ea0664750a12b4bc7344ccb2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4006d0cea128675c8885d879171204abffc5e11ea0664750a12b4bc7344ccb2e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:47:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:25Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:25 crc kubenswrapper[4846]: I1005 06:48:25.907162 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:25 crc kubenswrapper[4846]: I1005 06:48:25.907207 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:25 crc kubenswrapper[4846]: I1005 06:48:25.907217 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:25 crc kubenswrapper[4846]: I1005 06:48:25.907233 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:25 crc kubenswrapper[4846]: I1005 06:48:25.907245 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:25Z","lastTransitionTime":"2025-10-05T06:48:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:25 crc kubenswrapper[4846]: I1005 06:48:25.917420 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:25Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:25 crc kubenswrapper[4846]: I1005 06:48:25.932548 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dkhl7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dkhl7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-fscvf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:25Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:25 crc kubenswrapper[4846]: I1005 06:48:25.946104 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c9f812b-ecc8-4985-a670-53fa80fbb499\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2b5ee9781c01d6aec3ac778efef1067b6c3bfe4ed37431bdebe188fcde0165f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:47:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3fb4717692aa1fec0012b514ef8d18d1ca3707cfa4bb8248bb93841a0a40f11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:47:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0b6a7afe3094f1c92093c05c28344897a615d590c2fbb507c87e18a44c6c831\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://640ade2c733d7916ec213a201a1dd614d261c69eab654c80100efbc99a768995\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:47:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:25Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:25 crc kubenswrapper[4846]: I1005 06:48:25.959868 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c9f812b-ecc8-4985-a670-53fa80fbb499\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2b5ee9781c01d6aec3ac778efef1067b6c3bfe4ed37431bdebe188fcde0165f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:47:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3fb4717692aa1fec0012b514ef8d18d1ca3707cfa4bb8248bb93841a0a40f11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:47:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0b6a7afe3094f1c92093c05c28344897a615d590c2fbb507c87e18a44c6c831\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://640ade2c733d7916ec213a201a1dd614d261c69eab654c80100efbc99a768995\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:47:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:25Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:25 crc kubenswrapper[4846]: I1005 06:48:25.971467 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:25Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:25 crc kubenswrapper[4846]: I1005 06:48:25.982298 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e67d0e447a8a6a381259e25d25badc791194e1291f56b1cd77bd63989b20408\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dkhl7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://516f88f4990f5962d8b694a3d889259cc5508a309919870191c2e1331640c973\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dkhl7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-fscvf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:25Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:25 crc kubenswrapper[4846]: I1005 06:48:25.998539 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ff10c31459b3b3724ee7b769bd6f72f9ed80cce5419b6adcdf69a3239b70756\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:25Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:26 crc kubenswrapper[4846]: I1005 06:48:26.010918 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:26 crc kubenswrapper[4846]: I1005 06:48:26.010972 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:26 crc kubenswrapper[4846]: I1005 06:48:26.010985 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:26 crc kubenswrapper[4846]: I1005 06:48:26.011005 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:26 crc kubenswrapper[4846]: I1005 06:48:26.011018 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:26Z","lastTransitionTime":"2025-10-05T06:48:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:26 crc kubenswrapper[4846]: I1005 06:48:26.011629 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://347f900743b7bba3daf92f21c7eba0625e8c774f50057bc9472f0c82119325e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:26Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:26 crc kubenswrapper[4846]: I1005 06:48:26.027355 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56f1fe1f9f06f0249cf937a6963fda82467732dc056ca2b2189be6d8dd9d4a05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7212ce5f42c1fb55146abb190a557a504b43f16af6df84f90e1073567a329ec5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:26Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:26 crc kubenswrapper[4846]: I1005 06:48:26.039858 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-7wm5q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9bc7744f-d700-450f-8b8d-98140877fee9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90379e74c544fccca44f7acdc228ff378154390e211ac801b79f8d074012241e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h5p4q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-7wm5q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:26Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:26 crc kubenswrapper[4846]: I1005 06:48:26.053880 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"775b5702-a131-4d6c-852d-91bfa056cac5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0839c1765df2c68d2554e9c66d289505e559a9672caf8069feeff2c4fdfd2171\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15dbaf2445704da0987f256f81dff2ee95664988541fa48ce94898afeadab2e8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd3b3ab323c6d5153054aa04af2d91f7b9e8436ff1e2f75f3594b57831b2a416\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c52e076eddfd307110b1f6de00a169b81af8f2785af7becef2a820ce363a5861\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c52e076eddfd307110b1f6de00a169b81af8f2785af7becef2a820ce363a5861\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-05T06:48:14Z\\\",\\\"message\\\":\\\"le observer\\\\nW1005 06:48:14.337870 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1005 06:48:14.338086 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1005 06:48:14.339389 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1467752215/tls.crt::/tmp/serving-cert-1467752215/tls.key\\\\\\\"\\\\nI1005 06:48:14.864345 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1005 06:48:14.868087 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1005 06:48:14.868114 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1005 06:48:14.868140 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1005 06:48:14.868147 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1005 06:48:14.877354 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1005 06:48:14.877412 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1005 06:48:14.877456 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1005 06:48:14.877471 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1005 06:48:14.877490 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1005 06:48:14.877501 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1005 06:48:14.877507 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1005 06:48:14.877515 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1005 06:48:14.880326 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:13Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4a091716ff935adb42ae2e88048619fd347bd174f64e5f9c97a01f81690e57f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:01Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dc5411b642f99b13d514d2f21defad0c394aeeb82a966f72300076617e2586c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5dc5411b642f99b13d514d2f21defad0c394aeeb82a966f72300076617e2586c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:47:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:47:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:47:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:26Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:26 crc kubenswrapper[4846]: I1005 06:48:26.067900 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:26Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:26 crc kubenswrapper[4846]: I1005 06:48:26.079412 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-r7ztb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a5119cd9-af70-4324-a955-978306ab0b20\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06231a87429ec292cb57ebd88cda7ad547bcb32e80343b776833621a11c28ad1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jfcj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-r7ztb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:26Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:26 crc kubenswrapper[4846]: I1005 06:48:26.107725 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c482bb97-fd50-40cf-96ea-f806c3f01125\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bfe45d30e4fa141541aca1d46bc5cc1da0a7b4a5eb1fc9bb2caa5b7fdcb0b264\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://496a4782faa30a9a20ff63038141600cd635f12902d1cf17989cd067fcdfe929\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://281be077869a6aa1ca815770871a14413d4a97e97aeb550fd5cd215cba88b21f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6771c7215d2f5dd5838d17888375a4768b9d1755fc2e21bda1bbfbb2eb1dbb95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c17a711c2c7473b076a3944ffe7c700808d05cba582fbd55c5c2ac8ff73838a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44a9f740f0a06c8bf928e2a9d1d5979217558c31e73a333631baccebf00fd5d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://44a9f740f0a06c8bf928e2a9d1d5979217558c31e73a333631baccebf00fd5d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:47:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:47:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://48ca0ba6ca6092bfcb5b153220c054f02cc36fb22707c42e021f06d329e18e2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48ca0ba6ca6092bfcb5b153220c054f02cc36fb22707c42e021f06d329e18e2e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:00Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4006d0cea128675c8885d879171204abffc5e11ea0664750a12b4bc7344ccb2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4006d0cea128675c8885d879171204abffc5e11ea0664750a12b4bc7344ccb2e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:47:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:26Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:26 crc kubenswrapper[4846]: I1005 06:48:26.114495 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:26 crc kubenswrapper[4846]: I1005 06:48:26.114536 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:26 crc kubenswrapper[4846]: I1005 06:48:26.114548 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:26 crc kubenswrapper[4846]: I1005 06:48:26.114569 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:26 crc kubenswrapper[4846]: I1005 06:48:26.114583 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:26Z","lastTransitionTime":"2025-10-05T06:48:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:26 crc kubenswrapper[4846]: I1005 06:48:26.126600 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:26Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:26 crc kubenswrapper[4846]: I1005 06:48:26.141278 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-sfmxd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1dbb13ac-136c-4a1c-9229-b4e8ac4f9c71\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://226d977b596d07aeda0486452e2a8605949ed4d2453b21f56174937c395d6f36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://226d977b596d07aeda0486452e2a8605949ed4d2453b21f56174937c395d6f36\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-sfmxd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:26Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:26 crc kubenswrapper[4846]: I1005 06:48:26.160743 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-txzlk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"93a6ace4-8d64-44d7-9637-457d3af6543c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff3c6545ff3d2436da27bbd223b950b30b0c69a3e7f3034cb0e61380cb38d338\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ff3c6545ff3d2436da27bbd223b950b30b0c69a3e7f3034cb0e61380cb38d338\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:24Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-txzlk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:26Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:26 crc kubenswrapper[4846]: I1005 06:48:26.217340 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:26 crc kubenswrapper[4846]: I1005 06:48:26.217723 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:26 crc kubenswrapper[4846]: I1005 06:48:26.217737 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:26 crc kubenswrapper[4846]: I1005 06:48:26.217759 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:26 crc kubenswrapper[4846]: I1005 06:48:26.217776 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:26Z","lastTransitionTime":"2025-10-05T06:48:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:26 crc kubenswrapper[4846]: I1005 06:48:26.320960 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:26 crc kubenswrapper[4846]: I1005 06:48:26.321021 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:26 crc kubenswrapper[4846]: I1005 06:48:26.321035 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:26 crc kubenswrapper[4846]: I1005 06:48:26.321056 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:26 crc kubenswrapper[4846]: I1005 06:48:26.321071 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:26Z","lastTransitionTime":"2025-10-05T06:48:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:26 crc kubenswrapper[4846]: I1005 06:48:26.424346 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:26 crc kubenswrapper[4846]: I1005 06:48:26.424384 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:26 crc kubenswrapper[4846]: I1005 06:48:26.424394 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:26 crc kubenswrapper[4846]: I1005 06:48:26.424411 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:26 crc kubenswrapper[4846]: I1005 06:48:26.424424 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:26Z","lastTransitionTime":"2025-10-05T06:48:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:26 crc kubenswrapper[4846]: I1005 06:48:26.497307 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 05 06:48:26 crc kubenswrapper[4846]: E1005 06:48:26.497473 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 05 06:48:26 crc kubenswrapper[4846]: I1005 06:48:26.527217 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:26 crc kubenswrapper[4846]: I1005 06:48:26.527263 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:26 crc kubenswrapper[4846]: I1005 06:48:26.527280 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:26 crc kubenswrapper[4846]: I1005 06:48:26.527298 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:26 crc kubenswrapper[4846]: I1005 06:48:26.527310 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:26Z","lastTransitionTime":"2025-10-05T06:48:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:26 crc kubenswrapper[4846]: I1005 06:48:26.630584 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:26 crc kubenswrapper[4846]: I1005 06:48:26.630927 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:26 crc kubenswrapper[4846]: I1005 06:48:26.630937 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:26 crc kubenswrapper[4846]: I1005 06:48:26.630953 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:26 crc kubenswrapper[4846]: I1005 06:48:26.630962 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:26Z","lastTransitionTime":"2025-10-05T06:48:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:26 crc kubenswrapper[4846]: I1005 06:48:26.727019 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-sfmxd" event={"ID":"1dbb13ac-136c-4a1c-9229-b4e8ac4f9c71","Type":"ContainerStarted","Data":"99c07a4bdb3059d3a99ffdaf794d668d52cffd9e5e3bb8945204dd6d1f8c8328"} Oct 05 06:48:26 crc kubenswrapper[4846]: I1005 06:48:26.732238 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-txzlk" event={"ID":"93a6ace4-8d64-44d7-9637-457d3af6543c","Type":"ContainerStarted","Data":"af5e5a4da9ad7371d5a24633e8268f12e3a7d419a8c93b11ddbcc006d8339c6c"} Oct 05 06:48:26 crc kubenswrapper[4846]: I1005 06:48:26.732313 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-txzlk" event={"ID":"93a6ace4-8d64-44d7-9637-457d3af6543c","Type":"ContainerStarted","Data":"0a6f6b2c32dd3da5f9fd5e6971c1773e81a9a24184b7aa5b543cec879001ae53"} Oct 05 06:48:26 crc kubenswrapper[4846]: I1005 06:48:26.732328 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-txzlk" event={"ID":"93a6ace4-8d64-44d7-9637-457d3af6543c","Type":"ContainerStarted","Data":"1ad1940d07e37217089642478ece9ccdaca74784738a400929a9f5359d158c01"} Oct 05 06:48:26 crc kubenswrapper[4846]: I1005 06:48:26.732341 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-txzlk" event={"ID":"93a6ace4-8d64-44d7-9637-457d3af6543c","Type":"ContainerStarted","Data":"8da85222775abc7a772f6a8ff8255f1a0d5b15807ba177a7e3e2b2d01a4f24b9"} Oct 05 06:48:26 crc kubenswrapper[4846]: I1005 06:48:26.732354 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-txzlk" event={"ID":"93a6ace4-8d64-44d7-9637-457d3af6543c","Type":"ContainerStarted","Data":"1409651de8c897804f87be03d016c138f8acf27e662dce265a93e15c1848adcf"} Oct 05 06:48:26 crc kubenswrapper[4846]: I1005 06:48:26.733004 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:26 crc kubenswrapper[4846]: I1005 06:48:26.733039 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:26 crc kubenswrapper[4846]: I1005 06:48:26.733049 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:26 crc kubenswrapper[4846]: I1005 06:48:26.733063 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:26 crc kubenswrapper[4846]: I1005 06:48:26.733074 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:26Z","lastTransitionTime":"2025-10-05T06:48:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:26 crc kubenswrapper[4846]: I1005 06:48:26.809391 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c482bb97-fd50-40cf-96ea-f806c3f01125\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bfe45d30e4fa141541aca1d46bc5cc1da0a7b4a5eb1fc9bb2caa5b7fdcb0b264\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://496a4782faa30a9a20ff63038141600cd635f12902d1cf17989cd067fcdfe929\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://281be077869a6aa1ca815770871a14413d4a97e97aeb550fd5cd215cba88b21f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6771c7215d2f5dd5838d17888375a4768b9d1755fc2e21bda1bbfbb2eb1dbb95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c17a711c2c7473b076a3944ffe7c700808d05cba582fbd55c5c2ac8ff73838a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44a9f740f0a06c8bf928e2a9d1d5979217558c31e73a333631baccebf00fd5d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://44a9f740f0a06c8bf928e2a9d1d5979217558c31e73a333631baccebf00fd5d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:47:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:47:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://48ca0ba6ca6092bfcb5b153220c054f02cc36fb22707c42e021f06d329e18e2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48ca0ba6ca6092bfcb5b153220c054f02cc36fb22707c42e021f06d329e18e2e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:00Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4006d0cea128675c8885d879171204abffc5e11ea0664750a12b4bc7344ccb2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4006d0cea128675c8885d879171204abffc5e11ea0664750a12b4bc7344ccb2e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:47:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:26Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:26 crc kubenswrapper[4846]: I1005 06:48:26.823733 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:26Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:26 crc kubenswrapper[4846]: I1005 06:48:26.836267 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:26 crc kubenswrapper[4846]: I1005 06:48:26.836309 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:26 crc kubenswrapper[4846]: I1005 06:48:26.836320 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:26 crc kubenswrapper[4846]: I1005 06:48:26.836338 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:26 crc kubenswrapper[4846]: I1005 06:48:26.836349 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:26Z","lastTransitionTime":"2025-10-05T06:48:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:26 crc kubenswrapper[4846]: I1005 06:48:26.842153 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-sfmxd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1dbb13ac-136c-4a1c-9229-b4e8ac4f9c71\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://226d977b596d07aeda0486452e2a8605949ed4d2453b21f56174937c395d6f36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://226d977b596d07aeda0486452e2a8605949ed4d2453b21f56174937c395d6f36\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://99c07a4bdb3059d3a99ffdaf794d668d52cffd9e5e3bb8945204dd6d1f8c8328\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-sfmxd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:26Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:26 crc kubenswrapper[4846]: I1005 06:48:26.865096 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-txzlk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"93a6ace4-8d64-44d7-9637-457d3af6543c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff3c6545ff3d2436da27bbd223b950b30b0c69a3e7f3034cb0e61380cb38d338\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ff3c6545ff3d2436da27bbd223b950b30b0c69a3e7f3034cb0e61380cb38d338\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:24Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-txzlk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:26Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:26 crc kubenswrapper[4846]: I1005 06:48:26.878835 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c9f812b-ecc8-4985-a670-53fa80fbb499\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2b5ee9781c01d6aec3ac778efef1067b6c3bfe4ed37431bdebe188fcde0165f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:47:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3fb4717692aa1fec0012b514ef8d18d1ca3707cfa4bb8248bb93841a0a40f11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:47:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0b6a7afe3094f1c92093c05c28344897a615d590c2fbb507c87e18a44c6c831\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://640ade2c733d7916ec213a201a1dd614d261c69eab654c80100efbc99a768995\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:47:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:26Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:26 crc kubenswrapper[4846]: I1005 06:48:26.922672 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:26Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:26 crc kubenswrapper[4846]: I1005 06:48:26.938840 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:26 crc kubenswrapper[4846]: I1005 06:48:26.938873 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:26 crc kubenswrapper[4846]: I1005 06:48:26.938888 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:26 crc kubenswrapper[4846]: I1005 06:48:26.938902 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:26 crc kubenswrapper[4846]: I1005 06:48:26.938912 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:26Z","lastTransitionTime":"2025-10-05T06:48:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:26 crc kubenswrapper[4846]: I1005 06:48:26.946402 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e67d0e447a8a6a381259e25d25badc791194e1291f56b1cd77bd63989b20408\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dkhl7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://516f88f4990f5962d8b694a3d889259cc5508a309919870191c2e1331640c973\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dkhl7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-fscvf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:26Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:26 crc kubenswrapper[4846]: I1005 06:48:26.965906 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-7wm5q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9bc7744f-d700-450f-8b8d-98140877fee9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90379e74c544fccca44f7acdc228ff378154390e211ac801b79f8d074012241e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h5p4q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-7wm5q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:26Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:26 crc kubenswrapper[4846]: I1005 06:48:26.982041 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ff10c31459b3b3724ee7b769bd6f72f9ed80cce5419b6adcdf69a3239b70756\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:26Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:26 crc kubenswrapper[4846]: I1005 06:48:26.993210 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://347f900743b7bba3daf92f21c7eba0625e8c774f50057bc9472f0c82119325e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:26Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:27 crc kubenswrapper[4846]: I1005 06:48:27.006049 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56f1fe1f9f06f0249cf937a6963fda82467732dc056ca2b2189be6d8dd9d4a05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7212ce5f42c1fb55146abb190a557a504b43f16af6df84f90e1073567a329ec5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:27Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:27 crc kubenswrapper[4846]: I1005 06:48:27.020146 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"775b5702-a131-4d6c-852d-91bfa056cac5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0839c1765df2c68d2554e9c66d289505e559a9672caf8069feeff2c4fdfd2171\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15dbaf2445704da0987f256f81dff2ee95664988541fa48ce94898afeadab2e8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd3b3ab323c6d5153054aa04af2d91f7b9e8436ff1e2f75f3594b57831b2a416\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c52e076eddfd307110b1f6de00a169b81af8f2785af7becef2a820ce363a5861\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c52e076eddfd307110b1f6de00a169b81af8f2785af7becef2a820ce363a5861\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-05T06:48:14Z\\\",\\\"message\\\":\\\"le observer\\\\nW1005 06:48:14.337870 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1005 06:48:14.338086 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1005 06:48:14.339389 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1467752215/tls.crt::/tmp/serving-cert-1467752215/tls.key\\\\\\\"\\\\nI1005 06:48:14.864345 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1005 06:48:14.868087 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1005 06:48:14.868114 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1005 06:48:14.868140 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1005 06:48:14.868147 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1005 06:48:14.877354 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1005 06:48:14.877412 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1005 06:48:14.877456 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1005 06:48:14.877471 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1005 06:48:14.877490 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1005 06:48:14.877501 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1005 06:48:14.877507 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1005 06:48:14.877515 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1005 06:48:14.880326 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:13Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4a091716ff935adb42ae2e88048619fd347bd174f64e5f9c97a01f81690e57f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:01Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dc5411b642f99b13d514d2f21defad0c394aeeb82a966f72300076617e2586c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5dc5411b642f99b13d514d2f21defad0c394aeeb82a966f72300076617e2586c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:47:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:47:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:47:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:27Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:27 crc kubenswrapper[4846]: I1005 06:48:27.030711 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:27Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:27 crc kubenswrapper[4846]: I1005 06:48:27.041885 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:27 crc kubenswrapper[4846]: I1005 06:48:27.041928 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:27 crc kubenswrapper[4846]: I1005 06:48:27.041942 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:27 crc kubenswrapper[4846]: I1005 06:48:27.041964 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:27 crc kubenswrapper[4846]: I1005 06:48:27.041979 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:27Z","lastTransitionTime":"2025-10-05T06:48:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:27 crc kubenswrapper[4846]: I1005 06:48:27.046783 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-r7ztb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a5119cd9-af70-4324-a955-978306ab0b20\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06231a87429ec292cb57ebd88cda7ad547bcb32e80343b776833621a11c28ad1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jfcj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-r7ztb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:27Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:27 crc kubenswrapper[4846]: I1005 06:48:27.145335 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:27 crc kubenswrapper[4846]: I1005 06:48:27.145382 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:27 crc kubenswrapper[4846]: I1005 06:48:27.145395 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:27 crc kubenswrapper[4846]: I1005 06:48:27.145414 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:27 crc kubenswrapper[4846]: I1005 06:48:27.145427 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:27Z","lastTransitionTime":"2025-10-05T06:48:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:27 crc kubenswrapper[4846]: I1005 06:48:27.247893 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:27 crc kubenswrapper[4846]: I1005 06:48:27.247975 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:27 crc kubenswrapper[4846]: I1005 06:48:27.247996 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:27 crc kubenswrapper[4846]: I1005 06:48:27.248023 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:27 crc kubenswrapper[4846]: I1005 06:48:27.248043 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:27Z","lastTransitionTime":"2025-10-05T06:48:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:27 crc kubenswrapper[4846]: I1005 06:48:27.266856 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/node-ca-w2sj8"] Oct 05 06:48:27 crc kubenswrapper[4846]: I1005 06:48:27.267362 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-w2sj8" Oct 05 06:48:27 crc kubenswrapper[4846]: I1005 06:48:27.271004 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Oct 05 06:48:27 crc kubenswrapper[4846]: I1005 06:48:27.271142 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Oct 05 06:48:27 crc kubenswrapper[4846]: I1005 06:48:27.271570 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Oct 05 06:48:27 crc kubenswrapper[4846]: I1005 06:48:27.272792 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Oct 05 06:48:27 crc kubenswrapper[4846]: I1005 06:48:27.292875 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"775b5702-a131-4d6c-852d-91bfa056cac5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0839c1765df2c68d2554e9c66d289505e559a9672caf8069feeff2c4fdfd2171\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15dbaf2445704da0987f256f81dff2ee95664988541fa48ce94898afeadab2e8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd3b3ab323c6d5153054aa04af2d91f7b9e8436ff1e2f75f3594b57831b2a416\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c52e076eddfd307110b1f6de00a169b81af8f2785af7becef2a820ce363a5861\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c52e076eddfd307110b1f6de00a169b81af8f2785af7becef2a820ce363a5861\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-05T06:48:14Z\\\",\\\"message\\\":\\\"le observer\\\\nW1005 06:48:14.337870 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1005 06:48:14.338086 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1005 06:48:14.339389 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1467752215/tls.crt::/tmp/serving-cert-1467752215/tls.key\\\\\\\"\\\\nI1005 06:48:14.864345 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1005 06:48:14.868087 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1005 06:48:14.868114 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1005 06:48:14.868140 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1005 06:48:14.868147 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1005 06:48:14.877354 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1005 06:48:14.877412 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1005 06:48:14.877456 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1005 06:48:14.877471 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1005 06:48:14.877490 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1005 06:48:14.877501 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1005 06:48:14.877507 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1005 06:48:14.877515 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1005 06:48:14.880326 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:13Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4a091716ff935adb42ae2e88048619fd347bd174f64e5f9c97a01f81690e57f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:01Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dc5411b642f99b13d514d2f21defad0c394aeeb82a966f72300076617e2586c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5dc5411b642f99b13d514d2f21defad0c394aeeb82a966f72300076617e2586c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:47:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:47:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:47:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:27Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:27 crc kubenswrapper[4846]: I1005 06:48:27.313818 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:27Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:27 crc kubenswrapper[4846]: I1005 06:48:27.323646 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9rtxf\" (UniqueName: \"kubernetes.io/projected/32931fce-7374-4cab-a2e0-79d96a842274-kube-api-access-9rtxf\") pod \"node-ca-w2sj8\" (UID: \"32931fce-7374-4cab-a2e0-79d96a842274\") " pod="openshift-image-registry/node-ca-w2sj8" Oct 05 06:48:27 crc kubenswrapper[4846]: I1005 06:48:27.323739 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/32931fce-7374-4cab-a2e0-79d96a842274-host\") pod \"node-ca-w2sj8\" (UID: \"32931fce-7374-4cab-a2e0-79d96a842274\") " pod="openshift-image-registry/node-ca-w2sj8" Oct 05 06:48:27 crc kubenswrapper[4846]: I1005 06:48:27.323824 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/32931fce-7374-4cab-a2e0-79d96a842274-serviceca\") pod \"node-ca-w2sj8\" (UID: \"32931fce-7374-4cab-a2e0-79d96a842274\") " pod="openshift-image-registry/node-ca-w2sj8" Oct 05 06:48:27 crc kubenswrapper[4846]: I1005 06:48:27.330899 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-r7ztb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a5119cd9-af70-4324-a955-978306ab0b20\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06231a87429ec292cb57ebd88cda7ad547bcb32e80343b776833621a11c28ad1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jfcj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-r7ztb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:27Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:27 crc kubenswrapper[4846]: I1005 06:48:27.350671 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:27 crc kubenswrapper[4846]: I1005 06:48:27.350739 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:27 crc kubenswrapper[4846]: I1005 06:48:27.350758 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:27 crc kubenswrapper[4846]: I1005 06:48:27.350790 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:27 crc kubenswrapper[4846]: I1005 06:48:27.350851 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:27Z","lastTransitionTime":"2025-10-05T06:48:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:27 crc kubenswrapper[4846]: I1005 06:48:27.353108 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c482bb97-fd50-40cf-96ea-f806c3f01125\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bfe45d30e4fa141541aca1d46bc5cc1da0a7b4a5eb1fc9bb2caa5b7fdcb0b264\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://496a4782faa30a9a20ff63038141600cd635f12902d1cf17989cd067fcdfe929\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://281be077869a6aa1ca815770871a14413d4a97e97aeb550fd5cd215cba88b21f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6771c7215d2f5dd5838d17888375a4768b9d1755fc2e21bda1bbfbb2eb1dbb95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c17a711c2c7473b076a3944ffe7c700808d05cba582fbd55c5c2ac8ff73838a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44a9f740f0a06c8bf928e2a9d1d5979217558c31e73a333631baccebf00fd5d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://44a9f740f0a06c8bf928e2a9d1d5979217558c31e73a333631baccebf00fd5d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:47:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:47:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://48ca0ba6ca6092bfcb5b153220c054f02cc36fb22707c42e021f06d329e18e2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48ca0ba6ca6092bfcb5b153220c054f02cc36fb22707c42e021f06d329e18e2e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:00Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4006d0cea128675c8885d879171204abffc5e11ea0664750a12b4bc7344ccb2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4006d0cea128675c8885d879171204abffc5e11ea0664750a12b4bc7344ccb2e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:47:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:27Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:27 crc kubenswrapper[4846]: I1005 06:48:27.369303 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:27Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:27 crc kubenswrapper[4846]: I1005 06:48:27.394603 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-sfmxd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1dbb13ac-136c-4a1c-9229-b4e8ac4f9c71\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://226d977b596d07aeda0486452e2a8605949ed4d2453b21f56174937c395d6f36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://226d977b596d07aeda0486452e2a8605949ed4d2453b21f56174937c395d6f36\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://99c07a4bdb3059d3a99ffdaf794d668d52cffd9e5e3bb8945204dd6d1f8c8328\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-sfmxd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:27Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:27 crc kubenswrapper[4846]: I1005 06:48:27.420929 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-txzlk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"93a6ace4-8d64-44d7-9637-457d3af6543c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff3c6545ff3d2436da27bbd223b950b30b0c69a3e7f3034cb0e61380cb38d338\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ff3c6545ff3d2436da27bbd223b950b30b0c69a3e7f3034cb0e61380cb38d338\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:24Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-txzlk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:27Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:27 crc kubenswrapper[4846]: I1005 06:48:27.424887 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/32931fce-7374-4cab-a2e0-79d96a842274-host\") pod \"node-ca-w2sj8\" (UID: \"32931fce-7374-4cab-a2e0-79d96a842274\") " pod="openshift-image-registry/node-ca-w2sj8" Oct 05 06:48:27 crc kubenswrapper[4846]: I1005 06:48:27.424980 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/32931fce-7374-4cab-a2e0-79d96a842274-serviceca\") pod \"node-ca-w2sj8\" (UID: \"32931fce-7374-4cab-a2e0-79d96a842274\") " pod="openshift-image-registry/node-ca-w2sj8" Oct 05 06:48:27 crc kubenswrapper[4846]: I1005 06:48:27.425067 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9rtxf\" (UniqueName: \"kubernetes.io/projected/32931fce-7374-4cab-a2e0-79d96a842274-kube-api-access-9rtxf\") pod \"node-ca-w2sj8\" (UID: \"32931fce-7374-4cab-a2e0-79d96a842274\") " pod="openshift-image-registry/node-ca-w2sj8" Oct 05 06:48:27 crc kubenswrapper[4846]: I1005 06:48:27.425070 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/32931fce-7374-4cab-a2e0-79d96a842274-host\") pod \"node-ca-w2sj8\" (UID: \"32931fce-7374-4cab-a2e0-79d96a842274\") " pod="openshift-image-registry/node-ca-w2sj8" Oct 05 06:48:27 crc kubenswrapper[4846]: I1005 06:48:27.427353 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/32931fce-7374-4cab-a2e0-79d96a842274-serviceca\") pod \"node-ca-w2sj8\" (UID: \"32931fce-7374-4cab-a2e0-79d96a842274\") " pod="openshift-image-registry/node-ca-w2sj8" Oct 05 06:48:27 crc kubenswrapper[4846]: I1005 06:48:27.437119 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c9f812b-ecc8-4985-a670-53fa80fbb499\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2b5ee9781c01d6aec3ac778efef1067b6c3bfe4ed37431bdebe188fcde0165f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:47:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3fb4717692aa1fec0012b514ef8d18d1ca3707cfa4bb8248bb93841a0a40f11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:47:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0b6a7afe3094f1c92093c05c28344897a615d590c2fbb507c87e18a44c6c831\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://640ade2c733d7916ec213a201a1dd614d261c69eab654c80100efbc99a768995\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:47:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:27Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:27 crc kubenswrapper[4846]: I1005 06:48:27.453539 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:27Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:27 crc kubenswrapper[4846]: I1005 06:48:27.454136 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:27 crc kubenswrapper[4846]: I1005 06:48:27.454209 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:27 crc kubenswrapper[4846]: I1005 06:48:27.454227 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:27 crc kubenswrapper[4846]: I1005 06:48:27.454257 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:27 crc kubenswrapper[4846]: I1005 06:48:27.454291 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:27Z","lastTransitionTime":"2025-10-05T06:48:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:27 crc kubenswrapper[4846]: I1005 06:48:27.455012 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9rtxf\" (UniqueName: \"kubernetes.io/projected/32931fce-7374-4cab-a2e0-79d96a842274-kube-api-access-9rtxf\") pod \"node-ca-w2sj8\" (UID: \"32931fce-7374-4cab-a2e0-79d96a842274\") " pod="openshift-image-registry/node-ca-w2sj8" Oct 05 06:48:27 crc kubenswrapper[4846]: I1005 06:48:27.472244 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e67d0e447a8a6a381259e25d25badc791194e1291f56b1cd77bd63989b20408\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dkhl7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://516f88f4990f5962d8b694a3d889259cc5508a309919870191c2e1331640c973\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dkhl7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-fscvf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:27Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:27 crc kubenswrapper[4846]: I1005 06:48:27.488583 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-w2sj8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32931fce-7374-4cab-a2e0-79d96a842274\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:27Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:27Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:27Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9rtxf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:27Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-w2sj8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:27Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:27 crc kubenswrapper[4846]: I1005 06:48:27.496650 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 05 06:48:27 crc kubenswrapper[4846]: I1005 06:48:27.496683 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 05 06:48:27 crc kubenswrapper[4846]: E1005 06:48:27.496843 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 05 06:48:27 crc kubenswrapper[4846]: E1005 06:48:27.496956 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 05 06:48:27 crc kubenswrapper[4846]: I1005 06:48:27.505214 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ff10c31459b3b3724ee7b769bd6f72f9ed80cce5419b6adcdf69a3239b70756\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:27Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:27 crc kubenswrapper[4846]: I1005 06:48:27.519768 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://347f900743b7bba3daf92f21c7eba0625e8c774f50057bc9472f0c82119325e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:27Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:27 crc kubenswrapper[4846]: I1005 06:48:27.535600 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56f1fe1f9f06f0249cf937a6963fda82467732dc056ca2b2189be6d8dd9d4a05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7212ce5f42c1fb55146abb190a557a504b43f16af6df84f90e1073567a329ec5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:27Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:27 crc kubenswrapper[4846]: I1005 06:48:27.550410 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-7wm5q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9bc7744f-d700-450f-8b8d-98140877fee9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90379e74c544fccca44f7acdc228ff378154390e211ac801b79f8d074012241e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h5p4q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-7wm5q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:27Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:27 crc kubenswrapper[4846]: I1005 06:48:27.557610 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:27 crc kubenswrapper[4846]: I1005 06:48:27.557659 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:27 crc kubenswrapper[4846]: I1005 06:48:27.557672 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:27 crc kubenswrapper[4846]: I1005 06:48:27.557692 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:27 crc kubenswrapper[4846]: I1005 06:48:27.557706 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:27Z","lastTransitionTime":"2025-10-05T06:48:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:27 crc kubenswrapper[4846]: I1005 06:48:27.588849 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-w2sj8" Oct 05 06:48:27 crc kubenswrapper[4846]: W1005 06:48:27.611393 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod32931fce_7374_4cab_a2e0_79d96a842274.slice/crio-414bbf09365ff12ad647484259e5d25647200fca35ba6876975487eb131a3ef6 WatchSource:0}: Error finding container 414bbf09365ff12ad647484259e5d25647200fca35ba6876975487eb131a3ef6: Status 404 returned error can't find the container with id 414bbf09365ff12ad647484259e5d25647200fca35ba6876975487eb131a3ef6 Oct 05 06:48:27 crc kubenswrapper[4846]: I1005 06:48:27.661694 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:27 crc kubenswrapper[4846]: I1005 06:48:27.661748 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:27 crc kubenswrapper[4846]: I1005 06:48:27.661763 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:27 crc kubenswrapper[4846]: I1005 06:48:27.661789 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:27 crc kubenswrapper[4846]: I1005 06:48:27.661803 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:27Z","lastTransitionTime":"2025-10-05T06:48:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:27 crc kubenswrapper[4846]: I1005 06:48:27.736915 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-w2sj8" event={"ID":"32931fce-7374-4cab-a2e0-79d96a842274","Type":"ContainerStarted","Data":"414bbf09365ff12ad647484259e5d25647200fca35ba6876975487eb131a3ef6"} Oct 05 06:48:27 crc kubenswrapper[4846]: I1005 06:48:27.739393 4846 generic.go:334] "Generic (PLEG): container finished" podID="1dbb13ac-136c-4a1c-9229-b4e8ac4f9c71" containerID="99c07a4bdb3059d3a99ffdaf794d668d52cffd9e5e3bb8945204dd6d1f8c8328" exitCode=0 Oct 05 06:48:27 crc kubenswrapper[4846]: I1005 06:48:27.739466 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-sfmxd" event={"ID":"1dbb13ac-136c-4a1c-9229-b4e8ac4f9c71","Type":"ContainerDied","Data":"99c07a4bdb3059d3a99ffdaf794d668d52cffd9e5e3bb8945204dd6d1f8c8328"} Oct 05 06:48:27 crc kubenswrapper[4846]: I1005 06:48:27.750409 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-txzlk" event={"ID":"93a6ace4-8d64-44d7-9637-457d3af6543c","Type":"ContainerStarted","Data":"9e9b0219ecbc7e60905d6de3e67a2c8dc9af1017b4aed9538ee0c58c861f2438"} Oct 05 06:48:27 crc kubenswrapper[4846]: I1005 06:48:27.762956 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:27Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:27 crc kubenswrapper[4846]: I1005 06:48:27.764869 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:27 crc kubenswrapper[4846]: I1005 06:48:27.764904 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:27 crc kubenswrapper[4846]: I1005 06:48:27.764924 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:27 crc kubenswrapper[4846]: I1005 06:48:27.764952 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:27 crc kubenswrapper[4846]: I1005 06:48:27.764970 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:27Z","lastTransitionTime":"2025-10-05T06:48:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:27 crc kubenswrapper[4846]: I1005 06:48:27.782061 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e67d0e447a8a6a381259e25d25badc791194e1291f56b1cd77bd63989b20408\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dkhl7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://516f88f4990f5962d8b694a3d889259cc5508a309919870191c2e1331640c973\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dkhl7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-fscvf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:27Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:27 crc kubenswrapper[4846]: I1005 06:48:27.794598 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-w2sj8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32931fce-7374-4cab-a2e0-79d96a842274\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:27Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:27Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:27Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9rtxf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:27Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-w2sj8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:27Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:27 crc kubenswrapper[4846]: I1005 06:48:27.811815 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c9f812b-ecc8-4985-a670-53fa80fbb499\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2b5ee9781c01d6aec3ac778efef1067b6c3bfe4ed37431bdebe188fcde0165f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:47:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3fb4717692aa1fec0012b514ef8d18d1ca3707cfa4bb8248bb93841a0a40f11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:47:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0b6a7afe3094f1c92093c05c28344897a615d590c2fbb507c87e18a44c6c831\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://640ade2c733d7916ec213a201a1dd614d261c69eab654c80100efbc99a768995\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:47:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:27Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:27 crc kubenswrapper[4846]: I1005 06:48:27.828858 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ff10c31459b3b3724ee7b769bd6f72f9ed80cce5419b6adcdf69a3239b70756\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:27Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:27 crc kubenswrapper[4846]: I1005 06:48:27.844102 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://347f900743b7bba3daf92f21c7eba0625e8c774f50057bc9472f0c82119325e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:27Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:27 crc kubenswrapper[4846]: I1005 06:48:27.858774 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56f1fe1f9f06f0249cf937a6963fda82467732dc056ca2b2189be6d8dd9d4a05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7212ce5f42c1fb55146abb190a557a504b43f16af6df84f90e1073567a329ec5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:27Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:27 crc kubenswrapper[4846]: I1005 06:48:27.870400 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:27 crc kubenswrapper[4846]: I1005 06:48:27.870445 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:27 crc kubenswrapper[4846]: I1005 06:48:27.870488 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:27 crc kubenswrapper[4846]: I1005 06:48:27.870509 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:27 crc kubenswrapper[4846]: I1005 06:48:27.870590 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:27Z","lastTransitionTime":"2025-10-05T06:48:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:27 crc kubenswrapper[4846]: I1005 06:48:27.872100 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-7wm5q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9bc7744f-d700-450f-8b8d-98140877fee9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90379e74c544fccca44f7acdc228ff378154390e211ac801b79f8d074012241e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h5p4q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-7wm5q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:27Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:27 crc kubenswrapper[4846]: I1005 06:48:27.892771 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"775b5702-a131-4d6c-852d-91bfa056cac5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0839c1765df2c68d2554e9c66d289505e559a9672caf8069feeff2c4fdfd2171\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15dbaf2445704da0987f256f81dff2ee95664988541fa48ce94898afeadab2e8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd3b3ab323c6d5153054aa04af2d91f7b9e8436ff1e2f75f3594b57831b2a416\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c52e076eddfd307110b1f6de00a169b81af8f2785af7becef2a820ce363a5861\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c52e076eddfd307110b1f6de00a169b81af8f2785af7becef2a820ce363a5861\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-05T06:48:14Z\\\",\\\"message\\\":\\\"le observer\\\\nW1005 06:48:14.337870 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1005 06:48:14.338086 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1005 06:48:14.339389 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1467752215/tls.crt::/tmp/serving-cert-1467752215/tls.key\\\\\\\"\\\\nI1005 06:48:14.864345 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1005 06:48:14.868087 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1005 06:48:14.868114 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1005 06:48:14.868140 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1005 06:48:14.868147 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1005 06:48:14.877354 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1005 06:48:14.877412 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1005 06:48:14.877456 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1005 06:48:14.877471 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1005 06:48:14.877490 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1005 06:48:14.877501 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1005 06:48:14.877507 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1005 06:48:14.877515 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1005 06:48:14.880326 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:13Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4a091716ff935adb42ae2e88048619fd347bd174f64e5f9c97a01f81690e57f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:01Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dc5411b642f99b13d514d2f21defad0c394aeeb82a966f72300076617e2586c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5dc5411b642f99b13d514d2f21defad0c394aeeb82a966f72300076617e2586c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:47:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:47:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:47:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:27Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:27 crc kubenswrapper[4846]: I1005 06:48:27.905581 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:27Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:27 crc kubenswrapper[4846]: I1005 06:48:27.917683 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-r7ztb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a5119cd9-af70-4324-a955-978306ab0b20\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06231a87429ec292cb57ebd88cda7ad547bcb32e80343b776833621a11c28ad1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jfcj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-r7ztb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:27Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:27 crc kubenswrapper[4846]: I1005 06:48:27.933202 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:27Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:27 crc kubenswrapper[4846]: I1005 06:48:27.950646 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-sfmxd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1dbb13ac-136c-4a1c-9229-b4e8ac4f9c71\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://226d977b596d07aeda0486452e2a8605949ed4d2453b21f56174937c395d6f36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://226d977b596d07aeda0486452e2a8605949ed4d2453b21f56174937c395d6f36\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://99c07a4bdb3059d3a99ffdaf794d668d52cffd9e5e3bb8945204dd6d1f8c8328\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99c07a4bdb3059d3a99ffdaf794d668d52cffd9e5e3bb8945204dd6d1f8c8328\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-sfmxd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:27Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:27 crc kubenswrapper[4846]: I1005 06:48:27.972112 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-txzlk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"93a6ace4-8d64-44d7-9637-457d3af6543c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff3c6545ff3d2436da27bbd223b950b30b0c69a3e7f3034cb0e61380cb38d338\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ff3c6545ff3d2436da27bbd223b950b30b0c69a3e7f3034cb0e61380cb38d338\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:24Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-txzlk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:27Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:27 crc kubenswrapper[4846]: I1005 06:48:27.973836 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:27 crc kubenswrapper[4846]: I1005 06:48:27.973865 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:27 crc kubenswrapper[4846]: I1005 06:48:27.973877 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:27 crc kubenswrapper[4846]: I1005 06:48:27.973892 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:27 crc kubenswrapper[4846]: I1005 06:48:27.973904 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:27Z","lastTransitionTime":"2025-10-05T06:48:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:28 crc kubenswrapper[4846]: I1005 06:48:28.007299 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c482bb97-fd50-40cf-96ea-f806c3f01125\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bfe45d30e4fa141541aca1d46bc5cc1da0a7b4a5eb1fc9bb2caa5b7fdcb0b264\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://496a4782faa30a9a20ff63038141600cd635f12902d1cf17989cd067fcdfe929\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://281be077869a6aa1ca815770871a14413d4a97e97aeb550fd5cd215cba88b21f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6771c7215d2f5dd5838d17888375a4768b9d1755fc2e21bda1bbfbb2eb1dbb95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c17a711c2c7473b076a3944ffe7c700808d05cba582fbd55c5c2ac8ff73838a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44a9f740f0a06c8bf928e2a9d1d5979217558c31e73a333631baccebf00fd5d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://44a9f740f0a06c8bf928e2a9d1d5979217558c31e73a333631baccebf00fd5d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:47:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:47:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://48ca0ba6ca6092bfcb5b153220c054f02cc36fb22707c42e021f06d329e18e2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48ca0ba6ca6092bfcb5b153220c054f02cc36fb22707c42e021f06d329e18e2e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:00Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4006d0cea128675c8885d879171204abffc5e11ea0664750a12b4bc7344ccb2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4006d0cea128675c8885d879171204abffc5e11ea0664750a12b4bc7344ccb2e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:47:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:28Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:28 crc kubenswrapper[4846]: I1005 06:48:28.077029 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:28 crc kubenswrapper[4846]: I1005 06:48:28.077074 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:28 crc kubenswrapper[4846]: I1005 06:48:28.077089 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:28 crc kubenswrapper[4846]: I1005 06:48:28.077109 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:28 crc kubenswrapper[4846]: I1005 06:48:28.077127 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:28Z","lastTransitionTime":"2025-10-05T06:48:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:28 crc kubenswrapper[4846]: I1005 06:48:28.186084 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:28 crc kubenswrapper[4846]: I1005 06:48:28.186221 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:28 crc kubenswrapper[4846]: I1005 06:48:28.186242 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:28 crc kubenswrapper[4846]: I1005 06:48:28.186266 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:28 crc kubenswrapper[4846]: I1005 06:48:28.186317 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:28Z","lastTransitionTime":"2025-10-05T06:48:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:28 crc kubenswrapper[4846]: I1005 06:48:28.291261 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:28 crc kubenswrapper[4846]: I1005 06:48:28.291303 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:28 crc kubenswrapper[4846]: I1005 06:48:28.291313 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:28 crc kubenswrapper[4846]: I1005 06:48:28.291329 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:28 crc kubenswrapper[4846]: I1005 06:48:28.291341 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:28Z","lastTransitionTime":"2025-10-05T06:48:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:28 crc kubenswrapper[4846]: I1005 06:48:28.394471 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:28 crc kubenswrapper[4846]: I1005 06:48:28.394529 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:28 crc kubenswrapper[4846]: I1005 06:48:28.394542 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:28 crc kubenswrapper[4846]: I1005 06:48:28.394558 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:28 crc kubenswrapper[4846]: I1005 06:48:28.394569 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:28Z","lastTransitionTime":"2025-10-05T06:48:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:28 crc kubenswrapper[4846]: I1005 06:48:28.497382 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 05 06:48:28 crc kubenswrapper[4846]: E1005 06:48:28.497607 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 05 06:48:28 crc kubenswrapper[4846]: I1005 06:48:28.498084 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:28 crc kubenswrapper[4846]: I1005 06:48:28.498131 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:28 crc kubenswrapper[4846]: I1005 06:48:28.498149 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:28 crc kubenswrapper[4846]: I1005 06:48:28.498210 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:28 crc kubenswrapper[4846]: I1005 06:48:28.498239 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:28Z","lastTransitionTime":"2025-10-05T06:48:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:28 crc kubenswrapper[4846]: I1005 06:48:28.520461 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://347f900743b7bba3daf92f21c7eba0625e8c774f50057bc9472f0c82119325e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:28Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:28 crc kubenswrapper[4846]: I1005 06:48:28.541502 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56f1fe1f9f06f0249cf937a6963fda82467732dc056ca2b2189be6d8dd9d4a05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7212ce5f42c1fb55146abb190a557a504b43f16af6df84f90e1073567a329ec5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:28Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:28 crc kubenswrapper[4846]: I1005 06:48:28.562437 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-7wm5q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9bc7744f-d700-450f-8b8d-98140877fee9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90379e74c544fccca44f7acdc228ff378154390e211ac801b79f8d074012241e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h5p4q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-7wm5q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:28Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:28 crc kubenswrapper[4846]: I1005 06:48:28.583130 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ff10c31459b3b3724ee7b769bd6f72f9ed80cce5419b6adcdf69a3239b70756\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:28Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:28 crc kubenswrapper[4846]: I1005 06:48:28.600430 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:28 crc kubenswrapper[4846]: I1005 06:48:28.600484 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:28 crc kubenswrapper[4846]: I1005 06:48:28.600500 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:28 crc kubenswrapper[4846]: I1005 06:48:28.600525 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:28 crc kubenswrapper[4846]: I1005 06:48:28.600543 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:28Z","lastTransitionTime":"2025-10-05T06:48:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:28 crc kubenswrapper[4846]: I1005 06:48:28.614034 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:28Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:28 crc kubenswrapper[4846]: I1005 06:48:28.633770 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-r7ztb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a5119cd9-af70-4324-a955-978306ab0b20\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06231a87429ec292cb57ebd88cda7ad547bcb32e80343b776833621a11c28ad1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jfcj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-r7ztb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:28Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:28 crc kubenswrapper[4846]: I1005 06:48:28.656938 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"775b5702-a131-4d6c-852d-91bfa056cac5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0839c1765df2c68d2554e9c66d289505e559a9672caf8069feeff2c4fdfd2171\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15dbaf2445704da0987f256f81dff2ee95664988541fa48ce94898afeadab2e8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd3b3ab323c6d5153054aa04af2d91f7b9e8436ff1e2f75f3594b57831b2a416\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c52e076eddfd307110b1f6de00a169b81af8f2785af7becef2a820ce363a5861\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c52e076eddfd307110b1f6de00a169b81af8f2785af7becef2a820ce363a5861\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-05T06:48:14Z\\\",\\\"message\\\":\\\"le observer\\\\nW1005 06:48:14.337870 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1005 06:48:14.338086 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1005 06:48:14.339389 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1467752215/tls.crt::/tmp/serving-cert-1467752215/tls.key\\\\\\\"\\\\nI1005 06:48:14.864345 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1005 06:48:14.868087 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1005 06:48:14.868114 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1005 06:48:14.868140 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1005 06:48:14.868147 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1005 06:48:14.877354 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1005 06:48:14.877412 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1005 06:48:14.877456 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1005 06:48:14.877471 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1005 06:48:14.877490 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1005 06:48:14.877501 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1005 06:48:14.877507 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1005 06:48:14.877515 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1005 06:48:14.880326 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:13Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4a091716ff935adb42ae2e88048619fd347bd174f64e5f9c97a01f81690e57f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:01Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dc5411b642f99b13d514d2f21defad0c394aeeb82a966f72300076617e2586c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5dc5411b642f99b13d514d2f21defad0c394aeeb82a966f72300076617e2586c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:47:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:47:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:47:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:28Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:28 crc kubenswrapper[4846]: I1005 06:48:28.682610 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-sfmxd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1dbb13ac-136c-4a1c-9229-b4e8ac4f9c71\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://226d977b596d07aeda0486452e2a8605949ed4d2453b21f56174937c395d6f36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://226d977b596d07aeda0486452e2a8605949ed4d2453b21f56174937c395d6f36\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://99c07a4bdb3059d3a99ffdaf794d668d52cffd9e5e3bb8945204dd6d1f8c8328\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99c07a4bdb3059d3a99ffdaf794d668d52cffd9e5e3bb8945204dd6d1f8c8328\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-sfmxd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:28Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:28 crc kubenswrapper[4846]: I1005 06:48:28.703372 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:28 crc kubenswrapper[4846]: I1005 06:48:28.703437 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:28 crc kubenswrapper[4846]: I1005 06:48:28.703454 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:28 crc kubenswrapper[4846]: I1005 06:48:28.703483 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:28 crc kubenswrapper[4846]: I1005 06:48:28.703501 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:28Z","lastTransitionTime":"2025-10-05T06:48:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:28 crc kubenswrapper[4846]: I1005 06:48:28.720008 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-txzlk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"93a6ace4-8d64-44d7-9637-457d3af6543c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff3c6545ff3d2436da27bbd223b950b30b0c69a3e7f3034cb0e61380cb38d338\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ff3c6545ff3d2436da27bbd223b950b30b0c69a3e7f3034cb0e61380cb38d338\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:24Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-txzlk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:28Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:28 crc kubenswrapper[4846]: I1005 06:48:28.755879 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c482bb97-fd50-40cf-96ea-f806c3f01125\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bfe45d30e4fa141541aca1d46bc5cc1da0a7b4a5eb1fc9bb2caa5b7fdcb0b264\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://496a4782faa30a9a20ff63038141600cd635f12902d1cf17989cd067fcdfe929\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://281be077869a6aa1ca815770871a14413d4a97e97aeb550fd5cd215cba88b21f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6771c7215d2f5dd5838d17888375a4768b9d1755fc2e21bda1bbfbb2eb1dbb95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c17a711c2c7473b076a3944ffe7c700808d05cba582fbd55c5c2ac8ff73838a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44a9f740f0a06c8bf928e2a9d1d5979217558c31e73a333631baccebf00fd5d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://44a9f740f0a06c8bf928e2a9d1d5979217558c31e73a333631baccebf00fd5d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:47:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:47:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://48ca0ba6ca6092bfcb5b153220c054f02cc36fb22707c42e021f06d329e18e2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48ca0ba6ca6092bfcb5b153220c054f02cc36fb22707c42e021f06d329e18e2e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:00Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4006d0cea128675c8885d879171204abffc5e11ea0664750a12b4bc7344ccb2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4006d0cea128675c8885d879171204abffc5e11ea0664750a12b4bc7344ccb2e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:47:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:28Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:28 crc kubenswrapper[4846]: I1005 06:48:28.759964 4846 generic.go:334] "Generic (PLEG): container finished" podID="1dbb13ac-136c-4a1c-9229-b4e8ac4f9c71" containerID="27095272bf4f31d63a78a436636de0aae12f7103597da944f24ed7a74329ee3f" exitCode=0 Oct 05 06:48:28 crc kubenswrapper[4846]: I1005 06:48:28.760892 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-sfmxd" event={"ID":"1dbb13ac-136c-4a1c-9229-b4e8ac4f9c71","Type":"ContainerDied","Data":"27095272bf4f31d63a78a436636de0aae12f7103597da944f24ed7a74329ee3f"} Oct 05 06:48:28 crc kubenswrapper[4846]: I1005 06:48:28.763146 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-w2sj8" event={"ID":"32931fce-7374-4cab-a2e0-79d96a842274","Type":"ContainerStarted","Data":"0df535d65d39de6c9a3dfb931bafc4469adc22b72f7a7903beda4fef8cd48138"} Oct 05 06:48:28 crc kubenswrapper[4846]: I1005 06:48:28.778967 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:28Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:28 crc kubenswrapper[4846]: I1005 06:48:28.804754 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e67d0e447a8a6a381259e25d25badc791194e1291f56b1cd77bd63989b20408\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dkhl7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://516f88f4990f5962d8b694a3d889259cc5508a309919870191c2e1331640c973\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dkhl7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-fscvf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:28Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:28 crc kubenswrapper[4846]: I1005 06:48:28.811240 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:28 crc kubenswrapper[4846]: I1005 06:48:28.811302 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:28 crc kubenswrapper[4846]: I1005 06:48:28.811319 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:28 crc kubenswrapper[4846]: I1005 06:48:28.811345 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:28 crc kubenswrapper[4846]: I1005 06:48:28.811367 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:28Z","lastTransitionTime":"2025-10-05T06:48:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:28 crc kubenswrapper[4846]: I1005 06:48:28.825323 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-w2sj8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32931fce-7374-4cab-a2e0-79d96a842274\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:27Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:27Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:27Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9rtxf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:27Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-w2sj8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:28Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:28 crc kubenswrapper[4846]: I1005 06:48:28.849373 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c9f812b-ecc8-4985-a670-53fa80fbb499\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2b5ee9781c01d6aec3ac778efef1067b6c3bfe4ed37431bdebe188fcde0165f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:47:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3fb4717692aa1fec0012b514ef8d18d1ca3707cfa4bb8248bb93841a0a40f11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:47:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0b6a7afe3094f1c92093c05c28344897a615d590c2fbb507c87e18a44c6c831\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://640ade2c733d7916ec213a201a1dd614d261c69eab654c80100efbc99a768995\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:47:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:28Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:28 crc kubenswrapper[4846]: I1005 06:48:28.873478 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:28Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:28 crc kubenswrapper[4846]: I1005 06:48:28.909947 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-txzlk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"93a6ace4-8d64-44d7-9637-457d3af6543c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff3c6545ff3d2436da27bbd223b950b30b0c69a3e7f3034cb0e61380cb38d338\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ff3c6545ff3d2436da27bbd223b950b30b0c69a3e7f3034cb0e61380cb38d338\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:24Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-txzlk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:28Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:28 crc kubenswrapper[4846]: I1005 06:48:28.927078 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:28 crc kubenswrapper[4846]: I1005 06:48:28.927148 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:28 crc kubenswrapper[4846]: I1005 06:48:28.927172 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:28 crc kubenswrapper[4846]: I1005 06:48:28.927255 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:28 crc kubenswrapper[4846]: I1005 06:48:28.927282 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:28Z","lastTransitionTime":"2025-10-05T06:48:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:28 crc kubenswrapper[4846]: I1005 06:48:28.942142 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c482bb97-fd50-40cf-96ea-f806c3f01125\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bfe45d30e4fa141541aca1d46bc5cc1da0a7b4a5eb1fc9bb2caa5b7fdcb0b264\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://496a4782faa30a9a20ff63038141600cd635f12902d1cf17989cd067fcdfe929\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://281be077869a6aa1ca815770871a14413d4a97e97aeb550fd5cd215cba88b21f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6771c7215d2f5dd5838d17888375a4768b9d1755fc2e21bda1bbfbb2eb1dbb95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c17a711c2c7473b076a3944ffe7c700808d05cba582fbd55c5c2ac8ff73838a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44a9f740f0a06c8bf928e2a9d1d5979217558c31e73a333631baccebf00fd5d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://44a9f740f0a06c8bf928e2a9d1d5979217558c31e73a333631baccebf00fd5d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:47:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:47:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://48ca0ba6ca6092bfcb5b153220c054f02cc36fb22707c42e021f06d329e18e2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48ca0ba6ca6092bfcb5b153220c054f02cc36fb22707c42e021f06d329e18e2e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:00Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4006d0cea128675c8885d879171204abffc5e11ea0664750a12b4bc7344ccb2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4006d0cea128675c8885d879171204abffc5e11ea0664750a12b4bc7344ccb2e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:47:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:28Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:28 crc kubenswrapper[4846]: I1005 06:48:28.962408 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:28Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:28 crc kubenswrapper[4846]: I1005 06:48:28.982105 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-sfmxd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1dbb13ac-136c-4a1c-9229-b4e8ac4f9c71\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://226d977b596d07aeda0486452e2a8605949ed4d2453b21f56174937c395d6f36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://226d977b596d07aeda0486452e2a8605949ed4d2453b21f56174937c395d6f36\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://99c07a4bdb3059d3a99ffdaf794d668d52cffd9e5e3bb8945204dd6d1f8c8328\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99c07a4bdb3059d3a99ffdaf794d668d52cffd9e5e3bb8945204dd6d1f8c8328\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27095272bf4f31d63a78a436636de0aae12f7103597da944f24ed7a74329ee3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27095272bf4f31d63a78a436636de0aae12f7103597da944f24ed7a74329ee3f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-sfmxd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:28Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:28 crc kubenswrapper[4846]: I1005 06:48:28.998253 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-w2sj8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32931fce-7374-4cab-a2e0-79d96a842274\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0df535d65d39de6c9a3dfb931bafc4469adc22b72f7a7903beda4fef8cd48138\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9rtxf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:27Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-w2sj8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:28Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:29 crc kubenswrapper[4846]: I1005 06:48:29.020135 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c9f812b-ecc8-4985-a670-53fa80fbb499\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2b5ee9781c01d6aec3ac778efef1067b6c3bfe4ed37431bdebe188fcde0165f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:47:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3fb4717692aa1fec0012b514ef8d18d1ca3707cfa4bb8248bb93841a0a40f11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:47:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0b6a7afe3094f1c92093c05c28344897a615d590c2fbb507c87e18a44c6c831\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://640ade2c733d7916ec213a201a1dd614d261c69eab654c80100efbc99a768995\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:47:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:29Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:29 crc kubenswrapper[4846]: I1005 06:48:29.030409 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:29 crc kubenswrapper[4846]: I1005 06:48:29.030511 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:29 crc kubenswrapper[4846]: I1005 06:48:29.030524 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:29 crc kubenswrapper[4846]: I1005 06:48:29.030542 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:29 crc kubenswrapper[4846]: I1005 06:48:29.030555 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:29Z","lastTransitionTime":"2025-10-05T06:48:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:29 crc kubenswrapper[4846]: I1005 06:48:29.035050 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:29Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:29 crc kubenswrapper[4846]: I1005 06:48:29.052583 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e67d0e447a8a6a381259e25d25badc791194e1291f56b1cd77bd63989b20408\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dkhl7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://516f88f4990f5962d8b694a3d889259cc5508a309919870191c2e1331640c973\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dkhl7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-fscvf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:29Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:29 crc kubenswrapper[4846]: I1005 06:48:29.068721 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56f1fe1f9f06f0249cf937a6963fda82467732dc056ca2b2189be6d8dd9d4a05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7212ce5f42c1fb55146abb190a557a504b43f16af6df84f90e1073567a329ec5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:29Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:29 crc kubenswrapper[4846]: I1005 06:48:29.083586 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-7wm5q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9bc7744f-d700-450f-8b8d-98140877fee9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90379e74c544fccca44f7acdc228ff378154390e211ac801b79f8d074012241e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h5p4q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-7wm5q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:29Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:29 crc kubenswrapper[4846]: I1005 06:48:29.100850 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ff10c31459b3b3724ee7b769bd6f72f9ed80cce5419b6adcdf69a3239b70756\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:29Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:29 crc kubenswrapper[4846]: I1005 06:48:29.113850 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://347f900743b7bba3daf92f21c7eba0625e8c774f50057bc9472f0c82119325e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:29Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:29 crc kubenswrapper[4846]: I1005 06:48:29.125013 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-r7ztb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a5119cd9-af70-4324-a955-978306ab0b20\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06231a87429ec292cb57ebd88cda7ad547bcb32e80343b776833621a11c28ad1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jfcj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-r7ztb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:29Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:29 crc kubenswrapper[4846]: I1005 06:48:29.136740 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:29 crc kubenswrapper[4846]: I1005 06:48:29.136788 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:29 crc kubenswrapper[4846]: I1005 06:48:29.136804 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:29 crc kubenswrapper[4846]: I1005 06:48:29.136825 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:29 crc kubenswrapper[4846]: I1005 06:48:29.136838 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:29Z","lastTransitionTime":"2025-10-05T06:48:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:29 crc kubenswrapper[4846]: I1005 06:48:29.139576 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"775b5702-a131-4d6c-852d-91bfa056cac5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0839c1765df2c68d2554e9c66d289505e559a9672caf8069feeff2c4fdfd2171\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15dbaf2445704da0987f256f81dff2ee95664988541fa48ce94898afeadab2e8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd3b3ab323c6d5153054aa04af2d91f7b9e8436ff1e2f75f3594b57831b2a416\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c52e076eddfd307110b1f6de00a169b81af8f2785af7becef2a820ce363a5861\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c52e076eddfd307110b1f6de00a169b81af8f2785af7becef2a820ce363a5861\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-05T06:48:14Z\\\",\\\"message\\\":\\\"le observer\\\\nW1005 06:48:14.337870 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1005 06:48:14.338086 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1005 06:48:14.339389 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1467752215/tls.crt::/tmp/serving-cert-1467752215/tls.key\\\\\\\"\\\\nI1005 06:48:14.864345 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1005 06:48:14.868087 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1005 06:48:14.868114 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1005 06:48:14.868140 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1005 06:48:14.868147 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1005 06:48:14.877354 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1005 06:48:14.877412 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1005 06:48:14.877456 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1005 06:48:14.877471 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1005 06:48:14.877490 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1005 06:48:14.877501 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1005 06:48:14.877507 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1005 06:48:14.877515 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1005 06:48:14.880326 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:13Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4a091716ff935adb42ae2e88048619fd347bd174f64e5f9c97a01f81690e57f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:01Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dc5411b642f99b13d514d2f21defad0c394aeeb82a966f72300076617e2586c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5dc5411b642f99b13d514d2f21defad0c394aeeb82a966f72300076617e2586c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:47:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:47:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:47:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:29Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:29 crc kubenswrapper[4846]: I1005 06:48:29.158724 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:29Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:29 crc kubenswrapper[4846]: I1005 06:48:29.241004 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:29 crc kubenswrapper[4846]: I1005 06:48:29.241077 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:29 crc kubenswrapper[4846]: I1005 06:48:29.241096 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:29 crc kubenswrapper[4846]: I1005 06:48:29.241126 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:29 crc kubenswrapper[4846]: I1005 06:48:29.241148 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:29Z","lastTransitionTime":"2025-10-05T06:48:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:29 crc kubenswrapper[4846]: I1005 06:48:29.344717 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:29 crc kubenswrapper[4846]: I1005 06:48:29.344788 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:29 crc kubenswrapper[4846]: I1005 06:48:29.344807 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:29 crc kubenswrapper[4846]: I1005 06:48:29.344836 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:29 crc kubenswrapper[4846]: I1005 06:48:29.344855 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:29Z","lastTransitionTime":"2025-10-05T06:48:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:29 crc kubenswrapper[4846]: I1005 06:48:29.448331 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:29 crc kubenswrapper[4846]: I1005 06:48:29.448410 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:29 crc kubenswrapper[4846]: I1005 06:48:29.448437 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:29 crc kubenswrapper[4846]: I1005 06:48:29.448473 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:29 crc kubenswrapper[4846]: I1005 06:48:29.448513 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:29Z","lastTransitionTime":"2025-10-05T06:48:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:29 crc kubenswrapper[4846]: I1005 06:48:29.496811 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 05 06:48:29 crc kubenswrapper[4846]: I1005 06:48:29.496882 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 05 06:48:29 crc kubenswrapper[4846]: E1005 06:48:29.497036 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 05 06:48:29 crc kubenswrapper[4846]: E1005 06:48:29.497223 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 05 06:48:29 crc kubenswrapper[4846]: I1005 06:48:29.551654 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:29 crc kubenswrapper[4846]: I1005 06:48:29.551735 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:29 crc kubenswrapper[4846]: I1005 06:48:29.551756 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:29 crc kubenswrapper[4846]: I1005 06:48:29.551784 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:29 crc kubenswrapper[4846]: I1005 06:48:29.551804 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:29Z","lastTransitionTime":"2025-10-05T06:48:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:29 crc kubenswrapper[4846]: I1005 06:48:29.654974 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:29 crc kubenswrapper[4846]: I1005 06:48:29.655035 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:29 crc kubenswrapper[4846]: I1005 06:48:29.655053 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:29 crc kubenswrapper[4846]: I1005 06:48:29.655082 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:29 crc kubenswrapper[4846]: I1005 06:48:29.655101 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:29Z","lastTransitionTime":"2025-10-05T06:48:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:29 crc kubenswrapper[4846]: I1005 06:48:29.759214 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:29 crc kubenswrapper[4846]: I1005 06:48:29.759279 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:29 crc kubenswrapper[4846]: I1005 06:48:29.759296 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:29 crc kubenswrapper[4846]: I1005 06:48:29.759328 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:29 crc kubenswrapper[4846]: I1005 06:48:29.759348 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:29Z","lastTransitionTime":"2025-10-05T06:48:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:29 crc kubenswrapper[4846]: I1005 06:48:29.771506 4846 generic.go:334] "Generic (PLEG): container finished" podID="1dbb13ac-136c-4a1c-9229-b4e8ac4f9c71" containerID="1211571af66d8a94302b65feccf30768ba82b7cbad918e44dd405787dce484cb" exitCode=0 Oct 05 06:48:29 crc kubenswrapper[4846]: I1005 06:48:29.771571 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-sfmxd" event={"ID":"1dbb13ac-136c-4a1c-9229-b4e8ac4f9c71","Type":"ContainerDied","Data":"1211571af66d8a94302b65feccf30768ba82b7cbad918e44dd405787dce484cb"} Oct 05 06:48:29 crc kubenswrapper[4846]: I1005 06:48:29.780479 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-txzlk" event={"ID":"93a6ace4-8d64-44d7-9637-457d3af6543c","Type":"ContainerStarted","Data":"0df0ebcd87dc5af4b17736b43ec7b2395968b19ed16e775366af14e2b2d49c52"} Oct 05 06:48:29 crc kubenswrapper[4846]: I1005 06:48:29.796118 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e67d0e447a8a6a381259e25d25badc791194e1291f56b1cd77bd63989b20408\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dkhl7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://516f88f4990f5962d8b694a3d889259cc5508a309919870191c2e1331640c973\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dkhl7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-fscvf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:29Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:29 crc kubenswrapper[4846]: I1005 06:48:29.816060 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-w2sj8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32931fce-7374-4cab-a2e0-79d96a842274\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0df535d65d39de6c9a3dfb931bafc4469adc22b72f7a7903beda4fef8cd48138\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9rtxf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:27Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-w2sj8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:29Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:29 crc kubenswrapper[4846]: I1005 06:48:29.839111 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c9f812b-ecc8-4985-a670-53fa80fbb499\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2b5ee9781c01d6aec3ac778efef1067b6c3bfe4ed37431bdebe188fcde0165f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:47:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3fb4717692aa1fec0012b514ef8d18d1ca3707cfa4bb8248bb93841a0a40f11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:47:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0b6a7afe3094f1c92093c05c28344897a615d590c2fbb507c87e18a44c6c831\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://640ade2c733d7916ec213a201a1dd614d261c69eab654c80100efbc99a768995\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:47:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:29Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:29 crc kubenswrapper[4846]: I1005 06:48:29.858608 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:29Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:29 crc kubenswrapper[4846]: I1005 06:48:29.862699 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:29 crc kubenswrapper[4846]: I1005 06:48:29.862788 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:29 crc kubenswrapper[4846]: I1005 06:48:29.862817 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:29 crc kubenswrapper[4846]: I1005 06:48:29.862852 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:29 crc kubenswrapper[4846]: I1005 06:48:29.862882 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:29Z","lastTransitionTime":"2025-10-05T06:48:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:29 crc kubenswrapper[4846]: I1005 06:48:29.884965 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://347f900743b7bba3daf92f21c7eba0625e8c774f50057bc9472f0c82119325e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:29Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:29 crc kubenswrapper[4846]: I1005 06:48:29.904665 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56f1fe1f9f06f0249cf937a6963fda82467732dc056ca2b2189be6d8dd9d4a05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7212ce5f42c1fb55146abb190a557a504b43f16af6df84f90e1073567a329ec5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:29Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:29 crc kubenswrapper[4846]: I1005 06:48:29.928739 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-7wm5q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9bc7744f-d700-450f-8b8d-98140877fee9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90379e74c544fccca44f7acdc228ff378154390e211ac801b79f8d074012241e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h5p4q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-7wm5q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:29Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:29 crc kubenswrapper[4846]: I1005 06:48:29.954506 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ff10c31459b3b3724ee7b769bd6f72f9ed80cce5419b6adcdf69a3239b70756\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:29Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:29 crc kubenswrapper[4846]: I1005 06:48:29.968064 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:29 crc kubenswrapper[4846]: I1005 06:48:29.968130 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:29 crc kubenswrapper[4846]: I1005 06:48:29.968153 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:29 crc kubenswrapper[4846]: I1005 06:48:29.968211 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:29 crc kubenswrapper[4846]: I1005 06:48:29.968233 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:29Z","lastTransitionTime":"2025-10-05T06:48:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:29 crc kubenswrapper[4846]: I1005 06:48:29.972498 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:29Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:29 crc kubenswrapper[4846]: I1005 06:48:29.985129 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-r7ztb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a5119cd9-af70-4324-a955-978306ab0b20\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06231a87429ec292cb57ebd88cda7ad547bcb32e80343b776833621a11c28ad1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jfcj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-r7ztb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:29Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:30 crc kubenswrapper[4846]: I1005 06:48:30.009157 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"775b5702-a131-4d6c-852d-91bfa056cac5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0839c1765df2c68d2554e9c66d289505e559a9672caf8069feeff2c4fdfd2171\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15dbaf2445704da0987f256f81dff2ee95664988541fa48ce94898afeadab2e8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd3b3ab323c6d5153054aa04af2d91f7b9e8436ff1e2f75f3594b57831b2a416\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c52e076eddfd307110b1f6de00a169b81af8f2785af7becef2a820ce363a5861\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c52e076eddfd307110b1f6de00a169b81af8f2785af7becef2a820ce363a5861\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-05T06:48:14Z\\\",\\\"message\\\":\\\"le observer\\\\nW1005 06:48:14.337870 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1005 06:48:14.338086 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1005 06:48:14.339389 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1467752215/tls.crt::/tmp/serving-cert-1467752215/tls.key\\\\\\\"\\\\nI1005 06:48:14.864345 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1005 06:48:14.868087 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1005 06:48:14.868114 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1005 06:48:14.868140 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1005 06:48:14.868147 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1005 06:48:14.877354 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1005 06:48:14.877412 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1005 06:48:14.877456 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1005 06:48:14.877471 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1005 06:48:14.877490 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1005 06:48:14.877501 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1005 06:48:14.877507 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1005 06:48:14.877515 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1005 06:48:14.880326 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:13Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4a091716ff935adb42ae2e88048619fd347bd174f64e5f9c97a01f81690e57f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:01Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dc5411b642f99b13d514d2f21defad0c394aeeb82a966f72300076617e2586c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5dc5411b642f99b13d514d2f21defad0c394aeeb82a966f72300076617e2586c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:47:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:47:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:47:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:30Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:30 crc kubenswrapper[4846]: I1005 06:48:30.031933 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-sfmxd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1dbb13ac-136c-4a1c-9229-b4e8ac4f9c71\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://226d977b596d07aeda0486452e2a8605949ed4d2453b21f56174937c395d6f36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://226d977b596d07aeda0486452e2a8605949ed4d2453b21f56174937c395d6f36\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://99c07a4bdb3059d3a99ffdaf794d668d52cffd9e5e3bb8945204dd6d1f8c8328\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99c07a4bdb3059d3a99ffdaf794d668d52cffd9e5e3bb8945204dd6d1f8c8328\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27095272bf4f31d63a78a436636de0aae12f7103597da944f24ed7a74329ee3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27095272bf4f31d63a78a436636de0aae12f7103597da944f24ed7a74329ee3f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1211571af66d8a94302b65feccf30768ba82b7cbad918e44dd405787dce484cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1211571af66d8a94302b65feccf30768ba82b7cbad918e44dd405787dce484cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-sfmxd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:30Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:30 crc kubenswrapper[4846]: I1005 06:48:30.059656 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-txzlk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"93a6ace4-8d64-44d7-9637-457d3af6543c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff3c6545ff3d2436da27bbd223b950b30b0c69a3e7f3034cb0e61380cb38d338\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ff3c6545ff3d2436da27bbd223b950b30b0c69a3e7f3034cb0e61380cb38d338\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:24Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-txzlk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:30Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:30 crc kubenswrapper[4846]: I1005 06:48:30.072588 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:30 crc kubenswrapper[4846]: I1005 06:48:30.072683 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:30 crc kubenswrapper[4846]: I1005 06:48:30.072713 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:30 crc kubenswrapper[4846]: I1005 06:48:30.072748 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:30 crc kubenswrapper[4846]: I1005 06:48:30.072772 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:30Z","lastTransitionTime":"2025-10-05T06:48:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:30 crc kubenswrapper[4846]: I1005 06:48:30.092596 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c482bb97-fd50-40cf-96ea-f806c3f01125\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bfe45d30e4fa141541aca1d46bc5cc1da0a7b4a5eb1fc9bb2caa5b7fdcb0b264\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://496a4782faa30a9a20ff63038141600cd635f12902d1cf17989cd067fcdfe929\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://281be077869a6aa1ca815770871a14413d4a97e97aeb550fd5cd215cba88b21f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6771c7215d2f5dd5838d17888375a4768b9d1755fc2e21bda1bbfbb2eb1dbb95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c17a711c2c7473b076a3944ffe7c700808d05cba582fbd55c5c2ac8ff73838a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44a9f740f0a06c8bf928e2a9d1d5979217558c31e73a333631baccebf00fd5d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://44a9f740f0a06c8bf928e2a9d1d5979217558c31e73a333631baccebf00fd5d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:47:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:47:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://48ca0ba6ca6092bfcb5b153220c054f02cc36fb22707c42e021f06d329e18e2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48ca0ba6ca6092bfcb5b153220c054f02cc36fb22707c42e021f06d329e18e2e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:00Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4006d0cea128675c8885d879171204abffc5e11ea0664750a12b4bc7344ccb2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4006d0cea128675c8885d879171204abffc5e11ea0664750a12b4bc7344ccb2e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:47:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:30Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:30 crc kubenswrapper[4846]: I1005 06:48:30.114272 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:30Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:30 crc kubenswrapper[4846]: I1005 06:48:30.176427 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:30 crc kubenswrapper[4846]: I1005 06:48:30.176509 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:30 crc kubenswrapper[4846]: I1005 06:48:30.176527 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:30 crc kubenswrapper[4846]: I1005 06:48:30.176556 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:30 crc kubenswrapper[4846]: I1005 06:48:30.176574 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:30Z","lastTransitionTime":"2025-10-05T06:48:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:30 crc kubenswrapper[4846]: I1005 06:48:30.283102 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:30 crc kubenswrapper[4846]: I1005 06:48:30.284106 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:30 crc kubenswrapper[4846]: I1005 06:48:30.284322 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:30 crc kubenswrapper[4846]: I1005 06:48:30.284489 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:30 crc kubenswrapper[4846]: I1005 06:48:30.284681 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:30Z","lastTransitionTime":"2025-10-05T06:48:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:30 crc kubenswrapper[4846]: I1005 06:48:30.388264 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:30 crc kubenswrapper[4846]: I1005 06:48:30.388776 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:30 crc kubenswrapper[4846]: I1005 06:48:30.389276 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:30 crc kubenswrapper[4846]: I1005 06:48:30.389644 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:30 crc kubenswrapper[4846]: I1005 06:48:30.389981 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:30Z","lastTransitionTime":"2025-10-05T06:48:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:30 crc kubenswrapper[4846]: I1005 06:48:30.493886 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:30 crc kubenswrapper[4846]: I1005 06:48:30.493932 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:30 crc kubenswrapper[4846]: I1005 06:48:30.493944 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:30 crc kubenswrapper[4846]: I1005 06:48:30.493962 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:30 crc kubenswrapper[4846]: I1005 06:48:30.493975 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:30Z","lastTransitionTime":"2025-10-05T06:48:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:30 crc kubenswrapper[4846]: I1005 06:48:30.497230 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 05 06:48:30 crc kubenswrapper[4846]: E1005 06:48:30.497361 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 05 06:48:30 crc kubenswrapper[4846]: I1005 06:48:30.598267 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:30 crc kubenswrapper[4846]: I1005 06:48:30.598324 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:30 crc kubenswrapper[4846]: I1005 06:48:30.598337 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:30 crc kubenswrapper[4846]: I1005 06:48:30.598357 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:30 crc kubenswrapper[4846]: I1005 06:48:30.598375 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:30Z","lastTransitionTime":"2025-10-05T06:48:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:30 crc kubenswrapper[4846]: I1005 06:48:30.702508 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:30 crc kubenswrapper[4846]: I1005 06:48:30.702746 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:30 crc kubenswrapper[4846]: I1005 06:48:30.702837 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:30 crc kubenswrapper[4846]: I1005 06:48:30.702948 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:30 crc kubenswrapper[4846]: I1005 06:48:30.703030 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:30Z","lastTransitionTime":"2025-10-05T06:48:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:30 crc kubenswrapper[4846]: I1005 06:48:30.789321 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-sfmxd" event={"ID":"1dbb13ac-136c-4a1c-9229-b4e8ac4f9c71","Type":"ContainerStarted","Data":"27631f90d528900fa180c15cfce80e96d4cdba07743eb6d847e2e822d9917218"} Oct 05 06:48:30 crc kubenswrapper[4846]: I1005 06:48:30.806448 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:30 crc kubenswrapper[4846]: I1005 06:48:30.806521 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:30 crc kubenswrapper[4846]: I1005 06:48:30.806540 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:30 crc kubenswrapper[4846]: I1005 06:48:30.806566 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:30 crc kubenswrapper[4846]: I1005 06:48:30.806585 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:30Z","lastTransitionTime":"2025-10-05T06:48:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:30 crc kubenswrapper[4846]: I1005 06:48:30.811397 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ff10c31459b3b3724ee7b769bd6f72f9ed80cce5419b6adcdf69a3239b70756\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:30Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:30 crc kubenswrapper[4846]: I1005 06:48:30.830046 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://347f900743b7bba3daf92f21c7eba0625e8c774f50057bc9472f0c82119325e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:30Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:30 crc kubenswrapper[4846]: I1005 06:48:30.848079 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56f1fe1f9f06f0249cf937a6963fda82467732dc056ca2b2189be6d8dd9d4a05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7212ce5f42c1fb55146abb190a557a504b43f16af6df84f90e1073567a329ec5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:30Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:30 crc kubenswrapper[4846]: I1005 06:48:30.890814 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-7wm5q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9bc7744f-d700-450f-8b8d-98140877fee9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90379e74c544fccca44f7acdc228ff378154390e211ac801b79f8d074012241e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h5p4q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-7wm5q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:30Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:30 crc kubenswrapper[4846]: I1005 06:48:30.909980 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:30 crc kubenswrapper[4846]: I1005 06:48:30.910042 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:30 crc kubenswrapper[4846]: I1005 06:48:30.910058 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:30 crc kubenswrapper[4846]: I1005 06:48:30.910084 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:30 crc kubenswrapper[4846]: I1005 06:48:30.910107 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:30Z","lastTransitionTime":"2025-10-05T06:48:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:30 crc kubenswrapper[4846]: I1005 06:48:30.938483 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"775b5702-a131-4d6c-852d-91bfa056cac5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0839c1765df2c68d2554e9c66d289505e559a9672caf8069feeff2c4fdfd2171\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15dbaf2445704da0987f256f81dff2ee95664988541fa48ce94898afeadab2e8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd3b3ab323c6d5153054aa04af2d91f7b9e8436ff1e2f75f3594b57831b2a416\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c52e076eddfd307110b1f6de00a169b81af8f2785af7becef2a820ce363a5861\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c52e076eddfd307110b1f6de00a169b81af8f2785af7becef2a820ce363a5861\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-05T06:48:14Z\\\",\\\"message\\\":\\\"le observer\\\\nW1005 06:48:14.337870 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1005 06:48:14.338086 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1005 06:48:14.339389 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1467752215/tls.crt::/tmp/serving-cert-1467752215/tls.key\\\\\\\"\\\\nI1005 06:48:14.864345 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1005 06:48:14.868087 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1005 06:48:14.868114 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1005 06:48:14.868140 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1005 06:48:14.868147 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1005 06:48:14.877354 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1005 06:48:14.877412 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1005 06:48:14.877456 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1005 06:48:14.877471 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1005 06:48:14.877490 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1005 06:48:14.877501 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1005 06:48:14.877507 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1005 06:48:14.877515 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1005 06:48:14.880326 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:13Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4a091716ff935adb42ae2e88048619fd347bd174f64e5f9c97a01f81690e57f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:01Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dc5411b642f99b13d514d2f21defad0c394aeeb82a966f72300076617e2586c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5dc5411b642f99b13d514d2f21defad0c394aeeb82a966f72300076617e2586c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:47:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:47:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:47:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:30Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:30 crc kubenswrapper[4846]: I1005 06:48:30.969833 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:30Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:30 crc kubenswrapper[4846]: I1005 06:48:30.982894 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-r7ztb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a5119cd9-af70-4324-a955-978306ab0b20\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06231a87429ec292cb57ebd88cda7ad547bcb32e80343b776833621a11c28ad1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jfcj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-r7ztb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:30Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:30 crc kubenswrapper[4846]: I1005 06:48:30.997707 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:30Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:31 crc kubenswrapper[4846]: I1005 06:48:31.012518 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-sfmxd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1dbb13ac-136c-4a1c-9229-b4e8ac4f9c71\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://226d977b596d07aeda0486452e2a8605949ed4d2453b21f56174937c395d6f36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://226d977b596d07aeda0486452e2a8605949ed4d2453b21f56174937c395d6f36\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://99c07a4bdb3059d3a99ffdaf794d668d52cffd9e5e3bb8945204dd6d1f8c8328\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99c07a4bdb3059d3a99ffdaf794d668d52cffd9e5e3bb8945204dd6d1f8c8328\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27095272bf4f31d63a78a436636de0aae12f7103597da944f24ed7a74329ee3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27095272bf4f31d63a78a436636de0aae12f7103597da944f24ed7a74329ee3f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1211571af66d8a94302b65feccf30768ba82b7cbad918e44dd405787dce484cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1211571af66d8a94302b65feccf30768ba82b7cbad918e44dd405787dce484cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27631f90d528900fa180c15cfce80e96d4cdba07743eb6d847e2e822d9917218\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-sfmxd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:31Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:31 crc kubenswrapper[4846]: I1005 06:48:31.013463 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:31 crc kubenswrapper[4846]: I1005 06:48:31.013511 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:31 crc kubenswrapper[4846]: I1005 06:48:31.013525 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:31 crc kubenswrapper[4846]: I1005 06:48:31.013569 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:31 crc kubenswrapper[4846]: I1005 06:48:31.013581 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:31Z","lastTransitionTime":"2025-10-05T06:48:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:31 crc kubenswrapper[4846]: I1005 06:48:31.033386 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-txzlk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"93a6ace4-8d64-44d7-9637-457d3af6543c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff3c6545ff3d2436da27bbd223b950b30b0c69a3e7f3034cb0e61380cb38d338\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ff3c6545ff3d2436da27bbd223b950b30b0c69a3e7f3034cb0e61380cb38d338\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:24Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-txzlk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:31Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:31 crc kubenswrapper[4846]: I1005 06:48:31.066294 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c482bb97-fd50-40cf-96ea-f806c3f01125\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bfe45d30e4fa141541aca1d46bc5cc1da0a7b4a5eb1fc9bb2caa5b7fdcb0b264\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://496a4782faa30a9a20ff63038141600cd635f12902d1cf17989cd067fcdfe929\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://281be077869a6aa1ca815770871a14413d4a97e97aeb550fd5cd215cba88b21f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6771c7215d2f5dd5838d17888375a4768b9d1755fc2e21bda1bbfbb2eb1dbb95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c17a711c2c7473b076a3944ffe7c700808d05cba582fbd55c5c2ac8ff73838a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44a9f740f0a06c8bf928e2a9d1d5979217558c31e73a333631baccebf00fd5d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://44a9f740f0a06c8bf928e2a9d1d5979217558c31e73a333631baccebf00fd5d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:47:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:47:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://48ca0ba6ca6092bfcb5b153220c054f02cc36fb22707c42e021f06d329e18e2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48ca0ba6ca6092bfcb5b153220c054f02cc36fb22707c42e021f06d329e18e2e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:00Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4006d0cea128675c8885d879171204abffc5e11ea0664750a12b4bc7344ccb2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4006d0cea128675c8885d879171204abffc5e11ea0664750a12b4bc7344ccb2e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:47:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:31Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:31 crc kubenswrapper[4846]: I1005 06:48:31.083105 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:31Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:31 crc kubenswrapper[4846]: I1005 06:48:31.100670 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e67d0e447a8a6a381259e25d25badc791194e1291f56b1cd77bd63989b20408\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dkhl7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://516f88f4990f5962d8b694a3d889259cc5508a309919870191c2e1331640c973\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dkhl7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-fscvf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:31Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:31 crc kubenswrapper[4846]: I1005 06:48:31.117510 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:31 crc kubenswrapper[4846]: I1005 06:48:31.118097 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:31 crc kubenswrapper[4846]: I1005 06:48:31.118119 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:31 crc kubenswrapper[4846]: I1005 06:48:31.118150 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:31 crc kubenswrapper[4846]: I1005 06:48:31.118238 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:31Z","lastTransitionTime":"2025-10-05T06:48:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:31 crc kubenswrapper[4846]: I1005 06:48:31.121119 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-w2sj8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32931fce-7374-4cab-a2e0-79d96a842274\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0df535d65d39de6c9a3dfb931bafc4469adc22b72f7a7903beda4fef8cd48138\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9rtxf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:27Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-w2sj8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:31Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:31 crc kubenswrapper[4846]: I1005 06:48:31.135644 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c9f812b-ecc8-4985-a670-53fa80fbb499\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2b5ee9781c01d6aec3ac778efef1067b6c3bfe4ed37431bdebe188fcde0165f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:47:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3fb4717692aa1fec0012b514ef8d18d1ca3707cfa4bb8248bb93841a0a40f11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:47:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0b6a7afe3094f1c92093c05c28344897a615d590c2fbb507c87e18a44c6c831\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://640ade2c733d7916ec213a201a1dd614d261c69eab654c80100efbc99a768995\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:47:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:31Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:31 crc kubenswrapper[4846]: I1005 06:48:31.221055 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:31 crc kubenswrapper[4846]: I1005 06:48:31.221112 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:31 crc kubenswrapper[4846]: I1005 06:48:31.221127 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:31 crc kubenswrapper[4846]: I1005 06:48:31.221148 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:31 crc kubenswrapper[4846]: I1005 06:48:31.221162 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:31Z","lastTransitionTime":"2025-10-05T06:48:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:31 crc kubenswrapper[4846]: I1005 06:48:31.324430 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:31 crc kubenswrapper[4846]: I1005 06:48:31.324545 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:31 crc kubenswrapper[4846]: I1005 06:48:31.324566 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:31 crc kubenswrapper[4846]: I1005 06:48:31.324603 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:31 crc kubenswrapper[4846]: I1005 06:48:31.324623 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:31Z","lastTransitionTime":"2025-10-05T06:48:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:31 crc kubenswrapper[4846]: I1005 06:48:31.428391 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:31 crc kubenswrapper[4846]: I1005 06:48:31.428480 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:31 crc kubenswrapper[4846]: I1005 06:48:31.428512 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:31 crc kubenswrapper[4846]: I1005 06:48:31.428557 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:31 crc kubenswrapper[4846]: I1005 06:48:31.428583 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:31Z","lastTransitionTime":"2025-10-05T06:48:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:31 crc kubenswrapper[4846]: I1005 06:48:31.497436 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 05 06:48:31 crc kubenswrapper[4846]: E1005 06:48:31.497684 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 05 06:48:31 crc kubenswrapper[4846]: I1005 06:48:31.497893 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 05 06:48:31 crc kubenswrapper[4846]: E1005 06:48:31.498465 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 05 06:48:31 crc kubenswrapper[4846]: I1005 06:48:31.498945 4846 scope.go:117] "RemoveContainer" containerID="c52e076eddfd307110b1f6de00a169b81af8f2785af7becef2a820ce363a5861" Oct 05 06:48:31 crc kubenswrapper[4846]: I1005 06:48:31.533449 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:31 crc kubenswrapper[4846]: I1005 06:48:31.533547 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:31 crc kubenswrapper[4846]: I1005 06:48:31.533566 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:31 crc kubenswrapper[4846]: I1005 06:48:31.533589 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:31 crc kubenswrapper[4846]: I1005 06:48:31.533605 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:31Z","lastTransitionTime":"2025-10-05T06:48:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:31 crc kubenswrapper[4846]: I1005 06:48:31.637419 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:31 crc kubenswrapper[4846]: I1005 06:48:31.637486 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:31 crc kubenswrapper[4846]: I1005 06:48:31.637500 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:31 crc kubenswrapper[4846]: I1005 06:48:31.637522 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:31 crc kubenswrapper[4846]: I1005 06:48:31.637538 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:31Z","lastTransitionTime":"2025-10-05T06:48:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:31 crc kubenswrapper[4846]: I1005 06:48:31.740769 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:31 crc kubenswrapper[4846]: I1005 06:48:31.740831 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:31 crc kubenswrapper[4846]: I1005 06:48:31.740846 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:31 crc kubenswrapper[4846]: I1005 06:48:31.740869 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:31 crc kubenswrapper[4846]: I1005 06:48:31.740886 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:31Z","lastTransitionTime":"2025-10-05T06:48:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:31 crc kubenswrapper[4846]: I1005 06:48:31.798265 4846 generic.go:334] "Generic (PLEG): container finished" podID="1dbb13ac-136c-4a1c-9229-b4e8ac4f9c71" containerID="27631f90d528900fa180c15cfce80e96d4cdba07743eb6d847e2e822d9917218" exitCode=0 Oct 05 06:48:31 crc kubenswrapper[4846]: I1005 06:48:31.798329 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-sfmxd" event={"ID":"1dbb13ac-136c-4a1c-9229-b4e8ac4f9c71","Type":"ContainerDied","Data":"27631f90d528900fa180c15cfce80e96d4cdba07743eb6d847e2e822d9917218"} Oct 05 06:48:31 crc kubenswrapper[4846]: I1005 06:48:31.806771 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-txzlk" event={"ID":"93a6ace4-8d64-44d7-9637-457d3af6543c","Type":"ContainerStarted","Data":"c212f5d230fae82a701afeceb39a848a9c91a3354908d88b062b7e1a726240cb"} Oct 05 06:48:31 crc kubenswrapper[4846]: I1005 06:48:31.807698 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-txzlk" Oct 05 06:48:31 crc kubenswrapper[4846]: I1005 06:48:31.807792 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-txzlk" Oct 05 06:48:31 crc kubenswrapper[4846]: I1005 06:48:31.833353 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"775b5702-a131-4d6c-852d-91bfa056cac5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0839c1765df2c68d2554e9c66d289505e559a9672caf8069feeff2c4fdfd2171\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15dbaf2445704da0987f256f81dff2ee95664988541fa48ce94898afeadab2e8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd3b3ab323c6d5153054aa04af2d91f7b9e8436ff1e2f75f3594b57831b2a416\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c52e076eddfd307110b1f6de00a169b81af8f2785af7becef2a820ce363a5861\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c52e076eddfd307110b1f6de00a169b81af8f2785af7becef2a820ce363a5861\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-05T06:48:14Z\\\",\\\"message\\\":\\\"le observer\\\\nW1005 06:48:14.337870 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1005 06:48:14.338086 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1005 06:48:14.339389 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1467752215/tls.crt::/tmp/serving-cert-1467752215/tls.key\\\\\\\"\\\\nI1005 06:48:14.864345 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1005 06:48:14.868087 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1005 06:48:14.868114 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1005 06:48:14.868140 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1005 06:48:14.868147 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1005 06:48:14.877354 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1005 06:48:14.877412 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1005 06:48:14.877456 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1005 06:48:14.877471 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1005 06:48:14.877490 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1005 06:48:14.877501 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1005 06:48:14.877507 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1005 06:48:14.877515 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1005 06:48:14.880326 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:13Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4a091716ff935adb42ae2e88048619fd347bd174f64e5f9c97a01f81690e57f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:01Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dc5411b642f99b13d514d2f21defad0c394aeeb82a966f72300076617e2586c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5dc5411b642f99b13d514d2f21defad0c394aeeb82a966f72300076617e2586c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:47:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:47:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:47:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:31Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:31 crc kubenswrapper[4846]: I1005 06:48:31.853972 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:31 crc kubenswrapper[4846]: I1005 06:48:31.854025 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:31 crc kubenswrapper[4846]: I1005 06:48:31.854044 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:31 crc kubenswrapper[4846]: I1005 06:48:31.854070 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:31 crc kubenswrapper[4846]: I1005 06:48:31.854090 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:31Z","lastTransitionTime":"2025-10-05T06:48:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:31 crc kubenswrapper[4846]: I1005 06:48:31.856945 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:31Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:31 crc kubenswrapper[4846]: I1005 06:48:31.869027 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-txzlk" Oct 05 06:48:31 crc kubenswrapper[4846]: I1005 06:48:31.869142 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-txzlk" Oct 05 06:48:31 crc kubenswrapper[4846]: I1005 06:48:31.879357 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-r7ztb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a5119cd9-af70-4324-a955-978306ab0b20\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06231a87429ec292cb57ebd88cda7ad547bcb32e80343b776833621a11c28ad1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jfcj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-r7ztb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:31Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:31 crc kubenswrapper[4846]: I1005 06:48:31.911128 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c482bb97-fd50-40cf-96ea-f806c3f01125\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bfe45d30e4fa141541aca1d46bc5cc1da0a7b4a5eb1fc9bb2caa5b7fdcb0b264\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://496a4782faa30a9a20ff63038141600cd635f12902d1cf17989cd067fcdfe929\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://281be077869a6aa1ca815770871a14413d4a97e97aeb550fd5cd215cba88b21f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6771c7215d2f5dd5838d17888375a4768b9d1755fc2e21bda1bbfbb2eb1dbb95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c17a711c2c7473b076a3944ffe7c700808d05cba582fbd55c5c2ac8ff73838a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44a9f740f0a06c8bf928e2a9d1d5979217558c31e73a333631baccebf00fd5d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://44a9f740f0a06c8bf928e2a9d1d5979217558c31e73a333631baccebf00fd5d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:47:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:47:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://48ca0ba6ca6092bfcb5b153220c054f02cc36fb22707c42e021f06d329e18e2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48ca0ba6ca6092bfcb5b153220c054f02cc36fb22707c42e021f06d329e18e2e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:00Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4006d0cea128675c8885d879171204abffc5e11ea0664750a12b4bc7344ccb2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4006d0cea128675c8885d879171204abffc5e11ea0664750a12b4bc7344ccb2e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:47:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:31Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:31 crc kubenswrapper[4846]: I1005 06:48:31.931034 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:31Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:31 crc kubenswrapper[4846]: I1005 06:48:31.953101 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-sfmxd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1dbb13ac-136c-4a1c-9229-b4e8ac4f9c71\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://226d977b596d07aeda0486452e2a8605949ed4d2453b21f56174937c395d6f36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://226d977b596d07aeda0486452e2a8605949ed4d2453b21f56174937c395d6f36\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://99c07a4bdb3059d3a99ffdaf794d668d52cffd9e5e3bb8945204dd6d1f8c8328\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99c07a4bdb3059d3a99ffdaf794d668d52cffd9e5e3bb8945204dd6d1f8c8328\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27095272bf4f31d63a78a436636de0aae12f7103597da944f24ed7a74329ee3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27095272bf4f31d63a78a436636de0aae12f7103597da944f24ed7a74329ee3f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1211571af66d8a94302b65feccf30768ba82b7cbad918e44dd405787dce484cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1211571af66d8a94302b65feccf30768ba82b7cbad918e44dd405787dce484cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27631f90d528900fa180c15cfce80e96d4cdba07743eb6d847e2e822d9917218\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27631f90d528900fa180c15cfce80e96d4cdba07743eb6d847e2e822d9917218\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-sfmxd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:31Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:31 crc kubenswrapper[4846]: I1005 06:48:31.959327 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:31 crc kubenswrapper[4846]: I1005 06:48:31.959418 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:31 crc kubenswrapper[4846]: I1005 06:48:31.959454 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:31 crc kubenswrapper[4846]: I1005 06:48:31.959493 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:31 crc kubenswrapper[4846]: I1005 06:48:31.959520 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:31Z","lastTransitionTime":"2025-10-05T06:48:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:31 crc kubenswrapper[4846]: I1005 06:48:31.985668 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-txzlk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"93a6ace4-8d64-44d7-9637-457d3af6543c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff3c6545ff3d2436da27bbd223b950b30b0c69a3e7f3034cb0e61380cb38d338\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ff3c6545ff3d2436da27bbd223b950b30b0c69a3e7f3034cb0e61380cb38d338\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:24Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-txzlk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:31Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:32 crc kubenswrapper[4846]: I1005 06:48:32.008888 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c9f812b-ecc8-4985-a670-53fa80fbb499\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2b5ee9781c01d6aec3ac778efef1067b6c3bfe4ed37431bdebe188fcde0165f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:47:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3fb4717692aa1fec0012b514ef8d18d1ca3707cfa4bb8248bb93841a0a40f11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:47:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0b6a7afe3094f1c92093c05c28344897a615d590c2fbb507c87e18a44c6c831\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://640ade2c733d7916ec213a201a1dd614d261c69eab654c80100efbc99a768995\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:47:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:32Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:32 crc kubenswrapper[4846]: I1005 06:48:32.027237 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:32Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:32 crc kubenswrapper[4846]: I1005 06:48:32.046755 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e67d0e447a8a6a381259e25d25badc791194e1291f56b1cd77bd63989b20408\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dkhl7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://516f88f4990f5962d8b694a3d889259cc5508a309919870191c2e1331640c973\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dkhl7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-fscvf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:32Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:32 crc kubenswrapper[4846]: I1005 06:48:32.062073 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-w2sj8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32931fce-7374-4cab-a2e0-79d96a842274\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0df535d65d39de6c9a3dfb931bafc4469adc22b72f7a7903beda4fef8cd48138\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9rtxf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:27Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-w2sj8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:32Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:32 crc kubenswrapper[4846]: I1005 06:48:32.064027 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:32 crc kubenswrapper[4846]: I1005 06:48:32.064083 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:32 crc kubenswrapper[4846]: I1005 06:48:32.064103 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:32 crc kubenswrapper[4846]: I1005 06:48:32.064137 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:32 crc kubenswrapper[4846]: I1005 06:48:32.064158 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:32Z","lastTransitionTime":"2025-10-05T06:48:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:32 crc kubenswrapper[4846]: I1005 06:48:32.080162 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ff10c31459b3b3724ee7b769bd6f72f9ed80cce5419b6adcdf69a3239b70756\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:32Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:32 crc kubenswrapper[4846]: I1005 06:48:32.097253 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://347f900743b7bba3daf92f21c7eba0625e8c774f50057bc9472f0c82119325e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:32Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:32 crc kubenswrapper[4846]: I1005 06:48:32.115169 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56f1fe1f9f06f0249cf937a6963fda82467732dc056ca2b2189be6d8dd9d4a05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7212ce5f42c1fb55146abb190a557a504b43f16af6df84f90e1073567a329ec5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:32Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:32 crc kubenswrapper[4846]: I1005 06:48:32.136933 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-7wm5q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9bc7744f-d700-450f-8b8d-98140877fee9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90379e74c544fccca44f7acdc228ff378154390e211ac801b79f8d074012241e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h5p4q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-7wm5q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:32Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:32 crc kubenswrapper[4846]: I1005 06:48:32.154347 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://347f900743b7bba3daf92f21c7eba0625e8c774f50057bc9472f0c82119325e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:32Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:32 crc kubenswrapper[4846]: I1005 06:48:32.168288 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:32 crc kubenswrapper[4846]: I1005 06:48:32.168334 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:32 crc kubenswrapper[4846]: I1005 06:48:32.168345 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:32 crc kubenswrapper[4846]: I1005 06:48:32.168363 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:32 crc kubenswrapper[4846]: I1005 06:48:32.168375 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:32Z","lastTransitionTime":"2025-10-05T06:48:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:32 crc kubenswrapper[4846]: I1005 06:48:32.170563 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56f1fe1f9f06f0249cf937a6963fda82467732dc056ca2b2189be6d8dd9d4a05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7212ce5f42c1fb55146abb190a557a504b43f16af6df84f90e1073567a329ec5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:32Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:32 crc kubenswrapper[4846]: I1005 06:48:32.186465 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-7wm5q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9bc7744f-d700-450f-8b8d-98140877fee9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90379e74c544fccca44f7acdc228ff378154390e211ac801b79f8d074012241e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h5p4q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-7wm5q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:32Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:32 crc kubenswrapper[4846]: I1005 06:48:32.202831 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ff10c31459b3b3724ee7b769bd6f72f9ed80cce5419b6adcdf69a3239b70756\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:32Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:32 crc kubenswrapper[4846]: I1005 06:48:32.218672 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:32Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:32 crc kubenswrapper[4846]: I1005 06:48:32.235273 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-r7ztb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a5119cd9-af70-4324-a955-978306ab0b20\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06231a87429ec292cb57ebd88cda7ad547bcb32e80343b776833621a11c28ad1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jfcj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-r7ztb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:32Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:32 crc kubenswrapper[4846]: I1005 06:48:32.256232 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"775b5702-a131-4d6c-852d-91bfa056cac5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0839c1765df2c68d2554e9c66d289505e559a9672caf8069feeff2c4fdfd2171\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15dbaf2445704da0987f256f81dff2ee95664988541fa48ce94898afeadab2e8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd3b3ab323c6d5153054aa04af2d91f7b9e8436ff1e2f75f3594b57831b2a416\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c52e076eddfd307110b1f6de00a169b81af8f2785af7becef2a820ce363a5861\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c52e076eddfd307110b1f6de00a169b81af8f2785af7becef2a820ce363a5861\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-05T06:48:14Z\\\",\\\"message\\\":\\\"le observer\\\\nW1005 06:48:14.337870 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1005 06:48:14.338086 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1005 06:48:14.339389 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1467752215/tls.crt::/tmp/serving-cert-1467752215/tls.key\\\\\\\"\\\\nI1005 06:48:14.864345 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1005 06:48:14.868087 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1005 06:48:14.868114 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1005 06:48:14.868140 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1005 06:48:14.868147 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1005 06:48:14.877354 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1005 06:48:14.877412 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1005 06:48:14.877456 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1005 06:48:14.877471 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1005 06:48:14.877490 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1005 06:48:14.877501 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1005 06:48:14.877507 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1005 06:48:14.877515 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1005 06:48:14.880326 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:13Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4a091716ff935adb42ae2e88048619fd347bd174f64e5f9c97a01f81690e57f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:01Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dc5411b642f99b13d514d2f21defad0c394aeeb82a966f72300076617e2586c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5dc5411b642f99b13d514d2f21defad0c394aeeb82a966f72300076617e2586c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:47:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:47:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:47:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:32Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:32 crc kubenswrapper[4846]: I1005 06:48:32.271878 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:32 crc kubenswrapper[4846]: I1005 06:48:32.271944 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:32 crc kubenswrapper[4846]: I1005 06:48:32.271957 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:32 crc kubenswrapper[4846]: I1005 06:48:32.271978 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:32 crc kubenswrapper[4846]: I1005 06:48:32.271997 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:32Z","lastTransitionTime":"2025-10-05T06:48:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:32 crc kubenswrapper[4846]: I1005 06:48:32.284125 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-sfmxd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1dbb13ac-136c-4a1c-9229-b4e8ac4f9c71\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://226d977b596d07aeda0486452e2a8605949ed4d2453b21f56174937c395d6f36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://226d977b596d07aeda0486452e2a8605949ed4d2453b21f56174937c395d6f36\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://99c07a4bdb3059d3a99ffdaf794d668d52cffd9e5e3bb8945204dd6d1f8c8328\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99c07a4bdb3059d3a99ffdaf794d668d52cffd9e5e3bb8945204dd6d1f8c8328\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27095272bf4f31d63a78a436636de0aae12f7103597da944f24ed7a74329ee3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27095272bf4f31d63a78a436636de0aae12f7103597da944f24ed7a74329ee3f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1211571af66d8a94302b65feccf30768ba82b7cbad918e44dd405787dce484cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1211571af66d8a94302b65feccf30768ba82b7cbad918e44dd405787dce484cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27631f90d528900fa180c15cfce80e96d4cdba07743eb6d847e2e822d9917218\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27631f90d528900fa180c15cfce80e96d4cdba07743eb6d847e2e822d9917218\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-sfmxd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:32Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:32 crc kubenswrapper[4846]: I1005 06:48:32.315937 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-txzlk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"93a6ace4-8d64-44d7-9637-457d3af6543c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ad1940d07e37217089642478ece9ccdaca74784738a400929a9f5359d158c01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a6f6b2c32dd3da5f9fd5e6971c1773e81a9a24184b7aa5b543cec879001ae53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e9b0219ecbc7e60905d6de3e67a2c8dc9af1017b4aed9538ee0c58c861f2438\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af5e5a4da9ad7371d5a24633e8268f12e3a7d419a8c93b11ddbcc006d8339c6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8da85222775abc7a772f6a8ff8255f1a0d5b15807ba177a7e3e2b2d01a4f24b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1409651de8c897804f87be03d016c138f8acf27e662dce265a93e15c1848adcf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c212f5d230fae82a701afeceb39a848a9c91a3354908d88b062b7e1a726240cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0df0ebcd87dc5af4b17736b43ec7b2395968b19ed16e775366af14e2b2d49c52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff3c6545ff3d2436da27bbd223b950b30b0c69a3e7f3034cb0e61380cb38d338\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ff3c6545ff3d2436da27bbd223b950b30b0c69a3e7f3034cb0e61380cb38d338\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:24Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-txzlk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:32Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:32 crc kubenswrapper[4846]: I1005 06:48:32.355666 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c482bb97-fd50-40cf-96ea-f806c3f01125\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bfe45d30e4fa141541aca1d46bc5cc1da0a7b4a5eb1fc9bb2caa5b7fdcb0b264\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://496a4782faa30a9a20ff63038141600cd635f12902d1cf17989cd067fcdfe929\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://281be077869a6aa1ca815770871a14413d4a97e97aeb550fd5cd215cba88b21f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6771c7215d2f5dd5838d17888375a4768b9d1755fc2e21bda1bbfbb2eb1dbb95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c17a711c2c7473b076a3944ffe7c700808d05cba582fbd55c5c2ac8ff73838a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44a9f740f0a06c8bf928e2a9d1d5979217558c31e73a333631baccebf00fd5d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://44a9f740f0a06c8bf928e2a9d1d5979217558c31e73a333631baccebf00fd5d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:47:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:47:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://48ca0ba6ca6092bfcb5b153220c054f02cc36fb22707c42e021f06d329e18e2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48ca0ba6ca6092bfcb5b153220c054f02cc36fb22707c42e021f06d329e18e2e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:00Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4006d0cea128675c8885d879171204abffc5e11ea0664750a12b4bc7344ccb2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4006d0cea128675c8885d879171204abffc5e11ea0664750a12b4bc7344ccb2e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:47:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:32Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:32 crc kubenswrapper[4846]: I1005 06:48:32.374528 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:32Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:32 crc kubenswrapper[4846]: I1005 06:48:32.375401 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:32 crc kubenswrapper[4846]: I1005 06:48:32.375502 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:32 crc kubenswrapper[4846]: I1005 06:48:32.375519 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:32 crc kubenswrapper[4846]: I1005 06:48:32.375542 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:32 crc kubenswrapper[4846]: I1005 06:48:32.375558 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:32Z","lastTransitionTime":"2025-10-05T06:48:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:32 crc kubenswrapper[4846]: I1005 06:48:32.394908 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e67d0e447a8a6a381259e25d25badc791194e1291f56b1cd77bd63989b20408\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dkhl7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://516f88f4990f5962d8b694a3d889259cc5508a309919870191c2e1331640c973\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dkhl7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-fscvf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:32Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:32 crc kubenswrapper[4846]: I1005 06:48:32.412545 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-w2sj8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32931fce-7374-4cab-a2e0-79d96a842274\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0df535d65d39de6c9a3dfb931bafc4469adc22b72f7a7903beda4fef8cd48138\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9rtxf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:27Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-w2sj8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:32Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:32 crc kubenswrapper[4846]: I1005 06:48:32.433375 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c9f812b-ecc8-4985-a670-53fa80fbb499\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2b5ee9781c01d6aec3ac778efef1067b6c3bfe4ed37431bdebe188fcde0165f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:47:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3fb4717692aa1fec0012b514ef8d18d1ca3707cfa4bb8248bb93841a0a40f11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:47:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0b6a7afe3094f1c92093c05c28344897a615d590c2fbb507c87e18a44c6c831\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://640ade2c733d7916ec213a201a1dd614d261c69eab654c80100efbc99a768995\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:47:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:32Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:32 crc kubenswrapper[4846]: I1005 06:48:32.455218 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:32Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:32 crc kubenswrapper[4846]: I1005 06:48:32.479593 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:32 crc kubenswrapper[4846]: I1005 06:48:32.479673 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:32 crc kubenswrapper[4846]: I1005 06:48:32.479692 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:32 crc kubenswrapper[4846]: I1005 06:48:32.479722 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:32 crc kubenswrapper[4846]: I1005 06:48:32.479741 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:32Z","lastTransitionTime":"2025-10-05T06:48:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:32 crc kubenswrapper[4846]: I1005 06:48:32.497511 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 05 06:48:32 crc kubenswrapper[4846]: E1005 06:48:32.497763 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 05 06:48:32 crc kubenswrapper[4846]: I1005 06:48:32.582735 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:32 crc kubenswrapper[4846]: I1005 06:48:32.582809 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:32 crc kubenswrapper[4846]: I1005 06:48:32.582830 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:32 crc kubenswrapper[4846]: I1005 06:48:32.582860 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:32 crc kubenswrapper[4846]: I1005 06:48:32.582881 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:32Z","lastTransitionTime":"2025-10-05T06:48:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:32 crc kubenswrapper[4846]: I1005 06:48:32.686157 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:32 crc kubenswrapper[4846]: I1005 06:48:32.686277 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:32 crc kubenswrapper[4846]: I1005 06:48:32.686302 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:32 crc kubenswrapper[4846]: I1005 06:48:32.686338 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:32 crc kubenswrapper[4846]: I1005 06:48:32.686361 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:32Z","lastTransitionTime":"2025-10-05T06:48:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:32 crc kubenswrapper[4846]: I1005 06:48:32.789431 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:32 crc kubenswrapper[4846]: I1005 06:48:32.789500 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:32 crc kubenswrapper[4846]: I1005 06:48:32.789521 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:32 crc kubenswrapper[4846]: I1005 06:48:32.789546 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:32 crc kubenswrapper[4846]: I1005 06:48:32.789566 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:32Z","lastTransitionTime":"2025-10-05T06:48:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:32 crc kubenswrapper[4846]: I1005 06:48:32.817343 4846 generic.go:334] "Generic (PLEG): container finished" podID="1dbb13ac-136c-4a1c-9229-b4e8ac4f9c71" containerID="f345693e9aeb1f4dc196349d3354f3c833a2b0415a693298b1cfccfa2b0d4459" exitCode=0 Oct 05 06:48:32 crc kubenswrapper[4846]: I1005 06:48:32.817478 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-sfmxd" event={"ID":"1dbb13ac-136c-4a1c-9229-b4e8ac4f9c71","Type":"ContainerDied","Data":"f345693e9aeb1f4dc196349d3354f3c833a2b0415a693298b1cfccfa2b0d4459"} Oct 05 06:48:32 crc kubenswrapper[4846]: I1005 06:48:32.820699 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Oct 05 06:48:32 crc kubenswrapper[4846]: I1005 06:48:32.825421 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"67c46f5f0a3ace95bd2e9479c5561126664e18b148c5eceea5bf5efc27e81a67"} Oct 05 06:48:32 crc kubenswrapper[4846]: I1005 06:48:32.825554 4846 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 05 06:48:32 crc kubenswrapper[4846]: I1005 06:48:32.860513 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c482bb97-fd50-40cf-96ea-f806c3f01125\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bfe45d30e4fa141541aca1d46bc5cc1da0a7b4a5eb1fc9bb2caa5b7fdcb0b264\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://496a4782faa30a9a20ff63038141600cd635f12902d1cf17989cd067fcdfe929\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://281be077869a6aa1ca815770871a14413d4a97e97aeb550fd5cd215cba88b21f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6771c7215d2f5dd5838d17888375a4768b9d1755fc2e21bda1bbfbb2eb1dbb95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c17a711c2c7473b076a3944ffe7c700808d05cba582fbd55c5c2ac8ff73838a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44a9f740f0a06c8bf928e2a9d1d5979217558c31e73a333631baccebf00fd5d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://44a9f740f0a06c8bf928e2a9d1d5979217558c31e73a333631baccebf00fd5d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:47:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:47:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://48ca0ba6ca6092bfcb5b153220c054f02cc36fb22707c42e021f06d329e18e2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48ca0ba6ca6092bfcb5b153220c054f02cc36fb22707c42e021f06d329e18e2e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:00Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4006d0cea128675c8885d879171204abffc5e11ea0664750a12b4bc7344ccb2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4006d0cea128675c8885d879171204abffc5e11ea0664750a12b4bc7344ccb2e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:47:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:32Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:32 crc kubenswrapper[4846]: I1005 06:48:32.881502 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:32Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:32 crc kubenswrapper[4846]: I1005 06:48:32.893541 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:32 crc kubenswrapper[4846]: I1005 06:48:32.893592 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:32 crc kubenswrapper[4846]: I1005 06:48:32.893610 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:32 crc kubenswrapper[4846]: I1005 06:48:32.893636 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:32 crc kubenswrapper[4846]: I1005 06:48:32.893659 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:32Z","lastTransitionTime":"2025-10-05T06:48:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:32 crc kubenswrapper[4846]: I1005 06:48:32.907631 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-sfmxd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1dbb13ac-136c-4a1c-9229-b4e8ac4f9c71\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://226d977b596d07aeda0486452e2a8605949ed4d2453b21f56174937c395d6f36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://226d977b596d07aeda0486452e2a8605949ed4d2453b21f56174937c395d6f36\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://99c07a4bdb3059d3a99ffdaf794d668d52cffd9e5e3bb8945204dd6d1f8c8328\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99c07a4bdb3059d3a99ffdaf794d668d52cffd9e5e3bb8945204dd6d1f8c8328\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27095272bf4f31d63a78a436636de0aae12f7103597da944f24ed7a74329ee3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27095272bf4f31d63a78a436636de0aae12f7103597da944f24ed7a74329ee3f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1211571af66d8a94302b65feccf30768ba82b7cbad918e44dd405787dce484cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1211571af66d8a94302b65feccf30768ba82b7cbad918e44dd405787dce484cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27631f90d528900fa180c15cfce80e96d4cdba07743eb6d847e2e822d9917218\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27631f90d528900fa180c15cfce80e96d4cdba07743eb6d847e2e822d9917218\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f345693e9aeb1f4dc196349d3354f3c833a2b0415a693298b1cfccfa2b0d4459\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f345693e9aeb1f4dc196349d3354f3c833a2b0415a693298b1cfccfa2b0d4459\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-sfmxd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:32Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:32 crc kubenswrapper[4846]: I1005 06:48:32.963603 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-txzlk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"93a6ace4-8d64-44d7-9637-457d3af6543c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ad1940d07e37217089642478ece9ccdaca74784738a400929a9f5359d158c01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a6f6b2c32dd3da5f9fd5e6971c1773e81a9a24184b7aa5b543cec879001ae53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e9b0219ecbc7e60905d6de3e67a2c8dc9af1017b4aed9538ee0c58c861f2438\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af5e5a4da9ad7371d5a24633e8268f12e3a7d419a8c93b11ddbcc006d8339c6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8da85222775abc7a772f6a8ff8255f1a0d5b15807ba177a7e3e2b2d01a4f24b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1409651de8c897804f87be03d016c138f8acf27e662dce265a93e15c1848adcf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c212f5d230fae82a701afeceb39a848a9c91a3354908d88b062b7e1a726240cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0df0ebcd87dc5af4b17736b43ec7b2395968b19ed16e775366af14e2b2d49c52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff3c6545ff3d2436da27bbd223b950b30b0c69a3e7f3034cb0e61380cb38d338\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ff3c6545ff3d2436da27bbd223b950b30b0c69a3e7f3034cb0e61380cb38d338\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:24Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-txzlk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:32Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:32 crc kubenswrapper[4846]: I1005 06:48:32.996587 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:32 crc kubenswrapper[4846]: I1005 06:48:32.996653 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:32 crc kubenswrapper[4846]: I1005 06:48:32.996667 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:32 crc kubenswrapper[4846]: I1005 06:48:32.996692 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:32 crc kubenswrapper[4846]: I1005 06:48:32.996708 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:32Z","lastTransitionTime":"2025-10-05T06:48:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:32 crc kubenswrapper[4846]: I1005 06:48:32.997896 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c9f812b-ecc8-4985-a670-53fa80fbb499\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2b5ee9781c01d6aec3ac778efef1067b6c3bfe4ed37431bdebe188fcde0165f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:47:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3fb4717692aa1fec0012b514ef8d18d1ca3707cfa4bb8248bb93841a0a40f11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:47:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0b6a7afe3094f1c92093c05c28344897a615d590c2fbb507c87e18a44c6c831\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://640ade2c733d7916ec213a201a1dd614d261c69eab654c80100efbc99a768995\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:47:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:32Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:33 crc kubenswrapper[4846]: I1005 06:48:33.015870 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:33Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:33 crc kubenswrapper[4846]: I1005 06:48:33.034367 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e67d0e447a8a6a381259e25d25badc791194e1291f56b1cd77bd63989b20408\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dkhl7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://516f88f4990f5962d8b694a3d889259cc5508a309919870191c2e1331640c973\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dkhl7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-fscvf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:33Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:33 crc kubenswrapper[4846]: I1005 06:48:33.053499 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-w2sj8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32931fce-7374-4cab-a2e0-79d96a842274\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0df535d65d39de6c9a3dfb931bafc4469adc22b72f7a7903beda4fef8cd48138\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9rtxf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:27Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-w2sj8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:33Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:33 crc kubenswrapper[4846]: I1005 06:48:33.069435 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ff10c31459b3b3724ee7b769bd6f72f9ed80cce5419b6adcdf69a3239b70756\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:33Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:33 crc kubenswrapper[4846]: I1005 06:48:33.099850 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:33 crc kubenswrapper[4846]: I1005 06:48:33.099894 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:33 crc kubenswrapper[4846]: I1005 06:48:33.099907 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:33 crc kubenswrapper[4846]: I1005 06:48:33.099926 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:33 crc kubenswrapper[4846]: I1005 06:48:33.099939 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:33Z","lastTransitionTime":"2025-10-05T06:48:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:33 crc kubenswrapper[4846]: I1005 06:48:33.101134 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://347f900743b7bba3daf92f21c7eba0625e8c774f50057bc9472f0c82119325e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:33Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:33 crc kubenswrapper[4846]: I1005 06:48:33.122352 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56f1fe1f9f06f0249cf937a6963fda82467732dc056ca2b2189be6d8dd9d4a05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7212ce5f42c1fb55146abb190a557a504b43f16af6df84f90e1073567a329ec5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:33Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:33 crc kubenswrapper[4846]: I1005 06:48:33.138125 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-7wm5q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9bc7744f-d700-450f-8b8d-98140877fee9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90379e74c544fccca44f7acdc228ff378154390e211ac801b79f8d074012241e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h5p4q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-7wm5q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:33Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:33 crc kubenswrapper[4846]: I1005 06:48:33.162921 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"775b5702-a131-4d6c-852d-91bfa056cac5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0839c1765df2c68d2554e9c66d289505e559a9672caf8069feeff2c4fdfd2171\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15dbaf2445704da0987f256f81dff2ee95664988541fa48ce94898afeadab2e8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd3b3ab323c6d5153054aa04af2d91f7b9e8436ff1e2f75f3594b57831b2a416\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c52e076eddfd307110b1f6de00a169b81af8f2785af7becef2a820ce363a5861\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c52e076eddfd307110b1f6de00a169b81af8f2785af7becef2a820ce363a5861\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-05T06:48:14Z\\\",\\\"message\\\":\\\"le observer\\\\nW1005 06:48:14.337870 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1005 06:48:14.338086 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1005 06:48:14.339389 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1467752215/tls.crt::/tmp/serving-cert-1467752215/tls.key\\\\\\\"\\\\nI1005 06:48:14.864345 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1005 06:48:14.868087 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1005 06:48:14.868114 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1005 06:48:14.868140 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1005 06:48:14.868147 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1005 06:48:14.877354 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1005 06:48:14.877412 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1005 06:48:14.877456 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1005 06:48:14.877471 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1005 06:48:14.877490 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1005 06:48:14.877501 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1005 06:48:14.877507 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1005 06:48:14.877515 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1005 06:48:14.880326 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:13Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4a091716ff935adb42ae2e88048619fd347bd174f64e5f9c97a01f81690e57f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:01Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dc5411b642f99b13d514d2f21defad0c394aeeb82a966f72300076617e2586c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5dc5411b642f99b13d514d2f21defad0c394aeeb82a966f72300076617e2586c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:47:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:47:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:47:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:33Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:33 crc kubenswrapper[4846]: I1005 06:48:33.179290 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:33Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:33 crc kubenswrapper[4846]: I1005 06:48:33.190597 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-r7ztb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a5119cd9-af70-4324-a955-978306ab0b20\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06231a87429ec292cb57ebd88cda7ad547bcb32e80343b776833621a11c28ad1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jfcj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-r7ztb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:33Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:33 crc kubenswrapper[4846]: I1005 06:48:33.203635 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:33 crc kubenswrapper[4846]: I1005 06:48:33.203679 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:33 crc kubenswrapper[4846]: I1005 06:48:33.203693 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:33 crc kubenswrapper[4846]: I1005 06:48:33.203711 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:33 crc kubenswrapper[4846]: I1005 06:48:33.203723 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:33Z","lastTransitionTime":"2025-10-05T06:48:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:33 crc kubenswrapper[4846]: I1005 06:48:33.214794 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c482bb97-fd50-40cf-96ea-f806c3f01125\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bfe45d30e4fa141541aca1d46bc5cc1da0a7b4a5eb1fc9bb2caa5b7fdcb0b264\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://496a4782faa30a9a20ff63038141600cd635f12902d1cf17989cd067fcdfe929\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://281be077869a6aa1ca815770871a14413d4a97e97aeb550fd5cd215cba88b21f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6771c7215d2f5dd5838d17888375a4768b9d1755fc2e21bda1bbfbb2eb1dbb95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c17a711c2c7473b076a3944ffe7c700808d05cba582fbd55c5c2ac8ff73838a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44a9f740f0a06c8bf928e2a9d1d5979217558c31e73a333631baccebf00fd5d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://44a9f740f0a06c8bf928e2a9d1d5979217558c31e73a333631baccebf00fd5d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:47:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:47:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://48ca0ba6ca6092bfcb5b153220c054f02cc36fb22707c42e021f06d329e18e2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48ca0ba6ca6092bfcb5b153220c054f02cc36fb22707c42e021f06d329e18e2e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:00Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4006d0cea128675c8885d879171204abffc5e11ea0664750a12b4bc7344ccb2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4006d0cea128675c8885d879171204abffc5e11ea0664750a12b4bc7344ccb2e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:47:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:33Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:33 crc kubenswrapper[4846]: I1005 06:48:33.232732 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:33Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:33 crc kubenswrapper[4846]: I1005 06:48:33.251986 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-sfmxd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1dbb13ac-136c-4a1c-9229-b4e8ac4f9c71\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://226d977b596d07aeda0486452e2a8605949ed4d2453b21f56174937c395d6f36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://226d977b596d07aeda0486452e2a8605949ed4d2453b21f56174937c395d6f36\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://99c07a4bdb3059d3a99ffdaf794d668d52cffd9e5e3bb8945204dd6d1f8c8328\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99c07a4bdb3059d3a99ffdaf794d668d52cffd9e5e3bb8945204dd6d1f8c8328\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27095272bf4f31d63a78a436636de0aae12f7103597da944f24ed7a74329ee3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27095272bf4f31d63a78a436636de0aae12f7103597da944f24ed7a74329ee3f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1211571af66d8a94302b65feccf30768ba82b7cbad918e44dd405787dce484cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1211571af66d8a94302b65feccf30768ba82b7cbad918e44dd405787dce484cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27631f90d528900fa180c15cfce80e96d4cdba07743eb6d847e2e822d9917218\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27631f90d528900fa180c15cfce80e96d4cdba07743eb6d847e2e822d9917218\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f345693e9aeb1f4dc196349d3354f3c833a2b0415a693298b1cfccfa2b0d4459\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f345693e9aeb1f4dc196349d3354f3c833a2b0415a693298b1cfccfa2b0d4459\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-sfmxd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:33Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:33 crc kubenswrapper[4846]: I1005 06:48:33.276194 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-txzlk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"93a6ace4-8d64-44d7-9637-457d3af6543c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ad1940d07e37217089642478ece9ccdaca74784738a400929a9f5359d158c01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a6f6b2c32dd3da5f9fd5e6971c1773e81a9a24184b7aa5b543cec879001ae53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e9b0219ecbc7e60905d6de3e67a2c8dc9af1017b4aed9538ee0c58c861f2438\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af5e5a4da9ad7371d5a24633e8268f12e3a7d419a8c93b11ddbcc006d8339c6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8da85222775abc7a772f6a8ff8255f1a0d5b15807ba177a7e3e2b2d01a4f24b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1409651de8c897804f87be03d016c138f8acf27e662dce265a93e15c1848adcf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c212f5d230fae82a701afeceb39a848a9c91a3354908d88b062b7e1a726240cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0df0ebcd87dc5af4b17736b43ec7b2395968b19ed16e775366af14e2b2d49c52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff3c6545ff3d2436da27bbd223b950b30b0c69a3e7f3034cb0e61380cb38d338\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ff3c6545ff3d2436da27bbd223b950b30b0c69a3e7f3034cb0e61380cb38d338\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:24Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-txzlk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:33Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:33 crc kubenswrapper[4846]: I1005 06:48:33.292752 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c9f812b-ecc8-4985-a670-53fa80fbb499\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2b5ee9781c01d6aec3ac778efef1067b6c3bfe4ed37431bdebe188fcde0165f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:47:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3fb4717692aa1fec0012b514ef8d18d1ca3707cfa4bb8248bb93841a0a40f11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:47:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0b6a7afe3094f1c92093c05c28344897a615d590c2fbb507c87e18a44c6c831\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://640ade2c733d7916ec213a201a1dd614d261c69eab654c80100efbc99a768995\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:47:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:33Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:33 crc kubenswrapper[4846]: I1005 06:48:33.306834 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:33 crc kubenswrapper[4846]: I1005 06:48:33.306876 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:33 crc kubenswrapper[4846]: I1005 06:48:33.306889 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:33 crc kubenswrapper[4846]: I1005 06:48:33.306908 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:33 crc kubenswrapper[4846]: I1005 06:48:33.306923 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:33Z","lastTransitionTime":"2025-10-05T06:48:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:33 crc kubenswrapper[4846]: I1005 06:48:33.308818 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:33Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:33 crc kubenswrapper[4846]: I1005 06:48:33.323979 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e67d0e447a8a6a381259e25d25badc791194e1291f56b1cd77bd63989b20408\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dkhl7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://516f88f4990f5962d8b694a3d889259cc5508a309919870191c2e1331640c973\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dkhl7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-fscvf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:33Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:33 crc kubenswrapper[4846]: I1005 06:48:33.337458 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-w2sj8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32931fce-7374-4cab-a2e0-79d96a842274\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0df535d65d39de6c9a3dfb931bafc4469adc22b72f7a7903beda4fef8cd48138\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9rtxf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:27Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-w2sj8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:33Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:33 crc kubenswrapper[4846]: I1005 06:48:33.352721 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-7wm5q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9bc7744f-d700-450f-8b8d-98140877fee9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90379e74c544fccca44f7acdc228ff378154390e211ac801b79f8d074012241e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h5p4q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-7wm5q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:33Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:33 crc kubenswrapper[4846]: I1005 06:48:33.368291 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ff10c31459b3b3724ee7b769bd6f72f9ed80cce5419b6adcdf69a3239b70756\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:33Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:33 crc kubenswrapper[4846]: I1005 06:48:33.381610 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://347f900743b7bba3daf92f21c7eba0625e8c774f50057bc9472f0c82119325e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:33Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:33 crc kubenswrapper[4846]: I1005 06:48:33.407992 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56f1fe1f9f06f0249cf937a6963fda82467732dc056ca2b2189be6d8dd9d4a05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7212ce5f42c1fb55146abb190a557a504b43f16af6df84f90e1073567a329ec5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:33Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:33 crc kubenswrapper[4846]: I1005 06:48:33.409067 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 05 06:48:33 crc kubenswrapper[4846]: I1005 06:48:33.409203 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 05 06:48:33 crc kubenswrapper[4846]: I1005 06:48:33.409235 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 05 06:48:33 crc kubenswrapper[4846]: I1005 06:48:33.409274 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 05 06:48:33 crc kubenswrapper[4846]: I1005 06:48:33.409297 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 05 06:48:33 crc kubenswrapper[4846]: E1005 06:48:33.409445 4846 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 05 06:48:33 crc kubenswrapper[4846]: E1005 06:48:33.409470 4846 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 05 06:48:33 crc kubenswrapper[4846]: E1005 06:48:33.409483 4846 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 05 06:48:33 crc kubenswrapper[4846]: E1005 06:48:33.409534 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-05 06:48:49.409520151 +0000 UTC m=+51.650372926 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 05 06:48:33 crc kubenswrapper[4846]: E1005 06:48:33.409610 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-05 06:48:49.409604323 +0000 UTC m=+51.650457098 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 06:48:33 crc kubenswrapper[4846]: E1005 06:48:33.409660 4846 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 05 06:48:33 crc kubenswrapper[4846]: E1005 06:48:33.409685 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-05 06:48:49.409677095 +0000 UTC m=+51.650529870 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 05 06:48:33 crc kubenswrapper[4846]: E1005 06:48:33.409735 4846 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 05 06:48:33 crc kubenswrapper[4846]: E1005 06:48:33.409745 4846 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 05 06:48:33 crc kubenswrapper[4846]: E1005 06:48:33.409753 4846 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 05 06:48:33 crc kubenswrapper[4846]: E1005 06:48:33.409771 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-05 06:48:49.409765938 +0000 UTC m=+51.650618713 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 05 06:48:33 crc kubenswrapper[4846]: E1005 06:48:33.409804 4846 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 05 06:48:33 crc kubenswrapper[4846]: E1005 06:48:33.409824 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-05 06:48:49.40981866 +0000 UTC m=+51.650671435 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 05 06:48:33 crc kubenswrapper[4846]: I1005 06:48:33.410276 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:33 crc kubenswrapper[4846]: I1005 06:48:33.410308 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:33 crc kubenswrapper[4846]: I1005 06:48:33.410318 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:33 crc kubenswrapper[4846]: I1005 06:48:33.410336 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:33 crc kubenswrapper[4846]: I1005 06:48:33.410347 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:33Z","lastTransitionTime":"2025-10-05T06:48:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:33 crc kubenswrapper[4846]: I1005 06:48:33.423480 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"775b5702-a131-4d6c-852d-91bfa056cac5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0839c1765df2c68d2554e9c66d289505e559a9672caf8069feeff2c4fdfd2171\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15dbaf2445704da0987f256f81dff2ee95664988541fa48ce94898afeadab2e8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd3b3ab323c6d5153054aa04af2d91f7b9e8436ff1e2f75f3594b57831b2a416\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://67c46f5f0a3ace95bd2e9479c5561126664e18b148c5eceea5bf5efc27e81a67\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c52e076eddfd307110b1f6de00a169b81af8f2785af7becef2a820ce363a5861\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-05T06:48:14Z\\\",\\\"message\\\":\\\"le observer\\\\nW1005 06:48:14.337870 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1005 06:48:14.338086 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1005 06:48:14.339389 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1467752215/tls.crt::/tmp/serving-cert-1467752215/tls.key\\\\\\\"\\\\nI1005 06:48:14.864345 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1005 06:48:14.868087 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1005 06:48:14.868114 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1005 06:48:14.868140 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1005 06:48:14.868147 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1005 06:48:14.877354 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1005 06:48:14.877412 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1005 06:48:14.877456 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1005 06:48:14.877471 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1005 06:48:14.877490 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1005 06:48:14.877501 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1005 06:48:14.877507 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1005 06:48:14.877515 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1005 06:48:14.880326 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:13Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4a091716ff935adb42ae2e88048619fd347bd174f64e5f9c97a01f81690e57f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:01Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dc5411b642f99b13d514d2f21defad0c394aeeb82a966f72300076617e2586c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5dc5411b642f99b13d514d2f21defad0c394aeeb82a966f72300076617e2586c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:47:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:47:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:47:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:33Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:33 crc kubenswrapper[4846]: I1005 06:48:33.439095 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:33Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:33 crc kubenswrapper[4846]: I1005 06:48:33.449310 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-r7ztb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a5119cd9-af70-4324-a955-978306ab0b20\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06231a87429ec292cb57ebd88cda7ad547bcb32e80343b776833621a11c28ad1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jfcj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-r7ztb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:33Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:33 crc kubenswrapper[4846]: I1005 06:48:33.496640 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 05 06:48:33 crc kubenswrapper[4846]: I1005 06:48:33.496733 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 05 06:48:33 crc kubenswrapper[4846]: E1005 06:48:33.496792 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 05 06:48:33 crc kubenswrapper[4846]: E1005 06:48:33.496886 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 05 06:48:33 crc kubenswrapper[4846]: I1005 06:48:33.512458 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:33 crc kubenswrapper[4846]: I1005 06:48:33.512495 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:33 crc kubenswrapper[4846]: I1005 06:48:33.512508 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:33 crc kubenswrapper[4846]: I1005 06:48:33.512524 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:33 crc kubenswrapper[4846]: I1005 06:48:33.512539 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:33Z","lastTransitionTime":"2025-10-05T06:48:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:33 crc kubenswrapper[4846]: I1005 06:48:33.615724 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:33 crc kubenswrapper[4846]: I1005 06:48:33.615773 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:33 crc kubenswrapper[4846]: I1005 06:48:33.615789 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:33 crc kubenswrapper[4846]: I1005 06:48:33.615810 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:33 crc kubenswrapper[4846]: I1005 06:48:33.615826 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:33Z","lastTransitionTime":"2025-10-05T06:48:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:33 crc kubenswrapper[4846]: I1005 06:48:33.718351 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:33 crc kubenswrapper[4846]: I1005 06:48:33.718382 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:33 crc kubenswrapper[4846]: I1005 06:48:33.718390 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:33 crc kubenswrapper[4846]: I1005 06:48:33.718406 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:33 crc kubenswrapper[4846]: I1005 06:48:33.718417 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:33Z","lastTransitionTime":"2025-10-05T06:48:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:33 crc kubenswrapper[4846]: I1005 06:48:33.820493 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:33 crc kubenswrapper[4846]: I1005 06:48:33.820523 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:33 crc kubenswrapper[4846]: I1005 06:48:33.820530 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:33 crc kubenswrapper[4846]: I1005 06:48:33.820544 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:33 crc kubenswrapper[4846]: I1005 06:48:33.820553 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:33Z","lastTransitionTime":"2025-10-05T06:48:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:33 crc kubenswrapper[4846]: I1005 06:48:33.832345 4846 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 05 06:48:33 crc kubenswrapper[4846]: I1005 06:48:33.833439 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-sfmxd" event={"ID":"1dbb13ac-136c-4a1c-9229-b4e8ac4f9c71","Type":"ContainerStarted","Data":"2b0af6a6e735f7b7877c177c1d53d1a8dc12c4713a8059c4766fc858af31c9a4"} Oct 05 06:48:33 crc kubenswrapper[4846]: I1005 06:48:33.854328 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c482bb97-fd50-40cf-96ea-f806c3f01125\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bfe45d30e4fa141541aca1d46bc5cc1da0a7b4a5eb1fc9bb2caa5b7fdcb0b264\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://496a4782faa30a9a20ff63038141600cd635f12902d1cf17989cd067fcdfe929\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://281be077869a6aa1ca815770871a14413d4a97e97aeb550fd5cd215cba88b21f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6771c7215d2f5dd5838d17888375a4768b9d1755fc2e21bda1bbfbb2eb1dbb95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c17a711c2c7473b076a3944ffe7c700808d05cba582fbd55c5c2ac8ff73838a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44a9f740f0a06c8bf928e2a9d1d5979217558c31e73a333631baccebf00fd5d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://44a9f740f0a06c8bf928e2a9d1d5979217558c31e73a333631baccebf00fd5d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:47:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:47:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://48ca0ba6ca6092bfcb5b153220c054f02cc36fb22707c42e021f06d329e18e2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48ca0ba6ca6092bfcb5b153220c054f02cc36fb22707c42e021f06d329e18e2e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:00Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4006d0cea128675c8885d879171204abffc5e11ea0664750a12b4bc7344ccb2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4006d0cea128675c8885d879171204abffc5e11ea0664750a12b4bc7344ccb2e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:47:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:33Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:33 crc kubenswrapper[4846]: I1005 06:48:33.874758 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:33Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:33 crc kubenswrapper[4846]: I1005 06:48:33.896067 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-sfmxd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1dbb13ac-136c-4a1c-9229-b4e8ac4f9c71\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b0af6a6e735f7b7877c177c1d53d1a8dc12c4713a8059c4766fc858af31c9a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://226d977b596d07aeda0486452e2a8605949ed4d2453b21f56174937c395d6f36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://226d977b596d07aeda0486452e2a8605949ed4d2453b21f56174937c395d6f36\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://99c07a4bdb3059d3a99ffdaf794d668d52cffd9e5e3bb8945204dd6d1f8c8328\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99c07a4bdb3059d3a99ffdaf794d668d52cffd9e5e3bb8945204dd6d1f8c8328\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27095272bf4f31d63a78a436636de0aae12f7103597da944f24ed7a74329ee3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27095272bf4f31d63a78a436636de0aae12f7103597da944f24ed7a74329ee3f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1211571af66d8a94302b65feccf30768ba82b7cbad918e44dd405787dce484cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1211571af66d8a94302b65feccf30768ba82b7cbad918e44dd405787dce484cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27631f90d528900fa180c15cfce80e96d4cdba07743eb6d847e2e822d9917218\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27631f90d528900fa180c15cfce80e96d4cdba07743eb6d847e2e822d9917218\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f345693e9aeb1f4dc196349d3354f3c833a2b0415a693298b1cfccfa2b0d4459\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f345693e9aeb1f4dc196349d3354f3c833a2b0415a693298b1cfccfa2b0d4459\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-sfmxd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:33Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:33 crc kubenswrapper[4846]: I1005 06:48:33.923634 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:33 crc kubenswrapper[4846]: I1005 06:48:33.923723 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:33 crc kubenswrapper[4846]: I1005 06:48:33.923751 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:33 crc kubenswrapper[4846]: I1005 06:48:33.923785 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:33 crc kubenswrapper[4846]: I1005 06:48:33.923811 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:33Z","lastTransitionTime":"2025-10-05T06:48:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:33 crc kubenswrapper[4846]: I1005 06:48:33.925441 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-txzlk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"93a6ace4-8d64-44d7-9637-457d3af6543c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ad1940d07e37217089642478ece9ccdaca74784738a400929a9f5359d158c01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a6f6b2c32dd3da5f9fd5e6971c1773e81a9a24184b7aa5b543cec879001ae53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e9b0219ecbc7e60905d6de3e67a2c8dc9af1017b4aed9538ee0c58c861f2438\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af5e5a4da9ad7371d5a24633e8268f12e3a7d419a8c93b11ddbcc006d8339c6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8da85222775abc7a772f6a8ff8255f1a0d5b15807ba177a7e3e2b2d01a4f24b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1409651de8c897804f87be03d016c138f8acf27e662dce265a93e15c1848adcf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c212f5d230fae82a701afeceb39a848a9c91a3354908d88b062b7e1a726240cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0df0ebcd87dc5af4b17736b43ec7b2395968b19ed16e775366af14e2b2d49c52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff3c6545ff3d2436da27bbd223b950b30b0c69a3e7f3034cb0e61380cb38d338\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ff3c6545ff3d2436da27bbd223b950b30b0c69a3e7f3034cb0e61380cb38d338\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:24Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-txzlk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:33Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:33 crc kubenswrapper[4846]: I1005 06:48:33.942260 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c9f812b-ecc8-4985-a670-53fa80fbb499\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2b5ee9781c01d6aec3ac778efef1067b6c3bfe4ed37431bdebe188fcde0165f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:47:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3fb4717692aa1fec0012b514ef8d18d1ca3707cfa4bb8248bb93841a0a40f11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:47:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0b6a7afe3094f1c92093c05c28344897a615d590c2fbb507c87e18a44c6c831\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://640ade2c733d7916ec213a201a1dd614d261c69eab654c80100efbc99a768995\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:47:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:33Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:33 crc kubenswrapper[4846]: I1005 06:48:33.956573 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:33Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:33 crc kubenswrapper[4846]: I1005 06:48:33.976272 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e67d0e447a8a6a381259e25d25badc791194e1291f56b1cd77bd63989b20408\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dkhl7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://516f88f4990f5962d8b694a3d889259cc5508a309919870191c2e1331640c973\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dkhl7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-fscvf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:33Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:33 crc kubenswrapper[4846]: I1005 06:48:33.990339 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-w2sj8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32931fce-7374-4cab-a2e0-79d96a842274\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0df535d65d39de6c9a3dfb931bafc4469adc22b72f7a7903beda4fef8cd48138\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9rtxf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:27Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-w2sj8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:33Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:34 crc kubenswrapper[4846]: I1005 06:48:34.006247 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ff10c31459b3b3724ee7b769bd6f72f9ed80cce5419b6adcdf69a3239b70756\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:34Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:34 crc kubenswrapper[4846]: I1005 06:48:34.027370 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:34 crc kubenswrapper[4846]: I1005 06:48:34.027648 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://347f900743b7bba3daf92f21c7eba0625e8c774f50057bc9472f0c82119325e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:34Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:34 crc kubenswrapper[4846]: I1005 06:48:34.027734 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:34 crc kubenswrapper[4846]: I1005 06:48:34.027897 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:34 crc kubenswrapper[4846]: I1005 06:48:34.027925 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:34 crc kubenswrapper[4846]: I1005 06:48:34.027940 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:34Z","lastTransitionTime":"2025-10-05T06:48:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:34 crc kubenswrapper[4846]: I1005 06:48:34.049085 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56f1fe1f9f06f0249cf937a6963fda82467732dc056ca2b2189be6d8dd9d4a05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7212ce5f42c1fb55146abb190a557a504b43f16af6df84f90e1073567a329ec5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:34Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:34 crc kubenswrapper[4846]: I1005 06:48:34.073555 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-7wm5q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9bc7744f-d700-450f-8b8d-98140877fee9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90379e74c544fccca44f7acdc228ff378154390e211ac801b79f8d074012241e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h5p4q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-7wm5q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:34Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:34 crc kubenswrapper[4846]: I1005 06:48:34.090931 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"775b5702-a131-4d6c-852d-91bfa056cac5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0839c1765df2c68d2554e9c66d289505e559a9672caf8069feeff2c4fdfd2171\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15dbaf2445704da0987f256f81dff2ee95664988541fa48ce94898afeadab2e8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd3b3ab323c6d5153054aa04af2d91f7b9e8436ff1e2f75f3594b57831b2a416\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://67c46f5f0a3ace95bd2e9479c5561126664e18b148c5eceea5bf5efc27e81a67\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c52e076eddfd307110b1f6de00a169b81af8f2785af7becef2a820ce363a5861\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-05T06:48:14Z\\\",\\\"message\\\":\\\"le observer\\\\nW1005 06:48:14.337870 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1005 06:48:14.338086 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1005 06:48:14.339389 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1467752215/tls.crt::/tmp/serving-cert-1467752215/tls.key\\\\\\\"\\\\nI1005 06:48:14.864345 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1005 06:48:14.868087 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1005 06:48:14.868114 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1005 06:48:14.868140 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1005 06:48:14.868147 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1005 06:48:14.877354 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1005 06:48:14.877412 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1005 06:48:14.877456 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1005 06:48:14.877471 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1005 06:48:14.877490 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1005 06:48:14.877501 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1005 06:48:14.877507 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1005 06:48:14.877515 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1005 06:48:14.880326 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:13Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4a091716ff935adb42ae2e88048619fd347bd174f64e5f9c97a01f81690e57f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:01Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dc5411b642f99b13d514d2f21defad0c394aeeb82a966f72300076617e2586c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5dc5411b642f99b13d514d2f21defad0c394aeeb82a966f72300076617e2586c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:47:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:47:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:47:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:34Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:34 crc kubenswrapper[4846]: I1005 06:48:34.107436 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:34Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:34 crc kubenswrapper[4846]: I1005 06:48:34.121685 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-r7ztb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a5119cd9-af70-4324-a955-978306ab0b20\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06231a87429ec292cb57ebd88cda7ad547bcb32e80343b776833621a11c28ad1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jfcj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-r7ztb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:34Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:34 crc kubenswrapper[4846]: I1005 06:48:34.131471 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:34 crc kubenswrapper[4846]: I1005 06:48:34.131539 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:34 crc kubenswrapper[4846]: I1005 06:48:34.131559 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:34 crc kubenswrapper[4846]: I1005 06:48:34.131586 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:34 crc kubenswrapper[4846]: I1005 06:48:34.131604 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:34Z","lastTransitionTime":"2025-10-05T06:48:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:34 crc kubenswrapper[4846]: I1005 06:48:34.235229 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:34 crc kubenswrapper[4846]: I1005 06:48:34.235287 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:34 crc kubenswrapper[4846]: I1005 06:48:34.235306 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:34 crc kubenswrapper[4846]: I1005 06:48:34.235328 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:34 crc kubenswrapper[4846]: I1005 06:48:34.235341 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:34Z","lastTransitionTime":"2025-10-05T06:48:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:34 crc kubenswrapper[4846]: I1005 06:48:34.339327 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:34 crc kubenswrapper[4846]: I1005 06:48:34.339365 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:34 crc kubenswrapper[4846]: I1005 06:48:34.339381 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:34 crc kubenswrapper[4846]: I1005 06:48:34.339405 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:34 crc kubenswrapper[4846]: I1005 06:48:34.339422 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:34Z","lastTransitionTime":"2025-10-05T06:48:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:34 crc kubenswrapper[4846]: I1005 06:48:34.442939 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:34 crc kubenswrapper[4846]: I1005 06:48:34.443012 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:34 crc kubenswrapper[4846]: I1005 06:48:34.443032 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:34 crc kubenswrapper[4846]: I1005 06:48:34.443060 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:34 crc kubenswrapper[4846]: I1005 06:48:34.443079 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:34Z","lastTransitionTime":"2025-10-05T06:48:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:34 crc kubenswrapper[4846]: I1005 06:48:34.497299 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 05 06:48:34 crc kubenswrapper[4846]: E1005 06:48:34.497526 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 05 06:48:34 crc kubenswrapper[4846]: I1005 06:48:34.524991 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:34 crc kubenswrapper[4846]: I1005 06:48:34.525061 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:34 crc kubenswrapper[4846]: I1005 06:48:34.525081 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:34 crc kubenswrapper[4846]: I1005 06:48:34.525108 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:34 crc kubenswrapper[4846]: I1005 06:48:34.525129 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:34Z","lastTransitionTime":"2025-10-05T06:48:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:34 crc kubenswrapper[4846]: E1005 06:48:34.545487 4846 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:48:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:48:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:34Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:48:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:48:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:34Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"04ce628d-2f61-4931-aad6-615924ea53fb\\\",\\\"systemUUID\\\":\\\"a050a23b-e773-4ba9-989b-360aa53a0605\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:34Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:34 crc kubenswrapper[4846]: I1005 06:48:34.551706 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:34 crc kubenswrapper[4846]: I1005 06:48:34.551791 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:34 crc kubenswrapper[4846]: I1005 06:48:34.551813 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:34 crc kubenswrapper[4846]: I1005 06:48:34.551843 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:34 crc kubenswrapper[4846]: I1005 06:48:34.551865 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:34Z","lastTransitionTime":"2025-10-05T06:48:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:34 crc kubenswrapper[4846]: E1005 06:48:34.571788 4846 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:48:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:48:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:34Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:48:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:48:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:34Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"04ce628d-2f61-4931-aad6-615924ea53fb\\\",\\\"systemUUID\\\":\\\"a050a23b-e773-4ba9-989b-360aa53a0605\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:34Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:34 crc kubenswrapper[4846]: I1005 06:48:34.577523 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:34 crc kubenswrapper[4846]: I1005 06:48:34.577596 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:34 crc kubenswrapper[4846]: I1005 06:48:34.577613 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:34 crc kubenswrapper[4846]: I1005 06:48:34.577642 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:34 crc kubenswrapper[4846]: I1005 06:48:34.577662 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:34Z","lastTransitionTime":"2025-10-05T06:48:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:34 crc kubenswrapper[4846]: E1005 06:48:34.596970 4846 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:48:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:48:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:34Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:48:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:48:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:34Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"04ce628d-2f61-4931-aad6-615924ea53fb\\\",\\\"systemUUID\\\":\\\"a050a23b-e773-4ba9-989b-360aa53a0605\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:34Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:34 crc kubenswrapper[4846]: I1005 06:48:34.603319 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:34 crc kubenswrapper[4846]: I1005 06:48:34.603392 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:34 crc kubenswrapper[4846]: I1005 06:48:34.603416 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:34 crc kubenswrapper[4846]: I1005 06:48:34.603454 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:34 crc kubenswrapper[4846]: I1005 06:48:34.603480 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:34Z","lastTransitionTime":"2025-10-05T06:48:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:34 crc kubenswrapper[4846]: E1005 06:48:34.627045 4846 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:48:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:48:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:34Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:48:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:48:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:34Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"04ce628d-2f61-4931-aad6-615924ea53fb\\\",\\\"systemUUID\\\":\\\"a050a23b-e773-4ba9-989b-360aa53a0605\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:34Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:34 crc kubenswrapper[4846]: I1005 06:48:34.632259 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:34 crc kubenswrapper[4846]: I1005 06:48:34.632339 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:34 crc kubenswrapper[4846]: I1005 06:48:34.632365 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:34 crc kubenswrapper[4846]: I1005 06:48:34.632396 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:34 crc kubenswrapper[4846]: I1005 06:48:34.632418 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:34Z","lastTransitionTime":"2025-10-05T06:48:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:34 crc kubenswrapper[4846]: E1005 06:48:34.652607 4846 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:48:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:48:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:34Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:48:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:48:34Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:34Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"04ce628d-2f61-4931-aad6-615924ea53fb\\\",\\\"systemUUID\\\":\\\"a050a23b-e773-4ba9-989b-360aa53a0605\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:34Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:34 crc kubenswrapper[4846]: E1005 06:48:34.652853 4846 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 05 06:48:34 crc kubenswrapper[4846]: I1005 06:48:34.655592 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:34 crc kubenswrapper[4846]: I1005 06:48:34.655645 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:34 crc kubenswrapper[4846]: I1005 06:48:34.655670 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:34 crc kubenswrapper[4846]: I1005 06:48:34.655700 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:34 crc kubenswrapper[4846]: I1005 06:48:34.655719 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:34Z","lastTransitionTime":"2025-10-05T06:48:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:34 crc kubenswrapper[4846]: I1005 06:48:34.759486 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:34 crc kubenswrapper[4846]: I1005 06:48:34.759563 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:34 crc kubenswrapper[4846]: I1005 06:48:34.759584 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:34 crc kubenswrapper[4846]: I1005 06:48:34.759615 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:34 crc kubenswrapper[4846]: I1005 06:48:34.759634 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:34Z","lastTransitionTime":"2025-10-05T06:48:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:34 crc kubenswrapper[4846]: I1005 06:48:34.839815 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-txzlk_93a6ace4-8d64-44d7-9637-457d3af6543c/ovnkube-controller/0.log" Oct 05 06:48:34 crc kubenswrapper[4846]: I1005 06:48:34.850216 4846 generic.go:334] "Generic (PLEG): container finished" podID="93a6ace4-8d64-44d7-9637-457d3af6543c" containerID="c212f5d230fae82a701afeceb39a848a9c91a3354908d88b062b7e1a726240cb" exitCode=1 Oct 05 06:48:34 crc kubenswrapper[4846]: I1005 06:48:34.850250 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-txzlk" event={"ID":"93a6ace4-8d64-44d7-9637-457d3af6543c","Type":"ContainerDied","Data":"c212f5d230fae82a701afeceb39a848a9c91a3354908d88b062b7e1a726240cb"} Oct 05 06:48:34 crc kubenswrapper[4846]: I1005 06:48:34.851938 4846 scope.go:117] "RemoveContainer" containerID="c212f5d230fae82a701afeceb39a848a9c91a3354908d88b062b7e1a726240cb" Oct 05 06:48:34 crc kubenswrapper[4846]: I1005 06:48:34.862679 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:34 crc kubenswrapper[4846]: I1005 06:48:34.862719 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:34 crc kubenswrapper[4846]: I1005 06:48:34.862738 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:34 crc kubenswrapper[4846]: I1005 06:48:34.862763 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:34 crc kubenswrapper[4846]: I1005 06:48:34.862782 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:34Z","lastTransitionTime":"2025-10-05T06:48:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:34 crc kubenswrapper[4846]: I1005 06:48:34.873716 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c9f812b-ecc8-4985-a670-53fa80fbb499\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2b5ee9781c01d6aec3ac778efef1067b6c3bfe4ed37431bdebe188fcde0165f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:47:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3fb4717692aa1fec0012b514ef8d18d1ca3707cfa4bb8248bb93841a0a40f11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:47:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0b6a7afe3094f1c92093c05c28344897a615d590c2fbb507c87e18a44c6c831\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://640ade2c733d7916ec213a201a1dd614d261c69eab654c80100efbc99a768995\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:47:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:34Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:34 crc kubenswrapper[4846]: I1005 06:48:34.913658 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:34Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:34 crc kubenswrapper[4846]: I1005 06:48:34.946552 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e67d0e447a8a6a381259e25d25badc791194e1291f56b1cd77bd63989b20408\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dkhl7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://516f88f4990f5962d8b694a3d889259cc5508a309919870191c2e1331640c973\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dkhl7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-fscvf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:34Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:34 crc kubenswrapper[4846]: I1005 06:48:34.967309 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-w2sj8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32931fce-7374-4cab-a2e0-79d96a842274\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0df535d65d39de6c9a3dfb931bafc4469adc22b72f7a7903beda4fef8cd48138\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9rtxf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:27Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-w2sj8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:34Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:34 crc kubenswrapper[4846]: I1005 06:48:34.979937 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:34 crc kubenswrapper[4846]: I1005 06:48:34.979975 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:34 crc kubenswrapper[4846]: I1005 06:48:34.979983 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:34 crc kubenswrapper[4846]: I1005 06:48:34.979998 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:34 crc kubenswrapper[4846]: I1005 06:48:34.980008 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:34Z","lastTransitionTime":"2025-10-05T06:48:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:34 crc kubenswrapper[4846]: I1005 06:48:34.991848 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-7wm5q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9bc7744f-d700-450f-8b8d-98140877fee9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90379e74c544fccca44f7acdc228ff378154390e211ac801b79f8d074012241e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h5p4q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-7wm5q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:34Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:35 crc kubenswrapper[4846]: I1005 06:48:35.007910 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ff10c31459b3b3724ee7b769bd6f72f9ed80cce5419b6adcdf69a3239b70756\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:35Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:35 crc kubenswrapper[4846]: I1005 06:48:35.024608 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://347f900743b7bba3daf92f21c7eba0625e8c774f50057bc9472f0c82119325e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:35Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:35 crc kubenswrapper[4846]: I1005 06:48:35.042583 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56f1fe1f9f06f0249cf937a6963fda82467732dc056ca2b2189be6d8dd9d4a05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7212ce5f42c1fb55146abb190a557a504b43f16af6df84f90e1073567a329ec5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:35Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:35 crc kubenswrapper[4846]: I1005 06:48:35.059852 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"775b5702-a131-4d6c-852d-91bfa056cac5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0839c1765df2c68d2554e9c66d289505e559a9672caf8069feeff2c4fdfd2171\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15dbaf2445704da0987f256f81dff2ee95664988541fa48ce94898afeadab2e8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd3b3ab323c6d5153054aa04af2d91f7b9e8436ff1e2f75f3594b57831b2a416\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://67c46f5f0a3ace95bd2e9479c5561126664e18b148c5eceea5bf5efc27e81a67\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c52e076eddfd307110b1f6de00a169b81af8f2785af7becef2a820ce363a5861\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-05T06:48:14Z\\\",\\\"message\\\":\\\"le observer\\\\nW1005 06:48:14.337870 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1005 06:48:14.338086 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1005 06:48:14.339389 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1467752215/tls.crt::/tmp/serving-cert-1467752215/tls.key\\\\\\\"\\\\nI1005 06:48:14.864345 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1005 06:48:14.868087 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1005 06:48:14.868114 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1005 06:48:14.868140 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1005 06:48:14.868147 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1005 06:48:14.877354 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1005 06:48:14.877412 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1005 06:48:14.877456 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1005 06:48:14.877471 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1005 06:48:14.877490 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1005 06:48:14.877501 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1005 06:48:14.877507 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1005 06:48:14.877515 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1005 06:48:14.880326 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:13Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4a091716ff935adb42ae2e88048619fd347bd174f64e5f9c97a01f81690e57f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:01Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dc5411b642f99b13d514d2f21defad0c394aeeb82a966f72300076617e2586c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5dc5411b642f99b13d514d2f21defad0c394aeeb82a966f72300076617e2586c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:47:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:47:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:47:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:35Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:35 crc kubenswrapper[4846]: I1005 06:48:35.074753 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:35Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:35 crc kubenswrapper[4846]: I1005 06:48:35.083239 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:35 crc kubenswrapper[4846]: I1005 06:48:35.083293 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:35 crc kubenswrapper[4846]: I1005 06:48:35.083302 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:35 crc kubenswrapper[4846]: I1005 06:48:35.083321 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:35 crc kubenswrapper[4846]: I1005 06:48:35.083331 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:35Z","lastTransitionTime":"2025-10-05T06:48:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:35 crc kubenswrapper[4846]: I1005 06:48:35.096086 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-r7ztb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a5119cd9-af70-4324-a955-978306ab0b20\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06231a87429ec292cb57ebd88cda7ad547bcb32e80343b776833621a11c28ad1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jfcj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-r7ztb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:35Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:35 crc kubenswrapper[4846]: I1005 06:48:35.122637 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c482bb97-fd50-40cf-96ea-f806c3f01125\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bfe45d30e4fa141541aca1d46bc5cc1da0a7b4a5eb1fc9bb2caa5b7fdcb0b264\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://496a4782faa30a9a20ff63038141600cd635f12902d1cf17989cd067fcdfe929\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://281be077869a6aa1ca815770871a14413d4a97e97aeb550fd5cd215cba88b21f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6771c7215d2f5dd5838d17888375a4768b9d1755fc2e21bda1bbfbb2eb1dbb95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c17a711c2c7473b076a3944ffe7c700808d05cba582fbd55c5c2ac8ff73838a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44a9f740f0a06c8bf928e2a9d1d5979217558c31e73a333631baccebf00fd5d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://44a9f740f0a06c8bf928e2a9d1d5979217558c31e73a333631baccebf00fd5d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:47:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:47:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://48ca0ba6ca6092bfcb5b153220c054f02cc36fb22707c42e021f06d329e18e2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48ca0ba6ca6092bfcb5b153220c054f02cc36fb22707c42e021f06d329e18e2e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:00Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4006d0cea128675c8885d879171204abffc5e11ea0664750a12b4bc7344ccb2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4006d0cea128675c8885d879171204abffc5e11ea0664750a12b4bc7344ccb2e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:47:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:35Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:35 crc kubenswrapper[4846]: I1005 06:48:35.139911 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:35Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:35 crc kubenswrapper[4846]: I1005 06:48:35.165244 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-sfmxd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1dbb13ac-136c-4a1c-9229-b4e8ac4f9c71\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b0af6a6e735f7b7877c177c1d53d1a8dc12c4713a8059c4766fc858af31c9a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://226d977b596d07aeda0486452e2a8605949ed4d2453b21f56174937c395d6f36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://226d977b596d07aeda0486452e2a8605949ed4d2453b21f56174937c395d6f36\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://99c07a4bdb3059d3a99ffdaf794d668d52cffd9e5e3bb8945204dd6d1f8c8328\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99c07a4bdb3059d3a99ffdaf794d668d52cffd9e5e3bb8945204dd6d1f8c8328\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27095272bf4f31d63a78a436636de0aae12f7103597da944f24ed7a74329ee3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27095272bf4f31d63a78a436636de0aae12f7103597da944f24ed7a74329ee3f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1211571af66d8a94302b65feccf30768ba82b7cbad918e44dd405787dce484cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1211571af66d8a94302b65feccf30768ba82b7cbad918e44dd405787dce484cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27631f90d528900fa180c15cfce80e96d4cdba07743eb6d847e2e822d9917218\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27631f90d528900fa180c15cfce80e96d4cdba07743eb6d847e2e822d9917218\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f345693e9aeb1f4dc196349d3354f3c833a2b0415a693298b1cfccfa2b0d4459\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f345693e9aeb1f4dc196349d3354f3c833a2b0415a693298b1cfccfa2b0d4459\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-sfmxd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:35Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:35 crc kubenswrapper[4846]: I1005 06:48:35.186578 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:35 crc kubenswrapper[4846]: I1005 06:48:35.186651 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:35 crc kubenswrapper[4846]: I1005 06:48:35.186666 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:35 crc kubenswrapper[4846]: I1005 06:48:35.186693 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:35 crc kubenswrapper[4846]: I1005 06:48:35.186715 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:35Z","lastTransitionTime":"2025-10-05T06:48:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:35 crc kubenswrapper[4846]: I1005 06:48:35.192033 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-txzlk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"93a6ace4-8d64-44d7-9637-457d3af6543c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ad1940d07e37217089642478ece9ccdaca74784738a400929a9f5359d158c01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a6f6b2c32dd3da5f9fd5e6971c1773e81a9a24184b7aa5b543cec879001ae53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e9b0219ecbc7e60905d6de3e67a2c8dc9af1017b4aed9538ee0c58c861f2438\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af5e5a4da9ad7371d5a24633e8268f12e3a7d419a8c93b11ddbcc006d8339c6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8da85222775abc7a772f6a8ff8255f1a0d5b15807ba177a7e3e2b2d01a4f24b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1409651de8c897804f87be03d016c138f8acf27e662dce265a93e15c1848adcf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c212f5d230fae82a701afeceb39a848a9c91a3354908d88b062b7e1a726240cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c212f5d230fae82a701afeceb39a848a9c91a3354908d88b062b7e1a726240cb\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-05T06:48:34Z\\\",\\\"message\\\":\\\"g reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1005 06:48:34.496108 6146 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1005 06:48:34.496160 6146 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1005 06:48:34.496168 6146 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1005 06:48:34.496246 6146 factory.go:656] Stopping watch factory\\\\nI1005 06:48:34.496262 6146 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1005 06:48:34.496271 6146 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1005 06:48:34.496279 6146 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1005 06:48:34.496293 6146 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1005 06:48:34.496400 6146 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1005 06:48:34.496553 6146 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1005 06:48:34.496742 6146 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0df0ebcd87dc5af4b17736b43ec7b2395968b19ed16e775366af14e2b2d49c52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff3c6545ff3d2436da27bbd223b950b30b0c69a3e7f3034cb0e61380cb38d338\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ff3c6545ff3d2436da27bbd223b950b30b0c69a3e7f3034cb0e61380cb38d338\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:24Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-txzlk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:35Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:35 crc kubenswrapper[4846]: I1005 06:48:35.289856 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:35 crc kubenswrapper[4846]: I1005 06:48:35.289925 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:35 crc kubenswrapper[4846]: I1005 06:48:35.289940 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:35 crc kubenswrapper[4846]: I1005 06:48:35.289963 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:35 crc kubenswrapper[4846]: I1005 06:48:35.289983 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:35Z","lastTransitionTime":"2025-10-05T06:48:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:35 crc kubenswrapper[4846]: I1005 06:48:35.393696 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:35 crc kubenswrapper[4846]: I1005 06:48:35.393759 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:35 crc kubenswrapper[4846]: I1005 06:48:35.393773 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:35 crc kubenswrapper[4846]: I1005 06:48:35.393816 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:35 crc kubenswrapper[4846]: I1005 06:48:35.393833 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:35Z","lastTransitionTime":"2025-10-05T06:48:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:35 crc kubenswrapper[4846]: I1005 06:48:35.496988 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 05 06:48:35 crc kubenswrapper[4846]: I1005 06:48:35.497052 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 05 06:48:35 crc kubenswrapper[4846]: E1005 06:48:35.497169 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 05 06:48:35 crc kubenswrapper[4846]: I1005 06:48:35.497273 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:35 crc kubenswrapper[4846]: I1005 06:48:35.497310 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:35 crc kubenswrapper[4846]: I1005 06:48:35.497321 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:35 crc kubenswrapper[4846]: I1005 06:48:35.497341 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:35 crc kubenswrapper[4846]: E1005 06:48:35.497327 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 05 06:48:35 crc kubenswrapper[4846]: I1005 06:48:35.497353 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:35Z","lastTransitionTime":"2025-10-05T06:48:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:35 crc kubenswrapper[4846]: I1005 06:48:35.600310 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:35 crc kubenswrapper[4846]: I1005 06:48:35.600351 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:35 crc kubenswrapper[4846]: I1005 06:48:35.600361 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:35 crc kubenswrapper[4846]: I1005 06:48:35.600377 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:35 crc kubenswrapper[4846]: I1005 06:48:35.600389 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:35Z","lastTransitionTime":"2025-10-05T06:48:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:35 crc kubenswrapper[4846]: I1005 06:48:35.703729 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:35 crc kubenswrapper[4846]: I1005 06:48:35.703797 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:35 crc kubenswrapper[4846]: I1005 06:48:35.703808 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:35 crc kubenswrapper[4846]: I1005 06:48:35.703827 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:35 crc kubenswrapper[4846]: I1005 06:48:35.703851 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:35Z","lastTransitionTime":"2025-10-05T06:48:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:35 crc kubenswrapper[4846]: I1005 06:48:35.806659 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:35 crc kubenswrapper[4846]: I1005 06:48:35.806711 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:35 crc kubenswrapper[4846]: I1005 06:48:35.806721 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:35 crc kubenswrapper[4846]: I1005 06:48:35.806737 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:35 crc kubenswrapper[4846]: I1005 06:48:35.806747 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:35Z","lastTransitionTime":"2025-10-05T06:48:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:35 crc kubenswrapper[4846]: I1005 06:48:35.856522 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-txzlk_93a6ace4-8d64-44d7-9637-457d3af6543c/ovnkube-controller/0.log" Oct 05 06:48:35 crc kubenswrapper[4846]: I1005 06:48:35.859286 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-txzlk" event={"ID":"93a6ace4-8d64-44d7-9637-457d3af6543c","Type":"ContainerStarted","Data":"93149bae0504218735e0900b660ad9759ebbec3eb4dc214cc40ad349ac002d45"} Oct 05 06:48:35 crc kubenswrapper[4846]: I1005 06:48:35.859427 4846 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 05 06:48:35 crc kubenswrapper[4846]: I1005 06:48:35.904164 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c482bb97-fd50-40cf-96ea-f806c3f01125\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bfe45d30e4fa141541aca1d46bc5cc1da0a7b4a5eb1fc9bb2caa5b7fdcb0b264\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://496a4782faa30a9a20ff63038141600cd635f12902d1cf17989cd067fcdfe929\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://281be077869a6aa1ca815770871a14413d4a97e97aeb550fd5cd215cba88b21f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6771c7215d2f5dd5838d17888375a4768b9d1755fc2e21bda1bbfbb2eb1dbb95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c17a711c2c7473b076a3944ffe7c700808d05cba582fbd55c5c2ac8ff73838a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44a9f740f0a06c8bf928e2a9d1d5979217558c31e73a333631baccebf00fd5d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://44a9f740f0a06c8bf928e2a9d1d5979217558c31e73a333631baccebf00fd5d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:47:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:47:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://48ca0ba6ca6092bfcb5b153220c054f02cc36fb22707c42e021f06d329e18e2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48ca0ba6ca6092bfcb5b153220c054f02cc36fb22707c42e021f06d329e18e2e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:00Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4006d0cea128675c8885d879171204abffc5e11ea0664750a12b4bc7344ccb2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4006d0cea128675c8885d879171204abffc5e11ea0664750a12b4bc7344ccb2e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:47:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:35Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:35 crc kubenswrapper[4846]: I1005 06:48:35.909533 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:35 crc kubenswrapper[4846]: I1005 06:48:35.909612 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:35 crc kubenswrapper[4846]: I1005 06:48:35.909637 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:35 crc kubenswrapper[4846]: I1005 06:48:35.909668 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:35 crc kubenswrapper[4846]: I1005 06:48:35.909689 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:35Z","lastTransitionTime":"2025-10-05T06:48:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:35 crc kubenswrapper[4846]: I1005 06:48:35.931024 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:35Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:35 crc kubenswrapper[4846]: I1005 06:48:35.948908 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-sfmxd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1dbb13ac-136c-4a1c-9229-b4e8ac4f9c71\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b0af6a6e735f7b7877c177c1d53d1a8dc12c4713a8059c4766fc858af31c9a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://226d977b596d07aeda0486452e2a8605949ed4d2453b21f56174937c395d6f36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://226d977b596d07aeda0486452e2a8605949ed4d2453b21f56174937c395d6f36\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://99c07a4bdb3059d3a99ffdaf794d668d52cffd9e5e3bb8945204dd6d1f8c8328\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99c07a4bdb3059d3a99ffdaf794d668d52cffd9e5e3bb8945204dd6d1f8c8328\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27095272bf4f31d63a78a436636de0aae12f7103597da944f24ed7a74329ee3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27095272bf4f31d63a78a436636de0aae12f7103597da944f24ed7a74329ee3f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1211571af66d8a94302b65feccf30768ba82b7cbad918e44dd405787dce484cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1211571af66d8a94302b65feccf30768ba82b7cbad918e44dd405787dce484cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27631f90d528900fa180c15cfce80e96d4cdba07743eb6d847e2e822d9917218\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27631f90d528900fa180c15cfce80e96d4cdba07743eb6d847e2e822d9917218\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f345693e9aeb1f4dc196349d3354f3c833a2b0415a693298b1cfccfa2b0d4459\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f345693e9aeb1f4dc196349d3354f3c833a2b0415a693298b1cfccfa2b0d4459\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-sfmxd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:35Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:35 crc kubenswrapper[4846]: I1005 06:48:35.970075 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-txzlk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"93a6ace4-8d64-44d7-9637-457d3af6543c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ad1940d07e37217089642478ece9ccdaca74784738a400929a9f5359d158c01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a6f6b2c32dd3da5f9fd5e6971c1773e81a9a24184b7aa5b543cec879001ae53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e9b0219ecbc7e60905d6de3e67a2c8dc9af1017b4aed9538ee0c58c861f2438\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af5e5a4da9ad7371d5a24633e8268f12e3a7d419a8c93b11ddbcc006d8339c6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8da85222775abc7a772f6a8ff8255f1a0d5b15807ba177a7e3e2b2d01a4f24b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1409651de8c897804f87be03d016c138f8acf27e662dce265a93e15c1848adcf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://93149bae0504218735e0900b660ad9759ebbec3eb4dc214cc40ad349ac002d45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c212f5d230fae82a701afeceb39a848a9c91a3354908d88b062b7e1a726240cb\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-05T06:48:34Z\\\",\\\"message\\\":\\\"g reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1005 06:48:34.496108 6146 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1005 06:48:34.496160 6146 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1005 06:48:34.496168 6146 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1005 06:48:34.496246 6146 factory.go:656] Stopping watch factory\\\\nI1005 06:48:34.496262 6146 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1005 06:48:34.496271 6146 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1005 06:48:34.496279 6146 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1005 06:48:34.496293 6146 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1005 06:48:34.496400 6146 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1005 06:48:34.496553 6146 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1005 06:48:34.496742 6146 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:31Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0df0ebcd87dc5af4b17736b43ec7b2395968b19ed16e775366af14e2b2d49c52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff3c6545ff3d2436da27bbd223b950b30b0c69a3e7f3034cb0e61380cb38d338\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ff3c6545ff3d2436da27bbd223b950b30b0c69a3e7f3034cb0e61380cb38d338\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:24Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-txzlk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:35Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:35 crc kubenswrapper[4846]: I1005 06:48:35.986808 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c9f812b-ecc8-4985-a670-53fa80fbb499\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2b5ee9781c01d6aec3ac778efef1067b6c3bfe4ed37431bdebe188fcde0165f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:47:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3fb4717692aa1fec0012b514ef8d18d1ca3707cfa4bb8248bb93841a0a40f11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:47:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0b6a7afe3094f1c92093c05c28344897a615d590c2fbb507c87e18a44c6c831\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://640ade2c733d7916ec213a201a1dd614d261c69eab654c80100efbc99a768995\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:47:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:35Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:36 crc kubenswrapper[4846]: I1005 06:48:36.004469 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:36Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:36 crc kubenswrapper[4846]: I1005 06:48:36.013004 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:36 crc kubenswrapper[4846]: I1005 06:48:36.013063 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:36 crc kubenswrapper[4846]: I1005 06:48:36.013089 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:36 crc kubenswrapper[4846]: I1005 06:48:36.013121 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:36 crc kubenswrapper[4846]: I1005 06:48:36.013146 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:36Z","lastTransitionTime":"2025-10-05T06:48:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:36 crc kubenswrapper[4846]: I1005 06:48:36.021985 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e67d0e447a8a6a381259e25d25badc791194e1291f56b1cd77bd63989b20408\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dkhl7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://516f88f4990f5962d8b694a3d889259cc5508a309919870191c2e1331640c973\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dkhl7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-fscvf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:36Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:36 crc kubenswrapper[4846]: I1005 06:48:36.038939 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-w2sj8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32931fce-7374-4cab-a2e0-79d96a842274\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0df535d65d39de6c9a3dfb931bafc4469adc22b72f7a7903beda4fef8cd48138\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9rtxf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:27Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-w2sj8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:36Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:36 crc kubenswrapper[4846]: I1005 06:48:36.063690 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ff10c31459b3b3724ee7b769bd6f72f9ed80cce5419b6adcdf69a3239b70756\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:36Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:36 crc kubenswrapper[4846]: I1005 06:48:36.084299 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://347f900743b7bba3daf92f21c7eba0625e8c774f50057bc9472f0c82119325e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:36Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:36 crc kubenswrapper[4846]: I1005 06:48:36.105237 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56f1fe1f9f06f0249cf937a6963fda82467732dc056ca2b2189be6d8dd9d4a05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7212ce5f42c1fb55146abb190a557a504b43f16af6df84f90e1073567a329ec5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:36Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:36 crc kubenswrapper[4846]: I1005 06:48:36.116046 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:36 crc kubenswrapper[4846]: I1005 06:48:36.116390 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:36 crc kubenswrapper[4846]: I1005 06:48:36.116692 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:36 crc kubenswrapper[4846]: I1005 06:48:36.117003 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:36 crc kubenswrapper[4846]: I1005 06:48:36.117324 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:36Z","lastTransitionTime":"2025-10-05T06:48:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:36 crc kubenswrapper[4846]: I1005 06:48:36.125076 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-7wm5q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9bc7744f-d700-450f-8b8d-98140877fee9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90379e74c544fccca44f7acdc228ff378154390e211ac801b79f8d074012241e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h5p4q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-7wm5q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:36Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:36 crc kubenswrapper[4846]: I1005 06:48:36.146241 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"775b5702-a131-4d6c-852d-91bfa056cac5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0839c1765df2c68d2554e9c66d289505e559a9672caf8069feeff2c4fdfd2171\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15dbaf2445704da0987f256f81dff2ee95664988541fa48ce94898afeadab2e8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd3b3ab323c6d5153054aa04af2d91f7b9e8436ff1e2f75f3594b57831b2a416\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://67c46f5f0a3ace95bd2e9479c5561126664e18b148c5eceea5bf5efc27e81a67\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c52e076eddfd307110b1f6de00a169b81af8f2785af7becef2a820ce363a5861\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-05T06:48:14Z\\\",\\\"message\\\":\\\"le observer\\\\nW1005 06:48:14.337870 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1005 06:48:14.338086 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1005 06:48:14.339389 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1467752215/tls.crt::/tmp/serving-cert-1467752215/tls.key\\\\\\\"\\\\nI1005 06:48:14.864345 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1005 06:48:14.868087 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1005 06:48:14.868114 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1005 06:48:14.868140 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1005 06:48:14.868147 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1005 06:48:14.877354 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1005 06:48:14.877412 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1005 06:48:14.877456 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1005 06:48:14.877471 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1005 06:48:14.877490 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1005 06:48:14.877501 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1005 06:48:14.877507 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1005 06:48:14.877515 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1005 06:48:14.880326 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:13Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4a091716ff935adb42ae2e88048619fd347bd174f64e5f9c97a01f81690e57f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:01Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dc5411b642f99b13d514d2f21defad0c394aeeb82a966f72300076617e2586c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5dc5411b642f99b13d514d2f21defad0c394aeeb82a966f72300076617e2586c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:47:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:47:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:47:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:36Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:36 crc kubenswrapper[4846]: I1005 06:48:36.166533 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:36Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:36 crc kubenswrapper[4846]: I1005 06:48:36.185267 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-r7ztb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a5119cd9-af70-4324-a955-978306ab0b20\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06231a87429ec292cb57ebd88cda7ad547bcb32e80343b776833621a11c28ad1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jfcj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-r7ztb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:36Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:36 crc kubenswrapper[4846]: I1005 06:48:36.220156 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:36 crc kubenswrapper[4846]: I1005 06:48:36.220225 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:36 crc kubenswrapper[4846]: I1005 06:48:36.220236 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:36 crc kubenswrapper[4846]: I1005 06:48:36.220280 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:36 crc kubenswrapper[4846]: I1005 06:48:36.220293 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:36Z","lastTransitionTime":"2025-10-05T06:48:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:36 crc kubenswrapper[4846]: I1005 06:48:36.323081 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:36 crc kubenswrapper[4846]: I1005 06:48:36.323119 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:36 crc kubenswrapper[4846]: I1005 06:48:36.323127 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:36 crc kubenswrapper[4846]: I1005 06:48:36.323142 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:36 crc kubenswrapper[4846]: I1005 06:48:36.323151 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:36Z","lastTransitionTime":"2025-10-05T06:48:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:36 crc kubenswrapper[4846]: I1005 06:48:36.426083 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:36 crc kubenswrapper[4846]: I1005 06:48:36.426127 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:36 crc kubenswrapper[4846]: I1005 06:48:36.426135 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:36 crc kubenswrapper[4846]: I1005 06:48:36.426151 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:36 crc kubenswrapper[4846]: I1005 06:48:36.426162 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:36Z","lastTransitionTime":"2025-10-05T06:48:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:36 crc kubenswrapper[4846]: I1005 06:48:36.497420 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 05 06:48:36 crc kubenswrapper[4846]: E1005 06:48:36.497654 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 05 06:48:36 crc kubenswrapper[4846]: I1005 06:48:36.529958 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:36 crc kubenswrapper[4846]: I1005 06:48:36.530036 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:36 crc kubenswrapper[4846]: I1005 06:48:36.530056 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:36 crc kubenswrapper[4846]: I1005 06:48:36.530084 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:36 crc kubenswrapper[4846]: I1005 06:48:36.530107 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:36Z","lastTransitionTime":"2025-10-05T06:48:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:36 crc kubenswrapper[4846]: I1005 06:48:36.633498 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:36 crc kubenswrapper[4846]: I1005 06:48:36.633543 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:36 crc kubenswrapper[4846]: I1005 06:48:36.633554 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:36 crc kubenswrapper[4846]: I1005 06:48:36.633574 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:36 crc kubenswrapper[4846]: I1005 06:48:36.633587 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:36Z","lastTransitionTime":"2025-10-05T06:48:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:36 crc kubenswrapper[4846]: I1005 06:48:36.737116 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:36 crc kubenswrapper[4846]: I1005 06:48:36.737169 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:36 crc kubenswrapper[4846]: I1005 06:48:36.737200 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:36 crc kubenswrapper[4846]: I1005 06:48:36.737218 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:36 crc kubenswrapper[4846]: I1005 06:48:36.737231 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:36Z","lastTransitionTime":"2025-10-05T06:48:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:36 crc kubenswrapper[4846]: I1005 06:48:36.840602 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:36 crc kubenswrapper[4846]: I1005 06:48:36.840639 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:36 crc kubenswrapper[4846]: I1005 06:48:36.840650 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:36 crc kubenswrapper[4846]: I1005 06:48:36.840666 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:36 crc kubenswrapper[4846]: I1005 06:48:36.840676 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:36Z","lastTransitionTime":"2025-10-05T06:48:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:36 crc kubenswrapper[4846]: I1005 06:48:36.869491 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-txzlk_93a6ace4-8d64-44d7-9637-457d3af6543c/ovnkube-controller/1.log" Oct 05 06:48:36 crc kubenswrapper[4846]: I1005 06:48:36.870428 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-txzlk_93a6ace4-8d64-44d7-9637-457d3af6543c/ovnkube-controller/0.log" Oct 05 06:48:36 crc kubenswrapper[4846]: I1005 06:48:36.873979 4846 generic.go:334] "Generic (PLEG): container finished" podID="93a6ace4-8d64-44d7-9637-457d3af6543c" containerID="93149bae0504218735e0900b660ad9759ebbec3eb4dc214cc40ad349ac002d45" exitCode=1 Oct 05 06:48:36 crc kubenswrapper[4846]: I1005 06:48:36.874034 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-txzlk" event={"ID":"93a6ace4-8d64-44d7-9637-457d3af6543c","Type":"ContainerDied","Data":"93149bae0504218735e0900b660ad9759ebbec3eb4dc214cc40ad349ac002d45"} Oct 05 06:48:36 crc kubenswrapper[4846]: I1005 06:48:36.874073 4846 scope.go:117] "RemoveContainer" containerID="c212f5d230fae82a701afeceb39a848a9c91a3354908d88b062b7e1a726240cb" Oct 05 06:48:36 crc kubenswrapper[4846]: I1005 06:48:36.875557 4846 scope.go:117] "RemoveContainer" containerID="93149bae0504218735e0900b660ad9759ebbec3eb4dc214cc40ad349ac002d45" Oct 05 06:48:36 crc kubenswrapper[4846]: E1005 06:48:36.875997 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-txzlk_openshift-ovn-kubernetes(93a6ace4-8d64-44d7-9637-457d3af6543c)\"" pod="openshift-ovn-kubernetes/ovnkube-node-txzlk" podUID="93a6ace4-8d64-44d7-9637-457d3af6543c" Oct 05 06:48:36 crc kubenswrapper[4846]: I1005 06:48:36.895561 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"775b5702-a131-4d6c-852d-91bfa056cac5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0839c1765df2c68d2554e9c66d289505e559a9672caf8069feeff2c4fdfd2171\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15dbaf2445704da0987f256f81dff2ee95664988541fa48ce94898afeadab2e8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd3b3ab323c6d5153054aa04af2d91f7b9e8436ff1e2f75f3594b57831b2a416\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://67c46f5f0a3ace95bd2e9479c5561126664e18b148c5eceea5bf5efc27e81a67\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c52e076eddfd307110b1f6de00a169b81af8f2785af7becef2a820ce363a5861\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-05T06:48:14Z\\\",\\\"message\\\":\\\"le observer\\\\nW1005 06:48:14.337870 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1005 06:48:14.338086 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1005 06:48:14.339389 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1467752215/tls.crt::/tmp/serving-cert-1467752215/tls.key\\\\\\\"\\\\nI1005 06:48:14.864345 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1005 06:48:14.868087 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1005 06:48:14.868114 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1005 06:48:14.868140 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1005 06:48:14.868147 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1005 06:48:14.877354 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1005 06:48:14.877412 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1005 06:48:14.877456 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1005 06:48:14.877471 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1005 06:48:14.877490 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1005 06:48:14.877501 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1005 06:48:14.877507 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1005 06:48:14.877515 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1005 06:48:14.880326 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:13Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4a091716ff935adb42ae2e88048619fd347bd174f64e5f9c97a01f81690e57f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:01Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dc5411b642f99b13d514d2f21defad0c394aeeb82a966f72300076617e2586c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5dc5411b642f99b13d514d2f21defad0c394aeeb82a966f72300076617e2586c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:47:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:47:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:47:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:36Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:36 crc kubenswrapper[4846]: I1005 06:48:36.911711 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:36Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:36 crc kubenswrapper[4846]: I1005 06:48:36.925157 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-r7ztb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a5119cd9-af70-4324-a955-978306ab0b20\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06231a87429ec292cb57ebd88cda7ad547bcb32e80343b776833621a11c28ad1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jfcj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-r7ztb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:36Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:36 crc kubenswrapper[4846]: I1005 06:48:36.943134 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:36 crc kubenswrapper[4846]: I1005 06:48:36.943258 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:36 crc kubenswrapper[4846]: I1005 06:48:36.943278 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:36 crc kubenswrapper[4846]: I1005 06:48:36.943303 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:36 crc kubenswrapper[4846]: I1005 06:48:36.943322 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:36Z","lastTransitionTime":"2025-10-05T06:48:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:36 crc kubenswrapper[4846]: I1005 06:48:36.957687 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c482bb97-fd50-40cf-96ea-f806c3f01125\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bfe45d30e4fa141541aca1d46bc5cc1da0a7b4a5eb1fc9bb2caa5b7fdcb0b264\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://496a4782faa30a9a20ff63038141600cd635f12902d1cf17989cd067fcdfe929\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://281be077869a6aa1ca815770871a14413d4a97e97aeb550fd5cd215cba88b21f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6771c7215d2f5dd5838d17888375a4768b9d1755fc2e21bda1bbfbb2eb1dbb95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c17a711c2c7473b076a3944ffe7c700808d05cba582fbd55c5c2ac8ff73838a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44a9f740f0a06c8bf928e2a9d1d5979217558c31e73a333631baccebf00fd5d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://44a9f740f0a06c8bf928e2a9d1d5979217558c31e73a333631baccebf00fd5d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:47:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:47:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://48ca0ba6ca6092bfcb5b153220c054f02cc36fb22707c42e021f06d329e18e2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48ca0ba6ca6092bfcb5b153220c054f02cc36fb22707c42e021f06d329e18e2e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:00Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4006d0cea128675c8885d879171204abffc5e11ea0664750a12b4bc7344ccb2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4006d0cea128675c8885d879171204abffc5e11ea0664750a12b4bc7344ccb2e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:47:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:36Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:36 crc kubenswrapper[4846]: I1005 06:48:36.978607 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:36Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:37 crc kubenswrapper[4846]: I1005 06:48:37.000311 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-sfmxd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1dbb13ac-136c-4a1c-9229-b4e8ac4f9c71\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b0af6a6e735f7b7877c177c1d53d1a8dc12c4713a8059c4766fc858af31c9a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://226d977b596d07aeda0486452e2a8605949ed4d2453b21f56174937c395d6f36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://226d977b596d07aeda0486452e2a8605949ed4d2453b21f56174937c395d6f36\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://99c07a4bdb3059d3a99ffdaf794d668d52cffd9e5e3bb8945204dd6d1f8c8328\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99c07a4bdb3059d3a99ffdaf794d668d52cffd9e5e3bb8945204dd6d1f8c8328\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27095272bf4f31d63a78a436636de0aae12f7103597da944f24ed7a74329ee3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27095272bf4f31d63a78a436636de0aae12f7103597da944f24ed7a74329ee3f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1211571af66d8a94302b65feccf30768ba82b7cbad918e44dd405787dce484cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1211571af66d8a94302b65feccf30768ba82b7cbad918e44dd405787dce484cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27631f90d528900fa180c15cfce80e96d4cdba07743eb6d847e2e822d9917218\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27631f90d528900fa180c15cfce80e96d4cdba07743eb6d847e2e822d9917218\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f345693e9aeb1f4dc196349d3354f3c833a2b0415a693298b1cfccfa2b0d4459\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f345693e9aeb1f4dc196349d3354f3c833a2b0415a693298b1cfccfa2b0d4459\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-sfmxd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:36Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:37 crc kubenswrapper[4846]: I1005 06:48:37.030540 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-txzlk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"93a6ace4-8d64-44d7-9637-457d3af6543c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ad1940d07e37217089642478ece9ccdaca74784738a400929a9f5359d158c01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a6f6b2c32dd3da5f9fd5e6971c1773e81a9a24184b7aa5b543cec879001ae53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e9b0219ecbc7e60905d6de3e67a2c8dc9af1017b4aed9538ee0c58c861f2438\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af5e5a4da9ad7371d5a24633e8268f12e3a7d419a8c93b11ddbcc006d8339c6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8da85222775abc7a772f6a8ff8255f1a0d5b15807ba177a7e3e2b2d01a4f24b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1409651de8c897804f87be03d016c138f8acf27e662dce265a93e15c1848adcf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://93149bae0504218735e0900b660ad9759ebbec3eb4dc214cc40ad349ac002d45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c212f5d230fae82a701afeceb39a848a9c91a3354908d88b062b7e1a726240cb\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-05T06:48:34Z\\\",\\\"message\\\":\\\"g reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1005 06:48:34.496108 6146 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1005 06:48:34.496160 6146 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1005 06:48:34.496168 6146 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1005 06:48:34.496246 6146 factory.go:656] Stopping watch factory\\\\nI1005 06:48:34.496262 6146 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1005 06:48:34.496271 6146 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1005 06:48:34.496279 6146 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1005 06:48:34.496293 6146 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1005 06:48:34.496400 6146 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1005 06:48:34.496553 6146 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1005 06:48:34.496742 6146 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:31Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://93149bae0504218735e0900b660ad9759ebbec3eb4dc214cc40ad349ac002d45\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-05T06:48:36Z\\\",\\\"message\\\":\\\"ePorts:nil,LoadBalancerClass:nil,InternalTrafficPolicy:*Cluster,TrafficDistribution:nil,},Status:ServiceStatus{LoadBalancer:LoadBalancerStatus{Ingress:[]LoadBalancerIngress{},},Conditions:[]Condition{},},}\\\\nI1005 06:48:35.889902 6320 lb_config.go:1031] Cluster endpoints for openshift-kube-apiserver/apiserver for network=default are: map[]\\\\nI1005 06:48:35.889910 6320 services_controller.go:443] Built service openshift-kube-apiserver/apiserver LB cluster-wide configs for network=default: []services.lbConfig{services.lbConfig{vips:[]string{\\\\\\\"10.217.4.93\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:443, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI1005 06:48:35.889925 6320 services_controller.go:444] Built service openshift-kube-apiserver/apiserver LB per-node configs for network=default: []services.lbConfig(nil)\\\\nF1005 06:48:35.889931 6320 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: failed to add event handler: handler {0x1e60340 0x1e60020 0x1e5ffc0} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to se\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0df0ebcd87dc5af4b17736b43ec7b2395968b19ed16e775366af14e2b2d49c52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff3c6545ff3d2436da27bbd223b950b30b0c69a3e7f3034cb0e61380cb38d338\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ff3c6545ff3d2436da27bbd223b950b30b0c69a3e7f3034cb0e61380cb38d338\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:24Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-txzlk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:37Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:37 crc kubenswrapper[4846]: I1005 06:48:37.047080 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:37 crc kubenswrapper[4846]: I1005 06:48:37.047136 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:37 crc kubenswrapper[4846]: I1005 06:48:37.047151 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:37 crc kubenswrapper[4846]: I1005 06:48:37.047196 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:37 crc kubenswrapper[4846]: I1005 06:48:37.047214 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:37Z","lastTransitionTime":"2025-10-05T06:48:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:37 crc kubenswrapper[4846]: I1005 06:48:37.052206 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c9f812b-ecc8-4985-a670-53fa80fbb499\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2b5ee9781c01d6aec3ac778efef1067b6c3bfe4ed37431bdebe188fcde0165f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:47:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3fb4717692aa1fec0012b514ef8d18d1ca3707cfa4bb8248bb93841a0a40f11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:47:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0b6a7afe3094f1c92093c05c28344897a615d590c2fbb507c87e18a44c6c831\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://640ade2c733d7916ec213a201a1dd614d261c69eab654c80100efbc99a768995\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:47:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:37Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:37 crc kubenswrapper[4846]: I1005 06:48:37.071577 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:37Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:37 crc kubenswrapper[4846]: I1005 06:48:37.087984 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e67d0e447a8a6a381259e25d25badc791194e1291f56b1cd77bd63989b20408\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dkhl7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://516f88f4990f5962d8b694a3d889259cc5508a309919870191c2e1331640c973\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dkhl7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-fscvf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:37Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:37 crc kubenswrapper[4846]: I1005 06:48:37.103693 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-w2sj8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32931fce-7374-4cab-a2e0-79d96a842274\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0df535d65d39de6c9a3dfb931bafc4469adc22b72f7a7903beda4fef8cd48138\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9rtxf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:27Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-w2sj8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:37Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:37 crc kubenswrapper[4846]: I1005 06:48:37.129998 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ff10c31459b3b3724ee7b769bd6f72f9ed80cce5419b6adcdf69a3239b70756\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:37Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:37 crc kubenswrapper[4846]: I1005 06:48:37.147526 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://347f900743b7bba3daf92f21c7eba0625e8c774f50057bc9472f0c82119325e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:37Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:37 crc kubenswrapper[4846]: I1005 06:48:37.150349 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:37 crc kubenswrapper[4846]: I1005 06:48:37.150408 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:37 crc kubenswrapper[4846]: I1005 06:48:37.150428 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:37 crc kubenswrapper[4846]: I1005 06:48:37.150459 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:37 crc kubenswrapper[4846]: I1005 06:48:37.150479 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:37Z","lastTransitionTime":"2025-10-05T06:48:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:37 crc kubenswrapper[4846]: I1005 06:48:37.171588 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56f1fe1f9f06f0249cf937a6963fda82467732dc056ca2b2189be6d8dd9d4a05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7212ce5f42c1fb55146abb190a557a504b43f16af6df84f90e1073567a329ec5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:37Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:37 crc kubenswrapper[4846]: I1005 06:48:37.192993 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-7wm5q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9bc7744f-d700-450f-8b8d-98140877fee9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90379e74c544fccca44f7acdc228ff378154390e211ac801b79f8d074012241e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h5p4q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-7wm5q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:37Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:37 crc kubenswrapper[4846]: I1005 06:48:37.253778 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:37 crc kubenswrapper[4846]: I1005 06:48:37.253853 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:37 crc kubenswrapper[4846]: I1005 06:48:37.253871 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:37 crc kubenswrapper[4846]: I1005 06:48:37.253899 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:37 crc kubenswrapper[4846]: I1005 06:48:37.253919 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:37Z","lastTransitionTime":"2025-10-05T06:48:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:37 crc kubenswrapper[4846]: I1005 06:48:37.358058 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:37 crc kubenswrapper[4846]: I1005 06:48:37.358390 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:37 crc kubenswrapper[4846]: I1005 06:48:37.358501 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:37 crc kubenswrapper[4846]: I1005 06:48:37.358608 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:37 crc kubenswrapper[4846]: I1005 06:48:37.358691 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:37Z","lastTransitionTime":"2025-10-05T06:48:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:37 crc kubenswrapper[4846]: I1005 06:48:37.462121 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:37 crc kubenswrapper[4846]: I1005 06:48:37.462171 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:37 crc kubenswrapper[4846]: I1005 06:48:37.462194 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:37 crc kubenswrapper[4846]: I1005 06:48:37.462213 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:37 crc kubenswrapper[4846]: I1005 06:48:37.462224 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:37Z","lastTransitionTime":"2025-10-05T06:48:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:37 crc kubenswrapper[4846]: I1005 06:48:37.496751 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 05 06:48:37 crc kubenswrapper[4846]: I1005 06:48:37.496763 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 05 06:48:37 crc kubenswrapper[4846]: E1005 06:48:37.496934 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 05 06:48:37 crc kubenswrapper[4846]: E1005 06:48:37.497143 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 05 06:48:37 crc kubenswrapper[4846]: I1005 06:48:37.565793 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:37 crc kubenswrapper[4846]: I1005 06:48:37.565845 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:37 crc kubenswrapper[4846]: I1005 06:48:37.565863 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:37 crc kubenswrapper[4846]: I1005 06:48:37.565892 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:37 crc kubenswrapper[4846]: I1005 06:48:37.565913 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:37Z","lastTransitionTime":"2025-10-05T06:48:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:37 crc kubenswrapper[4846]: I1005 06:48:37.670085 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:37 crc kubenswrapper[4846]: I1005 06:48:37.670211 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:37 crc kubenswrapper[4846]: I1005 06:48:37.670243 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:37 crc kubenswrapper[4846]: I1005 06:48:37.670282 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:37 crc kubenswrapper[4846]: I1005 06:48:37.670308 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:37Z","lastTransitionTime":"2025-10-05T06:48:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:37 crc kubenswrapper[4846]: I1005 06:48:37.738099 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-gklb6"] Oct 05 06:48:37 crc kubenswrapper[4846]: I1005 06:48:37.739290 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-gklb6" Oct 05 06:48:37 crc kubenswrapper[4846]: I1005 06:48:37.742731 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Oct 05 06:48:37 crc kubenswrapper[4846]: I1005 06:48:37.742830 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Oct 05 06:48:37 crc kubenswrapper[4846]: I1005 06:48:37.761425 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://347f900743b7bba3daf92f21c7eba0625e8c774f50057bc9472f0c82119325e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:37Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:37 crc kubenswrapper[4846]: I1005 06:48:37.773557 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:37 crc kubenswrapper[4846]: I1005 06:48:37.773625 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:37 crc kubenswrapper[4846]: I1005 06:48:37.773645 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:37 crc kubenswrapper[4846]: I1005 06:48:37.773675 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:37 crc kubenswrapper[4846]: I1005 06:48:37.773696 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:37Z","lastTransitionTime":"2025-10-05T06:48:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:37 crc kubenswrapper[4846]: I1005 06:48:37.782959 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56f1fe1f9f06f0249cf937a6963fda82467732dc056ca2b2189be6d8dd9d4a05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7212ce5f42c1fb55146abb190a557a504b43f16af6df84f90e1073567a329ec5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:37Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:37 crc kubenswrapper[4846]: I1005 06:48:37.804304 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-7wm5q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9bc7744f-d700-450f-8b8d-98140877fee9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90379e74c544fccca44f7acdc228ff378154390e211ac801b79f8d074012241e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h5p4q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-7wm5q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:37Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:37 crc kubenswrapper[4846]: I1005 06:48:37.826314 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ff10c31459b3b3724ee7b769bd6f72f9ed80cce5419b6adcdf69a3239b70756\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:37Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:37 crc kubenswrapper[4846]: I1005 06:48:37.846976 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:37Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:37 crc kubenswrapper[4846]: I1005 06:48:37.856216 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 05 06:48:37 crc kubenswrapper[4846]: I1005 06:48:37.864011 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/698ab882-bb26-4b4c-81e9-8305a44fe068-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-gklb6\" (UID: \"698ab882-bb26-4b4c-81e9-8305a44fe068\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-gklb6" Oct 05 06:48:37 crc kubenswrapper[4846]: I1005 06:48:37.864063 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/698ab882-bb26-4b4c-81e9-8305a44fe068-env-overrides\") pod \"ovnkube-control-plane-749d76644c-gklb6\" (UID: \"698ab882-bb26-4b4c-81e9-8305a44fe068\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-gklb6" Oct 05 06:48:37 crc kubenswrapper[4846]: I1005 06:48:37.864158 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/698ab882-bb26-4b4c-81e9-8305a44fe068-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-gklb6\" (UID: \"698ab882-bb26-4b4c-81e9-8305a44fe068\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-gklb6" Oct 05 06:48:37 crc kubenswrapper[4846]: I1005 06:48:37.864228 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-r7ztb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a5119cd9-af70-4324-a955-978306ab0b20\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06231a87429ec292cb57ebd88cda7ad547bcb32e80343b776833621a11c28ad1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jfcj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-r7ztb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:37Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:37 crc kubenswrapper[4846]: I1005 06:48:37.864303 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x6d6d\" (UniqueName: \"kubernetes.io/projected/698ab882-bb26-4b4c-81e9-8305a44fe068-kube-api-access-x6d6d\") pod \"ovnkube-control-plane-749d76644c-gklb6\" (UID: \"698ab882-bb26-4b4c-81e9-8305a44fe068\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-gklb6" Oct 05 06:48:37 crc kubenswrapper[4846]: I1005 06:48:37.877272 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:37 crc kubenswrapper[4846]: I1005 06:48:37.877315 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:37 crc kubenswrapper[4846]: I1005 06:48:37.877338 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:37 crc kubenswrapper[4846]: I1005 06:48:37.877368 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:37 crc kubenswrapper[4846]: I1005 06:48:37.877388 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:37Z","lastTransitionTime":"2025-10-05T06:48:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:37 crc kubenswrapper[4846]: I1005 06:48:37.880576 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-txzlk_93a6ace4-8d64-44d7-9637-457d3af6543c/ovnkube-controller/1.log" Oct 05 06:48:37 crc kubenswrapper[4846]: I1005 06:48:37.889382 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"775b5702-a131-4d6c-852d-91bfa056cac5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0839c1765df2c68d2554e9c66d289505e559a9672caf8069feeff2c4fdfd2171\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15dbaf2445704da0987f256f81dff2ee95664988541fa48ce94898afeadab2e8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd3b3ab323c6d5153054aa04af2d91f7b9e8436ff1e2f75f3594b57831b2a416\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://67c46f5f0a3ace95bd2e9479c5561126664e18b148c5eceea5bf5efc27e81a67\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c52e076eddfd307110b1f6de00a169b81af8f2785af7becef2a820ce363a5861\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-05T06:48:14Z\\\",\\\"message\\\":\\\"le observer\\\\nW1005 06:48:14.337870 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1005 06:48:14.338086 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1005 06:48:14.339389 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1467752215/tls.crt::/tmp/serving-cert-1467752215/tls.key\\\\\\\"\\\\nI1005 06:48:14.864345 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1005 06:48:14.868087 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1005 06:48:14.868114 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1005 06:48:14.868140 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1005 06:48:14.868147 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1005 06:48:14.877354 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1005 06:48:14.877412 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1005 06:48:14.877456 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1005 06:48:14.877471 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1005 06:48:14.877490 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1005 06:48:14.877501 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1005 06:48:14.877507 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1005 06:48:14.877515 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1005 06:48:14.880326 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:13Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4a091716ff935adb42ae2e88048619fd347bd174f64e5f9c97a01f81690e57f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:01Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dc5411b642f99b13d514d2f21defad0c394aeeb82a966f72300076617e2586c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5dc5411b642f99b13d514d2f21defad0c394aeeb82a966f72300076617e2586c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:47:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:47:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:47:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:37Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:37 crc kubenswrapper[4846]: I1005 06:48:37.918142 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-sfmxd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1dbb13ac-136c-4a1c-9229-b4e8ac4f9c71\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b0af6a6e735f7b7877c177c1d53d1a8dc12c4713a8059c4766fc858af31c9a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://226d977b596d07aeda0486452e2a8605949ed4d2453b21f56174937c395d6f36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://226d977b596d07aeda0486452e2a8605949ed4d2453b21f56174937c395d6f36\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://99c07a4bdb3059d3a99ffdaf794d668d52cffd9e5e3bb8945204dd6d1f8c8328\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99c07a4bdb3059d3a99ffdaf794d668d52cffd9e5e3bb8945204dd6d1f8c8328\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27095272bf4f31d63a78a436636de0aae12f7103597da944f24ed7a74329ee3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27095272bf4f31d63a78a436636de0aae12f7103597da944f24ed7a74329ee3f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1211571af66d8a94302b65feccf30768ba82b7cbad918e44dd405787dce484cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1211571af66d8a94302b65feccf30768ba82b7cbad918e44dd405787dce484cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27631f90d528900fa180c15cfce80e96d4cdba07743eb6d847e2e822d9917218\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27631f90d528900fa180c15cfce80e96d4cdba07743eb6d847e2e822d9917218\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f345693e9aeb1f4dc196349d3354f3c833a2b0415a693298b1cfccfa2b0d4459\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f345693e9aeb1f4dc196349d3354f3c833a2b0415a693298b1cfccfa2b0d4459\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-sfmxd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:37Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:37 crc kubenswrapper[4846]: I1005 06:48:37.950542 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-txzlk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"93a6ace4-8d64-44d7-9637-457d3af6543c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ad1940d07e37217089642478ece9ccdaca74784738a400929a9f5359d158c01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a6f6b2c32dd3da5f9fd5e6971c1773e81a9a24184b7aa5b543cec879001ae53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e9b0219ecbc7e60905d6de3e67a2c8dc9af1017b4aed9538ee0c58c861f2438\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af5e5a4da9ad7371d5a24633e8268f12e3a7d419a8c93b11ddbcc006d8339c6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8da85222775abc7a772f6a8ff8255f1a0d5b15807ba177a7e3e2b2d01a4f24b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1409651de8c897804f87be03d016c138f8acf27e662dce265a93e15c1848adcf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://93149bae0504218735e0900b660ad9759ebbec3eb4dc214cc40ad349ac002d45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c212f5d230fae82a701afeceb39a848a9c91a3354908d88b062b7e1a726240cb\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-05T06:48:34Z\\\",\\\"message\\\":\\\"g reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1005 06:48:34.496108 6146 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1005 06:48:34.496160 6146 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1005 06:48:34.496168 6146 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1005 06:48:34.496246 6146 factory.go:656] Stopping watch factory\\\\nI1005 06:48:34.496262 6146 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1005 06:48:34.496271 6146 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1005 06:48:34.496279 6146 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1005 06:48:34.496293 6146 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1005 06:48:34.496400 6146 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1005 06:48:34.496553 6146 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1005 06:48:34.496742 6146 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:31Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://93149bae0504218735e0900b660ad9759ebbec3eb4dc214cc40ad349ac002d45\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-05T06:48:36Z\\\",\\\"message\\\":\\\"ePorts:nil,LoadBalancerClass:nil,InternalTrafficPolicy:*Cluster,TrafficDistribution:nil,},Status:ServiceStatus{LoadBalancer:LoadBalancerStatus{Ingress:[]LoadBalancerIngress{},},Conditions:[]Condition{},},}\\\\nI1005 06:48:35.889902 6320 lb_config.go:1031] Cluster endpoints for openshift-kube-apiserver/apiserver for network=default are: map[]\\\\nI1005 06:48:35.889910 6320 services_controller.go:443] Built service openshift-kube-apiserver/apiserver LB cluster-wide configs for network=default: []services.lbConfig{services.lbConfig{vips:[]string{\\\\\\\"10.217.4.93\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:443, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI1005 06:48:35.889925 6320 services_controller.go:444] Built service openshift-kube-apiserver/apiserver LB per-node configs for network=default: []services.lbConfig(nil)\\\\nF1005 06:48:35.889931 6320 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: failed to add event handler: handler {0x1e60340 0x1e60020 0x1e5ffc0} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to se\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0df0ebcd87dc5af4b17736b43ec7b2395968b19ed16e775366af14e2b2d49c52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff3c6545ff3d2436da27bbd223b950b30b0c69a3e7f3034cb0e61380cb38d338\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ff3c6545ff3d2436da27bbd223b950b30b0c69a3e7f3034cb0e61380cb38d338\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:24Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-txzlk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:37Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:37 crc kubenswrapper[4846]: I1005 06:48:37.965362 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x6d6d\" (UniqueName: \"kubernetes.io/projected/698ab882-bb26-4b4c-81e9-8305a44fe068-kube-api-access-x6d6d\") pod \"ovnkube-control-plane-749d76644c-gklb6\" (UID: \"698ab882-bb26-4b4c-81e9-8305a44fe068\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-gklb6" Oct 05 06:48:37 crc kubenswrapper[4846]: I1005 06:48:37.965460 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/698ab882-bb26-4b4c-81e9-8305a44fe068-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-gklb6\" (UID: \"698ab882-bb26-4b4c-81e9-8305a44fe068\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-gklb6" Oct 05 06:48:37 crc kubenswrapper[4846]: I1005 06:48:37.965512 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/698ab882-bb26-4b4c-81e9-8305a44fe068-env-overrides\") pod \"ovnkube-control-plane-749d76644c-gklb6\" (UID: \"698ab882-bb26-4b4c-81e9-8305a44fe068\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-gklb6" Oct 05 06:48:37 crc kubenswrapper[4846]: I1005 06:48:37.965614 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/698ab882-bb26-4b4c-81e9-8305a44fe068-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-gklb6\" (UID: \"698ab882-bb26-4b4c-81e9-8305a44fe068\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-gklb6" Oct 05 06:48:37 crc kubenswrapper[4846]: I1005 06:48:37.966745 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/698ab882-bb26-4b4c-81e9-8305a44fe068-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-gklb6\" (UID: \"698ab882-bb26-4b4c-81e9-8305a44fe068\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-gklb6" Oct 05 06:48:37 crc kubenswrapper[4846]: I1005 06:48:37.966969 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/698ab882-bb26-4b4c-81e9-8305a44fe068-env-overrides\") pod \"ovnkube-control-plane-749d76644c-gklb6\" (UID: \"698ab882-bb26-4b4c-81e9-8305a44fe068\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-gklb6" Oct 05 06:48:37 crc kubenswrapper[4846]: I1005 06:48:37.970098 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-gklb6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"698ab882-bb26-4b4c-81e9-8305a44fe068\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x6d6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x6d6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:37Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-gklb6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:37Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:37 crc kubenswrapper[4846]: I1005 06:48:37.980983 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:37 crc kubenswrapper[4846]: I1005 06:48:37.981071 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:37 crc kubenswrapper[4846]: I1005 06:48:37.981096 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:37 crc kubenswrapper[4846]: I1005 06:48:37.981133 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:37 crc kubenswrapper[4846]: I1005 06:48:37.981165 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:37Z","lastTransitionTime":"2025-10-05T06:48:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:37 crc kubenswrapper[4846]: I1005 06:48:37.988311 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/698ab882-bb26-4b4c-81e9-8305a44fe068-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-gklb6\" (UID: \"698ab882-bb26-4b4c-81e9-8305a44fe068\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-gklb6" Oct 05 06:48:37 crc kubenswrapper[4846]: I1005 06:48:37.991460 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x6d6d\" (UniqueName: \"kubernetes.io/projected/698ab882-bb26-4b4c-81e9-8305a44fe068-kube-api-access-x6d6d\") pod \"ovnkube-control-plane-749d76644c-gklb6\" (UID: \"698ab882-bb26-4b4c-81e9-8305a44fe068\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-gklb6" Oct 05 06:48:38 crc kubenswrapper[4846]: I1005 06:48:38.010269 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c482bb97-fd50-40cf-96ea-f806c3f01125\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bfe45d30e4fa141541aca1d46bc5cc1da0a7b4a5eb1fc9bb2caa5b7fdcb0b264\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://496a4782faa30a9a20ff63038141600cd635f12902d1cf17989cd067fcdfe929\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://281be077869a6aa1ca815770871a14413d4a97e97aeb550fd5cd215cba88b21f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6771c7215d2f5dd5838d17888375a4768b9d1755fc2e21bda1bbfbb2eb1dbb95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c17a711c2c7473b076a3944ffe7c700808d05cba582fbd55c5c2ac8ff73838a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44a9f740f0a06c8bf928e2a9d1d5979217558c31e73a333631baccebf00fd5d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://44a9f740f0a06c8bf928e2a9d1d5979217558c31e73a333631baccebf00fd5d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:47:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:47:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://48ca0ba6ca6092bfcb5b153220c054f02cc36fb22707c42e021f06d329e18e2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48ca0ba6ca6092bfcb5b153220c054f02cc36fb22707c42e021f06d329e18e2e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:00Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4006d0cea128675c8885d879171204abffc5e11ea0664750a12b4bc7344ccb2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4006d0cea128675c8885d879171204abffc5e11ea0664750a12b4bc7344ccb2e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:47:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:38Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:38 crc kubenswrapper[4846]: I1005 06:48:38.034974 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:38Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:38 crc kubenswrapper[4846]: I1005 06:48:38.054753 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e67d0e447a8a6a381259e25d25badc791194e1291f56b1cd77bd63989b20408\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dkhl7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://516f88f4990f5962d8b694a3d889259cc5508a309919870191c2e1331640c973\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dkhl7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-fscvf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:38Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:38 crc kubenswrapper[4846]: I1005 06:48:38.060643 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-gklb6" Oct 05 06:48:38 crc kubenswrapper[4846]: I1005 06:48:38.072572 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-w2sj8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32931fce-7374-4cab-a2e0-79d96a842274\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0df535d65d39de6c9a3dfb931bafc4469adc22b72f7a7903beda4fef8cd48138\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9rtxf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:27Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-w2sj8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:38Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:38 crc kubenswrapper[4846]: I1005 06:48:38.085994 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:38 crc kubenswrapper[4846]: I1005 06:48:38.086050 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:38 crc kubenswrapper[4846]: I1005 06:48:38.086069 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:38 crc kubenswrapper[4846]: I1005 06:48:38.086101 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:38 crc kubenswrapper[4846]: I1005 06:48:38.086126 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:38Z","lastTransitionTime":"2025-10-05T06:48:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:38 crc kubenswrapper[4846]: W1005 06:48:38.092308 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod698ab882_bb26_4b4c_81e9_8305a44fe068.slice/crio-301b44bc1f633a2031d1cee0e2d051b703f73fe39b5f6c5aea4c4a632464421f WatchSource:0}: Error finding container 301b44bc1f633a2031d1cee0e2d051b703f73fe39b5f6c5aea4c4a632464421f: Status 404 returned error can't find the container with id 301b44bc1f633a2031d1cee0e2d051b703f73fe39b5f6c5aea4c4a632464421f Oct 05 06:48:38 crc kubenswrapper[4846]: I1005 06:48:38.096084 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c9f812b-ecc8-4985-a670-53fa80fbb499\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2b5ee9781c01d6aec3ac778efef1067b6c3bfe4ed37431bdebe188fcde0165f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:47:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3fb4717692aa1fec0012b514ef8d18d1ca3707cfa4bb8248bb93841a0a40f11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:47:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0b6a7afe3094f1c92093c05c28344897a615d590c2fbb507c87e18a44c6c831\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://640ade2c733d7916ec213a201a1dd614d261c69eab654c80100efbc99a768995\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:47:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:38Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:38 crc kubenswrapper[4846]: I1005 06:48:38.121364 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:38Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:38 crc kubenswrapper[4846]: I1005 06:48:38.189745 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:38 crc kubenswrapper[4846]: I1005 06:48:38.189795 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:38 crc kubenswrapper[4846]: I1005 06:48:38.189811 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:38 crc kubenswrapper[4846]: I1005 06:48:38.189837 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:38 crc kubenswrapper[4846]: I1005 06:48:38.189853 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:38Z","lastTransitionTime":"2025-10-05T06:48:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:38 crc kubenswrapper[4846]: I1005 06:48:38.295057 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:38 crc kubenswrapper[4846]: I1005 06:48:38.295106 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:38 crc kubenswrapper[4846]: I1005 06:48:38.295119 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:38 crc kubenswrapper[4846]: I1005 06:48:38.295137 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:38 crc kubenswrapper[4846]: I1005 06:48:38.295150 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:38Z","lastTransitionTime":"2025-10-05T06:48:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:38 crc kubenswrapper[4846]: I1005 06:48:38.398238 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:38 crc kubenswrapper[4846]: I1005 06:48:38.398306 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:38 crc kubenswrapper[4846]: I1005 06:48:38.398326 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:38 crc kubenswrapper[4846]: I1005 06:48:38.398365 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:38 crc kubenswrapper[4846]: I1005 06:48:38.398385 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:38Z","lastTransitionTime":"2025-10-05T06:48:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:38 crc kubenswrapper[4846]: I1005 06:48:38.497809 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 05 06:48:38 crc kubenswrapper[4846]: E1005 06:48:38.498082 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 05 06:48:38 crc kubenswrapper[4846]: I1005 06:48:38.501511 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:38 crc kubenswrapper[4846]: I1005 06:48:38.501561 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:38 crc kubenswrapper[4846]: I1005 06:48:38.501663 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:38 crc kubenswrapper[4846]: I1005 06:48:38.501695 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:38 crc kubenswrapper[4846]: I1005 06:48:38.501736 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:38Z","lastTransitionTime":"2025-10-05T06:48:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:38 crc kubenswrapper[4846]: I1005 06:48:38.520283 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c9f812b-ecc8-4985-a670-53fa80fbb499\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2b5ee9781c01d6aec3ac778efef1067b6c3bfe4ed37431bdebe188fcde0165f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:47:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3fb4717692aa1fec0012b514ef8d18d1ca3707cfa4bb8248bb93841a0a40f11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:47:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0b6a7afe3094f1c92093c05c28344897a615d590c2fbb507c87e18a44c6c831\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://640ade2c733d7916ec213a201a1dd614d261c69eab654c80100efbc99a768995\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:47:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:38Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:38 crc kubenswrapper[4846]: I1005 06:48:38.536051 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:38Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:38 crc kubenswrapper[4846]: I1005 06:48:38.552296 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e67d0e447a8a6a381259e25d25badc791194e1291f56b1cd77bd63989b20408\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dkhl7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://516f88f4990f5962d8b694a3d889259cc5508a309919870191c2e1331640c973\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dkhl7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-fscvf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:38Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:38 crc kubenswrapper[4846]: I1005 06:48:38.567480 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-w2sj8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32931fce-7374-4cab-a2e0-79d96a842274\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0df535d65d39de6c9a3dfb931bafc4469adc22b72f7a7903beda4fef8cd48138\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9rtxf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:27Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-w2sj8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:38Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:38 crc kubenswrapper[4846]: I1005 06:48:38.592098 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ff10c31459b3b3724ee7b769bd6f72f9ed80cce5419b6adcdf69a3239b70756\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:38Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:38 crc kubenswrapper[4846]: I1005 06:48:38.605085 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:38 crc kubenswrapper[4846]: I1005 06:48:38.605156 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:38 crc kubenswrapper[4846]: I1005 06:48:38.605204 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:38 crc kubenswrapper[4846]: I1005 06:48:38.605235 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:38 crc kubenswrapper[4846]: I1005 06:48:38.605255 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:38Z","lastTransitionTime":"2025-10-05T06:48:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:38 crc kubenswrapper[4846]: I1005 06:48:38.609031 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://347f900743b7bba3daf92f21c7eba0625e8c774f50057bc9472f0c82119325e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:38Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:38 crc kubenswrapper[4846]: I1005 06:48:38.632586 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56f1fe1f9f06f0249cf937a6963fda82467732dc056ca2b2189be6d8dd9d4a05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7212ce5f42c1fb55146abb190a557a504b43f16af6df84f90e1073567a329ec5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:38Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:38 crc kubenswrapper[4846]: I1005 06:48:38.649638 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-7wm5q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9bc7744f-d700-450f-8b8d-98140877fee9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90379e74c544fccca44f7acdc228ff378154390e211ac801b79f8d074012241e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h5p4q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-7wm5q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:38Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:38 crc kubenswrapper[4846]: I1005 06:48:38.673448 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"775b5702-a131-4d6c-852d-91bfa056cac5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0839c1765df2c68d2554e9c66d289505e559a9672caf8069feeff2c4fdfd2171\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15dbaf2445704da0987f256f81dff2ee95664988541fa48ce94898afeadab2e8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd3b3ab323c6d5153054aa04af2d91f7b9e8436ff1e2f75f3594b57831b2a416\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://67c46f5f0a3ace95bd2e9479c5561126664e18b148c5eceea5bf5efc27e81a67\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c52e076eddfd307110b1f6de00a169b81af8f2785af7becef2a820ce363a5861\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-05T06:48:14Z\\\",\\\"message\\\":\\\"le observer\\\\nW1005 06:48:14.337870 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1005 06:48:14.338086 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1005 06:48:14.339389 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1467752215/tls.crt::/tmp/serving-cert-1467752215/tls.key\\\\\\\"\\\\nI1005 06:48:14.864345 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1005 06:48:14.868087 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1005 06:48:14.868114 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1005 06:48:14.868140 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1005 06:48:14.868147 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1005 06:48:14.877354 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1005 06:48:14.877412 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1005 06:48:14.877456 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1005 06:48:14.877471 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1005 06:48:14.877490 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1005 06:48:14.877501 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1005 06:48:14.877507 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1005 06:48:14.877515 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1005 06:48:14.880326 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:13Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4a091716ff935adb42ae2e88048619fd347bd174f64e5f9c97a01f81690e57f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:01Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dc5411b642f99b13d514d2f21defad0c394aeeb82a966f72300076617e2586c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5dc5411b642f99b13d514d2f21defad0c394aeeb82a966f72300076617e2586c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:47:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:47:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:47:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:38Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:38 crc kubenswrapper[4846]: I1005 06:48:38.693144 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:38Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:38 crc kubenswrapper[4846]: I1005 06:48:38.708505 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:38 crc kubenswrapper[4846]: I1005 06:48:38.708554 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:38 crc kubenswrapper[4846]: I1005 06:48:38.708567 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:38 crc kubenswrapper[4846]: I1005 06:48:38.708585 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:38 crc kubenswrapper[4846]: I1005 06:48:38.708600 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:38Z","lastTransitionTime":"2025-10-05T06:48:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:38 crc kubenswrapper[4846]: I1005 06:48:38.713535 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-r7ztb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a5119cd9-af70-4324-a955-978306ab0b20\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06231a87429ec292cb57ebd88cda7ad547bcb32e80343b776833621a11c28ad1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jfcj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-r7ztb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:38Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:38 crc kubenswrapper[4846]: I1005 06:48:38.749348 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c482bb97-fd50-40cf-96ea-f806c3f01125\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bfe45d30e4fa141541aca1d46bc5cc1da0a7b4a5eb1fc9bb2caa5b7fdcb0b264\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://496a4782faa30a9a20ff63038141600cd635f12902d1cf17989cd067fcdfe929\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://281be077869a6aa1ca815770871a14413d4a97e97aeb550fd5cd215cba88b21f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6771c7215d2f5dd5838d17888375a4768b9d1755fc2e21bda1bbfbb2eb1dbb95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c17a711c2c7473b076a3944ffe7c700808d05cba582fbd55c5c2ac8ff73838a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44a9f740f0a06c8bf928e2a9d1d5979217558c31e73a333631baccebf00fd5d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://44a9f740f0a06c8bf928e2a9d1d5979217558c31e73a333631baccebf00fd5d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:47:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:47:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://48ca0ba6ca6092bfcb5b153220c054f02cc36fb22707c42e021f06d329e18e2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48ca0ba6ca6092bfcb5b153220c054f02cc36fb22707c42e021f06d329e18e2e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:00Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4006d0cea128675c8885d879171204abffc5e11ea0664750a12b4bc7344ccb2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4006d0cea128675c8885d879171204abffc5e11ea0664750a12b4bc7344ccb2e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:47:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:38Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:38 crc kubenswrapper[4846]: I1005 06:48:38.765948 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:38Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:38 crc kubenswrapper[4846]: I1005 06:48:38.790442 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-sfmxd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1dbb13ac-136c-4a1c-9229-b4e8ac4f9c71\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b0af6a6e735f7b7877c177c1d53d1a8dc12c4713a8059c4766fc858af31c9a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://226d977b596d07aeda0486452e2a8605949ed4d2453b21f56174937c395d6f36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://226d977b596d07aeda0486452e2a8605949ed4d2453b21f56174937c395d6f36\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://99c07a4bdb3059d3a99ffdaf794d668d52cffd9e5e3bb8945204dd6d1f8c8328\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99c07a4bdb3059d3a99ffdaf794d668d52cffd9e5e3bb8945204dd6d1f8c8328\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27095272bf4f31d63a78a436636de0aae12f7103597da944f24ed7a74329ee3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27095272bf4f31d63a78a436636de0aae12f7103597da944f24ed7a74329ee3f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1211571af66d8a94302b65feccf30768ba82b7cbad918e44dd405787dce484cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1211571af66d8a94302b65feccf30768ba82b7cbad918e44dd405787dce484cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27631f90d528900fa180c15cfce80e96d4cdba07743eb6d847e2e822d9917218\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27631f90d528900fa180c15cfce80e96d4cdba07743eb6d847e2e822d9917218\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f345693e9aeb1f4dc196349d3354f3c833a2b0415a693298b1cfccfa2b0d4459\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f345693e9aeb1f4dc196349d3354f3c833a2b0415a693298b1cfccfa2b0d4459\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-sfmxd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:38Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:38 crc kubenswrapper[4846]: I1005 06:48:38.812268 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:38 crc kubenswrapper[4846]: I1005 06:48:38.812359 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:38 crc kubenswrapper[4846]: I1005 06:48:38.812378 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:38 crc kubenswrapper[4846]: I1005 06:48:38.812421 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:38 crc kubenswrapper[4846]: I1005 06:48:38.812439 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:38Z","lastTransitionTime":"2025-10-05T06:48:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:38 crc kubenswrapper[4846]: I1005 06:48:38.817365 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-txzlk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"93a6ace4-8d64-44d7-9637-457d3af6543c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ad1940d07e37217089642478ece9ccdaca74784738a400929a9f5359d158c01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a6f6b2c32dd3da5f9fd5e6971c1773e81a9a24184b7aa5b543cec879001ae53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e9b0219ecbc7e60905d6de3e67a2c8dc9af1017b4aed9538ee0c58c861f2438\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af5e5a4da9ad7371d5a24633e8268f12e3a7d419a8c93b11ddbcc006d8339c6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8da85222775abc7a772f6a8ff8255f1a0d5b15807ba177a7e3e2b2d01a4f24b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1409651de8c897804f87be03d016c138f8acf27e662dce265a93e15c1848adcf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://93149bae0504218735e0900b660ad9759ebbec3eb4dc214cc40ad349ac002d45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c212f5d230fae82a701afeceb39a848a9c91a3354908d88b062b7e1a726240cb\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-05T06:48:34Z\\\",\\\"message\\\":\\\"g reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1005 06:48:34.496108 6146 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1005 06:48:34.496160 6146 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1005 06:48:34.496168 6146 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1005 06:48:34.496246 6146 factory.go:656] Stopping watch factory\\\\nI1005 06:48:34.496262 6146 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1005 06:48:34.496271 6146 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1005 06:48:34.496279 6146 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1005 06:48:34.496293 6146 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1005 06:48:34.496400 6146 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1005 06:48:34.496553 6146 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1005 06:48:34.496742 6146 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:31Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://93149bae0504218735e0900b660ad9759ebbec3eb4dc214cc40ad349ac002d45\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-05T06:48:36Z\\\",\\\"message\\\":\\\"ePorts:nil,LoadBalancerClass:nil,InternalTrafficPolicy:*Cluster,TrafficDistribution:nil,},Status:ServiceStatus{LoadBalancer:LoadBalancerStatus{Ingress:[]LoadBalancerIngress{},},Conditions:[]Condition{},},}\\\\nI1005 06:48:35.889902 6320 lb_config.go:1031] Cluster endpoints for openshift-kube-apiserver/apiserver for network=default are: map[]\\\\nI1005 06:48:35.889910 6320 services_controller.go:443] Built service openshift-kube-apiserver/apiserver LB cluster-wide configs for network=default: []services.lbConfig{services.lbConfig{vips:[]string{\\\\\\\"10.217.4.93\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:443, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI1005 06:48:35.889925 6320 services_controller.go:444] Built service openshift-kube-apiserver/apiserver LB per-node configs for network=default: []services.lbConfig(nil)\\\\nF1005 06:48:35.889931 6320 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: failed to add event handler: handler {0x1e60340 0x1e60020 0x1e5ffc0} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to se\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0df0ebcd87dc5af4b17736b43ec7b2395968b19ed16e775366af14e2b2d49c52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff3c6545ff3d2436da27bbd223b950b30b0c69a3e7f3034cb0e61380cb38d338\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ff3c6545ff3d2436da27bbd223b950b30b0c69a3e7f3034cb0e61380cb38d338\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:24Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-txzlk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:38Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:38 crc kubenswrapper[4846]: I1005 06:48:38.832614 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-gklb6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"698ab882-bb26-4b4c-81e9-8305a44fe068\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:37Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x6d6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x6d6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:37Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-gklb6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:38Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:38 crc kubenswrapper[4846]: I1005 06:48:38.893319 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-gklb6" event={"ID":"698ab882-bb26-4b4c-81e9-8305a44fe068","Type":"ContainerStarted","Data":"3c191538ab797b8056820de7da1f8e93ea18873891a32382b8a59c9af762e103"} Oct 05 06:48:38 crc kubenswrapper[4846]: I1005 06:48:38.893396 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-gklb6" event={"ID":"698ab882-bb26-4b4c-81e9-8305a44fe068","Type":"ContainerStarted","Data":"7a6edca88ea244a58f8cf6cc431110a960c6bfd892812cd0d00fa3f97be3bc67"} Oct 05 06:48:38 crc kubenswrapper[4846]: I1005 06:48:38.893410 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-gklb6" event={"ID":"698ab882-bb26-4b4c-81e9-8305a44fe068","Type":"ContainerStarted","Data":"301b44bc1f633a2031d1cee0e2d051b703f73fe39b5f6c5aea4c4a632464421f"} Oct 05 06:48:38 crc kubenswrapper[4846]: I1005 06:48:38.897675 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/network-metrics-daemon-9r2kt"] Oct 05 06:48:38 crc kubenswrapper[4846]: I1005 06:48:38.898571 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9r2kt" Oct 05 06:48:38 crc kubenswrapper[4846]: E1005 06:48:38.898671 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9r2kt" podUID="59b83f4f-71f0-4b28-9275-56d5f933ec2c" Oct 05 06:48:38 crc kubenswrapper[4846]: I1005 06:48:38.915892 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:38 crc kubenswrapper[4846]: I1005 06:48:38.915936 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:38 crc kubenswrapper[4846]: I1005 06:48:38.915951 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:38 crc kubenswrapper[4846]: I1005 06:48:38.915970 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:38 crc kubenswrapper[4846]: I1005 06:48:38.915986 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:38Z","lastTransitionTime":"2025-10-05T06:48:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:38 crc kubenswrapper[4846]: I1005 06:48:38.918017 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ff10c31459b3b3724ee7b769bd6f72f9ed80cce5419b6adcdf69a3239b70756\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:38Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:38 crc kubenswrapper[4846]: I1005 06:48:38.937153 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://347f900743b7bba3daf92f21c7eba0625e8c774f50057bc9472f0c82119325e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:38Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:38 crc kubenswrapper[4846]: I1005 06:48:38.955838 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56f1fe1f9f06f0249cf937a6963fda82467732dc056ca2b2189be6d8dd9d4a05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7212ce5f42c1fb55146abb190a557a504b43f16af6df84f90e1073567a329ec5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:38Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:38 crc kubenswrapper[4846]: I1005 06:48:38.978283 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-7wm5q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9bc7744f-d700-450f-8b8d-98140877fee9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90379e74c544fccca44f7acdc228ff378154390e211ac801b79f8d074012241e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h5p4q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-7wm5q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:38Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:38 crc kubenswrapper[4846]: I1005 06:48:38.984646 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6tdwj\" (UniqueName: \"kubernetes.io/projected/59b83f4f-71f0-4b28-9275-56d5f933ec2c-kube-api-access-6tdwj\") pod \"network-metrics-daemon-9r2kt\" (UID: \"59b83f4f-71f0-4b28-9275-56d5f933ec2c\") " pod="openshift-multus/network-metrics-daemon-9r2kt" Oct 05 06:48:38 crc kubenswrapper[4846]: I1005 06:48:38.984805 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/59b83f4f-71f0-4b28-9275-56d5f933ec2c-metrics-certs\") pod \"network-metrics-daemon-9r2kt\" (UID: \"59b83f4f-71f0-4b28-9275-56d5f933ec2c\") " pod="openshift-multus/network-metrics-daemon-9r2kt" Oct 05 06:48:39 crc kubenswrapper[4846]: I1005 06:48:39.008285 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"775b5702-a131-4d6c-852d-91bfa056cac5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0839c1765df2c68d2554e9c66d289505e559a9672caf8069feeff2c4fdfd2171\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15dbaf2445704da0987f256f81dff2ee95664988541fa48ce94898afeadab2e8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd3b3ab323c6d5153054aa04af2d91f7b9e8436ff1e2f75f3594b57831b2a416\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://67c46f5f0a3ace95bd2e9479c5561126664e18b148c5eceea5bf5efc27e81a67\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c52e076eddfd307110b1f6de00a169b81af8f2785af7becef2a820ce363a5861\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-05T06:48:14Z\\\",\\\"message\\\":\\\"le observer\\\\nW1005 06:48:14.337870 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1005 06:48:14.338086 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1005 06:48:14.339389 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1467752215/tls.crt::/tmp/serving-cert-1467752215/tls.key\\\\\\\"\\\\nI1005 06:48:14.864345 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1005 06:48:14.868087 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1005 06:48:14.868114 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1005 06:48:14.868140 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1005 06:48:14.868147 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1005 06:48:14.877354 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1005 06:48:14.877412 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1005 06:48:14.877456 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1005 06:48:14.877471 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1005 06:48:14.877490 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1005 06:48:14.877501 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1005 06:48:14.877507 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1005 06:48:14.877515 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1005 06:48:14.880326 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:13Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4a091716ff935adb42ae2e88048619fd347bd174f64e5f9c97a01f81690e57f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:01Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dc5411b642f99b13d514d2f21defad0c394aeeb82a966f72300076617e2586c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5dc5411b642f99b13d514d2f21defad0c394aeeb82a966f72300076617e2586c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:47:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:47:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:47:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:39Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:39 crc kubenswrapper[4846]: I1005 06:48:39.020147 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:39 crc kubenswrapper[4846]: I1005 06:48:39.020271 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:39 crc kubenswrapper[4846]: I1005 06:48:39.020300 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:39 crc kubenswrapper[4846]: I1005 06:48:39.020337 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:39 crc kubenswrapper[4846]: I1005 06:48:39.020364 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:39Z","lastTransitionTime":"2025-10-05T06:48:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:39 crc kubenswrapper[4846]: I1005 06:48:39.030472 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:39Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:39 crc kubenswrapper[4846]: I1005 06:48:39.045857 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-r7ztb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a5119cd9-af70-4324-a955-978306ab0b20\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06231a87429ec292cb57ebd88cda7ad547bcb32e80343b776833621a11c28ad1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jfcj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-r7ztb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:39Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:39 crc kubenswrapper[4846]: I1005 06:48:39.079045 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c482bb97-fd50-40cf-96ea-f806c3f01125\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bfe45d30e4fa141541aca1d46bc5cc1da0a7b4a5eb1fc9bb2caa5b7fdcb0b264\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://496a4782faa30a9a20ff63038141600cd635f12902d1cf17989cd067fcdfe929\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://281be077869a6aa1ca815770871a14413d4a97e97aeb550fd5cd215cba88b21f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6771c7215d2f5dd5838d17888375a4768b9d1755fc2e21bda1bbfbb2eb1dbb95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c17a711c2c7473b076a3944ffe7c700808d05cba582fbd55c5c2ac8ff73838a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44a9f740f0a06c8bf928e2a9d1d5979217558c31e73a333631baccebf00fd5d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://44a9f740f0a06c8bf928e2a9d1d5979217558c31e73a333631baccebf00fd5d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:47:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:47:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://48ca0ba6ca6092bfcb5b153220c054f02cc36fb22707c42e021f06d329e18e2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48ca0ba6ca6092bfcb5b153220c054f02cc36fb22707c42e021f06d329e18e2e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:00Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4006d0cea128675c8885d879171204abffc5e11ea0664750a12b4bc7344ccb2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4006d0cea128675c8885d879171204abffc5e11ea0664750a12b4bc7344ccb2e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:47:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:39Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:39 crc kubenswrapper[4846]: I1005 06:48:39.086734 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/59b83f4f-71f0-4b28-9275-56d5f933ec2c-metrics-certs\") pod \"network-metrics-daemon-9r2kt\" (UID: \"59b83f4f-71f0-4b28-9275-56d5f933ec2c\") " pod="openshift-multus/network-metrics-daemon-9r2kt" Oct 05 06:48:39 crc kubenswrapper[4846]: I1005 06:48:39.086869 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6tdwj\" (UniqueName: \"kubernetes.io/projected/59b83f4f-71f0-4b28-9275-56d5f933ec2c-kube-api-access-6tdwj\") pod \"network-metrics-daemon-9r2kt\" (UID: \"59b83f4f-71f0-4b28-9275-56d5f933ec2c\") " pod="openshift-multus/network-metrics-daemon-9r2kt" Oct 05 06:48:39 crc kubenswrapper[4846]: E1005 06:48:39.087031 4846 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 05 06:48:39 crc kubenswrapper[4846]: E1005 06:48:39.087269 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/59b83f4f-71f0-4b28-9275-56d5f933ec2c-metrics-certs podName:59b83f4f-71f0-4b28-9275-56d5f933ec2c nodeName:}" failed. No retries permitted until 2025-10-05 06:48:39.5871608 +0000 UTC m=+41.828013745 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/59b83f4f-71f0-4b28-9275-56d5f933ec2c-metrics-certs") pod "network-metrics-daemon-9r2kt" (UID: "59b83f4f-71f0-4b28-9275-56d5f933ec2c") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 05 06:48:39 crc kubenswrapper[4846]: I1005 06:48:39.098109 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:39Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:39 crc kubenswrapper[4846]: I1005 06:48:39.109470 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6tdwj\" (UniqueName: \"kubernetes.io/projected/59b83f4f-71f0-4b28-9275-56d5f933ec2c-kube-api-access-6tdwj\") pod \"network-metrics-daemon-9r2kt\" (UID: \"59b83f4f-71f0-4b28-9275-56d5f933ec2c\") " pod="openshift-multus/network-metrics-daemon-9r2kt" Oct 05 06:48:39 crc kubenswrapper[4846]: I1005 06:48:39.117160 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-sfmxd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1dbb13ac-136c-4a1c-9229-b4e8ac4f9c71\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b0af6a6e735f7b7877c177c1d53d1a8dc12c4713a8059c4766fc858af31c9a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://226d977b596d07aeda0486452e2a8605949ed4d2453b21f56174937c395d6f36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://226d977b596d07aeda0486452e2a8605949ed4d2453b21f56174937c395d6f36\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://99c07a4bdb3059d3a99ffdaf794d668d52cffd9e5e3bb8945204dd6d1f8c8328\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99c07a4bdb3059d3a99ffdaf794d668d52cffd9e5e3bb8945204dd6d1f8c8328\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27095272bf4f31d63a78a436636de0aae12f7103597da944f24ed7a74329ee3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27095272bf4f31d63a78a436636de0aae12f7103597da944f24ed7a74329ee3f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1211571af66d8a94302b65feccf30768ba82b7cbad918e44dd405787dce484cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1211571af66d8a94302b65feccf30768ba82b7cbad918e44dd405787dce484cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27631f90d528900fa180c15cfce80e96d4cdba07743eb6d847e2e822d9917218\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27631f90d528900fa180c15cfce80e96d4cdba07743eb6d847e2e822d9917218\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f345693e9aeb1f4dc196349d3354f3c833a2b0415a693298b1cfccfa2b0d4459\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f345693e9aeb1f4dc196349d3354f3c833a2b0415a693298b1cfccfa2b0d4459\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-sfmxd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:39Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:39 crc kubenswrapper[4846]: I1005 06:48:39.123318 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:39 crc kubenswrapper[4846]: I1005 06:48:39.123371 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:39 crc kubenswrapper[4846]: I1005 06:48:39.123390 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:39 crc kubenswrapper[4846]: I1005 06:48:39.123416 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:39 crc kubenswrapper[4846]: I1005 06:48:39.123433 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:39Z","lastTransitionTime":"2025-10-05T06:48:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:39 crc kubenswrapper[4846]: I1005 06:48:39.144693 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-txzlk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"93a6ace4-8d64-44d7-9637-457d3af6543c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ad1940d07e37217089642478ece9ccdaca74784738a400929a9f5359d158c01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a6f6b2c32dd3da5f9fd5e6971c1773e81a9a24184b7aa5b543cec879001ae53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e9b0219ecbc7e60905d6de3e67a2c8dc9af1017b4aed9538ee0c58c861f2438\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af5e5a4da9ad7371d5a24633e8268f12e3a7d419a8c93b11ddbcc006d8339c6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8da85222775abc7a772f6a8ff8255f1a0d5b15807ba177a7e3e2b2d01a4f24b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1409651de8c897804f87be03d016c138f8acf27e662dce265a93e15c1848adcf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://93149bae0504218735e0900b660ad9759ebbec3eb4dc214cc40ad349ac002d45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c212f5d230fae82a701afeceb39a848a9c91a3354908d88b062b7e1a726240cb\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-05T06:48:34Z\\\",\\\"message\\\":\\\"g reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1005 06:48:34.496108 6146 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1005 06:48:34.496160 6146 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1005 06:48:34.496168 6146 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1005 06:48:34.496246 6146 factory.go:656] Stopping watch factory\\\\nI1005 06:48:34.496262 6146 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1005 06:48:34.496271 6146 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1005 06:48:34.496279 6146 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1005 06:48:34.496293 6146 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1005 06:48:34.496400 6146 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1005 06:48:34.496553 6146 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1005 06:48:34.496742 6146 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:31Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://93149bae0504218735e0900b660ad9759ebbec3eb4dc214cc40ad349ac002d45\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-05T06:48:36Z\\\",\\\"message\\\":\\\"ePorts:nil,LoadBalancerClass:nil,InternalTrafficPolicy:*Cluster,TrafficDistribution:nil,},Status:ServiceStatus{LoadBalancer:LoadBalancerStatus{Ingress:[]LoadBalancerIngress{},},Conditions:[]Condition{},},}\\\\nI1005 06:48:35.889902 6320 lb_config.go:1031] Cluster endpoints for openshift-kube-apiserver/apiserver for network=default are: map[]\\\\nI1005 06:48:35.889910 6320 services_controller.go:443] Built service openshift-kube-apiserver/apiserver LB cluster-wide configs for network=default: []services.lbConfig{services.lbConfig{vips:[]string{\\\\\\\"10.217.4.93\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:443, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI1005 06:48:35.889925 6320 services_controller.go:444] Built service openshift-kube-apiserver/apiserver LB per-node configs for network=default: []services.lbConfig(nil)\\\\nF1005 06:48:35.889931 6320 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: failed to add event handler: handler {0x1e60340 0x1e60020 0x1e5ffc0} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to se\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0df0ebcd87dc5af4b17736b43ec7b2395968b19ed16e775366af14e2b2d49c52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff3c6545ff3d2436da27bbd223b950b30b0c69a3e7f3034cb0e61380cb38d338\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ff3c6545ff3d2436da27bbd223b950b30b0c69a3e7f3034cb0e61380cb38d338\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:24Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-txzlk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:39Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:39 crc kubenswrapper[4846]: I1005 06:48:39.163439 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-gklb6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"698ab882-bb26-4b4c-81e9-8305a44fe068\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a6edca88ea244a58f8cf6cc431110a960c6bfd892812cd0d00fa3f97be3bc67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x6d6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c191538ab797b8056820de7da1f8e93ea18873891a32382b8a59c9af762e103\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x6d6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:37Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-gklb6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:39Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:39 crc kubenswrapper[4846]: I1005 06:48:39.178120 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c9f812b-ecc8-4985-a670-53fa80fbb499\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2b5ee9781c01d6aec3ac778efef1067b6c3bfe4ed37431bdebe188fcde0165f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:47:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3fb4717692aa1fec0012b514ef8d18d1ca3707cfa4bb8248bb93841a0a40f11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:47:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0b6a7afe3094f1c92093c05c28344897a615d590c2fbb507c87e18a44c6c831\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://640ade2c733d7916ec213a201a1dd614d261c69eab654c80100efbc99a768995\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:47:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:39Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:39 crc kubenswrapper[4846]: I1005 06:48:39.197760 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:39Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:39 crc kubenswrapper[4846]: I1005 06:48:39.211317 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e67d0e447a8a6a381259e25d25badc791194e1291f56b1cd77bd63989b20408\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dkhl7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://516f88f4990f5962d8b694a3d889259cc5508a309919870191c2e1331640c973\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dkhl7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-fscvf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:39Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:39 crc kubenswrapper[4846]: I1005 06:48:39.226401 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:39 crc kubenswrapper[4846]: I1005 06:48:39.226474 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:39 crc kubenswrapper[4846]: I1005 06:48:39.226493 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:39 crc kubenswrapper[4846]: I1005 06:48:39.226524 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:39 crc kubenswrapper[4846]: I1005 06:48:39.226546 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:39Z","lastTransitionTime":"2025-10-05T06:48:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:39 crc kubenswrapper[4846]: I1005 06:48:39.227751 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-w2sj8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32931fce-7374-4cab-a2e0-79d96a842274\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0df535d65d39de6c9a3dfb931bafc4469adc22b72f7a7903beda4fef8cd48138\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9rtxf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:27Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-w2sj8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:39Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:39 crc kubenswrapper[4846]: I1005 06:48:39.242832 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-gklb6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"698ab882-bb26-4b4c-81e9-8305a44fe068\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a6edca88ea244a58f8cf6cc431110a960c6bfd892812cd0d00fa3f97be3bc67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x6d6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c191538ab797b8056820de7da1f8e93ea18873891a32382b8a59c9af762e103\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x6d6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:37Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-gklb6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:39Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:39 crc kubenswrapper[4846]: I1005 06:48:39.274690 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c482bb97-fd50-40cf-96ea-f806c3f01125\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bfe45d30e4fa141541aca1d46bc5cc1da0a7b4a5eb1fc9bb2caa5b7fdcb0b264\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://496a4782faa30a9a20ff63038141600cd635f12902d1cf17989cd067fcdfe929\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://281be077869a6aa1ca815770871a14413d4a97e97aeb550fd5cd215cba88b21f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6771c7215d2f5dd5838d17888375a4768b9d1755fc2e21bda1bbfbb2eb1dbb95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c17a711c2c7473b076a3944ffe7c700808d05cba582fbd55c5c2ac8ff73838a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44a9f740f0a06c8bf928e2a9d1d5979217558c31e73a333631baccebf00fd5d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://44a9f740f0a06c8bf928e2a9d1d5979217558c31e73a333631baccebf00fd5d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:47:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:47:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://48ca0ba6ca6092bfcb5b153220c054f02cc36fb22707c42e021f06d329e18e2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48ca0ba6ca6092bfcb5b153220c054f02cc36fb22707c42e021f06d329e18e2e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:00Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4006d0cea128675c8885d879171204abffc5e11ea0664750a12b4bc7344ccb2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4006d0cea128675c8885d879171204abffc5e11ea0664750a12b4bc7344ccb2e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:47:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:39Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:39 crc kubenswrapper[4846]: I1005 06:48:39.292693 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:39Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:39 crc kubenswrapper[4846]: I1005 06:48:39.312246 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-sfmxd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1dbb13ac-136c-4a1c-9229-b4e8ac4f9c71\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b0af6a6e735f7b7877c177c1d53d1a8dc12c4713a8059c4766fc858af31c9a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://226d977b596d07aeda0486452e2a8605949ed4d2453b21f56174937c395d6f36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://226d977b596d07aeda0486452e2a8605949ed4d2453b21f56174937c395d6f36\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://99c07a4bdb3059d3a99ffdaf794d668d52cffd9e5e3bb8945204dd6d1f8c8328\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99c07a4bdb3059d3a99ffdaf794d668d52cffd9e5e3bb8945204dd6d1f8c8328\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27095272bf4f31d63a78a436636de0aae12f7103597da944f24ed7a74329ee3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27095272bf4f31d63a78a436636de0aae12f7103597da944f24ed7a74329ee3f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1211571af66d8a94302b65feccf30768ba82b7cbad918e44dd405787dce484cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1211571af66d8a94302b65feccf30768ba82b7cbad918e44dd405787dce484cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27631f90d528900fa180c15cfce80e96d4cdba07743eb6d847e2e822d9917218\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27631f90d528900fa180c15cfce80e96d4cdba07743eb6d847e2e822d9917218\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f345693e9aeb1f4dc196349d3354f3c833a2b0415a693298b1cfccfa2b0d4459\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f345693e9aeb1f4dc196349d3354f3c833a2b0415a693298b1cfccfa2b0d4459\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-sfmxd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:39Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:39 crc kubenswrapper[4846]: I1005 06:48:39.366954 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:39 crc kubenswrapper[4846]: I1005 06:48:39.367230 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:39 crc kubenswrapper[4846]: I1005 06:48:39.367253 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:39 crc kubenswrapper[4846]: I1005 06:48:39.367283 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:39 crc kubenswrapper[4846]: I1005 06:48:39.367969 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:39Z","lastTransitionTime":"2025-10-05T06:48:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:39 crc kubenswrapper[4846]: I1005 06:48:39.370586 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-txzlk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"93a6ace4-8d64-44d7-9637-457d3af6543c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ad1940d07e37217089642478ece9ccdaca74784738a400929a9f5359d158c01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a6f6b2c32dd3da5f9fd5e6971c1773e81a9a24184b7aa5b543cec879001ae53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e9b0219ecbc7e60905d6de3e67a2c8dc9af1017b4aed9538ee0c58c861f2438\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af5e5a4da9ad7371d5a24633e8268f12e3a7d419a8c93b11ddbcc006d8339c6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8da85222775abc7a772f6a8ff8255f1a0d5b15807ba177a7e3e2b2d01a4f24b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1409651de8c897804f87be03d016c138f8acf27e662dce265a93e15c1848adcf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://93149bae0504218735e0900b660ad9759ebbec3eb4dc214cc40ad349ac002d45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c212f5d230fae82a701afeceb39a848a9c91a3354908d88b062b7e1a726240cb\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-05T06:48:34Z\\\",\\\"message\\\":\\\"g reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1005 06:48:34.496108 6146 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1005 06:48:34.496160 6146 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1005 06:48:34.496168 6146 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1005 06:48:34.496246 6146 factory.go:656] Stopping watch factory\\\\nI1005 06:48:34.496262 6146 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1005 06:48:34.496271 6146 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1005 06:48:34.496279 6146 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1005 06:48:34.496293 6146 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1005 06:48:34.496400 6146 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1005 06:48:34.496553 6146 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1005 06:48:34.496742 6146 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:31Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://93149bae0504218735e0900b660ad9759ebbec3eb4dc214cc40ad349ac002d45\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-05T06:48:36Z\\\",\\\"message\\\":\\\"ePorts:nil,LoadBalancerClass:nil,InternalTrafficPolicy:*Cluster,TrafficDistribution:nil,},Status:ServiceStatus{LoadBalancer:LoadBalancerStatus{Ingress:[]LoadBalancerIngress{},},Conditions:[]Condition{},},}\\\\nI1005 06:48:35.889902 6320 lb_config.go:1031] Cluster endpoints for openshift-kube-apiserver/apiserver for network=default are: map[]\\\\nI1005 06:48:35.889910 6320 services_controller.go:443] Built service openshift-kube-apiserver/apiserver LB cluster-wide configs for network=default: []services.lbConfig{services.lbConfig{vips:[]string{\\\\\\\"10.217.4.93\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:443, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI1005 06:48:35.889925 6320 services_controller.go:444] Built service openshift-kube-apiserver/apiserver LB per-node configs for network=default: []services.lbConfig(nil)\\\\nF1005 06:48:35.889931 6320 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: failed to add event handler: handler {0x1e60340 0x1e60020 0x1e5ffc0} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to se\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0df0ebcd87dc5af4b17736b43ec7b2395968b19ed16e775366af14e2b2d49c52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff3c6545ff3d2436da27bbd223b950b30b0c69a3e7f3034cb0e61380cb38d338\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ff3c6545ff3d2436da27bbd223b950b30b0c69a3e7f3034cb0e61380cb38d338\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:24Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-txzlk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:39Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:39 crc kubenswrapper[4846]: I1005 06:48:39.391805 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c9f812b-ecc8-4985-a670-53fa80fbb499\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2b5ee9781c01d6aec3ac778efef1067b6c3bfe4ed37431bdebe188fcde0165f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:47:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3fb4717692aa1fec0012b514ef8d18d1ca3707cfa4bb8248bb93841a0a40f11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:47:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0b6a7afe3094f1c92093c05c28344897a615d590c2fbb507c87e18a44c6c831\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://640ade2c733d7916ec213a201a1dd614d261c69eab654c80100efbc99a768995\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:47:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:39Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:39 crc kubenswrapper[4846]: I1005 06:48:39.409270 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:39Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:39 crc kubenswrapper[4846]: I1005 06:48:39.425633 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e67d0e447a8a6a381259e25d25badc791194e1291f56b1cd77bd63989b20408\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dkhl7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://516f88f4990f5962d8b694a3d889259cc5508a309919870191c2e1331640c973\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dkhl7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-fscvf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:39Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:39 crc kubenswrapper[4846]: I1005 06:48:39.440382 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-w2sj8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32931fce-7374-4cab-a2e0-79d96a842274\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0df535d65d39de6c9a3dfb931bafc4469adc22b72f7a7903beda4fef8cd48138\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9rtxf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:27Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-w2sj8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:39Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:39 crc kubenswrapper[4846]: I1005 06:48:39.463148 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-7wm5q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9bc7744f-d700-450f-8b8d-98140877fee9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90379e74c544fccca44f7acdc228ff378154390e211ac801b79f8d074012241e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h5p4q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-7wm5q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:39Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:39 crc kubenswrapper[4846]: I1005 06:48:39.475079 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:39 crc kubenswrapper[4846]: I1005 06:48:39.475270 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:39 crc kubenswrapper[4846]: I1005 06:48:39.475297 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:39 crc kubenswrapper[4846]: I1005 06:48:39.475328 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:39 crc kubenswrapper[4846]: I1005 06:48:39.475360 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:39Z","lastTransitionTime":"2025-10-05T06:48:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:39 crc kubenswrapper[4846]: I1005 06:48:39.484744 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ff10c31459b3b3724ee7b769bd6f72f9ed80cce5419b6adcdf69a3239b70756\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:39Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:39 crc kubenswrapper[4846]: I1005 06:48:39.496487 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 05 06:48:39 crc kubenswrapper[4846]: E1005 06:48:39.496619 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 05 06:48:39 crc kubenswrapper[4846]: I1005 06:48:39.496503 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 05 06:48:39 crc kubenswrapper[4846]: E1005 06:48:39.496778 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 05 06:48:39 crc kubenswrapper[4846]: I1005 06:48:39.500894 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://347f900743b7bba3daf92f21c7eba0625e8c774f50057bc9472f0c82119325e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:39Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:39 crc kubenswrapper[4846]: I1005 06:48:39.514915 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56f1fe1f9f06f0249cf937a6963fda82467732dc056ca2b2189be6d8dd9d4a05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7212ce5f42c1fb55146abb190a557a504b43f16af6df84f90e1073567a329ec5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:39Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:39 crc kubenswrapper[4846]: I1005 06:48:39.526372 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-9r2kt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"59b83f4f-71f0-4b28-9275-56d5f933ec2c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tdwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tdwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:38Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-9r2kt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:39Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:39 crc kubenswrapper[4846]: I1005 06:48:39.545010 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"775b5702-a131-4d6c-852d-91bfa056cac5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0839c1765df2c68d2554e9c66d289505e559a9672caf8069feeff2c4fdfd2171\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15dbaf2445704da0987f256f81dff2ee95664988541fa48ce94898afeadab2e8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd3b3ab323c6d5153054aa04af2d91f7b9e8436ff1e2f75f3594b57831b2a416\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://67c46f5f0a3ace95bd2e9479c5561126664e18b148c5eceea5bf5efc27e81a67\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c52e076eddfd307110b1f6de00a169b81af8f2785af7becef2a820ce363a5861\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-05T06:48:14Z\\\",\\\"message\\\":\\\"le observer\\\\nW1005 06:48:14.337870 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1005 06:48:14.338086 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1005 06:48:14.339389 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1467752215/tls.crt::/tmp/serving-cert-1467752215/tls.key\\\\\\\"\\\\nI1005 06:48:14.864345 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1005 06:48:14.868087 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1005 06:48:14.868114 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1005 06:48:14.868140 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1005 06:48:14.868147 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1005 06:48:14.877354 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1005 06:48:14.877412 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1005 06:48:14.877456 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1005 06:48:14.877471 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1005 06:48:14.877490 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1005 06:48:14.877501 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1005 06:48:14.877507 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1005 06:48:14.877515 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1005 06:48:14.880326 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:13Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4a091716ff935adb42ae2e88048619fd347bd174f64e5f9c97a01f81690e57f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:01Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dc5411b642f99b13d514d2f21defad0c394aeeb82a966f72300076617e2586c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5dc5411b642f99b13d514d2f21defad0c394aeeb82a966f72300076617e2586c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:47:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:47:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:47:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:39Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:39 crc kubenswrapper[4846]: I1005 06:48:39.573824 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:39Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:39 crc kubenswrapper[4846]: I1005 06:48:39.579428 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:39 crc kubenswrapper[4846]: I1005 06:48:39.579502 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:39 crc kubenswrapper[4846]: I1005 06:48:39.579523 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:39 crc kubenswrapper[4846]: I1005 06:48:39.579553 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:39 crc kubenswrapper[4846]: I1005 06:48:39.579574 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:39Z","lastTransitionTime":"2025-10-05T06:48:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:39 crc kubenswrapper[4846]: I1005 06:48:39.593319 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/59b83f4f-71f0-4b28-9275-56d5f933ec2c-metrics-certs\") pod \"network-metrics-daemon-9r2kt\" (UID: \"59b83f4f-71f0-4b28-9275-56d5f933ec2c\") " pod="openshift-multus/network-metrics-daemon-9r2kt" Oct 05 06:48:39 crc kubenswrapper[4846]: E1005 06:48:39.593522 4846 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 05 06:48:39 crc kubenswrapper[4846]: E1005 06:48:39.593602 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/59b83f4f-71f0-4b28-9275-56d5f933ec2c-metrics-certs podName:59b83f4f-71f0-4b28-9275-56d5f933ec2c nodeName:}" failed. No retries permitted until 2025-10-05 06:48:40.59357988 +0000 UTC m=+42.834432655 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/59b83f4f-71f0-4b28-9275-56d5f933ec2c-metrics-certs") pod "network-metrics-daemon-9r2kt" (UID: "59b83f4f-71f0-4b28-9275-56d5f933ec2c") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 05 06:48:39 crc kubenswrapper[4846]: I1005 06:48:39.607269 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-r7ztb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a5119cd9-af70-4324-a955-978306ab0b20\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06231a87429ec292cb57ebd88cda7ad547bcb32e80343b776833621a11c28ad1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jfcj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-r7ztb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:39Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:39 crc kubenswrapper[4846]: I1005 06:48:39.683044 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:39 crc kubenswrapper[4846]: I1005 06:48:39.683089 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:39 crc kubenswrapper[4846]: I1005 06:48:39.683103 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:39 crc kubenswrapper[4846]: I1005 06:48:39.683121 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:39 crc kubenswrapper[4846]: I1005 06:48:39.683134 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:39Z","lastTransitionTime":"2025-10-05T06:48:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:39 crc kubenswrapper[4846]: I1005 06:48:39.787007 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:39 crc kubenswrapper[4846]: I1005 06:48:39.787074 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:39 crc kubenswrapper[4846]: I1005 06:48:39.787092 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:39 crc kubenswrapper[4846]: I1005 06:48:39.787118 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:39 crc kubenswrapper[4846]: I1005 06:48:39.787136 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:39Z","lastTransitionTime":"2025-10-05T06:48:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:39 crc kubenswrapper[4846]: I1005 06:48:39.891126 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:39 crc kubenswrapper[4846]: I1005 06:48:39.891273 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:39 crc kubenswrapper[4846]: I1005 06:48:39.891302 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:39 crc kubenswrapper[4846]: I1005 06:48:39.891336 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:39 crc kubenswrapper[4846]: I1005 06:48:39.891359 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:39Z","lastTransitionTime":"2025-10-05T06:48:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:39 crc kubenswrapper[4846]: I1005 06:48:39.994786 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:39 crc kubenswrapper[4846]: I1005 06:48:39.994844 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:39 crc kubenswrapper[4846]: I1005 06:48:39.994861 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:39 crc kubenswrapper[4846]: I1005 06:48:39.994886 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:39 crc kubenswrapper[4846]: I1005 06:48:39.994903 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:39Z","lastTransitionTime":"2025-10-05T06:48:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:40 crc kubenswrapper[4846]: I1005 06:48:40.098525 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:40 crc kubenswrapper[4846]: I1005 06:48:40.098612 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:40 crc kubenswrapper[4846]: I1005 06:48:40.098644 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:40 crc kubenswrapper[4846]: I1005 06:48:40.098679 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:40 crc kubenswrapper[4846]: I1005 06:48:40.098720 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:40Z","lastTransitionTime":"2025-10-05T06:48:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:40 crc kubenswrapper[4846]: I1005 06:48:40.202066 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:40 crc kubenswrapper[4846]: I1005 06:48:40.202139 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:40 crc kubenswrapper[4846]: I1005 06:48:40.202161 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:40 crc kubenswrapper[4846]: I1005 06:48:40.202215 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:40 crc kubenswrapper[4846]: I1005 06:48:40.202237 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:40Z","lastTransitionTime":"2025-10-05T06:48:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:40 crc kubenswrapper[4846]: I1005 06:48:40.305226 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:40 crc kubenswrapper[4846]: I1005 06:48:40.305303 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:40 crc kubenswrapper[4846]: I1005 06:48:40.305324 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:40 crc kubenswrapper[4846]: I1005 06:48:40.305352 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:40 crc kubenswrapper[4846]: I1005 06:48:40.305376 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:40Z","lastTransitionTime":"2025-10-05T06:48:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:40 crc kubenswrapper[4846]: I1005 06:48:40.408408 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:40 crc kubenswrapper[4846]: I1005 06:48:40.408490 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:40 crc kubenswrapper[4846]: I1005 06:48:40.408516 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:40 crc kubenswrapper[4846]: I1005 06:48:40.408547 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:40 crc kubenswrapper[4846]: I1005 06:48:40.408572 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:40Z","lastTransitionTime":"2025-10-05T06:48:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:40 crc kubenswrapper[4846]: I1005 06:48:40.497423 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 05 06:48:40 crc kubenswrapper[4846]: E1005 06:48:40.497732 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 05 06:48:40 crc kubenswrapper[4846]: I1005 06:48:40.497419 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9r2kt" Oct 05 06:48:40 crc kubenswrapper[4846]: E1005 06:48:40.498551 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9r2kt" podUID="59b83f4f-71f0-4b28-9275-56d5f933ec2c" Oct 05 06:48:40 crc kubenswrapper[4846]: I1005 06:48:40.512748 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:40 crc kubenswrapper[4846]: I1005 06:48:40.512835 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:40 crc kubenswrapper[4846]: I1005 06:48:40.512856 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:40 crc kubenswrapper[4846]: I1005 06:48:40.512890 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:40 crc kubenswrapper[4846]: I1005 06:48:40.512912 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:40Z","lastTransitionTime":"2025-10-05T06:48:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:40 crc kubenswrapper[4846]: I1005 06:48:40.604393 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/59b83f4f-71f0-4b28-9275-56d5f933ec2c-metrics-certs\") pod \"network-metrics-daemon-9r2kt\" (UID: \"59b83f4f-71f0-4b28-9275-56d5f933ec2c\") " pod="openshift-multus/network-metrics-daemon-9r2kt" Oct 05 06:48:40 crc kubenswrapper[4846]: E1005 06:48:40.604681 4846 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 05 06:48:40 crc kubenswrapper[4846]: E1005 06:48:40.604876 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/59b83f4f-71f0-4b28-9275-56d5f933ec2c-metrics-certs podName:59b83f4f-71f0-4b28-9275-56d5f933ec2c nodeName:}" failed. No retries permitted until 2025-10-05 06:48:42.604832043 +0000 UTC m=+44.845684848 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/59b83f4f-71f0-4b28-9275-56d5f933ec2c-metrics-certs") pod "network-metrics-daemon-9r2kt" (UID: "59b83f4f-71f0-4b28-9275-56d5f933ec2c") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 05 06:48:40 crc kubenswrapper[4846]: I1005 06:48:40.616933 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:40 crc kubenswrapper[4846]: I1005 06:48:40.617041 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:40 crc kubenswrapper[4846]: I1005 06:48:40.617063 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:40 crc kubenswrapper[4846]: I1005 06:48:40.617092 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:40 crc kubenswrapper[4846]: I1005 06:48:40.617112 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:40Z","lastTransitionTime":"2025-10-05T06:48:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:40 crc kubenswrapper[4846]: I1005 06:48:40.721578 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:40 crc kubenswrapper[4846]: I1005 06:48:40.721683 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:40 crc kubenswrapper[4846]: I1005 06:48:40.721704 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:40 crc kubenswrapper[4846]: I1005 06:48:40.721732 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:40 crc kubenswrapper[4846]: I1005 06:48:40.721754 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:40Z","lastTransitionTime":"2025-10-05T06:48:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:40 crc kubenswrapper[4846]: I1005 06:48:40.825887 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:40 crc kubenswrapper[4846]: I1005 06:48:40.825956 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:40 crc kubenswrapper[4846]: I1005 06:48:40.825973 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:40 crc kubenswrapper[4846]: I1005 06:48:40.826002 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:40 crc kubenswrapper[4846]: I1005 06:48:40.826020 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:40Z","lastTransitionTime":"2025-10-05T06:48:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:40 crc kubenswrapper[4846]: I1005 06:48:40.930145 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:40 crc kubenswrapper[4846]: I1005 06:48:40.930249 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:40 crc kubenswrapper[4846]: I1005 06:48:40.930267 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:40 crc kubenswrapper[4846]: I1005 06:48:40.930291 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:40 crc kubenswrapper[4846]: I1005 06:48:40.930314 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:40Z","lastTransitionTime":"2025-10-05T06:48:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:41 crc kubenswrapper[4846]: I1005 06:48:41.033797 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:41 crc kubenswrapper[4846]: I1005 06:48:41.033860 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:41 crc kubenswrapper[4846]: I1005 06:48:41.033872 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:41 crc kubenswrapper[4846]: I1005 06:48:41.033894 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:41 crc kubenswrapper[4846]: I1005 06:48:41.033910 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:41Z","lastTransitionTime":"2025-10-05T06:48:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:41 crc kubenswrapper[4846]: I1005 06:48:41.137619 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:41 crc kubenswrapper[4846]: I1005 06:48:41.137701 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:41 crc kubenswrapper[4846]: I1005 06:48:41.137722 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:41 crc kubenswrapper[4846]: I1005 06:48:41.137746 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:41 crc kubenswrapper[4846]: I1005 06:48:41.137763 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:41Z","lastTransitionTime":"2025-10-05T06:48:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:41 crc kubenswrapper[4846]: I1005 06:48:41.241627 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:41 crc kubenswrapper[4846]: I1005 06:48:41.241704 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:41 crc kubenswrapper[4846]: I1005 06:48:41.241720 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:41 crc kubenswrapper[4846]: I1005 06:48:41.241750 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:41 crc kubenswrapper[4846]: I1005 06:48:41.241771 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:41Z","lastTransitionTime":"2025-10-05T06:48:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:41 crc kubenswrapper[4846]: I1005 06:48:41.344869 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:41 crc kubenswrapper[4846]: I1005 06:48:41.344927 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:41 crc kubenswrapper[4846]: I1005 06:48:41.344944 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:41 crc kubenswrapper[4846]: I1005 06:48:41.344964 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:41 crc kubenswrapper[4846]: I1005 06:48:41.344977 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:41Z","lastTransitionTime":"2025-10-05T06:48:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:41 crc kubenswrapper[4846]: I1005 06:48:41.448225 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:41 crc kubenswrapper[4846]: I1005 06:48:41.448301 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:41 crc kubenswrapper[4846]: I1005 06:48:41.448324 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:41 crc kubenswrapper[4846]: I1005 06:48:41.448357 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:41 crc kubenswrapper[4846]: I1005 06:48:41.448380 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:41Z","lastTransitionTime":"2025-10-05T06:48:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:41 crc kubenswrapper[4846]: I1005 06:48:41.497328 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 05 06:48:41 crc kubenswrapper[4846]: E1005 06:48:41.497464 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 05 06:48:41 crc kubenswrapper[4846]: I1005 06:48:41.497794 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 05 06:48:41 crc kubenswrapper[4846]: E1005 06:48:41.497845 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 05 06:48:41 crc kubenswrapper[4846]: I1005 06:48:41.554769 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:41 crc kubenswrapper[4846]: I1005 06:48:41.554860 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:41 crc kubenswrapper[4846]: I1005 06:48:41.554912 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:41 crc kubenswrapper[4846]: I1005 06:48:41.554942 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:41 crc kubenswrapper[4846]: I1005 06:48:41.555027 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:41Z","lastTransitionTime":"2025-10-05T06:48:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:41 crc kubenswrapper[4846]: I1005 06:48:41.659205 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:41 crc kubenswrapper[4846]: I1005 06:48:41.659249 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:41 crc kubenswrapper[4846]: I1005 06:48:41.659257 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:41 crc kubenswrapper[4846]: I1005 06:48:41.659272 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:41 crc kubenswrapper[4846]: I1005 06:48:41.659281 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:41Z","lastTransitionTime":"2025-10-05T06:48:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:41 crc kubenswrapper[4846]: I1005 06:48:41.761911 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:41 crc kubenswrapper[4846]: I1005 06:48:41.762390 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:41 crc kubenswrapper[4846]: I1005 06:48:41.762599 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:41 crc kubenswrapper[4846]: I1005 06:48:41.762794 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:41 crc kubenswrapper[4846]: I1005 06:48:41.762987 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:41Z","lastTransitionTime":"2025-10-05T06:48:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:41 crc kubenswrapper[4846]: I1005 06:48:41.870722 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:41 crc kubenswrapper[4846]: I1005 06:48:41.870765 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:41 crc kubenswrapper[4846]: I1005 06:48:41.870783 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:41 crc kubenswrapper[4846]: I1005 06:48:41.870807 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:41 crc kubenswrapper[4846]: I1005 06:48:41.870824 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:41Z","lastTransitionTime":"2025-10-05T06:48:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:41 crc kubenswrapper[4846]: I1005 06:48:41.974886 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:41 crc kubenswrapper[4846]: I1005 06:48:41.974974 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:41 crc kubenswrapper[4846]: I1005 06:48:41.975030 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:41 crc kubenswrapper[4846]: I1005 06:48:41.975057 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:41 crc kubenswrapper[4846]: I1005 06:48:41.975074 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:41Z","lastTransitionTime":"2025-10-05T06:48:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:42 crc kubenswrapper[4846]: I1005 06:48:42.078943 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:42 crc kubenswrapper[4846]: I1005 06:48:42.079012 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:42 crc kubenswrapper[4846]: I1005 06:48:42.079030 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:42 crc kubenswrapper[4846]: I1005 06:48:42.079058 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:42 crc kubenswrapper[4846]: I1005 06:48:42.079082 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:42Z","lastTransitionTime":"2025-10-05T06:48:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:42 crc kubenswrapper[4846]: I1005 06:48:42.182214 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:42 crc kubenswrapper[4846]: I1005 06:48:42.182266 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:42 crc kubenswrapper[4846]: I1005 06:48:42.182278 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:42 crc kubenswrapper[4846]: I1005 06:48:42.182298 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:42 crc kubenswrapper[4846]: I1005 06:48:42.182312 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:42Z","lastTransitionTime":"2025-10-05T06:48:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:42 crc kubenswrapper[4846]: I1005 06:48:42.285577 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:42 crc kubenswrapper[4846]: I1005 06:48:42.285632 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:42 crc kubenswrapper[4846]: I1005 06:48:42.285645 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:42 crc kubenswrapper[4846]: I1005 06:48:42.285666 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:42 crc kubenswrapper[4846]: I1005 06:48:42.285680 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:42Z","lastTransitionTime":"2025-10-05T06:48:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:42 crc kubenswrapper[4846]: I1005 06:48:42.389277 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:42 crc kubenswrapper[4846]: I1005 06:48:42.389340 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:42 crc kubenswrapper[4846]: I1005 06:48:42.389357 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:42 crc kubenswrapper[4846]: I1005 06:48:42.389377 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:42 crc kubenswrapper[4846]: I1005 06:48:42.389391 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:42Z","lastTransitionTime":"2025-10-05T06:48:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:42 crc kubenswrapper[4846]: I1005 06:48:42.492608 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:42 crc kubenswrapper[4846]: I1005 06:48:42.492685 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:42 crc kubenswrapper[4846]: I1005 06:48:42.492704 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:42 crc kubenswrapper[4846]: I1005 06:48:42.492735 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:42 crc kubenswrapper[4846]: I1005 06:48:42.492755 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:42Z","lastTransitionTime":"2025-10-05T06:48:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:42 crc kubenswrapper[4846]: I1005 06:48:42.496857 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 05 06:48:42 crc kubenswrapper[4846]: I1005 06:48:42.497000 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9r2kt" Oct 05 06:48:42 crc kubenswrapper[4846]: E1005 06:48:42.497218 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 05 06:48:42 crc kubenswrapper[4846]: E1005 06:48:42.497443 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9r2kt" podUID="59b83f4f-71f0-4b28-9275-56d5f933ec2c" Oct 05 06:48:42 crc kubenswrapper[4846]: I1005 06:48:42.596008 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:42 crc kubenswrapper[4846]: I1005 06:48:42.596081 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:42 crc kubenswrapper[4846]: I1005 06:48:42.596097 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:42 crc kubenswrapper[4846]: I1005 06:48:42.596123 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:42 crc kubenswrapper[4846]: I1005 06:48:42.596142 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:42Z","lastTransitionTime":"2025-10-05T06:48:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:42 crc kubenswrapper[4846]: I1005 06:48:42.629824 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/59b83f4f-71f0-4b28-9275-56d5f933ec2c-metrics-certs\") pod \"network-metrics-daemon-9r2kt\" (UID: \"59b83f4f-71f0-4b28-9275-56d5f933ec2c\") " pod="openshift-multus/network-metrics-daemon-9r2kt" Oct 05 06:48:42 crc kubenswrapper[4846]: E1005 06:48:42.630049 4846 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 05 06:48:42 crc kubenswrapper[4846]: E1005 06:48:42.630148 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/59b83f4f-71f0-4b28-9275-56d5f933ec2c-metrics-certs podName:59b83f4f-71f0-4b28-9275-56d5f933ec2c nodeName:}" failed. No retries permitted until 2025-10-05 06:48:46.630114218 +0000 UTC m=+48.870967033 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/59b83f4f-71f0-4b28-9275-56d5f933ec2c-metrics-certs") pod "network-metrics-daemon-9r2kt" (UID: "59b83f4f-71f0-4b28-9275-56d5f933ec2c") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 05 06:48:42 crc kubenswrapper[4846]: I1005 06:48:42.700154 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:42 crc kubenswrapper[4846]: I1005 06:48:42.700246 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:42 crc kubenswrapper[4846]: I1005 06:48:42.700263 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:42 crc kubenswrapper[4846]: I1005 06:48:42.700294 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:42 crc kubenswrapper[4846]: I1005 06:48:42.700314 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:42Z","lastTransitionTime":"2025-10-05T06:48:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:42 crc kubenswrapper[4846]: I1005 06:48:42.803493 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:42 crc kubenswrapper[4846]: I1005 06:48:42.803595 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:42 crc kubenswrapper[4846]: I1005 06:48:42.803613 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:42 crc kubenswrapper[4846]: I1005 06:48:42.803643 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:42 crc kubenswrapper[4846]: I1005 06:48:42.803660 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:42Z","lastTransitionTime":"2025-10-05T06:48:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:42 crc kubenswrapper[4846]: I1005 06:48:42.907686 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:42 crc kubenswrapper[4846]: I1005 06:48:42.907957 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:42 crc kubenswrapper[4846]: I1005 06:48:42.908566 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:42 crc kubenswrapper[4846]: I1005 06:48:42.908624 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:42 crc kubenswrapper[4846]: I1005 06:48:42.908647 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:42Z","lastTransitionTime":"2025-10-05T06:48:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:43 crc kubenswrapper[4846]: I1005 06:48:43.012759 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:43 crc kubenswrapper[4846]: I1005 06:48:43.012826 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:43 crc kubenswrapper[4846]: I1005 06:48:43.012843 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:43 crc kubenswrapper[4846]: I1005 06:48:43.012869 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:43 crc kubenswrapper[4846]: I1005 06:48:43.012887 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:43Z","lastTransitionTime":"2025-10-05T06:48:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:43 crc kubenswrapper[4846]: I1005 06:48:43.116434 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:43 crc kubenswrapper[4846]: I1005 06:48:43.116500 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:43 crc kubenswrapper[4846]: I1005 06:48:43.116517 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:43 crc kubenswrapper[4846]: I1005 06:48:43.116543 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:43 crc kubenswrapper[4846]: I1005 06:48:43.116560 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:43Z","lastTransitionTime":"2025-10-05T06:48:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:43 crc kubenswrapper[4846]: I1005 06:48:43.219613 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:43 crc kubenswrapper[4846]: I1005 06:48:43.219697 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:43 crc kubenswrapper[4846]: I1005 06:48:43.219722 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:43 crc kubenswrapper[4846]: I1005 06:48:43.219753 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:43 crc kubenswrapper[4846]: I1005 06:48:43.219780 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:43Z","lastTransitionTime":"2025-10-05T06:48:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:43 crc kubenswrapper[4846]: I1005 06:48:43.323244 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:43 crc kubenswrapper[4846]: I1005 06:48:43.323318 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:43 crc kubenswrapper[4846]: I1005 06:48:43.323345 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:43 crc kubenswrapper[4846]: I1005 06:48:43.323377 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:43 crc kubenswrapper[4846]: I1005 06:48:43.323395 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:43Z","lastTransitionTime":"2025-10-05T06:48:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:43 crc kubenswrapper[4846]: I1005 06:48:43.427026 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:43 crc kubenswrapper[4846]: I1005 06:48:43.427089 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:43 crc kubenswrapper[4846]: I1005 06:48:43.427112 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:43 crc kubenswrapper[4846]: I1005 06:48:43.427142 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:43 crc kubenswrapper[4846]: I1005 06:48:43.427170 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:43Z","lastTransitionTime":"2025-10-05T06:48:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:43 crc kubenswrapper[4846]: I1005 06:48:43.496738 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 05 06:48:43 crc kubenswrapper[4846]: I1005 06:48:43.496770 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 05 06:48:43 crc kubenswrapper[4846]: E1005 06:48:43.496977 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 05 06:48:43 crc kubenswrapper[4846]: E1005 06:48:43.497150 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 05 06:48:43 crc kubenswrapper[4846]: I1005 06:48:43.530361 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:43 crc kubenswrapper[4846]: I1005 06:48:43.530418 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:43 crc kubenswrapper[4846]: I1005 06:48:43.530434 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:43 crc kubenswrapper[4846]: I1005 06:48:43.530455 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:43 crc kubenswrapper[4846]: I1005 06:48:43.530472 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:43Z","lastTransitionTime":"2025-10-05T06:48:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:43 crc kubenswrapper[4846]: I1005 06:48:43.633090 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:43 crc kubenswrapper[4846]: I1005 06:48:43.633262 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:43 crc kubenswrapper[4846]: I1005 06:48:43.633281 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:43 crc kubenswrapper[4846]: I1005 06:48:43.633306 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:43 crc kubenswrapper[4846]: I1005 06:48:43.633334 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:43Z","lastTransitionTime":"2025-10-05T06:48:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:43 crc kubenswrapper[4846]: I1005 06:48:43.735846 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:43 crc kubenswrapper[4846]: I1005 06:48:43.735897 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:43 crc kubenswrapper[4846]: I1005 06:48:43.735907 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:43 crc kubenswrapper[4846]: I1005 06:48:43.735922 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:43 crc kubenswrapper[4846]: I1005 06:48:43.735932 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:43Z","lastTransitionTime":"2025-10-05T06:48:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:43 crc kubenswrapper[4846]: I1005 06:48:43.837913 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:43 crc kubenswrapper[4846]: I1005 06:48:43.837954 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:43 crc kubenswrapper[4846]: I1005 06:48:43.837964 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:43 crc kubenswrapper[4846]: I1005 06:48:43.837980 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:43 crc kubenswrapper[4846]: I1005 06:48:43.837993 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:43Z","lastTransitionTime":"2025-10-05T06:48:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:43 crc kubenswrapper[4846]: I1005 06:48:43.947415 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:43 crc kubenswrapper[4846]: I1005 06:48:43.947483 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:43 crc kubenswrapper[4846]: I1005 06:48:43.947501 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:43 crc kubenswrapper[4846]: I1005 06:48:43.947526 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:43 crc kubenswrapper[4846]: I1005 06:48:43.947544 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:43Z","lastTransitionTime":"2025-10-05T06:48:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:44 crc kubenswrapper[4846]: I1005 06:48:44.050856 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:44 crc kubenswrapper[4846]: I1005 06:48:44.050928 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:44 crc kubenswrapper[4846]: I1005 06:48:44.050944 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:44 crc kubenswrapper[4846]: I1005 06:48:44.050966 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:44 crc kubenswrapper[4846]: I1005 06:48:44.050984 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:44Z","lastTransitionTime":"2025-10-05T06:48:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:44 crc kubenswrapper[4846]: I1005 06:48:44.154097 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:44 crc kubenswrapper[4846]: I1005 06:48:44.154146 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:44 crc kubenswrapper[4846]: I1005 06:48:44.154158 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:44 crc kubenswrapper[4846]: I1005 06:48:44.154194 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:44 crc kubenswrapper[4846]: I1005 06:48:44.154208 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:44Z","lastTransitionTime":"2025-10-05T06:48:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:44 crc kubenswrapper[4846]: I1005 06:48:44.257575 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:44 crc kubenswrapper[4846]: I1005 06:48:44.257633 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:44 crc kubenswrapper[4846]: I1005 06:48:44.257646 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:44 crc kubenswrapper[4846]: I1005 06:48:44.257662 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:44 crc kubenswrapper[4846]: I1005 06:48:44.257675 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:44Z","lastTransitionTime":"2025-10-05T06:48:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:44 crc kubenswrapper[4846]: I1005 06:48:44.359831 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:44 crc kubenswrapper[4846]: I1005 06:48:44.359877 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:44 crc kubenswrapper[4846]: I1005 06:48:44.359889 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:44 crc kubenswrapper[4846]: I1005 06:48:44.359905 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:44 crc kubenswrapper[4846]: I1005 06:48:44.359920 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:44Z","lastTransitionTime":"2025-10-05T06:48:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:44 crc kubenswrapper[4846]: I1005 06:48:44.463218 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:44 crc kubenswrapper[4846]: I1005 06:48:44.463297 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:44 crc kubenswrapper[4846]: I1005 06:48:44.463321 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:44 crc kubenswrapper[4846]: I1005 06:48:44.463355 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:44 crc kubenswrapper[4846]: I1005 06:48:44.463378 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:44Z","lastTransitionTime":"2025-10-05T06:48:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:44 crc kubenswrapper[4846]: I1005 06:48:44.496986 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9r2kt" Oct 05 06:48:44 crc kubenswrapper[4846]: I1005 06:48:44.497015 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 05 06:48:44 crc kubenswrapper[4846]: E1005 06:48:44.497232 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9r2kt" podUID="59b83f4f-71f0-4b28-9275-56d5f933ec2c" Oct 05 06:48:44 crc kubenswrapper[4846]: E1005 06:48:44.497426 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 05 06:48:44 crc kubenswrapper[4846]: I1005 06:48:44.565878 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:44 crc kubenswrapper[4846]: I1005 06:48:44.565960 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:44 crc kubenswrapper[4846]: I1005 06:48:44.565985 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:44 crc kubenswrapper[4846]: I1005 06:48:44.566025 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:44 crc kubenswrapper[4846]: I1005 06:48:44.566081 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:44Z","lastTransitionTime":"2025-10-05T06:48:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:44 crc kubenswrapper[4846]: I1005 06:48:44.669847 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:44 crc kubenswrapper[4846]: I1005 06:48:44.670398 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:44 crc kubenswrapper[4846]: I1005 06:48:44.670578 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:44 crc kubenswrapper[4846]: I1005 06:48:44.670842 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:44 crc kubenswrapper[4846]: I1005 06:48:44.671100 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:44Z","lastTransitionTime":"2025-10-05T06:48:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:44 crc kubenswrapper[4846]: I1005 06:48:44.774606 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:44 crc kubenswrapper[4846]: I1005 06:48:44.774695 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:44 crc kubenswrapper[4846]: I1005 06:48:44.774721 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:44 crc kubenswrapper[4846]: I1005 06:48:44.774750 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:44 crc kubenswrapper[4846]: I1005 06:48:44.774772 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:44Z","lastTransitionTime":"2025-10-05T06:48:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:44 crc kubenswrapper[4846]: I1005 06:48:44.878800 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:44 crc kubenswrapper[4846]: I1005 06:48:44.878930 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:44 crc kubenswrapper[4846]: I1005 06:48:44.878949 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:44 crc kubenswrapper[4846]: I1005 06:48:44.878974 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:44 crc kubenswrapper[4846]: I1005 06:48:44.878991 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:44Z","lastTransitionTime":"2025-10-05T06:48:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:44 crc kubenswrapper[4846]: I1005 06:48:44.982533 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:44 crc kubenswrapper[4846]: I1005 06:48:44.983418 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:44 crc kubenswrapper[4846]: I1005 06:48:44.983589 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:44 crc kubenswrapper[4846]: I1005 06:48:44.983744 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:44 crc kubenswrapper[4846]: I1005 06:48:44.983885 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:44Z","lastTransitionTime":"2025-10-05T06:48:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:45 crc kubenswrapper[4846]: I1005 06:48:45.037057 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:45 crc kubenswrapper[4846]: I1005 06:48:45.037410 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:45 crc kubenswrapper[4846]: I1005 06:48:45.037539 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:45 crc kubenswrapper[4846]: I1005 06:48:45.037696 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:45 crc kubenswrapper[4846]: I1005 06:48:45.037814 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:45Z","lastTransitionTime":"2025-10-05T06:48:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:45 crc kubenswrapper[4846]: E1005 06:48:45.061577 4846 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:48:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:48:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:45Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:48:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:48:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:45Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"04ce628d-2f61-4931-aad6-615924ea53fb\\\",\\\"systemUUID\\\":\\\"a050a23b-e773-4ba9-989b-360aa53a0605\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:45Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:45 crc kubenswrapper[4846]: I1005 06:48:45.068514 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:45 crc kubenswrapper[4846]: I1005 06:48:45.068911 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:45 crc kubenswrapper[4846]: I1005 06:48:45.069093 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:45 crc kubenswrapper[4846]: I1005 06:48:45.069354 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:45 crc kubenswrapper[4846]: I1005 06:48:45.069558 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:45Z","lastTransitionTime":"2025-10-05T06:48:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:45 crc kubenswrapper[4846]: E1005 06:48:45.091843 4846 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:48:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:48:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:45Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:48:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:48:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:45Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"04ce628d-2f61-4931-aad6-615924ea53fb\\\",\\\"systemUUID\\\":\\\"a050a23b-e773-4ba9-989b-360aa53a0605\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:45Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:45 crc kubenswrapper[4846]: I1005 06:48:45.098551 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:45 crc kubenswrapper[4846]: I1005 06:48:45.098865 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:45 crc kubenswrapper[4846]: I1005 06:48:45.099066 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:45 crc kubenswrapper[4846]: I1005 06:48:45.099255 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:45 crc kubenswrapper[4846]: I1005 06:48:45.099409 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:45Z","lastTransitionTime":"2025-10-05T06:48:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:45 crc kubenswrapper[4846]: E1005 06:48:45.121320 4846 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:48:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:48:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:45Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:48:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:48:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:45Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"04ce628d-2f61-4931-aad6-615924ea53fb\\\",\\\"systemUUID\\\":\\\"a050a23b-e773-4ba9-989b-360aa53a0605\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:45Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:45 crc kubenswrapper[4846]: I1005 06:48:45.126256 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:45 crc kubenswrapper[4846]: I1005 06:48:45.126310 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:45 crc kubenswrapper[4846]: I1005 06:48:45.126327 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:45 crc kubenswrapper[4846]: I1005 06:48:45.126356 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:45 crc kubenswrapper[4846]: I1005 06:48:45.126385 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:45Z","lastTransitionTime":"2025-10-05T06:48:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:45 crc kubenswrapper[4846]: E1005 06:48:45.147247 4846 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:48:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:48:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:45Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:48:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:48:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:45Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"04ce628d-2f61-4931-aad6-615924ea53fb\\\",\\\"systemUUID\\\":\\\"a050a23b-e773-4ba9-989b-360aa53a0605\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:45Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:45 crc kubenswrapper[4846]: I1005 06:48:45.152969 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:45 crc kubenswrapper[4846]: I1005 06:48:45.153029 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:45 crc kubenswrapper[4846]: I1005 06:48:45.153046 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:45 crc kubenswrapper[4846]: I1005 06:48:45.153076 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:45 crc kubenswrapper[4846]: I1005 06:48:45.153094 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:45Z","lastTransitionTime":"2025-10-05T06:48:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:45 crc kubenswrapper[4846]: E1005 06:48:45.174452 4846 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:48:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:48:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:45Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:48:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:48:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:45Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"04ce628d-2f61-4931-aad6-615924ea53fb\\\",\\\"systemUUID\\\":\\\"a050a23b-e773-4ba9-989b-360aa53a0605\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:45Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:45 crc kubenswrapper[4846]: E1005 06:48:45.174670 4846 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 05 06:48:45 crc kubenswrapper[4846]: I1005 06:48:45.176862 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:45 crc kubenswrapper[4846]: I1005 06:48:45.176919 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:45 crc kubenswrapper[4846]: I1005 06:48:45.176928 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:45 crc kubenswrapper[4846]: I1005 06:48:45.176945 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:45 crc kubenswrapper[4846]: I1005 06:48:45.176956 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:45Z","lastTransitionTime":"2025-10-05T06:48:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:45 crc kubenswrapper[4846]: I1005 06:48:45.280600 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:45 crc kubenswrapper[4846]: I1005 06:48:45.280670 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:45 crc kubenswrapper[4846]: I1005 06:48:45.280688 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:45 crc kubenswrapper[4846]: I1005 06:48:45.280715 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:45 crc kubenswrapper[4846]: I1005 06:48:45.280737 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:45Z","lastTransitionTime":"2025-10-05T06:48:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:45 crc kubenswrapper[4846]: I1005 06:48:45.384428 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:45 crc kubenswrapper[4846]: I1005 06:48:45.384488 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:45 crc kubenswrapper[4846]: I1005 06:48:45.384504 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:45 crc kubenswrapper[4846]: I1005 06:48:45.384531 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:45 crc kubenswrapper[4846]: I1005 06:48:45.384548 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:45Z","lastTransitionTime":"2025-10-05T06:48:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:45 crc kubenswrapper[4846]: I1005 06:48:45.488080 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:45 crc kubenswrapper[4846]: I1005 06:48:45.488147 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:45 crc kubenswrapper[4846]: I1005 06:48:45.488166 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:45 crc kubenswrapper[4846]: I1005 06:48:45.488225 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:45 crc kubenswrapper[4846]: I1005 06:48:45.488250 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:45Z","lastTransitionTime":"2025-10-05T06:48:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:45 crc kubenswrapper[4846]: I1005 06:48:45.497515 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 05 06:48:45 crc kubenswrapper[4846]: I1005 06:48:45.497515 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 05 06:48:45 crc kubenswrapper[4846]: E1005 06:48:45.497683 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 05 06:48:45 crc kubenswrapper[4846]: E1005 06:48:45.497800 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 05 06:48:45 crc kubenswrapper[4846]: I1005 06:48:45.591654 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:45 crc kubenswrapper[4846]: I1005 06:48:45.591738 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:45 crc kubenswrapper[4846]: I1005 06:48:45.591756 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:45 crc kubenswrapper[4846]: I1005 06:48:45.591782 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:45 crc kubenswrapper[4846]: I1005 06:48:45.591810 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:45Z","lastTransitionTime":"2025-10-05T06:48:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:45 crc kubenswrapper[4846]: I1005 06:48:45.696633 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:45 crc kubenswrapper[4846]: I1005 06:48:45.696730 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:45 crc kubenswrapper[4846]: I1005 06:48:45.696757 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:45 crc kubenswrapper[4846]: I1005 06:48:45.696788 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:45 crc kubenswrapper[4846]: I1005 06:48:45.696813 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:45Z","lastTransitionTime":"2025-10-05T06:48:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:45 crc kubenswrapper[4846]: I1005 06:48:45.800392 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:45 crc kubenswrapper[4846]: I1005 06:48:45.800612 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:45 crc kubenswrapper[4846]: I1005 06:48:45.800636 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:45 crc kubenswrapper[4846]: I1005 06:48:45.800697 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:45 crc kubenswrapper[4846]: I1005 06:48:45.800716 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:45Z","lastTransitionTime":"2025-10-05T06:48:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:45 crc kubenswrapper[4846]: I1005 06:48:45.904743 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:45 crc kubenswrapper[4846]: I1005 06:48:45.904956 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:45 crc kubenswrapper[4846]: I1005 06:48:45.904975 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:45 crc kubenswrapper[4846]: I1005 06:48:45.905003 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:45 crc kubenswrapper[4846]: I1005 06:48:45.905021 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:45Z","lastTransitionTime":"2025-10-05T06:48:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:46 crc kubenswrapper[4846]: I1005 06:48:46.007725 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:46 crc kubenswrapper[4846]: I1005 06:48:46.007784 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:46 crc kubenswrapper[4846]: I1005 06:48:46.007801 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:46 crc kubenswrapper[4846]: I1005 06:48:46.007824 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:46 crc kubenswrapper[4846]: I1005 06:48:46.007842 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:46Z","lastTransitionTime":"2025-10-05T06:48:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:46 crc kubenswrapper[4846]: I1005 06:48:46.111430 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:46 crc kubenswrapper[4846]: I1005 06:48:46.111482 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:46 crc kubenswrapper[4846]: I1005 06:48:46.111498 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:46 crc kubenswrapper[4846]: I1005 06:48:46.111523 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:46 crc kubenswrapper[4846]: I1005 06:48:46.111540 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:46Z","lastTransitionTime":"2025-10-05T06:48:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:46 crc kubenswrapper[4846]: I1005 06:48:46.215322 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:46 crc kubenswrapper[4846]: I1005 06:48:46.215418 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:46 crc kubenswrapper[4846]: I1005 06:48:46.215456 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:46 crc kubenswrapper[4846]: I1005 06:48:46.215489 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:46 crc kubenswrapper[4846]: I1005 06:48:46.215511 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:46Z","lastTransitionTime":"2025-10-05T06:48:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:46 crc kubenswrapper[4846]: I1005 06:48:46.319408 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:46 crc kubenswrapper[4846]: I1005 06:48:46.319492 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:46 crc kubenswrapper[4846]: I1005 06:48:46.319511 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:46 crc kubenswrapper[4846]: I1005 06:48:46.319541 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:46 crc kubenswrapper[4846]: I1005 06:48:46.319564 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:46Z","lastTransitionTime":"2025-10-05T06:48:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:46 crc kubenswrapper[4846]: I1005 06:48:46.423061 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:46 crc kubenswrapper[4846]: I1005 06:48:46.423139 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:46 crc kubenswrapper[4846]: I1005 06:48:46.423156 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:46 crc kubenswrapper[4846]: I1005 06:48:46.423213 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:46 crc kubenswrapper[4846]: I1005 06:48:46.423236 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:46Z","lastTransitionTime":"2025-10-05T06:48:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:46 crc kubenswrapper[4846]: I1005 06:48:46.496890 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9r2kt" Oct 05 06:48:46 crc kubenswrapper[4846]: I1005 06:48:46.497005 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 05 06:48:46 crc kubenswrapper[4846]: E1005 06:48:46.497223 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9r2kt" podUID="59b83f4f-71f0-4b28-9275-56d5f933ec2c" Oct 05 06:48:46 crc kubenswrapper[4846]: E1005 06:48:46.497350 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 05 06:48:46 crc kubenswrapper[4846]: I1005 06:48:46.526652 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:46 crc kubenswrapper[4846]: I1005 06:48:46.526711 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:46 crc kubenswrapper[4846]: I1005 06:48:46.526728 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:46 crc kubenswrapper[4846]: I1005 06:48:46.526751 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:46 crc kubenswrapper[4846]: I1005 06:48:46.526769 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:46Z","lastTransitionTime":"2025-10-05T06:48:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:46 crc kubenswrapper[4846]: I1005 06:48:46.630326 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:46 crc kubenswrapper[4846]: I1005 06:48:46.630412 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:46 crc kubenswrapper[4846]: I1005 06:48:46.630431 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:46 crc kubenswrapper[4846]: I1005 06:48:46.630459 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:46 crc kubenswrapper[4846]: I1005 06:48:46.630478 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:46Z","lastTransitionTime":"2025-10-05T06:48:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:46 crc kubenswrapper[4846]: I1005 06:48:46.673170 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/59b83f4f-71f0-4b28-9275-56d5f933ec2c-metrics-certs\") pod \"network-metrics-daemon-9r2kt\" (UID: \"59b83f4f-71f0-4b28-9275-56d5f933ec2c\") " pod="openshift-multus/network-metrics-daemon-9r2kt" Oct 05 06:48:46 crc kubenswrapper[4846]: E1005 06:48:46.673371 4846 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 05 06:48:46 crc kubenswrapper[4846]: E1005 06:48:46.673437 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/59b83f4f-71f0-4b28-9275-56d5f933ec2c-metrics-certs podName:59b83f4f-71f0-4b28-9275-56d5f933ec2c nodeName:}" failed. No retries permitted until 2025-10-05 06:48:54.67342089 +0000 UTC m=+56.914273665 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/59b83f4f-71f0-4b28-9275-56d5f933ec2c-metrics-certs") pod "network-metrics-daemon-9r2kt" (UID: "59b83f4f-71f0-4b28-9275-56d5f933ec2c") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 05 06:48:46 crc kubenswrapper[4846]: I1005 06:48:46.734151 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:46 crc kubenswrapper[4846]: I1005 06:48:46.734247 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:46 crc kubenswrapper[4846]: I1005 06:48:46.734261 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:46 crc kubenswrapper[4846]: I1005 06:48:46.734282 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:46 crc kubenswrapper[4846]: I1005 06:48:46.734295 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:46Z","lastTransitionTime":"2025-10-05T06:48:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:46 crc kubenswrapper[4846]: I1005 06:48:46.837347 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:46 crc kubenswrapper[4846]: I1005 06:48:46.837437 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:46 crc kubenswrapper[4846]: I1005 06:48:46.837567 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:46 crc kubenswrapper[4846]: I1005 06:48:46.837603 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:46 crc kubenswrapper[4846]: I1005 06:48:46.837628 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:46Z","lastTransitionTime":"2025-10-05T06:48:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:46 crc kubenswrapper[4846]: I1005 06:48:46.940380 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:46 crc kubenswrapper[4846]: I1005 06:48:46.940429 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:46 crc kubenswrapper[4846]: I1005 06:48:46.940438 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:46 crc kubenswrapper[4846]: I1005 06:48:46.940453 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:46 crc kubenswrapper[4846]: I1005 06:48:46.940464 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:46Z","lastTransitionTime":"2025-10-05T06:48:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:47 crc kubenswrapper[4846]: I1005 06:48:47.043336 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:47 crc kubenswrapper[4846]: I1005 06:48:47.043429 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:47 crc kubenswrapper[4846]: I1005 06:48:47.043448 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:47 crc kubenswrapper[4846]: I1005 06:48:47.043475 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:47 crc kubenswrapper[4846]: I1005 06:48:47.043497 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:47Z","lastTransitionTime":"2025-10-05T06:48:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:47 crc kubenswrapper[4846]: I1005 06:48:47.146323 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:47 crc kubenswrapper[4846]: I1005 06:48:47.146370 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:47 crc kubenswrapper[4846]: I1005 06:48:47.146379 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:47 crc kubenswrapper[4846]: I1005 06:48:47.146397 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:47 crc kubenswrapper[4846]: I1005 06:48:47.146411 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:47Z","lastTransitionTime":"2025-10-05T06:48:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:47 crc kubenswrapper[4846]: I1005 06:48:47.248793 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:47 crc kubenswrapper[4846]: I1005 06:48:47.248857 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:47 crc kubenswrapper[4846]: I1005 06:48:47.248884 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:47 crc kubenswrapper[4846]: I1005 06:48:47.248915 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:47 crc kubenswrapper[4846]: I1005 06:48:47.248934 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:47Z","lastTransitionTime":"2025-10-05T06:48:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:47 crc kubenswrapper[4846]: I1005 06:48:47.351701 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:47 crc kubenswrapper[4846]: I1005 06:48:47.351753 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:47 crc kubenswrapper[4846]: I1005 06:48:47.351769 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:47 crc kubenswrapper[4846]: I1005 06:48:47.351794 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:47 crc kubenswrapper[4846]: I1005 06:48:47.351813 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:47Z","lastTransitionTime":"2025-10-05T06:48:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:47 crc kubenswrapper[4846]: I1005 06:48:47.455068 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:47 crc kubenswrapper[4846]: I1005 06:48:47.455102 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:47 crc kubenswrapper[4846]: I1005 06:48:47.455113 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:47 crc kubenswrapper[4846]: I1005 06:48:47.455127 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:47 crc kubenswrapper[4846]: I1005 06:48:47.455136 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:47Z","lastTransitionTime":"2025-10-05T06:48:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:47 crc kubenswrapper[4846]: I1005 06:48:47.497339 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 05 06:48:47 crc kubenswrapper[4846]: I1005 06:48:47.497365 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 05 06:48:47 crc kubenswrapper[4846]: E1005 06:48:47.497554 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 05 06:48:47 crc kubenswrapper[4846]: E1005 06:48:47.497763 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 05 06:48:47 crc kubenswrapper[4846]: I1005 06:48:47.558836 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:47 crc kubenswrapper[4846]: I1005 06:48:47.558900 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:47 crc kubenswrapper[4846]: I1005 06:48:47.558917 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:47 crc kubenswrapper[4846]: I1005 06:48:47.558984 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:47 crc kubenswrapper[4846]: I1005 06:48:47.559006 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:47Z","lastTransitionTime":"2025-10-05T06:48:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:47 crc kubenswrapper[4846]: I1005 06:48:47.661997 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:47 crc kubenswrapper[4846]: I1005 06:48:47.662068 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:47 crc kubenswrapper[4846]: I1005 06:48:47.662085 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:47 crc kubenswrapper[4846]: I1005 06:48:47.662108 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:47 crc kubenswrapper[4846]: I1005 06:48:47.662126 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:47Z","lastTransitionTime":"2025-10-05T06:48:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:47 crc kubenswrapper[4846]: I1005 06:48:47.765756 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:47 crc kubenswrapper[4846]: I1005 06:48:47.765830 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:47 crc kubenswrapper[4846]: I1005 06:48:47.765846 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:47 crc kubenswrapper[4846]: I1005 06:48:47.765879 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:47 crc kubenswrapper[4846]: I1005 06:48:47.765897 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:47Z","lastTransitionTime":"2025-10-05T06:48:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:47 crc kubenswrapper[4846]: I1005 06:48:47.864566 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 05 06:48:47 crc kubenswrapper[4846]: I1005 06:48:47.869454 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:47 crc kubenswrapper[4846]: I1005 06:48:47.869500 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:47 crc kubenswrapper[4846]: I1005 06:48:47.869510 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:47 crc kubenswrapper[4846]: I1005 06:48:47.869526 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:47 crc kubenswrapper[4846]: I1005 06:48:47.869539 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:47Z","lastTransitionTime":"2025-10-05T06:48:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:47 crc kubenswrapper[4846]: I1005 06:48:47.887301 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56f1fe1f9f06f0249cf937a6963fda82467732dc056ca2b2189be6d8dd9d4a05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7212ce5f42c1fb55146abb190a557a504b43f16af6df84f90e1073567a329ec5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:47Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:47 crc kubenswrapper[4846]: I1005 06:48:47.906237 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-7wm5q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9bc7744f-d700-450f-8b8d-98140877fee9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90379e74c544fccca44f7acdc228ff378154390e211ac801b79f8d074012241e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h5p4q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-7wm5q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:47Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:47 crc kubenswrapper[4846]: I1005 06:48:47.925993 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ff10c31459b3b3724ee7b769bd6f72f9ed80cce5419b6adcdf69a3239b70756\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:47Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:47 crc kubenswrapper[4846]: I1005 06:48:47.944816 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://347f900743b7bba3daf92f21c7eba0625e8c774f50057bc9472f0c82119325e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:47Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:47 crc kubenswrapper[4846]: I1005 06:48:47.959792 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-r7ztb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a5119cd9-af70-4324-a955-978306ab0b20\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06231a87429ec292cb57ebd88cda7ad547bcb32e80343b776833621a11c28ad1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jfcj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-r7ztb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:47Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:47 crc kubenswrapper[4846]: I1005 06:48:47.972525 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:47 crc kubenswrapper[4846]: I1005 06:48:47.972575 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:47 crc kubenswrapper[4846]: I1005 06:48:47.972585 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:47 crc kubenswrapper[4846]: I1005 06:48:47.972601 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:47 crc kubenswrapper[4846]: I1005 06:48:47.972613 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:47Z","lastTransitionTime":"2025-10-05T06:48:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:47 crc kubenswrapper[4846]: I1005 06:48:47.976673 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-9r2kt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"59b83f4f-71f0-4b28-9275-56d5f933ec2c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tdwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tdwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:38Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-9r2kt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:47Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:48 crc kubenswrapper[4846]: I1005 06:48:48.000781 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"775b5702-a131-4d6c-852d-91bfa056cac5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0839c1765df2c68d2554e9c66d289505e559a9672caf8069feeff2c4fdfd2171\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15dbaf2445704da0987f256f81dff2ee95664988541fa48ce94898afeadab2e8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd3b3ab323c6d5153054aa04af2d91f7b9e8436ff1e2f75f3594b57831b2a416\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://67c46f5f0a3ace95bd2e9479c5561126664e18b148c5eceea5bf5efc27e81a67\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c52e076eddfd307110b1f6de00a169b81af8f2785af7becef2a820ce363a5861\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-05T06:48:14Z\\\",\\\"message\\\":\\\"le observer\\\\nW1005 06:48:14.337870 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1005 06:48:14.338086 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1005 06:48:14.339389 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1467752215/tls.crt::/tmp/serving-cert-1467752215/tls.key\\\\\\\"\\\\nI1005 06:48:14.864345 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1005 06:48:14.868087 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1005 06:48:14.868114 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1005 06:48:14.868140 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1005 06:48:14.868147 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1005 06:48:14.877354 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1005 06:48:14.877412 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1005 06:48:14.877456 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1005 06:48:14.877471 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1005 06:48:14.877490 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1005 06:48:14.877501 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1005 06:48:14.877507 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1005 06:48:14.877515 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1005 06:48:14.880326 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:13Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4a091716ff935adb42ae2e88048619fd347bd174f64e5f9c97a01f81690e57f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:01Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dc5411b642f99b13d514d2f21defad0c394aeeb82a966f72300076617e2586c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5dc5411b642f99b13d514d2f21defad0c394aeeb82a966f72300076617e2586c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:47:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:47:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:47:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:47Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:48 crc kubenswrapper[4846]: I1005 06:48:48.021408 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:48Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:48 crc kubenswrapper[4846]: I1005 06:48:48.047245 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-txzlk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"93a6ace4-8d64-44d7-9637-457d3af6543c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ad1940d07e37217089642478ece9ccdaca74784738a400929a9f5359d158c01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a6f6b2c32dd3da5f9fd5e6971c1773e81a9a24184b7aa5b543cec879001ae53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e9b0219ecbc7e60905d6de3e67a2c8dc9af1017b4aed9538ee0c58c861f2438\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af5e5a4da9ad7371d5a24633e8268f12e3a7d419a8c93b11ddbcc006d8339c6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8da85222775abc7a772f6a8ff8255f1a0d5b15807ba177a7e3e2b2d01a4f24b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1409651de8c897804f87be03d016c138f8acf27e662dce265a93e15c1848adcf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://93149bae0504218735e0900b660ad9759ebbec3eb4dc214cc40ad349ac002d45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c212f5d230fae82a701afeceb39a848a9c91a3354908d88b062b7e1a726240cb\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-05T06:48:34Z\\\",\\\"message\\\":\\\"g reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1005 06:48:34.496108 6146 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1005 06:48:34.496160 6146 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1005 06:48:34.496168 6146 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1005 06:48:34.496246 6146 factory.go:656] Stopping watch factory\\\\nI1005 06:48:34.496262 6146 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1005 06:48:34.496271 6146 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1005 06:48:34.496279 6146 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1005 06:48:34.496293 6146 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1005 06:48:34.496400 6146 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1005 06:48:34.496553 6146 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1005 06:48:34.496742 6146 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:31Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://93149bae0504218735e0900b660ad9759ebbec3eb4dc214cc40ad349ac002d45\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-05T06:48:36Z\\\",\\\"message\\\":\\\"ePorts:nil,LoadBalancerClass:nil,InternalTrafficPolicy:*Cluster,TrafficDistribution:nil,},Status:ServiceStatus{LoadBalancer:LoadBalancerStatus{Ingress:[]LoadBalancerIngress{},},Conditions:[]Condition{},},}\\\\nI1005 06:48:35.889902 6320 lb_config.go:1031] Cluster endpoints for openshift-kube-apiserver/apiserver for network=default are: map[]\\\\nI1005 06:48:35.889910 6320 services_controller.go:443] Built service openshift-kube-apiserver/apiserver LB cluster-wide configs for network=default: []services.lbConfig{services.lbConfig{vips:[]string{\\\\\\\"10.217.4.93\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:443, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI1005 06:48:35.889925 6320 services_controller.go:444] Built service openshift-kube-apiserver/apiserver LB per-node configs for network=default: []services.lbConfig(nil)\\\\nF1005 06:48:35.889931 6320 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: failed to add event handler: handler {0x1e60340 0x1e60020 0x1e5ffc0} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to se\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0df0ebcd87dc5af4b17736b43ec7b2395968b19ed16e775366af14e2b2d49c52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff3c6545ff3d2436da27bbd223b950b30b0c69a3e7f3034cb0e61380cb38d338\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ff3c6545ff3d2436da27bbd223b950b30b0c69a3e7f3034cb0e61380cb38d338\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:24Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-txzlk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:48Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:48 crc kubenswrapper[4846]: I1005 06:48:48.066406 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-gklb6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"698ab882-bb26-4b4c-81e9-8305a44fe068\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a6edca88ea244a58f8cf6cc431110a960c6bfd892812cd0d00fa3f97be3bc67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x6d6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c191538ab797b8056820de7da1f8e93ea18873891a32382b8a59c9af762e103\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x6d6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:37Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-gklb6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:48Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:48 crc kubenswrapper[4846]: I1005 06:48:48.075413 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:48 crc kubenswrapper[4846]: I1005 06:48:48.075689 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:48 crc kubenswrapper[4846]: I1005 06:48:48.075854 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:48 crc kubenswrapper[4846]: I1005 06:48:48.076017 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:48 crc kubenswrapper[4846]: I1005 06:48:48.076158 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:48Z","lastTransitionTime":"2025-10-05T06:48:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:48 crc kubenswrapper[4846]: I1005 06:48:48.100024 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c482bb97-fd50-40cf-96ea-f806c3f01125\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bfe45d30e4fa141541aca1d46bc5cc1da0a7b4a5eb1fc9bb2caa5b7fdcb0b264\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://496a4782faa30a9a20ff63038141600cd635f12902d1cf17989cd067fcdfe929\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://281be077869a6aa1ca815770871a14413d4a97e97aeb550fd5cd215cba88b21f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6771c7215d2f5dd5838d17888375a4768b9d1755fc2e21bda1bbfbb2eb1dbb95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c17a711c2c7473b076a3944ffe7c700808d05cba582fbd55c5c2ac8ff73838a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44a9f740f0a06c8bf928e2a9d1d5979217558c31e73a333631baccebf00fd5d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://44a9f740f0a06c8bf928e2a9d1d5979217558c31e73a333631baccebf00fd5d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:47:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:47:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://48ca0ba6ca6092bfcb5b153220c054f02cc36fb22707c42e021f06d329e18e2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48ca0ba6ca6092bfcb5b153220c054f02cc36fb22707c42e021f06d329e18e2e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:00Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4006d0cea128675c8885d879171204abffc5e11ea0664750a12b4bc7344ccb2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4006d0cea128675c8885d879171204abffc5e11ea0664750a12b4bc7344ccb2e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:47:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:48Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:48 crc kubenswrapper[4846]: I1005 06:48:48.123391 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:48Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:48 crc kubenswrapper[4846]: I1005 06:48:48.148509 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-sfmxd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1dbb13ac-136c-4a1c-9229-b4e8ac4f9c71\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b0af6a6e735f7b7877c177c1d53d1a8dc12c4713a8059c4766fc858af31c9a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://226d977b596d07aeda0486452e2a8605949ed4d2453b21f56174937c395d6f36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://226d977b596d07aeda0486452e2a8605949ed4d2453b21f56174937c395d6f36\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://99c07a4bdb3059d3a99ffdaf794d668d52cffd9e5e3bb8945204dd6d1f8c8328\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99c07a4bdb3059d3a99ffdaf794d668d52cffd9e5e3bb8945204dd6d1f8c8328\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27095272bf4f31d63a78a436636de0aae12f7103597da944f24ed7a74329ee3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27095272bf4f31d63a78a436636de0aae12f7103597da944f24ed7a74329ee3f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1211571af66d8a94302b65feccf30768ba82b7cbad918e44dd405787dce484cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1211571af66d8a94302b65feccf30768ba82b7cbad918e44dd405787dce484cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27631f90d528900fa180c15cfce80e96d4cdba07743eb6d847e2e822d9917218\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27631f90d528900fa180c15cfce80e96d4cdba07743eb6d847e2e822d9917218\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f345693e9aeb1f4dc196349d3354f3c833a2b0415a693298b1cfccfa2b0d4459\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f345693e9aeb1f4dc196349d3354f3c833a2b0415a693298b1cfccfa2b0d4459\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-sfmxd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:48Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:48 crc kubenswrapper[4846]: I1005 06:48:48.162591 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-w2sj8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32931fce-7374-4cab-a2e0-79d96a842274\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0df535d65d39de6c9a3dfb931bafc4469adc22b72f7a7903beda4fef8cd48138\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9rtxf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:27Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-w2sj8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:48Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:48 crc kubenswrapper[4846]: I1005 06:48:48.177248 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c9f812b-ecc8-4985-a670-53fa80fbb499\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2b5ee9781c01d6aec3ac778efef1067b6c3bfe4ed37431bdebe188fcde0165f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:47:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3fb4717692aa1fec0012b514ef8d18d1ca3707cfa4bb8248bb93841a0a40f11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:47:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0b6a7afe3094f1c92093c05c28344897a615d590c2fbb507c87e18a44c6c831\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://640ade2c733d7916ec213a201a1dd614d261c69eab654c80100efbc99a768995\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:47:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:48Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:48 crc kubenswrapper[4846]: I1005 06:48:48.179100 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:48 crc kubenswrapper[4846]: I1005 06:48:48.179381 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:48 crc kubenswrapper[4846]: I1005 06:48:48.179462 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:48 crc kubenswrapper[4846]: I1005 06:48:48.179569 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:48 crc kubenswrapper[4846]: I1005 06:48:48.179590 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:48Z","lastTransitionTime":"2025-10-05T06:48:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:48 crc kubenswrapper[4846]: I1005 06:48:48.193444 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:48Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:48 crc kubenswrapper[4846]: I1005 06:48:48.209077 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e67d0e447a8a6a381259e25d25badc791194e1291f56b1cd77bd63989b20408\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dkhl7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://516f88f4990f5962d8b694a3d889259cc5508a309919870191c2e1331640c973\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dkhl7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-fscvf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:48Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:48 crc kubenswrapper[4846]: I1005 06:48:48.282920 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:48 crc kubenswrapper[4846]: I1005 06:48:48.283527 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:48 crc kubenswrapper[4846]: I1005 06:48:48.283636 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:48 crc kubenswrapper[4846]: I1005 06:48:48.283729 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:48 crc kubenswrapper[4846]: I1005 06:48:48.283822 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:48Z","lastTransitionTime":"2025-10-05T06:48:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:48 crc kubenswrapper[4846]: I1005 06:48:48.387291 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:48 crc kubenswrapper[4846]: I1005 06:48:48.387727 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:48 crc kubenswrapper[4846]: I1005 06:48:48.387831 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:48 crc kubenswrapper[4846]: I1005 06:48:48.387940 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:48 crc kubenswrapper[4846]: I1005 06:48:48.388068 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:48Z","lastTransitionTime":"2025-10-05T06:48:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:48 crc kubenswrapper[4846]: I1005 06:48:48.491365 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:48 crc kubenswrapper[4846]: I1005 06:48:48.491432 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:48 crc kubenswrapper[4846]: I1005 06:48:48.491449 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:48 crc kubenswrapper[4846]: I1005 06:48:48.491473 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:48 crc kubenswrapper[4846]: I1005 06:48:48.491491 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:48Z","lastTransitionTime":"2025-10-05T06:48:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:48 crc kubenswrapper[4846]: I1005 06:48:48.496739 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 05 06:48:48 crc kubenswrapper[4846]: E1005 06:48:48.496976 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 05 06:48:48 crc kubenswrapper[4846]: I1005 06:48:48.496759 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9r2kt" Oct 05 06:48:48 crc kubenswrapper[4846]: E1005 06:48:48.497234 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9r2kt" podUID="59b83f4f-71f0-4b28-9275-56d5f933ec2c" Oct 05 06:48:48 crc kubenswrapper[4846]: I1005 06:48:48.521372 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-txzlk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"93a6ace4-8d64-44d7-9637-457d3af6543c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ad1940d07e37217089642478ece9ccdaca74784738a400929a9f5359d158c01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a6f6b2c32dd3da5f9fd5e6971c1773e81a9a24184b7aa5b543cec879001ae53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e9b0219ecbc7e60905d6de3e67a2c8dc9af1017b4aed9538ee0c58c861f2438\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af5e5a4da9ad7371d5a24633e8268f12e3a7d419a8c93b11ddbcc006d8339c6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8da85222775abc7a772f6a8ff8255f1a0d5b15807ba177a7e3e2b2d01a4f24b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1409651de8c897804f87be03d016c138f8acf27e662dce265a93e15c1848adcf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://93149bae0504218735e0900b660ad9759ebbec3eb4dc214cc40ad349ac002d45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c212f5d230fae82a701afeceb39a848a9c91a3354908d88b062b7e1a726240cb\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-05T06:48:34Z\\\",\\\"message\\\":\\\"g reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1005 06:48:34.496108 6146 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1005 06:48:34.496160 6146 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1005 06:48:34.496168 6146 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1005 06:48:34.496246 6146 factory.go:656] Stopping watch factory\\\\nI1005 06:48:34.496262 6146 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1005 06:48:34.496271 6146 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1005 06:48:34.496279 6146 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1005 06:48:34.496293 6146 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1005 06:48:34.496400 6146 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1005 06:48:34.496553 6146 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1005 06:48:34.496742 6146 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:31Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://93149bae0504218735e0900b660ad9759ebbec3eb4dc214cc40ad349ac002d45\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-05T06:48:36Z\\\",\\\"message\\\":\\\"ePorts:nil,LoadBalancerClass:nil,InternalTrafficPolicy:*Cluster,TrafficDistribution:nil,},Status:ServiceStatus{LoadBalancer:LoadBalancerStatus{Ingress:[]LoadBalancerIngress{},},Conditions:[]Condition{},},}\\\\nI1005 06:48:35.889902 6320 lb_config.go:1031] Cluster endpoints for openshift-kube-apiserver/apiserver for network=default are: map[]\\\\nI1005 06:48:35.889910 6320 services_controller.go:443] Built service openshift-kube-apiserver/apiserver LB cluster-wide configs for network=default: []services.lbConfig{services.lbConfig{vips:[]string{\\\\\\\"10.217.4.93\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:443, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI1005 06:48:35.889925 6320 services_controller.go:444] Built service openshift-kube-apiserver/apiserver LB per-node configs for network=default: []services.lbConfig(nil)\\\\nF1005 06:48:35.889931 6320 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: failed to add event handler: handler {0x1e60340 0x1e60020 0x1e5ffc0} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to se\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0df0ebcd87dc5af4b17736b43ec7b2395968b19ed16e775366af14e2b2d49c52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff3c6545ff3d2436da27bbd223b950b30b0c69a3e7f3034cb0e61380cb38d338\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ff3c6545ff3d2436da27bbd223b950b30b0c69a3e7f3034cb0e61380cb38d338\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:24Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-txzlk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:48Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:48 crc kubenswrapper[4846]: I1005 06:48:48.538309 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-gklb6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"698ab882-bb26-4b4c-81e9-8305a44fe068\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a6edca88ea244a58f8cf6cc431110a960c6bfd892812cd0d00fa3f97be3bc67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x6d6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c191538ab797b8056820de7da1f8e93ea18873891a32382b8a59c9af762e103\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x6d6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:37Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-gklb6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:48Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:48 crc kubenswrapper[4846]: I1005 06:48:48.573778 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c482bb97-fd50-40cf-96ea-f806c3f01125\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bfe45d30e4fa141541aca1d46bc5cc1da0a7b4a5eb1fc9bb2caa5b7fdcb0b264\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://496a4782faa30a9a20ff63038141600cd635f12902d1cf17989cd067fcdfe929\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://281be077869a6aa1ca815770871a14413d4a97e97aeb550fd5cd215cba88b21f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6771c7215d2f5dd5838d17888375a4768b9d1755fc2e21bda1bbfbb2eb1dbb95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c17a711c2c7473b076a3944ffe7c700808d05cba582fbd55c5c2ac8ff73838a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44a9f740f0a06c8bf928e2a9d1d5979217558c31e73a333631baccebf00fd5d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://44a9f740f0a06c8bf928e2a9d1d5979217558c31e73a333631baccebf00fd5d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:47:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:47:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://48ca0ba6ca6092bfcb5b153220c054f02cc36fb22707c42e021f06d329e18e2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48ca0ba6ca6092bfcb5b153220c054f02cc36fb22707c42e021f06d329e18e2e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:00Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4006d0cea128675c8885d879171204abffc5e11ea0664750a12b4bc7344ccb2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4006d0cea128675c8885d879171204abffc5e11ea0664750a12b4bc7344ccb2e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:47:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:48Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:48 crc kubenswrapper[4846]: I1005 06:48:48.594854 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:48 crc kubenswrapper[4846]: I1005 06:48:48.594933 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:48 crc kubenswrapper[4846]: I1005 06:48:48.594945 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:48 crc kubenswrapper[4846]: I1005 06:48:48.594966 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:48 crc kubenswrapper[4846]: I1005 06:48:48.594979 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:48Z","lastTransitionTime":"2025-10-05T06:48:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:48 crc kubenswrapper[4846]: I1005 06:48:48.596524 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:48Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:48 crc kubenswrapper[4846]: I1005 06:48:48.625644 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-sfmxd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1dbb13ac-136c-4a1c-9229-b4e8ac4f9c71\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b0af6a6e735f7b7877c177c1d53d1a8dc12c4713a8059c4766fc858af31c9a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://226d977b596d07aeda0486452e2a8605949ed4d2453b21f56174937c395d6f36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://226d977b596d07aeda0486452e2a8605949ed4d2453b21f56174937c395d6f36\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://99c07a4bdb3059d3a99ffdaf794d668d52cffd9e5e3bb8945204dd6d1f8c8328\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99c07a4bdb3059d3a99ffdaf794d668d52cffd9e5e3bb8945204dd6d1f8c8328\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27095272bf4f31d63a78a436636de0aae12f7103597da944f24ed7a74329ee3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27095272bf4f31d63a78a436636de0aae12f7103597da944f24ed7a74329ee3f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1211571af66d8a94302b65feccf30768ba82b7cbad918e44dd405787dce484cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1211571af66d8a94302b65feccf30768ba82b7cbad918e44dd405787dce484cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27631f90d528900fa180c15cfce80e96d4cdba07743eb6d847e2e822d9917218\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27631f90d528900fa180c15cfce80e96d4cdba07743eb6d847e2e822d9917218\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f345693e9aeb1f4dc196349d3354f3c833a2b0415a693298b1cfccfa2b0d4459\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f345693e9aeb1f4dc196349d3354f3c833a2b0415a693298b1cfccfa2b0d4459\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-sfmxd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:48Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:48 crc kubenswrapper[4846]: I1005 06:48:48.649928 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-w2sj8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32931fce-7374-4cab-a2e0-79d96a842274\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0df535d65d39de6c9a3dfb931bafc4469adc22b72f7a7903beda4fef8cd48138\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9rtxf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:27Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-w2sj8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:48Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:48 crc kubenswrapper[4846]: I1005 06:48:48.674513 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c9f812b-ecc8-4985-a670-53fa80fbb499\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2b5ee9781c01d6aec3ac778efef1067b6c3bfe4ed37431bdebe188fcde0165f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:47:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3fb4717692aa1fec0012b514ef8d18d1ca3707cfa4bb8248bb93841a0a40f11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:47:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0b6a7afe3094f1c92093c05c28344897a615d590c2fbb507c87e18a44c6c831\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://640ade2c733d7916ec213a201a1dd614d261c69eab654c80100efbc99a768995\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:47:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:48Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:48 crc kubenswrapper[4846]: I1005 06:48:48.696970 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:48Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:48 crc kubenswrapper[4846]: I1005 06:48:48.697778 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:48 crc kubenswrapper[4846]: I1005 06:48:48.697823 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:48 crc kubenswrapper[4846]: I1005 06:48:48.697834 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:48 crc kubenswrapper[4846]: I1005 06:48:48.697857 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:48 crc kubenswrapper[4846]: I1005 06:48:48.697869 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:48Z","lastTransitionTime":"2025-10-05T06:48:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:48 crc kubenswrapper[4846]: I1005 06:48:48.716368 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e67d0e447a8a6a381259e25d25badc791194e1291f56b1cd77bd63989b20408\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dkhl7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://516f88f4990f5962d8b694a3d889259cc5508a309919870191c2e1331640c973\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dkhl7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-fscvf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:48Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:48 crc kubenswrapper[4846]: I1005 06:48:48.735544 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56f1fe1f9f06f0249cf937a6963fda82467732dc056ca2b2189be6d8dd9d4a05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7212ce5f42c1fb55146abb190a557a504b43f16af6df84f90e1073567a329ec5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:48Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:48 crc kubenswrapper[4846]: I1005 06:48:48.752964 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-7wm5q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9bc7744f-d700-450f-8b8d-98140877fee9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90379e74c544fccca44f7acdc228ff378154390e211ac801b79f8d074012241e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h5p4q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-7wm5q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:48Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:48 crc kubenswrapper[4846]: I1005 06:48:48.773520 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ff10c31459b3b3724ee7b769bd6f72f9ed80cce5419b6adcdf69a3239b70756\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:48Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:48 crc kubenswrapper[4846]: I1005 06:48:48.789545 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://347f900743b7bba3daf92f21c7eba0625e8c774f50057bc9472f0c82119325e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:48Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:48 crc kubenswrapper[4846]: I1005 06:48:48.801573 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:48 crc kubenswrapper[4846]: I1005 06:48:48.801641 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:48 crc kubenswrapper[4846]: I1005 06:48:48.801661 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:48 crc kubenswrapper[4846]: I1005 06:48:48.801689 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:48 crc kubenswrapper[4846]: I1005 06:48:48.801709 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:48Z","lastTransitionTime":"2025-10-05T06:48:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:48 crc kubenswrapper[4846]: I1005 06:48:48.804326 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-r7ztb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a5119cd9-af70-4324-a955-978306ab0b20\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06231a87429ec292cb57ebd88cda7ad547bcb32e80343b776833621a11c28ad1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jfcj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-r7ztb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:48Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:48 crc kubenswrapper[4846]: I1005 06:48:48.821772 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-9r2kt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"59b83f4f-71f0-4b28-9275-56d5f933ec2c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tdwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tdwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:38Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-9r2kt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:48Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:48 crc kubenswrapper[4846]: I1005 06:48:48.842300 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"775b5702-a131-4d6c-852d-91bfa056cac5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0839c1765df2c68d2554e9c66d289505e559a9672caf8069feeff2c4fdfd2171\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15dbaf2445704da0987f256f81dff2ee95664988541fa48ce94898afeadab2e8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd3b3ab323c6d5153054aa04af2d91f7b9e8436ff1e2f75f3594b57831b2a416\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://67c46f5f0a3ace95bd2e9479c5561126664e18b148c5eceea5bf5efc27e81a67\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c52e076eddfd307110b1f6de00a169b81af8f2785af7becef2a820ce363a5861\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-05T06:48:14Z\\\",\\\"message\\\":\\\"le observer\\\\nW1005 06:48:14.337870 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1005 06:48:14.338086 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1005 06:48:14.339389 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1467752215/tls.crt::/tmp/serving-cert-1467752215/tls.key\\\\\\\"\\\\nI1005 06:48:14.864345 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1005 06:48:14.868087 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1005 06:48:14.868114 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1005 06:48:14.868140 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1005 06:48:14.868147 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1005 06:48:14.877354 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1005 06:48:14.877412 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1005 06:48:14.877456 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1005 06:48:14.877471 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1005 06:48:14.877490 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1005 06:48:14.877501 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1005 06:48:14.877507 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1005 06:48:14.877515 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1005 06:48:14.880326 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:13Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4a091716ff935adb42ae2e88048619fd347bd174f64e5f9c97a01f81690e57f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:01Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dc5411b642f99b13d514d2f21defad0c394aeeb82a966f72300076617e2586c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5dc5411b642f99b13d514d2f21defad0c394aeeb82a966f72300076617e2586c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:47:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:47:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:47:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:48Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:48 crc kubenswrapper[4846]: I1005 06:48:48.858066 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:48Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:48 crc kubenswrapper[4846]: I1005 06:48:48.905164 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:48 crc kubenswrapper[4846]: I1005 06:48:48.905240 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:48 crc kubenswrapper[4846]: I1005 06:48:48.905252 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:48 crc kubenswrapper[4846]: I1005 06:48:48.905268 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:48 crc kubenswrapper[4846]: I1005 06:48:48.905278 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:48Z","lastTransitionTime":"2025-10-05T06:48:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:49 crc kubenswrapper[4846]: I1005 06:48:49.008658 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:49 crc kubenswrapper[4846]: I1005 06:48:49.008724 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:49 crc kubenswrapper[4846]: I1005 06:48:49.008735 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:49 crc kubenswrapper[4846]: I1005 06:48:49.008754 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:49 crc kubenswrapper[4846]: I1005 06:48:49.008770 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:49Z","lastTransitionTime":"2025-10-05T06:48:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:49 crc kubenswrapper[4846]: I1005 06:48:49.111822 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:49 crc kubenswrapper[4846]: I1005 06:48:49.111881 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:49 crc kubenswrapper[4846]: I1005 06:48:49.111893 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:49 crc kubenswrapper[4846]: I1005 06:48:49.111914 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:49 crc kubenswrapper[4846]: I1005 06:48:49.111927 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:49Z","lastTransitionTime":"2025-10-05T06:48:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:49 crc kubenswrapper[4846]: I1005 06:48:49.215514 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:49 crc kubenswrapper[4846]: I1005 06:48:49.215557 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:49 crc kubenswrapper[4846]: I1005 06:48:49.215568 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:49 crc kubenswrapper[4846]: I1005 06:48:49.215588 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:49 crc kubenswrapper[4846]: I1005 06:48:49.215602 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:49Z","lastTransitionTime":"2025-10-05T06:48:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:49 crc kubenswrapper[4846]: I1005 06:48:49.318552 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:49 crc kubenswrapper[4846]: I1005 06:48:49.318641 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:49 crc kubenswrapper[4846]: I1005 06:48:49.318668 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:49 crc kubenswrapper[4846]: I1005 06:48:49.318700 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:49 crc kubenswrapper[4846]: I1005 06:48:49.318727 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:49Z","lastTransitionTime":"2025-10-05T06:48:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:49 crc kubenswrapper[4846]: I1005 06:48:49.422315 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:49 crc kubenswrapper[4846]: I1005 06:48:49.422383 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:49 crc kubenswrapper[4846]: I1005 06:48:49.422400 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:49 crc kubenswrapper[4846]: I1005 06:48:49.422426 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:49 crc kubenswrapper[4846]: I1005 06:48:49.422445 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:49Z","lastTransitionTime":"2025-10-05T06:48:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:49 crc kubenswrapper[4846]: I1005 06:48:49.497221 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 05 06:48:49 crc kubenswrapper[4846]: I1005 06:48:49.497771 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 05 06:48:49 crc kubenswrapper[4846]: E1005 06:48:49.497903 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 05 06:48:49 crc kubenswrapper[4846]: I1005 06:48:49.498616 4846 scope.go:117] "RemoveContainer" containerID="93149bae0504218735e0900b660ad9759ebbec3eb4dc214cc40ad349ac002d45" Oct 05 06:48:49 crc kubenswrapper[4846]: E1005 06:48:49.501363 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 05 06:48:49 crc kubenswrapper[4846]: I1005 06:48:49.504259 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 05 06:48:49 crc kubenswrapper[4846]: I1005 06:48:49.504426 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 05 06:48:49 crc kubenswrapper[4846]: E1005 06:48:49.504457 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-05 06:49:21.504421673 +0000 UTC m=+83.745274488 (durationBeforeRetry 32s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 06:48:49 crc kubenswrapper[4846]: I1005 06:48:49.504516 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 05 06:48:49 crc kubenswrapper[4846]: E1005 06:48:49.504586 4846 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 05 06:48:49 crc kubenswrapper[4846]: I1005 06:48:49.504633 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 05 06:48:49 crc kubenswrapper[4846]: E1005 06:48:49.504654 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-05 06:49:21.504633089 +0000 UTC m=+83.745485904 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 05 06:48:49 crc kubenswrapper[4846]: I1005 06:48:49.504679 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 05 06:48:49 crc kubenswrapper[4846]: E1005 06:48:49.504802 4846 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 05 06:48:49 crc kubenswrapper[4846]: E1005 06:48:49.504871 4846 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 05 06:48:49 crc kubenswrapper[4846]: E1005 06:48:49.504875 4846 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 05 06:48:49 crc kubenswrapper[4846]: E1005 06:48:49.504894 4846 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 05 06:48:49 crc kubenswrapper[4846]: E1005 06:48:49.504922 4846 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 05 06:48:49 crc kubenswrapper[4846]: E1005 06:48:49.504948 4846 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 05 06:48:49 crc kubenswrapper[4846]: E1005 06:48:49.504926 4846 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 05 06:48:49 crc kubenswrapper[4846]: E1005 06:48:49.504885 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-05 06:49:21.504859796 +0000 UTC m=+83.745712791 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 05 06:48:49 crc kubenswrapper[4846]: E1005 06:48:49.505314 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-05 06:49:21.505290228 +0000 UTC m=+83.746143183 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 05 06:48:49 crc kubenswrapper[4846]: E1005 06:48:49.505346 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-05 06:49:21.505330139 +0000 UTC m=+83.746183154 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 05 06:48:49 crc kubenswrapper[4846]: I1005 06:48:49.520473 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56f1fe1f9f06f0249cf937a6963fda82467732dc056ca2b2189be6d8dd9d4a05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7212ce5f42c1fb55146abb190a557a504b43f16af6df84f90e1073567a329ec5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:49Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:49 crc kubenswrapper[4846]: I1005 06:48:49.526198 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:49 crc kubenswrapper[4846]: I1005 06:48:49.526262 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:49 crc kubenswrapper[4846]: I1005 06:48:49.526283 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:49 crc kubenswrapper[4846]: I1005 06:48:49.526309 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:49 crc kubenswrapper[4846]: I1005 06:48:49.526327 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:49Z","lastTransitionTime":"2025-10-05T06:48:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:49 crc kubenswrapper[4846]: I1005 06:48:49.542234 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-7wm5q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9bc7744f-d700-450f-8b8d-98140877fee9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90379e74c544fccca44f7acdc228ff378154390e211ac801b79f8d074012241e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h5p4q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-7wm5q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:49Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:49 crc kubenswrapper[4846]: I1005 06:48:49.566584 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ff10c31459b3b3724ee7b769bd6f72f9ed80cce5419b6adcdf69a3239b70756\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:49Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:49 crc kubenswrapper[4846]: I1005 06:48:49.588053 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://347f900743b7bba3daf92f21c7eba0625e8c774f50057bc9472f0c82119325e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:49Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:49 crc kubenswrapper[4846]: I1005 06:48:49.606089 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-r7ztb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a5119cd9-af70-4324-a955-978306ab0b20\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06231a87429ec292cb57ebd88cda7ad547bcb32e80343b776833621a11c28ad1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jfcj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-r7ztb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:49Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:49 crc kubenswrapper[4846]: I1005 06:48:49.629209 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-9r2kt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"59b83f4f-71f0-4b28-9275-56d5f933ec2c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tdwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tdwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:38Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-9r2kt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:49Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:49 crc kubenswrapper[4846]: I1005 06:48:49.639046 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:49 crc kubenswrapper[4846]: I1005 06:48:49.639103 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:49 crc kubenswrapper[4846]: I1005 06:48:49.639121 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:49 crc kubenswrapper[4846]: I1005 06:48:49.639145 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:49 crc kubenswrapper[4846]: I1005 06:48:49.639162 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:49Z","lastTransitionTime":"2025-10-05T06:48:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:49 crc kubenswrapper[4846]: I1005 06:48:49.655509 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"775b5702-a131-4d6c-852d-91bfa056cac5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0839c1765df2c68d2554e9c66d289505e559a9672caf8069feeff2c4fdfd2171\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15dbaf2445704da0987f256f81dff2ee95664988541fa48ce94898afeadab2e8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd3b3ab323c6d5153054aa04af2d91f7b9e8436ff1e2f75f3594b57831b2a416\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://67c46f5f0a3ace95bd2e9479c5561126664e18b148c5eceea5bf5efc27e81a67\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c52e076eddfd307110b1f6de00a169b81af8f2785af7becef2a820ce363a5861\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-05T06:48:14Z\\\",\\\"message\\\":\\\"le observer\\\\nW1005 06:48:14.337870 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1005 06:48:14.338086 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1005 06:48:14.339389 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1467752215/tls.crt::/tmp/serving-cert-1467752215/tls.key\\\\\\\"\\\\nI1005 06:48:14.864345 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1005 06:48:14.868087 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1005 06:48:14.868114 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1005 06:48:14.868140 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1005 06:48:14.868147 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1005 06:48:14.877354 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1005 06:48:14.877412 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1005 06:48:14.877456 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1005 06:48:14.877471 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1005 06:48:14.877490 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1005 06:48:14.877501 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1005 06:48:14.877507 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1005 06:48:14.877515 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1005 06:48:14.880326 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:13Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4a091716ff935adb42ae2e88048619fd347bd174f64e5f9c97a01f81690e57f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:01Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dc5411b642f99b13d514d2f21defad0c394aeeb82a966f72300076617e2586c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5dc5411b642f99b13d514d2f21defad0c394aeeb82a966f72300076617e2586c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:47:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:47:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:47:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:49Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:49 crc kubenswrapper[4846]: I1005 06:48:49.675527 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:49Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:49 crc kubenswrapper[4846]: I1005 06:48:49.709799 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-txzlk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"93a6ace4-8d64-44d7-9637-457d3af6543c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ad1940d07e37217089642478ece9ccdaca74784738a400929a9f5359d158c01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a6f6b2c32dd3da5f9fd5e6971c1773e81a9a24184b7aa5b543cec879001ae53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e9b0219ecbc7e60905d6de3e67a2c8dc9af1017b4aed9538ee0c58c861f2438\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af5e5a4da9ad7371d5a24633e8268f12e3a7d419a8c93b11ddbcc006d8339c6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8da85222775abc7a772f6a8ff8255f1a0d5b15807ba177a7e3e2b2d01a4f24b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1409651de8c897804f87be03d016c138f8acf27e662dce265a93e15c1848adcf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://93149bae0504218735e0900b660ad9759ebbec3eb4dc214cc40ad349ac002d45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://93149bae0504218735e0900b660ad9759ebbec3eb4dc214cc40ad349ac002d45\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-05T06:48:36Z\\\",\\\"message\\\":\\\"ePorts:nil,LoadBalancerClass:nil,InternalTrafficPolicy:*Cluster,TrafficDistribution:nil,},Status:ServiceStatus{LoadBalancer:LoadBalancerStatus{Ingress:[]LoadBalancerIngress{},},Conditions:[]Condition{},},}\\\\nI1005 06:48:35.889902 6320 lb_config.go:1031] Cluster endpoints for openshift-kube-apiserver/apiserver for network=default are: map[]\\\\nI1005 06:48:35.889910 6320 services_controller.go:443] Built service openshift-kube-apiserver/apiserver LB cluster-wide configs for network=default: []services.lbConfig{services.lbConfig{vips:[]string{\\\\\\\"10.217.4.93\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:443, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI1005 06:48:35.889925 6320 services_controller.go:444] Built service openshift-kube-apiserver/apiserver LB per-node configs for network=default: []services.lbConfig(nil)\\\\nF1005 06:48:35.889931 6320 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: failed to add event handler: handler {0x1e60340 0x1e60020 0x1e5ffc0} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to se\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:35Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-txzlk_openshift-ovn-kubernetes(93a6ace4-8d64-44d7-9637-457d3af6543c)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0df0ebcd87dc5af4b17736b43ec7b2395968b19ed16e775366af14e2b2d49c52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff3c6545ff3d2436da27bbd223b950b30b0c69a3e7f3034cb0e61380cb38d338\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ff3c6545ff3d2436da27bbd223b950b30b0c69a3e7f3034cb0e61380cb38d338\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:24Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-txzlk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:49Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:49 crc kubenswrapper[4846]: I1005 06:48:49.730962 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-gklb6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"698ab882-bb26-4b4c-81e9-8305a44fe068\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a6edca88ea244a58f8cf6cc431110a960c6bfd892812cd0d00fa3f97be3bc67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x6d6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c191538ab797b8056820de7da1f8e93ea18873891a32382b8a59c9af762e103\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x6d6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:37Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-gklb6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:49Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:49 crc kubenswrapper[4846]: I1005 06:48:49.742904 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:49 crc kubenswrapper[4846]: I1005 06:48:49.742977 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:49 crc kubenswrapper[4846]: I1005 06:48:49.743002 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:49 crc kubenswrapper[4846]: I1005 06:48:49.743039 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:49 crc kubenswrapper[4846]: I1005 06:48:49.743062 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:49Z","lastTransitionTime":"2025-10-05T06:48:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:49 crc kubenswrapper[4846]: I1005 06:48:49.770226 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c482bb97-fd50-40cf-96ea-f806c3f01125\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bfe45d30e4fa141541aca1d46bc5cc1da0a7b4a5eb1fc9bb2caa5b7fdcb0b264\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://496a4782faa30a9a20ff63038141600cd635f12902d1cf17989cd067fcdfe929\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://281be077869a6aa1ca815770871a14413d4a97e97aeb550fd5cd215cba88b21f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6771c7215d2f5dd5838d17888375a4768b9d1755fc2e21bda1bbfbb2eb1dbb95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c17a711c2c7473b076a3944ffe7c700808d05cba582fbd55c5c2ac8ff73838a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44a9f740f0a06c8bf928e2a9d1d5979217558c31e73a333631baccebf00fd5d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://44a9f740f0a06c8bf928e2a9d1d5979217558c31e73a333631baccebf00fd5d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:47:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:47:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://48ca0ba6ca6092bfcb5b153220c054f02cc36fb22707c42e021f06d329e18e2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48ca0ba6ca6092bfcb5b153220c054f02cc36fb22707c42e021f06d329e18e2e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:00Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4006d0cea128675c8885d879171204abffc5e11ea0664750a12b4bc7344ccb2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4006d0cea128675c8885d879171204abffc5e11ea0664750a12b4bc7344ccb2e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:47:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:49Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:49 crc kubenswrapper[4846]: I1005 06:48:49.792076 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:49Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:49 crc kubenswrapper[4846]: I1005 06:48:49.821111 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-sfmxd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1dbb13ac-136c-4a1c-9229-b4e8ac4f9c71\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b0af6a6e735f7b7877c177c1d53d1a8dc12c4713a8059c4766fc858af31c9a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://226d977b596d07aeda0486452e2a8605949ed4d2453b21f56174937c395d6f36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://226d977b596d07aeda0486452e2a8605949ed4d2453b21f56174937c395d6f36\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://99c07a4bdb3059d3a99ffdaf794d668d52cffd9e5e3bb8945204dd6d1f8c8328\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99c07a4bdb3059d3a99ffdaf794d668d52cffd9e5e3bb8945204dd6d1f8c8328\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27095272bf4f31d63a78a436636de0aae12f7103597da944f24ed7a74329ee3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27095272bf4f31d63a78a436636de0aae12f7103597da944f24ed7a74329ee3f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1211571af66d8a94302b65feccf30768ba82b7cbad918e44dd405787dce484cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1211571af66d8a94302b65feccf30768ba82b7cbad918e44dd405787dce484cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27631f90d528900fa180c15cfce80e96d4cdba07743eb6d847e2e822d9917218\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27631f90d528900fa180c15cfce80e96d4cdba07743eb6d847e2e822d9917218\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f345693e9aeb1f4dc196349d3354f3c833a2b0415a693298b1cfccfa2b0d4459\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f345693e9aeb1f4dc196349d3354f3c833a2b0415a693298b1cfccfa2b0d4459\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-sfmxd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:49Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:49 crc kubenswrapper[4846]: I1005 06:48:49.841584 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-w2sj8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32931fce-7374-4cab-a2e0-79d96a842274\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0df535d65d39de6c9a3dfb931bafc4469adc22b72f7a7903beda4fef8cd48138\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9rtxf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:27Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-w2sj8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:49Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:49 crc kubenswrapper[4846]: I1005 06:48:49.847697 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:49 crc kubenswrapper[4846]: I1005 06:48:49.847785 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:49 crc kubenswrapper[4846]: I1005 06:48:49.847806 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:49 crc kubenswrapper[4846]: I1005 06:48:49.847833 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:49 crc kubenswrapper[4846]: I1005 06:48:49.847854 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:49Z","lastTransitionTime":"2025-10-05T06:48:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:49 crc kubenswrapper[4846]: I1005 06:48:49.863941 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c9f812b-ecc8-4985-a670-53fa80fbb499\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2b5ee9781c01d6aec3ac778efef1067b6c3bfe4ed37431bdebe188fcde0165f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:47:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3fb4717692aa1fec0012b514ef8d18d1ca3707cfa4bb8248bb93841a0a40f11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:47:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0b6a7afe3094f1c92093c05c28344897a615d590c2fbb507c87e18a44c6c831\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://640ade2c733d7916ec213a201a1dd614d261c69eab654c80100efbc99a768995\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:47:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:49Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:49 crc kubenswrapper[4846]: I1005 06:48:49.888608 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:49Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:49 crc kubenswrapper[4846]: I1005 06:48:49.910851 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e67d0e447a8a6a381259e25d25badc791194e1291f56b1cd77bd63989b20408\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dkhl7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://516f88f4990f5962d8b694a3d889259cc5508a309919870191c2e1331640c973\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dkhl7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-fscvf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:49Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:49 crc kubenswrapper[4846]: I1005 06:48:49.939980 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-txzlk_93a6ace4-8d64-44d7-9637-457d3af6543c/ovnkube-controller/1.log" Oct 05 06:48:49 crc kubenswrapper[4846]: I1005 06:48:49.943812 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-txzlk" event={"ID":"93a6ace4-8d64-44d7-9637-457d3af6543c","Type":"ContainerStarted","Data":"dd084bb9806ece6529db0866cee69e6a3152f0da7e0f13fc2e4adcbe2f55962c"} Oct 05 06:48:49 crc kubenswrapper[4846]: I1005 06:48:49.944114 4846 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 05 06:48:49 crc kubenswrapper[4846]: I1005 06:48:49.950315 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:49 crc kubenswrapper[4846]: I1005 06:48:49.950356 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:49 crc kubenswrapper[4846]: I1005 06:48:49.950369 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:49 crc kubenswrapper[4846]: I1005 06:48:49.950385 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:49 crc kubenswrapper[4846]: I1005 06:48:49.950399 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:49Z","lastTransitionTime":"2025-10-05T06:48:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:49 crc kubenswrapper[4846]: I1005 06:48:49.966586 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ff10c31459b3b3724ee7b769bd6f72f9ed80cce5419b6adcdf69a3239b70756\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:49Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:49 crc kubenswrapper[4846]: I1005 06:48:49.982805 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://347f900743b7bba3daf92f21c7eba0625e8c774f50057bc9472f0c82119325e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:49Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:50 crc kubenswrapper[4846]: I1005 06:48:50.002510 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56f1fe1f9f06f0249cf937a6963fda82467732dc056ca2b2189be6d8dd9d4a05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7212ce5f42c1fb55146abb190a557a504b43f16af6df84f90e1073567a329ec5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:50Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:50 crc kubenswrapper[4846]: I1005 06:48:50.023634 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-7wm5q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9bc7744f-d700-450f-8b8d-98140877fee9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90379e74c544fccca44f7acdc228ff378154390e211ac801b79f8d074012241e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h5p4q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-7wm5q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:50Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:50 crc kubenswrapper[4846]: I1005 06:48:50.045126 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"775b5702-a131-4d6c-852d-91bfa056cac5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0839c1765df2c68d2554e9c66d289505e559a9672caf8069feeff2c4fdfd2171\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15dbaf2445704da0987f256f81dff2ee95664988541fa48ce94898afeadab2e8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd3b3ab323c6d5153054aa04af2d91f7b9e8436ff1e2f75f3594b57831b2a416\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://67c46f5f0a3ace95bd2e9479c5561126664e18b148c5eceea5bf5efc27e81a67\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c52e076eddfd307110b1f6de00a169b81af8f2785af7becef2a820ce363a5861\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-05T06:48:14Z\\\",\\\"message\\\":\\\"le observer\\\\nW1005 06:48:14.337870 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1005 06:48:14.338086 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1005 06:48:14.339389 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1467752215/tls.crt::/tmp/serving-cert-1467752215/tls.key\\\\\\\"\\\\nI1005 06:48:14.864345 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1005 06:48:14.868087 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1005 06:48:14.868114 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1005 06:48:14.868140 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1005 06:48:14.868147 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1005 06:48:14.877354 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1005 06:48:14.877412 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1005 06:48:14.877456 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1005 06:48:14.877471 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1005 06:48:14.877490 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1005 06:48:14.877501 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1005 06:48:14.877507 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1005 06:48:14.877515 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1005 06:48:14.880326 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:13Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4a091716ff935adb42ae2e88048619fd347bd174f64e5f9c97a01f81690e57f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:01Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dc5411b642f99b13d514d2f21defad0c394aeeb82a966f72300076617e2586c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5dc5411b642f99b13d514d2f21defad0c394aeeb82a966f72300076617e2586c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:47:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:47:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:47:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:50Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:50 crc kubenswrapper[4846]: I1005 06:48:50.054015 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:50 crc kubenswrapper[4846]: I1005 06:48:50.054044 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:50 crc kubenswrapper[4846]: I1005 06:48:50.054054 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:50 crc kubenswrapper[4846]: I1005 06:48:50.054071 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:50 crc kubenswrapper[4846]: I1005 06:48:50.054083 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:50Z","lastTransitionTime":"2025-10-05T06:48:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:50 crc kubenswrapper[4846]: I1005 06:48:50.064681 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:50Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:50 crc kubenswrapper[4846]: I1005 06:48:50.080108 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-r7ztb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a5119cd9-af70-4324-a955-978306ab0b20\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06231a87429ec292cb57ebd88cda7ad547bcb32e80343b776833621a11c28ad1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jfcj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-r7ztb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:50Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:50 crc kubenswrapper[4846]: I1005 06:48:50.097508 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-9r2kt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"59b83f4f-71f0-4b28-9275-56d5f933ec2c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tdwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tdwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:38Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-9r2kt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:50Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:50 crc kubenswrapper[4846]: I1005 06:48:50.132568 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c482bb97-fd50-40cf-96ea-f806c3f01125\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bfe45d30e4fa141541aca1d46bc5cc1da0a7b4a5eb1fc9bb2caa5b7fdcb0b264\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://496a4782faa30a9a20ff63038141600cd635f12902d1cf17989cd067fcdfe929\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://281be077869a6aa1ca815770871a14413d4a97e97aeb550fd5cd215cba88b21f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6771c7215d2f5dd5838d17888375a4768b9d1755fc2e21bda1bbfbb2eb1dbb95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c17a711c2c7473b076a3944ffe7c700808d05cba582fbd55c5c2ac8ff73838a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44a9f740f0a06c8bf928e2a9d1d5979217558c31e73a333631baccebf00fd5d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://44a9f740f0a06c8bf928e2a9d1d5979217558c31e73a333631baccebf00fd5d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:47:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:47:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://48ca0ba6ca6092bfcb5b153220c054f02cc36fb22707c42e021f06d329e18e2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48ca0ba6ca6092bfcb5b153220c054f02cc36fb22707c42e021f06d329e18e2e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:00Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4006d0cea128675c8885d879171204abffc5e11ea0664750a12b4bc7344ccb2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4006d0cea128675c8885d879171204abffc5e11ea0664750a12b4bc7344ccb2e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:47:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:50Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:50 crc kubenswrapper[4846]: I1005 06:48:50.153368 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:50Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:50 crc kubenswrapper[4846]: I1005 06:48:50.157122 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:50 crc kubenswrapper[4846]: I1005 06:48:50.157213 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:50 crc kubenswrapper[4846]: I1005 06:48:50.157239 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:50 crc kubenswrapper[4846]: I1005 06:48:50.157270 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:50 crc kubenswrapper[4846]: I1005 06:48:50.157294 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:50Z","lastTransitionTime":"2025-10-05T06:48:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:50 crc kubenswrapper[4846]: I1005 06:48:50.184818 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-sfmxd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1dbb13ac-136c-4a1c-9229-b4e8ac4f9c71\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b0af6a6e735f7b7877c177c1d53d1a8dc12c4713a8059c4766fc858af31c9a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://226d977b596d07aeda0486452e2a8605949ed4d2453b21f56174937c395d6f36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://226d977b596d07aeda0486452e2a8605949ed4d2453b21f56174937c395d6f36\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://99c07a4bdb3059d3a99ffdaf794d668d52cffd9e5e3bb8945204dd6d1f8c8328\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99c07a4bdb3059d3a99ffdaf794d668d52cffd9e5e3bb8945204dd6d1f8c8328\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27095272bf4f31d63a78a436636de0aae12f7103597da944f24ed7a74329ee3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27095272bf4f31d63a78a436636de0aae12f7103597da944f24ed7a74329ee3f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1211571af66d8a94302b65feccf30768ba82b7cbad918e44dd405787dce484cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1211571af66d8a94302b65feccf30768ba82b7cbad918e44dd405787dce484cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27631f90d528900fa180c15cfce80e96d4cdba07743eb6d847e2e822d9917218\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27631f90d528900fa180c15cfce80e96d4cdba07743eb6d847e2e822d9917218\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f345693e9aeb1f4dc196349d3354f3c833a2b0415a693298b1cfccfa2b0d4459\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f345693e9aeb1f4dc196349d3354f3c833a2b0415a693298b1cfccfa2b0d4459\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-sfmxd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:50Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:50 crc kubenswrapper[4846]: I1005 06:48:50.224879 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-txzlk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"93a6ace4-8d64-44d7-9637-457d3af6543c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ad1940d07e37217089642478ece9ccdaca74784738a400929a9f5359d158c01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a6f6b2c32dd3da5f9fd5e6971c1773e81a9a24184b7aa5b543cec879001ae53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e9b0219ecbc7e60905d6de3e67a2c8dc9af1017b4aed9538ee0c58c861f2438\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af5e5a4da9ad7371d5a24633e8268f12e3a7d419a8c93b11ddbcc006d8339c6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8da85222775abc7a772f6a8ff8255f1a0d5b15807ba177a7e3e2b2d01a4f24b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1409651de8c897804f87be03d016c138f8acf27e662dce265a93e15c1848adcf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd084bb9806ece6529db0866cee69e6a3152f0da7e0f13fc2e4adcbe2f55962c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://93149bae0504218735e0900b660ad9759ebbec3eb4dc214cc40ad349ac002d45\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-05T06:48:36Z\\\",\\\"message\\\":\\\"ePorts:nil,LoadBalancerClass:nil,InternalTrafficPolicy:*Cluster,TrafficDistribution:nil,},Status:ServiceStatus{LoadBalancer:LoadBalancerStatus{Ingress:[]LoadBalancerIngress{},},Conditions:[]Condition{},},}\\\\nI1005 06:48:35.889902 6320 lb_config.go:1031] Cluster endpoints for openshift-kube-apiserver/apiserver for network=default are: map[]\\\\nI1005 06:48:35.889910 6320 services_controller.go:443] Built service openshift-kube-apiserver/apiserver LB cluster-wide configs for network=default: []services.lbConfig{services.lbConfig{vips:[]string{\\\\\\\"10.217.4.93\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:443, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI1005 06:48:35.889925 6320 services_controller.go:444] Built service openshift-kube-apiserver/apiserver LB per-node configs for network=default: []services.lbConfig(nil)\\\\nF1005 06:48:35.889931 6320 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: failed to add event handler: handler {0x1e60340 0x1e60020 0x1e5ffc0} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to se\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:35Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0df0ebcd87dc5af4b17736b43ec7b2395968b19ed16e775366af14e2b2d49c52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff3c6545ff3d2436da27bbd223b950b30b0c69a3e7f3034cb0e61380cb38d338\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ff3c6545ff3d2436da27bbd223b950b30b0c69a3e7f3034cb0e61380cb38d338\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:24Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-txzlk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:50Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:50 crc kubenswrapper[4846]: I1005 06:48:50.248241 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-gklb6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"698ab882-bb26-4b4c-81e9-8305a44fe068\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a6edca88ea244a58f8cf6cc431110a960c6bfd892812cd0d00fa3f97be3bc67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x6d6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c191538ab797b8056820de7da1f8e93ea18873891a32382b8a59c9af762e103\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x6d6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:37Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-gklb6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:50Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:50 crc kubenswrapper[4846]: I1005 06:48:50.260081 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:50 crc kubenswrapper[4846]: I1005 06:48:50.260150 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:50 crc kubenswrapper[4846]: I1005 06:48:50.260173 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:50 crc kubenswrapper[4846]: I1005 06:48:50.260241 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:50 crc kubenswrapper[4846]: I1005 06:48:50.260265 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:50Z","lastTransitionTime":"2025-10-05T06:48:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:50 crc kubenswrapper[4846]: I1005 06:48:50.273521 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c9f812b-ecc8-4985-a670-53fa80fbb499\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2b5ee9781c01d6aec3ac778efef1067b6c3bfe4ed37431bdebe188fcde0165f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:47:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3fb4717692aa1fec0012b514ef8d18d1ca3707cfa4bb8248bb93841a0a40f11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:47:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0b6a7afe3094f1c92093c05c28344897a615d590c2fbb507c87e18a44c6c831\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://640ade2c733d7916ec213a201a1dd614d261c69eab654c80100efbc99a768995\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:47:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:50Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:50 crc kubenswrapper[4846]: I1005 06:48:50.301514 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:50Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:50 crc kubenswrapper[4846]: I1005 06:48:50.339024 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e67d0e447a8a6a381259e25d25badc791194e1291f56b1cd77bd63989b20408\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dkhl7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://516f88f4990f5962d8b694a3d889259cc5508a309919870191c2e1331640c973\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dkhl7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-fscvf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:50Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:50 crc kubenswrapper[4846]: I1005 06:48:50.359583 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-w2sj8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32931fce-7374-4cab-a2e0-79d96a842274\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0df535d65d39de6c9a3dfb931bafc4469adc22b72f7a7903beda4fef8cd48138\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9rtxf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:27Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-w2sj8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:50Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:50 crc kubenswrapper[4846]: I1005 06:48:50.362289 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:50 crc kubenswrapper[4846]: I1005 06:48:50.362343 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:50 crc kubenswrapper[4846]: I1005 06:48:50.362360 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:50 crc kubenswrapper[4846]: I1005 06:48:50.362383 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:50 crc kubenswrapper[4846]: I1005 06:48:50.362400 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:50Z","lastTransitionTime":"2025-10-05T06:48:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:50 crc kubenswrapper[4846]: I1005 06:48:50.467087 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:50 crc kubenswrapper[4846]: I1005 06:48:50.467149 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:50 crc kubenswrapper[4846]: I1005 06:48:50.467172 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:50 crc kubenswrapper[4846]: I1005 06:48:50.467250 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:50 crc kubenswrapper[4846]: I1005 06:48:50.467275 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:50Z","lastTransitionTime":"2025-10-05T06:48:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:50 crc kubenswrapper[4846]: I1005 06:48:50.497776 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9r2kt" Oct 05 06:48:50 crc kubenswrapper[4846]: E1005 06:48:50.500542 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9r2kt" podUID="59b83f4f-71f0-4b28-9275-56d5f933ec2c" Oct 05 06:48:50 crc kubenswrapper[4846]: I1005 06:48:50.500661 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 05 06:48:50 crc kubenswrapper[4846]: E1005 06:48:50.500870 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 05 06:48:50 crc kubenswrapper[4846]: I1005 06:48:50.570025 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:50 crc kubenswrapper[4846]: I1005 06:48:50.570097 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:50 crc kubenswrapper[4846]: I1005 06:48:50.570116 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:50 crc kubenswrapper[4846]: I1005 06:48:50.570141 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:50 crc kubenswrapper[4846]: I1005 06:48:50.570161 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:50Z","lastTransitionTime":"2025-10-05T06:48:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:50 crc kubenswrapper[4846]: I1005 06:48:50.672852 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:50 crc kubenswrapper[4846]: I1005 06:48:50.672954 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:50 crc kubenswrapper[4846]: I1005 06:48:50.672973 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:50 crc kubenswrapper[4846]: I1005 06:48:50.673000 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:50 crc kubenswrapper[4846]: I1005 06:48:50.673016 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:50Z","lastTransitionTime":"2025-10-05T06:48:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:50 crc kubenswrapper[4846]: I1005 06:48:50.776400 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:50 crc kubenswrapper[4846]: I1005 06:48:50.776461 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:50 crc kubenswrapper[4846]: I1005 06:48:50.776478 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:50 crc kubenswrapper[4846]: I1005 06:48:50.776506 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:50 crc kubenswrapper[4846]: I1005 06:48:50.776525 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:50Z","lastTransitionTime":"2025-10-05T06:48:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:50 crc kubenswrapper[4846]: I1005 06:48:50.879776 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:50 crc kubenswrapper[4846]: I1005 06:48:50.879814 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:50 crc kubenswrapper[4846]: I1005 06:48:50.879830 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:50 crc kubenswrapper[4846]: I1005 06:48:50.879850 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:50 crc kubenswrapper[4846]: I1005 06:48:50.879863 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:50Z","lastTransitionTime":"2025-10-05T06:48:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:50 crc kubenswrapper[4846]: I1005 06:48:50.949451 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-txzlk_93a6ace4-8d64-44d7-9637-457d3af6543c/ovnkube-controller/2.log" Oct 05 06:48:50 crc kubenswrapper[4846]: I1005 06:48:50.950629 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-txzlk_93a6ace4-8d64-44d7-9637-457d3af6543c/ovnkube-controller/1.log" Oct 05 06:48:50 crc kubenswrapper[4846]: I1005 06:48:50.954161 4846 generic.go:334] "Generic (PLEG): container finished" podID="93a6ace4-8d64-44d7-9637-457d3af6543c" containerID="dd084bb9806ece6529db0866cee69e6a3152f0da7e0f13fc2e4adcbe2f55962c" exitCode=1 Oct 05 06:48:50 crc kubenswrapper[4846]: I1005 06:48:50.954211 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-txzlk" event={"ID":"93a6ace4-8d64-44d7-9637-457d3af6543c","Type":"ContainerDied","Data":"dd084bb9806ece6529db0866cee69e6a3152f0da7e0f13fc2e4adcbe2f55962c"} Oct 05 06:48:50 crc kubenswrapper[4846]: I1005 06:48:50.954250 4846 scope.go:117] "RemoveContainer" containerID="93149bae0504218735e0900b660ad9759ebbec3eb4dc214cc40ad349ac002d45" Oct 05 06:48:50 crc kubenswrapper[4846]: I1005 06:48:50.955692 4846 scope.go:117] "RemoveContainer" containerID="dd084bb9806ece6529db0866cee69e6a3152f0da7e0f13fc2e4adcbe2f55962c" Oct 05 06:48:50 crc kubenswrapper[4846]: E1005 06:48:50.956048 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-txzlk_openshift-ovn-kubernetes(93a6ace4-8d64-44d7-9637-457d3af6543c)\"" pod="openshift-ovn-kubernetes/ovnkube-node-txzlk" podUID="93a6ace4-8d64-44d7-9637-457d3af6543c" Oct 05 06:48:50 crc kubenswrapper[4846]: I1005 06:48:50.976876 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c9f812b-ecc8-4985-a670-53fa80fbb499\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2b5ee9781c01d6aec3ac778efef1067b6c3bfe4ed37431bdebe188fcde0165f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:47:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3fb4717692aa1fec0012b514ef8d18d1ca3707cfa4bb8248bb93841a0a40f11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:47:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0b6a7afe3094f1c92093c05c28344897a615d590c2fbb507c87e18a44c6c831\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://640ade2c733d7916ec213a201a1dd614d261c69eab654c80100efbc99a768995\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:47:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:50Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:50 crc kubenswrapper[4846]: I1005 06:48:50.982427 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:50 crc kubenswrapper[4846]: I1005 06:48:50.982471 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:50 crc kubenswrapper[4846]: I1005 06:48:50.982490 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:50 crc kubenswrapper[4846]: I1005 06:48:50.982512 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:50 crc kubenswrapper[4846]: I1005 06:48:50.982529 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:50Z","lastTransitionTime":"2025-10-05T06:48:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:50 crc kubenswrapper[4846]: I1005 06:48:50.998889 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:50Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:51 crc kubenswrapper[4846]: I1005 06:48:51.017376 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e67d0e447a8a6a381259e25d25badc791194e1291f56b1cd77bd63989b20408\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dkhl7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://516f88f4990f5962d8b694a3d889259cc5508a309919870191c2e1331640c973\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dkhl7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-fscvf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:51Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:51 crc kubenswrapper[4846]: I1005 06:48:51.035760 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-w2sj8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32931fce-7374-4cab-a2e0-79d96a842274\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0df535d65d39de6c9a3dfb931bafc4469adc22b72f7a7903beda4fef8cd48138\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9rtxf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:27Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-w2sj8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:51Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:51 crc kubenswrapper[4846]: I1005 06:48:51.058457 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ff10c31459b3b3724ee7b769bd6f72f9ed80cce5419b6adcdf69a3239b70756\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:51Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:51 crc kubenswrapper[4846]: I1005 06:48:51.076381 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://347f900743b7bba3daf92f21c7eba0625e8c774f50057bc9472f0c82119325e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:51Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:51 crc kubenswrapper[4846]: I1005 06:48:51.086584 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:51 crc kubenswrapper[4846]: I1005 06:48:51.086635 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:51 crc kubenswrapper[4846]: I1005 06:48:51.086651 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:51 crc kubenswrapper[4846]: I1005 06:48:51.086674 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:51 crc kubenswrapper[4846]: I1005 06:48:51.086691 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:51Z","lastTransitionTime":"2025-10-05T06:48:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:51 crc kubenswrapper[4846]: I1005 06:48:51.101949 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56f1fe1f9f06f0249cf937a6963fda82467732dc056ca2b2189be6d8dd9d4a05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7212ce5f42c1fb55146abb190a557a504b43f16af6df84f90e1073567a329ec5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:51Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:51 crc kubenswrapper[4846]: I1005 06:48:51.125117 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-7wm5q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9bc7744f-d700-450f-8b8d-98140877fee9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90379e74c544fccca44f7acdc228ff378154390e211ac801b79f8d074012241e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h5p4q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-7wm5q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:51Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:51 crc kubenswrapper[4846]: I1005 06:48:51.145695 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"775b5702-a131-4d6c-852d-91bfa056cac5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0839c1765df2c68d2554e9c66d289505e559a9672caf8069feeff2c4fdfd2171\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15dbaf2445704da0987f256f81dff2ee95664988541fa48ce94898afeadab2e8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd3b3ab323c6d5153054aa04af2d91f7b9e8436ff1e2f75f3594b57831b2a416\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://67c46f5f0a3ace95bd2e9479c5561126664e18b148c5eceea5bf5efc27e81a67\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c52e076eddfd307110b1f6de00a169b81af8f2785af7becef2a820ce363a5861\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-05T06:48:14Z\\\",\\\"message\\\":\\\"le observer\\\\nW1005 06:48:14.337870 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1005 06:48:14.338086 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1005 06:48:14.339389 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1467752215/tls.crt::/tmp/serving-cert-1467752215/tls.key\\\\\\\"\\\\nI1005 06:48:14.864345 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1005 06:48:14.868087 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1005 06:48:14.868114 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1005 06:48:14.868140 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1005 06:48:14.868147 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1005 06:48:14.877354 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1005 06:48:14.877412 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1005 06:48:14.877456 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1005 06:48:14.877471 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1005 06:48:14.877490 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1005 06:48:14.877501 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1005 06:48:14.877507 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1005 06:48:14.877515 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1005 06:48:14.880326 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:13Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4a091716ff935adb42ae2e88048619fd347bd174f64e5f9c97a01f81690e57f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:01Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dc5411b642f99b13d514d2f21defad0c394aeeb82a966f72300076617e2586c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5dc5411b642f99b13d514d2f21defad0c394aeeb82a966f72300076617e2586c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:47:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:47:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:47:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:51Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:51 crc kubenswrapper[4846]: I1005 06:48:51.165318 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:51Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:51 crc kubenswrapper[4846]: I1005 06:48:51.180752 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-r7ztb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a5119cd9-af70-4324-a955-978306ab0b20\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06231a87429ec292cb57ebd88cda7ad547bcb32e80343b776833621a11c28ad1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jfcj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-r7ztb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:51Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:51 crc kubenswrapper[4846]: I1005 06:48:51.189166 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:51 crc kubenswrapper[4846]: I1005 06:48:51.189255 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:51 crc kubenswrapper[4846]: I1005 06:48:51.189281 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:51 crc kubenswrapper[4846]: I1005 06:48:51.189313 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:51 crc kubenswrapper[4846]: I1005 06:48:51.189337 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:51Z","lastTransitionTime":"2025-10-05T06:48:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:51 crc kubenswrapper[4846]: I1005 06:48:51.197499 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-9r2kt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"59b83f4f-71f0-4b28-9275-56d5f933ec2c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tdwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tdwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:38Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-9r2kt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:51Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:51 crc kubenswrapper[4846]: I1005 06:48:51.231775 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c482bb97-fd50-40cf-96ea-f806c3f01125\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bfe45d30e4fa141541aca1d46bc5cc1da0a7b4a5eb1fc9bb2caa5b7fdcb0b264\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://496a4782faa30a9a20ff63038141600cd635f12902d1cf17989cd067fcdfe929\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://281be077869a6aa1ca815770871a14413d4a97e97aeb550fd5cd215cba88b21f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6771c7215d2f5dd5838d17888375a4768b9d1755fc2e21bda1bbfbb2eb1dbb95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c17a711c2c7473b076a3944ffe7c700808d05cba582fbd55c5c2ac8ff73838a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44a9f740f0a06c8bf928e2a9d1d5979217558c31e73a333631baccebf00fd5d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://44a9f740f0a06c8bf928e2a9d1d5979217558c31e73a333631baccebf00fd5d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:47:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:47:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://48ca0ba6ca6092bfcb5b153220c054f02cc36fb22707c42e021f06d329e18e2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48ca0ba6ca6092bfcb5b153220c054f02cc36fb22707c42e021f06d329e18e2e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:00Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4006d0cea128675c8885d879171204abffc5e11ea0664750a12b4bc7344ccb2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4006d0cea128675c8885d879171204abffc5e11ea0664750a12b4bc7344ccb2e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:47:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:51Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:51 crc kubenswrapper[4846]: I1005 06:48:51.254074 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:51Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:51 crc kubenswrapper[4846]: I1005 06:48:51.275516 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-sfmxd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1dbb13ac-136c-4a1c-9229-b4e8ac4f9c71\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b0af6a6e735f7b7877c177c1d53d1a8dc12c4713a8059c4766fc858af31c9a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://226d977b596d07aeda0486452e2a8605949ed4d2453b21f56174937c395d6f36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://226d977b596d07aeda0486452e2a8605949ed4d2453b21f56174937c395d6f36\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://99c07a4bdb3059d3a99ffdaf794d668d52cffd9e5e3bb8945204dd6d1f8c8328\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99c07a4bdb3059d3a99ffdaf794d668d52cffd9e5e3bb8945204dd6d1f8c8328\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27095272bf4f31d63a78a436636de0aae12f7103597da944f24ed7a74329ee3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27095272bf4f31d63a78a436636de0aae12f7103597da944f24ed7a74329ee3f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1211571af66d8a94302b65feccf30768ba82b7cbad918e44dd405787dce484cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1211571af66d8a94302b65feccf30768ba82b7cbad918e44dd405787dce484cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27631f90d528900fa180c15cfce80e96d4cdba07743eb6d847e2e822d9917218\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27631f90d528900fa180c15cfce80e96d4cdba07743eb6d847e2e822d9917218\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f345693e9aeb1f4dc196349d3354f3c833a2b0415a693298b1cfccfa2b0d4459\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f345693e9aeb1f4dc196349d3354f3c833a2b0415a693298b1cfccfa2b0d4459\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-sfmxd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:51Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:51 crc kubenswrapper[4846]: I1005 06:48:51.292270 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:51 crc kubenswrapper[4846]: I1005 06:48:51.292332 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:51 crc kubenswrapper[4846]: I1005 06:48:51.292349 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:51 crc kubenswrapper[4846]: I1005 06:48:51.292396 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:51 crc kubenswrapper[4846]: I1005 06:48:51.292415 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:51Z","lastTransitionTime":"2025-10-05T06:48:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:51 crc kubenswrapper[4846]: I1005 06:48:51.307979 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-txzlk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"93a6ace4-8d64-44d7-9637-457d3af6543c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ad1940d07e37217089642478ece9ccdaca74784738a400929a9f5359d158c01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a6f6b2c32dd3da5f9fd5e6971c1773e81a9a24184b7aa5b543cec879001ae53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e9b0219ecbc7e60905d6de3e67a2c8dc9af1017b4aed9538ee0c58c861f2438\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af5e5a4da9ad7371d5a24633e8268f12e3a7d419a8c93b11ddbcc006d8339c6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8da85222775abc7a772f6a8ff8255f1a0d5b15807ba177a7e3e2b2d01a4f24b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1409651de8c897804f87be03d016c138f8acf27e662dce265a93e15c1848adcf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd084bb9806ece6529db0866cee69e6a3152f0da7e0f13fc2e4adcbe2f55962c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://93149bae0504218735e0900b660ad9759ebbec3eb4dc214cc40ad349ac002d45\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-05T06:48:36Z\\\",\\\"message\\\":\\\"ePorts:nil,LoadBalancerClass:nil,InternalTrafficPolicy:*Cluster,TrafficDistribution:nil,},Status:ServiceStatus{LoadBalancer:LoadBalancerStatus{Ingress:[]LoadBalancerIngress{},},Conditions:[]Condition{},},}\\\\nI1005 06:48:35.889902 6320 lb_config.go:1031] Cluster endpoints for openshift-kube-apiserver/apiserver for network=default are: map[]\\\\nI1005 06:48:35.889910 6320 services_controller.go:443] Built service openshift-kube-apiserver/apiserver LB cluster-wide configs for network=default: []services.lbConfig{services.lbConfig{vips:[]string{\\\\\\\"10.217.4.93\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:443, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI1005 06:48:35.889925 6320 services_controller.go:444] Built service openshift-kube-apiserver/apiserver LB per-node configs for network=default: []services.lbConfig(nil)\\\\nF1005 06:48:35.889931 6320 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: failed to add event handler: handler {0x1e60340 0x1e60020 0x1e5ffc0} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to se\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:35Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd084bb9806ece6529db0866cee69e6a3152f0da7e0f13fc2e4adcbe2f55962c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-05T06:48:50Z\\\",\\\"message\\\":\\\":50.819394 6534 handler.go:208] Removed *v1.Node event handler 2\\\\nI1005 06:48:50.819434 6534 handler.go:208] Removed *v1.Node event handler 7\\\\nI1005 06:48:50.819451 6534 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1005 06:48:50.819624 6534 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1005 06:48:50.819659 6534 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1005 06:48:50.819711 6534 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1005 06:48:50.819735 6534 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1005 06:48:50.820556 6534 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1005 06:48:50.820605 6534 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1005 06:48:50.820649 6534 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1005 06:48:50.820664 6534 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1005 06:48:50.820706 6534 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1005 06:48:50.820711 6534 factory.go:656] Stopping watch factory\\\\nI1005 06:48:50.820731 6534 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1005 06:48:50.820750 6534 ovnkube.go:599] Stopped ovnkube\\\\nI1005 06:48:50.820796 6534 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1005 06:48:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0df0ebcd87dc5af4b17736b43ec7b2395968b19ed16e775366af14e2b2d49c52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff3c6545ff3d2436da27bbd223b950b30b0c69a3e7f3034cb0e61380cb38d338\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ff3c6545ff3d2436da27bbd223b950b30b0c69a3e7f3034cb0e61380cb38d338\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:24Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-txzlk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:51Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:51 crc kubenswrapper[4846]: I1005 06:48:51.330460 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-gklb6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"698ab882-bb26-4b4c-81e9-8305a44fe068\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a6edca88ea244a58f8cf6cc431110a960c6bfd892812cd0d00fa3f97be3bc67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x6d6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c191538ab797b8056820de7da1f8e93ea18873891a32382b8a59c9af762e103\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x6d6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:37Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-gklb6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:51Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:51 crc kubenswrapper[4846]: I1005 06:48:51.395285 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:51 crc kubenswrapper[4846]: I1005 06:48:51.395349 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:51 crc kubenswrapper[4846]: I1005 06:48:51.395367 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:51 crc kubenswrapper[4846]: I1005 06:48:51.395396 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:51 crc kubenswrapper[4846]: I1005 06:48:51.395414 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:51Z","lastTransitionTime":"2025-10-05T06:48:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:51 crc kubenswrapper[4846]: I1005 06:48:51.496704 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 05 06:48:51 crc kubenswrapper[4846]: I1005 06:48:51.496869 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 05 06:48:51 crc kubenswrapper[4846]: E1005 06:48:51.497040 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 05 06:48:51 crc kubenswrapper[4846]: E1005 06:48:51.497219 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 05 06:48:51 crc kubenswrapper[4846]: I1005 06:48:51.498465 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:51 crc kubenswrapper[4846]: I1005 06:48:51.498512 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:51 crc kubenswrapper[4846]: I1005 06:48:51.498528 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:51 crc kubenswrapper[4846]: I1005 06:48:51.498551 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:51 crc kubenswrapper[4846]: I1005 06:48:51.498567 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:51Z","lastTransitionTime":"2025-10-05T06:48:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:51 crc kubenswrapper[4846]: I1005 06:48:51.602667 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:51 crc kubenswrapper[4846]: I1005 06:48:51.602753 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:51 crc kubenswrapper[4846]: I1005 06:48:51.602778 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:51 crc kubenswrapper[4846]: I1005 06:48:51.602809 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:51 crc kubenswrapper[4846]: I1005 06:48:51.602836 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:51Z","lastTransitionTime":"2025-10-05T06:48:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:51 crc kubenswrapper[4846]: I1005 06:48:51.706736 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:51 crc kubenswrapper[4846]: I1005 06:48:51.706828 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:51 crc kubenswrapper[4846]: I1005 06:48:51.706847 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:51 crc kubenswrapper[4846]: I1005 06:48:51.706875 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:51 crc kubenswrapper[4846]: I1005 06:48:51.706899 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:51Z","lastTransitionTime":"2025-10-05T06:48:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:51 crc kubenswrapper[4846]: I1005 06:48:51.810450 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:51 crc kubenswrapper[4846]: I1005 06:48:51.810529 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:51 crc kubenswrapper[4846]: I1005 06:48:51.810553 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:51 crc kubenswrapper[4846]: I1005 06:48:51.810587 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:51 crc kubenswrapper[4846]: I1005 06:48:51.810607 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:51Z","lastTransitionTime":"2025-10-05T06:48:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:51 crc kubenswrapper[4846]: I1005 06:48:51.815915 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-txzlk" Oct 05 06:48:51 crc kubenswrapper[4846]: I1005 06:48:51.914366 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:51 crc kubenswrapper[4846]: I1005 06:48:51.914443 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:51 crc kubenswrapper[4846]: I1005 06:48:51.914466 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:51 crc kubenswrapper[4846]: I1005 06:48:51.914495 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:51 crc kubenswrapper[4846]: I1005 06:48:51.914518 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:51Z","lastTransitionTime":"2025-10-05T06:48:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:51 crc kubenswrapper[4846]: I1005 06:48:51.963278 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-txzlk_93a6ace4-8d64-44d7-9637-457d3af6543c/ovnkube-controller/2.log" Oct 05 06:48:51 crc kubenswrapper[4846]: I1005 06:48:51.969256 4846 scope.go:117] "RemoveContainer" containerID="dd084bb9806ece6529db0866cee69e6a3152f0da7e0f13fc2e4adcbe2f55962c" Oct 05 06:48:51 crc kubenswrapper[4846]: E1005 06:48:51.969537 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-txzlk_openshift-ovn-kubernetes(93a6ace4-8d64-44d7-9637-457d3af6543c)\"" pod="openshift-ovn-kubernetes/ovnkube-node-txzlk" podUID="93a6ace4-8d64-44d7-9637-457d3af6543c" Oct 05 06:48:51 crc kubenswrapper[4846]: I1005 06:48:51.995668 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ff10c31459b3b3724ee7b769bd6f72f9ed80cce5419b6adcdf69a3239b70756\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:51Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:52 crc kubenswrapper[4846]: I1005 06:48:52.016889 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://347f900743b7bba3daf92f21c7eba0625e8c774f50057bc9472f0c82119325e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:52Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:52 crc kubenswrapper[4846]: I1005 06:48:52.017831 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:52 crc kubenswrapper[4846]: I1005 06:48:52.017895 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:52 crc kubenswrapper[4846]: I1005 06:48:52.017912 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:52 crc kubenswrapper[4846]: I1005 06:48:52.017940 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:52 crc kubenswrapper[4846]: I1005 06:48:52.017958 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:52Z","lastTransitionTime":"2025-10-05T06:48:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:52 crc kubenswrapper[4846]: I1005 06:48:52.039890 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56f1fe1f9f06f0249cf937a6963fda82467732dc056ca2b2189be6d8dd9d4a05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7212ce5f42c1fb55146abb190a557a504b43f16af6df84f90e1073567a329ec5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:52Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:52 crc kubenswrapper[4846]: I1005 06:48:52.061633 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-7wm5q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9bc7744f-d700-450f-8b8d-98140877fee9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90379e74c544fccca44f7acdc228ff378154390e211ac801b79f8d074012241e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h5p4q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-7wm5q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:52Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:52 crc kubenswrapper[4846]: I1005 06:48:52.084582 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"775b5702-a131-4d6c-852d-91bfa056cac5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0839c1765df2c68d2554e9c66d289505e559a9672caf8069feeff2c4fdfd2171\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15dbaf2445704da0987f256f81dff2ee95664988541fa48ce94898afeadab2e8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd3b3ab323c6d5153054aa04af2d91f7b9e8436ff1e2f75f3594b57831b2a416\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://67c46f5f0a3ace95bd2e9479c5561126664e18b148c5eceea5bf5efc27e81a67\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c52e076eddfd307110b1f6de00a169b81af8f2785af7becef2a820ce363a5861\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-05T06:48:14Z\\\",\\\"message\\\":\\\"le observer\\\\nW1005 06:48:14.337870 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1005 06:48:14.338086 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1005 06:48:14.339389 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1467752215/tls.crt::/tmp/serving-cert-1467752215/tls.key\\\\\\\"\\\\nI1005 06:48:14.864345 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1005 06:48:14.868087 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1005 06:48:14.868114 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1005 06:48:14.868140 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1005 06:48:14.868147 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1005 06:48:14.877354 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1005 06:48:14.877412 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1005 06:48:14.877456 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1005 06:48:14.877471 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1005 06:48:14.877490 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1005 06:48:14.877501 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1005 06:48:14.877507 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1005 06:48:14.877515 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1005 06:48:14.880326 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:13Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4a091716ff935adb42ae2e88048619fd347bd174f64e5f9c97a01f81690e57f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:01Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dc5411b642f99b13d514d2f21defad0c394aeeb82a966f72300076617e2586c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5dc5411b642f99b13d514d2f21defad0c394aeeb82a966f72300076617e2586c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:47:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:47:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:47:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:52Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:52 crc kubenswrapper[4846]: I1005 06:48:52.104634 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:52Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:52 crc kubenswrapper[4846]: I1005 06:48:52.121277 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-r7ztb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a5119cd9-af70-4324-a955-978306ab0b20\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06231a87429ec292cb57ebd88cda7ad547bcb32e80343b776833621a11c28ad1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jfcj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-r7ztb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:52Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:52 crc kubenswrapper[4846]: I1005 06:48:52.121779 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:52 crc kubenswrapper[4846]: I1005 06:48:52.121838 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:52 crc kubenswrapper[4846]: I1005 06:48:52.121855 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:52 crc kubenswrapper[4846]: I1005 06:48:52.121881 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:52 crc kubenswrapper[4846]: I1005 06:48:52.121899 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:52Z","lastTransitionTime":"2025-10-05T06:48:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:52 crc kubenswrapper[4846]: I1005 06:48:52.137604 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-9r2kt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"59b83f4f-71f0-4b28-9275-56d5f933ec2c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tdwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tdwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:38Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-9r2kt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:52Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:52 crc kubenswrapper[4846]: I1005 06:48:52.171083 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c482bb97-fd50-40cf-96ea-f806c3f01125\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bfe45d30e4fa141541aca1d46bc5cc1da0a7b4a5eb1fc9bb2caa5b7fdcb0b264\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://496a4782faa30a9a20ff63038141600cd635f12902d1cf17989cd067fcdfe929\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://281be077869a6aa1ca815770871a14413d4a97e97aeb550fd5cd215cba88b21f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6771c7215d2f5dd5838d17888375a4768b9d1755fc2e21bda1bbfbb2eb1dbb95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c17a711c2c7473b076a3944ffe7c700808d05cba582fbd55c5c2ac8ff73838a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44a9f740f0a06c8bf928e2a9d1d5979217558c31e73a333631baccebf00fd5d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://44a9f740f0a06c8bf928e2a9d1d5979217558c31e73a333631baccebf00fd5d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:47:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:47:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://48ca0ba6ca6092bfcb5b153220c054f02cc36fb22707c42e021f06d329e18e2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48ca0ba6ca6092bfcb5b153220c054f02cc36fb22707c42e021f06d329e18e2e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:00Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4006d0cea128675c8885d879171204abffc5e11ea0664750a12b4bc7344ccb2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4006d0cea128675c8885d879171204abffc5e11ea0664750a12b4bc7344ccb2e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:47:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:52Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:52 crc kubenswrapper[4846]: I1005 06:48:52.193525 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:52Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:52 crc kubenswrapper[4846]: I1005 06:48:52.218679 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-sfmxd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1dbb13ac-136c-4a1c-9229-b4e8ac4f9c71\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b0af6a6e735f7b7877c177c1d53d1a8dc12c4713a8059c4766fc858af31c9a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://226d977b596d07aeda0486452e2a8605949ed4d2453b21f56174937c395d6f36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://226d977b596d07aeda0486452e2a8605949ed4d2453b21f56174937c395d6f36\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://99c07a4bdb3059d3a99ffdaf794d668d52cffd9e5e3bb8945204dd6d1f8c8328\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99c07a4bdb3059d3a99ffdaf794d668d52cffd9e5e3bb8945204dd6d1f8c8328\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27095272bf4f31d63a78a436636de0aae12f7103597da944f24ed7a74329ee3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27095272bf4f31d63a78a436636de0aae12f7103597da944f24ed7a74329ee3f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1211571af66d8a94302b65feccf30768ba82b7cbad918e44dd405787dce484cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1211571af66d8a94302b65feccf30768ba82b7cbad918e44dd405787dce484cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27631f90d528900fa180c15cfce80e96d4cdba07743eb6d847e2e822d9917218\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27631f90d528900fa180c15cfce80e96d4cdba07743eb6d847e2e822d9917218\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f345693e9aeb1f4dc196349d3354f3c833a2b0415a693298b1cfccfa2b0d4459\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f345693e9aeb1f4dc196349d3354f3c833a2b0415a693298b1cfccfa2b0d4459\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-sfmxd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:52Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:52 crc kubenswrapper[4846]: I1005 06:48:52.225383 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:52 crc kubenswrapper[4846]: I1005 06:48:52.225498 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:52 crc kubenswrapper[4846]: I1005 06:48:52.225519 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:52 crc kubenswrapper[4846]: I1005 06:48:52.225544 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:52 crc kubenswrapper[4846]: I1005 06:48:52.225568 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:52Z","lastTransitionTime":"2025-10-05T06:48:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:52 crc kubenswrapper[4846]: I1005 06:48:52.253763 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-txzlk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"93a6ace4-8d64-44d7-9637-457d3af6543c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ad1940d07e37217089642478ece9ccdaca74784738a400929a9f5359d158c01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a6f6b2c32dd3da5f9fd5e6971c1773e81a9a24184b7aa5b543cec879001ae53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e9b0219ecbc7e60905d6de3e67a2c8dc9af1017b4aed9538ee0c58c861f2438\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af5e5a4da9ad7371d5a24633e8268f12e3a7d419a8c93b11ddbcc006d8339c6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8da85222775abc7a772f6a8ff8255f1a0d5b15807ba177a7e3e2b2d01a4f24b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1409651de8c897804f87be03d016c138f8acf27e662dce265a93e15c1848adcf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd084bb9806ece6529db0866cee69e6a3152f0da7e0f13fc2e4adcbe2f55962c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd084bb9806ece6529db0866cee69e6a3152f0da7e0f13fc2e4adcbe2f55962c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-05T06:48:50Z\\\",\\\"message\\\":\\\":50.819394 6534 handler.go:208] Removed *v1.Node event handler 2\\\\nI1005 06:48:50.819434 6534 handler.go:208] Removed *v1.Node event handler 7\\\\nI1005 06:48:50.819451 6534 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1005 06:48:50.819624 6534 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1005 06:48:50.819659 6534 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1005 06:48:50.819711 6534 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1005 06:48:50.819735 6534 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1005 06:48:50.820556 6534 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1005 06:48:50.820605 6534 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1005 06:48:50.820649 6534 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1005 06:48:50.820664 6534 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1005 06:48:50.820706 6534 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1005 06:48:50.820711 6534 factory.go:656] Stopping watch factory\\\\nI1005 06:48:50.820731 6534 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1005 06:48:50.820750 6534 ovnkube.go:599] Stopped ovnkube\\\\nI1005 06:48:50.820796 6534 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1005 06:48:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:49Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-txzlk_openshift-ovn-kubernetes(93a6ace4-8d64-44d7-9637-457d3af6543c)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0df0ebcd87dc5af4b17736b43ec7b2395968b19ed16e775366af14e2b2d49c52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff3c6545ff3d2436da27bbd223b950b30b0c69a3e7f3034cb0e61380cb38d338\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ff3c6545ff3d2436da27bbd223b950b30b0c69a3e7f3034cb0e61380cb38d338\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:24Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-txzlk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:52Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:52 crc kubenswrapper[4846]: I1005 06:48:52.273715 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-gklb6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"698ab882-bb26-4b4c-81e9-8305a44fe068\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a6edca88ea244a58f8cf6cc431110a960c6bfd892812cd0d00fa3f97be3bc67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x6d6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c191538ab797b8056820de7da1f8e93ea18873891a32382b8a59c9af762e103\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x6d6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:37Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-gklb6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:52Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:52 crc kubenswrapper[4846]: I1005 06:48:52.296646 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c9f812b-ecc8-4985-a670-53fa80fbb499\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2b5ee9781c01d6aec3ac778efef1067b6c3bfe4ed37431bdebe188fcde0165f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:47:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3fb4717692aa1fec0012b514ef8d18d1ca3707cfa4bb8248bb93841a0a40f11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:47:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0b6a7afe3094f1c92093c05c28344897a615d590c2fbb507c87e18a44c6c831\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://640ade2c733d7916ec213a201a1dd614d261c69eab654c80100efbc99a768995\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:47:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:52Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:52 crc kubenswrapper[4846]: I1005 06:48:52.318409 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:52Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:52 crc kubenswrapper[4846]: I1005 06:48:52.328962 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:52 crc kubenswrapper[4846]: I1005 06:48:52.329032 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:52 crc kubenswrapper[4846]: I1005 06:48:52.329059 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:52 crc kubenswrapper[4846]: I1005 06:48:52.329093 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:52 crc kubenswrapper[4846]: I1005 06:48:52.329118 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:52Z","lastTransitionTime":"2025-10-05T06:48:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:52 crc kubenswrapper[4846]: I1005 06:48:52.340777 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e67d0e447a8a6a381259e25d25badc791194e1291f56b1cd77bd63989b20408\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dkhl7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://516f88f4990f5962d8b694a3d889259cc5508a309919870191c2e1331640c973\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dkhl7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-fscvf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:52Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:52 crc kubenswrapper[4846]: I1005 06:48:52.357305 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-w2sj8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32931fce-7374-4cab-a2e0-79d96a842274\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0df535d65d39de6c9a3dfb931bafc4469adc22b72f7a7903beda4fef8cd48138\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9rtxf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:27Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-w2sj8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:52Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:52 crc kubenswrapper[4846]: I1005 06:48:52.432267 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:52 crc kubenswrapper[4846]: I1005 06:48:52.432309 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:52 crc kubenswrapper[4846]: I1005 06:48:52.432321 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:52 crc kubenswrapper[4846]: I1005 06:48:52.432339 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:52 crc kubenswrapper[4846]: I1005 06:48:52.432356 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:52Z","lastTransitionTime":"2025-10-05T06:48:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:52 crc kubenswrapper[4846]: I1005 06:48:52.497429 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 05 06:48:52 crc kubenswrapper[4846]: I1005 06:48:52.497434 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9r2kt" Oct 05 06:48:52 crc kubenswrapper[4846]: E1005 06:48:52.497608 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 05 06:48:52 crc kubenswrapper[4846]: E1005 06:48:52.497753 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9r2kt" podUID="59b83f4f-71f0-4b28-9275-56d5f933ec2c" Oct 05 06:48:52 crc kubenswrapper[4846]: I1005 06:48:52.535075 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:52 crc kubenswrapper[4846]: I1005 06:48:52.535123 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:52 crc kubenswrapper[4846]: I1005 06:48:52.535136 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:52 crc kubenswrapper[4846]: I1005 06:48:52.535151 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:52 crc kubenswrapper[4846]: I1005 06:48:52.535165 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:52Z","lastTransitionTime":"2025-10-05T06:48:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:52 crc kubenswrapper[4846]: I1005 06:48:52.637753 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:52 crc kubenswrapper[4846]: I1005 06:48:52.637820 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:52 crc kubenswrapper[4846]: I1005 06:48:52.637832 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:52 crc kubenswrapper[4846]: I1005 06:48:52.637851 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:52 crc kubenswrapper[4846]: I1005 06:48:52.637862 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:52Z","lastTransitionTime":"2025-10-05T06:48:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:52 crc kubenswrapper[4846]: I1005 06:48:52.740101 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:52 crc kubenswrapper[4846]: I1005 06:48:52.740137 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:52 crc kubenswrapper[4846]: I1005 06:48:52.740145 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:52 crc kubenswrapper[4846]: I1005 06:48:52.740161 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:52 crc kubenswrapper[4846]: I1005 06:48:52.740171 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:52Z","lastTransitionTime":"2025-10-05T06:48:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:52 crc kubenswrapper[4846]: I1005 06:48:52.842286 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:52 crc kubenswrapper[4846]: I1005 06:48:52.842333 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:52 crc kubenswrapper[4846]: I1005 06:48:52.842345 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:52 crc kubenswrapper[4846]: I1005 06:48:52.842362 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:52 crc kubenswrapper[4846]: I1005 06:48:52.842375 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:52Z","lastTransitionTime":"2025-10-05T06:48:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:52 crc kubenswrapper[4846]: I1005 06:48:52.944692 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:52 crc kubenswrapper[4846]: I1005 06:48:52.944753 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:52 crc kubenswrapper[4846]: I1005 06:48:52.944775 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:52 crc kubenswrapper[4846]: I1005 06:48:52.944795 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:52 crc kubenswrapper[4846]: I1005 06:48:52.944809 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:52Z","lastTransitionTime":"2025-10-05T06:48:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:53 crc kubenswrapper[4846]: I1005 06:48:53.047385 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:53 crc kubenswrapper[4846]: I1005 06:48:53.047474 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:53 crc kubenswrapper[4846]: I1005 06:48:53.047492 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:53 crc kubenswrapper[4846]: I1005 06:48:53.047519 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:53 crc kubenswrapper[4846]: I1005 06:48:53.047537 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:53Z","lastTransitionTime":"2025-10-05T06:48:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:53 crc kubenswrapper[4846]: I1005 06:48:53.150022 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:53 crc kubenswrapper[4846]: I1005 06:48:53.150096 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:53 crc kubenswrapper[4846]: I1005 06:48:53.150114 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:53 crc kubenswrapper[4846]: I1005 06:48:53.150139 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:53 crc kubenswrapper[4846]: I1005 06:48:53.150156 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:53Z","lastTransitionTime":"2025-10-05T06:48:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:53 crc kubenswrapper[4846]: I1005 06:48:53.253431 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:53 crc kubenswrapper[4846]: I1005 06:48:53.253494 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:53 crc kubenswrapper[4846]: I1005 06:48:53.253510 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:53 crc kubenswrapper[4846]: I1005 06:48:53.253533 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:53 crc kubenswrapper[4846]: I1005 06:48:53.253552 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:53Z","lastTransitionTime":"2025-10-05T06:48:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:53 crc kubenswrapper[4846]: I1005 06:48:53.357578 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:53 crc kubenswrapper[4846]: I1005 06:48:53.357641 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:53 crc kubenswrapper[4846]: I1005 06:48:53.357653 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:53 crc kubenswrapper[4846]: I1005 06:48:53.357673 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:53 crc kubenswrapper[4846]: I1005 06:48:53.357732 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:53Z","lastTransitionTime":"2025-10-05T06:48:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:53 crc kubenswrapper[4846]: I1005 06:48:53.460740 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:53 crc kubenswrapper[4846]: I1005 06:48:53.460789 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:53 crc kubenswrapper[4846]: I1005 06:48:53.460799 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:53 crc kubenswrapper[4846]: I1005 06:48:53.460817 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:53 crc kubenswrapper[4846]: I1005 06:48:53.460828 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:53Z","lastTransitionTime":"2025-10-05T06:48:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:53 crc kubenswrapper[4846]: I1005 06:48:53.496965 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 05 06:48:53 crc kubenswrapper[4846]: I1005 06:48:53.497000 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 05 06:48:53 crc kubenswrapper[4846]: E1005 06:48:53.497133 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 05 06:48:53 crc kubenswrapper[4846]: E1005 06:48:53.497348 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 05 06:48:53 crc kubenswrapper[4846]: I1005 06:48:53.563748 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:53 crc kubenswrapper[4846]: I1005 06:48:53.563789 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:53 crc kubenswrapper[4846]: I1005 06:48:53.563797 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:53 crc kubenswrapper[4846]: I1005 06:48:53.563815 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:53 crc kubenswrapper[4846]: I1005 06:48:53.563829 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:53Z","lastTransitionTime":"2025-10-05T06:48:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:53 crc kubenswrapper[4846]: I1005 06:48:53.666569 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:53 crc kubenswrapper[4846]: I1005 06:48:53.666646 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:53 crc kubenswrapper[4846]: I1005 06:48:53.666669 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:53 crc kubenswrapper[4846]: I1005 06:48:53.666695 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:53 crc kubenswrapper[4846]: I1005 06:48:53.666713 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:53Z","lastTransitionTime":"2025-10-05T06:48:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:53 crc kubenswrapper[4846]: I1005 06:48:53.770558 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:53 crc kubenswrapper[4846]: I1005 06:48:53.770633 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:53 crc kubenswrapper[4846]: I1005 06:48:53.770653 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:53 crc kubenswrapper[4846]: I1005 06:48:53.770683 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:53 crc kubenswrapper[4846]: I1005 06:48:53.770709 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:53Z","lastTransitionTime":"2025-10-05T06:48:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:53 crc kubenswrapper[4846]: I1005 06:48:53.873370 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:53 crc kubenswrapper[4846]: I1005 06:48:53.873431 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:53 crc kubenswrapper[4846]: I1005 06:48:53.873447 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:53 crc kubenswrapper[4846]: I1005 06:48:53.873471 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:53 crc kubenswrapper[4846]: I1005 06:48:53.873488 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:53Z","lastTransitionTime":"2025-10-05T06:48:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:53 crc kubenswrapper[4846]: I1005 06:48:53.976685 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:53 crc kubenswrapper[4846]: I1005 06:48:53.976737 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:53 crc kubenswrapper[4846]: I1005 06:48:53.976750 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:53 crc kubenswrapper[4846]: I1005 06:48:53.976773 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:53 crc kubenswrapper[4846]: I1005 06:48:53.976788 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:53Z","lastTransitionTime":"2025-10-05T06:48:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:54 crc kubenswrapper[4846]: I1005 06:48:54.080065 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:54 crc kubenswrapper[4846]: I1005 06:48:54.080114 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:54 crc kubenswrapper[4846]: I1005 06:48:54.080126 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:54 crc kubenswrapper[4846]: I1005 06:48:54.080143 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:54 crc kubenswrapper[4846]: I1005 06:48:54.080156 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:54Z","lastTransitionTime":"2025-10-05T06:48:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:54 crc kubenswrapper[4846]: I1005 06:48:54.183355 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:54 crc kubenswrapper[4846]: I1005 06:48:54.183424 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:54 crc kubenswrapper[4846]: I1005 06:48:54.183443 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:54 crc kubenswrapper[4846]: I1005 06:48:54.183468 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:54 crc kubenswrapper[4846]: I1005 06:48:54.183490 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:54Z","lastTransitionTime":"2025-10-05T06:48:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:54 crc kubenswrapper[4846]: I1005 06:48:54.286086 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:54 crc kubenswrapper[4846]: I1005 06:48:54.286310 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:54 crc kubenswrapper[4846]: I1005 06:48:54.286336 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:54 crc kubenswrapper[4846]: I1005 06:48:54.286421 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:54 crc kubenswrapper[4846]: I1005 06:48:54.286447 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:54Z","lastTransitionTime":"2025-10-05T06:48:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:54 crc kubenswrapper[4846]: I1005 06:48:54.389219 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:54 crc kubenswrapper[4846]: I1005 06:48:54.389273 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:54 crc kubenswrapper[4846]: I1005 06:48:54.389286 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:54 crc kubenswrapper[4846]: I1005 06:48:54.389304 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:54 crc kubenswrapper[4846]: I1005 06:48:54.389319 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:54Z","lastTransitionTime":"2025-10-05T06:48:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:54 crc kubenswrapper[4846]: I1005 06:48:54.491594 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:54 crc kubenswrapper[4846]: I1005 06:48:54.491664 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:54 crc kubenswrapper[4846]: I1005 06:48:54.491683 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:54 crc kubenswrapper[4846]: I1005 06:48:54.491713 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:54 crc kubenswrapper[4846]: I1005 06:48:54.491739 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:54Z","lastTransitionTime":"2025-10-05T06:48:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:54 crc kubenswrapper[4846]: I1005 06:48:54.496766 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 05 06:48:54 crc kubenswrapper[4846]: I1005 06:48:54.496790 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9r2kt" Oct 05 06:48:54 crc kubenswrapper[4846]: E1005 06:48:54.496936 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 05 06:48:54 crc kubenswrapper[4846]: E1005 06:48:54.497050 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9r2kt" podUID="59b83f4f-71f0-4b28-9275-56d5f933ec2c" Oct 05 06:48:54 crc kubenswrapper[4846]: I1005 06:48:54.594530 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:54 crc kubenswrapper[4846]: I1005 06:48:54.594585 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:54 crc kubenswrapper[4846]: I1005 06:48:54.594597 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:54 crc kubenswrapper[4846]: I1005 06:48:54.594615 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:54 crc kubenswrapper[4846]: I1005 06:48:54.594629 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:54Z","lastTransitionTime":"2025-10-05T06:48:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:54 crc kubenswrapper[4846]: I1005 06:48:54.697037 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:54 crc kubenswrapper[4846]: I1005 06:48:54.697095 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:54 crc kubenswrapper[4846]: I1005 06:48:54.697104 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:54 crc kubenswrapper[4846]: I1005 06:48:54.697117 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:54 crc kubenswrapper[4846]: I1005 06:48:54.697127 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:54Z","lastTransitionTime":"2025-10-05T06:48:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:54 crc kubenswrapper[4846]: I1005 06:48:54.766366 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/59b83f4f-71f0-4b28-9275-56d5f933ec2c-metrics-certs\") pod \"network-metrics-daemon-9r2kt\" (UID: \"59b83f4f-71f0-4b28-9275-56d5f933ec2c\") " pod="openshift-multus/network-metrics-daemon-9r2kt" Oct 05 06:48:54 crc kubenswrapper[4846]: E1005 06:48:54.766595 4846 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 05 06:48:54 crc kubenswrapper[4846]: E1005 06:48:54.766754 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/59b83f4f-71f0-4b28-9275-56d5f933ec2c-metrics-certs podName:59b83f4f-71f0-4b28-9275-56d5f933ec2c nodeName:}" failed. No retries permitted until 2025-10-05 06:49:10.766731211 +0000 UTC m=+73.007583986 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/59b83f4f-71f0-4b28-9275-56d5f933ec2c-metrics-certs") pod "network-metrics-daemon-9r2kt" (UID: "59b83f4f-71f0-4b28-9275-56d5f933ec2c") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 05 06:48:54 crc kubenswrapper[4846]: I1005 06:48:54.799833 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:54 crc kubenswrapper[4846]: I1005 06:48:54.799879 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:54 crc kubenswrapper[4846]: I1005 06:48:54.799888 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:54 crc kubenswrapper[4846]: I1005 06:48:54.799909 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:54 crc kubenswrapper[4846]: I1005 06:48:54.799919 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:54Z","lastTransitionTime":"2025-10-05T06:48:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:54 crc kubenswrapper[4846]: I1005 06:48:54.902196 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:54 crc kubenswrapper[4846]: I1005 06:48:54.902243 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:54 crc kubenswrapper[4846]: I1005 06:48:54.902255 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:54 crc kubenswrapper[4846]: I1005 06:48:54.902275 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:54 crc kubenswrapper[4846]: I1005 06:48:54.902290 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:54Z","lastTransitionTime":"2025-10-05T06:48:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:54 crc kubenswrapper[4846]: I1005 06:48:54.987102 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 05 06:48:54 crc kubenswrapper[4846]: I1005 06:48:54.997038 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Oct 05 06:48:55 crc kubenswrapper[4846]: I1005 06:48:55.005343 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:55 crc kubenswrapper[4846]: I1005 06:48:55.005379 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:55 crc kubenswrapper[4846]: I1005 06:48:55.005392 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:55 crc kubenswrapper[4846]: I1005 06:48:55.005408 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:55 crc kubenswrapper[4846]: I1005 06:48:55.005420 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:55Z","lastTransitionTime":"2025-10-05T06:48:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:55 crc kubenswrapper[4846]: I1005 06:48:55.007355 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e67d0e447a8a6a381259e25d25badc791194e1291f56b1cd77bd63989b20408\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dkhl7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://516f88f4990f5962d8b694a3d889259cc5508a309919870191c2e1331640c973\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dkhl7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-fscvf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:55Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:55 crc kubenswrapper[4846]: I1005 06:48:55.019945 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-w2sj8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32931fce-7374-4cab-a2e0-79d96a842274\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0df535d65d39de6c9a3dfb931bafc4469adc22b72f7a7903beda4fef8cd48138\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9rtxf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:27Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-w2sj8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:55Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:55 crc kubenswrapper[4846]: I1005 06:48:55.034498 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c9f812b-ecc8-4985-a670-53fa80fbb499\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2b5ee9781c01d6aec3ac778efef1067b6c3bfe4ed37431bdebe188fcde0165f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:47:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3fb4717692aa1fec0012b514ef8d18d1ca3707cfa4bb8248bb93841a0a40f11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:47:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0b6a7afe3094f1c92093c05c28344897a615d590c2fbb507c87e18a44c6c831\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://640ade2c733d7916ec213a201a1dd614d261c69eab654c80100efbc99a768995\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:47:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:55Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:55 crc kubenswrapper[4846]: I1005 06:48:55.051365 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:55Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:55 crc kubenswrapper[4846]: I1005 06:48:55.078979 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://347f900743b7bba3daf92f21c7eba0625e8c774f50057bc9472f0c82119325e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:55Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:55 crc kubenswrapper[4846]: I1005 06:48:55.102265 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56f1fe1f9f06f0249cf937a6963fda82467732dc056ca2b2189be6d8dd9d4a05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7212ce5f42c1fb55146abb190a557a504b43f16af6df84f90e1073567a329ec5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:55Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:55 crc kubenswrapper[4846]: I1005 06:48:55.107573 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:55 crc kubenswrapper[4846]: I1005 06:48:55.107610 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:55 crc kubenswrapper[4846]: I1005 06:48:55.107620 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:55 crc kubenswrapper[4846]: I1005 06:48:55.107634 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:55 crc kubenswrapper[4846]: I1005 06:48:55.107645 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:55Z","lastTransitionTime":"2025-10-05T06:48:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:55 crc kubenswrapper[4846]: I1005 06:48:55.136679 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-7wm5q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9bc7744f-d700-450f-8b8d-98140877fee9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90379e74c544fccca44f7acdc228ff378154390e211ac801b79f8d074012241e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h5p4q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-7wm5q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:55Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:55 crc kubenswrapper[4846]: I1005 06:48:55.153558 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ff10c31459b3b3724ee7b769bd6f72f9ed80cce5419b6adcdf69a3239b70756\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:55Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:55 crc kubenswrapper[4846]: I1005 06:48:55.172139 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:55Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:55 crc kubenswrapper[4846]: I1005 06:48:55.181867 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-r7ztb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a5119cd9-af70-4324-a955-978306ab0b20\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06231a87429ec292cb57ebd88cda7ad547bcb32e80343b776833621a11c28ad1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jfcj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-r7ztb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:55Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:55 crc kubenswrapper[4846]: I1005 06:48:55.193560 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-9r2kt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"59b83f4f-71f0-4b28-9275-56d5f933ec2c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tdwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tdwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:38Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-9r2kt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:55Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:55 crc kubenswrapper[4846]: I1005 06:48:55.205899 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"775b5702-a131-4d6c-852d-91bfa056cac5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0839c1765df2c68d2554e9c66d289505e559a9672caf8069feeff2c4fdfd2171\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15dbaf2445704da0987f256f81dff2ee95664988541fa48ce94898afeadab2e8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd3b3ab323c6d5153054aa04af2d91f7b9e8436ff1e2f75f3594b57831b2a416\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://67c46f5f0a3ace95bd2e9479c5561126664e18b148c5eceea5bf5efc27e81a67\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c52e076eddfd307110b1f6de00a169b81af8f2785af7becef2a820ce363a5861\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-05T06:48:14Z\\\",\\\"message\\\":\\\"le observer\\\\nW1005 06:48:14.337870 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1005 06:48:14.338086 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1005 06:48:14.339389 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1467752215/tls.crt::/tmp/serving-cert-1467752215/tls.key\\\\\\\"\\\\nI1005 06:48:14.864345 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1005 06:48:14.868087 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1005 06:48:14.868114 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1005 06:48:14.868140 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1005 06:48:14.868147 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1005 06:48:14.877354 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1005 06:48:14.877412 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1005 06:48:14.877456 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1005 06:48:14.877471 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1005 06:48:14.877490 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1005 06:48:14.877501 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1005 06:48:14.877507 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1005 06:48:14.877515 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1005 06:48:14.880326 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:13Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4a091716ff935adb42ae2e88048619fd347bd174f64e5f9c97a01f81690e57f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:01Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dc5411b642f99b13d514d2f21defad0c394aeeb82a966f72300076617e2586c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5dc5411b642f99b13d514d2f21defad0c394aeeb82a966f72300076617e2586c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:47:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:47:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:47:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:55Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:55 crc kubenswrapper[4846]: I1005 06:48:55.209919 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:55 crc kubenswrapper[4846]: I1005 06:48:55.209951 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:55 crc kubenswrapper[4846]: I1005 06:48:55.209959 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:55 crc kubenswrapper[4846]: I1005 06:48:55.209974 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:55 crc kubenswrapper[4846]: I1005 06:48:55.209984 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:55Z","lastTransitionTime":"2025-10-05T06:48:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:55 crc kubenswrapper[4846]: I1005 06:48:55.220505 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-sfmxd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1dbb13ac-136c-4a1c-9229-b4e8ac4f9c71\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b0af6a6e735f7b7877c177c1d53d1a8dc12c4713a8059c4766fc858af31c9a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://226d977b596d07aeda0486452e2a8605949ed4d2453b21f56174937c395d6f36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://226d977b596d07aeda0486452e2a8605949ed4d2453b21f56174937c395d6f36\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://99c07a4bdb3059d3a99ffdaf794d668d52cffd9e5e3bb8945204dd6d1f8c8328\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99c07a4bdb3059d3a99ffdaf794d668d52cffd9e5e3bb8945204dd6d1f8c8328\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27095272bf4f31d63a78a436636de0aae12f7103597da944f24ed7a74329ee3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27095272bf4f31d63a78a436636de0aae12f7103597da944f24ed7a74329ee3f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1211571af66d8a94302b65feccf30768ba82b7cbad918e44dd405787dce484cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1211571af66d8a94302b65feccf30768ba82b7cbad918e44dd405787dce484cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27631f90d528900fa180c15cfce80e96d4cdba07743eb6d847e2e822d9917218\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27631f90d528900fa180c15cfce80e96d4cdba07743eb6d847e2e822d9917218\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f345693e9aeb1f4dc196349d3354f3c833a2b0415a693298b1cfccfa2b0d4459\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f345693e9aeb1f4dc196349d3354f3c833a2b0415a693298b1cfccfa2b0d4459\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-sfmxd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:55Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:55 crc kubenswrapper[4846]: I1005 06:48:55.237142 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-txzlk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"93a6ace4-8d64-44d7-9637-457d3af6543c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ad1940d07e37217089642478ece9ccdaca74784738a400929a9f5359d158c01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a6f6b2c32dd3da5f9fd5e6971c1773e81a9a24184b7aa5b543cec879001ae53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e9b0219ecbc7e60905d6de3e67a2c8dc9af1017b4aed9538ee0c58c861f2438\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af5e5a4da9ad7371d5a24633e8268f12e3a7d419a8c93b11ddbcc006d8339c6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8da85222775abc7a772f6a8ff8255f1a0d5b15807ba177a7e3e2b2d01a4f24b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1409651de8c897804f87be03d016c138f8acf27e662dce265a93e15c1848adcf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd084bb9806ece6529db0866cee69e6a3152f0da7e0f13fc2e4adcbe2f55962c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd084bb9806ece6529db0866cee69e6a3152f0da7e0f13fc2e4adcbe2f55962c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-05T06:48:50Z\\\",\\\"message\\\":\\\":50.819394 6534 handler.go:208] Removed *v1.Node event handler 2\\\\nI1005 06:48:50.819434 6534 handler.go:208] Removed *v1.Node event handler 7\\\\nI1005 06:48:50.819451 6534 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1005 06:48:50.819624 6534 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1005 06:48:50.819659 6534 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1005 06:48:50.819711 6534 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1005 06:48:50.819735 6534 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1005 06:48:50.820556 6534 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1005 06:48:50.820605 6534 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1005 06:48:50.820649 6534 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1005 06:48:50.820664 6534 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1005 06:48:50.820706 6534 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1005 06:48:50.820711 6534 factory.go:656] Stopping watch factory\\\\nI1005 06:48:50.820731 6534 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1005 06:48:50.820750 6534 ovnkube.go:599] Stopped ovnkube\\\\nI1005 06:48:50.820796 6534 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1005 06:48:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:49Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-txzlk_openshift-ovn-kubernetes(93a6ace4-8d64-44d7-9637-457d3af6543c)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0df0ebcd87dc5af4b17736b43ec7b2395968b19ed16e775366af14e2b2d49c52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff3c6545ff3d2436da27bbd223b950b30b0c69a3e7f3034cb0e61380cb38d338\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ff3c6545ff3d2436da27bbd223b950b30b0c69a3e7f3034cb0e61380cb38d338\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:24Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-txzlk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:55Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:55 crc kubenswrapper[4846]: I1005 06:48:55.247350 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-gklb6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"698ab882-bb26-4b4c-81e9-8305a44fe068\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a6edca88ea244a58f8cf6cc431110a960c6bfd892812cd0d00fa3f97be3bc67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x6d6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c191538ab797b8056820de7da1f8e93ea18873891a32382b8a59c9af762e103\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x6d6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:37Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-gklb6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:55Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:55 crc kubenswrapper[4846]: I1005 06:48:55.266046 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c482bb97-fd50-40cf-96ea-f806c3f01125\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bfe45d30e4fa141541aca1d46bc5cc1da0a7b4a5eb1fc9bb2caa5b7fdcb0b264\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://496a4782faa30a9a20ff63038141600cd635f12902d1cf17989cd067fcdfe929\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://281be077869a6aa1ca815770871a14413d4a97e97aeb550fd5cd215cba88b21f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6771c7215d2f5dd5838d17888375a4768b9d1755fc2e21bda1bbfbb2eb1dbb95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c17a711c2c7473b076a3944ffe7c700808d05cba582fbd55c5c2ac8ff73838a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44a9f740f0a06c8bf928e2a9d1d5979217558c31e73a333631baccebf00fd5d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://44a9f740f0a06c8bf928e2a9d1d5979217558c31e73a333631baccebf00fd5d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:47:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:47:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://48ca0ba6ca6092bfcb5b153220c054f02cc36fb22707c42e021f06d329e18e2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48ca0ba6ca6092bfcb5b153220c054f02cc36fb22707c42e021f06d329e18e2e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:00Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4006d0cea128675c8885d879171204abffc5e11ea0664750a12b4bc7344ccb2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4006d0cea128675c8885d879171204abffc5e11ea0664750a12b4bc7344ccb2e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:47:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:55Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:55 crc kubenswrapper[4846]: I1005 06:48:55.280684 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:55Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:55 crc kubenswrapper[4846]: I1005 06:48:55.313597 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:55 crc kubenswrapper[4846]: I1005 06:48:55.313665 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:55 crc kubenswrapper[4846]: I1005 06:48:55.313681 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:55 crc kubenswrapper[4846]: I1005 06:48:55.313705 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:55 crc kubenswrapper[4846]: I1005 06:48:55.313723 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:55Z","lastTransitionTime":"2025-10-05T06:48:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:55 crc kubenswrapper[4846]: I1005 06:48:55.384667 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:55 crc kubenswrapper[4846]: I1005 06:48:55.384714 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:55 crc kubenswrapper[4846]: I1005 06:48:55.384725 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:55 crc kubenswrapper[4846]: I1005 06:48:55.384748 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:55 crc kubenswrapper[4846]: I1005 06:48:55.384759 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:55Z","lastTransitionTime":"2025-10-05T06:48:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:55 crc kubenswrapper[4846]: E1005 06:48:55.401020 4846 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:48:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:48:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:55Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:48:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:48:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:55Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"04ce628d-2f61-4931-aad6-615924ea53fb\\\",\\\"systemUUID\\\":\\\"a050a23b-e773-4ba9-989b-360aa53a0605\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:55Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:55 crc kubenswrapper[4846]: I1005 06:48:55.405923 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:55 crc kubenswrapper[4846]: I1005 06:48:55.405974 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:55 crc kubenswrapper[4846]: I1005 06:48:55.405990 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:55 crc kubenswrapper[4846]: I1005 06:48:55.406015 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:55 crc kubenswrapper[4846]: I1005 06:48:55.406033 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:55Z","lastTransitionTime":"2025-10-05T06:48:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:55 crc kubenswrapper[4846]: E1005 06:48:55.424378 4846 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:48:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:48:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:55Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:48:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:48:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:55Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"04ce628d-2f61-4931-aad6-615924ea53fb\\\",\\\"systemUUID\\\":\\\"a050a23b-e773-4ba9-989b-360aa53a0605\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:55Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:55 crc kubenswrapper[4846]: I1005 06:48:55.429026 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:55 crc kubenswrapper[4846]: I1005 06:48:55.429079 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:55 crc kubenswrapper[4846]: I1005 06:48:55.429096 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:55 crc kubenswrapper[4846]: I1005 06:48:55.429120 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:55 crc kubenswrapper[4846]: I1005 06:48:55.429137 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:55Z","lastTransitionTime":"2025-10-05T06:48:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:55 crc kubenswrapper[4846]: E1005 06:48:55.446985 4846 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:48:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:48:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:55Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:48:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:48:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:55Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"04ce628d-2f61-4931-aad6-615924ea53fb\\\",\\\"systemUUID\\\":\\\"a050a23b-e773-4ba9-989b-360aa53a0605\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:55Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:55 crc kubenswrapper[4846]: I1005 06:48:55.452014 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:55 crc kubenswrapper[4846]: I1005 06:48:55.452053 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:55 crc kubenswrapper[4846]: I1005 06:48:55.452068 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:55 crc kubenswrapper[4846]: I1005 06:48:55.452087 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:55 crc kubenswrapper[4846]: I1005 06:48:55.452099 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:55Z","lastTransitionTime":"2025-10-05T06:48:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:55 crc kubenswrapper[4846]: E1005 06:48:55.468873 4846 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:48:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:48:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:55Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:48:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:48:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:55Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"04ce628d-2f61-4931-aad6-615924ea53fb\\\",\\\"systemUUID\\\":\\\"a050a23b-e773-4ba9-989b-360aa53a0605\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:55Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:55 crc kubenswrapper[4846]: I1005 06:48:55.472584 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:55 crc kubenswrapper[4846]: I1005 06:48:55.472621 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:55 crc kubenswrapper[4846]: I1005 06:48:55.472631 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:55 crc kubenswrapper[4846]: I1005 06:48:55.472650 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:55 crc kubenswrapper[4846]: I1005 06:48:55.472662 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:55Z","lastTransitionTime":"2025-10-05T06:48:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:55 crc kubenswrapper[4846]: E1005 06:48:55.489325 4846 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:48:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:48:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:55Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:48:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:48:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:55Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"04ce628d-2f61-4931-aad6-615924ea53fb\\\",\\\"systemUUID\\\":\\\"a050a23b-e773-4ba9-989b-360aa53a0605\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:55Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:55 crc kubenswrapper[4846]: E1005 06:48:55.489485 4846 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 05 06:48:55 crc kubenswrapper[4846]: I1005 06:48:55.491078 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:55 crc kubenswrapper[4846]: I1005 06:48:55.491116 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:55 crc kubenswrapper[4846]: I1005 06:48:55.491129 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:55 crc kubenswrapper[4846]: I1005 06:48:55.491147 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:55 crc kubenswrapper[4846]: I1005 06:48:55.491161 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:55Z","lastTransitionTime":"2025-10-05T06:48:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:55 crc kubenswrapper[4846]: I1005 06:48:55.496403 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 05 06:48:55 crc kubenswrapper[4846]: I1005 06:48:55.496483 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 05 06:48:55 crc kubenswrapper[4846]: E1005 06:48:55.496512 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 05 06:48:55 crc kubenswrapper[4846]: E1005 06:48:55.496660 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 05 06:48:55 crc kubenswrapper[4846]: I1005 06:48:55.594490 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:55 crc kubenswrapper[4846]: I1005 06:48:55.594547 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:55 crc kubenswrapper[4846]: I1005 06:48:55.594557 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:55 crc kubenswrapper[4846]: I1005 06:48:55.594576 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:55 crc kubenswrapper[4846]: I1005 06:48:55.594587 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:55Z","lastTransitionTime":"2025-10-05T06:48:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:55 crc kubenswrapper[4846]: I1005 06:48:55.697604 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:55 crc kubenswrapper[4846]: I1005 06:48:55.697682 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:55 crc kubenswrapper[4846]: I1005 06:48:55.697708 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:55 crc kubenswrapper[4846]: I1005 06:48:55.697751 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:55 crc kubenswrapper[4846]: I1005 06:48:55.697774 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:55Z","lastTransitionTime":"2025-10-05T06:48:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:55 crc kubenswrapper[4846]: I1005 06:48:55.801152 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:55 crc kubenswrapper[4846]: I1005 06:48:55.801269 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:55 crc kubenswrapper[4846]: I1005 06:48:55.801288 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:55 crc kubenswrapper[4846]: I1005 06:48:55.801314 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:55 crc kubenswrapper[4846]: I1005 06:48:55.801333 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:55Z","lastTransitionTime":"2025-10-05T06:48:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:55 crc kubenswrapper[4846]: I1005 06:48:55.903960 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:55 crc kubenswrapper[4846]: I1005 06:48:55.904014 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:55 crc kubenswrapper[4846]: I1005 06:48:55.904027 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:55 crc kubenswrapper[4846]: I1005 06:48:55.904048 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:55 crc kubenswrapper[4846]: I1005 06:48:55.904063 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:55Z","lastTransitionTime":"2025-10-05T06:48:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:56 crc kubenswrapper[4846]: I1005 06:48:56.007435 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:56 crc kubenswrapper[4846]: I1005 06:48:56.007526 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:56 crc kubenswrapper[4846]: I1005 06:48:56.007550 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:56 crc kubenswrapper[4846]: I1005 06:48:56.007589 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:56 crc kubenswrapper[4846]: I1005 06:48:56.007609 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:56Z","lastTransitionTime":"2025-10-05T06:48:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:56 crc kubenswrapper[4846]: I1005 06:48:56.110938 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:56 crc kubenswrapper[4846]: I1005 06:48:56.110981 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:56 crc kubenswrapper[4846]: I1005 06:48:56.110993 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:56 crc kubenswrapper[4846]: I1005 06:48:56.111010 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:56 crc kubenswrapper[4846]: I1005 06:48:56.111022 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:56Z","lastTransitionTime":"2025-10-05T06:48:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:56 crc kubenswrapper[4846]: I1005 06:48:56.214686 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:56 crc kubenswrapper[4846]: I1005 06:48:56.214752 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:56 crc kubenswrapper[4846]: I1005 06:48:56.214771 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:56 crc kubenswrapper[4846]: I1005 06:48:56.214801 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:56 crc kubenswrapper[4846]: I1005 06:48:56.214822 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:56Z","lastTransitionTime":"2025-10-05T06:48:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:56 crc kubenswrapper[4846]: I1005 06:48:56.318685 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:56 crc kubenswrapper[4846]: I1005 06:48:56.318756 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:56 crc kubenswrapper[4846]: I1005 06:48:56.318777 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:56 crc kubenswrapper[4846]: I1005 06:48:56.318809 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:56 crc kubenswrapper[4846]: I1005 06:48:56.318828 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:56Z","lastTransitionTime":"2025-10-05T06:48:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:56 crc kubenswrapper[4846]: I1005 06:48:56.422754 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:56 crc kubenswrapper[4846]: I1005 06:48:56.422836 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:56 crc kubenswrapper[4846]: I1005 06:48:56.422862 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:56 crc kubenswrapper[4846]: I1005 06:48:56.422893 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:56 crc kubenswrapper[4846]: I1005 06:48:56.422916 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:56Z","lastTransitionTime":"2025-10-05T06:48:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:56 crc kubenswrapper[4846]: I1005 06:48:56.497363 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 05 06:48:56 crc kubenswrapper[4846]: I1005 06:48:56.497348 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9r2kt" Oct 05 06:48:56 crc kubenswrapper[4846]: E1005 06:48:56.497617 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 05 06:48:56 crc kubenswrapper[4846]: E1005 06:48:56.497779 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9r2kt" podUID="59b83f4f-71f0-4b28-9275-56d5f933ec2c" Oct 05 06:48:56 crc kubenswrapper[4846]: I1005 06:48:56.526284 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:56 crc kubenswrapper[4846]: I1005 06:48:56.526330 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:56 crc kubenswrapper[4846]: I1005 06:48:56.526347 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:56 crc kubenswrapper[4846]: I1005 06:48:56.526372 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:56 crc kubenswrapper[4846]: I1005 06:48:56.526391 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:56Z","lastTransitionTime":"2025-10-05T06:48:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:56 crc kubenswrapper[4846]: I1005 06:48:56.629141 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:56 crc kubenswrapper[4846]: I1005 06:48:56.629260 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:56 crc kubenswrapper[4846]: I1005 06:48:56.629280 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:56 crc kubenswrapper[4846]: I1005 06:48:56.629302 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:56 crc kubenswrapper[4846]: I1005 06:48:56.629320 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:56Z","lastTransitionTime":"2025-10-05T06:48:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:56 crc kubenswrapper[4846]: I1005 06:48:56.732079 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:56 crc kubenswrapper[4846]: I1005 06:48:56.732143 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:56 crc kubenswrapper[4846]: I1005 06:48:56.732161 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:56 crc kubenswrapper[4846]: I1005 06:48:56.732227 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:56 crc kubenswrapper[4846]: I1005 06:48:56.732251 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:56Z","lastTransitionTime":"2025-10-05T06:48:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:56 crc kubenswrapper[4846]: I1005 06:48:56.835866 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:56 crc kubenswrapper[4846]: I1005 06:48:56.835936 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:56 crc kubenswrapper[4846]: I1005 06:48:56.835954 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:56 crc kubenswrapper[4846]: I1005 06:48:56.835983 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:56 crc kubenswrapper[4846]: I1005 06:48:56.836004 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:56Z","lastTransitionTime":"2025-10-05T06:48:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:56 crc kubenswrapper[4846]: I1005 06:48:56.939811 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:56 crc kubenswrapper[4846]: I1005 06:48:56.939874 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:56 crc kubenswrapper[4846]: I1005 06:48:56.939890 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:56 crc kubenswrapper[4846]: I1005 06:48:56.939913 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:56 crc kubenswrapper[4846]: I1005 06:48:56.939930 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:56Z","lastTransitionTime":"2025-10-05T06:48:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:57 crc kubenswrapper[4846]: I1005 06:48:57.042668 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:57 crc kubenswrapper[4846]: I1005 06:48:57.042999 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:57 crc kubenswrapper[4846]: I1005 06:48:57.043085 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:57 crc kubenswrapper[4846]: I1005 06:48:57.043172 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:57 crc kubenswrapper[4846]: I1005 06:48:57.043267 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:57Z","lastTransitionTime":"2025-10-05T06:48:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:57 crc kubenswrapper[4846]: I1005 06:48:57.145920 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:57 crc kubenswrapper[4846]: I1005 06:48:57.146425 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:57 crc kubenswrapper[4846]: I1005 06:48:57.146587 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:57 crc kubenswrapper[4846]: I1005 06:48:57.146734 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:57 crc kubenswrapper[4846]: I1005 06:48:57.146894 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:57Z","lastTransitionTime":"2025-10-05T06:48:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:57 crc kubenswrapper[4846]: I1005 06:48:57.250074 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:57 crc kubenswrapper[4846]: I1005 06:48:57.250162 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:57 crc kubenswrapper[4846]: I1005 06:48:57.250233 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:57 crc kubenswrapper[4846]: I1005 06:48:57.250266 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:57 crc kubenswrapper[4846]: I1005 06:48:57.250289 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:57Z","lastTransitionTime":"2025-10-05T06:48:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:57 crc kubenswrapper[4846]: I1005 06:48:57.353776 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:57 crc kubenswrapper[4846]: I1005 06:48:57.353821 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:57 crc kubenswrapper[4846]: I1005 06:48:57.353830 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:57 crc kubenswrapper[4846]: I1005 06:48:57.353847 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:57 crc kubenswrapper[4846]: I1005 06:48:57.353857 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:57Z","lastTransitionTime":"2025-10-05T06:48:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:57 crc kubenswrapper[4846]: I1005 06:48:57.457320 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:57 crc kubenswrapper[4846]: I1005 06:48:57.457364 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:57 crc kubenswrapper[4846]: I1005 06:48:57.457372 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:57 crc kubenswrapper[4846]: I1005 06:48:57.457385 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:57 crc kubenswrapper[4846]: I1005 06:48:57.457394 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:57Z","lastTransitionTime":"2025-10-05T06:48:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:57 crc kubenswrapper[4846]: I1005 06:48:57.497290 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 05 06:48:57 crc kubenswrapper[4846]: I1005 06:48:57.497291 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 05 06:48:57 crc kubenswrapper[4846]: E1005 06:48:57.497657 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 05 06:48:57 crc kubenswrapper[4846]: E1005 06:48:57.497513 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 05 06:48:57 crc kubenswrapper[4846]: I1005 06:48:57.560559 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:57 crc kubenswrapper[4846]: I1005 06:48:57.560622 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:57 crc kubenswrapper[4846]: I1005 06:48:57.560644 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:57 crc kubenswrapper[4846]: I1005 06:48:57.560672 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:57 crc kubenswrapper[4846]: I1005 06:48:57.560695 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:57Z","lastTransitionTime":"2025-10-05T06:48:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:57 crc kubenswrapper[4846]: I1005 06:48:57.663739 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:57 crc kubenswrapper[4846]: I1005 06:48:57.663800 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:57 crc kubenswrapper[4846]: I1005 06:48:57.663827 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:57 crc kubenswrapper[4846]: I1005 06:48:57.663871 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:57 crc kubenswrapper[4846]: I1005 06:48:57.663894 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:57Z","lastTransitionTime":"2025-10-05T06:48:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:57 crc kubenswrapper[4846]: I1005 06:48:57.767064 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:57 crc kubenswrapper[4846]: I1005 06:48:57.767103 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:57 crc kubenswrapper[4846]: I1005 06:48:57.767138 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:57 crc kubenswrapper[4846]: I1005 06:48:57.767156 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:57 crc kubenswrapper[4846]: I1005 06:48:57.767168 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:57Z","lastTransitionTime":"2025-10-05T06:48:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:57 crc kubenswrapper[4846]: I1005 06:48:57.870142 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:57 crc kubenswrapper[4846]: I1005 06:48:57.870207 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:57 crc kubenswrapper[4846]: I1005 06:48:57.870218 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:57 crc kubenswrapper[4846]: I1005 06:48:57.870234 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:57 crc kubenswrapper[4846]: I1005 06:48:57.870243 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:57Z","lastTransitionTime":"2025-10-05T06:48:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:57 crc kubenswrapper[4846]: I1005 06:48:57.973661 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:57 crc kubenswrapper[4846]: I1005 06:48:57.973736 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:57 crc kubenswrapper[4846]: I1005 06:48:57.973763 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:57 crc kubenswrapper[4846]: I1005 06:48:57.973793 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:57 crc kubenswrapper[4846]: I1005 06:48:57.973815 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:57Z","lastTransitionTime":"2025-10-05T06:48:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:58 crc kubenswrapper[4846]: I1005 06:48:58.076932 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:58 crc kubenswrapper[4846]: I1005 06:48:58.077000 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:58 crc kubenswrapper[4846]: I1005 06:48:58.077018 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:58 crc kubenswrapper[4846]: I1005 06:48:58.077046 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:58 crc kubenswrapper[4846]: I1005 06:48:58.077071 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:58Z","lastTransitionTime":"2025-10-05T06:48:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:58 crc kubenswrapper[4846]: I1005 06:48:58.180354 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:58 crc kubenswrapper[4846]: I1005 06:48:58.180426 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:58 crc kubenswrapper[4846]: I1005 06:48:58.180449 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:58 crc kubenswrapper[4846]: I1005 06:48:58.180482 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:58 crc kubenswrapper[4846]: I1005 06:48:58.180507 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:58Z","lastTransitionTime":"2025-10-05T06:48:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:58 crc kubenswrapper[4846]: I1005 06:48:58.284601 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:58 crc kubenswrapper[4846]: I1005 06:48:58.284660 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:58 crc kubenswrapper[4846]: I1005 06:48:58.284677 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:58 crc kubenswrapper[4846]: I1005 06:48:58.284696 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:58 crc kubenswrapper[4846]: I1005 06:48:58.284709 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:58Z","lastTransitionTime":"2025-10-05T06:48:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:58 crc kubenswrapper[4846]: I1005 06:48:58.387660 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:58 crc kubenswrapper[4846]: I1005 06:48:58.387722 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:58 crc kubenswrapper[4846]: I1005 06:48:58.387739 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:58 crc kubenswrapper[4846]: I1005 06:48:58.387764 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:58 crc kubenswrapper[4846]: I1005 06:48:58.387783 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:58Z","lastTransitionTime":"2025-10-05T06:48:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:58 crc kubenswrapper[4846]: I1005 06:48:58.491939 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:58 crc kubenswrapper[4846]: I1005 06:48:58.492309 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:58 crc kubenswrapper[4846]: I1005 06:48:58.492328 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:58 crc kubenswrapper[4846]: I1005 06:48:58.492356 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:58 crc kubenswrapper[4846]: I1005 06:48:58.492454 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:58Z","lastTransitionTime":"2025-10-05T06:48:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:58 crc kubenswrapper[4846]: I1005 06:48:58.496867 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9r2kt" Oct 05 06:48:58 crc kubenswrapper[4846]: I1005 06:48:58.496899 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 05 06:48:58 crc kubenswrapper[4846]: E1005 06:48:58.497144 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9r2kt" podUID="59b83f4f-71f0-4b28-9275-56d5f933ec2c" Oct 05 06:48:58 crc kubenswrapper[4846]: E1005 06:48:58.497331 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 05 06:48:58 crc kubenswrapper[4846]: I1005 06:48:58.516494 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-7wm5q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9bc7744f-d700-450f-8b8d-98140877fee9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90379e74c544fccca44f7acdc228ff378154390e211ac801b79f8d074012241e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h5p4q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-7wm5q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:58Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:58 crc kubenswrapper[4846]: I1005 06:48:58.535972 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"78b204fd-1034-4ea1-bcae-e5c876594a39\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b94cc50fe15d58f375f8492067811be16278cdf13ee3d80cfed4682279157b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5e5cf9125d1d413c91666744aeab06980b44420dbf0341b165b05787527944b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://70f0d5b436d18a78a2d8e2767143ba1a3104332d86bd63f91d23cd8e77a952aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c236b4ed7f039a221baac043246353eecbb6812c1e6bf6836a5fa96ec8f5abdb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c236b4ed7f039a221baac043246353eecbb6812c1e6bf6836a5fa96ec8f5abdb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:47:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:47:59Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:47:58Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:58Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:58 crc kubenswrapper[4846]: I1005 06:48:58.559848 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ff10c31459b3b3724ee7b769bd6f72f9ed80cce5419b6adcdf69a3239b70756\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:58Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:58 crc kubenswrapper[4846]: I1005 06:48:58.581782 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://347f900743b7bba3daf92f21c7eba0625e8c774f50057bc9472f0c82119325e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:58Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:58 crc kubenswrapper[4846]: I1005 06:48:58.603834 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56f1fe1f9f06f0249cf937a6963fda82467732dc056ca2b2189be6d8dd9d4a05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7212ce5f42c1fb55146abb190a557a504b43f16af6df84f90e1073567a329ec5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:58Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:58 crc kubenswrapper[4846]: I1005 06:48:58.608043 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:58 crc kubenswrapper[4846]: I1005 06:48:58.608077 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:58 crc kubenswrapper[4846]: I1005 06:48:58.608088 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:58 crc kubenswrapper[4846]: I1005 06:48:58.608107 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:58 crc kubenswrapper[4846]: I1005 06:48:58.608120 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:58Z","lastTransitionTime":"2025-10-05T06:48:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:58 crc kubenswrapper[4846]: I1005 06:48:58.623706 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-9r2kt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"59b83f4f-71f0-4b28-9275-56d5f933ec2c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tdwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tdwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:38Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-9r2kt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:58Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:58 crc kubenswrapper[4846]: I1005 06:48:58.643919 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"775b5702-a131-4d6c-852d-91bfa056cac5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0839c1765df2c68d2554e9c66d289505e559a9672caf8069feeff2c4fdfd2171\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15dbaf2445704da0987f256f81dff2ee95664988541fa48ce94898afeadab2e8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd3b3ab323c6d5153054aa04af2d91f7b9e8436ff1e2f75f3594b57831b2a416\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://67c46f5f0a3ace95bd2e9479c5561126664e18b148c5eceea5bf5efc27e81a67\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c52e076eddfd307110b1f6de00a169b81af8f2785af7becef2a820ce363a5861\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-05T06:48:14Z\\\",\\\"message\\\":\\\"le observer\\\\nW1005 06:48:14.337870 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1005 06:48:14.338086 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1005 06:48:14.339389 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1467752215/tls.crt::/tmp/serving-cert-1467752215/tls.key\\\\\\\"\\\\nI1005 06:48:14.864345 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1005 06:48:14.868087 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1005 06:48:14.868114 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1005 06:48:14.868140 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1005 06:48:14.868147 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1005 06:48:14.877354 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1005 06:48:14.877412 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1005 06:48:14.877456 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1005 06:48:14.877471 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1005 06:48:14.877490 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1005 06:48:14.877501 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1005 06:48:14.877507 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1005 06:48:14.877515 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1005 06:48:14.880326 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:13Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4a091716ff935adb42ae2e88048619fd347bd174f64e5f9c97a01f81690e57f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:01Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dc5411b642f99b13d514d2f21defad0c394aeeb82a966f72300076617e2586c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5dc5411b642f99b13d514d2f21defad0c394aeeb82a966f72300076617e2586c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:47:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:47:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:47:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:58Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:58 crc kubenswrapper[4846]: I1005 06:48:58.664473 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:58Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:58 crc kubenswrapper[4846]: I1005 06:48:58.680266 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-r7ztb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a5119cd9-af70-4324-a955-978306ab0b20\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06231a87429ec292cb57ebd88cda7ad547bcb32e80343b776833621a11c28ad1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jfcj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-r7ztb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:58Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:58 crc kubenswrapper[4846]: I1005 06:48:58.699625 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-gklb6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"698ab882-bb26-4b4c-81e9-8305a44fe068\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a6edca88ea244a58f8cf6cc431110a960c6bfd892812cd0d00fa3f97be3bc67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x6d6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c191538ab797b8056820de7da1f8e93ea18873891a32382b8a59c9af762e103\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x6d6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:37Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-gklb6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:58Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:58 crc kubenswrapper[4846]: I1005 06:48:58.711568 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:58 crc kubenswrapper[4846]: I1005 06:48:58.711611 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:58 crc kubenswrapper[4846]: I1005 06:48:58.711629 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:58 crc kubenswrapper[4846]: I1005 06:48:58.711655 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:58 crc kubenswrapper[4846]: I1005 06:48:58.711673 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:58Z","lastTransitionTime":"2025-10-05T06:48:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:58 crc kubenswrapper[4846]: I1005 06:48:58.737101 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c482bb97-fd50-40cf-96ea-f806c3f01125\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bfe45d30e4fa141541aca1d46bc5cc1da0a7b4a5eb1fc9bb2caa5b7fdcb0b264\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://496a4782faa30a9a20ff63038141600cd635f12902d1cf17989cd067fcdfe929\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://281be077869a6aa1ca815770871a14413d4a97e97aeb550fd5cd215cba88b21f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6771c7215d2f5dd5838d17888375a4768b9d1755fc2e21bda1bbfbb2eb1dbb95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c17a711c2c7473b076a3944ffe7c700808d05cba582fbd55c5c2ac8ff73838a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44a9f740f0a06c8bf928e2a9d1d5979217558c31e73a333631baccebf00fd5d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://44a9f740f0a06c8bf928e2a9d1d5979217558c31e73a333631baccebf00fd5d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:47:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:47:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://48ca0ba6ca6092bfcb5b153220c054f02cc36fb22707c42e021f06d329e18e2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48ca0ba6ca6092bfcb5b153220c054f02cc36fb22707c42e021f06d329e18e2e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:00Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4006d0cea128675c8885d879171204abffc5e11ea0664750a12b4bc7344ccb2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4006d0cea128675c8885d879171204abffc5e11ea0664750a12b4bc7344ccb2e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:47:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:58Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:58 crc kubenswrapper[4846]: I1005 06:48:58.756299 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:58Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:58 crc kubenswrapper[4846]: I1005 06:48:58.777993 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-sfmxd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1dbb13ac-136c-4a1c-9229-b4e8ac4f9c71\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b0af6a6e735f7b7877c177c1d53d1a8dc12c4713a8059c4766fc858af31c9a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://226d977b596d07aeda0486452e2a8605949ed4d2453b21f56174937c395d6f36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://226d977b596d07aeda0486452e2a8605949ed4d2453b21f56174937c395d6f36\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://99c07a4bdb3059d3a99ffdaf794d668d52cffd9e5e3bb8945204dd6d1f8c8328\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99c07a4bdb3059d3a99ffdaf794d668d52cffd9e5e3bb8945204dd6d1f8c8328\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27095272bf4f31d63a78a436636de0aae12f7103597da944f24ed7a74329ee3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27095272bf4f31d63a78a436636de0aae12f7103597da944f24ed7a74329ee3f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1211571af66d8a94302b65feccf30768ba82b7cbad918e44dd405787dce484cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1211571af66d8a94302b65feccf30768ba82b7cbad918e44dd405787dce484cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27631f90d528900fa180c15cfce80e96d4cdba07743eb6d847e2e822d9917218\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27631f90d528900fa180c15cfce80e96d4cdba07743eb6d847e2e822d9917218\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f345693e9aeb1f4dc196349d3354f3c833a2b0415a693298b1cfccfa2b0d4459\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f345693e9aeb1f4dc196349d3354f3c833a2b0415a693298b1cfccfa2b0d4459\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-sfmxd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:58Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:58 crc kubenswrapper[4846]: I1005 06:48:58.811984 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-txzlk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"93a6ace4-8d64-44d7-9637-457d3af6543c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ad1940d07e37217089642478ece9ccdaca74784738a400929a9f5359d158c01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a6f6b2c32dd3da5f9fd5e6971c1773e81a9a24184b7aa5b543cec879001ae53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e9b0219ecbc7e60905d6de3e67a2c8dc9af1017b4aed9538ee0c58c861f2438\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af5e5a4da9ad7371d5a24633e8268f12e3a7d419a8c93b11ddbcc006d8339c6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8da85222775abc7a772f6a8ff8255f1a0d5b15807ba177a7e3e2b2d01a4f24b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1409651de8c897804f87be03d016c138f8acf27e662dce265a93e15c1848adcf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd084bb9806ece6529db0866cee69e6a3152f0da7e0f13fc2e4adcbe2f55962c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd084bb9806ece6529db0866cee69e6a3152f0da7e0f13fc2e4adcbe2f55962c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-05T06:48:50Z\\\",\\\"message\\\":\\\":50.819394 6534 handler.go:208] Removed *v1.Node event handler 2\\\\nI1005 06:48:50.819434 6534 handler.go:208] Removed *v1.Node event handler 7\\\\nI1005 06:48:50.819451 6534 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1005 06:48:50.819624 6534 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1005 06:48:50.819659 6534 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1005 06:48:50.819711 6534 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1005 06:48:50.819735 6534 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1005 06:48:50.820556 6534 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1005 06:48:50.820605 6534 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1005 06:48:50.820649 6534 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1005 06:48:50.820664 6534 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1005 06:48:50.820706 6534 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1005 06:48:50.820711 6534 factory.go:656] Stopping watch factory\\\\nI1005 06:48:50.820731 6534 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1005 06:48:50.820750 6534 ovnkube.go:599] Stopped ovnkube\\\\nI1005 06:48:50.820796 6534 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1005 06:48:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:49Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-txzlk_openshift-ovn-kubernetes(93a6ace4-8d64-44d7-9637-457d3af6543c)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0df0ebcd87dc5af4b17736b43ec7b2395968b19ed16e775366af14e2b2d49c52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff3c6545ff3d2436da27bbd223b950b30b0c69a3e7f3034cb0e61380cb38d338\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ff3c6545ff3d2436da27bbd223b950b30b0c69a3e7f3034cb0e61380cb38d338\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:24Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-txzlk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:58Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:58 crc kubenswrapper[4846]: I1005 06:48:58.820976 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:58 crc kubenswrapper[4846]: I1005 06:48:58.821053 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:58 crc kubenswrapper[4846]: I1005 06:48:58.821065 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:58 crc kubenswrapper[4846]: I1005 06:48:58.821085 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:58 crc kubenswrapper[4846]: I1005 06:48:58.821098 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:58Z","lastTransitionTime":"2025-10-05T06:48:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:58 crc kubenswrapper[4846]: I1005 06:48:58.838431 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c9f812b-ecc8-4985-a670-53fa80fbb499\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2b5ee9781c01d6aec3ac778efef1067b6c3bfe4ed37431bdebe188fcde0165f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:47:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3fb4717692aa1fec0012b514ef8d18d1ca3707cfa4bb8248bb93841a0a40f11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:47:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0b6a7afe3094f1c92093c05c28344897a615d590c2fbb507c87e18a44c6c831\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://640ade2c733d7916ec213a201a1dd614d261c69eab654c80100efbc99a768995\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:47:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:58Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:58 crc kubenswrapper[4846]: I1005 06:48:58.861534 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:58Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:58 crc kubenswrapper[4846]: I1005 06:48:58.881835 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e67d0e447a8a6a381259e25d25badc791194e1291f56b1cd77bd63989b20408\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dkhl7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://516f88f4990f5962d8b694a3d889259cc5508a309919870191c2e1331640c973\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dkhl7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-fscvf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:58Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:58 crc kubenswrapper[4846]: I1005 06:48:58.900529 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-w2sj8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32931fce-7374-4cab-a2e0-79d96a842274\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0df535d65d39de6c9a3dfb931bafc4469adc22b72f7a7903beda4fef8cd48138\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9rtxf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:27Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-w2sj8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:48:58Z is after 2025-08-24T17:21:41Z" Oct 05 06:48:58 crc kubenswrapper[4846]: I1005 06:48:58.925315 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:58 crc kubenswrapper[4846]: I1005 06:48:58.925386 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:58 crc kubenswrapper[4846]: I1005 06:48:58.925406 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:58 crc kubenswrapper[4846]: I1005 06:48:58.925433 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:58 crc kubenswrapper[4846]: I1005 06:48:58.925451 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:58Z","lastTransitionTime":"2025-10-05T06:48:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:59 crc kubenswrapper[4846]: I1005 06:48:59.029355 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:59 crc kubenswrapper[4846]: I1005 06:48:59.029420 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:59 crc kubenswrapper[4846]: I1005 06:48:59.029431 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:59 crc kubenswrapper[4846]: I1005 06:48:59.029471 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:59 crc kubenswrapper[4846]: I1005 06:48:59.029485 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:59Z","lastTransitionTime":"2025-10-05T06:48:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:59 crc kubenswrapper[4846]: I1005 06:48:59.132638 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:59 crc kubenswrapper[4846]: I1005 06:48:59.132688 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:59 crc kubenswrapper[4846]: I1005 06:48:59.132697 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:59 crc kubenswrapper[4846]: I1005 06:48:59.132712 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:59 crc kubenswrapper[4846]: I1005 06:48:59.132741 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:59Z","lastTransitionTime":"2025-10-05T06:48:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:59 crc kubenswrapper[4846]: I1005 06:48:59.236806 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:59 crc kubenswrapper[4846]: I1005 06:48:59.236885 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:59 crc kubenswrapper[4846]: I1005 06:48:59.236904 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:59 crc kubenswrapper[4846]: I1005 06:48:59.236938 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:59 crc kubenswrapper[4846]: I1005 06:48:59.236965 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:59Z","lastTransitionTime":"2025-10-05T06:48:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:59 crc kubenswrapper[4846]: I1005 06:48:59.339564 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:59 crc kubenswrapper[4846]: I1005 06:48:59.339641 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:59 crc kubenswrapper[4846]: I1005 06:48:59.339665 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:59 crc kubenswrapper[4846]: I1005 06:48:59.339696 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:59 crc kubenswrapper[4846]: I1005 06:48:59.339732 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:59Z","lastTransitionTime":"2025-10-05T06:48:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:59 crc kubenswrapper[4846]: I1005 06:48:59.443174 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:59 crc kubenswrapper[4846]: I1005 06:48:59.443275 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:59 crc kubenswrapper[4846]: I1005 06:48:59.443292 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:59 crc kubenswrapper[4846]: I1005 06:48:59.443320 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:59 crc kubenswrapper[4846]: I1005 06:48:59.443338 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:59Z","lastTransitionTime":"2025-10-05T06:48:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:59 crc kubenswrapper[4846]: I1005 06:48:59.497424 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 05 06:48:59 crc kubenswrapper[4846]: I1005 06:48:59.497433 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 05 06:48:59 crc kubenswrapper[4846]: E1005 06:48:59.497690 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 05 06:48:59 crc kubenswrapper[4846]: E1005 06:48:59.497870 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 05 06:48:59 crc kubenswrapper[4846]: I1005 06:48:59.546304 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:59 crc kubenswrapper[4846]: I1005 06:48:59.546360 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:59 crc kubenswrapper[4846]: I1005 06:48:59.546422 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:59 crc kubenswrapper[4846]: I1005 06:48:59.546454 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:59 crc kubenswrapper[4846]: I1005 06:48:59.546521 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:59Z","lastTransitionTime":"2025-10-05T06:48:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:59 crc kubenswrapper[4846]: I1005 06:48:59.650151 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:59 crc kubenswrapper[4846]: I1005 06:48:59.650262 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:59 crc kubenswrapper[4846]: I1005 06:48:59.650290 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:59 crc kubenswrapper[4846]: I1005 06:48:59.650324 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:59 crc kubenswrapper[4846]: I1005 06:48:59.650348 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:59Z","lastTransitionTime":"2025-10-05T06:48:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:59 crc kubenswrapper[4846]: I1005 06:48:59.753496 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:59 crc kubenswrapper[4846]: I1005 06:48:59.753581 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:59 crc kubenswrapper[4846]: I1005 06:48:59.753603 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:59 crc kubenswrapper[4846]: I1005 06:48:59.753633 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:59 crc kubenswrapper[4846]: I1005 06:48:59.753654 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:59Z","lastTransitionTime":"2025-10-05T06:48:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:59 crc kubenswrapper[4846]: I1005 06:48:59.857579 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:59 crc kubenswrapper[4846]: I1005 06:48:59.857629 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:59 crc kubenswrapper[4846]: I1005 06:48:59.857646 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:59 crc kubenswrapper[4846]: I1005 06:48:59.857669 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:59 crc kubenswrapper[4846]: I1005 06:48:59.857687 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:59Z","lastTransitionTime":"2025-10-05T06:48:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:48:59 crc kubenswrapper[4846]: I1005 06:48:59.960888 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:48:59 crc kubenswrapper[4846]: I1005 06:48:59.960947 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:48:59 crc kubenswrapper[4846]: I1005 06:48:59.960966 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:48:59 crc kubenswrapper[4846]: I1005 06:48:59.960989 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:48:59 crc kubenswrapper[4846]: I1005 06:48:59.961006 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:48:59Z","lastTransitionTime":"2025-10-05T06:48:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:00 crc kubenswrapper[4846]: I1005 06:49:00.064552 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:00 crc kubenswrapper[4846]: I1005 06:49:00.064628 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:00 crc kubenswrapper[4846]: I1005 06:49:00.064647 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:00 crc kubenswrapper[4846]: I1005 06:49:00.064674 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:00 crc kubenswrapper[4846]: I1005 06:49:00.064693 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:00Z","lastTransitionTime":"2025-10-05T06:49:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:00 crc kubenswrapper[4846]: I1005 06:49:00.168588 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:00 crc kubenswrapper[4846]: I1005 06:49:00.168855 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:00 crc kubenswrapper[4846]: I1005 06:49:00.168875 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:00 crc kubenswrapper[4846]: I1005 06:49:00.168900 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:00 crc kubenswrapper[4846]: I1005 06:49:00.168923 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:00Z","lastTransitionTime":"2025-10-05T06:49:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:00 crc kubenswrapper[4846]: I1005 06:49:00.271738 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:00 crc kubenswrapper[4846]: I1005 06:49:00.271800 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:00 crc kubenswrapper[4846]: I1005 06:49:00.271820 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:00 crc kubenswrapper[4846]: I1005 06:49:00.271846 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:00 crc kubenswrapper[4846]: I1005 06:49:00.271863 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:00Z","lastTransitionTime":"2025-10-05T06:49:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:00 crc kubenswrapper[4846]: I1005 06:49:00.374783 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:00 crc kubenswrapper[4846]: I1005 06:49:00.374861 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:00 crc kubenswrapper[4846]: I1005 06:49:00.374879 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:00 crc kubenswrapper[4846]: I1005 06:49:00.374916 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:00 crc kubenswrapper[4846]: I1005 06:49:00.374936 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:00Z","lastTransitionTime":"2025-10-05T06:49:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:00 crc kubenswrapper[4846]: I1005 06:49:00.480648 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:00 crc kubenswrapper[4846]: I1005 06:49:00.480725 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:00 crc kubenswrapper[4846]: I1005 06:49:00.480746 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:00 crc kubenswrapper[4846]: I1005 06:49:00.480778 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:00 crc kubenswrapper[4846]: I1005 06:49:00.480804 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:00Z","lastTransitionTime":"2025-10-05T06:49:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:00 crc kubenswrapper[4846]: I1005 06:49:00.497530 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9r2kt" Oct 05 06:49:00 crc kubenswrapper[4846]: I1005 06:49:00.497580 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 05 06:49:00 crc kubenswrapper[4846]: E1005 06:49:00.497775 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9r2kt" podUID="59b83f4f-71f0-4b28-9275-56d5f933ec2c" Oct 05 06:49:00 crc kubenswrapper[4846]: E1005 06:49:00.497962 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 05 06:49:00 crc kubenswrapper[4846]: I1005 06:49:00.583924 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:00 crc kubenswrapper[4846]: I1005 06:49:00.583974 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:00 crc kubenswrapper[4846]: I1005 06:49:00.583984 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:00 crc kubenswrapper[4846]: I1005 06:49:00.584000 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:00 crc kubenswrapper[4846]: I1005 06:49:00.584013 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:00Z","lastTransitionTime":"2025-10-05T06:49:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:00 crc kubenswrapper[4846]: I1005 06:49:00.687056 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:00 crc kubenswrapper[4846]: I1005 06:49:00.687089 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:00 crc kubenswrapper[4846]: I1005 06:49:00.687098 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:00 crc kubenswrapper[4846]: I1005 06:49:00.687113 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:00 crc kubenswrapper[4846]: I1005 06:49:00.687122 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:00Z","lastTransitionTime":"2025-10-05T06:49:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:00 crc kubenswrapper[4846]: I1005 06:49:00.790003 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:00 crc kubenswrapper[4846]: I1005 06:49:00.790054 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:00 crc kubenswrapper[4846]: I1005 06:49:00.790067 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:00 crc kubenswrapper[4846]: I1005 06:49:00.790089 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:00 crc kubenswrapper[4846]: I1005 06:49:00.790103 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:00Z","lastTransitionTime":"2025-10-05T06:49:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:00 crc kubenswrapper[4846]: I1005 06:49:00.893362 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:00 crc kubenswrapper[4846]: I1005 06:49:00.893415 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:00 crc kubenswrapper[4846]: I1005 06:49:00.893426 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:00 crc kubenswrapper[4846]: I1005 06:49:00.893445 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:00 crc kubenswrapper[4846]: I1005 06:49:00.893459 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:00Z","lastTransitionTime":"2025-10-05T06:49:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:00 crc kubenswrapper[4846]: I1005 06:49:00.996069 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:00 crc kubenswrapper[4846]: I1005 06:49:00.996132 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:00 crc kubenswrapper[4846]: I1005 06:49:00.996150 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:00 crc kubenswrapper[4846]: I1005 06:49:00.996201 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:00 crc kubenswrapper[4846]: I1005 06:49:00.996220 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:00Z","lastTransitionTime":"2025-10-05T06:49:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:01 crc kubenswrapper[4846]: I1005 06:49:01.099102 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:01 crc kubenswrapper[4846]: I1005 06:49:01.099148 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:01 crc kubenswrapper[4846]: I1005 06:49:01.099158 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:01 crc kubenswrapper[4846]: I1005 06:49:01.099186 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:01 crc kubenswrapper[4846]: I1005 06:49:01.099198 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:01Z","lastTransitionTime":"2025-10-05T06:49:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:01 crc kubenswrapper[4846]: I1005 06:49:01.202838 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:01 crc kubenswrapper[4846]: I1005 06:49:01.202906 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:01 crc kubenswrapper[4846]: I1005 06:49:01.202937 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:01 crc kubenswrapper[4846]: I1005 06:49:01.202974 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:01 crc kubenswrapper[4846]: I1005 06:49:01.202987 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:01Z","lastTransitionTime":"2025-10-05T06:49:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:01 crc kubenswrapper[4846]: I1005 06:49:01.305893 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:01 crc kubenswrapper[4846]: I1005 06:49:01.305993 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:01 crc kubenswrapper[4846]: I1005 06:49:01.306011 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:01 crc kubenswrapper[4846]: I1005 06:49:01.306037 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:01 crc kubenswrapper[4846]: I1005 06:49:01.306056 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:01Z","lastTransitionTime":"2025-10-05T06:49:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:01 crc kubenswrapper[4846]: I1005 06:49:01.409252 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:01 crc kubenswrapper[4846]: I1005 06:49:01.409324 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:01 crc kubenswrapper[4846]: I1005 06:49:01.409348 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:01 crc kubenswrapper[4846]: I1005 06:49:01.409380 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:01 crc kubenswrapper[4846]: I1005 06:49:01.409399 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:01Z","lastTransitionTime":"2025-10-05T06:49:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:01 crc kubenswrapper[4846]: I1005 06:49:01.497596 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 05 06:49:01 crc kubenswrapper[4846]: I1005 06:49:01.497679 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 05 06:49:01 crc kubenswrapper[4846]: E1005 06:49:01.497873 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 05 06:49:01 crc kubenswrapper[4846]: E1005 06:49:01.498076 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 05 06:49:01 crc kubenswrapper[4846]: I1005 06:49:01.512304 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:01 crc kubenswrapper[4846]: I1005 06:49:01.512449 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:01 crc kubenswrapper[4846]: I1005 06:49:01.512489 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:01 crc kubenswrapper[4846]: I1005 06:49:01.512525 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:01 crc kubenswrapper[4846]: I1005 06:49:01.512552 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:01Z","lastTransitionTime":"2025-10-05T06:49:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:01 crc kubenswrapper[4846]: I1005 06:49:01.615686 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:01 crc kubenswrapper[4846]: I1005 06:49:01.615754 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:01 crc kubenswrapper[4846]: I1005 06:49:01.615772 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:01 crc kubenswrapper[4846]: I1005 06:49:01.615796 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:01 crc kubenswrapper[4846]: I1005 06:49:01.615815 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:01Z","lastTransitionTime":"2025-10-05T06:49:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:01 crc kubenswrapper[4846]: I1005 06:49:01.719035 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:01 crc kubenswrapper[4846]: I1005 06:49:01.719083 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:01 crc kubenswrapper[4846]: I1005 06:49:01.719094 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:01 crc kubenswrapper[4846]: I1005 06:49:01.719112 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:01 crc kubenswrapper[4846]: I1005 06:49:01.719123 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:01Z","lastTransitionTime":"2025-10-05T06:49:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:01 crc kubenswrapper[4846]: I1005 06:49:01.821812 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:01 crc kubenswrapper[4846]: I1005 06:49:01.821875 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:01 crc kubenswrapper[4846]: I1005 06:49:01.821892 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:01 crc kubenswrapper[4846]: I1005 06:49:01.821916 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:01 crc kubenswrapper[4846]: I1005 06:49:01.821934 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:01Z","lastTransitionTime":"2025-10-05T06:49:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:01 crc kubenswrapper[4846]: I1005 06:49:01.925369 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:01 crc kubenswrapper[4846]: I1005 06:49:01.925436 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:01 crc kubenswrapper[4846]: I1005 06:49:01.925454 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:01 crc kubenswrapper[4846]: I1005 06:49:01.925479 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:01 crc kubenswrapper[4846]: I1005 06:49:01.925523 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:01Z","lastTransitionTime":"2025-10-05T06:49:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:02 crc kubenswrapper[4846]: I1005 06:49:02.029300 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:02 crc kubenswrapper[4846]: I1005 06:49:02.029356 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:02 crc kubenswrapper[4846]: I1005 06:49:02.029368 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:02 crc kubenswrapper[4846]: I1005 06:49:02.029388 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:02 crc kubenswrapper[4846]: I1005 06:49:02.029405 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:02Z","lastTransitionTime":"2025-10-05T06:49:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:02 crc kubenswrapper[4846]: I1005 06:49:02.132673 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:02 crc kubenswrapper[4846]: I1005 06:49:02.132748 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:02 crc kubenswrapper[4846]: I1005 06:49:02.132769 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:02 crc kubenswrapper[4846]: I1005 06:49:02.132797 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:02 crc kubenswrapper[4846]: I1005 06:49:02.132817 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:02Z","lastTransitionTime":"2025-10-05T06:49:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:02 crc kubenswrapper[4846]: I1005 06:49:02.235756 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:02 crc kubenswrapper[4846]: I1005 06:49:02.235818 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:02 crc kubenswrapper[4846]: I1005 06:49:02.235835 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:02 crc kubenswrapper[4846]: I1005 06:49:02.235861 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:02 crc kubenswrapper[4846]: I1005 06:49:02.235880 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:02Z","lastTransitionTime":"2025-10-05T06:49:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:02 crc kubenswrapper[4846]: I1005 06:49:02.338978 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:02 crc kubenswrapper[4846]: I1005 06:49:02.339024 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:02 crc kubenswrapper[4846]: I1005 06:49:02.339045 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:02 crc kubenswrapper[4846]: I1005 06:49:02.339067 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:02 crc kubenswrapper[4846]: I1005 06:49:02.339083 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:02Z","lastTransitionTime":"2025-10-05T06:49:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:02 crc kubenswrapper[4846]: I1005 06:49:02.442468 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:02 crc kubenswrapper[4846]: I1005 06:49:02.442517 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:02 crc kubenswrapper[4846]: I1005 06:49:02.442528 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:02 crc kubenswrapper[4846]: I1005 06:49:02.442544 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:02 crc kubenswrapper[4846]: I1005 06:49:02.442558 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:02Z","lastTransitionTime":"2025-10-05T06:49:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:02 crc kubenswrapper[4846]: I1005 06:49:02.496795 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 05 06:49:02 crc kubenswrapper[4846]: I1005 06:49:02.496991 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9r2kt" Oct 05 06:49:02 crc kubenswrapper[4846]: E1005 06:49:02.497117 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 05 06:49:02 crc kubenswrapper[4846]: E1005 06:49:02.497315 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9r2kt" podUID="59b83f4f-71f0-4b28-9275-56d5f933ec2c" Oct 05 06:49:02 crc kubenswrapper[4846]: I1005 06:49:02.545880 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:02 crc kubenswrapper[4846]: I1005 06:49:02.545932 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:02 crc kubenswrapper[4846]: I1005 06:49:02.545947 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:02 crc kubenswrapper[4846]: I1005 06:49:02.545967 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:02 crc kubenswrapper[4846]: I1005 06:49:02.545981 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:02Z","lastTransitionTime":"2025-10-05T06:49:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:02 crc kubenswrapper[4846]: I1005 06:49:02.650171 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:02 crc kubenswrapper[4846]: I1005 06:49:02.650283 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:02 crc kubenswrapper[4846]: I1005 06:49:02.650302 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:02 crc kubenswrapper[4846]: I1005 06:49:02.650327 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:02 crc kubenswrapper[4846]: I1005 06:49:02.650343 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:02Z","lastTransitionTime":"2025-10-05T06:49:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:02 crc kubenswrapper[4846]: I1005 06:49:02.753898 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:02 crc kubenswrapper[4846]: I1005 06:49:02.753946 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:02 crc kubenswrapper[4846]: I1005 06:49:02.753961 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:02 crc kubenswrapper[4846]: I1005 06:49:02.753985 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:02 crc kubenswrapper[4846]: I1005 06:49:02.754009 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:02Z","lastTransitionTime":"2025-10-05T06:49:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:02 crc kubenswrapper[4846]: I1005 06:49:02.856887 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:02 crc kubenswrapper[4846]: I1005 06:49:02.856939 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:02 crc kubenswrapper[4846]: I1005 06:49:02.856955 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:02 crc kubenswrapper[4846]: I1005 06:49:02.856979 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:02 crc kubenswrapper[4846]: I1005 06:49:02.857000 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:02Z","lastTransitionTime":"2025-10-05T06:49:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:02 crc kubenswrapper[4846]: I1005 06:49:02.960238 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:02 crc kubenswrapper[4846]: I1005 06:49:02.960276 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:02 crc kubenswrapper[4846]: I1005 06:49:02.960287 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:02 crc kubenswrapper[4846]: I1005 06:49:02.960306 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:02 crc kubenswrapper[4846]: I1005 06:49:02.960320 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:02Z","lastTransitionTime":"2025-10-05T06:49:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:03 crc kubenswrapper[4846]: I1005 06:49:03.064427 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:03 crc kubenswrapper[4846]: I1005 06:49:03.064484 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:03 crc kubenswrapper[4846]: I1005 06:49:03.064505 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:03 crc kubenswrapper[4846]: I1005 06:49:03.064528 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:03 crc kubenswrapper[4846]: I1005 06:49:03.064546 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:03Z","lastTransitionTime":"2025-10-05T06:49:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:03 crc kubenswrapper[4846]: I1005 06:49:03.167940 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:03 crc kubenswrapper[4846]: I1005 06:49:03.167966 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:03 crc kubenswrapper[4846]: I1005 06:49:03.167976 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:03 crc kubenswrapper[4846]: I1005 06:49:03.167991 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:03 crc kubenswrapper[4846]: I1005 06:49:03.168003 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:03Z","lastTransitionTime":"2025-10-05T06:49:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:03 crc kubenswrapper[4846]: I1005 06:49:03.272389 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:03 crc kubenswrapper[4846]: I1005 06:49:03.272455 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:03 crc kubenswrapper[4846]: I1005 06:49:03.272473 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:03 crc kubenswrapper[4846]: I1005 06:49:03.272501 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:03 crc kubenswrapper[4846]: I1005 06:49:03.272516 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:03Z","lastTransitionTime":"2025-10-05T06:49:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:03 crc kubenswrapper[4846]: I1005 06:49:03.375742 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:03 crc kubenswrapper[4846]: I1005 06:49:03.375808 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:03 crc kubenswrapper[4846]: I1005 06:49:03.375826 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:03 crc kubenswrapper[4846]: I1005 06:49:03.375852 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:03 crc kubenswrapper[4846]: I1005 06:49:03.375873 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:03Z","lastTransitionTime":"2025-10-05T06:49:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:03 crc kubenswrapper[4846]: I1005 06:49:03.479325 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:03 crc kubenswrapper[4846]: I1005 06:49:03.479356 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:03 crc kubenswrapper[4846]: I1005 06:49:03.479364 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:03 crc kubenswrapper[4846]: I1005 06:49:03.479380 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:03 crc kubenswrapper[4846]: I1005 06:49:03.479390 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:03Z","lastTransitionTime":"2025-10-05T06:49:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:03 crc kubenswrapper[4846]: I1005 06:49:03.497571 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 05 06:49:03 crc kubenswrapper[4846]: I1005 06:49:03.497689 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 05 06:49:03 crc kubenswrapper[4846]: E1005 06:49:03.497801 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 05 06:49:03 crc kubenswrapper[4846]: E1005 06:49:03.497878 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 05 06:49:03 crc kubenswrapper[4846]: I1005 06:49:03.582787 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:03 crc kubenswrapper[4846]: I1005 06:49:03.582833 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:03 crc kubenswrapper[4846]: I1005 06:49:03.582844 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:03 crc kubenswrapper[4846]: I1005 06:49:03.582862 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:03 crc kubenswrapper[4846]: I1005 06:49:03.582877 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:03Z","lastTransitionTime":"2025-10-05T06:49:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:03 crc kubenswrapper[4846]: I1005 06:49:03.685928 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:03 crc kubenswrapper[4846]: I1005 06:49:03.686099 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:03 crc kubenswrapper[4846]: I1005 06:49:03.686122 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:03 crc kubenswrapper[4846]: I1005 06:49:03.686156 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:03 crc kubenswrapper[4846]: I1005 06:49:03.686206 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:03Z","lastTransitionTime":"2025-10-05T06:49:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:03 crc kubenswrapper[4846]: I1005 06:49:03.790649 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:03 crc kubenswrapper[4846]: I1005 06:49:03.790718 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:03 crc kubenswrapper[4846]: I1005 06:49:03.790734 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:03 crc kubenswrapper[4846]: I1005 06:49:03.790758 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:03 crc kubenswrapper[4846]: I1005 06:49:03.790773 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:03Z","lastTransitionTime":"2025-10-05T06:49:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:03 crc kubenswrapper[4846]: I1005 06:49:03.894452 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:03 crc kubenswrapper[4846]: I1005 06:49:03.894518 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:03 crc kubenswrapper[4846]: I1005 06:49:03.894529 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:03 crc kubenswrapper[4846]: I1005 06:49:03.894549 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:03 crc kubenswrapper[4846]: I1005 06:49:03.894565 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:03Z","lastTransitionTime":"2025-10-05T06:49:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:03 crc kubenswrapper[4846]: I1005 06:49:03.997880 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:03 crc kubenswrapper[4846]: I1005 06:49:03.997959 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:03 crc kubenswrapper[4846]: I1005 06:49:03.997979 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:03 crc kubenswrapper[4846]: I1005 06:49:03.998008 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:03 crc kubenswrapper[4846]: I1005 06:49:03.998030 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:03Z","lastTransitionTime":"2025-10-05T06:49:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:04 crc kubenswrapper[4846]: I1005 06:49:04.101135 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:04 crc kubenswrapper[4846]: I1005 06:49:04.101199 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:04 crc kubenswrapper[4846]: I1005 06:49:04.101213 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:04 crc kubenswrapper[4846]: I1005 06:49:04.101231 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:04 crc kubenswrapper[4846]: I1005 06:49:04.101246 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:04Z","lastTransitionTime":"2025-10-05T06:49:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:04 crc kubenswrapper[4846]: I1005 06:49:04.204253 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:04 crc kubenswrapper[4846]: I1005 06:49:04.204300 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:04 crc kubenswrapper[4846]: I1005 06:49:04.204311 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:04 crc kubenswrapper[4846]: I1005 06:49:04.204325 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:04 crc kubenswrapper[4846]: I1005 06:49:04.204338 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:04Z","lastTransitionTime":"2025-10-05T06:49:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:04 crc kubenswrapper[4846]: I1005 06:49:04.307428 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:04 crc kubenswrapper[4846]: I1005 06:49:04.307498 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:04 crc kubenswrapper[4846]: I1005 06:49:04.307517 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:04 crc kubenswrapper[4846]: I1005 06:49:04.307545 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:04 crc kubenswrapper[4846]: I1005 06:49:04.307566 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:04Z","lastTransitionTime":"2025-10-05T06:49:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:04 crc kubenswrapper[4846]: I1005 06:49:04.411011 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:04 crc kubenswrapper[4846]: I1005 06:49:04.411060 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:04 crc kubenswrapper[4846]: I1005 06:49:04.411069 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:04 crc kubenswrapper[4846]: I1005 06:49:04.411086 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:04 crc kubenswrapper[4846]: I1005 06:49:04.411096 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:04Z","lastTransitionTime":"2025-10-05T06:49:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:04 crc kubenswrapper[4846]: I1005 06:49:04.496451 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9r2kt" Oct 05 06:49:04 crc kubenswrapper[4846]: I1005 06:49:04.496713 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 05 06:49:04 crc kubenswrapper[4846]: E1005 06:49:04.496873 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 05 06:49:04 crc kubenswrapper[4846]: E1005 06:49:04.497360 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9r2kt" podUID="59b83f4f-71f0-4b28-9275-56d5f933ec2c" Oct 05 06:49:04 crc kubenswrapper[4846]: I1005 06:49:04.497920 4846 scope.go:117] "RemoveContainer" containerID="dd084bb9806ece6529db0866cee69e6a3152f0da7e0f13fc2e4adcbe2f55962c" Oct 05 06:49:04 crc kubenswrapper[4846]: E1005 06:49:04.498139 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-txzlk_openshift-ovn-kubernetes(93a6ace4-8d64-44d7-9637-457d3af6543c)\"" pod="openshift-ovn-kubernetes/ovnkube-node-txzlk" podUID="93a6ace4-8d64-44d7-9637-457d3af6543c" Oct 05 06:49:04 crc kubenswrapper[4846]: I1005 06:49:04.513407 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:04 crc kubenswrapper[4846]: I1005 06:49:04.513448 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:04 crc kubenswrapper[4846]: I1005 06:49:04.513470 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:04 crc kubenswrapper[4846]: I1005 06:49:04.513486 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:04 crc kubenswrapper[4846]: I1005 06:49:04.513501 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:04Z","lastTransitionTime":"2025-10-05T06:49:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:04 crc kubenswrapper[4846]: I1005 06:49:04.616078 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:04 crc kubenswrapper[4846]: I1005 06:49:04.616148 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:04 crc kubenswrapper[4846]: I1005 06:49:04.616168 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:04 crc kubenswrapper[4846]: I1005 06:49:04.616234 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:04 crc kubenswrapper[4846]: I1005 06:49:04.616255 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:04Z","lastTransitionTime":"2025-10-05T06:49:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:04 crc kubenswrapper[4846]: I1005 06:49:04.719530 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:04 crc kubenswrapper[4846]: I1005 06:49:04.719580 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:04 crc kubenswrapper[4846]: I1005 06:49:04.719595 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:04 crc kubenswrapper[4846]: I1005 06:49:04.719616 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:04 crc kubenswrapper[4846]: I1005 06:49:04.719629 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:04Z","lastTransitionTime":"2025-10-05T06:49:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:04 crc kubenswrapper[4846]: I1005 06:49:04.821852 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:04 crc kubenswrapper[4846]: I1005 06:49:04.821900 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:04 crc kubenswrapper[4846]: I1005 06:49:04.821911 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:04 crc kubenswrapper[4846]: I1005 06:49:04.821928 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:04 crc kubenswrapper[4846]: I1005 06:49:04.821944 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:04Z","lastTransitionTime":"2025-10-05T06:49:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:04 crc kubenswrapper[4846]: I1005 06:49:04.923677 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:04 crc kubenswrapper[4846]: I1005 06:49:04.923720 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:04 crc kubenswrapper[4846]: I1005 06:49:04.923733 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:04 crc kubenswrapper[4846]: I1005 06:49:04.923750 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:04 crc kubenswrapper[4846]: I1005 06:49:04.923760 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:04Z","lastTransitionTime":"2025-10-05T06:49:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:05 crc kubenswrapper[4846]: I1005 06:49:05.025608 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:05 crc kubenswrapper[4846]: I1005 06:49:05.025638 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:05 crc kubenswrapper[4846]: I1005 06:49:05.025646 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:05 crc kubenswrapper[4846]: I1005 06:49:05.025660 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:05 crc kubenswrapper[4846]: I1005 06:49:05.025671 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:05Z","lastTransitionTime":"2025-10-05T06:49:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:05 crc kubenswrapper[4846]: I1005 06:49:05.128309 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:05 crc kubenswrapper[4846]: I1005 06:49:05.128376 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:05 crc kubenswrapper[4846]: I1005 06:49:05.128399 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:05 crc kubenswrapper[4846]: I1005 06:49:05.128427 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:05 crc kubenswrapper[4846]: I1005 06:49:05.128447 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:05Z","lastTransitionTime":"2025-10-05T06:49:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:05 crc kubenswrapper[4846]: I1005 06:49:05.231578 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:05 crc kubenswrapper[4846]: I1005 06:49:05.231654 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:05 crc kubenswrapper[4846]: I1005 06:49:05.231673 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:05 crc kubenswrapper[4846]: I1005 06:49:05.231703 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:05 crc kubenswrapper[4846]: I1005 06:49:05.231721 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:05Z","lastTransitionTime":"2025-10-05T06:49:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:05 crc kubenswrapper[4846]: I1005 06:49:05.337683 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:05 crc kubenswrapper[4846]: I1005 06:49:05.337739 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:05 crc kubenswrapper[4846]: I1005 06:49:05.337753 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:05 crc kubenswrapper[4846]: I1005 06:49:05.337783 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:05 crc kubenswrapper[4846]: I1005 06:49:05.337798 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:05Z","lastTransitionTime":"2025-10-05T06:49:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:05 crc kubenswrapper[4846]: I1005 06:49:05.440199 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:05 crc kubenswrapper[4846]: I1005 06:49:05.440239 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:05 crc kubenswrapper[4846]: I1005 06:49:05.440250 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:05 crc kubenswrapper[4846]: I1005 06:49:05.440267 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:05 crc kubenswrapper[4846]: I1005 06:49:05.440278 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:05Z","lastTransitionTime":"2025-10-05T06:49:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:05 crc kubenswrapper[4846]: I1005 06:49:05.497158 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 05 06:49:05 crc kubenswrapper[4846]: I1005 06:49:05.497235 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 05 06:49:05 crc kubenswrapper[4846]: E1005 06:49:05.497392 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 05 06:49:05 crc kubenswrapper[4846]: E1005 06:49:05.497511 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 05 06:49:05 crc kubenswrapper[4846]: I1005 06:49:05.543087 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:05 crc kubenswrapper[4846]: I1005 06:49:05.543144 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:05 crc kubenswrapper[4846]: I1005 06:49:05.543158 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:05 crc kubenswrapper[4846]: I1005 06:49:05.543194 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:05 crc kubenswrapper[4846]: I1005 06:49:05.543208 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:05Z","lastTransitionTime":"2025-10-05T06:49:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:05 crc kubenswrapper[4846]: I1005 06:49:05.591993 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:05 crc kubenswrapper[4846]: I1005 06:49:05.592035 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:05 crc kubenswrapper[4846]: I1005 06:49:05.592049 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:05 crc kubenswrapper[4846]: I1005 06:49:05.592067 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:05 crc kubenswrapper[4846]: I1005 06:49:05.592079 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:05Z","lastTransitionTime":"2025-10-05T06:49:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:05 crc kubenswrapper[4846]: E1005 06:49:05.607207 4846 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:49:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:49:05Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:49:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:49:05Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:49:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:49:05Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:49:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:49:05Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"04ce628d-2f61-4931-aad6-615924ea53fb\\\",\\\"systemUUID\\\":\\\"a050a23b-e773-4ba9-989b-360aa53a0605\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:49:05Z is after 2025-08-24T17:21:41Z" Oct 05 06:49:05 crc kubenswrapper[4846]: I1005 06:49:05.611618 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:05 crc kubenswrapper[4846]: I1005 06:49:05.611662 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:05 crc kubenswrapper[4846]: I1005 06:49:05.611674 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:05 crc kubenswrapper[4846]: I1005 06:49:05.611693 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:05 crc kubenswrapper[4846]: I1005 06:49:05.611705 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:05Z","lastTransitionTime":"2025-10-05T06:49:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:05 crc kubenswrapper[4846]: E1005 06:49:05.627365 4846 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:49:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:49:05Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:49:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:49:05Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:49:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:49:05Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:49:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:49:05Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"04ce628d-2f61-4931-aad6-615924ea53fb\\\",\\\"systemUUID\\\":\\\"a050a23b-e773-4ba9-989b-360aa53a0605\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:49:05Z is after 2025-08-24T17:21:41Z" Oct 05 06:49:05 crc kubenswrapper[4846]: I1005 06:49:05.631748 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:05 crc kubenswrapper[4846]: I1005 06:49:05.631801 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:05 crc kubenswrapper[4846]: I1005 06:49:05.631811 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:05 crc kubenswrapper[4846]: I1005 06:49:05.631828 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:05 crc kubenswrapper[4846]: I1005 06:49:05.631838 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:05Z","lastTransitionTime":"2025-10-05T06:49:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:05 crc kubenswrapper[4846]: E1005 06:49:05.643439 4846 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:49:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:49:05Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:49:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:49:05Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:49:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:49:05Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:49:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:49:05Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"04ce628d-2f61-4931-aad6-615924ea53fb\\\",\\\"systemUUID\\\":\\\"a050a23b-e773-4ba9-989b-360aa53a0605\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:49:05Z is after 2025-08-24T17:21:41Z" Oct 05 06:49:05 crc kubenswrapper[4846]: I1005 06:49:05.646937 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:05 crc kubenswrapper[4846]: I1005 06:49:05.646971 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:05 crc kubenswrapper[4846]: I1005 06:49:05.646980 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:05 crc kubenswrapper[4846]: I1005 06:49:05.646996 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:05 crc kubenswrapper[4846]: I1005 06:49:05.647006 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:05Z","lastTransitionTime":"2025-10-05T06:49:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:05 crc kubenswrapper[4846]: E1005 06:49:05.658352 4846 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:49:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:49:05Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:49:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:49:05Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:49:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:49:05Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:49:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:49:05Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"04ce628d-2f61-4931-aad6-615924ea53fb\\\",\\\"systemUUID\\\":\\\"a050a23b-e773-4ba9-989b-360aa53a0605\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:49:05Z is after 2025-08-24T17:21:41Z" Oct 05 06:49:05 crc kubenswrapper[4846]: I1005 06:49:05.661915 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:05 crc kubenswrapper[4846]: I1005 06:49:05.661940 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:05 crc kubenswrapper[4846]: I1005 06:49:05.661975 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:05 crc kubenswrapper[4846]: I1005 06:49:05.661993 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:05 crc kubenswrapper[4846]: I1005 06:49:05.662005 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:05Z","lastTransitionTime":"2025-10-05T06:49:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:05 crc kubenswrapper[4846]: E1005 06:49:05.675436 4846 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:49:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:49:05Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:49:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:49:05Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:49:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:49:05Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:49:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:49:05Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"04ce628d-2f61-4931-aad6-615924ea53fb\\\",\\\"systemUUID\\\":\\\"a050a23b-e773-4ba9-989b-360aa53a0605\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:49:05Z is after 2025-08-24T17:21:41Z" Oct 05 06:49:05 crc kubenswrapper[4846]: E1005 06:49:05.676007 4846 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 05 06:49:05 crc kubenswrapper[4846]: I1005 06:49:05.677449 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:05 crc kubenswrapper[4846]: I1005 06:49:05.677469 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:05 crc kubenswrapper[4846]: I1005 06:49:05.677478 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:05 crc kubenswrapper[4846]: I1005 06:49:05.677488 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:05 crc kubenswrapper[4846]: I1005 06:49:05.677497 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:05Z","lastTransitionTime":"2025-10-05T06:49:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:05 crc kubenswrapper[4846]: I1005 06:49:05.780358 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:05 crc kubenswrapper[4846]: I1005 06:49:05.780392 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:05 crc kubenswrapper[4846]: I1005 06:49:05.780400 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:05 crc kubenswrapper[4846]: I1005 06:49:05.780412 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:05 crc kubenswrapper[4846]: I1005 06:49:05.780421 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:05Z","lastTransitionTime":"2025-10-05T06:49:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:05 crc kubenswrapper[4846]: I1005 06:49:05.883651 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:05 crc kubenswrapper[4846]: I1005 06:49:05.883703 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:05 crc kubenswrapper[4846]: I1005 06:49:05.883716 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:05 crc kubenswrapper[4846]: I1005 06:49:05.883738 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:05 crc kubenswrapper[4846]: I1005 06:49:05.883752 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:05Z","lastTransitionTime":"2025-10-05T06:49:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:05 crc kubenswrapper[4846]: I1005 06:49:05.986587 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:05 crc kubenswrapper[4846]: I1005 06:49:05.986659 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:05 crc kubenswrapper[4846]: I1005 06:49:05.986681 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:05 crc kubenswrapper[4846]: I1005 06:49:05.986715 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:05 crc kubenswrapper[4846]: I1005 06:49:05.986735 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:05Z","lastTransitionTime":"2025-10-05T06:49:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:06 crc kubenswrapper[4846]: I1005 06:49:06.089733 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:06 crc kubenswrapper[4846]: I1005 06:49:06.089776 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:06 crc kubenswrapper[4846]: I1005 06:49:06.089790 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:06 crc kubenswrapper[4846]: I1005 06:49:06.089811 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:06 crc kubenswrapper[4846]: I1005 06:49:06.089823 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:06Z","lastTransitionTime":"2025-10-05T06:49:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:06 crc kubenswrapper[4846]: I1005 06:49:06.193741 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:06 crc kubenswrapper[4846]: I1005 06:49:06.193795 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:06 crc kubenswrapper[4846]: I1005 06:49:06.193812 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:06 crc kubenswrapper[4846]: I1005 06:49:06.193829 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:06 crc kubenswrapper[4846]: I1005 06:49:06.193842 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:06Z","lastTransitionTime":"2025-10-05T06:49:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:06 crc kubenswrapper[4846]: I1005 06:49:06.297940 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:06 crc kubenswrapper[4846]: I1005 06:49:06.297975 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:06 crc kubenswrapper[4846]: I1005 06:49:06.298014 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:06 crc kubenswrapper[4846]: I1005 06:49:06.298031 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:06 crc kubenswrapper[4846]: I1005 06:49:06.298045 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:06Z","lastTransitionTime":"2025-10-05T06:49:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:06 crc kubenswrapper[4846]: I1005 06:49:06.401517 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:06 crc kubenswrapper[4846]: I1005 06:49:06.401586 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:06 crc kubenswrapper[4846]: I1005 06:49:06.401607 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:06 crc kubenswrapper[4846]: I1005 06:49:06.401634 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:06 crc kubenswrapper[4846]: I1005 06:49:06.401656 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:06Z","lastTransitionTime":"2025-10-05T06:49:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:06 crc kubenswrapper[4846]: I1005 06:49:06.497109 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 05 06:49:06 crc kubenswrapper[4846]: E1005 06:49:06.497441 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 05 06:49:06 crc kubenswrapper[4846]: I1005 06:49:06.497595 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9r2kt" Oct 05 06:49:06 crc kubenswrapper[4846]: E1005 06:49:06.497846 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9r2kt" podUID="59b83f4f-71f0-4b28-9275-56d5f933ec2c" Oct 05 06:49:06 crc kubenswrapper[4846]: I1005 06:49:06.504564 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:06 crc kubenswrapper[4846]: I1005 06:49:06.504601 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:06 crc kubenswrapper[4846]: I1005 06:49:06.504612 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:06 crc kubenswrapper[4846]: I1005 06:49:06.504624 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:06 crc kubenswrapper[4846]: I1005 06:49:06.504635 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:06Z","lastTransitionTime":"2025-10-05T06:49:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:06 crc kubenswrapper[4846]: I1005 06:49:06.607406 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:06 crc kubenswrapper[4846]: I1005 06:49:06.607453 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:06 crc kubenswrapper[4846]: I1005 06:49:06.607467 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:06 crc kubenswrapper[4846]: I1005 06:49:06.607486 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:06 crc kubenswrapper[4846]: I1005 06:49:06.607500 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:06Z","lastTransitionTime":"2025-10-05T06:49:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:06 crc kubenswrapper[4846]: I1005 06:49:06.709575 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:06 crc kubenswrapper[4846]: I1005 06:49:06.709603 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:06 crc kubenswrapper[4846]: I1005 06:49:06.709612 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:06 crc kubenswrapper[4846]: I1005 06:49:06.709626 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:06 crc kubenswrapper[4846]: I1005 06:49:06.709637 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:06Z","lastTransitionTime":"2025-10-05T06:49:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:06 crc kubenswrapper[4846]: I1005 06:49:06.812433 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:06 crc kubenswrapper[4846]: I1005 06:49:06.812483 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:06 crc kubenswrapper[4846]: I1005 06:49:06.812497 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:06 crc kubenswrapper[4846]: I1005 06:49:06.812517 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:06 crc kubenswrapper[4846]: I1005 06:49:06.812535 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:06Z","lastTransitionTime":"2025-10-05T06:49:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:06 crc kubenswrapper[4846]: I1005 06:49:06.915766 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:06 crc kubenswrapper[4846]: I1005 06:49:06.915830 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:06 crc kubenswrapper[4846]: I1005 06:49:06.915844 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:06 crc kubenswrapper[4846]: I1005 06:49:06.915865 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:06 crc kubenswrapper[4846]: I1005 06:49:06.915877 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:06Z","lastTransitionTime":"2025-10-05T06:49:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:07 crc kubenswrapper[4846]: I1005 06:49:07.022663 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:07 crc kubenswrapper[4846]: I1005 06:49:07.022744 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:07 crc kubenswrapper[4846]: I1005 06:49:07.022766 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:07 crc kubenswrapper[4846]: I1005 06:49:07.022797 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:07 crc kubenswrapper[4846]: I1005 06:49:07.022817 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:07Z","lastTransitionTime":"2025-10-05T06:49:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:07 crc kubenswrapper[4846]: I1005 06:49:07.125357 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:07 crc kubenswrapper[4846]: I1005 06:49:07.125411 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:07 crc kubenswrapper[4846]: I1005 06:49:07.125421 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:07 crc kubenswrapper[4846]: I1005 06:49:07.125440 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:07 crc kubenswrapper[4846]: I1005 06:49:07.125453 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:07Z","lastTransitionTime":"2025-10-05T06:49:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:07 crc kubenswrapper[4846]: I1005 06:49:07.228817 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:07 crc kubenswrapper[4846]: I1005 06:49:07.228887 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:07 crc kubenswrapper[4846]: I1005 06:49:07.228909 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:07 crc kubenswrapper[4846]: I1005 06:49:07.228935 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:07 crc kubenswrapper[4846]: I1005 06:49:07.228954 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:07Z","lastTransitionTime":"2025-10-05T06:49:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:07 crc kubenswrapper[4846]: I1005 06:49:07.332414 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:07 crc kubenswrapper[4846]: I1005 06:49:07.332472 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:07 crc kubenswrapper[4846]: I1005 06:49:07.332488 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:07 crc kubenswrapper[4846]: I1005 06:49:07.332508 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:07 crc kubenswrapper[4846]: I1005 06:49:07.332524 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:07Z","lastTransitionTime":"2025-10-05T06:49:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:07 crc kubenswrapper[4846]: I1005 06:49:07.435256 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:07 crc kubenswrapper[4846]: I1005 06:49:07.435324 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:07 crc kubenswrapper[4846]: I1005 06:49:07.435343 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:07 crc kubenswrapper[4846]: I1005 06:49:07.435370 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:07 crc kubenswrapper[4846]: I1005 06:49:07.435389 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:07Z","lastTransitionTime":"2025-10-05T06:49:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:07 crc kubenswrapper[4846]: I1005 06:49:07.496977 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 05 06:49:07 crc kubenswrapper[4846]: I1005 06:49:07.497122 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 05 06:49:07 crc kubenswrapper[4846]: E1005 06:49:07.497163 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 05 06:49:07 crc kubenswrapper[4846]: E1005 06:49:07.497431 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 05 06:49:07 crc kubenswrapper[4846]: I1005 06:49:07.515025 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Oct 05 06:49:07 crc kubenswrapper[4846]: I1005 06:49:07.538825 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:07 crc kubenswrapper[4846]: I1005 06:49:07.538873 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:07 crc kubenswrapper[4846]: I1005 06:49:07.538883 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:07 crc kubenswrapper[4846]: I1005 06:49:07.538902 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:07 crc kubenswrapper[4846]: I1005 06:49:07.538913 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:07Z","lastTransitionTime":"2025-10-05T06:49:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:07 crc kubenswrapper[4846]: I1005 06:49:07.641394 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:07 crc kubenswrapper[4846]: I1005 06:49:07.641438 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:07 crc kubenswrapper[4846]: I1005 06:49:07.641451 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:07 crc kubenswrapper[4846]: I1005 06:49:07.641469 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:07 crc kubenswrapper[4846]: I1005 06:49:07.641484 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:07Z","lastTransitionTime":"2025-10-05T06:49:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:07 crc kubenswrapper[4846]: I1005 06:49:07.744331 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:07 crc kubenswrapper[4846]: I1005 06:49:07.744388 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:07 crc kubenswrapper[4846]: I1005 06:49:07.744400 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:07 crc kubenswrapper[4846]: I1005 06:49:07.744421 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:07 crc kubenswrapper[4846]: I1005 06:49:07.744434 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:07Z","lastTransitionTime":"2025-10-05T06:49:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:07 crc kubenswrapper[4846]: I1005 06:49:07.847814 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:07 crc kubenswrapper[4846]: I1005 06:49:07.847861 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:07 crc kubenswrapper[4846]: I1005 06:49:07.847874 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:07 crc kubenswrapper[4846]: I1005 06:49:07.847892 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:07 crc kubenswrapper[4846]: I1005 06:49:07.847905 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:07Z","lastTransitionTime":"2025-10-05T06:49:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:07 crc kubenswrapper[4846]: I1005 06:49:07.950879 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:07 crc kubenswrapper[4846]: I1005 06:49:07.950923 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:07 crc kubenswrapper[4846]: I1005 06:49:07.950939 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:07 crc kubenswrapper[4846]: I1005 06:49:07.950958 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:07 crc kubenswrapper[4846]: I1005 06:49:07.950972 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:07Z","lastTransitionTime":"2025-10-05T06:49:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:08 crc kubenswrapper[4846]: I1005 06:49:08.054032 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:08 crc kubenswrapper[4846]: I1005 06:49:08.054071 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:08 crc kubenswrapper[4846]: I1005 06:49:08.054081 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:08 crc kubenswrapper[4846]: I1005 06:49:08.054095 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:08 crc kubenswrapper[4846]: I1005 06:49:08.054116 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:08Z","lastTransitionTime":"2025-10-05T06:49:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:08 crc kubenswrapper[4846]: I1005 06:49:08.158525 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:08 crc kubenswrapper[4846]: I1005 06:49:08.158575 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:08 crc kubenswrapper[4846]: I1005 06:49:08.158586 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:08 crc kubenswrapper[4846]: I1005 06:49:08.158607 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:08 crc kubenswrapper[4846]: I1005 06:49:08.158619 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:08Z","lastTransitionTime":"2025-10-05T06:49:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:08 crc kubenswrapper[4846]: I1005 06:49:08.262084 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:08 crc kubenswrapper[4846]: I1005 06:49:08.262444 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:08 crc kubenswrapper[4846]: I1005 06:49:08.262507 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:08 crc kubenswrapper[4846]: I1005 06:49:08.262570 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:08 crc kubenswrapper[4846]: I1005 06:49:08.262641 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:08Z","lastTransitionTime":"2025-10-05T06:49:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:08 crc kubenswrapper[4846]: I1005 06:49:08.369507 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:08 crc kubenswrapper[4846]: I1005 06:49:08.369553 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:08 crc kubenswrapper[4846]: I1005 06:49:08.369570 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:08 crc kubenswrapper[4846]: I1005 06:49:08.369796 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:08 crc kubenswrapper[4846]: I1005 06:49:08.369814 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:08Z","lastTransitionTime":"2025-10-05T06:49:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:08 crc kubenswrapper[4846]: I1005 06:49:08.472635 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:08 crc kubenswrapper[4846]: I1005 06:49:08.472681 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:08 crc kubenswrapper[4846]: I1005 06:49:08.472689 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:08 crc kubenswrapper[4846]: I1005 06:49:08.472706 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:08 crc kubenswrapper[4846]: I1005 06:49:08.472717 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:08Z","lastTransitionTime":"2025-10-05T06:49:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:08 crc kubenswrapper[4846]: I1005 06:49:08.497376 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9r2kt" Oct 05 06:49:08 crc kubenswrapper[4846]: I1005 06:49:08.497621 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 05 06:49:08 crc kubenswrapper[4846]: E1005 06:49:08.497857 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9r2kt" podUID="59b83f4f-71f0-4b28-9275-56d5f933ec2c" Oct 05 06:49:08 crc kubenswrapper[4846]: E1005 06:49:08.497999 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 05 06:49:08 crc kubenswrapper[4846]: I1005 06:49:08.511826 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ff10c31459b3b3724ee7b769bd6f72f9ed80cce5419b6adcdf69a3239b70756\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:49:08Z is after 2025-08-24T17:21:41Z" Oct 05 06:49:08 crc kubenswrapper[4846]: I1005 06:49:08.524946 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://347f900743b7bba3daf92f21c7eba0625e8c774f50057bc9472f0c82119325e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:49:08Z is after 2025-08-24T17:21:41Z" Oct 05 06:49:08 crc kubenswrapper[4846]: I1005 06:49:08.543285 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56f1fe1f9f06f0249cf937a6963fda82467732dc056ca2b2189be6d8dd9d4a05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7212ce5f42c1fb55146abb190a557a504b43f16af6df84f90e1073567a329ec5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:49:08Z is after 2025-08-24T17:21:41Z" Oct 05 06:49:08 crc kubenswrapper[4846]: I1005 06:49:08.556035 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-7wm5q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9bc7744f-d700-450f-8b8d-98140877fee9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90379e74c544fccca44f7acdc228ff378154390e211ac801b79f8d074012241e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h5p4q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-7wm5q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:49:08Z is after 2025-08-24T17:21:41Z" Oct 05 06:49:08 crc kubenswrapper[4846]: I1005 06:49:08.571730 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"78b204fd-1034-4ea1-bcae-e5c876594a39\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b94cc50fe15d58f375f8492067811be16278cdf13ee3d80cfed4682279157b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5e5cf9125d1d413c91666744aeab06980b44420dbf0341b165b05787527944b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://70f0d5b436d18a78a2d8e2767143ba1a3104332d86bd63f91d23cd8e77a952aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c236b4ed7f039a221baac043246353eecbb6812c1e6bf6836a5fa96ec8f5abdb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c236b4ed7f039a221baac043246353eecbb6812c1e6bf6836a5fa96ec8f5abdb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:47:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:47:59Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:47:58Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:49:08Z is after 2025-08-24T17:21:41Z" Oct 05 06:49:08 crc kubenswrapper[4846]: I1005 06:49:08.575424 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:08 crc kubenswrapper[4846]: I1005 06:49:08.575474 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:08 crc kubenswrapper[4846]: I1005 06:49:08.575492 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:08 crc kubenswrapper[4846]: I1005 06:49:08.575516 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:08 crc kubenswrapper[4846]: I1005 06:49:08.575535 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:08Z","lastTransitionTime":"2025-10-05T06:49:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:08 crc kubenswrapper[4846]: I1005 06:49:08.591374 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"775b5702-a131-4d6c-852d-91bfa056cac5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0839c1765df2c68d2554e9c66d289505e559a9672caf8069feeff2c4fdfd2171\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15dbaf2445704da0987f256f81dff2ee95664988541fa48ce94898afeadab2e8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd3b3ab323c6d5153054aa04af2d91f7b9e8436ff1e2f75f3594b57831b2a416\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://67c46f5f0a3ace95bd2e9479c5561126664e18b148c5eceea5bf5efc27e81a67\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c52e076eddfd307110b1f6de00a169b81af8f2785af7becef2a820ce363a5861\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-05T06:48:14Z\\\",\\\"message\\\":\\\"le observer\\\\nW1005 06:48:14.337870 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1005 06:48:14.338086 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1005 06:48:14.339389 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1467752215/tls.crt::/tmp/serving-cert-1467752215/tls.key\\\\\\\"\\\\nI1005 06:48:14.864345 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1005 06:48:14.868087 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1005 06:48:14.868114 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1005 06:48:14.868140 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1005 06:48:14.868147 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1005 06:48:14.877354 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1005 06:48:14.877412 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1005 06:48:14.877456 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1005 06:48:14.877471 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1005 06:48:14.877490 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1005 06:48:14.877501 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1005 06:48:14.877507 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1005 06:48:14.877515 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1005 06:48:14.880326 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:13Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4a091716ff935adb42ae2e88048619fd347bd174f64e5f9c97a01f81690e57f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:01Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dc5411b642f99b13d514d2f21defad0c394aeeb82a966f72300076617e2586c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5dc5411b642f99b13d514d2f21defad0c394aeeb82a966f72300076617e2586c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:47:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:47:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:47:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:49:08Z is after 2025-08-24T17:21:41Z" Oct 05 06:49:08 crc kubenswrapper[4846]: I1005 06:49:08.606126 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:49:08Z is after 2025-08-24T17:21:41Z" Oct 05 06:49:08 crc kubenswrapper[4846]: I1005 06:49:08.622537 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-r7ztb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a5119cd9-af70-4324-a955-978306ab0b20\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06231a87429ec292cb57ebd88cda7ad547bcb32e80343b776833621a11c28ad1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jfcj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-r7ztb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:49:08Z is after 2025-08-24T17:21:41Z" Oct 05 06:49:08 crc kubenswrapper[4846]: I1005 06:49:08.634771 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-9r2kt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"59b83f4f-71f0-4b28-9275-56d5f933ec2c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tdwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tdwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:38Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-9r2kt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:49:08Z is after 2025-08-24T17:21:41Z" Oct 05 06:49:08 crc kubenswrapper[4846]: I1005 06:49:08.646747 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c21154c6-7c11-47e7-b1f0-6a5fe33d16b1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e13964315903099b3d7e3dc8c143fe58bb9f9654214702ba51581ac4044414d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef573b754a80e5148b8caa71b4f22c461f57dfa415394a5bccda9771b7684e4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef573b754a80e5148b8caa71b4f22c461f57dfa415394a5bccda9771b7684e4c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:47:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:47:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:47:58Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:49:08Z is after 2025-08-24T17:21:41Z" Oct 05 06:49:08 crc kubenswrapper[4846]: I1005 06:49:08.659473 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:49:08Z is after 2025-08-24T17:21:41Z" Oct 05 06:49:08 crc kubenswrapper[4846]: I1005 06:49:08.673341 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-sfmxd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1dbb13ac-136c-4a1c-9229-b4e8ac4f9c71\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b0af6a6e735f7b7877c177c1d53d1a8dc12c4713a8059c4766fc858af31c9a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://226d977b596d07aeda0486452e2a8605949ed4d2453b21f56174937c395d6f36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://226d977b596d07aeda0486452e2a8605949ed4d2453b21f56174937c395d6f36\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://99c07a4bdb3059d3a99ffdaf794d668d52cffd9e5e3bb8945204dd6d1f8c8328\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99c07a4bdb3059d3a99ffdaf794d668d52cffd9e5e3bb8945204dd6d1f8c8328\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27095272bf4f31d63a78a436636de0aae12f7103597da944f24ed7a74329ee3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27095272bf4f31d63a78a436636de0aae12f7103597da944f24ed7a74329ee3f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1211571af66d8a94302b65feccf30768ba82b7cbad918e44dd405787dce484cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1211571af66d8a94302b65feccf30768ba82b7cbad918e44dd405787dce484cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27631f90d528900fa180c15cfce80e96d4cdba07743eb6d847e2e822d9917218\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27631f90d528900fa180c15cfce80e96d4cdba07743eb6d847e2e822d9917218\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f345693e9aeb1f4dc196349d3354f3c833a2b0415a693298b1cfccfa2b0d4459\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f345693e9aeb1f4dc196349d3354f3c833a2b0415a693298b1cfccfa2b0d4459\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-sfmxd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:49:08Z is after 2025-08-24T17:21:41Z" Oct 05 06:49:08 crc kubenswrapper[4846]: I1005 06:49:08.681287 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:08 crc kubenswrapper[4846]: I1005 06:49:08.681330 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:08 crc kubenswrapper[4846]: I1005 06:49:08.681465 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:08 crc kubenswrapper[4846]: I1005 06:49:08.681490 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:08 crc kubenswrapper[4846]: I1005 06:49:08.681501 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:08Z","lastTransitionTime":"2025-10-05T06:49:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:08 crc kubenswrapper[4846]: I1005 06:49:08.696378 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-txzlk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"93a6ace4-8d64-44d7-9637-457d3af6543c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ad1940d07e37217089642478ece9ccdaca74784738a400929a9f5359d158c01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a6f6b2c32dd3da5f9fd5e6971c1773e81a9a24184b7aa5b543cec879001ae53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e9b0219ecbc7e60905d6de3e67a2c8dc9af1017b4aed9538ee0c58c861f2438\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af5e5a4da9ad7371d5a24633e8268f12e3a7d419a8c93b11ddbcc006d8339c6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8da85222775abc7a772f6a8ff8255f1a0d5b15807ba177a7e3e2b2d01a4f24b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1409651de8c897804f87be03d016c138f8acf27e662dce265a93e15c1848adcf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd084bb9806ece6529db0866cee69e6a3152f0da7e0f13fc2e4adcbe2f55962c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd084bb9806ece6529db0866cee69e6a3152f0da7e0f13fc2e4adcbe2f55962c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-05T06:48:50Z\\\",\\\"message\\\":\\\":50.819394 6534 handler.go:208] Removed *v1.Node event handler 2\\\\nI1005 06:48:50.819434 6534 handler.go:208] Removed *v1.Node event handler 7\\\\nI1005 06:48:50.819451 6534 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1005 06:48:50.819624 6534 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1005 06:48:50.819659 6534 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1005 06:48:50.819711 6534 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1005 06:48:50.819735 6534 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1005 06:48:50.820556 6534 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1005 06:48:50.820605 6534 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1005 06:48:50.820649 6534 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1005 06:48:50.820664 6534 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1005 06:48:50.820706 6534 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1005 06:48:50.820711 6534 factory.go:656] Stopping watch factory\\\\nI1005 06:48:50.820731 6534 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1005 06:48:50.820750 6534 ovnkube.go:599] Stopped ovnkube\\\\nI1005 06:48:50.820796 6534 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1005 06:48:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:49Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-txzlk_openshift-ovn-kubernetes(93a6ace4-8d64-44d7-9637-457d3af6543c)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0df0ebcd87dc5af4b17736b43ec7b2395968b19ed16e775366af14e2b2d49c52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff3c6545ff3d2436da27bbd223b950b30b0c69a3e7f3034cb0e61380cb38d338\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ff3c6545ff3d2436da27bbd223b950b30b0c69a3e7f3034cb0e61380cb38d338\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:24Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-txzlk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:49:08Z is after 2025-08-24T17:21:41Z" Oct 05 06:49:08 crc kubenswrapper[4846]: I1005 06:49:08.711540 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-gklb6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"698ab882-bb26-4b4c-81e9-8305a44fe068\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a6edca88ea244a58f8cf6cc431110a960c6bfd892812cd0d00fa3f97be3bc67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x6d6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c191538ab797b8056820de7da1f8e93ea18873891a32382b8a59c9af762e103\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x6d6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:37Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-gklb6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:49:08Z is after 2025-08-24T17:21:41Z" Oct 05 06:49:08 crc kubenswrapper[4846]: I1005 06:49:08.734380 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c482bb97-fd50-40cf-96ea-f806c3f01125\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bfe45d30e4fa141541aca1d46bc5cc1da0a7b4a5eb1fc9bb2caa5b7fdcb0b264\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://496a4782faa30a9a20ff63038141600cd635f12902d1cf17989cd067fcdfe929\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://281be077869a6aa1ca815770871a14413d4a97e97aeb550fd5cd215cba88b21f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6771c7215d2f5dd5838d17888375a4768b9d1755fc2e21bda1bbfbb2eb1dbb95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c17a711c2c7473b076a3944ffe7c700808d05cba582fbd55c5c2ac8ff73838a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44a9f740f0a06c8bf928e2a9d1d5979217558c31e73a333631baccebf00fd5d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://44a9f740f0a06c8bf928e2a9d1d5979217558c31e73a333631baccebf00fd5d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:47:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:47:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://48ca0ba6ca6092bfcb5b153220c054f02cc36fb22707c42e021f06d329e18e2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48ca0ba6ca6092bfcb5b153220c054f02cc36fb22707c42e021f06d329e18e2e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:00Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4006d0cea128675c8885d879171204abffc5e11ea0664750a12b4bc7344ccb2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4006d0cea128675c8885d879171204abffc5e11ea0664750a12b4bc7344ccb2e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:47:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:49:08Z is after 2025-08-24T17:21:41Z" Oct 05 06:49:08 crc kubenswrapper[4846]: I1005 06:49:08.749857 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:49:08Z is after 2025-08-24T17:21:41Z" Oct 05 06:49:08 crc kubenswrapper[4846]: I1005 06:49:08.762873 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e67d0e447a8a6a381259e25d25badc791194e1291f56b1cd77bd63989b20408\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dkhl7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://516f88f4990f5962d8b694a3d889259cc5508a309919870191c2e1331640c973\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dkhl7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-fscvf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:49:08Z is after 2025-08-24T17:21:41Z" Oct 05 06:49:08 crc kubenswrapper[4846]: I1005 06:49:08.774280 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-w2sj8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32931fce-7374-4cab-a2e0-79d96a842274\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0df535d65d39de6c9a3dfb931bafc4469adc22b72f7a7903beda4fef8cd48138\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9rtxf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:27Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-w2sj8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:49:08Z is after 2025-08-24T17:21:41Z" Oct 05 06:49:08 crc kubenswrapper[4846]: I1005 06:49:08.784050 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:08 crc kubenswrapper[4846]: I1005 06:49:08.784097 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:08 crc kubenswrapper[4846]: I1005 06:49:08.784108 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:08 crc kubenswrapper[4846]: I1005 06:49:08.784124 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:08 crc kubenswrapper[4846]: I1005 06:49:08.784135 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:08Z","lastTransitionTime":"2025-10-05T06:49:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:08 crc kubenswrapper[4846]: I1005 06:49:08.790743 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c9f812b-ecc8-4985-a670-53fa80fbb499\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2b5ee9781c01d6aec3ac778efef1067b6c3bfe4ed37431bdebe188fcde0165f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:47:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3fb4717692aa1fec0012b514ef8d18d1ca3707cfa4bb8248bb93841a0a40f11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:47:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0b6a7afe3094f1c92093c05c28344897a615d590c2fbb507c87e18a44c6c831\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://640ade2c733d7916ec213a201a1dd614d261c69eab654c80100efbc99a768995\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:47:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:49:08Z is after 2025-08-24T17:21:41Z" Oct 05 06:49:08 crc kubenswrapper[4846]: I1005 06:49:08.887686 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:08 crc kubenswrapper[4846]: I1005 06:49:08.887800 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:08 crc kubenswrapper[4846]: I1005 06:49:08.887819 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:08 crc kubenswrapper[4846]: I1005 06:49:08.887844 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:08 crc kubenswrapper[4846]: I1005 06:49:08.887864 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:08Z","lastTransitionTime":"2025-10-05T06:49:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:08 crc kubenswrapper[4846]: I1005 06:49:08.990374 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:08 crc kubenswrapper[4846]: I1005 06:49:08.990428 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:08 crc kubenswrapper[4846]: I1005 06:49:08.990440 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:08 crc kubenswrapper[4846]: I1005 06:49:08.990458 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:08 crc kubenswrapper[4846]: I1005 06:49:08.990472 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:08Z","lastTransitionTime":"2025-10-05T06:49:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:09 crc kubenswrapper[4846]: I1005 06:49:09.094189 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:09 crc kubenswrapper[4846]: I1005 06:49:09.094232 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:09 crc kubenswrapper[4846]: I1005 06:49:09.094242 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:09 crc kubenswrapper[4846]: I1005 06:49:09.094258 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:09 crc kubenswrapper[4846]: I1005 06:49:09.094268 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:09Z","lastTransitionTime":"2025-10-05T06:49:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:09 crc kubenswrapper[4846]: I1005 06:49:09.196956 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:09 crc kubenswrapper[4846]: I1005 06:49:09.196999 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:09 crc kubenswrapper[4846]: I1005 06:49:09.197010 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:09 crc kubenswrapper[4846]: I1005 06:49:09.197027 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:09 crc kubenswrapper[4846]: I1005 06:49:09.197038 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:09Z","lastTransitionTime":"2025-10-05T06:49:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:09 crc kubenswrapper[4846]: I1005 06:49:09.300389 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:09 crc kubenswrapper[4846]: I1005 06:49:09.300437 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:09 crc kubenswrapper[4846]: I1005 06:49:09.300449 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:09 crc kubenswrapper[4846]: I1005 06:49:09.300474 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:09 crc kubenswrapper[4846]: I1005 06:49:09.300488 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:09Z","lastTransitionTime":"2025-10-05T06:49:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:09 crc kubenswrapper[4846]: I1005 06:49:09.403144 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:09 crc kubenswrapper[4846]: I1005 06:49:09.403210 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:09 crc kubenswrapper[4846]: I1005 06:49:09.403227 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:09 crc kubenswrapper[4846]: I1005 06:49:09.403245 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:09 crc kubenswrapper[4846]: I1005 06:49:09.403256 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:09Z","lastTransitionTime":"2025-10-05T06:49:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:09 crc kubenswrapper[4846]: I1005 06:49:09.497463 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 05 06:49:09 crc kubenswrapper[4846]: I1005 06:49:09.497545 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 05 06:49:09 crc kubenswrapper[4846]: E1005 06:49:09.497617 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 05 06:49:09 crc kubenswrapper[4846]: E1005 06:49:09.497749 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 05 06:49:09 crc kubenswrapper[4846]: I1005 06:49:09.505683 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:09 crc kubenswrapper[4846]: I1005 06:49:09.505715 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:09 crc kubenswrapper[4846]: I1005 06:49:09.505724 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:09 crc kubenswrapper[4846]: I1005 06:49:09.505737 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:09 crc kubenswrapper[4846]: I1005 06:49:09.505747 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:09Z","lastTransitionTime":"2025-10-05T06:49:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:09 crc kubenswrapper[4846]: I1005 06:49:09.609168 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:09 crc kubenswrapper[4846]: I1005 06:49:09.609293 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:09 crc kubenswrapper[4846]: I1005 06:49:09.609322 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:09 crc kubenswrapper[4846]: I1005 06:49:09.609356 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:09 crc kubenswrapper[4846]: I1005 06:49:09.609382 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:09Z","lastTransitionTime":"2025-10-05T06:49:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:09 crc kubenswrapper[4846]: I1005 06:49:09.711764 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:09 crc kubenswrapper[4846]: I1005 06:49:09.711821 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:09 crc kubenswrapper[4846]: I1005 06:49:09.711839 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:09 crc kubenswrapper[4846]: I1005 06:49:09.711863 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:09 crc kubenswrapper[4846]: I1005 06:49:09.711882 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:09Z","lastTransitionTime":"2025-10-05T06:49:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:09 crc kubenswrapper[4846]: I1005 06:49:09.814897 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:09 crc kubenswrapper[4846]: I1005 06:49:09.814945 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:09 crc kubenswrapper[4846]: I1005 06:49:09.814957 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:09 crc kubenswrapper[4846]: I1005 06:49:09.814972 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:09 crc kubenswrapper[4846]: I1005 06:49:09.814983 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:09Z","lastTransitionTime":"2025-10-05T06:49:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:09 crc kubenswrapper[4846]: I1005 06:49:09.917775 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:09 crc kubenswrapper[4846]: I1005 06:49:09.917832 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:09 crc kubenswrapper[4846]: I1005 06:49:09.917847 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:09 crc kubenswrapper[4846]: I1005 06:49:09.917870 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:09 crc kubenswrapper[4846]: I1005 06:49:09.917887 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:09Z","lastTransitionTime":"2025-10-05T06:49:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:10 crc kubenswrapper[4846]: I1005 06:49:10.020031 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:10 crc kubenswrapper[4846]: I1005 06:49:10.020077 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:10 crc kubenswrapper[4846]: I1005 06:49:10.020087 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:10 crc kubenswrapper[4846]: I1005 06:49:10.020105 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:10 crc kubenswrapper[4846]: I1005 06:49:10.020116 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:10Z","lastTransitionTime":"2025-10-05T06:49:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:10 crc kubenswrapper[4846]: I1005 06:49:10.123451 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:10 crc kubenswrapper[4846]: I1005 06:49:10.123494 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:10 crc kubenswrapper[4846]: I1005 06:49:10.123504 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:10 crc kubenswrapper[4846]: I1005 06:49:10.123519 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:10 crc kubenswrapper[4846]: I1005 06:49:10.123534 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:10Z","lastTransitionTime":"2025-10-05T06:49:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:10 crc kubenswrapper[4846]: I1005 06:49:10.226689 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:10 crc kubenswrapper[4846]: I1005 06:49:10.226733 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:10 crc kubenswrapper[4846]: I1005 06:49:10.226742 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:10 crc kubenswrapper[4846]: I1005 06:49:10.226781 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:10 crc kubenswrapper[4846]: I1005 06:49:10.226792 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:10Z","lastTransitionTime":"2025-10-05T06:49:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:10 crc kubenswrapper[4846]: I1005 06:49:10.328947 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:10 crc kubenswrapper[4846]: I1005 06:49:10.329006 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:10 crc kubenswrapper[4846]: I1005 06:49:10.329017 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:10 crc kubenswrapper[4846]: I1005 06:49:10.329031 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:10 crc kubenswrapper[4846]: I1005 06:49:10.329041 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:10Z","lastTransitionTime":"2025-10-05T06:49:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:10 crc kubenswrapper[4846]: I1005 06:49:10.431467 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:10 crc kubenswrapper[4846]: I1005 06:49:10.431529 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:10 crc kubenswrapper[4846]: I1005 06:49:10.431544 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:10 crc kubenswrapper[4846]: I1005 06:49:10.431572 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:10 crc kubenswrapper[4846]: I1005 06:49:10.431591 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:10Z","lastTransitionTime":"2025-10-05T06:49:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:10 crc kubenswrapper[4846]: I1005 06:49:10.496759 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 05 06:49:10 crc kubenswrapper[4846]: I1005 06:49:10.496841 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9r2kt" Oct 05 06:49:10 crc kubenswrapper[4846]: E1005 06:49:10.497504 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 05 06:49:10 crc kubenswrapper[4846]: E1005 06:49:10.501937 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9r2kt" podUID="59b83f4f-71f0-4b28-9275-56d5f933ec2c" Oct 05 06:49:10 crc kubenswrapper[4846]: I1005 06:49:10.534308 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:10 crc kubenswrapper[4846]: I1005 06:49:10.534369 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:10 crc kubenswrapper[4846]: I1005 06:49:10.534386 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:10 crc kubenswrapper[4846]: I1005 06:49:10.534408 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:10 crc kubenswrapper[4846]: I1005 06:49:10.534426 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:10Z","lastTransitionTime":"2025-10-05T06:49:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:10 crc kubenswrapper[4846]: I1005 06:49:10.638517 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:10 crc kubenswrapper[4846]: I1005 06:49:10.638580 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:10 crc kubenswrapper[4846]: I1005 06:49:10.638600 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:10 crc kubenswrapper[4846]: I1005 06:49:10.638622 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:10 crc kubenswrapper[4846]: I1005 06:49:10.638642 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:10Z","lastTransitionTime":"2025-10-05T06:49:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:10 crc kubenswrapper[4846]: I1005 06:49:10.741253 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:10 crc kubenswrapper[4846]: I1005 06:49:10.741306 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:10 crc kubenswrapper[4846]: I1005 06:49:10.741317 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:10 crc kubenswrapper[4846]: I1005 06:49:10.741337 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:10 crc kubenswrapper[4846]: I1005 06:49:10.741350 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:10Z","lastTransitionTime":"2025-10-05T06:49:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:10 crc kubenswrapper[4846]: I1005 06:49:10.844071 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:10 crc kubenswrapper[4846]: I1005 06:49:10.844135 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:10 crc kubenswrapper[4846]: I1005 06:49:10.844153 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:10 crc kubenswrapper[4846]: I1005 06:49:10.844210 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:10 crc kubenswrapper[4846]: I1005 06:49:10.844235 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:10Z","lastTransitionTime":"2025-10-05T06:49:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:10 crc kubenswrapper[4846]: I1005 06:49:10.855112 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/59b83f4f-71f0-4b28-9275-56d5f933ec2c-metrics-certs\") pod \"network-metrics-daemon-9r2kt\" (UID: \"59b83f4f-71f0-4b28-9275-56d5f933ec2c\") " pod="openshift-multus/network-metrics-daemon-9r2kt" Oct 05 06:49:10 crc kubenswrapper[4846]: E1005 06:49:10.855387 4846 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 05 06:49:10 crc kubenswrapper[4846]: E1005 06:49:10.855512 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/59b83f4f-71f0-4b28-9275-56d5f933ec2c-metrics-certs podName:59b83f4f-71f0-4b28-9275-56d5f933ec2c nodeName:}" failed. No retries permitted until 2025-10-05 06:49:42.855477112 +0000 UTC m=+105.096329917 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/59b83f4f-71f0-4b28-9275-56d5f933ec2c-metrics-certs") pod "network-metrics-daemon-9r2kt" (UID: "59b83f4f-71f0-4b28-9275-56d5f933ec2c") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 05 06:49:10 crc kubenswrapper[4846]: I1005 06:49:10.947985 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:10 crc kubenswrapper[4846]: I1005 06:49:10.948064 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:10 crc kubenswrapper[4846]: I1005 06:49:10.948085 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:10 crc kubenswrapper[4846]: I1005 06:49:10.948111 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:10 crc kubenswrapper[4846]: I1005 06:49:10.948128 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:10Z","lastTransitionTime":"2025-10-05T06:49:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:11 crc kubenswrapper[4846]: I1005 06:49:11.050661 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:11 crc kubenswrapper[4846]: I1005 06:49:11.050713 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:11 crc kubenswrapper[4846]: I1005 06:49:11.050726 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:11 crc kubenswrapper[4846]: I1005 06:49:11.050749 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:11 crc kubenswrapper[4846]: I1005 06:49:11.050768 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:11Z","lastTransitionTime":"2025-10-05T06:49:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:11 crc kubenswrapper[4846]: I1005 06:49:11.155256 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:11 crc kubenswrapper[4846]: I1005 06:49:11.155317 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:11 crc kubenswrapper[4846]: I1005 06:49:11.155340 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:11 crc kubenswrapper[4846]: I1005 06:49:11.155371 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:11 crc kubenswrapper[4846]: I1005 06:49:11.155392 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:11Z","lastTransitionTime":"2025-10-05T06:49:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:11 crc kubenswrapper[4846]: I1005 06:49:11.259261 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:11 crc kubenswrapper[4846]: I1005 06:49:11.259338 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:11 crc kubenswrapper[4846]: I1005 06:49:11.259355 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:11 crc kubenswrapper[4846]: I1005 06:49:11.259381 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:11 crc kubenswrapper[4846]: I1005 06:49:11.259399 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:11Z","lastTransitionTime":"2025-10-05T06:49:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:11 crc kubenswrapper[4846]: I1005 06:49:11.362653 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:11 crc kubenswrapper[4846]: I1005 06:49:11.362697 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:11 crc kubenswrapper[4846]: I1005 06:49:11.362707 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:11 crc kubenswrapper[4846]: I1005 06:49:11.362727 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:11 crc kubenswrapper[4846]: I1005 06:49:11.362738 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:11Z","lastTransitionTime":"2025-10-05T06:49:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:11 crc kubenswrapper[4846]: I1005 06:49:11.465003 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:11 crc kubenswrapper[4846]: I1005 06:49:11.465046 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:11 crc kubenswrapper[4846]: I1005 06:49:11.465054 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:11 crc kubenswrapper[4846]: I1005 06:49:11.465068 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:11 crc kubenswrapper[4846]: I1005 06:49:11.465078 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:11Z","lastTransitionTime":"2025-10-05T06:49:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:11 crc kubenswrapper[4846]: I1005 06:49:11.497395 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 05 06:49:11 crc kubenswrapper[4846]: I1005 06:49:11.497395 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 05 06:49:11 crc kubenswrapper[4846]: E1005 06:49:11.497708 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 05 06:49:11 crc kubenswrapper[4846]: E1005 06:49:11.497781 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 05 06:49:11 crc kubenswrapper[4846]: I1005 06:49:11.567031 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:11 crc kubenswrapper[4846]: I1005 06:49:11.567087 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:11 crc kubenswrapper[4846]: I1005 06:49:11.567099 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:11 crc kubenswrapper[4846]: I1005 06:49:11.567141 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:11 crc kubenswrapper[4846]: I1005 06:49:11.567154 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:11Z","lastTransitionTime":"2025-10-05T06:49:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:11 crc kubenswrapper[4846]: I1005 06:49:11.670456 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:11 crc kubenswrapper[4846]: I1005 06:49:11.670506 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:11 crc kubenswrapper[4846]: I1005 06:49:11.670518 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:11 crc kubenswrapper[4846]: I1005 06:49:11.670538 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:11 crc kubenswrapper[4846]: I1005 06:49:11.670551 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:11Z","lastTransitionTime":"2025-10-05T06:49:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:11 crc kubenswrapper[4846]: I1005 06:49:11.773743 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:11 crc kubenswrapper[4846]: I1005 06:49:11.773846 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:11 crc kubenswrapper[4846]: I1005 06:49:11.773872 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:11 crc kubenswrapper[4846]: I1005 06:49:11.773901 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:11 crc kubenswrapper[4846]: I1005 06:49:11.773922 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:11Z","lastTransitionTime":"2025-10-05T06:49:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:11 crc kubenswrapper[4846]: I1005 06:49:11.877357 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:11 crc kubenswrapper[4846]: I1005 06:49:11.877454 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:11 crc kubenswrapper[4846]: I1005 06:49:11.877474 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:11 crc kubenswrapper[4846]: I1005 06:49:11.877502 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:11 crc kubenswrapper[4846]: I1005 06:49:11.877521 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:11Z","lastTransitionTime":"2025-10-05T06:49:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:11 crc kubenswrapper[4846]: I1005 06:49:11.980779 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:11 crc kubenswrapper[4846]: I1005 06:49:11.980839 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:11 crc kubenswrapper[4846]: I1005 06:49:11.980855 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:11 crc kubenswrapper[4846]: I1005 06:49:11.980880 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:11 crc kubenswrapper[4846]: I1005 06:49:11.980899 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:11Z","lastTransitionTime":"2025-10-05T06:49:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:12 crc kubenswrapper[4846]: I1005 06:49:12.043501 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-7wm5q_9bc7744f-d700-450f-8b8d-98140877fee9/kube-multus/0.log" Oct 05 06:49:12 crc kubenswrapper[4846]: I1005 06:49:12.043576 4846 generic.go:334] "Generic (PLEG): container finished" podID="9bc7744f-d700-450f-8b8d-98140877fee9" containerID="90379e74c544fccca44f7acdc228ff378154390e211ac801b79f8d074012241e" exitCode=1 Oct 05 06:49:12 crc kubenswrapper[4846]: I1005 06:49:12.043627 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-7wm5q" event={"ID":"9bc7744f-d700-450f-8b8d-98140877fee9","Type":"ContainerDied","Data":"90379e74c544fccca44f7acdc228ff378154390e211ac801b79f8d074012241e"} Oct 05 06:49:12 crc kubenswrapper[4846]: I1005 06:49:12.044629 4846 scope.go:117] "RemoveContainer" containerID="90379e74c544fccca44f7acdc228ff378154390e211ac801b79f8d074012241e" Oct 05 06:49:12 crc kubenswrapper[4846]: I1005 06:49:12.063385 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-9r2kt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"59b83f4f-71f0-4b28-9275-56d5f933ec2c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tdwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tdwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:38Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-9r2kt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:49:12Z is after 2025-08-24T17:21:41Z" Oct 05 06:49:12 crc kubenswrapper[4846]: I1005 06:49:12.079076 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c21154c6-7c11-47e7-b1f0-6a5fe33d16b1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e13964315903099b3d7e3dc8c143fe58bb9f9654214702ba51581ac4044414d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef573b754a80e5148b8caa71b4f22c461f57dfa415394a5bccda9771b7684e4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef573b754a80e5148b8caa71b4f22c461f57dfa415394a5bccda9771b7684e4c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:47:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:47:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:47:58Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:49:12Z is after 2025-08-24T17:21:41Z" Oct 05 06:49:12 crc kubenswrapper[4846]: I1005 06:49:12.084393 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:12 crc kubenswrapper[4846]: I1005 06:49:12.084456 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:12 crc kubenswrapper[4846]: I1005 06:49:12.084475 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:12 crc kubenswrapper[4846]: I1005 06:49:12.084501 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:12 crc kubenswrapper[4846]: I1005 06:49:12.084521 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:12Z","lastTransitionTime":"2025-10-05T06:49:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:12 crc kubenswrapper[4846]: I1005 06:49:12.097491 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"775b5702-a131-4d6c-852d-91bfa056cac5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0839c1765df2c68d2554e9c66d289505e559a9672caf8069feeff2c4fdfd2171\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15dbaf2445704da0987f256f81dff2ee95664988541fa48ce94898afeadab2e8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd3b3ab323c6d5153054aa04af2d91f7b9e8436ff1e2f75f3594b57831b2a416\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://67c46f5f0a3ace95bd2e9479c5561126664e18b148c5eceea5bf5efc27e81a67\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c52e076eddfd307110b1f6de00a169b81af8f2785af7becef2a820ce363a5861\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-05T06:48:14Z\\\",\\\"message\\\":\\\"le observer\\\\nW1005 06:48:14.337870 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1005 06:48:14.338086 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1005 06:48:14.339389 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1467752215/tls.crt::/tmp/serving-cert-1467752215/tls.key\\\\\\\"\\\\nI1005 06:48:14.864345 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1005 06:48:14.868087 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1005 06:48:14.868114 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1005 06:48:14.868140 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1005 06:48:14.868147 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1005 06:48:14.877354 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1005 06:48:14.877412 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1005 06:48:14.877456 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1005 06:48:14.877471 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1005 06:48:14.877490 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1005 06:48:14.877501 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1005 06:48:14.877507 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1005 06:48:14.877515 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1005 06:48:14.880326 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:13Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4a091716ff935adb42ae2e88048619fd347bd174f64e5f9c97a01f81690e57f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:01Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dc5411b642f99b13d514d2f21defad0c394aeeb82a966f72300076617e2586c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5dc5411b642f99b13d514d2f21defad0c394aeeb82a966f72300076617e2586c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:47:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:47:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:47:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:49:12Z is after 2025-08-24T17:21:41Z" Oct 05 06:49:12 crc kubenswrapper[4846]: I1005 06:49:12.115146 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:49:12Z is after 2025-08-24T17:21:41Z" Oct 05 06:49:12 crc kubenswrapper[4846]: I1005 06:49:12.136983 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-r7ztb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a5119cd9-af70-4324-a955-978306ab0b20\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06231a87429ec292cb57ebd88cda7ad547bcb32e80343b776833621a11c28ad1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jfcj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-r7ztb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:49:12Z is after 2025-08-24T17:21:41Z" Oct 05 06:49:12 crc kubenswrapper[4846]: I1005 06:49:12.156870 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-gklb6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"698ab882-bb26-4b4c-81e9-8305a44fe068\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a6edca88ea244a58f8cf6cc431110a960c6bfd892812cd0d00fa3f97be3bc67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x6d6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c191538ab797b8056820de7da1f8e93ea18873891a32382b8a59c9af762e103\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x6d6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:37Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-gklb6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:49:12Z is after 2025-08-24T17:21:41Z" Oct 05 06:49:12 crc kubenswrapper[4846]: I1005 06:49:12.187552 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:12 crc kubenswrapper[4846]: I1005 06:49:12.187711 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:12 crc kubenswrapper[4846]: I1005 06:49:12.187802 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:12 crc kubenswrapper[4846]: I1005 06:49:12.187883 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:12 crc kubenswrapper[4846]: I1005 06:49:12.187852 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c482bb97-fd50-40cf-96ea-f806c3f01125\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bfe45d30e4fa141541aca1d46bc5cc1da0a7b4a5eb1fc9bb2caa5b7fdcb0b264\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://496a4782faa30a9a20ff63038141600cd635f12902d1cf17989cd067fcdfe929\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://281be077869a6aa1ca815770871a14413d4a97e97aeb550fd5cd215cba88b21f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6771c7215d2f5dd5838d17888375a4768b9d1755fc2e21bda1bbfbb2eb1dbb95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c17a711c2c7473b076a3944ffe7c700808d05cba582fbd55c5c2ac8ff73838a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44a9f740f0a06c8bf928e2a9d1d5979217558c31e73a333631baccebf00fd5d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://44a9f740f0a06c8bf928e2a9d1d5979217558c31e73a333631baccebf00fd5d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:47:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:47:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://48ca0ba6ca6092bfcb5b153220c054f02cc36fb22707c42e021f06d329e18e2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48ca0ba6ca6092bfcb5b153220c054f02cc36fb22707c42e021f06d329e18e2e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:00Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4006d0cea128675c8885d879171204abffc5e11ea0664750a12b4bc7344ccb2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4006d0cea128675c8885d879171204abffc5e11ea0664750a12b4bc7344ccb2e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:47:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:49:12Z is after 2025-08-24T17:21:41Z" Oct 05 06:49:12 crc kubenswrapper[4846]: I1005 06:49:12.187973 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:12Z","lastTransitionTime":"2025-10-05T06:49:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:12 crc kubenswrapper[4846]: I1005 06:49:12.204690 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:49:12Z is after 2025-08-24T17:21:41Z" Oct 05 06:49:12 crc kubenswrapper[4846]: I1005 06:49:12.231320 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-sfmxd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1dbb13ac-136c-4a1c-9229-b4e8ac4f9c71\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b0af6a6e735f7b7877c177c1d53d1a8dc12c4713a8059c4766fc858af31c9a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://226d977b596d07aeda0486452e2a8605949ed4d2453b21f56174937c395d6f36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://226d977b596d07aeda0486452e2a8605949ed4d2453b21f56174937c395d6f36\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://99c07a4bdb3059d3a99ffdaf794d668d52cffd9e5e3bb8945204dd6d1f8c8328\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99c07a4bdb3059d3a99ffdaf794d668d52cffd9e5e3bb8945204dd6d1f8c8328\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27095272bf4f31d63a78a436636de0aae12f7103597da944f24ed7a74329ee3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27095272bf4f31d63a78a436636de0aae12f7103597da944f24ed7a74329ee3f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1211571af66d8a94302b65feccf30768ba82b7cbad918e44dd405787dce484cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1211571af66d8a94302b65feccf30768ba82b7cbad918e44dd405787dce484cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27631f90d528900fa180c15cfce80e96d4cdba07743eb6d847e2e822d9917218\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27631f90d528900fa180c15cfce80e96d4cdba07743eb6d847e2e822d9917218\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f345693e9aeb1f4dc196349d3354f3c833a2b0415a693298b1cfccfa2b0d4459\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f345693e9aeb1f4dc196349d3354f3c833a2b0415a693298b1cfccfa2b0d4459\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-sfmxd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:49:12Z is after 2025-08-24T17:21:41Z" Oct 05 06:49:12 crc kubenswrapper[4846]: I1005 06:49:12.257293 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-txzlk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"93a6ace4-8d64-44d7-9637-457d3af6543c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ad1940d07e37217089642478ece9ccdaca74784738a400929a9f5359d158c01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a6f6b2c32dd3da5f9fd5e6971c1773e81a9a24184b7aa5b543cec879001ae53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e9b0219ecbc7e60905d6de3e67a2c8dc9af1017b4aed9538ee0c58c861f2438\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af5e5a4da9ad7371d5a24633e8268f12e3a7d419a8c93b11ddbcc006d8339c6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8da85222775abc7a772f6a8ff8255f1a0d5b15807ba177a7e3e2b2d01a4f24b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1409651de8c897804f87be03d016c138f8acf27e662dce265a93e15c1848adcf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd084bb9806ece6529db0866cee69e6a3152f0da7e0f13fc2e4adcbe2f55962c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd084bb9806ece6529db0866cee69e6a3152f0da7e0f13fc2e4adcbe2f55962c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-05T06:48:50Z\\\",\\\"message\\\":\\\":50.819394 6534 handler.go:208] Removed *v1.Node event handler 2\\\\nI1005 06:48:50.819434 6534 handler.go:208] Removed *v1.Node event handler 7\\\\nI1005 06:48:50.819451 6534 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1005 06:48:50.819624 6534 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1005 06:48:50.819659 6534 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1005 06:48:50.819711 6534 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1005 06:48:50.819735 6534 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1005 06:48:50.820556 6534 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1005 06:48:50.820605 6534 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1005 06:48:50.820649 6534 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1005 06:48:50.820664 6534 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1005 06:48:50.820706 6534 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1005 06:48:50.820711 6534 factory.go:656] Stopping watch factory\\\\nI1005 06:48:50.820731 6534 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1005 06:48:50.820750 6534 ovnkube.go:599] Stopped ovnkube\\\\nI1005 06:48:50.820796 6534 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1005 06:48:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:49Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-txzlk_openshift-ovn-kubernetes(93a6ace4-8d64-44d7-9637-457d3af6543c)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0df0ebcd87dc5af4b17736b43ec7b2395968b19ed16e775366af14e2b2d49c52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff3c6545ff3d2436da27bbd223b950b30b0c69a3e7f3034cb0e61380cb38d338\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ff3c6545ff3d2436da27bbd223b950b30b0c69a3e7f3034cb0e61380cb38d338\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:24Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-txzlk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:49:12Z is after 2025-08-24T17:21:41Z" Oct 05 06:49:12 crc kubenswrapper[4846]: I1005 06:49:12.273462 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c9f812b-ecc8-4985-a670-53fa80fbb499\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2b5ee9781c01d6aec3ac778efef1067b6c3bfe4ed37431bdebe188fcde0165f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:47:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3fb4717692aa1fec0012b514ef8d18d1ca3707cfa4bb8248bb93841a0a40f11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:47:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0b6a7afe3094f1c92093c05c28344897a615d590c2fbb507c87e18a44c6c831\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://640ade2c733d7916ec213a201a1dd614d261c69eab654c80100efbc99a768995\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:47:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:49:12Z is after 2025-08-24T17:21:41Z" Oct 05 06:49:12 crc kubenswrapper[4846]: I1005 06:49:12.287744 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:49:12Z is after 2025-08-24T17:21:41Z" Oct 05 06:49:12 crc kubenswrapper[4846]: I1005 06:49:12.290480 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:12 crc kubenswrapper[4846]: I1005 06:49:12.290592 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:12 crc kubenswrapper[4846]: I1005 06:49:12.290683 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:12 crc kubenswrapper[4846]: I1005 06:49:12.290761 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:12 crc kubenswrapper[4846]: I1005 06:49:12.290820 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:12Z","lastTransitionTime":"2025-10-05T06:49:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:12 crc kubenswrapper[4846]: I1005 06:49:12.302216 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e67d0e447a8a6a381259e25d25badc791194e1291f56b1cd77bd63989b20408\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dkhl7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://516f88f4990f5962d8b694a3d889259cc5508a309919870191c2e1331640c973\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dkhl7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-fscvf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:49:12Z is after 2025-08-24T17:21:41Z" Oct 05 06:49:12 crc kubenswrapper[4846]: I1005 06:49:12.312281 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-w2sj8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32931fce-7374-4cab-a2e0-79d96a842274\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0df535d65d39de6c9a3dfb931bafc4469adc22b72f7a7903beda4fef8cd48138\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9rtxf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:27Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-w2sj8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:49:12Z is after 2025-08-24T17:21:41Z" Oct 05 06:49:12 crc kubenswrapper[4846]: I1005 06:49:12.326791 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-7wm5q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9bc7744f-d700-450f-8b8d-98140877fee9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:49:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:49:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://90379e74c544fccca44f7acdc228ff378154390e211ac801b79f8d074012241e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://90379e74c544fccca44f7acdc228ff378154390e211ac801b79f8d074012241e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-05T06:49:11Z\\\",\\\"message\\\":\\\"2025-10-05T06:48:26+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_0ecc6292-0274-4088-90d5-740e638af9d4\\\\n2025-10-05T06:48:26+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_0ecc6292-0274-4088-90d5-740e638af9d4 to /host/opt/cni/bin/\\\\n2025-10-05T06:48:26Z [verbose] multus-daemon started\\\\n2025-10-05T06:48:26Z [verbose] Readiness Indicator file check\\\\n2025-10-05T06:49:11Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h5p4q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-7wm5q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:49:12Z is after 2025-08-24T17:21:41Z" Oct 05 06:49:12 crc kubenswrapper[4846]: I1005 06:49:12.340140 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"78b204fd-1034-4ea1-bcae-e5c876594a39\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b94cc50fe15d58f375f8492067811be16278cdf13ee3d80cfed4682279157b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5e5cf9125d1d413c91666744aeab06980b44420dbf0341b165b05787527944b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://70f0d5b436d18a78a2d8e2767143ba1a3104332d86bd63f91d23cd8e77a952aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c236b4ed7f039a221baac043246353eecbb6812c1e6bf6836a5fa96ec8f5abdb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c236b4ed7f039a221baac043246353eecbb6812c1e6bf6836a5fa96ec8f5abdb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:47:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:47:59Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:47:58Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:49:12Z is after 2025-08-24T17:21:41Z" Oct 05 06:49:12 crc kubenswrapper[4846]: I1005 06:49:12.355243 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ff10c31459b3b3724ee7b769bd6f72f9ed80cce5419b6adcdf69a3239b70756\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:49:12Z is after 2025-08-24T17:21:41Z" Oct 05 06:49:12 crc kubenswrapper[4846]: I1005 06:49:12.370770 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://347f900743b7bba3daf92f21c7eba0625e8c774f50057bc9472f0c82119325e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:49:12Z is after 2025-08-24T17:21:41Z" Oct 05 06:49:12 crc kubenswrapper[4846]: I1005 06:49:12.385038 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56f1fe1f9f06f0249cf937a6963fda82467732dc056ca2b2189be6d8dd9d4a05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7212ce5f42c1fb55146abb190a557a504b43f16af6df84f90e1073567a329ec5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:49:12Z is after 2025-08-24T17:21:41Z" Oct 05 06:49:12 crc kubenswrapper[4846]: I1005 06:49:12.394071 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:12 crc kubenswrapper[4846]: I1005 06:49:12.394329 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:12 crc kubenswrapper[4846]: I1005 06:49:12.394405 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:12 crc kubenswrapper[4846]: I1005 06:49:12.394491 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:12 crc kubenswrapper[4846]: I1005 06:49:12.394558 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:12Z","lastTransitionTime":"2025-10-05T06:49:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:12 crc kubenswrapper[4846]: I1005 06:49:12.496634 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 05 06:49:12 crc kubenswrapper[4846]: I1005 06:49:12.496807 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9r2kt" Oct 05 06:49:12 crc kubenswrapper[4846]: E1005 06:49:12.497345 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 05 06:49:12 crc kubenswrapper[4846]: E1005 06:49:12.497488 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9r2kt" podUID="59b83f4f-71f0-4b28-9275-56d5f933ec2c" Oct 05 06:49:12 crc kubenswrapper[4846]: I1005 06:49:12.498878 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:12 crc kubenswrapper[4846]: I1005 06:49:12.498934 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:12 crc kubenswrapper[4846]: I1005 06:49:12.498946 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:12 crc kubenswrapper[4846]: I1005 06:49:12.498965 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:12 crc kubenswrapper[4846]: I1005 06:49:12.498976 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:12Z","lastTransitionTime":"2025-10-05T06:49:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:12 crc kubenswrapper[4846]: I1005 06:49:12.601965 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:12 crc kubenswrapper[4846]: I1005 06:49:12.602295 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:12 crc kubenswrapper[4846]: I1005 06:49:12.602365 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:12 crc kubenswrapper[4846]: I1005 06:49:12.602436 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:12 crc kubenswrapper[4846]: I1005 06:49:12.602492 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:12Z","lastTransitionTime":"2025-10-05T06:49:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:12 crc kubenswrapper[4846]: I1005 06:49:12.705645 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:12 crc kubenswrapper[4846]: I1005 06:49:12.705704 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:12 crc kubenswrapper[4846]: I1005 06:49:12.705720 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:12 crc kubenswrapper[4846]: I1005 06:49:12.705739 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:12 crc kubenswrapper[4846]: I1005 06:49:12.705752 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:12Z","lastTransitionTime":"2025-10-05T06:49:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:12 crc kubenswrapper[4846]: I1005 06:49:12.809442 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:12 crc kubenswrapper[4846]: I1005 06:49:12.809491 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:12 crc kubenswrapper[4846]: I1005 06:49:12.809504 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:12 crc kubenswrapper[4846]: I1005 06:49:12.809521 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:12 crc kubenswrapper[4846]: I1005 06:49:12.809537 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:12Z","lastTransitionTime":"2025-10-05T06:49:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:12 crc kubenswrapper[4846]: I1005 06:49:12.912996 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:12 crc kubenswrapper[4846]: I1005 06:49:12.913055 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:12 crc kubenswrapper[4846]: I1005 06:49:12.913066 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:12 crc kubenswrapper[4846]: I1005 06:49:12.913084 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:12 crc kubenswrapper[4846]: I1005 06:49:12.913097 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:12Z","lastTransitionTime":"2025-10-05T06:49:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:13 crc kubenswrapper[4846]: I1005 06:49:13.015988 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:13 crc kubenswrapper[4846]: I1005 06:49:13.017417 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:13 crc kubenswrapper[4846]: I1005 06:49:13.017569 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:13 crc kubenswrapper[4846]: I1005 06:49:13.017713 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:13 crc kubenswrapper[4846]: I1005 06:49:13.017845 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:13Z","lastTransitionTime":"2025-10-05T06:49:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:13 crc kubenswrapper[4846]: I1005 06:49:13.049630 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-7wm5q_9bc7744f-d700-450f-8b8d-98140877fee9/kube-multus/0.log" Oct 05 06:49:13 crc kubenswrapper[4846]: I1005 06:49:13.049697 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-7wm5q" event={"ID":"9bc7744f-d700-450f-8b8d-98140877fee9","Type":"ContainerStarted","Data":"9d8631ba978f8fa7677feeeff42cabc454bfd51e5c5c21c78519fa9dfd85ec3c"} Oct 05 06:49:13 crc kubenswrapper[4846]: I1005 06:49:13.077628 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c482bb97-fd50-40cf-96ea-f806c3f01125\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bfe45d30e4fa141541aca1d46bc5cc1da0a7b4a5eb1fc9bb2caa5b7fdcb0b264\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://496a4782faa30a9a20ff63038141600cd635f12902d1cf17989cd067fcdfe929\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://281be077869a6aa1ca815770871a14413d4a97e97aeb550fd5cd215cba88b21f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6771c7215d2f5dd5838d17888375a4768b9d1755fc2e21bda1bbfbb2eb1dbb95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c17a711c2c7473b076a3944ffe7c700808d05cba582fbd55c5c2ac8ff73838a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44a9f740f0a06c8bf928e2a9d1d5979217558c31e73a333631baccebf00fd5d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://44a9f740f0a06c8bf928e2a9d1d5979217558c31e73a333631baccebf00fd5d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:47:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:47:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://48ca0ba6ca6092bfcb5b153220c054f02cc36fb22707c42e021f06d329e18e2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48ca0ba6ca6092bfcb5b153220c054f02cc36fb22707c42e021f06d329e18e2e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:00Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4006d0cea128675c8885d879171204abffc5e11ea0664750a12b4bc7344ccb2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4006d0cea128675c8885d879171204abffc5e11ea0664750a12b4bc7344ccb2e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:47:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:49:13Z is after 2025-08-24T17:21:41Z" Oct 05 06:49:13 crc kubenswrapper[4846]: I1005 06:49:13.100302 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:49:13Z is after 2025-08-24T17:21:41Z" Oct 05 06:49:13 crc kubenswrapper[4846]: I1005 06:49:13.121138 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:13 crc kubenswrapper[4846]: I1005 06:49:13.121510 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:13 crc kubenswrapper[4846]: I1005 06:49:13.121663 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:13 crc kubenswrapper[4846]: I1005 06:49:13.121851 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:13 crc kubenswrapper[4846]: I1005 06:49:13.121996 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:13Z","lastTransitionTime":"2025-10-05T06:49:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:13 crc kubenswrapper[4846]: I1005 06:49:13.127681 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-sfmxd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1dbb13ac-136c-4a1c-9229-b4e8ac4f9c71\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b0af6a6e735f7b7877c177c1d53d1a8dc12c4713a8059c4766fc858af31c9a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://226d977b596d07aeda0486452e2a8605949ed4d2453b21f56174937c395d6f36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://226d977b596d07aeda0486452e2a8605949ed4d2453b21f56174937c395d6f36\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://99c07a4bdb3059d3a99ffdaf794d668d52cffd9e5e3bb8945204dd6d1f8c8328\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99c07a4bdb3059d3a99ffdaf794d668d52cffd9e5e3bb8945204dd6d1f8c8328\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27095272bf4f31d63a78a436636de0aae12f7103597da944f24ed7a74329ee3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27095272bf4f31d63a78a436636de0aae12f7103597da944f24ed7a74329ee3f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1211571af66d8a94302b65feccf30768ba82b7cbad918e44dd405787dce484cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1211571af66d8a94302b65feccf30768ba82b7cbad918e44dd405787dce484cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27631f90d528900fa180c15cfce80e96d4cdba07743eb6d847e2e822d9917218\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27631f90d528900fa180c15cfce80e96d4cdba07743eb6d847e2e822d9917218\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f345693e9aeb1f4dc196349d3354f3c833a2b0415a693298b1cfccfa2b0d4459\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f345693e9aeb1f4dc196349d3354f3c833a2b0415a693298b1cfccfa2b0d4459\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-sfmxd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:49:13Z is after 2025-08-24T17:21:41Z" Oct 05 06:49:13 crc kubenswrapper[4846]: I1005 06:49:13.170445 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-txzlk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"93a6ace4-8d64-44d7-9637-457d3af6543c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ad1940d07e37217089642478ece9ccdaca74784738a400929a9f5359d158c01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a6f6b2c32dd3da5f9fd5e6971c1773e81a9a24184b7aa5b543cec879001ae53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e9b0219ecbc7e60905d6de3e67a2c8dc9af1017b4aed9538ee0c58c861f2438\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af5e5a4da9ad7371d5a24633e8268f12e3a7d419a8c93b11ddbcc006d8339c6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8da85222775abc7a772f6a8ff8255f1a0d5b15807ba177a7e3e2b2d01a4f24b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1409651de8c897804f87be03d016c138f8acf27e662dce265a93e15c1848adcf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd084bb9806ece6529db0866cee69e6a3152f0da7e0f13fc2e4adcbe2f55962c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd084bb9806ece6529db0866cee69e6a3152f0da7e0f13fc2e4adcbe2f55962c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-05T06:48:50Z\\\",\\\"message\\\":\\\":50.819394 6534 handler.go:208] Removed *v1.Node event handler 2\\\\nI1005 06:48:50.819434 6534 handler.go:208] Removed *v1.Node event handler 7\\\\nI1005 06:48:50.819451 6534 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1005 06:48:50.819624 6534 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1005 06:48:50.819659 6534 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1005 06:48:50.819711 6534 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1005 06:48:50.819735 6534 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1005 06:48:50.820556 6534 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1005 06:48:50.820605 6534 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1005 06:48:50.820649 6534 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1005 06:48:50.820664 6534 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1005 06:48:50.820706 6534 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1005 06:48:50.820711 6534 factory.go:656] Stopping watch factory\\\\nI1005 06:48:50.820731 6534 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1005 06:48:50.820750 6534 ovnkube.go:599] Stopped ovnkube\\\\nI1005 06:48:50.820796 6534 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1005 06:48:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:49Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-txzlk_openshift-ovn-kubernetes(93a6ace4-8d64-44d7-9637-457d3af6543c)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0df0ebcd87dc5af4b17736b43ec7b2395968b19ed16e775366af14e2b2d49c52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff3c6545ff3d2436da27bbd223b950b30b0c69a3e7f3034cb0e61380cb38d338\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ff3c6545ff3d2436da27bbd223b950b30b0c69a3e7f3034cb0e61380cb38d338\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:24Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-txzlk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:49:13Z is after 2025-08-24T17:21:41Z" Oct 05 06:49:13 crc kubenswrapper[4846]: I1005 06:49:13.192169 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-gklb6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"698ab882-bb26-4b4c-81e9-8305a44fe068\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a6edca88ea244a58f8cf6cc431110a960c6bfd892812cd0d00fa3f97be3bc67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x6d6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c191538ab797b8056820de7da1f8e93ea18873891a32382b8a59c9af762e103\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x6d6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:37Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-gklb6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:49:13Z is after 2025-08-24T17:21:41Z" Oct 05 06:49:13 crc kubenswrapper[4846]: I1005 06:49:13.214386 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c9f812b-ecc8-4985-a670-53fa80fbb499\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2b5ee9781c01d6aec3ac778efef1067b6c3bfe4ed37431bdebe188fcde0165f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:47:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3fb4717692aa1fec0012b514ef8d18d1ca3707cfa4bb8248bb93841a0a40f11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:47:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0b6a7afe3094f1c92093c05c28344897a615d590c2fbb507c87e18a44c6c831\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://640ade2c733d7916ec213a201a1dd614d261c69eab654c80100efbc99a768995\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:47:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:49:13Z is after 2025-08-24T17:21:41Z" Oct 05 06:49:13 crc kubenswrapper[4846]: I1005 06:49:13.227085 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:13 crc kubenswrapper[4846]: I1005 06:49:13.227149 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:13 crc kubenswrapper[4846]: I1005 06:49:13.227168 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:13 crc kubenswrapper[4846]: I1005 06:49:13.227219 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:13 crc kubenswrapper[4846]: I1005 06:49:13.227237 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:13Z","lastTransitionTime":"2025-10-05T06:49:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:13 crc kubenswrapper[4846]: I1005 06:49:13.245883 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:49:13Z is after 2025-08-24T17:21:41Z" Oct 05 06:49:13 crc kubenswrapper[4846]: I1005 06:49:13.266557 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e67d0e447a8a6a381259e25d25badc791194e1291f56b1cd77bd63989b20408\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dkhl7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://516f88f4990f5962d8b694a3d889259cc5508a309919870191c2e1331640c973\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dkhl7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-fscvf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:49:13Z is after 2025-08-24T17:21:41Z" Oct 05 06:49:13 crc kubenswrapper[4846]: I1005 06:49:13.283859 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-w2sj8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32931fce-7374-4cab-a2e0-79d96a842274\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0df535d65d39de6c9a3dfb931bafc4469adc22b72f7a7903beda4fef8cd48138\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9rtxf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:27Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-w2sj8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:49:13Z is after 2025-08-24T17:21:41Z" Oct 05 06:49:13 crc kubenswrapper[4846]: I1005 06:49:13.303166 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"78b204fd-1034-4ea1-bcae-e5c876594a39\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b94cc50fe15d58f375f8492067811be16278cdf13ee3d80cfed4682279157b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5e5cf9125d1d413c91666744aeab06980b44420dbf0341b165b05787527944b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://70f0d5b436d18a78a2d8e2767143ba1a3104332d86bd63f91d23cd8e77a952aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c236b4ed7f039a221baac043246353eecbb6812c1e6bf6836a5fa96ec8f5abdb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c236b4ed7f039a221baac043246353eecbb6812c1e6bf6836a5fa96ec8f5abdb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:47:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:47:59Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:47:58Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:49:13Z is after 2025-08-24T17:21:41Z" Oct 05 06:49:13 crc kubenswrapper[4846]: I1005 06:49:13.325445 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ff10c31459b3b3724ee7b769bd6f72f9ed80cce5419b6adcdf69a3239b70756\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:49:13Z is after 2025-08-24T17:21:41Z" Oct 05 06:49:13 crc kubenswrapper[4846]: I1005 06:49:13.331017 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:13 crc kubenswrapper[4846]: I1005 06:49:13.331261 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:13 crc kubenswrapper[4846]: I1005 06:49:13.331453 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:13 crc kubenswrapper[4846]: I1005 06:49:13.331659 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:13 crc kubenswrapper[4846]: I1005 06:49:13.331837 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:13Z","lastTransitionTime":"2025-10-05T06:49:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:13 crc kubenswrapper[4846]: I1005 06:49:13.346791 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://347f900743b7bba3daf92f21c7eba0625e8c774f50057bc9472f0c82119325e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:49:13Z is after 2025-08-24T17:21:41Z" Oct 05 06:49:13 crc kubenswrapper[4846]: I1005 06:49:13.368524 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56f1fe1f9f06f0249cf937a6963fda82467732dc056ca2b2189be6d8dd9d4a05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7212ce5f42c1fb55146abb190a557a504b43f16af6df84f90e1073567a329ec5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:49:13Z is after 2025-08-24T17:21:41Z" Oct 05 06:49:13 crc kubenswrapper[4846]: I1005 06:49:13.392343 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-7wm5q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9bc7744f-d700-450f-8b8d-98140877fee9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:49:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:49:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9d8631ba978f8fa7677feeeff42cabc454bfd51e5c5c21c78519fa9dfd85ec3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://90379e74c544fccca44f7acdc228ff378154390e211ac801b79f8d074012241e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-05T06:49:11Z\\\",\\\"message\\\":\\\"2025-10-05T06:48:26+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_0ecc6292-0274-4088-90d5-740e638af9d4\\\\n2025-10-05T06:48:26+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_0ecc6292-0274-4088-90d5-740e638af9d4 to /host/opt/cni/bin/\\\\n2025-10-05T06:48:26Z [verbose] multus-daemon started\\\\n2025-10-05T06:48:26Z [verbose] Readiness Indicator file check\\\\n2025-10-05T06:49:11Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:24Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:49:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h5p4q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-7wm5q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:49:13Z is after 2025-08-24T17:21:41Z" Oct 05 06:49:13 crc kubenswrapper[4846]: I1005 06:49:13.410581 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c21154c6-7c11-47e7-b1f0-6a5fe33d16b1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e13964315903099b3d7e3dc8c143fe58bb9f9654214702ba51581ac4044414d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef573b754a80e5148b8caa71b4f22c461f57dfa415394a5bccda9771b7684e4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef573b754a80e5148b8caa71b4f22c461f57dfa415394a5bccda9771b7684e4c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:47:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:47:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:47:58Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:49:13Z is after 2025-08-24T17:21:41Z" Oct 05 06:49:13 crc kubenswrapper[4846]: I1005 06:49:13.433809 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"775b5702-a131-4d6c-852d-91bfa056cac5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0839c1765df2c68d2554e9c66d289505e559a9672caf8069feeff2c4fdfd2171\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15dbaf2445704da0987f256f81dff2ee95664988541fa48ce94898afeadab2e8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd3b3ab323c6d5153054aa04af2d91f7b9e8436ff1e2f75f3594b57831b2a416\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://67c46f5f0a3ace95bd2e9479c5561126664e18b148c5eceea5bf5efc27e81a67\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c52e076eddfd307110b1f6de00a169b81af8f2785af7becef2a820ce363a5861\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-05T06:48:14Z\\\",\\\"message\\\":\\\"le observer\\\\nW1005 06:48:14.337870 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1005 06:48:14.338086 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1005 06:48:14.339389 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1467752215/tls.crt::/tmp/serving-cert-1467752215/tls.key\\\\\\\"\\\\nI1005 06:48:14.864345 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1005 06:48:14.868087 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1005 06:48:14.868114 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1005 06:48:14.868140 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1005 06:48:14.868147 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1005 06:48:14.877354 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1005 06:48:14.877412 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1005 06:48:14.877456 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1005 06:48:14.877471 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1005 06:48:14.877490 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1005 06:48:14.877501 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1005 06:48:14.877507 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1005 06:48:14.877515 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1005 06:48:14.880326 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:13Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4a091716ff935adb42ae2e88048619fd347bd174f64e5f9c97a01f81690e57f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:01Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dc5411b642f99b13d514d2f21defad0c394aeeb82a966f72300076617e2586c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5dc5411b642f99b13d514d2f21defad0c394aeeb82a966f72300076617e2586c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:47:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:47:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:47:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:49:13Z is after 2025-08-24T17:21:41Z" Oct 05 06:49:13 crc kubenswrapper[4846]: I1005 06:49:13.436270 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:13 crc kubenswrapper[4846]: I1005 06:49:13.436350 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:13 crc kubenswrapper[4846]: I1005 06:49:13.436370 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:13 crc kubenswrapper[4846]: I1005 06:49:13.436403 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:13 crc kubenswrapper[4846]: I1005 06:49:13.436427 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:13Z","lastTransitionTime":"2025-10-05T06:49:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:13 crc kubenswrapper[4846]: I1005 06:49:13.460362 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:49:13Z is after 2025-08-24T17:21:41Z" Oct 05 06:49:13 crc kubenswrapper[4846]: I1005 06:49:13.479286 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-r7ztb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a5119cd9-af70-4324-a955-978306ab0b20\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06231a87429ec292cb57ebd88cda7ad547bcb32e80343b776833621a11c28ad1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jfcj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-r7ztb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:49:13Z is after 2025-08-24T17:21:41Z" Oct 05 06:49:13 crc kubenswrapper[4846]: I1005 06:49:13.496765 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 05 06:49:13 crc kubenswrapper[4846]: E1005 06:49:13.497003 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 05 06:49:13 crc kubenswrapper[4846]: I1005 06:49:13.496765 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 05 06:49:13 crc kubenswrapper[4846]: I1005 06:49:13.497329 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-9r2kt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"59b83f4f-71f0-4b28-9275-56d5f933ec2c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tdwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tdwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:38Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-9r2kt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:49:13Z is after 2025-08-24T17:21:41Z" Oct 05 06:49:13 crc kubenswrapper[4846]: E1005 06:49:13.497883 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 05 06:49:13 crc kubenswrapper[4846]: I1005 06:49:13.539898 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:13 crc kubenswrapper[4846]: I1005 06:49:13.539969 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:13 crc kubenswrapper[4846]: I1005 06:49:13.539987 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:13 crc kubenswrapper[4846]: I1005 06:49:13.540016 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:13 crc kubenswrapper[4846]: I1005 06:49:13.540039 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:13Z","lastTransitionTime":"2025-10-05T06:49:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:13 crc kubenswrapper[4846]: I1005 06:49:13.642409 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:13 crc kubenswrapper[4846]: I1005 06:49:13.642511 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:13 crc kubenswrapper[4846]: I1005 06:49:13.642536 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:13 crc kubenswrapper[4846]: I1005 06:49:13.642565 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:13 crc kubenswrapper[4846]: I1005 06:49:13.642587 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:13Z","lastTransitionTime":"2025-10-05T06:49:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:13 crc kubenswrapper[4846]: I1005 06:49:13.745274 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:13 crc kubenswrapper[4846]: I1005 06:49:13.745328 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:13 crc kubenswrapper[4846]: I1005 06:49:13.745346 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:13 crc kubenswrapper[4846]: I1005 06:49:13.745370 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:13 crc kubenswrapper[4846]: I1005 06:49:13.745388 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:13Z","lastTransitionTime":"2025-10-05T06:49:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:13 crc kubenswrapper[4846]: I1005 06:49:13.848728 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:13 crc kubenswrapper[4846]: I1005 06:49:13.848807 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:13 crc kubenswrapper[4846]: I1005 06:49:13.848828 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:13 crc kubenswrapper[4846]: I1005 06:49:13.848861 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:13 crc kubenswrapper[4846]: I1005 06:49:13.848881 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:13Z","lastTransitionTime":"2025-10-05T06:49:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:13 crc kubenswrapper[4846]: I1005 06:49:13.952608 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:13 crc kubenswrapper[4846]: I1005 06:49:13.952668 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:13 crc kubenswrapper[4846]: I1005 06:49:13.952688 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:13 crc kubenswrapper[4846]: I1005 06:49:13.952714 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:13 crc kubenswrapper[4846]: I1005 06:49:13.952733 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:13Z","lastTransitionTime":"2025-10-05T06:49:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:14 crc kubenswrapper[4846]: I1005 06:49:14.054980 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:14 crc kubenswrapper[4846]: I1005 06:49:14.055016 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:14 crc kubenswrapper[4846]: I1005 06:49:14.055027 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:14 crc kubenswrapper[4846]: I1005 06:49:14.055043 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:14 crc kubenswrapper[4846]: I1005 06:49:14.055053 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:14Z","lastTransitionTime":"2025-10-05T06:49:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:14 crc kubenswrapper[4846]: I1005 06:49:14.158782 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:14 crc kubenswrapper[4846]: I1005 06:49:14.158843 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:14 crc kubenswrapper[4846]: I1005 06:49:14.158861 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:14 crc kubenswrapper[4846]: I1005 06:49:14.158892 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:14 crc kubenswrapper[4846]: I1005 06:49:14.158911 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:14Z","lastTransitionTime":"2025-10-05T06:49:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:14 crc kubenswrapper[4846]: I1005 06:49:14.261609 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:14 crc kubenswrapper[4846]: I1005 06:49:14.261675 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:14 crc kubenswrapper[4846]: I1005 06:49:14.261757 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:14 crc kubenswrapper[4846]: I1005 06:49:14.261787 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:14 crc kubenswrapper[4846]: I1005 06:49:14.261807 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:14Z","lastTransitionTime":"2025-10-05T06:49:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:14 crc kubenswrapper[4846]: I1005 06:49:14.367071 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:14 crc kubenswrapper[4846]: I1005 06:49:14.367134 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:14 crc kubenswrapper[4846]: I1005 06:49:14.367152 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:14 crc kubenswrapper[4846]: I1005 06:49:14.367200 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:14 crc kubenswrapper[4846]: I1005 06:49:14.367219 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:14Z","lastTransitionTime":"2025-10-05T06:49:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:14 crc kubenswrapper[4846]: I1005 06:49:14.471436 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:14 crc kubenswrapper[4846]: I1005 06:49:14.471505 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:14 crc kubenswrapper[4846]: I1005 06:49:14.471522 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:14 crc kubenswrapper[4846]: I1005 06:49:14.471548 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:14 crc kubenswrapper[4846]: I1005 06:49:14.471566 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:14Z","lastTransitionTime":"2025-10-05T06:49:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:14 crc kubenswrapper[4846]: I1005 06:49:14.497460 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9r2kt" Oct 05 06:49:14 crc kubenswrapper[4846]: I1005 06:49:14.497504 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 05 06:49:14 crc kubenswrapper[4846]: E1005 06:49:14.497751 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9r2kt" podUID="59b83f4f-71f0-4b28-9275-56d5f933ec2c" Oct 05 06:49:14 crc kubenswrapper[4846]: E1005 06:49:14.497940 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 05 06:49:14 crc kubenswrapper[4846]: I1005 06:49:14.574545 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:14 crc kubenswrapper[4846]: I1005 06:49:14.574631 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:14 crc kubenswrapper[4846]: I1005 06:49:14.574655 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:14 crc kubenswrapper[4846]: I1005 06:49:14.574689 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:14 crc kubenswrapper[4846]: I1005 06:49:14.574718 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:14Z","lastTransitionTime":"2025-10-05T06:49:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:14 crc kubenswrapper[4846]: I1005 06:49:14.677617 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:14 crc kubenswrapper[4846]: I1005 06:49:14.677724 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:14 crc kubenswrapper[4846]: I1005 06:49:14.677747 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:14 crc kubenswrapper[4846]: I1005 06:49:14.677771 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:14 crc kubenswrapper[4846]: I1005 06:49:14.677789 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:14Z","lastTransitionTime":"2025-10-05T06:49:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:14 crc kubenswrapper[4846]: I1005 06:49:14.780923 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:14 crc kubenswrapper[4846]: I1005 06:49:14.780974 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:14 crc kubenswrapper[4846]: I1005 06:49:14.780986 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:14 crc kubenswrapper[4846]: I1005 06:49:14.781005 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:14 crc kubenswrapper[4846]: I1005 06:49:14.781018 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:14Z","lastTransitionTime":"2025-10-05T06:49:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:14 crc kubenswrapper[4846]: I1005 06:49:14.884366 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:14 crc kubenswrapper[4846]: I1005 06:49:14.884414 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:14 crc kubenswrapper[4846]: I1005 06:49:14.884427 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:14 crc kubenswrapper[4846]: I1005 06:49:14.884444 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:14 crc kubenswrapper[4846]: I1005 06:49:14.884457 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:14Z","lastTransitionTime":"2025-10-05T06:49:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:14 crc kubenswrapper[4846]: I1005 06:49:14.989500 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:14 crc kubenswrapper[4846]: I1005 06:49:14.989555 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:14 crc kubenswrapper[4846]: I1005 06:49:14.989569 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:14 crc kubenswrapper[4846]: I1005 06:49:14.989590 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:14 crc kubenswrapper[4846]: I1005 06:49:14.989600 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:14Z","lastTransitionTime":"2025-10-05T06:49:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:15 crc kubenswrapper[4846]: I1005 06:49:15.092981 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:15 crc kubenswrapper[4846]: I1005 06:49:15.093028 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:15 crc kubenswrapper[4846]: I1005 06:49:15.093043 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:15 crc kubenswrapper[4846]: I1005 06:49:15.093061 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:15 crc kubenswrapper[4846]: I1005 06:49:15.093074 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:15Z","lastTransitionTime":"2025-10-05T06:49:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:15 crc kubenswrapper[4846]: I1005 06:49:15.196079 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:15 crc kubenswrapper[4846]: I1005 06:49:15.196137 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:15 crc kubenswrapper[4846]: I1005 06:49:15.196147 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:15 crc kubenswrapper[4846]: I1005 06:49:15.196194 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:15 crc kubenswrapper[4846]: I1005 06:49:15.196236 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:15Z","lastTransitionTime":"2025-10-05T06:49:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:15 crc kubenswrapper[4846]: I1005 06:49:15.299745 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:15 crc kubenswrapper[4846]: I1005 06:49:15.299795 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:15 crc kubenswrapper[4846]: I1005 06:49:15.299804 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:15 crc kubenswrapper[4846]: I1005 06:49:15.299820 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:15 crc kubenswrapper[4846]: I1005 06:49:15.299831 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:15Z","lastTransitionTime":"2025-10-05T06:49:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:15 crc kubenswrapper[4846]: I1005 06:49:15.402373 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:15 crc kubenswrapper[4846]: I1005 06:49:15.402428 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:15 crc kubenswrapper[4846]: I1005 06:49:15.402456 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:15 crc kubenswrapper[4846]: I1005 06:49:15.402480 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:15 crc kubenswrapper[4846]: I1005 06:49:15.402491 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:15Z","lastTransitionTime":"2025-10-05T06:49:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:15 crc kubenswrapper[4846]: I1005 06:49:15.497482 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 05 06:49:15 crc kubenswrapper[4846]: I1005 06:49:15.497538 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 05 06:49:15 crc kubenswrapper[4846]: E1005 06:49:15.497720 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 05 06:49:15 crc kubenswrapper[4846]: E1005 06:49:15.497930 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 05 06:49:15 crc kubenswrapper[4846]: I1005 06:49:15.505401 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:15 crc kubenswrapper[4846]: I1005 06:49:15.505459 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:15 crc kubenswrapper[4846]: I1005 06:49:15.505480 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:15 crc kubenswrapper[4846]: I1005 06:49:15.505507 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:15 crc kubenswrapper[4846]: I1005 06:49:15.505548 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:15Z","lastTransitionTime":"2025-10-05T06:49:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:15 crc kubenswrapper[4846]: I1005 06:49:15.608743 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:15 crc kubenswrapper[4846]: I1005 06:49:15.608806 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:15 crc kubenswrapper[4846]: I1005 06:49:15.608824 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:15 crc kubenswrapper[4846]: I1005 06:49:15.608849 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:15 crc kubenswrapper[4846]: I1005 06:49:15.608868 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:15Z","lastTransitionTime":"2025-10-05T06:49:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:15 crc kubenswrapper[4846]: I1005 06:49:15.712361 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:15 crc kubenswrapper[4846]: I1005 06:49:15.712450 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:15 crc kubenswrapper[4846]: I1005 06:49:15.712470 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:15 crc kubenswrapper[4846]: I1005 06:49:15.712500 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:15 crc kubenswrapper[4846]: I1005 06:49:15.712522 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:15Z","lastTransitionTime":"2025-10-05T06:49:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:15 crc kubenswrapper[4846]: I1005 06:49:15.815784 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:15 crc kubenswrapper[4846]: I1005 06:49:15.815855 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:15 crc kubenswrapper[4846]: I1005 06:49:15.815872 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:15 crc kubenswrapper[4846]: I1005 06:49:15.815900 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:15 crc kubenswrapper[4846]: I1005 06:49:15.815921 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:15Z","lastTransitionTime":"2025-10-05T06:49:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:15 crc kubenswrapper[4846]: I1005 06:49:15.918571 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:15 crc kubenswrapper[4846]: I1005 06:49:15.918632 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:15 crc kubenswrapper[4846]: I1005 06:49:15.918650 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:15 crc kubenswrapper[4846]: I1005 06:49:15.918680 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:15 crc kubenswrapper[4846]: I1005 06:49:15.918699 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:15Z","lastTransitionTime":"2025-10-05T06:49:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:16 crc kubenswrapper[4846]: I1005 06:49:16.022263 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:16 crc kubenswrapper[4846]: I1005 06:49:16.022347 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:16 crc kubenswrapper[4846]: I1005 06:49:16.022372 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:16 crc kubenswrapper[4846]: I1005 06:49:16.022405 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:16 crc kubenswrapper[4846]: I1005 06:49:16.022429 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:16Z","lastTransitionTime":"2025-10-05T06:49:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:16 crc kubenswrapper[4846]: I1005 06:49:16.049673 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:16 crc kubenswrapper[4846]: I1005 06:49:16.049755 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:16 crc kubenswrapper[4846]: I1005 06:49:16.049772 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:16 crc kubenswrapper[4846]: I1005 06:49:16.049801 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:16 crc kubenswrapper[4846]: I1005 06:49:16.049820 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:16Z","lastTransitionTime":"2025-10-05T06:49:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:16 crc kubenswrapper[4846]: E1005 06:49:16.072741 4846 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:49:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:49:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:49:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:49:16Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:49:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:49:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:49:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:49:16Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"04ce628d-2f61-4931-aad6-615924ea53fb\\\",\\\"systemUUID\\\":\\\"a050a23b-e773-4ba9-989b-360aa53a0605\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:49:16Z is after 2025-08-24T17:21:41Z" Oct 05 06:49:16 crc kubenswrapper[4846]: I1005 06:49:16.078538 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:16 crc kubenswrapper[4846]: I1005 06:49:16.078671 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:16 crc kubenswrapper[4846]: I1005 06:49:16.078698 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:16 crc kubenswrapper[4846]: I1005 06:49:16.078730 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:16 crc kubenswrapper[4846]: I1005 06:49:16.078753 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:16Z","lastTransitionTime":"2025-10-05T06:49:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:16 crc kubenswrapper[4846]: E1005 06:49:16.099829 4846 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:49:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:49:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:49:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:49:16Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:49:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:49:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:49:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:49:16Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"04ce628d-2f61-4931-aad6-615924ea53fb\\\",\\\"systemUUID\\\":\\\"a050a23b-e773-4ba9-989b-360aa53a0605\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:49:16Z is after 2025-08-24T17:21:41Z" Oct 05 06:49:16 crc kubenswrapper[4846]: I1005 06:49:16.106150 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:16 crc kubenswrapper[4846]: I1005 06:49:16.106255 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:16 crc kubenswrapper[4846]: I1005 06:49:16.106282 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:16 crc kubenswrapper[4846]: I1005 06:49:16.106313 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:16 crc kubenswrapper[4846]: I1005 06:49:16.106337 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:16Z","lastTransitionTime":"2025-10-05T06:49:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:16 crc kubenswrapper[4846]: E1005 06:49:16.126695 4846 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:49:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:49:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:49:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:49:16Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:49:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:49:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:49:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:49:16Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"04ce628d-2f61-4931-aad6-615924ea53fb\\\",\\\"systemUUID\\\":\\\"a050a23b-e773-4ba9-989b-360aa53a0605\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:49:16Z is after 2025-08-24T17:21:41Z" Oct 05 06:49:16 crc kubenswrapper[4846]: I1005 06:49:16.134912 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:16 crc kubenswrapper[4846]: I1005 06:49:16.134984 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:16 crc kubenswrapper[4846]: I1005 06:49:16.135002 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:16 crc kubenswrapper[4846]: I1005 06:49:16.135031 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:16 crc kubenswrapper[4846]: I1005 06:49:16.135046 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:16Z","lastTransitionTime":"2025-10-05T06:49:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:16 crc kubenswrapper[4846]: E1005 06:49:16.158300 4846 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:49:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:49:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:49:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:49:16Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:49:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:49:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:49:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:49:16Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"04ce628d-2f61-4931-aad6-615924ea53fb\\\",\\\"systemUUID\\\":\\\"a050a23b-e773-4ba9-989b-360aa53a0605\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:49:16Z is after 2025-08-24T17:21:41Z" Oct 05 06:49:16 crc kubenswrapper[4846]: I1005 06:49:16.162647 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:16 crc kubenswrapper[4846]: I1005 06:49:16.162682 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:16 crc kubenswrapper[4846]: I1005 06:49:16.162692 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:16 crc kubenswrapper[4846]: I1005 06:49:16.162707 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:16 crc kubenswrapper[4846]: I1005 06:49:16.162717 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:16Z","lastTransitionTime":"2025-10-05T06:49:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:16 crc kubenswrapper[4846]: E1005 06:49:16.177433 4846 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:49:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:49:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:49:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:49:16Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:49:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:49:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:49:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:49:16Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"04ce628d-2f61-4931-aad6-615924ea53fb\\\",\\\"systemUUID\\\":\\\"a050a23b-e773-4ba9-989b-360aa53a0605\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:49:16Z is after 2025-08-24T17:21:41Z" Oct 05 06:49:16 crc kubenswrapper[4846]: E1005 06:49:16.177538 4846 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 05 06:49:16 crc kubenswrapper[4846]: I1005 06:49:16.179048 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:16 crc kubenswrapper[4846]: I1005 06:49:16.179078 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:16 crc kubenswrapper[4846]: I1005 06:49:16.179086 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:16 crc kubenswrapper[4846]: I1005 06:49:16.179099 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:16 crc kubenswrapper[4846]: I1005 06:49:16.179110 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:16Z","lastTransitionTime":"2025-10-05T06:49:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:16 crc kubenswrapper[4846]: I1005 06:49:16.282731 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:16 crc kubenswrapper[4846]: I1005 06:49:16.282787 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:16 crc kubenswrapper[4846]: I1005 06:49:16.282800 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:16 crc kubenswrapper[4846]: I1005 06:49:16.282821 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:16 crc kubenswrapper[4846]: I1005 06:49:16.282838 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:16Z","lastTransitionTime":"2025-10-05T06:49:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:16 crc kubenswrapper[4846]: I1005 06:49:16.386416 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:16 crc kubenswrapper[4846]: I1005 06:49:16.386485 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:16 crc kubenswrapper[4846]: I1005 06:49:16.386503 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:16 crc kubenswrapper[4846]: I1005 06:49:16.386530 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:16 crc kubenswrapper[4846]: I1005 06:49:16.386550 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:16Z","lastTransitionTime":"2025-10-05T06:49:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:16 crc kubenswrapper[4846]: I1005 06:49:16.488688 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:16 crc kubenswrapper[4846]: I1005 06:49:16.488728 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:16 crc kubenswrapper[4846]: I1005 06:49:16.488740 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:16 crc kubenswrapper[4846]: I1005 06:49:16.488755 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:16 crc kubenswrapper[4846]: I1005 06:49:16.488766 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:16Z","lastTransitionTime":"2025-10-05T06:49:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:16 crc kubenswrapper[4846]: I1005 06:49:16.497447 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9r2kt" Oct 05 06:49:16 crc kubenswrapper[4846]: E1005 06:49:16.497561 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9r2kt" podUID="59b83f4f-71f0-4b28-9275-56d5f933ec2c" Oct 05 06:49:16 crc kubenswrapper[4846]: I1005 06:49:16.497642 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 05 06:49:16 crc kubenswrapper[4846]: E1005 06:49:16.497858 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 05 06:49:16 crc kubenswrapper[4846]: I1005 06:49:16.592136 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:16 crc kubenswrapper[4846]: I1005 06:49:16.592232 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:16 crc kubenswrapper[4846]: I1005 06:49:16.592252 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:16 crc kubenswrapper[4846]: I1005 06:49:16.592278 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:16 crc kubenswrapper[4846]: I1005 06:49:16.592298 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:16Z","lastTransitionTime":"2025-10-05T06:49:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:16 crc kubenswrapper[4846]: I1005 06:49:16.696035 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:16 crc kubenswrapper[4846]: I1005 06:49:16.696110 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:16 crc kubenswrapper[4846]: I1005 06:49:16.696127 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:16 crc kubenswrapper[4846]: I1005 06:49:16.696153 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:16 crc kubenswrapper[4846]: I1005 06:49:16.696173 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:16Z","lastTransitionTime":"2025-10-05T06:49:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:16 crc kubenswrapper[4846]: I1005 06:49:16.799139 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:16 crc kubenswrapper[4846]: I1005 06:49:16.799199 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:16 crc kubenswrapper[4846]: I1005 06:49:16.799212 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:16 crc kubenswrapper[4846]: I1005 06:49:16.799229 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:16 crc kubenswrapper[4846]: I1005 06:49:16.799241 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:16Z","lastTransitionTime":"2025-10-05T06:49:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:16 crc kubenswrapper[4846]: I1005 06:49:16.902549 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:16 crc kubenswrapper[4846]: I1005 06:49:16.902615 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:16 crc kubenswrapper[4846]: I1005 06:49:16.902637 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:16 crc kubenswrapper[4846]: I1005 06:49:16.902669 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:16 crc kubenswrapper[4846]: I1005 06:49:16.902732 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:16Z","lastTransitionTime":"2025-10-05T06:49:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:17 crc kubenswrapper[4846]: I1005 06:49:17.006385 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:17 crc kubenswrapper[4846]: I1005 06:49:17.006471 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:17 crc kubenswrapper[4846]: I1005 06:49:17.006498 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:17 crc kubenswrapper[4846]: I1005 06:49:17.006523 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:17 crc kubenswrapper[4846]: I1005 06:49:17.006541 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:17Z","lastTransitionTime":"2025-10-05T06:49:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:17 crc kubenswrapper[4846]: I1005 06:49:17.109509 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:17 crc kubenswrapper[4846]: I1005 06:49:17.109573 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:17 crc kubenswrapper[4846]: I1005 06:49:17.109597 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:17 crc kubenswrapper[4846]: I1005 06:49:17.109628 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:17 crc kubenswrapper[4846]: I1005 06:49:17.109649 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:17Z","lastTransitionTime":"2025-10-05T06:49:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:17 crc kubenswrapper[4846]: I1005 06:49:17.212892 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:17 crc kubenswrapper[4846]: I1005 06:49:17.212961 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:17 crc kubenswrapper[4846]: I1005 06:49:17.212983 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:17 crc kubenswrapper[4846]: I1005 06:49:17.213007 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:17 crc kubenswrapper[4846]: I1005 06:49:17.213023 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:17Z","lastTransitionTime":"2025-10-05T06:49:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:17 crc kubenswrapper[4846]: I1005 06:49:17.317020 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:17 crc kubenswrapper[4846]: I1005 06:49:17.317077 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:17 crc kubenswrapper[4846]: I1005 06:49:17.317101 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:17 crc kubenswrapper[4846]: I1005 06:49:17.317148 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:17 crc kubenswrapper[4846]: I1005 06:49:17.317171 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:17Z","lastTransitionTime":"2025-10-05T06:49:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:17 crc kubenswrapper[4846]: I1005 06:49:17.420348 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:17 crc kubenswrapper[4846]: I1005 06:49:17.420417 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:17 crc kubenswrapper[4846]: I1005 06:49:17.420439 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:17 crc kubenswrapper[4846]: I1005 06:49:17.420464 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:17 crc kubenswrapper[4846]: I1005 06:49:17.420481 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:17Z","lastTransitionTime":"2025-10-05T06:49:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:17 crc kubenswrapper[4846]: I1005 06:49:17.496534 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 05 06:49:17 crc kubenswrapper[4846]: I1005 06:49:17.496647 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 05 06:49:17 crc kubenswrapper[4846]: E1005 06:49:17.496746 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 05 06:49:17 crc kubenswrapper[4846]: E1005 06:49:17.497444 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 05 06:49:17 crc kubenswrapper[4846]: I1005 06:49:17.498075 4846 scope.go:117] "RemoveContainer" containerID="dd084bb9806ece6529db0866cee69e6a3152f0da7e0f13fc2e4adcbe2f55962c" Oct 05 06:49:17 crc kubenswrapper[4846]: I1005 06:49:17.523833 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:17 crc kubenswrapper[4846]: I1005 06:49:17.523906 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:17 crc kubenswrapper[4846]: I1005 06:49:17.523931 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:17 crc kubenswrapper[4846]: I1005 06:49:17.523979 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:17 crc kubenswrapper[4846]: I1005 06:49:17.524011 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:17Z","lastTransitionTime":"2025-10-05T06:49:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:17 crc kubenswrapper[4846]: I1005 06:49:17.626991 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:17 crc kubenswrapper[4846]: I1005 06:49:17.627057 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:17 crc kubenswrapper[4846]: I1005 06:49:17.627078 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:17 crc kubenswrapper[4846]: I1005 06:49:17.627118 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:17 crc kubenswrapper[4846]: I1005 06:49:17.627136 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:17Z","lastTransitionTime":"2025-10-05T06:49:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:17 crc kubenswrapper[4846]: I1005 06:49:17.729711 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:17 crc kubenswrapper[4846]: I1005 06:49:17.729745 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:17 crc kubenswrapper[4846]: I1005 06:49:17.729757 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:17 crc kubenswrapper[4846]: I1005 06:49:17.729773 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:17 crc kubenswrapper[4846]: I1005 06:49:17.729788 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:17Z","lastTransitionTime":"2025-10-05T06:49:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:17 crc kubenswrapper[4846]: I1005 06:49:17.832731 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:17 crc kubenswrapper[4846]: I1005 06:49:17.832803 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:17 crc kubenswrapper[4846]: I1005 06:49:17.832820 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:17 crc kubenswrapper[4846]: I1005 06:49:17.832848 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:17 crc kubenswrapper[4846]: I1005 06:49:17.832870 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:17Z","lastTransitionTime":"2025-10-05T06:49:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:17 crc kubenswrapper[4846]: I1005 06:49:17.947288 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:17 crc kubenswrapper[4846]: I1005 06:49:17.950604 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:17 crc kubenswrapper[4846]: I1005 06:49:17.950620 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:17 crc kubenswrapper[4846]: I1005 06:49:17.950641 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:17 crc kubenswrapper[4846]: I1005 06:49:17.950654 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:17Z","lastTransitionTime":"2025-10-05T06:49:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:18 crc kubenswrapper[4846]: I1005 06:49:18.052902 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:18 crc kubenswrapper[4846]: I1005 06:49:18.052958 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:18 crc kubenswrapper[4846]: I1005 06:49:18.052970 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:18 crc kubenswrapper[4846]: I1005 06:49:18.052989 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:18 crc kubenswrapper[4846]: I1005 06:49:18.053003 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:18Z","lastTransitionTime":"2025-10-05T06:49:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:18 crc kubenswrapper[4846]: I1005 06:49:18.070941 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-txzlk_93a6ace4-8d64-44d7-9637-457d3af6543c/ovnkube-controller/2.log" Oct 05 06:49:18 crc kubenswrapper[4846]: I1005 06:49:18.073686 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-txzlk" event={"ID":"93a6ace4-8d64-44d7-9637-457d3af6543c","Type":"ContainerStarted","Data":"59fa020f7bfea4547e132058cf3804a406ee0c06cdc4e78a08ae43a51f46633f"} Oct 05 06:49:18 crc kubenswrapper[4846]: I1005 06:49:18.074151 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-txzlk" Oct 05 06:49:18 crc kubenswrapper[4846]: I1005 06:49:18.085870 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c21154c6-7c11-47e7-b1f0-6a5fe33d16b1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e13964315903099b3d7e3dc8c143fe58bb9f9654214702ba51581ac4044414d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef573b754a80e5148b8caa71b4f22c461f57dfa415394a5bccda9771b7684e4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef573b754a80e5148b8caa71b4f22c461f57dfa415394a5bccda9771b7684e4c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:47:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:47:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:47:58Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:49:18Z is after 2025-08-24T17:21:41Z" Oct 05 06:49:18 crc kubenswrapper[4846]: I1005 06:49:18.099824 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"775b5702-a131-4d6c-852d-91bfa056cac5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0839c1765df2c68d2554e9c66d289505e559a9672caf8069feeff2c4fdfd2171\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15dbaf2445704da0987f256f81dff2ee95664988541fa48ce94898afeadab2e8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd3b3ab323c6d5153054aa04af2d91f7b9e8436ff1e2f75f3594b57831b2a416\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://67c46f5f0a3ace95bd2e9479c5561126664e18b148c5eceea5bf5efc27e81a67\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c52e076eddfd307110b1f6de00a169b81af8f2785af7becef2a820ce363a5861\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-05T06:48:14Z\\\",\\\"message\\\":\\\"le observer\\\\nW1005 06:48:14.337870 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1005 06:48:14.338086 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1005 06:48:14.339389 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1467752215/tls.crt::/tmp/serving-cert-1467752215/tls.key\\\\\\\"\\\\nI1005 06:48:14.864345 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1005 06:48:14.868087 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1005 06:48:14.868114 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1005 06:48:14.868140 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1005 06:48:14.868147 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1005 06:48:14.877354 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1005 06:48:14.877412 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1005 06:48:14.877456 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1005 06:48:14.877471 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1005 06:48:14.877490 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1005 06:48:14.877501 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1005 06:48:14.877507 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1005 06:48:14.877515 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1005 06:48:14.880326 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:13Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4a091716ff935adb42ae2e88048619fd347bd174f64e5f9c97a01f81690e57f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:01Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dc5411b642f99b13d514d2f21defad0c394aeeb82a966f72300076617e2586c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5dc5411b642f99b13d514d2f21defad0c394aeeb82a966f72300076617e2586c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:47:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:47:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:47:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:49:18Z is after 2025-08-24T17:21:41Z" Oct 05 06:49:18 crc kubenswrapper[4846]: I1005 06:49:18.116956 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:49:18Z is after 2025-08-24T17:21:41Z" Oct 05 06:49:18 crc kubenswrapper[4846]: I1005 06:49:18.128154 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-r7ztb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a5119cd9-af70-4324-a955-978306ab0b20\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06231a87429ec292cb57ebd88cda7ad547bcb32e80343b776833621a11c28ad1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jfcj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-r7ztb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:49:18Z is after 2025-08-24T17:21:41Z" Oct 05 06:49:18 crc kubenswrapper[4846]: I1005 06:49:18.138954 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-9r2kt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"59b83f4f-71f0-4b28-9275-56d5f933ec2c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tdwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tdwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:38Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-9r2kt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:49:18Z is after 2025-08-24T17:21:41Z" Oct 05 06:49:18 crc kubenswrapper[4846]: I1005 06:49:18.155577 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:18 crc kubenswrapper[4846]: I1005 06:49:18.155612 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:18 crc kubenswrapper[4846]: I1005 06:49:18.155623 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:18 crc kubenswrapper[4846]: I1005 06:49:18.155640 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:18 crc kubenswrapper[4846]: I1005 06:49:18.155650 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:18Z","lastTransitionTime":"2025-10-05T06:49:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:18 crc kubenswrapper[4846]: I1005 06:49:18.165108 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c482bb97-fd50-40cf-96ea-f806c3f01125\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bfe45d30e4fa141541aca1d46bc5cc1da0a7b4a5eb1fc9bb2caa5b7fdcb0b264\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://496a4782faa30a9a20ff63038141600cd635f12902d1cf17989cd067fcdfe929\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://281be077869a6aa1ca815770871a14413d4a97e97aeb550fd5cd215cba88b21f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6771c7215d2f5dd5838d17888375a4768b9d1755fc2e21bda1bbfbb2eb1dbb95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c17a711c2c7473b076a3944ffe7c700808d05cba582fbd55c5c2ac8ff73838a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44a9f740f0a06c8bf928e2a9d1d5979217558c31e73a333631baccebf00fd5d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://44a9f740f0a06c8bf928e2a9d1d5979217558c31e73a333631baccebf00fd5d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:47:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:47:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://48ca0ba6ca6092bfcb5b153220c054f02cc36fb22707c42e021f06d329e18e2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48ca0ba6ca6092bfcb5b153220c054f02cc36fb22707c42e021f06d329e18e2e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:00Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4006d0cea128675c8885d879171204abffc5e11ea0664750a12b4bc7344ccb2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4006d0cea128675c8885d879171204abffc5e11ea0664750a12b4bc7344ccb2e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:47:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:49:18Z is after 2025-08-24T17:21:41Z" Oct 05 06:49:18 crc kubenswrapper[4846]: I1005 06:49:18.190145 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:49:18Z is after 2025-08-24T17:21:41Z" Oct 05 06:49:18 crc kubenswrapper[4846]: I1005 06:49:18.204257 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-sfmxd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1dbb13ac-136c-4a1c-9229-b4e8ac4f9c71\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b0af6a6e735f7b7877c177c1d53d1a8dc12c4713a8059c4766fc858af31c9a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://226d977b596d07aeda0486452e2a8605949ed4d2453b21f56174937c395d6f36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://226d977b596d07aeda0486452e2a8605949ed4d2453b21f56174937c395d6f36\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://99c07a4bdb3059d3a99ffdaf794d668d52cffd9e5e3bb8945204dd6d1f8c8328\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99c07a4bdb3059d3a99ffdaf794d668d52cffd9e5e3bb8945204dd6d1f8c8328\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27095272bf4f31d63a78a436636de0aae12f7103597da944f24ed7a74329ee3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27095272bf4f31d63a78a436636de0aae12f7103597da944f24ed7a74329ee3f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1211571af66d8a94302b65feccf30768ba82b7cbad918e44dd405787dce484cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1211571af66d8a94302b65feccf30768ba82b7cbad918e44dd405787dce484cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27631f90d528900fa180c15cfce80e96d4cdba07743eb6d847e2e822d9917218\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27631f90d528900fa180c15cfce80e96d4cdba07743eb6d847e2e822d9917218\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f345693e9aeb1f4dc196349d3354f3c833a2b0415a693298b1cfccfa2b0d4459\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f345693e9aeb1f4dc196349d3354f3c833a2b0415a693298b1cfccfa2b0d4459\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-sfmxd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:49:18Z is after 2025-08-24T17:21:41Z" Oct 05 06:49:18 crc kubenswrapper[4846]: I1005 06:49:18.240819 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-txzlk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"93a6ace4-8d64-44d7-9637-457d3af6543c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ad1940d07e37217089642478ece9ccdaca74784738a400929a9f5359d158c01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a6f6b2c32dd3da5f9fd5e6971c1773e81a9a24184b7aa5b543cec879001ae53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e9b0219ecbc7e60905d6de3e67a2c8dc9af1017b4aed9538ee0c58c861f2438\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af5e5a4da9ad7371d5a24633e8268f12e3a7d419a8c93b11ddbcc006d8339c6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8da85222775abc7a772f6a8ff8255f1a0d5b15807ba177a7e3e2b2d01a4f24b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1409651de8c897804f87be03d016c138f8acf27e662dce265a93e15c1848adcf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://59fa020f7bfea4547e132058cf3804a406ee0c06cdc4e78a08ae43a51f46633f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd084bb9806ece6529db0866cee69e6a3152f0da7e0f13fc2e4adcbe2f55962c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-05T06:48:50Z\\\",\\\"message\\\":\\\":50.819394 6534 handler.go:208] Removed *v1.Node event handler 2\\\\nI1005 06:48:50.819434 6534 handler.go:208] Removed *v1.Node event handler 7\\\\nI1005 06:48:50.819451 6534 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1005 06:48:50.819624 6534 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1005 06:48:50.819659 6534 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1005 06:48:50.819711 6534 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1005 06:48:50.819735 6534 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1005 06:48:50.820556 6534 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1005 06:48:50.820605 6534 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1005 06:48:50.820649 6534 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1005 06:48:50.820664 6534 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1005 06:48:50.820706 6534 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1005 06:48:50.820711 6534 factory.go:656] Stopping watch factory\\\\nI1005 06:48:50.820731 6534 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1005 06:48:50.820750 6534 ovnkube.go:599] Stopped ovnkube\\\\nI1005 06:48:50.820796 6534 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1005 06:48:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:49Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:49:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0df0ebcd87dc5af4b17736b43ec7b2395968b19ed16e775366af14e2b2d49c52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff3c6545ff3d2436da27bbd223b950b30b0c69a3e7f3034cb0e61380cb38d338\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ff3c6545ff3d2436da27bbd223b950b30b0c69a3e7f3034cb0e61380cb38d338\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:24Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-txzlk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:49:18Z is after 2025-08-24T17:21:41Z" Oct 05 06:49:18 crc kubenswrapper[4846]: I1005 06:49:18.256568 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-gklb6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"698ab882-bb26-4b4c-81e9-8305a44fe068\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a6edca88ea244a58f8cf6cc431110a960c6bfd892812cd0d00fa3f97be3bc67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x6d6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c191538ab797b8056820de7da1f8e93ea18873891a32382b8a59c9af762e103\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x6d6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:37Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-gklb6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:49:18Z is after 2025-08-24T17:21:41Z" Oct 05 06:49:18 crc kubenswrapper[4846]: I1005 06:49:18.258369 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:18 crc kubenswrapper[4846]: I1005 06:49:18.258398 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:18 crc kubenswrapper[4846]: I1005 06:49:18.258406 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:18 crc kubenswrapper[4846]: I1005 06:49:18.258419 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:18 crc kubenswrapper[4846]: I1005 06:49:18.258428 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:18Z","lastTransitionTime":"2025-10-05T06:49:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:18 crc kubenswrapper[4846]: I1005 06:49:18.270752 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c9f812b-ecc8-4985-a670-53fa80fbb499\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2b5ee9781c01d6aec3ac778efef1067b6c3bfe4ed37431bdebe188fcde0165f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:47:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3fb4717692aa1fec0012b514ef8d18d1ca3707cfa4bb8248bb93841a0a40f11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:47:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0b6a7afe3094f1c92093c05c28344897a615d590c2fbb507c87e18a44c6c831\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://640ade2c733d7916ec213a201a1dd614d261c69eab654c80100efbc99a768995\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:47:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:49:18Z is after 2025-08-24T17:21:41Z" Oct 05 06:49:18 crc kubenswrapper[4846]: I1005 06:49:18.282108 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:49:18Z is after 2025-08-24T17:21:41Z" Oct 05 06:49:18 crc kubenswrapper[4846]: I1005 06:49:18.294476 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e67d0e447a8a6a381259e25d25badc791194e1291f56b1cd77bd63989b20408\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dkhl7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://516f88f4990f5962d8b694a3d889259cc5508a309919870191c2e1331640c973\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dkhl7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-fscvf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:49:18Z is after 2025-08-24T17:21:41Z" Oct 05 06:49:18 crc kubenswrapper[4846]: I1005 06:49:18.304492 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-w2sj8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32931fce-7374-4cab-a2e0-79d96a842274\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0df535d65d39de6c9a3dfb931bafc4469adc22b72f7a7903beda4fef8cd48138\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9rtxf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:27Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-w2sj8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:49:18Z is after 2025-08-24T17:21:41Z" Oct 05 06:49:18 crc kubenswrapper[4846]: I1005 06:49:18.314968 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"78b204fd-1034-4ea1-bcae-e5c876594a39\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b94cc50fe15d58f375f8492067811be16278cdf13ee3d80cfed4682279157b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5e5cf9125d1d413c91666744aeab06980b44420dbf0341b165b05787527944b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://70f0d5b436d18a78a2d8e2767143ba1a3104332d86bd63f91d23cd8e77a952aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c236b4ed7f039a221baac043246353eecbb6812c1e6bf6836a5fa96ec8f5abdb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c236b4ed7f039a221baac043246353eecbb6812c1e6bf6836a5fa96ec8f5abdb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:47:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:47:59Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:47:58Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:49:18Z is after 2025-08-24T17:21:41Z" Oct 05 06:49:18 crc kubenswrapper[4846]: I1005 06:49:18.326826 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ff10c31459b3b3724ee7b769bd6f72f9ed80cce5419b6adcdf69a3239b70756\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:49:18Z is after 2025-08-24T17:21:41Z" Oct 05 06:49:18 crc kubenswrapper[4846]: I1005 06:49:18.337662 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://347f900743b7bba3daf92f21c7eba0625e8c774f50057bc9472f0c82119325e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:49:18Z is after 2025-08-24T17:21:41Z" Oct 05 06:49:18 crc kubenswrapper[4846]: I1005 06:49:18.349468 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56f1fe1f9f06f0249cf937a6963fda82467732dc056ca2b2189be6d8dd9d4a05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7212ce5f42c1fb55146abb190a557a504b43f16af6df84f90e1073567a329ec5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:49:18Z is after 2025-08-24T17:21:41Z" Oct 05 06:49:18 crc kubenswrapper[4846]: I1005 06:49:18.360753 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:18 crc kubenswrapper[4846]: I1005 06:49:18.360799 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:18 crc kubenswrapper[4846]: I1005 06:49:18.360812 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:18 crc kubenswrapper[4846]: I1005 06:49:18.360830 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:18 crc kubenswrapper[4846]: I1005 06:49:18.360844 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:18Z","lastTransitionTime":"2025-10-05T06:49:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:18 crc kubenswrapper[4846]: I1005 06:49:18.363192 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-7wm5q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9bc7744f-d700-450f-8b8d-98140877fee9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:49:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:49:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9d8631ba978f8fa7677feeeff42cabc454bfd51e5c5c21c78519fa9dfd85ec3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://90379e74c544fccca44f7acdc228ff378154390e211ac801b79f8d074012241e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-05T06:49:11Z\\\",\\\"message\\\":\\\"2025-10-05T06:48:26+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_0ecc6292-0274-4088-90d5-740e638af9d4\\\\n2025-10-05T06:48:26+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_0ecc6292-0274-4088-90d5-740e638af9d4 to /host/opt/cni/bin/\\\\n2025-10-05T06:48:26Z [verbose] multus-daemon started\\\\n2025-10-05T06:48:26Z [verbose] Readiness Indicator file check\\\\n2025-10-05T06:49:11Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:24Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:49:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h5p4q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-7wm5q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:49:18Z is after 2025-08-24T17:21:41Z" Oct 05 06:49:18 crc kubenswrapper[4846]: I1005 06:49:18.465013 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:18 crc kubenswrapper[4846]: I1005 06:49:18.465052 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:18 crc kubenswrapper[4846]: I1005 06:49:18.465059 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:18 crc kubenswrapper[4846]: I1005 06:49:18.465076 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:18 crc kubenswrapper[4846]: I1005 06:49:18.465088 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:18Z","lastTransitionTime":"2025-10-05T06:49:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:18 crc kubenswrapper[4846]: I1005 06:49:18.496953 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 05 06:49:18 crc kubenswrapper[4846]: I1005 06:49:18.497034 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9r2kt" Oct 05 06:49:18 crc kubenswrapper[4846]: E1005 06:49:18.497768 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9r2kt" podUID="59b83f4f-71f0-4b28-9275-56d5f933ec2c" Oct 05 06:49:18 crc kubenswrapper[4846]: E1005 06:49:18.497776 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 05 06:49:18 crc kubenswrapper[4846]: I1005 06:49:18.512806 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c9f812b-ecc8-4985-a670-53fa80fbb499\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2b5ee9781c01d6aec3ac778efef1067b6c3bfe4ed37431bdebe188fcde0165f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:47:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3fb4717692aa1fec0012b514ef8d18d1ca3707cfa4bb8248bb93841a0a40f11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:47:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0b6a7afe3094f1c92093c05c28344897a615d590c2fbb507c87e18a44c6c831\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://640ade2c733d7916ec213a201a1dd614d261c69eab654c80100efbc99a768995\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:47:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:49:18Z is after 2025-08-24T17:21:41Z" Oct 05 06:49:18 crc kubenswrapper[4846]: I1005 06:49:18.527321 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:49:18Z is after 2025-08-24T17:21:41Z" Oct 05 06:49:18 crc kubenswrapper[4846]: I1005 06:49:18.539551 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e67d0e447a8a6a381259e25d25badc791194e1291f56b1cd77bd63989b20408\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dkhl7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://516f88f4990f5962d8b694a3d889259cc5508a309919870191c2e1331640c973\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dkhl7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-fscvf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:49:18Z is after 2025-08-24T17:21:41Z" Oct 05 06:49:18 crc kubenswrapper[4846]: I1005 06:49:18.551232 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-w2sj8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32931fce-7374-4cab-a2e0-79d96a842274\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0df535d65d39de6c9a3dfb931bafc4469adc22b72f7a7903beda4fef8cd48138\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9rtxf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:27Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-w2sj8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:49:18Z is after 2025-08-24T17:21:41Z" Oct 05 06:49:18 crc kubenswrapper[4846]: I1005 06:49:18.563421 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"78b204fd-1034-4ea1-bcae-e5c876594a39\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b94cc50fe15d58f375f8492067811be16278cdf13ee3d80cfed4682279157b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5e5cf9125d1d413c91666744aeab06980b44420dbf0341b165b05787527944b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://70f0d5b436d18a78a2d8e2767143ba1a3104332d86bd63f91d23cd8e77a952aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c236b4ed7f039a221baac043246353eecbb6812c1e6bf6836a5fa96ec8f5abdb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c236b4ed7f039a221baac043246353eecbb6812c1e6bf6836a5fa96ec8f5abdb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:47:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:47:59Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:47:58Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:49:18Z is after 2025-08-24T17:21:41Z" Oct 05 06:49:18 crc kubenswrapper[4846]: I1005 06:49:18.567981 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:18 crc kubenswrapper[4846]: I1005 06:49:18.568033 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:18 crc kubenswrapper[4846]: I1005 06:49:18.568047 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:18 crc kubenswrapper[4846]: I1005 06:49:18.568068 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:18 crc kubenswrapper[4846]: I1005 06:49:18.568083 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:18Z","lastTransitionTime":"2025-10-05T06:49:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:18 crc kubenswrapper[4846]: I1005 06:49:18.579113 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ff10c31459b3b3724ee7b769bd6f72f9ed80cce5419b6adcdf69a3239b70756\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:49:18Z is after 2025-08-24T17:21:41Z" Oct 05 06:49:18 crc kubenswrapper[4846]: I1005 06:49:18.591419 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://347f900743b7bba3daf92f21c7eba0625e8c774f50057bc9472f0c82119325e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:49:18Z is after 2025-08-24T17:21:41Z" Oct 05 06:49:18 crc kubenswrapper[4846]: I1005 06:49:18.605434 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56f1fe1f9f06f0249cf937a6963fda82467732dc056ca2b2189be6d8dd9d4a05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7212ce5f42c1fb55146abb190a557a504b43f16af6df84f90e1073567a329ec5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:49:18Z is after 2025-08-24T17:21:41Z" Oct 05 06:49:18 crc kubenswrapper[4846]: I1005 06:49:18.630760 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-7wm5q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9bc7744f-d700-450f-8b8d-98140877fee9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:49:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:49:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9d8631ba978f8fa7677feeeff42cabc454bfd51e5c5c21c78519fa9dfd85ec3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://90379e74c544fccca44f7acdc228ff378154390e211ac801b79f8d074012241e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-05T06:49:11Z\\\",\\\"message\\\":\\\"2025-10-05T06:48:26+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_0ecc6292-0274-4088-90d5-740e638af9d4\\\\n2025-10-05T06:48:26+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_0ecc6292-0274-4088-90d5-740e638af9d4 to /host/opt/cni/bin/\\\\n2025-10-05T06:48:26Z [verbose] multus-daemon started\\\\n2025-10-05T06:48:26Z [verbose] Readiness Indicator file check\\\\n2025-10-05T06:49:11Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:24Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:49:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h5p4q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-7wm5q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:49:18Z is after 2025-08-24T17:21:41Z" Oct 05 06:49:18 crc kubenswrapper[4846]: I1005 06:49:18.645754 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c21154c6-7c11-47e7-b1f0-6a5fe33d16b1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e13964315903099b3d7e3dc8c143fe58bb9f9654214702ba51581ac4044414d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef573b754a80e5148b8caa71b4f22c461f57dfa415394a5bccda9771b7684e4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef573b754a80e5148b8caa71b4f22c461f57dfa415394a5bccda9771b7684e4c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:47:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:47:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:47:58Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:49:18Z is after 2025-08-24T17:21:41Z" Oct 05 06:49:18 crc kubenswrapper[4846]: I1005 06:49:18.664874 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"775b5702-a131-4d6c-852d-91bfa056cac5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0839c1765df2c68d2554e9c66d289505e559a9672caf8069feeff2c4fdfd2171\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15dbaf2445704da0987f256f81dff2ee95664988541fa48ce94898afeadab2e8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd3b3ab323c6d5153054aa04af2d91f7b9e8436ff1e2f75f3594b57831b2a416\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://67c46f5f0a3ace95bd2e9479c5561126664e18b148c5eceea5bf5efc27e81a67\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c52e076eddfd307110b1f6de00a169b81af8f2785af7becef2a820ce363a5861\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-05T06:48:14Z\\\",\\\"message\\\":\\\"le observer\\\\nW1005 06:48:14.337870 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1005 06:48:14.338086 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1005 06:48:14.339389 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1467752215/tls.crt::/tmp/serving-cert-1467752215/tls.key\\\\\\\"\\\\nI1005 06:48:14.864345 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1005 06:48:14.868087 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1005 06:48:14.868114 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1005 06:48:14.868140 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1005 06:48:14.868147 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1005 06:48:14.877354 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1005 06:48:14.877412 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1005 06:48:14.877456 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1005 06:48:14.877471 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1005 06:48:14.877490 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1005 06:48:14.877501 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1005 06:48:14.877507 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1005 06:48:14.877515 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1005 06:48:14.880326 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:13Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4a091716ff935adb42ae2e88048619fd347bd174f64e5f9c97a01f81690e57f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:01Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dc5411b642f99b13d514d2f21defad0c394aeeb82a966f72300076617e2586c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5dc5411b642f99b13d514d2f21defad0c394aeeb82a966f72300076617e2586c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:47:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:47:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:47:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:49:18Z is after 2025-08-24T17:21:41Z" Oct 05 06:49:18 crc kubenswrapper[4846]: I1005 06:49:18.670932 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:18 crc kubenswrapper[4846]: I1005 06:49:18.670970 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:18 crc kubenswrapper[4846]: I1005 06:49:18.670981 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:18 crc kubenswrapper[4846]: I1005 06:49:18.671000 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:18 crc kubenswrapper[4846]: I1005 06:49:18.671010 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:18Z","lastTransitionTime":"2025-10-05T06:49:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:18 crc kubenswrapper[4846]: I1005 06:49:18.683343 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:49:18Z is after 2025-08-24T17:21:41Z" Oct 05 06:49:18 crc kubenswrapper[4846]: I1005 06:49:18.695385 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-r7ztb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a5119cd9-af70-4324-a955-978306ab0b20\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06231a87429ec292cb57ebd88cda7ad547bcb32e80343b776833621a11c28ad1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jfcj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-r7ztb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:49:18Z is after 2025-08-24T17:21:41Z" Oct 05 06:49:18 crc kubenswrapper[4846]: I1005 06:49:18.710548 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-9r2kt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"59b83f4f-71f0-4b28-9275-56d5f933ec2c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tdwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tdwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:38Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-9r2kt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:49:18Z is after 2025-08-24T17:21:41Z" Oct 05 06:49:18 crc kubenswrapper[4846]: I1005 06:49:18.741712 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c482bb97-fd50-40cf-96ea-f806c3f01125\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bfe45d30e4fa141541aca1d46bc5cc1da0a7b4a5eb1fc9bb2caa5b7fdcb0b264\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://496a4782faa30a9a20ff63038141600cd635f12902d1cf17989cd067fcdfe929\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://281be077869a6aa1ca815770871a14413d4a97e97aeb550fd5cd215cba88b21f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6771c7215d2f5dd5838d17888375a4768b9d1755fc2e21bda1bbfbb2eb1dbb95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c17a711c2c7473b076a3944ffe7c700808d05cba582fbd55c5c2ac8ff73838a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44a9f740f0a06c8bf928e2a9d1d5979217558c31e73a333631baccebf00fd5d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://44a9f740f0a06c8bf928e2a9d1d5979217558c31e73a333631baccebf00fd5d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:47:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:47:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://48ca0ba6ca6092bfcb5b153220c054f02cc36fb22707c42e021f06d329e18e2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48ca0ba6ca6092bfcb5b153220c054f02cc36fb22707c42e021f06d329e18e2e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:00Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4006d0cea128675c8885d879171204abffc5e11ea0664750a12b4bc7344ccb2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4006d0cea128675c8885d879171204abffc5e11ea0664750a12b4bc7344ccb2e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:47:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:49:18Z is after 2025-08-24T17:21:41Z" Oct 05 06:49:18 crc kubenswrapper[4846]: I1005 06:49:18.761877 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:49:18Z is after 2025-08-24T17:21:41Z" Oct 05 06:49:18 crc kubenswrapper[4846]: I1005 06:49:18.773587 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:18 crc kubenswrapper[4846]: I1005 06:49:18.773654 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:18 crc kubenswrapper[4846]: I1005 06:49:18.773671 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:18 crc kubenswrapper[4846]: I1005 06:49:18.773698 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:18 crc kubenswrapper[4846]: I1005 06:49:18.773720 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:18Z","lastTransitionTime":"2025-10-05T06:49:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:18 crc kubenswrapper[4846]: I1005 06:49:18.781726 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-sfmxd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1dbb13ac-136c-4a1c-9229-b4e8ac4f9c71\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b0af6a6e735f7b7877c177c1d53d1a8dc12c4713a8059c4766fc858af31c9a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://226d977b596d07aeda0486452e2a8605949ed4d2453b21f56174937c395d6f36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://226d977b596d07aeda0486452e2a8605949ed4d2453b21f56174937c395d6f36\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://99c07a4bdb3059d3a99ffdaf794d668d52cffd9e5e3bb8945204dd6d1f8c8328\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99c07a4bdb3059d3a99ffdaf794d668d52cffd9e5e3bb8945204dd6d1f8c8328\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27095272bf4f31d63a78a436636de0aae12f7103597da944f24ed7a74329ee3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27095272bf4f31d63a78a436636de0aae12f7103597da944f24ed7a74329ee3f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1211571af66d8a94302b65feccf30768ba82b7cbad918e44dd405787dce484cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1211571af66d8a94302b65feccf30768ba82b7cbad918e44dd405787dce484cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27631f90d528900fa180c15cfce80e96d4cdba07743eb6d847e2e822d9917218\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27631f90d528900fa180c15cfce80e96d4cdba07743eb6d847e2e822d9917218\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f345693e9aeb1f4dc196349d3354f3c833a2b0415a693298b1cfccfa2b0d4459\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f345693e9aeb1f4dc196349d3354f3c833a2b0415a693298b1cfccfa2b0d4459\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-sfmxd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:49:18Z is after 2025-08-24T17:21:41Z" Oct 05 06:49:18 crc kubenswrapper[4846]: I1005 06:49:18.812785 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-txzlk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"93a6ace4-8d64-44d7-9637-457d3af6543c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ad1940d07e37217089642478ece9ccdaca74784738a400929a9f5359d158c01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a6f6b2c32dd3da5f9fd5e6971c1773e81a9a24184b7aa5b543cec879001ae53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e9b0219ecbc7e60905d6de3e67a2c8dc9af1017b4aed9538ee0c58c861f2438\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af5e5a4da9ad7371d5a24633e8268f12e3a7d419a8c93b11ddbcc006d8339c6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8da85222775abc7a772f6a8ff8255f1a0d5b15807ba177a7e3e2b2d01a4f24b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1409651de8c897804f87be03d016c138f8acf27e662dce265a93e15c1848adcf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://59fa020f7bfea4547e132058cf3804a406ee0c06cdc4e78a08ae43a51f46633f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd084bb9806ece6529db0866cee69e6a3152f0da7e0f13fc2e4adcbe2f55962c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-05T06:48:50Z\\\",\\\"message\\\":\\\":50.819394 6534 handler.go:208] Removed *v1.Node event handler 2\\\\nI1005 06:48:50.819434 6534 handler.go:208] Removed *v1.Node event handler 7\\\\nI1005 06:48:50.819451 6534 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1005 06:48:50.819624 6534 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1005 06:48:50.819659 6534 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1005 06:48:50.819711 6534 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1005 06:48:50.819735 6534 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1005 06:48:50.820556 6534 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1005 06:48:50.820605 6534 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1005 06:48:50.820649 6534 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1005 06:48:50.820664 6534 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1005 06:48:50.820706 6534 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1005 06:48:50.820711 6534 factory.go:656] Stopping watch factory\\\\nI1005 06:48:50.820731 6534 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1005 06:48:50.820750 6534 ovnkube.go:599] Stopped ovnkube\\\\nI1005 06:48:50.820796 6534 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1005 06:48:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:49Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:49:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0df0ebcd87dc5af4b17736b43ec7b2395968b19ed16e775366af14e2b2d49c52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff3c6545ff3d2436da27bbd223b950b30b0c69a3e7f3034cb0e61380cb38d338\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ff3c6545ff3d2436da27bbd223b950b30b0c69a3e7f3034cb0e61380cb38d338\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:24Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-txzlk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:49:18Z is after 2025-08-24T17:21:41Z" Oct 05 06:49:18 crc kubenswrapper[4846]: I1005 06:49:18.833581 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-gklb6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"698ab882-bb26-4b4c-81e9-8305a44fe068\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a6edca88ea244a58f8cf6cc431110a960c6bfd892812cd0d00fa3f97be3bc67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x6d6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c191538ab797b8056820de7da1f8e93ea18873891a32382b8a59c9af762e103\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x6d6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:37Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-gklb6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:49:18Z is after 2025-08-24T17:21:41Z" Oct 05 06:49:18 crc kubenswrapper[4846]: I1005 06:49:18.877446 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:18 crc kubenswrapper[4846]: I1005 06:49:18.877500 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:18 crc kubenswrapper[4846]: I1005 06:49:18.877518 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:18 crc kubenswrapper[4846]: I1005 06:49:18.877547 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:18 crc kubenswrapper[4846]: I1005 06:49:18.877567 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:18Z","lastTransitionTime":"2025-10-05T06:49:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:18 crc kubenswrapper[4846]: I1005 06:49:18.981753 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:18 crc kubenswrapper[4846]: I1005 06:49:18.981806 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:18 crc kubenswrapper[4846]: I1005 06:49:18.981818 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:18 crc kubenswrapper[4846]: I1005 06:49:18.981837 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:18 crc kubenswrapper[4846]: I1005 06:49:18.981851 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:18Z","lastTransitionTime":"2025-10-05T06:49:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:19 crc kubenswrapper[4846]: I1005 06:49:19.082526 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-txzlk_93a6ace4-8d64-44d7-9637-457d3af6543c/ovnkube-controller/3.log" Oct 05 06:49:19 crc kubenswrapper[4846]: I1005 06:49:19.083797 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-txzlk_93a6ace4-8d64-44d7-9637-457d3af6543c/ovnkube-controller/2.log" Oct 05 06:49:19 crc kubenswrapper[4846]: I1005 06:49:19.084150 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:19 crc kubenswrapper[4846]: I1005 06:49:19.084229 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:19 crc kubenswrapper[4846]: I1005 06:49:19.084246 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:19 crc kubenswrapper[4846]: I1005 06:49:19.084269 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:19 crc kubenswrapper[4846]: I1005 06:49:19.084290 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:19Z","lastTransitionTime":"2025-10-05T06:49:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:19 crc kubenswrapper[4846]: I1005 06:49:19.090016 4846 generic.go:334] "Generic (PLEG): container finished" podID="93a6ace4-8d64-44d7-9637-457d3af6543c" containerID="59fa020f7bfea4547e132058cf3804a406ee0c06cdc4e78a08ae43a51f46633f" exitCode=1 Oct 05 06:49:19 crc kubenswrapper[4846]: I1005 06:49:19.090098 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-txzlk" event={"ID":"93a6ace4-8d64-44d7-9637-457d3af6543c","Type":"ContainerDied","Data":"59fa020f7bfea4547e132058cf3804a406ee0c06cdc4e78a08ae43a51f46633f"} Oct 05 06:49:19 crc kubenswrapper[4846]: I1005 06:49:19.090171 4846 scope.go:117] "RemoveContainer" containerID="dd084bb9806ece6529db0866cee69e6a3152f0da7e0f13fc2e4adcbe2f55962c" Oct 05 06:49:19 crc kubenswrapper[4846]: I1005 06:49:19.091146 4846 scope.go:117] "RemoveContainer" containerID="59fa020f7bfea4547e132058cf3804a406ee0c06cdc4e78a08ae43a51f46633f" Oct 05 06:49:19 crc kubenswrapper[4846]: E1005 06:49:19.091437 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-txzlk_openshift-ovn-kubernetes(93a6ace4-8d64-44d7-9637-457d3af6543c)\"" pod="openshift-ovn-kubernetes/ovnkube-node-txzlk" podUID="93a6ace4-8d64-44d7-9637-457d3af6543c" Oct 05 06:49:19 crc kubenswrapper[4846]: I1005 06:49:19.130345 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c482bb97-fd50-40cf-96ea-f806c3f01125\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bfe45d30e4fa141541aca1d46bc5cc1da0a7b4a5eb1fc9bb2caa5b7fdcb0b264\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://496a4782faa30a9a20ff63038141600cd635f12902d1cf17989cd067fcdfe929\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://281be077869a6aa1ca815770871a14413d4a97e97aeb550fd5cd215cba88b21f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6771c7215d2f5dd5838d17888375a4768b9d1755fc2e21bda1bbfbb2eb1dbb95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c17a711c2c7473b076a3944ffe7c700808d05cba582fbd55c5c2ac8ff73838a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44a9f740f0a06c8bf928e2a9d1d5979217558c31e73a333631baccebf00fd5d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://44a9f740f0a06c8bf928e2a9d1d5979217558c31e73a333631baccebf00fd5d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:47:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:47:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://48ca0ba6ca6092bfcb5b153220c054f02cc36fb22707c42e021f06d329e18e2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48ca0ba6ca6092bfcb5b153220c054f02cc36fb22707c42e021f06d329e18e2e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:00Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4006d0cea128675c8885d879171204abffc5e11ea0664750a12b4bc7344ccb2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4006d0cea128675c8885d879171204abffc5e11ea0664750a12b4bc7344ccb2e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:47:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:49:19Z is after 2025-08-24T17:21:41Z" Oct 05 06:49:19 crc kubenswrapper[4846]: I1005 06:49:19.154095 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:49:19Z is after 2025-08-24T17:21:41Z" Oct 05 06:49:19 crc kubenswrapper[4846]: I1005 06:49:19.181307 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-sfmxd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1dbb13ac-136c-4a1c-9229-b4e8ac4f9c71\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b0af6a6e735f7b7877c177c1d53d1a8dc12c4713a8059c4766fc858af31c9a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://226d977b596d07aeda0486452e2a8605949ed4d2453b21f56174937c395d6f36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://226d977b596d07aeda0486452e2a8605949ed4d2453b21f56174937c395d6f36\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://99c07a4bdb3059d3a99ffdaf794d668d52cffd9e5e3bb8945204dd6d1f8c8328\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99c07a4bdb3059d3a99ffdaf794d668d52cffd9e5e3bb8945204dd6d1f8c8328\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27095272bf4f31d63a78a436636de0aae12f7103597da944f24ed7a74329ee3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27095272bf4f31d63a78a436636de0aae12f7103597da944f24ed7a74329ee3f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1211571af66d8a94302b65feccf30768ba82b7cbad918e44dd405787dce484cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1211571af66d8a94302b65feccf30768ba82b7cbad918e44dd405787dce484cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27631f90d528900fa180c15cfce80e96d4cdba07743eb6d847e2e822d9917218\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27631f90d528900fa180c15cfce80e96d4cdba07743eb6d847e2e822d9917218\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f345693e9aeb1f4dc196349d3354f3c833a2b0415a693298b1cfccfa2b0d4459\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f345693e9aeb1f4dc196349d3354f3c833a2b0415a693298b1cfccfa2b0d4459\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-sfmxd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:49:19Z is after 2025-08-24T17:21:41Z" Oct 05 06:49:19 crc kubenswrapper[4846]: I1005 06:49:19.187968 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:19 crc kubenswrapper[4846]: I1005 06:49:19.188050 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:19 crc kubenswrapper[4846]: I1005 06:49:19.188071 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:19 crc kubenswrapper[4846]: I1005 06:49:19.188105 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:19 crc kubenswrapper[4846]: I1005 06:49:19.188130 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:19Z","lastTransitionTime":"2025-10-05T06:49:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:19 crc kubenswrapper[4846]: I1005 06:49:19.215359 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-txzlk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"93a6ace4-8d64-44d7-9637-457d3af6543c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ad1940d07e37217089642478ece9ccdaca74784738a400929a9f5359d158c01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a6f6b2c32dd3da5f9fd5e6971c1773e81a9a24184b7aa5b543cec879001ae53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e9b0219ecbc7e60905d6de3e67a2c8dc9af1017b4aed9538ee0c58c861f2438\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af5e5a4da9ad7371d5a24633e8268f12e3a7d419a8c93b11ddbcc006d8339c6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8da85222775abc7a772f6a8ff8255f1a0d5b15807ba177a7e3e2b2d01a4f24b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1409651de8c897804f87be03d016c138f8acf27e662dce265a93e15c1848adcf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://59fa020f7bfea4547e132058cf3804a406ee0c06cdc4e78a08ae43a51f46633f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dd084bb9806ece6529db0866cee69e6a3152f0da7e0f13fc2e4adcbe2f55962c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-05T06:48:50Z\\\",\\\"message\\\":\\\":50.819394 6534 handler.go:208] Removed *v1.Node event handler 2\\\\nI1005 06:48:50.819434 6534 handler.go:208] Removed *v1.Node event handler 7\\\\nI1005 06:48:50.819451 6534 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1005 06:48:50.819624 6534 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1005 06:48:50.819659 6534 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1005 06:48:50.819711 6534 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1005 06:48:50.819735 6534 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1005 06:48:50.820556 6534 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1005 06:48:50.820605 6534 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1005 06:48:50.820649 6534 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1005 06:48:50.820664 6534 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1005 06:48:50.820706 6534 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1005 06:48:50.820711 6534 factory.go:656] Stopping watch factory\\\\nI1005 06:48:50.820731 6534 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1005 06:48:50.820750 6534 ovnkube.go:599] Stopped ovnkube\\\\nI1005 06:48:50.820796 6534 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1005 06:48:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:49Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://59fa020f7bfea4547e132058cf3804a406ee0c06cdc4e78a08ae43a51f46633f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-05T06:49:18Z\\\",\\\"message\\\":\\\" start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:49:18Z is after 2025-08-24T17:21:41Z]\\\\nI1005 06:49:18.451492 6882 services_controller.go:473] Services do not match for network=default, existing lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-marketplace/redhat-marketplace_TCP_cluster\\\\\\\", UUID:\\\\\\\"97b6e7b0-06ca-455e-8259-06895040cb0c\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-marketplace/redhat-marketplace\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T06:49:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0df0ebcd87dc5af4b17736b43ec7b2395968b19ed16e775366af14e2b2d49c52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff3c6545ff3d2436da27bbd223b950b30b0c69a3e7f3034cb0e61380cb38d338\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ff3c6545ff3d2436da27bbd223b950b30b0c69a3e7f3034cb0e61380cb38d338\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:24Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-txzlk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:49:19Z is after 2025-08-24T17:21:41Z" Oct 05 06:49:19 crc kubenswrapper[4846]: I1005 06:49:19.235933 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-gklb6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"698ab882-bb26-4b4c-81e9-8305a44fe068\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a6edca88ea244a58f8cf6cc431110a960c6bfd892812cd0d00fa3f97be3bc67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x6d6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c191538ab797b8056820de7da1f8e93ea18873891a32382b8a59c9af762e103\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x6d6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:37Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-gklb6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:49:19Z is after 2025-08-24T17:21:41Z" Oct 05 06:49:19 crc kubenswrapper[4846]: I1005 06:49:19.256565 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c9f812b-ecc8-4985-a670-53fa80fbb499\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2b5ee9781c01d6aec3ac778efef1067b6c3bfe4ed37431bdebe188fcde0165f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:47:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3fb4717692aa1fec0012b514ef8d18d1ca3707cfa4bb8248bb93841a0a40f11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:47:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0b6a7afe3094f1c92093c05c28344897a615d590c2fbb507c87e18a44c6c831\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://640ade2c733d7916ec213a201a1dd614d261c69eab654c80100efbc99a768995\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:47:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:49:19Z is after 2025-08-24T17:21:41Z" Oct 05 06:49:19 crc kubenswrapper[4846]: I1005 06:49:19.272843 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:49:19Z is after 2025-08-24T17:21:41Z" Oct 05 06:49:19 crc kubenswrapper[4846]: I1005 06:49:19.287435 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e67d0e447a8a6a381259e25d25badc791194e1291f56b1cd77bd63989b20408\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dkhl7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://516f88f4990f5962d8b694a3d889259cc5508a309919870191c2e1331640c973\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dkhl7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-fscvf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:49:19Z is after 2025-08-24T17:21:41Z" Oct 05 06:49:19 crc kubenswrapper[4846]: I1005 06:49:19.291412 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:19 crc kubenswrapper[4846]: I1005 06:49:19.291454 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:19 crc kubenswrapper[4846]: I1005 06:49:19.291469 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:19 crc kubenswrapper[4846]: I1005 06:49:19.291492 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:19 crc kubenswrapper[4846]: I1005 06:49:19.291508 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:19Z","lastTransitionTime":"2025-10-05T06:49:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:19 crc kubenswrapper[4846]: I1005 06:49:19.300612 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-w2sj8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32931fce-7374-4cab-a2e0-79d96a842274\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0df535d65d39de6c9a3dfb931bafc4469adc22b72f7a7903beda4fef8cd48138\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9rtxf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:27Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-w2sj8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:49:19Z is after 2025-08-24T17:21:41Z" Oct 05 06:49:19 crc kubenswrapper[4846]: I1005 06:49:19.314211 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"78b204fd-1034-4ea1-bcae-e5c876594a39\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b94cc50fe15d58f375f8492067811be16278cdf13ee3d80cfed4682279157b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5e5cf9125d1d413c91666744aeab06980b44420dbf0341b165b05787527944b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://70f0d5b436d18a78a2d8e2767143ba1a3104332d86bd63f91d23cd8e77a952aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c236b4ed7f039a221baac043246353eecbb6812c1e6bf6836a5fa96ec8f5abdb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c236b4ed7f039a221baac043246353eecbb6812c1e6bf6836a5fa96ec8f5abdb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:47:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:47:59Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:47:58Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:49:19Z is after 2025-08-24T17:21:41Z" Oct 05 06:49:19 crc kubenswrapper[4846]: I1005 06:49:19.328555 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ff10c31459b3b3724ee7b769bd6f72f9ed80cce5419b6adcdf69a3239b70756\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:49:19Z is after 2025-08-24T17:21:41Z" Oct 05 06:49:19 crc kubenswrapper[4846]: I1005 06:49:19.346084 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://347f900743b7bba3daf92f21c7eba0625e8c774f50057bc9472f0c82119325e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:49:19Z is after 2025-08-24T17:21:41Z" Oct 05 06:49:19 crc kubenswrapper[4846]: I1005 06:49:19.362452 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56f1fe1f9f06f0249cf937a6963fda82467732dc056ca2b2189be6d8dd9d4a05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7212ce5f42c1fb55146abb190a557a504b43f16af6df84f90e1073567a329ec5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:49:19Z is after 2025-08-24T17:21:41Z" Oct 05 06:49:19 crc kubenswrapper[4846]: I1005 06:49:19.376142 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-7wm5q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9bc7744f-d700-450f-8b8d-98140877fee9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:49:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:49:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9d8631ba978f8fa7677feeeff42cabc454bfd51e5c5c21c78519fa9dfd85ec3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://90379e74c544fccca44f7acdc228ff378154390e211ac801b79f8d074012241e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-05T06:49:11Z\\\",\\\"message\\\":\\\"2025-10-05T06:48:26+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_0ecc6292-0274-4088-90d5-740e638af9d4\\\\n2025-10-05T06:48:26+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_0ecc6292-0274-4088-90d5-740e638af9d4 to /host/opt/cni/bin/\\\\n2025-10-05T06:48:26Z [verbose] multus-daemon started\\\\n2025-10-05T06:48:26Z [verbose] Readiness Indicator file check\\\\n2025-10-05T06:49:11Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:24Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:49:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h5p4q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-7wm5q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:49:19Z is after 2025-08-24T17:21:41Z" Oct 05 06:49:19 crc kubenswrapper[4846]: I1005 06:49:19.393707 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:19 crc kubenswrapper[4846]: I1005 06:49:19.393756 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:19 crc kubenswrapper[4846]: I1005 06:49:19.393768 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:19 crc kubenswrapper[4846]: I1005 06:49:19.393784 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:19 crc kubenswrapper[4846]: I1005 06:49:19.393796 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:19Z","lastTransitionTime":"2025-10-05T06:49:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:19 crc kubenswrapper[4846]: I1005 06:49:19.396738 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c21154c6-7c11-47e7-b1f0-6a5fe33d16b1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e13964315903099b3d7e3dc8c143fe58bb9f9654214702ba51581ac4044414d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef573b754a80e5148b8caa71b4f22c461f57dfa415394a5bccda9771b7684e4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef573b754a80e5148b8caa71b4f22c461f57dfa415394a5bccda9771b7684e4c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:47:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:47:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:47:58Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:49:19Z is after 2025-08-24T17:21:41Z" Oct 05 06:49:19 crc kubenswrapper[4846]: I1005 06:49:19.410446 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"775b5702-a131-4d6c-852d-91bfa056cac5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0839c1765df2c68d2554e9c66d289505e559a9672caf8069feeff2c4fdfd2171\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15dbaf2445704da0987f256f81dff2ee95664988541fa48ce94898afeadab2e8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd3b3ab323c6d5153054aa04af2d91f7b9e8436ff1e2f75f3594b57831b2a416\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://67c46f5f0a3ace95bd2e9479c5561126664e18b148c5eceea5bf5efc27e81a67\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c52e076eddfd307110b1f6de00a169b81af8f2785af7becef2a820ce363a5861\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-05T06:48:14Z\\\",\\\"message\\\":\\\"le observer\\\\nW1005 06:48:14.337870 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1005 06:48:14.338086 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1005 06:48:14.339389 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1467752215/tls.crt::/tmp/serving-cert-1467752215/tls.key\\\\\\\"\\\\nI1005 06:48:14.864345 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1005 06:48:14.868087 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1005 06:48:14.868114 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1005 06:48:14.868140 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1005 06:48:14.868147 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1005 06:48:14.877354 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1005 06:48:14.877412 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1005 06:48:14.877456 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1005 06:48:14.877471 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1005 06:48:14.877490 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1005 06:48:14.877501 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1005 06:48:14.877507 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1005 06:48:14.877515 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1005 06:48:14.880326 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:13Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4a091716ff935adb42ae2e88048619fd347bd174f64e5f9c97a01f81690e57f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:01Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dc5411b642f99b13d514d2f21defad0c394aeeb82a966f72300076617e2586c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5dc5411b642f99b13d514d2f21defad0c394aeeb82a966f72300076617e2586c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:47:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:47:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:47:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:49:19Z is after 2025-08-24T17:21:41Z" Oct 05 06:49:19 crc kubenswrapper[4846]: I1005 06:49:19.426988 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:49:19Z is after 2025-08-24T17:21:41Z" Oct 05 06:49:19 crc kubenswrapper[4846]: I1005 06:49:19.441441 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-r7ztb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a5119cd9-af70-4324-a955-978306ab0b20\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06231a87429ec292cb57ebd88cda7ad547bcb32e80343b776833621a11c28ad1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jfcj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-r7ztb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:49:19Z is after 2025-08-24T17:21:41Z" Oct 05 06:49:19 crc kubenswrapper[4846]: I1005 06:49:19.452740 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-9r2kt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"59b83f4f-71f0-4b28-9275-56d5f933ec2c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tdwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tdwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:38Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-9r2kt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:49:19Z is after 2025-08-24T17:21:41Z" Oct 05 06:49:19 crc kubenswrapper[4846]: I1005 06:49:19.496417 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 05 06:49:19 crc kubenswrapper[4846]: E1005 06:49:19.496524 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 05 06:49:19 crc kubenswrapper[4846]: I1005 06:49:19.496702 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:19 crc kubenswrapper[4846]: I1005 06:49:19.496722 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:19 crc kubenswrapper[4846]: I1005 06:49:19.496729 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:19 crc kubenswrapper[4846]: I1005 06:49:19.496741 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:19 crc kubenswrapper[4846]: I1005 06:49:19.496751 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:19Z","lastTransitionTime":"2025-10-05T06:49:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:19 crc kubenswrapper[4846]: I1005 06:49:19.497089 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 05 06:49:19 crc kubenswrapper[4846]: E1005 06:49:19.497143 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 05 06:49:19 crc kubenswrapper[4846]: I1005 06:49:19.600387 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:19 crc kubenswrapper[4846]: I1005 06:49:19.600484 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:19 crc kubenswrapper[4846]: I1005 06:49:19.600512 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:19 crc kubenswrapper[4846]: I1005 06:49:19.600548 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:19 crc kubenswrapper[4846]: I1005 06:49:19.600579 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:19Z","lastTransitionTime":"2025-10-05T06:49:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:19 crc kubenswrapper[4846]: I1005 06:49:19.703884 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:19 crc kubenswrapper[4846]: I1005 06:49:19.703961 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:19 crc kubenswrapper[4846]: I1005 06:49:19.703982 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:19 crc kubenswrapper[4846]: I1005 06:49:19.704049 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:19 crc kubenswrapper[4846]: I1005 06:49:19.704068 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:19Z","lastTransitionTime":"2025-10-05T06:49:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:19 crc kubenswrapper[4846]: I1005 06:49:19.807886 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:19 crc kubenswrapper[4846]: I1005 06:49:19.807957 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:19 crc kubenswrapper[4846]: I1005 06:49:19.807977 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:19 crc kubenswrapper[4846]: I1005 06:49:19.808009 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:19 crc kubenswrapper[4846]: I1005 06:49:19.808034 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:19Z","lastTransitionTime":"2025-10-05T06:49:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:19 crc kubenswrapper[4846]: I1005 06:49:19.911079 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:19 crc kubenswrapper[4846]: I1005 06:49:19.911139 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:19 crc kubenswrapper[4846]: I1005 06:49:19.911156 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:19 crc kubenswrapper[4846]: I1005 06:49:19.911193 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:19 crc kubenswrapper[4846]: I1005 06:49:19.911207 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:19Z","lastTransitionTime":"2025-10-05T06:49:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:20 crc kubenswrapper[4846]: I1005 06:49:20.014492 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:20 crc kubenswrapper[4846]: I1005 06:49:20.014559 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:20 crc kubenswrapper[4846]: I1005 06:49:20.014578 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:20 crc kubenswrapper[4846]: I1005 06:49:20.014605 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:20 crc kubenswrapper[4846]: I1005 06:49:20.014624 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:20Z","lastTransitionTime":"2025-10-05T06:49:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:20 crc kubenswrapper[4846]: I1005 06:49:20.096022 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-txzlk_93a6ace4-8d64-44d7-9637-457d3af6543c/ovnkube-controller/3.log" Oct 05 06:49:20 crc kubenswrapper[4846]: I1005 06:49:20.102559 4846 scope.go:117] "RemoveContainer" containerID="59fa020f7bfea4547e132058cf3804a406ee0c06cdc4e78a08ae43a51f46633f" Oct 05 06:49:20 crc kubenswrapper[4846]: E1005 06:49:20.102869 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-txzlk_openshift-ovn-kubernetes(93a6ace4-8d64-44d7-9637-457d3af6543c)\"" pod="openshift-ovn-kubernetes/ovnkube-node-txzlk" podUID="93a6ace4-8d64-44d7-9637-457d3af6543c" Oct 05 06:49:20 crc kubenswrapper[4846]: I1005 06:49:20.117279 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:20 crc kubenswrapper[4846]: I1005 06:49:20.117321 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:20 crc kubenswrapper[4846]: I1005 06:49:20.117339 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:20 crc kubenswrapper[4846]: I1005 06:49:20.117361 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:20 crc kubenswrapper[4846]: I1005 06:49:20.117421 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:20Z","lastTransitionTime":"2025-10-05T06:49:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:20 crc kubenswrapper[4846]: I1005 06:49:20.134373 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c482bb97-fd50-40cf-96ea-f806c3f01125\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bfe45d30e4fa141541aca1d46bc5cc1da0a7b4a5eb1fc9bb2caa5b7fdcb0b264\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://496a4782faa30a9a20ff63038141600cd635f12902d1cf17989cd067fcdfe929\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://281be077869a6aa1ca815770871a14413d4a97e97aeb550fd5cd215cba88b21f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6771c7215d2f5dd5838d17888375a4768b9d1755fc2e21bda1bbfbb2eb1dbb95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c17a711c2c7473b076a3944ffe7c700808d05cba582fbd55c5c2ac8ff73838a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44a9f740f0a06c8bf928e2a9d1d5979217558c31e73a333631baccebf00fd5d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://44a9f740f0a06c8bf928e2a9d1d5979217558c31e73a333631baccebf00fd5d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:47:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:47:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://48ca0ba6ca6092bfcb5b153220c054f02cc36fb22707c42e021f06d329e18e2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48ca0ba6ca6092bfcb5b153220c054f02cc36fb22707c42e021f06d329e18e2e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:00Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4006d0cea128675c8885d879171204abffc5e11ea0664750a12b4bc7344ccb2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4006d0cea128675c8885d879171204abffc5e11ea0664750a12b4bc7344ccb2e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:47:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:49:20Z is after 2025-08-24T17:21:41Z" Oct 05 06:49:20 crc kubenswrapper[4846]: I1005 06:49:20.155291 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:49:20Z is after 2025-08-24T17:21:41Z" Oct 05 06:49:20 crc kubenswrapper[4846]: I1005 06:49:20.184747 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-sfmxd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1dbb13ac-136c-4a1c-9229-b4e8ac4f9c71\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b0af6a6e735f7b7877c177c1d53d1a8dc12c4713a8059c4766fc858af31c9a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://226d977b596d07aeda0486452e2a8605949ed4d2453b21f56174937c395d6f36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://226d977b596d07aeda0486452e2a8605949ed4d2453b21f56174937c395d6f36\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://99c07a4bdb3059d3a99ffdaf794d668d52cffd9e5e3bb8945204dd6d1f8c8328\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99c07a4bdb3059d3a99ffdaf794d668d52cffd9e5e3bb8945204dd6d1f8c8328\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27095272bf4f31d63a78a436636de0aae12f7103597da944f24ed7a74329ee3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27095272bf4f31d63a78a436636de0aae12f7103597da944f24ed7a74329ee3f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1211571af66d8a94302b65feccf30768ba82b7cbad918e44dd405787dce484cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1211571af66d8a94302b65feccf30768ba82b7cbad918e44dd405787dce484cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27631f90d528900fa180c15cfce80e96d4cdba07743eb6d847e2e822d9917218\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27631f90d528900fa180c15cfce80e96d4cdba07743eb6d847e2e822d9917218\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f345693e9aeb1f4dc196349d3354f3c833a2b0415a693298b1cfccfa2b0d4459\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f345693e9aeb1f4dc196349d3354f3c833a2b0415a693298b1cfccfa2b0d4459\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-sfmxd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:49:20Z is after 2025-08-24T17:21:41Z" Oct 05 06:49:20 crc kubenswrapper[4846]: I1005 06:49:20.218618 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-txzlk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"93a6ace4-8d64-44d7-9637-457d3af6543c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ad1940d07e37217089642478ece9ccdaca74784738a400929a9f5359d158c01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a6f6b2c32dd3da5f9fd5e6971c1773e81a9a24184b7aa5b543cec879001ae53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e9b0219ecbc7e60905d6de3e67a2c8dc9af1017b4aed9538ee0c58c861f2438\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af5e5a4da9ad7371d5a24633e8268f12e3a7d419a8c93b11ddbcc006d8339c6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8da85222775abc7a772f6a8ff8255f1a0d5b15807ba177a7e3e2b2d01a4f24b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1409651de8c897804f87be03d016c138f8acf27e662dce265a93e15c1848adcf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://59fa020f7bfea4547e132058cf3804a406ee0c06cdc4e78a08ae43a51f46633f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://59fa020f7bfea4547e132058cf3804a406ee0c06cdc4e78a08ae43a51f46633f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-05T06:49:18Z\\\",\\\"message\\\":\\\" start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:49:18Z is after 2025-08-24T17:21:41Z]\\\\nI1005 06:49:18.451492 6882 services_controller.go:473] Services do not match for network=default, existing lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-marketplace/redhat-marketplace_TCP_cluster\\\\\\\", UUID:\\\\\\\"97b6e7b0-06ca-455e-8259-06895040cb0c\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-marketplace/redhat-marketplace\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T06:49:17Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-txzlk_openshift-ovn-kubernetes(93a6ace4-8d64-44d7-9637-457d3af6543c)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0df0ebcd87dc5af4b17736b43ec7b2395968b19ed16e775366af14e2b2d49c52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff3c6545ff3d2436da27bbd223b950b30b0c69a3e7f3034cb0e61380cb38d338\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ff3c6545ff3d2436da27bbd223b950b30b0c69a3e7f3034cb0e61380cb38d338\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:24Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-txzlk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:49:20Z is after 2025-08-24T17:21:41Z" Oct 05 06:49:20 crc kubenswrapper[4846]: I1005 06:49:20.220575 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:20 crc kubenswrapper[4846]: I1005 06:49:20.220625 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:20 crc kubenswrapper[4846]: I1005 06:49:20.220642 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:20 crc kubenswrapper[4846]: I1005 06:49:20.220667 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:20 crc kubenswrapper[4846]: I1005 06:49:20.220684 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:20Z","lastTransitionTime":"2025-10-05T06:49:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:20 crc kubenswrapper[4846]: I1005 06:49:20.239453 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-gklb6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"698ab882-bb26-4b4c-81e9-8305a44fe068\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a6edca88ea244a58f8cf6cc431110a960c6bfd892812cd0d00fa3f97be3bc67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x6d6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c191538ab797b8056820de7da1f8e93ea18873891a32382b8a59c9af762e103\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x6d6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:37Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-gklb6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:49:20Z is after 2025-08-24T17:21:41Z" Oct 05 06:49:20 crc kubenswrapper[4846]: I1005 06:49:20.261547 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c9f812b-ecc8-4985-a670-53fa80fbb499\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2b5ee9781c01d6aec3ac778efef1067b6c3bfe4ed37431bdebe188fcde0165f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:47:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3fb4717692aa1fec0012b514ef8d18d1ca3707cfa4bb8248bb93841a0a40f11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:47:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0b6a7afe3094f1c92093c05c28344897a615d590c2fbb507c87e18a44c6c831\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://640ade2c733d7916ec213a201a1dd614d261c69eab654c80100efbc99a768995\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:47:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:49:20Z is after 2025-08-24T17:21:41Z" Oct 05 06:49:20 crc kubenswrapper[4846]: I1005 06:49:20.280883 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:49:20Z is after 2025-08-24T17:21:41Z" Oct 05 06:49:20 crc kubenswrapper[4846]: I1005 06:49:20.298448 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e67d0e447a8a6a381259e25d25badc791194e1291f56b1cd77bd63989b20408\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dkhl7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://516f88f4990f5962d8b694a3d889259cc5508a309919870191c2e1331640c973\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dkhl7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-fscvf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:49:20Z is after 2025-08-24T17:21:41Z" Oct 05 06:49:20 crc kubenswrapper[4846]: I1005 06:49:20.317149 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-w2sj8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32931fce-7374-4cab-a2e0-79d96a842274\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0df535d65d39de6c9a3dfb931bafc4469adc22b72f7a7903beda4fef8cd48138\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9rtxf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:27Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-w2sj8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:49:20Z is after 2025-08-24T17:21:41Z" Oct 05 06:49:20 crc kubenswrapper[4846]: I1005 06:49:20.325450 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:20 crc kubenswrapper[4846]: I1005 06:49:20.325532 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:20 crc kubenswrapper[4846]: I1005 06:49:20.325555 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:20 crc kubenswrapper[4846]: I1005 06:49:20.325585 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:20 crc kubenswrapper[4846]: I1005 06:49:20.325606 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:20Z","lastTransitionTime":"2025-10-05T06:49:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:20 crc kubenswrapper[4846]: I1005 06:49:20.338421 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"78b204fd-1034-4ea1-bcae-e5c876594a39\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b94cc50fe15d58f375f8492067811be16278cdf13ee3d80cfed4682279157b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5e5cf9125d1d413c91666744aeab06980b44420dbf0341b165b05787527944b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://70f0d5b436d18a78a2d8e2767143ba1a3104332d86bd63f91d23cd8e77a952aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c236b4ed7f039a221baac043246353eecbb6812c1e6bf6836a5fa96ec8f5abdb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c236b4ed7f039a221baac043246353eecbb6812c1e6bf6836a5fa96ec8f5abdb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:47:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:47:59Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:47:58Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:49:20Z is after 2025-08-24T17:21:41Z" Oct 05 06:49:20 crc kubenswrapper[4846]: I1005 06:49:20.360704 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ff10c31459b3b3724ee7b769bd6f72f9ed80cce5419b6adcdf69a3239b70756\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:49:20Z is after 2025-08-24T17:21:41Z" Oct 05 06:49:20 crc kubenswrapper[4846]: I1005 06:49:20.380157 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://347f900743b7bba3daf92f21c7eba0625e8c774f50057bc9472f0c82119325e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:49:20Z is after 2025-08-24T17:21:41Z" Oct 05 06:49:20 crc kubenswrapper[4846]: I1005 06:49:20.402570 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56f1fe1f9f06f0249cf937a6963fda82467732dc056ca2b2189be6d8dd9d4a05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7212ce5f42c1fb55146abb190a557a504b43f16af6df84f90e1073567a329ec5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:49:20Z is after 2025-08-24T17:21:41Z" Oct 05 06:49:20 crc kubenswrapper[4846]: I1005 06:49:20.422676 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-7wm5q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9bc7744f-d700-450f-8b8d-98140877fee9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:49:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:49:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9d8631ba978f8fa7677feeeff42cabc454bfd51e5c5c21c78519fa9dfd85ec3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://90379e74c544fccca44f7acdc228ff378154390e211ac801b79f8d074012241e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-05T06:49:11Z\\\",\\\"message\\\":\\\"2025-10-05T06:48:26+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_0ecc6292-0274-4088-90d5-740e638af9d4\\\\n2025-10-05T06:48:26+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_0ecc6292-0274-4088-90d5-740e638af9d4 to /host/opt/cni/bin/\\\\n2025-10-05T06:48:26Z [verbose] multus-daemon started\\\\n2025-10-05T06:48:26Z [verbose] Readiness Indicator file check\\\\n2025-10-05T06:49:11Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:24Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:49:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h5p4q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-7wm5q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:49:20Z is after 2025-08-24T17:21:41Z" Oct 05 06:49:20 crc kubenswrapper[4846]: I1005 06:49:20.430144 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:20 crc kubenswrapper[4846]: I1005 06:49:20.430312 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:20 crc kubenswrapper[4846]: I1005 06:49:20.430386 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:20 crc kubenswrapper[4846]: I1005 06:49:20.430415 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:20 crc kubenswrapper[4846]: I1005 06:49:20.430478 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:20Z","lastTransitionTime":"2025-10-05T06:49:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:20 crc kubenswrapper[4846]: I1005 06:49:20.441039 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c21154c6-7c11-47e7-b1f0-6a5fe33d16b1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e13964315903099b3d7e3dc8c143fe58bb9f9654214702ba51581ac4044414d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef573b754a80e5148b8caa71b4f22c461f57dfa415394a5bccda9771b7684e4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef573b754a80e5148b8caa71b4f22c461f57dfa415394a5bccda9771b7684e4c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:47:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:47:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:47:58Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:49:20Z is after 2025-08-24T17:21:41Z" Oct 05 06:49:20 crc kubenswrapper[4846]: I1005 06:49:20.466449 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"775b5702-a131-4d6c-852d-91bfa056cac5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0839c1765df2c68d2554e9c66d289505e559a9672caf8069feeff2c4fdfd2171\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15dbaf2445704da0987f256f81dff2ee95664988541fa48ce94898afeadab2e8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd3b3ab323c6d5153054aa04af2d91f7b9e8436ff1e2f75f3594b57831b2a416\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://67c46f5f0a3ace95bd2e9479c5561126664e18b148c5eceea5bf5efc27e81a67\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c52e076eddfd307110b1f6de00a169b81af8f2785af7becef2a820ce363a5861\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-05T06:48:14Z\\\",\\\"message\\\":\\\"le observer\\\\nW1005 06:48:14.337870 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1005 06:48:14.338086 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1005 06:48:14.339389 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1467752215/tls.crt::/tmp/serving-cert-1467752215/tls.key\\\\\\\"\\\\nI1005 06:48:14.864345 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1005 06:48:14.868087 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1005 06:48:14.868114 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1005 06:48:14.868140 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1005 06:48:14.868147 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1005 06:48:14.877354 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1005 06:48:14.877412 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1005 06:48:14.877456 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1005 06:48:14.877471 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1005 06:48:14.877490 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1005 06:48:14.877501 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1005 06:48:14.877507 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1005 06:48:14.877515 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1005 06:48:14.880326 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:13Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4a091716ff935adb42ae2e88048619fd347bd174f64e5f9c97a01f81690e57f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:01Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dc5411b642f99b13d514d2f21defad0c394aeeb82a966f72300076617e2586c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5dc5411b642f99b13d514d2f21defad0c394aeeb82a966f72300076617e2586c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:47:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:47:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:47:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:49:20Z is after 2025-08-24T17:21:41Z" Oct 05 06:49:20 crc kubenswrapper[4846]: I1005 06:49:20.488265 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:49:20Z is after 2025-08-24T17:21:41Z" Oct 05 06:49:20 crc kubenswrapper[4846]: I1005 06:49:20.497350 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 05 06:49:20 crc kubenswrapper[4846]: I1005 06:49:20.497364 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9r2kt" Oct 05 06:49:20 crc kubenswrapper[4846]: E1005 06:49:20.498003 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 05 06:49:20 crc kubenswrapper[4846]: E1005 06:49:20.498097 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9r2kt" podUID="59b83f4f-71f0-4b28-9275-56d5f933ec2c" Oct 05 06:49:20 crc kubenswrapper[4846]: I1005 06:49:20.506221 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-r7ztb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a5119cd9-af70-4324-a955-978306ab0b20\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06231a87429ec292cb57ebd88cda7ad547bcb32e80343b776833621a11c28ad1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jfcj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-r7ztb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:49:20Z is after 2025-08-24T17:21:41Z" Oct 05 06:49:20 crc kubenswrapper[4846]: I1005 06:49:20.524771 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-9r2kt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"59b83f4f-71f0-4b28-9275-56d5f933ec2c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tdwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tdwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:38Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-9r2kt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:49:20Z is after 2025-08-24T17:21:41Z" Oct 05 06:49:20 crc kubenswrapper[4846]: I1005 06:49:20.534882 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:20 crc kubenswrapper[4846]: I1005 06:49:20.534954 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:20 crc kubenswrapper[4846]: I1005 06:49:20.534973 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:20 crc kubenswrapper[4846]: I1005 06:49:20.534996 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:20 crc kubenswrapper[4846]: I1005 06:49:20.535018 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:20Z","lastTransitionTime":"2025-10-05T06:49:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:20 crc kubenswrapper[4846]: I1005 06:49:20.638965 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:20 crc kubenswrapper[4846]: I1005 06:49:20.639408 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:20 crc kubenswrapper[4846]: I1005 06:49:20.639551 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:20 crc kubenswrapper[4846]: I1005 06:49:20.639680 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:20 crc kubenswrapper[4846]: I1005 06:49:20.639824 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:20Z","lastTransitionTime":"2025-10-05T06:49:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:20 crc kubenswrapper[4846]: I1005 06:49:20.743482 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:20 crc kubenswrapper[4846]: I1005 06:49:20.743566 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:20 crc kubenswrapper[4846]: I1005 06:49:20.743588 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:20 crc kubenswrapper[4846]: I1005 06:49:20.743617 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:20 crc kubenswrapper[4846]: I1005 06:49:20.743636 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:20Z","lastTransitionTime":"2025-10-05T06:49:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:20 crc kubenswrapper[4846]: I1005 06:49:20.847476 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:20 crc kubenswrapper[4846]: I1005 06:49:20.847818 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:20 crc kubenswrapper[4846]: I1005 06:49:20.848033 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:20 crc kubenswrapper[4846]: I1005 06:49:20.848301 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:20 crc kubenswrapper[4846]: I1005 06:49:20.848530 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:20Z","lastTransitionTime":"2025-10-05T06:49:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:20 crc kubenswrapper[4846]: I1005 06:49:20.951851 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:20 crc kubenswrapper[4846]: I1005 06:49:20.951941 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:20 crc kubenswrapper[4846]: I1005 06:49:20.951966 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:20 crc kubenswrapper[4846]: I1005 06:49:20.952009 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:20 crc kubenswrapper[4846]: I1005 06:49:20.952033 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:20Z","lastTransitionTime":"2025-10-05T06:49:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:21 crc kubenswrapper[4846]: I1005 06:49:21.054701 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:21 crc kubenswrapper[4846]: I1005 06:49:21.054791 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:21 crc kubenswrapper[4846]: I1005 06:49:21.054815 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:21 crc kubenswrapper[4846]: I1005 06:49:21.054845 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:21 crc kubenswrapper[4846]: I1005 06:49:21.054868 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:21Z","lastTransitionTime":"2025-10-05T06:49:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:21 crc kubenswrapper[4846]: I1005 06:49:21.157428 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:21 crc kubenswrapper[4846]: I1005 06:49:21.157486 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:21 crc kubenswrapper[4846]: I1005 06:49:21.157497 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:21 crc kubenswrapper[4846]: I1005 06:49:21.157517 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:21 crc kubenswrapper[4846]: I1005 06:49:21.157532 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:21Z","lastTransitionTime":"2025-10-05T06:49:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:21 crc kubenswrapper[4846]: I1005 06:49:21.260881 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:21 crc kubenswrapper[4846]: I1005 06:49:21.260955 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:21 crc kubenswrapper[4846]: I1005 06:49:21.260976 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:21 crc kubenswrapper[4846]: I1005 06:49:21.261005 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:21 crc kubenswrapper[4846]: I1005 06:49:21.261028 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:21Z","lastTransitionTime":"2025-10-05T06:49:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:21 crc kubenswrapper[4846]: I1005 06:49:21.364533 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:21 crc kubenswrapper[4846]: I1005 06:49:21.364606 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:21 crc kubenswrapper[4846]: I1005 06:49:21.364624 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:21 crc kubenswrapper[4846]: I1005 06:49:21.364654 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:21 crc kubenswrapper[4846]: I1005 06:49:21.364672 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:21Z","lastTransitionTime":"2025-10-05T06:49:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:21 crc kubenswrapper[4846]: I1005 06:49:21.467742 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:21 crc kubenswrapper[4846]: I1005 06:49:21.467956 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:21 crc kubenswrapper[4846]: I1005 06:49:21.467982 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:21 crc kubenswrapper[4846]: I1005 06:49:21.468050 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:21 crc kubenswrapper[4846]: I1005 06:49:21.468072 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:21Z","lastTransitionTime":"2025-10-05T06:49:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:21 crc kubenswrapper[4846]: I1005 06:49:21.496706 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 05 06:49:21 crc kubenswrapper[4846]: I1005 06:49:21.496748 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 05 06:49:21 crc kubenswrapper[4846]: E1005 06:49:21.496923 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 05 06:49:21 crc kubenswrapper[4846]: E1005 06:49:21.497147 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 05 06:49:21 crc kubenswrapper[4846]: I1005 06:49:21.572158 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:21 crc kubenswrapper[4846]: I1005 06:49:21.572302 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:21 crc kubenswrapper[4846]: I1005 06:49:21.572344 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:21 crc kubenswrapper[4846]: I1005 06:49:21.572380 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:21 crc kubenswrapper[4846]: I1005 06:49:21.572404 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:21Z","lastTransitionTime":"2025-10-05T06:49:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:21 crc kubenswrapper[4846]: I1005 06:49:21.592544 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 05 06:49:21 crc kubenswrapper[4846]: I1005 06:49:21.592769 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 05 06:49:21 crc kubenswrapper[4846]: E1005 06:49:21.592815 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-05 06:50:25.592767057 +0000 UTC m=+147.833619982 (durationBeforeRetry 1m4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 06:49:21 crc kubenswrapper[4846]: I1005 06:49:21.592897 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 05 06:49:21 crc kubenswrapper[4846]: E1005 06:49:21.592917 4846 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 05 06:49:21 crc kubenswrapper[4846]: E1005 06:49:21.593012 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-05 06:50:25.592978893 +0000 UTC m=+147.833831698 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 05 06:49:21 crc kubenswrapper[4846]: I1005 06:49:21.593077 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 05 06:49:21 crc kubenswrapper[4846]: I1005 06:49:21.593129 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 05 06:49:21 crc kubenswrapper[4846]: E1005 06:49:21.593335 4846 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 05 06:49:21 crc kubenswrapper[4846]: E1005 06:49:21.593375 4846 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 05 06:49:21 crc kubenswrapper[4846]: E1005 06:49:21.593381 4846 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 05 06:49:21 crc kubenswrapper[4846]: E1005 06:49:21.593402 4846 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 05 06:49:21 crc kubenswrapper[4846]: E1005 06:49:21.593417 4846 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 05 06:49:21 crc kubenswrapper[4846]: E1005 06:49:21.593417 4846 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 05 06:49:21 crc kubenswrapper[4846]: E1005 06:49:21.593506 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-05 06:50:25.593482857 +0000 UTC m=+147.834335802 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 05 06:49:21 crc kubenswrapper[4846]: E1005 06:49:21.593547 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-05 06:50:25.593529729 +0000 UTC m=+147.834382544 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 05 06:49:21 crc kubenswrapper[4846]: E1005 06:49:21.593441 4846 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 05 06:49:21 crc kubenswrapper[4846]: E1005 06:49:21.593625 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-05 06:50:25.593606921 +0000 UTC m=+147.834459736 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 05 06:49:21 crc kubenswrapper[4846]: I1005 06:49:21.675611 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:21 crc kubenswrapper[4846]: I1005 06:49:21.675687 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:21 crc kubenswrapper[4846]: I1005 06:49:21.675707 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:21 crc kubenswrapper[4846]: I1005 06:49:21.675736 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:21 crc kubenswrapper[4846]: I1005 06:49:21.675766 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:21Z","lastTransitionTime":"2025-10-05T06:49:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:21 crc kubenswrapper[4846]: I1005 06:49:21.780480 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:21 crc kubenswrapper[4846]: I1005 06:49:21.780574 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:21 crc kubenswrapper[4846]: I1005 06:49:21.780604 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:21 crc kubenswrapper[4846]: I1005 06:49:21.780638 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:21 crc kubenswrapper[4846]: I1005 06:49:21.780663 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:21Z","lastTransitionTime":"2025-10-05T06:49:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:21 crc kubenswrapper[4846]: I1005 06:49:21.885312 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:21 crc kubenswrapper[4846]: I1005 06:49:21.885380 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:21 crc kubenswrapper[4846]: I1005 06:49:21.885398 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:21 crc kubenswrapper[4846]: I1005 06:49:21.885425 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:21 crc kubenswrapper[4846]: I1005 06:49:21.885444 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:21Z","lastTransitionTime":"2025-10-05T06:49:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:21 crc kubenswrapper[4846]: I1005 06:49:21.989788 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:21 crc kubenswrapper[4846]: I1005 06:49:21.989865 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:21 crc kubenswrapper[4846]: I1005 06:49:21.989887 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:21 crc kubenswrapper[4846]: I1005 06:49:21.989916 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:21 crc kubenswrapper[4846]: I1005 06:49:21.989937 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:21Z","lastTransitionTime":"2025-10-05T06:49:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:22 crc kubenswrapper[4846]: I1005 06:49:22.093278 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:22 crc kubenswrapper[4846]: I1005 06:49:22.093339 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:22 crc kubenswrapper[4846]: I1005 06:49:22.093357 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:22 crc kubenswrapper[4846]: I1005 06:49:22.093386 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:22 crc kubenswrapper[4846]: I1005 06:49:22.093405 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:22Z","lastTransitionTime":"2025-10-05T06:49:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:22 crc kubenswrapper[4846]: I1005 06:49:22.197454 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:22 crc kubenswrapper[4846]: I1005 06:49:22.197555 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:22 crc kubenswrapper[4846]: I1005 06:49:22.197575 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:22 crc kubenswrapper[4846]: I1005 06:49:22.197653 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:22 crc kubenswrapper[4846]: I1005 06:49:22.197683 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:22Z","lastTransitionTime":"2025-10-05T06:49:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:22 crc kubenswrapper[4846]: I1005 06:49:22.301343 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:22 crc kubenswrapper[4846]: I1005 06:49:22.301425 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:22 crc kubenswrapper[4846]: I1005 06:49:22.301444 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:22 crc kubenswrapper[4846]: I1005 06:49:22.301474 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:22 crc kubenswrapper[4846]: I1005 06:49:22.301491 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:22Z","lastTransitionTime":"2025-10-05T06:49:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:22 crc kubenswrapper[4846]: I1005 06:49:22.404147 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:22 crc kubenswrapper[4846]: I1005 06:49:22.404263 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:22 crc kubenswrapper[4846]: I1005 06:49:22.404284 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:22 crc kubenswrapper[4846]: I1005 06:49:22.404311 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:22 crc kubenswrapper[4846]: I1005 06:49:22.404329 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:22Z","lastTransitionTime":"2025-10-05T06:49:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:22 crc kubenswrapper[4846]: I1005 06:49:22.496712 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9r2kt" Oct 05 06:49:22 crc kubenswrapper[4846]: E1005 06:49:22.497431 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9r2kt" podUID="59b83f4f-71f0-4b28-9275-56d5f933ec2c" Oct 05 06:49:22 crc kubenswrapper[4846]: I1005 06:49:22.497501 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 05 06:49:22 crc kubenswrapper[4846]: E1005 06:49:22.497786 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 05 06:49:22 crc kubenswrapper[4846]: I1005 06:49:22.507772 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:22 crc kubenswrapper[4846]: I1005 06:49:22.507850 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:22 crc kubenswrapper[4846]: I1005 06:49:22.507872 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:22 crc kubenswrapper[4846]: I1005 06:49:22.507899 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:22 crc kubenswrapper[4846]: I1005 06:49:22.507920 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:22Z","lastTransitionTime":"2025-10-05T06:49:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:22 crc kubenswrapper[4846]: I1005 06:49:22.611745 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:22 crc kubenswrapper[4846]: I1005 06:49:22.611861 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:22 crc kubenswrapper[4846]: I1005 06:49:22.611881 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:22 crc kubenswrapper[4846]: I1005 06:49:22.611907 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:22 crc kubenswrapper[4846]: I1005 06:49:22.611925 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:22Z","lastTransitionTime":"2025-10-05T06:49:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:22 crc kubenswrapper[4846]: I1005 06:49:22.715683 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:22 crc kubenswrapper[4846]: I1005 06:49:22.715792 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:22 crc kubenswrapper[4846]: I1005 06:49:22.715816 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:22 crc kubenswrapper[4846]: I1005 06:49:22.715848 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:22 crc kubenswrapper[4846]: I1005 06:49:22.715874 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:22Z","lastTransitionTime":"2025-10-05T06:49:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:22 crc kubenswrapper[4846]: I1005 06:49:22.819080 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:22 crc kubenswrapper[4846]: I1005 06:49:22.819134 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:22 crc kubenswrapper[4846]: I1005 06:49:22.819142 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:22 crc kubenswrapper[4846]: I1005 06:49:22.819159 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:22 crc kubenswrapper[4846]: I1005 06:49:22.819170 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:22Z","lastTransitionTime":"2025-10-05T06:49:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:22 crc kubenswrapper[4846]: I1005 06:49:22.922915 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:22 crc kubenswrapper[4846]: I1005 06:49:22.922981 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:22 crc kubenswrapper[4846]: I1005 06:49:22.922999 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:22 crc kubenswrapper[4846]: I1005 06:49:22.923023 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:22 crc kubenswrapper[4846]: I1005 06:49:22.923050 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:22Z","lastTransitionTime":"2025-10-05T06:49:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:23 crc kubenswrapper[4846]: I1005 06:49:23.033988 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:23 crc kubenswrapper[4846]: I1005 06:49:23.034076 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:23 crc kubenswrapper[4846]: I1005 06:49:23.034101 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:23 crc kubenswrapper[4846]: I1005 06:49:23.034134 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:23 crc kubenswrapper[4846]: I1005 06:49:23.034160 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:23Z","lastTransitionTime":"2025-10-05T06:49:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:23 crc kubenswrapper[4846]: I1005 06:49:23.137517 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:23 crc kubenswrapper[4846]: I1005 06:49:23.137598 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:23 crc kubenswrapper[4846]: I1005 06:49:23.137623 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:23 crc kubenswrapper[4846]: I1005 06:49:23.137653 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:23 crc kubenswrapper[4846]: I1005 06:49:23.137677 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:23Z","lastTransitionTime":"2025-10-05T06:49:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:23 crc kubenswrapper[4846]: I1005 06:49:23.241245 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:23 crc kubenswrapper[4846]: I1005 06:49:23.241305 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:23 crc kubenswrapper[4846]: I1005 06:49:23.241322 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:23 crc kubenswrapper[4846]: I1005 06:49:23.241350 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:23 crc kubenswrapper[4846]: I1005 06:49:23.241369 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:23Z","lastTransitionTime":"2025-10-05T06:49:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:23 crc kubenswrapper[4846]: I1005 06:49:23.344119 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:23 crc kubenswrapper[4846]: I1005 06:49:23.344210 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:23 crc kubenswrapper[4846]: I1005 06:49:23.344228 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:23 crc kubenswrapper[4846]: I1005 06:49:23.344253 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:23 crc kubenswrapper[4846]: I1005 06:49:23.344275 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:23Z","lastTransitionTime":"2025-10-05T06:49:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:23 crc kubenswrapper[4846]: I1005 06:49:23.447495 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:23 crc kubenswrapper[4846]: I1005 06:49:23.447561 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:23 crc kubenswrapper[4846]: I1005 06:49:23.447583 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:23 crc kubenswrapper[4846]: I1005 06:49:23.447612 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:23 crc kubenswrapper[4846]: I1005 06:49:23.447635 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:23Z","lastTransitionTime":"2025-10-05T06:49:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:23 crc kubenswrapper[4846]: I1005 06:49:23.497152 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 05 06:49:23 crc kubenswrapper[4846]: I1005 06:49:23.497162 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 05 06:49:23 crc kubenswrapper[4846]: E1005 06:49:23.497426 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 05 06:49:23 crc kubenswrapper[4846]: E1005 06:49:23.497463 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 05 06:49:23 crc kubenswrapper[4846]: I1005 06:49:23.551233 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:23 crc kubenswrapper[4846]: I1005 06:49:23.551306 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:23 crc kubenswrapper[4846]: I1005 06:49:23.551325 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:23 crc kubenswrapper[4846]: I1005 06:49:23.551352 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:23 crc kubenswrapper[4846]: I1005 06:49:23.551370 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:23Z","lastTransitionTime":"2025-10-05T06:49:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:23 crc kubenswrapper[4846]: I1005 06:49:23.654080 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:23 crc kubenswrapper[4846]: I1005 06:49:23.654140 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:23 crc kubenswrapper[4846]: I1005 06:49:23.654159 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:23 crc kubenswrapper[4846]: I1005 06:49:23.654218 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:23 crc kubenswrapper[4846]: I1005 06:49:23.654242 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:23Z","lastTransitionTime":"2025-10-05T06:49:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:23 crc kubenswrapper[4846]: I1005 06:49:23.757437 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:23 crc kubenswrapper[4846]: I1005 06:49:23.757496 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:23 crc kubenswrapper[4846]: I1005 06:49:23.757512 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:23 crc kubenswrapper[4846]: I1005 06:49:23.757534 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:23 crc kubenswrapper[4846]: I1005 06:49:23.757555 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:23Z","lastTransitionTime":"2025-10-05T06:49:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:23 crc kubenswrapper[4846]: I1005 06:49:23.861309 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:23 crc kubenswrapper[4846]: I1005 06:49:23.861387 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:23 crc kubenswrapper[4846]: I1005 06:49:23.861406 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:23 crc kubenswrapper[4846]: I1005 06:49:23.861435 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:23 crc kubenswrapper[4846]: I1005 06:49:23.861459 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:23Z","lastTransitionTime":"2025-10-05T06:49:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:23 crc kubenswrapper[4846]: I1005 06:49:23.965697 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:23 crc kubenswrapper[4846]: I1005 06:49:23.965764 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:23 crc kubenswrapper[4846]: I1005 06:49:23.965782 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:23 crc kubenswrapper[4846]: I1005 06:49:23.965807 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:23 crc kubenswrapper[4846]: I1005 06:49:23.965828 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:23Z","lastTransitionTime":"2025-10-05T06:49:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:24 crc kubenswrapper[4846]: I1005 06:49:24.069795 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:24 crc kubenswrapper[4846]: I1005 06:49:24.069888 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:24 crc kubenswrapper[4846]: I1005 06:49:24.069914 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:24 crc kubenswrapper[4846]: I1005 06:49:24.069945 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:24 crc kubenswrapper[4846]: I1005 06:49:24.069969 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:24Z","lastTransitionTime":"2025-10-05T06:49:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:24 crc kubenswrapper[4846]: I1005 06:49:24.172736 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:24 crc kubenswrapper[4846]: I1005 06:49:24.172812 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:24 crc kubenswrapper[4846]: I1005 06:49:24.172830 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:24 crc kubenswrapper[4846]: I1005 06:49:24.172859 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:24 crc kubenswrapper[4846]: I1005 06:49:24.172876 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:24Z","lastTransitionTime":"2025-10-05T06:49:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:24 crc kubenswrapper[4846]: I1005 06:49:24.275720 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:24 crc kubenswrapper[4846]: I1005 06:49:24.275801 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:24 crc kubenswrapper[4846]: I1005 06:49:24.275819 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:24 crc kubenswrapper[4846]: I1005 06:49:24.275847 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:24 crc kubenswrapper[4846]: I1005 06:49:24.275866 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:24Z","lastTransitionTime":"2025-10-05T06:49:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:24 crc kubenswrapper[4846]: I1005 06:49:24.379517 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:24 crc kubenswrapper[4846]: I1005 06:49:24.379598 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:24 crc kubenswrapper[4846]: I1005 06:49:24.379616 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:24 crc kubenswrapper[4846]: I1005 06:49:24.379649 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:24 crc kubenswrapper[4846]: I1005 06:49:24.379672 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:24Z","lastTransitionTime":"2025-10-05T06:49:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:24 crc kubenswrapper[4846]: I1005 06:49:24.488071 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:24 crc kubenswrapper[4846]: I1005 06:49:24.488144 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:24 crc kubenswrapper[4846]: I1005 06:49:24.488164 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:24 crc kubenswrapper[4846]: I1005 06:49:24.488226 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:24 crc kubenswrapper[4846]: I1005 06:49:24.488258 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:24Z","lastTransitionTime":"2025-10-05T06:49:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:24 crc kubenswrapper[4846]: I1005 06:49:24.497623 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 05 06:49:24 crc kubenswrapper[4846]: I1005 06:49:24.497708 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9r2kt" Oct 05 06:49:24 crc kubenswrapper[4846]: E1005 06:49:24.497917 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 05 06:49:24 crc kubenswrapper[4846]: E1005 06:49:24.498036 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9r2kt" podUID="59b83f4f-71f0-4b28-9275-56d5f933ec2c" Oct 05 06:49:24 crc kubenswrapper[4846]: I1005 06:49:24.591362 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:24 crc kubenswrapper[4846]: I1005 06:49:24.591434 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:24 crc kubenswrapper[4846]: I1005 06:49:24.591454 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:24 crc kubenswrapper[4846]: I1005 06:49:24.591488 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:24 crc kubenswrapper[4846]: I1005 06:49:24.591515 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:24Z","lastTransitionTime":"2025-10-05T06:49:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:24 crc kubenswrapper[4846]: I1005 06:49:24.694988 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:24 crc kubenswrapper[4846]: I1005 06:49:24.695067 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:24 crc kubenswrapper[4846]: I1005 06:49:24.695085 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:24 crc kubenswrapper[4846]: I1005 06:49:24.695113 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:24 crc kubenswrapper[4846]: I1005 06:49:24.695135 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:24Z","lastTransitionTime":"2025-10-05T06:49:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:24 crc kubenswrapper[4846]: I1005 06:49:24.798530 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:24 crc kubenswrapper[4846]: I1005 06:49:24.798601 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:24 crc kubenswrapper[4846]: I1005 06:49:24.798618 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:24 crc kubenswrapper[4846]: I1005 06:49:24.798643 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:24 crc kubenswrapper[4846]: I1005 06:49:24.798662 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:24Z","lastTransitionTime":"2025-10-05T06:49:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:24 crc kubenswrapper[4846]: I1005 06:49:24.902477 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:24 crc kubenswrapper[4846]: I1005 06:49:24.902582 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:24 crc kubenswrapper[4846]: I1005 06:49:24.902642 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:24 crc kubenswrapper[4846]: I1005 06:49:24.902670 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:24 crc kubenswrapper[4846]: I1005 06:49:24.902725 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:24Z","lastTransitionTime":"2025-10-05T06:49:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:25 crc kubenswrapper[4846]: I1005 06:49:25.005948 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:25 crc kubenswrapper[4846]: I1005 06:49:25.006027 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:25 crc kubenswrapper[4846]: I1005 06:49:25.006045 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:25 crc kubenswrapper[4846]: I1005 06:49:25.006072 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:25 crc kubenswrapper[4846]: I1005 06:49:25.006090 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:25Z","lastTransitionTime":"2025-10-05T06:49:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:25 crc kubenswrapper[4846]: I1005 06:49:25.109236 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:25 crc kubenswrapper[4846]: I1005 06:49:25.109289 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:25 crc kubenswrapper[4846]: I1005 06:49:25.109302 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:25 crc kubenswrapper[4846]: I1005 06:49:25.109322 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:25 crc kubenswrapper[4846]: I1005 06:49:25.109336 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:25Z","lastTransitionTime":"2025-10-05T06:49:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:25 crc kubenswrapper[4846]: I1005 06:49:25.213622 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:25 crc kubenswrapper[4846]: I1005 06:49:25.213694 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:25 crc kubenswrapper[4846]: I1005 06:49:25.213713 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:25 crc kubenswrapper[4846]: I1005 06:49:25.213741 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:25 crc kubenswrapper[4846]: I1005 06:49:25.213761 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:25Z","lastTransitionTime":"2025-10-05T06:49:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:25 crc kubenswrapper[4846]: I1005 06:49:25.317329 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:25 crc kubenswrapper[4846]: I1005 06:49:25.317393 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:25 crc kubenswrapper[4846]: I1005 06:49:25.317407 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:25 crc kubenswrapper[4846]: I1005 06:49:25.317426 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:25 crc kubenswrapper[4846]: I1005 06:49:25.317439 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:25Z","lastTransitionTime":"2025-10-05T06:49:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:25 crc kubenswrapper[4846]: I1005 06:49:25.420681 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:25 crc kubenswrapper[4846]: I1005 06:49:25.420732 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:25 crc kubenswrapper[4846]: I1005 06:49:25.420742 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:25 crc kubenswrapper[4846]: I1005 06:49:25.420762 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:25 crc kubenswrapper[4846]: I1005 06:49:25.420782 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:25Z","lastTransitionTime":"2025-10-05T06:49:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:25 crc kubenswrapper[4846]: I1005 06:49:25.496593 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 05 06:49:25 crc kubenswrapper[4846]: E1005 06:49:25.496851 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 05 06:49:25 crc kubenswrapper[4846]: I1005 06:49:25.496999 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 05 06:49:25 crc kubenswrapper[4846]: E1005 06:49:25.497124 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 05 06:49:25 crc kubenswrapper[4846]: I1005 06:49:25.524172 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:25 crc kubenswrapper[4846]: I1005 06:49:25.524265 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:25 crc kubenswrapper[4846]: I1005 06:49:25.524284 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:25 crc kubenswrapper[4846]: I1005 06:49:25.524314 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:25 crc kubenswrapper[4846]: I1005 06:49:25.524333 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:25Z","lastTransitionTime":"2025-10-05T06:49:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:25 crc kubenswrapper[4846]: I1005 06:49:25.627411 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:25 crc kubenswrapper[4846]: I1005 06:49:25.627478 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:25 crc kubenswrapper[4846]: I1005 06:49:25.627498 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:25 crc kubenswrapper[4846]: I1005 06:49:25.627527 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:25 crc kubenswrapper[4846]: I1005 06:49:25.627547 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:25Z","lastTransitionTime":"2025-10-05T06:49:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:25 crc kubenswrapper[4846]: I1005 06:49:25.731018 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:25 crc kubenswrapper[4846]: I1005 06:49:25.731093 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:25 crc kubenswrapper[4846]: I1005 06:49:25.731119 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:25 crc kubenswrapper[4846]: I1005 06:49:25.731158 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:25 crc kubenswrapper[4846]: I1005 06:49:25.731219 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:25Z","lastTransitionTime":"2025-10-05T06:49:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:25 crc kubenswrapper[4846]: I1005 06:49:25.834390 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:25 crc kubenswrapper[4846]: I1005 06:49:25.834469 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:25 crc kubenswrapper[4846]: I1005 06:49:25.834496 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:25 crc kubenswrapper[4846]: I1005 06:49:25.834529 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:25 crc kubenswrapper[4846]: I1005 06:49:25.834548 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:25Z","lastTransitionTime":"2025-10-05T06:49:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:25 crc kubenswrapper[4846]: I1005 06:49:25.937634 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:25 crc kubenswrapper[4846]: I1005 06:49:25.937702 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:25 crc kubenswrapper[4846]: I1005 06:49:25.937721 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:25 crc kubenswrapper[4846]: I1005 06:49:25.937746 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:25 crc kubenswrapper[4846]: I1005 06:49:25.937764 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:25Z","lastTransitionTime":"2025-10-05T06:49:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:26 crc kubenswrapper[4846]: I1005 06:49:26.041007 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:26 crc kubenswrapper[4846]: I1005 06:49:26.041078 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:26 crc kubenswrapper[4846]: I1005 06:49:26.041093 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:26 crc kubenswrapper[4846]: I1005 06:49:26.041117 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:26 crc kubenswrapper[4846]: I1005 06:49:26.041132 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:26Z","lastTransitionTime":"2025-10-05T06:49:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:26 crc kubenswrapper[4846]: I1005 06:49:26.144468 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:26 crc kubenswrapper[4846]: I1005 06:49:26.144567 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:26 crc kubenswrapper[4846]: I1005 06:49:26.144598 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:26 crc kubenswrapper[4846]: I1005 06:49:26.144636 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:26 crc kubenswrapper[4846]: I1005 06:49:26.144668 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:26Z","lastTransitionTime":"2025-10-05T06:49:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:26 crc kubenswrapper[4846]: I1005 06:49:26.248067 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:26 crc kubenswrapper[4846]: I1005 06:49:26.248130 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:26 crc kubenswrapper[4846]: I1005 06:49:26.248147 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:26 crc kubenswrapper[4846]: I1005 06:49:26.248172 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:26 crc kubenswrapper[4846]: I1005 06:49:26.248224 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:26Z","lastTransitionTime":"2025-10-05T06:49:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:26 crc kubenswrapper[4846]: I1005 06:49:26.250282 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:26 crc kubenswrapper[4846]: I1005 06:49:26.250362 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:26 crc kubenswrapper[4846]: I1005 06:49:26.250385 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:26 crc kubenswrapper[4846]: I1005 06:49:26.250414 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:26 crc kubenswrapper[4846]: I1005 06:49:26.250436 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:26Z","lastTransitionTime":"2025-10-05T06:49:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:26 crc kubenswrapper[4846]: E1005 06:49:26.272273 4846 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:49:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:49:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:49:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:49:26Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:49:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:49:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:49:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:49:26Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"04ce628d-2f61-4931-aad6-615924ea53fb\\\",\\\"systemUUID\\\":\\\"a050a23b-e773-4ba9-989b-360aa53a0605\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:49:26Z is after 2025-08-24T17:21:41Z" Oct 05 06:49:26 crc kubenswrapper[4846]: I1005 06:49:26.278691 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:26 crc kubenswrapper[4846]: I1005 06:49:26.278752 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:26 crc kubenswrapper[4846]: I1005 06:49:26.278773 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:26 crc kubenswrapper[4846]: I1005 06:49:26.278797 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:26 crc kubenswrapper[4846]: I1005 06:49:26.278814 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:26Z","lastTransitionTime":"2025-10-05T06:49:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:26 crc kubenswrapper[4846]: E1005 06:49:26.297040 4846 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:49:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:49:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:49:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:49:26Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:49:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:49:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:49:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:49:26Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"04ce628d-2f61-4931-aad6-615924ea53fb\\\",\\\"systemUUID\\\":\\\"a050a23b-e773-4ba9-989b-360aa53a0605\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:49:26Z is after 2025-08-24T17:21:41Z" Oct 05 06:49:26 crc kubenswrapper[4846]: I1005 06:49:26.301651 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:26 crc kubenswrapper[4846]: I1005 06:49:26.301719 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:26 crc kubenswrapper[4846]: I1005 06:49:26.301740 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:26 crc kubenswrapper[4846]: I1005 06:49:26.301769 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:26 crc kubenswrapper[4846]: I1005 06:49:26.301793 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:26Z","lastTransitionTime":"2025-10-05T06:49:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:26 crc kubenswrapper[4846]: E1005 06:49:26.318331 4846 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:49:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:49:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:49:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:49:26Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:49:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:49:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:49:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:49:26Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"04ce628d-2f61-4931-aad6-615924ea53fb\\\",\\\"systemUUID\\\":\\\"a050a23b-e773-4ba9-989b-360aa53a0605\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:49:26Z is after 2025-08-24T17:21:41Z" Oct 05 06:49:26 crc kubenswrapper[4846]: I1005 06:49:26.323389 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:26 crc kubenswrapper[4846]: I1005 06:49:26.323441 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:26 crc kubenswrapper[4846]: I1005 06:49:26.323460 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:26 crc kubenswrapper[4846]: I1005 06:49:26.323483 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:26 crc kubenswrapper[4846]: I1005 06:49:26.323501 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:26Z","lastTransitionTime":"2025-10-05T06:49:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:26 crc kubenswrapper[4846]: E1005 06:49:26.343127 4846 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:49:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:49:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:49:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:49:26Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:49:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:49:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:49:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:49:26Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"04ce628d-2f61-4931-aad6-615924ea53fb\\\",\\\"systemUUID\\\":\\\"a050a23b-e773-4ba9-989b-360aa53a0605\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:49:26Z is after 2025-08-24T17:21:41Z" Oct 05 06:49:26 crc kubenswrapper[4846]: I1005 06:49:26.348213 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:26 crc kubenswrapper[4846]: I1005 06:49:26.348262 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:26 crc kubenswrapper[4846]: I1005 06:49:26.348271 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:26 crc kubenswrapper[4846]: I1005 06:49:26.348290 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:26 crc kubenswrapper[4846]: I1005 06:49:26.348305 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:26Z","lastTransitionTime":"2025-10-05T06:49:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:26 crc kubenswrapper[4846]: E1005 06:49:26.367342 4846 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:49:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:49:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:49:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:49:26Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:49:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:49:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-05T06:49:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-05T06:49:26Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"04ce628d-2f61-4931-aad6-615924ea53fb\\\",\\\"systemUUID\\\":\\\"a050a23b-e773-4ba9-989b-360aa53a0605\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:49:26Z is after 2025-08-24T17:21:41Z" Oct 05 06:49:26 crc kubenswrapper[4846]: E1005 06:49:26.367590 4846 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 05 06:49:26 crc kubenswrapper[4846]: I1005 06:49:26.369850 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:26 crc kubenswrapper[4846]: I1005 06:49:26.369909 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:26 crc kubenswrapper[4846]: I1005 06:49:26.369928 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:26 crc kubenswrapper[4846]: I1005 06:49:26.369959 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:26 crc kubenswrapper[4846]: I1005 06:49:26.369980 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:26Z","lastTransitionTime":"2025-10-05T06:49:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:26 crc kubenswrapper[4846]: I1005 06:49:26.473459 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:26 crc kubenswrapper[4846]: I1005 06:49:26.473503 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:26 crc kubenswrapper[4846]: I1005 06:49:26.473513 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:26 crc kubenswrapper[4846]: I1005 06:49:26.473531 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:26 crc kubenswrapper[4846]: I1005 06:49:26.473545 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:26Z","lastTransitionTime":"2025-10-05T06:49:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:26 crc kubenswrapper[4846]: I1005 06:49:26.496539 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9r2kt" Oct 05 06:49:26 crc kubenswrapper[4846]: I1005 06:49:26.496581 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 05 06:49:26 crc kubenswrapper[4846]: E1005 06:49:26.496703 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9r2kt" podUID="59b83f4f-71f0-4b28-9275-56d5f933ec2c" Oct 05 06:49:26 crc kubenswrapper[4846]: E1005 06:49:26.496815 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 05 06:49:26 crc kubenswrapper[4846]: I1005 06:49:26.576606 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:26 crc kubenswrapper[4846]: I1005 06:49:26.576653 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:26 crc kubenswrapper[4846]: I1005 06:49:26.576666 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:26 crc kubenswrapper[4846]: I1005 06:49:26.576686 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:26 crc kubenswrapper[4846]: I1005 06:49:26.576702 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:26Z","lastTransitionTime":"2025-10-05T06:49:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:26 crc kubenswrapper[4846]: I1005 06:49:26.680396 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:26 crc kubenswrapper[4846]: I1005 06:49:26.680460 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:26 crc kubenswrapper[4846]: I1005 06:49:26.680477 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:26 crc kubenswrapper[4846]: I1005 06:49:26.680506 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:26 crc kubenswrapper[4846]: I1005 06:49:26.680523 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:26Z","lastTransitionTime":"2025-10-05T06:49:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:26 crc kubenswrapper[4846]: I1005 06:49:26.784027 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:26 crc kubenswrapper[4846]: I1005 06:49:26.784100 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:26 crc kubenswrapper[4846]: I1005 06:49:26.784119 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:26 crc kubenswrapper[4846]: I1005 06:49:26.784147 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:26 crc kubenswrapper[4846]: I1005 06:49:26.784170 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:26Z","lastTransitionTime":"2025-10-05T06:49:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:26 crc kubenswrapper[4846]: I1005 06:49:26.887880 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:26 crc kubenswrapper[4846]: I1005 06:49:26.888429 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:26 crc kubenswrapper[4846]: I1005 06:49:26.888457 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:26 crc kubenswrapper[4846]: I1005 06:49:26.888486 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:26 crc kubenswrapper[4846]: I1005 06:49:26.888505 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:26Z","lastTransitionTime":"2025-10-05T06:49:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:26 crc kubenswrapper[4846]: I1005 06:49:26.998390 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:26 crc kubenswrapper[4846]: I1005 06:49:26.998458 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:26 crc kubenswrapper[4846]: I1005 06:49:26.998477 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:26 crc kubenswrapper[4846]: I1005 06:49:26.998503 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:26 crc kubenswrapper[4846]: I1005 06:49:26.998521 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:26Z","lastTransitionTime":"2025-10-05T06:49:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:27 crc kubenswrapper[4846]: I1005 06:49:27.102092 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:27 crc kubenswrapper[4846]: I1005 06:49:27.102145 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:27 crc kubenswrapper[4846]: I1005 06:49:27.102159 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:27 crc kubenswrapper[4846]: I1005 06:49:27.102205 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:27 crc kubenswrapper[4846]: I1005 06:49:27.102226 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:27Z","lastTransitionTime":"2025-10-05T06:49:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:27 crc kubenswrapper[4846]: I1005 06:49:27.205663 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:27 crc kubenswrapper[4846]: I1005 06:49:27.205748 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:27 crc kubenswrapper[4846]: I1005 06:49:27.205771 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:27 crc kubenswrapper[4846]: I1005 06:49:27.205797 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:27 crc kubenswrapper[4846]: I1005 06:49:27.205815 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:27Z","lastTransitionTime":"2025-10-05T06:49:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:27 crc kubenswrapper[4846]: I1005 06:49:27.309542 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:27 crc kubenswrapper[4846]: I1005 06:49:27.309617 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:27 crc kubenswrapper[4846]: I1005 06:49:27.309640 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:27 crc kubenswrapper[4846]: I1005 06:49:27.309667 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:27 crc kubenswrapper[4846]: I1005 06:49:27.309685 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:27Z","lastTransitionTime":"2025-10-05T06:49:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:27 crc kubenswrapper[4846]: I1005 06:49:27.413424 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:27 crc kubenswrapper[4846]: I1005 06:49:27.413497 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:27 crc kubenswrapper[4846]: I1005 06:49:27.413515 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:27 crc kubenswrapper[4846]: I1005 06:49:27.413542 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:27 crc kubenswrapper[4846]: I1005 06:49:27.413560 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:27Z","lastTransitionTime":"2025-10-05T06:49:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:27 crc kubenswrapper[4846]: I1005 06:49:27.497457 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 05 06:49:27 crc kubenswrapper[4846]: I1005 06:49:27.497578 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 05 06:49:27 crc kubenswrapper[4846]: E1005 06:49:27.497706 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 05 06:49:27 crc kubenswrapper[4846]: E1005 06:49:27.497842 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 05 06:49:27 crc kubenswrapper[4846]: I1005 06:49:27.517270 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:27 crc kubenswrapper[4846]: I1005 06:49:27.517332 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:27 crc kubenswrapper[4846]: I1005 06:49:27.517349 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:27 crc kubenswrapper[4846]: I1005 06:49:27.517381 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:27 crc kubenswrapper[4846]: I1005 06:49:27.517399 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:27Z","lastTransitionTime":"2025-10-05T06:49:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:27 crc kubenswrapper[4846]: I1005 06:49:27.621266 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:27 crc kubenswrapper[4846]: I1005 06:49:27.621340 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:27 crc kubenswrapper[4846]: I1005 06:49:27.621362 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:27 crc kubenswrapper[4846]: I1005 06:49:27.621389 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:27 crc kubenswrapper[4846]: I1005 06:49:27.621412 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:27Z","lastTransitionTime":"2025-10-05T06:49:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:27 crc kubenswrapper[4846]: I1005 06:49:27.725796 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:27 crc kubenswrapper[4846]: I1005 06:49:27.725876 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:27 crc kubenswrapper[4846]: I1005 06:49:27.725896 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:27 crc kubenswrapper[4846]: I1005 06:49:27.725928 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:27 crc kubenswrapper[4846]: I1005 06:49:27.725952 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:27Z","lastTransitionTime":"2025-10-05T06:49:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:27 crc kubenswrapper[4846]: I1005 06:49:27.830341 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:27 crc kubenswrapper[4846]: I1005 06:49:27.830438 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:27 crc kubenswrapper[4846]: I1005 06:49:27.830463 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:27 crc kubenswrapper[4846]: I1005 06:49:27.830493 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:27 crc kubenswrapper[4846]: I1005 06:49:27.830512 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:27Z","lastTransitionTime":"2025-10-05T06:49:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:27 crc kubenswrapper[4846]: I1005 06:49:27.933588 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:27 crc kubenswrapper[4846]: I1005 06:49:27.933722 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:27 crc kubenswrapper[4846]: I1005 06:49:27.933747 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:27 crc kubenswrapper[4846]: I1005 06:49:27.933770 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:27 crc kubenswrapper[4846]: I1005 06:49:27.933792 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:27Z","lastTransitionTime":"2025-10-05T06:49:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:28 crc kubenswrapper[4846]: I1005 06:49:28.038387 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:28 crc kubenswrapper[4846]: I1005 06:49:28.038461 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:28 crc kubenswrapper[4846]: I1005 06:49:28.038479 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:28 crc kubenswrapper[4846]: I1005 06:49:28.038505 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:28 crc kubenswrapper[4846]: I1005 06:49:28.038527 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:28Z","lastTransitionTime":"2025-10-05T06:49:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:28 crc kubenswrapper[4846]: I1005 06:49:28.142989 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:28 crc kubenswrapper[4846]: I1005 06:49:28.143475 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:28 crc kubenswrapper[4846]: I1005 06:49:28.143496 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:28 crc kubenswrapper[4846]: I1005 06:49:28.143527 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:28 crc kubenswrapper[4846]: I1005 06:49:28.143550 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:28Z","lastTransitionTime":"2025-10-05T06:49:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:28 crc kubenswrapper[4846]: I1005 06:49:28.247951 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:28 crc kubenswrapper[4846]: I1005 06:49:28.248035 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:28 crc kubenswrapper[4846]: I1005 06:49:28.248057 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:28 crc kubenswrapper[4846]: I1005 06:49:28.248086 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:28 crc kubenswrapper[4846]: I1005 06:49:28.248104 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:28Z","lastTransitionTime":"2025-10-05T06:49:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:28 crc kubenswrapper[4846]: I1005 06:49:28.352059 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:28 crc kubenswrapper[4846]: I1005 06:49:28.352104 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:28 crc kubenswrapper[4846]: I1005 06:49:28.352116 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:28 crc kubenswrapper[4846]: I1005 06:49:28.352132 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:28 crc kubenswrapper[4846]: I1005 06:49:28.352143 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:28Z","lastTransitionTime":"2025-10-05T06:49:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:28 crc kubenswrapper[4846]: I1005 06:49:28.455090 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:28 crc kubenswrapper[4846]: I1005 06:49:28.455164 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:28 crc kubenswrapper[4846]: I1005 06:49:28.455207 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:28 crc kubenswrapper[4846]: I1005 06:49:28.455234 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:28 crc kubenswrapper[4846]: I1005 06:49:28.455251 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:28Z","lastTransitionTime":"2025-10-05T06:49:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:28 crc kubenswrapper[4846]: I1005 06:49:28.496840 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9r2kt" Oct 05 06:49:28 crc kubenswrapper[4846]: I1005 06:49:28.496976 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 05 06:49:28 crc kubenswrapper[4846]: E1005 06:49:28.497042 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9r2kt" podUID="59b83f4f-71f0-4b28-9275-56d5f933ec2c" Oct 05 06:49:28 crc kubenswrapper[4846]: E1005 06:49:28.497244 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 05 06:49:28 crc kubenswrapper[4846]: I1005 06:49:28.522289 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-sfmxd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1dbb13ac-136c-4a1c-9229-b4e8ac4f9c71\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b0af6a6e735f7b7877c177c1d53d1a8dc12c4713a8059c4766fc858af31c9a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://226d977b596d07aeda0486452e2a8605949ed4d2453b21f56174937c395d6f36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://226d977b596d07aeda0486452e2a8605949ed4d2453b21f56174937c395d6f36\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://99c07a4bdb3059d3a99ffdaf794d668d52cffd9e5e3bb8945204dd6d1f8c8328\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99c07a4bdb3059d3a99ffdaf794d668d52cffd9e5e3bb8945204dd6d1f8c8328\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27095272bf4f31d63a78a436636de0aae12f7103597da944f24ed7a74329ee3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27095272bf4f31d63a78a436636de0aae12f7103597da944f24ed7a74329ee3f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:28Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1211571af66d8a94302b65feccf30768ba82b7cbad918e44dd405787dce484cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1211571af66d8a94302b65feccf30768ba82b7cbad918e44dd405787dce484cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27631f90d528900fa180c15cfce80e96d4cdba07743eb6d847e2e822d9917218\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27631f90d528900fa180c15cfce80e96d4cdba07743eb6d847e2e822d9917218\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f345693e9aeb1f4dc196349d3354f3c833a2b0415a693298b1cfccfa2b0d4459\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f345693e9aeb1f4dc196349d3354f3c833a2b0415a693298b1cfccfa2b0d4459\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dp2wt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-sfmxd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:49:28Z is after 2025-08-24T17:21:41Z" Oct 05 06:49:28 crc kubenswrapper[4846]: I1005 06:49:28.556096 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-txzlk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"93a6ace4-8d64-44d7-9637-457d3af6543c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ad1940d07e37217089642478ece9ccdaca74784738a400929a9f5359d158c01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a6f6b2c32dd3da5f9fd5e6971c1773e81a9a24184b7aa5b543cec879001ae53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e9b0219ecbc7e60905d6de3e67a2c8dc9af1017b4aed9538ee0c58c861f2438\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af5e5a4da9ad7371d5a24633e8268f12e3a7d419a8c93b11ddbcc006d8339c6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8da85222775abc7a772f6a8ff8255f1a0d5b15807ba177a7e3e2b2d01a4f24b9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1409651de8c897804f87be03d016c138f8acf27e662dce265a93e15c1848adcf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://59fa020f7bfea4547e132058cf3804a406ee0c06cdc4e78a08ae43a51f46633f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://59fa020f7bfea4547e132058cf3804a406ee0c06cdc4e78a08ae43a51f46633f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-05T06:49:18Z\\\",\\\"message\\\":\\\" start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:49:18Z is after 2025-08-24T17:21:41Z]\\\\nI1005 06:49:18.451492 6882 services_controller.go:473] Services do not match for network=default, existing lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-marketplace/redhat-marketplace_TCP_cluster\\\\\\\", UUID:\\\\\\\"97b6e7b0-06ca-455e-8259-06895040cb0c\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-marketplace/redhat-marketplace\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T06:49:17Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-txzlk_openshift-ovn-kubernetes(93a6ace4-8d64-44d7-9637-457d3af6543c)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0df0ebcd87dc5af4b17736b43ec7b2395968b19ed16e775366af14e2b2d49c52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff3c6545ff3d2436da27bbd223b950b30b0c69a3e7f3034cb0e61380cb38d338\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ff3c6545ff3d2436da27bbd223b950b30b0c69a3e7f3034cb0e61380cb38d338\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-96h7b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:24Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-txzlk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:49:28Z is after 2025-08-24T17:21:41Z" Oct 05 06:49:28 crc kubenswrapper[4846]: I1005 06:49:28.558158 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:28 crc kubenswrapper[4846]: I1005 06:49:28.558259 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:28 crc kubenswrapper[4846]: I1005 06:49:28.558282 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:28 crc kubenswrapper[4846]: I1005 06:49:28.558314 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:28 crc kubenswrapper[4846]: I1005 06:49:28.558337 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:28Z","lastTransitionTime":"2025-10-05T06:49:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:28 crc kubenswrapper[4846]: I1005 06:49:28.570773 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-gklb6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"698ab882-bb26-4b4c-81e9-8305a44fe068\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a6edca88ea244a58f8cf6cc431110a960c6bfd892812cd0d00fa3f97be3bc67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x6d6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c191538ab797b8056820de7da1f8e93ea18873891a32382b8a59c9af762e103\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x6d6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:37Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-gklb6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:49:28Z is after 2025-08-24T17:21:41Z" Oct 05 06:49:28 crc kubenswrapper[4846]: I1005 06:49:28.602142 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c482bb97-fd50-40cf-96ea-f806c3f01125\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bfe45d30e4fa141541aca1d46bc5cc1da0a7b4a5eb1fc9bb2caa5b7fdcb0b264\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://496a4782faa30a9a20ff63038141600cd635f12902d1cf17989cd067fcdfe929\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://281be077869a6aa1ca815770871a14413d4a97e97aeb550fd5cd215cba88b21f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6771c7215d2f5dd5838d17888375a4768b9d1755fc2e21bda1bbfbb2eb1dbb95\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c17a711c2c7473b076a3944ffe7c700808d05cba582fbd55c5c2ac8ff73838a4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://44a9f740f0a06c8bf928e2a9d1d5979217558c31e73a333631baccebf00fd5d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://44a9f740f0a06c8bf928e2a9d1d5979217558c31e73a333631baccebf00fd5d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:47:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:47:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://48ca0ba6ca6092bfcb5b153220c054f02cc36fb22707c42e021f06d329e18e2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://48ca0ba6ca6092bfcb5b153220c054f02cc36fb22707c42e021f06d329e18e2e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:00Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://4006d0cea128675c8885d879171204abffc5e11ea0664750a12b4bc7344ccb2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4006d0cea128675c8885d879171204abffc5e11ea0664750a12b4bc7344ccb2e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:48:01Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:47:58Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:49:28Z is after 2025-08-24T17:21:41Z" Oct 05 06:49:28 crc kubenswrapper[4846]: I1005 06:49:28.623945 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:49:28Z is after 2025-08-24T17:21:41Z" Oct 05 06:49:28 crc kubenswrapper[4846]: I1005 06:49:28.644772 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e67d0e447a8a6a381259e25d25badc791194e1291f56b1cd77bd63989b20408\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dkhl7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://516f88f4990f5962d8b694a3d889259cc5508a309919870191c2e1331640c973\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dkhl7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-fscvf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:49:28Z is after 2025-08-24T17:21:41Z" Oct 05 06:49:28 crc kubenswrapper[4846]: I1005 06:49:28.660465 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-w2sj8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32931fce-7374-4cab-a2e0-79d96a842274\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0df535d65d39de6c9a3dfb931bafc4469adc22b72f7a7903beda4fef8cd48138\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9rtxf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:27Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-w2sj8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:49:28Z is after 2025-08-24T17:21:41Z" Oct 05 06:49:28 crc kubenswrapper[4846]: I1005 06:49:28.662273 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:28 crc kubenswrapper[4846]: I1005 06:49:28.662470 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:28 crc kubenswrapper[4846]: I1005 06:49:28.662485 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:28 crc kubenswrapper[4846]: I1005 06:49:28.662543 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:28 crc kubenswrapper[4846]: I1005 06:49:28.662561 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:28Z","lastTransitionTime":"2025-10-05T06:49:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:28 crc kubenswrapper[4846]: I1005 06:49:28.677473 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c9f812b-ecc8-4985-a670-53fa80fbb499\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f2b5ee9781c01d6aec3ac778efef1067b6c3bfe4ed37431bdebe188fcde0165f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:47:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3fb4717692aa1fec0012b514ef8d18d1ca3707cfa4bb8248bb93841a0a40f11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:47:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0b6a7afe3094f1c92093c05c28344897a615d590c2fbb507c87e18a44c6c831\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://640ade2c733d7916ec213a201a1dd614d261c69eab654c80100efbc99a768995\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:47:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:49:28Z is after 2025-08-24T17:21:41Z" Oct 05 06:49:28 crc kubenswrapper[4846]: I1005 06:49:28.705255 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:49:28Z is after 2025-08-24T17:21:41Z" Oct 05 06:49:28 crc kubenswrapper[4846]: I1005 06:49:28.722494 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:21Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:21Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://347f900743b7bba3daf92f21c7eba0625e8c774f50057bc9472f0c82119325e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:49:28Z is after 2025-08-24T17:21:41Z" Oct 05 06:49:28 crc kubenswrapper[4846]: I1005 06:49:28.736931 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://56f1fe1f9f06f0249cf937a6963fda82467732dc056ca2b2189be6d8dd9d4a05\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7212ce5f42c1fb55146abb190a557a504b43f16af6df84f90e1073567a329ec5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:49:28Z is after 2025-08-24T17:21:41Z" Oct 05 06:49:28 crc kubenswrapper[4846]: I1005 06:49:28.754765 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-7wm5q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9bc7744f-d700-450f-8b8d-98140877fee9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:49:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:49:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9d8631ba978f8fa7677feeeff42cabc454bfd51e5c5c21c78519fa9dfd85ec3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://90379e74c544fccca44f7acdc228ff378154390e211ac801b79f8d074012241e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-05T06:49:11Z\\\",\\\"message\\\":\\\"2025-10-05T06:48:26+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_0ecc6292-0274-4088-90d5-740e638af9d4\\\\n2025-10-05T06:48:26+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_0ecc6292-0274-4088-90d5-740e638af9d4 to /host/opt/cni/bin/\\\\n2025-10-05T06:48:26Z [verbose] multus-daemon started\\\\n2025-10-05T06:48:26Z [verbose] Readiness Indicator file check\\\\n2025-10-05T06:49:11Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:24Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:49:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h5p4q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:24Z\\\"}}\" for pod \"openshift-multus\"/\"multus-7wm5q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:49:28Z is after 2025-08-24T17:21:41Z" Oct 05 06:49:28 crc kubenswrapper[4846]: I1005 06:49:28.767066 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:28 crc kubenswrapper[4846]: I1005 06:49:28.767137 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:28 crc kubenswrapper[4846]: I1005 06:49:28.767149 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:28 crc kubenswrapper[4846]: I1005 06:49:28.767165 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:28 crc kubenswrapper[4846]: I1005 06:49:28.767210 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:28Z","lastTransitionTime":"2025-10-05T06:49:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:28 crc kubenswrapper[4846]: I1005 06:49:28.771383 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"78b204fd-1034-4ea1-bcae-e5c876594a39\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2b94cc50fe15d58f375f8492067811be16278cdf13ee3d80cfed4682279157b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5e5cf9125d1d413c91666744aeab06980b44420dbf0341b165b05787527944b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://70f0d5b436d18a78a2d8e2767143ba1a3104332d86bd63f91d23cd8e77a952aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c236b4ed7f039a221baac043246353eecbb6812c1e6bf6836a5fa96ec8f5abdb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c236b4ed7f039a221baac043246353eecbb6812c1e6bf6836a5fa96ec8f5abdb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:47:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:47:59Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:47:58Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:49:28Z is after 2025-08-24T17:21:41Z" Oct 05 06:49:28 crc kubenswrapper[4846]: I1005 06:49:28.789554 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7ff10c31459b3b3724ee7b769bd6f72f9ed80cce5419b6adcdf69a3239b70756\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:49:28Z is after 2025-08-24T17:21:41Z" Oct 05 06:49:28 crc kubenswrapper[4846]: I1005 06:49:28.807061 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:17Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:49:28Z is after 2025-08-24T17:21:41Z" Oct 05 06:49:28 crc kubenswrapper[4846]: I1005 06:49:28.820253 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-r7ztb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a5119cd9-af70-4324-a955-978306ab0b20\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06231a87429ec292cb57ebd88cda7ad547bcb32e80343b776833621a11c28ad1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jfcj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-r7ztb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:49:28Z is after 2025-08-24T17:21:41Z" Oct 05 06:49:28 crc kubenswrapper[4846]: I1005 06:49:28.834710 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-9r2kt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"59b83f4f-71f0-4b28-9275-56d5f933ec2c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:38Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tdwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6tdwj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:48:38Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-9r2kt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:49:28Z is after 2025-08-24T17:21:41Z" Oct 05 06:49:28 crc kubenswrapper[4846]: I1005 06:49:28.850058 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c21154c6-7c11-47e7-b1f0-6a5fe33d16b1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e13964315903099b3d7e3dc8c143fe58bb9f9654214702ba51581ac4044414d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef573b754a80e5148b8caa71b4f22c461f57dfa415394a5bccda9771b7684e4c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef573b754a80e5148b8caa71b4f22c461f57dfa415394a5bccda9771b7684e4c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:47:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:47:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:47:58Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:49:28Z is after 2025-08-24T17:21:41Z" Oct 05 06:49:28 crc kubenswrapper[4846]: I1005 06:49:28.870736 4846 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"775b5702-a131-4d6c-852d-91bfa056cac5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T06:47:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0839c1765df2c68d2554e9c66d289505e559a9672caf8069feeff2c4fdfd2171\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://15dbaf2445704da0987f256f81dff2ee95664988541fa48ce94898afeadab2e8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cd3b3ab323c6d5153054aa04af2d91f7b9e8436ff1e2f75f3594b57831b2a416\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://67c46f5f0a3ace95bd2e9479c5561126664e18b148c5eceea5bf5efc27e81a67\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c52e076eddfd307110b1f6de00a169b81af8f2785af7becef2a820ce363a5861\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-05T06:48:14Z\\\",\\\"message\\\":\\\"le observer\\\\nW1005 06:48:14.337870 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1005 06:48:14.338086 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1005 06:48:14.339389 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1467752215/tls.crt::/tmp/serving-cert-1467752215/tls.key\\\\\\\"\\\\nI1005 06:48:14.864345 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1005 06:48:14.868087 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1005 06:48:14.868114 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1005 06:48:14.868140 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1005 06:48:14.868147 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1005 06:48:14.877354 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI1005 06:48:14.877412 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1005 06:48:14.877456 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1005 06:48:14.877471 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1005 06:48:14.877490 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1005 06:48:14.877501 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1005 06:48:14.877507 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1005 06:48:14.877515 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1005 06:48:14.880326 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-05T06:48:13Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f4a091716ff935adb42ae2e88048619fd347bd174f64e5f9c97a01f81690e57f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-05T06:48:01Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5dc5411b642f99b13d514d2f21defad0c394aeeb82a966f72300076617e2586c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5dc5411b642f99b13d514d2f21defad0c394aeeb82a966f72300076617e2586c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-05T06:47:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-05T06:47:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T06:47:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-05T06:49:28Z is after 2025-08-24T17:21:41Z" Oct 05 06:49:28 crc kubenswrapper[4846]: I1005 06:49:28.870975 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:28 crc kubenswrapper[4846]: I1005 06:49:28.871029 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:28 crc kubenswrapper[4846]: I1005 06:49:28.871043 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:28 crc kubenswrapper[4846]: I1005 06:49:28.871063 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:28 crc kubenswrapper[4846]: I1005 06:49:28.871076 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:28Z","lastTransitionTime":"2025-10-05T06:49:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:28 crc kubenswrapper[4846]: I1005 06:49:28.973865 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:28 crc kubenswrapper[4846]: I1005 06:49:28.973924 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:28 crc kubenswrapper[4846]: I1005 06:49:28.973943 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:28 crc kubenswrapper[4846]: I1005 06:49:28.973967 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:28 crc kubenswrapper[4846]: I1005 06:49:28.973985 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:28Z","lastTransitionTime":"2025-10-05T06:49:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:29 crc kubenswrapper[4846]: I1005 06:49:29.076972 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:29 crc kubenswrapper[4846]: I1005 06:49:29.077050 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:29 crc kubenswrapper[4846]: I1005 06:49:29.077073 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:29 crc kubenswrapper[4846]: I1005 06:49:29.077103 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:29 crc kubenswrapper[4846]: I1005 06:49:29.077124 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:29Z","lastTransitionTime":"2025-10-05T06:49:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:29 crc kubenswrapper[4846]: I1005 06:49:29.180410 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:29 crc kubenswrapper[4846]: I1005 06:49:29.180455 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:29 crc kubenswrapper[4846]: I1005 06:49:29.180467 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:29 crc kubenswrapper[4846]: I1005 06:49:29.180485 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:29 crc kubenswrapper[4846]: I1005 06:49:29.180497 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:29Z","lastTransitionTime":"2025-10-05T06:49:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:29 crc kubenswrapper[4846]: I1005 06:49:29.283363 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:29 crc kubenswrapper[4846]: I1005 06:49:29.283412 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:29 crc kubenswrapper[4846]: I1005 06:49:29.283424 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:29 crc kubenswrapper[4846]: I1005 06:49:29.283439 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:29 crc kubenswrapper[4846]: I1005 06:49:29.283449 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:29Z","lastTransitionTime":"2025-10-05T06:49:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:29 crc kubenswrapper[4846]: I1005 06:49:29.387735 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:29 crc kubenswrapper[4846]: I1005 06:49:29.387781 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:29 crc kubenswrapper[4846]: I1005 06:49:29.387792 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:29 crc kubenswrapper[4846]: I1005 06:49:29.387810 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:29 crc kubenswrapper[4846]: I1005 06:49:29.387823 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:29Z","lastTransitionTime":"2025-10-05T06:49:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:29 crc kubenswrapper[4846]: I1005 06:49:29.492032 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:29 crc kubenswrapper[4846]: I1005 06:49:29.492088 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:29 crc kubenswrapper[4846]: I1005 06:49:29.492106 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:29 crc kubenswrapper[4846]: I1005 06:49:29.492137 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:29 crc kubenswrapper[4846]: I1005 06:49:29.492159 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:29Z","lastTransitionTime":"2025-10-05T06:49:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:29 crc kubenswrapper[4846]: I1005 06:49:29.497460 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 05 06:49:29 crc kubenswrapper[4846]: E1005 06:49:29.497634 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 05 06:49:29 crc kubenswrapper[4846]: I1005 06:49:29.497737 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 05 06:49:29 crc kubenswrapper[4846]: E1005 06:49:29.497837 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 05 06:49:29 crc kubenswrapper[4846]: I1005 06:49:29.595530 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:29 crc kubenswrapper[4846]: I1005 06:49:29.595596 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:29 crc kubenswrapper[4846]: I1005 06:49:29.595614 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:29 crc kubenswrapper[4846]: I1005 06:49:29.595639 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:29 crc kubenswrapper[4846]: I1005 06:49:29.595658 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:29Z","lastTransitionTime":"2025-10-05T06:49:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:29 crc kubenswrapper[4846]: I1005 06:49:29.699268 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:29 crc kubenswrapper[4846]: I1005 06:49:29.699345 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:29 crc kubenswrapper[4846]: I1005 06:49:29.699364 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:29 crc kubenswrapper[4846]: I1005 06:49:29.699396 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:29 crc kubenswrapper[4846]: I1005 06:49:29.699415 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:29Z","lastTransitionTime":"2025-10-05T06:49:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:29 crc kubenswrapper[4846]: I1005 06:49:29.803087 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:29 crc kubenswrapper[4846]: I1005 06:49:29.803155 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:29 crc kubenswrapper[4846]: I1005 06:49:29.803173 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:29 crc kubenswrapper[4846]: I1005 06:49:29.803316 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:29 crc kubenswrapper[4846]: I1005 06:49:29.803390 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:29Z","lastTransitionTime":"2025-10-05T06:49:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:29 crc kubenswrapper[4846]: I1005 06:49:29.906341 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:29 crc kubenswrapper[4846]: I1005 06:49:29.906411 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:29 crc kubenswrapper[4846]: I1005 06:49:29.906428 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:29 crc kubenswrapper[4846]: I1005 06:49:29.906455 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:29 crc kubenswrapper[4846]: I1005 06:49:29.906477 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:29Z","lastTransitionTime":"2025-10-05T06:49:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:30 crc kubenswrapper[4846]: I1005 06:49:30.009137 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:30 crc kubenswrapper[4846]: I1005 06:49:30.009218 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:30 crc kubenswrapper[4846]: I1005 06:49:30.009234 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:30 crc kubenswrapper[4846]: I1005 06:49:30.009260 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:30 crc kubenswrapper[4846]: I1005 06:49:30.009277 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:30Z","lastTransitionTime":"2025-10-05T06:49:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:30 crc kubenswrapper[4846]: I1005 06:49:30.112603 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:30 crc kubenswrapper[4846]: I1005 06:49:30.112687 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:30 crc kubenswrapper[4846]: I1005 06:49:30.112707 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:30 crc kubenswrapper[4846]: I1005 06:49:30.112738 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:30 crc kubenswrapper[4846]: I1005 06:49:30.112761 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:30Z","lastTransitionTime":"2025-10-05T06:49:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:30 crc kubenswrapper[4846]: I1005 06:49:30.216225 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:30 crc kubenswrapper[4846]: I1005 06:49:30.216303 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:30 crc kubenswrapper[4846]: I1005 06:49:30.216321 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:30 crc kubenswrapper[4846]: I1005 06:49:30.216363 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:30 crc kubenswrapper[4846]: I1005 06:49:30.216385 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:30Z","lastTransitionTime":"2025-10-05T06:49:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:30 crc kubenswrapper[4846]: I1005 06:49:30.320380 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:30 crc kubenswrapper[4846]: I1005 06:49:30.320450 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:30 crc kubenswrapper[4846]: I1005 06:49:30.320468 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:30 crc kubenswrapper[4846]: I1005 06:49:30.320494 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:30 crc kubenswrapper[4846]: I1005 06:49:30.320520 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:30Z","lastTransitionTime":"2025-10-05T06:49:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:30 crc kubenswrapper[4846]: I1005 06:49:30.424747 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:30 crc kubenswrapper[4846]: I1005 06:49:30.424819 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:30 crc kubenswrapper[4846]: I1005 06:49:30.424832 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:30 crc kubenswrapper[4846]: I1005 06:49:30.424855 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:30 crc kubenswrapper[4846]: I1005 06:49:30.424872 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:30Z","lastTransitionTime":"2025-10-05T06:49:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:30 crc kubenswrapper[4846]: I1005 06:49:30.497386 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 05 06:49:30 crc kubenswrapper[4846]: E1005 06:49:30.497587 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 05 06:49:30 crc kubenswrapper[4846]: I1005 06:49:30.497664 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9r2kt" Oct 05 06:49:30 crc kubenswrapper[4846]: E1005 06:49:30.497912 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9r2kt" podUID="59b83f4f-71f0-4b28-9275-56d5f933ec2c" Oct 05 06:49:30 crc kubenswrapper[4846]: I1005 06:49:30.528246 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:30 crc kubenswrapper[4846]: I1005 06:49:30.528310 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:30 crc kubenswrapper[4846]: I1005 06:49:30.528327 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:30 crc kubenswrapper[4846]: I1005 06:49:30.528354 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:30 crc kubenswrapper[4846]: I1005 06:49:30.528373 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:30Z","lastTransitionTime":"2025-10-05T06:49:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:30 crc kubenswrapper[4846]: I1005 06:49:30.631912 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:30 crc kubenswrapper[4846]: I1005 06:49:30.631958 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:30 crc kubenswrapper[4846]: I1005 06:49:30.631968 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:30 crc kubenswrapper[4846]: I1005 06:49:30.631983 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:30 crc kubenswrapper[4846]: I1005 06:49:30.631993 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:30Z","lastTransitionTime":"2025-10-05T06:49:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:30 crc kubenswrapper[4846]: I1005 06:49:30.736235 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:30 crc kubenswrapper[4846]: I1005 06:49:30.736312 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:30 crc kubenswrapper[4846]: I1005 06:49:30.736331 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:30 crc kubenswrapper[4846]: I1005 06:49:30.736363 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:30 crc kubenswrapper[4846]: I1005 06:49:30.736381 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:30Z","lastTransitionTime":"2025-10-05T06:49:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:30 crc kubenswrapper[4846]: I1005 06:49:30.839419 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:30 crc kubenswrapper[4846]: I1005 06:49:30.839505 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:30 crc kubenswrapper[4846]: I1005 06:49:30.839535 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:30 crc kubenswrapper[4846]: I1005 06:49:30.839563 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:30 crc kubenswrapper[4846]: I1005 06:49:30.839584 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:30Z","lastTransitionTime":"2025-10-05T06:49:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:30 crc kubenswrapper[4846]: I1005 06:49:30.943603 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:30 crc kubenswrapper[4846]: I1005 06:49:30.943672 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:30 crc kubenswrapper[4846]: I1005 06:49:30.943689 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:30 crc kubenswrapper[4846]: I1005 06:49:30.943716 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:30 crc kubenswrapper[4846]: I1005 06:49:30.943734 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:30Z","lastTransitionTime":"2025-10-05T06:49:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:31 crc kubenswrapper[4846]: I1005 06:49:31.047620 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:31 crc kubenswrapper[4846]: I1005 06:49:31.047707 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:31 crc kubenswrapper[4846]: I1005 06:49:31.047729 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:31 crc kubenswrapper[4846]: I1005 06:49:31.047758 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:31 crc kubenswrapper[4846]: I1005 06:49:31.047782 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:31Z","lastTransitionTime":"2025-10-05T06:49:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:31 crc kubenswrapper[4846]: I1005 06:49:31.150638 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:31 crc kubenswrapper[4846]: I1005 06:49:31.150708 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:31 crc kubenswrapper[4846]: I1005 06:49:31.150727 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:31 crc kubenswrapper[4846]: I1005 06:49:31.150757 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:31 crc kubenswrapper[4846]: I1005 06:49:31.150777 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:31Z","lastTransitionTime":"2025-10-05T06:49:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:31 crc kubenswrapper[4846]: I1005 06:49:31.254207 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:31 crc kubenswrapper[4846]: I1005 06:49:31.254286 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:31 crc kubenswrapper[4846]: I1005 06:49:31.254306 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:31 crc kubenswrapper[4846]: I1005 06:49:31.254336 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:31 crc kubenswrapper[4846]: I1005 06:49:31.254356 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:31Z","lastTransitionTime":"2025-10-05T06:49:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:31 crc kubenswrapper[4846]: I1005 06:49:31.358088 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:31 crc kubenswrapper[4846]: I1005 06:49:31.358162 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:31 crc kubenswrapper[4846]: I1005 06:49:31.358215 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:31 crc kubenswrapper[4846]: I1005 06:49:31.358245 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:31 crc kubenswrapper[4846]: I1005 06:49:31.358264 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:31Z","lastTransitionTime":"2025-10-05T06:49:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:31 crc kubenswrapper[4846]: I1005 06:49:31.461861 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:31 crc kubenswrapper[4846]: I1005 06:49:31.461942 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:31 crc kubenswrapper[4846]: I1005 06:49:31.461958 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:31 crc kubenswrapper[4846]: I1005 06:49:31.461986 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:31 crc kubenswrapper[4846]: I1005 06:49:31.462006 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:31Z","lastTransitionTime":"2025-10-05T06:49:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:31 crc kubenswrapper[4846]: I1005 06:49:31.496713 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 05 06:49:31 crc kubenswrapper[4846]: I1005 06:49:31.496888 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 05 06:49:31 crc kubenswrapper[4846]: E1005 06:49:31.497328 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 05 06:49:31 crc kubenswrapper[4846]: E1005 06:49:31.497481 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 05 06:49:31 crc kubenswrapper[4846]: I1005 06:49:31.565813 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:31 crc kubenswrapper[4846]: I1005 06:49:31.565882 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:31 crc kubenswrapper[4846]: I1005 06:49:31.565903 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:31 crc kubenswrapper[4846]: I1005 06:49:31.565936 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:31 crc kubenswrapper[4846]: I1005 06:49:31.565959 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:31Z","lastTransitionTime":"2025-10-05T06:49:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:31 crc kubenswrapper[4846]: I1005 06:49:31.670466 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:31 crc kubenswrapper[4846]: I1005 06:49:31.670578 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:31 crc kubenswrapper[4846]: I1005 06:49:31.670603 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:31 crc kubenswrapper[4846]: I1005 06:49:31.670669 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:31 crc kubenswrapper[4846]: I1005 06:49:31.670692 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:31Z","lastTransitionTime":"2025-10-05T06:49:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:31 crc kubenswrapper[4846]: I1005 06:49:31.774507 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:31 crc kubenswrapper[4846]: I1005 06:49:31.774580 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:31 crc kubenswrapper[4846]: I1005 06:49:31.774601 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:31 crc kubenswrapper[4846]: I1005 06:49:31.774630 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:31 crc kubenswrapper[4846]: I1005 06:49:31.774649 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:31Z","lastTransitionTime":"2025-10-05T06:49:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:31 crc kubenswrapper[4846]: I1005 06:49:31.878275 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:31 crc kubenswrapper[4846]: I1005 06:49:31.878350 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:31 crc kubenswrapper[4846]: I1005 06:49:31.878374 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:31 crc kubenswrapper[4846]: I1005 06:49:31.878406 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:31 crc kubenswrapper[4846]: I1005 06:49:31.878429 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:31Z","lastTransitionTime":"2025-10-05T06:49:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:31 crc kubenswrapper[4846]: I1005 06:49:31.981854 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:31 crc kubenswrapper[4846]: I1005 06:49:31.981909 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:31 crc kubenswrapper[4846]: I1005 06:49:31.981927 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:31 crc kubenswrapper[4846]: I1005 06:49:31.981956 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:31 crc kubenswrapper[4846]: I1005 06:49:31.981973 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:31Z","lastTransitionTime":"2025-10-05T06:49:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:32 crc kubenswrapper[4846]: I1005 06:49:32.085513 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:32 crc kubenswrapper[4846]: I1005 06:49:32.085601 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:32 crc kubenswrapper[4846]: I1005 06:49:32.085621 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:32 crc kubenswrapper[4846]: I1005 06:49:32.085657 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:32 crc kubenswrapper[4846]: I1005 06:49:32.085738 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:32Z","lastTransitionTime":"2025-10-05T06:49:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:32 crc kubenswrapper[4846]: I1005 06:49:32.189866 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:32 crc kubenswrapper[4846]: I1005 06:49:32.190295 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:32 crc kubenswrapper[4846]: I1005 06:49:32.190443 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:32 crc kubenswrapper[4846]: I1005 06:49:32.190584 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:32 crc kubenswrapper[4846]: I1005 06:49:32.190709 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:32Z","lastTransitionTime":"2025-10-05T06:49:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:32 crc kubenswrapper[4846]: I1005 06:49:32.293892 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:32 crc kubenswrapper[4846]: I1005 06:49:32.293982 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:32 crc kubenswrapper[4846]: I1005 06:49:32.294005 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:32 crc kubenswrapper[4846]: I1005 06:49:32.294072 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:32 crc kubenswrapper[4846]: I1005 06:49:32.294118 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:32Z","lastTransitionTime":"2025-10-05T06:49:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:32 crc kubenswrapper[4846]: I1005 06:49:32.397368 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:32 crc kubenswrapper[4846]: I1005 06:49:32.397432 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:32 crc kubenswrapper[4846]: I1005 06:49:32.397452 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:32 crc kubenswrapper[4846]: I1005 06:49:32.397476 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:32 crc kubenswrapper[4846]: I1005 06:49:32.397492 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:32Z","lastTransitionTime":"2025-10-05T06:49:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:32 crc kubenswrapper[4846]: I1005 06:49:32.497327 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 05 06:49:32 crc kubenswrapper[4846]: I1005 06:49:32.497348 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9r2kt" Oct 05 06:49:32 crc kubenswrapper[4846]: E1005 06:49:32.497547 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 05 06:49:32 crc kubenswrapper[4846]: E1005 06:49:32.497691 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9r2kt" podUID="59b83f4f-71f0-4b28-9275-56d5f933ec2c" Oct 05 06:49:32 crc kubenswrapper[4846]: I1005 06:49:32.500528 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:32 crc kubenswrapper[4846]: I1005 06:49:32.500583 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:32 crc kubenswrapper[4846]: I1005 06:49:32.500595 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:32 crc kubenswrapper[4846]: I1005 06:49:32.500618 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:32 crc kubenswrapper[4846]: I1005 06:49:32.500631 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:32Z","lastTransitionTime":"2025-10-05T06:49:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:32 crc kubenswrapper[4846]: I1005 06:49:32.604513 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:32 crc kubenswrapper[4846]: I1005 06:49:32.604576 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:32 crc kubenswrapper[4846]: I1005 06:49:32.604595 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:32 crc kubenswrapper[4846]: I1005 06:49:32.604623 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:32 crc kubenswrapper[4846]: I1005 06:49:32.604647 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:32Z","lastTransitionTime":"2025-10-05T06:49:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:32 crc kubenswrapper[4846]: I1005 06:49:32.712101 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:32 crc kubenswrapper[4846]: I1005 06:49:32.712202 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:32 crc kubenswrapper[4846]: I1005 06:49:32.712224 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:32 crc kubenswrapper[4846]: I1005 06:49:32.712254 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:32 crc kubenswrapper[4846]: I1005 06:49:32.712284 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:32Z","lastTransitionTime":"2025-10-05T06:49:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:32 crc kubenswrapper[4846]: I1005 06:49:32.815435 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:32 crc kubenswrapper[4846]: I1005 06:49:32.815501 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:32 crc kubenswrapper[4846]: I1005 06:49:32.815520 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:32 crc kubenswrapper[4846]: I1005 06:49:32.815548 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:32 crc kubenswrapper[4846]: I1005 06:49:32.815570 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:32Z","lastTransitionTime":"2025-10-05T06:49:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:32 crc kubenswrapper[4846]: I1005 06:49:32.918268 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:32 crc kubenswrapper[4846]: I1005 06:49:32.918332 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:32 crc kubenswrapper[4846]: I1005 06:49:32.918352 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:32 crc kubenswrapper[4846]: I1005 06:49:32.918376 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:32 crc kubenswrapper[4846]: I1005 06:49:32.918394 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:32Z","lastTransitionTime":"2025-10-05T06:49:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:33 crc kubenswrapper[4846]: I1005 06:49:33.021706 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:33 crc kubenswrapper[4846]: I1005 06:49:33.021768 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:33 crc kubenswrapper[4846]: I1005 06:49:33.021785 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:33 crc kubenswrapper[4846]: I1005 06:49:33.021810 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:33 crc kubenswrapper[4846]: I1005 06:49:33.021829 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:33Z","lastTransitionTime":"2025-10-05T06:49:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:33 crc kubenswrapper[4846]: I1005 06:49:33.125163 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:33 crc kubenswrapper[4846]: I1005 06:49:33.125269 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:33 crc kubenswrapper[4846]: I1005 06:49:33.125292 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:33 crc kubenswrapper[4846]: I1005 06:49:33.125328 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:33 crc kubenswrapper[4846]: I1005 06:49:33.125353 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:33Z","lastTransitionTime":"2025-10-05T06:49:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:33 crc kubenswrapper[4846]: I1005 06:49:33.229097 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:33 crc kubenswrapper[4846]: I1005 06:49:33.229159 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:33 crc kubenswrapper[4846]: I1005 06:49:33.229206 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:33 crc kubenswrapper[4846]: I1005 06:49:33.229236 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:33 crc kubenswrapper[4846]: I1005 06:49:33.229256 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:33Z","lastTransitionTime":"2025-10-05T06:49:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:33 crc kubenswrapper[4846]: I1005 06:49:33.332591 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:33 crc kubenswrapper[4846]: I1005 06:49:33.332653 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:33 crc kubenswrapper[4846]: I1005 06:49:33.332672 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:33 crc kubenswrapper[4846]: I1005 06:49:33.332699 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:33 crc kubenswrapper[4846]: I1005 06:49:33.332717 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:33Z","lastTransitionTime":"2025-10-05T06:49:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:33 crc kubenswrapper[4846]: I1005 06:49:33.436102 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:33 crc kubenswrapper[4846]: I1005 06:49:33.436211 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:33 crc kubenswrapper[4846]: I1005 06:49:33.436232 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:33 crc kubenswrapper[4846]: I1005 06:49:33.436265 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:33 crc kubenswrapper[4846]: I1005 06:49:33.436286 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:33Z","lastTransitionTime":"2025-10-05T06:49:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:33 crc kubenswrapper[4846]: I1005 06:49:33.497100 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 05 06:49:33 crc kubenswrapper[4846]: I1005 06:49:33.497147 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 05 06:49:33 crc kubenswrapper[4846]: E1005 06:49:33.497298 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 05 06:49:33 crc kubenswrapper[4846]: E1005 06:49:33.497408 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 05 06:49:33 crc kubenswrapper[4846]: I1005 06:49:33.539250 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:33 crc kubenswrapper[4846]: I1005 06:49:33.539321 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:33 crc kubenswrapper[4846]: I1005 06:49:33.539341 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:33 crc kubenswrapper[4846]: I1005 06:49:33.539373 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:33 crc kubenswrapper[4846]: I1005 06:49:33.539392 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:33Z","lastTransitionTime":"2025-10-05T06:49:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:33 crc kubenswrapper[4846]: I1005 06:49:33.641888 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:33 crc kubenswrapper[4846]: I1005 06:49:33.641999 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:33 crc kubenswrapper[4846]: I1005 06:49:33.642027 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:33 crc kubenswrapper[4846]: I1005 06:49:33.642064 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:33 crc kubenswrapper[4846]: I1005 06:49:33.642091 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:33Z","lastTransitionTime":"2025-10-05T06:49:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:33 crc kubenswrapper[4846]: I1005 06:49:33.745640 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:33 crc kubenswrapper[4846]: I1005 06:49:33.745704 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:33 crc kubenswrapper[4846]: I1005 06:49:33.745716 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:33 crc kubenswrapper[4846]: I1005 06:49:33.745741 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:33 crc kubenswrapper[4846]: I1005 06:49:33.745754 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:33Z","lastTransitionTime":"2025-10-05T06:49:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:33 crc kubenswrapper[4846]: I1005 06:49:33.849199 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:33 crc kubenswrapper[4846]: I1005 06:49:33.849236 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:33 crc kubenswrapper[4846]: I1005 06:49:33.849250 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:33 crc kubenswrapper[4846]: I1005 06:49:33.849265 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:33 crc kubenswrapper[4846]: I1005 06:49:33.849275 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:33Z","lastTransitionTime":"2025-10-05T06:49:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:33 crc kubenswrapper[4846]: I1005 06:49:33.951431 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:33 crc kubenswrapper[4846]: I1005 06:49:33.951478 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:33 crc kubenswrapper[4846]: I1005 06:49:33.951491 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:33 crc kubenswrapper[4846]: I1005 06:49:33.951510 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:33 crc kubenswrapper[4846]: I1005 06:49:33.951524 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:33Z","lastTransitionTime":"2025-10-05T06:49:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:34 crc kubenswrapper[4846]: I1005 06:49:34.053991 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:34 crc kubenswrapper[4846]: I1005 06:49:34.054099 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:34 crc kubenswrapper[4846]: I1005 06:49:34.054118 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:34 crc kubenswrapper[4846]: I1005 06:49:34.054144 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:34 crc kubenswrapper[4846]: I1005 06:49:34.054163 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:34Z","lastTransitionTime":"2025-10-05T06:49:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:34 crc kubenswrapper[4846]: I1005 06:49:34.157876 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:34 crc kubenswrapper[4846]: I1005 06:49:34.157975 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:34 crc kubenswrapper[4846]: I1005 06:49:34.157984 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:34 crc kubenswrapper[4846]: I1005 06:49:34.158003 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:34 crc kubenswrapper[4846]: I1005 06:49:34.158016 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:34Z","lastTransitionTime":"2025-10-05T06:49:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:34 crc kubenswrapper[4846]: I1005 06:49:34.260270 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:34 crc kubenswrapper[4846]: I1005 06:49:34.260352 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:34 crc kubenswrapper[4846]: I1005 06:49:34.260371 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:34 crc kubenswrapper[4846]: I1005 06:49:34.260400 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:34 crc kubenswrapper[4846]: I1005 06:49:34.260419 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:34Z","lastTransitionTime":"2025-10-05T06:49:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:34 crc kubenswrapper[4846]: I1005 06:49:34.363557 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:34 crc kubenswrapper[4846]: I1005 06:49:34.363624 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:34 crc kubenswrapper[4846]: I1005 06:49:34.363644 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:34 crc kubenswrapper[4846]: I1005 06:49:34.363673 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:34 crc kubenswrapper[4846]: I1005 06:49:34.363696 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:34Z","lastTransitionTime":"2025-10-05T06:49:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:34 crc kubenswrapper[4846]: I1005 06:49:34.466213 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:34 crc kubenswrapper[4846]: I1005 06:49:34.466281 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:34 crc kubenswrapper[4846]: I1005 06:49:34.466300 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:34 crc kubenswrapper[4846]: I1005 06:49:34.466352 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:34 crc kubenswrapper[4846]: I1005 06:49:34.466372 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:34Z","lastTransitionTime":"2025-10-05T06:49:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:34 crc kubenswrapper[4846]: I1005 06:49:34.497211 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 05 06:49:34 crc kubenswrapper[4846]: I1005 06:49:34.497381 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9r2kt" Oct 05 06:49:34 crc kubenswrapper[4846]: E1005 06:49:34.497582 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 05 06:49:34 crc kubenswrapper[4846]: E1005 06:49:34.497873 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9r2kt" podUID="59b83f4f-71f0-4b28-9275-56d5f933ec2c" Oct 05 06:49:34 crc kubenswrapper[4846]: I1005 06:49:34.569919 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:34 crc kubenswrapper[4846]: I1005 06:49:34.570025 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:34 crc kubenswrapper[4846]: I1005 06:49:34.570047 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:34 crc kubenswrapper[4846]: I1005 06:49:34.570076 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:34 crc kubenswrapper[4846]: I1005 06:49:34.570095 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:34Z","lastTransitionTime":"2025-10-05T06:49:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:34 crc kubenswrapper[4846]: I1005 06:49:34.673827 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:34 crc kubenswrapper[4846]: I1005 06:49:34.673895 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:34 crc kubenswrapper[4846]: I1005 06:49:34.673916 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:34 crc kubenswrapper[4846]: I1005 06:49:34.673943 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:34 crc kubenswrapper[4846]: I1005 06:49:34.673961 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:34Z","lastTransitionTime":"2025-10-05T06:49:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:34 crc kubenswrapper[4846]: I1005 06:49:34.777073 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:34 crc kubenswrapper[4846]: I1005 06:49:34.777134 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:34 crc kubenswrapper[4846]: I1005 06:49:34.777151 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:34 crc kubenswrapper[4846]: I1005 06:49:34.777207 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:34 crc kubenswrapper[4846]: I1005 06:49:34.777233 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:34Z","lastTransitionTime":"2025-10-05T06:49:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:34 crc kubenswrapper[4846]: I1005 06:49:34.880568 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:34 crc kubenswrapper[4846]: I1005 06:49:34.880624 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:34 crc kubenswrapper[4846]: I1005 06:49:34.880643 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:34 crc kubenswrapper[4846]: I1005 06:49:34.880664 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:34 crc kubenswrapper[4846]: I1005 06:49:34.880684 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:34Z","lastTransitionTime":"2025-10-05T06:49:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:34 crc kubenswrapper[4846]: I1005 06:49:34.985359 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:34 crc kubenswrapper[4846]: I1005 06:49:34.985424 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:34 crc kubenswrapper[4846]: I1005 06:49:34.985443 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:34 crc kubenswrapper[4846]: I1005 06:49:34.985468 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:34 crc kubenswrapper[4846]: I1005 06:49:34.985491 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:34Z","lastTransitionTime":"2025-10-05T06:49:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:35 crc kubenswrapper[4846]: I1005 06:49:35.089397 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:35 crc kubenswrapper[4846]: I1005 06:49:35.089484 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:35 crc kubenswrapper[4846]: I1005 06:49:35.089510 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:35 crc kubenswrapper[4846]: I1005 06:49:35.089549 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:35 crc kubenswrapper[4846]: I1005 06:49:35.089574 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:35Z","lastTransitionTime":"2025-10-05T06:49:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:35 crc kubenswrapper[4846]: I1005 06:49:35.193357 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:35 crc kubenswrapper[4846]: I1005 06:49:35.193426 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:35 crc kubenswrapper[4846]: I1005 06:49:35.193449 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:35 crc kubenswrapper[4846]: I1005 06:49:35.193486 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:35 crc kubenswrapper[4846]: I1005 06:49:35.193512 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:35Z","lastTransitionTime":"2025-10-05T06:49:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:35 crc kubenswrapper[4846]: I1005 06:49:35.304036 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:35 crc kubenswrapper[4846]: I1005 06:49:35.304101 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:35 crc kubenswrapper[4846]: I1005 06:49:35.304120 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:35 crc kubenswrapper[4846]: I1005 06:49:35.304147 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:35 crc kubenswrapper[4846]: I1005 06:49:35.304167 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:35Z","lastTransitionTime":"2025-10-05T06:49:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:35 crc kubenswrapper[4846]: I1005 06:49:35.407037 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:35 crc kubenswrapper[4846]: I1005 06:49:35.407276 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:35 crc kubenswrapper[4846]: I1005 06:49:35.407297 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:35 crc kubenswrapper[4846]: I1005 06:49:35.407324 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:35 crc kubenswrapper[4846]: I1005 06:49:35.407345 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:35Z","lastTransitionTime":"2025-10-05T06:49:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:35 crc kubenswrapper[4846]: I1005 06:49:35.497155 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 05 06:49:35 crc kubenswrapper[4846]: I1005 06:49:35.497320 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 05 06:49:35 crc kubenswrapper[4846]: E1005 06:49:35.497551 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 05 06:49:35 crc kubenswrapper[4846]: E1005 06:49:35.498612 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 05 06:49:35 crc kubenswrapper[4846]: I1005 06:49:35.498710 4846 scope.go:117] "RemoveContainer" containerID="59fa020f7bfea4547e132058cf3804a406ee0c06cdc4e78a08ae43a51f46633f" Oct 05 06:49:35 crc kubenswrapper[4846]: E1005 06:49:35.498967 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-txzlk_openshift-ovn-kubernetes(93a6ace4-8d64-44d7-9637-457d3af6543c)\"" pod="openshift-ovn-kubernetes/ovnkube-node-txzlk" podUID="93a6ace4-8d64-44d7-9637-457d3af6543c" Oct 05 06:49:35 crc kubenswrapper[4846]: I1005 06:49:35.510615 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:35 crc kubenswrapper[4846]: I1005 06:49:35.510668 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:35 crc kubenswrapper[4846]: I1005 06:49:35.510688 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:35 crc kubenswrapper[4846]: I1005 06:49:35.510715 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:35 crc kubenswrapper[4846]: I1005 06:49:35.510734 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:35Z","lastTransitionTime":"2025-10-05T06:49:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:35 crc kubenswrapper[4846]: I1005 06:49:35.614325 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:35 crc kubenswrapper[4846]: I1005 06:49:35.614388 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:35 crc kubenswrapper[4846]: I1005 06:49:35.614408 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:35 crc kubenswrapper[4846]: I1005 06:49:35.614436 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:35 crc kubenswrapper[4846]: I1005 06:49:35.614457 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:35Z","lastTransitionTime":"2025-10-05T06:49:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:35 crc kubenswrapper[4846]: I1005 06:49:35.717636 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:35 crc kubenswrapper[4846]: I1005 06:49:35.717707 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:35 crc kubenswrapper[4846]: I1005 06:49:35.717731 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:35 crc kubenswrapper[4846]: I1005 06:49:35.717836 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:35 crc kubenswrapper[4846]: I1005 06:49:35.717863 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:35Z","lastTransitionTime":"2025-10-05T06:49:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:35 crc kubenswrapper[4846]: I1005 06:49:35.821883 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:35 crc kubenswrapper[4846]: I1005 06:49:35.821947 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:35 crc kubenswrapper[4846]: I1005 06:49:35.821969 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:35 crc kubenswrapper[4846]: I1005 06:49:35.821995 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:35 crc kubenswrapper[4846]: I1005 06:49:35.822015 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:35Z","lastTransitionTime":"2025-10-05T06:49:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:35 crc kubenswrapper[4846]: I1005 06:49:35.925138 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:35 crc kubenswrapper[4846]: I1005 06:49:35.925254 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:35 crc kubenswrapper[4846]: I1005 06:49:35.925275 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:35 crc kubenswrapper[4846]: I1005 06:49:35.925304 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:35 crc kubenswrapper[4846]: I1005 06:49:35.925329 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:35Z","lastTransitionTime":"2025-10-05T06:49:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:36 crc kubenswrapper[4846]: I1005 06:49:36.028786 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:36 crc kubenswrapper[4846]: I1005 06:49:36.029032 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:36 crc kubenswrapper[4846]: I1005 06:49:36.029048 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:36 crc kubenswrapper[4846]: I1005 06:49:36.029076 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:36 crc kubenswrapper[4846]: I1005 06:49:36.029093 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:36Z","lastTransitionTime":"2025-10-05T06:49:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:36 crc kubenswrapper[4846]: I1005 06:49:36.132950 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:36 crc kubenswrapper[4846]: I1005 06:49:36.133014 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:36 crc kubenswrapper[4846]: I1005 06:49:36.133031 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:36 crc kubenswrapper[4846]: I1005 06:49:36.133068 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:36 crc kubenswrapper[4846]: I1005 06:49:36.133088 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:36Z","lastTransitionTime":"2025-10-05T06:49:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:36 crc kubenswrapper[4846]: I1005 06:49:36.237537 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:36 crc kubenswrapper[4846]: I1005 06:49:36.237594 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:36 crc kubenswrapper[4846]: I1005 06:49:36.237611 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:36 crc kubenswrapper[4846]: I1005 06:49:36.237642 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:36 crc kubenswrapper[4846]: I1005 06:49:36.237662 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:36Z","lastTransitionTime":"2025-10-05T06:49:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:36 crc kubenswrapper[4846]: I1005 06:49:36.341904 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:36 crc kubenswrapper[4846]: I1005 06:49:36.341969 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:36 crc kubenswrapper[4846]: I1005 06:49:36.341989 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:36 crc kubenswrapper[4846]: I1005 06:49:36.342016 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:36 crc kubenswrapper[4846]: I1005 06:49:36.342036 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:36Z","lastTransitionTime":"2025-10-05T06:49:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:36 crc kubenswrapper[4846]: I1005 06:49:36.446211 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:36 crc kubenswrapper[4846]: I1005 06:49:36.446291 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:36 crc kubenswrapper[4846]: I1005 06:49:36.446311 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:36 crc kubenswrapper[4846]: I1005 06:49:36.446342 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:36 crc kubenswrapper[4846]: I1005 06:49:36.446363 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:36Z","lastTransitionTime":"2025-10-05T06:49:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:36 crc kubenswrapper[4846]: I1005 06:49:36.497173 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 05 06:49:36 crc kubenswrapper[4846]: I1005 06:49:36.497316 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9r2kt" Oct 05 06:49:36 crc kubenswrapper[4846]: E1005 06:49:36.497395 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 05 06:49:36 crc kubenswrapper[4846]: E1005 06:49:36.497546 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9r2kt" podUID="59b83f4f-71f0-4b28-9275-56d5f933ec2c" Oct 05 06:49:36 crc kubenswrapper[4846]: I1005 06:49:36.550022 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:36 crc kubenswrapper[4846]: I1005 06:49:36.550082 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:36 crc kubenswrapper[4846]: I1005 06:49:36.550099 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:36 crc kubenswrapper[4846]: I1005 06:49:36.550127 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:36 crc kubenswrapper[4846]: I1005 06:49:36.550146 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:36Z","lastTransitionTime":"2025-10-05T06:49:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:36 crc kubenswrapper[4846]: I1005 06:49:36.557624 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 05 06:49:36 crc kubenswrapper[4846]: I1005 06:49:36.557708 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 05 06:49:36 crc kubenswrapper[4846]: I1005 06:49:36.557728 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 05 06:49:36 crc kubenswrapper[4846]: I1005 06:49:36.557760 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 05 06:49:36 crc kubenswrapper[4846]: I1005 06:49:36.557779 4846 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-05T06:49:36Z","lastTransitionTime":"2025-10-05T06:49:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 05 06:49:36 crc kubenswrapper[4846]: I1005 06:49:36.637664 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-k9bkk"] Oct 05 06:49:36 crc kubenswrapper[4846]: I1005 06:49:36.638280 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-k9bkk" Oct 05 06:49:36 crc kubenswrapper[4846]: I1005 06:49:36.643417 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Oct 05 06:49:36 crc kubenswrapper[4846]: I1005 06:49:36.643532 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Oct 05 06:49:36 crc kubenswrapper[4846]: I1005 06:49:36.643436 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Oct 05 06:49:36 crc kubenswrapper[4846]: I1005 06:49:36.643993 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Oct 05 06:49:36 crc kubenswrapper[4846]: I1005 06:49:36.689153 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=79.689118995 podStartE2EDuration="1m19.689118995s" podCreationTimestamp="2025-10-05 06:48:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 06:49:36.687261731 +0000 UTC m=+98.928114556" watchObservedRunningTime="2025-10-05 06:49:36.689118995 +0000 UTC m=+98.929971810" Oct 05 06:49:36 crc kubenswrapper[4846]: I1005 06:49:36.747572 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-sfmxd" podStartSLOduration=72.747537159 podStartE2EDuration="1m12.747537159s" podCreationTimestamp="2025-10-05 06:48:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 06:49:36.737039518 +0000 UTC m=+98.977892323" watchObservedRunningTime="2025-10-05 06:49:36.747537159 +0000 UTC m=+98.988389974" Oct 05 06:49:36 crc kubenswrapper[4846]: I1005 06:49:36.811458 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-gklb6" podStartSLOduration=72.81143573 podStartE2EDuration="1m12.81143573s" podCreationTimestamp="2025-10-05 06:48:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 06:49:36.81072791 +0000 UTC m=+99.051580685" watchObservedRunningTime="2025-10-05 06:49:36.81143573 +0000 UTC m=+99.052288515" Oct 05 06:49:36 crc kubenswrapper[4846]: I1005 06:49:36.812934 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/8558e25b-c6db-448f-8c6f-c877d6661bc1-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-k9bkk\" (UID: \"8558e25b-c6db-448f-8c6f-c877d6661bc1\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-k9bkk" Oct 05 06:49:36 crc kubenswrapper[4846]: I1005 06:49:36.813025 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/8558e25b-c6db-448f-8c6f-c877d6661bc1-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-k9bkk\" (UID: \"8558e25b-c6db-448f-8c6f-c877d6661bc1\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-k9bkk" Oct 05 06:49:36 crc kubenswrapper[4846]: I1005 06:49:36.813087 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/8558e25b-c6db-448f-8c6f-c877d6661bc1-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-k9bkk\" (UID: \"8558e25b-c6db-448f-8c6f-c877d6661bc1\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-k9bkk" Oct 05 06:49:36 crc kubenswrapper[4846]: I1005 06:49:36.813228 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/8558e25b-c6db-448f-8c6f-c877d6661bc1-service-ca\") pod \"cluster-version-operator-5c965bbfc6-k9bkk\" (UID: \"8558e25b-c6db-448f-8c6f-c877d6661bc1\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-k9bkk" Oct 05 06:49:36 crc kubenswrapper[4846]: I1005 06:49:36.813288 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8558e25b-c6db-448f-8c6f-c877d6661bc1-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-k9bkk\" (UID: \"8558e25b-c6db-448f-8c6f-c877d6661bc1\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-k9bkk" Oct 05 06:49:36 crc kubenswrapper[4846]: I1005 06:49:36.859359 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=79.859337333 podStartE2EDuration="1m19.859337333s" podCreationTimestamp="2025-10-05 06:48:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 06:49:36.859049605 +0000 UTC m=+99.099902420" watchObservedRunningTime="2025-10-05 06:49:36.859337333 +0000 UTC m=+99.100190108" Oct 05 06:49:36 crc kubenswrapper[4846]: I1005 06:49:36.891630 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podStartSLOduration=73.891594538 podStartE2EDuration="1m13.891594538s" podCreationTimestamp="2025-10-05 06:48:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 06:49:36.890875777 +0000 UTC m=+99.131728562" watchObservedRunningTime="2025-10-05 06:49:36.891594538 +0000 UTC m=+99.132447353" Oct 05 06:49:36 crc kubenswrapper[4846]: I1005 06:49:36.907276 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-w2sj8" podStartSLOduration=73.907249496 podStartE2EDuration="1m13.907249496s" podCreationTimestamp="2025-10-05 06:48:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 06:49:36.907148813 +0000 UTC m=+99.148001618" watchObservedRunningTime="2025-10-05 06:49:36.907249496 +0000 UTC m=+99.148102281" Oct 05 06:49:36 crc kubenswrapper[4846]: I1005 06:49:36.914260 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/8558e25b-c6db-448f-8c6f-c877d6661bc1-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-k9bkk\" (UID: \"8558e25b-c6db-448f-8c6f-c877d6661bc1\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-k9bkk" Oct 05 06:49:36 crc kubenswrapper[4846]: I1005 06:49:36.914329 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/8558e25b-c6db-448f-8c6f-c877d6661bc1-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-k9bkk\" (UID: \"8558e25b-c6db-448f-8c6f-c877d6661bc1\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-k9bkk" Oct 05 06:49:36 crc kubenswrapper[4846]: I1005 06:49:36.914411 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/8558e25b-c6db-448f-8c6f-c877d6661bc1-service-ca\") pod \"cluster-version-operator-5c965bbfc6-k9bkk\" (UID: \"8558e25b-c6db-448f-8c6f-c877d6661bc1\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-k9bkk" Oct 05 06:49:36 crc kubenswrapper[4846]: I1005 06:49:36.914447 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8558e25b-c6db-448f-8c6f-c877d6661bc1-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-k9bkk\" (UID: \"8558e25b-c6db-448f-8c6f-c877d6661bc1\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-k9bkk" Oct 05 06:49:36 crc kubenswrapper[4846]: I1005 06:49:36.914460 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/8558e25b-c6db-448f-8c6f-c877d6661bc1-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-k9bkk\" (UID: \"8558e25b-c6db-448f-8c6f-c877d6661bc1\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-k9bkk" Oct 05 06:49:36 crc kubenswrapper[4846]: I1005 06:49:36.914561 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/8558e25b-c6db-448f-8c6f-c877d6661bc1-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-k9bkk\" (UID: \"8558e25b-c6db-448f-8c6f-c877d6661bc1\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-k9bkk" Oct 05 06:49:36 crc kubenswrapper[4846]: I1005 06:49:36.914660 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/8558e25b-c6db-448f-8c6f-c877d6661bc1-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-k9bkk\" (UID: \"8558e25b-c6db-448f-8c6f-c877d6661bc1\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-k9bkk" Oct 05 06:49:36 crc kubenswrapper[4846]: I1005 06:49:36.915602 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/8558e25b-c6db-448f-8c6f-c877d6661bc1-service-ca\") pod \"cluster-version-operator-5c965bbfc6-k9bkk\" (UID: \"8558e25b-c6db-448f-8c6f-c877d6661bc1\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-k9bkk" Oct 05 06:49:36 crc kubenswrapper[4846]: I1005 06:49:36.925032 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8558e25b-c6db-448f-8c6f-c877d6661bc1-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-k9bkk\" (UID: \"8558e25b-c6db-448f-8c6f-c877d6661bc1\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-k9bkk" Oct 05 06:49:36 crc kubenswrapper[4846]: I1005 06:49:36.926571 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=42.92655688 podStartE2EDuration="42.92655688s" podCreationTimestamp="2025-10-05 06:48:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 06:49:36.92620594 +0000 UTC m=+99.167058755" watchObservedRunningTime="2025-10-05 06:49:36.92655688 +0000 UTC m=+99.167409655" Oct 05 06:49:36 crc kubenswrapper[4846]: I1005 06:49:36.935485 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/8558e25b-c6db-448f-8c6f-c877d6661bc1-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-k9bkk\" (UID: \"8558e25b-c6db-448f-8c6f-c877d6661bc1\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-k9bkk" Oct 05 06:49:36 crc kubenswrapper[4846]: I1005 06:49:36.967311 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-k9bkk" Oct 05 06:49:37 crc kubenswrapper[4846]: I1005 06:49:37.019273 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-7wm5q" podStartSLOduration=73.019251066 podStartE2EDuration="1m13.019251066s" podCreationTimestamp="2025-10-05 06:48:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 06:49:37.006641605 +0000 UTC m=+99.247494400" watchObservedRunningTime="2025-10-05 06:49:37.019251066 +0000 UTC m=+99.260103841" Oct 05 06:49:37 crc kubenswrapper[4846]: I1005 06:49:37.040475 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=30.040444854 podStartE2EDuration="30.040444854s" podCreationTimestamp="2025-10-05 06:49:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 06:49:37.020736049 +0000 UTC m=+99.261588824" watchObservedRunningTime="2025-10-05 06:49:37.040444854 +0000 UTC m=+99.281297659" Oct 05 06:49:37 crc kubenswrapper[4846]: I1005 06:49:37.041548 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=80.041532925 podStartE2EDuration="1m20.041532925s" podCreationTimestamp="2025-10-05 06:48:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 06:49:37.039283391 +0000 UTC m=+99.280136166" watchObservedRunningTime="2025-10-05 06:49:37.041532925 +0000 UTC m=+99.282385730" Oct 05 06:49:37 crc kubenswrapper[4846]: I1005 06:49:37.070060 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-r7ztb" podStartSLOduration=74.070032422 podStartE2EDuration="1m14.070032422s" podCreationTimestamp="2025-10-05 06:48:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 06:49:37.068810037 +0000 UTC m=+99.309662822" watchObservedRunningTime="2025-10-05 06:49:37.070032422 +0000 UTC m=+99.310885237" Oct 05 06:49:37 crc kubenswrapper[4846]: I1005 06:49:37.174443 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-k9bkk" event={"ID":"8558e25b-c6db-448f-8c6f-c877d6661bc1","Type":"ContainerStarted","Data":"e58381f0b7f00d5c5cfa7c0d31d5820cb1b0a50272c4ea9eedc54bac78fa9529"} Oct 05 06:49:37 crc kubenswrapper[4846]: I1005 06:49:37.175031 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-k9bkk" event={"ID":"8558e25b-c6db-448f-8c6f-c877d6661bc1","Type":"ContainerStarted","Data":"9058c0a7d6bb201f37d5198ac7d07679567f494c007091a207fa3aae4c415b8c"} Oct 05 06:49:37 crc kubenswrapper[4846]: I1005 06:49:37.497032 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 05 06:49:37 crc kubenswrapper[4846]: I1005 06:49:37.497132 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 05 06:49:37 crc kubenswrapper[4846]: E1005 06:49:37.497310 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 05 06:49:37 crc kubenswrapper[4846]: E1005 06:49:37.497460 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 05 06:49:38 crc kubenswrapper[4846]: I1005 06:49:38.497023 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9r2kt" Oct 05 06:49:38 crc kubenswrapper[4846]: I1005 06:49:38.497057 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 05 06:49:38 crc kubenswrapper[4846]: E1005 06:49:38.498936 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9r2kt" podUID="59b83f4f-71f0-4b28-9275-56d5f933ec2c" Oct 05 06:49:38 crc kubenswrapper[4846]: E1005 06:49:38.499361 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 05 06:49:39 crc kubenswrapper[4846]: I1005 06:49:39.496770 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 05 06:49:39 crc kubenswrapper[4846]: I1005 06:49:39.496802 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 05 06:49:39 crc kubenswrapper[4846]: E1005 06:49:39.496979 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 05 06:49:39 crc kubenswrapper[4846]: E1005 06:49:39.497134 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 05 06:49:40 crc kubenswrapper[4846]: I1005 06:49:40.497130 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9r2kt" Oct 05 06:49:40 crc kubenswrapper[4846]: I1005 06:49:40.497130 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 05 06:49:40 crc kubenswrapper[4846]: E1005 06:49:40.497365 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9r2kt" podUID="59b83f4f-71f0-4b28-9275-56d5f933ec2c" Oct 05 06:49:40 crc kubenswrapper[4846]: E1005 06:49:40.497506 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 05 06:49:41 crc kubenswrapper[4846]: I1005 06:49:41.497513 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 05 06:49:41 crc kubenswrapper[4846]: I1005 06:49:41.497596 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 05 06:49:41 crc kubenswrapper[4846]: E1005 06:49:41.497730 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 05 06:49:41 crc kubenswrapper[4846]: E1005 06:49:41.497953 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 05 06:49:42 crc kubenswrapper[4846]: I1005 06:49:42.496954 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9r2kt" Oct 05 06:49:42 crc kubenswrapper[4846]: I1005 06:49:42.496992 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 05 06:49:42 crc kubenswrapper[4846]: E1005 06:49:42.497257 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9r2kt" podUID="59b83f4f-71f0-4b28-9275-56d5f933ec2c" Oct 05 06:49:42 crc kubenswrapper[4846]: E1005 06:49:42.497442 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 05 06:49:42 crc kubenswrapper[4846]: I1005 06:49:42.889219 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/59b83f4f-71f0-4b28-9275-56d5f933ec2c-metrics-certs\") pod \"network-metrics-daemon-9r2kt\" (UID: \"59b83f4f-71f0-4b28-9275-56d5f933ec2c\") " pod="openshift-multus/network-metrics-daemon-9r2kt" Oct 05 06:49:42 crc kubenswrapper[4846]: E1005 06:49:42.889475 4846 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 05 06:49:42 crc kubenswrapper[4846]: E1005 06:49:42.889623 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/59b83f4f-71f0-4b28-9275-56d5f933ec2c-metrics-certs podName:59b83f4f-71f0-4b28-9275-56d5f933ec2c nodeName:}" failed. No retries permitted until 2025-10-05 06:50:46.889586044 +0000 UTC m=+169.130438859 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/59b83f4f-71f0-4b28-9275-56d5f933ec2c-metrics-certs") pod "network-metrics-daemon-9r2kt" (UID: "59b83f4f-71f0-4b28-9275-56d5f933ec2c") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 05 06:49:43 crc kubenswrapper[4846]: I1005 06:49:43.496927 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 05 06:49:43 crc kubenswrapper[4846]: I1005 06:49:43.496939 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 05 06:49:43 crc kubenswrapper[4846]: E1005 06:49:43.497163 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 05 06:49:43 crc kubenswrapper[4846]: E1005 06:49:43.497351 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 05 06:49:44 crc kubenswrapper[4846]: I1005 06:49:44.497645 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 05 06:49:44 crc kubenswrapper[4846]: I1005 06:49:44.497645 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9r2kt" Oct 05 06:49:44 crc kubenswrapper[4846]: E1005 06:49:44.497958 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 05 06:49:44 crc kubenswrapper[4846]: E1005 06:49:44.498236 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9r2kt" podUID="59b83f4f-71f0-4b28-9275-56d5f933ec2c" Oct 05 06:49:45 crc kubenswrapper[4846]: I1005 06:49:45.497565 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 05 06:49:45 crc kubenswrapper[4846]: I1005 06:49:45.497676 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 05 06:49:45 crc kubenswrapper[4846]: E1005 06:49:45.498246 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 05 06:49:45 crc kubenswrapper[4846]: E1005 06:49:45.498472 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 05 06:49:46 crc kubenswrapper[4846]: I1005 06:49:46.496947 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 05 06:49:46 crc kubenswrapper[4846]: I1005 06:49:46.497335 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9r2kt" Oct 05 06:49:46 crc kubenswrapper[4846]: E1005 06:49:46.497376 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 05 06:49:46 crc kubenswrapper[4846]: E1005 06:49:46.497602 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9r2kt" podUID="59b83f4f-71f0-4b28-9275-56d5f933ec2c" Oct 05 06:49:47 crc kubenswrapper[4846]: I1005 06:49:47.496990 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 05 06:49:47 crc kubenswrapper[4846]: I1005 06:49:47.497033 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 05 06:49:47 crc kubenswrapper[4846]: E1005 06:49:47.497265 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 05 06:49:47 crc kubenswrapper[4846]: E1005 06:49:47.497431 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 05 06:49:48 crc kubenswrapper[4846]: I1005 06:49:48.497054 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9r2kt" Oct 05 06:49:48 crc kubenswrapper[4846]: I1005 06:49:48.497154 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 05 06:49:48 crc kubenswrapper[4846]: E1005 06:49:48.499111 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9r2kt" podUID="59b83f4f-71f0-4b28-9275-56d5f933ec2c" Oct 05 06:49:48 crc kubenswrapper[4846]: E1005 06:49:48.499742 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 05 06:49:49 crc kubenswrapper[4846]: I1005 06:49:49.497271 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 05 06:49:49 crc kubenswrapper[4846]: I1005 06:49:49.497303 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 05 06:49:49 crc kubenswrapper[4846]: E1005 06:49:49.498526 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 05 06:49:49 crc kubenswrapper[4846]: I1005 06:49:49.498569 4846 scope.go:117] "RemoveContainer" containerID="59fa020f7bfea4547e132058cf3804a406ee0c06cdc4e78a08ae43a51f46633f" Oct 05 06:49:49 crc kubenswrapper[4846]: E1005 06:49:49.498629 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 05 06:49:49 crc kubenswrapper[4846]: E1005 06:49:49.498816 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-txzlk_openshift-ovn-kubernetes(93a6ace4-8d64-44d7-9637-457d3af6543c)\"" pod="openshift-ovn-kubernetes/ovnkube-node-txzlk" podUID="93a6ace4-8d64-44d7-9637-457d3af6543c" Oct 05 06:49:50 crc kubenswrapper[4846]: I1005 06:49:50.497062 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 05 06:49:50 crc kubenswrapper[4846]: I1005 06:49:50.497230 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9r2kt" Oct 05 06:49:50 crc kubenswrapper[4846]: E1005 06:49:50.497300 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 05 06:49:50 crc kubenswrapper[4846]: E1005 06:49:50.497460 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9r2kt" podUID="59b83f4f-71f0-4b28-9275-56d5f933ec2c" Oct 05 06:49:51 crc kubenswrapper[4846]: I1005 06:49:51.496943 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 05 06:49:51 crc kubenswrapper[4846]: I1005 06:49:51.497001 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 05 06:49:51 crc kubenswrapper[4846]: E1005 06:49:51.497144 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 05 06:49:51 crc kubenswrapper[4846]: E1005 06:49:51.497302 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 05 06:49:52 crc kubenswrapper[4846]: I1005 06:49:52.496847 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 05 06:49:52 crc kubenswrapper[4846]: I1005 06:49:52.496896 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9r2kt" Oct 05 06:49:52 crc kubenswrapper[4846]: E1005 06:49:52.497072 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 05 06:49:52 crc kubenswrapper[4846]: E1005 06:49:52.497278 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9r2kt" podUID="59b83f4f-71f0-4b28-9275-56d5f933ec2c" Oct 05 06:49:53 crc kubenswrapper[4846]: I1005 06:49:53.497454 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 05 06:49:53 crc kubenswrapper[4846]: E1005 06:49:53.497674 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 05 06:49:53 crc kubenswrapper[4846]: I1005 06:49:53.497469 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 05 06:49:53 crc kubenswrapper[4846]: E1005 06:49:53.497836 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 05 06:49:54 crc kubenswrapper[4846]: I1005 06:49:54.497097 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 05 06:49:54 crc kubenswrapper[4846]: I1005 06:49:54.497289 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9r2kt" Oct 05 06:49:54 crc kubenswrapper[4846]: E1005 06:49:54.497983 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9r2kt" podUID="59b83f4f-71f0-4b28-9275-56d5f933ec2c" Oct 05 06:49:54 crc kubenswrapper[4846]: E1005 06:49:54.498706 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 05 06:49:55 crc kubenswrapper[4846]: I1005 06:49:55.497253 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 05 06:49:55 crc kubenswrapper[4846]: I1005 06:49:55.497352 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 05 06:49:55 crc kubenswrapper[4846]: E1005 06:49:55.497468 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 05 06:49:55 crc kubenswrapper[4846]: E1005 06:49:55.498011 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 05 06:49:56 crc kubenswrapper[4846]: I1005 06:49:56.496781 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 05 06:49:56 crc kubenswrapper[4846]: I1005 06:49:56.496793 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9r2kt" Oct 05 06:49:56 crc kubenswrapper[4846]: E1005 06:49:56.497045 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 05 06:49:56 crc kubenswrapper[4846]: E1005 06:49:56.497215 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9r2kt" podUID="59b83f4f-71f0-4b28-9275-56d5f933ec2c" Oct 05 06:49:57 crc kubenswrapper[4846]: I1005 06:49:57.498072 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 05 06:49:57 crc kubenswrapper[4846]: I1005 06:49:57.498288 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 05 06:49:57 crc kubenswrapper[4846]: E1005 06:49:57.498456 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 05 06:49:57 crc kubenswrapper[4846]: E1005 06:49:57.500256 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 05 06:49:58 crc kubenswrapper[4846]: I1005 06:49:58.262236 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-7wm5q_9bc7744f-d700-450f-8b8d-98140877fee9/kube-multus/1.log" Oct 05 06:49:58 crc kubenswrapper[4846]: I1005 06:49:58.263131 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-7wm5q_9bc7744f-d700-450f-8b8d-98140877fee9/kube-multus/0.log" Oct 05 06:49:58 crc kubenswrapper[4846]: I1005 06:49:58.263371 4846 generic.go:334] "Generic (PLEG): container finished" podID="9bc7744f-d700-450f-8b8d-98140877fee9" containerID="9d8631ba978f8fa7677feeeff42cabc454bfd51e5c5c21c78519fa9dfd85ec3c" exitCode=1 Oct 05 06:49:58 crc kubenswrapper[4846]: I1005 06:49:58.263519 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-7wm5q" event={"ID":"9bc7744f-d700-450f-8b8d-98140877fee9","Type":"ContainerDied","Data":"9d8631ba978f8fa7677feeeff42cabc454bfd51e5c5c21c78519fa9dfd85ec3c"} Oct 05 06:49:58 crc kubenswrapper[4846]: I1005 06:49:58.263646 4846 scope.go:117] "RemoveContainer" containerID="90379e74c544fccca44f7acdc228ff378154390e211ac801b79f8d074012241e" Oct 05 06:49:58 crc kubenswrapper[4846]: I1005 06:49:58.264752 4846 scope.go:117] "RemoveContainer" containerID="9d8631ba978f8fa7677feeeff42cabc454bfd51e5c5c21c78519fa9dfd85ec3c" Oct 05 06:49:58 crc kubenswrapper[4846]: E1005 06:49:58.265652 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-multus pod=multus-7wm5q_openshift-multus(9bc7744f-d700-450f-8b8d-98140877fee9)\"" pod="openshift-multus/multus-7wm5q" podUID="9bc7744f-d700-450f-8b8d-98140877fee9" Oct 05 06:49:58 crc kubenswrapper[4846]: I1005 06:49:58.296919 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-k9bkk" podStartSLOduration=95.296883455 podStartE2EDuration="1m35.296883455s" podCreationTimestamp="2025-10-05 06:48:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 06:49:37.196950519 +0000 UTC m=+99.437803324" watchObservedRunningTime="2025-10-05 06:49:58.296883455 +0000 UTC m=+120.537736270" Oct 05 06:49:58 crc kubenswrapper[4846]: E1005 06:49:58.451662 4846 kubelet_node_status.go:497] "Node not becoming ready in time after startup" Oct 05 06:49:58 crc kubenswrapper[4846]: I1005 06:49:58.496862 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 05 06:49:58 crc kubenswrapper[4846]: E1005 06:49:58.497854 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 05 06:49:58 crc kubenswrapper[4846]: I1005 06:49:58.497961 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9r2kt" Oct 05 06:49:58 crc kubenswrapper[4846]: E1005 06:49:58.498132 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9r2kt" podUID="59b83f4f-71f0-4b28-9275-56d5f933ec2c" Oct 05 06:49:58 crc kubenswrapper[4846]: E1005 06:49:58.620968 4846 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Oct 05 06:49:59 crc kubenswrapper[4846]: I1005 06:49:59.271363 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-7wm5q_9bc7744f-d700-450f-8b8d-98140877fee9/kube-multus/1.log" Oct 05 06:49:59 crc kubenswrapper[4846]: I1005 06:49:59.497364 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 05 06:49:59 crc kubenswrapper[4846]: I1005 06:49:59.497363 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 05 06:49:59 crc kubenswrapper[4846]: E1005 06:49:59.497574 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 05 06:49:59 crc kubenswrapper[4846]: E1005 06:49:59.497665 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 05 06:50:00 crc kubenswrapper[4846]: I1005 06:50:00.497543 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 05 06:50:00 crc kubenswrapper[4846]: I1005 06:50:00.497660 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9r2kt" Oct 05 06:50:00 crc kubenswrapper[4846]: E1005 06:50:00.497837 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 05 06:50:00 crc kubenswrapper[4846]: E1005 06:50:00.498144 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9r2kt" podUID="59b83f4f-71f0-4b28-9275-56d5f933ec2c" Oct 05 06:50:01 crc kubenswrapper[4846]: I1005 06:50:01.496769 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 05 06:50:01 crc kubenswrapper[4846]: E1005 06:50:01.497013 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 05 06:50:01 crc kubenswrapper[4846]: I1005 06:50:01.496783 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 05 06:50:01 crc kubenswrapper[4846]: E1005 06:50:01.497885 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 05 06:50:02 crc kubenswrapper[4846]: I1005 06:50:02.497683 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9r2kt" Oct 05 06:50:02 crc kubenswrapper[4846]: I1005 06:50:02.497725 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 05 06:50:02 crc kubenswrapper[4846]: E1005 06:50:02.498340 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 05 06:50:02 crc kubenswrapper[4846]: E1005 06:50:02.499575 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9r2kt" podUID="59b83f4f-71f0-4b28-9275-56d5f933ec2c" Oct 05 06:50:03 crc kubenswrapper[4846]: I1005 06:50:03.496817 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 05 06:50:03 crc kubenswrapper[4846]: I1005 06:50:03.496820 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 05 06:50:03 crc kubenswrapper[4846]: E1005 06:50:03.497037 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 05 06:50:03 crc kubenswrapper[4846]: E1005 06:50:03.497215 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 05 06:50:03 crc kubenswrapper[4846]: I1005 06:50:03.498418 4846 scope.go:117] "RemoveContainer" containerID="59fa020f7bfea4547e132058cf3804a406ee0c06cdc4e78a08ae43a51f46633f" Oct 05 06:50:03 crc kubenswrapper[4846]: E1005 06:50:03.623619 4846 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Oct 05 06:50:04 crc kubenswrapper[4846]: I1005 06:50:04.293375 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-txzlk_93a6ace4-8d64-44d7-9637-457d3af6543c/ovnkube-controller/3.log" Oct 05 06:50:04 crc kubenswrapper[4846]: I1005 06:50:04.296558 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-txzlk" event={"ID":"93a6ace4-8d64-44d7-9637-457d3af6543c","Type":"ContainerStarted","Data":"32a57312adddc39f79d4dbc9dfb0197ec33241adfb654ed64db6bc10d1d479c8"} Oct 05 06:50:04 crc kubenswrapper[4846]: I1005 06:50:04.297330 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-txzlk" Oct 05 06:50:04 crc kubenswrapper[4846]: I1005 06:50:04.374123 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-txzlk" podStartSLOduration=100.374094819 podStartE2EDuration="1m40.374094819s" podCreationTimestamp="2025-10-05 06:48:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 06:50:04.372460802 +0000 UTC m=+126.613313627" watchObservedRunningTime="2025-10-05 06:50:04.374094819 +0000 UTC m=+126.614947604" Oct 05 06:50:04 crc kubenswrapper[4846]: I1005 06:50:04.497410 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9r2kt" Oct 05 06:50:04 crc kubenswrapper[4846]: I1005 06:50:04.497435 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 05 06:50:04 crc kubenswrapper[4846]: E1005 06:50:04.497711 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9r2kt" podUID="59b83f4f-71f0-4b28-9275-56d5f933ec2c" Oct 05 06:50:04 crc kubenswrapper[4846]: E1005 06:50:04.498027 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 05 06:50:04 crc kubenswrapper[4846]: I1005 06:50:04.728711 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-9r2kt"] Oct 05 06:50:05 crc kubenswrapper[4846]: I1005 06:50:05.300792 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9r2kt" Oct 05 06:50:05 crc kubenswrapper[4846]: E1005 06:50:05.301619 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9r2kt" podUID="59b83f4f-71f0-4b28-9275-56d5f933ec2c" Oct 05 06:50:05 crc kubenswrapper[4846]: I1005 06:50:05.497275 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 05 06:50:05 crc kubenswrapper[4846]: I1005 06:50:05.497296 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 05 06:50:05 crc kubenswrapper[4846]: E1005 06:50:05.497561 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 05 06:50:05 crc kubenswrapper[4846]: E1005 06:50:05.497756 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 05 06:50:06 crc kubenswrapper[4846]: I1005 06:50:06.496985 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 05 06:50:06 crc kubenswrapper[4846]: I1005 06:50:06.497050 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9r2kt" Oct 05 06:50:06 crc kubenswrapper[4846]: E1005 06:50:06.497245 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 05 06:50:06 crc kubenswrapper[4846]: E1005 06:50:06.497500 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9r2kt" podUID="59b83f4f-71f0-4b28-9275-56d5f933ec2c" Oct 05 06:50:07 crc kubenswrapper[4846]: I1005 06:50:07.496804 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 05 06:50:07 crc kubenswrapper[4846]: I1005 06:50:07.496897 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 05 06:50:07 crc kubenswrapper[4846]: E1005 06:50:07.497036 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 05 06:50:07 crc kubenswrapper[4846]: E1005 06:50:07.497284 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 05 06:50:08 crc kubenswrapper[4846]: I1005 06:50:08.496732 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 05 06:50:08 crc kubenswrapper[4846]: I1005 06:50:08.496922 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9r2kt" Oct 05 06:50:08 crc kubenswrapper[4846]: E1005 06:50:08.498850 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 05 06:50:08 crc kubenswrapper[4846]: E1005 06:50:08.499071 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9r2kt" podUID="59b83f4f-71f0-4b28-9275-56d5f933ec2c" Oct 05 06:50:08 crc kubenswrapper[4846]: E1005 06:50:08.625444 4846 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Oct 05 06:50:09 crc kubenswrapper[4846]: I1005 06:50:09.497526 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 05 06:50:09 crc kubenswrapper[4846]: I1005 06:50:09.497551 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 05 06:50:09 crc kubenswrapper[4846]: E1005 06:50:09.497742 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 05 06:50:09 crc kubenswrapper[4846]: E1005 06:50:09.497943 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 05 06:50:10 crc kubenswrapper[4846]: I1005 06:50:10.497573 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 05 06:50:10 crc kubenswrapper[4846]: I1005 06:50:10.497706 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9r2kt" Oct 05 06:50:10 crc kubenswrapper[4846]: E1005 06:50:10.497792 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 05 06:50:10 crc kubenswrapper[4846]: E1005 06:50:10.497955 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9r2kt" podUID="59b83f4f-71f0-4b28-9275-56d5f933ec2c" Oct 05 06:50:11 crc kubenswrapper[4846]: I1005 06:50:11.497609 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 05 06:50:11 crc kubenswrapper[4846]: I1005 06:50:11.497930 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 05 06:50:11 crc kubenswrapper[4846]: E1005 06:50:11.498071 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 05 06:50:11 crc kubenswrapper[4846]: E1005 06:50:11.498242 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 05 06:50:11 crc kubenswrapper[4846]: I1005 06:50:11.498318 4846 scope.go:117] "RemoveContainer" containerID="9d8631ba978f8fa7677feeeff42cabc454bfd51e5c5c21c78519fa9dfd85ec3c" Oct 05 06:50:12 crc kubenswrapper[4846]: I1005 06:50:12.332909 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-7wm5q_9bc7744f-d700-450f-8b8d-98140877fee9/kube-multus/1.log" Oct 05 06:50:12 crc kubenswrapper[4846]: I1005 06:50:12.333016 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-7wm5q" event={"ID":"9bc7744f-d700-450f-8b8d-98140877fee9","Type":"ContainerStarted","Data":"75ffecdf3834699d0ea47cca69028168c32263e5b5ddc5043cc0c1dc1d7bf74d"} Oct 05 06:50:12 crc kubenswrapper[4846]: I1005 06:50:12.497142 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 05 06:50:12 crc kubenswrapper[4846]: E1005 06:50:12.497400 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 05 06:50:12 crc kubenswrapper[4846]: I1005 06:50:12.497570 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9r2kt" Oct 05 06:50:12 crc kubenswrapper[4846]: E1005 06:50:12.497829 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-9r2kt" podUID="59b83f4f-71f0-4b28-9275-56d5f933ec2c" Oct 05 06:50:13 crc kubenswrapper[4846]: I1005 06:50:13.497400 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 05 06:50:13 crc kubenswrapper[4846]: I1005 06:50:13.497435 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 05 06:50:13 crc kubenswrapper[4846]: E1005 06:50:13.498648 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 05 06:50:13 crc kubenswrapper[4846]: E1005 06:50:13.498765 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 05 06:50:14 crc kubenswrapper[4846]: I1005 06:50:14.497110 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9r2kt" Oct 05 06:50:14 crc kubenswrapper[4846]: I1005 06:50:14.497152 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 05 06:50:14 crc kubenswrapper[4846]: I1005 06:50:14.501689 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Oct 05 06:50:14 crc kubenswrapper[4846]: I1005 06:50:14.501893 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Oct 05 06:50:14 crc kubenswrapper[4846]: I1005 06:50:14.502009 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Oct 05 06:50:14 crc kubenswrapper[4846]: I1005 06:50:14.504898 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Oct 05 06:50:15 crc kubenswrapper[4846]: I1005 06:50:15.496810 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 05 06:50:15 crc kubenswrapper[4846]: I1005 06:50:15.496963 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 05 06:50:15 crc kubenswrapper[4846]: I1005 06:50:15.500635 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Oct 05 06:50:15 crc kubenswrapper[4846]: I1005 06:50:15.500862 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.490677 4846 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.538650 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-6bpvf"] Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.539322 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-6bpvf" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.540775 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-vwx77"] Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.541917 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-vwx77" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.542439 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-m467d"] Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.543211 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-m467d" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.544070 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-stjw8"] Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.545103 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-stjw8" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.546413 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.547506 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-vwktr"] Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.548158 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-pglrh"] Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.548845 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-pglrh" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.549989 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.550315 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.550619 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.550888 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.551386 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-vwktr" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.554767 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.556734 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.556788 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.557555 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.557602 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.561255 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-rcnlv"] Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.562288 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-rcnlv" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.564014 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.564793 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-9hcbc"] Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.565580 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-9hcbc" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.566100 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-6t2mt"] Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.566663 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-6t2mt" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.568299 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.568733 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.571524 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.571681 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.571804 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.572042 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.572159 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.573052 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-wcg6b"] Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.574359 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-wcg6b" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.576653 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.578290 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-c6cxs"] Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.578986 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-c6cxs" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.593774 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.594536 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.594766 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.594974 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.595094 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.595159 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.595381 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.595544 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.595591 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.595675 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.595731 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.595829 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.595924 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.595984 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.596052 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.595546 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.596128 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.596204 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.594546 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.596425 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.595597 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.596513 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.596701 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.596959 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.597087 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.597281 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.610587 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.610964 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.611455 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.611793 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.612372 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.613638 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.615782 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.637812 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.638512 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.638830 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.638924 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.639065 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.639446 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.639585 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.639932 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.641199 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.641307 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.641521 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-k87kc"] Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.641656 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.642229 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-t8f4h"] Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.642779 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-t8f4h" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.642786 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-k87kc" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.646881 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.647327 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.647665 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.647830 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.647867 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.652823 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.656228 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-2pd47"] Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.656542 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.657119 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-2pd47" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.658097 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.658472 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.658856 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.658928 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.659169 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.659262 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.659191 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.660845 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.661494 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.661784 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.671313 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.671586 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.672251 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.675026 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-vjmbx"] Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.675859 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-vjmbx" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.676586 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-9jtcc"] Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.677246 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.677749 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-9jtcc" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.678298 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.689448 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-vwx77"] Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.689842 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.690423 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.691438 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-526w4"] Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.691886 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.692305 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.692766 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.693088 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.694351 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-526w4" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.702210 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.711424 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.711700 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.712311 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-5qwb9"] Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.713065 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-rf9l4"] Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.713119 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.713429 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-rf9l4" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.713844 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-5qwb9" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.714487 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.717390 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.723908 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.723942 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.724602 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.724805 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.726357 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.726489 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.728564 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-8vjkd"] Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.729500 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-7jflr"] Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.729719 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-8vjkd" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.730266 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-7jflr" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.732309 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-skh8h"] Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.732889 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-skh8h" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.734933 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.735283 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.737318 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.739706 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/e4ea5d4f-3481-4e72-9bd8-e5d8183bc687-encryption-config\") pod \"apiserver-7bbb656c7d-stjw8\" (UID: \"e4ea5d4f-3481-4e72-9bd8-e5d8183bc687\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-stjw8" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.739748 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/53fe640c-6cb7-43a3-9a84-467eb36fe66b-available-featuregates\") pod \"openshift-config-operator-7777fb866f-rcnlv\" (UID: \"53fe640c-6cb7-43a3-9a84-467eb36fe66b\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-rcnlv" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.739777 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/74f16025-fe11-4a23-8c1a-ff3da2a3d751-service-ca\") pod \"console-f9d7485db-9hcbc\" (UID: \"74f16025-fe11-4a23-8c1a-ff3da2a3d751\") " pod="openshift-console/console-f9d7485db-9hcbc" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.739797 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5bddce7c-b835-4a1f-a3af-db5a1d3771f2-config\") pod \"machine-approver-56656f9798-pglrh\" (UID: \"5bddce7c-b835-4a1f-a3af-db5a1d3771f2\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-pglrh" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.739816 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e4ea5d4f-3481-4e72-9bd8-e5d8183bc687-serving-cert\") pod \"apiserver-7bbb656c7d-stjw8\" (UID: \"e4ea5d4f-3481-4e72-9bd8-e5d8183bc687\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-stjw8" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.739891 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/770ae4fc-21dd-446d-9a41-26e484687787-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-6t2mt\" (UID: \"770ae4fc-21dd-446d-9a41-26e484687787\") " pod="openshift-authentication/oauth-openshift-558db77b4-6t2mt" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.739944 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/74f16025-fe11-4a23-8c1a-ff3da2a3d751-console-config\") pod \"console-f9d7485db-9hcbc\" (UID: \"74f16025-fe11-4a23-8c1a-ff3da2a3d751\") " pod="openshift-console/console-f9d7485db-9hcbc" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.739971 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r8vnl\" (UniqueName: \"kubernetes.io/projected/74f16025-fe11-4a23-8c1a-ff3da2a3d751-kube-api-access-r8vnl\") pod \"console-f9d7485db-9hcbc\" (UID: \"74f16025-fe11-4a23-8c1a-ff3da2a3d751\") " pod="openshift-console/console-f9d7485db-9hcbc" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.739997 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/75d0c362-7aa3-4919-937d-8849f425466e-trusted-ca-bundle\") pod \"apiserver-76f77b778f-6bpvf\" (UID: \"75d0c362-7aa3-4919-937d-8849f425466e\") " pod="openshift-apiserver/apiserver-76f77b778f-6bpvf" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.740022 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/e4ea5d4f-3481-4e72-9bd8-e5d8183bc687-audit-dir\") pod \"apiserver-7bbb656c7d-stjw8\" (UID: \"e4ea5d4f-3481-4e72-9bd8-e5d8183bc687\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-stjw8" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.740074 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/75d0c362-7aa3-4919-937d-8849f425466e-audit\") pod \"apiserver-76f77b778f-6bpvf\" (UID: \"75d0c362-7aa3-4919-937d-8849f425466e\") " pod="openshift-apiserver/apiserver-76f77b778f-6bpvf" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.740115 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/beec0ea6-0c57-44d6-bfef-d00e5b3724ad-serving-cert\") pod \"route-controller-manager-6576b87f9c-vwktr\" (UID: \"beec0ea6-0c57-44d6-bfef-d00e5b3724ad\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-vwktr" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.740133 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/75d0c362-7aa3-4919-937d-8849f425466e-config\") pod \"apiserver-76f77b778f-6bpvf\" (UID: \"75d0c362-7aa3-4919-937d-8849f425466e\") " pod="openshift-apiserver/apiserver-76f77b778f-6bpvf" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.740152 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/770ae4fc-21dd-446d-9a41-26e484687787-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-6t2mt\" (UID: \"770ae4fc-21dd-446d-9a41-26e484687787\") " pod="openshift-authentication/oauth-openshift-558db77b4-6t2mt" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.740168 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/770ae4fc-21dd-446d-9a41-26e484687787-audit-dir\") pod \"oauth-openshift-558db77b4-6t2mt\" (UID: \"770ae4fc-21dd-446d-9a41-26e484687787\") " pod="openshift-authentication/oauth-openshift-558db77b4-6t2mt" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.740203 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/75d0c362-7aa3-4919-937d-8849f425466e-encryption-config\") pod \"apiserver-76f77b778f-6bpvf\" (UID: \"75d0c362-7aa3-4919-937d-8849f425466e\") " pod="openshift-apiserver/apiserver-76f77b778f-6bpvf" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.740221 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pmhpw\" (UniqueName: \"kubernetes.io/projected/beec0ea6-0c57-44d6-bfef-d00e5b3724ad-kube-api-access-pmhpw\") pod \"route-controller-manager-6576b87f9c-vwktr\" (UID: \"beec0ea6-0c57-44d6-bfef-d00e5b3724ad\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-vwktr" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.740237 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/e4ea5d4f-3481-4e72-9bd8-e5d8183bc687-etcd-client\") pod \"apiserver-7bbb656c7d-stjw8\" (UID: \"e4ea5d4f-3481-4e72-9bd8-e5d8183bc687\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-stjw8" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.741854 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.746703 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-m467d"] Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.748240 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-gqn9s"] Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.748897 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-vwktr"] Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.748998 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-gqn9s" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.749858 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hdh6p\" (UniqueName: \"kubernetes.io/projected/5bddce7c-b835-4a1f-a3af-db5a1d3771f2-kube-api-access-hdh6p\") pod \"machine-approver-56656f9798-pglrh\" (UID: \"5bddce7c-b835-4a1f-a3af-db5a1d3771f2\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-pglrh" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.749905 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/75d0c362-7aa3-4919-937d-8849f425466e-image-import-ca\") pod \"apiserver-76f77b778f-6bpvf\" (UID: \"75d0c362-7aa3-4919-937d-8849f425466e\") " pod="openshift-apiserver/apiserver-76f77b778f-6bpvf" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.749930 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v5gbg\" (UniqueName: \"kubernetes.io/projected/f743f03b-3f08-4da0-b53e-1cb33ab516b9-kube-api-access-v5gbg\") pod \"machine-api-operator-5694c8668f-vwx77\" (UID: \"f743f03b-3f08-4da0-b53e-1cb33ab516b9\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-vwx77" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.749947 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/770ae4fc-21dd-446d-9a41-26e484687787-audit-policies\") pod \"oauth-openshift-558db77b4-6t2mt\" (UID: \"770ae4fc-21dd-446d-9a41-26e484687787\") " pod="openshift-authentication/oauth-openshift-558db77b4-6t2mt" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.749973 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/74f16025-fe11-4a23-8c1a-ff3da2a3d751-oauth-serving-cert\") pod \"console-f9d7485db-9hcbc\" (UID: \"74f16025-fe11-4a23-8c1a-ff3da2a3d751\") " pod="openshift-console/console-f9d7485db-9hcbc" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.749992 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b1cd661a-c160-4e14-81cf-842d3b9ae01d-config\") pod \"openshift-apiserver-operator-796bbdcf4f-m467d\" (UID: \"b1cd661a-c160-4e14-81cf-842d3b9ae01d\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-m467d" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.750008 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/75d0c362-7aa3-4919-937d-8849f425466e-node-pullsecrets\") pod \"apiserver-76f77b778f-6bpvf\" (UID: \"75d0c362-7aa3-4919-937d-8849f425466e\") " pod="openshift-apiserver/apiserver-76f77b778f-6bpvf" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.750042 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/f743f03b-3f08-4da0-b53e-1cb33ab516b9-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-vwx77\" (UID: \"f743f03b-3f08-4da0-b53e-1cb33ab516b9\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-vwx77" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.750061 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/beec0ea6-0c57-44d6-bfef-d00e5b3724ad-config\") pod \"route-controller-manager-6576b87f9c-vwktr\" (UID: \"beec0ea6-0c57-44d6-bfef-d00e5b3724ad\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-vwktr" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.750078 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n8cd4\" (UniqueName: \"kubernetes.io/projected/475e1ca4-a6fa-4f77-8c89-0619b2e9681e-kube-api-access-n8cd4\") pod \"downloads-7954f5f757-c6cxs\" (UID: \"475e1ca4-a6fa-4f77-8c89-0619b2e9681e\") " pod="openshift-console/downloads-7954f5f757-c6cxs" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.750099 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/770ae4fc-21dd-446d-9a41-26e484687787-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-6t2mt\" (UID: \"770ae4fc-21dd-446d-9a41-26e484687787\") " pod="openshift-authentication/oauth-openshift-558db77b4-6t2mt" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.750117 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/5bddce7c-b835-4a1f-a3af-db5a1d3771f2-machine-approver-tls\") pod \"machine-approver-56656f9798-pglrh\" (UID: \"5bddce7c-b835-4a1f-a3af-db5a1d3771f2\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-pglrh" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.750125 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.750136 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b1cd661a-c160-4e14-81cf-842d3b9ae01d-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-m467d\" (UID: \"b1cd661a-c160-4e14-81cf-842d3b9ae01d\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-m467d" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.750153 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/e4ea5d4f-3481-4e72-9bd8-e5d8183bc687-audit-policies\") pod \"apiserver-7bbb656c7d-stjw8\" (UID: \"e4ea5d4f-3481-4e72-9bd8-e5d8183bc687\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-stjw8" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.750168 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/e4ea5d4f-3481-4e72-9bd8-e5d8183bc687-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-stjw8\" (UID: \"e4ea5d4f-3481-4e72-9bd8-e5d8183bc687\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-stjw8" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.750199 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/5bddce7c-b835-4a1f-a3af-db5a1d3771f2-auth-proxy-config\") pod \"machine-approver-56656f9798-pglrh\" (UID: \"5bddce7c-b835-4a1f-a3af-db5a1d3771f2\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-pglrh" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.750220 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/770ae4fc-21dd-446d-9a41-26e484687787-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-6t2mt\" (UID: \"770ae4fc-21dd-446d-9a41-26e484687787\") " pod="openshift-authentication/oauth-openshift-558db77b4-6t2mt" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.750241 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/770ae4fc-21dd-446d-9a41-26e484687787-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-6t2mt\" (UID: \"770ae4fc-21dd-446d-9a41-26e484687787\") " pod="openshift-authentication/oauth-openshift-558db77b4-6t2mt" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.750257 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/75d0c362-7aa3-4919-937d-8849f425466e-serving-cert\") pod \"apiserver-76f77b778f-6bpvf\" (UID: \"75d0c362-7aa3-4919-937d-8849f425466e\") " pod="openshift-apiserver/apiserver-76f77b778f-6bpvf" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.750272 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n8bmr\" (UniqueName: \"kubernetes.io/projected/75d0c362-7aa3-4919-937d-8849f425466e-kube-api-access-n8bmr\") pod \"apiserver-76f77b778f-6bpvf\" (UID: \"75d0c362-7aa3-4919-937d-8849f425466e\") " pod="openshift-apiserver/apiserver-76f77b778f-6bpvf" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.750297 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/74f16025-fe11-4a23-8c1a-ff3da2a3d751-trusted-ca-bundle\") pod \"console-f9d7485db-9hcbc\" (UID: \"74f16025-fe11-4a23-8c1a-ff3da2a3d751\") " pod="openshift-console/console-f9d7485db-9hcbc" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.750315 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/85542c44-c09c-4c39-851f-fd320a0a36b2-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-wcg6b\" (UID: \"85542c44-c09c-4c39-851f-fd320a0a36b2\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-wcg6b" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.750335 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f743f03b-3f08-4da0-b53e-1cb33ab516b9-config\") pod \"machine-api-operator-5694c8668f-vwx77\" (UID: \"f743f03b-3f08-4da0-b53e-1cb33ab516b9\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-vwx77" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.750367 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/74f16025-fe11-4a23-8c1a-ff3da2a3d751-console-oauth-config\") pod \"console-f9d7485db-9hcbc\" (UID: \"74f16025-fe11-4a23-8c1a-ff3da2a3d751\") " pod="openshift-console/console-f9d7485db-9hcbc" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.750384 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k8gmq\" (UniqueName: \"kubernetes.io/projected/b1cd661a-c160-4e14-81cf-842d3b9ae01d-kube-api-access-k8gmq\") pod \"openshift-apiserver-operator-796bbdcf4f-m467d\" (UID: \"b1cd661a-c160-4e14-81cf-842d3b9ae01d\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-m467d" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.750399 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/75d0c362-7aa3-4919-937d-8849f425466e-etcd-client\") pod \"apiserver-76f77b778f-6bpvf\" (UID: \"75d0c362-7aa3-4919-937d-8849f425466e\") " pod="openshift-apiserver/apiserver-76f77b778f-6bpvf" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.750416 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k8whk\" (UniqueName: \"kubernetes.io/projected/e4ea5d4f-3481-4e72-9bd8-e5d8183bc687-kube-api-access-k8whk\") pod \"apiserver-7bbb656c7d-stjw8\" (UID: \"e4ea5d4f-3481-4e72-9bd8-e5d8183bc687\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-stjw8" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.750436 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-drmkm\" (UniqueName: \"kubernetes.io/projected/770ae4fc-21dd-446d-9a41-26e484687787-kube-api-access-drmkm\") pod \"oauth-openshift-558db77b4-6t2mt\" (UID: \"770ae4fc-21dd-446d-9a41-26e484687787\") " pod="openshift-authentication/oauth-openshift-558db77b4-6t2mt" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.750453 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/74f16025-fe11-4a23-8c1a-ff3da2a3d751-console-serving-cert\") pod \"console-f9d7485db-9hcbc\" (UID: \"74f16025-fe11-4a23-8c1a-ff3da2a3d751\") " pod="openshift-console/console-f9d7485db-9hcbc" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.750476 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/f743f03b-3f08-4da0-b53e-1cb33ab516b9-images\") pod \"machine-api-operator-5694c8668f-vwx77\" (UID: \"f743f03b-3f08-4da0-b53e-1cb33ab516b9\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-vwx77" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.751835 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/770ae4fc-21dd-446d-9a41-26e484687787-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-6t2mt\" (UID: \"770ae4fc-21dd-446d-9a41-26e484687787\") " pod="openshift-authentication/oauth-openshift-558db77b4-6t2mt" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.751866 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e4ea5d4f-3481-4e72-9bd8-e5d8183bc687-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-stjw8\" (UID: \"e4ea5d4f-3481-4e72-9bd8-e5d8183bc687\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-stjw8" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.751887 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/75d0c362-7aa3-4919-937d-8849f425466e-audit-dir\") pod \"apiserver-76f77b778f-6bpvf\" (UID: \"75d0c362-7aa3-4919-937d-8849f425466e\") " pod="openshift-apiserver/apiserver-76f77b778f-6bpvf" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.751905 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-49qp9\" (UniqueName: \"kubernetes.io/projected/53fe640c-6cb7-43a3-9a84-467eb36fe66b-kube-api-access-49qp9\") pod \"openshift-config-operator-7777fb866f-rcnlv\" (UID: \"53fe640c-6cb7-43a3-9a84-467eb36fe66b\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-rcnlv" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.751939 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/770ae4fc-21dd-446d-9a41-26e484687787-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-6t2mt\" (UID: \"770ae4fc-21dd-446d-9a41-26e484687787\") " pod="openshift-authentication/oauth-openshift-558db77b4-6t2mt" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.751957 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/770ae4fc-21dd-446d-9a41-26e484687787-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-6t2mt\" (UID: \"770ae4fc-21dd-446d-9a41-26e484687787\") " pod="openshift-authentication/oauth-openshift-558db77b4-6t2mt" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.751974 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/85542c44-c09c-4c39-851f-fd320a0a36b2-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-wcg6b\" (UID: \"85542c44-c09c-4c39-851f-fd320a0a36b2\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-wcg6b" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.751995 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/770ae4fc-21dd-446d-9a41-26e484687787-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-6t2mt\" (UID: \"770ae4fc-21dd-446d-9a41-26e484687787\") " pod="openshift-authentication/oauth-openshift-558db77b4-6t2mt" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.752012 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/75d0c362-7aa3-4919-937d-8849f425466e-etcd-serving-ca\") pod \"apiserver-76f77b778f-6bpvf\" (UID: \"75d0c362-7aa3-4919-937d-8849f425466e\") " pod="openshift-apiserver/apiserver-76f77b778f-6bpvf" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.752033 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xw5w5\" (UniqueName: \"kubernetes.io/projected/85542c44-c09c-4c39-851f-fd320a0a36b2-kube-api-access-xw5w5\") pod \"cluster-image-registry-operator-dc59b4c8b-wcg6b\" (UID: \"85542c44-c09c-4c39-851f-fd320a0a36b2\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-wcg6b" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.752049 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/53fe640c-6cb7-43a3-9a84-467eb36fe66b-serving-cert\") pod \"openshift-config-operator-7777fb866f-rcnlv\" (UID: \"53fe640c-6cb7-43a3-9a84-467eb36fe66b\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-rcnlv" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.752068 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/770ae4fc-21dd-446d-9a41-26e484687787-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-6t2mt\" (UID: \"770ae4fc-21dd-446d-9a41-26e484687787\") " pod="openshift-authentication/oauth-openshift-558db77b4-6t2mt" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.752084 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/85542c44-c09c-4c39-851f-fd320a0a36b2-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-wcg6b\" (UID: \"85542c44-c09c-4c39-851f-fd320a0a36b2\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-wcg6b" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.752120 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/770ae4fc-21dd-446d-9a41-26e484687787-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-6t2mt\" (UID: \"770ae4fc-21dd-446d-9a41-26e484687787\") " pod="openshift-authentication/oauth-openshift-558db77b4-6t2mt" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.752199 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/beec0ea6-0c57-44d6-bfef-d00e5b3724ad-client-ca\") pod \"route-controller-manager-6576b87f9c-vwktr\" (UID: \"beec0ea6-0c57-44d6-bfef-d00e5b3724ad\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-vwktr" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.752350 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-x5n4h"] Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.753226 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-8bpqg"] Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.753344 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-x5n4h" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.754140 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-8bpqg" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.754497 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-4pvwg"] Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.754898 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-4pvwg" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.755490 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-m7gph"] Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.756004 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-m7gph" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.756440 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-58hwl"] Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.757196 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-58hwl" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.762337 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-hmzsn"] Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.763008 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-hmzsn" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.764221 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.764443 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-zf4p8"] Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.765078 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-zf4p8" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.765228 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-2rc5n"] Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.766335 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-2rc5n" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.768724 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-pqm4r"] Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.769588 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-pqm4r" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.769730 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-2x9nl"] Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.773397 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-dtbh5"] Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.774134 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-bpbtf"] Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.774791 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-bpbtf" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.775314 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-2x9nl" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.775588 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-dtbh5" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.777306 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-5rj2f"] Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.795683 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-5rj2f" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.795981 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-6bpvf"] Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.797961 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-wcg6b"] Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.799790 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-wnmjw"] Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.801385 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-9hcbc"] Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.801575 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-wnmjw" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.801954 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.803031 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.803049 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29327445-dn5mf"] Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.804259 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29327445-dn5mf" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.805509 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-k87kc"] Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.808489 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-6t2mt"] Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.810471 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-t8f4h"] Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.811504 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-c6cxs"] Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.812487 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-4cxsw"] Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.813240 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-4cxsw" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.813536 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-stjw8"] Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.814556 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-5qwb9"] Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.815565 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-2pd47"] Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.816565 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-rcnlv"] Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.817571 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-vjmbx"] Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.818943 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-4pvwg"] Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.820348 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-skh8h"] Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.821134 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.821776 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-8vjkd"] Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.823067 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-hmzsn"] Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.824409 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-526w4"] Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.825096 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-rf9l4"] Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.826091 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-gqn9s"] Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.827120 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-9jtcc"] Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.828374 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-2x9nl"] Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.829221 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-dtbh5"] Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.830170 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-dsqhc"] Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.831738 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-jm4gb"] Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.831857 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-dsqhc" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.833962 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-7jflr"] Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.833984 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-x5n4h"] Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.834084 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-jm4gb" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.834119 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29327445-dn5mf"] Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.837230 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-pqm4r"] Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.837416 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-4cxsw"] Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.838551 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-zf4p8"] Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.839544 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-8bpqg"] Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.840523 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-m7gph"] Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.841403 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.841530 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-2rc5n"] Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.842964 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-jm4gb"] Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.843998 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-wnmjw"] Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.845012 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-5rj2f"] Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.846015 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-bpbtf"] Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.847035 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-dsqhc"] Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.847988 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-2g2rj"] Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.848688 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-2g2rj" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.853073 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/74f16025-fe11-4a23-8c1a-ff3da2a3d751-oauth-serving-cert\") pod \"console-f9d7485db-9hcbc\" (UID: \"74f16025-fe11-4a23-8c1a-ff3da2a3d751\") " pod="openshift-console/console-f9d7485db-9hcbc" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.853103 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b1cd661a-c160-4e14-81cf-842d3b9ae01d-config\") pod \"openshift-apiserver-operator-796bbdcf4f-m467d\" (UID: \"b1cd661a-c160-4e14-81cf-842d3b9ae01d\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-m467d" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.853123 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/75d0c362-7aa3-4919-937d-8849f425466e-node-pullsecrets\") pod \"apiserver-76f77b778f-6bpvf\" (UID: \"75d0c362-7aa3-4919-937d-8849f425466e\") " pod="openshift-apiserver/apiserver-76f77b778f-6bpvf" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.853143 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/beec0ea6-0c57-44d6-bfef-d00e5b3724ad-config\") pod \"route-controller-manager-6576b87f9c-vwktr\" (UID: \"beec0ea6-0c57-44d6-bfef-d00e5b3724ad\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-vwktr" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.853158 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n8cd4\" (UniqueName: \"kubernetes.io/projected/475e1ca4-a6fa-4f77-8c89-0619b2e9681e-kube-api-access-n8cd4\") pod \"downloads-7954f5f757-c6cxs\" (UID: \"475e1ca4-a6fa-4f77-8c89-0619b2e9681e\") " pod="openshift-console/downloads-7954f5f757-c6cxs" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.853262 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/75d0c362-7aa3-4919-937d-8849f425466e-node-pullsecrets\") pod \"apiserver-76f77b778f-6bpvf\" (UID: \"75d0c362-7aa3-4919-937d-8849f425466e\") " pod="openshift-apiserver/apiserver-76f77b778f-6bpvf" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.853401 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/f743f03b-3f08-4da0-b53e-1cb33ab516b9-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-vwx77\" (UID: \"f743f03b-3f08-4da0-b53e-1cb33ab516b9\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-vwx77" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.853435 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hnbqg\" (UniqueName: \"kubernetes.io/projected/7b82343c-28e6-4e0b-9731-5131cb0014f2-kube-api-access-hnbqg\") pod \"ingress-operator-5b745b69d9-9jtcc\" (UID: \"7b82343c-28e6-4e0b-9731-5131cb0014f2\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-9jtcc" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.853467 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/770ae4fc-21dd-446d-9a41-26e484687787-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-6t2mt\" (UID: \"770ae4fc-21dd-446d-9a41-26e484687787\") " pod="openshift-authentication/oauth-openshift-558db77b4-6t2mt" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.853485 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/5bddce7c-b835-4a1f-a3af-db5a1d3771f2-machine-approver-tls\") pod \"machine-approver-56656f9798-pglrh\" (UID: \"5bddce7c-b835-4a1f-a3af-db5a1d3771f2\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-pglrh" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.853503 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/5bddce7c-b835-4a1f-a3af-db5a1d3771f2-auth-proxy-config\") pod \"machine-approver-56656f9798-pglrh\" (UID: \"5bddce7c-b835-4a1f-a3af-db5a1d3771f2\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-pglrh" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.853520 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b1cd661a-c160-4e14-81cf-842d3b9ae01d-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-m467d\" (UID: \"b1cd661a-c160-4e14-81cf-842d3b9ae01d\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-m467d" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.853538 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/e4ea5d4f-3481-4e72-9bd8-e5d8183bc687-audit-policies\") pod \"apiserver-7bbb656c7d-stjw8\" (UID: \"e4ea5d4f-3481-4e72-9bd8-e5d8183bc687\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-stjw8" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.853556 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/e4ea5d4f-3481-4e72-9bd8-e5d8183bc687-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-stjw8\" (UID: \"e4ea5d4f-3481-4e72-9bd8-e5d8183bc687\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-stjw8" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.853571 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/75d0c362-7aa3-4919-937d-8849f425466e-serving-cert\") pod \"apiserver-76f77b778f-6bpvf\" (UID: \"75d0c362-7aa3-4919-937d-8849f425466e\") " pod="openshift-apiserver/apiserver-76f77b778f-6bpvf" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.853588 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n8bmr\" (UniqueName: \"kubernetes.io/projected/75d0c362-7aa3-4919-937d-8849f425466e-kube-api-access-n8bmr\") pod \"apiserver-76f77b778f-6bpvf\" (UID: \"75d0c362-7aa3-4919-937d-8849f425466e\") " pod="openshift-apiserver/apiserver-76f77b778f-6bpvf" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.853607 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/770ae4fc-21dd-446d-9a41-26e484687787-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-6t2mt\" (UID: \"770ae4fc-21dd-446d-9a41-26e484687787\") " pod="openshift-authentication/oauth-openshift-558db77b4-6t2mt" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.853625 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/770ae4fc-21dd-446d-9a41-26e484687787-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-6t2mt\" (UID: \"770ae4fc-21dd-446d-9a41-26e484687787\") " pod="openshift-authentication/oauth-openshift-558db77b4-6t2mt" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.853646 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/74f16025-fe11-4a23-8c1a-ff3da2a3d751-trusted-ca-bundle\") pod \"console-f9d7485db-9hcbc\" (UID: \"74f16025-fe11-4a23-8c1a-ff3da2a3d751\") " pod="openshift-console/console-f9d7485db-9hcbc" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.853687 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/85542c44-c09c-4c39-851f-fd320a0a36b2-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-wcg6b\" (UID: \"85542c44-c09c-4c39-851f-fd320a0a36b2\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-wcg6b" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.853719 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k8whk\" (UniqueName: \"kubernetes.io/projected/e4ea5d4f-3481-4e72-9bd8-e5d8183bc687-kube-api-access-k8whk\") pod \"apiserver-7bbb656c7d-stjw8\" (UID: \"e4ea5d4f-3481-4e72-9bd8-e5d8183bc687\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-stjw8" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.853742 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/7b82343c-28e6-4e0b-9731-5131cb0014f2-bound-sa-token\") pod \"ingress-operator-5b745b69d9-9jtcc\" (UID: \"7b82343c-28e6-4e0b-9731-5131cb0014f2\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-9jtcc" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.853771 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f743f03b-3f08-4da0-b53e-1cb33ab516b9-config\") pod \"machine-api-operator-5694c8668f-vwx77\" (UID: \"f743f03b-3f08-4da0-b53e-1cb33ab516b9\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-vwx77" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.853788 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/74f16025-fe11-4a23-8c1a-ff3da2a3d751-console-oauth-config\") pod \"console-f9d7485db-9hcbc\" (UID: \"74f16025-fe11-4a23-8c1a-ff3da2a3d751\") " pod="openshift-console/console-f9d7485db-9hcbc" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.853804 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k8gmq\" (UniqueName: \"kubernetes.io/projected/b1cd661a-c160-4e14-81cf-842d3b9ae01d-kube-api-access-k8gmq\") pod \"openshift-apiserver-operator-796bbdcf4f-m467d\" (UID: \"b1cd661a-c160-4e14-81cf-842d3b9ae01d\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-m467d" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.853820 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/75d0c362-7aa3-4919-937d-8849f425466e-etcd-client\") pod \"apiserver-76f77b778f-6bpvf\" (UID: \"75d0c362-7aa3-4919-937d-8849f425466e\") " pod="openshift-apiserver/apiserver-76f77b778f-6bpvf" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.853837 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-drmkm\" (UniqueName: \"kubernetes.io/projected/770ae4fc-21dd-446d-9a41-26e484687787-kube-api-access-drmkm\") pod \"oauth-openshift-558db77b4-6t2mt\" (UID: \"770ae4fc-21dd-446d-9a41-26e484687787\") " pod="openshift-authentication/oauth-openshift-558db77b4-6t2mt" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.853855 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/74f16025-fe11-4a23-8c1a-ff3da2a3d751-console-serving-cert\") pod \"console-f9d7485db-9hcbc\" (UID: \"74f16025-fe11-4a23-8c1a-ff3da2a3d751\") " pod="openshift-console/console-f9d7485db-9hcbc" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.853874 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/f743f03b-3f08-4da0-b53e-1cb33ab516b9-images\") pod \"machine-api-operator-5694c8668f-vwx77\" (UID: \"f743f03b-3f08-4da0-b53e-1cb33ab516b9\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-vwx77" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.853892 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/770ae4fc-21dd-446d-9a41-26e484687787-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-6t2mt\" (UID: \"770ae4fc-21dd-446d-9a41-26e484687787\") " pod="openshift-authentication/oauth-openshift-558db77b4-6t2mt" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.853910 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e4ea5d4f-3481-4e72-9bd8-e5d8183bc687-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-stjw8\" (UID: \"e4ea5d4f-3481-4e72-9bd8-e5d8183bc687\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-stjw8" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.853929 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/75d0c362-7aa3-4919-937d-8849f425466e-audit-dir\") pod \"apiserver-76f77b778f-6bpvf\" (UID: \"75d0c362-7aa3-4919-937d-8849f425466e\") " pod="openshift-apiserver/apiserver-76f77b778f-6bpvf" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.853943 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b1cd661a-c160-4e14-81cf-842d3b9ae01d-config\") pod \"openshift-apiserver-operator-796bbdcf4f-m467d\" (UID: \"b1cd661a-c160-4e14-81cf-842d3b9ae01d\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-m467d" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.853954 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-49qp9\" (UniqueName: \"kubernetes.io/projected/53fe640c-6cb7-43a3-9a84-467eb36fe66b-kube-api-access-49qp9\") pod \"openshift-config-operator-7777fb866f-rcnlv\" (UID: \"53fe640c-6cb7-43a3-9a84-467eb36fe66b\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-rcnlv" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.853999 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/85542c44-c09c-4c39-851f-fd320a0a36b2-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-wcg6b\" (UID: \"85542c44-c09c-4c39-851f-fd320a0a36b2\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-wcg6b" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.854042 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/770ae4fc-21dd-446d-9a41-26e484687787-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-6t2mt\" (UID: \"770ae4fc-21dd-446d-9a41-26e484687787\") " pod="openshift-authentication/oauth-openshift-558db77b4-6t2mt" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.854062 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/770ae4fc-21dd-446d-9a41-26e484687787-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-6t2mt\" (UID: \"770ae4fc-21dd-446d-9a41-26e484687787\") " pod="openshift-authentication/oauth-openshift-558db77b4-6t2mt" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.854081 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/770ae4fc-21dd-446d-9a41-26e484687787-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-6t2mt\" (UID: \"770ae4fc-21dd-446d-9a41-26e484687787\") " pod="openshift-authentication/oauth-openshift-558db77b4-6t2mt" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.854100 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/75d0c362-7aa3-4919-937d-8849f425466e-etcd-serving-ca\") pod \"apiserver-76f77b778f-6bpvf\" (UID: \"75d0c362-7aa3-4919-937d-8849f425466e\") " pod="openshift-apiserver/apiserver-76f77b778f-6bpvf" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.854124 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xw5w5\" (UniqueName: \"kubernetes.io/projected/85542c44-c09c-4c39-851f-fd320a0a36b2-kube-api-access-xw5w5\") pod \"cluster-image-registry-operator-dc59b4c8b-wcg6b\" (UID: \"85542c44-c09c-4c39-851f-fd320a0a36b2\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-wcg6b" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.854145 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/53fe640c-6cb7-43a3-9a84-467eb36fe66b-serving-cert\") pod \"openshift-config-operator-7777fb866f-rcnlv\" (UID: \"53fe640c-6cb7-43a3-9a84-467eb36fe66b\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-rcnlv" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.854163 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/770ae4fc-21dd-446d-9a41-26e484687787-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-6t2mt\" (UID: \"770ae4fc-21dd-446d-9a41-26e484687787\") " pod="openshift-authentication/oauth-openshift-558db77b4-6t2mt" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.854204 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/85542c44-c09c-4c39-851f-fd320a0a36b2-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-wcg6b\" (UID: \"85542c44-c09c-4c39-851f-fd320a0a36b2\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-wcg6b" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.854220 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/74f16025-fe11-4a23-8c1a-ff3da2a3d751-oauth-serving-cert\") pod \"console-f9d7485db-9hcbc\" (UID: \"74f16025-fe11-4a23-8c1a-ff3da2a3d751\") " pod="openshift-console/console-f9d7485db-9hcbc" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.854224 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/770ae4fc-21dd-446d-9a41-26e484687787-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-6t2mt\" (UID: \"770ae4fc-21dd-446d-9a41-26e484687787\") " pod="openshift-authentication/oauth-openshift-558db77b4-6t2mt" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.854284 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/beec0ea6-0c57-44d6-bfef-d00e5b3724ad-client-ca\") pod \"route-controller-manager-6576b87f9c-vwktr\" (UID: \"beec0ea6-0c57-44d6-bfef-d00e5b3724ad\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-vwktr" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.854309 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/74f16025-fe11-4a23-8c1a-ff3da2a3d751-service-ca\") pod \"console-f9d7485db-9hcbc\" (UID: \"74f16025-fe11-4a23-8c1a-ff3da2a3d751\") " pod="openshift-console/console-f9d7485db-9hcbc" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.854331 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/e4ea5d4f-3481-4e72-9bd8-e5d8183bc687-encryption-config\") pod \"apiserver-7bbb656c7d-stjw8\" (UID: \"e4ea5d4f-3481-4e72-9bd8-e5d8183bc687\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-stjw8" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.854350 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/53fe640c-6cb7-43a3-9a84-467eb36fe66b-available-featuregates\") pod \"openshift-config-operator-7777fb866f-rcnlv\" (UID: \"53fe640c-6cb7-43a3-9a84-467eb36fe66b\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-rcnlv" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.854376 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5bddce7c-b835-4a1f-a3af-db5a1d3771f2-config\") pod \"machine-approver-56656f9798-pglrh\" (UID: \"5bddce7c-b835-4a1f-a3af-db5a1d3771f2\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-pglrh" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.854394 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e4ea5d4f-3481-4e72-9bd8-e5d8183bc687-serving-cert\") pod \"apiserver-7bbb656c7d-stjw8\" (UID: \"e4ea5d4f-3481-4e72-9bd8-e5d8183bc687\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-stjw8" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.854413 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/770ae4fc-21dd-446d-9a41-26e484687787-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-6t2mt\" (UID: \"770ae4fc-21dd-446d-9a41-26e484687787\") " pod="openshift-authentication/oauth-openshift-558db77b4-6t2mt" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.854454 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/e4ea5d4f-3481-4e72-9bd8-e5d8183bc687-audit-dir\") pod \"apiserver-7bbb656c7d-stjw8\" (UID: \"e4ea5d4f-3481-4e72-9bd8-e5d8183bc687\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-stjw8" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.854473 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/74f16025-fe11-4a23-8c1a-ff3da2a3d751-console-config\") pod \"console-f9d7485db-9hcbc\" (UID: \"74f16025-fe11-4a23-8c1a-ff3da2a3d751\") " pod="openshift-console/console-f9d7485db-9hcbc" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.854489 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r8vnl\" (UniqueName: \"kubernetes.io/projected/74f16025-fe11-4a23-8c1a-ff3da2a3d751-kube-api-access-r8vnl\") pod \"console-f9d7485db-9hcbc\" (UID: \"74f16025-fe11-4a23-8c1a-ff3da2a3d751\") " pod="openshift-console/console-f9d7485db-9hcbc" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.854505 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/75d0c362-7aa3-4919-937d-8849f425466e-trusted-ca-bundle\") pod \"apiserver-76f77b778f-6bpvf\" (UID: \"75d0c362-7aa3-4919-937d-8849f425466e\") " pod="openshift-apiserver/apiserver-76f77b778f-6bpvf" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.854525 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/75d0c362-7aa3-4919-937d-8849f425466e-audit\") pod \"apiserver-76f77b778f-6bpvf\" (UID: \"75d0c362-7aa3-4919-937d-8849f425466e\") " pod="openshift-apiserver/apiserver-76f77b778f-6bpvf" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.854552 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/beec0ea6-0c57-44d6-bfef-d00e5b3724ad-serving-cert\") pod \"route-controller-manager-6576b87f9c-vwktr\" (UID: \"beec0ea6-0c57-44d6-bfef-d00e5b3724ad\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-vwktr" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.854568 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/75d0c362-7aa3-4919-937d-8849f425466e-config\") pod \"apiserver-76f77b778f-6bpvf\" (UID: \"75d0c362-7aa3-4919-937d-8849f425466e\") " pod="openshift-apiserver/apiserver-76f77b778f-6bpvf" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.854584 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/770ae4fc-21dd-446d-9a41-26e484687787-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-6t2mt\" (UID: \"770ae4fc-21dd-446d-9a41-26e484687787\") " pod="openshift-authentication/oauth-openshift-558db77b4-6t2mt" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.854603 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/770ae4fc-21dd-446d-9a41-26e484687787-audit-dir\") pod \"oauth-openshift-558db77b4-6t2mt\" (UID: \"770ae4fc-21dd-446d-9a41-26e484687787\") " pod="openshift-authentication/oauth-openshift-558db77b4-6t2mt" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.854623 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/75d0c362-7aa3-4919-937d-8849f425466e-encryption-config\") pod \"apiserver-76f77b778f-6bpvf\" (UID: \"75d0c362-7aa3-4919-937d-8849f425466e\") " pod="openshift-apiserver/apiserver-76f77b778f-6bpvf" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.854642 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pmhpw\" (UniqueName: \"kubernetes.io/projected/beec0ea6-0c57-44d6-bfef-d00e5b3724ad-kube-api-access-pmhpw\") pod \"route-controller-manager-6576b87f9c-vwktr\" (UID: \"beec0ea6-0c57-44d6-bfef-d00e5b3724ad\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-vwktr" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.854657 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/e4ea5d4f-3481-4e72-9bd8-e5d8183bc687-etcd-client\") pod \"apiserver-7bbb656c7d-stjw8\" (UID: \"e4ea5d4f-3481-4e72-9bd8-e5d8183bc687\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-stjw8" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.854678 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/7b82343c-28e6-4e0b-9731-5131cb0014f2-trusted-ca\") pod \"ingress-operator-5b745b69d9-9jtcc\" (UID: \"7b82343c-28e6-4e0b-9731-5131cb0014f2\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-9jtcc" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.854705 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hdh6p\" (UniqueName: \"kubernetes.io/projected/5bddce7c-b835-4a1f-a3af-db5a1d3771f2-kube-api-access-hdh6p\") pod \"machine-approver-56656f9798-pglrh\" (UID: \"5bddce7c-b835-4a1f-a3af-db5a1d3771f2\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-pglrh" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.854711 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/770ae4fc-21dd-446d-9a41-26e484687787-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-6t2mt\" (UID: \"770ae4fc-21dd-446d-9a41-26e484687787\") " pod="openshift-authentication/oauth-openshift-558db77b4-6t2mt" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.854712 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/beec0ea6-0c57-44d6-bfef-d00e5b3724ad-config\") pod \"route-controller-manager-6576b87f9c-vwktr\" (UID: \"beec0ea6-0c57-44d6-bfef-d00e5b3724ad\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-vwktr" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.854724 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/75d0c362-7aa3-4919-937d-8849f425466e-image-import-ca\") pod \"apiserver-76f77b778f-6bpvf\" (UID: \"75d0c362-7aa3-4919-937d-8849f425466e\") " pod="openshift-apiserver/apiserver-76f77b778f-6bpvf" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.854743 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/7b82343c-28e6-4e0b-9731-5131cb0014f2-metrics-tls\") pod \"ingress-operator-5b745b69d9-9jtcc\" (UID: \"7b82343c-28e6-4e0b-9731-5131cb0014f2\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-9jtcc" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.854765 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v5gbg\" (UniqueName: \"kubernetes.io/projected/f743f03b-3f08-4da0-b53e-1cb33ab516b9-kube-api-access-v5gbg\") pod \"machine-api-operator-5694c8668f-vwx77\" (UID: \"f743f03b-3f08-4da0-b53e-1cb33ab516b9\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-vwx77" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.854783 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/770ae4fc-21dd-446d-9a41-26e484687787-audit-policies\") pod \"oauth-openshift-558db77b4-6t2mt\" (UID: \"770ae4fc-21dd-446d-9a41-26e484687787\") " pod="openshift-authentication/oauth-openshift-558db77b4-6t2mt" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.855332 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/770ae4fc-21dd-446d-9a41-26e484687787-audit-policies\") pod \"oauth-openshift-558db77b4-6t2mt\" (UID: \"770ae4fc-21dd-446d-9a41-26e484687787\") " pod="openshift-authentication/oauth-openshift-558db77b4-6t2mt" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.855845 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/beec0ea6-0c57-44d6-bfef-d00e5b3724ad-client-ca\") pod \"route-controller-manager-6576b87f9c-vwktr\" (UID: \"beec0ea6-0c57-44d6-bfef-d00e5b3724ad\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-vwktr" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.857076 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/74f16025-fe11-4a23-8c1a-ff3da2a3d751-service-ca\") pod \"console-f9d7485db-9hcbc\" (UID: \"74f16025-fe11-4a23-8c1a-ff3da2a3d751\") " pod="openshift-console/console-f9d7485db-9hcbc" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.857141 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/75d0c362-7aa3-4919-937d-8849f425466e-etcd-serving-ca\") pod \"apiserver-76f77b778f-6bpvf\" (UID: \"75d0c362-7aa3-4919-937d-8849f425466e\") " pod="openshift-apiserver/apiserver-76f77b778f-6bpvf" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.858064 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f743f03b-3f08-4da0-b53e-1cb33ab516b9-config\") pod \"machine-api-operator-5694c8668f-vwx77\" (UID: \"f743f03b-3f08-4da0-b53e-1cb33ab516b9\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-vwx77" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.858614 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/53fe640c-6cb7-43a3-9a84-467eb36fe66b-available-featuregates\") pod \"openshift-config-operator-7777fb866f-rcnlv\" (UID: \"53fe640c-6cb7-43a3-9a84-467eb36fe66b\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-rcnlv" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.859208 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/f743f03b-3f08-4da0-b53e-1cb33ab516b9-images\") pod \"machine-api-operator-5694c8668f-vwx77\" (UID: \"f743f03b-3f08-4da0-b53e-1cb33ab516b9\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-vwx77" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.859524 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/770ae4fc-21dd-446d-9a41-26e484687787-audit-dir\") pod \"oauth-openshift-558db77b4-6t2mt\" (UID: \"770ae4fc-21dd-446d-9a41-26e484687787\") " pod="openshift-authentication/oauth-openshift-558db77b4-6t2mt" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.859708 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5bddce7c-b835-4a1f-a3af-db5a1d3771f2-config\") pod \"machine-approver-56656f9798-pglrh\" (UID: \"5bddce7c-b835-4a1f-a3af-db5a1d3771f2\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-pglrh" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.861137 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/e4ea5d4f-3481-4e72-9bd8-e5d8183bc687-audit-dir\") pod \"apiserver-7bbb656c7d-stjw8\" (UID: \"e4ea5d4f-3481-4e72-9bd8-e5d8183bc687\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-stjw8" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.861603 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e4ea5d4f-3481-4e72-9bd8-e5d8183bc687-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-stjw8\" (UID: \"e4ea5d4f-3481-4e72-9bd8-e5d8183bc687\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-stjw8" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.862029 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/770ae4fc-21dd-446d-9a41-26e484687787-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-6t2mt\" (UID: \"770ae4fc-21dd-446d-9a41-26e484687787\") " pod="openshift-authentication/oauth-openshift-558db77b4-6t2mt" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.862103 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/75d0c362-7aa3-4919-937d-8849f425466e-etcd-client\") pod \"apiserver-76f77b778f-6bpvf\" (UID: \"75d0c362-7aa3-4919-937d-8849f425466e\") " pod="openshift-apiserver/apiserver-76f77b778f-6bpvf" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.862642 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/75d0c362-7aa3-4919-937d-8849f425466e-audit-dir\") pod \"apiserver-76f77b778f-6bpvf\" (UID: \"75d0c362-7aa3-4919-937d-8849f425466e\") " pod="openshift-apiserver/apiserver-76f77b778f-6bpvf" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.863269 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/e4ea5d4f-3481-4e72-9bd8-e5d8183bc687-encryption-config\") pod \"apiserver-7bbb656c7d-stjw8\" (UID: \"e4ea5d4f-3481-4e72-9bd8-e5d8183bc687\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-stjw8" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.863340 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/770ae4fc-21dd-446d-9a41-26e484687787-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-6t2mt\" (UID: \"770ae4fc-21dd-446d-9a41-26e484687787\") " pod="openshift-authentication/oauth-openshift-558db77b4-6t2mt" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.863365 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/85542c44-c09c-4c39-851f-fd320a0a36b2-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-wcg6b\" (UID: \"85542c44-c09c-4c39-851f-fd320a0a36b2\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-wcg6b" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.864145 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/74f16025-fe11-4a23-8c1a-ff3da2a3d751-console-config\") pod \"console-f9d7485db-9hcbc\" (UID: \"74f16025-fe11-4a23-8c1a-ff3da2a3d751\") " pod="openshift-console/console-f9d7485db-9hcbc" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.864212 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/5bddce7c-b835-4a1f-a3af-db5a1d3771f2-auth-proxy-config\") pod \"machine-approver-56656f9798-pglrh\" (UID: \"5bddce7c-b835-4a1f-a3af-db5a1d3771f2\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-pglrh" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.864294 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.864409 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/e4ea5d4f-3481-4e72-9bd8-e5d8183bc687-audit-policies\") pod \"apiserver-7bbb656c7d-stjw8\" (UID: \"e4ea5d4f-3481-4e72-9bd8-e5d8183bc687\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-stjw8" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.864542 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/770ae4fc-21dd-446d-9a41-26e484687787-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-6t2mt\" (UID: \"770ae4fc-21dd-446d-9a41-26e484687787\") " pod="openshift-authentication/oauth-openshift-558db77b4-6t2mt" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.865249 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/770ae4fc-21dd-446d-9a41-26e484687787-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-6t2mt\" (UID: \"770ae4fc-21dd-446d-9a41-26e484687787\") " pod="openshift-authentication/oauth-openshift-558db77b4-6t2mt" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.865356 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/75d0c362-7aa3-4919-937d-8849f425466e-config\") pod \"apiserver-76f77b778f-6bpvf\" (UID: \"75d0c362-7aa3-4919-937d-8849f425466e\") " pod="openshift-apiserver/apiserver-76f77b778f-6bpvf" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.865356 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/e4ea5d4f-3481-4e72-9bd8-e5d8183bc687-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-stjw8\" (UID: \"e4ea5d4f-3481-4e72-9bd8-e5d8183bc687\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-stjw8" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.866068 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/53fe640c-6cb7-43a3-9a84-467eb36fe66b-serving-cert\") pod \"openshift-config-operator-7777fb866f-rcnlv\" (UID: \"53fe640c-6cb7-43a3-9a84-467eb36fe66b\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-rcnlv" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.866614 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/5bddce7c-b835-4a1f-a3af-db5a1d3771f2-machine-approver-tls\") pod \"machine-approver-56656f9798-pglrh\" (UID: \"5bddce7c-b835-4a1f-a3af-db5a1d3771f2\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-pglrh" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.866653 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/770ae4fc-21dd-446d-9a41-26e484687787-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-6t2mt\" (UID: \"770ae4fc-21dd-446d-9a41-26e484687787\") " pod="openshift-authentication/oauth-openshift-558db77b4-6t2mt" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.866713 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/f743f03b-3f08-4da0-b53e-1cb33ab516b9-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-vwx77\" (UID: \"f743f03b-3f08-4da0-b53e-1cb33ab516b9\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-vwx77" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.866637 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/770ae4fc-21dd-446d-9a41-26e484687787-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-6t2mt\" (UID: \"770ae4fc-21dd-446d-9a41-26e484687787\") " pod="openshift-authentication/oauth-openshift-558db77b4-6t2mt" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.866934 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/74f16025-fe11-4a23-8c1a-ff3da2a3d751-trusted-ca-bundle\") pod \"console-f9d7485db-9hcbc\" (UID: \"74f16025-fe11-4a23-8c1a-ff3da2a3d751\") " pod="openshift-console/console-f9d7485db-9hcbc" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.867249 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/75d0c362-7aa3-4919-937d-8849f425466e-encryption-config\") pod \"apiserver-76f77b778f-6bpvf\" (UID: \"75d0c362-7aa3-4919-937d-8849f425466e\") " pod="openshift-apiserver/apiserver-76f77b778f-6bpvf" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.867502 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/75d0c362-7aa3-4919-937d-8849f425466e-audit\") pod \"apiserver-76f77b778f-6bpvf\" (UID: \"75d0c362-7aa3-4919-937d-8849f425466e\") " pod="openshift-apiserver/apiserver-76f77b778f-6bpvf" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.867801 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/75d0c362-7aa3-4919-937d-8849f425466e-trusted-ca-bundle\") pod \"apiserver-76f77b778f-6bpvf\" (UID: \"75d0c362-7aa3-4919-937d-8849f425466e\") " pod="openshift-apiserver/apiserver-76f77b778f-6bpvf" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.867885 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e4ea5d4f-3481-4e72-9bd8-e5d8183bc687-serving-cert\") pod \"apiserver-7bbb656c7d-stjw8\" (UID: \"e4ea5d4f-3481-4e72-9bd8-e5d8183bc687\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-stjw8" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.868088 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/770ae4fc-21dd-446d-9a41-26e484687787-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-6t2mt\" (UID: \"770ae4fc-21dd-446d-9a41-26e484687787\") " pod="openshift-authentication/oauth-openshift-558db77b4-6t2mt" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.868409 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/85542c44-c09c-4c39-851f-fd320a0a36b2-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-wcg6b\" (UID: \"85542c44-c09c-4c39-851f-fd320a0a36b2\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-wcg6b" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.868666 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/74f16025-fe11-4a23-8c1a-ff3da2a3d751-console-serving-cert\") pod \"console-f9d7485db-9hcbc\" (UID: \"74f16025-fe11-4a23-8c1a-ff3da2a3d751\") " pod="openshift-console/console-f9d7485db-9hcbc" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.868756 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/75d0c362-7aa3-4919-937d-8849f425466e-image-import-ca\") pod \"apiserver-76f77b778f-6bpvf\" (UID: \"75d0c362-7aa3-4919-937d-8849f425466e\") " pod="openshift-apiserver/apiserver-76f77b778f-6bpvf" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.869000 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/75d0c362-7aa3-4919-937d-8849f425466e-serving-cert\") pod \"apiserver-76f77b778f-6bpvf\" (UID: \"75d0c362-7aa3-4919-937d-8849f425466e\") " pod="openshift-apiserver/apiserver-76f77b778f-6bpvf" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.869074 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b1cd661a-c160-4e14-81cf-842d3b9ae01d-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-m467d\" (UID: \"b1cd661a-c160-4e14-81cf-842d3b9ae01d\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-m467d" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.869938 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/e4ea5d4f-3481-4e72-9bd8-e5d8183bc687-etcd-client\") pod \"apiserver-7bbb656c7d-stjw8\" (UID: \"e4ea5d4f-3481-4e72-9bd8-e5d8183bc687\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-stjw8" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.870018 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/770ae4fc-21dd-446d-9a41-26e484687787-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-6t2mt\" (UID: \"770ae4fc-21dd-446d-9a41-26e484687787\") " pod="openshift-authentication/oauth-openshift-558db77b4-6t2mt" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.873089 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/beec0ea6-0c57-44d6-bfef-d00e5b3724ad-serving-cert\") pod \"route-controller-manager-6576b87f9c-vwktr\" (UID: \"beec0ea6-0c57-44d6-bfef-d00e5b3724ad\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-vwktr" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.873318 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/770ae4fc-21dd-446d-9a41-26e484687787-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-6t2mt\" (UID: \"770ae4fc-21dd-446d-9a41-26e484687787\") " pod="openshift-authentication/oauth-openshift-558db77b4-6t2mt" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.873516 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/770ae4fc-21dd-446d-9a41-26e484687787-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-6t2mt\" (UID: \"770ae4fc-21dd-446d-9a41-26e484687787\") " pod="openshift-authentication/oauth-openshift-558db77b4-6t2mt" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.877246 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/74f16025-fe11-4a23-8c1a-ff3da2a3d751-console-oauth-config\") pod \"console-f9d7485db-9hcbc\" (UID: \"74f16025-fe11-4a23-8c1a-ff3da2a3d751\") " pod="openshift-console/console-f9d7485db-9hcbc" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.881694 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.901399 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.941441 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.955519 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/7b82343c-28e6-4e0b-9731-5131cb0014f2-trusted-ca\") pod \"ingress-operator-5b745b69d9-9jtcc\" (UID: \"7b82343c-28e6-4e0b-9731-5131cb0014f2\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-9jtcc" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.957753 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/7b82343c-28e6-4e0b-9731-5131cb0014f2-metrics-tls\") pod \"ingress-operator-5b745b69d9-9jtcc\" (UID: \"7b82343c-28e6-4e0b-9731-5131cb0014f2\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-9jtcc" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.957857 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hnbqg\" (UniqueName: \"kubernetes.io/projected/7b82343c-28e6-4e0b-9731-5131cb0014f2-kube-api-access-hnbqg\") pod \"ingress-operator-5b745b69d9-9jtcc\" (UID: \"7b82343c-28e6-4e0b-9731-5131cb0014f2\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-9jtcc" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.957964 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/7b82343c-28e6-4e0b-9731-5131cb0014f2-bound-sa-token\") pod \"ingress-operator-5b745b69d9-9jtcc\" (UID: \"7b82343c-28e6-4e0b-9731-5131cb0014f2\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-9jtcc" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.959575 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/7b82343c-28e6-4e0b-9731-5131cb0014f2-trusted-ca\") pod \"ingress-operator-5b745b69d9-9jtcc\" (UID: \"7b82343c-28e6-4e0b-9731-5131cb0014f2\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-9jtcc" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.961982 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.967932 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/7b82343c-28e6-4e0b-9731-5131cb0014f2-metrics-tls\") pod \"ingress-operator-5b745b69d9-9jtcc\" (UID: \"7b82343c-28e6-4e0b-9731-5131cb0014f2\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-9jtcc" Oct 05 06:50:17 crc kubenswrapper[4846]: I1005 06:50:17.983066 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Oct 05 06:50:18 crc kubenswrapper[4846]: I1005 06:50:18.002281 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Oct 05 06:50:18 crc kubenswrapper[4846]: I1005 06:50:18.023198 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Oct 05 06:50:18 crc kubenswrapper[4846]: I1005 06:50:18.041718 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Oct 05 06:50:18 crc kubenswrapper[4846]: I1005 06:50:18.080760 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Oct 05 06:50:18 crc kubenswrapper[4846]: I1005 06:50:18.088620 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Oct 05 06:50:18 crc kubenswrapper[4846]: I1005 06:50:18.112712 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Oct 05 06:50:18 crc kubenswrapper[4846]: I1005 06:50:18.123622 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Oct 05 06:50:18 crc kubenswrapper[4846]: I1005 06:50:18.144069 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Oct 05 06:50:18 crc kubenswrapper[4846]: I1005 06:50:18.161427 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Oct 05 06:50:18 crc kubenswrapper[4846]: I1005 06:50:18.182563 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Oct 05 06:50:18 crc kubenswrapper[4846]: I1005 06:50:18.201312 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Oct 05 06:50:18 crc kubenswrapper[4846]: I1005 06:50:18.222155 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Oct 05 06:50:18 crc kubenswrapper[4846]: I1005 06:50:18.262372 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Oct 05 06:50:18 crc kubenswrapper[4846]: I1005 06:50:18.281818 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Oct 05 06:50:18 crc kubenswrapper[4846]: I1005 06:50:18.302751 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Oct 05 06:50:18 crc kubenswrapper[4846]: I1005 06:50:18.322969 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Oct 05 06:50:18 crc kubenswrapper[4846]: I1005 06:50:18.342484 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Oct 05 06:50:18 crc kubenswrapper[4846]: I1005 06:50:18.363135 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Oct 05 06:50:18 crc kubenswrapper[4846]: I1005 06:50:18.383094 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Oct 05 06:50:18 crc kubenswrapper[4846]: I1005 06:50:18.403373 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Oct 05 06:50:18 crc kubenswrapper[4846]: I1005 06:50:18.421382 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Oct 05 06:50:18 crc kubenswrapper[4846]: I1005 06:50:18.441751 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Oct 05 06:50:18 crc kubenswrapper[4846]: I1005 06:50:18.462220 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Oct 05 06:50:18 crc kubenswrapper[4846]: I1005 06:50:18.485225 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Oct 05 06:50:18 crc kubenswrapper[4846]: I1005 06:50:18.502114 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Oct 05 06:50:18 crc kubenswrapper[4846]: I1005 06:50:18.522066 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Oct 05 06:50:18 crc kubenswrapper[4846]: I1005 06:50:18.542055 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Oct 05 06:50:18 crc kubenswrapper[4846]: I1005 06:50:18.561730 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Oct 05 06:50:18 crc kubenswrapper[4846]: I1005 06:50:18.582548 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Oct 05 06:50:18 crc kubenswrapper[4846]: I1005 06:50:18.604543 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Oct 05 06:50:18 crc kubenswrapper[4846]: I1005 06:50:18.623453 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Oct 05 06:50:18 crc kubenswrapper[4846]: I1005 06:50:18.646512 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Oct 05 06:50:18 crc kubenswrapper[4846]: I1005 06:50:18.662531 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Oct 05 06:50:18 crc kubenswrapper[4846]: I1005 06:50:18.682309 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Oct 05 06:50:18 crc kubenswrapper[4846]: I1005 06:50:18.702304 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Oct 05 06:50:18 crc kubenswrapper[4846]: I1005 06:50:18.722556 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Oct 05 06:50:18 crc kubenswrapper[4846]: I1005 06:50:18.742350 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Oct 05 06:50:18 crc kubenswrapper[4846]: I1005 06:50:18.760054 4846 request.go:700] Waited for 1.002445246s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-ingress/configmaps?fieldSelector=metadata.name%3Dopenshift-service-ca.crt&limit=500&resourceVersion=0 Oct 05 06:50:18 crc kubenswrapper[4846]: I1005 06:50:18.762027 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Oct 05 06:50:18 crc kubenswrapper[4846]: I1005 06:50:18.782240 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Oct 05 06:50:18 crc kubenswrapper[4846]: I1005 06:50:18.801729 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Oct 05 06:50:18 crc kubenswrapper[4846]: I1005 06:50:18.822477 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Oct 05 06:50:18 crc kubenswrapper[4846]: I1005 06:50:18.841476 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Oct 05 06:50:18 crc kubenswrapper[4846]: I1005 06:50:18.862407 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Oct 05 06:50:18 crc kubenswrapper[4846]: I1005 06:50:18.893434 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Oct 05 06:50:18 crc kubenswrapper[4846]: I1005 06:50:18.902451 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Oct 05 06:50:18 crc kubenswrapper[4846]: I1005 06:50:18.922444 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Oct 05 06:50:18 crc kubenswrapper[4846]: I1005 06:50:18.941925 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Oct 05 06:50:18 crc kubenswrapper[4846]: I1005 06:50:18.962386 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Oct 05 06:50:18 crc kubenswrapper[4846]: I1005 06:50:18.982323 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Oct 05 06:50:19 crc kubenswrapper[4846]: I1005 06:50:19.002066 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Oct 05 06:50:19 crc kubenswrapper[4846]: I1005 06:50:19.022063 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Oct 05 06:50:19 crc kubenswrapper[4846]: I1005 06:50:19.042928 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Oct 05 06:50:19 crc kubenswrapper[4846]: I1005 06:50:19.063457 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Oct 05 06:50:19 crc kubenswrapper[4846]: I1005 06:50:19.082716 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Oct 05 06:50:19 crc kubenswrapper[4846]: I1005 06:50:19.102804 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Oct 05 06:50:19 crc kubenswrapper[4846]: I1005 06:50:19.126424 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Oct 05 06:50:19 crc kubenswrapper[4846]: I1005 06:50:19.141108 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Oct 05 06:50:19 crc kubenswrapper[4846]: I1005 06:50:19.161410 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Oct 05 06:50:19 crc kubenswrapper[4846]: I1005 06:50:19.180905 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Oct 05 06:50:19 crc kubenswrapper[4846]: I1005 06:50:19.201762 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Oct 05 06:50:19 crc kubenswrapper[4846]: I1005 06:50:19.221558 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Oct 05 06:50:19 crc kubenswrapper[4846]: I1005 06:50:19.242121 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Oct 05 06:50:19 crc kubenswrapper[4846]: I1005 06:50:19.261816 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Oct 05 06:50:19 crc kubenswrapper[4846]: I1005 06:50:19.280640 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Oct 05 06:50:19 crc kubenswrapper[4846]: I1005 06:50:19.301800 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 05 06:50:19 crc kubenswrapper[4846]: I1005 06:50:19.322038 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 05 06:50:19 crc kubenswrapper[4846]: I1005 06:50:19.340839 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Oct 05 06:50:19 crc kubenswrapper[4846]: I1005 06:50:19.362130 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Oct 05 06:50:19 crc kubenswrapper[4846]: I1005 06:50:19.381496 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Oct 05 06:50:19 crc kubenswrapper[4846]: I1005 06:50:19.401773 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Oct 05 06:50:19 crc kubenswrapper[4846]: I1005 06:50:19.421653 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Oct 05 06:50:19 crc kubenswrapper[4846]: I1005 06:50:19.441688 4846 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Oct 05 06:50:19 crc kubenswrapper[4846]: I1005 06:50:19.461016 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Oct 05 06:50:19 crc kubenswrapper[4846]: I1005 06:50:19.481665 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Oct 05 06:50:19 crc kubenswrapper[4846]: I1005 06:50:19.502373 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Oct 05 06:50:19 crc kubenswrapper[4846]: I1005 06:50:19.521315 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Oct 05 06:50:19 crc kubenswrapper[4846]: I1005 06:50:19.540973 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Oct 05 06:50:19 crc kubenswrapper[4846]: I1005 06:50:19.561559 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Oct 05 06:50:19 crc kubenswrapper[4846]: I1005 06:50:19.582212 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Oct 05 06:50:19 crc kubenswrapper[4846]: I1005 06:50:19.619815 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n8cd4\" (UniqueName: \"kubernetes.io/projected/475e1ca4-a6fa-4f77-8c89-0619b2e9681e-kube-api-access-n8cd4\") pod \"downloads-7954f5f757-c6cxs\" (UID: \"475e1ca4-a6fa-4f77-8c89-0619b2e9681e\") " pod="openshift-console/downloads-7954f5f757-c6cxs" Oct 05 06:50:19 crc kubenswrapper[4846]: I1005 06:50:19.646258 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-49qp9\" (UniqueName: \"kubernetes.io/projected/53fe640c-6cb7-43a3-9a84-467eb36fe66b-kube-api-access-49qp9\") pod \"openshift-config-operator-7777fb866f-rcnlv\" (UID: \"53fe640c-6cb7-43a3-9a84-467eb36fe66b\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-rcnlv" Oct 05 06:50:19 crc kubenswrapper[4846]: I1005 06:50:19.657048 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/85542c44-c09c-4c39-851f-fd320a0a36b2-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-wcg6b\" (UID: \"85542c44-c09c-4c39-851f-fd320a0a36b2\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-wcg6b" Oct 05 06:50:19 crc kubenswrapper[4846]: I1005 06:50:19.688497 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xw5w5\" (UniqueName: \"kubernetes.io/projected/85542c44-c09c-4c39-851f-fd320a0a36b2-kube-api-access-xw5w5\") pod \"cluster-image-registry-operator-dc59b4c8b-wcg6b\" (UID: \"85542c44-c09c-4c39-851f-fd320a0a36b2\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-wcg6b" Oct 05 06:50:19 crc kubenswrapper[4846]: I1005 06:50:19.703872 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k8whk\" (UniqueName: \"kubernetes.io/projected/e4ea5d4f-3481-4e72-9bd8-e5d8183bc687-kube-api-access-k8whk\") pod \"apiserver-7bbb656c7d-stjw8\" (UID: \"e4ea5d4f-3481-4e72-9bd8-e5d8183bc687\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-stjw8" Oct 05 06:50:19 crc kubenswrapper[4846]: I1005 06:50:19.719911 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-drmkm\" (UniqueName: \"kubernetes.io/projected/770ae4fc-21dd-446d-9a41-26e484687787-kube-api-access-drmkm\") pod \"oauth-openshift-558db77b4-6t2mt\" (UID: \"770ae4fc-21dd-446d-9a41-26e484687787\") " pod="openshift-authentication/oauth-openshift-558db77b4-6t2mt" Oct 05 06:50:19 crc kubenswrapper[4846]: I1005 06:50:19.739396 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k8gmq\" (UniqueName: \"kubernetes.io/projected/b1cd661a-c160-4e14-81cf-842d3b9ae01d-kube-api-access-k8gmq\") pod \"openshift-apiserver-operator-796bbdcf4f-m467d\" (UID: \"b1cd661a-c160-4e14-81cf-842d3b9ae01d\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-m467d" Oct 05 06:50:19 crc kubenswrapper[4846]: I1005 06:50:19.752563 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-m467d" Oct 05 06:50:19 crc kubenswrapper[4846]: I1005 06:50:19.756439 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pmhpw\" (UniqueName: \"kubernetes.io/projected/beec0ea6-0c57-44d6-bfef-d00e5b3724ad-kube-api-access-pmhpw\") pod \"route-controller-manager-6576b87f9c-vwktr\" (UID: \"beec0ea6-0c57-44d6-bfef-d00e5b3724ad\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-vwktr" Oct 05 06:50:19 crc kubenswrapper[4846]: I1005 06:50:19.760344 4846 request.go:700] Waited for 1.89771839s due to client-side throttling, not priority and fairness, request: POST:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-apiserver/serviceaccounts/openshift-apiserver-sa/token Oct 05 06:50:19 crc kubenswrapper[4846]: I1005 06:50:19.767793 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-stjw8" Oct 05 06:50:19 crc kubenswrapper[4846]: I1005 06:50:19.779423 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n8bmr\" (UniqueName: \"kubernetes.io/projected/75d0c362-7aa3-4919-937d-8849f425466e-kube-api-access-n8bmr\") pod \"apiserver-76f77b778f-6bpvf\" (UID: \"75d0c362-7aa3-4919-937d-8849f425466e\") " pod="openshift-apiserver/apiserver-76f77b778f-6bpvf" Oct 05 06:50:19 crc kubenswrapper[4846]: I1005 06:50:19.800953 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r8vnl\" (UniqueName: \"kubernetes.io/projected/74f16025-fe11-4a23-8c1a-ff3da2a3d751-kube-api-access-r8vnl\") pod \"console-f9d7485db-9hcbc\" (UID: \"74f16025-fe11-4a23-8c1a-ff3da2a3d751\") " pod="openshift-console/console-f9d7485db-9hcbc" Oct 05 06:50:19 crc kubenswrapper[4846]: I1005 06:50:19.814548 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-vwktr" Oct 05 06:50:19 crc kubenswrapper[4846]: I1005 06:50:19.826985 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hdh6p\" (UniqueName: \"kubernetes.io/projected/5bddce7c-b835-4a1f-a3af-db5a1d3771f2-kube-api-access-hdh6p\") pod \"machine-approver-56656f9798-pglrh\" (UID: \"5bddce7c-b835-4a1f-a3af-db5a1d3771f2\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-pglrh" Oct 05 06:50:19 crc kubenswrapper[4846]: I1005 06:50:19.827227 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-rcnlv" Oct 05 06:50:19 crc kubenswrapper[4846]: I1005 06:50:19.838088 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-9hcbc" Oct 05 06:50:19 crc kubenswrapper[4846]: I1005 06:50:19.852501 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-6t2mt" Oct 05 06:50:19 crc kubenswrapper[4846]: I1005 06:50:19.853782 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v5gbg\" (UniqueName: \"kubernetes.io/projected/f743f03b-3f08-4da0-b53e-1cb33ab516b9-kube-api-access-v5gbg\") pod \"machine-api-operator-5694c8668f-vwx77\" (UID: \"f743f03b-3f08-4da0-b53e-1cb33ab516b9\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-vwx77" Oct 05 06:50:19 crc kubenswrapper[4846]: I1005 06:50:19.865999 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-wcg6b" Oct 05 06:50:19 crc kubenswrapper[4846]: I1005 06:50:19.886315 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/7b82343c-28e6-4e0b-9731-5131cb0014f2-bound-sa-token\") pod \"ingress-operator-5b745b69d9-9jtcc\" (UID: \"7b82343c-28e6-4e0b-9731-5131cb0014f2\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-9jtcc" Oct 05 06:50:19 crc kubenswrapper[4846]: I1005 06:50:19.895880 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-c6cxs" Oct 05 06:50:19 crc kubenswrapper[4846]: I1005 06:50:19.902670 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hnbqg\" (UniqueName: \"kubernetes.io/projected/7b82343c-28e6-4e0b-9731-5131cb0014f2-kube-api-access-hnbqg\") pod \"ingress-operator-5b745b69d9-9jtcc\" (UID: \"7b82343c-28e6-4e0b-9731-5131cb0014f2\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-9jtcc" Oct 05 06:50:19 crc kubenswrapper[4846]: I1005 06:50:19.940158 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-9jtcc" Oct 05 06:50:19 crc kubenswrapper[4846]: I1005 06:50:19.990856 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7a19cb1c-5274-48e7-ad86-3c2ff85ecf8c-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-7jflr\" (UID: \"7a19cb1c-5274-48e7-ad86-3c2ff85ecf8c\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-7jflr" Oct 05 06:50:19 crc kubenswrapper[4846]: I1005 06:50:19.990927 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/18507a98-2ef8-43b1-b94d-51106f9bdd3b-trusted-ca\") pod \"console-operator-58897d9998-t8f4h\" (UID: \"18507a98-2ef8-43b1-b94d-51106f9bdd3b\") " pod="openshift-console-operator/console-operator-58897d9998-t8f4h" Oct 05 06:50:19 crc kubenswrapper[4846]: I1005 06:50:19.990945 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cb184e9f-f0d9-4166-b0da-213c5c7aba5f-serving-cert\") pod \"etcd-operator-b45778765-8vjkd\" (UID: \"cb184e9f-f0d9-4166-b0da-213c5c7aba5f\") " pod="openshift-etcd-operator/etcd-operator-b45778765-8vjkd" Oct 05 06:50:19 crc kubenswrapper[4846]: I1005 06:50:19.990962 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6be3f74a-b215-4588-8b0f-735eba580bbd-serving-cert\") pod \"controller-manager-879f6c89f-vjmbx\" (UID: \"6be3f74a-b215-4588-8b0f-735eba580bbd\") " pod="openshift-controller-manager/controller-manager-879f6c89f-vjmbx" Oct 05 06:50:19 crc kubenswrapper[4846]: I1005 06:50:19.991003 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/53168eca-95fe-4303-b401-380995e5adcb-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-526w4\" (UID: \"53168eca-95fe-4303-b401-380995e5adcb\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-526w4" Oct 05 06:50:19 crc kubenswrapper[4846]: I1005 06:50:19.991071 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/18507a98-2ef8-43b1-b94d-51106f9bdd3b-config\") pod \"console-operator-58897d9998-t8f4h\" (UID: \"18507a98-2ef8-43b1-b94d-51106f9bdd3b\") " pod="openshift-console-operator/console-operator-58897d9998-t8f4h" Oct 05 06:50:19 crc kubenswrapper[4846]: I1005 06:50:19.991122 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6ec3075e-f780-435f-8efa-30d2fb4855a1-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-2pd47\" (UID: \"6ec3075e-f780-435f-8efa-30d2fb4855a1\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-2pd47" Oct 05 06:50:19 crc kubenswrapper[4846]: I1005 06:50:19.991139 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7a19cb1c-5274-48e7-ad86-3c2ff85ecf8c-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-7jflr\" (UID: \"7a19cb1c-5274-48e7-ad86-3c2ff85ecf8c\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-7jflr" Oct 05 06:50:19 crc kubenswrapper[4846]: I1005 06:50:19.991158 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gbd7g\" (UniqueName: \"kubernetes.io/projected/6be3f74a-b215-4588-8b0f-735eba580bbd-kube-api-access-gbd7g\") pod \"controller-manager-879f6c89f-vjmbx\" (UID: \"6be3f74a-b215-4588-8b0f-735eba580bbd\") " pod="openshift-controller-manager/controller-manager-879f6c89f-vjmbx" Oct 05 06:50:19 crc kubenswrapper[4846]: I1005 06:50:19.991220 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/bc560c3d-6558-43e6-86ca-312c9fa264ea-metrics-tls\") pod \"dns-operator-744455d44c-5qwb9\" (UID: \"bc560c3d-6558-43e6-86ca-312c9fa264ea\") " pod="openshift-dns-operator/dns-operator-744455d44c-5qwb9" Oct 05 06:50:19 crc kubenswrapper[4846]: I1005 06:50:19.991275 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/53168eca-95fe-4303-b401-380995e5adcb-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-526w4\" (UID: \"53168eca-95fe-4303-b401-380995e5adcb\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-526w4" Oct 05 06:50:19 crc kubenswrapper[4846]: I1005 06:50:19.991301 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/78d3e7b1-7258-4508-ad38-373f21bb5d96-registry-certificates\") pod \"image-registry-697d97f7c8-rf9l4\" (UID: \"78d3e7b1-7258-4508-ad38-373f21bb5d96\") " pod="openshift-image-registry/image-registry-697d97f7c8-rf9l4" Oct 05 06:50:19 crc kubenswrapper[4846]: I1005 06:50:19.991354 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9tjtl\" (UniqueName: \"kubernetes.io/projected/54f93640-64fd-4728-a83b-499aee1ffa38-kube-api-access-9tjtl\") pod \"migrator-59844c95c7-skh8h\" (UID: \"54f93640-64fd-4728-a83b-499aee1ffa38\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-skh8h" Oct 05 06:50:19 crc kubenswrapper[4846]: I1005 06:50:19.991370 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6ec3075e-f780-435f-8efa-30d2fb4855a1-config\") pod \"authentication-operator-69f744f599-2pd47\" (UID: \"6ec3075e-f780-435f-8efa-30d2fb4855a1\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-2pd47" Oct 05 06:50:19 crc kubenswrapper[4846]: I1005 06:50:19.991385 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sgkh2\" (UniqueName: \"kubernetes.io/projected/6ec3075e-f780-435f-8efa-30d2fb4855a1-kube-api-access-sgkh2\") pod \"authentication-operator-69f744f599-2pd47\" (UID: \"6ec3075e-f780-435f-8efa-30d2fb4855a1\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-2pd47" Oct 05 06:50:19 crc kubenswrapper[4846]: I1005 06:50:19.991435 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/cb184e9f-f0d9-4166-b0da-213c5c7aba5f-etcd-ca\") pod \"etcd-operator-b45778765-8vjkd\" (UID: \"cb184e9f-f0d9-4166-b0da-213c5c7aba5f\") " pod="openshift-etcd-operator/etcd-operator-b45778765-8vjkd" Oct 05 06:50:19 crc kubenswrapper[4846]: I1005 06:50:19.991458 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/cb184e9f-f0d9-4166-b0da-213c5c7aba5f-etcd-service-ca\") pod \"etcd-operator-b45778765-8vjkd\" (UID: \"cb184e9f-f0d9-4166-b0da-213c5c7aba5f\") " pod="openshift-etcd-operator/etcd-operator-b45778765-8vjkd" Oct 05 06:50:19 crc kubenswrapper[4846]: I1005 06:50:19.991478 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9vncv\" (UniqueName: \"kubernetes.io/projected/bc560c3d-6558-43e6-86ca-312c9fa264ea-kube-api-access-9vncv\") pod \"dns-operator-744455d44c-5qwb9\" (UID: \"bc560c3d-6558-43e6-86ca-312c9fa264ea\") " pod="openshift-dns-operator/dns-operator-744455d44c-5qwb9" Oct 05 06:50:19 crc kubenswrapper[4846]: I1005 06:50:19.991565 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/6be3f74a-b215-4588-8b0f-735eba580bbd-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-vjmbx\" (UID: \"6be3f74a-b215-4588-8b0f-735eba580bbd\") " pod="openshift-controller-manager/controller-manager-879f6c89f-vjmbx" Oct 05 06:50:19 crc kubenswrapper[4846]: I1005 06:50:19.991618 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/78d3e7b1-7258-4508-ad38-373f21bb5d96-installation-pull-secrets\") pod \"image-registry-697d97f7c8-rf9l4\" (UID: \"78d3e7b1-7258-4508-ad38-373f21bb5d96\") " pod="openshift-image-registry/image-registry-697d97f7c8-rf9l4" Oct 05 06:50:19 crc kubenswrapper[4846]: I1005 06:50:19.991665 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/78d3e7b1-7258-4508-ad38-373f21bb5d96-ca-trust-extracted\") pod \"image-registry-697d97f7c8-rf9l4\" (UID: \"78d3e7b1-7258-4508-ad38-373f21bb5d96\") " pod="openshift-image-registry/image-registry-697d97f7c8-rf9l4" Oct 05 06:50:19 crc kubenswrapper[4846]: I1005 06:50:19.991736 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rf9l4\" (UID: \"78d3e7b1-7258-4508-ad38-373f21bb5d96\") " pod="openshift-image-registry/image-registry-697d97f7c8-rf9l4" Oct 05 06:50:19 crc kubenswrapper[4846]: I1005 06:50:19.991760 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/78d3e7b1-7258-4508-ad38-373f21bb5d96-registry-tls\") pod \"image-registry-697d97f7c8-rf9l4\" (UID: \"78d3e7b1-7258-4508-ad38-373f21bb5d96\") " pod="openshift-image-registry/image-registry-697d97f7c8-rf9l4" Oct 05 06:50:19 crc kubenswrapper[4846]: I1005 06:50:19.991775 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/78d3e7b1-7258-4508-ad38-373f21bb5d96-trusted-ca\") pod \"image-registry-697d97f7c8-rf9l4\" (UID: \"78d3e7b1-7258-4508-ad38-373f21bb5d96\") " pod="openshift-image-registry/image-registry-697d97f7c8-rf9l4" Oct 05 06:50:19 crc kubenswrapper[4846]: I1005 06:50:19.991790 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6ec3075e-f780-435f-8efa-30d2fb4855a1-service-ca-bundle\") pod \"authentication-operator-69f744f599-2pd47\" (UID: \"6ec3075e-f780-435f-8efa-30d2fb4855a1\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-2pd47" Oct 05 06:50:19 crc kubenswrapper[4846]: I1005 06:50:19.991827 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/cb184e9f-f0d9-4166-b0da-213c5c7aba5f-etcd-client\") pod \"etcd-operator-b45778765-8vjkd\" (UID: \"cb184e9f-f0d9-4166-b0da-213c5c7aba5f\") " pod="openshift-etcd-operator/etcd-operator-b45778765-8vjkd" Oct 05 06:50:19 crc kubenswrapper[4846]: I1005 06:50:19.991843 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mtl7j\" (UniqueName: \"kubernetes.io/projected/78d3e7b1-7258-4508-ad38-373f21bb5d96-kube-api-access-mtl7j\") pod \"image-registry-697d97f7c8-rf9l4\" (UID: \"78d3e7b1-7258-4508-ad38-373f21bb5d96\") " pod="openshift-image-registry/image-registry-697d97f7c8-rf9l4" Oct 05 06:50:19 crc kubenswrapper[4846]: I1005 06:50:19.991858 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cb184e9f-f0d9-4166-b0da-213c5c7aba5f-config\") pod \"etcd-operator-b45778765-8vjkd\" (UID: \"cb184e9f-f0d9-4166-b0da-213c5c7aba5f\") " pod="openshift-etcd-operator/etcd-operator-b45778765-8vjkd" Oct 05 06:50:19 crc kubenswrapper[4846]: I1005 06:50:19.991874 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/6be3f74a-b215-4588-8b0f-735eba580bbd-client-ca\") pod \"controller-manager-879f6c89f-vjmbx\" (UID: \"6be3f74a-b215-4588-8b0f-735eba580bbd\") " pod="openshift-controller-manager/controller-manager-879f6c89f-vjmbx" Oct 05 06:50:19 crc kubenswrapper[4846]: I1005 06:50:19.991899 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7a19cb1c-5274-48e7-ad86-3c2ff85ecf8c-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-7jflr\" (UID: \"7a19cb1c-5274-48e7-ad86-3c2ff85ecf8c\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-7jflr" Oct 05 06:50:19 crc kubenswrapper[4846]: I1005 06:50:19.991922 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b55x6\" (UniqueName: \"kubernetes.io/projected/cb184e9f-f0d9-4166-b0da-213c5c7aba5f-kube-api-access-b55x6\") pod \"etcd-operator-b45778765-8vjkd\" (UID: \"cb184e9f-f0d9-4166-b0da-213c5c7aba5f\") " pod="openshift-etcd-operator/etcd-operator-b45778765-8vjkd" Oct 05 06:50:19 crc kubenswrapper[4846]: I1005 06:50:19.991913 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-6bpvf" Oct 05 06:50:19 crc kubenswrapper[4846]: I1005 06:50:19.991946 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/78d3e7b1-7258-4508-ad38-373f21bb5d96-bound-sa-token\") pod \"image-registry-697d97f7c8-rf9l4\" (UID: \"78d3e7b1-7258-4508-ad38-373f21bb5d96\") " pod="openshift-image-registry/image-registry-697d97f7c8-rf9l4" Oct 05 06:50:19 crc kubenswrapper[4846]: I1005 06:50:19.991964 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/da8eba09-f287-4704-b990-348ff69f0173-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-k87kc\" (UID: \"da8eba09-f287-4704-b990-348ff69f0173\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-k87kc" Oct 05 06:50:19 crc kubenswrapper[4846]: I1005 06:50:19.991979 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vw2sb\" (UniqueName: \"kubernetes.io/projected/18507a98-2ef8-43b1-b94d-51106f9bdd3b-kube-api-access-vw2sb\") pod \"console-operator-58897d9998-t8f4h\" (UID: \"18507a98-2ef8-43b1-b94d-51106f9bdd3b\") " pod="openshift-console-operator/console-operator-58897d9998-t8f4h" Oct 05 06:50:19 crc kubenswrapper[4846]: I1005 06:50:19.992029 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6gnmv\" (UniqueName: \"kubernetes.io/projected/53168eca-95fe-4303-b401-380995e5adcb-kube-api-access-6gnmv\") pod \"openshift-controller-manager-operator-756b6f6bc6-526w4\" (UID: \"53168eca-95fe-4303-b401-380995e5adcb\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-526w4" Oct 05 06:50:19 crc kubenswrapper[4846]: I1005 06:50:19.992048 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6ec3075e-f780-435f-8efa-30d2fb4855a1-serving-cert\") pod \"authentication-operator-69f744f599-2pd47\" (UID: \"6ec3075e-f780-435f-8efa-30d2fb4855a1\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-2pd47" Oct 05 06:50:19 crc kubenswrapper[4846]: I1005 06:50:19.992092 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2tvs4\" (UniqueName: \"kubernetes.io/projected/da8eba09-f287-4704-b990-348ff69f0173-kube-api-access-2tvs4\") pod \"cluster-samples-operator-665b6dd947-k87kc\" (UID: \"da8eba09-f287-4704-b990-348ff69f0173\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-k87kc" Oct 05 06:50:19 crc kubenswrapper[4846]: I1005 06:50:19.992108 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6be3f74a-b215-4588-8b0f-735eba580bbd-config\") pod \"controller-manager-879f6c89f-vjmbx\" (UID: \"6be3f74a-b215-4588-8b0f-735eba580bbd\") " pod="openshift-controller-manager/controller-manager-879f6c89f-vjmbx" Oct 05 06:50:19 crc kubenswrapper[4846]: I1005 06:50:19.992143 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/18507a98-2ef8-43b1-b94d-51106f9bdd3b-serving-cert\") pod \"console-operator-58897d9998-t8f4h\" (UID: \"18507a98-2ef8-43b1-b94d-51106f9bdd3b\") " pod="openshift-console-operator/console-operator-58897d9998-t8f4h" Oct 05 06:50:19 crc kubenswrapper[4846]: E1005 06:50:19.994132 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-05 06:50:20.494118727 +0000 UTC m=+142.734971492 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rf9l4" (UID: "78d3e7b1-7258-4508-ad38-373f21bb5d96") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.018971 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-vwx77" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.089250 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-pglrh" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.094799 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.095043 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2tvs4\" (UniqueName: \"kubernetes.io/projected/da8eba09-f287-4704-b990-348ff69f0173-kube-api-access-2tvs4\") pod \"cluster-samples-operator-665b6dd947-k87kc\" (UID: \"da8eba09-f287-4704-b990-348ff69f0173\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-k87kc" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.095070 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6be3f74a-b215-4588-8b0f-735eba580bbd-config\") pod \"controller-manager-879f6c89f-vjmbx\" (UID: \"6be3f74a-b215-4588-8b0f-735eba580bbd\") " pod="openshift-controller-manager/controller-manager-879f6c89f-vjmbx" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.095101 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/beeea3e9-d3c3-4280-bf04-85cee4378cb0-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-8bpqg\" (UID: \"beeea3e9-d3c3-4280-bf04-85cee4378cb0\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-8bpqg" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.095116 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fv4fz\" (UniqueName: \"kubernetes.io/projected/788472f5-2ad1-49d0-8e31-6a4fe14c3c85-kube-api-access-fv4fz\") pod \"machine-config-operator-74547568cd-wnmjw\" (UID: \"788472f5-2ad1-49d0-8e31-6a4fe14c3c85\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-wnmjw" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.095132 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9dm98\" (UniqueName: \"kubernetes.io/projected/74822779-bdcd-4aae-bc01-9cbcbb0c1d83-kube-api-access-9dm98\") pod \"kube-storage-version-migrator-operator-b67b599dd-x5n4h\" (UID: \"74822779-bdcd-4aae-bc01-9cbcbb0c1d83\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-x5n4h" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.095160 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cb184e9f-f0d9-4166-b0da-213c5c7aba5f-serving-cert\") pod \"etcd-operator-b45778765-8vjkd\" (UID: \"cb184e9f-f0d9-4166-b0da-213c5c7aba5f\") " pod="openshift-etcd-operator/etcd-operator-b45778765-8vjkd" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.095193 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/53168eca-95fe-4303-b401-380995e5adcb-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-526w4\" (UID: \"53168eca-95fe-4303-b401-380995e5adcb\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-526w4" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.095212 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/c041548e-b7c2-4f61-aaea-0b76e298ff22-apiservice-cert\") pod \"packageserver-d55dfcdfc-4pvwg\" (UID: \"c041548e-b7c2-4f61-aaea-0b76e298ff22\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-4pvwg" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.095230 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sl5t6\" (UniqueName: \"kubernetes.io/projected/0cf52841-edf1-4daf-8e2c-76f5f6b93104-kube-api-access-sl5t6\") pod \"olm-operator-6b444d44fb-zf4p8\" (UID: \"0cf52841-edf1-4daf-8e2c-76f5f6b93104\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-zf4p8" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.095248 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/80c88b0b-c4cf-478e-8e05-4cde023d15ee-node-bootstrap-token\") pod \"machine-config-server-2g2rj\" (UID: \"80c88b0b-c4cf-478e-8e05-4cde023d15ee\") " pod="openshift-machine-config-operator/machine-config-server-2g2rj" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.095274 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/8c2f8349-aa93-4ac8-a369-5b4b2b9cd6e2-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-2rc5n\" (UID: \"8c2f8349-aa93-4ac8-a369-5b4b2b9cd6e2\") " pod="openshift-marketplace/marketplace-operator-79b997595-2rc5n" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.095292 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/18507a98-2ef8-43b1-b94d-51106f9bdd3b-config\") pod \"console-operator-58897d9998-t8f4h\" (UID: \"18507a98-2ef8-43b1-b94d-51106f9bdd3b\") " pod="openshift-console-operator/console-operator-58897d9998-t8f4h" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.095308 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mxbkd\" (UniqueName: \"kubernetes.io/projected/2ecfaba9-9034-4ac7-bd51-e3f64b41d825-kube-api-access-mxbkd\") pod \"router-default-5444994796-58hwl\" (UID: \"2ecfaba9-9034-4ac7-bd51-e3f64b41d825\") " pod="openshift-ingress/router-default-5444994796-58hwl" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.095334 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ppwzm\" (UniqueName: \"kubernetes.io/projected/31d92551-cb79-4d9c-ad1c-c5090142b82c-kube-api-access-ppwzm\") pod \"service-ca-9c57cc56f-bpbtf\" (UID: \"31d92551-cb79-4d9c-ad1c-c5090142b82c\") " pod="openshift-service-ca/service-ca-9c57cc56f-bpbtf" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.095351 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4d7ch\" (UniqueName: \"kubernetes.io/projected/80c88b0b-c4cf-478e-8e05-4cde023d15ee-kube-api-access-4d7ch\") pod \"machine-config-server-2g2rj\" (UID: \"80c88b0b-c4cf-478e-8e05-4cde023d15ee\") " pod="openshift-machine-config-operator/machine-config-server-2g2rj" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.095366 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8b7fb5f2-4b5e-4fc3-9c97-dfbe36646871-serving-cert\") pod \"service-ca-operator-777779d784-5rj2f\" (UID: \"8b7fb5f2-4b5e-4fc3-9c97-dfbe36646871\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-5rj2f" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.095382 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gbd7g\" (UniqueName: \"kubernetes.io/projected/6be3f74a-b215-4588-8b0f-735eba580bbd-kube-api-access-gbd7g\") pod \"controller-manager-879f6c89f-vjmbx\" (UID: \"6be3f74a-b215-4588-8b0f-735eba580bbd\") " pod="openshift-controller-manager/controller-manager-879f6c89f-vjmbx" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.095400 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2ecfaba9-9034-4ac7-bd51-e3f64b41d825-service-ca-bundle\") pod \"router-default-5444994796-58hwl\" (UID: \"2ecfaba9-9034-4ac7-bd51-e3f64b41d825\") " pod="openshift-ingress/router-default-5444994796-58hwl" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.095421 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/c65d29d9-66e3-41ce-acb2-cec28c52fb29-metrics-tls\") pod \"dns-default-jm4gb\" (UID: \"c65d29d9-66e3-41ce-acb2-cec28c52fb29\") " pod="openshift-dns/dns-default-jm4gb" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.095441 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wth4v\" (UniqueName: \"kubernetes.io/projected/e532fab3-55a6-4c42-a532-9291452d5d82-kube-api-access-wth4v\") pod \"ingress-canary-4cxsw\" (UID: \"e532fab3-55a6-4c42-a532-9291452d5d82\") " pod="openshift-ingress-canary/ingress-canary-4cxsw" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.095465 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/bc560c3d-6558-43e6-86ca-312c9fa264ea-metrics-tls\") pod \"dns-operator-744455d44c-5qwb9\" (UID: \"bc560c3d-6558-43e6-86ca-312c9fa264ea\") " pod="openshift-dns-operator/dns-operator-744455d44c-5qwb9" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.095487 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/57e551f8-d9da-4692-b8ae-f7ff6829fe25-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-2x9nl\" (UID: \"57e551f8-d9da-4692-b8ae-f7ff6829fe25\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-2x9nl" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.095508 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/94a883ac-bee2-403e-8747-a4045e7defea-config\") pod \"kube-controller-manager-operator-78b949d7b-gqn9s\" (UID: \"94a883ac-bee2-403e-8747-a4045e7defea\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-gqn9s" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.095528 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s8g9n\" (UniqueName: \"kubernetes.io/projected/0755bb69-c94a-4b36-97ee-8a3dfd972c79-kube-api-access-s8g9n\") pod \"multus-admission-controller-857f4d67dd-dtbh5\" (UID: \"0755bb69-c94a-4b36-97ee-8a3dfd972c79\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-dtbh5" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.095554 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/2ecfaba9-9034-4ac7-bd51-e3f64b41d825-default-certificate\") pod \"router-default-5444994796-58hwl\" (UID: \"2ecfaba9-9034-4ac7-bd51-e3f64b41d825\") " pod="openshift-ingress/router-default-5444994796-58hwl" Oct 05 06:50:20 crc kubenswrapper[4846]: E1005 06:50:20.095595 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-05 06:50:20.595552639 +0000 UTC m=+142.836405414 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.095632 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9mht8\" (UniqueName: \"kubernetes.io/projected/b21ccaf7-40a5-4939-ad8b-4c7efdcf4306-kube-api-access-9mht8\") pod \"machine-config-controller-84d6567774-m7gph\" (UID: \"b21ccaf7-40a5-4939-ad8b-4c7efdcf4306\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-m7gph" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.095662 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/788472f5-2ad1-49d0-8e31-6a4fe14c3c85-auth-proxy-config\") pod \"machine-config-operator-74547568cd-wnmjw\" (UID: \"788472f5-2ad1-49d0-8e31-6a4fe14c3c85\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-wnmjw" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.095682 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/beeea3e9-d3c3-4280-bf04-85cee4378cb0-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-8bpqg\" (UID: \"beeea3e9-d3c3-4280-bf04-85cee4378cb0\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-8bpqg" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.095701 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7ml2d\" (UniqueName: \"kubernetes.io/projected/8b7fb5f2-4b5e-4fc3-9c97-dfbe36646871-kube-api-access-7ml2d\") pod \"service-ca-operator-777779d784-5rj2f\" (UID: \"8b7fb5f2-4b5e-4fc3-9c97-dfbe36646871\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-5rj2f" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.095722 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/31d92551-cb79-4d9c-ad1c-c5090142b82c-signing-key\") pod \"service-ca-9c57cc56f-bpbtf\" (UID: \"31d92551-cb79-4d9c-ad1c-c5090142b82c\") " pod="openshift-service-ca/service-ca-9c57cc56f-bpbtf" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.095768 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6ec3075e-f780-435f-8efa-30d2fb4855a1-config\") pod \"authentication-operator-69f744f599-2pd47\" (UID: \"6ec3075e-f780-435f-8efa-30d2fb4855a1\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-2pd47" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.095791 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sgkh2\" (UniqueName: \"kubernetes.io/projected/6ec3075e-f780-435f-8efa-30d2fb4855a1-kube-api-access-sgkh2\") pod \"authentication-operator-69f744f599-2pd47\" (UID: \"6ec3075e-f780-435f-8efa-30d2fb4855a1\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-2pd47" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.095809 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6srxd\" (UniqueName: \"kubernetes.io/projected/c041548e-b7c2-4f61-aaea-0b76e298ff22-kube-api-access-6srxd\") pod \"packageserver-d55dfcdfc-4pvwg\" (UID: \"c041548e-b7c2-4f61-aaea-0b76e298ff22\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-4pvwg" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.095830 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/cb184e9f-f0d9-4166-b0da-213c5c7aba5f-etcd-ca\") pod \"etcd-operator-b45778765-8vjkd\" (UID: \"cb184e9f-f0d9-4166-b0da-213c5c7aba5f\") " pod="openshift-etcd-operator/etcd-operator-b45778765-8vjkd" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.095849 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/cb184e9f-f0d9-4166-b0da-213c5c7aba5f-etcd-service-ca\") pod \"etcd-operator-b45778765-8vjkd\" (UID: \"cb184e9f-f0d9-4166-b0da-213c5c7aba5f\") " pod="openshift-etcd-operator/etcd-operator-b45778765-8vjkd" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.096590 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/6be3f74a-b215-4588-8b0f-735eba580bbd-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-vjmbx\" (UID: \"6be3f74a-b215-4588-8b0f-735eba580bbd\") " pod="openshift-controller-manager/controller-manager-879f6c89f-vjmbx" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.103645 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/18507a98-2ef8-43b1-b94d-51106f9bdd3b-config\") pod \"console-operator-58897d9998-t8f4h\" (UID: \"18507a98-2ef8-43b1-b94d-51106f9bdd3b\") " pod="openshift-console-operator/console-operator-58897d9998-t8f4h" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.104950 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/0cf52841-edf1-4daf-8e2c-76f5f6b93104-profile-collector-cert\") pod \"olm-operator-6b444d44fb-zf4p8\" (UID: \"0cf52841-edf1-4daf-8e2c-76f5f6b93104\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-zf4p8" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.105048 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/94a883ac-bee2-403e-8747-a4045e7defea-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-gqn9s\" (UID: \"94a883ac-bee2-403e-8747-a4045e7defea\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-gqn9s" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.105079 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/b21ccaf7-40a5-4939-ad8b-4c7efdcf4306-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-m7gph\" (UID: \"b21ccaf7-40a5-4939-ad8b-4c7efdcf4306\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-m7gph" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.106379 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6be3f74a-b215-4588-8b0f-735eba580bbd-config\") pod \"controller-manager-879f6c89f-vjmbx\" (UID: \"6be3f74a-b215-4588-8b0f-735eba580bbd\") " pod="openshift-controller-manager/controller-manager-879f6c89f-vjmbx" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.107560 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/6be3f74a-b215-4588-8b0f-735eba580bbd-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-vjmbx\" (UID: \"6be3f74a-b215-4588-8b0f-735eba580bbd\") " pod="openshift-controller-manager/controller-manager-879f6c89f-vjmbx" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.107835 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cb184e9f-f0d9-4166-b0da-213c5c7aba5f-serving-cert\") pod \"etcd-operator-b45778765-8vjkd\" (UID: \"cb184e9f-f0d9-4166-b0da-213c5c7aba5f\") " pod="openshift-etcd-operator/etcd-operator-b45778765-8vjkd" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.108110 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/c041548e-b7c2-4f61-aaea-0b76e298ff22-webhook-cert\") pod \"packageserver-d55dfcdfc-4pvwg\" (UID: \"c041548e-b7c2-4f61-aaea-0b76e298ff22\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-4pvwg" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.108347 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/78d3e7b1-7258-4508-ad38-373f21bb5d96-ca-trust-extracted\") pod \"image-registry-697d97f7c8-rf9l4\" (UID: \"78d3e7b1-7258-4508-ad38-373f21bb5d96\") " pod="openshift-image-registry/image-registry-697d97f7c8-rf9l4" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.108397 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/2ecfaba9-9034-4ac7-bd51-e3f64b41d825-metrics-certs\") pod \"router-default-5444994796-58hwl\" (UID: \"2ecfaba9-9034-4ac7-bd51-e3f64b41d825\") " pod="openshift-ingress/router-default-5444994796-58hwl" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.108700 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/78d3e7b1-7258-4508-ad38-373f21bb5d96-ca-trust-extracted\") pod \"image-registry-697d97f7c8-rf9l4\" (UID: \"78d3e7b1-7258-4508-ad38-373f21bb5d96\") " pod="openshift-image-registry/image-registry-697d97f7c8-rf9l4" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.108795 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/78d3e7b1-7258-4508-ad38-373f21bb5d96-trusted-ca\") pod \"image-registry-697d97f7c8-rf9l4\" (UID: \"78d3e7b1-7258-4508-ad38-373f21bb5d96\") " pod="openshift-image-registry/image-registry-697d97f7c8-rf9l4" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.108822 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/74822779-bdcd-4aae-bc01-9cbcbb0c1d83-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-x5n4h\" (UID: \"74822779-bdcd-4aae-bc01-9cbcbb0c1d83\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-x5n4h" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.108870 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9h8b2\" (UniqueName: \"kubernetes.io/projected/57e551f8-d9da-4692-b8ae-f7ff6829fe25-kube-api-access-9h8b2\") pod \"control-plane-machine-set-operator-78cbb6b69f-2x9nl\" (UID: \"57e551f8-d9da-4692-b8ae-f7ff6829fe25\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-2x9nl" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.108892 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/788472f5-2ad1-49d0-8e31-6a4fe14c3c85-proxy-tls\") pod \"machine-config-operator-74547568cd-wnmjw\" (UID: \"788472f5-2ad1-49d0-8e31-6a4fe14c3c85\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-wnmjw" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.108900 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6ec3075e-f780-435f-8efa-30d2fb4855a1-config\") pod \"authentication-operator-69f744f599-2pd47\" (UID: \"6ec3075e-f780-435f-8efa-30d2fb4855a1\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-2pd47" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.108928 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/cb184e9f-f0d9-4166-b0da-213c5c7aba5f-etcd-client\") pod \"etcd-operator-b45778765-8vjkd\" (UID: \"cb184e9f-f0d9-4166-b0da-213c5c7aba5f\") " pod="openshift-etcd-operator/etcd-operator-b45778765-8vjkd" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.108955 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ghg5r\" (UniqueName: \"kubernetes.io/projected/8049dd16-5d6f-4718-94b1-fb5a322c2530-kube-api-access-ghg5r\") pod \"package-server-manager-789f6589d5-pqm4r\" (UID: \"8049dd16-5d6f-4718-94b1-fb5a322c2530\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-pqm4r" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.108976 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/80ceab29-9389-4191-a909-8a6c40bba1cd-srv-cert\") pod \"catalog-operator-68c6474976-hmzsn\" (UID: \"80ceab29-9389-4191-a909-8a6c40bba1cd\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-hmzsn" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.111277 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/cb184e9f-f0d9-4166-b0da-213c5c7aba5f-etcd-service-ca\") pod \"etcd-operator-b45778765-8vjkd\" (UID: \"cb184e9f-f0d9-4166-b0da-213c5c7aba5f\") " pod="openshift-etcd-operator/etcd-operator-b45778765-8vjkd" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.111844 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cb184e9f-f0d9-4166-b0da-213c5c7aba5f-config\") pod \"etcd-operator-b45778765-8vjkd\" (UID: \"cb184e9f-f0d9-4166-b0da-213c5c7aba5f\") " pod="openshift-etcd-operator/etcd-operator-b45778765-8vjkd" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.113593 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/6be3f74a-b215-4588-8b0f-735eba580bbd-client-ca\") pod \"controller-manager-879f6c89f-vjmbx\" (UID: \"6be3f74a-b215-4588-8b0f-735eba580bbd\") " pod="openshift-controller-manager/controller-manager-879f6c89f-vjmbx" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.113719 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/80c88b0b-c4cf-478e-8e05-4cde023d15ee-certs\") pod \"machine-config-server-2g2rj\" (UID: \"80c88b0b-c4cf-478e-8e05-4cde023d15ee\") " pod="openshift-machine-config-operator/machine-config-server-2g2rj" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.113989 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/c041548e-b7c2-4f61-aaea-0b76e298ff22-tmpfs\") pod \"packageserver-d55dfcdfc-4pvwg\" (UID: \"c041548e-b7c2-4f61-aaea-0b76e298ff22\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-4pvwg" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.114614 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/cb184e9f-f0d9-4166-b0da-213c5c7aba5f-etcd-client\") pod \"etcd-operator-b45778765-8vjkd\" (UID: \"cb184e9f-f0d9-4166-b0da-213c5c7aba5f\") " pod="openshift-etcd-operator/etcd-operator-b45778765-8vjkd" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.113458 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/bc560c3d-6558-43e6-86ca-312c9fa264ea-metrics-tls\") pod \"dns-operator-744455d44c-5qwb9\" (UID: \"bc560c3d-6558-43e6-86ca-312c9fa264ea\") " pod="openshift-dns-operator/dns-operator-744455d44c-5qwb9" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.112472 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cb184e9f-f0d9-4166-b0da-213c5c7aba5f-config\") pod \"etcd-operator-b45778765-8vjkd\" (UID: \"cb184e9f-f0d9-4166-b0da-213c5c7aba5f\") " pod="openshift-etcd-operator/etcd-operator-b45778765-8vjkd" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.114905 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/6be3f74a-b215-4588-8b0f-735eba580bbd-client-ca\") pod \"controller-manager-879f6c89f-vjmbx\" (UID: \"6be3f74a-b215-4588-8b0f-735eba580bbd\") " pod="openshift-controller-manager/controller-manager-879f6c89f-vjmbx" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.114678 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b55x6\" (UniqueName: \"kubernetes.io/projected/cb184e9f-f0d9-4166-b0da-213c5c7aba5f-kube-api-access-b55x6\") pod \"etcd-operator-b45778765-8vjkd\" (UID: \"cb184e9f-f0d9-4166-b0da-213c5c7aba5f\") " pod="openshift-etcd-operator/etcd-operator-b45778765-8vjkd" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.115026 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/78d3e7b1-7258-4508-ad38-373f21bb5d96-bound-sa-token\") pod \"image-registry-697d97f7c8-rf9l4\" (UID: \"78d3e7b1-7258-4508-ad38-373f21bb5d96\") " pod="openshift-image-registry/image-registry-697d97f7c8-rf9l4" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.115059 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/2ecfaba9-9034-4ac7-bd51-e3f64b41d825-stats-auth\") pod \"router-default-5444994796-58hwl\" (UID: \"2ecfaba9-9034-4ac7-bd51-e3f64b41d825\") " pod="openshift-ingress/router-default-5444994796-58hwl" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.115146 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-rcnlv"] Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.115160 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/80ceab29-9389-4191-a909-8a6c40bba1cd-profile-collector-cert\") pod \"catalog-operator-68c6474976-hmzsn\" (UID: \"80ceab29-9389-4191-a909-8a6c40bba1cd\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-hmzsn" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.115242 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/18507a98-2ef8-43b1-b94d-51106f9bdd3b-serving-cert\") pod \"console-operator-58897d9998-t8f4h\" (UID: \"18507a98-2ef8-43b1-b94d-51106f9bdd3b\") " pod="openshift-console-operator/console-operator-58897d9998-t8f4h" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.115270 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/41aa056a-34d5-463b-8352-c71e0f0db1a4-secret-volume\") pod \"collect-profiles-29327445-dn5mf\" (UID: \"41aa056a-34d5-463b-8352-c71e0f0db1a4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327445-dn5mf" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.115315 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wsj78\" (UniqueName: \"kubernetes.io/projected/c65d29d9-66e3-41ce-acb2-cec28c52fb29-kube-api-access-wsj78\") pod \"dns-default-jm4gb\" (UID: \"c65d29d9-66e3-41ce-acb2-cec28c52fb29\") " pod="openshift-dns/dns-default-jm4gb" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.110273 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/78d3e7b1-7258-4508-ad38-373f21bb5d96-trusted-ca\") pod \"image-registry-697d97f7c8-rf9l4\" (UID: \"78d3e7b1-7258-4508-ad38-373f21bb5d96\") " pod="openshift-image-registry/image-registry-697d97f7c8-rf9l4" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.115740 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/94a883ac-bee2-403e-8747-a4045e7defea-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-gqn9s\" (UID: \"94a883ac-bee2-403e-8747-a4045e7defea\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-gqn9s" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.116076 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7a19cb1c-5274-48e7-ad86-3c2ff85ecf8c-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-7jflr\" (UID: \"7a19cb1c-5274-48e7-ad86-3c2ff85ecf8c\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-7jflr" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.116219 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/18507a98-2ef8-43b1-b94d-51106f9bdd3b-trusted-ca\") pod \"console-operator-58897d9998-t8f4h\" (UID: \"18507a98-2ef8-43b1-b94d-51106f9bdd3b\") " pod="openshift-console-operator/console-operator-58897d9998-t8f4h" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.116999 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6be3f74a-b215-4588-8b0f-735eba580bbd-serving-cert\") pod \"controller-manager-879f6c89f-vjmbx\" (UID: \"6be3f74a-b215-4588-8b0f-735eba580bbd\") " pod="openshift-controller-manager/controller-manager-879f6c89f-vjmbx" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.117055 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/5fed566e-ecb6-4801-84b3-ad1143a03f08-csi-data-dir\") pod \"csi-hostpathplugin-dsqhc\" (UID: \"5fed566e-ecb6-4801-84b3-ad1143a03f08\") " pod="hostpath-provisioner/csi-hostpathplugin-dsqhc" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.117096 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8c2f8349-aa93-4ac8-a369-5b4b2b9cd6e2-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-2rc5n\" (UID: \"8c2f8349-aa93-4ac8-a369-5b4b2b9cd6e2\") " pod="openshift-marketplace/marketplace-operator-79b997595-2rc5n" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.117134 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6ec3075e-f780-435f-8efa-30d2fb4855a1-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-2pd47\" (UID: \"6ec3075e-f780-435f-8efa-30d2fb4855a1\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-2pd47" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.117202 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7a19cb1c-5274-48e7-ad86-3c2ff85ecf8c-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-7jflr\" (UID: \"7a19cb1c-5274-48e7-ad86-3c2ff85ecf8c\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-7jflr" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.117256 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/53168eca-95fe-4303-b401-380995e5adcb-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-526w4\" (UID: \"53168eca-95fe-4303-b401-380995e5adcb\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-526w4" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.117294 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/78d3e7b1-7258-4508-ad38-373f21bb5d96-registry-certificates\") pod \"image-registry-697d97f7c8-rf9l4\" (UID: \"78d3e7b1-7258-4508-ad38-373f21bb5d96\") " pod="openshift-image-registry/image-registry-697d97f7c8-rf9l4" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.117348 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6hlbm\" (UniqueName: \"kubernetes.io/projected/80ceab29-9389-4191-a909-8a6c40bba1cd-kube-api-access-6hlbm\") pod \"catalog-operator-68c6474976-hmzsn\" (UID: \"80ceab29-9389-4191-a909-8a6c40bba1cd\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-hmzsn" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.117386 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/41aa056a-34d5-463b-8352-c71e0f0db1a4-config-volume\") pod \"collect-profiles-29327445-dn5mf\" (UID: \"41aa056a-34d5-463b-8352-c71e0f0db1a4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327445-dn5mf" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.117429 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/e532fab3-55a6-4c42-a532-9291452d5d82-cert\") pod \"ingress-canary-4cxsw\" (UID: \"e532fab3-55a6-4c42-a532-9291452d5d82\") " pod="openshift-ingress-canary/ingress-canary-4cxsw" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.117456 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-trr7q\" (UniqueName: \"kubernetes.io/projected/8c2f8349-aa93-4ac8-a369-5b4b2b9cd6e2-kube-api-access-trr7q\") pod \"marketplace-operator-79b997595-2rc5n\" (UID: \"8c2f8349-aa93-4ac8-a369-5b4b2b9cd6e2\") " pod="openshift-marketplace/marketplace-operator-79b997595-2rc5n" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.117480 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5422n\" (UniqueName: \"kubernetes.io/projected/5fed566e-ecb6-4801-84b3-ad1143a03f08-kube-api-access-5422n\") pod \"csi-hostpathplugin-dsqhc\" (UID: \"5fed566e-ecb6-4801-84b3-ad1143a03f08\") " pod="hostpath-provisioner/csi-hostpathplugin-dsqhc" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.117519 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9tjtl\" (UniqueName: \"kubernetes.io/projected/54f93640-64fd-4728-a83b-499aee1ffa38-kube-api-access-9tjtl\") pod \"migrator-59844c95c7-skh8h\" (UID: \"54f93640-64fd-4728-a83b-499aee1ffa38\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-skh8h" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.117569 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/5fed566e-ecb6-4801-84b3-ad1143a03f08-socket-dir\") pod \"csi-hostpathplugin-dsqhc\" (UID: \"5fed566e-ecb6-4801-84b3-ad1143a03f08\") " pod="hostpath-provisioner/csi-hostpathplugin-dsqhc" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.117601 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9vncv\" (UniqueName: \"kubernetes.io/projected/bc560c3d-6558-43e6-86ca-312c9fa264ea-kube-api-access-9vncv\") pod \"dns-operator-744455d44c-5qwb9\" (UID: \"bc560c3d-6558-43e6-86ca-312c9fa264ea\") " pod="openshift-dns-operator/dns-operator-744455d44c-5qwb9" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.117636 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/5fed566e-ecb6-4801-84b3-ad1143a03f08-registration-dir\") pod \"csi-hostpathplugin-dsqhc\" (UID: \"5fed566e-ecb6-4801-84b3-ad1143a03f08\") " pod="hostpath-provisioner/csi-hostpathplugin-dsqhc" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.117660 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/beeea3e9-d3c3-4280-bf04-85cee4378cb0-config\") pod \"kube-apiserver-operator-766d6c64bb-8bpqg\" (UID: \"beeea3e9-d3c3-4280-bf04-85cee4378cb0\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-8bpqg" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.117785 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zwxrp\" (UniqueName: \"kubernetes.io/projected/41aa056a-34d5-463b-8352-c71e0f0db1a4-kube-api-access-zwxrp\") pod \"collect-profiles-29327445-dn5mf\" (UID: \"41aa056a-34d5-463b-8352-c71e0f0db1a4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327445-dn5mf" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.117814 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/78d3e7b1-7258-4508-ad38-373f21bb5d96-installation-pull-secrets\") pod \"image-registry-697d97f7c8-rf9l4\" (UID: \"78d3e7b1-7258-4508-ad38-373f21bb5d96\") " pod="openshift-image-registry/image-registry-697d97f7c8-rf9l4" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.118476 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/18507a98-2ef8-43b1-b94d-51106f9bdd3b-trusted-ca\") pod \"console-operator-58897d9998-t8f4h\" (UID: \"18507a98-2ef8-43b1-b94d-51106f9bdd3b\") " pod="openshift-console-operator/console-operator-58897d9998-t8f4h" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.117835 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/8049dd16-5d6f-4718-94b1-fb5a322c2530-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-pqm4r\" (UID: \"8049dd16-5d6f-4718-94b1-fb5a322c2530\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-pqm4r" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.118725 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/5fed566e-ecb6-4801-84b3-ad1143a03f08-mountpoint-dir\") pod \"csi-hostpathplugin-dsqhc\" (UID: \"5fed566e-ecb6-4801-84b3-ad1143a03f08\") " pod="hostpath-provisioner/csi-hostpathplugin-dsqhc" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.118768 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/0cf52841-edf1-4daf-8e2c-76f5f6b93104-srv-cert\") pod \"olm-operator-6b444d44fb-zf4p8\" (UID: \"0cf52841-edf1-4daf-8e2c-76f5f6b93104\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-zf4p8" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.118822 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rf9l4\" (UID: \"78d3e7b1-7258-4508-ad38-373f21bb5d96\") " pod="openshift-image-registry/image-registry-697d97f7c8-rf9l4" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.118878 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/78d3e7b1-7258-4508-ad38-373f21bb5d96-registry-tls\") pod \"image-registry-697d97f7c8-rf9l4\" (UID: \"78d3e7b1-7258-4508-ad38-373f21bb5d96\") " pod="openshift-image-registry/image-registry-697d97f7c8-rf9l4" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.118901 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6ec3075e-f780-435f-8efa-30d2fb4855a1-service-ca-bundle\") pod \"authentication-operator-69f744f599-2pd47\" (UID: \"6ec3075e-f780-435f-8efa-30d2fb4855a1\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-2pd47" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.118971 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/5fed566e-ecb6-4801-84b3-ad1143a03f08-plugins-dir\") pod \"csi-hostpathplugin-dsqhc\" (UID: \"5fed566e-ecb6-4801-84b3-ad1143a03f08\") " pod="hostpath-provisioner/csi-hostpathplugin-dsqhc" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.119706 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/18507a98-2ef8-43b1-b94d-51106f9bdd3b-serving-cert\") pod \"console-operator-58897d9998-t8f4h\" (UID: \"18507a98-2ef8-43b1-b94d-51106f9bdd3b\") " pod="openshift-console-operator/console-operator-58897d9998-t8f4h" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.118998 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mtl7j\" (UniqueName: \"kubernetes.io/projected/78d3e7b1-7258-4508-ad38-373f21bb5d96-kube-api-access-mtl7j\") pod \"image-registry-697d97f7c8-rf9l4\" (UID: \"78d3e7b1-7258-4508-ad38-373f21bb5d96\") " pod="openshift-image-registry/image-registry-697d97f7c8-rf9l4" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.120658 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/0755bb69-c94a-4b36-97ee-8a3dfd972c79-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-dtbh5\" (UID: \"0755bb69-c94a-4b36-97ee-8a3dfd972c79\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-dtbh5" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.120710 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/74822779-bdcd-4aae-bc01-9cbcbb0c1d83-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-x5n4h\" (UID: \"74822779-bdcd-4aae-bc01-9cbcbb0c1d83\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-x5n4h" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.120755 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8b7fb5f2-4b5e-4fc3-9c97-dfbe36646871-config\") pod \"service-ca-operator-777779d784-5rj2f\" (UID: \"8b7fb5f2-4b5e-4fc3-9c97-dfbe36646871\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-5rj2f" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.120808 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c65d29d9-66e3-41ce-acb2-cec28c52fb29-config-volume\") pod \"dns-default-jm4gb\" (UID: \"c65d29d9-66e3-41ce-acb2-cec28c52fb29\") " pod="openshift-dns/dns-default-jm4gb" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.120836 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7a19cb1c-5274-48e7-ad86-3c2ff85ecf8c-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-7jflr\" (UID: \"7a19cb1c-5274-48e7-ad86-3c2ff85ecf8c\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-7jflr" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.120877 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/31d92551-cb79-4d9c-ad1c-c5090142b82c-signing-cabundle\") pod \"service-ca-9c57cc56f-bpbtf\" (UID: \"31d92551-cb79-4d9c-ad1c-c5090142b82c\") " pod="openshift-service-ca/service-ca-9c57cc56f-bpbtf" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.120901 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/b21ccaf7-40a5-4939-ad8b-4c7efdcf4306-proxy-tls\") pod \"machine-config-controller-84d6567774-m7gph\" (UID: \"b21ccaf7-40a5-4939-ad8b-4c7efdcf4306\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-m7gph" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.120929 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/da8eba09-f287-4704-b990-348ff69f0173-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-k87kc\" (UID: \"da8eba09-f287-4704-b990-348ff69f0173\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-k87kc" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.121137 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vw2sb\" (UniqueName: \"kubernetes.io/projected/18507a98-2ef8-43b1-b94d-51106f9bdd3b-kube-api-access-vw2sb\") pod \"console-operator-58897d9998-t8f4h\" (UID: \"18507a98-2ef8-43b1-b94d-51106f9bdd3b\") " pod="openshift-console-operator/console-operator-58897d9998-t8f4h" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.121221 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/788472f5-2ad1-49d0-8e31-6a4fe14c3c85-images\") pod \"machine-config-operator-74547568cd-wnmjw\" (UID: \"788472f5-2ad1-49d0-8e31-6a4fe14c3c85\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-wnmjw" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.121284 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6gnmv\" (UniqueName: \"kubernetes.io/projected/53168eca-95fe-4303-b401-380995e5adcb-kube-api-access-6gnmv\") pod \"openshift-controller-manager-operator-756b6f6bc6-526w4\" (UID: \"53168eca-95fe-4303-b401-380995e5adcb\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-526w4" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.121312 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6ec3075e-f780-435f-8efa-30d2fb4855a1-serving-cert\") pod \"authentication-operator-69f744f599-2pd47\" (UID: \"6ec3075e-f780-435f-8efa-30d2fb4855a1\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-2pd47" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.121315 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7a19cb1c-5274-48e7-ad86-3c2ff85ecf8c-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-7jflr\" (UID: \"7a19cb1c-5274-48e7-ad86-3c2ff85ecf8c\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-7jflr" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.121942 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/53168eca-95fe-4303-b401-380995e5adcb-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-526w4\" (UID: \"53168eca-95fe-4303-b401-380995e5adcb\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-526w4" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.123744 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/53168eca-95fe-4303-b401-380995e5adcb-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-526w4\" (UID: \"53168eca-95fe-4303-b401-380995e5adcb\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-526w4" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.124129 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/cb184e9f-f0d9-4166-b0da-213c5c7aba5f-etcd-ca\") pod \"etcd-operator-b45778765-8vjkd\" (UID: \"cb184e9f-f0d9-4166-b0da-213c5c7aba5f\") " pod="openshift-etcd-operator/etcd-operator-b45778765-8vjkd" Oct 05 06:50:20 crc kubenswrapper[4846]: E1005 06:50:20.124736 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-05 06:50:20.624716214 +0000 UTC m=+142.865568989 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rf9l4" (UID: "78d3e7b1-7258-4508-ad38-373f21bb5d96") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.124874 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6ec3075e-f780-435f-8efa-30d2fb4855a1-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-2pd47\" (UID: \"6ec3075e-f780-435f-8efa-30d2fb4855a1\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-2pd47" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.125643 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6ec3075e-f780-435f-8efa-30d2fb4855a1-service-ca-bundle\") pod \"authentication-operator-69f744f599-2pd47\" (UID: \"6ec3075e-f780-435f-8efa-30d2fb4855a1\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-2pd47" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.128045 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/78d3e7b1-7258-4508-ad38-373f21bb5d96-registry-tls\") pod \"image-registry-697d97f7c8-rf9l4\" (UID: \"78d3e7b1-7258-4508-ad38-373f21bb5d96\") " pod="openshift-image-registry/image-registry-697d97f7c8-rf9l4" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.129128 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6ec3075e-f780-435f-8efa-30d2fb4855a1-serving-cert\") pod \"authentication-operator-69f744f599-2pd47\" (UID: \"6ec3075e-f780-435f-8efa-30d2fb4855a1\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-2pd47" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.129131 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/78d3e7b1-7258-4508-ad38-373f21bb5d96-registry-certificates\") pod \"image-registry-697d97f7c8-rf9l4\" (UID: \"78d3e7b1-7258-4508-ad38-373f21bb5d96\") " pod="openshift-image-registry/image-registry-697d97f7c8-rf9l4" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.130587 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/da8eba09-f287-4704-b990-348ff69f0173-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-k87kc\" (UID: \"da8eba09-f287-4704-b990-348ff69f0173\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-k87kc" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.130807 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/78d3e7b1-7258-4508-ad38-373f21bb5d96-installation-pull-secrets\") pod \"image-registry-697d97f7c8-rf9l4\" (UID: \"78d3e7b1-7258-4508-ad38-373f21bb5d96\") " pod="openshift-image-registry/image-registry-697d97f7c8-rf9l4" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.134903 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6be3f74a-b215-4588-8b0f-735eba580bbd-serving-cert\") pod \"controller-manager-879f6c89f-vjmbx\" (UID: \"6be3f74a-b215-4588-8b0f-735eba580bbd\") " pod="openshift-controller-manager/controller-manager-879f6c89f-vjmbx" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.138621 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gbd7g\" (UniqueName: \"kubernetes.io/projected/6be3f74a-b215-4588-8b0f-735eba580bbd-kube-api-access-gbd7g\") pod \"controller-manager-879f6c89f-vjmbx\" (UID: \"6be3f74a-b215-4588-8b0f-735eba580bbd\") " pod="openshift-controller-manager/controller-manager-879f6c89f-vjmbx" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.141672 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7a19cb1c-5274-48e7-ad86-3c2ff85ecf8c-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-7jflr\" (UID: \"7a19cb1c-5274-48e7-ad86-3c2ff85ecf8c\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-7jflr" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.162800 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2tvs4\" (UniqueName: \"kubernetes.io/projected/da8eba09-f287-4704-b990-348ff69f0173-kube-api-access-2tvs4\") pod \"cluster-samples-operator-665b6dd947-k87kc\" (UID: \"da8eba09-f287-4704-b990-348ff69f0173\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-k87kc" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.179347 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sgkh2\" (UniqueName: \"kubernetes.io/projected/6ec3075e-f780-435f-8efa-30d2fb4855a1-kube-api-access-sgkh2\") pod \"authentication-operator-69f744f599-2pd47\" (UID: \"6ec3075e-f780-435f-8efa-30d2fb4855a1\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-2pd47" Oct 05 06:50:20 crc kubenswrapper[4846]: W1005 06:50:20.182878 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5bddce7c_b835_4a1f_a3af_db5a1d3771f2.slice/crio-a7844d2c8a706845d237cc7e460f547241d5b84ce18cd3365e6234685157a3fc WatchSource:0}: Error finding container a7844d2c8a706845d237cc7e460f547241d5b84ce18cd3365e6234685157a3fc: Status 404 returned error can't find the container with id a7844d2c8a706845d237cc7e460f547241d5b84ce18cd3365e6234685157a3fc Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.197759 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/78d3e7b1-7258-4508-ad38-373f21bb5d96-bound-sa-token\") pod \"image-registry-697d97f7c8-rf9l4\" (UID: \"78d3e7b1-7258-4508-ad38-373f21bb5d96\") " pod="openshift-image-registry/image-registry-697d97f7c8-rf9l4" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.216863 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-k87kc" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.222351 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.222485 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/5fed566e-ecb6-4801-84b3-ad1143a03f08-socket-dir\") pod \"csi-hostpathplugin-dsqhc\" (UID: \"5fed566e-ecb6-4801-84b3-ad1143a03f08\") " pod="hostpath-provisioner/csi-hostpathplugin-dsqhc" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.222511 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/5fed566e-ecb6-4801-84b3-ad1143a03f08-registration-dir\") pod \"csi-hostpathplugin-dsqhc\" (UID: \"5fed566e-ecb6-4801-84b3-ad1143a03f08\") " pod="hostpath-provisioner/csi-hostpathplugin-dsqhc" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.222527 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/beeea3e9-d3c3-4280-bf04-85cee4378cb0-config\") pod \"kube-apiserver-operator-766d6c64bb-8bpqg\" (UID: \"beeea3e9-d3c3-4280-bf04-85cee4378cb0\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-8bpqg" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.222546 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zwxrp\" (UniqueName: \"kubernetes.io/projected/41aa056a-34d5-463b-8352-c71e0f0db1a4-kube-api-access-zwxrp\") pod \"collect-profiles-29327445-dn5mf\" (UID: \"41aa056a-34d5-463b-8352-c71e0f0db1a4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327445-dn5mf" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.222564 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/8049dd16-5d6f-4718-94b1-fb5a322c2530-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-pqm4r\" (UID: \"8049dd16-5d6f-4718-94b1-fb5a322c2530\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-pqm4r" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.222579 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/5fed566e-ecb6-4801-84b3-ad1143a03f08-mountpoint-dir\") pod \"csi-hostpathplugin-dsqhc\" (UID: \"5fed566e-ecb6-4801-84b3-ad1143a03f08\") " pod="hostpath-provisioner/csi-hostpathplugin-dsqhc" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.222593 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/0cf52841-edf1-4daf-8e2c-76f5f6b93104-srv-cert\") pod \"olm-operator-6b444d44fb-zf4p8\" (UID: \"0cf52841-edf1-4daf-8e2c-76f5f6b93104\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-zf4p8" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.222635 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/5fed566e-ecb6-4801-84b3-ad1143a03f08-plugins-dir\") pod \"csi-hostpathplugin-dsqhc\" (UID: \"5fed566e-ecb6-4801-84b3-ad1143a03f08\") " pod="hostpath-provisioner/csi-hostpathplugin-dsqhc" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.222654 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/0755bb69-c94a-4b36-97ee-8a3dfd972c79-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-dtbh5\" (UID: \"0755bb69-c94a-4b36-97ee-8a3dfd972c79\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-dtbh5" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.222672 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/74822779-bdcd-4aae-bc01-9cbcbb0c1d83-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-x5n4h\" (UID: \"74822779-bdcd-4aae-bc01-9cbcbb0c1d83\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-x5n4h" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.222687 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8b7fb5f2-4b5e-4fc3-9c97-dfbe36646871-config\") pod \"service-ca-operator-777779d784-5rj2f\" (UID: \"8b7fb5f2-4b5e-4fc3-9c97-dfbe36646871\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-5rj2f" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.222701 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c65d29d9-66e3-41ce-acb2-cec28c52fb29-config-volume\") pod \"dns-default-jm4gb\" (UID: \"c65d29d9-66e3-41ce-acb2-cec28c52fb29\") " pod="openshift-dns/dns-default-jm4gb" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.222720 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/31d92551-cb79-4d9c-ad1c-c5090142b82c-signing-cabundle\") pod \"service-ca-9c57cc56f-bpbtf\" (UID: \"31d92551-cb79-4d9c-ad1c-c5090142b82c\") " pod="openshift-service-ca/service-ca-9c57cc56f-bpbtf" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.222735 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/b21ccaf7-40a5-4939-ad8b-4c7efdcf4306-proxy-tls\") pod \"machine-config-controller-84d6567774-m7gph\" (UID: \"b21ccaf7-40a5-4939-ad8b-4c7efdcf4306\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-m7gph" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.222759 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/788472f5-2ad1-49d0-8e31-6a4fe14c3c85-images\") pod \"machine-config-operator-74547568cd-wnmjw\" (UID: \"788472f5-2ad1-49d0-8e31-6a4fe14c3c85\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-wnmjw" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.222784 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/beeea3e9-d3c3-4280-bf04-85cee4378cb0-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-8bpqg\" (UID: \"beeea3e9-d3c3-4280-bf04-85cee4378cb0\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-8bpqg" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.222799 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fv4fz\" (UniqueName: \"kubernetes.io/projected/788472f5-2ad1-49d0-8e31-6a4fe14c3c85-kube-api-access-fv4fz\") pod \"machine-config-operator-74547568cd-wnmjw\" (UID: \"788472f5-2ad1-49d0-8e31-6a4fe14c3c85\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-wnmjw" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.222821 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9dm98\" (UniqueName: \"kubernetes.io/projected/74822779-bdcd-4aae-bc01-9cbcbb0c1d83-kube-api-access-9dm98\") pod \"kube-storage-version-migrator-operator-b67b599dd-x5n4h\" (UID: \"74822779-bdcd-4aae-bc01-9cbcbb0c1d83\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-x5n4h" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.222837 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/c041548e-b7c2-4f61-aaea-0b76e298ff22-apiservice-cert\") pod \"packageserver-d55dfcdfc-4pvwg\" (UID: \"c041548e-b7c2-4f61-aaea-0b76e298ff22\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-4pvwg" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.222853 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sl5t6\" (UniqueName: \"kubernetes.io/projected/0cf52841-edf1-4daf-8e2c-76f5f6b93104-kube-api-access-sl5t6\") pod \"olm-operator-6b444d44fb-zf4p8\" (UID: \"0cf52841-edf1-4daf-8e2c-76f5f6b93104\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-zf4p8" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.222873 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/80c88b0b-c4cf-478e-8e05-4cde023d15ee-node-bootstrap-token\") pod \"machine-config-server-2g2rj\" (UID: \"80c88b0b-c4cf-478e-8e05-4cde023d15ee\") " pod="openshift-machine-config-operator/machine-config-server-2g2rj" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.222899 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/8c2f8349-aa93-4ac8-a369-5b4b2b9cd6e2-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-2rc5n\" (UID: \"8c2f8349-aa93-4ac8-a369-5b4b2b9cd6e2\") " pod="openshift-marketplace/marketplace-operator-79b997595-2rc5n" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.222916 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mxbkd\" (UniqueName: \"kubernetes.io/projected/2ecfaba9-9034-4ac7-bd51-e3f64b41d825-kube-api-access-mxbkd\") pod \"router-default-5444994796-58hwl\" (UID: \"2ecfaba9-9034-4ac7-bd51-e3f64b41d825\") " pod="openshift-ingress/router-default-5444994796-58hwl" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.222940 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ppwzm\" (UniqueName: \"kubernetes.io/projected/31d92551-cb79-4d9c-ad1c-c5090142b82c-kube-api-access-ppwzm\") pod \"service-ca-9c57cc56f-bpbtf\" (UID: \"31d92551-cb79-4d9c-ad1c-c5090142b82c\") " pod="openshift-service-ca/service-ca-9c57cc56f-bpbtf" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.222961 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4d7ch\" (UniqueName: \"kubernetes.io/projected/80c88b0b-c4cf-478e-8e05-4cde023d15ee-kube-api-access-4d7ch\") pod \"machine-config-server-2g2rj\" (UID: \"80c88b0b-c4cf-478e-8e05-4cde023d15ee\") " pod="openshift-machine-config-operator/machine-config-server-2g2rj" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.222978 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8b7fb5f2-4b5e-4fc3-9c97-dfbe36646871-serving-cert\") pod \"service-ca-operator-777779d784-5rj2f\" (UID: \"8b7fb5f2-4b5e-4fc3-9c97-dfbe36646871\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-5rj2f" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.222994 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2ecfaba9-9034-4ac7-bd51-e3f64b41d825-service-ca-bundle\") pod \"router-default-5444994796-58hwl\" (UID: \"2ecfaba9-9034-4ac7-bd51-e3f64b41d825\") " pod="openshift-ingress/router-default-5444994796-58hwl" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.223010 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/c65d29d9-66e3-41ce-acb2-cec28c52fb29-metrics-tls\") pod \"dns-default-jm4gb\" (UID: \"c65d29d9-66e3-41ce-acb2-cec28c52fb29\") " pod="openshift-dns/dns-default-jm4gb" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.223025 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wth4v\" (UniqueName: \"kubernetes.io/projected/e532fab3-55a6-4c42-a532-9291452d5d82-kube-api-access-wth4v\") pod \"ingress-canary-4cxsw\" (UID: \"e532fab3-55a6-4c42-a532-9291452d5d82\") " pod="openshift-ingress-canary/ingress-canary-4cxsw" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.223043 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/57e551f8-d9da-4692-b8ae-f7ff6829fe25-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-2x9nl\" (UID: \"57e551f8-d9da-4692-b8ae-f7ff6829fe25\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-2x9nl" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.223060 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/94a883ac-bee2-403e-8747-a4045e7defea-config\") pod \"kube-controller-manager-operator-78b949d7b-gqn9s\" (UID: \"94a883ac-bee2-403e-8747-a4045e7defea\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-gqn9s" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.223077 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s8g9n\" (UniqueName: \"kubernetes.io/projected/0755bb69-c94a-4b36-97ee-8a3dfd972c79-kube-api-access-s8g9n\") pod \"multus-admission-controller-857f4d67dd-dtbh5\" (UID: \"0755bb69-c94a-4b36-97ee-8a3dfd972c79\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-dtbh5" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.223093 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/2ecfaba9-9034-4ac7-bd51-e3f64b41d825-default-certificate\") pod \"router-default-5444994796-58hwl\" (UID: \"2ecfaba9-9034-4ac7-bd51-e3f64b41d825\") " pod="openshift-ingress/router-default-5444994796-58hwl" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.223111 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9mht8\" (UniqueName: \"kubernetes.io/projected/b21ccaf7-40a5-4939-ad8b-4c7efdcf4306-kube-api-access-9mht8\") pod \"machine-config-controller-84d6567774-m7gph\" (UID: \"b21ccaf7-40a5-4939-ad8b-4c7efdcf4306\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-m7gph" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.223134 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/788472f5-2ad1-49d0-8e31-6a4fe14c3c85-auth-proxy-config\") pod \"machine-config-operator-74547568cd-wnmjw\" (UID: \"788472f5-2ad1-49d0-8e31-6a4fe14c3c85\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-wnmjw" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.223157 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/beeea3e9-d3c3-4280-bf04-85cee4378cb0-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-8bpqg\" (UID: \"beeea3e9-d3c3-4280-bf04-85cee4378cb0\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-8bpqg" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.223206 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7ml2d\" (UniqueName: \"kubernetes.io/projected/8b7fb5f2-4b5e-4fc3-9c97-dfbe36646871-kube-api-access-7ml2d\") pod \"service-ca-operator-777779d784-5rj2f\" (UID: \"8b7fb5f2-4b5e-4fc3-9c97-dfbe36646871\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-5rj2f" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.223234 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/31d92551-cb79-4d9c-ad1c-c5090142b82c-signing-key\") pod \"service-ca-9c57cc56f-bpbtf\" (UID: \"31d92551-cb79-4d9c-ad1c-c5090142b82c\") " pod="openshift-service-ca/service-ca-9c57cc56f-bpbtf" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.223251 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6srxd\" (UniqueName: \"kubernetes.io/projected/c041548e-b7c2-4f61-aaea-0b76e298ff22-kube-api-access-6srxd\") pod \"packageserver-d55dfcdfc-4pvwg\" (UID: \"c041548e-b7c2-4f61-aaea-0b76e298ff22\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-4pvwg" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.223267 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/0cf52841-edf1-4daf-8e2c-76f5f6b93104-profile-collector-cert\") pod \"olm-operator-6b444d44fb-zf4p8\" (UID: \"0cf52841-edf1-4daf-8e2c-76f5f6b93104\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-zf4p8" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.223284 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/94a883ac-bee2-403e-8747-a4045e7defea-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-gqn9s\" (UID: \"94a883ac-bee2-403e-8747-a4045e7defea\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-gqn9s" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.223299 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/b21ccaf7-40a5-4939-ad8b-4c7efdcf4306-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-m7gph\" (UID: \"b21ccaf7-40a5-4939-ad8b-4c7efdcf4306\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-m7gph" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.223319 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/c041548e-b7c2-4f61-aaea-0b76e298ff22-webhook-cert\") pod \"packageserver-d55dfcdfc-4pvwg\" (UID: \"c041548e-b7c2-4f61-aaea-0b76e298ff22\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-4pvwg" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.223341 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/2ecfaba9-9034-4ac7-bd51-e3f64b41d825-metrics-certs\") pod \"router-default-5444994796-58hwl\" (UID: \"2ecfaba9-9034-4ac7-bd51-e3f64b41d825\") " pod="openshift-ingress/router-default-5444994796-58hwl" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.223366 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/74822779-bdcd-4aae-bc01-9cbcbb0c1d83-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-x5n4h\" (UID: \"74822779-bdcd-4aae-bc01-9cbcbb0c1d83\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-x5n4h" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.223386 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9h8b2\" (UniqueName: \"kubernetes.io/projected/57e551f8-d9da-4692-b8ae-f7ff6829fe25-kube-api-access-9h8b2\") pod \"control-plane-machine-set-operator-78cbb6b69f-2x9nl\" (UID: \"57e551f8-d9da-4692-b8ae-f7ff6829fe25\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-2x9nl" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.223408 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/788472f5-2ad1-49d0-8e31-6a4fe14c3c85-proxy-tls\") pod \"machine-config-operator-74547568cd-wnmjw\" (UID: \"788472f5-2ad1-49d0-8e31-6a4fe14c3c85\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-wnmjw" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.223432 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ghg5r\" (UniqueName: \"kubernetes.io/projected/8049dd16-5d6f-4718-94b1-fb5a322c2530-kube-api-access-ghg5r\") pod \"package-server-manager-789f6589d5-pqm4r\" (UID: \"8049dd16-5d6f-4718-94b1-fb5a322c2530\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-pqm4r" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.223453 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/80ceab29-9389-4191-a909-8a6c40bba1cd-srv-cert\") pod \"catalog-operator-68c6474976-hmzsn\" (UID: \"80ceab29-9389-4191-a909-8a6c40bba1cd\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-hmzsn" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.223468 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/80c88b0b-c4cf-478e-8e05-4cde023d15ee-certs\") pod \"machine-config-server-2g2rj\" (UID: \"80c88b0b-c4cf-478e-8e05-4cde023d15ee\") " pod="openshift-machine-config-operator/machine-config-server-2g2rj" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.223482 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/c041548e-b7c2-4f61-aaea-0b76e298ff22-tmpfs\") pod \"packageserver-d55dfcdfc-4pvwg\" (UID: \"c041548e-b7c2-4f61-aaea-0b76e298ff22\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-4pvwg" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.223501 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/2ecfaba9-9034-4ac7-bd51-e3f64b41d825-stats-auth\") pod \"router-default-5444994796-58hwl\" (UID: \"2ecfaba9-9034-4ac7-bd51-e3f64b41d825\") " pod="openshift-ingress/router-default-5444994796-58hwl" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.223528 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/80ceab29-9389-4191-a909-8a6c40bba1cd-profile-collector-cert\") pod \"catalog-operator-68c6474976-hmzsn\" (UID: \"80ceab29-9389-4191-a909-8a6c40bba1cd\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-hmzsn" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.223550 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/41aa056a-34d5-463b-8352-c71e0f0db1a4-secret-volume\") pod \"collect-profiles-29327445-dn5mf\" (UID: \"41aa056a-34d5-463b-8352-c71e0f0db1a4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327445-dn5mf" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.223571 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wsj78\" (UniqueName: \"kubernetes.io/projected/c65d29d9-66e3-41ce-acb2-cec28c52fb29-kube-api-access-wsj78\") pod \"dns-default-jm4gb\" (UID: \"c65d29d9-66e3-41ce-acb2-cec28c52fb29\") " pod="openshift-dns/dns-default-jm4gb" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.223589 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/94a883ac-bee2-403e-8747-a4045e7defea-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-gqn9s\" (UID: \"94a883ac-bee2-403e-8747-a4045e7defea\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-gqn9s" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.223610 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/5fed566e-ecb6-4801-84b3-ad1143a03f08-csi-data-dir\") pod \"csi-hostpathplugin-dsqhc\" (UID: \"5fed566e-ecb6-4801-84b3-ad1143a03f08\") " pod="hostpath-provisioner/csi-hostpathplugin-dsqhc" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.223632 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8c2f8349-aa93-4ac8-a369-5b4b2b9cd6e2-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-2rc5n\" (UID: \"8c2f8349-aa93-4ac8-a369-5b4b2b9cd6e2\") " pod="openshift-marketplace/marketplace-operator-79b997595-2rc5n" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.223658 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6hlbm\" (UniqueName: \"kubernetes.io/projected/80ceab29-9389-4191-a909-8a6c40bba1cd-kube-api-access-6hlbm\") pod \"catalog-operator-68c6474976-hmzsn\" (UID: \"80ceab29-9389-4191-a909-8a6c40bba1cd\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-hmzsn" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.223673 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/41aa056a-34d5-463b-8352-c71e0f0db1a4-config-volume\") pod \"collect-profiles-29327445-dn5mf\" (UID: \"41aa056a-34d5-463b-8352-c71e0f0db1a4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327445-dn5mf" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.223688 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/e532fab3-55a6-4c42-a532-9291452d5d82-cert\") pod \"ingress-canary-4cxsw\" (UID: \"e532fab3-55a6-4c42-a532-9291452d5d82\") " pod="openshift-ingress-canary/ingress-canary-4cxsw" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.223703 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-trr7q\" (UniqueName: \"kubernetes.io/projected/8c2f8349-aa93-4ac8-a369-5b4b2b9cd6e2-kube-api-access-trr7q\") pod \"marketplace-operator-79b997595-2rc5n\" (UID: \"8c2f8349-aa93-4ac8-a369-5b4b2b9cd6e2\") " pod="openshift-marketplace/marketplace-operator-79b997595-2rc5n" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.223718 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5422n\" (UniqueName: \"kubernetes.io/projected/5fed566e-ecb6-4801-84b3-ad1143a03f08-kube-api-access-5422n\") pod \"csi-hostpathplugin-dsqhc\" (UID: \"5fed566e-ecb6-4801-84b3-ad1143a03f08\") " pod="hostpath-provisioner/csi-hostpathplugin-dsqhc" Oct 05 06:50:20 crc kubenswrapper[4846]: E1005 06:50:20.223922 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-05 06:50:20.723907157 +0000 UTC m=+142.964759922 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.224155 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/5fed566e-ecb6-4801-84b3-ad1143a03f08-socket-dir\") pod \"csi-hostpathplugin-dsqhc\" (UID: \"5fed566e-ecb6-4801-84b3-ad1143a03f08\") " pod="hostpath-provisioner/csi-hostpathplugin-dsqhc" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.224215 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/5fed566e-ecb6-4801-84b3-ad1143a03f08-registration-dir\") pod \"csi-hostpathplugin-dsqhc\" (UID: \"5fed566e-ecb6-4801-84b3-ad1143a03f08\") " pod="hostpath-provisioner/csi-hostpathplugin-dsqhc" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.224725 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/beeea3e9-d3c3-4280-bf04-85cee4378cb0-config\") pod \"kube-apiserver-operator-766d6c64bb-8bpqg\" (UID: \"beeea3e9-d3c3-4280-bf04-85cee4378cb0\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-8bpqg" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.225824 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/5fed566e-ecb6-4801-84b3-ad1143a03f08-mountpoint-dir\") pod \"csi-hostpathplugin-dsqhc\" (UID: \"5fed566e-ecb6-4801-84b3-ad1143a03f08\") " pod="hostpath-provisioner/csi-hostpathplugin-dsqhc" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.226042 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b55x6\" (UniqueName: \"kubernetes.io/projected/cb184e9f-f0d9-4166-b0da-213c5c7aba5f-kube-api-access-b55x6\") pod \"etcd-operator-b45778765-8vjkd\" (UID: \"cb184e9f-f0d9-4166-b0da-213c5c7aba5f\") " pod="openshift-etcd-operator/etcd-operator-b45778765-8vjkd" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.227907 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/8049dd16-5d6f-4718-94b1-fb5a322c2530-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-pqm4r\" (UID: \"8049dd16-5d6f-4718-94b1-fb5a322c2530\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-pqm4r" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.229720 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-2pd47" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.230551 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/788472f5-2ad1-49d0-8e31-6a4fe14c3c85-auth-proxy-config\") pod \"machine-config-operator-74547568cd-wnmjw\" (UID: \"788472f5-2ad1-49d0-8e31-6a4fe14c3c85\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-wnmjw" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.230597 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/5fed566e-ecb6-4801-84b3-ad1143a03f08-plugins-dir\") pod \"csi-hostpathplugin-dsqhc\" (UID: \"5fed566e-ecb6-4801-84b3-ad1143a03f08\") " pod="hostpath-provisioner/csi-hostpathplugin-dsqhc" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.232651 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/0cf52841-edf1-4daf-8e2c-76f5f6b93104-srv-cert\") pod \"olm-operator-6b444d44fb-zf4p8\" (UID: \"0cf52841-edf1-4daf-8e2c-76f5f6b93104\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-zf4p8" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.233427 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-vjmbx" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.233713 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/788472f5-2ad1-49d0-8e31-6a4fe14c3c85-images\") pod \"machine-config-operator-74547568cd-wnmjw\" (UID: \"788472f5-2ad1-49d0-8e31-6a4fe14c3c85\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-wnmjw" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.234223 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/74822779-bdcd-4aae-bc01-9cbcbb0c1d83-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-x5n4h\" (UID: \"74822779-bdcd-4aae-bc01-9cbcbb0c1d83\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-x5n4h" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.234697 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8b7fb5f2-4b5e-4fc3-9c97-dfbe36646871-config\") pod \"service-ca-operator-777779d784-5rj2f\" (UID: \"8b7fb5f2-4b5e-4fc3-9c97-dfbe36646871\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-5rj2f" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.234910 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/31d92551-cb79-4d9c-ad1c-c5090142b82c-signing-cabundle\") pod \"service-ca-9c57cc56f-bpbtf\" (UID: \"31d92551-cb79-4d9c-ad1c-c5090142b82c\") " pod="openshift-service-ca/service-ca-9c57cc56f-bpbtf" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.235962 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c65d29d9-66e3-41ce-acb2-cec28c52fb29-config-volume\") pod \"dns-default-jm4gb\" (UID: \"c65d29d9-66e3-41ce-acb2-cec28c52fb29\") " pod="openshift-dns/dns-default-jm4gb" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.236281 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/beeea3e9-d3c3-4280-bf04-85cee4378cb0-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-8bpqg\" (UID: \"beeea3e9-d3c3-4280-bf04-85cee4378cb0\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-8bpqg" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.236557 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/0755bb69-c94a-4b36-97ee-8a3dfd972c79-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-dtbh5\" (UID: \"0755bb69-c94a-4b36-97ee-8a3dfd972c79\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-dtbh5" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.240041 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/5fed566e-ecb6-4801-84b3-ad1143a03f08-csi-data-dir\") pod \"csi-hostpathplugin-dsqhc\" (UID: \"5fed566e-ecb6-4801-84b3-ad1143a03f08\") " pod="hostpath-provisioner/csi-hostpathplugin-dsqhc" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.240461 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/c041548e-b7c2-4f61-aaea-0b76e298ff22-apiservice-cert\") pod \"packageserver-d55dfcdfc-4pvwg\" (UID: \"c041548e-b7c2-4f61-aaea-0b76e298ff22\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-4pvwg" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.241129 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/b21ccaf7-40a5-4939-ad8b-4c7efdcf4306-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-m7gph\" (UID: \"b21ccaf7-40a5-4939-ad8b-4c7efdcf4306\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-m7gph" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.241212 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2ecfaba9-9034-4ac7-bd51-e3f64b41d825-service-ca-bundle\") pod \"router-default-5444994796-58hwl\" (UID: \"2ecfaba9-9034-4ac7-bd51-e3f64b41d825\") " pod="openshift-ingress/router-default-5444994796-58hwl" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.241550 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8c2f8349-aa93-4ac8-a369-5b4b2b9cd6e2-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-2rc5n\" (UID: \"8c2f8349-aa93-4ac8-a369-5b4b2b9cd6e2\") " pod="openshift-marketplace/marketplace-operator-79b997595-2rc5n" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.243162 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/c65d29d9-66e3-41ce-acb2-cec28c52fb29-metrics-tls\") pod \"dns-default-jm4gb\" (UID: \"c65d29d9-66e3-41ce-acb2-cec28c52fb29\") " pod="openshift-dns/dns-default-jm4gb" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.243606 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-9hcbc"] Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.243760 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/94a883ac-bee2-403e-8747-a4045e7defea-config\") pod \"kube-controller-manager-operator-78b949d7b-gqn9s\" (UID: \"94a883ac-bee2-403e-8747-a4045e7defea\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-gqn9s" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.243798 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/41aa056a-34d5-463b-8352-c71e0f0db1a4-config-volume\") pod \"collect-profiles-29327445-dn5mf\" (UID: \"41aa056a-34d5-463b-8352-c71e0f0db1a4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327445-dn5mf" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.244531 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/b21ccaf7-40a5-4939-ad8b-4c7efdcf4306-proxy-tls\") pod \"machine-config-controller-84d6567774-m7gph\" (UID: \"b21ccaf7-40a5-4939-ad8b-4c7efdcf4306\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-m7gph" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.245214 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/c041548e-b7c2-4f61-aaea-0b76e298ff22-tmpfs\") pod \"packageserver-d55dfcdfc-4pvwg\" (UID: \"c041548e-b7c2-4f61-aaea-0b76e298ff22\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-4pvwg" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.248708 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/2ecfaba9-9034-4ac7-bd51-e3f64b41d825-stats-auth\") pod \"router-default-5444994796-58hwl\" (UID: \"2ecfaba9-9034-4ac7-bd51-e3f64b41d825\") " pod="openshift-ingress/router-default-5444994796-58hwl" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.250661 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/2ecfaba9-9034-4ac7-bd51-e3f64b41d825-default-certificate\") pod \"router-default-5444994796-58hwl\" (UID: \"2ecfaba9-9034-4ac7-bd51-e3f64b41d825\") " pod="openshift-ingress/router-default-5444994796-58hwl" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.250816 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/c041548e-b7c2-4f61-aaea-0b76e298ff22-webhook-cert\") pod \"packageserver-d55dfcdfc-4pvwg\" (UID: \"c041548e-b7c2-4f61-aaea-0b76e298ff22\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-4pvwg" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.252599 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/80c88b0b-c4cf-478e-8e05-4cde023d15ee-certs\") pod \"machine-config-server-2g2rj\" (UID: \"80c88b0b-c4cf-478e-8e05-4cde023d15ee\") " pod="openshift-machine-config-operator/machine-config-server-2g2rj" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.252927 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/80ceab29-9389-4191-a909-8a6c40bba1cd-srv-cert\") pod \"catalog-operator-68c6474976-hmzsn\" (UID: \"80ceab29-9389-4191-a909-8a6c40bba1cd\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-hmzsn" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.253804 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/2ecfaba9-9034-4ac7-bd51-e3f64b41d825-metrics-certs\") pod \"router-default-5444994796-58hwl\" (UID: \"2ecfaba9-9034-4ac7-bd51-e3f64b41d825\") " pod="openshift-ingress/router-default-5444994796-58hwl" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.253828 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8b7fb5f2-4b5e-4fc3-9c97-dfbe36646871-serving-cert\") pod \"service-ca-operator-777779d784-5rj2f\" (UID: \"8b7fb5f2-4b5e-4fc3-9c97-dfbe36646871\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-5rj2f" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.254218 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/94a883ac-bee2-403e-8747-a4045e7defea-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-gqn9s\" (UID: \"94a883ac-bee2-403e-8747-a4045e7defea\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-gqn9s" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.254259 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/0cf52841-edf1-4daf-8e2c-76f5f6b93104-profile-collector-cert\") pod \"olm-operator-6b444d44fb-zf4p8\" (UID: \"0cf52841-edf1-4daf-8e2c-76f5f6b93104\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-zf4p8" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.254409 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/57e551f8-d9da-4692-b8ae-f7ff6829fe25-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-2x9nl\" (UID: \"57e551f8-d9da-4692-b8ae-f7ff6829fe25\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-2x9nl" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.254425 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/8c2f8349-aa93-4ac8-a369-5b4b2b9cd6e2-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-2rc5n\" (UID: \"8c2f8349-aa93-4ac8-a369-5b4b2b9cd6e2\") " pod="openshift-marketplace/marketplace-operator-79b997595-2rc5n" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.254628 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/80c88b0b-c4cf-478e-8e05-4cde023d15ee-node-bootstrap-token\") pod \"machine-config-server-2g2rj\" (UID: \"80c88b0b-c4cf-478e-8e05-4cde023d15ee\") " pod="openshift-machine-config-operator/machine-config-server-2g2rj" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.254683 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/80ceab29-9389-4191-a909-8a6c40bba1cd-profile-collector-cert\") pod \"catalog-operator-68c6474976-hmzsn\" (UID: \"80ceab29-9389-4191-a909-8a6c40bba1cd\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-hmzsn" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.254932 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/41aa056a-34d5-463b-8352-c71e0f0db1a4-secret-volume\") pod \"collect-profiles-29327445-dn5mf\" (UID: \"41aa056a-34d5-463b-8352-c71e0f0db1a4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327445-dn5mf" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.255222 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/e532fab3-55a6-4c42-a532-9291452d5d82-cert\") pod \"ingress-canary-4cxsw\" (UID: \"e532fab3-55a6-4c42-a532-9291452d5d82\") " pod="openshift-ingress-canary/ingress-canary-4cxsw" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.256821 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/788472f5-2ad1-49d0-8e31-6a4fe14c3c85-proxy-tls\") pod \"machine-config-operator-74547568cd-wnmjw\" (UID: \"788472f5-2ad1-49d0-8e31-6a4fe14c3c85\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-wnmjw" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.256921 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/74822779-bdcd-4aae-bc01-9cbcbb0c1d83-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-x5n4h\" (UID: \"74822779-bdcd-4aae-bc01-9cbcbb0c1d83\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-x5n4h" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.257744 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/31d92551-cb79-4d9c-ad1c-c5090142b82c-signing-key\") pod \"service-ca-9c57cc56f-bpbtf\" (UID: \"31d92551-cb79-4d9c-ad1c-c5090142b82c\") " pod="openshift-service-ca/service-ca-9c57cc56f-bpbtf" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.261474 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9tjtl\" (UniqueName: \"kubernetes.io/projected/54f93640-64fd-4728-a83b-499aee1ffa38-kube-api-access-9tjtl\") pod \"migrator-59844c95c7-skh8h\" (UID: \"54f93640-64fd-4728-a83b-499aee1ffa38\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-skh8h" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.266785 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-m467d"] Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.280156 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-stjw8"] Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.280522 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mtl7j\" (UniqueName: \"kubernetes.io/projected/78d3e7b1-7258-4508-ad38-373f21bb5d96-kube-api-access-mtl7j\") pod \"image-registry-697d97f7c8-rf9l4\" (UID: \"78d3e7b1-7258-4508-ad38-373f21bb5d96\") " pod="openshift-image-registry/image-registry-697d97f7c8-rf9l4" Oct 05 06:50:20 crc kubenswrapper[4846]: W1005 06:50:20.291285 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod74f16025_fe11_4a23_8c1a_ff3da2a3d751.slice/crio-6a2010efd88b09b2d8cf8c289be1ddd4504bd85b79adc3f42557ed0bb350b4f8 WatchSource:0}: Error finding container 6a2010efd88b09b2d8cf8c289be1ddd4504bd85b79adc3f42557ed0bb350b4f8: Status 404 returned error can't find the container with id 6a2010efd88b09b2d8cf8c289be1ddd4504bd85b79adc3f42557ed0bb350b4f8 Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.299378 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9vncv\" (UniqueName: \"kubernetes.io/projected/bc560c3d-6558-43e6-86ca-312c9fa264ea-kube-api-access-9vncv\") pod \"dns-operator-744455d44c-5qwb9\" (UID: \"bc560c3d-6558-43e6-86ca-312c9fa264ea\") " pod="openshift-dns-operator/dns-operator-744455d44c-5qwb9" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.304241 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-5qwb9" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.314778 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-8vjkd" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.317836 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vw2sb\" (UniqueName: \"kubernetes.io/projected/18507a98-2ef8-43b1-b94d-51106f9bdd3b-kube-api-access-vw2sb\") pod \"console-operator-58897d9998-t8f4h\" (UID: \"18507a98-2ef8-43b1-b94d-51106f9bdd3b\") " pod="openshift-console-operator/console-operator-58897d9998-t8f4h" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.326947 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rf9l4\" (UID: \"78d3e7b1-7258-4508-ad38-373f21bb5d96\") " pod="openshift-image-registry/image-registry-697d97f7c8-rf9l4" Oct 05 06:50:20 crc kubenswrapper[4846]: E1005 06:50:20.327333 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-05 06:50:20.827317661 +0000 UTC m=+143.068170436 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rf9l4" (UID: "78d3e7b1-7258-4508-ad38-373f21bb5d96") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.327365 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-skh8h" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.336570 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7a19cb1c-5274-48e7-ad86-3c2ff85ecf8c-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-7jflr\" (UID: \"7a19cb1c-5274-48e7-ad86-3c2ff85ecf8c\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-7jflr" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.358339 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-vwktr"] Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.361683 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-6t2mt"] Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.368863 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6gnmv\" (UniqueName: \"kubernetes.io/projected/53168eca-95fe-4303-b401-380995e5adcb-kube-api-access-6gnmv\") pod \"openshift-controller-manager-operator-756b6f6bc6-526w4\" (UID: \"53168eca-95fe-4303-b401-380995e5adcb\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-526w4" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.371752 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-9hcbc" event={"ID":"74f16025-fe11-4a23-8c1a-ff3da2a3d751","Type":"ContainerStarted","Data":"6a2010efd88b09b2d8cf8c289be1ddd4504bd85b79adc3f42557ed0bb350b4f8"} Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.372348 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-wcg6b"] Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.375042 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-rcnlv" event={"ID":"53fe640c-6cb7-43a3-9a84-467eb36fe66b","Type":"ContainerStarted","Data":"380db708741aded66247576c0551e6baa89c356523a3af3d9de632ecc6c2b8d4"} Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.378645 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-stjw8" event={"ID":"e4ea5d4f-3481-4e72-9bd8-e5d8183bc687","Type":"ContainerStarted","Data":"4a7df794ded57327e1296f18bedcfaeeb0e41ef92a1125c7651ecb7a9d563216"} Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.382279 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-pglrh" event={"ID":"5bddce7c-b835-4a1f-a3af-db5a1d3771f2","Type":"ContainerStarted","Data":"a7844d2c8a706845d237cc7e460f547241d5b84ce18cd3365e6234685157a3fc"} Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.386084 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5422n\" (UniqueName: \"kubernetes.io/projected/5fed566e-ecb6-4801-84b3-ad1143a03f08-kube-api-access-5422n\") pod \"csi-hostpathplugin-dsqhc\" (UID: \"5fed566e-ecb6-4801-84b3-ad1143a03f08\") " pod="hostpath-provisioner/csi-hostpathplugin-dsqhc" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.386900 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-m467d" event={"ID":"b1cd661a-c160-4e14-81cf-842d3b9ae01d","Type":"ContainerStarted","Data":"1f62953c1e0f209ee2fa15585e68ecb9a72d608dc32338e571eb2c20639b7beb"} Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.405357 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zwxrp\" (UniqueName: \"kubernetes.io/projected/41aa056a-34d5-463b-8352-c71e0f0db1a4-kube-api-access-zwxrp\") pod \"collect-profiles-29327445-dn5mf\" (UID: \"41aa056a-34d5-463b-8352-c71e0f0db1a4\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327445-dn5mf" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.427713 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 05 06:50:20 crc kubenswrapper[4846]: E1005 06:50:20.431079 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-05 06:50:20.931054633 +0000 UTC m=+143.171907408 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.452074 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9mht8\" (UniqueName: \"kubernetes.io/projected/b21ccaf7-40a5-4939-ad8b-4c7efdcf4306-kube-api-access-9mht8\") pod \"machine-config-controller-84d6567774-m7gph\" (UID: \"b21ccaf7-40a5-4939-ad8b-4c7efdcf4306\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-m7gph" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.460999 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9dm98\" (UniqueName: \"kubernetes.io/projected/74822779-bdcd-4aae-bc01-9cbcbb0c1d83-kube-api-access-9dm98\") pod \"kube-storage-version-migrator-operator-b67b599dd-x5n4h\" (UID: \"74822779-bdcd-4aae-bc01-9cbcbb0c1d83\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-x5n4h" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.485938 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-9jtcc"] Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.486967 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/beeea3e9-d3c3-4280-bf04-85cee4378cb0-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-8bpqg\" (UID: \"beeea3e9-d3c3-4280-bf04-85cee4378cb0\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-8bpqg" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.505025 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-vwx77"] Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.507005 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-6bpvf"] Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.507428 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-t8f4h" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.509336 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fv4fz\" (UniqueName: \"kubernetes.io/projected/788472f5-2ad1-49d0-8e31-6a4fe14c3c85-kube-api-access-fv4fz\") pod \"machine-config-operator-74547568cd-wnmjw\" (UID: \"788472f5-2ad1-49d0-8e31-6a4fe14c3c85\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-wnmjw" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.513003 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-c6cxs"] Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.521740 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sl5t6\" (UniqueName: \"kubernetes.io/projected/0cf52841-edf1-4daf-8e2c-76f5f6b93104-kube-api-access-sl5t6\") pod \"olm-operator-6b444d44fb-zf4p8\" (UID: \"0cf52841-edf1-4daf-8e2c-76f5f6b93104\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-zf4p8" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.526375 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-2pd47"] Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.529326 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rf9l4\" (UID: \"78d3e7b1-7258-4508-ad38-373f21bb5d96\") " pod="openshift-image-registry/image-registry-697d97f7c8-rf9l4" Oct 05 06:50:20 crc kubenswrapper[4846]: E1005 06:50:20.529779 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-05 06:50:21.029764264 +0000 UTC m=+143.270617039 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rf9l4" (UID: "78d3e7b1-7258-4508-ad38-373f21bb5d96") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.539476 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wth4v\" (UniqueName: \"kubernetes.io/projected/e532fab3-55a6-4c42-a532-9291452d5d82-kube-api-access-wth4v\") pod \"ingress-canary-4cxsw\" (UID: \"e532fab3-55a6-4c42-a532-9291452d5d82\") " pod="openshift-ingress-canary/ingress-canary-4cxsw" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.545140 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-wnmjw" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.549854 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-526w4" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.556824 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29327445-dn5mf" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.562281 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6srxd\" (UniqueName: \"kubernetes.io/projected/c041548e-b7c2-4f61-aaea-0b76e298ff22-kube-api-access-6srxd\") pod \"packageserver-d55dfcdfc-4pvwg\" (UID: \"c041548e-b7c2-4f61-aaea-0b76e298ff22\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-4pvwg" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.570810 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-4cxsw" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.578691 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-vjmbx"] Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.587033 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mxbkd\" (UniqueName: \"kubernetes.io/projected/2ecfaba9-9034-4ac7-bd51-e3f64b41d825-kube-api-access-mxbkd\") pod \"router-default-5444994796-58hwl\" (UID: \"2ecfaba9-9034-4ac7-bd51-e3f64b41d825\") " pod="openshift-ingress/router-default-5444994796-58hwl" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.599404 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-dsqhc" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.617402 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6hlbm\" (UniqueName: \"kubernetes.io/projected/80ceab29-9389-4191-a909-8a6c40bba1cd-kube-api-access-6hlbm\") pod \"catalog-operator-68c6474976-hmzsn\" (UID: \"80ceab29-9389-4191-a909-8a6c40bba1cd\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-hmzsn" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.619907 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7ml2d\" (UniqueName: \"kubernetes.io/projected/8b7fb5f2-4b5e-4fc3-9c97-dfbe36646871-kube-api-access-7ml2d\") pod \"service-ca-operator-777779d784-5rj2f\" (UID: \"8b7fb5f2-4b5e-4fc3-9c97-dfbe36646871\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-5rj2f" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.620510 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-7jflr" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.630523 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 05 06:50:20 crc kubenswrapper[4846]: E1005 06:50:20.630720 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-05 06:50:21.130686432 +0000 UTC m=+143.371539197 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.631076 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rf9l4\" (UID: \"78d3e7b1-7258-4508-ad38-373f21bb5d96\") " pod="openshift-image-registry/image-registry-697d97f7c8-rf9l4" Oct 05 06:50:20 crc kubenswrapper[4846]: E1005 06:50:20.631433 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-05 06:50:21.131417871 +0000 UTC m=+143.372270646 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rf9l4" (UID: "78d3e7b1-7258-4508-ad38-373f21bb5d96") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.641970 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ppwzm\" (UniqueName: \"kubernetes.io/projected/31d92551-cb79-4d9c-ad1c-c5090142b82c-kube-api-access-ppwzm\") pod \"service-ca-9c57cc56f-bpbtf\" (UID: \"31d92551-cb79-4d9c-ad1c-c5090142b82c\") " pod="openshift-service-ca/service-ca-9c57cc56f-bpbtf" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.650339 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-k87kc"] Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.651641 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-x5n4h" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.664911 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/94a883ac-bee2-403e-8747-a4045e7defea-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-gqn9s\" (UID: \"94a883ac-bee2-403e-8747-a4045e7defea\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-gqn9s" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.678061 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-8bpqg" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.687716 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-4pvwg" Oct 05 06:50:20 crc kubenswrapper[4846]: W1005 06:50:20.694365 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod75d0c362_7aa3_4919_937d_8849f425466e.slice/crio-a0632a8deb90454689a3d420c13cd01fbf7e8f006e6a331e692b82717c03e024 WatchSource:0}: Error finding container a0632a8deb90454689a3d420c13cd01fbf7e8f006e6a331e692b82717c03e024: Status 404 returned error can't find the container with id a0632a8deb90454689a3d420c13cd01fbf7e8f006e6a331e692b82717c03e024 Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.697824 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4d7ch\" (UniqueName: \"kubernetes.io/projected/80c88b0b-c4cf-478e-8e05-4cde023d15ee-kube-api-access-4d7ch\") pod \"machine-config-server-2g2rj\" (UID: \"80c88b0b-c4cf-478e-8e05-4cde023d15ee\") " pod="openshift-machine-config-operator/machine-config-server-2g2rj" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.698373 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-trr7q\" (UniqueName: \"kubernetes.io/projected/8c2f8349-aa93-4ac8-a369-5b4b2b9cd6e2-kube-api-access-trr7q\") pod \"marketplace-operator-79b997595-2rc5n\" (UID: \"8c2f8349-aa93-4ac8-a369-5b4b2b9cd6e2\") " pod="openshift-marketplace/marketplace-operator-79b997595-2rc5n" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.699775 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-m7gph" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.706912 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-58hwl" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.718833 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-hmzsn" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.721957 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-5qwb9"] Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.723550 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9h8b2\" (UniqueName: \"kubernetes.io/projected/57e551f8-d9da-4692-b8ae-f7ff6829fe25-kube-api-access-9h8b2\") pod \"control-plane-machine-set-operator-78cbb6b69f-2x9nl\" (UID: \"57e551f8-d9da-4692-b8ae-f7ff6829fe25\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-2x9nl" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.728367 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-zf4p8" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.731626 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 05 06:50:20 crc kubenswrapper[4846]: E1005 06:50:20.732135 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-05 06:50:21.232117004 +0000 UTC m=+143.472969779 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.737569 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-2rc5n" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.754540 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-bpbtf" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.760962 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s8g9n\" (UniqueName: \"kubernetes.io/projected/0755bb69-c94a-4b36-97ee-8a3dfd972c79-kube-api-access-s8g9n\") pod \"multus-admission-controller-857f4d67dd-dtbh5\" (UID: \"0755bb69-c94a-4b36-97ee-8a3dfd972c79\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-dtbh5" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.768976 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ghg5r\" (UniqueName: \"kubernetes.io/projected/8049dd16-5d6f-4718-94b1-fb5a322c2530-kube-api-access-ghg5r\") pod \"package-server-manager-789f6589d5-pqm4r\" (UID: \"8049dd16-5d6f-4718-94b1-fb5a322c2530\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-pqm4r" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.777124 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-8vjkd"] Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.781825 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-2x9nl" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.783382 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wsj78\" (UniqueName: \"kubernetes.io/projected/c65d29d9-66e3-41ce-acb2-cec28c52fb29-kube-api-access-wsj78\") pod \"dns-default-jm4gb\" (UID: \"c65d29d9-66e3-41ce-acb2-cec28c52fb29\") " pod="openshift-dns/dns-default-jm4gb" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.819651 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-dtbh5" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.833599 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rf9l4\" (UID: \"78d3e7b1-7258-4508-ad38-373f21bb5d96\") " pod="openshift-image-registry/image-registry-697d97f7c8-rf9l4" Oct 05 06:50:20 crc kubenswrapper[4846]: E1005 06:50:20.833948 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-05 06:50:21.333934416 +0000 UTC m=+143.574787181 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rf9l4" (UID: "78d3e7b1-7258-4508-ad38-373f21bb5d96") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.834425 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-5rj2f" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.835051 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-skh8h"] Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.875451 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-t8f4h"] Oct 05 06:50:20 crc kubenswrapper[4846]: W1005 06:50:20.890440 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbc560c3d_6558_43e6_86ca_312c9fa264ea.slice/crio-c37daf0575e737d41228b08f840fdb022377e89dbdca02d5c83751e31b1d79f5 WatchSource:0}: Error finding container c37daf0575e737d41228b08f840fdb022377e89dbdca02d5c83751e31b1d79f5: Status 404 returned error can't find the container with id c37daf0575e737d41228b08f840fdb022377e89dbdca02d5c83751e31b1d79f5 Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.911161 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-jm4gb" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.915811 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-2g2rj" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.934205 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 05 06:50:20 crc kubenswrapper[4846]: E1005 06:50:20.934607 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-05 06:50:21.434589727 +0000 UTC m=+143.675442502 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 06:50:20 crc kubenswrapper[4846]: W1005 06:50:20.936361 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod54f93640_64fd_4728_a83b_499aee1ffa38.slice/crio-71255268760fbe7c89d7fe8df2c3dc0e6df586c59c9cfe0e63da53b00ca74825 WatchSource:0}: Error finding container 71255268760fbe7c89d7fe8df2c3dc0e6df586c59c9cfe0e63da53b00ca74825: Status 404 returned error can't find the container with id 71255268760fbe7c89d7fe8df2c3dc0e6df586c59c9cfe0e63da53b00ca74825 Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.936509 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-gqn9s" Oct 05 06:50:20 crc kubenswrapper[4846]: I1005 06:50:20.958887 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-wnmjw"] Oct 05 06:50:21 crc kubenswrapper[4846]: W1005 06:50:21.023399 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod788472f5_2ad1_49d0_8e31_6a4fe14c3c85.slice/crio-96cfb5f979e3b2329c6cf52635aa5883711019f5164324d730ccec1393650a5f WatchSource:0}: Error finding container 96cfb5f979e3b2329c6cf52635aa5883711019f5164324d730ccec1393650a5f: Status 404 returned error can't find the container with id 96cfb5f979e3b2329c6cf52635aa5883711019f5164324d730ccec1393650a5f Oct 05 06:50:21 crc kubenswrapper[4846]: I1005 06:50:21.038843 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rf9l4\" (UID: \"78d3e7b1-7258-4508-ad38-373f21bb5d96\") " pod="openshift-image-registry/image-registry-697d97f7c8-rf9l4" Oct 05 06:50:21 crc kubenswrapper[4846]: E1005 06:50:21.039242 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-05 06:50:21.539228113 +0000 UTC m=+143.780080888 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rf9l4" (UID: "78d3e7b1-7258-4508-ad38-373f21bb5d96") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 06:50:21 crc kubenswrapper[4846]: I1005 06:50:21.042740 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29327445-dn5mf"] Oct 05 06:50:21 crc kubenswrapper[4846]: I1005 06:50:21.045382 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-pqm4r" Oct 05 06:50:21 crc kubenswrapper[4846]: I1005 06:50:21.085317 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-4cxsw"] Oct 05 06:50:21 crc kubenswrapper[4846]: I1005 06:50:21.139907 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 05 06:50:21 crc kubenswrapper[4846]: E1005 06:50:21.140081 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-05 06:50:21.640044259 +0000 UTC m=+143.880897034 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 06:50:21 crc kubenswrapper[4846]: I1005 06:50:21.140295 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rf9l4\" (UID: \"78d3e7b1-7258-4508-ad38-373f21bb5d96\") " pod="openshift-image-registry/image-registry-697d97f7c8-rf9l4" Oct 05 06:50:21 crc kubenswrapper[4846]: E1005 06:50:21.140662 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-05 06:50:21.640642995 +0000 UTC m=+143.881495780 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rf9l4" (UID: "78d3e7b1-7258-4508-ad38-373f21bb5d96") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 06:50:21 crc kubenswrapper[4846]: I1005 06:50:21.167323 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-526w4"] Oct 05 06:50:21 crc kubenswrapper[4846]: I1005 06:50:21.240902 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 05 06:50:21 crc kubenswrapper[4846]: E1005 06:50:21.241337 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-05 06:50:21.741317767 +0000 UTC m=+143.982170552 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 06:50:21 crc kubenswrapper[4846]: I1005 06:50:21.343583 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rf9l4\" (UID: \"78d3e7b1-7258-4508-ad38-373f21bb5d96\") " pod="openshift-image-registry/image-registry-697d97f7c8-rf9l4" Oct 05 06:50:21 crc kubenswrapper[4846]: E1005 06:50:21.344907 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-05 06:50:21.844870334 +0000 UTC m=+144.085723099 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rf9l4" (UID: "78d3e7b1-7258-4508-ad38-373f21bb5d96") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 06:50:21 crc kubenswrapper[4846]: I1005 06:50:21.366187 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-7jflr"] Oct 05 06:50:21 crc kubenswrapper[4846]: I1005 06:50:21.403916 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-8vjkd" event={"ID":"cb184e9f-f0d9-4166-b0da-213c5c7aba5f","Type":"ContainerStarted","Data":"10e6bdd74da43987e221d6dd2b42979689456509ca1d9453dfa2f68bcce064ba"} Oct 05 06:50:21 crc kubenswrapper[4846]: I1005 06:50:21.407429 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-vwx77" event={"ID":"f743f03b-3f08-4da0-b53e-1cb33ab516b9","Type":"ContainerStarted","Data":"5a354f2e3c97a1ea3cc6caee1593b2ac944bce6789e9f51c51b1e01e65973eee"} Oct 05 06:50:21 crc kubenswrapper[4846]: I1005 06:50:21.411146 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-9jtcc" event={"ID":"7b82343c-28e6-4e0b-9731-5131cb0014f2","Type":"ContainerStarted","Data":"eac7ea19d1339b9947130fc204fa3b9a36cd1ecc9e143ffe946bf3ce777e2d3f"} Oct 05 06:50:21 crc kubenswrapper[4846]: I1005 06:50:21.411303 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-9jtcc" event={"ID":"7b82343c-28e6-4e0b-9731-5131cb0014f2","Type":"ContainerStarted","Data":"5fcce8a0d40f1619b5d8e1b6b84851a76d0c0799b171f7001abe393df79b740f"} Oct 05 06:50:21 crc kubenswrapper[4846]: I1005 06:50:21.412714 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-6bpvf" event={"ID":"75d0c362-7aa3-4919-937d-8849f425466e","Type":"ContainerStarted","Data":"a0632a8deb90454689a3d420c13cd01fbf7e8f006e6a331e692b82717c03e024"} Oct 05 06:50:21 crc kubenswrapper[4846]: I1005 06:50:21.422489 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-6t2mt" event={"ID":"770ae4fc-21dd-446d-9a41-26e484687787","Type":"ContainerStarted","Data":"81328a495c32d2ab0a60dd37759f9f0b77921a3f7df6bef74ba81ecf88e84e5d"} Oct 05 06:50:21 crc kubenswrapper[4846]: I1005 06:50:21.422911 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-6t2mt" event={"ID":"770ae4fc-21dd-446d-9a41-26e484687787","Type":"ContainerStarted","Data":"eab4cdc88060cdf416b1f6d3d9f5545c5bc5357725f772908b1c8cd0490ed45a"} Oct 05 06:50:21 crc kubenswrapper[4846]: I1005 06:50:21.422929 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-6t2mt" Oct 05 06:50:21 crc kubenswrapper[4846]: I1005 06:50:21.431132 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-2pd47" event={"ID":"6ec3075e-f780-435f-8efa-30d2fb4855a1","Type":"ContainerStarted","Data":"f71ac859cc6bf3200d6d39013648bb5bab2cc5a7b0bd8b17135ea52208de831a"} Oct 05 06:50:21 crc kubenswrapper[4846]: I1005 06:50:21.436778 4846 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-6t2mt container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.10:6443/healthz\": dial tcp 10.217.0.10:6443: connect: connection refused" start-of-body= Oct 05 06:50:21 crc kubenswrapper[4846]: I1005 06:50:21.436857 4846 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-6t2mt" podUID="770ae4fc-21dd-446d-9a41-26e484687787" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.10:6443/healthz\": dial tcp 10.217.0.10:6443: connect: connection refused" Oct 05 06:50:21 crc kubenswrapper[4846]: I1005 06:50:21.439255 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-dsqhc"] Oct 05 06:50:21 crc kubenswrapper[4846]: I1005 06:50:21.440367 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-wcg6b" event={"ID":"85542c44-c09c-4c39-851f-fd320a0a36b2","Type":"ContainerStarted","Data":"6f16f569942da4e6ea0d8f4131f267f33dee8bc96f5c2f63d8a8f6734e9eea2b"} Oct 05 06:50:21 crc kubenswrapper[4846]: I1005 06:50:21.440398 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-wcg6b" event={"ID":"85542c44-c09c-4c39-851f-fd320a0a36b2","Type":"ContainerStarted","Data":"641de3ba288a2e242ab51ddf40a5b80d00782d89059505a4fdb2fd2e55eea8bf"} Oct 05 06:50:21 crc kubenswrapper[4846]: I1005 06:50:21.442421 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-pglrh" event={"ID":"5bddce7c-b835-4a1f-a3af-db5a1d3771f2","Type":"ContainerStarted","Data":"005daad82c298f6bb8fd0e78a27cfaf98d83d9cb719756c5873f319c0dcb287b"} Oct 05 06:50:21 crc kubenswrapper[4846]: I1005 06:50:21.443064 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29327445-dn5mf" event={"ID":"41aa056a-34d5-463b-8352-c71e0f0db1a4","Type":"ContainerStarted","Data":"e2a9a7d69bb7c87fcf725e311147d7f32065f9ed22017993686f8049925fc55d"} Oct 05 06:50:21 crc kubenswrapper[4846]: I1005 06:50:21.444975 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 05 06:50:21 crc kubenswrapper[4846]: I1005 06:50:21.445000 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-stjw8" event={"ID":"e4ea5d4f-3481-4e72-9bd8-e5d8183bc687","Type":"ContainerDied","Data":"ece9d7aacd2b229f40fa2f9bf2944a5d6782374b27705cfa4b6408c2b4e0b7e1"} Oct 05 06:50:21 crc kubenswrapper[4846]: I1005 06:50:21.444964 4846 generic.go:334] "Generic (PLEG): container finished" podID="e4ea5d4f-3481-4e72-9bd8-e5d8183bc687" containerID="ece9d7aacd2b229f40fa2f9bf2944a5d6782374b27705cfa4b6408c2b4e0b7e1" exitCode=0 Oct 05 06:50:21 crc kubenswrapper[4846]: E1005 06:50:21.445504 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-05 06:50:21.945485494 +0000 UTC m=+144.186338269 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 06:50:21 crc kubenswrapper[4846]: I1005 06:50:21.448360 4846 generic.go:334] "Generic (PLEG): container finished" podID="53fe640c-6cb7-43a3-9a84-467eb36fe66b" containerID="522f9924601ef9012752f0a8052f5fa03b417afd028de13c6017e3b2414d38f0" exitCode=0 Oct 05 06:50:21 crc kubenswrapper[4846]: I1005 06:50:21.448921 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-rcnlv" event={"ID":"53fe640c-6cb7-43a3-9a84-467eb36fe66b","Type":"ContainerDied","Data":"522f9924601ef9012752f0a8052f5fa03b417afd028de13c6017e3b2414d38f0"} Oct 05 06:50:21 crc kubenswrapper[4846]: I1005 06:50:21.470038 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-k87kc" event={"ID":"da8eba09-f287-4704-b990-348ff69f0173","Type":"ContainerStarted","Data":"8fe871cd18ddce7d8e9f41b427779b2723a816dcd12d49ceecc245fb09668223"} Oct 05 06:50:21 crc kubenswrapper[4846]: I1005 06:50:21.476714 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-m467d" event={"ID":"b1cd661a-c160-4e14-81cf-842d3b9ae01d","Type":"ContainerStarted","Data":"73fba6698c81c29383f2fed216fba199cf0a298e2e3421120ae9403bbd938fc3"} Oct 05 06:50:21 crc kubenswrapper[4846]: I1005 06:50:21.487026 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-wnmjw" event={"ID":"788472f5-2ad1-49d0-8e31-6a4fe14c3c85","Type":"ContainerStarted","Data":"96cfb5f979e3b2329c6cf52635aa5883711019f5164324d730ccec1393650a5f"} Oct 05 06:50:21 crc kubenswrapper[4846]: I1005 06:50:21.512672 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-c6cxs" event={"ID":"475e1ca4-a6fa-4f77-8c89-0619b2e9681e","Type":"ContainerStarted","Data":"160f176dda448ff5392064cf1df6cda816c8335a84daeffad6d71c8b0a37ddf2"} Oct 05 06:50:21 crc kubenswrapper[4846]: I1005 06:50:21.512729 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-c6cxs" event={"ID":"475e1ca4-a6fa-4f77-8c89-0619b2e9681e","Type":"ContainerStarted","Data":"20ac664fcb9b96ed51c70aee8524cd88f5eec0b78bec80cd40964c2a577fd8e4"} Oct 05 06:50:21 crc kubenswrapper[4846]: I1005 06:50:21.513375 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-c6cxs" Oct 05 06:50:21 crc kubenswrapper[4846]: I1005 06:50:21.536240 4846 patch_prober.go:28] interesting pod/downloads-7954f5f757-c6cxs container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.12:8080/\": dial tcp 10.217.0.12:8080: connect: connection refused" start-of-body= Oct 05 06:50:21 crc kubenswrapper[4846]: I1005 06:50:21.536298 4846 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-c6cxs" podUID="475e1ca4-a6fa-4f77-8c89-0619b2e9681e" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.12:8080/\": dial tcp 10.217.0.12:8080: connect: connection refused" Oct 05 06:50:21 crc kubenswrapper[4846]: I1005 06:50:21.547399 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rf9l4\" (UID: \"78d3e7b1-7258-4508-ad38-373f21bb5d96\") " pod="openshift-image-registry/image-registry-697d97f7c8-rf9l4" Oct 05 06:50:21 crc kubenswrapper[4846]: E1005 06:50:21.550697 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-05 06:50:22.050680515 +0000 UTC m=+144.291533290 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rf9l4" (UID: "78d3e7b1-7258-4508-ad38-373f21bb5d96") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 06:50:21 crc kubenswrapper[4846]: I1005 06:50:21.588894 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-9hcbc" event={"ID":"74f16025-fe11-4a23-8c1a-ff3da2a3d751","Type":"ContainerStarted","Data":"8b88162ddf15a27840637668117eb2e0ff0a349ff0b37277d499b60fec10c880"} Oct 05 06:50:21 crc kubenswrapper[4846]: I1005 06:50:21.603056 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-4cxsw" event={"ID":"e532fab3-55a6-4c42-a532-9291452d5d82","Type":"ContainerStarted","Data":"fd23108f2ac2cd8440bc91003b316725ce461ed8c35e22496276638fd07f48d4"} Oct 05 06:50:21 crc kubenswrapper[4846]: I1005 06:50:21.606780 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-5qwb9" event={"ID":"bc560c3d-6558-43e6-86ca-312c9fa264ea","Type":"ContainerStarted","Data":"c37daf0575e737d41228b08f840fdb022377e89dbdca02d5c83751e31b1d79f5"} Oct 05 06:50:21 crc kubenswrapper[4846]: I1005 06:50:21.624246 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-vwktr" event={"ID":"beec0ea6-0c57-44d6-bfef-d00e5b3724ad","Type":"ContainerStarted","Data":"572da8558f11938b557fcd55ba9ad07a9e2c86156f81abfce221a75fa0f165ab"} Oct 05 06:50:21 crc kubenswrapper[4846]: I1005 06:50:21.624294 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-vwktr" event={"ID":"beec0ea6-0c57-44d6-bfef-d00e5b3724ad","Type":"ContainerStarted","Data":"d48285c00b668451b292a1fae7b40af066195b9dc894206d63635b4698949e65"} Oct 05 06:50:21 crc kubenswrapper[4846]: I1005 06:50:21.624930 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-vwktr" Oct 05 06:50:21 crc kubenswrapper[4846]: I1005 06:50:21.626548 4846 patch_prober.go:28] interesting pod/route-controller-manager-6576b87f9c-vwktr container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.7:8443/healthz\": dial tcp 10.217.0.7:8443: connect: connection refused" start-of-body= Oct 05 06:50:21 crc kubenswrapper[4846]: I1005 06:50:21.626642 4846 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-vwktr" podUID="beec0ea6-0c57-44d6-bfef-d00e5b3724ad" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.7:8443/healthz\": dial tcp 10.217.0.7:8443: connect: connection refused" Oct 05 06:50:21 crc kubenswrapper[4846]: I1005 06:50:21.632414 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-x5n4h"] Oct 05 06:50:21 crc kubenswrapper[4846]: I1005 06:50:21.638339 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-skh8h" event={"ID":"54f93640-64fd-4728-a83b-499aee1ffa38","Type":"ContainerStarted","Data":"71255268760fbe7c89d7fe8df2c3dc0e6df586c59c9cfe0e63da53b00ca74825"} Oct 05 06:50:21 crc kubenswrapper[4846]: I1005 06:50:21.671896 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 05 06:50:21 crc kubenswrapper[4846]: E1005 06:50:21.672154 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-05 06:50:22.172135812 +0000 UTC m=+144.412988587 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 06:50:21 crc kubenswrapper[4846]: I1005 06:50:21.672347 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rf9l4\" (UID: \"78d3e7b1-7258-4508-ad38-373f21bb5d96\") " pod="openshift-image-registry/image-registry-697d97f7c8-rf9l4" Oct 05 06:50:21 crc kubenswrapper[4846]: E1005 06:50:21.674974 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-05 06:50:22.174964777 +0000 UTC m=+144.415817542 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rf9l4" (UID: "78d3e7b1-7258-4508-ad38-373f21bb5d96") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 06:50:21 crc kubenswrapper[4846]: I1005 06:50:21.675516 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-t8f4h" event={"ID":"18507a98-2ef8-43b1-b94d-51106f9bdd3b","Type":"ContainerStarted","Data":"c051895128f8333206944a82794a927cc17129f4246923af942e635214ae8f74"} Oct 05 06:50:21 crc kubenswrapper[4846]: I1005 06:50:21.682272 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-vjmbx" event={"ID":"6be3f74a-b215-4588-8b0f-735eba580bbd","Type":"ContainerStarted","Data":"38cfaabdd0e1e1e2fe9b14b6b4553dc224a05211452f25e0ccce1f49c8e68578"} Oct 05 06:50:21 crc kubenswrapper[4846]: I1005 06:50:21.682310 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-vjmbx" event={"ID":"6be3f74a-b215-4588-8b0f-735eba580bbd","Type":"ContainerStarted","Data":"87a40d3be84539a20530df2487213814433c61ad32b51f566c8a99041eb5102f"} Oct 05 06:50:21 crc kubenswrapper[4846]: I1005 06:50:21.683084 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-vjmbx" Oct 05 06:50:21 crc kubenswrapper[4846]: I1005 06:50:21.685012 4846 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-vjmbx container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.18:8443/healthz\": dial tcp 10.217.0.18:8443: connect: connection refused" start-of-body= Oct 05 06:50:21 crc kubenswrapper[4846]: I1005 06:50:21.685051 4846 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-vjmbx" podUID="6be3f74a-b215-4588-8b0f-735eba580bbd" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.18:8443/healthz\": dial tcp 10.217.0.18:8443: connect: connection refused" Oct 05 06:50:21 crc kubenswrapper[4846]: I1005 06:50:21.685155 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-526w4" event={"ID":"53168eca-95fe-4303-b401-380995e5adcb","Type":"ContainerStarted","Data":"ec9fb7f6b025cb230a968dddefc618da0569c3f871c57a6166fd0d55884637b8"} Oct 05 06:50:21 crc kubenswrapper[4846]: I1005 06:50:21.691796 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-2x9nl"] Oct 05 06:50:21 crc kubenswrapper[4846]: I1005 06:50:21.713529 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-zf4p8"] Oct 05 06:50:21 crc kubenswrapper[4846]: I1005 06:50:21.722933 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-8bpqg"] Oct 05 06:50:21 crc kubenswrapper[4846]: I1005 06:50:21.773564 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 05 06:50:21 crc kubenswrapper[4846]: E1005 06:50:21.774468 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-05 06:50:22.274434907 +0000 UTC m=+144.515287682 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 06:50:21 crc kubenswrapper[4846]: W1005 06:50:21.813708 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod74822779_bdcd_4aae_bc01_9cbcbb0c1d83.slice/crio-0fe693d1cbc0a346699c582a4acd50d6367c5c0efbcb096cee3f5fea7b3e8873 WatchSource:0}: Error finding container 0fe693d1cbc0a346699c582a4acd50d6367c5c0efbcb096cee3f5fea7b3e8873: Status 404 returned error can't find the container with id 0fe693d1cbc0a346699c582a4acd50d6367c5c0efbcb096cee3f5fea7b3e8873 Oct 05 06:50:21 crc kubenswrapper[4846]: I1005 06:50:21.865502 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-txzlk" Oct 05 06:50:21 crc kubenswrapper[4846]: I1005 06:50:21.895040 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-2rc5n"] Oct 05 06:50:21 crc kubenswrapper[4846]: I1005 06:50:21.895929 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rf9l4\" (UID: \"78d3e7b1-7258-4508-ad38-373f21bb5d96\") " pod="openshift-image-registry/image-registry-697d97f7c8-rf9l4" Oct 05 06:50:21 crc kubenswrapper[4846]: E1005 06:50:21.897775 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-05 06:50:22.397760163 +0000 UTC m=+144.638612938 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rf9l4" (UID: "78d3e7b1-7258-4508-ad38-373f21bb5d96") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 06:50:21 crc kubenswrapper[4846]: I1005 06:50:21.903001 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-dtbh5"] Oct 05 06:50:21 crc kubenswrapper[4846]: I1005 06:50:21.934491 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-m7gph"] Oct 05 06:50:21 crc kubenswrapper[4846]: I1005 06:50:21.962641 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-hmzsn"] Oct 05 06:50:21 crc kubenswrapper[4846]: I1005 06:50:21.981623 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-5rj2f"] Oct 05 06:50:21 crc kubenswrapper[4846]: I1005 06:50:21.997410 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 05 06:50:21 crc kubenswrapper[4846]: E1005 06:50:21.997624 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-05 06:50:22.497589892 +0000 UTC m=+144.738442667 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 06:50:21 crc kubenswrapper[4846]: I1005 06:50:21.997741 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rf9l4\" (UID: \"78d3e7b1-7258-4508-ad38-373f21bb5d96\") " pod="openshift-image-registry/image-registry-697d97f7c8-rf9l4" Oct 05 06:50:21 crc kubenswrapper[4846]: E1005 06:50:21.998087 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-05 06:50:22.498073735 +0000 UTC m=+144.738926510 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rf9l4" (UID: "78d3e7b1-7258-4508-ad38-373f21bb5d96") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 06:50:22 crc kubenswrapper[4846]: I1005 06:50:22.098474 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 05 06:50:22 crc kubenswrapper[4846]: E1005 06:50:22.098730 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-05 06:50:22.598699255 +0000 UTC m=+144.839552030 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 06:50:22 crc kubenswrapper[4846]: I1005 06:50:22.099006 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rf9l4\" (UID: \"78d3e7b1-7258-4508-ad38-373f21bb5d96\") " pod="openshift-image-registry/image-registry-697d97f7c8-rf9l4" Oct 05 06:50:22 crc kubenswrapper[4846]: E1005 06:50:22.099411 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-05 06:50:22.599397724 +0000 UTC m=+144.840250499 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rf9l4" (UID: "78d3e7b1-7258-4508-ad38-373f21bb5d96") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 06:50:22 crc kubenswrapper[4846]: I1005 06:50:22.099875 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-bpbtf"] Oct 05 06:50:22 crc kubenswrapper[4846]: I1005 06:50:22.141407 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-gqn9s"] Oct 05 06:50:22 crc kubenswrapper[4846]: W1005 06:50:22.167935 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod31d92551_cb79_4d9c_ad1c_c5090142b82c.slice/crio-53c2655a7be0812b3cc58419ce286f6fc420318d04842bc31374851e1a8334d8 WatchSource:0}: Error finding container 53c2655a7be0812b3cc58419ce286f6fc420318d04842bc31374851e1a8334d8: Status 404 returned error can't find the container with id 53c2655a7be0812b3cc58419ce286f6fc420318d04842bc31374851e1a8334d8 Oct 05 06:50:22 crc kubenswrapper[4846]: I1005 06:50:22.191340 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-wcg6b" podStartSLOduration=118.191322006 podStartE2EDuration="1m58.191322006s" podCreationTimestamp="2025-10-05 06:48:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 06:50:22.190675819 +0000 UTC m=+144.431528594" watchObservedRunningTime="2025-10-05 06:50:22.191322006 +0000 UTC m=+144.432174781" Oct 05 06:50:22 crc kubenswrapper[4846]: I1005 06:50:22.203708 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 05 06:50:22 crc kubenswrapper[4846]: E1005 06:50:22.203835 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-05 06:50:22.703807174 +0000 UTC m=+144.944659949 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 06:50:22 crc kubenswrapper[4846]: I1005 06:50:22.203983 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rf9l4\" (UID: \"78d3e7b1-7258-4508-ad38-373f21bb5d96\") " pod="openshift-image-registry/image-registry-697d97f7c8-rf9l4" Oct 05 06:50:22 crc kubenswrapper[4846]: E1005 06:50:22.204371 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-05 06:50:22.704357318 +0000 UTC m=+144.945210093 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rf9l4" (UID: "78d3e7b1-7258-4508-ad38-373f21bb5d96") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 06:50:22 crc kubenswrapper[4846]: I1005 06:50:22.223771 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-vjmbx" podStartSLOduration=118.223739047 podStartE2EDuration="1m58.223739047s" podCreationTimestamp="2025-10-05 06:48:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 06:50:22.215566792 +0000 UTC m=+144.456419567" watchObservedRunningTime="2025-10-05 06:50:22.223739047 +0000 UTC m=+144.464591822" Oct 05 06:50:22 crc kubenswrapper[4846]: I1005 06:50:22.246848 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-pqm4r"] Oct 05 06:50:22 crc kubenswrapper[4846]: I1005 06:50:22.250812 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-jm4gb"] Oct 05 06:50:22 crc kubenswrapper[4846]: I1005 06:50:22.283023 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-4pvwg"] Oct 05 06:50:22 crc kubenswrapper[4846]: I1005 06:50:22.305635 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 05 06:50:22 crc kubenswrapper[4846]: E1005 06:50:22.306046 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-05 06:50:22.806022766 +0000 UTC m=+145.046875541 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 06:50:22 crc kubenswrapper[4846]: I1005 06:50:22.381374 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-6t2mt" podStartSLOduration=119.381355533 podStartE2EDuration="1m59.381355533s" podCreationTimestamp="2025-10-05 06:48:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 06:50:22.378161799 +0000 UTC m=+144.619014584" watchObservedRunningTime="2025-10-05 06:50:22.381355533 +0000 UTC m=+144.622208308" Oct 05 06:50:22 crc kubenswrapper[4846]: W1005 06:50:22.392103 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc041548e_b7c2_4f61_aaea_0b76e298ff22.slice/crio-5de7edbe641eb58d6f0f5638695eee181311d2489717378ab7cce3233f2346b0 WatchSource:0}: Error finding container 5de7edbe641eb58d6f0f5638695eee181311d2489717378ab7cce3233f2346b0: Status 404 returned error can't find the container with id 5de7edbe641eb58d6f0f5638695eee181311d2489717378ab7cce3233f2346b0 Oct 05 06:50:22 crc kubenswrapper[4846]: I1005 06:50:22.406882 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rf9l4\" (UID: \"78d3e7b1-7258-4508-ad38-373f21bb5d96\") " pod="openshift-image-registry/image-registry-697d97f7c8-rf9l4" Oct 05 06:50:22 crc kubenswrapper[4846]: E1005 06:50:22.407291 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-05 06:50:22.907272313 +0000 UTC m=+145.148125088 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rf9l4" (UID: "78d3e7b1-7258-4508-ad38-373f21bb5d96") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 06:50:22 crc kubenswrapper[4846]: I1005 06:50:22.418097 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-m467d" podStartSLOduration=119.418064646 podStartE2EDuration="1m59.418064646s" podCreationTimestamp="2025-10-05 06:48:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 06:50:22.417781439 +0000 UTC m=+144.658634214" watchObservedRunningTime="2025-10-05 06:50:22.418064646 +0000 UTC m=+144.658917421" Oct 05 06:50:22 crc kubenswrapper[4846]: I1005 06:50:22.506406 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-vwktr" podStartSLOduration=118.506386874 podStartE2EDuration="1m58.506386874s" podCreationTimestamp="2025-10-05 06:48:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 06:50:22.502319777 +0000 UTC m=+144.743172562" watchObservedRunningTime="2025-10-05 06:50:22.506386874 +0000 UTC m=+144.747239649" Oct 05 06:50:22 crc kubenswrapper[4846]: I1005 06:50:22.507720 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 05 06:50:22 crc kubenswrapper[4846]: E1005 06:50:22.508048 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-05 06:50:23.008024297 +0000 UTC m=+145.248877072 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 06:50:22 crc kubenswrapper[4846]: I1005 06:50:22.540000 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-9hcbc" podStartSLOduration=119.539978766 podStartE2EDuration="1m59.539978766s" podCreationTimestamp="2025-10-05 06:48:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 06:50:22.53901328 +0000 UTC m=+144.779866055" watchObservedRunningTime="2025-10-05 06:50:22.539978766 +0000 UTC m=+144.780831551" Oct 05 06:50:22 crc kubenswrapper[4846]: I1005 06:50:22.563904 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-c6cxs" podStartSLOduration=118.563880393 podStartE2EDuration="1m58.563880393s" podCreationTimestamp="2025-10-05 06:48:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 06:50:22.563801771 +0000 UTC m=+144.804654546" watchObservedRunningTime="2025-10-05 06:50:22.563880393 +0000 UTC m=+144.804733158" Oct 05 06:50:22 crc kubenswrapper[4846]: I1005 06:50:22.609035 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rf9l4\" (UID: \"78d3e7b1-7258-4508-ad38-373f21bb5d96\") " pod="openshift-image-registry/image-registry-697d97f7c8-rf9l4" Oct 05 06:50:22 crc kubenswrapper[4846]: E1005 06:50:22.610813 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-05 06:50:23.110791184 +0000 UTC m=+145.351643959 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rf9l4" (UID: "78d3e7b1-7258-4508-ad38-373f21bb5d96") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 06:50:22 crc kubenswrapper[4846]: I1005 06:50:22.711709 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 05 06:50:22 crc kubenswrapper[4846]: E1005 06:50:22.711927 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-05 06:50:23.211902927 +0000 UTC m=+145.452755702 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 06:50:22 crc kubenswrapper[4846]: I1005 06:50:22.713698 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rf9l4\" (UID: \"78d3e7b1-7258-4508-ad38-373f21bb5d96\") " pod="openshift-image-registry/image-registry-697d97f7c8-rf9l4" Oct 05 06:50:22 crc kubenswrapper[4846]: E1005 06:50:22.714128 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-05 06:50:23.214111465 +0000 UTC m=+145.454964240 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rf9l4" (UID: "78d3e7b1-7258-4508-ad38-373f21bb5d96") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 06:50:22 crc kubenswrapper[4846]: I1005 06:50:22.736784 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-pglrh" event={"ID":"5bddce7c-b835-4a1f-a3af-db5a1d3771f2","Type":"ContainerStarted","Data":"c65785d38b97895674eefbb4f7bcf074f2579ed1e387e596cc4809a7c4f54cd2"} Oct 05 06:50:22 crc kubenswrapper[4846]: I1005 06:50:22.738573 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-hmzsn" event={"ID":"80ceab29-9389-4191-a909-8a6c40bba1cd","Type":"ContainerStarted","Data":"0a148ea246b2fa1f9539c0cd68609776ce35a86114dd7017822c256f0c29c834"} Oct 05 06:50:22 crc kubenswrapper[4846]: I1005 06:50:22.746634 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-pqm4r" event={"ID":"8049dd16-5d6f-4718-94b1-fb5a322c2530","Type":"ContainerStarted","Data":"67f32177fa0bca27fe7bc3220882ea4cc3b556f11baf00a250efc88c90d9a378"} Oct 05 06:50:22 crc kubenswrapper[4846]: I1005 06:50:22.751729 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-bpbtf" event={"ID":"31d92551-cb79-4d9c-ad1c-c5090142b82c","Type":"ContainerStarted","Data":"53c2655a7be0812b3cc58419ce286f6fc420318d04842bc31374851e1a8334d8"} Oct 05 06:50:22 crc kubenswrapper[4846]: I1005 06:50:22.752846 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-4cxsw" event={"ID":"e532fab3-55a6-4c42-a532-9291452d5d82","Type":"ContainerStarted","Data":"84fdbc77f1e6510edec59c5ec7976171d515bd889ba774de3115ced650a41d85"} Oct 05 06:50:22 crc kubenswrapper[4846]: I1005 06:50:22.756097 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-wnmjw" event={"ID":"788472f5-2ad1-49d0-8e31-6a4fe14c3c85","Type":"ContainerStarted","Data":"bc89fb18ae33a788335f053bafa5cb5a994937c6a1acd72a5a5c1b448c3b9804"} Oct 05 06:50:22 crc kubenswrapper[4846]: I1005 06:50:22.756118 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-wnmjw" event={"ID":"788472f5-2ad1-49d0-8e31-6a4fe14c3c85","Type":"ContainerStarted","Data":"96b75f3072c2c4c745264be699a6eb4ff4ceb59b681a96938a2211a3430f635f"} Oct 05 06:50:22 crc kubenswrapper[4846]: I1005 06:50:22.759402 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-2rc5n" event={"ID":"8c2f8349-aa93-4ac8-a369-5b4b2b9cd6e2","Type":"ContainerStarted","Data":"b9d7ffdddc7410abfb4858bc3fd11d1af0497c11ac8e83b72de8664c8a6dab25"} Oct 05 06:50:22 crc kubenswrapper[4846]: I1005 06:50:22.776020 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-2pd47" event={"ID":"6ec3075e-f780-435f-8efa-30d2fb4855a1","Type":"ContainerStarted","Data":"c7d4ea6c620d8666bab89f607e7d141465bf6142e8f7f39719f9cb5a1752d0b8"} Oct 05 06:50:22 crc kubenswrapper[4846]: I1005 06:50:22.791515 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-t8f4h" event={"ID":"18507a98-2ef8-43b1-b94d-51106f9bdd3b","Type":"ContainerStarted","Data":"1f3fc758862da456200f7939c8ed21c3f7e3b190f8231e9887937ce2fd87a2ec"} Oct 05 06:50:22 crc kubenswrapper[4846]: I1005 06:50:22.792517 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-t8f4h" Oct 05 06:50:22 crc kubenswrapper[4846]: I1005 06:50:22.793870 4846 patch_prober.go:28] interesting pod/console-operator-58897d9998-t8f4h container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.14:8443/readyz\": dial tcp 10.217.0.14:8443: connect: connection refused" start-of-body= Oct 05 06:50:22 crc kubenswrapper[4846]: I1005 06:50:22.793906 4846 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-t8f4h" podUID="18507a98-2ef8-43b1-b94d-51106f9bdd3b" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.14:8443/readyz\": dial tcp 10.217.0.14:8443: connect: connection refused" Oct 05 06:50:22 crc kubenswrapper[4846]: I1005 06:50:22.797953 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-2g2rj" event={"ID":"80c88b0b-c4cf-478e-8e05-4cde023d15ee","Type":"ContainerStarted","Data":"81a413f306cfab5793124d29a9f8bf5ba5b1fd708f60fc4929b70c52f309de19"} Oct 05 06:50:22 crc kubenswrapper[4846]: I1005 06:50:22.797978 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-2g2rj" event={"ID":"80c88b0b-c4cf-478e-8e05-4cde023d15ee","Type":"ContainerStarted","Data":"97323e26e6dfc7129bdcca593a881f79614ec35204b547af0715140fc4ee9c3c"} Oct 05 06:50:22 crc kubenswrapper[4846]: I1005 06:50:22.803587 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-5rj2f" event={"ID":"8b7fb5f2-4b5e-4fc3-9c97-dfbe36646871","Type":"ContainerStarted","Data":"f5caa23450268cdca3d8897adc3142dbe02ae24dcf4983264001ca2b492e72a8"} Oct 05 06:50:22 crc kubenswrapper[4846]: I1005 06:50:22.807224 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-8vjkd" event={"ID":"cb184e9f-f0d9-4166-b0da-213c5c7aba5f","Type":"ContainerStarted","Data":"cfb0e2b187c5c00f4bf9a79ef0d98e00985800605d9d1704e942a25a163ce452"} Oct 05 06:50:22 crc kubenswrapper[4846]: I1005 06:50:22.812548 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-gqn9s" event={"ID":"94a883ac-bee2-403e-8747-a4045e7defea","Type":"ContainerStarted","Data":"b575263ea96b8acb92fffe1514d2350e2826b64d82e57f83b8f913b6fb644aa9"} Oct 05 06:50:22 crc kubenswrapper[4846]: I1005 06:50:22.816306 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 05 06:50:22 crc kubenswrapper[4846]: E1005 06:50:22.816525 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-05 06:50:23.316493212 +0000 UTC m=+145.557345987 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 06:50:22 crc kubenswrapper[4846]: I1005 06:50:22.816731 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rf9l4\" (UID: \"78d3e7b1-7258-4508-ad38-373f21bb5d96\") " pod="openshift-image-registry/image-registry-697d97f7c8-rf9l4" Oct 05 06:50:22 crc kubenswrapper[4846]: E1005 06:50:22.823467 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-05 06:50:23.323442304 +0000 UTC m=+145.564295289 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rf9l4" (UID: "78d3e7b1-7258-4508-ad38-373f21bb5d96") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 06:50:22 crc kubenswrapper[4846]: I1005 06:50:22.855069 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-9jtcc" event={"ID":"7b82343c-28e6-4e0b-9731-5131cb0014f2","Type":"ContainerStarted","Data":"8706719b8160199c08992485dcaa045050da56016f047bba88d0ca0ef0b9150f"} Oct 05 06:50:22 crc kubenswrapper[4846]: I1005 06:50:22.857406 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-4pvwg" event={"ID":"c041548e-b7c2-4f61-aaea-0b76e298ff22","Type":"ContainerStarted","Data":"5de7edbe641eb58d6f0f5638695eee181311d2489717378ab7cce3233f2346b0"} Oct 05 06:50:22 crc kubenswrapper[4846]: I1005 06:50:22.869084 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-7jflr" event={"ID":"7a19cb1c-5274-48e7-ad86-3c2ff85ecf8c","Type":"ContainerStarted","Data":"9933c3d5e382e2048720d7bb0b6b8958aaa6c886c48c28becd807ac718e4503d"} Oct 05 06:50:22 crc kubenswrapper[4846]: I1005 06:50:22.869143 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-7jflr" event={"ID":"7a19cb1c-5274-48e7-ad86-3c2ff85ecf8c","Type":"ContainerStarted","Data":"3d30a24ddf0d7d0208ef9eaf420819aceb66f29e27389260afc0d0ea5f1eb4b7"} Oct 05 06:50:22 crc kubenswrapper[4846]: I1005 06:50:22.875493 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-m7gph" event={"ID":"b21ccaf7-40a5-4939-ad8b-4c7efdcf4306","Type":"ContainerStarted","Data":"7737e135624f89b76624a94df84c1b4b0bc18607c8195150a035222ca1b9bd12"} Oct 05 06:50:22 crc kubenswrapper[4846]: I1005 06:50:22.888822 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-dsqhc" event={"ID":"5fed566e-ecb6-4801-84b3-ad1143a03f08","Type":"ContainerStarted","Data":"c55da91c67258201bfd09a7bb06a346fa567a526ac7a21e9c4b3a31ed38949d8"} Oct 05 06:50:22 crc kubenswrapper[4846]: I1005 06:50:22.892022 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-526w4" event={"ID":"53168eca-95fe-4303-b401-380995e5adcb","Type":"ContainerStarted","Data":"07637f36fa0001957ba6a046df222ba20eeb11494d0fecf0cf2d02ded652bd11"} Oct 05 06:50:22 crc kubenswrapper[4846]: I1005 06:50:22.893709 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-2x9nl" event={"ID":"57e551f8-d9da-4692-b8ae-f7ff6829fe25","Type":"ContainerStarted","Data":"fdeedc91094d838c7b6c4096b74a36075fae15ffc977a5fef6271bb47b20330f"} Oct 05 06:50:22 crc kubenswrapper[4846]: I1005 06:50:22.901596 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-skh8h" event={"ID":"54f93640-64fd-4728-a83b-499aee1ffa38","Type":"ContainerStarted","Data":"d067446958ded5f1176b8cc118c2fd9cefc54f7b01ff888a678986c177d1e82e"} Oct 05 06:50:22 crc kubenswrapper[4846]: I1005 06:50:22.905236 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29327445-dn5mf" event={"ID":"41aa056a-34d5-463b-8352-c71e0f0db1a4","Type":"ContainerStarted","Data":"e46a0ae8f0a310b6735443855f8119ee77787d11058d4f9af366550a7374d17d"} Oct 05 06:50:22 crc kubenswrapper[4846]: I1005 06:50:22.907931 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-vwx77" event={"ID":"f743f03b-3f08-4da0-b53e-1cb33ab516b9","Type":"ContainerStarted","Data":"c54cf2e61e733658a1fe8e5d19f3573c060acee76bfcac6f2db3554155474f9d"} Oct 05 06:50:22 crc kubenswrapper[4846]: I1005 06:50:22.907963 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-vwx77" event={"ID":"f743f03b-3f08-4da0-b53e-1cb33ab516b9","Type":"ContainerStarted","Data":"32d02ac5f7620b4f98c0d8429aec53878adec66dad413aa9d8ca5d0772307f6a"} Oct 05 06:50:22 crc kubenswrapper[4846]: I1005 06:50:22.912431 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-rcnlv" event={"ID":"53fe640c-6cb7-43a3-9a84-467eb36fe66b","Type":"ContainerStarted","Data":"894650799f4762f76202057507e0bdd89056924f9b320107e2e8215bc1fedec0"} Oct 05 06:50:22 crc kubenswrapper[4846]: I1005 06:50:22.912886 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-rcnlv" Oct 05 06:50:22 crc kubenswrapper[4846]: I1005 06:50:22.917373 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-58hwl" event={"ID":"2ecfaba9-9034-4ac7-bd51-e3f64b41d825","Type":"ContainerStarted","Data":"9d3479598930d9f3bb400e32cbae4dd0e5179e194bd97736ed9142617eeea2d0"} Oct 05 06:50:22 crc kubenswrapper[4846]: I1005 06:50:22.917425 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-58hwl" event={"ID":"2ecfaba9-9034-4ac7-bd51-e3f64b41d825","Type":"ContainerStarted","Data":"b1c74b5951fa8d74b2f14c9c7761afff53fe1495df4fd5b5e38e43f27cf9688a"} Oct 05 06:50:22 crc kubenswrapper[4846]: I1005 06:50:22.917502 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 05 06:50:22 crc kubenswrapper[4846]: E1005 06:50:22.917737 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-05 06:50:23.417702168 +0000 UTC m=+145.658554973 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 06:50:22 crc kubenswrapper[4846]: I1005 06:50:22.917918 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rf9l4\" (UID: \"78d3e7b1-7258-4508-ad38-373f21bb5d96\") " pod="openshift-image-registry/image-registry-697d97f7c8-rf9l4" Oct 05 06:50:22 crc kubenswrapper[4846]: E1005 06:50:22.918745 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-05 06:50:23.418725075 +0000 UTC m=+145.659578080 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rf9l4" (UID: "78d3e7b1-7258-4508-ad38-373f21bb5d96") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 06:50:22 crc kubenswrapper[4846]: I1005 06:50:22.920657 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-5qwb9" event={"ID":"bc560c3d-6558-43e6-86ca-312c9fa264ea","Type":"ContainerStarted","Data":"d93c0ed7871bb9c5a35e89cf1add8b7f56da18dbccf5a4d959df44e51f93f4e9"} Oct 05 06:50:22 crc kubenswrapper[4846]: I1005 06:50:22.922770 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-8bpqg" event={"ID":"beeea3e9-d3c3-4280-bf04-85cee4378cb0","Type":"ContainerStarted","Data":"4d2eefcd89069dcddbadef7f87ad2a05ecff38e6d546ee68cd52fc4519a81787"} Oct 05 06:50:22 crc kubenswrapper[4846]: I1005 06:50:22.924283 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-zf4p8" event={"ID":"0cf52841-edf1-4daf-8e2c-76f5f6b93104","Type":"ContainerStarted","Data":"123585e48bbf82d9d071352f1de6090b88f8bb9502c93dd3cbcdc092bdf31caf"} Oct 05 06:50:22 crc kubenswrapper[4846]: I1005 06:50:22.925767 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-x5n4h" event={"ID":"74822779-bdcd-4aae-bc01-9cbcbb0c1d83","Type":"ContainerStarted","Data":"7590f1ade8e9a82d5c3c533d02a5a11405175d1e95f019ca456a6784e4c3853e"} Oct 05 06:50:22 crc kubenswrapper[4846]: I1005 06:50:22.925794 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-x5n4h" event={"ID":"74822779-bdcd-4aae-bc01-9cbcbb0c1d83","Type":"ContainerStarted","Data":"0fe693d1cbc0a346699c582a4acd50d6367c5c0efbcb096cee3f5fea7b3e8873"} Oct 05 06:50:22 crc kubenswrapper[4846]: I1005 06:50:22.927692 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-dtbh5" event={"ID":"0755bb69-c94a-4b36-97ee-8a3dfd972c79","Type":"ContainerStarted","Data":"a0256c6d3ab94315a7e62a41260ac202604c5a887e52c2b9b5ccec95a620b2fd"} Oct 05 06:50:22 crc kubenswrapper[4846]: I1005 06:50:22.932509 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-jm4gb" event={"ID":"c65d29d9-66e3-41ce-acb2-cec28c52fb29","Type":"ContainerStarted","Data":"38a5f3e1db192bda49b5fd51b1b223e1f3d255c57de3a76d6cf4564a1560a0cc"} Oct 05 06:50:22 crc kubenswrapper[4846]: I1005 06:50:22.935006 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-k87kc" event={"ID":"da8eba09-f287-4704-b990-348ff69f0173","Type":"ContainerStarted","Data":"fb09742cefe47dec453dd6a841e4de0ee3218524b7b0e2752de8c99b38704f0b"} Oct 05 06:50:22 crc kubenswrapper[4846]: I1005 06:50:22.936961 4846 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-vjmbx container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.18:8443/healthz\": dial tcp 10.217.0.18:8443: connect: connection refused" start-of-body= Oct 05 06:50:22 crc kubenswrapper[4846]: I1005 06:50:22.937007 4846 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-vjmbx" podUID="6be3f74a-b215-4588-8b0f-735eba580bbd" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.18:8443/healthz\": dial tcp 10.217.0.18:8443: connect: connection refused" Oct 05 06:50:22 crc kubenswrapper[4846]: I1005 06:50:22.937078 4846 patch_prober.go:28] interesting pod/downloads-7954f5f757-c6cxs container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.12:8080/\": dial tcp 10.217.0.12:8080: connect: connection refused" start-of-body= Oct 05 06:50:22 crc kubenswrapper[4846]: I1005 06:50:22.937126 4846 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-c6cxs" podUID="475e1ca4-a6fa-4f77-8c89-0619b2e9681e" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.12:8080/\": dial tcp 10.217.0.12:8080: connect: connection refused" Oct 05 06:50:22 crc kubenswrapper[4846]: I1005 06:50:22.941170 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-vwktr" Oct 05 06:50:23 crc kubenswrapper[4846]: I1005 06:50:23.034676 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 05 06:50:23 crc kubenswrapper[4846]: E1005 06:50:23.034896 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-05 06:50:23.534862393 +0000 UTC m=+145.775715178 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 06:50:23 crc kubenswrapper[4846]: I1005 06:50:23.035218 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rf9l4\" (UID: \"78d3e7b1-7258-4508-ad38-373f21bb5d96\") " pod="openshift-image-registry/image-registry-697d97f7c8-rf9l4" Oct 05 06:50:23 crc kubenswrapper[4846]: E1005 06:50:23.046398 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-05 06:50:23.546377855 +0000 UTC m=+145.787230860 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rf9l4" (UID: "78d3e7b1-7258-4508-ad38-373f21bb5d96") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 06:50:23 crc kubenswrapper[4846]: I1005 06:50:23.137714 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 05 06:50:23 crc kubenswrapper[4846]: E1005 06:50:23.138102 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-05 06:50:23.638085851 +0000 UTC m=+145.878938626 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 06:50:23 crc kubenswrapper[4846]: I1005 06:50:23.192125 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-6t2mt" Oct 05 06:50:23 crc kubenswrapper[4846]: I1005 06:50:23.234776 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-9jtcc" podStartSLOduration=119.234753828 podStartE2EDuration="1m59.234753828s" podCreationTimestamp="2025-10-05 06:48:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 06:50:23.229340146 +0000 UTC m=+145.470192921" watchObservedRunningTime="2025-10-05 06:50:23.234753828 +0000 UTC m=+145.475606603" Oct 05 06:50:23 crc kubenswrapper[4846]: I1005 06:50:23.238917 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rf9l4\" (UID: \"78d3e7b1-7258-4508-ad38-373f21bb5d96\") " pod="openshift-image-registry/image-registry-697d97f7c8-rf9l4" Oct 05 06:50:23 crc kubenswrapper[4846]: E1005 06:50:23.239450 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-05 06:50:23.739435161 +0000 UTC m=+145.980287936 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rf9l4" (UID: "78d3e7b1-7258-4508-ad38-373f21bb5d96") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 06:50:23 crc kubenswrapper[4846]: I1005 06:50:23.272432 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-rcnlv" podStartSLOduration=120.272394846 podStartE2EDuration="2m0.272394846s" podCreationTimestamp="2025-10-05 06:48:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 06:50:23.265932266 +0000 UTC m=+145.506785041" watchObservedRunningTime="2025-10-05 06:50:23.272394846 +0000 UTC m=+145.513247641" Oct 05 06:50:23 crc kubenswrapper[4846]: I1005 06:50:23.321386 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-x5n4h" podStartSLOduration=119.321365601 podStartE2EDuration="1m59.321365601s" podCreationTimestamp="2025-10-05 06:48:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 06:50:23.289742461 +0000 UTC m=+145.530595236" watchObservedRunningTime="2025-10-05 06:50:23.321365601 +0000 UTC m=+145.562218376" Oct 05 06:50:23 crc kubenswrapper[4846]: I1005 06:50:23.321518 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-wnmjw" podStartSLOduration=119.321513635 podStartE2EDuration="1m59.321513635s" podCreationTimestamp="2025-10-05 06:48:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 06:50:23.319964624 +0000 UTC m=+145.560817399" watchObservedRunningTime="2025-10-05 06:50:23.321513635 +0000 UTC m=+145.562366410" Oct 05 06:50:23 crc kubenswrapper[4846]: I1005 06:50:23.340645 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 05 06:50:23 crc kubenswrapper[4846]: E1005 06:50:23.341105 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-05 06:50:23.841069078 +0000 UTC m=+146.081921853 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 06:50:23 crc kubenswrapper[4846]: I1005 06:50:23.349554 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-vwx77" podStartSLOduration=119.34953341 podStartE2EDuration="1m59.34953341s" podCreationTimestamp="2025-10-05 06:48:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 06:50:23.343909623 +0000 UTC m=+145.584762398" watchObservedRunningTime="2025-10-05 06:50:23.34953341 +0000 UTC m=+145.590386175" Oct 05 06:50:23 crc kubenswrapper[4846]: I1005 06:50:23.426850 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-58hwl" podStartSLOduration=119.426825869 podStartE2EDuration="1m59.426825869s" podCreationTimestamp="2025-10-05 06:48:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 06:50:23.423572133 +0000 UTC m=+145.664424908" watchObservedRunningTime="2025-10-05 06:50:23.426825869 +0000 UTC m=+145.667678654" Oct 05 06:50:23 crc kubenswrapper[4846]: I1005 06:50:23.442438 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rf9l4\" (UID: \"78d3e7b1-7258-4508-ad38-373f21bb5d96\") " pod="openshift-image-registry/image-registry-697d97f7c8-rf9l4" Oct 05 06:50:23 crc kubenswrapper[4846]: E1005 06:50:23.442828 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-05 06:50:23.942813298 +0000 UTC m=+146.183666073 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rf9l4" (UID: "78d3e7b1-7258-4508-ad38-373f21bb5d96") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 06:50:23 crc kubenswrapper[4846]: I1005 06:50:23.481858 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-4cxsw" podStartSLOduration=6.481836402 podStartE2EDuration="6.481836402s" podCreationTimestamp="2025-10-05 06:50:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 06:50:23.47603116 +0000 UTC m=+145.716883935" watchObservedRunningTime="2025-10-05 06:50:23.481836402 +0000 UTC m=+145.722689177" Oct 05 06:50:23 crc kubenswrapper[4846]: I1005 06:50:23.512509 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-8vjkd" podStartSLOduration=119.512491847 podStartE2EDuration="1m59.512491847s" podCreationTimestamp="2025-10-05 06:48:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 06:50:23.510247198 +0000 UTC m=+145.751099983" watchObservedRunningTime="2025-10-05 06:50:23.512491847 +0000 UTC m=+145.753344622" Oct 05 06:50:23 crc kubenswrapper[4846]: I1005 06:50:23.543050 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 05 06:50:23 crc kubenswrapper[4846]: E1005 06:50:23.543346 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-05 06:50:24.043307955 +0000 UTC m=+146.284160730 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 06:50:23 crc kubenswrapper[4846]: I1005 06:50:23.544806 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rf9l4\" (UID: \"78d3e7b1-7258-4508-ad38-373f21bb5d96\") " pod="openshift-image-registry/image-registry-697d97f7c8-rf9l4" Oct 05 06:50:23 crc kubenswrapper[4846]: E1005 06:50:23.547518 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-05 06:50:24.047499815 +0000 UTC m=+146.288352590 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rf9l4" (UID: "78d3e7b1-7258-4508-ad38-373f21bb5d96") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 06:50:23 crc kubenswrapper[4846]: I1005 06:50:23.602246 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-526w4" podStartSLOduration=119.602215301 podStartE2EDuration="1m59.602215301s" podCreationTimestamp="2025-10-05 06:48:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 06:50:23.575719956 +0000 UTC m=+145.816572731" watchObservedRunningTime="2025-10-05 06:50:23.602215301 +0000 UTC m=+145.843068076" Oct 05 06:50:23 crc kubenswrapper[4846]: I1005 06:50:23.623576 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29327445-dn5mf" podStartSLOduration=120.623547371 podStartE2EDuration="2m0.623547371s" podCreationTimestamp="2025-10-05 06:48:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 06:50:23.620752388 +0000 UTC m=+145.861605163" watchObservedRunningTime="2025-10-05 06:50:23.623547371 +0000 UTC m=+145.864400146" Oct 05 06:50:23 crc kubenswrapper[4846]: I1005 06:50:23.653788 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 05 06:50:23 crc kubenswrapper[4846]: E1005 06:50:23.654467 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-05 06:50:24.154449912 +0000 UTC m=+146.395302677 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 06:50:23 crc kubenswrapper[4846]: I1005 06:50:23.668580 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-t8f4h" podStartSLOduration=120.668549642 podStartE2EDuration="2m0.668549642s" podCreationTimestamp="2025-10-05 06:48:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 06:50:23.665155093 +0000 UTC m=+145.906007868" watchObservedRunningTime="2025-10-05 06:50:23.668549642 +0000 UTC m=+145.909402417" Oct 05 06:50:23 crc kubenswrapper[4846]: I1005 06:50:23.707666 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-58hwl" Oct 05 06:50:23 crc kubenswrapper[4846]: I1005 06:50:23.711468 4846 patch_prober.go:28] interesting pod/router-default-5444994796-58hwl container/router namespace/openshift-ingress: Startup probe status=failure output="Get \"http://localhost:1936/healthz/ready\": dial tcp [::1]:1936: connect: connection refused" start-of-body= Oct 05 06:50:23 crc kubenswrapper[4846]: I1005 06:50:23.711511 4846 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-58hwl" podUID="2ecfaba9-9034-4ac7-bd51-e3f64b41d825" containerName="router" probeResult="failure" output="Get \"http://localhost:1936/healthz/ready\": dial tcp [::1]:1936: connect: connection refused" Oct 05 06:50:23 crc kubenswrapper[4846]: I1005 06:50:23.743920 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-pglrh" podStartSLOduration=120.743898659 podStartE2EDuration="2m0.743898659s" podCreationTimestamp="2025-10-05 06:48:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 06:50:23.742423601 +0000 UTC m=+145.983276376" watchObservedRunningTime="2025-10-05 06:50:23.743898659 +0000 UTC m=+145.984751434" Oct 05 06:50:23 crc kubenswrapper[4846]: I1005 06:50:23.746023 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-2pd47" podStartSLOduration=120.746009835 podStartE2EDuration="2m0.746009835s" podCreationTimestamp="2025-10-05 06:48:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 06:50:23.704543687 +0000 UTC m=+145.945396482" watchObservedRunningTime="2025-10-05 06:50:23.746009835 +0000 UTC m=+145.986862610" Oct 05 06:50:23 crc kubenswrapper[4846]: I1005 06:50:23.755996 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rf9l4\" (UID: \"78d3e7b1-7258-4508-ad38-373f21bb5d96\") " pod="openshift-image-registry/image-registry-697d97f7c8-rf9l4" Oct 05 06:50:23 crc kubenswrapper[4846]: E1005 06:50:23.756882 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-05 06:50:24.25686859 +0000 UTC m=+146.497721355 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rf9l4" (UID: "78d3e7b1-7258-4508-ad38-373f21bb5d96") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 06:50:23 crc kubenswrapper[4846]: I1005 06:50:23.794725 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-7jflr" podStartSLOduration=119.794703723 podStartE2EDuration="1m59.794703723s" podCreationTimestamp="2025-10-05 06:48:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 06:50:23.79384077 +0000 UTC m=+146.034693545" watchObservedRunningTime="2025-10-05 06:50:23.794703723 +0000 UTC m=+146.035556498" Oct 05 06:50:23 crc kubenswrapper[4846]: I1005 06:50:23.819395 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-2g2rj" podStartSLOduration=6.81937499 podStartE2EDuration="6.81937499s" podCreationTimestamp="2025-10-05 06:50:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 06:50:23.817744197 +0000 UTC m=+146.058596982" watchObservedRunningTime="2025-10-05 06:50:23.81937499 +0000 UTC m=+146.060227775" Oct 05 06:50:23 crc kubenswrapper[4846]: I1005 06:50:23.858830 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 05 06:50:23 crc kubenswrapper[4846]: E1005 06:50:23.858979 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-05 06:50:24.358954839 +0000 UTC m=+146.599807614 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 06:50:23 crc kubenswrapper[4846]: I1005 06:50:23.859096 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rf9l4\" (UID: \"78d3e7b1-7258-4508-ad38-373f21bb5d96\") " pod="openshift-image-registry/image-registry-697d97f7c8-rf9l4" Oct 05 06:50:23 crc kubenswrapper[4846]: E1005 06:50:23.859397 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-05 06:50:24.35938817 +0000 UTC m=+146.600240945 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rf9l4" (UID: "78d3e7b1-7258-4508-ad38-373f21bb5d96") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 06:50:23 crc kubenswrapper[4846]: I1005 06:50:23.955466 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-8bpqg" event={"ID":"beeea3e9-d3c3-4280-bf04-85cee4378cb0","Type":"ContainerStarted","Data":"b621feb16f93ff27ee99a89c8b38c0c3c6c24a693305db052fdd5f3a72358567"} Oct 05 06:50:23 crc kubenswrapper[4846]: I1005 06:50:23.962590 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 05 06:50:23 crc kubenswrapper[4846]: E1005 06:50:23.962919 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-05 06:50:24.462897396 +0000 UTC m=+146.703750191 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 06:50:23 crc kubenswrapper[4846]: I1005 06:50:23.969542 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-stjw8" event={"ID":"e4ea5d4f-3481-4e72-9bd8-e5d8183bc687","Type":"ContainerStarted","Data":"122aae43cdd6b1c4af0f6cf74f9ca5ce10ae0fac3df4977f5c37987c348dd471"} Oct 05 06:50:23 crc kubenswrapper[4846]: I1005 06:50:23.979548 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-8bpqg" podStartSLOduration=119.979530783 podStartE2EDuration="1m59.979530783s" podCreationTimestamp="2025-10-05 06:48:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 06:50:23.974645895 +0000 UTC m=+146.215498670" watchObservedRunningTime="2025-10-05 06:50:23.979530783 +0000 UTC m=+146.220383548" Oct 05 06:50:23 crc kubenswrapper[4846]: I1005 06:50:23.986317 4846 generic.go:334] "Generic (PLEG): container finished" podID="75d0c362-7aa3-4919-937d-8849f425466e" containerID="f641c0fc563fba7ae07b5ba4e2f2ab0f111f434e2ad94277bb68e26ba0e7e33c" exitCode=0 Oct 05 06:50:23 crc kubenswrapper[4846]: I1005 06:50:23.986460 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-6bpvf" event={"ID":"75d0c362-7aa3-4919-937d-8849f425466e","Type":"ContainerDied","Data":"f641c0fc563fba7ae07b5ba4e2f2ab0f111f434e2ad94277bb68e26ba0e7e33c"} Oct 05 06:50:24 crc kubenswrapper[4846]: I1005 06:50:24.012775 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-stjw8" podStartSLOduration=120.012757665 podStartE2EDuration="2m0.012757665s" podCreationTimestamp="2025-10-05 06:48:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 06:50:24.01220094 +0000 UTC m=+146.253053715" watchObservedRunningTime="2025-10-05 06:50:24.012757665 +0000 UTC m=+146.253610440" Oct 05 06:50:24 crc kubenswrapper[4846]: I1005 06:50:24.017155 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-5rj2f" event={"ID":"8b7fb5f2-4b5e-4fc3-9c97-dfbe36646871","Type":"ContainerStarted","Data":"7333ec19f4516e1ee1891d48c20b3c4b81fca68db8f72564733a30655e04b60a"} Oct 05 06:50:24 crc kubenswrapper[4846]: I1005 06:50:24.022905 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-4pvwg" event={"ID":"c041548e-b7c2-4f61-aaea-0b76e298ff22","Type":"ContainerStarted","Data":"99a37c2ed9f01df55d44a743ce056a0d19887923c1827ba03dc047bac4a7fd58"} Oct 05 06:50:24 crc kubenswrapper[4846]: I1005 06:50:24.023980 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-4pvwg" Oct 05 06:50:24 crc kubenswrapper[4846]: I1005 06:50:24.036264 4846 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-4pvwg container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.22:5443/healthz\": dial tcp 10.217.0.22:5443: connect: connection refused" start-of-body= Oct 05 06:50:24 crc kubenswrapper[4846]: I1005 06:50:24.036332 4846 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-4pvwg" podUID="c041548e-b7c2-4f61-aaea-0b76e298ff22" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.22:5443/healthz\": dial tcp 10.217.0.22:5443: connect: connection refused" Oct 05 06:50:24 crc kubenswrapper[4846]: I1005 06:50:24.036842 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-5qwb9" event={"ID":"bc560c3d-6558-43e6-86ca-312c9fa264ea","Type":"ContainerStarted","Data":"95dc8e11f82a7b5124c5669308a93033edc497eea20fb63eebcac7afd735fc3d"} Oct 05 06:50:24 crc kubenswrapper[4846]: I1005 06:50:24.059603 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-pqm4r" event={"ID":"8049dd16-5d6f-4718-94b1-fb5a322c2530","Type":"ContainerStarted","Data":"73752502bde466a66997b6e9c282b5f182efe5107bbfbe87e3495bd309ef806f"} Oct 05 06:50:24 crc kubenswrapper[4846]: I1005 06:50:24.059675 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-pqm4r" event={"ID":"8049dd16-5d6f-4718-94b1-fb5a322c2530","Type":"ContainerStarted","Data":"48b789b8ce3d1676a18f83a8596b8b90277b75cac0656a15385df0c64391bcdb"} Oct 05 06:50:24 crc kubenswrapper[4846]: I1005 06:50:24.060476 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-pqm4r" Oct 05 06:50:24 crc kubenswrapper[4846]: I1005 06:50:24.071924 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rf9l4\" (UID: \"78d3e7b1-7258-4508-ad38-373f21bb5d96\") " pod="openshift-image-registry/image-registry-697d97f7c8-rf9l4" Oct 05 06:50:24 crc kubenswrapper[4846]: E1005 06:50:24.072358 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-05 06:50:24.572323468 +0000 UTC m=+146.813176243 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rf9l4" (UID: "78d3e7b1-7258-4508-ad38-373f21bb5d96") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 06:50:24 crc kubenswrapper[4846]: I1005 06:50:24.073099 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-m7gph" event={"ID":"b21ccaf7-40a5-4939-ad8b-4c7efdcf4306","Type":"ContainerStarted","Data":"0f182a323e687374b263c2625601f703b3317eea13804a1878b5c76bc0f07099"} Oct 05 06:50:24 crc kubenswrapper[4846]: I1005 06:50:24.073161 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-m7gph" event={"ID":"b21ccaf7-40a5-4939-ad8b-4c7efdcf4306","Type":"ContainerStarted","Data":"fee92f945739b84472ac650ae2306e02146dbe20dbc5e73c2bc4ec0f7dc3a9a3"} Oct 05 06:50:24 crc kubenswrapper[4846]: I1005 06:50:24.084257 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-zf4p8" event={"ID":"0cf52841-edf1-4daf-8e2c-76f5f6b93104","Type":"ContainerStarted","Data":"60d05839976fee2996526e4f569967c703a59e4479584c8189cd57d5b36438ff"} Oct 05 06:50:24 crc kubenswrapper[4846]: I1005 06:50:24.085557 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-zf4p8" Oct 05 06:50:24 crc kubenswrapper[4846]: I1005 06:50:24.090492 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-5rj2f" podStartSLOduration=120.090467904 podStartE2EDuration="2m0.090467904s" podCreationTimestamp="2025-10-05 06:48:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 06:50:24.08383226 +0000 UTC m=+146.324685035" watchObservedRunningTime="2025-10-05 06:50:24.090467904 +0000 UTC m=+146.331320679" Oct 05 06:50:24 crc kubenswrapper[4846]: I1005 06:50:24.117765 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-dtbh5" event={"ID":"0755bb69-c94a-4b36-97ee-8a3dfd972c79","Type":"ContainerStarted","Data":"747f1372c4bcffb17b039b87f2b599cac8126e1be275057e1c89b1f121322b1b"} Oct 05 06:50:24 crc kubenswrapper[4846]: I1005 06:50:24.120733 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-gqn9s" event={"ID":"94a883ac-bee2-403e-8747-a4045e7defea","Type":"ContainerStarted","Data":"6c7d33b45d1323e83aba7cb0c3743b56c545653066a0300c549fa68b131a13a3"} Oct 05 06:50:24 crc kubenswrapper[4846]: I1005 06:50:24.121737 4846 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-zf4p8 container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.35:8443/healthz\": dial tcp 10.217.0.35:8443: connect: connection refused" start-of-body= Oct 05 06:50:24 crc kubenswrapper[4846]: I1005 06:50:24.121793 4846 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-zf4p8" podUID="0cf52841-edf1-4daf-8e2c-76f5f6b93104" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.35:8443/healthz\": dial tcp 10.217.0.35:8443: connect: connection refused" Oct 05 06:50:24 crc kubenswrapper[4846]: I1005 06:50:24.130354 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-skh8h" event={"ID":"54f93640-64fd-4728-a83b-499aee1ffa38","Type":"ContainerStarted","Data":"a0cd9a3bd2decf4afe8e48b98a5596dab311c66674e4c10983a44621fba0e8f0"} Oct 05 06:50:24 crc kubenswrapper[4846]: I1005 06:50:24.137726 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-4pvwg" podStartSLOduration=120.137705834 podStartE2EDuration="2m0.137705834s" podCreationTimestamp="2025-10-05 06:48:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 06:50:24.136434611 +0000 UTC m=+146.377287386" watchObservedRunningTime="2025-10-05 06:50:24.137705834 +0000 UTC m=+146.378558609" Oct 05 06:50:24 crc kubenswrapper[4846]: I1005 06:50:24.144577 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-2x9nl" event={"ID":"57e551f8-d9da-4692-b8ae-f7ff6829fe25","Type":"ContainerStarted","Data":"a6262c598436ffbc0fcfb0eba9cce00885465a287763e953f04af795a137bd0c"} Oct 05 06:50:24 crc kubenswrapper[4846]: I1005 06:50:24.158550 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-5qwb9" podStartSLOduration=120.1585327 podStartE2EDuration="2m0.1585327s" podCreationTimestamp="2025-10-05 06:48:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 06:50:24.15853016 +0000 UTC m=+146.399382935" watchObservedRunningTime="2025-10-05 06:50:24.1585327 +0000 UTC m=+146.399385475" Oct 05 06:50:24 crc kubenswrapper[4846]: I1005 06:50:24.168125 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-bpbtf" event={"ID":"31d92551-cb79-4d9c-ad1c-c5090142b82c","Type":"ContainerStarted","Data":"79276118a5ea830f658df5d41f1e1588c02adae45edd58fad2a3ef34f2301761"} Oct 05 06:50:24 crc kubenswrapper[4846]: I1005 06:50:24.176258 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 05 06:50:24 crc kubenswrapper[4846]: E1005 06:50:24.176658 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-05 06:50:24.676632015 +0000 UTC m=+146.917484790 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 06:50:24 crc kubenswrapper[4846]: I1005 06:50:24.176985 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rf9l4\" (UID: \"78d3e7b1-7258-4508-ad38-373f21bb5d96\") " pod="openshift-image-registry/image-registry-697d97f7c8-rf9l4" Oct 05 06:50:24 crc kubenswrapper[4846]: E1005 06:50:24.179553 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-05 06:50:24.679544472 +0000 UTC m=+146.920397247 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rf9l4" (UID: "78d3e7b1-7258-4508-ad38-373f21bb5d96") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 06:50:24 crc kubenswrapper[4846]: I1005 06:50:24.218901 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-jm4gb" event={"ID":"c65d29d9-66e3-41ce-acb2-cec28c52fb29","Type":"ContainerStarted","Data":"d939a244297727e8b36b94558a6c7dd0027325c9399909c90653a495198894ed"} Oct 05 06:50:24 crc kubenswrapper[4846]: I1005 06:50:24.220900 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-k87kc" event={"ID":"da8eba09-f287-4704-b990-348ff69f0173","Type":"ContainerStarted","Data":"aa3d70b869841523b622a712b7f404baadad8a1e0cd80d57da7e314aa4992f08"} Oct 05 06:50:24 crc kubenswrapper[4846]: I1005 06:50:24.222862 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-hmzsn" event={"ID":"80ceab29-9389-4191-a909-8a6c40bba1cd","Type":"ContainerStarted","Data":"537e0c7e7a6e16fca899ac951f81eaf922cd55bca1936a65398babcef1c8718c"} Oct 05 06:50:24 crc kubenswrapper[4846]: I1005 06:50:24.223493 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-hmzsn" Oct 05 06:50:24 crc kubenswrapper[4846]: I1005 06:50:24.229600 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-2x9nl" podStartSLOduration=120.229581575 podStartE2EDuration="2m0.229581575s" podCreationTimestamp="2025-10-05 06:48:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 06:50:24.195694166 +0000 UTC m=+146.436546941" watchObservedRunningTime="2025-10-05 06:50:24.229581575 +0000 UTC m=+146.470434350" Oct 05 06:50:24 crc kubenswrapper[4846]: I1005 06:50:24.230591 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-gqn9s" podStartSLOduration=120.230585731 podStartE2EDuration="2m0.230585731s" podCreationTimestamp="2025-10-05 06:48:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 06:50:24.228281641 +0000 UTC m=+146.469134416" watchObservedRunningTime="2025-10-05 06:50:24.230585731 +0000 UTC m=+146.471438496" Oct 05 06:50:24 crc kubenswrapper[4846]: I1005 06:50:24.239577 4846 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-hmzsn container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.23:8443/healthz\": dial tcp 10.217.0.23:8443: connect: connection refused" start-of-body= Oct 05 06:50:24 crc kubenswrapper[4846]: I1005 06:50:24.239630 4846 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-hmzsn" podUID="80ceab29-9389-4191-a909-8a6c40bba1cd" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.23:8443/healthz\": dial tcp 10.217.0.23:8443: connect: connection refused" Oct 05 06:50:24 crc kubenswrapper[4846]: I1005 06:50:24.242111 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-2rc5n" event={"ID":"8c2f8349-aa93-4ac8-a369-5b4b2b9cd6e2","Type":"ContainerStarted","Data":"f702ca5fddb2d55c48f6f9d3494d5c98b925185899945a42f5693b66af2b7020"} Oct 05 06:50:24 crc kubenswrapper[4846]: I1005 06:50:24.242141 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-2rc5n" Oct 05 06:50:24 crc kubenswrapper[4846]: I1005 06:50:24.244551 4846 patch_prober.go:28] interesting pod/console-operator-58897d9998-t8f4h container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.14:8443/readyz\": dial tcp 10.217.0.14:8443: connect: connection refused" start-of-body= Oct 05 06:50:24 crc kubenswrapper[4846]: I1005 06:50:24.244600 4846 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-t8f4h" podUID="18507a98-2ef8-43b1-b94d-51106f9bdd3b" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.14:8443/readyz\": dial tcp 10.217.0.14:8443: connect: connection refused" Oct 05 06:50:24 crc kubenswrapper[4846]: I1005 06:50:24.244811 4846 patch_prober.go:28] interesting pod/downloads-7954f5f757-c6cxs container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.12:8080/\": dial tcp 10.217.0.12:8080: connect: connection refused" start-of-body= Oct 05 06:50:24 crc kubenswrapper[4846]: I1005 06:50:24.244839 4846 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-c6cxs" podUID="475e1ca4-a6fa-4f77-8c89-0619b2e9681e" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.12:8080/\": dial tcp 10.217.0.12:8080: connect: connection refused" Oct 05 06:50:24 crc kubenswrapper[4846]: I1005 06:50:24.273613 4846 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-2rc5n container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.38:8080/healthz\": dial tcp 10.217.0.38:8080: connect: connection refused" start-of-body= Oct 05 06:50:24 crc kubenswrapper[4846]: I1005 06:50:24.273656 4846 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-2rc5n" podUID="8c2f8349-aa93-4ac8-a369-5b4b2b9cd6e2" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.38:8080/healthz\": dial tcp 10.217.0.38:8080: connect: connection refused" Oct 05 06:50:24 crc kubenswrapper[4846]: I1005 06:50:24.280913 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 05 06:50:24 crc kubenswrapper[4846]: E1005 06:50:24.288188 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-05 06:50:24.787913836 +0000 UTC m=+147.028766611 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 06:50:24 crc kubenswrapper[4846]: I1005 06:50:24.291856 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-zf4p8" podStartSLOduration=120.291828898 podStartE2EDuration="2m0.291828898s" podCreationTimestamp="2025-10-05 06:48:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 06:50:24.271625918 +0000 UTC m=+146.512478693" watchObservedRunningTime="2025-10-05 06:50:24.291828898 +0000 UTC m=+146.532681673" Oct 05 06:50:24 crc kubenswrapper[4846]: I1005 06:50:24.299137 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-vjmbx" Oct 05 06:50:24 crc kubenswrapper[4846]: I1005 06:50:24.359787 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-pqm4r" podStartSLOduration=120.359753491 podStartE2EDuration="2m0.359753491s" podCreationTimestamp="2025-10-05 06:48:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 06:50:24.353498957 +0000 UTC m=+146.594351722" watchObservedRunningTime="2025-10-05 06:50:24.359753491 +0000 UTC m=+146.600606416" Oct 05 06:50:24 crc kubenswrapper[4846]: I1005 06:50:24.379915 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-skh8h" podStartSLOduration=120.37989606 podStartE2EDuration="2m0.37989606s" podCreationTimestamp="2025-10-05 06:48:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 06:50:24.378465622 +0000 UTC m=+146.619318397" watchObservedRunningTime="2025-10-05 06:50:24.37989606 +0000 UTC m=+146.620748835" Oct 05 06:50:24 crc kubenswrapper[4846]: I1005 06:50:24.384552 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rf9l4\" (UID: \"78d3e7b1-7258-4508-ad38-373f21bb5d96\") " pod="openshift-image-registry/image-registry-697d97f7c8-rf9l4" Oct 05 06:50:24 crc kubenswrapper[4846]: E1005 06:50:24.385299 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-05 06:50:24.88523667 +0000 UTC m=+147.126089555 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rf9l4" (UID: "78d3e7b1-7258-4508-ad38-373f21bb5d96") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 06:50:24 crc kubenswrapper[4846]: I1005 06:50:24.402541 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-m7gph" podStartSLOduration=120.402517523 podStartE2EDuration="2m0.402517523s" podCreationTimestamp="2025-10-05 06:48:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 06:50:24.399018181 +0000 UTC m=+146.639870966" watchObservedRunningTime="2025-10-05 06:50:24.402517523 +0000 UTC m=+146.643370308" Oct 05 06:50:24 crc kubenswrapper[4846]: I1005 06:50:24.425596 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-hmzsn" podStartSLOduration=120.425575548 podStartE2EDuration="2m0.425575548s" podCreationTimestamp="2025-10-05 06:48:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 06:50:24.425500266 +0000 UTC m=+146.666353041" watchObservedRunningTime="2025-10-05 06:50:24.425575548 +0000 UTC m=+146.666428323" Oct 05 06:50:24 crc kubenswrapper[4846]: I1005 06:50:24.451838 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-k87kc" podStartSLOduration=121.451821287 podStartE2EDuration="2m1.451821287s" podCreationTimestamp="2025-10-05 06:48:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 06:50:24.450745939 +0000 UTC m=+146.691598724" watchObservedRunningTime="2025-10-05 06:50:24.451821287 +0000 UTC m=+146.692674052" Oct 05 06:50:24 crc kubenswrapper[4846]: I1005 06:50:24.487720 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 05 06:50:24 crc kubenswrapper[4846]: E1005 06:50:24.488198 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-05 06:50:24.988160691 +0000 UTC m=+147.229013466 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 06:50:24 crc kubenswrapper[4846]: I1005 06:50:24.498088 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-2rc5n" podStartSLOduration=120.498067351 podStartE2EDuration="2m0.498067351s" podCreationTimestamp="2025-10-05 06:48:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 06:50:24.496554201 +0000 UTC m=+146.737406976" watchObservedRunningTime="2025-10-05 06:50:24.498067351 +0000 UTC m=+146.738920126" Oct 05 06:50:24 crc kubenswrapper[4846]: I1005 06:50:24.522337 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-bpbtf" podStartSLOduration=120.522312977 podStartE2EDuration="2m0.522312977s" podCreationTimestamp="2025-10-05 06:48:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 06:50:24.514515952 +0000 UTC m=+146.755368727" watchObservedRunningTime="2025-10-05 06:50:24.522312977 +0000 UTC m=+146.763165752" Oct 05 06:50:24 crc kubenswrapper[4846]: I1005 06:50:24.590348 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rf9l4\" (UID: \"78d3e7b1-7258-4508-ad38-373f21bb5d96\") " pod="openshift-image-registry/image-registry-697d97f7c8-rf9l4" Oct 05 06:50:24 crc kubenswrapper[4846]: E1005 06:50:24.590903 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-05 06:50:25.090885536 +0000 UTC m=+147.331738331 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rf9l4" (UID: "78d3e7b1-7258-4508-ad38-373f21bb5d96") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 06:50:24 crc kubenswrapper[4846]: I1005 06:50:24.691365 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 05 06:50:24 crc kubenswrapper[4846]: E1005 06:50:24.691565 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-05 06:50:25.191527707 +0000 UTC m=+147.432380482 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 06:50:24 crc kubenswrapper[4846]: I1005 06:50:24.691685 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rf9l4\" (UID: \"78d3e7b1-7258-4508-ad38-373f21bb5d96\") " pod="openshift-image-registry/image-registry-697d97f7c8-rf9l4" Oct 05 06:50:24 crc kubenswrapper[4846]: E1005 06:50:24.692033 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-05 06:50:25.19201868 +0000 UTC m=+147.432871455 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rf9l4" (UID: "78d3e7b1-7258-4508-ad38-373f21bb5d96") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 06:50:24 crc kubenswrapper[4846]: I1005 06:50:24.720661 4846 patch_prober.go:28] interesting pod/router-default-5444994796-58hwl container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 05 06:50:24 crc kubenswrapper[4846]: [-]has-synced failed: reason withheld Oct 05 06:50:24 crc kubenswrapper[4846]: [+]process-running ok Oct 05 06:50:24 crc kubenswrapper[4846]: healthz check failed Oct 05 06:50:24 crc kubenswrapper[4846]: I1005 06:50:24.720735 4846 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-58hwl" podUID="2ecfaba9-9034-4ac7-bd51-e3f64b41d825" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 05 06:50:24 crc kubenswrapper[4846]: I1005 06:50:24.767938 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-stjw8" Oct 05 06:50:24 crc kubenswrapper[4846]: I1005 06:50:24.768138 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-stjw8" Oct 05 06:50:24 crc kubenswrapper[4846]: I1005 06:50:24.769641 4846 patch_prober.go:28] interesting pod/apiserver-7bbb656c7d-stjw8 container/oauth-apiserver namespace/openshift-oauth-apiserver: Startup probe status=failure output="Get \"https://10.217.0.15:8443/livez\": dial tcp 10.217.0.15:8443: connect: connection refused" start-of-body= Oct 05 06:50:24 crc kubenswrapper[4846]: I1005 06:50:24.769682 4846 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-stjw8" podUID="e4ea5d4f-3481-4e72-9bd8-e5d8183bc687" containerName="oauth-apiserver" probeResult="failure" output="Get \"https://10.217.0.15:8443/livez\": dial tcp 10.217.0.15:8443: connect: connection refused" Oct 05 06:50:24 crc kubenswrapper[4846]: I1005 06:50:24.792926 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 05 06:50:24 crc kubenswrapper[4846]: E1005 06:50:24.793128 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-05 06:50:25.293094013 +0000 UTC m=+147.533946788 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 06:50:24 crc kubenswrapper[4846]: I1005 06:50:24.793380 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rf9l4\" (UID: \"78d3e7b1-7258-4508-ad38-373f21bb5d96\") " pod="openshift-image-registry/image-registry-697d97f7c8-rf9l4" Oct 05 06:50:24 crc kubenswrapper[4846]: E1005 06:50:24.793698 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-05 06:50:25.293690048 +0000 UTC m=+147.534542823 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rf9l4" (UID: "78d3e7b1-7258-4508-ad38-373f21bb5d96") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 06:50:24 crc kubenswrapper[4846]: I1005 06:50:24.894200 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 05 06:50:24 crc kubenswrapper[4846]: E1005 06:50:24.894415 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-05 06:50:25.394379681 +0000 UTC m=+147.635232456 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 06:50:24 crc kubenswrapper[4846]: I1005 06:50:24.894828 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rf9l4\" (UID: \"78d3e7b1-7258-4508-ad38-373f21bb5d96\") " pod="openshift-image-registry/image-registry-697d97f7c8-rf9l4" Oct 05 06:50:24 crc kubenswrapper[4846]: E1005 06:50:24.895274 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-05 06:50:25.395255074 +0000 UTC m=+147.636107839 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rf9l4" (UID: "78d3e7b1-7258-4508-ad38-373f21bb5d96") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 06:50:24 crc kubenswrapper[4846]: I1005 06:50:24.996471 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 05 06:50:24 crc kubenswrapper[4846]: E1005 06:50:24.996907 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-05 06:50:25.496886421 +0000 UTC m=+147.737739206 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 06:50:25 crc kubenswrapper[4846]: I1005 06:50:25.098578 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rf9l4\" (UID: \"78d3e7b1-7258-4508-ad38-373f21bb5d96\") " pod="openshift-image-registry/image-registry-697d97f7c8-rf9l4" Oct 05 06:50:25 crc kubenswrapper[4846]: E1005 06:50:25.099033 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-05 06:50:25.599017081 +0000 UTC m=+147.839869856 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rf9l4" (UID: "78d3e7b1-7258-4508-ad38-373f21bb5d96") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 06:50:25 crc kubenswrapper[4846]: I1005 06:50:25.199833 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 05 06:50:25 crc kubenswrapper[4846]: E1005 06:50:25.200380 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-05 06:50:25.70035801 +0000 UTC m=+147.941210785 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 06:50:25 crc kubenswrapper[4846]: I1005 06:50:25.248681 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-jm4gb" event={"ID":"c65d29d9-66e3-41ce-acb2-cec28c52fb29","Type":"ContainerStarted","Data":"7039e5aee9dbf6ea382cf6ca87aec852afcc6a55e8487c428f7533f5d9add645"} Oct 05 06:50:25 crc kubenswrapper[4846]: I1005 06:50:25.248841 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-jm4gb" Oct 05 06:50:25 crc kubenswrapper[4846]: I1005 06:50:25.251676 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-6bpvf" event={"ID":"75d0c362-7aa3-4919-937d-8849f425466e","Type":"ContainerStarted","Data":"7177fd0bc242130a5bfbad7a9448837c0cbd2ffbd24b3bb6c309d3ce247fae26"} Oct 05 06:50:25 crc kubenswrapper[4846]: I1005 06:50:25.251718 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-6bpvf" event={"ID":"75d0c362-7aa3-4919-937d-8849f425466e","Type":"ContainerStarted","Data":"1636ac6f235724cf1098a65f21f45733e1e88175c9cc6c9f7aa106960ab30e40"} Oct 05 06:50:25 crc kubenswrapper[4846]: I1005 06:50:25.254297 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-dtbh5" event={"ID":"0755bb69-c94a-4b36-97ee-8a3dfd972c79","Type":"ContainerStarted","Data":"b0bfa7c125862ab6e3b98822420bf1a0b44385c77f551d25893adc00a18d91c0"} Oct 05 06:50:25 crc kubenswrapper[4846]: I1005 06:50:25.255626 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-dsqhc" event={"ID":"5fed566e-ecb6-4801-84b3-ad1143a03f08","Type":"ContainerStarted","Data":"92cc78df67ca92aef1ff0735627bc60bbb55dea7278cfdf8262be82d97905adb"} Oct 05 06:50:25 crc kubenswrapper[4846]: I1005 06:50:25.256159 4846 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-2rc5n container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.38:8080/healthz\": dial tcp 10.217.0.38:8080: connect: connection refused" start-of-body= Oct 05 06:50:25 crc kubenswrapper[4846]: I1005 06:50:25.256232 4846 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-2rc5n" podUID="8c2f8349-aa93-4ac8-a369-5b4b2b9cd6e2" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.38:8080/healthz\": dial tcp 10.217.0.38:8080: connect: connection refused" Oct 05 06:50:25 crc kubenswrapper[4846]: I1005 06:50:25.267539 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-hmzsn" Oct 05 06:50:25 crc kubenswrapper[4846]: I1005 06:50:25.276388 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-zf4p8" Oct 05 06:50:25 crc kubenswrapper[4846]: I1005 06:50:25.302219 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rf9l4\" (UID: \"78d3e7b1-7258-4508-ad38-373f21bb5d96\") " pod="openshift-image-registry/image-registry-697d97f7c8-rf9l4" Oct 05 06:50:25 crc kubenswrapper[4846]: E1005 06:50:25.304828 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-05 06:50:25.804810491 +0000 UTC m=+148.045663266 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rf9l4" (UID: "78d3e7b1-7258-4508-ad38-373f21bb5d96") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 06:50:25 crc kubenswrapper[4846]: I1005 06:50:25.346127 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-jm4gb" podStartSLOduration=8.346102585 podStartE2EDuration="8.346102585s" podCreationTimestamp="2025-10-05 06:50:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 06:50:25.296794991 +0000 UTC m=+147.537647766" watchObservedRunningTime="2025-10-05 06:50:25.346102585 +0000 UTC m=+147.586955360" Oct 05 06:50:25 crc kubenswrapper[4846]: I1005 06:50:25.346803 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-dtbh5" podStartSLOduration=121.346796763 podStartE2EDuration="2m1.346796763s" podCreationTimestamp="2025-10-05 06:48:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 06:50:25.343483536 +0000 UTC m=+147.584336311" watchObservedRunningTime="2025-10-05 06:50:25.346796763 +0000 UTC m=+147.587649538" Oct 05 06:50:25 crc kubenswrapper[4846]: I1005 06:50:25.404447 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 05 06:50:25 crc kubenswrapper[4846]: E1005 06:50:25.404630 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-05 06:50:25.90460347 +0000 UTC m=+148.145456245 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 06:50:25 crc kubenswrapper[4846]: I1005 06:50:25.404771 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rf9l4\" (UID: \"78d3e7b1-7258-4508-ad38-373f21bb5d96\") " pod="openshift-image-registry/image-registry-697d97f7c8-rf9l4" Oct 05 06:50:25 crc kubenswrapper[4846]: E1005 06:50:25.405157 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-05 06:50:25.905149454 +0000 UTC m=+148.146002229 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rf9l4" (UID: "78d3e7b1-7258-4508-ad38-373f21bb5d96") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 06:50:25 crc kubenswrapper[4846]: I1005 06:50:25.505490 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 05 06:50:25 crc kubenswrapper[4846]: E1005 06:50:25.505779 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-05 06:50:26.005761155 +0000 UTC m=+148.246613920 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 06:50:25 crc kubenswrapper[4846]: I1005 06:50:25.542492 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-6bpvf" podStartSLOduration=122.542474588 podStartE2EDuration="2m2.542474588s" podCreationTimestamp="2025-10-05 06:48:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 06:50:25.540425154 +0000 UTC m=+147.781277929" watchObservedRunningTime="2025-10-05 06:50:25.542474588 +0000 UTC m=+147.783327363" Oct 05 06:50:25 crc kubenswrapper[4846]: I1005 06:50:25.607035 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 05 06:50:25 crc kubenswrapper[4846]: I1005 06:50:25.607085 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 05 06:50:25 crc kubenswrapper[4846]: I1005 06:50:25.607111 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 05 06:50:25 crc kubenswrapper[4846]: I1005 06:50:25.607141 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rf9l4\" (UID: \"78d3e7b1-7258-4508-ad38-373f21bb5d96\") " pod="openshift-image-registry/image-registry-697d97f7c8-rf9l4" Oct 05 06:50:25 crc kubenswrapper[4846]: I1005 06:50:25.607190 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 05 06:50:25 crc kubenswrapper[4846]: E1005 06:50:25.607765 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-05 06:50:26.10773944 +0000 UTC m=+148.348592215 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rf9l4" (UID: "78d3e7b1-7258-4508-ad38-373f21bb5d96") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 06:50:25 crc kubenswrapper[4846]: I1005 06:50:25.608128 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 05 06:50:25 crc kubenswrapper[4846]: I1005 06:50:25.613143 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 05 06:50:25 crc kubenswrapper[4846]: I1005 06:50:25.613974 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 05 06:50:25 crc kubenswrapper[4846]: I1005 06:50:25.626452 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 05 06:50:25 crc kubenswrapper[4846]: I1005 06:50:25.628535 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 05 06:50:25 crc kubenswrapper[4846]: I1005 06:50:25.711570 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 05 06:50:25 crc kubenswrapper[4846]: E1005 06:50:25.711959 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-05 06:50:26.211938454 +0000 UTC m=+148.452791229 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 06:50:25 crc kubenswrapper[4846]: I1005 06:50:25.721673 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 05 06:50:25 crc kubenswrapper[4846]: I1005 06:50:25.724790 4846 patch_prober.go:28] interesting pod/router-default-5444994796-58hwl container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 05 06:50:25 crc kubenswrapper[4846]: [-]has-synced failed: reason withheld Oct 05 06:50:25 crc kubenswrapper[4846]: [+]process-running ok Oct 05 06:50:25 crc kubenswrapper[4846]: healthz check failed Oct 05 06:50:25 crc kubenswrapper[4846]: I1005 06:50:25.724827 4846 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-58hwl" podUID="2ecfaba9-9034-4ac7-bd51-e3f64b41d825" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 05 06:50:25 crc kubenswrapper[4846]: I1005 06:50:25.734657 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 05 06:50:25 crc kubenswrapper[4846]: I1005 06:50:25.814939 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rf9l4\" (UID: \"78d3e7b1-7258-4508-ad38-373f21bb5d96\") " pod="openshift-image-registry/image-registry-697d97f7c8-rf9l4" Oct 05 06:50:25 crc kubenswrapper[4846]: E1005 06:50:25.815345 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-05 06:50:26.315328228 +0000 UTC m=+148.556181003 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rf9l4" (UID: "78d3e7b1-7258-4508-ad38-373f21bb5d96") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 06:50:25 crc kubenswrapper[4846]: I1005 06:50:25.917979 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 05 06:50:25 crc kubenswrapper[4846]: E1005 06:50:25.918319 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-05 06:50:26.41829942 +0000 UTC m=+148.659152195 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 06:50:25 crc kubenswrapper[4846]: I1005 06:50:25.983811 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-rcnlv" Oct 05 06:50:25 crc kubenswrapper[4846]: I1005 06:50:25.995433 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-4pvwg" Oct 05 06:50:26 crc kubenswrapper[4846]: I1005 06:50:26.022206 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rf9l4\" (UID: \"78d3e7b1-7258-4508-ad38-373f21bb5d96\") " pod="openshift-image-registry/image-registry-697d97f7c8-rf9l4" Oct 05 06:50:26 crc kubenswrapper[4846]: E1005 06:50:26.022579 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-05 06:50:26.522564356 +0000 UTC m=+148.763417131 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rf9l4" (UID: "78d3e7b1-7258-4508-ad38-373f21bb5d96") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 06:50:26 crc kubenswrapper[4846]: I1005 06:50:26.127642 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 05 06:50:26 crc kubenswrapper[4846]: E1005 06:50:26.128491 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-05 06:50:26.628467555 +0000 UTC m=+148.869320320 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 06:50:26 crc kubenswrapper[4846]: I1005 06:50:26.231870 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rf9l4\" (UID: \"78d3e7b1-7258-4508-ad38-373f21bb5d96\") " pod="openshift-image-registry/image-registry-697d97f7c8-rf9l4" Oct 05 06:50:26 crc kubenswrapper[4846]: E1005 06:50:26.232234 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-05 06:50:26.732221118 +0000 UTC m=+148.973073893 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rf9l4" (UID: "78d3e7b1-7258-4508-ad38-373f21bb5d96") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 06:50:26 crc kubenswrapper[4846]: I1005 06:50:26.334833 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 05 06:50:26 crc kubenswrapper[4846]: E1005 06:50:26.335058 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-05 06:50:26.835031996 +0000 UTC m=+149.075884771 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 06:50:26 crc kubenswrapper[4846]: I1005 06:50:26.335165 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rf9l4\" (UID: \"78d3e7b1-7258-4508-ad38-373f21bb5d96\") " pod="openshift-image-registry/image-registry-697d97f7c8-rf9l4" Oct 05 06:50:26 crc kubenswrapper[4846]: E1005 06:50:26.335518 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-05 06:50:26.835511669 +0000 UTC m=+149.076364444 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rf9l4" (UID: "78d3e7b1-7258-4508-ad38-373f21bb5d96") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 06:50:26 crc kubenswrapper[4846]: I1005 06:50:26.356419 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-dsqhc" event={"ID":"5fed566e-ecb6-4801-84b3-ad1143a03f08","Type":"ContainerStarted","Data":"636214208db4df14723b9b17495641b7cc62009ca2ba98d85ce2ca87ea55fddc"} Oct 05 06:50:26 crc kubenswrapper[4846]: I1005 06:50:26.435943 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 05 06:50:26 crc kubenswrapper[4846]: E1005 06:50:26.436383 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-05 06:50:26.936364805 +0000 UTC m=+149.177217580 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 06:50:26 crc kubenswrapper[4846]: I1005 06:50:26.542993 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rf9l4\" (UID: \"78d3e7b1-7258-4508-ad38-373f21bb5d96\") " pod="openshift-image-registry/image-registry-697d97f7c8-rf9l4" Oct 05 06:50:26 crc kubenswrapper[4846]: E1005 06:50:26.546063 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-05 06:50:27.046044243 +0000 UTC m=+149.286897018 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rf9l4" (UID: "78d3e7b1-7258-4508-ad38-373f21bb5d96") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 06:50:26 crc kubenswrapper[4846]: I1005 06:50:26.643599 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 05 06:50:26 crc kubenswrapper[4846]: E1005 06:50:26.644247 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-05 06:50:27.14423049 +0000 UTC m=+149.385083255 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 06:50:26 crc kubenswrapper[4846]: I1005 06:50:26.726375 4846 patch_prober.go:28] interesting pod/router-default-5444994796-58hwl container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 05 06:50:26 crc kubenswrapper[4846]: [-]has-synced failed: reason withheld Oct 05 06:50:26 crc kubenswrapper[4846]: [+]process-running ok Oct 05 06:50:26 crc kubenswrapper[4846]: healthz check failed Oct 05 06:50:26 crc kubenswrapper[4846]: I1005 06:50:26.726444 4846 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-58hwl" podUID="2ecfaba9-9034-4ac7-bd51-e3f64b41d825" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 05 06:50:26 crc kubenswrapper[4846]: I1005 06:50:26.747039 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rf9l4\" (UID: \"78d3e7b1-7258-4508-ad38-373f21bb5d96\") " pod="openshift-image-registry/image-registry-697d97f7c8-rf9l4" Oct 05 06:50:26 crc kubenswrapper[4846]: E1005 06:50:26.747457 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-05 06:50:27.247441959 +0000 UTC m=+149.488294734 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rf9l4" (UID: "78d3e7b1-7258-4508-ad38-373f21bb5d96") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 06:50:26 crc kubenswrapper[4846]: I1005 06:50:26.747816 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-mzc7c"] Oct 05 06:50:26 crc kubenswrapper[4846]: I1005 06:50:26.749213 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mzc7c" Oct 05 06:50:26 crc kubenswrapper[4846]: I1005 06:50:26.767507 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Oct 05 06:50:26 crc kubenswrapper[4846]: I1005 06:50:26.768695 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-mzc7c"] Oct 05 06:50:26 crc kubenswrapper[4846]: I1005 06:50:26.852220 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 05 06:50:26 crc kubenswrapper[4846]: I1005 06:50:26.852342 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/da62c2a4-387f-4d5a-a9f2-04bd1e110d20-utilities\") pod \"certified-operators-mzc7c\" (UID: \"da62c2a4-387f-4d5a-a9f2-04bd1e110d20\") " pod="openshift-marketplace/certified-operators-mzc7c" Oct 05 06:50:26 crc kubenswrapper[4846]: I1005 06:50:26.852369 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/da62c2a4-387f-4d5a-a9f2-04bd1e110d20-catalog-content\") pod \"certified-operators-mzc7c\" (UID: \"da62c2a4-387f-4d5a-a9f2-04bd1e110d20\") " pod="openshift-marketplace/certified-operators-mzc7c" Oct 05 06:50:26 crc kubenswrapper[4846]: I1005 06:50:26.852418 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hmh5t\" (UniqueName: \"kubernetes.io/projected/da62c2a4-387f-4d5a-a9f2-04bd1e110d20-kube-api-access-hmh5t\") pod \"certified-operators-mzc7c\" (UID: \"da62c2a4-387f-4d5a-a9f2-04bd1e110d20\") " pod="openshift-marketplace/certified-operators-mzc7c" Oct 05 06:50:26 crc kubenswrapper[4846]: E1005 06:50:26.852545 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-05 06:50:27.352526306 +0000 UTC m=+149.593379071 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 06:50:26 crc kubenswrapper[4846]: I1005 06:50:26.953412 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/da62c2a4-387f-4d5a-a9f2-04bd1e110d20-utilities\") pod \"certified-operators-mzc7c\" (UID: \"da62c2a4-387f-4d5a-a9f2-04bd1e110d20\") " pod="openshift-marketplace/certified-operators-mzc7c" Oct 05 06:50:26 crc kubenswrapper[4846]: I1005 06:50:26.953781 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/da62c2a4-387f-4d5a-a9f2-04bd1e110d20-catalog-content\") pod \"certified-operators-mzc7c\" (UID: \"da62c2a4-387f-4d5a-a9f2-04bd1e110d20\") " pod="openshift-marketplace/certified-operators-mzc7c" Oct 05 06:50:26 crc kubenswrapper[4846]: I1005 06:50:26.953815 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rf9l4\" (UID: \"78d3e7b1-7258-4508-ad38-373f21bb5d96\") " pod="openshift-image-registry/image-registry-697d97f7c8-rf9l4" Oct 05 06:50:26 crc kubenswrapper[4846]: I1005 06:50:26.953847 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hmh5t\" (UniqueName: \"kubernetes.io/projected/da62c2a4-387f-4d5a-a9f2-04bd1e110d20-kube-api-access-hmh5t\") pod \"certified-operators-mzc7c\" (UID: \"da62c2a4-387f-4d5a-a9f2-04bd1e110d20\") " pod="openshift-marketplace/certified-operators-mzc7c" Oct 05 06:50:26 crc kubenswrapper[4846]: I1005 06:50:26.953953 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/da62c2a4-387f-4d5a-a9f2-04bd1e110d20-utilities\") pod \"certified-operators-mzc7c\" (UID: \"da62c2a4-387f-4d5a-a9f2-04bd1e110d20\") " pod="openshift-marketplace/certified-operators-mzc7c" Oct 05 06:50:26 crc kubenswrapper[4846]: E1005 06:50:26.954310 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-05 06:50:27.454292417 +0000 UTC m=+149.695145192 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rf9l4" (UID: "78d3e7b1-7258-4508-ad38-373f21bb5d96") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 06:50:26 crc kubenswrapper[4846]: I1005 06:50:26.954329 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/da62c2a4-387f-4d5a-a9f2-04bd1e110d20-catalog-content\") pod \"certified-operators-mzc7c\" (UID: \"da62c2a4-387f-4d5a-a9f2-04bd1e110d20\") " pod="openshift-marketplace/certified-operators-mzc7c" Oct 05 06:50:26 crc kubenswrapper[4846]: I1005 06:50:26.983103 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hmh5t\" (UniqueName: \"kubernetes.io/projected/da62c2a4-387f-4d5a-a9f2-04bd1e110d20-kube-api-access-hmh5t\") pod \"certified-operators-mzc7c\" (UID: \"da62c2a4-387f-4d5a-a9f2-04bd1e110d20\") " pod="openshift-marketplace/certified-operators-mzc7c" Oct 05 06:50:27 crc kubenswrapper[4846]: I1005 06:50:27.056349 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 05 06:50:27 crc kubenswrapper[4846]: E1005 06:50:27.056965 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-05 06:50:27.55691993 +0000 UTC m=+149.797772705 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 06:50:27 crc kubenswrapper[4846]: I1005 06:50:27.113041 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-s9f8z"] Oct 05 06:50:27 crc kubenswrapper[4846]: I1005 06:50:27.114273 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-s9f8z" Oct 05 06:50:27 crc kubenswrapper[4846]: I1005 06:50:27.135624 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mzc7c" Oct 05 06:50:27 crc kubenswrapper[4846]: I1005 06:50:27.139218 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-s9f8z"] Oct 05 06:50:27 crc kubenswrapper[4846]: I1005 06:50:27.172142 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/39351673-4e84-4ce0-b669-e2ca9072a443-catalog-content\") pod \"certified-operators-s9f8z\" (UID: \"39351673-4e84-4ce0-b669-e2ca9072a443\") " pod="openshift-marketplace/certified-operators-s9f8z" Oct 05 06:50:27 crc kubenswrapper[4846]: I1005 06:50:27.172210 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/39351673-4e84-4ce0-b669-e2ca9072a443-utilities\") pod \"certified-operators-s9f8z\" (UID: \"39351673-4e84-4ce0-b669-e2ca9072a443\") " pod="openshift-marketplace/certified-operators-s9f8z" Oct 05 06:50:27 crc kubenswrapper[4846]: I1005 06:50:27.172242 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7dh2q\" (UniqueName: \"kubernetes.io/projected/39351673-4e84-4ce0-b669-e2ca9072a443-kube-api-access-7dh2q\") pod \"certified-operators-s9f8z\" (UID: \"39351673-4e84-4ce0-b669-e2ca9072a443\") " pod="openshift-marketplace/certified-operators-s9f8z" Oct 05 06:50:27 crc kubenswrapper[4846]: I1005 06:50:27.172272 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rf9l4\" (UID: \"78d3e7b1-7258-4508-ad38-373f21bb5d96\") " pod="openshift-image-registry/image-registry-697d97f7c8-rf9l4" Oct 05 06:50:27 crc kubenswrapper[4846]: E1005 06:50:27.172610 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-05 06:50:27.672591606 +0000 UTC m=+149.913444391 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rf9l4" (UID: "78d3e7b1-7258-4508-ad38-373f21bb5d96") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 06:50:27 crc kubenswrapper[4846]: I1005 06:50:27.272973 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 05 06:50:27 crc kubenswrapper[4846]: E1005 06:50:27.273194 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-05 06:50:27.773153235 +0000 UTC m=+150.014006020 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 06:50:27 crc kubenswrapper[4846]: I1005 06:50:27.273555 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/39351673-4e84-4ce0-b669-e2ca9072a443-utilities\") pod \"certified-operators-s9f8z\" (UID: \"39351673-4e84-4ce0-b669-e2ca9072a443\") " pod="openshift-marketplace/certified-operators-s9f8z" Oct 05 06:50:27 crc kubenswrapper[4846]: I1005 06:50:27.273594 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7dh2q\" (UniqueName: \"kubernetes.io/projected/39351673-4e84-4ce0-b669-e2ca9072a443-kube-api-access-7dh2q\") pod \"certified-operators-s9f8z\" (UID: \"39351673-4e84-4ce0-b669-e2ca9072a443\") " pod="openshift-marketplace/certified-operators-s9f8z" Oct 05 06:50:27 crc kubenswrapper[4846]: I1005 06:50:27.273916 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rf9l4\" (UID: \"78d3e7b1-7258-4508-ad38-373f21bb5d96\") " pod="openshift-image-registry/image-registry-697d97f7c8-rf9l4" Oct 05 06:50:27 crc kubenswrapper[4846]: I1005 06:50:27.273999 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/39351673-4e84-4ce0-b669-e2ca9072a443-catalog-content\") pod \"certified-operators-s9f8z\" (UID: \"39351673-4e84-4ce0-b669-e2ca9072a443\") " pod="openshift-marketplace/certified-operators-s9f8z" Oct 05 06:50:27 crc kubenswrapper[4846]: E1005 06:50:27.274254 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-05 06:50:27.774228603 +0000 UTC m=+150.015081378 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rf9l4" (UID: "78d3e7b1-7258-4508-ad38-373f21bb5d96") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 06:50:27 crc kubenswrapper[4846]: I1005 06:50:27.274456 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/39351673-4e84-4ce0-b669-e2ca9072a443-utilities\") pod \"certified-operators-s9f8z\" (UID: \"39351673-4e84-4ce0-b669-e2ca9072a443\") " pod="openshift-marketplace/certified-operators-s9f8z" Oct 05 06:50:27 crc kubenswrapper[4846]: I1005 06:50:27.274504 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/39351673-4e84-4ce0-b669-e2ca9072a443-catalog-content\") pod \"certified-operators-s9f8z\" (UID: \"39351673-4e84-4ce0-b669-e2ca9072a443\") " pod="openshift-marketplace/certified-operators-s9f8z" Oct 05 06:50:27 crc kubenswrapper[4846]: I1005 06:50:27.305717 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-76jjh"] Oct 05 06:50:27 crc kubenswrapper[4846]: I1005 06:50:27.306777 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-76jjh" Oct 05 06:50:27 crc kubenswrapper[4846]: I1005 06:50:27.309774 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Oct 05 06:50:27 crc kubenswrapper[4846]: I1005 06:50:27.330109 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-76jjh"] Oct 05 06:50:27 crc kubenswrapper[4846]: I1005 06:50:27.334285 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7dh2q\" (UniqueName: \"kubernetes.io/projected/39351673-4e84-4ce0-b669-e2ca9072a443-kube-api-access-7dh2q\") pod \"certified-operators-s9f8z\" (UID: \"39351673-4e84-4ce0-b669-e2ca9072a443\") " pod="openshift-marketplace/certified-operators-s9f8z" Oct 05 06:50:27 crc kubenswrapper[4846]: I1005 06:50:27.375699 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 05 06:50:27 crc kubenswrapper[4846]: E1005 06:50:27.376023 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-05 06:50:27.876005464 +0000 UTC m=+150.116858239 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 06:50:27 crc kubenswrapper[4846]: I1005 06:50:27.385119 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-dsqhc" event={"ID":"5fed566e-ecb6-4801-84b3-ad1143a03f08","Type":"ContainerStarted","Data":"e2d9c21bb6b0fc2b3b1b55b2ce3fa1d7120cc3384e147385e97d9951c3b8e484"} Oct 05 06:50:27 crc kubenswrapper[4846]: I1005 06:50:27.393455 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"925f00ac02f72a44f9475d4fdce55fad0245200951f3b56f1656e1014bd937cc"} Oct 05 06:50:27 crc kubenswrapper[4846]: I1005 06:50:27.393506 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"6daa603a018a802def6bc61ad935dff4de81c5722c39df0ae3beb42c507d2bc7"} Oct 05 06:50:27 crc kubenswrapper[4846]: I1005 06:50:27.411574 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"8222a201d70677d5bd0496321511767e3b6f301bc28d1a4d714cfd62fbfd2561"} Oct 05 06:50:27 crc kubenswrapper[4846]: I1005 06:50:27.411628 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"e7f66faf13434d275f35c568629ce9b578773e522c783d236bd6f7054c60f234"} Oct 05 06:50:27 crc kubenswrapper[4846]: I1005 06:50:27.412295 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 05 06:50:27 crc kubenswrapper[4846]: I1005 06:50:27.426577 4846 generic.go:334] "Generic (PLEG): container finished" podID="41aa056a-34d5-463b-8352-c71e0f0db1a4" containerID="e46a0ae8f0a310b6735443855f8119ee77787d11058d4f9af366550a7374d17d" exitCode=0 Oct 05 06:50:27 crc kubenswrapper[4846]: I1005 06:50:27.426659 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29327445-dn5mf" event={"ID":"41aa056a-34d5-463b-8352-c71e0f0db1a4","Type":"ContainerDied","Data":"e46a0ae8f0a310b6735443855f8119ee77787d11058d4f9af366550a7374d17d"} Oct 05 06:50:27 crc kubenswrapper[4846]: I1005 06:50:27.446051 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"004c54b5d2f101622c2d9955ca0e5c68dfb729251970e129854102878a4debb1"} Oct 05 06:50:27 crc kubenswrapper[4846]: I1005 06:50:27.446106 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"11b39665d042da42c8ca8f70d97c127715a599030afc150f0a1876ce7d541b84"} Oct 05 06:50:27 crc kubenswrapper[4846]: I1005 06:50:27.447944 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-s9f8z" Oct 05 06:50:27 crc kubenswrapper[4846]: I1005 06:50:27.477127 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/73ca18b5-0a68-4990-9e7c-1685f87c8dcb-utilities\") pod \"community-operators-76jjh\" (UID: \"73ca18b5-0a68-4990-9e7c-1685f87c8dcb\") " pod="openshift-marketplace/community-operators-76jjh" Oct 05 06:50:27 crc kubenswrapper[4846]: I1005 06:50:27.477476 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kjdm9\" (UniqueName: \"kubernetes.io/projected/73ca18b5-0a68-4990-9e7c-1685f87c8dcb-kube-api-access-kjdm9\") pod \"community-operators-76jjh\" (UID: \"73ca18b5-0a68-4990-9e7c-1685f87c8dcb\") " pod="openshift-marketplace/community-operators-76jjh" Oct 05 06:50:27 crc kubenswrapper[4846]: I1005 06:50:27.477517 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rf9l4\" (UID: \"78d3e7b1-7258-4508-ad38-373f21bb5d96\") " pod="openshift-image-registry/image-registry-697d97f7c8-rf9l4" Oct 05 06:50:27 crc kubenswrapper[4846]: I1005 06:50:27.477572 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/73ca18b5-0a68-4990-9e7c-1685f87c8dcb-catalog-content\") pod \"community-operators-76jjh\" (UID: \"73ca18b5-0a68-4990-9e7c-1685f87c8dcb\") " pod="openshift-marketplace/community-operators-76jjh" Oct 05 06:50:27 crc kubenswrapper[4846]: E1005 06:50:27.477996 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-05 06:50:27.9779799 +0000 UTC m=+150.218832675 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rf9l4" (UID: "78d3e7b1-7258-4508-ad38-373f21bb5d96") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 06:50:27 crc kubenswrapper[4846]: I1005 06:50:27.524740 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-brf8j"] Oct 05 06:50:27 crc kubenswrapper[4846]: I1005 06:50:27.538005 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-brf8j" Oct 05 06:50:27 crc kubenswrapper[4846]: I1005 06:50:27.543979 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-brf8j"] Oct 05 06:50:27 crc kubenswrapper[4846]: I1005 06:50:27.586000 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 05 06:50:27 crc kubenswrapper[4846]: E1005 06:50:27.587288 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-05 06:50:28.087261558 +0000 UTC m=+150.328114333 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 06:50:27 crc kubenswrapper[4846]: I1005 06:50:27.594438 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8bdbd298-f264-43ea-bb5e-625b1e5ab562-utilities\") pod \"community-operators-brf8j\" (UID: \"8bdbd298-f264-43ea-bb5e-625b1e5ab562\") " pod="openshift-marketplace/community-operators-brf8j" Oct 05 06:50:27 crc kubenswrapper[4846]: I1005 06:50:27.594659 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/73ca18b5-0a68-4990-9e7c-1685f87c8dcb-utilities\") pod \"community-operators-76jjh\" (UID: \"73ca18b5-0a68-4990-9e7c-1685f87c8dcb\") " pod="openshift-marketplace/community-operators-76jjh" Oct 05 06:50:27 crc kubenswrapper[4846]: I1005 06:50:27.594742 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kjdm9\" (UniqueName: \"kubernetes.io/projected/73ca18b5-0a68-4990-9e7c-1685f87c8dcb-kube-api-access-kjdm9\") pod \"community-operators-76jjh\" (UID: \"73ca18b5-0a68-4990-9e7c-1685f87c8dcb\") " pod="openshift-marketplace/community-operators-76jjh" Oct 05 06:50:27 crc kubenswrapper[4846]: I1005 06:50:27.594782 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rf9l4\" (UID: \"78d3e7b1-7258-4508-ad38-373f21bb5d96\") " pod="openshift-image-registry/image-registry-697d97f7c8-rf9l4" Oct 05 06:50:27 crc kubenswrapper[4846]: I1005 06:50:27.594878 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dvncs\" (UniqueName: \"kubernetes.io/projected/8bdbd298-f264-43ea-bb5e-625b1e5ab562-kube-api-access-dvncs\") pod \"community-operators-brf8j\" (UID: \"8bdbd298-f264-43ea-bb5e-625b1e5ab562\") " pod="openshift-marketplace/community-operators-brf8j" Oct 05 06:50:27 crc kubenswrapper[4846]: I1005 06:50:27.595048 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/73ca18b5-0a68-4990-9e7c-1685f87c8dcb-catalog-content\") pod \"community-operators-76jjh\" (UID: \"73ca18b5-0a68-4990-9e7c-1685f87c8dcb\") " pod="openshift-marketplace/community-operators-76jjh" Oct 05 06:50:27 crc kubenswrapper[4846]: I1005 06:50:27.595159 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8bdbd298-f264-43ea-bb5e-625b1e5ab562-catalog-content\") pod \"community-operators-brf8j\" (UID: \"8bdbd298-f264-43ea-bb5e-625b1e5ab562\") " pod="openshift-marketplace/community-operators-brf8j" Oct 05 06:50:27 crc kubenswrapper[4846]: E1005 06:50:27.597370 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-05 06:50:28.097355942 +0000 UTC m=+150.338208717 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rf9l4" (UID: "78d3e7b1-7258-4508-ad38-373f21bb5d96") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 06:50:27 crc kubenswrapper[4846]: I1005 06:50:27.598676 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/73ca18b5-0a68-4990-9e7c-1685f87c8dcb-utilities\") pod \"community-operators-76jjh\" (UID: \"73ca18b5-0a68-4990-9e7c-1685f87c8dcb\") " pod="openshift-marketplace/community-operators-76jjh" Oct 05 06:50:27 crc kubenswrapper[4846]: I1005 06:50:27.602913 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/73ca18b5-0a68-4990-9e7c-1685f87c8dcb-catalog-content\") pod \"community-operators-76jjh\" (UID: \"73ca18b5-0a68-4990-9e7c-1685f87c8dcb\") " pod="openshift-marketplace/community-operators-76jjh" Oct 05 06:50:27 crc kubenswrapper[4846]: I1005 06:50:27.633615 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kjdm9\" (UniqueName: \"kubernetes.io/projected/73ca18b5-0a68-4990-9e7c-1685f87c8dcb-kube-api-access-kjdm9\") pod \"community-operators-76jjh\" (UID: \"73ca18b5-0a68-4990-9e7c-1685f87c8dcb\") " pod="openshift-marketplace/community-operators-76jjh" Oct 05 06:50:27 crc kubenswrapper[4846]: I1005 06:50:27.696683 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 05 06:50:27 crc kubenswrapper[4846]: I1005 06:50:27.696962 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8bdbd298-f264-43ea-bb5e-625b1e5ab562-catalog-content\") pod \"community-operators-brf8j\" (UID: \"8bdbd298-f264-43ea-bb5e-625b1e5ab562\") " pod="openshift-marketplace/community-operators-brf8j" Oct 05 06:50:27 crc kubenswrapper[4846]: I1005 06:50:27.697034 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8bdbd298-f264-43ea-bb5e-625b1e5ab562-utilities\") pod \"community-operators-brf8j\" (UID: \"8bdbd298-f264-43ea-bb5e-625b1e5ab562\") " pod="openshift-marketplace/community-operators-brf8j" Oct 05 06:50:27 crc kubenswrapper[4846]: I1005 06:50:27.697090 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dvncs\" (UniqueName: \"kubernetes.io/projected/8bdbd298-f264-43ea-bb5e-625b1e5ab562-kube-api-access-dvncs\") pod \"community-operators-brf8j\" (UID: \"8bdbd298-f264-43ea-bb5e-625b1e5ab562\") " pod="openshift-marketplace/community-operators-brf8j" Oct 05 06:50:27 crc kubenswrapper[4846]: I1005 06:50:27.698122 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8bdbd298-f264-43ea-bb5e-625b1e5ab562-utilities\") pod \"community-operators-brf8j\" (UID: \"8bdbd298-f264-43ea-bb5e-625b1e5ab562\") " pod="openshift-marketplace/community-operators-brf8j" Oct 05 06:50:27 crc kubenswrapper[4846]: E1005 06:50:27.698218 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-05 06:50:28.198201339 +0000 UTC m=+150.439054114 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 06:50:27 crc kubenswrapper[4846]: I1005 06:50:27.698531 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8bdbd298-f264-43ea-bb5e-625b1e5ab562-catalog-content\") pod \"community-operators-brf8j\" (UID: \"8bdbd298-f264-43ea-bb5e-625b1e5ab562\") " pod="openshift-marketplace/community-operators-brf8j" Oct 05 06:50:27 crc kubenswrapper[4846]: I1005 06:50:27.721117 4846 patch_prober.go:28] interesting pod/router-default-5444994796-58hwl container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 05 06:50:27 crc kubenswrapper[4846]: [-]has-synced failed: reason withheld Oct 05 06:50:27 crc kubenswrapper[4846]: [+]process-running ok Oct 05 06:50:27 crc kubenswrapper[4846]: healthz check failed Oct 05 06:50:27 crc kubenswrapper[4846]: I1005 06:50:27.721481 4846 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-58hwl" podUID="2ecfaba9-9034-4ac7-bd51-e3f64b41d825" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 05 06:50:27 crc kubenswrapper[4846]: I1005 06:50:27.738248 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dvncs\" (UniqueName: \"kubernetes.io/projected/8bdbd298-f264-43ea-bb5e-625b1e5ab562-kube-api-access-dvncs\") pod \"community-operators-brf8j\" (UID: \"8bdbd298-f264-43ea-bb5e-625b1e5ab562\") " pod="openshift-marketplace/community-operators-brf8j" Oct 05 06:50:27 crc kubenswrapper[4846]: I1005 06:50:27.766784 4846 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Oct 05 06:50:27 crc kubenswrapper[4846]: I1005 06:50:27.798932 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rf9l4\" (UID: \"78d3e7b1-7258-4508-ad38-373f21bb5d96\") " pod="openshift-image-registry/image-registry-697d97f7c8-rf9l4" Oct 05 06:50:27 crc kubenswrapper[4846]: E1005 06:50:27.799381 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-05 06:50:28.299365194 +0000 UTC m=+150.540217969 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rf9l4" (UID: "78d3e7b1-7258-4508-ad38-373f21bb5d96") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 06:50:27 crc kubenswrapper[4846]: I1005 06:50:27.854600 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-mzc7c"] Oct 05 06:50:27 crc kubenswrapper[4846]: I1005 06:50:27.888566 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-brf8j" Oct 05 06:50:27 crc kubenswrapper[4846]: I1005 06:50:27.900835 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 05 06:50:27 crc kubenswrapper[4846]: E1005 06:50:27.901265 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-05 06:50:28.401243687 +0000 UTC m=+150.642096462 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 06:50:27 crc kubenswrapper[4846]: I1005 06:50:27.927910 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-76jjh" Oct 05 06:50:28 crc kubenswrapper[4846]: I1005 06:50:28.001870 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rf9l4\" (UID: \"78d3e7b1-7258-4508-ad38-373f21bb5d96\") " pod="openshift-image-registry/image-registry-697d97f7c8-rf9l4" Oct 05 06:50:28 crc kubenswrapper[4846]: E1005 06:50:28.002538 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-05 06:50:28.502525285 +0000 UTC m=+150.743378060 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rf9l4" (UID: "78d3e7b1-7258-4508-ad38-373f21bb5d96") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 06:50:28 crc kubenswrapper[4846]: I1005 06:50:28.003076 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-s9f8z"] Oct 05 06:50:28 crc kubenswrapper[4846]: W1005 06:50:28.056019 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod39351673_4e84_4ce0_b669_e2ca9072a443.slice/crio-a2dc26496c2aa7176cf68db65f7d800f161d0cf870355071be2166fe1bbd900a WatchSource:0}: Error finding container a2dc26496c2aa7176cf68db65f7d800f161d0cf870355071be2166fe1bbd900a: Status 404 returned error can't find the container with id a2dc26496c2aa7176cf68db65f7d800f161d0cf870355071be2166fe1bbd900a Oct 05 06:50:28 crc kubenswrapper[4846]: I1005 06:50:28.105675 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 05 06:50:28 crc kubenswrapper[4846]: E1005 06:50:28.105867 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-05 06:50:28.605840436 +0000 UTC m=+150.846693211 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 06:50:28 crc kubenswrapper[4846]: I1005 06:50:28.106011 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rf9l4\" (UID: \"78d3e7b1-7258-4508-ad38-373f21bb5d96\") " pod="openshift-image-registry/image-registry-697d97f7c8-rf9l4" Oct 05 06:50:28 crc kubenswrapper[4846]: E1005 06:50:28.106346 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-05 06:50:28.606336639 +0000 UTC m=+150.847189414 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rf9l4" (UID: "78d3e7b1-7258-4508-ad38-373f21bb5d96") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 06:50:28 crc kubenswrapper[4846]: I1005 06:50:28.207635 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 05 06:50:28 crc kubenswrapper[4846]: E1005 06:50:28.207753 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-05 06:50:28.70773569 +0000 UTC m=+150.948588465 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 06:50:28 crc kubenswrapper[4846]: I1005 06:50:28.208045 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rf9l4\" (UID: \"78d3e7b1-7258-4508-ad38-373f21bb5d96\") " pod="openshift-image-registry/image-registry-697d97f7c8-rf9l4" Oct 05 06:50:28 crc kubenswrapper[4846]: E1005 06:50:28.208382 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-05 06:50:28.708374837 +0000 UTC m=+150.949227612 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rf9l4" (UID: "78d3e7b1-7258-4508-ad38-373f21bb5d96") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 06:50:28 crc kubenswrapper[4846]: I1005 06:50:28.309100 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 05 06:50:28 crc kubenswrapper[4846]: E1005 06:50:28.309382 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-05 06:50:28.809365977 +0000 UTC m=+151.050218742 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 06:50:28 crc kubenswrapper[4846]: I1005 06:50:28.410785 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rf9l4\" (UID: \"78d3e7b1-7258-4508-ad38-373f21bb5d96\") " pod="openshift-image-registry/image-registry-697d97f7c8-rf9l4" Oct 05 06:50:28 crc kubenswrapper[4846]: E1005 06:50:28.411738 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-05 06:50:28.911720763 +0000 UTC m=+151.152573538 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rf9l4" (UID: "78d3e7b1-7258-4508-ad38-373f21bb5d96") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 06:50:28 crc kubenswrapper[4846]: I1005 06:50:28.447971 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-brf8j"] Oct 05 06:50:28 crc kubenswrapper[4846]: W1005 06:50:28.467264 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8bdbd298_f264_43ea_bb5e_625b1e5ab562.slice/crio-4f8b6e87a091660159594171adf0f5860837cfc2cade653d3d35e120ea202d44 WatchSource:0}: Error finding container 4f8b6e87a091660159594171adf0f5860837cfc2cade653d3d35e120ea202d44: Status 404 returned error can't find the container with id 4f8b6e87a091660159594171adf0f5860837cfc2cade653d3d35e120ea202d44 Oct 05 06:50:28 crc kubenswrapper[4846]: I1005 06:50:28.476375 4846 generic.go:334] "Generic (PLEG): container finished" podID="39351673-4e84-4ce0-b669-e2ca9072a443" containerID="af2013d70507bd5ee301dc25d61ac8f1069d697bc61e97f665e7d47ffdac0d7f" exitCode=0 Oct 05 06:50:28 crc kubenswrapper[4846]: I1005 06:50:28.476751 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-s9f8z" event={"ID":"39351673-4e84-4ce0-b669-e2ca9072a443","Type":"ContainerDied","Data":"af2013d70507bd5ee301dc25d61ac8f1069d697bc61e97f665e7d47ffdac0d7f"} Oct 05 06:50:28 crc kubenswrapper[4846]: I1005 06:50:28.476808 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-s9f8z" event={"ID":"39351673-4e84-4ce0-b669-e2ca9072a443","Type":"ContainerStarted","Data":"a2dc26496c2aa7176cf68db65f7d800f161d0cf870355071be2166fe1bbd900a"} Oct 05 06:50:28 crc kubenswrapper[4846]: I1005 06:50:28.481567 4846 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 05 06:50:28 crc kubenswrapper[4846]: I1005 06:50:28.506986 4846 generic.go:334] "Generic (PLEG): container finished" podID="da62c2a4-387f-4d5a-a9f2-04bd1e110d20" containerID="33ceecc33b378346cadfa69cd4c623a07f6709f95d9e297d39cf73e1f5800c82" exitCode=0 Oct 05 06:50:28 crc kubenswrapper[4846]: I1005 06:50:28.513733 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-76jjh"] Oct 05 06:50:28 crc kubenswrapper[4846]: I1005 06:50:28.513779 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mzc7c" event={"ID":"da62c2a4-387f-4d5a-a9f2-04bd1e110d20","Type":"ContainerDied","Data":"33ceecc33b378346cadfa69cd4c623a07f6709f95d9e297d39cf73e1f5800c82"} Oct 05 06:50:28 crc kubenswrapper[4846]: I1005 06:50:28.513811 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mzc7c" event={"ID":"da62c2a4-387f-4d5a-a9f2-04bd1e110d20","Type":"ContainerStarted","Data":"65593e4de55ad1d75bd3fc517aae979c3bb9241f569305d666ed25f257fa6902"} Oct 05 06:50:28 crc kubenswrapper[4846]: I1005 06:50:28.514571 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 05 06:50:28 crc kubenswrapper[4846]: E1005 06:50:28.514833 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-05 06:50:29.014723926 +0000 UTC m=+151.255576701 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 06:50:28 crc kubenswrapper[4846]: I1005 06:50:28.515075 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rf9l4\" (UID: \"78d3e7b1-7258-4508-ad38-373f21bb5d96\") " pod="openshift-image-registry/image-registry-697d97f7c8-rf9l4" Oct 05 06:50:28 crc kubenswrapper[4846]: E1005 06:50:28.515458 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-05 06:50:29.015443095 +0000 UTC m=+151.256295870 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rf9l4" (UID: "78d3e7b1-7258-4508-ad38-373f21bb5d96") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 06:50:28 crc kubenswrapper[4846]: I1005 06:50:28.516038 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-dsqhc" event={"ID":"5fed566e-ecb6-4801-84b3-ad1143a03f08","Type":"ContainerStarted","Data":"1eb8afc53fbb4a17aa584da10a3dd9aca4b18f667050c22e519860b1a229285c"} Oct 05 06:50:28 crc kubenswrapper[4846]: W1005 06:50:28.525395 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod73ca18b5_0a68_4990_9e7c_1685f87c8dcb.slice/crio-a88ea6505c818bdd1a13a8907c540d83ad51db690c5dd647ba5641f8fabc33ef WatchSource:0}: Error finding container a88ea6505c818bdd1a13a8907c540d83ad51db690c5dd647ba5641f8fabc33ef: Status 404 returned error can't find the container with id a88ea6505c818bdd1a13a8907c540d83ad51db690c5dd647ba5641f8fabc33ef Oct 05 06:50:28 crc kubenswrapper[4846]: I1005 06:50:28.616015 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 05 06:50:28 crc kubenswrapper[4846]: E1005 06:50:28.616191 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-05 06:50:29.116150328 +0000 UTC m=+151.357003103 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 06:50:28 crc kubenswrapper[4846]: I1005 06:50:28.617848 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rf9l4\" (UID: \"78d3e7b1-7258-4508-ad38-373f21bb5d96\") " pod="openshift-image-registry/image-registry-697d97f7c8-rf9l4" Oct 05 06:50:28 crc kubenswrapper[4846]: E1005 06:50:28.618138 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-05 06:50:29.11812716 +0000 UTC m=+151.358979935 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rf9l4" (UID: "78d3e7b1-7258-4508-ad38-373f21bb5d96") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 05 06:50:28 crc kubenswrapper[4846]: I1005 06:50:28.617519 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-dsqhc" podStartSLOduration=11.617507634 podStartE2EDuration="11.617507634s" podCreationTimestamp="2025-10-05 06:50:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 06:50:28.61661198 +0000 UTC m=+150.857464755" watchObservedRunningTime="2025-10-05 06:50:28.617507634 +0000 UTC m=+150.858360409" Oct 05 06:50:28 crc kubenswrapper[4846]: I1005 06:50:28.685396 4846 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2025-10-05T06:50:27.76681944Z","Handler":null,"Name":""} Oct 05 06:50:28 crc kubenswrapper[4846]: I1005 06:50:28.713057 4846 patch_prober.go:28] interesting pod/router-default-5444994796-58hwl container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 05 06:50:28 crc kubenswrapper[4846]: [-]has-synced failed: reason withheld Oct 05 06:50:28 crc kubenswrapper[4846]: [+]process-running ok Oct 05 06:50:28 crc kubenswrapper[4846]: healthz check failed Oct 05 06:50:28 crc kubenswrapper[4846]: I1005 06:50:28.713113 4846 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-58hwl" podUID="2ecfaba9-9034-4ac7-bd51-e3f64b41d825" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 05 06:50:28 crc kubenswrapper[4846]: I1005 06:50:28.713241 4846 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Oct 05 06:50:28 crc kubenswrapper[4846]: I1005 06:50:28.713304 4846 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Oct 05 06:50:28 crc kubenswrapper[4846]: I1005 06:50:28.721627 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 05 06:50:28 crc kubenswrapper[4846]: I1005 06:50:28.734833 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Oct 05 06:50:28 crc kubenswrapper[4846]: I1005 06:50:28.822901 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rf9l4\" (UID: \"78d3e7b1-7258-4508-ad38-373f21bb5d96\") " pod="openshift-image-registry/image-registry-697d97f7c8-rf9l4" Oct 05 06:50:28 crc kubenswrapper[4846]: I1005 06:50:28.829774 4846 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 05 06:50:28 crc kubenswrapper[4846]: I1005 06:50:28.829810 4846 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rf9l4\" (UID: \"78d3e7b1-7258-4508-ad38-373f21bb5d96\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-rf9l4" Oct 05 06:50:28 crc kubenswrapper[4846]: I1005 06:50:28.855297 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rf9l4\" (UID: \"78d3e7b1-7258-4508-ad38-373f21bb5d96\") " pod="openshift-image-registry/image-registry-697d97f7c8-rf9l4" Oct 05 06:50:28 crc kubenswrapper[4846]: I1005 06:50:28.866750 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29327445-dn5mf" Oct 05 06:50:28 crc kubenswrapper[4846]: I1005 06:50:28.925385 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/41aa056a-34d5-463b-8352-c71e0f0db1a4-config-volume\") pod \"41aa056a-34d5-463b-8352-c71e0f0db1a4\" (UID: \"41aa056a-34d5-463b-8352-c71e0f0db1a4\") " Oct 05 06:50:28 crc kubenswrapper[4846]: I1005 06:50:28.925517 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/41aa056a-34d5-463b-8352-c71e0f0db1a4-secret-volume\") pod \"41aa056a-34d5-463b-8352-c71e0f0db1a4\" (UID: \"41aa056a-34d5-463b-8352-c71e0f0db1a4\") " Oct 05 06:50:28 crc kubenswrapper[4846]: I1005 06:50:28.925583 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zwxrp\" (UniqueName: \"kubernetes.io/projected/41aa056a-34d5-463b-8352-c71e0f0db1a4-kube-api-access-zwxrp\") pod \"41aa056a-34d5-463b-8352-c71e0f0db1a4\" (UID: \"41aa056a-34d5-463b-8352-c71e0f0db1a4\") " Oct 05 06:50:28 crc kubenswrapper[4846]: I1005 06:50:28.928119 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/41aa056a-34d5-463b-8352-c71e0f0db1a4-config-volume" (OuterVolumeSpecName: "config-volume") pod "41aa056a-34d5-463b-8352-c71e0f0db1a4" (UID: "41aa056a-34d5-463b-8352-c71e0f0db1a4"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 06:50:28 crc kubenswrapper[4846]: I1005 06:50:28.933953 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/41aa056a-34d5-463b-8352-c71e0f0db1a4-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "41aa056a-34d5-463b-8352-c71e0f0db1a4" (UID: "41aa056a-34d5-463b-8352-c71e0f0db1a4"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 06:50:28 crc kubenswrapper[4846]: I1005 06:50:28.934037 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/41aa056a-34d5-463b-8352-c71e0f0db1a4-kube-api-access-zwxrp" (OuterVolumeSpecName: "kube-api-access-zwxrp") pod "41aa056a-34d5-463b-8352-c71e0f0db1a4" (UID: "41aa056a-34d5-463b-8352-c71e0f0db1a4"). InnerVolumeSpecName "kube-api-access-zwxrp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 06:50:28 crc kubenswrapper[4846]: I1005 06:50:28.956527 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-rf9l4" Oct 05 06:50:29 crc kubenswrapper[4846]: I1005 06:50:29.027009 4846 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/41aa056a-34d5-463b-8352-c71e0f0db1a4-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 05 06:50:29 crc kubenswrapper[4846]: I1005 06:50:29.027051 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zwxrp\" (UniqueName: \"kubernetes.io/projected/41aa056a-34d5-463b-8352-c71e0f0db1a4-kube-api-access-zwxrp\") on node \"crc\" DevicePath \"\"" Oct 05 06:50:29 crc kubenswrapper[4846]: I1005 06:50:29.027061 4846 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/41aa056a-34d5-463b-8352-c71e0f0db1a4-config-volume\") on node \"crc\" DevicePath \"\"" Oct 05 06:50:29 crc kubenswrapper[4846]: I1005 06:50:29.172694 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-rf9l4"] Oct 05 06:50:29 crc kubenswrapper[4846]: W1005 06:50:29.185308 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod78d3e7b1_7258_4508_ad38_373f21bb5d96.slice/crio-678cdf83030f0336db44da72ff49da71992545f53d49850a215977954586614c WatchSource:0}: Error finding container 678cdf83030f0336db44da72ff49da71992545f53d49850a215977954586614c: Status 404 returned error can't find the container with id 678cdf83030f0336db44da72ff49da71992545f53d49850a215977954586614c Oct 05 06:50:29 crc kubenswrapper[4846]: I1005 06:50:29.281201 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-dczxn"] Oct 05 06:50:29 crc kubenswrapper[4846]: E1005 06:50:29.281422 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="41aa056a-34d5-463b-8352-c71e0f0db1a4" containerName="collect-profiles" Oct 05 06:50:29 crc kubenswrapper[4846]: I1005 06:50:29.281443 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="41aa056a-34d5-463b-8352-c71e0f0db1a4" containerName="collect-profiles" Oct 05 06:50:29 crc kubenswrapper[4846]: I1005 06:50:29.281556 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="41aa056a-34d5-463b-8352-c71e0f0db1a4" containerName="collect-profiles" Oct 05 06:50:29 crc kubenswrapper[4846]: I1005 06:50:29.282242 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dczxn" Oct 05 06:50:29 crc kubenswrapper[4846]: I1005 06:50:29.286625 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Oct 05 06:50:29 crc kubenswrapper[4846]: I1005 06:50:29.297103 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-dczxn"] Oct 05 06:50:29 crc kubenswrapper[4846]: I1005 06:50:29.332403 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-brxrf\" (UniqueName: \"kubernetes.io/projected/b6446bcc-5edd-46bd-a2dc-e8b4773f7248-kube-api-access-brxrf\") pod \"redhat-marketplace-dczxn\" (UID: \"b6446bcc-5edd-46bd-a2dc-e8b4773f7248\") " pod="openshift-marketplace/redhat-marketplace-dczxn" Oct 05 06:50:29 crc kubenswrapper[4846]: I1005 06:50:29.332487 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b6446bcc-5edd-46bd-a2dc-e8b4773f7248-utilities\") pod \"redhat-marketplace-dczxn\" (UID: \"b6446bcc-5edd-46bd-a2dc-e8b4773f7248\") " pod="openshift-marketplace/redhat-marketplace-dczxn" Oct 05 06:50:29 crc kubenswrapper[4846]: I1005 06:50:29.332562 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b6446bcc-5edd-46bd-a2dc-e8b4773f7248-catalog-content\") pod \"redhat-marketplace-dczxn\" (UID: \"b6446bcc-5edd-46bd-a2dc-e8b4773f7248\") " pod="openshift-marketplace/redhat-marketplace-dczxn" Oct 05 06:50:29 crc kubenswrapper[4846]: I1005 06:50:29.434121 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b6446bcc-5edd-46bd-a2dc-e8b4773f7248-utilities\") pod \"redhat-marketplace-dczxn\" (UID: \"b6446bcc-5edd-46bd-a2dc-e8b4773f7248\") " pod="openshift-marketplace/redhat-marketplace-dczxn" Oct 05 06:50:29 crc kubenswrapper[4846]: I1005 06:50:29.434281 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b6446bcc-5edd-46bd-a2dc-e8b4773f7248-catalog-content\") pod \"redhat-marketplace-dczxn\" (UID: \"b6446bcc-5edd-46bd-a2dc-e8b4773f7248\") " pod="openshift-marketplace/redhat-marketplace-dczxn" Oct 05 06:50:29 crc kubenswrapper[4846]: I1005 06:50:29.434326 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-brxrf\" (UniqueName: \"kubernetes.io/projected/b6446bcc-5edd-46bd-a2dc-e8b4773f7248-kube-api-access-brxrf\") pod \"redhat-marketplace-dczxn\" (UID: \"b6446bcc-5edd-46bd-a2dc-e8b4773f7248\") " pod="openshift-marketplace/redhat-marketplace-dczxn" Oct 05 06:50:29 crc kubenswrapper[4846]: I1005 06:50:29.434769 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b6446bcc-5edd-46bd-a2dc-e8b4773f7248-utilities\") pod \"redhat-marketplace-dczxn\" (UID: \"b6446bcc-5edd-46bd-a2dc-e8b4773f7248\") " pod="openshift-marketplace/redhat-marketplace-dczxn" Oct 05 06:50:29 crc kubenswrapper[4846]: I1005 06:50:29.435015 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b6446bcc-5edd-46bd-a2dc-e8b4773f7248-catalog-content\") pod \"redhat-marketplace-dczxn\" (UID: \"b6446bcc-5edd-46bd-a2dc-e8b4773f7248\") " pod="openshift-marketplace/redhat-marketplace-dczxn" Oct 05 06:50:29 crc kubenswrapper[4846]: I1005 06:50:29.454272 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-brxrf\" (UniqueName: \"kubernetes.io/projected/b6446bcc-5edd-46bd-a2dc-e8b4773f7248-kube-api-access-brxrf\") pod \"redhat-marketplace-dczxn\" (UID: \"b6446bcc-5edd-46bd-a2dc-e8b4773f7248\") " pod="openshift-marketplace/redhat-marketplace-dczxn" Oct 05 06:50:29 crc kubenswrapper[4846]: I1005 06:50:29.543313 4846 generic.go:334] "Generic (PLEG): container finished" podID="8bdbd298-f264-43ea-bb5e-625b1e5ab562" containerID="c8501678cb72712ce3f8d7c9c7c848aa787d2ddd138114782292e24dd8fde2d1" exitCode=0 Oct 05 06:50:29 crc kubenswrapper[4846]: I1005 06:50:29.544097 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-brf8j" event={"ID":"8bdbd298-f264-43ea-bb5e-625b1e5ab562","Type":"ContainerDied","Data":"c8501678cb72712ce3f8d7c9c7c848aa787d2ddd138114782292e24dd8fde2d1"} Oct 05 06:50:29 crc kubenswrapper[4846]: I1005 06:50:29.544134 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-brf8j" event={"ID":"8bdbd298-f264-43ea-bb5e-625b1e5ab562","Type":"ContainerStarted","Data":"4f8b6e87a091660159594171adf0f5860837cfc2cade653d3d35e120ea202d44"} Oct 05 06:50:29 crc kubenswrapper[4846]: I1005 06:50:29.547689 4846 generic.go:334] "Generic (PLEG): container finished" podID="73ca18b5-0a68-4990-9e7c-1685f87c8dcb" containerID="5743c92de46b094ed9889d55ed47b2a35abcb849affca8436f485549017a93eb" exitCode=0 Oct 05 06:50:29 crc kubenswrapper[4846]: I1005 06:50:29.547823 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-76jjh" event={"ID":"73ca18b5-0a68-4990-9e7c-1685f87c8dcb","Type":"ContainerDied","Data":"5743c92de46b094ed9889d55ed47b2a35abcb849affca8436f485549017a93eb"} Oct 05 06:50:29 crc kubenswrapper[4846]: I1005 06:50:29.547863 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-76jjh" event={"ID":"73ca18b5-0a68-4990-9e7c-1685f87c8dcb","Type":"ContainerStarted","Data":"a88ea6505c818bdd1a13a8907c540d83ad51db690c5dd647ba5641f8fabc33ef"} Oct 05 06:50:29 crc kubenswrapper[4846]: I1005 06:50:29.558686 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29327445-dn5mf" event={"ID":"41aa056a-34d5-463b-8352-c71e0f0db1a4","Type":"ContainerDied","Data":"e2a9a7d69bb7c87fcf725e311147d7f32065f9ed22017993686f8049925fc55d"} Oct 05 06:50:29 crc kubenswrapper[4846]: I1005 06:50:29.558726 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e2a9a7d69bb7c87fcf725e311147d7f32065f9ed22017993686f8049925fc55d" Oct 05 06:50:29 crc kubenswrapper[4846]: I1005 06:50:29.558796 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29327445-dn5mf" Oct 05 06:50:29 crc kubenswrapper[4846]: I1005 06:50:29.611017 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dczxn" Oct 05 06:50:29 crc kubenswrapper[4846]: I1005 06:50:29.621400 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-rf9l4" event={"ID":"78d3e7b1-7258-4508-ad38-373f21bb5d96","Type":"ContainerStarted","Data":"3edcc13d0073ef87f5bc93cf466592a38ca27007b40995acebf6d1c84edde195"} Oct 05 06:50:29 crc kubenswrapper[4846]: I1005 06:50:29.628791 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-rf9l4" Oct 05 06:50:29 crc kubenswrapper[4846]: I1005 06:50:29.628809 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-rf9l4" event={"ID":"78d3e7b1-7258-4508-ad38-373f21bb5d96","Type":"ContainerStarted","Data":"678cdf83030f0336db44da72ff49da71992545f53d49850a215977954586614c"} Oct 05 06:50:29 crc kubenswrapper[4846]: I1005 06:50:29.675410 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-rf9l4" podStartSLOduration=125.675389024 podStartE2EDuration="2m5.675389024s" podCreationTimestamp="2025-10-05 06:48:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 06:50:29.668085432 +0000 UTC m=+151.908938227" watchObservedRunningTime="2025-10-05 06:50:29.675389024 +0000 UTC m=+151.916241789" Oct 05 06:50:29 crc kubenswrapper[4846]: I1005 06:50:29.690340 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-z858q"] Oct 05 06:50:29 crc kubenswrapper[4846]: I1005 06:50:29.695011 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-z858q" Oct 05 06:50:29 crc kubenswrapper[4846]: I1005 06:50:29.708438 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-z858q"] Oct 05 06:50:29 crc kubenswrapper[4846]: I1005 06:50:29.720509 4846 patch_prober.go:28] interesting pod/router-default-5444994796-58hwl container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 05 06:50:29 crc kubenswrapper[4846]: [-]has-synced failed: reason withheld Oct 05 06:50:29 crc kubenswrapper[4846]: [+]process-running ok Oct 05 06:50:29 crc kubenswrapper[4846]: healthz check failed Oct 05 06:50:29 crc kubenswrapper[4846]: I1005 06:50:29.720575 4846 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-58hwl" podUID="2ecfaba9-9034-4ac7-bd51-e3f64b41d825" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 05 06:50:29 crc kubenswrapper[4846]: I1005 06:50:29.739402 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3df540e4-b545-4c4a-88c8-b20c6005bd84-utilities\") pod \"redhat-marketplace-z858q\" (UID: \"3df540e4-b545-4c4a-88c8-b20c6005bd84\") " pod="openshift-marketplace/redhat-marketplace-z858q" Oct 05 06:50:29 crc kubenswrapper[4846]: I1005 06:50:29.739451 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3df540e4-b545-4c4a-88c8-b20c6005bd84-catalog-content\") pod \"redhat-marketplace-z858q\" (UID: \"3df540e4-b545-4c4a-88c8-b20c6005bd84\") " pod="openshift-marketplace/redhat-marketplace-z858q" Oct 05 06:50:29 crc kubenswrapper[4846]: I1005 06:50:29.739534 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fjsks\" (UniqueName: \"kubernetes.io/projected/3df540e4-b545-4c4a-88c8-b20c6005bd84-kube-api-access-fjsks\") pod \"redhat-marketplace-z858q\" (UID: \"3df540e4-b545-4c4a-88c8-b20c6005bd84\") " pod="openshift-marketplace/redhat-marketplace-z858q" Oct 05 06:50:29 crc kubenswrapper[4846]: I1005 06:50:29.779970 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-stjw8" Oct 05 06:50:29 crc kubenswrapper[4846]: I1005 06:50:29.786663 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-stjw8" Oct 05 06:50:29 crc kubenswrapper[4846]: I1005 06:50:29.798254 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Oct 05 06:50:29 crc kubenswrapper[4846]: I1005 06:50:29.798987 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 05 06:50:29 crc kubenswrapper[4846]: I1005 06:50:29.803395 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Oct 05 06:50:29 crc kubenswrapper[4846]: I1005 06:50:29.803530 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Oct 05 06:50:29 crc kubenswrapper[4846]: I1005 06:50:29.825615 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Oct 05 06:50:29 crc kubenswrapper[4846]: I1005 06:50:29.838301 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-9hcbc" Oct 05 06:50:29 crc kubenswrapper[4846]: I1005 06:50:29.838339 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-9hcbc" Oct 05 06:50:29 crc kubenswrapper[4846]: I1005 06:50:29.840207 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3df540e4-b545-4c4a-88c8-b20c6005bd84-utilities\") pod \"redhat-marketplace-z858q\" (UID: \"3df540e4-b545-4c4a-88c8-b20c6005bd84\") " pod="openshift-marketplace/redhat-marketplace-z858q" Oct 05 06:50:29 crc kubenswrapper[4846]: I1005 06:50:29.840244 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3df540e4-b545-4c4a-88c8-b20c6005bd84-catalog-content\") pod \"redhat-marketplace-z858q\" (UID: \"3df540e4-b545-4c4a-88c8-b20c6005bd84\") " pod="openshift-marketplace/redhat-marketplace-z858q" Oct 05 06:50:29 crc kubenswrapper[4846]: I1005 06:50:29.840283 4846 patch_prober.go:28] interesting pod/console-f9d7485db-9hcbc container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.9:8443/health\": dial tcp 10.217.0.9:8443: connect: connection refused" start-of-body= Oct 05 06:50:29 crc kubenswrapper[4846]: I1005 06:50:29.841022 4846 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-9hcbc" podUID="74f16025-fe11-4a23-8c1a-ff3da2a3d751" containerName="console" probeResult="failure" output="Get \"https://10.217.0.9:8443/health\": dial tcp 10.217.0.9:8443: connect: connection refused" Oct 05 06:50:29 crc kubenswrapper[4846]: I1005 06:50:29.840859 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3df540e4-b545-4c4a-88c8-b20c6005bd84-catalog-content\") pod \"redhat-marketplace-z858q\" (UID: \"3df540e4-b545-4c4a-88c8-b20c6005bd84\") " pod="openshift-marketplace/redhat-marketplace-z858q" Oct 05 06:50:29 crc kubenswrapper[4846]: I1005 06:50:29.840329 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d42d9769-b723-4790-b5b5-9cb03fe1cc04-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"d42d9769-b723-4790-b5b5-9cb03fe1cc04\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 05 06:50:29 crc kubenswrapper[4846]: I1005 06:50:29.841198 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fjsks\" (UniqueName: \"kubernetes.io/projected/3df540e4-b545-4c4a-88c8-b20c6005bd84-kube-api-access-fjsks\") pod \"redhat-marketplace-z858q\" (UID: \"3df540e4-b545-4c4a-88c8-b20c6005bd84\") " pod="openshift-marketplace/redhat-marketplace-z858q" Oct 05 06:50:29 crc kubenswrapper[4846]: I1005 06:50:29.841352 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/d42d9769-b723-4790-b5b5-9cb03fe1cc04-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"d42d9769-b723-4790-b5b5-9cb03fe1cc04\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 05 06:50:29 crc kubenswrapper[4846]: I1005 06:50:29.841682 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3df540e4-b545-4c4a-88c8-b20c6005bd84-utilities\") pod \"redhat-marketplace-z858q\" (UID: \"3df540e4-b545-4c4a-88c8-b20c6005bd84\") " pod="openshift-marketplace/redhat-marketplace-z858q" Oct 05 06:50:29 crc kubenswrapper[4846]: I1005 06:50:29.889525 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fjsks\" (UniqueName: \"kubernetes.io/projected/3df540e4-b545-4c4a-88c8-b20c6005bd84-kube-api-access-fjsks\") pod \"redhat-marketplace-z858q\" (UID: \"3df540e4-b545-4c4a-88c8-b20c6005bd84\") " pod="openshift-marketplace/redhat-marketplace-z858q" Oct 05 06:50:29 crc kubenswrapper[4846]: I1005 06:50:29.900345 4846 patch_prober.go:28] interesting pod/downloads-7954f5f757-c6cxs container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.12:8080/\": dial tcp 10.217.0.12:8080: connect: connection refused" start-of-body= Oct 05 06:50:29 crc kubenswrapper[4846]: I1005 06:50:29.900407 4846 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-c6cxs" podUID="475e1ca4-a6fa-4f77-8c89-0619b2e9681e" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.12:8080/\": dial tcp 10.217.0.12:8080: connect: connection refused" Oct 05 06:50:29 crc kubenswrapper[4846]: I1005 06:50:29.902369 4846 patch_prober.go:28] interesting pod/downloads-7954f5f757-c6cxs container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.12:8080/\": dial tcp 10.217.0.12:8080: connect: connection refused" start-of-body= Oct 05 06:50:29 crc kubenswrapper[4846]: I1005 06:50:29.902449 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-c6cxs" podUID="475e1ca4-a6fa-4f77-8c89-0619b2e9681e" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.12:8080/\": dial tcp 10.217.0.12:8080: connect: connection refused" Oct 05 06:50:29 crc kubenswrapper[4846]: I1005 06:50:29.944839 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/d42d9769-b723-4790-b5b5-9cb03fe1cc04-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"d42d9769-b723-4790-b5b5-9cb03fe1cc04\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 05 06:50:29 crc kubenswrapper[4846]: I1005 06:50:29.945002 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d42d9769-b723-4790-b5b5-9cb03fe1cc04-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"d42d9769-b723-4790-b5b5-9cb03fe1cc04\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 05 06:50:29 crc kubenswrapper[4846]: I1005 06:50:29.945005 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/d42d9769-b723-4790-b5b5-9cb03fe1cc04-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"d42d9769-b723-4790-b5b5-9cb03fe1cc04\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 05 06:50:29 crc kubenswrapper[4846]: I1005 06:50:29.981832 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d42d9769-b723-4790-b5b5-9cb03fe1cc04-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"d42d9769-b723-4790-b5b5-9cb03fe1cc04\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 05 06:50:29 crc kubenswrapper[4846]: I1005 06:50:29.992479 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-6bpvf" Oct 05 06:50:29 crc kubenswrapper[4846]: I1005 06:50:29.992925 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-6bpvf" Oct 05 06:50:30 crc kubenswrapper[4846]: I1005 06:50:30.002641 4846 patch_prober.go:28] interesting pod/apiserver-76f77b778f-6bpvf container/openshift-apiserver namespace/openshift-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[+]ping ok Oct 05 06:50:30 crc kubenswrapper[4846]: [+]log ok Oct 05 06:50:30 crc kubenswrapper[4846]: [+]etcd ok Oct 05 06:50:30 crc kubenswrapper[4846]: [+]poststarthook/start-apiserver-admission-initializer ok Oct 05 06:50:30 crc kubenswrapper[4846]: [+]poststarthook/generic-apiserver-start-informers ok Oct 05 06:50:30 crc kubenswrapper[4846]: [+]poststarthook/max-in-flight-filter ok Oct 05 06:50:30 crc kubenswrapper[4846]: [+]poststarthook/storage-object-count-tracker-hook ok Oct 05 06:50:30 crc kubenswrapper[4846]: [+]poststarthook/image.openshift.io-apiserver-caches ok Oct 05 06:50:30 crc kubenswrapper[4846]: [-]poststarthook/authorization.openshift.io-bootstrapclusterroles failed: reason withheld Oct 05 06:50:30 crc kubenswrapper[4846]: [+]poststarthook/authorization.openshift.io-ensurenodebootstrap-sa ok Oct 05 06:50:30 crc kubenswrapper[4846]: [+]poststarthook/project.openshift.io-projectcache ok Oct 05 06:50:30 crc kubenswrapper[4846]: [+]poststarthook/project.openshift.io-projectauthorizationcache ok Oct 05 06:50:30 crc kubenswrapper[4846]: [+]poststarthook/openshift.io-startinformers ok Oct 05 06:50:30 crc kubenswrapper[4846]: [+]poststarthook/openshift.io-restmapperupdater ok Oct 05 06:50:30 crc kubenswrapper[4846]: [+]poststarthook/quota.openshift.io-clusterquotamapping ok Oct 05 06:50:30 crc kubenswrapper[4846]: livez check failed Oct 05 06:50:30 crc kubenswrapper[4846]: I1005 06:50:30.002720 4846 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-apiserver/apiserver-76f77b778f-6bpvf" podUID="75d0c362-7aa3-4919-937d-8849f425466e" containerName="openshift-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 05 06:50:30 crc kubenswrapper[4846]: I1005 06:50:30.021885 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-z858q" Oct 05 06:50:30 crc kubenswrapper[4846]: I1005 06:50:30.087609 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-knslg"] Oct 05 06:50:30 crc kubenswrapper[4846]: I1005 06:50:30.089752 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-knslg" Oct 05 06:50:30 crc kubenswrapper[4846]: I1005 06:50:30.092232 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Oct 05 06:50:30 crc kubenswrapper[4846]: I1005 06:50:30.102507 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-knslg"] Oct 05 06:50:30 crc kubenswrapper[4846]: I1005 06:50:30.128133 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 05 06:50:30 crc kubenswrapper[4846]: I1005 06:50:30.153272 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0984e21c-5c51-4493-80b9-6e0fadf7537d-catalog-content\") pod \"redhat-operators-knslg\" (UID: \"0984e21c-5c51-4493-80b9-6e0fadf7537d\") " pod="openshift-marketplace/redhat-operators-knslg" Oct 05 06:50:30 crc kubenswrapper[4846]: I1005 06:50:30.153324 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0984e21c-5c51-4493-80b9-6e0fadf7537d-utilities\") pod \"redhat-operators-knslg\" (UID: \"0984e21c-5c51-4493-80b9-6e0fadf7537d\") " pod="openshift-marketplace/redhat-operators-knslg" Oct 05 06:50:30 crc kubenswrapper[4846]: I1005 06:50:30.153440 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fqsk2\" (UniqueName: \"kubernetes.io/projected/0984e21c-5c51-4493-80b9-6e0fadf7537d-kube-api-access-fqsk2\") pod \"redhat-operators-knslg\" (UID: \"0984e21c-5c51-4493-80b9-6e0fadf7537d\") " pod="openshift-marketplace/redhat-operators-knslg" Oct 05 06:50:30 crc kubenswrapper[4846]: I1005 06:50:30.198008 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-dczxn"] Oct 05 06:50:30 crc kubenswrapper[4846]: W1005 06:50:30.213461 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb6446bcc_5edd_46bd_a2dc_e8b4773f7248.slice/crio-f3837f8f1c0c83c2d5e4933cd8760a5061a16e7b36865ebf44f231a16a8d163e WatchSource:0}: Error finding container f3837f8f1c0c83c2d5e4933cd8760a5061a16e7b36865ebf44f231a16a8d163e: Status 404 returned error can't find the container with id f3837f8f1c0c83c2d5e4933cd8760a5061a16e7b36865ebf44f231a16a8d163e Oct 05 06:50:30 crc kubenswrapper[4846]: I1005 06:50:30.257730 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0984e21c-5c51-4493-80b9-6e0fadf7537d-catalog-content\") pod \"redhat-operators-knslg\" (UID: \"0984e21c-5c51-4493-80b9-6e0fadf7537d\") " pod="openshift-marketplace/redhat-operators-knslg" Oct 05 06:50:30 crc kubenswrapper[4846]: I1005 06:50:30.257810 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0984e21c-5c51-4493-80b9-6e0fadf7537d-utilities\") pod \"redhat-operators-knslg\" (UID: \"0984e21c-5c51-4493-80b9-6e0fadf7537d\") " pod="openshift-marketplace/redhat-operators-knslg" Oct 05 06:50:30 crc kubenswrapper[4846]: I1005 06:50:30.257925 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fqsk2\" (UniqueName: \"kubernetes.io/projected/0984e21c-5c51-4493-80b9-6e0fadf7537d-kube-api-access-fqsk2\") pod \"redhat-operators-knslg\" (UID: \"0984e21c-5c51-4493-80b9-6e0fadf7537d\") " pod="openshift-marketplace/redhat-operators-knslg" Oct 05 06:50:30 crc kubenswrapper[4846]: I1005 06:50:30.258481 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0984e21c-5c51-4493-80b9-6e0fadf7537d-catalog-content\") pod \"redhat-operators-knslg\" (UID: \"0984e21c-5c51-4493-80b9-6e0fadf7537d\") " pod="openshift-marketplace/redhat-operators-knslg" Oct 05 06:50:30 crc kubenswrapper[4846]: I1005 06:50:30.258695 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0984e21c-5c51-4493-80b9-6e0fadf7537d-utilities\") pod \"redhat-operators-knslg\" (UID: \"0984e21c-5c51-4493-80b9-6e0fadf7537d\") " pod="openshift-marketplace/redhat-operators-knslg" Oct 05 06:50:30 crc kubenswrapper[4846]: I1005 06:50:30.282408 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fqsk2\" (UniqueName: \"kubernetes.io/projected/0984e21c-5c51-4493-80b9-6e0fadf7537d-kube-api-access-fqsk2\") pod \"redhat-operators-knslg\" (UID: \"0984e21c-5c51-4493-80b9-6e0fadf7537d\") " pod="openshift-marketplace/redhat-operators-knslg" Oct 05 06:50:30 crc kubenswrapper[4846]: I1005 06:50:30.296831 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-w424l"] Oct 05 06:50:30 crc kubenswrapper[4846]: I1005 06:50:30.298715 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-w424l" Oct 05 06:50:30 crc kubenswrapper[4846]: I1005 06:50:30.303515 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-w424l"] Oct 05 06:50:30 crc kubenswrapper[4846]: I1005 06:50:30.321612 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-z858q"] Oct 05 06:50:30 crc kubenswrapper[4846]: I1005 06:50:30.362868 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/63ae883f-8fb0-4270-923b-1ecad63773b0-catalog-content\") pod \"redhat-operators-w424l\" (UID: \"63ae883f-8fb0-4270-923b-1ecad63773b0\") " pod="openshift-marketplace/redhat-operators-w424l" Oct 05 06:50:30 crc kubenswrapper[4846]: I1005 06:50:30.363347 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/63ae883f-8fb0-4270-923b-1ecad63773b0-utilities\") pod \"redhat-operators-w424l\" (UID: \"63ae883f-8fb0-4270-923b-1ecad63773b0\") " pod="openshift-marketplace/redhat-operators-w424l" Oct 05 06:50:30 crc kubenswrapper[4846]: I1005 06:50:30.363437 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n29f4\" (UniqueName: \"kubernetes.io/projected/63ae883f-8fb0-4270-923b-1ecad63773b0-kube-api-access-n29f4\") pod \"redhat-operators-w424l\" (UID: \"63ae883f-8fb0-4270-923b-1ecad63773b0\") " pod="openshift-marketplace/redhat-operators-w424l" Oct 05 06:50:30 crc kubenswrapper[4846]: I1005 06:50:30.419336 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-knslg" Oct 05 06:50:30 crc kubenswrapper[4846]: I1005 06:50:30.466824 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/63ae883f-8fb0-4270-923b-1ecad63773b0-catalog-content\") pod \"redhat-operators-w424l\" (UID: \"63ae883f-8fb0-4270-923b-1ecad63773b0\") " pod="openshift-marketplace/redhat-operators-w424l" Oct 05 06:50:30 crc kubenswrapper[4846]: I1005 06:50:30.466970 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/63ae883f-8fb0-4270-923b-1ecad63773b0-utilities\") pod \"redhat-operators-w424l\" (UID: \"63ae883f-8fb0-4270-923b-1ecad63773b0\") " pod="openshift-marketplace/redhat-operators-w424l" Oct 05 06:50:30 crc kubenswrapper[4846]: I1005 06:50:30.467037 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n29f4\" (UniqueName: \"kubernetes.io/projected/63ae883f-8fb0-4270-923b-1ecad63773b0-kube-api-access-n29f4\") pod \"redhat-operators-w424l\" (UID: \"63ae883f-8fb0-4270-923b-1ecad63773b0\") " pod="openshift-marketplace/redhat-operators-w424l" Oct 05 06:50:30 crc kubenswrapper[4846]: I1005 06:50:30.468440 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/63ae883f-8fb0-4270-923b-1ecad63773b0-catalog-content\") pod \"redhat-operators-w424l\" (UID: \"63ae883f-8fb0-4270-923b-1ecad63773b0\") " pod="openshift-marketplace/redhat-operators-w424l" Oct 05 06:50:30 crc kubenswrapper[4846]: I1005 06:50:30.468674 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/63ae883f-8fb0-4270-923b-1ecad63773b0-utilities\") pod \"redhat-operators-w424l\" (UID: \"63ae883f-8fb0-4270-923b-1ecad63773b0\") " pod="openshift-marketplace/redhat-operators-w424l" Oct 05 06:50:30 crc kubenswrapper[4846]: I1005 06:50:30.495974 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n29f4\" (UniqueName: \"kubernetes.io/projected/63ae883f-8fb0-4270-923b-1ecad63773b0-kube-api-access-n29f4\") pod \"redhat-operators-w424l\" (UID: \"63ae883f-8fb0-4270-923b-1ecad63773b0\") " pod="openshift-marketplace/redhat-operators-w424l" Oct 05 06:50:30 crc kubenswrapper[4846]: I1005 06:50:30.530257 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Oct 05 06:50:30 crc kubenswrapper[4846]: I1005 06:50:30.531148 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-t8f4h" Oct 05 06:50:30 crc kubenswrapper[4846]: I1005 06:50:30.576798 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Oct 05 06:50:30 crc kubenswrapper[4846]: I1005 06:50:30.596992 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 05 06:50:30 crc kubenswrapper[4846]: I1005 06:50:30.600804 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Oct 05 06:50:30 crc kubenswrapper[4846]: I1005 06:50:30.601115 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Oct 05 06:50:30 crc kubenswrapper[4846]: I1005 06:50:30.602474 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Oct 05 06:50:30 crc kubenswrapper[4846]: I1005 06:50:30.650753 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-z858q" event={"ID":"3df540e4-b545-4c4a-88c8-b20c6005bd84","Type":"ContainerStarted","Data":"d6179db76f42e86b59b5c683d0bed1e4bde99212e2fadd1629d7f680caf7c317"} Oct 05 06:50:30 crc kubenswrapper[4846]: I1005 06:50:30.654904 4846 generic.go:334] "Generic (PLEG): container finished" podID="b6446bcc-5edd-46bd-a2dc-e8b4773f7248" containerID="7c900c6109b9cf97634b9dae519a95243a4b08eb406782cac904939033b704ca" exitCode=0 Oct 05 06:50:30 crc kubenswrapper[4846]: I1005 06:50:30.656066 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dczxn" event={"ID":"b6446bcc-5edd-46bd-a2dc-e8b4773f7248","Type":"ContainerDied","Data":"7c900c6109b9cf97634b9dae519a95243a4b08eb406782cac904939033b704ca"} Oct 05 06:50:30 crc kubenswrapper[4846]: I1005 06:50:30.656088 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dczxn" event={"ID":"b6446bcc-5edd-46bd-a2dc-e8b4773f7248","Type":"ContainerStarted","Data":"f3837f8f1c0c83c2d5e4933cd8760a5061a16e7b36865ebf44f231a16a8d163e"} Oct 05 06:50:30 crc kubenswrapper[4846]: I1005 06:50:30.659223 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-w424l" Oct 05 06:50:30 crc kubenswrapper[4846]: I1005 06:50:30.673662 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/7ce1a1f8-e7e8-4f26-851c-1d04367a0083-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"7ce1a1f8-e7e8-4f26-851c-1d04367a0083\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 05 06:50:30 crc kubenswrapper[4846]: I1005 06:50:30.674116 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7ce1a1f8-e7e8-4f26-851c-1d04367a0083-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"7ce1a1f8-e7e8-4f26-851c-1d04367a0083\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 05 06:50:30 crc kubenswrapper[4846]: I1005 06:50:30.706306 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Oct 05 06:50:30 crc kubenswrapper[4846]: I1005 06:50:30.707805 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-58hwl" Oct 05 06:50:30 crc kubenswrapper[4846]: I1005 06:50:30.715906 4846 patch_prober.go:28] interesting pod/router-default-5444994796-58hwl container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 05 06:50:30 crc kubenswrapper[4846]: [-]has-synced failed: reason withheld Oct 05 06:50:30 crc kubenswrapper[4846]: [+]process-running ok Oct 05 06:50:30 crc kubenswrapper[4846]: healthz check failed Oct 05 06:50:30 crc kubenswrapper[4846]: I1005 06:50:30.716240 4846 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-58hwl" podUID="2ecfaba9-9034-4ac7-bd51-e3f64b41d825" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 05 06:50:30 crc kubenswrapper[4846]: I1005 06:50:30.738441 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-knslg"] Oct 05 06:50:30 crc kubenswrapper[4846]: I1005 06:50:30.745806 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-2rc5n" Oct 05 06:50:30 crc kubenswrapper[4846]: I1005 06:50:30.776372 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/7ce1a1f8-e7e8-4f26-851c-1d04367a0083-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"7ce1a1f8-e7e8-4f26-851c-1d04367a0083\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 05 06:50:30 crc kubenswrapper[4846]: I1005 06:50:30.776454 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7ce1a1f8-e7e8-4f26-851c-1d04367a0083-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"7ce1a1f8-e7e8-4f26-851c-1d04367a0083\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 05 06:50:30 crc kubenswrapper[4846]: I1005 06:50:30.779782 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/7ce1a1f8-e7e8-4f26-851c-1d04367a0083-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"7ce1a1f8-e7e8-4f26-851c-1d04367a0083\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 05 06:50:30 crc kubenswrapper[4846]: I1005 06:50:30.799031 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7ce1a1f8-e7e8-4f26-851c-1d04367a0083-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"7ce1a1f8-e7e8-4f26-851c-1d04367a0083\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 05 06:50:30 crc kubenswrapper[4846]: I1005 06:50:30.939188 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 05 06:50:31 crc kubenswrapper[4846]: I1005 06:50:31.075098 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-w424l"] Oct 05 06:50:31 crc kubenswrapper[4846]: W1005 06:50:31.135732 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod63ae883f_8fb0_4270_923b_1ecad63773b0.slice/crio-fd87f185be80388544f1dd5aaa8d22762a1e9ec99ffb58f5726dc864fd935592 WatchSource:0}: Error finding container fd87f185be80388544f1dd5aaa8d22762a1e9ec99ffb58f5726dc864fd935592: Status 404 returned error can't find the container with id fd87f185be80388544f1dd5aaa8d22762a1e9ec99ffb58f5726dc864fd935592 Oct 05 06:50:31 crc kubenswrapper[4846]: I1005 06:50:31.222468 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Oct 05 06:50:31 crc kubenswrapper[4846]: W1005 06:50:31.226189 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod7ce1a1f8_e7e8_4f26_851c_1d04367a0083.slice/crio-d06fa06f0d5ec92c47c4899ebe5902a625530c8b1832eaf3414f7c3d9a82469a WatchSource:0}: Error finding container d06fa06f0d5ec92c47c4899ebe5902a625530c8b1832eaf3414f7c3d9a82469a: Status 404 returned error can't find the container with id d06fa06f0d5ec92c47c4899ebe5902a625530c8b1832eaf3414f7c3d9a82469a Oct 05 06:50:31 crc kubenswrapper[4846]: I1005 06:50:31.669846 4846 generic.go:334] "Generic (PLEG): container finished" podID="0984e21c-5c51-4493-80b9-6e0fadf7537d" containerID="0ff82a1bd96990b795985d4afdcb71edac3ab2c578e28afa371c340f30686a3d" exitCode=0 Oct 05 06:50:31 crc kubenswrapper[4846]: I1005 06:50:31.669921 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-knslg" event={"ID":"0984e21c-5c51-4493-80b9-6e0fadf7537d","Type":"ContainerDied","Data":"0ff82a1bd96990b795985d4afdcb71edac3ab2c578e28afa371c340f30686a3d"} Oct 05 06:50:31 crc kubenswrapper[4846]: I1005 06:50:31.670861 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-knslg" event={"ID":"0984e21c-5c51-4493-80b9-6e0fadf7537d","Type":"ContainerStarted","Data":"cc6e538b18c6545419212022250380cd4afeb8f2212a05ada67319a9bc928a73"} Oct 05 06:50:31 crc kubenswrapper[4846]: I1005 06:50:31.678085 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"d42d9769-b723-4790-b5b5-9cb03fe1cc04","Type":"ContainerStarted","Data":"850abc4bc09c715c281055924d63c4ada488a19f78cb64b1857e898f956ae181"} Oct 05 06:50:31 crc kubenswrapper[4846]: I1005 06:50:31.678126 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"d42d9769-b723-4790-b5b5-9cb03fe1cc04","Type":"ContainerStarted","Data":"a7b9dce148b59a413d51cc2996f2db61d3596a0f701d1e76e442f6e1113cf6a3"} Oct 05 06:50:31 crc kubenswrapper[4846]: I1005 06:50:31.682552 4846 generic.go:334] "Generic (PLEG): container finished" podID="3df540e4-b545-4c4a-88c8-b20c6005bd84" containerID="6e59d14c271b5f97931e8fbee43e336919fbb247169aec39b18ee044443d96a7" exitCode=0 Oct 05 06:50:31 crc kubenswrapper[4846]: I1005 06:50:31.682945 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-z858q" event={"ID":"3df540e4-b545-4c4a-88c8-b20c6005bd84","Type":"ContainerDied","Data":"6e59d14c271b5f97931e8fbee43e336919fbb247169aec39b18ee044443d96a7"} Oct 05 06:50:31 crc kubenswrapper[4846]: I1005 06:50:31.685594 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"7ce1a1f8-e7e8-4f26-851c-1d04367a0083","Type":"ContainerStarted","Data":"d06fa06f0d5ec92c47c4899ebe5902a625530c8b1832eaf3414f7c3d9a82469a"} Oct 05 06:50:31 crc kubenswrapper[4846]: I1005 06:50:31.695402 4846 generic.go:334] "Generic (PLEG): container finished" podID="63ae883f-8fb0-4270-923b-1ecad63773b0" containerID="76c946a77e6ddf901939bd92827d436efda3d52f75028018db04751e8e7238e3" exitCode=0 Oct 05 06:50:31 crc kubenswrapper[4846]: I1005 06:50:31.695473 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-w424l" event={"ID":"63ae883f-8fb0-4270-923b-1ecad63773b0","Type":"ContainerDied","Data":"76c946a77e6ddf901939bd92827d436efda3d52f75028018db04751e8e7238e3"} Oct 05 06:50:31 crc kubenswrapper[4846]: I1005 06:50:31.695729 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-w424l" event={"ID":"63ae883f-8fb0-4270-923b-1ecad63773b0","Type":"ContainerStarted","Data":"fd87f185be80388544f1dd5aaa8d22762a1e9ec99ffb58f5726dc864fd935592"} Oct 05 06:50:31 crc kubenswrapper[4846]: I1005 06:50:31.719594 4846 patch_prober.go:28] interesting pod/router-default-5444994796-58hwl container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 05 06:50:31 crc kubenswrapper[4846]: [-]has-synced failed: reason withheld Oct 05 06:50:31 crc kubenswrapper[4846]: [+]process-running ok Oct 05 06:50:31 crc kubenswrapper[4846]: healthz check failed Oct 05 06:50:31 crc kubenswrapper[4846]: I1005 06:50:31.719660 4846 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-58hwl" podUID="2ecfaba9-9034-4ac7-bd51-e3f64b41d825" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 05 06:50:31 crc kubenswrapper[4846]: I1005 06:50:31.739446 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/revision-pruner-9-crc" podStartSLOduration=2.739406859 podStartE2EDuration="2.739406859s" podCreationTimestamp="2025-10-05 06:50:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 06:50:31.728863702 +0000 UTC m=+153.969716477" watchObservedRunningTime="2025-10-05 06:50:31.739406859 +0000 UTC m=+153.980259764" Oct 05 06:50:32 crc kubenswrapper[4846]: I1005 06:50:32.708455 4846 generic.go:334] "Generic (PLEG): container finished" podID="d42d9769-b723-4790-b5b5-9cb03fe1cc04" containerID="850abc4bc09c715c281055924d63c4ada488a19f78cb64b1857e898f956ae181" exitCode=0 Oct 05 06:50:32 crc kubenswrapper[4846]: I1005 06:50:32.708534 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"d42d9769-b723-4790-b5b5-9cb03fe1cc04","Type":"ContainerDied","Data":"850abc4bc09c715c281055924d63c4ada488a19f78cb64b1857e898f956ae181"} Oct 05 06:50:32 crc kubenswrapper[4846]: I1005 06:50:32.709629 4846 patch_prober.go:28] interesting pod/router-default-5444994796-58hwl container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 05 06:50:32 crc kubenswrapper[4846]: [-]has-synced failed: reason withheld Oct 05 06:50:32 crc kubenswrapper[4846]: [+]process-running ok Oct 05 06:50:32 crc kubenswrapper[4846]: healthz check failed Oct 05 06:50:32 crc kubenswrapper[4846]: I1005 06:50:32.709689 4846 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-58hwl" podUID="2ecfaba9-9034-4ac7-bd51-e3f64b41d825" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 05 06:50:32 crc kubenswrapper[4846]: I1005 06:50:32.715765 4846 generic.go:334] "Generic (PLEG): container finished" podID="7ce1a1f8-e7e8-4f26-851c-1d04367a0083" containerID="f5bbf04651b75a014589741a77637c38edc1d381ae6052961826aa6bb00c94aa" exitCode=0 Oct 05 06:50:32 crc kubenswrapper[4846]: I1005 06:50:32.715826 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"7ce1a1f8-e7e8-4f26-851c-1d04367a0083","Type":"ContainerDied","Data":"f5bbf04651b75a014589741a77637c38edc1d381ae6052961826aa6bb00c94aa"} Oct 05 06:50:33 crc kubenswrapper[4846]: I1005 06:50:33.710365 4846 patch_prober.go:28] interesting pod/router-default-5444994796-58hwl container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 05 06:50:33 crc kubenswrapper[4846]: [-]has-synced failed: reason withheld Oct 05 06:50:33 crc kubenswrapper[4846]: [+]process-running ok Oct 05 06:50:33 crc kubenswrapper[4846]: healthz check failed Oct 05 06:50:33 crc kubenswrapper[4846]: I1005 06:50:33.710440 4846 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-58hwl" podUID="2ecfaba9-9034-4ac7-bd51-e3f64b41d825" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 05 06:50:34 crc kubenswrapper[4846]: I1005 06:50:34.710467 4846 patch_prober.go:28] interesting pod/router-default-5444994796-58hwl container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 05 06:50:34 crc kubenswrapper[4846]: [-]has-synced failed: reason withheld Oct 05 06:50:34 crc kubenswrapper[4846]: [+]process-running ok Oct 05 06:50:34 crc kubenswrapper[4846]: healthz check failed Oct 05 06:50:34 crc kubenswrapper[4846]: I1005 06:50:34.710892 4846 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-58hwl" podUID="2ecfaba9-9034-4ac7-bd51-e3f64b41d825" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 05 06:50:34 crc kubenswrapper[4846]: I1005 06:50:34.997925 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-6bpvf" Oct 05 06:50:35 crc kubenswrapper[4846]: I1005 06:50:35.003633 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-6bpvf" Oct 05 06:50:35 crc kubenswrapper[4846]: I1005 06:50:35.711687 4846 patch_prober.go:28] interesting pod/router-default-5444994796-58hwl container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 05 06:50:35 crc kubenswrapper[4846]: [-]has-synced failed: reason withheld Oct 05 06:50:35 crc kubenswrapper[4846]: [+]process-running ok Oct 05 06:50:35 crc kubenswrapper[4846]: healthz check failed Oct 05 06:50:35 crc kubenswrapper[4846]: I1005 06:50:35.712917 4846 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-58hwl" podUID="2ecfaba9-9034-4ac7-bd51-e3f64b41d825" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 05 06:50:35 crc kubenswrapper[4846]: I1005 06:50:35.919170 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-jm4gb" Oct 05 06:50:36 crc kubenswrapper[4846]: I1005 06:50:36.711188 4846 patch_prober.go:28] interesting pod/router-default-5444994796-58hwl container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 05 06:50:36 crc kubenswrapper[4846]: [-]has-synced failed: reason withheld Oct 05 06:50:36 crc kubenswrapper[4846]: [+]process-running ok Oct 05 06:50:36 crc kubenswrapper[4846]: healthz check failed Oct 05 06:50:36 crc kubenswrapper[4846]: I1005 06:50:36.711265 4846 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-58hwl" podUID="2ecfaba9-9034-4ac7-bd51-e3f64b41d825" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 05 06:50:37 crc kubenswrapper[4846]: I1005 06:50:37.710468 4846 patch_prober.go:28] interesting pod/router-default-5444994796-58hwl container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 05 06:50:37 crc kubenswrapper[4846]: [-]has-synced failed: reason withheld Oct 05 06:50:37 crc kubenswrapper[4846]: [+]process-running ok Oct 05 06:50:37 crc kubenswrapper[4846]: healthz check failed Oct 05 06:50:37 crc kubenswrapper[4846]: I1005 06:50:37.710544 4846 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-58hwl" podUID="2ecfaba9-9034-4ac7-bd51-e3f64b41d825" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 05 06:50:37 crc kubenswrapper[4846]: I1005 06:50:37.965807 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 05 06:50:38 crc kubenswrapper[4846]: I1005 06:50:38.009278 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/7ce1a1f8-e7e8-4f26-851c-1d04367a0083-kubelet-dir\") pod \"7ce1a1f8-e7e8-4f26-851c-1d04367a0083\" (UID: \"7ce1a1f8-e7e8-4f26-851c-1d04367a0083\") " Oct 05 06:50:38 crc kubenswrapper[4846]: I1005 06:50:38.009371 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7ce1a1f8-e7e8-4f26-851c-1d04367a0083-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "7ce1a1f8-e7e8-4f26-851c-1d04367a0083" (UID: "7ce1a1f8-e7e8-4f26-851c-1d04367a0083"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 05 06:50:38 crc kubenswrapper[4846]: I1005 06:50:38.009558 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7ce1a1f8-e7e8-4f26-851c-1d04367a0083-kube-api-access\") pod \"7ce1a1f8-e7e8-4f26-851c-1d04367a0083\" (UID: \"7ce1a1f8-e7e8-4f26-851c-1d04367a0083\") " Oct 05 06:50:38 crc kubenswrapper[4846]: I1005 06:50:38.010951 4846 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/7ce1a1f8-e7e8-4f26-851c-1d04367a0083-kubelet-dir\") on node \"crc\" DevicePath \"\"" Oct 05 06:50:38 crc kubenswrapper[4846]: I1005 06:50:38.020429 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7ce1a1f8-e7e8-4f26-851c-1d04367a0083-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "7ce1a1f8-e7e8-4f26-851c-1d04367a0083" (UID: "7ce1a1f8-e7e8-4f26-851c-1d04367a0083"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 06:50:38 crc kubenswrapper[4846]: I1005 06:50:38.111610 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7ce1a1f8-e7e8-4f26-851c-1d04367a0083-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 05 06:50:38 crc kubenswrapper[4846]: I1005 06:50:38.710329 4846 patch_prober.go:28] interesting pod/router-default-5444994796-58hwl container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 05 06:50:38 crc kubenswrapper[4846]: [-]has-synced failed: reason withheld Oct 05 06:50:38 crc kubenswrapper[4846]: [+]process-running ok Oct 05 06:50:38 crc kubenswrapper[4846]: healthz check failed Oct 05 06:50:38 crc kubenswrapper[4846]: I1005 06:50:38.710681 4846 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-58hwl" podUID="2ecfaba9-9034-4ac7-bd51-e3f64b41d825" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 05 06:50:38 crc kubenswrapper[4846]: I1005 06:50:38.769026 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"7ce1a1f8-e7e8-4f26-851c-1d04367a0083","Type":"ContainerDied","Data":"d06fa06f0d5ec92c47c4899ebe5902a625530c8b1832eaf3414f7c3d9a82469a"} Oct 05 06:50:38 crc kubenswrapper[4846]: I1005 06:50:38.769069 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d06fa06f0d5ec92c47c4899ebe5902a625530c8b1832eaf3414f7c3d9a82469a" Oct 05 06:50:38 crc kubenswrapper[4846]: I1005 06:50:38.769142 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 05 06:50:39 crc kubenswrapper[4846]: I1005 06:50:39.483302 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 05 06:50:39 crc kubenswrapper[4846]: I1005 06:50:39.537299 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/d42d9769-b723-4790-b5b5-9cb03fe1cc04-kubelet-dir\") pod \"d42d9769-b723-4790-b5b5-9cb03fe1cc04\" (UID: \"d42d9769-b723-4790-b5b5-9cb03fe1cc04\") " Oct 05 06:50:39 crc kubenswrapper[4846]: I1005 06:50:39.537734 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d42d9769-b723-4790-b5b5-9cb03fe1cc04-kube-api-access\") pod \"d42d9769-b723-4790-b5b5-9cb03fe1cc04\" (UID: \"d42d9769-b723-4790-b5b5-9cb03fe1cc04\") " Oct 05 06:50:39 crc kubenswrapper[4846]: I1005 06:50:39.537505 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d42d9769-b723-4790-b5b5-9cb03fe1cc04-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "d42d9769-b723-4790-b5b5-9cb03fe1cc04" (UID: "d42d9769-b723-4790-b5b5-9cb03fe1cc04"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 05 06:50:39 crc kubenswrapper[4846]: I1005 06:50:39.538359 4846 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/d42d9769-b723-4790-b5b5-9cb03fe1cc04-kubelet-dir\") on node \"crc\" DevicePath \"\"" Oct 05 06:50:39 crc kubenswrapper[4846]: I1005 06:50:39.541519 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d42d9769-b723-4790-b5b5-9cb03fe1cc04-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "d42d9769-b723-4790-b5b5-9cb03fe1cc04" (UID: "d42d9769-b723-4790-b5b5-9cb03fe1cc04"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 06:50:39 crc kubenswrapper[4846]: I1005 06:50:39.639517 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d42d9769-b723-4790-b5b5-9cb03fe1cc04-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 05 06:50:39 crc kubenswrapper[4846]: I1005 06:50:39.710198 4846 patch_prober.go:28] interesting pod/router-default-5444994796-58hwl container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 05 06:50:39 crc kubenswrapper[4846]: [-]has-synced failed: reason withheld Oct 05 06:50:39 crc kubenswrapper[4846]: [+]process-running ok Oct 05 06:50:39 crc kubenswrapper[4846]: healthz check failed Oct 05 06:50:39 crc kubenswrapper[4846]: I1005 06:50:39.710274 4846 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-58hwl" podUID="2ecfaba9-9034-4ac7-bd51-e3f64b41d825" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 05 06:50:39 crc kubenswrapper[4846]: I1005 06:50:39.781560 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"d42d9769-b723-4790-b5b5-9cb03fe1cc04","Type":"ContainerDied","Data":"a7b9dce148b59a413d51cc2996f2db61d3596a0f701d1e76e442f6e1113cf6a3"} Oct 05 06:50:39 crc kubenswrapper[4846]: I1005 06:50:39.781618 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a7b9dce148b59a413d51cc2996f2db61d3596a0f701d1e76e442f6e1113cf6a3" Oct 05 06:50:39 crc kubenswrapper[4846]: I1005 06:50:39.781692 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 05 06:50:39 crc kubenswrapper[4846]: E1005 06:50:39.805747 4846 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-podd42d9769_b723_4790_b5b5_9cb03fe1cc04.slice/crio-a7b9dce148b59a413d51cc2996f2db61d3596a0f701d1e76e442f6e1113cf6a3\": RecentStats: unable to find data in memory cache]" Oct 05 06:50:39 crc kubenswrapper[4846]: I1005 06:50:39.838871 4846 patch_prober.go:28] interesting pod/console-f9d7485db-9hcbc container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.9:8443/health\": dial tcp 10.217.0.9:8443: connect: connection refused" start-of-body= Oct 05 06:50:39 crc kubenswrapper[4846]: I1005 06:50:39.838938 4846 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-9hcbc" podUID="74f16025-fe11-4a23-8c1a-ff3da2a3d751" containerName="console" probeResult="failure" output="Get \"https://10.217.0.9:8443/health\": dial tcp 10.217.0.9:8443: connect: connection refused" Oct 05 06:50:39 crc kubenswrapper[4846]: I1005 06:50:39.903986 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-c6cxs" Oct 05 06:50:40 crc kubenswrapper[4846]: I1005 06:50:40.709473 4846 patch_prober.go:28] interesting pod/router-default-5444994796-58hwl container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 05 06:50:40 crc kubenswrapper[4846]: [-]has-synced failed: reason withheld Oct 05 06:50:40 crc kubenswrapper[4846]: [+]process-running ok Oct 05 06:50:40 crc kubenswrapper[4846]: healthz check failed Oct 05 06:50:40 crc kubenswrapper[4846]: I1005 06:50:40.709732 4846 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-58hwl" podUID="2ecfaba9-9034-4ac7-bd51-e3f64b41d825" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 05 06:50:41 crc kubenswrapper[4846]: I1005 06:50:41.710681 4846 patch_prober.go:28] interesting pod/router-default-5444994796-58hwl container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 05 06:50:41 crc kubenswrapper[4846]: [-]has-synced failed: reason withheld Oct 05 06:50:41 crc kubenswrapper[4846]: [+]process-running ok Oct 05 06:50:41 crc kubenswrapper[4846]: healthz check failed Oct 05 06:50:41 crc kubenswrapper[4846]: I1005 06:50:41.710778 4846 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-58hwl" podUID="2ecfaba9-9034-4ac7-bd51-e3f64b41d825" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 05 06:50:42 crc kubenswrapper[4846]: I1005 06:50:42.709993 4846 patch_prober.go:28] interesting pod/router-default-5444994796-58hwl container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 05 06:50:42 crc kubenswrapper[4846]: [-]has-synced failed: reason withheld Oct 05 06:50:42 crc kubenswrapper[4846]: [+]process-running ok Oct 05 06:50:42 crc kubenswrapper[4846]: healthz check failed Oct 05 06:50:42 crc kubenswrapper[4846]: I1005 06:50:42.710075 4846 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-58hwl" podUID="2ecfaba9-9034-4ac7-bd51-e3f64b41d825" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 05 06:50:43 crc kubenswrapper[4846]: I1005 06:50:43.709262 4846 patch_prober.go:28] interesting pod/router-default-5444994796-58hwl container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 05 06:50:43 crc kubenswrapper[4846]: [-]has-synced failed: reason withheld Oct 05 06:50:43 crc kubenswrapper[4846]: [+]process-running ok Oct 05 06:50:43 crc kubenswrapper[4846]: healthz check failed Oct 05 06:50:43 crc kubenswrapper[4846]: I1005 06:50:43.709350 4846 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-58hwl" podUID="2ecfaba9-9034-4ac7-bd51-e3f64b41d825" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 05 06:50:44 crc kubenswrapper[4846]: I1005 06:50:44.710097 4846 patch_prober.go:28] interesting pod/router-default-5444994796-58hwl container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 05 06:50:44 crc kubenswrapper[4846]: [+]has-synced ok Oct 05 06:50:44 crc kubenswrapper[4846]: [+]process-running ok Oct 05 06:50:44 crc kubenswrapper[4846]: healthz check failed Oct 05 06:50:44 crc kubenswrapper[4846]: I1005 06:50:44.710575 4846 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-58hwl" podUID="2ecfaba9-9034-4ac7-bd51-e3f64b41d825" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 05 06:50:45 crc kubenswrapper[4846]: I1005 06:50:45.712450 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-58hwl" Oct 05 06:50:45 crc kubenswrapper[4846]: I1005 06:50:45.718467 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-58hwl" Oct 05 06:50:46 crc kubenswrapper[4846]: I1005 06:50:46.946351 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/59b83f4f-71f0-4b28-9275-56d5f933ec2c-metrics-certs\") pod \"network-metrics-daemon-9r2kt\" (UID: \"59b83f4f-71f0-4b28-9275-56d5f933ec2c\") " pod="openshift-multus/network-metrics-daemon-9r2kt" Oct 05 06:50:46 crc kubenswrapper[4846]: I1005 06:50:46.952480 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/59b83f4f-71f0-4b28-9275-56d5f933ec2c-metrics-certs\") pod \"network-metrics-daemon-9r2kt\" (UID: \"59b83f4f-71f0-4b28-9275-56d5f933ec2c\") " pod="openshift-multus/network-metrics-daemon-9r2kt" Oct 05 06:50:47 crc kubenswrapper[4846]: I1005 06:50:47.237264 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-9r2kt" Oct 05 06:50:48 crc kubenswrapper[4846]: I1005 06:50:48.964534 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-rf9l4" Oct 05 06:50:49 crc kubenswrapper[4846]: I1005 06:50:49.844265 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-9hcbc" Oct 05 06:50:49 crc kubenswrapper[4846]: I1005 06:50:49.847960 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-9hcbc" Oct 05 06:50:53 crc kubenswrapper[4846]: I1005 06:50:53.325024 4846 patch_prober.go:28] interesting pod/machine-config-daemon-fscvf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 06:50:53 crc kubenswrapper[4846]: I1005 06:50:53.325564 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 06:51:01 crc kubenswrapper[4846]: I1005 06:51:01.049980 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-pqm4r" Oct 05 06:51:05 crc kubenswrapper[4846]: I1005 06:51:05.806761 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 05 06:51:06 crc kubenswrapper[4846]: E1005 06:51:06.142128 4846 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Oct 05 06:51:06 crc kubenswrapper[4846]: E1005 06:51:06.142467 4846 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-dvncs,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-brf8j_openshift-marketplace(8bdbd298-f264-43ea-bb5e-625b1e5ab562): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 05 06:51:06 crc kubenswrapper[4846]: E1005 06:51:06.143877 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-brf8j" podUID="8bdbd298-f264-43ea-bb5e-625b1e5ab562" Oct 05 06:51:11 crc kubenswrapper[4846]: E1005 06:51:11.097214 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-brf8j" podUID="8bdbd298-f264-43ea-bb5e-625b1e5ab562" Oct 05 06:51:11 crc kubenswrapper[4846]: E1005 06:51:11.228822 4846 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Oct 05 06:51:11 crc kubenswrapper[4846]: E1005 06:51:11.229744 4846 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-hmh5t,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-mzc7c_openshift-marketplace(da62c2a4-387f-4d5a-a9f2-04bd1e110d20): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 05 06:51:11 crc kubenswrapper[4846]: E1005 06:51:11.231343 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-mzc7c" podUID="da62c2a4-387f-4d5a-a9f2-04bd1e110d20" Oct 05 06:51:11 crc kubenswrapper[4846]: E1005 06:51:11.326389 4846 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Oct 05 06:51:11 crc kubenswrapper[4846]: E1005 06:51:11.326602 4846 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-kjdm9,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-76jjh_openshift-marketplace(73ca18b5-0a68-4990-9e7c-1685f87c8dcb): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 05 06:51:11 crc kubenswrapper[4846]: E1005 06:51:11.327834 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-76jjh" podUID="73ca18b5-0a68-4990-9e7c-1685f87c8dcb" Oct 05 06:51:12 crc kubenswrapper[4846]: E1005 06:51:12.030368 4846 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Oct 05 06:51:12 crc kubenswrapper[4846]: E1005 06:51:12.030931 4846 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-brxrf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-dczxn_openshift-marketplace(b6446bcc-5edd-46bd-a2dc-e8b4773f7248): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 05 06:51:12 crc kubenswrapper[4846]: E1005 06:51:12.034310 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-dczxn" podUID="b6446bcc-5edd-46bd-a2dc-e8b4773f7248" Oct 05 06:51:15 crc kubenswrapper[4846]: E1005 06:51:15.713565 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-76jjh" podUID="73ca18b5-0a68-4990-9e7c-1685f87c8dcb" Oct 05 06:51:15 crc kubenswrapper[4846]: E1005 06:51:15.713656 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-mzc7c" podUID="da62c2a4-387f-4d5a-a9f2-04bd1e110d20" Oct 05 06:51:15 crc kubenswrapper[4846]: E1005 06:51:15.713806 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-dczxn" podUID="b6446bcc-5edd-46bd-a2dc-e8b4773f7248" Oct 05 06:51:15 crc kubenswrapper[4846]: E1005 06:51:15.897240 4846 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Oct 05 06:51:15 crc kubenswrapper[4846]: E1005 06:51:15.897411 4846 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-fqsk2,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-knslg_openshift-marketplace(0984e21c-5c51-4493-80b9-6e0fadf7537d): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 05 06:51:15 crc kubenswrapper[4846]: E1005 06:51:15.899136 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-knslg" podUID="0984e21c-5c51-4493-80b9-6e0fadf7537d" Oct 05 06:51:15 crc kubenswrapper[4846]: E1005 06:51:15.981838 4846 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Oct 05 06:51:15 crc kubenswrapper[4846]: E1005 06:51:15.982009 4846 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-7dh2q,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-s9f8z_openshift-marketplace(39351673-4e84-4ce0-b669-e2ca9072a443): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 05 06:51:15 crc kubenswrapper[4846]: E1005 06:51:15.983229 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-s9f8z" podUID="39351673-4e84-4ce0-b669-e2ca9072a443" Oct 05 06:51:16 crc kubenswrapper[4846]: E1005 06:51:16.009329 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-knslg" podUID="0984e21c-5c51-4493-80b9-6e0fadf7537d" Oct 05 06:51:16 crc kubenswrapper[4846]: E1005 06:51:16.009670 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-s9f8z" podUID="39351673-4e84-4ce0-b669-e2ca9072a443" Oct 05 06:51:16 crc kubenswrapper[4846]: E1005 06:51:16.072996 4846 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Oct 05 06:51:16 crc kubenswrapper[4846]: E1005 06:51:16.073404 4846 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-n29f4,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-w424l_openshift-marketplace(63ae883f-8fb0-4270-923b-1ecad63773b0): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 05 06:51:16 crc kubenswrapper[4846]: E1005 06:51:16.074725 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-w424l" podUID="63ae883f-8fb0-4270-923b-1ecad63773b0" Oct 05 06:51:16 crc kubenswrapper[4846]: I1005 06:51:16.141781 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-9r2kt"] Oct 05 06:51:16 crc kubenswrapper[4846]: W1005 06:51:16.146100 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod59b83f4f_71f0_4b28_9275_56d5f933ec2c.slice/crio-dcc5399819796ca8c79757189f8ad757f51cbdc76a588cdc421196d44b2a7349 WatchSource:0}: Error finding container dcc5399819796ca8c79757189f8ad757f51cbdc76a588cdc421196d44b2a7349: Status 404 returned error can't find the container with id dcc5399819796ca8c79757189f8ad757f51cbdc76a588cdc421196d44b2a7349 Oct 05 06:51:16 crc kubenswrapper[4846]: E1005 06:51:16.285531 4846 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Oct 05 06:51:16 crc kubenswrapper[4846]: E1005 06:51:16.285768 4846 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-fjsks,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-z858q_openshift-marketplace(3df540e4-b545-4c4a-88c8-b20c6005bd84): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 05 06:51:16 crc kubenswrapper[4846]: E1005 06:51:16.287526 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-z858q" podUID="3df540e4-b545-4c4a-88c8-b20c6005bd84" Oct 05 06:51:17 crc kubenswrapper[4846]: I1005 06:51:17.017714 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-9r2kt" event={"ID":"59b83f4f-71f0-4b28-9275-56d5f933ec2c","Type":"ContainerStarted","Data":"766e3536f4c95383b4f6b810c29741be68fdc532c1cf12fd3e26764e07ea34e7"} Oct 05 06:51:17 crc kubenswrapper[4846]: I1005 06:51:17.018222 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-9r2kt" event={"ID":"59b83f4f-71f0-4b28-9275-56d5f933ec2c","Type":"ContainerStarted","Data":"dcc5399819796ca8c79757189f8ad757f51cbdc76a588cdc421196d44b2a7349"} Oct 05 06:51:17 crc kubenswrapper[4846]: E1005 06:51:17.020406 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-z858q" podUID="3df540e4-b545-4c4a-88c8-b20c6005bd84" Oct 05 06:51:17 crc kubenswrapper[4846]: E1005 06:51:17.021204 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-w424l" podUID="63ae883f-8fb0-4270-923b-1ecad63773b0" Oct 05 06:51:18 crc kubenswrapper[4846]: I1005 06:51:18.028400 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-9r2kt" event={"ID":"59b83f4f-71f0-4b28-9275-56d5f933ec2c","Type":"ContainerStarted","Data":"8eb4a5ba99c5f2f4997f4c03040575f48cb4e7954feaeab555068fe90545adac"} Oct 05 06:51:18 crc kubenswrapper[4846]: I1005 06:51:18.058281 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-9r2kt" podStartSLOduration=174.058249562 podStartE2EDuration="2m54.058249562s" podCreationTimestamp="2025-10-05 06:48:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 06:51:18.055387797 +0000 UTC m=+200.296240642" watchObservedRunningTime="2025-10-05 06:51:18.058249562 +0000 UTC m=+200.299102377" Oct 05 06:51:23 crc kubenswrapper[4846]: I1005 06:51:23.325288 4846 patch_prober.go:28] interesting pod/machine-config-daemon-fscvf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 06:51:23 crc kubenswrapper[4846]: I1005 06:51:23.326562 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 06:51:28 crc kubenswrapper[4846]: I1005 06:51:28.094555 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-brf8j" event={"ID":"8bdbd298-f264-43ea-bb5e-625b1e5ab562","Type":"ContainerStarted","Data":"e18916b3e527e2b20584f6cd25b307850205b8e2944915603fe1e57692a58734"} Oct 05 06:51:29 crc kubenswrapper[4846]: I1005 06:51:29.109064 4846 generic.go:334] "Generic (PLEG): container finished" podID="8bdbd298-f264-43ea-bb5e-625b1e5ab562" containerID="e18916b3e527e2b20584f6cd25b307850205b8e2944915603fe1e57692a58734" exitCode=0 Oct 05 06:51:29 crc kubenswrapper[4846]: I1005 06:51:29.109149 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-brf8j" event={"ID":"8bdbd298-f264-43ea-bb5e-625b1e5ab562","Type":"ContainerDied","Data":"e18916b3e527e2b20584f6cd25b307850205b8e2944915603fe1e57692a58734"} Oct 05 06:51:31 crc kubenswrapper[4846]: I1005 06:51:31.127735 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mzc7c" event={"ID":"da62c2a4-387f-4d5a-a9f2-04bd1e110d20","Type":"ContainerStarted","Data":"1bd92726c1452bf8f513f6467bd996e359d2e8eaf05ccdff4e933df05df839ec"} Oct 05 06:51:31 crc kubenswrapper[4846]: I1005 06:51:31.137389 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dczxn" event={"ID":"b6446bcc-5edd-46bd-a2dc-e8b4773f7248","Type":"ContainerStarted","Data":"39e16332d58703e73090ebf4dcff7a2dbb1e92e1320890b62c8d0c1391f81b4f"} Oct 05 06:51:32 crc kubenswrapper[4846]: I1005 06:51:32.144405 4846 generic.go:334] "Generic (PLEG): container finished" podID="b6446bcc-5edd-46bd-a2dc-e8b4773f7248" containerID="39e16332d58703e73090ebf4dcff7a2dbb1e92e1320890b62c8d0c1391f81b4f" exitCode=0 Oct 05 06:51:32 crc kubenswrapper[4846]: I1005 06:51:32.144486 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dczxn" event={"ID":"b6446bcc-5edd-46bd-a2dc-e8b4773f7248","Type":"ContainerDied","Data":"39e16332d58703e73090ebf4dcff7a2dbb1e92e1320890b62c8d0c1391f81b4f"} Oct 05 06:51:32 crc kubenswrapper[4846]: I1005 06:51:32.146578 4846 generic.go:334] "Generic (PLEG): container finished" podID="da62c2a4-387f-4d5a-a9f2-04bd1e110d20" containerID="1bd92726c1452bf8f513f6467bd996e359d2e8eaf05ccdff4e933df05df839ec" exitCode=0 Oct 05 06:51:32 crc kubenswrapper[4846]: I1005 06:51:32.146623 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mzc7c" event={"ID":"da62c2a4-387f-4d5a-a9f2-04bd1e110d20","Type":"ContainerDied","Data":"1bd92726c1452bf8f513f6467bd996e359d2e8eaf05ccdff4e933df05df839ec"} Oct 05 06:51:34 crc kubenswrapper[4846]: I1005 06:51:34.167819 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-brf8j" event={"ID":"8bdbd298-f264-43ea-bb5e-625b1e5ab562","Type":"ContainerStarted","Data":"05dc68992e82ca3e0d5b965865a8dd455139010d570c6ccd0cd86b742caf577b"} Oct 05 06:51:36 crc kubenswrapper[4846]: I1005 06:51:36.216630 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-brf8j" podStartSLOduration=7.292707757 podStartE2EDuration="1m9.21660664s" podCreationTimestamp="2025-10-05 06:50:27 +0000 UTC" firstStartedPulling="2025-10-05 06:50:29.546527792 +0000 UTC m=+151.787380567" lastFinishedPulling="2025-10-05 06:51:31.470426635 +0000 UTC m=+213.711279450" observedRunningTime="2025-10-05 06:51:36.205874309 +0000 UTC m=+218.446727124" watchObservedRunningTime="2025-10-05 06:51:36.21660664 +0000 UTC m=+218.457459425" Oct 05 06:51:37 crc kubenswrapper[4846]: I1005 06:51:37.889606 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-brf8j" Oct 05 06:51:37 crc kubenswrapper[4846]: I1005 06:51:37.894498 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-brf8j" Oct 05 06:51:39 crc kubenswrapper[4846]: I1005 06:51:39.631606 4846 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-brf8j" podUID="8bdbd298-f264-43ea-bb5e-625b1e5ab562" containerName="registry-server" probeResult="failure" output=< Oct 05 06:51:39 crc kubenswrapper[4846]: timeout: failed to connect service ":50051" within 1s Oct 05 06:51:39 crc kubenswrapper[4846]: > Oct 05 06:51:47 crc kubenswrapper[4846]: I1005 06:51:47.274415 4846 generic.go:334] "Generic (PLEG): container finished" podID="39351673-4e84-4ce0-b669-e2ca9072a443" containerID="1cadd90a51fa4f86dfe9e6212a68bbf446a83369a4be2b8d5c35e850b631670d" exitCode=0 Oct 05 06:51:47 crc kubenswrapper[4846]: I1005 06:51:47.274502 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-s9f8z" event={"ID":"39351673-4e84-4ce0-b669-e2ca9072a443","Type":"ContainerDied","Data":"1cadd90a51fa4f86dfe9e6212a68bbf446a83369a4be2b8d5c35e850b631670d"} Oct 05 06:51:47 crc kubenswrapper[4846]: I1005 06:51:47.278597 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-76jjh" event={"ID":"73ca18b5-0a68-4990-9e7c-1685f87c8dcb","Type":"ContainerStarted","Data":"90b774a7005c4431444506a41a786f922e160117a560487bb571077a73443327"} Oct 05 06:51:47 crc kubenswrapper[4846]: I1005 06:51:47.289698 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-knslg" event={"ID":"0984e21c-5c51-4493-80b9-6e0fadf7537d","Type":"ContainerStarted","Data":"8764d0318b784edf4871c6b83bf36b77a3a9539dd3aa682ebd544172e9dd882d"} Oct 05 06:51:47 crc kubenswrapper[4846]: I1005 06:51:47.293289 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mzc7c" event={"ID":"da62c2a4-387f-4d5a-a9f2-04bd1e110d20","Type":"ContainerStarted","Data":"30cb1644498c9f7df6b72282eca59d86cf0d32592b4adb1858f5ef1453b4bc80"} Oct 05 06:51:47 crc kubenswrapper[4846]: I1005 06:51:47.295727 4846 generic.go:334] "Generic (PLEG): container finished" podID="3df540e4-b545-4c4a-88c8-b20c6005bd84" containerID="9150a82905d8e2ccbdb1cd6682e9411505f6e1cb8b480fc4cfb47ef54a6e67e4" exitCode=0 Oct 05 06:51:47 crc kubenswrapper[4846]: I1005 06:51:47.295790 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-z858q" event={"ID":"3df540e4-b545-4c4a-88c8-b20c6005bd84","Type":"ContainerDied","Data":"9150a82905d8e2ccbdb1cd6682e9411505f6e1cb8b480fc4cfb47ef54a6e67e4"} Oct 05 06:51:47 crc kubenswrapper[4846]: I1005 06:51:47.298233 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dczxn" event={"ID":"b6446bcc-5edd-46bd-a2dc-e8b4773f7248","Type":"ContainerStarted","Data":"c90a269f8eaa15bb7b13ad740330f0936c8254ce793c00b727e5fbad6dd371f5"} Oct 05 06:51:47 crc kubenswrapper[4846]: I1005 06:51:47.300149 4846 generic.go:334] "Generic (PLEG): container finished" podID="63ae883f-8fb0-4270-923b-1ecad63773b0" containerID="de1f861bacffbed3c51c8dd7cdb953edb5491269d9fcadf8accde95daf0506e8" exitCode=0 Oct 05 06:51:47 crc kubenswrapper[4846]: I1005 06:51:47.300190 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-w424l" event={"ID":"63ae883f-8fb0-4270-923b-1ecad63773b0","Type":"ContainerDied","Data":"de1f861bacffbed3c51c8dd7cdb953edb5491269d9fcadf8accde95daf0506e8"} Oct 05 06:51:47 crc kubenswrapper[4846]: I1005 06:51:47.350792 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-mzc7c" podStartSLOduration=4.229995616 podStartE2EDuration="1m21.350770027s" podCreationTimestamp="2025-10-05 06:50:26 +0000 UTC" firstStartedPulling="2025-10-05 06:50:28.515764424 +0000 UTC m=+150.756617199" lastFinishedPulling="2025-10-05 06:51:45.636538775 +0000 UTC m=+227.877391610" observedRunningTime="2025-10-05 06:51:47.332711265 +0000 UTC m=+229.573564050" watchObservedRunningTime="2025-10-05 06:51:47.350770027 +0000 UTC m=+229.591622802" Oct 05 06:51:47 crc kubenswrapper[4846]: I1005 06:51:47.391842 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-dczxn" podStartSLOduration=3.41116001 podStartE2EDuration="1m18.39181797s" podCreationTimestamp="2025-10-05 06:50:29 +0000 UTC" firstStartedPulling="2025-10-05 06:50:30.658580995 +0000 UTC m=+152.899433770" lastFinishedPulling="2025-10-05 06:51:45.639238915 +0000 UTC m=+227.880091730" observedRunningTime="2025-10-05 06:51:47.391448301 +0000 UTC m=+229.632301066" watchObservedRunningTime="2025-10-05 06:51:47.39181797 +0000 UTC m=+229.632670745" Oct 05 06:51:47 crc kubenswrapper[4846]: I1005 06:51:47.953365 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-brf8j" Oct 05 06:51:48 crc kubenswrapper[4846]: I1005 06:51:48.030921 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-brf8j" Oct 05 06:51:48 crc kubenswrapper[4846]: I1005 06:51:48.306455 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-w424l" event={"ID":"63ae883f-8fb0-4270-923b-1ecad63773b0","Type":"ContainerStarted","Data":"0c7c749f8b6922eab76557e0fe7dcc92783d402f4a317565a2472d1c9c2b3d56"} Oct 05 06:51:48 crc kubenswrapper[4846]: I1005 06:51:48.310848 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-s9f8z" event={"ID":"39351673-4e84-4ce0-b669-e2ca9072a443","Type":"ContainerStarted","Data":"96394660b5ae8069fea21708a9ad5973e9fc889b0c4241674ad871e2bfe27c78"} Oct 05 06:51:48 crc kubenswrapper[4846]: I1005 06:51:48.312949 4846 generic.go:334] "Generic (PLEG): container finished" podID="73ca18b5-0a68-4990-9e7c-1685f87c8dcb" containerID="90b774a7005c4431444506a41a786f922e160117a560487bb571077a73443327" exitCode=0 Oct 05 06:51:48 crc kubenswrapper[4846]: I1005 06:51:48.313014 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-76jjh" event={"ID":"73ca18b5-0a68-4990-9e7c-1685f87c8dcb","Type":"ContainerDied","Data":"90b774a7005c4431444506a41a786f922e160117a560487bb571077a73443327"} Oct 05 06:51:48 crc kubenswrapper[4846]: I1005 06:51:48.315437 4846 generic.go:334] "Generic (PLEG): container finished" podID="0984e21c-5c51-4493-80b9-6e0fadf7537d" containerID="8764d0318b784edf4871c6b83bf36b77a3a9539dd3aa682ebd544172e9dd882d" exitCode=0 Oct 05 06:51:48 crc kubenswrapper[4846]: I1005 06:51:48.315488 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-knslg" event={"ID":"0984e21c-5c51-4493-80b9-6e0fadf7537d","Type":"ContainerDied","Data":"8764d0318b784edf4871c6b83bf36b77a3a9539dd3aa682ebd544172e9dd882d"} Oct 05 06:51:48 crc kubenswrapper[4846]: I1005 06:51:48.319637 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-z858q" event={"ID":"3df540e4-b545-4c4a-88c8-b20c6005bd84","Type":"ContainerStarted","Data":"232c3f8283d741faccd033b340a724340f556272a9cc80c5fa2377b70b1b1210"} Oct 05 06:51:48 crc kubenswrapper[4846]: I1005 06:51:48.326514 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-w424l" podStartSLOduration=2.011201478 podStartE2EDuration="1m18.326497878s" podCreationTimestamp="2025-10-05 06:50:30 +0000 UTC" firstStartedPulling="2025-10-05 06:50:31.703538397 +0000 UTC m=+153.944391162" lastFinishedPulling="2025-10-05 06:51:48.018834767 +0000 UTC m=+230.259687562" observedRunningTime="2025-10-05 06:51:48.323448418 +0000 UTC m=+230.564301193" watchObservedRunningTime="2025-10-05 06:51:48.326497878 +0000 UTC m=+230.567350653" Oct 05 06:51:48 crc kubenswrapper[4846]: I1005 06:51:48.345469 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-s9f8z" podStartSLOduration=1.948015486 podStartE2EDuration="1m21.345451193s" podCreationTimestamp="2025-10-05 06:50:27 +0000 UTC" firstStartedPulling="2025-10-05 06:50:28.481272878 +0000 UTC m=+150.722125653" lastFinishedPulling="2025-10-05 06:51:47.878708585 +0000 UTC m=+230.119561360" observedRunningTime="2025-10-05 06:51:48.344063887 +0000 UTC m=+230.584916662" watchObservedRunningTime="2025-10-05 06:51:48.345451193 +0000 UTC m=+230.586303968" Oct 05 06:51:48 crc kubenswrapper[4846]: I1005 06:51:48.364193 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-z858q" podStartSLOduration=3.106540008 podStartE2EDuration="1m19.364155162s" podCreationTimestamp="2025-10-05 06:50:29 +0000 UTC" firstStartedPulling="2025-10-05 06:50:31.684627731 +0000 UTC m=+153.925480506" lastFinishedPulling="2025-10-05 06:51:47.942242875 +0000 UTC m=+230.183095660" observedRunningTime="2025-10-05 06:51:48.361953764 +0000 UTC m=+230.602806539" watchObservedRunningTime="2025-10-05 06:51:48.364155162 +0000 UTC m=+230.605007937" Oct 05 06:51:49 crc kubenswrapper[4846]: I1005 06:51:49.327993 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-knslg" event={"ID":"0984e21c-5c51-4493-80b9-6e0fadf7537d","Type":"ContainerStarted","Data":"6ef045a66699331e787f02a0f98e95b9b5d68476955be9f1777ac3a1e593d959"} Oct 05 06:51:49 crc kubenswrapper[4846]: I1005 06:51:49.330627 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-76jjh" event={"ID":"73ca18b5-0a68-4990-9e7c-1685f87c8dcb","Type":"ContainerStarted","Data":"3e426226198ad8c4a4ed011f91f55edfcd595dc146977d4d9c4feb6e4eb28be7"} Oct 05 06:51:49 crc kubenswrapper[4846]: I1005 06:51:49.351977 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-knslg" podStartSLOduration=2.2320311569999998 podStartE2EDuration="1m19.351950279s" podCreationTimestamp="2025-10-05 06:50:30 +0000 UTC" firstStartedPulling="2025-10-05 06:50:31.671729883 +0000 UTC m=+153.912582658" lastFinishedPulling="2025-10-05 06:51:48.791649005 +0000 UTC m=+231.032501780" observedRunningTime="2025-10-05 06:51:49.349050153 +0000 UTC m=+231.589902928" watchObservedRunningTime="2025-10-05 06:51:49.351950279 +0000 UTC m=+231.592803054" Oct 05 06:51:49 crc kubenswrapper[4846]: I1005 06:51:49.611980 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-dczxn" Oct 05 06:51:49 crc kubenswrapper[4846]: I1005 06:51:49.612047 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-dczxn" Oct 05 06:51:49 crc kubenswrapper[4846]: I1005 06:51:49.664610 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-dczxn" Oct 05 06:51:49 crc kubenswrapper[4846]: I1005 06:51:49.686105 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-76jjh" podStartSLOduration=3.5594541509999997 podStartE2EDuration="1m22.686080401s" podCreationTimestamp="2025-10-05 06:50:27 +0000 UTC" firstStartedPulling="2025-10-05 06:50:29.599474612 +0000 UTC m=+151.840327387" lastFinishedPulling="2025-10-05 06:51:48.726100862 +0000 UTC m=+230.966953637" observedRunningTime="2025-10-05 06:51:49.3707482 +0000 UTC m=+231.611600985" watchObservedRunningTime="2025-10-05 06:51:49.686080401 +0000 UTC m=+231.926933176" Oct 05 06:51:50 crc kubenswrapper[4846]: I1005 06:51:50.023142 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-z858q" Oct 05 06:51:50 crc kubenswrapper[4846]: I1005 06:51:50.023444 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-z858q" Oct 05 06:51:50 crc kubenswrapper[4846]: I1005 06:51:50.070270 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-z858q" Oct 05 06:51:50 crc kubenswrapper[4846]: I1005 06:51:50.420299 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-knslg" Oct 05 06:51:50 crc kubenswrapper[4846]: I1005 06:51:50.420977 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-knslg" Oct 05 06:51:50 crc kubenswrapper[4846]: I1005 06:51:50.660009 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-w424l" Oct 05 06:51:50 crc kubenswrapper[4846]: I1005 06:51:50.660367 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-w424l" Oct 05 06:51:50 crc kubenswrapper[4846]: I1005 06:51:50.923212 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-brf8j"] Oct 05 06:51:50 crc kubenswrapper[4846]: I1005 06:51:50.923561 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-brf8j" podUID="8bdbd298-f264-43ea-bb5e-625b1e5ab562" containerName="registry-server" containerID="cri-o://05dc68992e82ca3e0d5b965865a8dd455139010d570c6ccd0cd86b742caf577b" gracePeriod=2 Oct 05 06:51:51 crc kubenswrapper[4846]: I1005 06:51:51.312653 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-brf8j" Oct 05 06:51:51 crc kubenswrapper[4846]: I1005 06:51:51.355307 4846 generic.go:334] "Generic (PLEG): container finished" podID="8bdbd298-f264-43ea-bb5e-625b1e5ab562" containerID="05dc68992e82ca3e0d5b965865a8dd455139010d570c6ccd0cd86b742caf577b" exitCode=0 Oct 05 06:51:51 crc kubenswrapper[4846]: I1005 06:51:51.356117 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-brf8j" Oct 05 06:51:51 crc kubenswrapper[4846]: I1005 06:51:51.356167 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-brf8j" event={"ID":"8bdbd298-f264-43ea-bb5e-625b1e5ab562","Type":"ContainerDied","Data":"05dc68992e82ca3e0d5b965865a8dd455139010d570c6ccd0cd86b742caf577b"} Oct 05 06:51:51 crc kubenswrapper[4846]: I1005 06:51:51.356227 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-brf8j" event={"ID":"8bdbd298-f264-43ea-bb5e-625b1e5ab562","Type":"ContainerDied","Data":"4f8b6e87a091660159594171adf0f5860837cfc2cade653d3d35e120ea202d44"} Oct 05 06:51:51 crc kubenswrapper[4846]: I1005 06:51:51.356249 4846 scope.go:117] "RemoveContainer" containerID="05dc68992e82ca3e0d5b965865a8dd455139010d570c6ccd0cd86b742caf577b" Oct 05 06:51:51 crc kubenswrapper[4846]: I1005 06:51:51.375922 4846 scope.go:117] "RemoveContainer" containerID="e18916b3e527e2b20584f6cd25b307850205b8e2944915603fe1e57692a58734" Oct 05 06:51:51 crc kubenswrapper[4846]: I1005 06:51:51.393103 4846 scope.go:117] "RemoveContainer" containerID="c8501678cb72712ce3f8d7c9c7c848aa787d2ddd138114782292e24dd8fde2d1" Oct 05 06:51:51 crc kubenswrapper[4846]: I1005 06:51:51.408386 4846 scope.go:117] "RemoveContainer" containerID="05dc68992e82ca3e0d5b965865a8dd455139010d570c6ccd0cd86b742caf577b" Oct 05 06:51:51 crc kubenswrapper[4846]: E1005 06:51:51.414699 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"05dc68992e82ca3e0d5b965865a8dd455139010d570c6ccd0cd86b742caf577b\": container with ID starting with 05dc68992e82ca3e0d5b965865a8dd455139010d570c6ccd0cd86b742caf577b not found: ID does not exist" containerID="05dc68992e82ca3e0d5b965865a8dd455139010d570c6ccd0cd86b742caf577b" Oct 05 06:51:51 crc kubenswrapper[4846]: I1005 06:51:51.414740 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"05dc68992e82ca3e0d5b965865a8dd455139010d570c6ccd0cd86b742caf577b"} err="failed to get container status \"05dc68992e82ca3e0d5b965865a8dd455139010d570c6ccd0cd86b742caf577b\": rpc error: code = NotFound desc = could not find container \"05dc68992e82ca3e0d5b965865a8dd455139010d570c6ccd0cd86b742caf577b\": container with ID starting with 05dc68992e82ca3e0d5b965865a8dd455139010d570c6ccd0cd86b742caf577b not found: ID does not exist" Oct 05 06:51:51 crc kubenswrapper[4846]: I1005 06:51:51.414791 4846 scope.go:117] "RemoveContainer" containerID="e18916b3e527e2b20584f6cd25b307850205b8e2944915603fe1e57692a58734" Oct 05 06:51:51 crc kubenswrapper[4846]: E1005 06:51:51.415204 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e18916b3e527e2b20584f6cd25b307850205b8e2944915603fe1e57692a58734\": container with ID starting with e18916b3e527e2b20584f6cd25b307850205b8e2944915603fe1e57692a58734 not found: ID does not exist" containerID="e18916b3e527e2b20584f6cd25b307850205b8e2944915603fe1e57692a58734" Oct 05 06:51:51 crc kubenswrapper[4846]: I1005 06:51:51.415227 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e18916b3e527e2b20584f6cd25b307850205b8e2944915603fe1e57692a58734"} err="failed to get container status \"e18916b3e527e2b20584f6cd25b307850205b8e2944915603fe1e57692a58734\": rpc error: code = NotFound desc = could not find container \"e18916b3e527e2b20584f6cd25b307850205b8e2944915603fe1e57692a58734\": container with ID starting with e18916b3e527e2b20584f6cd25b307850205b8e2944915603fe1e57692a58734 not found: ID does not exist" Oct 05 06:51:51 crc kubenswrapper[4846]: I1005 06:51:51.415240 4846 scope.go:117] "RemoveContainer" containerID="c8501678cb72712ce3f8d7c9c7c848aa787d2ddd138114782292e24dd8fde2d1" Oct 05 06:51:51 crc kubenswrapper[4846]: E1005 06:51:51.415476 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c8501678cb72712ce3f8d7c9c7c848aa787d2ddd138114782292e24dd8fde2d1\": container with ID starting with c8501678cb72712ce3f8d7c9c7c848aa787d2ddd138114782292e24dd8fde2d1 not found: ID does not exist" containerID="c8501678cb72712ce3f8d7c9c7c848aa787d2ddd138114782292e24dd8fde2d1" Oct 05 06:51:51 crc kubenswrapper[4846]: I1005 06:51:51.415497 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c8501678cb72712ce3f8d7c9c7c848aa787d2ddd138114782292e24dd8fde2d1"} err="failed to get container status \"c8501678cb72712ce3f8d7c9c7c848aa787d2ddd138114782292e24dd8fde2d1\": rpc error: code = NotFound desc = could not find container \"c8501678cb72712ce3f8d7c9c7c848aa787d2ddd138114782292e24dd8fde2d1\": container with ID starting with c8501678cb72712ce3f8d7c9c7c848aa787d2ddd138114782292e24dd8fde2d1 not found: ID does not exist" Oct 05 06:51:51 crc kubenswrapper[4846]: I1005 06:51:51.455816 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8bdbd298-f264-43ea-bb5e-625b1e5ab562-catalog-content\") pod \"8bdbd298-f264-43ea-bb5e-625b1e5ab562\" (UID: \"8bdbd298-f264-43ea-bb5e-625b1e5ab562\") " Oct 05 06:51:51 crc kubenswrapper[4846]: I1005 06:51:51.455893 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dvncs\" (UniqueName: \"kubernetes.io/projected/8bdbd298-f264-43ea-bb5e-625b1e5ab562-kube-api-access-dvncs\") pod \"8bdbd298-f264-43ea-bb5e-625b1e5ab562\" (UID: \"8bdbd298-f264-43ea-bb5e-625b1e5ab562\") " Oct 05 06:51:51 crc kubenswrapper[4846]: I1005 06:51:51.455966 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8bdbd298-f264-43ea-bb5e-625b1e5ab562-utilities\") pod \"8bdbd298-f264-43ea-bb5e-625b1e5ab562\" (UID: \"8bdbd298-f264-43ea-bb5e-625b1e5ab562\") " Oct 05 06:51:51 crc kubenswrapper[4846]: I1005 06:51:51.456681 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8bdbd298-f264-43ea-bb5e-625b1e5ab562-utilities" (OuterVolumeSpecName: "utilities") pod "8bdbd298-f264-43ea-bb5e-625b1e5ab562" (UID: "8bdbd298-f264-43ea-bb5e-625b1e5ab562"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 06:51:51 crc kubenswrapper[4846]: I1005 06:51:51.462386 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8bdbd298-f264-43ea-bb5e-625b1e5ab562-kube-api-access-dvncs" (OuterVolumeSpecName: "kube-api-access-dvncs") pod "8bdbd298-f264-43ea-bb5e-625b1e5ab562" (UID: "8bdbd298-f264-43ea-bb5e-625b1e5ab562"). InnerVolumeSpecName "kube-api-access-dvncs". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 06:51:51 crc kubenswrapper[4846]: I1005 06:51:51.464721 4846 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-knslg" podUID="0984e21c-5c51-4493-80b9-6e0fadf7537d" containerName="registry-server" probeResult="failure" output=< Oct 05 06:51:51 crc kubenswrapper[4846]: timeout: failed to connect service ":50051" within 1s Oct 05 06:51:51 crc kubenswrapper[4846]: > Oct 05 06:51:51 crc kubenswrapper[4846]: I1005 06:51:51.502444 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8bdbd298-f264-43ea-bb5e-625b1e5ab562-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8bdbd298-f264-43ea-bb5e-625b1e5ab562" (UID: "8bdbd298-f264-43ea-bb5e-625b1e5ab562"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 06:51:51 crc kubenswrapper[4846]: I1005 06:51:51.557907 4846 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8bdbd298-f264-43ea-bb5e-625b1e5ab562-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 05 06:51:51 crc kubenswrapper[4846]: I1005 06:51:51.557952 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dvncs\" (UniqueName: \"kubernetes.io/projected/8bdbd298-f264-43ea-bb5e-625b1e5ab562-kube-api-access-dvncs\") on node \"crc\" DevicePath \"\"" Oct 05 06:51:51 crc kubenswrapper[4846]: I1005 06:51:51.557970 4846 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8bdbd298-f264-43ea-bb5e-625b1e5ab562-utilities\") on node \"crc\" DevicePath \"\"" Oct 05 06:51:51 crc kubenswrapper[4846]: I1005 06:51:51.688222 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-brf8j"] Oct 05 06:51:51 crc kubenswrapper[4846]: I1005 06:51:51.690915 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-brf8j"] Oct 05 06:51:51 crc kubenswrapper[4846]: I1005 06:51:51.705209 4846 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-w424l" podUID="63ae883f-8fb0-4270-923b-1ecad63773b0" containerName="registry-server" probeResult="failure" output=< Oct 05 06:51:51 crc kubenswrapper[4846]: timeout: failed to connect service ":50051" within 1s Oct 05 06:51:51 crc kubenswrapper[4846]: > Oct 05 06:51:52 crc kubenswrapper[4846]: I1005 06:51:52.506239 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8bdbd298-f264-43ea-bb5e-625b1e5ab562" path="/var/lib/kubelet/pods/8bdbd298-f264-43ea-bb5e-625b1e5ab562/volumes" Oct 05 06:51:53 crc kubenswrapper[4846]: I1005 06:51:53.325083 4846 patch_prober.go:28] interesting pod/machine-config-daemon-fscvf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 06:51:53 crc kubenswrapper[4846]: I1005 06:51:53.325191 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 06:51:53 crc kubenswrapper[4846]: I1005 06:51:53.325260 4846 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" Oct 05 06:51:53 crc kubenswrapper[4846]: I1005 06:51:53.326166 4846 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"516f88f4990f5962d8b694a3d889259cc5508a309919870191c2e1331640c973"} pod="openshift-machine-config-operator/machine-config-daemon-fscvf" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 05 06:51:53 crc kubenswrapper[4846]: I1005 06:51:53.326265 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" containerName="machine-config-daemon" containerID="cri-o://516f88f4990f5962d8b694a3d889259cc5508a309919870191c2e1331640c973" gracePeriod=600 Oct 05 06:51:54 crc kubenswrapper[4846]: I1005 06:51:54.382436 4846 generic.go:334] "Generic (PLEG): container finished" podID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" containerID="516f88f4990f5962d8b694a3d889259cc5508a309919870191c2e1331640c973" exitCode=0 Oct 05 06:51:54 crc kubenswrapper[4846]: I1005 06:51:54.382509 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" event={"ID":"ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7","Type":"ContainerDied","Data":"516f88f4990f5962d8b694a3d889259cc5508a309919870191c2e1331640c973"} Oct 05 06:51:55 crc kubenswrapper[4846]: I1005 06:51:55.391083 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" event={"ID":"ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7","Type":"ContainerStarted","Data":"555aa2226a521510ee0c88d4d95100dced793271cffe5d0ca2a8cf181778e4a6"} Oct 05 06:51:57 crc kubenswrapper[4846]: I1005 06:51:57.136826 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-mzc7c" Oct 05 06:51:57 crc kubenswrapper[4846]: I1005 06:51:57.137065 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-mzc7c" Oct 05 06:51:57 crc kubenswrapper[4846]: I1005 06:51:57.183618 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-mzc7c" Oct 05 06:51:57 crc kubenswrapper[4846]: I1005 06:51:57.439130 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-mzc7c" Oct 05 06:51:57 crc kubenswrapper[4846]: I1005 06:51:57.449611 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-s9f8z" Oct 05 06:51:57 crc kubenswrapper[4846]: I1005 06:51:57.449643 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-s9f8z" Oct 05 06:51:57 crc kubenswrapper[4846]: I1005 06:51:57.490372 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-s9f8z" Oct 05 06:51:57 crc kubenswrapper[4846]: I1005 06:51:57.929636 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-76jjh" Oct 05 06:51:57 crc kubenswrapper[4846]: I1005 06:51:57.929707 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-76jjh" Oct 05 06:51:57 crc kubenswrapper[4846]: I1005 06:51:57.977240 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-76jjh" Oct 05 06:51:58 crc kubenswrapper[4846]: I1005 06:51:58.464038 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-s9f8z" Oct 05 06:51:58 crc kubenswrapper[4846]: I1005 06:51:58.466241 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-76jjh" Oct 05 06:51:59 crc kubenswrapper[4846]: I1005 06:51:59.668025 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-dczxn" Oct 05 06:52:00 crc kubenswrapper[4846]: I1005 06:52:00.078068 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-z858q" Oct 05 06:52:00 crc kubenswrapper[4846]: I1005 06:52:00.461605 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-knslg" Oct 05 06:52:00 crc kubenswrapper[4846]: I1005 06:52:00.509448 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-knslg" Oct 05 06:52:00 crc kubenswrapper[4846]: I1005 06:52:00.701061 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-w424l" Oct 05 06:52:00 crc kubenswrapper[4846]: I1005 06:52:00.749796 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-w424l" Oct 05 06:52:00 crc kubenswrapper[4846]: I1005 06:52:00.815668 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-s9f8z"] Oct 05 06:52:00 crc kubenswrapper[4846]: I1005 06:52:00.816012 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-s9f8z" podUID="39351673-4e84-4ce0-b669-e2ca9072a443" containerName="registry-server" containerID="cri-o://96394660b5ae8069fea21708a9ad5973e9fc889b0c4241674ad871e2bfe27c78" gracePeriod=2 Oct 05 06:52:02 crc kubenswrapper[4846]: I1005 06:52:02.213236 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-z858q"] Oct 05 06:52:02 crc kubenswrapper[4846]: I1005 06:52:02.213995 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-z858q" podUID="3df540e4-b545-4c4a-88c8-b20c6005bd84" containerName="registry-server" containerID="cri-o://232c3f8283d741faccd033b340a724340f556272a9cc80c5fa2377b70b1b1210" gracePeriod=2 Oct 05 06:52:02 crc kubenswrapper[4846]: I1005 06:52:02.440931 4846 generic.go:334] "Generic (PLEG): container finished" podID="39351673-4e84-4ce0-b669-e2ca9072a443" containerID="96394660b5ae8069fea21708a9ad5973e9fc889b0c4241674ad871e2bfe27c78" exitCode=0 Oct 05 06:52:02 crc kubenswrapper[4846]: I1005 06:52:02.440991 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-s9f8z" event={"ID":"39351673-4e84-4ce0-b669-e2ca9072a443","Type":"ContainerDied","Data":"96394660b5ae8069fea21708a9ad5973e9fc889b0c4241674ad871e2bfe27c78"} Oct 05 06:52:02 crc kubenswrapper[4846]: I1005 06:52:02.854002 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-s9f8z" Oct 05 06:52:02 crc kubenswrapper[4846]: I1005 06:52:02.921840 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/39351673-4e84-4ce0-b669-e2ca9072a443-catalog-content\") pod \"39351673-4e84-4ce0-b669-e2ca9072a443\" (UID: \"39351673-4e84-4ce0-b669-e2ca9072a443\") " Oct 05 06:52:02 crc kubenswrapper[4846]: I1005 06:52:02.921922 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7dh2q\" (UniqueName: \"kubernetes.io/projected/39351673-4e84-4ce0-b669-e2ca9072a443-kube-api-access-7dh2q\") pod \"39351673-4e84-4ce0-b669-e2ca9072a443\" (UID: \"39351673-4e84-4ce0-b669-e2ca9072a443\") " Oct 05 06:52:02 crc kubenswrapper[4846]: I1005 06:52:02.921980 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/39351673-4e84-4ce0-b669-e2ca9072a443-utilities\") pod \"39351673-4e84-4ce0-b669-e2ca9072a443\" (UID: \"39351673-4e84-4ce0-b669-e2ca9072a443\") " Oct 05 06:52:02 crc kubenswrapper[4846]: I1005 06:52:02.923436 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/39351673-4e84-4ce0-b669-e2ca9072a443-utilities" (OuterVolumeSpecName: "utilities") pod "39351673-4e84-4ce0-b669-e2ca9072a443" (UID: "39351673-4e84-4ce0-b669-e2ca9072a443"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 06:52:02 crc kubenswrapper[4846]: I1005 06:52:02.931354 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/39351673-4e84-4ce0-b669-e2ca9072a443-kube-api-access-7dh2q" (OuterVolumeSpecName: "kube-api-access-7dh2q") pod "39351673-4e84-4ce0-b669-e2ca9072a443" (UID: "39351673-4e84-4ce0-b669-e2ca9072a443"). InnerVolumeSpecName "kube-api-access-7dh2q". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 06:52:02 crc kubenswrapper[4846]: I1005 06:52:02.971447 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/39351673-4e84-4ce0-b669-e2ca9072a443-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "39351673-4e84-4ce0-b669-e2ca9072a443" (UID: "39351673-4e84-4ce0-b669-e2ca9072a443"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 06:52:03 crc kubenswrapper[4846]: I1005 06:52:03.024759 4846 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/39351673-4e84-4ce0-b669-e2ca9072a443-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 05 06:52:03 crc kubenswrapper[4846]: I1005 06:52:03.024832 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7dh2q\" (UniqueName: \"kubernetes.io/projected/39351673-4e84-4ce0-b669-e2ca9072a443-kube-api-access-7dh2q\") on node \"crc\" DevicePath \"\"" Oct 05 06:52:03 crc kubenswrapper[4846]: I1005 06:52:03.024859 4846 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/39351673-4e84-4ce0-b669-e2ca9072a443-utilities\") on node \"crc\" DevicePath \"\"" Oct 05 06:52:03 crc kubenswrapper[4846]: I1005 06:52:03.448801 4846 generic.go:334] "Generic (PLEG): container finished" podID="3df540e4-b545-4c4a-88c8-b20c6005bd84" containerID="232c3f8283d741faccd033b340a724340f556272a9cc80c5fa2377b70b1b1210" exitCode=0 Oct 05 06:52:03 crc kubenswrapper[4846]: I1005 06:52:03.448879 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-z858q" event={"ID":"3df540e4-b545-4c4a-88c8-b20c6005bd84","Type":"ContainerDied","Data":"232c3f8283d741faccd033b340a724340f556272a9cc80c5fa2377b70b1b1210"} Oct 05 06:52:03 crc kubenswrapper[4846]: I1005 06:52:03.451249 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-s9f8z" event={"ID":"39351673-4e84-4ce0-b669-e2ca9072a443","Type":"ContainerDied","Data":"a2dc26496c2aa7176cf68db65f7d800f161d0cf870355071be2166fe1bbd900a"} Oct 05 06:52:03 crc kubenswrapper[4846]: I1005 06:52:03.451328 4846 scope.go:117] "RemoveContainer" containerID="96394660b5ae8069fea21708a9ad5973e9fc889b0c4241674ad871e2bfe27c78" Oct 05 06:52:03 crc kubenswrapper[4846]: I1005 06:52:03.451408 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-s9f8z" Oct 05 06:52:03 crc kubenswrapper[4846]: I1005 06:52:03.466828 4846 scope.go:117] "RemoveContainer" containerID="1cadd90a51fa4f86dfe9e6212a68bbf446a83369a4be2b8d5c35e850b631670d" Oct 05 06:52:03 crc kubenswrapper[4846]: I1005 06:52:03.486338 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-s9f8z"] Oct 05 06:52:03 crc kubenswrapper[4846]: I1005 06:52:03.489290 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-s9f8z"] Oct 05 06:52:03 crc kubenswrapper[4846]: I1005 06:52:03.495655 4846 scope.go:117] "RemoveContainer" containerID="af2013d70507bd5ee301dc25d61ac8f1069d697bc61e97f665e7d47ffdac0d7f" Oct 05 06:52:04 crc kubenswrapper[4846]: I1005 06:52:04.511961 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="39351673-4e84-4ce0-b669-e2ca9072a443" path="/var/lib/kubelet/pods/39351673-4e84-4ce0-b669-e2ca9072a443/volumes" Oct 05 06:52:04 crc kubenswrapper[4846]: I1005 06:52:04.614090 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-w424l"] Oct 05 06:52:04 crc kubenswrapper[4846]: I1005 06:52:04.614388 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-w424l" podUID="63ae883f-8fb0-4270-923b-1ecad63773b0" containerName="registry-server" containerID="cri-o://0c7c749f8b6922eab76557e0fe7dcc92783d402f4a317565a2472d1c9c2b3d56" gracePeriod=2 Oct 05 06:52:04 crc kubenswrapper[4846]: I1005 06:52:04.789371 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-z858q" Oct 05 06:52:04 crc kubenswrapper[4846]: I1005 06:52:04.955778 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3df540e4-b545-4c4a-88c8-b20c6005bd84-catalog-content\") pod \"3df540e4-b545-4c4a-88c8-b20c6005bd84\" (UID: \"3df540e4-b545-4c4a-88c8-b20c6005bd84\") " Oct 05 06:52:04 crc kubenswrapper[4846]: I1005 06:52:04.955826 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3df540e4-b545-4c4a-88c8-b20c6005bd84-utilities\") pod \"3df540e4-b545-4c4a-88c8-b20c6005bd84\" (UID: \"3df540e4-b545-4c4a-88c8-b20c6005bd84\") " Oct 05 06:52:04 crc kubenswrapper[4846]: I1005 06:52:04.955857 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fjsks\" (UniqueName: \"kubernetes.io/projected/3df540e4-b545-4c4a-88c8-b20c6005bd84-kube-api-access-fjsks\") pod \"3df540e4-b545-4c4a-88c8-b20c6005bd84\" (UID: \"3df540e4-b545-4c4a-88c8-b20c6005bd84\") " Oct 05 06:52:04 crc kubenswrapper[4846]: I1005 06:52:04.957927 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3df540e4-b545-4c4a-88c8-b20c6005bd84-utilities" (OuterVolumeSpecName: "utilities") pod "3df540e4-b545-4c4a-88c8-b20c6005bd84" (UID: "3df540e4-b545-4c4a-88c8-b20c6005bd84"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 06:52:04 crc kubenswrapper[4846]: I1005 06:52:04.961254 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3df540e4-b545-4c4a-88c8-b20c6005bd84-kube-api-access-fjsks" (OuterVolumeSpecName: "kube-api-access-fjsks") pod "3df540e4-b545-4c4a-88c8-b20c6005bd84" (UID: "3df540e4-b545-4c4a-88c8-b20c6005bd84"). InnerVolumeSpecName "kube-api-access-fjsks". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 06:52:04 crc kubenswrapper[4846]: I1005 06:52:04.974242 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3df540e4-b545-4c4a-88c8-b20c6005bd84-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3df540e4-b545-4c4a-88c8-b20c6005bd84" (UID: "3df540e4-b545-4c4a-88c8-b20c6005bd84"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 06:52:05 crc kubenswrapper[4846]: I1005 06:52:05.057590 4846 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3df540e4-b545-4c4a-88c8-b20c6005bd84-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 05 06:52:05 crc kubenswrapper[4846]: I1005 06:52:05.057639 4846 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3df540e4-b545-4c4a-88c8-b20c6005bd84-utilities\") on node \"crc\" DevicePath \"\"" Oct 05 06:52:05 crc kubenswrapper[4846]: I1005 06:52:05.057651 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fjsks\" (UniqueName: \"kubernetes.io/projected/3df540e4-b545-4c4a-88c8-b20c6005bd84-kube-api-access-fjsks\") on node \"crc\" DevicePath \"\"" Oct 05 06:52:05 crc kubenswrapper[4846]: I1005 06:52:05.469294 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-z858q" event={"ID":"3df540e4-b545-4c4a-88c8-b20c6005bd84","Type":"ContainerDied","Data":"d6179db76f42e86b59b5c683d0bed1e4bde99212e2fadd1629d7f680caf7c317"} Oct 05 06:52:05 crc kubenswrapper[4846]: I1005 06:52:05.469663 4846 scope.go:117] "RemoveContainer" containerID="232c3f8283d741faccd033b340a724340f556272a9cc80c5fa2377b70b1b1210" Oct 05 06:52:05 crc kubenswrapper[4846]: I1005 06:52:05.469383 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-z858q" Oct 05 06:52:05 crc kubenswrapper[4846]: I1005 06:52:05.488808 4846 scope.go:117] "RemoveContainer" containerID="9150a82905d8e2ccbdb1cd6682e9411505f6e1cb8b480fc4cfb47ef54a6e67e4" Oct 05 06:52:05 crc kubenswrapper[4846]: I1005 06:52:05.500129 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-z858q"] Oct 05 06:52:05 crc kubenswrapper[4846]: I1005 06:52:05.506880 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-z858q"] Oct 05 06:52:05 crc kubenswrapper[4846]: I1005 06:52:05.511966 4846 scope.go:117] "RemoveContainer" containerID="6e59d14c271b5f97931e8fbee43e336919fbb247169aec39b18ee044443d96a7" Oct 05 06:52:05 crc kubenswrapper[4846]: I1005 06:52:05.928243 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-w424l" Oct 05 06:52:06 crc kubenswrapper[4846]: I1005 06:52:06.072139 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/63ae883f-8fb0-4270-923b-1ecad63773b0-utilities\") pod \"63ae883f-8fb0-4270-923b-1ecad63773b0\" (UID: \"63ae883f-8fb0-4270-923b-1ecad63773b0\") " Oct 05 06:52:06 crc kubenswrapper[4846]: I1005 06:52:06.072223 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/63ae883f-8fb0-4270-923b-1ecad63773b0-catalog-content\") pod \"63ae883f-8fb0-4270-923b-1ecad63773b0\" (UID: \"63ae883f-8fb0-4270-923b-1ecad63773b0\") " Oct 05 06:52:06 crc kubenswrapper[4846]: I1005 06:52:06.072253 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n29f4\" (UniqueName: \"kubernetes.io/projected/63ae883f-8fb0-4270-923b-1ecad63773b0-kube-api-access-n29f4\") pod \"63ae883f-8fb0-4270-923b-1ecad63773b0\" (UID: \"63ae883f-8fb0-4270-923b-1ecad63773b0\") " Oct 05 06:52:06 crc kubenswrapper[4846]: I1005 06:52:06.072963 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/63ae883f-8fb0-4270-923b-1ecad63773b0-utilities" (OuterVolumeSpecName: "utilities") pod "63ae883f-8fb0-4270-923b-1ecad63773b0" (UID: "63ae883f-8fb0-4270-923b-1ecad63773b0"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 06:52:06 crc kubenswrapper[4846]: I1005 06:52:06.076165 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/63ae883f-8fb0-4270-923b-1ecad63773b0-kube-api-access-n29f4" (OuterVolumeSpecName: "kube-api-access-n29f4") pod "63ae883f-8fb0-4270-923b-1ecad63773b0" (UID: "63ae883f-8fb0-4270-923b-1ecad63773b0"). InnerVolumeSpecName "kube-api-access-n29f4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 06:52:06 crc kubenswrapper[4846]: I1005 06:52:06.147803 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/63ae883f-8fb0-4270-923b-1ecad63773b0-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "63ae883f-8fb0-4270-923b-1ecad63773b0" (UID: "63ae883f-8fb0-4270-923b-1ecad63773b0"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 06:52:06 crc kubenswrapper[4846]: I1005 06:52:06.173540 4846 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/63ae883f-8fb0-4270-923b-1ecad63773b0-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 05 06:52:06 crc kubenswrapper[4846]: I1005 06:52:06.173606 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n29f4\" (UniqueName: \"kubernetes.io/projected/63ae883f-8fb0-4270-923b-1ecad63773b0-kube-api-access-n29f4\") on node \"crc\" DevicePath \"\"" Oct 05 06:52:06 crc kubenswrapper[4846]: I1005 06:52:06.173622 4846 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/63ae883f-8fb0-4270-923b-1ecad63773b0-utilities\") on node \"crc\" DevicePath \"\"" Oct 05 06:52:06 crc kubenswrapper[4846]: I1005 06:52:06.479266 4846 generic.go:334] "Generic (PLEG): container finished" podID="63ae883f-8fb0-4270-923b-1ecad63773b0" containerID="0c7c749f8b6922eab76557e0fe7dcc92783d402f4a317565a2472d1c9c2b3d56" exitCode=0 Oct 05 06:52:06 crc kubenswrapper[4846]: I1005 06:52:06.479342 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-w424l" Oct 05 06:52:06 crc kubenswrapper[4846]: I1005 06:52:06.479358 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-w424l" event={"ID":"63ae883f-8fb0-4270-923b-1ecad63773b0","Type":"ContainerDied","Data":"0c7c749f8b6922eab76557e0fe7dcc92783d402f4a317565a2472d1c9c2b3d56"} Oct 05 06:52:06 crc kubenswrapper[4846]: I1005 06:52:06.480049 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-w424l" event={"ID":"63ae883f-8fb0-4270-923b-1ecad63773b0","Type":"ContainerDied","Data":"fd87f185be80388544f1dd5aaa8d22762a1e9ec99ffb58f5726dc864fd935592"} Oct 05 06:52:06 crc kubenswrapper[4846]: I1005 06:52:06.480092 4846 scope.go:117] "RemoveContainer" containerID="0c7c749f8b6922eab76557e0fe7dcc92783d402f4a317565a2472d1c9c2b3d56" Oct 05 06:52:06 crc kubenswrapper[4846]: I1005 06:52:06.495599 4846 scope.go:117] "RemoveContainer" containerID="de1f861bacffbed3c51c8dd7cdb953edb5491269d9fcadf8accde95daf0506e8" Oct 05 06:52:06 crc kubenswrapper[4846]: I1005 06:52:06.505686 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3df540e4-b545-4c4a-88c8-b20c6005bd84" path="/var/lib/kubelet/pods/3df540e4-b545-4c4a-88c8-b20c6005bd84/volumes" Oct 05 06:52:06 crc kubenswrapper[4846]: I1005 06:52:06.511055 4846 scope.go:117] "RemoveContainer" containerID="76c946a77e6ddf901939bd92827d436efda3d52f75028018db04751e8e7238e3" Oct 05 06:52:06 crc kubenswrapper[4846]: I1005 06:52:06.523254 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-w424l"] Oct 05 06:52:06 crc kubenswrapper[4846]: I1005 06:52:06.527263 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-w424l"] Oct 05 06:52:06 crc kubenswrapper[4846]: I1005 06:52:06.537127 4846 scope.go:117] "RemoveContainer" containerID="0c7c749f8b6922eab76557e0fe7dcc92783d402f4a317565a2472d1c9c2b3d56" Oct 05 06:52:06 crc kubenswrapper[4846]: E1005 06:52:06.537772 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0c7c749f8b6922eab76557e0fe7dcc92783d402f4a317565a2472d1c9c2b3d56\": container with ID starting with 0c7c749f8b6922eab76557e0fe7dcc92783d402f4a317565a2472d1c9c2b3d56 not found: ID does not exist" containerID="0c7c749f8b6922eab76557e0fe7dcc92783d402f4a317565a2472d1c9c2b3d56" Oct 05 06:52:06 crc kubenswrapper[4846]: I1005 06:52:06.537809 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0c7c749f8b6922eab76557e0fe7dcc92783d402f4a317565a2472d1c9c2b3d56"} err="failed to get container status \"0c7c749f8b6922eab76557e0fe7dcc92783d402f4a317565a2472d1c9c2b3d56\": rpc error: code = NotFound desc = could not find container \"0c7c749f8b6922eab76557e0fe7dcc92783d402f4a317565a2472d1c9c2b3d56\": container with ID starting with 0c7c749f8b6922eab76557e0fe7dcc92783d402f4a317565a2472d1c9c2b3d56 not found: ID does not exist" Oct 05 06:52:06 crc kubenswrapper[4846]: I1005 06:52:06.537838 4846 scope.go:117] "RemoveContainer" containerID="de1f861bacffbed3c51c8dd7cdb953edb5491269d9fcadf8accde95daf0506e8" Oct 05 06:52:06 crc kubenswrapper[4846]: E1005 06:52:06.538391 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"de1f861bacffbed3c51c8dd7cdb953edb5491269d9fcadf8accde95daf0506e8\": container with ID starting with de1f861bacffbed3c51c8dd7cdb953edb5491269d9fcadf8accde95daf0506e8 not found: ID does not exist" containerID="de1f861bacffbed3c51c8dd7cdb953edb5491269d9fcadf8accde95daf0506e8" Oct 05 06:52:06 crc kubenswrapper[4846]: I1005 06:52:06.538417 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"de1f861bacffbed3c51c8dd7cdb953edb5491269d9fcadf8accde95daf0506e8"} err="failed to get container status \"de1f861bacffbed3c51c8dd7cdb953edb5491269d9fcadf8accde95daf0506e8\": rpc error: code = NotFound desc = could not find container \"de1f861bacffbed3c51c8dd7cdb953edb5491269d9fcadf8accde95daf0506e8\": container with ID starting with de1f861bacffbed3c51c8dd7cdb953edb5491269d9fcadf8accde95daf0506e8 not found: ID does not exist" Oct 05 06:52:06 crc kubenswrapper[4846]: I1005 06:52:06.538453 4846 scope.go:117] "RemoveContainer" containerID="76c946a77e6ddf901939bd92827d436efda3d52f75028018db04751e8e7238e3" Oct 05 06:52:06 crc kubenswrapper[4846]: E1005 06:52:06.538705 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"76c946a77e6ddf901939bd92827d436efda3d52f75028018db04751e8e7238e3\": container with ID starting with 76c946a77e6ddf901939bd92827d436efda3d52f75028018db04751e8e7238e3 not found: ID does not exist" containerID="76c946a77e6ddf901939bd92827d436efda3d52f75028018db04751e8e7238e3" Oct 05 06:52:06 crc kubenswrapper[4846]: I1005 06:52:06.538737 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"76c946a77e6ddf901939bd92827d436efda3d52f75028018db04751e8e7238e3"} err="failed to get container status \"76c946a77e6ddf901939bd92827d436efda3d52f75028018db04751e8e7238e3\": rpc error: code = NotFound desc = could not find container \"76c946a77e6ddf901939bd92827d436efda3d52f75028018db04751e8e7238e3\": container with ID starting with 76c946a77e6ddf901939bd92827d436efda3d52f75028018db04751e8e7238e3 not found: ID does not exist" Oct 05 06:52:08 crc kubenswrapper[4846]: I1005 06:52:08.517337 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="63ae883f-8fb0-4270-923b-1ecad63773b0" path="/var/lib/kubelet/pods/63ae883f-8fb0-4270-923b-1ecad63773b0/volumes" Oct 05 06:52:53 crc kubenswrapper[4846]: I1005 06:52:53.912267 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-mzc7c"] Oct 05 06:52:53 crc kubenswrapper[4846]: I1005 06:52:53.921997 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-mzc7c" podUID="da62c2a4-387f-4d5a-a9f2-04bd1e110d20" containerName="registry-server" containerID="cri-o://30cb1644498c9f7df6b72282eca59d86cf0d32592b4adb1858f5ef1453b4bc80" gracePeriod=30 Oct 05 06:52:53 crc kubenswrapper[4846]: I1005 06:52:53.927781 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-76jjh"] Oct 05 06:52:53 crc kubenswrapper[4846]: I1005 06:52:53.928243 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-76jjh" podUID="73ca18b5-0a68-4990-9e7c-1685f87c8dcb" containerName="registry-server" containerID="cri-o://3e426226198ad8c4a4ed011f91f55edfcd595dc146977d4d9c4feb6e4eb28be7" gracePeriod=30 Oct 05 06:52:53 crc kubenswrapper[4846]: I1005 06:52:53.945466 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-2rc5n"] Oct 05 06:52:53 crc kubenswrapper[4846]: I1005 06:52:53.945811 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-2rc5n" podUID="8c2f8349-aa93-4ac8-a369-5b4b2b9cd6e2" containerName="marketplace-operator" containerID="cri-o://f702ca5fddb2d55c48f6f9d3494d5c98b925185899945a42f5693b66af2b7020" gracePeriod=30 Oct 05 06:52:53 crc kubenswrapper[4846]: I1005 06:52:53.951405 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-dczxn"] Oct 05 06:52:53 crc kubenswrapper[4846]: I1005 06:52:53.951737 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-dczxn" podUID="b6446bcc-5edd-46bd-a2dc-e8b4773f7248" containerName="registry-server" containerID="cri-o://c90a269f8eaa15bb7b13ad740330f0936c8254ce793c00b727e5fbad6dd371f5" gracePeriod=30 Oct 05 06:52:53 crc kubenswrapper[4846]: I1005 06:52:53.968618 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-4kl8r"] Oct 05 06:52:53 crc kubenswrapper[4846]: E1005 06:52:53.968918 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3df540e4-b545-4c4a-88c8-b20c6005bd84" containerName="extract-content" Oct 05 06:52:53 crc kubenswrapper[4846]: I1005 06:52:53.968936 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="3df540e4-b545-4c4a-88c8-b20c6005bd84" containerName="extract-content" Oct 05 06:52:53 crc kubenswrapper[4846]: E1005 06:52:53.968955 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3df540e4-b545-4c4a-88c8-b20c6005bd84" containerName="registry-server" Oct 05 06:52:53 crc kubenswrapper[4846]: I1005 06:52:53.968965 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="3df540e4-b545-4c4a-88c8-b20c6005bd84" containerName="registry-server" Oct 05 06:52:53 crc kubenswrapper[4846]: E1005 06:52:53.968977 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="63ae883f-8fb0-4270-923b-1ecad63773b0" containerName="extract-utilities" Oct 05 06:52:53 crc kubenswrapper[4846]: I1005 06:52:53.968987 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="63ae883f-8fb0-4270-923b-1ecad63773b0" containerName="extract-utilities" Oct 05 06:52:53 crc kubenswrapper[4846]: E1005 06:52:53.969010 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8bdbd298-f264-43ea-bb5e-625b1e5ab562" containerName="extract-content" Oct 05 06:52:53 crc kubenswrapper[4846]: I1005 06:52:53.969021 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="8bdbd298-f264-43ea-bb5e-625b1e5ab562" containerName="extract-content" Oct 05 06:52:53 crc kubenswrapper[4846]: E1005 06:52:53.969032 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="39351673-4e84-4ce0-b669-e2ca9072a443" containerName="extract-content" Oct 05 06:52:53 crc kubenswrapper[4846]: I1005 06:52:53.969040 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="39351673-4e84-4ce0-b669-e2ca9072a443" containerName="extract-content" Oct 05 06:52:53 crc kubenswrapper[4846]: E1005 06:52:53.969051 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="63ae883f-8fb0-4270-923b-1ecad63773b0" containerName="registry-server" Oct 05 06:52:53 crc kubenswrapper[4846]: I1005 06:52:53.969059 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="63ae883f-8fb0-4270-923b-1ecad63773b0" containerName="registry-server" Oct 05 06:52:53 crc kubenswrapper[4846]: E1005 06:52:53.969069 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8bdbd298-f264-43ea-bb5e-625b1e5ab562" containerName="extract-utilities" Oct 05 06:52:53 crc kubenswrapper[4846]: I1005 06:52:53.969079 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="8bdbd298-f264-43ea-bb5e-625b1e5ab562" containerName="extract-utilities" Oct 05 06:52:53 crc kubenswrapper[4846]: E1005 06:52:53.969090 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d42d9769-b723-4790-b5b5-9cb03fe1cc04" containerName="pruner" Oct 05 06:52:53 crc kubenswrapper[4846]: I1005 06:52:53.969098 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="d42d9769-b723-4790-b5b5-9cb03fe1cc04" containerName="pruner" Oct 05 06:52:53 crc kubenswrapper[4846]: E1005 06:52:53.969117 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8bdbd298-f264-43ea-bb5e-625b1e5ab562" containerName="registry-server" Oct 05 06:52:53 crc kubenswrapper[4846]: I1005 06:52:53.969126 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="8bdbd298-f264-43ea-bb5e-625b1e5ab562" containerName="registry-server" Oct 05 06:52:53 crc kubenswrapper[4846]: E1005 06:52:53.969139 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="39351673-4e84-4ce0-b669-e2ca9072a443" containerName="extract-utilities" Oct 05 06:52:53 crc kubenswrapper[4846]: I1005 06:52:53.969147 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="39351673-4e84-4ce0-b669-e2ca9072a443" containerName="extract-utilities" Oct 05 06:52:53 crc kubenswrapper[4846]: E1005 06:52:53.969165 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7ce1a1f8-e7e8-4f26-851c-1d04367a0083" containerName="pruner" Oct 05 06:52:53 crc kubenswrapper[4846]: I1005 06:52:53.969201 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="7ce1a1f8-e7e8-4f26-851c-1d04367a0083" containerName="pruner" Oct 05 06:52:53 crc kubenswrapper[4846]: E1005 06:52:53.969220 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="39351673-4e84-4ce0-b669-e2ca9072a443" containerName="registry-server" Oct 05 06:52:53 crc kubenswrapper[4846]: I1005 06:52:53.969233 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="39351673-4e84-4ce0-b669-e2ca9072a443" containerName="registry-server" Oct 05 06:52:53 crc kubenswrapper[4846]: E1005 06:52:53.969248 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="63ae883f-8fb0-4270-923b-1ecad63773b0" containerName="extract-content" Oct 05 06:52:53 crc kubenswrapper[4846]: I1005 06:52:53.969256 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="63ae883f-8fb0-4270-923b-1ecad63773b0" containerName="extract-content" Oct 05 06:52:53 crc kubenswrapper[4846]: E1005 06:52:53.969267 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3df540e4-b545-4c4a-88c8-b20c6005bd84" containerName="extract-utilities" Oct 05 06:52:53 crc kubenswrapper[4846]: I1005 06:52:53.969278 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="3df540e4-b545-4c4a-88c8-b20c6005bd84" containerName="extract-utilities" Oct 05 06:52:53 crc kubenswrapper[4846]: I1005 06:52:53.969419 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="8bdbd298-f264-43ea-bb5e-625b1e5ab562" containerName="registry-server" Oct 05 06:52:53 crc kubenswrapper[4846]: I1005 06:52:53.969432 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="7ce1a1f8-e7e8-4f26-851c-1d04367a0083" containerName="pruner" Oct 05 06:52:53 crc kubenswrapper[4846]: I1005 06:52:53.969450 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="39351673-4e84-4ce0-b669-e2ca9072a443" containerName="registry-server" Oct 05 06:52:53 crc kubenswrapper[4846]: I1005 06:52:53.969464 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="3df540e4-b545-4c4a-88c8-b20c6005bd84" containerName="registry-server" Oct 05 06:52:53 crc kubenswrapper[4846]: I1005 06:52:53.969474 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="d42d9769-b723-4790-b5b5-9cb03fe1cc04" containerName="pruner" Oct 05 06:52:53 crc kubenswrapper[4846]: I1005 06:52:53.969483 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="63ae883f-8fb0-4270-923b-1ecad63773b0" containerName="registry-server" Oct 05 06:52:53 crc kubenswrapper[4846]: I1005 06:52:53.970023 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-4kl8r" Oct 05 06:52:53 crc kubenswrapper[4846]: I1005 06:52:53.976997 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-4kl8r"] Oct 05 06:52:53 crc kubenswrapper[4846]: I1005 06:52:53.982658 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-knslg"] Oct 05 06:52:53 crc kubenswrapper[4846]: I1005 06:52:53.982956 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-knslg" podUID="0984e21c-5c51-4493-80b9-6e0fadf7537d" containerName="registry-server" containerID="cri-o://6ef045a66699331e787f02a0f98e95b9b5d68476955be9f1777ac3a1e593d959" gracePeriod=30 Oct 05 06:52:54 crc kubenswrapper[4846]: I1005 06:52:54.020944 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zv97k\" (UniqueName: \"kubernetes.io/projected/b5257dc4-0faa-4e06-aeea-b25504581b7e-kube-api-access-zv97k\") pod \"marketplace-operator-79b997595-4kl8r\" (UID: \"b5257dc4-0faa-4e06-aeea-b25504581b7e\") " pod="openshift-marketplace/marketplace-operator-79b997595-4kl8r" Oct 05 06:52:54 crc kubenswrapper[4846]: I1005 06:52:54.021296 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b5257dc4-0faa-4e06-aeea-b25504581b7e-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-4kl8r\" (UID: \"b5257dc4-0faa-4e06-aeea-b25504581b7e\") " pod="openshift-marketplace/marketplace-operator-79b997595-4kl8r" Oct 05 06:52:54 crc kubenswrapper[4846]: I1005 06:52:54.021416 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b5257dc4-0faa-4e06-aeea-b25504581b7e-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-4kl8r\" (UID: \"b5257dc4-0faa-4e06-aeea-b25504581b7e\") " pod="openshift-marketplace/marketplace-operator-79b997595-4kl8r" Oct 05 06:52:54 crc kubenswrapper[4846]: I1005 06:52:54.122367 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zv97k\" (UniqueName: \"kubernetes.io/projected/b5257dc4-0faa-4e06-aeea-b25504581b7e-kube-api-access-zv97k\") pod \"marketplace-operator-79b997595-4kl8r\" (UID: \"b5257dc4-0faa-4e06-aeea-b25504581b7e\") " pod="openshift-marketplace/marketplace-operator-79b997595-4kl8r" Oct 05 06:52:54 crc kubenswrapper[4846]: I1005 06:52:54.122430 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b5257dc4-0faa-4e06-aeea-b25504581b7e-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-4kl8r\" (UID: \"b5257dc4-0faa-4e06-aeea-b25504581b7e\") " pod="openshift-marketplace/marketplace-operator-79b997595-4kl8r" Oct 05 06:52:54 crc kubenswrapper[4846]: I1005 06:52:54.122455 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b5257dc4-0faa-4e06-aeea-b25504581b7e-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-4kl8r\" (UID: \"b5257dc4-0faa-4e06-aeea-b25504581b7e\") " pod="openshift-marketplace/marketplace-operator-79b997595-4kl8r" Oct 05 06:52:54 crc kubenswrapper[4846]: I1005 06:52:54.123851 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b5257dc4-0faa-4e06-aeea-b25504581b7e-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-4kl8r\" (UID: \"b5257dc4-0faa-4e06-aeea-b25504581b7e\") " pod="openshift-marketplace/marketplace-operator-79b997595-4kl8r" Oct 05 06:52:54 crc kubenswrapper[4846]: I1005 06:52:54.136612 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b5257dc4-0faa-4e06-aeea-b25504581b7e-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-4kl8r\" (UID: \"b5257dc4-0faa-4e06-aeea-b25504581b7e\") " pod="openshift-marketplace/marketplace-operator-79b997595-4kl8r" Oct 05 06:52:54 crc kubenswrapper[4846]: I1005 06:52:54.146560 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zv97k\" (UniqueName: \"kubernetes.io/projected/b5257dc4-0faa-4e06-aeea-b25504581b7e-kube-api-access-zv97k\") pod \"marketplace-operator-79b997595-4kl8r\" (UID: \"b5257dc4-0faa-4e06-aeea-b25504581b7e\") " pod="openshift-marketplace/marketplace-operator-79b997595-4kl8r" Oct 05 06:52:54 crc kubenswrapper[4846]: I1005 06:52:54.296613 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-4kl8r" Oct 05 06:52:54 crc kubenswrapper[4846]: I1005 06:52:54.388703 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-2rc5n" Oct 05 06:52:54 crc kubenswrapper[4846]: I1005 06:52:54.389671 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mzc7c" Oct 05 06:52:54 crc kubenswrapper[4846]: I1005 06:52:54.398922 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-76jjh" Oct 05 06:52:54 crc kubenswrapper[4846]: I1005 06:52:54.402546 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dczxn" Oct 05 06:52:54 crc kubenswrapper[4846]: I1005 06:52:54.452703 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-knslg" Oct 05 06:52:54 crc kubenswrapper[4846]: I1005 06:52:54.528649 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-trr7q\" (UniqueName: \"kubernetes.io/projected/8c2f8349-aa93-4ac8-a369-5b4b2b9cd6e2-kube-api-access-trr7q\") pod \"8c2f8349-aa93-4ac8-a369-5b4b2b9cd6e2\" (UID: \"8c2f8349-aa93-4ac8-a369-5b4b2b9cd6e2\") " Oct 05 06:52:54 crc kubenswrapper[4846]: I1005 06:52:54.528729 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b6446bcc-5edd-46bd-a2dc-e8b4773f7248-utilities\") pod \"b6446bcc-5edd-46bd-a2dc-e8b4773f7248\" (UID: \"b6446bcc-5edd-46bd-a2dc-e8b4773f7248\") " Oct 05 06:52:54 crc kubenswrapper[4846]: I1005 06:52:54.528775 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/8c2f8349-aa93-4ac8-a369-5b4b2b9cd6e2-marketplace-operator-metrics\") pod \"8c2f8349-aa93-4ac8-a369-5b4b2b9cd6e2\" (UID: \"8c2f8349-aa93-4ac8-a369-5b4b2b9cd6e2\") " Oct 05 06:52:54 crc kubenswrapper[4846]: I1005 06:52:54.528793 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/da62c2a4-387f-4d5a-a9f2-04bd1e110d20-utilities\") pod \"da62c2a4-387f-4d5a-a9f2-04bd1e110d20\" (UID: \"da62c2a4-387f-4d5a-a9f2-04bd1e110d20\") " Oct 05 06:52:54 crc kubenswrapper[4846]: I1005 06:52:54.528811 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kjdm9\" (UniqueName: \"kubernetes.io/projected/73ca18b5-0a68-4990-9e7c-1685f87c8dcb-kube-api-access-kjdm9\") pod \"73ca18b5-0a68-4990-9e7c-1685f87c8dcb\" (UID: \"73ca18b5-0a68-4990-9e7c-1685f87c8dcb\") " Oct 05 06:52:54 crc kubenswrapper[4846]: I1005 06:52:54.528856 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/da62c2a4-387f-4d5a-a9f2-04bd1e110d20-catalog-content\") pod \"da62c2a4-387f-4d5a-a9f2-04bd1e110d20\" (UID: \"da62c2a4-387f-4d5a-a9f2-04bd1e110d20\") " Oct 05 06:52:54 crc kubenswrapper[4846]: I1005 06:52:54.528882 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hmh5t\" (UniqueName: \"kubernetes.io/projected/da62c2a4-387f-4d5a-a9f2-04bd1e110d20-kube-api-access-hmh5t\") pod \"da62c2a4-387f-4d5a-a9f2-04bd1e110d20\" (UID: \"da62c2a4-387f-4d5a-a9f2-04bd1e110d20\") " Oct 05 06:52:54 crc kubenswrapper[4846]: I1005 06:52:54.528898 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b6446bcc-5edd-46bd-a2dc-e8b4773f7248-catalog-content\") pod \"b6446bcc-5edd-46bd-a2dc-e8b4773f7248\" (UID: \"b6446bcc-5edd-46bd-a2dc-e8b4773f7248\") " Oct 05 06:52:54 crc kubenswrapper[4846]: I1005 06:52:54.528928 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8c2f8349-aa93-4ac8-a369-5b4b2b9cd6e2-marketplace-trusted-ca\") pod \"8c2f8349-aa93-4ac8-a369-5b4b2b9cd6e2\" (UID: \"8c2f8349-aa93-4ac8-a369-5b4b2b9cd6e2\") " Oct 05 06:52:54 crc kubenswrapper[4846]: I1005 06:52:54.528965 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-brxrf\" (UniqueName: \"kubernetes.io/projected/b6446bcc-5edd-46bd-a2dc-e8b4773f7248-kube-api-access-brxrf\") pod \"b6446bcc-5edd-46bd-a2dc-e8b4773f7248\" (UID: \"b6446bcc-5edd-46bd-a2dc-e8b4773f7248\") " Oct 05 06:52:54 crc kubenswrapper[4846]: I1005 06:52:54.528992 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/73ca18b5-0a68-4990-9e7c-1685f87c8dcb-utilities\") pod \"73ca18b5-0a68-4990-9e7c-1685f87c8dcb\" (UID: \"73ca18b5-0a68-4990-9e7c-1685f87c8dcb\") " Oct 05 06:52:54 crc kubenswrapper[4846]: I1005 06:52:54.529009 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/73ca18b5-0a68-4990-9e7c-1685f87c8dcb-catalog-content\") pod \"73ca18b5-0a68-4990-9e7c-1685f87c8dcb\" (UID: \"73ca18b5-0a68-4990-9e7c-1685f87c8dcb\") " Oct 05 06:52:54 crc kubenswrapper[4846]: I1005 06:52:54.533956 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8c2f8349-aa93-4ac8-a369-5b4b2b9cd6e2-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "8c2f8349-aa93-4ac8-a369-5b4b2b9cd6e2" (UID: "8c2f8349-aa93-4ac8-a369-5b4b2b9cd6e2"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 06:52:54 crc kubenswrapper[4846]: I1005 06:52:54.535860 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/da62c2a4-387f-4d5a-a9f2-04bd1e110d20-utilities" (OuterVolumeSpecName: "utilities") pod "da62c2a4-387f-4d5a-a9f2-04bd1e110d20" (UID: "da62c2a4-387f-4d5a-a9f2-04bd1e110d20"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 06:52:54 crc kubenswrapper[4846]: I1005 06:52:54.536577 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/73ca18b5-0a68-4990-9e7c-1685f87c8dcb-utilities" (OuterVolumeSpecName: "utilities") pod "73ca18b5-0a68-4990-9e7c-1685f87c8dcb" (UID: "73ca18b5-0a68-4990-9e7c-1685f87c8dcb"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 06:52:54 crc kubenswrapper[4846]: I1005 06:52:54.537109 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b6446bcc-5edd-46bd-a2dc-e8b4773f7248-utilities" (OuterVolumeSpecName: "utilities") pod "b6446bcc-5edd-46bd-a2dc-e8b4773f7248" (UID: "b6446bcc-5edd-46bd-a2dc-e8b4773f7248"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 06:52:54 crc kubenswrapper[4846]: I1005 06:52:54.538695 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/da62c2a4-387f-4d5a-a9f2-04bd1e110d20-kube-api-access-hmh5t" (OuterVolumeSpecName: "kube-api-access-hmh5t") pod "da62c2a4-387f-4d5a-a9f2-04bd1e110d20" (UID: "da62c2a4-387f-4d5a-a9f2-04bd1e110d20"). InnerVolumeSpecName "kube-api-access-hmh5t". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 06:52:54 crc kubenswrapper[4846]: I1005 06:52:54.539473 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8c2f8349-aa93-4ac8-a369-5b4b2b9cd6e2-kube-api-access-trr7q" (OuterVolumeSpecName: "kube-api-access-trr7q") pod "8c2f8349-aa93-4ac8-a369-5b4b2b9cd6e2" (UID: "8c2f8349-aa93-4ac8-a369-5b4b2b9cd6e2"). InnerVolumeSpecName "kube-api-access-trr7q". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 06:52:54 crc kubenswrapper[4846]: I1005 06:52:54.539902 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6446bcc-5edd-46bd-a2dc-e8b4773f7248-kube-api-access-brxrf" (OuterVolumeSpecName: "kube-api-access-brxrf") pod "b6446bcc-5edd-46bd-a2dc-e8b4773f7248" (UID: "b6446bcc-5edd-46bd-a2dc-e8b4773f7248"). InnerVolumeSpecName "kube-api-access-brxrf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 06:52:54 crc kubenswrapper[4846]: I1005 06:52:54.539967 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/73ca18b5-0a68-4990-9e7c-1685f87c8dcb-kube-api-access-kjdm9" (OuterVolumeSpecName: "kube-api-access-kjdm9") pod "73ca18b5-0a68-4990-9e7c-1685f87c8dcb" (UID: "73ca18b5-0a68-4990-9e7c-1685f87c8dcb"). InnerVolumeSpecName "kube-api-access-kjdm9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 06:52:54 crc kubenswrapper[4846]: I1005 06:52:54.540405 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8c2f8349-aa93-4ac8-a369-5b4b2b9cd6e2-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "8c2f8349-aa93-4ac8-a369-5b4b2b9cd6e2" (UID: "8c2f8349-aa93-4ac8-a369-5b4b2b9cd6e2"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 06:52:54 crc kubenswrapper[4846]: I1005 06:52:54.552166 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-4kl8r"] Oct 05 06:52:54 crc kubenswrapper[4846]: I1005 06:52:54.570914 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b6446bcc-5edd-46bd-a2dc-e8b4773f7248-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b6446bcc-5edd-46bd-a2dc-e8b4773f7248" (UID: "b6446bcc-5edd-46bd-a2dc-e8b4773f7248"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 06:52:54 crc kubenswrapper[4846]: I1005 06:52:54.592683 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/da62c2a4-387f-4d5a-a9f2-04bd1e110d20-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "da62c2a4-387f-4d5a-a9f2-04bd1e110d20" (UID: "da62c2a4-387f-4d5a-a9f2-04bd1e110d20"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 06:52:54 crc kubenswrapper[4846]: I1005 06:52:54.607396 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/73ca18b5-0a68-4990-9e7c-1685f87c8dcb-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "73ca18b5-0a68-4990-9e7c-1685f87c8dcb" (UID: "73ca18b5-0a68-4990-9e7c-1685f87c8dcb"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 06:52:54 crc kubenswrapper[4846]: I1005 06:52:54.630313 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0984e21c-5c51-4493-80b9-6e0fadf7537d-utilities\") pod \"0984e21c-5c51-4493-80b9-6e0fadf7537d\" (UID: \"0984e21c-5c51-4493-80b9-6e0fadf7537d\") " Oct 05 06:52:54 crc kubenswrapper[4846]: I1005 06:52:54.630544 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsk2\" (UniqueName: \"kubernetes.io/projected/0984e21c-5c51-4493-80b9-6e0fadf7537d-kube-api-access-fqsk2\") pod \"0984e21c-5c51-4493-80b9-6e0fadf7537d\" (UID: \"0984e21c-5c51-4493-80b9-6e0fadf7537d\") " Oct 05 06:52:54 crc kubenswrapper[4846]: I1005 06:52:54.630715 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0984e21c-5c51-4493-80b9-6e0fadf7537d-catalog-content\") pod \"0984e21c-5c51-4493-80b9-6e0fadf7537d\" (UID: \"0984e21c-5c51-4493-80b9-6e0fadf7537d\") " Oct 05 06:52:54 crc kubenswrapper[4846]: I1005 06:52:54.631103 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-brxrf\" (UniqueName: \"kubernetes.io/projected/b6446bcc-5edd-46bd-a2dc-e8b4773f7248-kube-api-access-brxrf\") on node \"crc\" DevicePath \"\"" Oct 05 06:52:54 crc kubenswrapper[4846]: I1005 06:52:54.631218 4846 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/73ca18b5-0a68-4990-9e7c-1685f87c8dcb-utilities\") on node \"crc\" DevicePath \"\"" Oct 05 06:52:54 crc kubenswrapper[4846]: I1005 06:52:54.631282 4846 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/73ca18b5-0a68-4990-9e7c-1685f87c8dcb-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 05 06:52:54 crc kubenswrapper[4846]: I1005 06:52:54.631352 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-trr7q\" (UniqueName: \"kubernetes.io/projected/8c2f8349-aa93-4ac8-a369-5b4b2b9cd6e2-kube-api-access-trr7q\") on node \"crc\" DevicePath \"\"" Oct 05 06:52:54 crc kubenswrapper[4846]: I1005 06:52:54.631422 4846 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b6446bcc-5edd-46bd-a2dc-e8b4773f7248-utilities\") on node \"crc\" DevicePath \"\"" Oct 05 06:52:54 crc kubenswrapper[4846]: I1005 06:52:54.631480 4846 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/8c2f8349-aa93-4ac8-a369-5b4b2b9cd6e2-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Oct 05 06:52:54 crc kubenswrapper[4846]: I1005 06:52:54.631536 4846 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/da62c2a4-387f-4d5a-a9f2-04bd1e110d20-utilities\") on node \"crc\" DevicePath \"\"" Oct 05 06:52:54 crc kubenswrapper[4846]: I1005 06:52:54.632125 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kjdm9\" (UniqueName: \"kubernetes.io/projected/73ca18b5-0a68-4990-9e7c-1685f87c8dcb-kube-api-access-kjdm9\") on node \"crc\" DevicePath \"\"" Oct 05 06:52:54 crc kubenswrapper[4846]: I1005 06:52:54.632257 4846 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/da62c2a4-387f-4d5a-a9f2-04bd1e110d20-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 05 06:52:54 crc kubenswrapper[4846]: I1005 06:52:54.632335 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hmh5t\" (UniqueName: \"kubernetes.io/projected/da62c2a4-387f-4d5a-a9f2-04bd1e110d20-kube-api-access-hmh5t\") on node \"crc\" DevicePath \"\"" Oct 05 06:52:54 crc kubenswrapper[4846]: I1005 06:52:54.632449 4846 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b6446bcc-5edd-46bd-a2dc-e8b4773f7248-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 05 06:52:54 crc kubenswrapper[4846]: I1005 06:52:54.632517 4846 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8c2f8349-aa93-4ac8-a369-5b4b2b9cd6e2-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 05 06:52:54 crc kubenswrapper[4846]: I1005 06:52:54.632386 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0984e21c-5c51-4493-80b9-6e0fadf7537d-utilities" (OuterVolumeSpecName: "utilities") pod "0984e21c-5c51-4493-80b9-6e0fadf7537d" (UID: "0984e21c-5c51-4493-80b9-6e0fadf7537d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 06:52:54 crc kubenswrapper[4846]: I1005 06:52:54.635622 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0984e21c-5c51-4493-80b9-6e0fadf7537d-kube-api-access-fqsk2" (OuterVolumeSpecName: "kube-api-access-fqsk2") pod "0984e21c-5c51-4493-80b9-6e0fadf7537d" (UID: "0984e21c-5c51-4493-80b9-6e0fadf7537d"). InnerVolumeSpecName "kube-api-access-fqsk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 06:52:54 crc kubenswrapper[4846]: I1005 06:52:54.723565 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0984e21c-5c51-4493-80b9-6e0fadf7537d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0984e21c-5c51-4493-80b9-6e0fadf7537d" (UID: "0984e21c-5c51-4493-80b9-6e0fadf7537d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 06:52:54 crc kubenswrapper[4846]: I1005 06:52:54.733700 4846 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0984e21c-5c51-4493-80b9-6e0fadf7537d-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 05 06:52:54 crc kubenswrapper[4846]: I1005 06:52:54.733732 4846 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0984e21c-5c51-4493-80b9-6e0fadf7537d-utilities\") on node \"crc\" DevicePath \"\"" Oct 05 06:52:54 crc kubenswrapper[4846]: I1005 06:52:54.733744 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsk2\" (UniqueName: \"kubernetes.io/projected/0984e21c-5c51-4493-80b9-6e0fadf7537d-kube-api-access-fqsk2\") on node \"crc\" DevicePath \"\"" Oct 05 06:52:54 crc kubenswrapper[4846]: I1005 06:52:54.792720 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-4kl8r" event={"ID":"b5257dc4-0faa-4e06-aeea-b25504581b7e","Type":"ContainerStarted","Data":"e1c19b3189a12affa673f73646e833172932e3e4f1cada26ab5bd9d6e4534ace"} Oct 05 06:52:54 crc kubenswrapper[4846]: I1005 06:52:54.792780 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-4kl8r" event={"ID":"b5257dc4-0faa-4e06-aeea-b25504581b7e","Type":"ContainerStarted","Data":"97f0d34e152efba68655d3fcc8e086047ce111aa56acf2a920d5503b9588eabd"} Oct 05 06:52:54 crc kubenswrapper[4846]: I1005 06:52:54.793644 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-4kl8r" Oct 05 06:52:54 crc kubenswrapper[4846]: I1005 06:52:54.795120 4846 generic.go:334] "Generic (PLEG): container finished" podID="73ca18b5-0a68-4990-9e7c-1685f87c8dcb" containerID="3e426226198ad8c4a4ed011f91f55edfcd595dc146977d4d9c4feb6e4eb28be7" exitCode=0 Oct 05 06:52:54 crc kubenswrapper[4846]: I1005 06:52:54.795159 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-76jjh" event={"ID":"73ca18b5-0a68-4990-9e7c-1685f87c8dcb","Type":"ContainerDied","Data":"3e426226198ad8c4a4ed011f91f55edfcd595dc146977d4d9c4feb6e4eb28be7"} Oct 05 06:52:54 crc kubenswrapper[4846]: I1005 06:52:54.795193 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-76jjh" event={"ID":"73ca18b5-0a68-4990-9e7c-1685f87c8dcb","Type":"ContainerDied","Data":"a88ea6505c818bdd1a13a8907c540d83ad51db690c5dd647ba5641f8fabc33ef"} Oct 05 06:52:54 crc kubenswrapper[4846]: I1005 06:52:54.795211 4846 scope.go:117] "RemoveContainer" containerID="3e426226198ad8c4a4ed011f91f55edfcd595dc146977d4d9c4feb6e4eb28be7" Oct 05 06:52:54 crc kubenswrapper[4846]: I1005 06:52:54.795311 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-76jjh" Oct 05 06:52:54 crc kubenswrapper[4846]: I1005 06:52:54.796150 4846 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-4kl8r container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.54:8080/healthz\": dial tcp 10.217.0.54:8080: connect: connection refused" start-of-body= Oct 05 06:52:54 crc kubenswrapper[4846]: I1005 06:52:54.796207 4846 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-4kl8r" podUID="b5257dc4-0faa-4e06-aeea-b25504581b7e" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.54:8080/healthz\": dial tcp 10.217.0.54:8080: connect: connection refused" Oct 05 06:52:54 crc kubenswrapper[4846]: I1005 06:52:54.799258 4846 generic.go:334] "Generic (PLEG): container finished" podID="0984e21c-5c51-4493-80b9-6e0fadf7537d" containerID="6ef045a66699331e787f02a0f98e95b9b5d68476955be9f1777ac3a1e593d959" exitCode=0 Oct 05 06:52:54 crc kubenswrapper[4846]: I1005 06:52:54.799335 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-knslg" Oct 05 06:52:54 crc kubenswrapper[4846]: I1005 06:52:54.799436 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-knslg" event={"ID":"0984e21c-5c51-4493-80b9-6e0fadf7537d","Type":"ContainerDied","Data":"6ef045a66699331e787f02a0f98e95b9b5d68476955be9f1777ac3a1e593d959"} Oct 05 06:52:54 crc kubenswrapper[4846]: I1005 06:52:54.799454 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-knslg" event={"ID":"0984e21c-5c51-4493-80b9-6e0fadf7537d","Type":"ContainerDied","Data":"cc6e538b18c6545419212022250380cd4afeb8f2212a05ada67319a9bc928a73"} Oct 05 06:52:54 crc kubenswrapper[4846]: I1005 06:52:54.801367 4846 generic.go:334] "Generic (PLEG): container finished" podID="da62c2a4-387f-4d5a-a9f2-04bd1e110d20" containerID="30cb1644498c9f7df6b72282eca59d86cf0d32592b4adb1858f5ef1453b4bc80" exitCode=0 Oct 05 06:52:54 crc kubenswrapper[4846]: I1005 06:52:54.801409 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mzc7c" event={"ID":"da62c2a4-387f-4d5a-a9f2-04bd1e110d20","Type":"ContainerDied","Data":"30cb1644498c9f7df6b72282eca59d86cf0d32592b4adb1858f5ef1453b4bc80"} Oct 05 06:52:54 crc kubenswrapper[4846]: I1005 06:52:54.801425 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mzc7c" event={"ID":"da62c2a4-387f-4d5a-a9f2-04bd1e110d20","Type":"ContainerDied","Data":"65593e4de55ad1d75bd3fc517aae979c3bb9241f569305d666ed25f257fa6902"} Oct 05 06:52:54 crc kubenswrapper[4846]: I1005 06:52:54.801471 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mzc7c" Oct 05 06:52:54 crc kubenswrapper[4846]: I1005 06:52:54.805871 4846 generic.go:334] "Generic (PLEG): container finished" podID="8c2f8349-aa93-4ac8-a369-5b4b2b9cd6e2" containerID="f702ca5fddb2d55c48f6f9d3494d5c98b925185899945a42f5693b66af2b7020" exitCode=0 Oct 05 06:52:54 crc kubenswrapper[4846]: I1005 06:52:54.805939 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-2rc5n" Oct 05 06:52:54 crc kubenswrapper[4846]: I1005 06:52:54.805998 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-2rc5n" event={"ID":"8c2f8349-aa93-4ac8-a369-5b4b2b9cd6e2","Type":"ContainerDied","Data":"f702ca5fddb2d55c48f6f9d3494d5c98b925185899945a42f5693b66af2b7020"} Oct 05 06:52:54 crc kubenswrapper[4846]: I1005 06:52:54.806031 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-2rc5n" event={"ID":"8c2f8349-aa93-4ac8-a369-5b4b2b9cd6e2","Type":"ContainerDied","Data":"b9d7ffdddc7410abfb4858bc3fd11d1af0497c11ac8e83b72de8664c8a6dab25"} Oct 05 06:52:54 crc kubenswrapper[4846]: I1005 06:52:54.808325 4846 generic.go:334] "Generic (PLEG): container finished" podID="b6446bcc-5edd-46bd-a2dc-e8b4773f7248" containerID="c90a269f8eaa15bb7b13ad740330f0936c8254ce793c00b727e5fbad6dd371f5" exitCode=0 Oct 05 06:52:54 crc kubenswrapper[4846]: I1005 06:52:54.808394 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dczxn" event={"ID":"b6446bcc-5edd-46bd-a2dc-e8b4773f7248","Type":"ContainerDied","Data":"c90a269f8eaa15bb7b13ad740330f0936c8254ce793c00b727e5fbad6dd371f5"} Oct 05 06:52:54 crc kubenswrapper[4846]: I1005 06:52:54.808432 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dczxn" event={"ID":"b6446bcc-5edd-46bd-a2dc-e8b4773f7248","Type":"ContainerDied","Data":"f3837f8f1c0c83c2d5e4933cd8760a5061a16e7b36865ebf44f231a16a8d163e"} Oct 05 06:52:54 crc kubenswrapper[4846]: I1005 06:52:54.808558 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dczxn" Oct 05 06:52:54 crc kubenswrapper[4846]: I1005 06:52:54.814455 4846 scope.go:117] "RemoveContainer" containerID="90b774a7005c4431444506a41a786f922e160117a560487bb571077a73443327" Oct 05 06:52:54 crc kubenswrapper[4846]: I1005 06:52:54.815346 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-4kl8r" podStartSLOduration=1.815335551 podStartE2EDuration="1.815335551s" podCreationTimestamp="2025-10-05 06:52:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 06:52:54.814986262 +0000 UTC m=+297.055839047" watchObservedRunningTime="2025-10-05 06:52:54.815335551 +0000 UTC m=+297.056188316" Oct 05 06:52:54 crc kubenswrapper[4846]: I1005 06:52:54.846425 4846 scope.go:117] "RemoveContainer" containerID="5743c92de46b094ed9889d55ed47b2a35abcb849affca8436f485549017a93eb" Oct 05 06:52:54 crc kubenswrapper[4846]: I1005 06:52:54.853803 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-knslg"] Oct 05 06:52:54 crc kubenswrapper[4846]: I1005 06:52:54.858048 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-knslg"] Oct 05 06:52:54 crc kubenswrapper[4846]: I1005 06:52:54.861403 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-76jjh"] Oct 05 06:52:54 crc kubenswrapper[4846]: I1005 06:52:54.868098 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-76jjh"] Oct 05 06:52:54 crc kubenswrapper[4846]: I1005 06:52:54.873580 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-2rc5n"] Oct 05 06:52:54 crc kubenswrapper[4846]: I1005 06:52:54.878446 4846 scope.go:117] "RemoveContainer" containerID="3e426226198ad8c4a4ed011f91f55edfcd595dc146977d4d9c4feb6e4eb28be7" Oct 05 06:52:54 crc kubenswrapper[4846]: E1005 06:52:54.879223 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3e426226198ad8c4a4ed011f91f55edfcd595dc146977d4d9c4feb6e4eb28be7\": container with ID starting with 3e426226198ad8c4a4ed011f91f55edfcd595dc146977d4d9c4feb6e4eb28be7 not found: ID does not exist" containerID="3e426226198ad8c4a4ed011f91f55edfcd595dc146977d4d9c4feb6e4eb28be7" Oct 05 06:52:54 crc kubenswrapper[4846]: I1005 06:52:54.879266 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3e426226198ad8c4a4ed011f91f55edfcd595dc146977d4d9c4feb6e4eb28be7"} err="failed to get container status \"3e426226198ad8c4a4ed011f91f55edfcd595dc146977d4d9c4feb6e4eb28be7\": rpc error: code = NotFound desc = could not find container \"3e426226198ad8c4a4ed011f91f55edfcd595dc146977d4d9c4feb6e4eb28be7\": container with ID starting with 3e426226198ad8c4a4ed011f91f55edfcd595dc146977d4d9c4feb6e4eb28be7 not found: ID does not exist" Oct 05 06:52:54 crc kubenswrapper[4846]: I1005 06:52:54.879299 4846 scope.go:117] "RemoveContainer" containerID="90b774a7005c4431444506a41a786f922e160117a560487bb571077a73443327" Oct 05 06:52:54 crc kubenswrapper[4846]: E1005 06:52:54.879650 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"90b774a7005c4431444506a41a786f922e160117a560487bb571077a73443327\": container with ID starting with 90b774a7005c4431444506a41a786f922e160117a560487bb571077a73443327 not found: ID does not exist" containerID="90b774a7005c4431444506a41a786f922e160117a560487bb571077a73443327" Oct 05 06:52:54 crc kubenswrapper[4846]: I1005 06:52:54.879697 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"90b774a7005c4431444506a41a786f922e160117a560487bb571077a73443327"} err="failed to get container status \"90b774a7005c4431444506a41a786f922e160117a560487bb571077a73443327\": rpc error: code = NotFound desc = could not find container \"90b774a7005c4431444506a41a786f922e160117a560487bb571077a73443327\": container with ID starting with 90b774a7005c4431444506a41a786f922e160117a560487bb571077a73443327 not found: ID does not exist" Oct 05 06:52:54 crc kubenswrapper[4846]: I1005 06:52:54.879727 4846 scope.go:117] "RemoveContainer" containerID="5743c92de46b094ed9889d55ed47b2a35abcb849affca8436f485549017a93eb" Oct 05 06:52:54 crc kubenswrapper[4846]: I1005 06:52:54.879723 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-2rc5n"] Oct 05 06:52:54 crc kubenswrapper[4846]: E1005 06:52:54.880150 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5743c92de46b094ed9889d55ed47b2a35abcb849affca8436f485549017a93eb\": container with ID starting with 5743c92de46b094ed9889d55ed47b2a35abcb849affca8436f485549017a93eb not found: ID does not exist" containerID="5743c92de46b094ed9889d55ed47b2a35abcb849affca8436f485549017a93eb" Oct 05 06:52:54 crc kubenswrapper[4846]: I1005 06:52:54.880196 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5743c92de46b094ed9889d55ed47b2a35abcb849affca8436f485549017a93eb"} err="failed to get container status \"5743c92de46b094ed9889d55ed47b2a35abcb849affca8436f485549017a93eb\": rpc error: code = NotFound desc = could not find container \"5743c92de46b094ed9889d55ed47b2a35abcb849affca8436f485549017a93eb\": container with ID starting with 5743c92de46b094ed9889d55ed47b2a35abcb849affca8436f485549017a93eb not found: ID does not exist" Oct 05 06:52:54 crc kubenswrapper[4846]: I1005 06:52:54.880219 4846 scope.go:117] "RemoveContainer" containerID="6ef045a66699331e787f02a0f98e95b9b5d68476955be9f1777ac3a1e593d959" Oct 05 06:52:54 crc kubenswrapper[4846]: I1005 06:52:54.886877 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-dczxn"] Oct 05 06:52:54 crc kubenswrapper[4846]: I1005 06:52:54.903331 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-dczxn"] Oct 05 06:52:54 crc kubenswrapper[4846]: I1005 06:52:54.910484 4846 scope.go:117] "RemoveContainer" containerID="8764d0318b784edf4871c6b83bf36b77a3a9539dd3aa682ebd544172e9dd882d" Oct 05 06:52:54 crc kubenswrapper[4846]: I1005 06:52:54.911819 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-mzc7c"] Oct 05 06:52:54 crc kubenswrapper[4846]: I1005 06:52:54.920710 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-mzc7c"] Oct 05 06:52:54 crc kubenswrapper[4846]: I1005 06:52:54.936217 4846 scope.go:117] "RemoveContainer" containerID="0ff82a1bd96990b795985d4afdcb71edac3ab2c578e28afa371c340f30686a3d" Oct 05 06:52:54 crc kubenswrapper[4846]: I1005 06:52:54.960564 4846 scope.go:117] "RemoveContainer" containerID="6ef045a66699331e787f02a0f98e95b9b5d68476955be9f1777ac3a1e593d959" Oct 05 06:52:54 crc kubenswrapper[4846]: E1005 06:52:54.961267 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6ef045a66699331e787f02a0f98e95b9b5d68476955be9f1777ac3a1e593d959\": container with ID starting with 6ef045a66699331e787f02a0f98e95b9b5d68476955be9f1777ac3a1e593d959 not found: ID does not exist" containerID="6ef045a66699331e787f02a0f98e95b9b5d68476955be9f1777ac3a1e593d959" Oct 05 06:52:54 crc kubenswrapper[4846]: I1005 06:52:54.961336 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6ef045a66699331e787f02a0f98e95b9b5d68476955be9f1777ac3a1e593d959"} err="failed to get container status \"6ef045a66699331e787f02a0f98e95b9b5d68476955be9f1777ac3a1e593d959\": rpc error: code = NotFound desc = could not find container \"6ef045a66699331e787f02a0f98e95b9b5d68476955be9f1777ac3a1e593d959\": container with ID starting with 6ef045a66699331e787f02a0f98e95b9b5d68476955be9f1777ac3a1e593d959 not found: ID does not exist" Oct 05 06:52:54 crc kubenswrapper[4846]: I1005 06:52:54.961382 4846 scope.go:117] "RemoveContainer" containerID="8764d0318b784edf4871c6b83bf36b77a3a9539dd3aa682ebd544172e9dd882d" Oct 05 06:52:54 crc kubenswrapper[4846]: E1005 06:52:54.961921 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8764d0318b784edf4871c6b83bf36b77a3a9539dd3aa682ebd544172e9dd882d\": container with ID starting with 8764d0318b784edf4871c6b83bf36b77a3a9539dd3aa682ebd544172e9dd882d not found: ID does not exist" containerID="8764d0318b784edf4871c6b83bf36b77a3a9539dd3aa682ebd544172e9dd882d" Oct 05 06:52:54 crc kubenswrapper[4846]: I1005 06:52:54.961988 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8764d0318b784edf4871c6b83bf36b77a3a9539dd3aa682ebd544172e9dd882d"} err="failed to get container status \"8764d0318b784edf4871c6b83bf36b77a3a9539dd3aa682ebd544172e9dd882d\": rpc error: code = NotFound desc = could not find container \"8764d0318b784edf4871c6b83bf36b77a3a9539dd3aa682ebd544172e9dd882d\": container with ID starting with 8764d0318b784edf4871c6b83bf36b77a3a9539dd3aa682ebd544172e9dd882d not found: ID does not exist" Oct 05 06:52:54 crc kubenswrapper[4846]: I1005 06:52:54.962029 4846 scope.go:117] "RemoveContainer" containerID="0ff82a1bd96990b795985d4afdcb71edac3ab2c578e28afa371c340f30686a3d" Oct 05 06:52:54 crc kubenswrapper[4846]: E1005 06:52:54.964214 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0ff82a1bd96990b795985d4afdcb71edac3ab2c578e28afa371c340f30686a3d\": container with ID starting with 0ff82a1bd96990b795985d4afdcb71edac3ab2c578e28afa371c340f30686a3d not found: ID does not exist" containerID="0ff82a1bd96990b795985d4afdcb71edac3ab2c578e28afa371c340f30686a3d" Oct 05 06:52:54 crc kubenswrapper[4846]: I1005 06:52:54.964341 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0ff82a1bd96990b795985d4afdcb71edac3ab2c578e28afa371c340f30686a3d"} err="failed to get container status \"0ff82a1bd96990b795985d4afdcb71edac3ab2c578e28afa371c340f30686a3d\": rpc error: code = NotFound desc = could not find container \"0ff82a1bd96990b795985d4afdcb71edac3ab2c578e28afa371c340f30686a3d\": container with ID starting with 0ff82a1bd96990b795985d4afdcb71edac3ab2c578e28afa371c340f30686a3d not found: ID does not exist" Oct 05 06:52:54 crc kubenswrapper[4846]: I1005 06:52:54.964381 4846 scope.go:117] "RemoveContainer" containerID="30cb1644498c9f7df6b72282eca59d86cf0d32592b4adb1858f5ef1453b4bc80" Oct 05 06:52:54 crc kubenswrapper[4846]: I1005 06:52:54.984029 4846 scope.go:117] "RemoveContainer" containerID="1bd92726c1452bf8f513f6467bd996e359d2e8eaf05ccdff4e933df05df839ec" Oct 05 06:52:55 crc kubenswrapper[4846]: I1005 06:52:55.009352 4846 scope.go:117] "RemoveContainer" containerID="33ceecc33b378346cadfa69cd4c623a07f6709f95d9e297d39cf73e1f5800c82" Oct 05 06:52:55 crc kubenswrapper[4846]: I1005 06:52:55.023284 4846 scope.go:117] "RemoveContainer" containerID="30cb1644498c9f7df6b72282eca59d86cf0d32592b4adb1858f5ef1453b4bc80" Oct 05 06:52:55 crc kubenswrapper[4846]: E1005 06:52:55.023740 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"30cb1644498c9f7df6b72282eca59d86cf0d32592b4adb1858f5ef1453b4bc80\": container with ID starting with 30cb1644498c9f7df6b72282eca59d86cf0d32592b4adb1858f5ef1453b4bc80 not found: ID does not exist" containerID="30cb1644498c9f7df6b72282eca59d86cf0d32592b4adb1858f5ef1453b4bc80" Oct 05 06:52:55 crc kubenswrapper[4846]: I1005 06:52:55.023770 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"30cb1644498c9f7df6b72282eca59d86cf0d32592b4adb1858f5ef1453b4bc80"} err="failed to get container status \"30cb1644498c9f7df6b72282eca59d86cf0d32592b4adb1858f5ef1453b4bc80\": rpc error: code = NotFound desc = could not find container \"30cb1644498c9f7df6b72282eca59d86cf0d32592b4adb1858f5ef1453b4bc80\": container with ID starting with 30cb1644498c9f7df6b72282eca59d86cf0d32592b4adb1858f5ef1453b4bc80 not found: ID does not exist" Oct 05 06:52:55 crc kubenswrapper[4846]: I1005 06:52:55.023838 4846 scope.go:117] "RemoveContainer" containerID="1bd92726c1452bf8f513f6467bd996e359d2e8eaf05ccdff4e933df05df839ec" Oct 05 06:52:55 crc kubenswrapper[4846]: E1005 06:52:55.024428 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1bd92726c1452bf8f513f6467bd996e359d2e8eaf05ccdff4e933df05df839ec\": container with ID starting with 1bd92726c1452bf8f513f6467bd996e359d2e8eaf05ccdff4e933df05df839ec not found: ID does not exist" containerID="1bd92726c1452bf8f513f6467bd996e359d2e8eaf05ccdff4e933df05df839ec" Oct 05 06:52:55 crc kubenswrapper[4846]: I1005 06:52:55.024446 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1bd92726c1452bf8f513f6467bd996e359d2e8eaf05ccdff4e933df05df839ec"} err="failed to get container status \"1bd92726c1452bf8f513f6467bd996e359d2e8eaf05ccdff4e933df05df839ec\": rpc error: code = NotFound desc = could not find container \"1bd92726c1452bf8f513f6467bd996e359d2e8eaf05ccdff4e933df05df839ec\": container with ID starting with 1bd92726c1452bf8f513f6467bd996e359d2e8eaf05ccdff4e933df05df839ec not found: ID does not exist" Oct 05 06:52:55 crc kubenswrapper[4846]: I1005 06:52:55.024460 4846 scope.go:117] "RemoveContainer" containerID="33ceecc33b378346cadfa69cd4c623a07f6709f95d9e297d39cf73e1f5800c82" Oct 05 06:52:55 crc kubenswrapper[4846]: E1005 06:52:55.025261 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"33ceecc33b378346cadfa69cd4c623a07f6709f95d9e297d39cf73e1f5800c82\": container with ID starting with 33ceecc33b378346cadfa69cd4c623a07f6709f95d9e297d39cf73e1f5800c82 not found: ID does not exist" containerID="33ceecc33b378346cadfa69cd4c623a07f6709f95d9e297d39cf73e1f5800c82" Oct 05 06:52:55 crc kubenswrapper[4846]: I1005 06:52:55.025286 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"33ceecc33b378346cadfa69cd4c623a07f6709f95d9e297d39cf73e1f5800c82"} err="failed to get container status \"33ceecc33b378346cadfa69cd4c623a07f6709f95d9e297d39cf73e1f5800c82\": rpc error: code = NotFound desc = could not find container \"33ceecc33b378346cadfa69cd4c623a07f6709f95d9e297d39cf73e1f5800c82\": container with ID starting with 33ceecc33b378346cadfa69cd4c623a07f6709f95d9e297d39cf73e1f5800c82 not found: ID does not exist" Oct 05 06:52:55 crc kubenswrapper[4846]: I1005 06:52:55.025302 4846 scope.go:117] "RemoveContainer" containerID="f702ca5fddb2d55c48f6f9d3494d5c98b925185899945a42f5693b66af2b7020" Oct 05 06:52:55 crc kubenswrapper[4846]: I1005 06:52:55.106397 4846 scope.go:117] "RemoveContainer" containerID="f702ca5fddb2d55c48f6f9d3494d5c98b925185899945a42f5693b66af2b7020" Oct 05 06:52:55 crc kubenswrapper[4846]: E1005 06:52:55.106953 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f702ca5fddb2d55c48f6f9d3494d5c98b925185899945a42f5693b66af2b7020\": container with ID starting with f702ca5fddb2d55c48f6f9d3494d5c98b925185899945a42f5693b66af2b7020 not found: ID does not exist" containerID="f702ca5fddb2d55c48f6f9d3494d5c98b925185899945a42f5693b66af2b7020" Oct 05 06:52:55 crc kubenswrapper[4846]: I1005 06:52:55.107005 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f702ca5fddb2d55c48f6f9d3494d5c98b925185899945a42f5693b66af2b7020"} err="failed to get container status \"f702ca5fddb2d55c48f6f9d3494d5c98b925185899945a42f5693b66af2b7020\": rpc error: code = NotFound desc = could not find container \"f702ca5fddb2d55c48f6f9d3494d5c98b925185899945a42f5693b66af2b7020\": container with ID starting with f702ca5fddb2d55c48f6f9d3494d5c98b925185899945a42f5693b66af2b7020 not found: ID does not exist" Oct 05 06:52:55 crc kubenswrapper[4846]: I1005 06:52:55.107040 4846 scope.go:117] "RemoveContainer" containerID="c90a269f8eaa15bb7b13ad740330f0936c8254ce793c00b727e5fbad6dd371f5" Oct 05 06:52:55 crc kubenswrapper[4846]: I1005 06:52:55.122577 4846 scope.go:117] "RemoveContainer" containerID="39e16332d58703e73090ebf4dcff7a2dbb1e92e1320890b62c8d0c1391f81b4f" Oct 05 06:52:55 crc kubenswrapper[4846]: I1005 06:52:55.148336 4846 scope.go:117] "RemoveContainer" containerID="7c900c6109b9cf97634b9dae519a95243a4b08eb406782cac904939033b704ca" Oct 05 06:52:55 crc kubenswrapper[4846]: I1005 06:52:55.159171 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-6t2mt"] Oct 05 06:52:55 crc kubenswrapper[4846]: I1005 06:52:55.168665 4846 scope.go:117] "RemoveContainer" containerID="c90a269f8eaa15bb7b13ad740330f0936c8254ce793c00b727e5fbad6dd371f5" Oct 05 06:52:55 crc kubenswrapper[4846]: E1005 06:52:55.171450 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c90a269f8eaa15bb7b13ad740330f0936c8254ce793c00b727e5fbad6dd371f5\": container with ID starting with c90a269f8eaa15bb7b13ad740330f0936c8254ce793c00b727e5fbad6dd371f5 not found: ID does not exist" containerID="c90a269f8eaa15bb7b13ad740330f0936c8254ce793c00b727e5fbad6dd371f5" Oct 05 06:52:55 crc kubenswrapper[4846]: I1005 06:52:55.171512 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c90a269f8eaa15bb7b13ad740330f0936c8254ce793c00b727e5fbad6dd371f5"} err="failed to get container status \"c90a269f8eaa15bb7b13ad740330f0936c8254ce793c00b727e5fbad6dd371f5\": rpc error: code = NotFound desc = could not find container \"c90a269f8eaa15bb7b13ad740330f0936c8254ce793c00b727e5fbad6dd371f5\": container with ID starting with c90a269f8eaa15bb7b13ad740330f0936c8254ce793c00b727e5fbad6dd371f5 not found: ID does not exist" Oct 05 06:52:55 crc kubenswrapper[4846]: I1005 06:52:55.171548 4846 scope.go:117] "RemoveContainer" containerID="39e16332d58703e73090ebf4dcff7a2dbb1e92e1320890b62c8d0c1391f81b4f" Oct 05 06:52:55 crc kubenswrapper[4846]: E1005 06:52:55.172273 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"39e16332d58703e73090ebf4dcff7a2dbb1e92e1320890b62c8d0c1391f81b4f\": container with ID starting with 39e16332d58703e73090ebf4dcff7a2dbb1e92e1320890b62c8d0c1391f81b4f not found: ID does not exist" containerID="39e16332d58703e73090ebf4dcff7a2dbb1e92e1320890b62c8d0c1391f81b4f" Oct 05 06:52:55 crc kubenswrapper[4846]: I1005 06:52:55.172305 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"39e16332d58703e73090ebf4dcff7a2dbb1e92e1320890b62c8d0c1391f81b4f"} err="failed to get container status \"39e16332d58703e73090ebf4dcff7a2dbb1e92e1320890b62c8d0c1391f81b4f\": rpc error: code = NotFound desc = could not find container \"39e16332d58703e73090ebf4dcff7a2dbb1e92e1320890b62c8d0c1391f81b4f\": container with ID starting with 39e16332d58703e73090ebf4dcff7a2dbb1e92e1320890b62c8d0c1391f81b4f not found: ID does not exist" Oct 05 06:52:55 crc kubenswrapper[4846]: I1005 06:52:55.172324 4846 scope.go:117] "RemoveContainer" containerID="7c900c6109b9cf97634b9dae519a95243a4b08eb406782cac904939033b704ca" Oct 05 06:52:55 crc kubenswrapper[4846]: E1005 06:52:55.172581 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7c900c6109b9cf97634b9dae519a95243a4b08eb406782cac904939033b704ca\": container with ID starting with 7c900c6109b9cf97634b9dae519a95243a4b08eb406782cac904939033b704ca not found: ID does not exist" containerID="7c900c6109b9cf97634b9dae519a95243a4b08eb406782cac904939033b704ca" Oct 05 06:52:55 crc kubenswrapper[4846]: I1005 06:52:55.172602 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7c900c6109b9cf97634b9dae519a95243a4b08eb406782cac904939033b704ca"} err="failed to get container status \"7c900c6109b9cf97634b9dae519a95243a4b08eb406782cac904939033b704ca\": rpc error: code = NotFound desc = could not find container \"7c900c6109b9cf97634b9dae519a95243a4b08eb406782cac904939033b704ca\": container with ID starting with 7c900c6109b9cf97634b9dae519a95243a4b08eb406782cac904939033b704ca not found: ID does not exist" Oct 05 06:52:55 crc kubenswrapper[4846]: I1005 06:52:55.827265 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-4kl8r" Oct 05 06:52:56 crc kubenswrapper[4846]: I1005 06:52:56.129055 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-mkttk"] Oct 05 06:52:56 crc kubenswrapper[4846]: E1005 06:52:56.129302 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="73ca18b5-0a68-4990-9e7c-1685f87c8dcb" containerName="registry-server" Oct 05 06:52:56 crc kubenswrapper[4846]: I1005 06:52:56.129321 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="73ca18b5-0a68-4990-9e7c-1685f87c8dcb" containerName="registry-server" Oct 05 06:52:56 crc kubenswrapper[4846]: E1005 06:52:56.129339 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8c2f8349-aa93-4ac8-a369-5b4b2b9cd6e2" containerName="marketplace-operator" Oct 05 06:52:56 crc kubenswrapper[4846]: I1005 06:52:56.129349 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="8c2f8349-aa93-4ac8-a369-5b4b2b9cd6e2" containerName="marketplace-operator" Oct 05 06:52:56 crc kubenswrapper[4846]: E1005 06:52:56.129359 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0984e21c-5c51-4493-80b9-6e0fadf7537d" containerName="extract-content" Oct 05 06:52:56 crc kubenswrapper[4846]: I1005 06:52:56.129367 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="0984e21c-5c51-4493-80b9-6e0fadf7537d" containerName="extract-content" Oct 05 06:52:56 crc kubenswrapper[4846]: E1005 06:52:56.129384 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b6446bcc-5edd-46bd-a2dc-e8b4773f7248" containerName="registry-server" Oct 05 06:52:56 crc kubenswrapper[4846]: I1005 06:52:56.129392 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="b6446bcc-5edd-46bd-a2dc-e8b4773f7248" containerName="registry-server" Oct 05 06:52:56 crc kubenswrapper[4846]: E1005 06:52:56.129401 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0984e21c-5c51-4493-80b9-6e0fadf7537d" containerName="registry-server" Oct 05 06:52:56 crc kubenswrapper[4846]: I1005 06:52:56.129409 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="0984e21c-5c51-4493-80b9-6e0fadf7537d" containerName="registry-server" Oct 05 06:52:56 crc kubenswrapper[4846]: E1005 06:52:56.129423 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="da62c2a4-387f-4d5a-a9f2-04bd1e110d20" containerName="registry-server" Oct 05 06:52:56 crc kubenswrapper[4846]: I1005 06:52:56.129430 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="da62c2a4-387f-4d5a-a9f2-04bd1e110d20" containerName="registry-server" Oct 05 06:52:56 crc kubenswrapper[4846]: E1005 06:52:56.129443 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="73ca18b5-0a68-4990-9e7c-1685f87c8dcb" containerName="extract-utilities" Oct 05 06:52:56 crc kubenswrapper[4846]: I1005 06:52:56.129451 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="73ca18b5-0a68-4990-9e7c-1685f87c8dcb" containerName="extract-utilities" Oct 05 06:52:56 crc kubenswrapper[4846]: E1005 06:52:56.129459 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="73ca18b5-0a68-4990-9e7c-1685f87c8dcb" containerName="extract-content" Oct 05 06:52:56 crc kubenswrapper[4846]: I1005 06:52:56.129467 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="73ca18b5-0a68-4990-9e7c-1685f87c8dcb" containerName="extract-content" Oct 05 06:52:56 crc kubenswrapper[4846]: E1005 06:52:56.129478 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b6446bcc-5edd-46bd-a2dc-e8b4773f7248" containerName="extract-utilities" Oct 05 06:52:56 crc kubenswrapper[4846]: I1005 06:52:56.129485 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="b6446bcc-5edd-46bd-a2dc-e8b4773f7248" containerName="extract-utilities" Oct 05 06:52:56 crc kubenswrapper[4846]: E1005 06:52:56.129496 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="da62c2a4-387f-4d5a-a9f2-04bd1e110d20" containerName="extract-utilities" Oct 05 06:52:56 crc kubenswrapper[4846]: I1005 06:52:56.129504 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="da62c2a4-387f-4d5a-a9f2-04bd1e110d20" containerName="extract-utilities" Oct 05 06:52:56 crc kubenswrapper[4846]: E1005 06:52:56.129516 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="da62c2a4-387f-4d5a-a9f2-04bd1e110d20" containerName="extract-content" Oct 05 06:52:56 crc kubenswrapper[4846]: I1005 06:52:56.129524 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="da62c2a4-387f-4d5a-a9f2-04bd1e110d20" containerName="extract-content" Oct 05 06:52:56 crc kubenswrapper[4846]: E1005 06:52:56.129532 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b6446bcc-5edd-46bd-a2dc-e8b4773f7248" containerName="extract-content" Oct 05 06:52:56 crc kubenswrapper[4846]: I1005 06:52:56.129541 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="b6446bcc-5edd-46bd-a2dc-e8b4773f7248" containerName="extract-content" Oct 05 06:52:56 crc kubenswrapper[4846]: E1005 06:52:56.129552 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0984e21c-5c51-4493-80b9-6e0fadf7537d" containerName="extract-utilities" Oct 05 06:52:56 crc kubenswrapper[4846]: I1005 06:52:56.129560 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="0984e21c-5c51-4493-80b9-6e0fadf7537d" containerName="extract-utilities" Oct 05 06:52:56 crc kubenswrapper[4846]: I1005 06:52:56.129667 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="73ca18b5-0a68-4990-9e7c-1685f87c8dcb" containerName="registry-server" Oct 05 06:52:56 crc kubenswrapper[4846]: I1005 06:52:56.129682 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="0984e21c-5c51-4493-80b9-6e0fadf7537d" containerName="registry-server" Oct 05 06:52:56 crc kubenswrapper[4846]: I1005 06:52:56.129701 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="da62c2a4-387f-4d5a-a9f2-04bd1e110d20" containerName="registry-server" Oct 05 06:52:56 crc kubenswrapper[4846]: I1005 06:52:56.129713 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="8c2f8349-aa93-4ac8-a369-5b4b2b9cd6e2" containerName="marketplace-operator" Oct 05 06:52:56 crc kubenswrapper[4846]: I1005 06:52:56.129724 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="b6446bcc-5edd-46bd-a2dc-e8b4773f7248" containerName="registry-server" Oct 05 06:52:56 crc kubenswrapper[4846]: I1005 06:52:56.130599 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mkttk" Oct 05 06:52:56 crc kubenswrapper[4846]: I1005 06:52:56.134707 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Oct 05 06:52:56 crc kubenswrapper[4846]: I1005 06:52:56.149346 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-mkttk"] Oct 05 06:52:56 crc kubenswrapper[4846]: I1005 06:52:56.256644 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/01709201-a99b-4e92-adc8-868c6d272ac8-utilities\") pod \"redhat-marketplace-mkttk\" (UID: \"01709201-a99b-4e92-adc8-868c6d272ac8\") " pod="openshift-marketplace/redhat-marketplace-mkttk" Oct 05 06:52:56 crc kubenswrapper[4846]: I1005 06:52:56.256700 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-brp9d\" (UniqueName: \"kubernetes.io/projected/01709201-a99b-4e92-adc8-868c6d272ac8-kube-api-access-brp9d\") pod \"redhat-marketplace-mkttk\" (UID: \"01709201-a99b-4e92-adc8-868c6d272ac8\") " pod="openshift-marketplace/redhat-marketplace-mkttk" Oct 05 06:52:56 crc kubenswrapper[4846]: I1005 06:52:56.256750 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/01709201-a99b-4e92-adc8-868c6d272ac8-catalog-content\") pod \"redhat-marketplace-mkttk\" (UID: \"01709201-a99b-4e92-adc8-868c6d272ac8\") " pod="openshift-marketplace/redhat-marketplace-mkttk" Oct 05 06:52:56 crc kubenswrapper[4846]: I1005 06:52:56.331728 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-7fskx"] Oct 05 06:52:56 crc kubenswrapper[4846]: I1005 06:52:56.332919 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7fskx" Oct 05 06:52:56 crc kubenswrapper[4846]: I1005 06:52:56.338362 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Oct 05 06:52:56 crc kubenswrapper[4846]: I1005 06:52:56.346406 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-7fskx"] Oct 05 06:52:56 crc kubenswrapper[4846]: I1005 06:52:56.441595 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/04f12f2c-eb1a-4b2a-a5d4-bb930a6e278f-utilities\") pod \"certified-operators-7fskx\" (UID: \"04f12f2c-eb1a-4b2a-a5d4-bb930a6e278f\") " pod="openshift-marketplace/certified-operators-7fskx" Oct 05 06:52:56 crc kubenswrapper[4846]: I1005 06:52:56.441655 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/04f12f2c-eb1a-4b2a-a5d4-bb930a6e278f-catalog-content\") pod \"certified-operators-7fskx\" (UID: \"04f12f2c-eb1a-4b2a-a5d4-bb930a6e278f\") " pod="openshift-marketplace/certified-operators-7fskx" Oct 05 06:52:56 crc kubenswrapper[4846]: I1005 06:52:56.441733 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/01709201-a99b-4e92-adc8-868c6d272ac8-utilities\") pod \"redhat-marketplace-mkttk\" (UID: \"01709201-a99b-4e92-adc8-868c6d272ac8\") " pod="openshift-marketplace/redhat-marketplace-mkttk" Oct 05 06:52:56 crc kubenswrapper[4846]: I1005 06:52:56.441761 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-brp9d\" (UniqueName: \"kubernetes.io/projected/01709201-a99b-4e92-adc8-868c6d272ac8-kube-api-access-brp9d\") pod \"redhat-marketplace-mkttk\" (UID: \"01709201-a99b-4e92-adc8-868c6d272ac8\") " pod="openshift-marketplace/redhat-marketplace-mkttk" Oct 05 06:52:56 crc kubenswrapper[4846]: I1005 06:52:56.441807 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/01709201-a99b-4e92-adc8-868c6d272ac8-catalog-content\") pod \"redhat-marketplace-mkttk\" (UID: \"01709201-a99b-4e92-adc8-868c6d272ac8\") " pod="openshift-marketplace/redhat-marketplace-mkttk" Oct 05 06:52:56 crc kubenswrapper[4846]: I1005 06:52:56.441827 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m6phk\" (UniqueName: \"kubernetes.io/projected/04f12f2c-eb1a-4b2a-a5d4-bb930a6e278f-kube-api-access-m6phk\") pod \"certified-operators-7fskx\" (UID: \"04f12f2c-eb1a-4b2a-a5d4-bb930a6e278f\") " pod="openshift-marketplace/certified-operators-7fskx" Oct 05 06:52:56 crc kubenswrapper[4846]: I1005 06:52:56.442330 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/01709201-a99b-4e92-adc8-868c6d272ac8-utilities\") pod \"redhat-marketplace-mkttk\" (UID: \"01709201-a99b-4e92-adc8-868c6d272ac8\") " pod="openshift-marketplace/redhat-marketplace-mkttk" Oct 05 06:52:56 crc kubenswrapper[4846]: I1005 06:52:56.442670 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/01709201-a99b-4e92-adc8-868c6d272ac8-catalog-content\") pod \"redhat-marketplace-mkttk\" (UID: \"01709201-a99b-4e92-adc8-868c6d272ac8\") " pod="openshift-marketplace/redhat-marketplace-mkttk" Oct 05 06:52:56 crc kubenswrapper[4846]: I1005 06:52:56.464831 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-brp9d\" (UniqueName: \"kubernetes.io/projected/01709201-a99b-4e92-adc8-868c6d272ac8-kube-api-access-brp9d\") pod \"redhat-marketplace-mkttk\" (UID: \"01709201-a99b-4e92-adc8-868c6d272ac8\") " pod="openshift-marketplace/redhat-marketplace-mkttk" Oct 05 06:52:56 crc kubenswrapper[4846]: I1005 06:52:56.506412 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0984e21c-5c51-4493-80b9-6e0fadf7537d" path="/var/lib/kubelet/pods/0984e21c-5c51-4493-80b9-6e0fadf7537d/volumes" Oct 05 06:52:56 crc kubenswrapper[4846]: I1005 06:52:56.507034 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="73ca18b5-0a68-4990-9e7c-1685f87c8dcb" path="/var/lib/kubelet/pods/73ca18b5-0a68-4990-9e7c-1685f87c8dcb/volumes" Oct 05 06:52:56 crc kubenswrapper[4846]: I1005 06:52:56.507693 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8c2f8349-aa93-4ac8-a369-5b4b2b9cd6e2" path="/var/lib/kubelet/pods/8c2f8349-aa93-4ac8-a369-5b4b2b9cd6e2/volumes" Oct 05 06:52:56 crc kubenswrapper[4846]: I1005 06:52:56.510289 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6446bcc-5edd-46bd-a2dc-e8b4773f7248" path="/var/lib/kubelet/pods/b6446bcc-5edd-46bd-a2dc-e8b4773f7248/volumes" Oct 05 06:52:56 crc kubenswrapper[4846]: I1005 06:52:56.511691 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="da62c2a4-387f-4d5a-a9f2-04bd1e110d20" path="/var/lib/kubelet/pods/da62c2a4-387f-4d5a-a9f2-04bd1e110d20/volumes" Oct 05 06:52:56 crc kubenswrapper[4846]: I1005 06:52:56.542740 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m6phk\" (UniqueName: \"kubernetes.io/projected/04f12f2c-eb1a-4b2a-a5d4-bb930a6e278f-kube-api-access-m6phk\") pod \"certified-operators-7fskx\" (UID: \"04f12f2c-eb1a-4b2a-a5d4-bb930a6e278f\") " pod="openshift-marketplace/certified-operators-7fskx" Oct 05 06:52:56 crc kubenswrapper[4846]: I1005 06:52:56.542821 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/04f12f2c-eb1a-4b2a-a5d4-bb930a6e278f-utilities\") pod \"certified-operators-7fskx\" (UID: \"04f12f2c-eb1a-4b2a-a5d4-bb930a6e278f\") " pod="openshift-marketplace/certified-operators-7fskx" Oct 05 06:52:56 crc kubenswrapper[4846]: I1005 06:52:56.542853 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/04f12f2c-eb1a-4b2a-a5d4-bb930a6e278f-catalog-content\") pod \"certified-operators-7fskx\" (UID: \"04f12f2c-eb1a-4b2a-a5d4-bb930a6e278f\") " pod="openshift-marketplace/certified-operators-7fskx" Oct 05 06:52:56 crc kubenswrapper[4846]: I1005 06:52:56.545109 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/04f12f2c-eb1a-4b2a-a5d4-bb930a6e278f-utilities\") pod \"certified-operators-7fskx\" (UID: \"04f12f2c-eb1a-4b2a-a5d4-bb930a6e278f\") " pod="openshift-marketplace/certified-operators-7fskx" Oct 05 06:52:56 crc kubenswrapper[4846]: I1005 06:52:56.545529 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/04f12f2c-eb1a-4b2a-a5d4-bb930a6e278f-catalog-content\") pod \"certified-operators-7fskx\" (UID: \"04f12f2c-eb1a-4b2a-a5d4-bb930a6e278f\") " pod="openshift-marketplace/certified-operators-7fskx" Oct 05 06:52:56 crc kubenswrapper[4846]: I1005 06:52:56.563515 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m6phk\" (UniqueName: \"kubernetes.io/projected/04f12f2c-eb1a-4b2a-a5d4-bb930a6e278f-kube-api-access-m6phk\") pod \"certified-operators-7fskx\" (UID: \"04f12f2c-eb1a-4b2a-a5d4-bb930a6e278f\") " pod="openshift-marketplace/certified-operators-7fskx" Oct 05 06:52:56 crc kubenswrapper[4846]: I1005 06:52:56.647962 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7fskx" Oct 05 06:52:56 crc kubenswrapper[4846]: I1005 06:52:56.758040 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mkttk" Oct 05 06:52:56 crc kubenswrapper[4846]: I1005 06:52:56.848159 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-7fskx"] Oct 05 06:52:56 crc kubenswrapper[4846]: W1005 06:52:56.888094 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod04f12f2c_eb1a_4b2a_a5d4_bb930a6e278f.slice/crio-af917b6dcdb54f598920653e9fc6dbf4032b4a624ab0ceab71642e7d796fa22c WatchSource:0}: Error finding container af917b6dcdb54f598920653e9fc6dbf4032b4a624ab0ceab71642e7d796fa22c: Status 404 returned error can't find the container with id af917b6dcdb54f598920653e9fc6dbf4032b4a624ab0ceab71642e7d796fa22c Oct 05 06:52:56 crc kubenswrapper[4846]: I1005 06:52:56.937227 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-mkttk"] Oct 05 06:52:56 crc kubenswrapper[4846]: W1005 06:52:56.942720 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod01709201_a99b_4e92_adc8_868c6d272ac8.slice/crio-8489668e56a7d513215486a1ddc6b27c5b7325284842df99313f629e686cc6f1 WatchSource:0}: Error finding container 8489668e56a7d513215486a1ddc6b27c5b7325284842df99313f629e686cc6f1: Status 404 returned error can't find the container with id 8489668e56a7d513215486a1ddc6b27c5b7325284842df99313f629e686cc6f1 Oct 05 06:52:57 crc kubenswrapper[4846]: I1005 06:52:57.859634 4846 generic.go:334] "Generic (PLEG): container finished" podID="04f12f2c-eb1a-4b2a-a5d4-bb930a6e278f" containerID="18e2bd9a541f943a89c1980f03c7ef54246f805b0bc6422755243849bcb064ef" exitCode=0 Oct 05 06:52:57 crc kubenswrapper[4846]: I1005 06:52:57.859754 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7fskx" event={"ID":"04f12f2c-eb1a-4b2a-a5d4-bb930a6e278f","Type":"ContainerDied","Data":"18e2bd9a541f943a89c1980f03c7ef54246f805b0bc6422755243849bcb064ef"} Oct 05 06:52:57 crc kubenswrapper[4846]: I1005 06:52:57.860062 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7fskx" event={"ID":"04f12f2c-eb1a-4b2a-a5d4-bb930a6e278f","Type":"ContainerStarted","Data":"af917b6dcdb54f598920653e9fc6dbf4032b4a624ab0ceab71642e7d796fa22c"} Oct 05 06:52:57 crc kubenswrapper[4846]: I1005 06:52:57.863509 4846 generic.go:334] "Generic (PLEG): container finished" podID="01709201-a99b-4e92-adc8-868c6d272ac8" containerID="820923a013375a4c246ba686341ae963205d2636710c2b5b7c77b8ddc61ac007" exitCode=0 Oct 05 06:52:57 crc kubenswrapper[4846]: I1005 06:52:57.863546 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mkttk" event={"ID":"01709201-a99b-4e92-adc8-868c6d272ac8","Type":"ContainerDied","Data":"820923a013375a4c246ba686341ae963205d2636710c2b5b7c77b8ddc61ac007"} Oct 05 06:52:57 crc kubenswrapper[4846]: I1005 06:52:57.863575 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mkttk" event={"ID":"01709201-a99b-4e92-adc8-868c6d272ac8","Type":"ContainerStarted","Data":"8489668e56a7d513215486a1ddc6b27c5b7325284842df99313f629e686cc6f1"} Oct 05 06:52:58 crc kubenswrapper[4846]: I1005 06:52:58.536477 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-2mvqk"] Oct 05 06:52:58 crc kubenswrapper[4846]: I1005 06:52:58.539400 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-2mvqk" Oct 05 06:52:58 crc kubenswrapper[4846]: I1005 06:52:58.544423 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Oct 05 06:52:58 crc kubenswrapper[4846]: I1005 06:52:58.552526 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-2mvqk"] Oct 05 06:52:58 crc kubenswrapper[4846]: I1005 06:52:58.570200 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8bgrh\" (UniqueName: \"kubernetes.io/projected/e6edda1b-9b24-48db-829d-9fd38c6faadb-kube-api-access-8bgrh\") pod \"redhat-operators-2mvqk\" (UID: \"e6edda1b-9b24-48db-829d-9fd38c6faadb\") " pod="openshift-marketplace/redhat-operators-2mvqk" Oct 05 06:52:58 crc kubenswrapper[4846]: I1005 06:52:58.570240 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e6edda1b-9b24-48db-829d-9fd38c6faadb-utilities\") pod \"redhat-operators-2mvqk\" (UID: \"e6edda1b-9b24-48db-829d-9fd38c6faadb\") " pod="openshift-marketplace/redhat-operators-2mvqk" Oct 05 06:52:58 crc kubenswrapper[4846]: I1005 06:52:58.570263 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e6edda1b-9b24-48db-829d-9fd38c6faadb-catalog-content\") pod \"redhat-operators-2mvqk\" (UID: \"e6edda1b-9b24-48db-829d-9fd38c6faadb\") " pod="openshift-marketplace/redhat-operators-2mvqk" Oct 05 06:52:58 crc kubenswrapper[4846]: I1005 06:52:58.671066 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8bgrh\" (UniqueName: \"kubernetes.io/projected/e6edda1b-9b24-48db-829d-9fd38c6faadb-kube-api-access-8bgrh\") pod \"redhat-operators-2mvqk\" (UID: \"e6edda1b-9b24-48db-829d-9fd38c6faadb\") " pod="openshift-marketplace/redhat-operators-2mvqk" Oct 05 06:52:58 crc kubenswrapper[4846]: I1005 06:52:58.671128 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e6edda1b-9b24-48db-829d-9fd38c6faadb-utilities\") pod \"redhat-operators-2mvqk\" (UID: \"e6edda1b-9b24-48db-829d-9fd38c6faadb\") " pod="openshift-marketplace/redhat-operators-2mvqk" Oct 05 06:52:58 crc kubenswrapper[4846]: I1005 06:52:58.671157 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e6edda1b-9b24-48db-829d-9fd38c6faadb-catalog-content\") pod \"redhat-operators-2mvqk\" (UID: \"e6edda1b-9b24-48db-829d-9fd38c6faadb\") " pod="openshift-marketplace/redhat-operators-2mvqk" Oct 05 06:52:58 crc kubenswrapper[4846]: I1005 06:52:58.671922 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e6edda1b-9b24-48db-829d-9fd38c6faadb-utilities\") pod \"redhat-operators-2mvqk\" (UID: \"e6edda1b-9b24-48db-829d-9fd38c6faadb\") " pod="openshift-marketplace/redhat-operators-2mvqk" Oct 05 06:52:58 crc kubenswrapper[4846]: I1005 06:52:58.673633 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e6edda1b-9b24-48db-829d-9fd38c6faadb-catalog-content\") pod \"redhat-operators-2mvqk\" (UID: \"e6edda1b-9b24-48db-829d-9fd38c6faadb\") " pod="openshift-marketplace/redhat-operators-2mvqk" Oct 05 06:52:58 crc kubenswrapper[4846]: I1005 06:52:58.696219 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8bgrh\" (UniqueName: \"kubernetes.io/projected/e6edda1b-9b24-48db-829d-9fd38c6faadb-kube-api-access-8bgrh\") pod \"redhat-operators-2mvqk\" (UID: \"e6edda1b-9b24-48db-829d-9fd38c6faadb\") " pod="openshift-marketplace/redhat-operators-2mvqk" Oct 05 06:52:58 crc kubenswrapper[4846]: I1005 06:52:58.730415 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-2cz9z"] Oct 05 06:52:58 crc kubenswrapper[4846]: I1005 06:52:58.731497 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-2cz9z" Oct 05 06:52:58 crc kubenswrapper[4846]: I1005 06:52:58.734123 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Oct 05 06:52:58 crc kubenswrapper[4846]: I1005 06:52:58.742085 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-2cz9z"] Oct 05 06:52:58 crc kubenswrapper[4846]: I1005 06:52:58.772277 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qq2xd\" (UniqueName: \"kubernetes.io/projected/78cd470d-a948-4332-a64f-6a0da13f295c-kube-api-access-qq2xd\") pod \"community-operators-2cz9z\" (UID: \"78cd470d-a948-4332-a64f-6a0da13f295c\") " pod="openshift-marketplace/community-operators-2cz9z" Oct 05 06:52:58 crc kubenswrapper[4846]: I1005 06:52:58.772355 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/78cd470d-a948-4332-a64f-6a0da13f295c-catalog-content\") pod \"community-operators-2cz9z\" (UID: \"78cd470d-a948-4332-a64f-6a0da13f295c\") " pod="openshift-marketplace/community-operators-2cz9z" Oct 05 06:52:58 crc kubenswrapper[4846]: I1005 06:52:58.772383 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/78cd470d-a948-4332-a64f-6a0da13f295c-utilities\") pod \"community-operators-2cz9z\" (UID: \"78cd470d-a948-4332-a64f-6a0da13f295c\") " pod="openshift-marketplace/community-operators-2cz9z" Oct 05 06:52:58 crc kubenswrapper[4846]: I1005 06:52:58.859571 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-2mvqk" Oct 05 06:52:58 crc kubenswrapper[4846]: I1005 06:52:58.873433 4846 generic.go:334] "Generic (PLEG): container finished" podID="04f12f2c-eb1a-4b2a-a5d4-bb930a6e278f" containerID="71892543adf1b17661583d1354c6942a4ee38203b16f4d82948de9a7a4bdff0d" exitCode=0 Oct 05 06:52:58 crc kubenswrapper[4846]: I1005 06:52:58.873538 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qq2xd\" (UniqueName: \"kubernetes.io/projected/78cd470d-a948-4332-a64f-6a0da13f295c-kube-api-access-qq2xd\") pod \"community-operators-2cz9z\" (UID: \"78cd470d-a948-4332-a64f-6a0da13f295c\") " pod="openshift-marketplace/community-operators-2cz9z" Oct 05 06:52:58 crc kubenswrapper[4846]: I1005 06:52:58.873632 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7fskx" event={"ID":"04f12f2c-eb1a-4b2a-a5d4-bb930a6e278f","Type":"ContainerDied","Data":"71892543adf1b17661583d1354c6942a4ee38203b16f4d82948de9a7a4bdff0d"} Oct 05 06:52:58 crc kubenswrapper[4846]: I1005 06:52:58.875203 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/78cd470d-a948-4332-a64f-6a0da13f295c-catalog-content\") pod \"community-operators-2cz9z\" (UID: \"78cd470d-a948-4332-a64f-6a0da13f295c\") " pod="openshift-marketplace/community-operators-2cz9z" Oct 05 06:52:58 crc kubenswrapper[4846]: I1005 06:52:58.875656 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/78cd470d-a948-4332-a64f-6a0da13f295c-utilities\") pod \"community-operators-2cz9z\" (UID: \"78cd470d-a948-4332-a64f-6a0da13f295c\") " pod="openshift-marketplace/community-operators-2cz9z" Oct 05 06:52:58 crc kubenswrapper[4846]: I1005 06:52:58.876507 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/78cd470d-a948-4332-a64f-6a0da13f295c-catalog-content\") pod \"community-operators-2cz9z\" (UID: \"78cd470d-a948-4332-a64f-6a0da13f295c\") " pod="openshift-marketplace/community-operators-2cz9z" Oct 05 06:52:58 crc kubenswrapper[4846]: I1005 06:52:58.877803 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/78cd470d-a948-4332-a64f-6a0da13f295c-utilities\") pod \"community-operators-2cz9z\" (UID: \"78cd470d-a948-4332-a64f-6a0da13f295c\") " pod="openshift-marketplace/community-operators-2cz9z" Oct 05 06:52:58 crc kubenswrapper[4846]: I1005 06:52:58.896923 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qq2xd\" (UniqueName: \"kubernetes.io/projected/78cd470d-a948-4332-a64f-6a0da13f295c-kube-api-access-qq2xd\") pod \"community-operators-2cz9z\" (UID: \"78cd470d-a948-4332-a64f-6a0da13f295c\") " pod="openshift-marketplace/community-operators-2cz9z" Oct 05 06:52:59 crc kubenswrapper[4846]: I1005 06:52:59.067603 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-2mvqk"] Oct 05 06:52:59 crc kubenswrapper[4846]: W1005 06:52:59.080785 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode6edda1b_9b24_48db_829d_9fd38c6faadb.slice/crio-14f860a82d7daa10aeb274123c3825fc25d6ee8d115280b4b1e6ab25c77f1d85 WatchSource:0}: Error finding container 14f860a82d7daa10aeb274123c3825fc25d6ee8d115280b4b1e6ab25c77f1d85: Status 404 returned error can't find the container with id 14f860a82d7daa10aeb274123c3825fc25d6ee8d115280b4b1e6ab25c77f1d85 Oct 05 06:52:59 crc kubenswrapper[4846]: I1005 06:52:59.088884 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-2cz9z" Oct 05 06:52:59 crc kubenswrapper[4846]: I1005 06:52:59.306721 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-2cz9z"] Oct 05 06:52:59 crc kubenswrapper[4846]: W1005 06:52:59.328823 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod78cd470d_a948_4332_a64f_6a0da13f295c.slice/crio-af82c3fd417783854a9716d971d03633881e1604afb1e44fed890b63f4a7d2d0 WatchSource:0}: Error finding container af82c3fd417783854a9716d971d03633881e1604afb1e44fed890b63f4a7d2d0: Status 404 returned error can't find the container with id af82c3fd417783854a9716d971d03633881e1604afb1e44fed890b63f4a7d2d0 Oct 05 06:52:59 crc kubenswrapper[4846]: I1005 06:52:59.881945 4846 generic.go:334] "Generic (PLEG): container finished" podID="78cd470d-a948-4332-a64f-6a0da13f295c" containerID="0b72a69222833298f68e2bd7d3182d6941ea62ec0ca7bff47fbe55d19c83a802" exitCode=0 Oct 05 06:52:59 crc kubenswrapper[4846]: I1005 06:52:59.882111 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2cz9z" event={"ID":"78cd470d-a948-4332-a64f-6a0da13f295c","Type":"ContainerDied","Data":"0b72a69222833298f68e2bd7d3182d6941ea62ec0ca7bff47fbe55d19c83a802"} Oct 05 06:52:59 crc kubenswrapper[4846]: I1005 06:52:59.882371 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2cz9z" event={"ID":"78cd470d-a948-4332-a64f-6a0da13f295c","Type":"ContainerStarted","Data":"af82c3fd417783854a9716d971d03633881e1604afb1e44fed890b63f4a7d2d0"} Oct 05 06:52:59 crc kubenswrapper[4846]: I1005 06:52:59.884140 4846 generic.go:334] "Generic (PLEG): container finished" podID="e6edda1b-9b24-48db-829d-9fd38c6faadb" containerID="7baeb0bff880dede7bcdb34351794ed3302f86b69673f1daa311af942b16ad36" exitCode=0 Oct 05 06:52:59 crc kubenswrapper[4846]: I1005 06:52:59.884210 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2mvqk" event={"ID":"e6edda1b-9b24-48db-829d-9fd38c6faadb","Type":"ContainerDied","Data":"7baeb0bff880dede7bcdb34351794ed3302f86b69673f1daa311af942b16ad36"} Oct 05 06:52:59 crc kubenswrapper[4846]: I1005 06:52:59.884230 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2mvqk" event={"ID":"e6edda1b-9b24-48db-829d-9fd38c6faadb","Type":"ContainerStarted","Data":"14f860a82d7daa10aeb274123c3825fc25d6ee8d115280b4b1e6ab25c77f1d85"} Oct 05 06:52:59 crc kubenswrapper[4846]: I1005 06:52:59.891399 4846 generic.go:334] "Generic (PLEG): container finished" podID="01709201-a99b-4e92-adc8-868c6d272ac8" containerID="f54d83ff074aaa416a6945b726a8cffb8829ade216b3a201896208ce5267b62c" exitCode=0 Oct 05 06:52:59 crc kubenswrapper[4846]: I1005 06:52:59.891496 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mkttk" event={"ID":"01709201-a99b-4e92-adc8-868c6d272ac8","Type":"ContainerDied","Data":"f54d83ff074aaa416a6945b726a8cffb8829ade216b3a201896208ce5267b62c"} Oct 05 06:52:59 crc kubenswrapper[4846]: I1005 06:52:59.896940 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7fskx" event={"ID":"04f12f2c-eb1a-4b2a-a5d4-bb930a6e278f","Type":"ContainerStarted","Data":"9fe83c84c385bf9375c9ce6f89f774cc32c6289fe57ec7e0d194f9294daca165"} Oct 05 06:52:59 crc kubenswrapper[4846]: I1005 06:52:59.932784 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-7fskx" podStartSLOduration=2.297882677 podStartE2EDuration="3.932759909s" podCreationTimestamp="2025-10-05 06:52:56 +0000 UTC" firstStartedPulling="2025-10-05 06:52:57.861613771 +0000 UTC m=+300.102466546" lastFinishedPulling="2025-10-05 06:52:59.496491003 +0000 UTC m=+301.737343778" observedRunningTime="2025-10-05 06:52:59.929093534 +0000 UTC m=+302.169946309" watchObservedRunningTime="2025-10-05 06:52:59.932759909 +0000 UTC m=+302.173612684" Oct 05 06:53:00 crc kubenswrapper[4846]: I1005 06:53:00.904543 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2mvqk" event={"ID":"e6edda1b-9b24-48db-829d-9fd38c6faadb","Type":"ContainerStarted","Data":"37744ba209759538b28126f757bf852e6e8d313148827dd7c046640679e5cf5b"} Oct 05 06:53:00 crc kubenswrapper[4846]: I1005 06:53:00.907301 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mkttk" event={"ID":"01709201-a99b-4e92-adc8-868c6d272ac8","Type":"ContainerStarted","Data":"0f6d72fe5fee43ce336d1bf49607efa0dc2f8f593c5cb1e0fd89f957a142f46e"} Oct 05 06:53:00 crc kubenswrapper[4846]: I1005 06:53:00.909365 4846 generic.go:334] "Generic (PLEG): container finished" podID="78cd470d-a948-4332-a64f-6a0da13f295c" containerID="e6c97dad2a04d143942bb6bbafe7bddd756825642ac16455deb77fc06f46c7af" exitCode=0 Oct 05 06:53:00 crc kubenswrapper[4846]: I1005 06:53:00.910199 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2cz9z" event={"ID":"78cd470d-a948-4332-a64f-6a0da13f295c","Type":"ContainerDied","Data":"e6c97dad2a04d143942bb6bbafe7bddd756825642ac16455deb77fc06f46c7af"} Oct 05 06:53:00 crc kubenswrapper[4846]: I1005 06:53:00.956909 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-mkttk" podStartSLOduration=2.55170829 podStartE2EDuration="4.956892282s" podCreationTimestamp="2025-10-05 06:52:56 +0000 UTC" firstStartedPulling="2025-10-05 06:52:57.865135232 +0000 UTC m=+300.105988007" lastFinishedPulling="2025-10-05 06:53:00.270319224 +0000 UTC m=+302.511171999" observedRunningTime="2025-10-05 06:53:00.953448292 +0000 UTC m=+303.194301077" watchObservedRunningTime="2025-10-05 06:53:00.956892282 +0000 UTC m=+303.197745057" Oct 05 06:53:01 crc kubenswrapper[4846]: I1005 06:53:01.917491 4846 generic.go:334] "Generic (PLEG): container finished" podID="e6edda1b-9b24-48db-829d-9fd38c6faadb" containerID="37744ba209759538b28126f757bf852e6e8d313148827dd7c046640679e5cf5b" exitCode=0 Oct 05 06:53:01 crc kubenswrapper[4846]: I1005 06:53:01.917586 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2mvqk" event={"ID":"e6edda1b-9b24-48db-829d-9fd38c6faadb","Type":"ContainerDied","Data":"37744ba209759538b28126f757bf852e6e8d313148827dd7c046640679e5cf5b"} Oct 05 06:53:02 crc kubenswrapper[4846]: I1005 06:53:02.926803 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2mvqk" event={"ID":"e6edda1b-9b24-48db-829d-9fd38c6faadb","Type":"ContainerStarted","Data":"30ec36765dbedd939351f654e7525a44bdcf19a9de50cb5582630ef5157403e3"} Oct 05 06:53:02 crc kubenswrapper[4846]: I1005 06:53:02.930664 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2cz9z" event={"ID":"78cd470d-a948-4332-a64f-6a0da13f295c","Type":"ContainerStarted","Data":"4b4ef477d788cd75b5d33a98c846b738894b9a36e82d9009dd6bae2007951190"} Oct 05 06:53:02 crc kubenswrapper[4846]: I1005 06:53:02.953025 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-2mvqk" podStartSLOduration=2.5040425649999998 podStartE2EDuration="4.953003284s" podCreationTimestamp="2025-10-05 06:52:58 +0000 UTC" firstStartedPulling="2025-10-05 06:52:59.885707598 +0000 UTC m=+302.126560373" lastFinishedPulling="2025-10-05 06:53:02.334668317 +0000 UTC m=+304.575521092" observedRunningTime="2025-10-05 06:53:02.94860942 +0000 UTC m=+305.189462185" watchObservedRunningTime="2025-10-05 06:53:02.953003284 +0000 UTC m=+305.193856069" Oct 05 06:53:02 crc kubenswrapper[4846]: I1005 06:53:02.967435 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-2cz9z" podStartSLOduration=3.536433073 podStartE2EDuration="4.967408058s" podCreationTimestamp="2025-10-05 06:52:58 +0000 UTC" firstStartedPulling="2025-10-05 06:52:59.884024475 +0000 UTC m=+302.124877270" lastFinishedPulling="2025-10-05 06:53:01.31499948 +0000 UTC m=+303.555852255" observedRunningTime="2025-10-05 06:53:02.966410352 +0000 UTC m=+305.207263127" watchObservedRunningTime="2025-10-05 06:53:02.967408058 +0000 UTC m=+305.208260833" Oct 05 06:53:06 crc kubenswrapper[4846]: I1005 06:53:06.649087 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-7fskx" Oct 05 06:53:06 crc kubenswrapper[4846]: I1005 06:53:06.649723 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-7fskx" Oct 05 06:53:06 crc kubenswrapper[4846]: I1005 06:53:06.707433 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-7fskx" Oct 05 06:53:06 crc kubenswrapper[4846]: I1005 06:53:06.759318 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-mkttk" Oct 05 06:53:06 crc kubenswrapper[4846]: I1005 06:53:06.759687 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-mkttk" Oct 05 06:53:06 crc kubenswrapper[4846]: I1005 06:53:06.803160 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-mkttk" Oct 05 06:53:06 crc kubenswrapper[4846]: I1005 06:53:06.993703 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-mkttk" Oct 05 06:53:06 crc kubenswrapper[4846]: I1005 06:53:06.994793 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-7fskx" Oct 05 06:53:08 crc kubenswrapper[4846]: I1005 06:53:08.860311 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-2mvqk" Oct 05 06:53:08 crc kubenswrapper[4846]: I1005 06:53:08.860399 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-2mvqk" Oct 05 06:53:08 crc kubenswrapper[4846]: I1005 06:53:08.954045 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-2mvqk" Oct 05 06:53:09 crc kubenswrapper[4846]: I1005 06:53:09.017690 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-2mvqk" Oct 05 06:53:09 crc kubenswrapper[4846]: I1005 06:53:09.090435 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-2cz9z" Oct 05 06:53:09 crc kubenswrapper[4846]: I1005 06:53:09.090497 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-2cz9z" Oct 05 06:53:09 crc kubenswrapper[4846]: I1005 06:53:09.143117 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-2cz9z" Oct 05 06:53:10 crc kubenswrapper[4846]: I1005 06:53:10.020169 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-2cz9z" Oct 05 06:53:20 crc kubenswrapper[4846]: I1005 06:53:20.188024 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-6t2mt" podUID="770ae4fc-21dd-446d-9a41-26e484687787" containerName="oauth-openshift" containerID="cri-o://81328a495c32d2ab0a60dd37759f9f0b77921a3f7df6bef74ba81ecf88e84e5d" gracePeriod=15 Oct 05 06:53:21 crc kubenswrapper[4846]: I1005 06:53:21.040507 4846 generic.go:334] "Generic (PLEG): container finished" podID="770ae4fc-21dd-446d-9a41-26e484687787" containerID="81328a495c32d2ab0a60dd37759f9f0b77921a3f7df6bef74ba81ecf88e84e5d" exitCode=0 Oct 05 06:53:21 crc kubenswrapper[4846]: I1005 06:53:21.040619 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-6t2mt" event={"ID":"770ae4fc-21dd-446d-9a41-26e484687787","Type":"ContainerDied","Data":"81328a495c32d2ab0a60dd37759f9f0b77921a3f7df6bef74ba81ecf88e84e5d"} Oct 05 06:53:21 crc kubenswrapper[4846]: I1005 06:53:21.169274 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-6t2mt" Oct 05 06:53:21 crc kubenswrapper[4846]: I1005 06:53:21.208313 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/770ae4fc-21dd-446d-9a41-26e484687787-v4-0-config-system-session\") pod \"770ae4fc-21dd-446d-9a41-26e484687787\" (UID: \"770ae4fc-21dd-446d-9a41-26e484687787\") " Oct 05 06:53:21 crc kubenswrapper[4846]: I1005 06:53:21.208421 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/770ae4fc-21dd-446d-9a41-26e484687787-audit-dir\") pod \"770ae4fc-21dd-446d-9a41-26e484687787\" (UID: \"770ae4fc-21dd-446d-9a41-26e484687787\") " Oct 05 06:53:21 crc kubenswrapper[4846]: I1005 06:53:21.208462 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/770ae4fc-21dd-446d-9a41-26e484687787-v4-0-config-system-service-ca\") pod \"770ae4fc-21dd-446d-9a41-26e484687787\" (UID: \"770ae4fc-21dd-446d-9a41-26e484687787\") " Oct 05 06:53:21 crc kubenswrapper[4846]: I1005 06:53:21.208534 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/770ae4fc-21dd-446d-9a41-26e484687787-v4-0-config-system-serving-cert\") pod \"770ae4fc-21dd-446d-9a41-26e484687787\" (UID: \"770ae4fc-21dd-446d-9a41-26e484687787\") " Oct 05 06:53:21 crc kubenswrapper[4846]: I1005 06:53:21.208574 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/770ae4fc-21dd-446d-9a41-26e484687787-v4-0-config-user-template-provider-selection\") pod \"770ae4fc-21dd-446d-9a41-26e484687787\" (UID: \"770ae4fc-21dd-446d-9a41-26e484687787\") " Oct 05 06:53:21 crc kubenswrapper[4846]: I1005 06:53:21.208652 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/770ae4fc-21dd-446d-9a41-26e484687787-v4-0-config-system-ocp-branding-template\") pod \"770ae4fc-21dd-446d-9a41-26e484687787\" (UID: \"770ae4fc-21dd-446d-9a41-26e484687787\") " Oct 05 06:53:21 crc kubenswrapper[4846]: I1005 06:53:21.208658 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/770ae4fc-21dd-446d-9a41-26e484687787-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "770ae4fc-21dd-446d-9a41-26e484687787" (UID: "770ae4fc-21dd-446d-9a41-26e484687787"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 05 06:53:21 crc kubenswrapper[4846]: I1005 06:53:21.208755 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/770ae4fc-21dd-446d-9a41-26e484687787-v4-0-config-system-cliconfig\") pod \"770ae4fc-21dd-446d-9a41-26e484687787\" (UID: \"770ae4fc-21dd-446d-9a41-26e484687787\") " Oct 05 06:53:21 crc kubenswrapper[4846]: I1005 06:53:21.209240 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/770ae4fc-21dd-446d-9a41-26e484687787-audit-policies\") pod \"770ae4fc-21dd-446d-9a41-26e484687787\" (UID: \"770ae4fc-21dd-446d-9a41-26e484687787\") " Oct 05 06:53:21 crc kubenswrapper[4846]: I1005 06:53:21.210007 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/770ae4fc-21dd-446d-9a41-26e484687787-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "770ae4fc-21dd-446d-9a41-26e484687787" (UID: "770ae4fc-21dd-446d-9a41-26e484687787"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 06:53:21 crc kubenswrapper[4846]: I1005 06:53:21.210574 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/770ae4fc-21dd-446d-9a41-26e484687787-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "770ae4fc-21dd-446d-9a41-26e484687787" (UID: "770ae4fc-21dd-446d-9a41-26e484687787"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 06:53:21 crc kubenswrapper[4846]: I1005 06:53:21.210637 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/770ae4fc-21dd-446d-9a41-26e484687787-v4-0-config-user-template-error\") pod \"770ae4fc-21dd-446d-9a41-26e484687787\" (UID: \"770ae4fc-21dd-446d-9a41-26e484687787\") " Oct 05 06:53:21 crc kubenswrapper[4846]: I1005 06:53:21.210732 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/770ae4fc-21dd-446d-9a41-26e484687787-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "770ae4fc-21dd-446d-9a41-26e484687787" (UID: "770ae4fc-21dd-446d-9a41-26e484687787"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 06:53:21 crc kubenswrapper[4846]: I1005 06:53:21.211811 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/770ae4fc-21dd-446d-9a41-26e484687787-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "770ae4fc-21dd-446d-9a41-26e484687787" (UID: "770ae4fc-21dd-446d-9a41-26e484687787"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 06:53:21 crc kubenswrapper[4846]: I1005 06:53:21.211923 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/770ae4fc-21dd-446d-9a41-26e484687787-v4-0-config-system-trusted-ca-bundle\") pod \"770ae4fc-21dd-446d-9a41-26e484687787\" (UID: \"770ae4fc-21dd-446d-9a41-26e484687787\") " Oct 05 06:53:21 crc kubenswrapper[4846]: I1005 06:53:21.212031 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/770ae4fc-21dd-446d-9a41-26e484687787-v4-0-config-user-template-login\") pod \"770ae4fc-21dd-446d-9a41-26e484687787\" (UID: \"770ae4fc-21dd-446d-9a41-26e484687787\") " Oct 05 06:53:21 crc kubenswrapper[4846]: I1005 06:53:21.212074 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/770ae4fc-21dd-446d-9a41-26e484687787-v4-0-config-user-idp-0-file-data\") pod \"770ae4fc-21dd-446d-9a41-26e484687787\" (UID: \"770ae4fc-21dd-446d-9a41-26e484687787\") " Oct 05 06:53:21 crc kubenswrapper[4846]: I1005 06:53:21.212111 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/770ae4fc-21dd-446d-9a41-26e484687787-v4-0-config-system-router-certs\") pod \"770ae4fc-21dd-446d-9a41-26e484687787\" (UID: \"770ae4fc-21dd-446d-9a41-26e484687787\") " Oct 05 06:53:21 crc kubenswrapper[4846]: I1005 06:53:21.212221 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-drmkm\" (UniqueName: \"kubernetes.io/projected/770ae4fc-21dd-446d-9a41-26e484687787-kube-api-access-drmkm\") pod \"770ae4fc-21dd-446d-9a41-26e484687787\" (UID: \"770ae4fc-21dd-446d-9a41-26e484687787\") " Oct 05 06:53:21 crc kubenswrapper[4846]: I1005 06:53:21.212580 4846 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/770ae4fc-21dd-446d-9a41-26e484687787-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Oct 05 06:53:21 crc kubenswrapper[4846]: I1005 06:53:21.212607 4846 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/770ae4fc-21dd-446d-9a41-26e484687787-audit-policies\") on node \"crc\" DevicePath \"\"" Oct 05 06:53:21 crc kubenswrapper[4846]: I1005 06:53:21.212626 4846 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/770ae4fc-21dd-446d-9a41-26e484687787-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 06:53:21 crc kubenswrapper[4846]: I1005 06:53:21.212647 4846 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/770ae4fc-21dd-446d-9a41-26e484687787-audit-dir\") on node \"crc\" DevicePath \"\"" Oct 05 06:53:21 crc kubenswrapper[4846]: I1005 06:53:21.212666 4846 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/770ae4fc-21dd-446d-9a41-26e484687787-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Oct 05 06:53:21 crc kubenswrapper[4846]: I1005 06:53:21.212915 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-666545c866-f2jr6"] Oct 05 06:53:21 crc kubenswrapper[4846]: E1005 06:53:21.213264 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="770ae4fc-21dd-446d-9a41-26e484687787" containerName="oauth-openshift" Oct 05 06:53:21 crc kubenswrapper[4846]: I1005 06:53:21.213297 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="770ae4fc-21dd-446d-9a41-26e484687787" containerName="oauth-openshift" Oct 05 06:53:21 crc kubenswrapper[4846]: I1005 06:53:21.213442 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="770ae4fc-21dd-446d-9a41-26e484687787" containerName="oauth-openshift" Oct 05 06:53:21 crc kubenswrapper[4846]: I1005 06:53:21.214105 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-666545c866-f2jr6" Oct 05 06:53:21 crc kubenswrapper[4846]: I1005 06:53:21.220077 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/770ae4fc-21dd-446d-9a41-26e484687787-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "770ae4fc-21dd-446d-9a41-26e484687787" (UID: "770ae4fc-21dd-446d-9a41-26e484687787"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 06:53:21 crc kubenswrapper[4846]: I1005 06:53:21.220657 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/770ae4fc-21dd-446d-9a41-26e484687787-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "770ae4fc-21dd-446d-9a41-26e484687787" (UID: "770ae4fc-21dd-446d-9a41-26e484687787"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 06:53:21 crc kubenswrapper[4846]: I1005 06:53:21.221452 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/770ae4fc-21dd-446d-9a41-26e484687787-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "770ae4fc-21dd-446d-9a41-26e484687787" (UID: "770ae4fc-21dd-446d-9a41-26e484687787"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 06:53:21 crc kubenswrapper[4846]: I1005 06:53:21.222752 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/770ae4fc-21dd-446d-9a41-26e484687787-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "770ae4fc-21dd-446d-9a41-26e484687787" (UID: "770ae4fc-21dd-446d-9a41-26e484687787"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 06:53:21 crc kubenswrapper[4846]: I1005 06:53:21.224403 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/770ae4fc-21dd-446d-9a41-26e484687787-kube-api-access-drmkm" (OuterVolumeSpecName: "kube-api-access-drmkm") pod "770ae4fc-21dd-446d-9a41-26e484687787" (UID: "770ae4fc-21dd-446d-9a41-26e484687787"). InnerVolumeSpecName "kube-api-access-drmkm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 06:53:21 crc kubenswrapper[4846]: I1005 06:53:21.229758 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-666545c866-f2jr6"] Oct 05 06:53:21 crc kubenswrapper[4846]: I1005 06:53:21.234558 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/770ae4fc-21dd-446d-9a41-26e484687787-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "770ae4fc-21dd-446d-9a41-26e484687787" (UID: "770ae4fc-21dd-446d-9a41-26e484687787"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 06:53:21 crc kubenswrapper[4846]: I1005 06:53:21.256608 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/770ae4fc-21dd-446d-9a41-26e484687787-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "770ae4fc-21dd-446d-9a41-26e484687787" (UID: "770ae4fc-21dd-446d-9a41-26e484687787"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 06:53:21 crc kubenswrapper[4846]: I1005 06:53:21.265477 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/770ae4fc-21dd-446d-9a41-26e484687787-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "770ae4fc-21dd-446d-9a41-26e484687787" (UID: "770ae4fc-21dd-446d-9a41-26e484687787"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 06:53:21 crc kubenswrapper[4846]: I1005 06:53:21.266125 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/770ae4fc-21dd-446d-9a41-26e484687787-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "770ae4fc-21dd-446d-9a41-26e484687787" (UID: "770ae4fc-21dd-446d-9a41-26e484687787"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 06:53:21 crc kubenswrapper[4846]: I1005 06:53:21.314315 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/cbeeaaca-fc59-49d2-9b83-0fa357d4b628-v4-0-config-system-session\") pod \"oauth-openshift-666545c866-f2jr6\" (UID: \"cbeeaaca-fc59-49d2-9b83-0fa357d4b628\") " pod="openshift-authentication/oauth-openshift-666545c866-f2jr6" Oct 05 06:53:21 crc kubenswrapper[4846]: I1005 06:53:21.314406 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/cbeeaaca-fc59-49d2-9b83-0fa357d4b628-v4-0-config-system-cliconfig\") pod \"oauth-openshift-666545c866-f2jr6\" (UID: \"cbeeaaca-fc59-49d2-9b83-0fa357d4b628\") " pod="openshift-authentication/oauth-openshift-666545c866-f2jr6" Oct 05 06:53:21 crc kubenswrapper[4846]: I1005 06:53:21.314434 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/cbeeaaca-fc59-49d2-9b83-0fa357d4b628-v4-0-config-system-router-certs\") pod \"oauth-openshift-666545c866-f2jr6\" (UID: \"cbeeaaca-fc59-49d2-9b83-0fa357d4b628\") " pod="openshift-authentication/oauth-openshift-666545c866-f2jr6" Oct 05 06:53:21 crc kubenswrapper[4846]: I1005 06:53:21.314459 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/cbeeaaca-fc59-49d2-9b83-0fa357d4b628-v4-0-config-user-template-error\") pod \"oauth-openshift-666545c866-f2jr6\" (UID: \"cbeeaaca-fc59-49d2-9b83-0fa357d4b628\") " pod="openshift-authentication/oauth-openshift-666545c866-f2jr6" Oct 05 06:53:21 crc kubenswrapper[4846]: I1005 06:53:21.314549 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/cbeeaaca-fc59-49d2-9b83-0fa357d4b628-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-666545c866-f2jr6\" (UID: \"cbeeaaca-fc59-49d2-9b83-0fa357d4b628\") " pod="openshift-authentication/oauth-openshift-666545c866-f2jr6" Oct 05 06:53:21 crc kubenswrapper[4846]: I1005 06:53:21.314619 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/cbeeaaca-fc59-49d2-9b83-0fa357d4b628-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-666545c866-f2jr6\" (UID: \"cbeeaaca-fc59-49d2-9b83-0fa357d4b628\") " pod="openshift-authentication/oauth-openshift-666545c866-f2jr6" Oct 05 06:53:21 crc kubenswrapper[4846]: I1005 06:53:21.314684 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/cbeeaaca-fc59-49d2-9b83-0fa357d4b628-audit-dir\") pod \"oauth-openshift-666545c866-f2jr6\" (UID: \"cbeeaaca-fc59-49d2-9b83-0fa357d4b628\") " pod="openshift-authentication/oauth-openshift-666545c866-f2jr6" Oct 05 06:53:21 crc kubenswrapper[4846]: I1005 06:53:21.314706 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/cbeeaaca-fc59-49d2-9b83-0fa357d4b628-v4-0-config-system-serving-cert\") pod \"oauth-openshift-666545c866-f2jr6\" (UID: \"cbeeaaca-fc59-49d2-9b83-0fa357d4b628\") " pod="openshift-authentication/oauth-openshift-666545c866-f2jr6" Oct 05 06:53:21 crc kubenswrapper[4846]: I1005 06:53:21.314765 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/cbeeaaca-fc59-49d2-9b83-0fa357d4b628-v4-0-config-system-service-ca\") pod \"oauth-openshift-666545c866-f2jr6\" (UID: \"cbeeaaca-fc59-49d2-9b83-0fa357d4b628\") " pod="openshift-authentication/oauth-openshift-666545c866-f2jr6" Oct 05 06:53:21 crc kubenswrapper[4846]: I1005 06:53:21.314793 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/cbeeaaca-fc59-49d2-9b83-0fa357d4b628-v4-0-config-user-template-login\") pod \"oauth-openshift-666545c866-f2jr6\" (UID: \"cbeeaaca-fc59-49d2-9b83-0fa357d4b628\") " pod="openshift-authentication/oauth-openshift-666545c866-f2jr6" Oct 05 06:53:21 crc kubenswrapper[4846]: I1005 06:53:21.314879 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/cbeeaaca-fc59-49d2-9b83-0fa357d4b628-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-666545c866-f2jr6\" (UID: \"cbeeaaca-fc59-49d2-9b83-0fa357d4b628\") " pod="openshift-authentication/oauth-openshift-666545c866-f2jr6" Oct 05 06:53:21 crc kubenswrapper[4846]: I1005 06:53:21.314905 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/cbeeaaca-fc59-49d2-9b83-0fa357d4b628-audit-policies\") pod \"oauth-openshift-666545c866-f2jr6\" (UID: \"cbeeaaca-fc59-49d2-9b83-0fa357d4b628\") " pod="openshift-authentication/oauth-openshift-666545c866-f2jr6" Oct 05 06:53:21 crc kubenswrapper[4846]: I1005 06:53:21.314960 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qsxwt\" (UniqueName: \"kubernetes.io/projected/cbeeaaca-fc59-49d2-9b83-0fa357d4b628-kube-api-access-qsxwt\") pod \"oauth-openshift-666545c866-f2jr6\" (UID: \"cbeeaaca-fc59-49d2-9b83-0fa357d4b628\") " pod="openshift-authentication/oauth-openshift-666545c866-f2jr6" Oct 05 06:53:21 crc kubenswrapper[4846]: I1005 06:53:21.315029 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/cbeeaaca-fc59-49d2-9b83-0fa357d4b628-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-666545c866-f2jr6\" (UID: \"cbeeaaca-fc59-49d2-9b83-0fa357d4b628\") " pod="openshift-authentication/oauth-openshift-666545c866-f2jr6" Oct 05 06:53:21 crc kubenswrapper[4846]: I1005 06:53:21.315134 4846 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/770ae4fc-21dd-446d-9a41-26e484687787-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Oct 05 06:53:21 crc kubenswrapper[4846]: I1005 06:53:21.315150 4846 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/770ae4fc-21dd-446d-9a41-26e484687787-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Oct 05 06:53:21 crc kubenswrapper[4846]: I1005 06:53:21.315163 4846 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/770ae4fc-21dd-446d-9a41-26e484687787-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Oct 05 06:53:21 crc kubenswrapper[4846]: I1005 06:53:21.315195 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-drmkm\" (UniqueName: \"kubernetes.io/projected/770ae4fc-21dd-446d-9a41-26e484687787-kube-api-access-drmkm\") on node \"crc\" DevicePath \"\"" Oct 05 06:53:21 crc kubenswrapper[4846]: I1005 06:53:21.315208 4846 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/770ae4fc-21dd-446d-9a41-26e484687787-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Oct 05 06:53:21 crc kubenswrapper[4846]: I1005 06:53:21.315220 4846 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/770ae4fc-21dd-446d-9a41-26e484687787-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 05 06:53:21 crc kubenswrapper[4846]: I1005 06:53:21.315236 4846 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/770ae4fc-21dd-446d-9a41-26e484687787-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Oct 05 06:53:21 crc kubenswrapper[4846]: I1005 06:53:21.315250 4846 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/770ae4fc-21dd-446d-9a41-26e484687787-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Oct 05 06:53:21 crc kubenswrapper[4846]: I1005 06:53:21.315283 4846 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/770ae4fc-21dd-446d-9a41-26e484687787-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Oct 05 06:53:21 crc kubenswrapper[4846]: I1005 06:53:21.417135 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/cbeeaaca-fc59-49d2-9b83-0fa357d4b628-audit-dir\") pod \"oauth-openshift-666545c866-f2jr6\" (UID: \"cbeeaaca-fc59-49d2-9b83-0fa357d4b628\") " pod="openshift-authentication/oauth-openshift-666545c866-f2jr6" Oct 05 06:53:21 crc kubenswrapper[4846]: I1005 06:53:21.417255 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/cbeeaaca-fc59-49d2-9b83-0fa357d4b628-v4-0-config-system-serving-cert\") pod \"oauth-openshift-666545c866-f2jr6\" (UID: \"cbeeaaca-fc59-49d2-9b83-0fa357d4b628\") " pod="openshift-authentication/oauth-openshift-666545c866-f2jr6" Oct 05 06:53:21 crc kubenswrapper[4846]: I1005 06:53:21.417301 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/cbeeaaca-fc59-49d2-9b83-0fa357d4b628-v4-0-config-system-service-ca\") pod \"oauth-openshift-666545c866-f2jr6\" (UID: \"cbeeaaca-fc59-49d2-9b83-0fa357d4b628\") " pod="openshift-authentication/oauth-openshift-666545c866-f2jr6" Oct 05 06:53:21 crc kubenswrapper[4846]: I1005 06:53:21.417357 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/cbeeaaca-fc59-49d2-9b83-0fa357d4b628-v4-0-config-user-template-login\") pod \"oauth-openshift-666545c866-f2jr6\" (UID: \"cbeeaaca-fc59-49d2-9b83-0fa357d4b628\") " pod="openshift-authentication/oauth-openshift-666545c866-f2jr6" Oct 05 06:53:21 crc kubenswrapper[4846]: I1005 06:53:21.417382 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/cbeeaaca-fc59-49d2-9b83-0fa357d4b628-audit-dir\") pod \"oauth-openshift-666545c866-f2jr6\" (UID: \"cbeeaaca-fc59-49d2-9b83-0fa357d4b628\") " pod="openshift-authentication/oauth-openshift-666545c866-f2jr6" Oct 05 06:53:21 crc kubenswrapper[4846]: I1005 06:53:21.417410 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/cbeeaaca-fc59-49d2-9b83-0fa357d4b628-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-666545c866-f2jr6\" (UID: \"cbeeaaca-fc59-49d2-9b83-0fa357d4b628\") " pod="openshift-authentication/oauth-openshift-666545c866-f2jr6" Oct 05 06:53:21 crc kubenswrapper[4846]: I1005 06:53:21.417570 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/cbeeaaca-fc59-49d2-9b83-0fa357d4b628-audit-policies\") pod \"oauth-openshift-666545c866-f2jr6\" (UID: \"cbeeaaca-fc59-49d2-9b83-0fa357d4b628\") " pod="openshift-authentication/oauth-openshift-666545c866-f2jr6" Oct 05 06:53:21 crc kubenswrapper[4846]: I1005 06:53:21.417723 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qsxwt\" (UniqueName: \"kubernetes.io/projected/cbeeaaca-fc59-49d2-9b83-0fa357d4b628-kube-api-access-qsxwt\") pod \"oauth-openshift-666545c866-f2jr6\" (UID: \"cbeeaaca-fc59-49d2-9b83-0fa357d4b628\") " pod="openshift-authentication/oauth-openshift-666545c866-f2jr6" Oct 05 06:53:21 crc kubenswrapper[4846]: I1005 06:53:21.417822 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/cbeeaaca-fc59-49d2-9b83-0fa357d4b628-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-666545c866-f2jr6\" (UID: \"cbeeaaca-fc59-49d2-9b83-0fa357d4b628\") " pod="openshift-authentication/oauth-openshift-666545c866-f2jr6" Oct 05 06:53:21 crc kubenswrapper[4846]: I1005 06:53:21.418026 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/cbeeaaca-fc59-49d2-9b83-0fa357d4b628-v4-0-config-system-session\") pod \"oauth-openshift-666545c866-f2jr6\" (UID: \"cbeeaaca-fc59-49d2-9b83-0fa357d4b628\") " pod="openshift-authentication/oauth-openshift-666545c866-f2jr6" Oct 05 06:53:21 crc kubenswrapper[4846]: I1005 06:53:21.418098 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/cbeeaaca-fc59-49d2-9b83-0fa357d4b628-v4-0-config-system-cliconfig\") pod \"oauth-openshift-666545c866-f2jr6\" (UID: \"cbeeaaca-fc59-49d2-9b83-0fa357d4b628\") " pod="openshift-authentication/oauth-openshift-666545c866-f2jr6" Oct 05 06:53:21 crc kubenswrapper[4846]: I1005 06:53:21.418138 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/cbeeaaca-fc59-49d2-9b83-0fa357d4b628-v4-0-config-system-router-certs\") pod \"oauth-openshift-666545c866-f2jr6\" (UID: \"cbeeaaca-fc59-49d2-9b83-0fa357d4b628\") " pod="openshift-authentication/oauth-openshift-666545c866-f2jr6" Oct 05 06:53:21 crc kubenswrapper[4846]: I1005 06:53:21.418239 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/cbeeaaca-fc59-49d2-9b83-0fa357d4b628-v4-0-config-user-template-error\") pod \"oauth-openshift-666545c866-f2jr6\" (UID: \"cbeeaaca-fc59-49d2-9b83-0fa357d4b628\") " pod="openshift-authentication/oauth-openshift-666545c866-f2jr6" Oct 05 06:53:21 crc kubenswrapper[4846]: I1005 06:53:21.418285 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/cbeeaaca-fc59-49d2-9b83-0fa357d4b628-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-666545c866-f2jr6\" (UID: \"cbeeaaca-fc59-49d2-9b83-0fa357d4b628\") " pod="openshift-authentication/oauth-openshift-666545c866-f2jr6" Oct 05 06:53:21 crc kubenswrapper[4846]: I1005 06:53:21.418395 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/cbeeaaca-fc59-49d2-9b83-0fa357d4b628-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-666545c866-f2jr6\" (UID: \"cbeeaaca-fc59-49d2-9b83-0fa357d4b628\") " pod="openshift-authentication/oauth-openshift-666545c866-f2jr6" Oct 05 06:53:21 crc kubenswrapper[4846]: I1005 06:53:21.418556 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/cbeeaaca-fc59-49d2-9b83-0fa357d4b628-v4-0-config-system-service-ca\") pod \"oauth-openshift-666545c866-f2jr6\" (UID: \"cbeeaaca-fc59-49d2-9b83-0fa357d4b628\") " pod="openshift-authentication/oauth-openshift-666545c866-f2jr6" Oct 05 06:53:21 crc kubenswrapper[4846]: I1005 06:53:21.420000 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/cbeeaaca-fc59-49d2-9b83-0fa357d4b628-audit-policies\") pod \"oauth-openshift-666545c866-f2jr6\" (UID: \"cbeeaaca-fc59-49d2-9b83-0fa357d4b628\") " pod="openshift-authentication/oauth-openshift-666545c866-f2jr6" Oct 05 06:53:21 crc kubenswrapper[4846]: I1005 06:53:21.420604 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/cbeeaaca-fc59-49d2-9b83-0fa357d4b628-v4-0-config-system-cliconfig\") pod \"oauth-openshift-666545c866-f2jr6\" (UID: \"cbeeaaca-fc59-49d2-9b83-0fa357d4b628\") " pod="openshift-authentication/oauth-openshift-666545c866-f2jr6" Oct 05 06:53:21 crc kubenswrapper[4846]: I1005 06:53:21.420919 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/cbeeaaca-fc59-49d2-9b83-0fa357d4b628-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-666545c866-f2jr6\" (UID: \"cbeeaaca-fc59-49d2-9b83-0fa357d4b628\") " pod="openshift-authentication/oauth-openshift-666545c866-f2jr6" Oct 05 06:53:21 crc kubenswrapper[4846]: I1005 06:53:21.421931 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/cbeeaaca-fc59-49d2-9b83-0fa357d4b628-v4-0-config-system-serving-cert\") pod \"oauth-openshift-666545c866-f2jr6\" (UID: \"cbeeaaca-fc59-49d2-9b83-0fa357d4b628\") " pod="openshift-authentication/oauth-openshift-666545c866-f2jr6" Oct 05 06:53:21 crc kubenswrapper[4846]: I1005 06:53:21.423267 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/cbeeaaca-fc59-49d2-9b83-0fa357d4b628-v4-0-config-system-session\") pod \"oauth-openshift-666545c866-f2jr6\" (UID: \"cbeeaaca-fc59-49d2-9b83-0fa357d4b628\") " pod="openshift-authentication/oauth-openshift-666545c866-f2jr6" Oct 05 06:53:21 crc kubenswrapper[4846]: I1005 06:53:21.423350 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/cbeeaaca-fc59-49d2-9b83-0fa357d4b628-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-666545c866-f2jr6\" (UID: \"cbeeaaca-fc59-49d2-9b83-0fa357d4b628\") " pod="openshift-authentication/oauth-openshift-666545c866-f2jr6" Oct 05 06:53:21 crc kubenswrapper[4846]: I1005 06:53:21.427022 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/cbeeaaca-fc59-49d2-9b83-0fa357d4b628-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-666545c866-f2jr6\" (UID: \"cbeeaaca-fc59-49d2-9b83-0fa357d4b628\") " pod="openshift-authentication/oauth-openshift-666545c866-f2jr6" Oct 05 06:53:21 crc kubenswrapper[4846]: I1005 06:53:21.427381 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/cbeeaaca-fc59-49d2-9b83-0fa357d4b628-v4-0-config-user-template-error\") pod \"oauth-openshift-666545c866-f2jr6\" (UID: \"cbeeaaca-fc59-49d2-9b83-0fa357d4b628\") " pod="openshift-authentication/oauth-openshift-666545c866-f2jr6" Oct 05 06:53:21 crc kubenswrapper[4846]: I1005 06:53:21.427644 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/cbeeaaca-fc59-49d2-9b83-0fa357d4b628-v4-0-config-system-router-certs\") pod \"oauth-openshift-666545c866-f2jr6\" (UID: \"cbeeaaca-fc59-49d2-9b83-0fa357d4b628\") " pod="openshift-authentication/oauth-openshift-666545c866-f2jr6" Oct 05 06:53:21 crc kubenswrapper[4846]: I1005 06:53:21.429040 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/cbeeaaca-fc59-49d2-9b83-0fa357d4b628-v4-0-config-user-template-login\") pod \"oauth-openshift-666545c866-f2jr6\" (UID: \"cbeeaaca-fc59-49d2-9b83-0fa357d4b628\") " pod="openshift-authentication/oauth-openshift-666545c866-f2jr6" Oct 05 06:53:21 crc kubenswrapper[4846]: I1005 06:53:21.429632 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/cbeeaaca-fc59-49d2-9b83-0fa357d4b628-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-666545c866-f2jr6\" (UID: \"cbeeaaca-fc59-49d2-9b83-0fa357d4b628\") " pod="openshift-authentication/oauth-openshift-666545c866-f2jr6" Oct 05 06:53:21 crc kubenswrapper[4846]: I1005 06:53:21.441653 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qsxwt\" (UniqueName: \"kubernetes.io/projected/cbeeaaca-fc59-49d2-9b83-0fa357d4b628-kube-api-access-qsxwt\") pod \"oauth-openshift-666545c866-f2jr6\" (UID: \"cbeeaaca-fc59-49d2-9b83-0fa357d4b628\") " pod="openshift-authentication/oauth-openshift-666545c866-f2jr6" Oct 05 06:53:21 crc kubenswrapper[4846]: I1005 06:53:21.585247 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-666545c866-f2jr6" Oct 05 06:53:21 crc kubenswrapper[4846]: I1005 06:53:21.876814 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-666545c866-f2jr6"] Oct 05 06:53:21 crc kubenswrapper[4846]: W1005 06:53:21.887382 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcbeeaaca_fc59_49d2_9b83_0fa357d4b628.slice/crio-0b3a72cc73c9a1ae5d856a80ccc19e0e6ec3c9d39e880ccb59836c14673b4400 WatchSource:0}: Error finding container 0b3a72cc73c9a1ae5d856a80ccc19e0e6ec3c9d39e880ccb59836c14673b4400: Status 404 returned error can't find the container with id 0b3a72cc73c9a1ae5d856a80ccc19e0e6ec3c9d39e880ccb59836c14673b4400 Oct 05 06:53:22 crc kubenswrapper[4846]: I1005 06:53:22.049162 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-6t2mt" event={"ID":"770ae4fc-21dd-446d-9a41-26e484687787","Type":"ContainerDied","Data":"eab4cdc88060cdf416b1f6d3d9f5545c5bc5357725f772908b1c8cd0490ed45a"} Oct 05 06:53:22 crc kubenswrapper[4846]: I1005 06:53:22.049338 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-6t2mt" Oct 05 06:53:22 crc kubenswrapper[4846]: I1005 06:53:22.049753 4846 scope.go:117] "RemoveContainer" containerID="81328a495c32d2ab0a60dd37759f9f0b77921a3f7df6bef74ba81ecf88e84e5d" Oct 05 06:53:22 crc kubenswrapper[4846]: I1005 06:53:22.053746 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-666545c866-f2jr6" event={"ID":"cbeeaaca-fc59-49d2-9b83-0fa357d4b628","Type":"ContainerStarted","Data":"0b3a72cc73c9a1ae5d856a80ccc19e0e6ec3c9d39e880ccb59836c14673b4400"} Oct 05 06:53:22 crc kubenswrapper[4846]: I1005 06:53:22.119269 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-6t2mt"] Oct 05 06:53:22 crc kubenswrapper[4846]: I1005 06:53:22.125392 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-6t2mt"] Oct 05 06:53:22 crc kubenswrapper[4846]: I1005 06:53:22.504462 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="770ae4fc-21dd-446d-9a41-26e484687787" path="/var/lib/kubelet/pods/770ae4fc-21dd-446d-9a41-26e484687787/volumes" Oct 05 06:53:23 crc kubenswrapper[4846]: I1005 06:53:23.067481 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-666545c866-f2jr6" event={"ID":"cbeeaaca-fc59-49d2-9b83-0fa357d4b628","Type":"ContainerStarted","Data":"6017233e91cc3b615a0268707d7b13fc46df24b4960ce02f6b4ae25e5f87084e"} Oct 05 06:53:23 crc kubenswrapper[4846]: I1005 06:53:23.068072 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-666545c866-f2jr6" Oct 05 06:53:23 crc kubenswrapper[4846]: I1005 06:53:23.079016 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-666545c866-f2jr6" Oct 05 06:53:23 crc kubenswrapper[4846]: I1005 06:53:23.106093 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-666545c866-f2jr6" podStartSLOduration=28.106066283 podStartE2EDuration="28.106066283s" podCreationTimestamp="2025-10-05 06:52:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 06:53:23.10135535 +0000 UTC m=+325.342208165" watchObservedRunningTime="2025-10-05 06:53:23.106066283 +0000 UTC m=+325.346919058" Oct 05 06:54:23 crc kubenswrapper[4846]: I1005 06:54:23.325123 4846 patch_prober.go:28] interesting pod/machine-config-daemon-fscvf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 06:54:23 crc kubenswrapper[4846]: I1005 06:54:23.325776 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 06:54:53 crc kubenswrapper[4846]: I1005 06:54:53.324716 4846 patch_prober.go:28] interesting pod/machine-config-daemon-fscvf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 06:54:53 crc kubenswrapper[4846]: I1005 06:54:53.325634 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 06:55:23 crc kubenswrapper[4846]: I1005 06:55:23.325523 4846 patch_prober.go:28] interesting pod/machine-config-daemon-fscvf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 06:55:23 crc kubenswrapper[4846]: I1005 06:55:23.327557 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 06:55:23 crc kubenswrapper[4846]: I1005 06:55:23.327749 4846 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" Oct 05 06:55:23 crc kubenswrapper[4846]: I1005 06:55:23.328575 4846 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"555aa2226a521510ee0c88d4d95100dced793271cffe5d0ca2a8cf181778e4a6"} pod="openshift-machine-config-operator/machine-config-daemon-fscvf" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 05 06:55:23 crc kubenswrapper[4846]: I1005 06:55:23.328766 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" containerName="machine-config-daemon" containerID="cri-o://555aa2226a521510ee0c88d4d95100dced793271cffe5d0ca2a8cf181778e4a6" gracePeriod=600 Oct 05 06:55:23 crc kubenswrapper[4846]: I1005 06:55:23.950531 4846 generic.go:334] "Generic (PLEG): container finished" podID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" containerID="555aa2226a521510ee0c88d4d95100dced793271cffe5d0ca2a8cf181778e4a6" exitCode=0 Oct 05 06:55:23 crc kubenswrapper[4846]: I1005 06:55:23.951436 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" event={"ID":"ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7","Type":"ContainerDied","Data":"555aa2226a521510ee0c88d4d95100dced793271cffe5d0ca2a8cf181778e4a6"} Oct 05 06:55:23 crc kubenswrapper[4846]: I1005 06:55:23.951473 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" event={"ID":"ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7","Type":"ContainerStarted","Data":"3b7026879bf093ff49269f1b47a66adf5f31a2fca4f61fcc9b2ba0d01156c507"} Oct 05 06:55:23 crc kubenswrapper[4846]: I1005 06:55:23.951492 4846 scope.go:117] "RemoveContainer" containerID="516f88f4990f5962d8b694a3d889259cc5508a309919870191c2e1331640c973" Oct 05 06:55:47 crc kubenswrapper[4846]: I1005 06:55:47.321020 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-swkwx"] Oct 05 06:55:47 crc kubenswrapper[4846]: I1005 06:55:47.322738 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-swkwx" Oct 05 06:55:47 crc kubenswrapper[4846]: I1005 06:55:47.341026 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-swkwx"] Oct 05 06:55:47 crc kubenswrapper[4846]: I1005 06:55:47.435703 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ntt9f\" (UniqueName: \"kubernetes.io/projected/f3f92d0d-6a1a-475d-8be1-1cbb859555da-kube-api-access-ntt9f\") pod \"image-registry-66df7c8f76-swkwx\" (UID: \"f3f92d0d-6a1a-475d-8be1-1cbb859555da\") " pod="openshift-image-registry/image-registry-66df7c8f76-swkwx" Oct 05 06:55:47 crc kubenswrapper[4846]: I1005 06:55:47.435792 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/f3f92d0d-6a1a-475d-8be1-1cbb859555da-ca-trust-extracted\") pod \"image-registry-66df7c8f76-swkwx\" (UID: \"f3f92d0d-6a1a-475d-8be1-1cbb859555da\") " pod="openshift-image-registry/image-registry-66df7c8f76-swkwx" Oct 05 06:55:47 crc kubenswrapper[4846]: I1005 06:55:47.435823 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/f3f92d0d-6a1a-475d-8be1-1cbb859555da-registry-certificates\") pod \"image-registry-66df7c8f76-swkwx\" (UID: \"f3f92d0d-6a1a-475d-8be1-1cbb859555da\") " pod="openshift-image-registry/image-registry-66df7c8f76-swkwx" Oct 05 06:55:47 crc kubenswrapper[4846]: I1005 06:55:47.435885 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-swkwx\" (UID: \"f3f92d0d-6a1a-475d-8be1-1cbb859555da\") " pod="openshift-image-registry/image-registry-66df7c8f76-swkwx" Oct 05 06:55:47 crc kubenswrapper[4846]: I1005 06:55:47.435945 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f3f92d0d-6a1a-475d-8be1-1cbb859555da-trusted-ca\") pod \"image-registry-66df7c8f76-swkwx\" (UID: \"f3f92d0d-6a1a-475d-8be1-1cbb859555da\") " pod="openshift-image-registry/image-registry-66df7c8f76-swkwx" Oct 05 06:55:47 crc kubenswrapper[4846]: I1005 06:55:47.435995 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/f3f92d0d-6a1a-475d-8be1-1cbb859555da-bound-sa-token\") pod \"image-registry-66df7c8f76-swkwx\" (UID: \"f3f92d0d-6a1a-475d-8be1-1cbb859555da\") " pod="openshift-image-registry/image-registry-66df7c8f76-swkwx" Oct 05 06:55:47 crc kubenswrapper[4846]: I1005 06:55:47.436399 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/f3f92d0d-6a1a-475d-8be1-1cbb859555da-registry-tls\") pod \"image-registry-66df7c8f76-swkwx\" (UID: \"f3f92d0d-6a1a-475d-8be1-1cbb859555da\") " pod="openshift-image-registry/image-registry-66df7c8f76-swkwx" Oct 05 06:55:47 crc kubenswrapper[4846]: I1005 06:55:47.436504 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/f3f92d0d-6a1a-475d-8be1-1cbb859555da-installation-pull-secrets\") pod \"image-registry-66df7c8f76-swkwx\" (UID: \"f3f92d0d-6a1a-475d-8be1-1cbb859555da\") " pod="openshift-image-registry/image-registry-66df7c8f76-swkwx" Oct 05 06:55:47 crc kubenswrapper[4846]: I1005 06:55:47.459581 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-swkwx\" (UID: \"f3f92d0d-6a1a-475d-8be1-1cbb859555da\") " pod="openshift-image-registry/image-registry-66df7c8f76-swkwx" Oct 05 06:55:47 crc kubenswrapper[4846]: I1005 06:55:47.538446 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/f3f92d0d-6a1a-475d-8be1-1cbb859555da-registry-tls\") pod \"image-registry-66df7c8f76-swkwx\" (UID: \"f3f92d0d-6a1a-475d-8be1-1cbb859555da\") " pod="openshift-image-registry/image-registry-66df7c8f76-swkwx" Oct 05 06:55:47 crc kubenswrapper[4846]: I1005 06:55:47.538985 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/f3f92d0d-6a1a-475d-8be1-1cbb859555da-installation-pull-secrets\") pod \"image-registry-66df7c8f76-swkwx\" (UID: \"f3f92d0d-6a1a-475d-8be1-1cbb859555da\") " pod="openshift-image-registry/image-registry-66df7c8f76-swkwx" Oct 05 06:55:47 crc kubenswrapper[4846]: I1005 06:55:47.539015 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ntt9f\" (UniqueName: \"kubernetes.io/projected/f3f92d0d-6a1a-475d-8be1-1cbb859555da-kube-api-access-ntt9f\") pod \"image-registry-66df7c8f76-swkwx\" (UID: \"f3f92d0d-6a1a-475d-8be1-1cbb859555da\") " pod="openshift-image-registry/image-registry-66df7c8f76-swkwx" Oct 05 06:55:47 crc kubenswrapper[4846]: I1005 06:55:47.539040 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/f3f92d0d-6a1a-475d-8be1-1cbb859555da-ca-trust-extracted\") pod \"image-registry-66df7c8f76-swkwx\" (UID: \"f3f92d0d-6a1a-475d-8be1-1cbb859555da\") " pod="openshift-image-registry/image-registry-66df7c8f76-swkwx" Oct 05 06:55:47 crc kubenswrapper[4846]: I1005 06:55:47.539059 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/f3f92d0d-6a1a-475d-8be1-1cbb859555da-registry-certificates\") pod \"image-registry-66df7c8f76-swkwx\" (UID: \"f3f92d0d-6a1a-475d-8be1-1cbb859555da\") " pod="openshift-image-registry/image-registry-66df7c8f76-swkwx" Oct 05 06:55:47 crc kubenswrapper[4846]: I1005 06:55:47.539089 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f3f92d0d-6a1a-475d-8be1-1cbb859555da-trusted-ca\") pod \"image-registry-66df7c8f76-swkwx\" (UID: \"f3f92d0d-6a1a-475d-8be1-1cbb859555da\") " pod="openshift-image-registry/image-registry-66df7c8f76-swkwx" Oct 05 06:55:47 crc kubenswrapper[4846]: I1005 06:55:47.539113 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/f3f92d0d-6a1a-475d-8be1-1cbb859555da-bound-sa-token\") pod \"image-registry-66df7c8f76-swkwx\" (UID: \"f3f92d0d-6a1a-475d-8be1-1cbb859555da\") " pod="openshift-image-registry/image-registry-66df7c8f76-swkwx" Oct 05 06:55:47 crc kubenswrapper[4846]: I1005 06:55:47.540497 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/f3f92d0d-6a1a-475d-8be1-1cbb859555da-ca-trust-extracted\") pod \"image-registry-66df7c8f76-swkwx\" (UID: \"f3f92d0d-6a1a-475d-8be1-1cbb859555da\") " pod="openshift-image-registry/image-registry-66df7c8f76-swkwx" Oct 05 06:55:47 crc kubenswrapper[4846]: I1005 06:55:47.541230 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f3f92d0d-6a1a-475d-8be1-1cbb859555da-trusted-ca\") pod \"image-registry-66df7c8f76-swkwx\" (UID: \"f3f92d0d-6a1a-475d-8be1-1cbb859555da\") " pod="openshift-image-registry/image-registry-66df7c8f76-swkwx" Oct 05 06:55:47 crc kubenswrapper[4846]: I1005 06:55:47.541400 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/f3f92d0d-6a1a-475d-8be1-1cbb859555da-registry-certificates\") pod \"image-registry-66df7c8f76-swkwx\" (UID: \"f3f92d0d-6a1a-475d-8be1-1cbb859555da\") " pod="openshift-image-registry/image-registry-66df7c8f76-swkwx" Oct 05 06:55:47 crc kubenswrapper[4846]: I1005 06:55:47.550168 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/f3f92d0d-6a1a-475d-8be1-1cbb859555da-registry-tls\") pod \"image-registry-66df7c8f76-swkwx\" (UID: \"f3f92d0d-6a1a-475d-8be1-1cbb859555da\") " pod="openshift-image-registry/image-registry-66df7c8f76-swkwx" Oct 05 06:55:47 crc kubenswrapper[4846]: I1005 06:55:47.554351 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/f3f92d0d-6a1a-475d-8be1-1cbb859555da-installation-pull-secrets\") pod \"image-registry-66df7c8f76-swkwx\" (UID: \"f3f92d0d-6a1a-475d-8be1-1cbb859555da\") " pod="openshift-image-registry/image-registry-66df7c8f76-swkwx" Oct 05 06:55:47 crc kubenswrapper[4846]: I1005 06:55:47.562052 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/f3f92d0d-6a1a-475d-8be1-1cbb859555da-bound-sa-token\") pod \"image-registry-66df7c8f76-swkwx\" (UID: \"f3f92d0d-6a1a-475d-8be1-1cbb859555da\") " pod="openshift-image-registry/image-registry-66df7c8f76-swkwx" Oct 05 06:55:47 crc kubenswrapper[4846]: I1005 06:55:47.563479 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ntt9f\" (UniqueName: \"kubernetes.io/projected/f3f92d0d-6a1a-475d-8be1-1cbb859555da-kube-api-access-ntt9f\") pod \"image-registry-66df7c8f76-swkwx\" (UID: \"f3f92d0d-6a1a-475d-8be1-1cbb859555da\") " pod="openshift-image-registry/image-registry-66df7c8f76-swkwx" Oct 05 06:55:47 crc kubenswrapper[4846]: I1005 06:55:47.643530 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-swkwx" Oct 05 06:55:47 crc kubenswrapper[4846]: I1005 06:55:47.900454 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-swkwx"] Oct 05 06:55:48 crc kubenswrapper[4846]: I1005 06:55:48.149813 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-swkwx" event={"ID":"f3f92d0d-6a1a-475d-8be1-1cbb859555da","Type":"ContainerStarted","Data":"a2bdbdcad35d115c0e00cff2f77aeee79c1a24b09902526f5b24f35c2fad32ef"} Oct 05 06:55:49 crc kubenswrapper[4846]: I1005 06:55:49.157838 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-swkwx" event={"ID":"f3f92d0d-6a1a-475d-8be1-1cbb859555da","Type":"ContainerStarted","Data":"0cc0ac8610fe9296f1dd42cc72eede83458a944b74699fd1b8383f499259a35d"} Oct 05 06:55:49 crc kubenswrapper[4846]: I1005 06:55:49.159171 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-swkwx" Oct 05 06:55:49 crc kubenswrapper[4846]: I1005 06:55:49.197228 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-swkwx" podStartSLOduration=2.197080854 podStartE2EDuration="2.197080854s" podCreationTimestamp="2025-10-05 06:55:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 06:55:49.196787386 +0000 UTC m=+471.437640181" watchObservedRunningTime="2025-10-05 06:55:49.197080854 +0000 UTC m=+471.437933669" Oct 05 06:56:07 crc kubenswrapper[4846]: I1005 06:56:07.649008 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-swkwx" Oct 05 06:56:07 crc kubenswrapper[4846]: I1005 06:56:07.720861 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-rf9l4"] Oct 05 06:56:32 crc kubenswrapper[4846]: I1005 06:56:32.775393 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-rf9l4" podUID="78d3e7b1-7258-4508-ad38-373f21bb5d96" containerName="registry" containerID="cri-o://3edcc13d0073ef87f5bc93cf466592a38ca27007b40995acebf6d1c84edde195" gracePeriod=30 Oct 05 06:56:33 crc kubenswrapper[4846]: I1005 06:56:33.252853 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-rf9l4" Oct 05 06:56:33 crc kubenswrapper[4846]: I1005 06:56:33.378167 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/78d3e7b1-7258-4508-ad38-373f21bb5d96-registry-certificates\") pod \"78d3e7b1-7258-4508-ad38-373f21bb5d96\" (UID: \"78d3e7b1-7258-4508-ad38-373f21bb5d96\") " Oct 05 06:56:33 crc kubenswrapper[4846]: I1005 06:56:33.379333 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/78d3e7b1-7258-4508-ad38-373f21bb5d96-registry-tls\") pod \"78d3e7b1-7258-4508-ad38-373f21bb5d96\" (UID: \"78d3e7b1-7258-4508-ad38-373f21bb5d96\") " Oct 05 06:56:33 crc kubenswrapper[4846]: I1005 06:56:33.379429 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/78d3e7b1-7258-4508-ad38-373f21bb5d96-installation-pull-secrets\") pod \"78d3e7b1-7258-4508-ad38-373f21bb5d96\" (UID: \"78d3e7b1-7258-4508-ad38-373f21bb5d96\") " Oct 05 06:56:33 crc kubenswrapper[4846]: I1005 06:56:33.379512 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/78d3e7b1-7258-4508-ad38-373f21bb5d96-trusted-ca\") pod \"78d3e7b1-7258-4508-ad38-373f21bb5d96\" (UID: \"78d3e7b1-7258-4508-ad38-373f21bb5d96\") " Oct 05 06:56:33 crc kubenswrapper[4846]: I1005 06:56:33.379779 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"78d3e7b1-7258-4508-ad38-373f21bb5d96\" (UID: \"78d3e7b1-7258-4508-ad38-373f21bb5d96\") " Oct 05 06:56:33 crc kubenswrapper[4846]: I1005 06:56:33.379860 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/78d3e7b1-7258-4508-ad38-373f21bb5d96-bound-sa-token\") pod \"78d3e7b1-7258-4508-ad38-373f21bb5d96\" (UID: \"78d3e7b1-7258-4508-ad38-373f21bb5d96\") " Oct 05 06:56:33 crc kubenswrapper[4846]: I1005 06:56:33.379962 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/78d3e7b1-7258-4508-ad38-373f21bb5d96-ca-trust-extracted\") pod \"78d3e7b1-7258-4508-ad38-373f21bb5d96\" (UID: \"78d3e7b1-7258-4508-ad38-373f21bb5d96\") " Oct 05 06:56:33 crc kubenswrapper[4846]: I1005 06:56:33.380081 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mtl7j\" (UniqueName: \"kubernetes.io/projected/78d3e7b1-7258-4508-ad38-373f21bb5d96-kube-api-access-mtl7j\") pod \"78d3e7b1-7258-4508-ad38-373f21bb5d96\" (UID: \"78d3e7b1-7258-4508-ad38-373f21bb5d96\") " Oct 05 06:56:33 crc kubenswrapper[4846]: I1005 06:56:33.380302 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/78d3e7b1-7258-4508-ad38-373f21bb5d96-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "78d3e7b1-7258-4508-ad38-373f21bb5d96" (UID: "78d3e7b1-7258-4508-ad38-373f21bb5d96"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 06:56:33 crc kubenswrapper[4846]: I1005 06:56:33.380638 4846 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/78d3e7b1-7258-4508-ad38-373f21bb5d96-registry-certificates\") on node \"crc\" DevicePath \"\"" Oct 05 06:56:33 crc kubenswrapper[4846]: I1005 06:56:33.381688 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/78d3e7b1-7258-4508-ad38-373f21bb5d96-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "78d3e7b1-7258-4508-ad38-373f21bb5d96" (UID: "78d3e7b1-7258-4508-ad38-373f21bb5d96"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 06:56:33 crc kubenswrapper[4846]: I1005 06:56:33.389534 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/78d3e7b1-7258-4508-ad38-373f21bb5d96-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "78d3e7b1-7258-4508-ad38-373f21bb5d96" (UID: "78d3e7b1-7258-4508-ad38-373f21bb5d96"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 06:56:33 crc kubenswrapper[4846]: I1005 06:56:33.391526 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/78d3e7b1-7258-4508-ad38-373f21bb5d96-kube-api-access-mtl7j" (OuterVolumeSpecName: "kube-api-access-mtl7j") pod "78d3e7b1-7258-4508-ad38-373f21bb5d96" (UID: "78d3e7b1-7258-4508-ad38-373f21bb5d96"). InnerVolumeSpecName "kube-api-access-mtl7j". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 06:56:33 crc kubenswrapper[4846]: I1005 06:56:33.391886 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/78d3e7b1-7258-4508-ad38-373f21bb5d96-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "78d3e7b1-7258-4508-ad38-373f21bb5d96" (UID: "78d3e7b1-7258-4508-ad38-373f21bb5d96"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 06:56:33 crc kubenswrapper[4846]: I1005 06:56:33.392080 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/78d3e7b1-7258-4508-ad38-373f21bb5d96-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "78d3e7b1-7258-4508-ad38-373f21bb5d96" (UID: "78d3e7b1-7258-4508-ad38-373f21bb5d96"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 06:56:33 crc kubenswrapper[4846]: I1005 06:56:33.397825 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/78d3e7b1-7258-4508-ad38-373f21bb5d96-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "78d3e7b1-7258-4508-ad38-373f21bb5d96" (UID: "78d3e7b1-7258-4508-ad38-373f21bb5d96"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 06:56:33 crc kubenswrapper[4846]: I1005 06:56:33.400822 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "78d3e7b1-7258-4508-ad38-373f21bb5d96" (UID: "78d3e7b1-7258-4508-ad38-373f21bb5d96"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Oct 05 06:56:33 crc kubenswrapper[4846]: I1005 06:56:33.481697 4846 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/78d3e7b1-7258-4508-ad38-373f21bb5d96-bound-sa-token\") on node \"crc\" DevicePath \"\"" Oct 05 06:56:33 crc kubenswrapper[4846]: I1005 06:56:33.481731 4846 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/78d3e7b1-7258-4508-ad38-373f21bb5d96-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Oct 05 06:56:33 crc kubenswrapper[4846]: I1005 06:56:33.481741 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mtl7j\" (UniqueName: \"kubernetes.io/projected/78d3e7b1-7258-4508-ad38-373f21bb5d96-kube-api-access-mtl7j\") on node \"crc\" DevicePath \"\"" Oct 05 06:56:33 crc kubenswrapper[4846]: I1005 06:56:33.481754 4846 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/78d3e7b1-7258-4508-ad38-373f21bb5d96-registry-tls\") on node \"crc\" DevicePath \"\"" Oct 05 06:56:33 crc kubenswrapper[4846]: I1005 06:56:33.481765 4846 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/78d3e7b1-7258-4508-ad38-373f21bb5d96-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Oct 05 06:56:33 crc kubenswrapper[4846]: I1005 06:56:33.481776 4846 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/78d3e7b1-7258-4508-ad38-373f21bb5d96-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 05 06:56:33 crc kubenswrapper[4846]: I1005 06:56:33.508395 4846 generic.go:334] "Generic (PLEG): container finished" podID="78d3e7b1-7258-4508-ad38-373f21bb5d96" containerID="3edcc13d0073ef87f5bc93cf466592a38ca27007b40995acebf6d1c84edde195" exitCode=0 Oct 05 06:56:33 crc kubenswrapper[4846]: I1005 06:56:33.508474 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-rf9l4" event={"ID":"78d3e7b1-7258-4508-ad38-373f21bb5d96","Type":"ContainerDied","Data":"3edcc13d0073ef87f5bc93cf466592a38ca27007b40995acebf6d1c84edde195"} Oct 05 06:56:33 crc kubenswrapper[4846]: I1005 06:56:33.508526 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-rf9l4" event={"ID":"78d3e7b1-7258-4508-ad38-373f21bb5d96","Type":"ContainerDied","Data":"678cdf83030f0336db44da72ff49da71992545f53d49850a215977954586614c"} Oct 05 06:56:33 crc kubenswrapper[4846]: I1005 06:56:33.508555 4846 scope.go:117] "RemoveContainer" containerID="3edcc13d0073ef87f5bc93cf466592a38ca27007b40995acebf6d1c84edde195" Oct 05 06:56:33 crc kubenswrapper[4846]: I1005 06:56:33.509125 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-rf9l4" Oct 05 06:56:33 crc kubenswrapper[4846]: I1005 06:56:33.532127 4846 scope.go:117] "RemoveContainer" containerID="3edcc13d0073ef87f5bc93cf466592a38ca27007b40995acebf6d1c84edde195" Oct 05 06:56:33 crc kubenswrapper[4846]: E1005 06:56:33.532924 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3edcc13d0073ef87f5bc93cf466592a38ca27007b40995acebf6d1c84edde195\": container with ID starting with 3edcc13d0073ef87f5bc93cf466592a38ca27007b40995acebf6d1c84edde195 not found: ID does not exist" containerID="3edcc13d0073ef87f5bc93cf466592a38ca27007b40995acebf6d1c84edde195" Oct 05 06:56:33 crc kubenswrapper[4846]: I1005 06:56:33.533033 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3edcc13d0073ef87f5bc93cf466592a38ca27007b40995acebf6d1c84edde195"} err="failed to get container status \"3edcc13d0073ef87f5bc93cf466592a38ca27007b40995acebf6d1c84edde195\": rpc error: code = NotFound desc = could not find container \"3edcc13d0073ef87f5bc93cf466592a38ca27007b40995acebf6d1c84edde195\": container with ID starting with 3edcc13d0073ef87f5bc93cf466592a38ca27007b40995acebf6d1c84edde195 not found: ID does not exist" Oct 05 06:56:33 crc kubenswrapper[4846]: I1005 06:56:33.573597 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-rf9l4"] Oct 05 06:56:33 crc kubenswrapper[4846]: I1005 06:56:33.586567 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-rf9l4"] Oct 05 06:56:34 crc kubenswrapper[4846]: I1005 06:56:34.509675 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="78d3e7b1-7258-4508-ad38-373f21bb5d96" path="/var/lib/kubelet/pods/78d3e7b1-7258-4508-ad38-373f21bb5d96/volumes" Oct 05 06:57:23 crc kubenswrapper[4846]: I1005 06:57:23.325439 4846 patch_prober.go:28] interesting pod/machine-config-daemon-fscvf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 06:57:23 crc kubenswrapper[4846]: I1005 06:57:23.326617 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 06:57:53 crc kubenswrapper[4846]: I1005 06:57:53.325892 4846 patch_prober.go:28] interesting pod/machine-config-daemon-fscvf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 06:57:53 crc kubenswrapper[4846]: I1005 06:57:53.326839 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 06:58:23 crc kubenswrapper[4846]: I1005 06:58:23.325459 4846 patch_prober.go:28] interesting pod/machine-config-daemon-fscvf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 06:58:23 crc kubenswrapper[4846]: I1005 06:58:23.327814 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 06:58:23 crc kubenswrapper[4846]: I1005 06:58:23.328129 4846 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" Oct 05 06:58:23 crc kubenswrapper[4846]: I1005 06:58:23.329463 4846 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"3b7026879bf093ff49269f1b47a66adf5f31a2fca4f61fcc9b2ba0d01156c507"} pod="openshift-machine-config-operator/machine-config-daemon-fscvf" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 05 06:58:23 crc kubenswrapper[4846]: I1005 06:58:23.329733 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" containerName="machine-config-daemon" containerID="cri-o://3b7026879bf093ff49269f1b47a66adf5f31a2fca4f61fcc9b2ba0d01156c507" gracePeriod=600 Oct 05 06:58:24 crc kubenswrapper[4846]: I1005 06:58:24.389707 4846 generic.go:334] "Generic (PLEG): container finished" podID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" containerID="3b7026879bf093ff49269f1b47a66adf5f31a2fca4f61fcc9b2ba0d01156c507" exitCode=0 Oct 05 06:58:24 crc kubenswrapper[4846]: I1005 06:58:24.389784 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" event={"ID":"ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7","Type":"ContainerDied","Data":"3b7026879bf093ff49269f1b47a66adf5f31a2fca4f61fcc9b2ba0d01156c507"} Oct 05 06:58:24 crc kubenswrapper[4846]: I1005 06:58:24.390641 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" event={"ID":"ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7","Type":"ContainerStarted","Data":"468d3aa8613beb034776fd6c8b08e9b13031340dca2680a10051a40fbbb92cf2"} Oct 05 06:58:24 crc kubenswrapper[4846]: I1005 06:58:24.390675 4846 scope.go:117] "RemoveContainer" containerID="555aa2226a521510ee0c88d4d95100dced793271cffe5d0ca2a8cf181778e4a6" Oct 05 06:59:42 crc kubenswrapper[4846]: I1005 06:59:42.341046 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-txzlk"] Oct 05 06:59:42 crc kubenswrapper[4846]: I1005 06:59:42.342352 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-txzlk" podUID="93a6ace4-8d64-44d7-9637-457d3af6543c" containerName="ovn-controller" containerID="cri-o://1409651de8c897804f87be03d016c138f8acf27e662dce265a93e15c1848adcf" gracePeriod=30 Oct 05 06:59:42 crc kubenswrapper[4846]: I1005 06:59:42.342446 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-txzlk" podUID="93a6ace4-8d64-44d7-9637-457d3af6543c" containerName="sbdb" containerID="cri-o://0df0ebcd87dc5af4b17736b43ec7b2395968b19ed16e775366af14e2b2d49c52" gracePeriod=30 Oct 05 06:59:42 crc kubenswrapper[4846]: I1005 06:59:42.342505 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-txzlk" podUID="93a6ace4-8d64-44d7-9637-457d3af6543c" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://0a6f6b2c32dd3da5f9fd5e6971c1773e81a9a24184b7aa5b543cec879001ae53" gracePeriod=30 Oct 05 06:59:42 crc kubenswrapper[4846]: I1005 06:59:42.342588 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-txzlk" podUID="93a6ace4-8d64-44d7-9637-457d3af6543c" containerName="kube-rbac-proxy-node" containerID="cri-o://1ad1940d07e37217089642478ece9ccdaca74784738a400929a9f5359d158c01" gracePeriod=30 Oct 05 06:59:42 crc kubenswrapper[4846]: I1005 06:59:42.342633 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-txzlk" podUID="93a6ace4-8d64-44d7-9637-457d3af6543c" containerName="ovn-acl-logging" containerID="cri-o://8da85222775abc7a772f6a8ff8255f1a0d5b15807ba177a7e3e2b2d01a4f24b9" gracePeriod=30 Oct 05 06:59:42 crc kubenswrapper[4846]: I1005 06:59:42.342796 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-txzlk" podUID="93a6ace4-8d64-44d7-9637-457d3af6543c" containerName="nbdb" containerID="cri-o://9e9b0219ecbc7e60905d6de3e67a2c8dc9af1017b4aed9538ee0c58c861f2438" gracePeriod=30 Oct 05 06:59:42 crc kubenswrapper[4846]: I1005 06:59:42.342884 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-txzlk" podUID="93a6ace4-8d64-44d7-9637-457d3af6543c" containerName="northd" containerID="cri-o://af5e5a4da9ad7371d5a24633e8268f12e3a7d419a8c93b11ddbcc006d8339c6c" gracePeriod=30 Oct 05 06:59:42 crc kubenswrapper[4846]: I1005 06:59:42.417925 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-txzlk" podUID="93a6ace4-8d64-44d7-9637-457d3af6543c" containerName="ovnkube-controller" containerID="cri-o://32a57312adddc39f79d4dbc9dfb0197ec33241adfb654ed64db6bc10d1d479c8" gracePeriod=30 Oct 05 06:59:42 crc kubenswrapper[4846]: I1005 06:59:42.697823 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-txzlk_93a6ace4-8d64-44d7-9637-457d3af6543c/ovnkube-controller/3.log" Oct 05 06:59:42 crc kubenswrapper[4846]: I1005 06:59:42.700055 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-txzlk_93a6ace4-8d64-44d7-9637-457d3af6543c/ovn-acl-logging/0.log" Oct 05 06:59:42 crc kubenswrapper[4846]: I1005 06:59:42.700537 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-txzlk_93a6ace4-8d64-44d7-9637-457d3af6543c/ovn-controller/0.log" Oct 05 06:59:42 crc kubenswrapper[4846]: I1005 06:59:42.701040 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-txzlk" Oct 05 06:59:42 crc kubenswrapper[4846]: I1005 06:59:42.807379 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-q224c"] Oct 05 06:59:42 crc kubenswrapper[4846]: E1005 06:59:42.807962 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="93a6ace4-8d64-44d7-9637-457d3af6543c" containerName="ovn-acl-logging" Oct 05 06:59:42 crc kubenswrapper[4846]: I1005 06:59:42.807977 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="93a6ace4-8d64-44d7-9637-457d3af6543c" containerName="ovn-acl-logging" Oct 05 06:59:42 crc kubenswrapper[4846]: E1005 06:59:42.807991 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="93a6ace4-8d64-44d7-9637-457d3af6543c" containerName="kubecfg-setup" Oct 05 06:59:42 crc kubenswrapper[4846]: I1005 06:59:42.807998 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="93a6ace4-8d64-44d7-9637-457d3af6543c" containerName="kubecfg-setup" Oct 05 06:59:42 crc kubenswrapper[4846]: E1005 06:59:42.808008 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="93a6ace4-8d64-44d7-9637-457d3af6543c" containerName="ovn-controller" Oct 05 06:59:42 crc kubenswrapper[4846]: I1005 06:59:42.808014 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="93a6ace4-8d64-44d7-9637-457d3af6543c" containerName="ovn-controller" Oct 05 06:59:42 crc kubenswrapper[4846]: E1005 06:59:42.808024 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="93a6ace4-8d64-44d7-9637-457d3af6543c" containerName="ovnkube-controller" Oct 05 06:59:42 crc kubenswrapper[4846]: I1005 06:59:42.808029 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="93a6ace4-8d64-44d7-9637-457d3af6543c" containerName="ovnkube-controller" Oct 05 06:59:42 crc kubenswrapper[4846]: E1005 06:59:42.808036 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="78d3e7b1-7258-4508-ad38-373f21bb5d96" containerName="registry" Oct 05 06:59:42 crc kubenswrapper[4846]: I1005 06:59:42.808043 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="78d3e7b1-7258-4508-ad38-373f21bb5d96" containerName="registry" Oct 05 06:59:42 crc kubenswrapper[4846]: E1005 06:59:42.808056 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="93a6ace4-8d64-44d7-9637-457d3af6543c" containerName="ovnkube-controller" Oct 05 06:59:42 crc kubenswrapper[4846]: I1005 06:59:42.808064 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="93a6ace4-8d64-44d7-9637-457d3af6543c" containerName="ovnkube-controller" Oct 05 06:59:42 crc kubenswrapper[4846]: E1005 06:59:42.808073 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="93a6ace4-8d64-44d7-9637-457d3af6543c" containerName="ovnkube-controller" Oct 05 06:59:42 crc kubenswrapper[4846]: I1005 06:59:42.808079 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="93a6ace4-8d64-44d7-9637-457d3af6543c" containerName="ovnkube-controller" Oct 05 06:59:42 crc kubenswrapper[4846]: E1005 06:59:42.808090 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="93a6ace4-8d64-44d7-9637-457d3af6543c" containerName="sbdb" Oct 05 06:59:42 crc kubenswrapper[4846]: I1005 06:59:42.808096 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="93a6ace4-8d64-44d7-9637-457d3af6543c" containerName="sbdb" Oct 05 06:59:42 crc kubenswrapper[4846]: E1005 06:59:42.808106 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="93a6ace4-8d64-44d7-9637-457d3af6543c" containerName="kube-rbac-proxy-ovn-metrics" Oct 05 06:59:42 crc kubenswrapper[4846]: I1005 06:59:42.808112 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="93a6ace4-8d64-44d7-9637-457d3af6543c" containerName="kube-rbac-proxy-ovn-metrics" Oct 05 06:59:42 crc kubenswrapper[4846]: E1005 06:59:42.808123 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="93a6ace4-8d64-44d7-9637-457d3af6543c" containerName="kube-rbac-proxy-node" Oct 05 06:59:42 crc kubenswrapper[4846]: I1005 06:59:42.808132 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="93a6ace4-8d64-44d7-9637-457d3af6543c" containerName="kube-rbac-proxy-node" Oct 05 06:59:42 crc kubenswrapper[4846]: E1005 06:59:42.808140 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="93a6ace4-8d64-44d7-9637-457d3af6543c" containerName="northd" Oct 05 06:59:42 crc kubenswrapper[4846]: I1005 06:59:42.808146 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="93a6ace4-8d64-44d7-9637-457d3af6543c" containerName="northd" Oct 05 06:59:42 crc kubenswrapper[4846]: E1005 06:59:42.808153 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="93a6ace4-8d64-44d7-9637-457d3af6543c" containerName="nbdb" Oct 05 06:59:42 crc kubenswrapper[4846]: I1005 06:59:42.808158 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="93a6ace4-8d64-44d7-9637-457d3af6543c" containerName="nbdb" Oct 05 06:59:42 crc kubenswrapper[4846]: I1005 06:59:42.808272 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="93a6ace4-8d64-44d7-9637-457d3af6543c" containerName="ovn-controller" Oct 05 06:59:42 crc kubenswrapper[4846]: I1005 06:59:42.808286 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="93a6ace4-8d64-44d7-9637-457d3af6543c" containerName="nbdb" Oct 05 06:59:42 crc kubenswrapper[4846]: I1005 06:59:42.808292 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="93a6ace4-8d64-44d7-9637-457d3af6543c" containerName="ovnkube-controller" Oct 05 06:59:42 crc kubenswrapper[4846]: I1005 06:59:42.808299 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="93a6ace4-8d64-44d7-9637-457d3af6543c" containerName="ovnkube-controller" Oct 05 06:59:42 crc kubenswrapper[4846]: I1005 06:59:42.808308 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="78d3e7b1-7258-4508-ad38-373f21bb5d96" containerName="registry" Oct 05 06:59:42 crc kubenswrapper[4846]: I1005 06:59:42.808316 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="93a6ace4-8d64-44d7-9637-457d3af6543c" containerName="ovn-acl-logging" Oct 05 06:59:42 crc kubenswrapper[4846]: I1005 06:59:42.808324 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="93a6ace4-8d64-44d7-9637-457d3af6543c" containerName="ovnkube-controller" Oct 05 06:59:42 crc kubenswrapper[4846]: I1005 06:59:42.808333 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="93a6ace4-8d64-44d7-9637-457d3af6543c" containerName="northd" Oct 05 06:59:42 crc kubenswrapper[4846]: I1005 06:59:42.808342 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="93a6ace4-8d64-44d7-9637-457d3af6543c" containerName="kube-rbac-proxy-node" Oct 05 06:59:42 crc kubenswrapper[4846]: I1005 06:59:42.808348 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="93a6ace4-8d64-44d7-9637-457d3af6543c" containerName="kube-rbac-proxy-ovn-metrics" Oct 05 06:59:42 crc kubenswrapper[4846]: I1005 06:59:42.808357 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="93a6ace4-8d64-44d7-9637-457d3af6543c" containerName="sbdb" Oct 05 06:59:42 crc kubenswrapper[4846]: E1005 06:59:42.808446 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="93a6ace4-8d64-44d7-9637-457d3af6543c" containerName="ovnkube-controller" Oct 05 06:59:42 crc kubenswrapper[4846]: I1005 06:59:42.808454 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="93a6ace4-8d64-44d7-9637-457d3af6543c" containerName="ovnkube-controller" Oct 05 06:59:42 crc kubenswrapper[4846]: I1005 06:59:42.808548 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="93a6ace4-8d64-44d7-9637-457d3af6543c" containerName="ovnkube-controller" Oct 05 06:59:42 crc kubenswrapper[4846]: I1005 06:59:42.808555 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="93a6ace4-8d64-44d7-9637-457d3af6543c" containerName="ovnkube-controller" Oct 05 06:59:42 crc kubenswrapper[4846]: E1005 06:59:42.808635 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="93a6ace4-8d64-44d7-9637-457d3af6543c" containerName="ovnkube-controller" Oct 05 06:59:42 crc kubenswrapper[4846]: I1005 06:59:42.808641 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="93a6ace4-8d64-44d7-9637-457d3af6543c" containerName="ovnkube-controller" Oct 05 06:59:42 crc kubenswrapper[4846]: I1005 06:59:42.818918 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-q224c" Oct 05 06:59:42 crc kubenswrapper[4846]: I1005 06:59:42.871100 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/93a6ace4-8d64-44d7-9637-457d3af6543c-host-run-netns\") pod \"93a6ace4-8d64-44d7-9637-457d3af6543c\" (UID: \"93a6ace4-8d64-44d7-9637-457d3af6543c\") " Oct 05 06:59:42 crc kubenswrapper[4846]: I1005 06:59:42.871150 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/93a6ace4-8d64-44d7-9637-457d3af6543c-host-slash\") pod \"93a6ace4-8d64-44d7-9637-457d3af6543c\" (UID: \"93a6ace4-8d64-44d7-9637-457d3af6543c\") " Oct 05 06:59:42 crc kubenswrapper[4846]: I1005 06:59:42.871185 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/93a6ace4-8d64-44d7-9637-457d3af6543c-host-var-lib-cni-networks-ovn-kubernetes\") pod \"93a6ace4-8d64-44d7-9637-457d3af6543c\" (UID: \"93a6ace4-8d64-44d7-9637-457d3af6543c\") " Oct 05 06:59:42 crc kubenswrapper[4846]: I1005 06:59:42.871233 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/93a6ace4-8d64-44d7-9637-457d3af6543c-host-kubelet\") pod \"93a6ace4-8d64-44d7-9637-457d3af6543c\" (UID: \"93a6ace4-8d64-44d7-9637-457d3af6543c\") " Oct 05 06:59:42 crc kubenswrapper[4846]: I1005 06:59:42.871271 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/93a6ace4-8d64-44d7-9637-457d3af6543c-ovnkube-script-lib\") pod \"93a6ace4-8d64-44d7-9637-457d3af6543c\" (UID: \"93a6ace4-8d64-44d7-9637-457d3af6543c\") " Oct 05 06:59:42 crc kubenswrapper[4846]: I1005 06:59:42.871300 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/93a6ace4-8d64-44d7-9637-457d3af6543c-etc-openvswitch\") pod \"93a6ace4-8d64-44d7-9637-457d3af6543c\" (UID: \"93a6ace4-8d64-44d7-9637-457d3af6543c\") " Oct 05 06:59:42 crc kubenswrapper[4846]: I1005 06:59:42.871336 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/93a6ace4-8d64-44d7-9637-457d3af6543c-ovnkube-config\") pod \"93a6ace4-8d64-44d7-9637-457d3af6543c\" (UID: \"93a6ace4-8d64-44d7-9637-457d3af6543c\") " Oct 05 06:59:42 crc kubenswrapper[4846]: I1005 06:59:42.871345 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/93a6ace4-8d64-44d7-9637-457d3af6543c-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "93a6ace4-8d64-44d7-9637-457d3af6543c" (UID: "93a6ace4-8d64-44d7-9637-457d3af6543c"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 05 06:59:42 crc kubenswrapper[4846]: I1005 06:59:42.871428 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/93a6ace4-8d64-44d7-9637-457d3af6543c-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "93a6ace4-8d64-44d7-9637-457d3af6543c" (UID: "93a6ace4-8d64-44d7-9637-457d3af6543c"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 05 06:59:42 crc kubenswrapper[4846]: I1005 06:59:42.871437 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/93a6ace4-8d64-44d7-9637-457d3af6543c-run-systemd\") pod \"93a6ace4-8d64-44d7-9637-457d3af6543c\" (UID: \"93a6ace4-8d64-44d7-9637-457d3af6543c\") " Oct 05 06:59:42 crc kubenswrapper[4846]: I1005 06:59:42.871454 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/93a6ace4-8d64-44d7-9637-457d3af6543c-host-slash" (OuterVolumeSpecName: "host-slash") pod "93a6ace4-8d64-44d7-9637-457d3af6543c" (UID: "93a6ace4-8d64-44d7-9637-457d3af6543c"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 05 06:59:42 crc kubenswrapper[4846]: I1005 06:59:42.871475 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/93a6ace4-8d64-44d7-9637-457d3af6543c-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "93a6ace4-8d64-44d7-9637-457d3af6543c" (UID: "93a6ace4-8d64-44d7-9637-457d3af6543c"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 05 06:59:42 crc kubenswrapper[4846]: I1005 06:59:42.871525 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/93a6ace4-8d64-44d7-9637-457d3af6543c-ovn-node-metrics-cert\") pod \"93a6ace4-8d64-44d7-9637-457d3af6543c\" (UID: \"93a6ace4-8d64-44d7-9637-457d3af6543c\") " Oct 05 06:59:42 crc kubenswrapper[4846]: I1005 06:59:42.871844 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/93a6ace4-8d64-44d7-9637-457d3af6543c-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "93a6ace4-8d64-44d7-9637-457d3af6543c" (UID: "93a6ace4-8d64-44d7-9637-457d3af6543c"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 05 06:59:42 crc kubenswrapper[4846]: I1005 06:59:42.871894 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/93a6ace4-8d64-44d7-9637-457d3af6543c-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "93a6ace4-8d64-44d7-9637-457d3af6543c" (UID: "93a6ace4-8d64-44d7-9637-457d3af6543c"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 06:59:42 crc kubenswrapper[4846]: I1005 06:59:42.872040 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/93a6ace4-8d64-44d7-9637-457d3af6543c-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "93a6ace4-8d64-44d7-9637-457d3af6543c" (UID: "93a6ace4-8d64-44d7-9637-457d3af6543c"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 06:59:42 crc kubenswrapper[4846]: I1005 06:59:42.872233 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/93a6ace4-8d64-44d7-9637-457d3af6543c-systemd-units\") pod \"93a6ace4-8d64-44d7-9637-457d3af6543c\" (UID: \"93a6ace4-8d64-44d7-9637-457d3af6543c\") " Oct 05 06:59:42 crc kubenswrapper[4846]: I1005 06:59:42.872376 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/93a6ace4-8d64-44d7-9637-457d3af6543c-run-openvswitch\") pod \"93a6ace4-8d64-44d7-9637-457d3af6543c\" (UID: \"93a6ace4-8d64-44d7-9637-457d3af6543c\") " Oct 05 06:59:42 crc kubenswrapper[4846]: I1005 06:59:42.872489 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/93a6ace4-8d64-44d7-9637-457d3af6543c-run-ovn\") pod \"93a6ace4-8d64-44d7-9637-457d3af6543c\" (UID: \"93a6ace4-8d64-44d7-9637-457d3af6543c\") " Oct 05 06:59:42 crc kubenswrapper[4846]: I1005 06:59:42.872691 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/93a6ace4-8d64-44d7-9637-457d3af6543c-host-cni-netd\") pod \"93a6ace4-8d64-44d7-9637-457d3af6543c\" (UID: \"93a6ace4-8d64-44d7-9637-457d3af6543c\") " Oct 05 06:59:42 crc kubenswrapper[4846]: I1005 06:59:42.872841 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/93a6ace4-8d64-44d7-9637-457d3af6543c-env-overrides\") pod \"93a6ace4-8d64-44d7-9637-457d3af6543c\" (UID: \"93a6ace4-8d64-44d7-9637-457d3af6543c\") " Oct 05 06:59:42 crc kubenswrapper[4846]: I1005 06:59:42.872957 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/93a6ace4-8d64-44d7-9637-457d3af6543c-var-lib-openvswitch\") pod \"93a6ace4-8d64-44d7-9637-457d3af6543c\" (UID: \"93a6ace4-8d64-44d7-9637-457d3af6543c\") " Oct 05 06:59:42 crc kubenswrapper[4846]: I1005 06:59:42.872320 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/93a6ace4-8d64-44d7-9637-457d3af6543c-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "93a6ace4-8d64-44d7-9637-457d3af6543c" (UID: "93a6ace4-8d64-44d7-9637-457d3af6543c"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 05 06:59:42 crc kubenswrapper[4846]: I1005 06:59:42.872455 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/93a6ace4-8d64-44d7-9637-457d3af6543c-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "93a6ace4-8d64-44d7-9637-457d3af6543c" (UID: "93a6ace4-8d64-44d7-9637-457d3af6543c"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 05 06:59:42 crc kubenswrapper[4846]: I1005 06:59:42.872597 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/93a6ace4-8d64-44d7-9637-457d3af6543c-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "93a6ace4-8d64-44d7-9637-457d3af6543c" (UID: "93a6ace4-8d64-44d7-9637-457d3af6543c"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 05 06:59:42 crc kubenswrapper[4846]: I1005 06:59:42.872886 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/93a6ace4-8d64-44d7-9637-457d3af6543c-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "93a6ace4-8d64-44d7-9637-457d3af6543c" (UID: "93a6ace4-8d64-44d7-9637-457d3af6543c"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 05 06:59:42 crc kubenswrapper[4846]: I1005 06:59:42.873058 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/93a6ace4-8d64-44d7-9637-457d3af6543c-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "93a6ace4-8d64-44d7-9637-457d3af6543c" (UID: "93a6ace4-8d64-44d7-9637-457d3af6543c"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 05 06:59:42 crc kubenswrapper[4846]: I1005 06:59:42.873085 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-96h7b\" (UniqueName: \"kubernetes.io/projected/93a6ace4-8d64-44d7-9637-457d3af6543c-kube-api-access-96h7b\") pod \"93a6ace4-8d64-44d7-9637-457d3af6543c\" (UID: \"93a6ace4-8d64-44d7-9637-457d3af6543c\") " Oct 05 06:59:42 crc kubenswrapper[4846]: I1005 06:59:42.873226 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/93a6ace4-8d64-44d7-9637-457d3af6543c-log-socket\") pod \"93a6ace4-8d64-44d7-9637-457d3af6543c\" (UID: \"93a6ace4-8d64-44d7-9637-457d3af6543c\") " Oct 05 06:59:42 crc kubenswrapper[4846]: I1005 06:59:42.873257 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/93a6ace4-8d64-44d7-9637-457d3af6543c-host-cni-bin\") pod \"93a6ace4-8d64-44d7-9637-457d3af6543c\" (UID: \"93a6ace4-8d64-44d7-9637-457d3af6543c\") " Oct 05 06:59:42 crc kubenswrapper[4846]: I1005 06:59:42.873269 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/93a6ace4-8d64-44d7-9637-457d3af6543c-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "93a6ace4-8d64-44d7-9637-457d3af6543c" (UID: "93a6ace4-8d64-44d7-9637-457d3af6543c"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 06:59:42 crc kubenswrapper[4846]: I1005 06:59:42.873301 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/93a6ace4-8d64-44d7-9637-457d3af6543c-host-run-ovn-kubernetes\") pod \"93a6ace4-8d64-44d7-9637-457d3af6543c\" (UID: \"93a6ace4-8d64-44d7-9637-457d3af6543c\") " Oct 05 06:59:42 crc kubenswrapper[4846]: I1005 06:59:42.873306 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/93a6ace4-8d64-44d7-9637-457d3af6543c-log-socket" (OuterVolumeSpecName: "log-socket") pod "93a6ace4-8d64-44d7-9637-457d3af6543c" (UID: "93a6ace4-8d64-44d7-9637-457d3af6543c"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 05 06:59:42 crc kubenswrapper[4846]: I1005 06:59:42.873325 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/93a6ace4-8d64-44d7-9637-457d3af6543c-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "93a6ace4-8d64-44d7-9637-457d3af6543c" (UID: "93a6ace4-8d64-44d7-9637-457d3af6543c"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 05 06:59:42 crc kubenswrapper[4846]: I1005 06:59:42.873334 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/93a6ace4-8d64-44d7-9637-457d3af6543c-node-log\") pod \"93a6ace4-8d64-44d7-9637-457d3af6543c\" (UID: \"93a6ace4-8d64-44d7-9637-457d3af6543c\") " Oct 05 06:59:42 crc kubenswrapper[4846]: I1005 06:59:42.873366 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/93a6ace4-8d64-44d7-9637-457d3af6543c-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "93a6ace4-8d64-44d7-9637-457d3af6543c" (UID: "93a6ace4-8d64-44d7-9637-457d3af6543c"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 05 06:59:42 crc kubenswrapper[4846]: I1005 06:59:42.873430 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/93a6ace4-8d64-44d7-9637-457d3af6543c-node-log" (OuterVolumeSpecName: "node-log") pod "93a6ace4-8d64-44d7-9637-457d3af6543c" (UID: "93a6ace4-8d64-44d7-9637-457d3af6543c"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 05 06:59:42 crc kubenswrapper[4846]: I1005 06:59:42.873455 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/778242f5-89b8-4f86-9f58-0a81e2d70569-host-kubelet\") pod \"ovnkube-node-q224c\" (UID: \"778242f5-89b8-4f86-9f58-0a81e2d70569\") " pod="openshift-ovn-kubernetes/ovnkube-node-q224c" Oct 05 06:59:42 crc kubenswrapper[4846]: I1005 06:59:42.873486 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/778242f5-89b8-4f86-9f58-0a81e2d70569-run-openvswitch\") pod \"ovnkube-node-q224c\" (UID: \"778242f5-89b8-4f86-9f58-0a81e2d70569\") " pod="openshift-ovn-kubernetes/ovnkube-node-q224c" Oct 05 06:59:42 crc kubenswrapper[4846]: I1005 06:59:42.873515 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tscwh\" (UniqueName: \"kubernetes.io/projected/778242f5-89b8-4f86-9f58-0a81e2d70569-kube-api-access-tscwh\") pod \"ovnkube-node-q224c\" (UID: \"778242f5-89b8-4f86-9f58-0a81e2d70569\") " pod="openshift-ovn-kubernetes/ovnkube-node-q224c" Oct 05 06:59:42 crc kubenswrapper[4846]: I1005 06:59:42.873579 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/778242f5-89b8-4f86-9f58-0a81e2d70569-log-socket\") pod \"ovnkube-node-q224c\" (UID: \"778242f5-89b8-4f86-9f58-0a81e2d70569\") " pod="openshift-ovn-kubernetes/ovnkube-node-q224c" Oct 05 06:59:42 crc kubenswrapper[4846]: I1005 06:59:42.873615 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/778242f5-89b8-4f86-9f58-0a81e2d70569-node-log\") pod \"ovnkube-node-q224c\" (UID: \"778242f5-89b8-4f86-9f58-0a81e2d70569\") " pod="openshift-ovn-kubernetes/ovnkube-node-q224c" Oct 05 06:59:42 crc kubenswrapper[4846]: I1005 06:59:42.873645 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/778242f5-89b8-4f86-9f58-0a81e2d70569-ovn-node-metrics-cert\") pod \"ovnkube-node-q224c\" (UID: \"778242f5-89b8-4f86-9f58-0a81e2d70569\") " pod="openshift-ovn-kubernetes/ovnkube-node-q224c" Oct 05 06:59:42 crc kubenswrapper[4846]: I1005 06:59:42.873674 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/778242f5-89b8-4f86-9f58-0a81e2d70569-env-overrides\") pod \"ovnkube-node-q224c\" (UID: \"778242f5-89b8-4f86-9f58-0a81e2d70569\") " pod="openshift-ovn-kubernetes/ovnkube-node-q224c" Oct 05 06:59:42 crc kubenswrapper[4846]: I1005 06:59:42.873731 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/778242f5-89b8-4f86-9f58-0a81e2d70569-run-ovn\") pod \"ovnkube-node-q224c\" (UID: \"778242f5-89b8-4f86-9f58-0a81e2d70569\") " pod="openshift-ovn-kubernetes/ovnkube-node-q224c" Oct 05 06:59:42 crc kubenswrapper[4846]: I1005 06:59:42.873755 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/778242f5-89b8-4f86-9f58-0a81e2d70569-host-slash\") pod \"ovnkube-node-q224c\" (UID: \"778242f5-89b8-4f86-9f58-0a81e2d70569\") " pod="openshift-ovn-kubernetes/ovnkube-node-q224c" Oct 05 06:59:42 crc kubenswrapper[4846]: I1005 06:59:42.873775 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/778242f5-89b8-4f86-9f58-0a81e2d70569-host-cni-netd\") pod \"ovnkube-node-q224c\" (UID: \"778242f5-89b8-4f86-9f58-0a81e2d70569\") " pod="openshift-ovn-kubernetes/ovnkube-node-q224c" Oct 05 06:59:42 crc kubenswrapper[4846]: I1005 06:59:42.873798 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/778242f5-89b8-4f86-9f58-0a81e2d70569-systemd-units\") pod \"ovnkube-node-q224c\" (UID: \"778242f5-89b8-4f86-9f58-0a81e2d70569\") " pod="openshift-ovn-kubernetes/ovnkube-node-q224c" Oct 05 06:59:42 crc kubenswrapper[4846]: I1005 06:59:42.873826 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/778242f5-89b8-4f86-9f58-0a81e2d70569-var-lib-openvswitch\") pod \"ovnkube-node-q224c\" (UID: \"778242f5-89b8-4f86-9f58-0a81e2d70569\") " pod="openshift-ovn-kubernetes/ovnkube-node-q224c" Oct 05 06:59:42 crc kubenswrapper[4846]: I1005 06:59:42.873899 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/778242f5-89b8-4f86-9f58-0a81e2d70569-ovnkube-config\") pod \"ovnkube-node-q224c\" (UID: \"778242f5-89b8-4f86-9f58-0a81e2d70569\") " pod="openshift-ovn-kubernetes/ovnkube-node-q224c" Oct 05 06:59:42 crc kubenswrapper[4846]: I1005 06:59:42.873969 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/778242f5-89b8-4f86-9f58-0a81e2d70569-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-q224c\" (UID: \"778242f5-89b8-4f86-9f58-0a81e2d70569\") " pod="openshift-ovn-kubernetes/ovnkube-node-q224c" Oct 05 06:59:42 crc kubenswrapper[4846]: I1005 06:59:42.873996 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/778242f5-89b8-4f86-9f58-0a81e2d70569-host-run-netns\") pod \"ovnkube-node-q224c\" (UID: \"778242f5-89b8-4f86-9f58-0a81e2d70569\") " pod="openshift-ovn-kubernetes/ovnkube-node-q224c" Oct 05 06:59:42 crc kubenswrapper[4846]: I1005 06:59:42.874023 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/778242f5-89b8-4f86-9f58-0a81e2d70569-host-cni-bin\") pod \"ovnkube-node-q224c\" (UID: \"778242f5-89b8-4f86-9f58-0a81e2d70569\") " pod="openshift-ovn-kubernetes/ovnkube-node-q224c" Oct 05 06:59:42 crc kubenswrapper[4846]: I1005 06:59:42.874047 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/778242f5-89b8-4f86-9f58-0a81e2d70569-ovnkube-script-lib\") pod \"ovnkube-node-q224c\" (UID: \"778242f5-89b8-4f86-9f58-0a81e2d70569\") " pod="openshift-ovn-kubernetes/ovnkube-node-q224c" Oct 05 06:59:42 crc kubenswrapper[4846]: I1005 06:59:42.874078 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/778242f5-89b8-4f86-9f58-0a81e2d70569-etc-openvswitch\") pod \"ovnkube-node-q224c\" (UID: \"778242f5-89b8-4f86-9f58-0a81e2d70569\") " pod="openshift-ovn-kubernetes/ovnkube-node-q224c" Oct 05 06:59:42 crc kubenswrapper[4846]: I1005 06:59:42.874103 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/778242f5-89b8-4f86-9f58-0a81e2d70569-host-run-ovn-kubernetes\") pod \"ovnkube-node-q224c\" (UID: \"778242f5-89b8-4f86-9f58-0a81e2d70569\") " pod="openshift-ovn-kubernetes/ovnkube-node-q224c" Oct 05 06:59:42 crc kubenswrapper[4846]: I1005 06:59:42.874124 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/778242f5-89b8-4f86-9f58-0a81e2d70569-run-systemd\") pod \"ovnkube-node-q224c\" (UID: \"778242f5-89b8-4f86-9f58-0a81e2d70569\") " pod="openshift-ovn-kubernetes/ovnkube-node-q224c" Oct 05 06:59:42 crc kubenswrapper[4846]: I1005 06:59:42.874167 4846 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/93a6ace4-8d64-44d7-9637-457d3af6543c-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Oct 05 06:59:42 crc kubenswrapper[4846]: I1005 06:59:42.874197 4846 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/93a6ace4-8d64-44d7-9637-457d3af6543c-ovnkube-config\") on node \"crc\" DevicePath \"\"" Oct 05 06:59:42 crc kubenswrapper[4846]: I1005 06:59:42.874209 4846 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/93a6ace4-8d64-44d7-9637-457d3af6543c-systemd-units\") on node \"crc\" DevicePath \"\"" Oct 05 06:59:42 crc kubenswrapper[4846]: I1005 06:59:42.874217 4846 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/93a6ace4-8d64-44d7-9637-457d3af6543c-run-openvswitch\") on node \"crc\" DevicePath \"\"" Oct 05 06:59:42 crc kubenswrapper[4846]: I1005 06:59:42.874227 4846 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/93a6ace4-8d64-44d7-9637-457d3af6543c-run-ovn\") on node \"crc\" DevicePath \"\"" Oct 05 06:59:42 crc kubenswrapper[4846]: I1005 06:59:42.874239 4846 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/93a6ace4-8d64-44d7-9637-457d3af6543c-host-cni-netd\") on node \"crc\" DevicePath \"\"" Oct 05 06:59:42 crc kubenswrapper[4846]: I1005 06:59:42.874248 4846 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/93a6ace4-8d64-44d7-9637-457d3af6543c-env-overrides\") on node \"crc\" DevicePath \"\"" Oct 05 06:59:42 crc kubenswrapper[4846]: I1005 06:59:42.874258 4846 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/93a6ace4-8d64-44d7-9637-457d3af6543c-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Oct 05 06:59:42 crc kubenswrapper[4846]: I1005 06:59:42.874268 4846 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/93a6ace4-8d64-44d7-9637-457d3af6543c-log-socket\") on node \"crc\" DevicePath \"\"" Oct 05 06:59:42 crc kubenswrapper[4846]: I1005 06:59:42.874279 4846 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/93a6ace4-8d64-44d7-9637-457d3af6543c-host-cni-bin\") on node \"crc\" DevicePath \"\"" Oct 05 06:59:42 crc kubenswrapper[4846]: I1005 06:59:42.874289 4846 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/93a6ace4-8d64-44d7-9637-457d3af6543c-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Oct 05 06:59:42 crc kubenswrapper[4846]: I1005 06:59:42.874298 4846 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/93a6ace4-8d64-44d7-9637-457d3af6543c-node-log\") on node \"crc\" DevicePath \"\"" Oct 05 06:59:42 crc kubenswrapper[4846]: I1005 06:59:42.874307 4846 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/93a6ace4-8d64-44d7-9637-457d3af6543c-host-run-netns\") on node \"crc\" DevicePath \"\"" Oct 05 06:59:42 crc kubenswrapper[4846]: I1005 06:59:42.874315 4846 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/93a6ace4-8d64-44d7-9637-457d3af6543c-host-slash\") on node \"crc\" DevicePath \"\"" Oct 05 06:59:42 crc kubenswrapper[4846]: I1005 06:59:42.874325 4846 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/93a6ace4-8d64-44d7-9637-457d3af6543c-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Oct 05 06:59:42 crc kubenswrapper[4846]: I1005 06:59:42.874334 4846 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/93a6ace4-8d64-44d7-9637-457d3af6543c-host-kubelet\") on node \"crc\" DevicePath \"\"" Oct 05 06:59:42 crc kubenswrapper[4846]: I1005 06:59:42.874343 4846 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/93a6ace4-8d64-44d7-9637-457d3af6543c-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Oct 05 06:59:42 crc kubenswrapper[4846]: I1005 06:59:42.879253 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/93a6ace4-8d64-44d7-9637-457d3af6543c-kube-api-access-96h7b" (OuterVolumeSpecName: "kube-api-access-96h7b") pod "93a6ace4-8d64-44d7-9637-457d3af6543c" (UID: "93a6ace4-8d64-44d7-9637-457d3af6543c"). InnerVolumeSpecName "kube-api-access-96h7b". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 06:59:42 crc kubenswrapper[4846]: I1005 06:59:42.879260 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/93a6ace4-8d64-44d7-9637-457d3af6543c-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "93a6ace4-8d64-44d7-9637-457d3af6543c" (UID: "93a6ace4-8d64-44d7-9637-457d3af6543c"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 06:59:42 crc kubenswrapper[4846]: I1005 06:59:42.886870 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/93a6ace4-8d64-44d7-9637-457d3af6543c-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "93a6ace4-8d64-44d7-9637-457d3af6543c" (UID: "93a6ace4-8d64-44d7-9637-457d3af6543c"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 05 06:59:42 crc kubenswrapper[4846]: I1005 06:59:42.975621 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/778242f5-89b8-4f86-9f58-0a81e2d70569-systemd-units\") pod \"ovnkube-node-q224c\" (UID: \"778242f5-89b8-4f86-9f58-0a81e2d70569\") " pod="openshift-ovn-kubernetes/ovnkube-node-q224c" Oct 05 06:59:42 crc kubenswrapper[4846]: I1005 06:59:42.975730 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/778242f5-89b8-4f86-9f58-0a81e2d70569-var-lib-openvswitch\") pod \"ovnkube-node-q224c\" (UID: \"778242f5-89b8-4f86-9f58-0a81e2d70569\") " pod="openshift-ovn-kubernetes/ovnkube-node-q224c" Oct 05 06:59:42 crc kubenswrapper[4846]: I1005 06:59:42.975782 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/778242f5-89b8-4f86-9f58-0a81e2d70569-ovnkube-config\") pod \"ovnkube-node-q224c\" (UID: \"778242f5-89b8-4f86-9f58-0a81e2d70569\") " pod="openshift-ovn-kubernetes/ovnkube-node-q224c" Oct 05 06:59:42 crc kubenswrapper[4846]: I1005 06:59:42.975863 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/778242f5-89b8-4f86-9f58-0a81e2d70569-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-q224c\" (UID: \"778242f5-89b8-4f86-9f58-0a81e2d70569\") " pod="openshift-ovn-kubernetes/ovnkube-node-q224c" Oct 05 06:59:42 crc kubenswrapper[4846]: I1005 06:59:42.975920 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/778242f5-89b8-4f86-9f58-0a81e2d70569-host-run-netns\") pod \"ovnkube-node-q224c\" (UID: \"778242f5-89b8-4f86-9f58-0a81e2d70569\") " pod="openshift-ovn-kubernetes/ovnkube-node-q224c" Oct 05 06:59:42 crc kubenswrapper[4846]: I1005 06:59:42.975971 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/778242f5-89b8-4f86-9f58-0a81e2d70569-host-cni-bin\") pod \"ovnkube-node-q224c\" (UID: \"778242f5-89b8-4f86-9f58-0a81e2d70569\") " pod="openshift-ovn-kubernetes/ovnkube-node-q224c" Oct 05 06:59:42 crc kubenswrapper[4846]: I1005 06:59:42.976038 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/778242f5-89b8-4f86-9f58-0a81e2d70569-ovnkube-script-lib\") pod \"ovnkube-node-q224c\" (UID: \"778242f5-89b8-4f86-9f58-0a81e2d70569\") " pod="openshift-ovn-kubernetes/ovnkube-node-q224c" Oct 05 06:59:42 crc kubenswrapper[4846]: I1005 06:59:42.976278 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/778242f5-89b8-4f86-9f58-0a81e2d70569-var-lib-openvswitch\") pod \"ovnkube-node-q224c\" (UID: \"778242f5-89b8-4f86-9f58-0a81e2d70569\") " pod="openshift-ovn-kubernetes/ovnkube-node-q224c" Oct 05 06:59:42 crc kubenswrapper[4846]: I1005 06:59:42.976452 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/778242f5-89b8-4f86-9f58-0a81e2d70569-host-run-netns\") pod \"ovnkube-node-q224c\" (UID: \"778242f5-89b8-4f86-9f58-0a81e2d70569\") " pod="openshift-ovn-kubernetes/ovnkube-node-q224c" Oct 05 06:59:42 crc kubenswrapper[4846]: I1005 06:59:42.976986 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/778242f5-89b8-4f86-9f58-0a81e2d70569-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-q224c\" (UID: \"778242f5-89b8-4f86-9f58-0a81e2d70569\") " pod="openshift-ovn-kubernetes/ovnkube-node-q224c" Oct 05 06:59:42 crc kubenswrapper[4846]: I1005 06:59:42.977008 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/778242f5-89b8-4f86-9f58-0a81e2d70569-host-cni-bin\") pod \"ovnkube-node-q224c\" (UID: \"778242f5-89b8-4f86-9f58-0a81e2d70569\") " pod="openshift-ovn-kubernetes/ovnkube-node-q224c" Oct 05 06:59:42 crc kubenswrapper[4846]: I1005 06:59:42.977498 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/778242f5-89b8-4f86-9f58-0a81e2d70569-systemd-units\") pod \"ovnkube-node-q224c\" (UID: \"778242f5-89b8-4f86-9f58-0a81e2d70569\") " pod="openshift-ovn-kubernetes/ovnkube-node-q224c" Oct 05 06:59:42 crc kubenswrapper[4846]: I1005 06:59:42.977611 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/778242f5-89b8-4f86-9f58-0a81e2d70569-etc-openvswitch\") pod \"ovnkube-node-q224c\" (UID: \"778242f5-89b8-4f86-9f58-0a81e2d70569\") " pod="openshift-ovn-kubernetes/ovnkube-node-q224c" Oct 05 06:59:42 crc kubenswrapper[4846]: I1005 06:59:42.976329 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/778242f5-89b8-4f86-9f58-0a81e2d70569-etc-openvswitch\") pod \"ovnkube-node-q224c\" (UID: \"778242f5-89b8-4f86-9f58-0a81e2d70569\") " pod="openshift-ovn-kubernetes/ovnkube-node-q224c" Oct 05 06:59:42 crc kubenswrapper[4846]: I1005 06:59:42.978457 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/778242f5-89b8-4f86-9f58-0a81e2d70569-host-run-ovn-kubernetes\") pod \"ovnkube-node-q224c\" (UID: \"778242f5-89b8-4f86-9f58-0a81e2d70569\") " pod="openshift-ovn-kubernetes/ovnkube-node-q224c" Oct 05 06:59:42 crc kubenswrapper[4846]: I1005 06:59:42.978509 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/778242f5-89b8-4f86-9f58-0a81e2d70569-run-systemd\") pod \"ovnkube-node-q224c\" (UID: \"778242f5-89b8-4f86-9f58-0a81e2d70569\") " pod="openshift-ovn-kubernetes/ovnkube-node-q224c" Oct 05 06:59:42 crc kubenswrapper[4846]: I1005 06:59:42.978558 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/778242f5-89b8-4f86-9f58-0a81e2d70569-host-kubelet\") pod \"ovnkube-node-q224c\" (UID: \"778242f5-89b8-4f86-9f58-0a81e2d70569\") " pod="openshift-ovn-kubernetes/ovnkube-node-q224c" Oct 05 06:59:42 crc kubenswrapper[4846]: I1005 06:59:42.978600 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/778242f5-89b8-4f86-9f58-0a81e2d70569-run-openvswitch\") pod \"ovnkube-node-q224c\" (UID: \"778242f5-89b8-4f86-9f58-0a81e2d70569\") " pod="openshift-ovn-kubernetes/ovnkube-node-q224c" Oct 05 06:59:42 crc kubenswrapper[4846]: I1005 06:59:42.978653 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tscwh\" (UniqueName: \"kubernetes.io/projected/778242f5-89b8-4f86-9f58-0a81e2d70569-kube-api-access-tscwh\") pod \"ovnkube-node-q224c\" (UID: \"778242f5-89b8-4f86-9f58-0a81e2d70569\") " pod="openshift-ovn-kubernetes/ovnkube-node-q224c" Oct 05 06:59:42 crc kubenswrapper[4846]: I1005 06:59:42.978708 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/778242f5-89b8-4f86-9f58-0a81e2d70569-log-socket\") pod \"ovnkube-node-q224c\" (UID: \"778242f5-89b8-4f86-9f58-0a81e2d70569\") " pod="openshift-ovn-kubernetes/ovnkube-node-q224c" Oct 05 06:59:42 crc kubenswrapper[4846]: I1005 06:59:42.978732 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/778242f5-89b8-4f86-9f58-0a81e2d70569-host-kubelet\") pod \"ovnkube-node-q224c\" (UID: \"778242f5-89b8-4f86-9f58-0a81e2d70569\") " pod="openshift-ovn-kubernetes/ovnkube-node-q224c" Oct 05 06:59:42 crc kubenswrapper[4846]: I1005 06:59:42.978759 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/778242f5-89b8-4f86-9f58-0a81e2d70569-node-log\") pod \"ovnkube-node-q224c\" (UID: \"778242f5-89b8-4f86-9f58-0a81e2d70569\") " pod="openshift-ovn-kubernetes/ovnkube-node-q224c" Oct 05 06:59:42 crc kubenswrapper[4846]: I1005 06:59:42.978781 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/778242f5-89b8-4f86-9f58-0a81e2d70569-host-run-ovn-kubernetes\") pod \"ovnkube-node-q224c\" (UID: \"778242f5-89b8-4f86-9f58-0a81e2d70569\") " pod="openshift-ovn-kubernetes/ovnkube-node-q224c" Oct 05 06:59:42 crc kubenswrapper[4846]: I1005 06:59:42.978804 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/778242f5-89b8-4f86-9f58-0a81e2d70569-ovn-node-metrics-cert\") pod \"ovnkube-node-q224c\" (UID: \"778242f5-89b8-4f86-9f58-0a81e2d70569\") " pod="openshift-ovn-kubernetes/ovnkube-node-q224c" Oct 05 06:59:42 crc kubenswrapper[4846]: I1005 06:59:42.978846 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/778242f5-89b8-4f86-9f58-0a81e2d70569-log-socket\") pod \"ovnkube-node-q224c\" (UID: \"778242f5-89b8-4f86-9f58-0a81e2d70569\") " pod="openshift-ovn-kubernetes/ovnkube-node-q224c" Oct 05 06:59:42 crc kubenswrapper[4846]: I1005 06:59:42.978878 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/778242f5-89b8-4f86-9f58-0a81e2d70569-run-openvswitch\") pod \"ovnkube-node-q224c\" (UID: \"778242f5-89b8-4f86-9f58-0a81e2d70569\") " pod="openshift-ovn-kubernetes/ovnkube-node-q224c" Oct 05 06:59:42 crc kubenswrapper[4846]: I1005 06:59:42.978875 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/778242f5-89b8-4f86-9f58-0a81e2d70569-env-overrides\") pod \"ovnkube-node-q224c\" (UID: \"778242f5-89b8-4f86-9f58-0a81e2d70569\") " pod="openshift-ovn-kubernetes/ovnkube-node-q224c" Oct 05 06:59:42 crc kubenswrapper[4846]: I1005 06:59:42.978934 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/778242f5-89b8-4f86-9f58-0a81e2d70569-host-slash\") pod \"ovnkube-node-q224c\" (UID: \"778242f5-89b8-4f86-9f58-0a81e2d70569\") " pod="openshift-ovn-kubernetes/ovnkube-node-q224c" Oct 05 06:59:42 crc kubenswrapper[4846]: I1005 06:59:42.978968 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/778242f5-89b8-4f86-9f58-0a81e2d70569-run-ovn\") pod \"ovnkube-node-q224c\" (UID: \"778242f5-89b8-4f86-9f58-0a81e2d70569\") " pod="openshift-ovn-kubernetes/ovnkube-node-q224c" Oct 05 06:59:42 crc kubenswrapper[4846]: I1005 06:59:42.978970 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/778242f5-89b8-4f86-9f58-0a81e2d70569-ovnkube-script-lib\") pod \"ovnkube-node-q224c\" (UID: \"778242f5-89b8-4f86-9f58-0a81e2d70569\") " pod="openshift-ovn-kubernetes/ovnkube-node-q224c" Oct 05 06:59:42 crc kubenswrapper[4846]: I1005 06:59:42.979008 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/778242f5-89b8-4f86-9f58-0a81e2d70569-host-cni-netd\") pod \"ovnkube-node-q224c\" (UID: \"778242f5-89b8-4f86-9f58-0a81e2d70569\") " pod="openshift-ovn-kubernetes/ovnkube-node-q224c" Oct 05 06:59:42 crc kubenswrapper[4846]: I1005 06:59:42.979054 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/778242f5-89b8-4f86-9f58-0a81e2d70569-host-cni-netd\") pod \"ovnkube-node-q224c\" (UID: \"778242f5-89b8-4f86-9f58-0a81e2d70569\") " pod="openshift-ovn-kubernetes/ovnkube-node-q224c" Oct 05 06:59:42 crc kubenswrapper[4846]: I1005 06:59:42.978820 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/778242f5-89b8-4f86-9f58-0a81e2d70569-run-systemd\") pod \"ovnkube-node-q224c\" (UID: \"778242f5-89b8-4f86-9f58-0a81e2d70569\") " pod="openshift-ovn-kubernetes/ovnkube-node-q224c" Oct 05 06:59:42 crc kubenswrapper[4846]: I1005 06:59:42.979966 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/778242f5-89b8-4f86-9f58-0a81e2d70569-node-log\") pod \"ovnkube-node-q224c\" (UID: \"778242f5-89b8-4f86-9f58-0a81e2d70569\") " pod="openshift-ovn-kubernetes/ovnkube-node-q224c" Oct 05 06:59:42 crc kubenswrapper[4846]: I1005 06:59:42.980049 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/778242f5-89b8-4f86-9f58-0a81e2d70569-host-slash\") pod \"ovnkube-node-q224c\" (UID: \"778242f5-89b8-4f86-9f58-0a81e2d70569\") " pod="openshift-ovn-kubernetes/ovnkube-node-q224c" Oct 05 06:59:42 crc kubenswrapper[4846]: I1005 06:59:42.980106 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/778242f5-89b8-4f86-9f58-0a81e2d70569-run-ovn\") pod \"ovnkube-node-q224c\" (UID: \"778242f5-89b8-4f86-9f58-0a81e2d70569\") " pod="openshift-ovn-kubernetes/ovnkube-node-q224c" Oct 05 06:59:42 crc kubenswrapper[4846]: I1005 06:59:42.980209 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-96h7b\" (UniqueName: \"kubernetes.io/projected/93a6ace4-8d64-44d7-9637-457d3af6543c-kube-api-access-96h7b\") on node \"crc\" DevicePath \"\"" Oct 05 06:59:42 crc kubenswrapper[4846]: I1005 06:59:42.980250 4846 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/93a6ace4-8d64-44d7-9637-457d3af6543c-run-systemd\") on node \"crc\" DevicePath \"\"" Oct 05 06:59:42 crc kubenswrapper[4846]: I1005 06:59:42.980276 4846 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/93a6ace4-8d64-44d7-9637-457d3af6543c-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Oct 05 06:59:42 crc kubenswrapper[4846]: I1005 06:59:42.981343 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/778242f5-89b8-4f86-9f58-0a81e2d70569-ovnkube-config\") pod \"ovnkube-node-q224c\" (UID: \"778242f5-89b8-4f86-9f58-0a81e2d70569\") " pod="openshift-ovn-kubernetes/ovnkube-node-q224c" Oct 05 06:59:42 crc kubenswrapper[4846]: I1005 06:59:42.983516 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/778242f5-89b8-4f86-9f58-0a81e2d70569-env-overrides\") pod \"ovnkube-node-q224c\" (UID: \"778242f5-89b8-4f86-9f58-0a81e2d70569\") " pod="openshift-ovn-kubernetes/ovnkube-node-q224c" Oct 05 06:59:42 crc kubenswrapper[4846]: I1005 06:59:42.989153 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/778242f5-89b8-4f86-9f58-0a81e2d70569-ovn-node-metrics-cert\") pod \"ovnkube-node-q224c\" (UID: \"778242f5-89b8-4f86-9f58-0a81e2d70569\") " pod="openshift-ovn-kubernetes/ovnkube-node-q224c" Oct 05 06:59:43 crc kubenswrapper[4846]: I1005 06:59:43.008343 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tscwh\" (UniqueName: \"kubernetes.io/projected/778242f5-89b8-4f86-9f58-0a81e2d70569-kube-api-access-tscwh\") pod \"ovnkube-node-q224c\" (UID: \"778242f5-89b8-4f86-9f58-0a81e2d70569\") " pod="openshift-ovn-kubernetes/ovnkube-node-q224c" Oct 05 06:59:43 crc kubenswrapper[4846]: I1005 06:59:43.023226 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-7wm5q_9bc7744f-d700-450f-8b8d-98140877fee9/kube-multus/2.log" Oct 05 06:59:43 crc kubenswrapper[4846]: I1005 06:59:43.023987 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-7wm5q_9bc7744f-d700-450f-8b8d-98140877fee9/kube-multus/1.log" Oct 05 06:59:43 crc kubenswrapper[4846]: I1005 06:59:43.024065 4846 generic.go:334] "Generic (PLEG): container finished" podID="9bc7744f-d700-450f-8b8d-98140877fee9" containerID="75ffecdf3834699d0ea47cca69028168c32263e5b5ddc5043cc0c1dc1d7bf74d" exitCode=2 Oct 05 06:59:43 crc kubenswrapper[4846]: I1005 06:59:43.024169 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-7wm5q" event={"ID":"9bc7744f-d700-450f-8b8d-98140877fee9","Type":"ContainerDied","Data":"75ffecdf3834699d0ea47cca69028168c32263e5b5ddc5043cc0c1dc1d7bf74d"} Oct 05 06:59:43 crc kubenswrapper[4846]: I1005 06:59:43.024275 4846 scope.go:117] "RemoveContainer" containerID="9d8631ba978f8fa7677feeeff42cabc454bfd51e5c5c21c78519fa9dfd85ec3c" Oct 05 06:59:43 crc kubenswrapper[4846]: I1005 06:59:43.025494 4846 scope.go:117] "RemoveContainer" containerID="75ffecdf3834699d0ea47cca69028168c32263e5b5ddc5043cc0c1dc1d7bf74d" Oct 05 06:59:43 crc kubenswrapper[4846]: E1005 06:59:43.026067 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-7wm5q_openshift-multus(9bc7744f-d700-450f-8b8d-98140877fee9)\"" pod="openshift-multus/multus-7wm5q" podUID="9bc7744f-d700-450f-8b8d-98140877fee9" Oct 05 06:59:43 crc kubenswrapper[4846]: I1005 06:59:43.029912 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-txzlk_93a6ace4-8d64-44d7-9637-457d3af6543c/ovnkube-controller/3.log" Oct 05 06:59:43 crc kubenswrapper[4846]: I1005 06:59:43.034623 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-txzlk_93a6ace4-8d64-44d7-9637-457d3af6543c/ovn-acl-logging/0.log" Oct 05 06:59:43 crc kubenswrapper[4846]: I1005 06:59:43.035655 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-txzlk_93a6ace4-8d64-44d7-9637-457d3af6543c/ovn-controller/0.log" Oct 05 06:59:43 crc kubenswrapper[4846]: I1005 06:59:43.036381 4846 generic.go:334] "Generic (PLEG): container finished" podID="93a6ace4-8d64-44d7-9637-457d3af6543c" containerID="32a57312adddc39f79d4dbc9dfb0197ec33241adfb654ed64db6bc10d1d479c8" exitCode=0 Oct 05 06:59:43 crc kubenswrapper[4846]: I1005 06:59:43.036417 4846 generic.go:334] "Generic (PLEG): container finished" podID="93a6ace4-8d64-44d7-9637-457d3af6543c" containerID="0df0ebcd87dc5af4b17736b43ec7b2395968b19ed16e775366af14e2b2d49c52" exitCode=0 Oct 05 06:59:43 crc kubenswrapper[4846]: I1005 06:59:43.036434 4846 generic.go:334] "Generic (PLEG): container finished" podID="93a6ace4-8d64-44d7-9637-457d3af6543c" containerID="9e9b0219ecbc7e60905d6de3e67a2c8dc9af1017b4aed9538ee0c58c861f2438" exitCode=0 Oct 05 06:59:43 crc kubenswrapper[4846]: I1005 06:59:43.036450 4846 generic.go:334] "Generic (PLEG): container finished" podID="93a6ace4-8d64-44d7-9637-457d3af6543c" containerID="af5e5a4da9ad7371d5a24633e8268f12e3a7d419a8c93b11ddbcc006d8339c6c" exitCode=0 Oct 05 06:59:43 crc kubenswrapper[4846]: I1005 06:59:43.036466 4846 generic.go:334] "Generic (PLEG): container finished" podID="93a6ace4-8d64-44d7-9637-457d3af6543c" containerID="0a6f6b2c32dd3da5f9fd5e6971c1773e81a9a24184b7aa5b543cec879001ae53" exitCode=0 Oct 05 06:59:43 crc kubenswrapper[4846]: I1005 06:59:43.036468 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-txzlk" event={"ID":"93a6ace4-8d64-44d7-9637-457d3af6543c","Type":"ContainerDied","Data":"32a57312adddc39f79d4dbc9dfb0197ec33241adfb654ed64db6bc10d1d479c8"} Oct 05 06:59:43 crc kubenswrapper[4846]: I1005 06:59:43.036549 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-txzlk" event={"ID":"93a6ace4-8d64-44d7-9637-457d3af6543c","Type":"ContainerDied","Data":"0df0ebcd87dc5af4b17736b43ec7b2395968b19ed16e775366af14e2b2d49c52"} Oct 05 06:59:43 crc kubenswrapper[4846]: I1005 06:59:43.036575 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-txzlk" event={"ID":"93a6ace4-8d64-44d7-9637-457d3af6543c","Type":"ContainerDied","Data":"9e9b0219ecbc7e60905d6de3e67a2c8dc9af1017b4aed9538ee0c58c861f2438"} Oct 05 06:59:43 crc kubenswrapper[4846]: I1005 06:59:43.036624 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-txzlk" event={"ID":"93a6ace4-8d64-44d7-9637-457d3af6543c","Type":"ContainerDied","Data":"af5e5a4da9ad7371d5a24633e8268f12e3a7d419a8c93b11ddbcc006d8339c6c"} Oct 05 06:59:43 crc kubenswrapper[4846]: I1005 06:59:43.036621 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-txzlk" Oct 05 06:59:43 crc kubenswrapper[4846]: I1005 06:59:43.036644 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-txzlk" event={"ID":"93a6ace4-8d64-44d7-9637-457d3af6543c","Type":"ContainerDied","Data":"0a6f6b2c32dd3da5f9fd5e6971c1773e81a9a24184b7aa5b543cec879001ae53"} Oct 05 06:59:43 crc kubenswrapper[4846]: I1005 06:59:43.036766 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-txzlk" event={"ID":"93a6ace4-8d64-44d7-9637-457d3af6543c","Type":"ContainerDied","Data":"1ad1940d07e37217089642478ece9ccdaca74784738a400929a9f5359d158c01"} Oct 05 06:59:43 crc kubenswrapper[4846]: I1005 06:59:43.036797 4846 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"32a57312adddc39f79d4dbc9dfb0197ec33241adfb654ed64db6bc10d1d479c8"} Oct 05 06:59:43 crc kubenswrapper[4846]: I1005 06:59:43.036819 4846 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"59fa020f7bfea4547e132058cf3804a406ee0c06cdc4e78a08ae43a51f46633f"} Oct 05 06:59:43 crc kubenswrapper[4846]: I1005 06:59:43.036832 4846 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"0df0ebcd87dc5af4b17736b43ec7b2395968b19ed16e775366af14e2b2d49c52"} Oct 05 06:59:43 crc kubenswrapper[4846]: I1005 06:59:43.036843 4846 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"9e9b0219ecbc7e60905d6de3e67a2c8dc9af1017b4aed9538ee0c58c861f2438"} Oct 05 06:59:43 crc kubenswrapper[4846]: I1005 06:59:43.036855 4846 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"af5e5a4da9ad7371d5a24633e8268f12e3a7d419a8c93b11ddbcc006d8339c6c"} Oct 05 06:59:43 crc kubenswrapper[4846]: I1005 06:59:43.036867 4846 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"0a6f6b2c32dd3da5f9fd5e6971c1773e81a9a24184b7aa5b543cec879001ae53"} Oct 05 06:59:43 crc kubenswrapper[4846]: I1005 06:59:43.036878 4846 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"1ad1940d07e37217089642478ece9ccdaca74784738a400929a9f5359d158c01"} Oct 05 06:59:43 crc kubenswrapper[4846]: I1005 06:59:43.036890 4846 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"8da85222775abc7a772f6a8ff8255f1a0d5b15807ba177a7e3e2b2d01a4f24b9"} Oct 05 06:59:43 crc kubenswrapper[4846]: I1005 06:59:43.036902 4846 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"1409651de8c897804f87be03d016c138f8acf27e662dce265a93e15c1848adcf"} Oct 05 06:59:43 crc kubenswrapper[4846]: I1005 06:59:43.036914 4846 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ff3c6545ff3d2436da27bbd223b950b30b0c69a3e7f3034cb0e61380cb38d338"} Oct 05 06:59:43 crc kubenswrapper[4846]: I1005 06:59:43.036481 4846 generic.go:334] "Generic (PLEG): container finished" podID="93a6ace4-8d64-44d7-9637-457d3af6543c" containerID="1ad1940d07e37217089642478ece9ccdaca74784738a400929a9f5359d158c01" exitCode=0 Oct 05 06:59:43 crc kubenswrapper[4846]: I1005 06:59:43.036949 4846 generic.go:334] "Generic (PLEG): container finished" podID="93a6ace4-8d64-44d7-9637-457d3af6543c" containerID="8da85222775abc7a772f6a8ff8255f1a0d5b15807ba177a7e3e2b2d01a4f24b9" exitCode=143 Oct 05 06:59:43 crc kubenswrapper[4846]: I1005 06:59:43.036968 4846 generic.go:334] "Generic (PLEG): container finished" podID="93a6ace4-8d64-44d7-9637-457d3af6543c" containerID="1409651de8c897804f87be03d016c138f8acf27e662dce265a93e15c1848adcf" exitCode=143 Oct 05 06:59:43 crc kubenswrapper[4846]: I1005 06:59:43.036994 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-txzlk" event={"ID":"93a6ace4-8d64-44d7-9637-457d3af6543c","Type":"ContainerDied","Data":"8da85222775abc7a772f6a8ff8255f1a0d5b15807ba177a7e3e2b2d01a4f24b9"} Oct 05 06:59:43 crc kubenswrapper[4846]: I1005 06:59:43.037013 4846 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"32a57312adddc39f79d4dbc9dfb0197ec33241adfb654ed64db6bc10d1d479c8"} Oct 05 06:59:43 crc kubenswrapper[4846]: I1005 06:59:43.037026 4846 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"59fa020f7bfea4547e132058cf3804a406ee0c06cdc4e78a08ae43a51f46633f"} Oct 05 06:59:43 crc kubenswrapper[4846]: I1005 06:59:43.037039 4846 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"0df0ebcd87dc5af4b17736b43ec7b2395968b19ed16e775366af14e2b2d49c52"} Oct 05 06:59:43 crc kubenswrapper[4846]: I1005 06:59:43.037052 4846 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"9e9b0219ecbc7e60905d6de3e67a2c8dc9af1017b4aed9538ee0c58c861f2438"} Oct 05 06:59:43 crc kubenswrapper[4846]: I1005 06:59:43.037064 4846 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"af5e5a4da9ad7371d5a24633e8268f12e3a7d419a8c93b11ddbcc006d8339c6c"} Oct 05 06:59:43 crc kubenswrapper[4846]: I1005 06:59:43.037075 4846 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"0a6f6b2c32dd3da5f9fd5e6971c1773e81a9a24184b7aa5b543cec879001ae53"} Oct 05 06:59:43 crc kubenswrapper[4846]: I1005 06:59:43.037087 4846 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"1ad1940d07e37217089642478ece9ccdaca74784738a400929a9f5359d158c01"} Oct 05 06:59:43 crc kubenswrapper[4846]: I1005 06:59:43.037099 4846 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"8da85222775abc7a772f6a8ff8255f1a0d5b15807ba177a7e3e2b2d01a4f24b9"} Oct 05 06:59:43 crc kubenswrapper[4846]: I1005 06:59:43.038246 4846 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"1409651de8c897804f87be03d016c138f8acf27e662dce265a93e15c1848adcf"} Oct 05 06:59:43 crc kubenswrapper[4846]: I1005 06:59:43.038298 4846 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ff3c6545ff3d2436da27bbd223b950b30b0c69a3e7f3034cb0e61380cb38d338"} Oct 05 06:59:43 crc kubenswrapper[4846]: I1005 06:59:43.038325 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-txzlk" event={"ID":"93a6ace4-8d64-44d7-9637-457d3af6543c","Type":"ContainerDied","Data":"1409651de8c897804f87be03d016c138f8acf27e662dce265a93e15c1848adcf"} Oct 05 06:59:43 crc kubenswrapper[4846]: I1005 06:59:43.038359 4846 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"32a57312adddc39f79d4dbc9dfb0197ec33241adfb654ed64db6bc10d1d479c8"} Oct 05 06:59:43 crc kubenswrapper[4846]: I1005 06:59:43.038374 4846 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"59fa020f7bfea4547e132058cf3804a406ee0c06cdc4e78a08ae43a51f46633f"} Oct 05 06:59:43 crc kubenswrapper[4846]: I1005 06:59:43.038383 4846 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"0df0ebcd87dc5af4b17736b43ec7b2395968b19ed16e775366af14e2b2d49c52"} Oct 05 06:59:43 crc kubenswrapper[4846]: I1005 06:59:43.038390 4846 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"9e9b0219ecbc7e60905d6de3e67a2c8dc9af1017b4aed9538ee0c58c861f2438"} Oct 05 06:59:43 crc kubenswrapper[4846]: I1005 06:59:43.038397 4846 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"af5e5a4da9ad7371d5a24633e8268f12e3a7d419a8c93b11ddbcc006d8339c6c"} Oct 05 06:59:43 crc kubenswrapper[4846]: I1005 06:59:43.038403 4846 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"0a6f6b2c32dd3da5f9fd5e6971c1773e81a9a24184b7aa5b543cec879001ae53"} Oct 05 06:59:43 crc kubenswrapper[4846]: I1005 06:59:43.038408 4846 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"1ad1940d07e37217089642478ece9ccdaca74784738a400929a9f5359d158c01"} Oct 05 06:59:43 crc kubenswrapper[4846]: I1005 06:59:43.038414 4846 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"8da85222775abc7a772f6a8ff8255f1a0d5b15807ba177a7e3e2b2d01a4f24b9"} Oct 05 06:59:43 crc kubenswrapper[4846]: I1005 06:59:43.038420 4846 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"1409651de8c897804f87be03d016c138f8acf27e662dce265a93e15c1848adcf"} Oct 05 06:59:43 crc kubenswrapper[4846]: I1005 06:59:43.038425 4846 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ff3c6545ff3d2436da27bbd223b950b30b0c69a3e7f3034cb0e61380cb38d338"} Oct 05 06:59:43 crc kubenswrapper[4846]: I1005 06:59:43.038433 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-txzlk" event={"ID":"93a6ace4-8d64-44d7-9637-457d3af6543c","Type":"ContainerDied","Data":"96612529b2d28247403966c25096e92a88b39c2bef12601d90018451dbf8a6e7"} Oct 05 06:59:43 crc kubenswrapper[4846]: I1005 06:59:43.038444 4846 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"32a57312adddc39f79d4dbc9dfb0197ec33241adfb654ed64db6bc10d1d479c8"} Oct 05 06:59:43 crc kubenswrapper[4846]: I1005 06:59:43.038451 4846 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"59fa020f7bfea4547e132058cf3804a406ee0c06cdc4e78a08ae43a51f46633f"} Oct 05 06:59:43 crc kubenswrapper[4846]: I1005 06:59:43.038457 4846 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"0df0ebcd87dc5af4b17736b43ec7b2395968b19ed16e775366af14e2b2d49c52"} Oct 05 06:59:43 crc kubenswrapper[4846]: I1005 06:59:43.038462 4846 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"9e9b0219ecbc7e60905d6de3e67a2c8dc9af1017b4aed9538ee0c58c861f2438"} Oct 05 06:59:43 crc kubenswrapper[4846]: I1005 06:59:43.038467 4846 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"af5e5a4da9ad7371d5a24633e8268f12e3a7d419a8c93b11ddbcc006d8339c6c"} Oct 05 06:59:43 crc kubenswrapper[4846]: I1005 06:59:43.038472 4846 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"0a6f6b2c32dd3da5f9fd5e6971c1773e81a9a24184b7aa5b543cec879001ae53"} Oct 05 06:59:43 crc kubenswrapper[4846]: I1005 06:59:43.038477 4846 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"1ad1940d07e37217089642478ece9ccdaca74784738a400929a9f5359d158c01"} Oct 05 06:59:43 crc kubenswrapper[4846]: I1005 06:59:43.038482 4846 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"8da85222775abc7a772f6a8ff8255f1a0d5b15807ba177a7e3e2b2d01a4f24b9"} Oct 05 06:59:43 crc kubenswrapper[4846]: I1005 06:59:43.038487 4846 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"1409651de8c897804f87be03d016c138f8acf27e662dce265a93e15c1848adcf"} Oct 05 06:59:43 crc kubenswrapper[4846]: I1005 06:59:43.038493 4846 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ff3c6545ff3d2436da27bbd223b950b30b0c69a3e7f3034cb0e61380cb38d338"} Oct 05 06:59:43 crc kubenswrapper[4846]: I1005 06:59:43.076312 4846 scope.go:117] "RemoveContainer" containerID="32a57312adddc39f79d4dbc9dfb0197ec33241adfb654ed64db6bc10d1d479c8" Oct 05 06:59:43 crc kubenswrapper[4846]: I1005 06:59:43.089507 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-txzlk"] Oct 05 06:59:43 crc kubenswrapper[4846]: I1005 06:59:43.095742 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-txzlk"] Oct 05 06:59:43 crc kubenswrapper[4846]: I1005 06:59:43.115488 4846 scope.go:117] "RemoveContainer" containerID="59fa020f7bfea4547e132058cf3804a406ee0c06cdc4e78a08ae43a51f46633f" Oct 05 06:59:43 crc kubenswrapper[4846]: I1005 06:59:43.143038 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-q224c" Oct 05 06:59:43 crc kubenswrapper[4846]: I1005 06:59:43.148315 4846 scope.go:117] "RemoveContainer" containerID="0df0ebcd87dc5af4b17736b43ec7b2395968b19ed16e775366af14e2b2d49c52" Oct 05 06:59:43 crc kubenswrapper[4846]: I1005 06:59:43.180863 4846 scope.go:117] "RemoveContainer" containerID="9e9b0219ecbc7e60905d6de3e67a2c8dc9af1017b4aed9538ee0c58c861f2438" Oct 05 06:59:43 crc kubenswrapper[4846]: I1005 06:59:43.202959 4846 scope.go:117] "RemoveContainer" containerID="af5e5a4da9ad7371d5a24633e8268f12e3a7d419a8c93b11ddbcc006d8339c6c" Oct 05 06:59:43 crc kubenswrapper[4846]: I1005 06:59:43.230927 4846 scope.go:117] "RemoveContainer" containerID="0a6f6b2c32dd3da5f9fd5e6971c1773e81a9a24184b7aa5b543cec879001ae53" Oct 05 06:59:43 crc kubenswrapper[4846]: I1005 06:59:43.255555 4846 scope.go:117] "RemoveContainer" containerID="1ad1940d07e37217089642478ece9ccdaca74784738a400929a9f5359d158c01" Oct 05 06:59:43 crc kubenswrapper[4846]: I1005 06:59:43.272259 4846 scope.go:117] "RemoveContainer" containerID="8da85222775abc7a772f6a8ff8255f1a0d5b15807ba177a7e3e2b2d01a4f24b9" Oct 05 06:59:43 crc kubenswrapper[4846]: I1005 06:59:43.290361 4846 scope.go:117] "RemoveContainer" containerID="1409651de8c897804f87be03d016c138f8acf27e662dce265a93e15c1848adcf" Oct 05 06:59:43 crc kubenswrapper[4846]: I1005 06:59:43.311136 4846 scope.go:117] "RemoveContainer" containerID="ff3c6545ff3d2436da27bbd223b950b30b0c69a3e7f3034cb0e61380cb38d338" Oct 05 06:59:43 crc kubenswrapper[4846]: I1005 06:59:43.349948 4846 scope.go:117] "RemoveContainer" containerID="32a57312adddc39f79d4dbc9dfb0197ec33241adfb654ed64db6bc10d1d479c8" Oct 05 06:59:43 crc kubenswrapper[4846]: E1005 06:59:43.350504 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"32a57312adddc39f79d4dbc9dfb0197ec33241adfb654ed64db6bc10d1d479c8\": container with ID starting with 32a57312adddc39f79d4dbc9dfb0197ec33241adfb654ed64db6bc10d1d479c8 not found: ID does not exist" containerID="32a57312adddc39f79d4dbc9dfb0197ec33241adfb654ed64db6bc10d1d479c8" Oct 05 06:59:43 crc kubenswrapper[4846]: I1005 06:59:43.350549 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"32a57312adddc39f79d4dbc9dfb0197ec33241adfb654ed64db6bc10d1d479c8"} err="failed to get container status \"32a57312adddc39f79d4dbc9dfb0197ec33241adfb654ed64db6bc10d1d479c8\": rpc error: code = NotFound desc = could not find container \"32a57312adddc39f79d4dbc9dfb0197ec33241adfb654ed64db6bc10d1d479c8\": container with ID starting with 32a57312adddc39f79d4dbc9dfb0197ec33241adfb654ed64db6bc10d1d479c8 not found: ID does not exist" Oct 05 06:59:43 crc kubenswrapper[4846]: I1005 06:59:43.350586 4846 scope.go:117] "RemoveContainer" containerID="59fa020f7bfea4547e132058cf3804a406ee0c06cdc4e78a08ae43a51f46633f" Oct 05 06:59:43 crc kubenswrapper[4846]: E1005 06:59:43.351505 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"59fa020f7bfea4547e132058cf3804a406ee0c06cdc4e78a08ae43a51f46633f\": container with ID starting with 59fa020f7bfea4547e132058cf3804a406ee0c06cdc4e78a08ae43a51f46633f not found: ID does not exist" containerID="59fa020f7bfea4547e132058cf3804a406ee0c06cdc4e78a08ae43a51f46633f" Oct 05 06:59:43 crc kubenswrapper[4846]: I1005 06:59:43.351663 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"59fa020f7bfea4547e132058cf3804a406ee0c06cdc4e78a08ae43a51f46633f"} err="failed to get container status \"59fa020f7bfea4547e132058cf3804a406ee0c06cdc4e78a08ae43a51f46633f\": rpc error: code = NotFound desc = could not find container \"59fa020f7bfea4547e132058cf3804a406ee0c06cdc4e78a08ae43a51f46633f\": container with ID starting with 59fa020f7bfea4547e132058cf3804a406ee0c06cdc4e78a08ae43a51f46633f not found: ID does not exist" Oct 05 06:59:43 crc kubenswrapper[4846]: I1005 06:59:43.351730 4846 scope.go:117] "RemoveContainer" containerID="0df0ebcd87dc5af4b17736b43ec7b2395968b19ed16e775366af14e2b2d49c52" Oct 05 06:59:43 crc kubenswrapper[4846]: E1005 06:59:43.353226 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0df0ebcd87dc5af4b17736b43ec7b2395968b19ed16e775366af14e2b2d49c52\": container with ID starting with 0df0ebcd87dc5af4b17736b43ec7b2395968b19ed16e775366af14e2b2d49c52 not found: ID does not exist" containerID="0df0ebcd87dc5af4b17736b43ec7b2395968b19ed16e775366af14e2b2d49c52" Oct 05 06:59:43 crc kubenswrapper[4846]: I1005 06:59:43.353301 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0df0ebcd87dc5af4b17736b43ec7b2395968b19ed16e775366af14e2b2d49c52"} err="failed to get container status \"0df0ebcd87dc5af4b17736b43ec7b2395968b19ed16e775366af14e2b2d49c52\": rpc error: code = NotFound desc = could not find container \"0df0ebcd87dc5af4b17736b43ec7b2395968b19ed16e775366af14e2b2d49c52\": container with ID starting with 0df0ebcd87dc5af4b17736b43ec7b2395968b19ed16e775366af14e2b2d49c52 not found: ID does not exist" Oct 05 06:59:43 crc kubenswrapper[4846]: I1005 06:59:43.353347 4846 scope.go:117] "RemoveContainer" containerID="9e9b0219ecbc7e60905d6de3e67a2c8dc9af1017b4aed9538ee0c58c861f2438" Oct 05 06:59:43 crc kubenswrapper[4846]: E1005 06:59:43.353958 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9e9b0219ecbc7e60905d6de3e67a2c8dc9af1017b4aed9538ee0c58c861f2438\": container with ID starting with 9e9b0219ecbc7e60905d6de3e67a2c8dc9af1017b4aed9538ee0c58c861f2438 not found: ID does not exist" containerID="9e9b0219ecbc7e60905d6de3e67a2c8dc9af1017b4aed9538ee0c58c861f2438" Oct 05 06:59:43 crc kubenswrapper[4846]: I1005 06:59:43.354029 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9e9b0219ecbc7e60905d6de3e67a2c8dc9af1017b4aed9538ee0c58c861f2438"} err="failed to get container status \"9e9b0219ecbc7e60905d6de3e67a2c8dc9af1017b4aed9538ee0c58c861f2438\": rpc error: code = NotFound desc = could not find container \"9e9b0219ecbc7e60905d6de3e67a2c8dc9af1017b4aed9538ee0c58c861f2438\": container with ID starting with 9e9b0219ecbc7e60905d6de3e67a2c8dc9af1017b4aed9538ee0c58c861f2438 not found: ID does not exist" Oct 05 06:59:43 crc kubenswrapper[4846]: I1005 06:59:43.354065 4846 scope.go:117] "RemoveContainer" containerID="af5e5a4da9ad7371d5a24633e8268f12e3a7d419a8c93b11ddbcc006d8339c6c" Oct 05 06:59:43 crc kubenswrapper[4846]: E1005 06:59:43.354694 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"af5e5a4da9ad7371d5a24633e8268f12e3a7d419a8c93b11ddbcc006d8339c6c\": container with ID starting with af5e5a4da9ad7371d5a24633e8268f12e3a7d419a8c93b11ddbcc006d8339c6c not found: ID does not exist" containerID="af5e5a4da9ad7371d5a24633e8268f12e3a7d419a8c93b11ddbcc006d8339c6c" Oct 05 06:59:43 crc kubenswrapper[4846]: I1005 06:59:43.354767 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"af5e5a4da9ad7371d5a24633e8268f12e3a7d419a8c93b11ddbcc006d8339c6c"} err="failed to get container status \"af5e5a4da9ad7371d5a24633e8268f12e3a7d419a8c93b11ddbcc006d8339c6c\": rpc error: code = NotFound desc = could not find container \"af5e5a4da9ad7371d5a24633e8268f12e3a7d419a8c93b11ddbcc006d8339c6c\": container with ID starting with af5e5a4da9ad7371d5a24633e8268f12e3a7d419a8c93b11ddbcc006d8339c6c not found: ID does not exist" Oct 05 06:59:43 crc kubenswrapper[4846]: I1005 06:59:43.354810 4846 scope.go:117] "RemoveContainer" containerID="0a6f6b2c32dd3da5f9fd5e6971c1773e81a9a24184b7aa5b543cec879001ae53" Oct 05 06:59:43 crc kubenswrapper[4846]: E1005 06:59:43.355467 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0a6f6b2c32dd3da5f9fd5e6971c1773e81a9a24184b7aa5b543cec879001ae53\": container with ID starting with 0a6f6b2c32dd3da5f9fd5e6971c1773e81a9a24184b7aa5b543cec879001ae53 not found: ID does not exist" containerID="0a6f6b2c32dd3da5f9fd5e6971c1773e81a9a24184b7aa5b543cec879001ae53" Oct 05 06:59:43 crc kubenswrapper[4846]: I1005 06:59:43.355548 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0a6f6b2c32dd3da5f9fd5e6971c1773e81a9a24184b7aa5b543cec879001ae53"} err="failed to get container status \"0a6f6b2c32dd3da5f9fd5e6971c1773e81a9a24184b7aa5b543cec879001ae53\": rpc error: code = NotFound desc = could not find container \"0a6f6b2c32dd3da5f9fd5e6971c1773e81a9a24184b7aa5b543cec879001ae53\": container with ID starting with 0a6f6b2c32dd3da5f9fd5e6971c1773e81a9a24184b7aa5b543cec879001ae53 not found: ID does not exist" Oct 05 06:59:43 crc kubenswrapper[4846]: I1005 06:59:43.355624 4846 scope.go:117] "RemoveContainer" containerID="1ad1940d07e37217089642478ece9ccdaca74784738a400929a9f5359d158c01" Oct 05 06:59:43 crc kubenswrapper[4846]: E1005 06:59:43.356691 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1ad1940d07e37217089642478ece9ccdaca74784738a400929a9f5359d158c01\": container with ID starting with 1ad1940d07e37217089642478ece9ccdaca74784738a400929a9f5359d158c01 not found: ID does not exist" containerID="1ad1940d07e37217089642478ece9ccdaca74784738a400929a9f5359d158c01" Oct 05 06:59:43 crc kubenswrapper[4846]: I1005 06:59:43.356758 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1ad1940d07e37217089642478ece9ccdaca74784738a400929a9f5359d158c01"} err="failed to get container status \"1ad1940d07e37217089642478ece9ccdaca74784738a400929a9f5359d158c01\": rpc error: code = NotFound desc = could not find container \"1ad1940d07e37217089642478ece9ccdaca74784738a400929a9f5359d158c01\": container with ID starting with 1ad1940d07e37217089642478ece9ccdaca74784738a400929a9f5359d158c01 not found: ID does not exist" Oct 05 06:59:43 crc kubenswrapper[4846]: I1005 06:59:43.356802 4846 scope.go:117] "RemoveContainer" containerID="8da85222775abc7a772f6a8ff8255f1a0d5b15807ba177a7e3e2b2d01a4f24b9" Oct 05 06:59:43 crc kubenswrapper[4846]: E1005 06:59:43.357662 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8da85222775abc7a772f6a8ff8255f1a0d5b15807ba177a7e3e2b2d01a4f24b9\": container with ID starting with 8da85222775abc7a772f6a8ff8255f1a0d5b15807ba177a7e3e2b2d01a4f24b9 not found: ID does not exist" containerID="8da85222775abc7a772f6a8ff8255f1a0d5b15807ba177a7e3e2b2d01a4f24b9" Oct 05 06:59:43 crc kubenswrapper[4846]: I1005 06:59:43.357724 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8da85222775abc7a772f6a8ff8255f1a0d5b15807ba177a7e3e2b2d01a4f24b9"} err="failed to get container status \"8da85222775abc7a772f6a8ff8255f1a0d5b15807ba177a7e3e2b2d01a4f24b9\": rpc error: code = NotFound desc = could not find container \"8da85222775abc7a772f6a8ff8255f1a0d5b15807ba177a7e3e2b2d01a4f24b9\": container with ID starting with 8da85222775abc7a772f6a8ff8255f1a0d5b15807ba177a7e3e2b2d01a4f24b9 not found: ID does not exist" Oct 05 06:59:43 crc kubenswrapper[4846]: I1005 06:59:43.357767 4846 scope.go:117] "RemoveContainer" containerID="1409651de8c897804f87be03d016c138f8acf27e662dce265a93e15c1848adcf" Oct 05 06:59:43 crc kubenswrapper[4846]: E1005 06:59:43.358569 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1409651de8c897804f87be03d016c138f8acf27e662dce265a93e15c1848adcf\": container with ID starting with 1409651de8c897804f87be03d016c138f8acf27e662dce265a93e15c1848adcf not found: ID does not exist" containerID="1409651de8c897804f87be03d016c138f8acf27e662dce265a93e15c1848adcf" Oct 05 06:59:43 crc kubenswrapper[4846]: I1005 06:59:43.358680 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1409651de8c897804f87be03d016c138f8acf27e662dce265a93e15c1848adcf"} err="failed to get container status \"1409651de8c897804f87be03d016c138f8acf27e662dce265a93e15c1848adcf\": rpc error: code = NotFound desc = could not find container \"1409651de8c897804f87be03d016c138f8acf27e662dce265a93e15c1848adcf\": container with ID starting with 1409651de8c897804f87be03d016c138f8acf27e662dce265a93e15c1848adcf not found: ID does not exist" Oct 05 06:59:43 crc kubenswrapper[4846]: I1005 06:59:43.358754 4846 scope.go:117] "RemoveContainer" containerID="ff3c6545ff3d2436da27bbd223b950b30b0c69a3e7f3034cb0e61380cb38d338" Oct 05 06:59:43 crc kubenswrapper[4846]: E1005 06:59:43.359348 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ff3c6545ff3d2436da27bbd223b950b30b0c69a3e7f3034cb0e61380cb38d338\": container with ID starting with ff3c6545ff3d2436da27bbd223b950b30b0c69a3e7f3034cb0e61380cb38d338 not found: ID does not exist" containerID="ff3c6545ff3d2436da27bbd223b950b30b0c69a3e7f3034cb0e61380cb38d338" Oct 05 06:59:43 crc kubenswrapper[4846]: I1005 06:59:43.359393 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ff3c6545ff3d2436da27bbd223b950b30b0c69a3e7f3034cb0e61380cb38d338"} err="failed to get container status \"ff3c6545ff3d2436da27bbd223b950b30b0c69a3e7f3034cb0e61380cb38d338\": rpc error: code = NotFound desc = could not find container \"ff3c6545ff3d2436da27bbd223b950b30b0c69a3e7f3034cb0e61380cb38d338\": container with ID starting with ff3c6545ff3d2436da27bbd223b950b30b0c69a3e7f3034cb0e61380cb38d338 not found: ID does not exist" Oct 05 06:59:43 crc kubenswrapper[4846]: I1005 06:59:43.359442 4846 scope.go:117] "RemoveContainer" containerID="32a57312adddc39f79d4dbc9dfb0197ec33241adfb654ed64db6bc10d1d479c8" Oct 05 06:59:43 crc kubenswrapper[4846]: I1005 06:59:43.360690 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"32a57312adddc39f79d4dbc9dfb0197ec33241adfb654ed64db6bc10d1d479c8"} err="failed to get container status \"32a57312adddc39f79d4dbc9dfb0197ec33241adfb654ed64db6bc10d1d479c8\": rpc error: code = NotFound desc = could not find container \"32a57312adddc39f79d4dbc9dfb0197ec33241adfb654ed64db6bc10d1d479c8\": container with ID starting with 32a57312adddc39f79d4dbc9dfb0197ec33241adfb654ed64db6bc10d1d479c8 not found: ID does not exist" Oct 05 06:59:43 crc kubenswrapper[4846]: I1005 06:59:43.360740 4846 scope.go:117] "RemoveContainer" containerID="59fa020f7bfea4547e132058cf3804a406ee0c06cdc4e78a08ae43a51f46633f" Oct 05 06:59:43 crc kubenswrapper[4846]: I1005 06:59:43.361232 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"59fa020f7bfea4547e132058cf3804a406ee0c06cdc4e78a08ae43a51f46633f"} err="failed to get container status \"59fa020f7bfea4547e132058cf3804a406ee0c06cdc4e78a08ae43a51f46633f\": rpc error: code = NotFound desc = could not find container \"59fa020f7bfea4547e132058cf3804a406ee0c06cdc4e78a08ae43a51f46633f\": container with ID starting with 59fa020f7bfea4547e132058cf3804a406ee0c06cdc4e78a08ae43a51f46633f not found: ID does not exist" Oct 05 06:59:43 crc kubenswrapper[4846]: I1005 06:59:43.361288 4846 scope.go:117] "RemoveContainer" containerID="0df0ebcd87dc5af4b17736b43ec7b2395968b19ed16e775366af14e2b2d49c52" Oct 05 06:59:43 crc kubenswrapper[4846]: I1005 06:59:43.361738 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0df0ebcd87dc5af4b17736b43ec7b2395968b19ed16e775366af14e2b2d49c52"} err="failed to get container status \"0df0ebcd87dc5af4b17736b43ec7b2395968b19ed16e775366af14e2b2d49c52\": rpc error: code = NotFound desc = could not find container \"0df0ebcd87dc5af4b17736b43ec7b2395968b19ed16e775366af14e2b2d49c52\": container with ID starting with 0df0ebcd87dc5af4b17736b43ec7b2395968b19ed16e775366af14e2b2d49c52 not found: ID does not exist" Oct 05 06:59:43 crc kubenswrapper[4846]: I1005 06:59:43.361789 4846 scope.go:117] "RemoveContainer" containerID="9e9b0219ecbc7e60905d6de3e67a2c8dc9af1017b4aed9538ee0c58c861f2438" Oct 05 06:59:43 crc kubenswrapper[4846]: I1005 06:59:43.362255 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9e9b0219ecbc7e60905d6de3e67a2c8dc9af1017b4aed9538ee0c58c861f2438"} err="failed to get container status \"9e9b0219ecbc7e60905d6de3e67a2c8dc9af1017b4aed9538ee0c58c861f2438\": rpc error: code = NotFound desc = could not find container \"9e9b0219ecbc7e60905d6de3e67a2c8dc9af1017b4aed9538ee0c58c861f2438\": container with ID starting with 9e9b0219ecbc7e60905d6de3e67a2c8dc9af1017b4aed9538ee0c58c861f2438 not found: ID does not exist" Oct 05 06:59:43 crc kubenswrapper[4846]: I1005 06:59:43.362312 4846 scope.go:117] "RemoveContainer" containerID="af5e5a4da9ad7371d5a24633e8268f12e3a7d419a8c93b11ddbcc006d8339c6c" Oct 05 06:59:43 crc kubenswrapper[4846]: I1005 06:59:43.362905 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"af5e5a4da9ad7371d5a24633e8268f12e3a7d419a8c93b11ddbcc006d8339c6c"} err="failed to get container status \"af5e5a4da9ad7371d5a24633e8268f12e3a7d419a8c93b11ddbcc006d8339c6c\": rpc error: code = NotFound desc = could not find container \"af5e5a4da9ad7371d5a24633e8268f12e3a7d419a8c93b11ddbcc006d8339c6c\": container with ID starting with af5e5a4da9ad7371d5a24633e8268f12e3a7d419a8c93b11ddbcc006d8339c6c not found: ID does not exist" Oct 05 06:59:43 crc kubenswrapper[4846]: I1005 06:59:43.363504 4846 scope.go:117] "RemoveContainer" containerID="0a6f6b2c32dd3da5f9fd5e6971c1773e81a9a24184b7aa5b543cec879001ae53" Oct 05 06:59:43 crc kubenswrapper[4846]: I1005 06:59:43.364767 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0a6f6b2c32dd3da5f9fd5e6971c1773e81a9a24184b7aa5b543cec879001ae53"} err="failed to get container status \"0a6f6b2c32dd3da5f9fd5e6971c1773e81a9a24184b7aa5b543cec879001ae53\": rpc error: code = NotFound desc = could not find container \"0a6f6b2c32dd3da5f9fd5e6971c1773e81a9a24184b7aa5b543cec879001ae53\": container with ID starting with 0a6f6b2c32dd3da5f9fd5e6971c1773e81a9a24184b7aa5b543cec879001ae53 not found: ID does not exist" Oct 05 06:59:43 crc kubenswrapper[4846]: I1005 06:59:43.364819 4846 scope.go:117] "RemoveContainer" containerID="1ad1940d07e37217089642478ece9ccdaca74784738a400929a9f5359d158c01" Oct 05 06:59:43 crc kubenswrapper[4846]: I1005 06:59:43.365352 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1ad1940d07e37217089642478ece9ccdaca74784738a400929a9f5359d158c01"} err="failed to get container status \"1ad1940d07e37217089642478ece9ccdaca74784738a400929a9f5359d158c01\": rpc error: code = NotFound desc = could not find container \"1ad1940d07e37217089642478ece9ccdaca74784738a400929a9f5359d158c01\": container with ID starting with 1ad1940d07e37217089642478ece9ccdaca74784738a400929a9f5359d158c01 not found: ID does not exist" Oct 05 06:59:43 crc kubenswrapper[4846]: I1005 06:59:43.365409 4846 scope.go:117] "RemoveContainer" containerID="8da85222775abc7a772f6a8ff8255f1a0d5b15807ba177a7e3e2b2d01a4f24b9" Oct 05 06:59:43 crc kubenswrapper[4846]: I1005 06:59:43.365958 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8da85222775abc7a772f6a8ff8255f1a0d5b15807ba177a7e3e2b2d01a4f24b9"} err="failed to get container status \"8da85222775abc7a772f6a8ff8255f1a0d5b15807ba177a7e3e2b2d01a4f24b9\": rpc error: code = NotFound desc = could not find container \"8da85222775abc7a772f6a8ff8255f1a0d5b15807ba177a7e3e2b2d01a4f24b9\": container with ID starting with 8da85222775abc7a772f6a8ff8255f1a0d5b15807ba177a7e3e2b2d01a4f24b9 not found: ID does not exist" Oct 05 06:59:43 crc kubenswrapper[4846]: I1005 06:59:43.366014 4846 scope.go:117] "RemoveContainer" containerID="1409651de8c897804f87be03d016c138f8acf27e662dce265a93e15c1848adcf" Oct 05 06:59:43 crc kubenswrapper[4846]: I1005 06:59:43.370640 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1409651de8c897804f87be03d016c138f8acf27e662dce265a93e15c1848adcf"} err="failed to get container status \"1409651de8c897804f87be03d016c138f8acf27e662dce265a93e15c1848adcf\": rpc error: code = NotFound desc = could not find container \"1409651de8c897804f87be03d016c138f8acf27e662dce265a93e15c1848adcf\": container with ID starting with 1409651de8c897804f87be03d016c138f8acf27e662dce265a93e15c1848adcf not found: ID does not exist" Oct 05 06:59:43 crc kubenswrapper[4846]: I1005 06:59:43.370704 4846 scope.go:117] "RemoveContainer" containerID="ff3c6545ff3d2436da27bbd223b950b30b0c69a3e7f3034cb0e61380cb38d338" Oct 05 06:59:43 crc kubenswrapper[4846]: I1005 06:59:43.371563 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ff3c6545ff3d2436da27bbd223b950b30b0c69a3e7f3034cb0e61380cb38d338"} err="failed to get container status \"ff3c6545ff3d2436da27bbd223b950b30b0c69a3e7f3034cb0e61380cb38d338\": rpc error: code = NotFound desc = could not find container \"ff3c6545ff3d2436da27bbd223b950b30b0c69a3e7f3034cb0e61380cb38d338\": container with ID starting with ff3c6545ff3d2436da27bbd223b950b30b0c69a3e7f3034cb0e61380cb38d338 not found: ID does not exist" Oct 05 06:59:43 crc kubenswrapper[4846]: I1005 06:59:43.371669 4846 scope.go:117] "RemoveContainer" containerID="32a57312adddc39f79d4dbc9dfb0197ec33241adfb654ed64db6bc10d1d479c8" Oct 05 06:59:43 crc kubenswrapper[4846]: I1005 06:59:43.372347 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"32a57312adddc39f79d4dbc9dfb0197ec33241adfb654ed64db6bc10d1d479c8"} err="failed to get container status \"32a57312adddc39f79d4dbc9dfb0197ec33241adfb654ed64db6bc10d1d479c8\": rpc error: code = NotFound desc = could not find container \"32a57312adddc39f79d4dbc9dfb0197ec33241adfb654ed64db6bc10d1d479c8\": container with ID starting with 32a57312adddc39f79d4dbc9dfb0197ec33241adfb654ed64db6bc10d1d479c8 not found: ID does not exist" Oct 05 06:59:43 crc kubenswrapper[4846]: I1005 06:59:43.372392 4846 scope.go:117] "RemoveContainer" containerID="59fa020f7bfea4547e132058cf3804a406ee0c06cdc4e78a08ae43a51f46633f" Oct 05 06:59:43 crc kubenswrapper[4846]: I1005 06:59:43.372868 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"59fa020f7bfea4547e132058cf3804a406ee0c06cdc4e78a08ae43a51f46633f"} err="failed to get container status \"59fa020f7bfea4547e132058cf3804a406ee0c06cdc4e78a08ae43a51f46633f\": rpc error: code = NotFound desc = could not find container \"59fa020f7bfea4547e132058cf3804a406ee0c06cdc4e78a08ae43a51f46633f\": container with ID starting with 59fa020f7bfea4547e132058cf3804a406ee0c06cdc4e78a08ae43a51f46633f not found: ID does not exist" Oct 05 06:59:43 crc kubenswrapper[4846]: I1005 06:59:43.372913 4846 scope.go:117] "RemoveContainer" containerID="0df0ebcd87dc5af4b17736b43ec7b2395968b19ed16e775366af14e2b2d49c52" Oct 05 06:59:43 crc kubenswrapper[4846]: I1005 06:59:43.373629 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0df0ebcd87dc5af4b17736b43ec7b2395968b19ed16e775366af14e2b2d49c52"} err="failed to get container status \"0df0ebcd87dc5af4b17736b43ec7b2395968b19ed16e775366af14e2b2d49c52\": rpc error: code = NotFound desc = could not find container \"0df0ebcd87dc5af4b17736b43ec7b2395968b19ed16e775366af14e2b2d49c52\": container with ID starting with 0df0ebcd87dc5af4b17736b43ec7b2395968b19ed16e775366af14e2b2d49c52 not found: ID does not exist" Oct 05 06:59:43 crc kubenswrapper[4846]: I1005 06:59:43.373680 4846 scope.go:117] "RemoveContainer" containerID="9e9b0219ecbc7e60905d6de3e67a2c8dc9af1017b4aed9538ee0c58c861f2438" Oct 05 06:59:43 crc kubenswrapper[4846]: I1005 06:59:43.374160 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9e9b0219ecbc7e60905d6de3e67a2c8dc9af1017b4aed9538ee0c58c861f2438"} err="failed to get container status \"9e9b0219ecbc7e60905d6de3e67a2c8dc9af1017b4aed9538ee0c58c861f2438\": rpc error: code = NotFound desc = could not find container \"9e9b0219ecbc7e60905d6de3e67a2c8dc9af1017b4aed9538ee0c58c861f2438\": container with ID starting with 9e9b0219ecbc7e60905d6de3e67a2c8dc9af1017b4aed9538ee0c58c861f2438 not found: ID does not exist" Oct 05 06:59:43 crc kubenswrapper[4846]: I1005 06:59:43.374246 4846 scope.go:117] "RemoveContainer" containerID="af5e5a4da9ad7371d5a24633e8268f12e3a7d419a8c93b11ddbcc006d8339c6c" Oct 05 06:59:43 crc kubenswrapper[4846]: I1005 06:59:43.377280 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"af5e5a4da9ad7371d5a24633e8268f12e3a7d419a8c93b11ddbcc006d8339c6c"} err="failed to get container status \"af5e5a4da9ad7371d5a24633e8268f12e3a7d419a8c93b11ddbcc006d8339c6c\": rpc error: code = NotFound desc = could not find container \"af5e5a4da9ad7371d5a24633e8268f12e3a7d419a8c93b11ddbcc006d8339c6c\": container with ID starting with af5e5a4da9ad7371d5a24633e8268f12e3a7d419a8c93b11ddbcc006d8339c6c not found: ID does not exist" Oct 05 06:59:43 crc kubenswrapper[4846]: I1005 06:59:43.377339 4846 scope.go:117] "RemoveContainer" containerID="0a6f6b2c32dd3da5f9fd5e6971c1773e81a9a24184b7aa5b543cec879001ae53" Oct 05 06:59:43 crc kubenswrapper[4846]: I1005 06:59:43.377864 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0a6f6b2c32dd3da5f9fd5e6971c1773e81a9a24184b7aa5b543cec879001ae53"} err="failed to get container status \"0a6f6b2c32dd3da5f9fd5e6971c1773e81a9a24184b7aa5b543cec879001ae53\": rpc error: code = NotFound desc = could not find container \"0a6f6b2c32dd3da5f9fd5e6971c1773e81a9a24184b7aa5b543cec879001ae53\": container with ID starting with 0a6f6b2c32dd3da5f9fd5e6971c1773e81a9a24184b7aa5b543cec879001ae53 not found: ID does not exist" Oct 05 06:59:43 crc kubenswrapper[4846]: I1005 06:59:43.377913 4846 scope.go:117] "RemoveContainer" containerID="1ad1940d07e37217089642478ece9ccdaca74784738a400929a9f5359d158c01" Oct 05 06:59:43 crc kubenswrapper[4846]: I1005 06:59:43.381446 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1ad1940d07e37217089642478ece9ccdaca74784738a400929a9f5359d158c01"} err="failed to get container status \"1ad1940d07e37217089642478ece9ccdaca74784738a400929a9f5359d158c01\": rpc error: code = NotFound desc = could not find container \"1ad1940d07e37217089642478ece9ccdaca74784738a400929a9f5359d158c01\": container with ID starting with 1ad1940d07e37217089642478ece9ccdaca74784738a400929a9f5359d158c01 not found: ID does not exist" Oct 05 06:59:43 crc kubenswrapper[4846]: I1005 06:59:43.381476 4846 scope.go:117] "RemoveContainer" containerID="8da85222775abc7a772f6a8ff8255f1a0d5b15807ba177a7e3e2b2d01a4f24b9" Oct 05 06:59:43 crc kubenswrapper[4846]: I1005 06:59:43.381817 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8da85222775abc7a772f6a8ff8255f1a0d5b15807ba177a7e3e2b2d01a4f24b9"} err="failed to get container status \"8da85222775abc7a772f6a8ff8255f1a0d5b15807ba177a7e3e2b2d01a4f24b9\": rpc error: code = NotFound desc = could not find container \"8da85222775abc7a772f6a8ff8255f1a0d5b15807ba177a7e3e2b2d01a4f24b9\": container with ID starting with 8da85222775abc7a772f6a8ff8255f1a0d5b15807ba177a7e3e2b2d01a4f24b9 not found: ID does not exist" Oct 05 06:59:43 crc kubenswrapper[4846]: I1005 06:59:43.381832 4846 scope.go:117] "RemoveContainer" containerID="1409651de8c897804f87be03d016c138f8acf27e662dce265a93e15c1848adcf" Oct 05 06:59:43 crc kubenswrapper[4846]: I1005 06:59:43.382147 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1409651de8c897804f87be03d016c138f8acf27e662dce265a93e15c1848adcf"} err="failed to get container status \"1409651de8c897804f87be03d016c138f8acf27e662dce265a93e15c1848adcf\": rpc error: code = NotFound desc = could not find container \"1409651de8c897804f87be03d016c138f8acf27e662dce265a93e15c1848adcf\": container with ID starting with 1409651de8c897804f87be03d016c138f8acf27e662dce265a93e15c1848adcf not found: ID does not exist" Oct 05 06:59:43 crc kubenswrapper[4846]: I1005 06:59:43.382162 4846 scope.go:117] "RemoveContainer" containerID="ff3c6545ff3d2436da27bbd223b950b30b0c69a3e7f3034cb0e61380cb38d338" Oct 05 06:59:43 crc kubenswrapper[4846]: I1005 06:59:43.382644 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ff3c6545ff3d2436da27bbd223b950b30b0c69a3e7f3034cb0e61380cb38d338"} err="failed to get container status \"ff3c6545ff3d2436da27bbd223b950b30b0c69a3e7f3034cb0e61380cb38d338\": rpc error: code = NotFound desc = could not find container \"ff3c6545ff3d2436da27bbd223b950b30b0c69a3e7f3034cb0e61380cb38d338\": container with ID starting with ff3c6545ff3d2436da27bbd223b950b30b0c69a3e7f3034cb0e61380cb38d338 not found: ID does not exist" Oct 05 06:59:43 crc kubenswrapper[4846]: I1005 06:59:43.382659 4846 scope.go:117] "RemoveContainer" containerID="32a57312adddc39f79d4dbc9dfb0197ec33241adfb654ed64db6bc10d1d479c8" Oct 05 06:59:43 crc kubenswrapper[4846]: I1005 06:59:43.382910 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"32a57312adddc39f79d4dbc9dfb0197ec33241adfb654ed64db6bc10d1d479c8"} err="failed to get container status \"32a57312adddc39f79d4dbc9dfb0197ec33241adfb654ed64db6bc10d1d479c8\": rpc error: code = NotFound desc = could not find container \"32a57312adddc39f79d4dbc9dfb0197ec33241adfb654ed64db6bc10d1d479c8\": container with ID starting with 32a57312adddc39f79d4dbc9dfb0197ec33241adfb654ed64db6bc10d1d479c8 not found: ID does not exist" Oct 05 06:59:43 crc kubenswrapper[4846]: I1005 06:59:43.382922 4846 scope.go:117] "RemoveContainer" containerID="59fa020f7bfea4547e132058cf3804a406ee0c06cdc4e78a08ae43a51f46633f" Oct 05 06:59:43 crc kubenswrapper[4846]: I1005 06:59:43.383153 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"59fa020f7bfea4547e132058cf3804a406ee0c06cdc4e78a08ae43a51f46633f"} err="failed to get container status \"59fa020f7bfea4547e132058cf3804a406ee0c06cdc4e78a08ae43a51f46633f\": rpc error: code = NotFound desc = could not find container \"59fa020f7bfea4547e132058cf3804a406ee0c06cdc4e78a08ae43a51f46633f\": container with ID starting with 59fa020f7bfea4547e132058cf3804a406ee0c06cdc4e78a08ae43a51f46633f not found: ID does not exist" Oct 05 06:59:43 crc kubenswrapper[4846]: I1005 06:59:43.383166 4846 scope.go:117] "RemoveContainer" containerID="0df0ebcd87dc5af4b17736b43ec7b2395968b19ed16e775366af14e2b2d49c52" Oct 05 06:59:43 crc kubenswrapper[4846]: I1005 06:59:43.383436 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0df0ebcd87dc5af4b17736b43ec7b2395968b19ed16e775366af14e2b2d49c52"} err="failed to get container status \"0df0ebcd87dc5af4b17736b43ec7b2395968b19ed16e775366af14e2b2d49c52\": rpc error: code = NotFound desc = could not find container \"0df0ebcd87dc5af4b17736b43ec7b2395968b19ed16e775366af14e2b2d49c52\": container with ID starting with 0df0ebcd87dc5af4b17736b43ec7b2395968b19ed16e775366af14e2b2d49c52 not found: ID does not exist" Oct 05 06:59:43 crc kubenswrapper[4846]: I1005 06:59:43.383451 4846 scope.go:117] "RemoveContainer" containerID="9e9b0219ecbc7e60905d6de3e67a2c8dc9af1017b4aed9538ee0c58c861f2438" Oct 05 06:59:43 crc kubenswrapper[4846]: I1005 06:59:43.383754 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9e9b0219ecbc7e60905d6de3e67a2c8dc9af1017b4aed9538ee0c58c861f2438"} err="failed to get container status \"9e9b0219ecbc7e60905d6de3e67a2c8dc9af1017b4aed9538ee0c58c861f2438\": rpc error: code = NotFound desc = could not find container \"9e9b0219ecbc7e60905d6de3e67a2c8dc9af1017b4aed9538ee0c58c861f2438\": container with ID starting with 9e9b0219ecbc7e60905d6de3e67a2c8dc9af1017b4aed9538ee0c58c861f2438 not found: ID does not exist" Oct 05 06:59:43 crc kubenswrapper[4846]: I1005 06:59:43.383772 4846 scope.go:117] "RemoveContainer" containerID="af5e5a4da9ad7371d5a24633e8268f12e3a7d419a8c93b11ddbcc006d8339c6c" Oct 05 06:59:43 crc kubenswrapper[4846]: I1005 06:59:43.384012 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"af5e5a4da9ad7371d5a24633e8268f12e3a7d419a8c93b11ddbcc006d8339c6c"} err="failed to get container status \"af5e5a4da9ad7371d5a24633e8268f12e3a7d419a8c93b11ddbcc006d8339c6c\": rpc error: code = NotFound desc = could not find container \"af5e5a4da9ad7371d5a24633e8268f12e3a7d419a8c93b11ddbcc006d8339c6c\": container with ID starting with af5e5a4da9ad7371d5a24633e8268f12e3a7d419a8c93b11ddbcc006d8339c6c not found: ID does not exist" Oct 05 06:59:43 crc kubenswrapper[4846]: I1005 06:59:43.384031 4846 scope.go:117] "RemoveContainer" containerID="0a6f6b2c32dd3da5f9fd5e6971c1773e81a9a24184b7aa5b543cec879001ae53" Oct 05 06:59:43 crc kubenswrapper[4846]: I1005 06:59:43.384340 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0a6f6b2c32dd3da5f9fd5e6971c1773e81a9a24184b7aa5b543cec879001ae53"} err="failed to get container status \"0a6f6b2c32dd3da5f9fd5e6971c1773e81a9a24184b7aa5b543cec879001ae53\": rpc error: code = NotFound desc = could not find container \"0a6f6b2c32dd3da5f9fd5e6971c1773e81a9a24184b7aa5b543cec879001ae53\": container with ID starting with 0a6f6b2c32dd3da5f9fd5e6971c1773e81a9a24184b7aa5b543cec879001ae53 not found: ID does not exist" Oct 05 06:59:43 crc kubenswrapper[4846]: I1005 06:59:43.384355 4846 scope.go:117] "RemoveContainer" containerID="1ad1940d07e37217089642478ece9ccdaca74784738a400929a9f5359d158c01" Oct 05 06:59:43 crc kubenswrapper[4846]: I1005 06:59:43.384802 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1ad1940d07e37217089642478ece9ccdaca74784738a400929a9f5359d158c01"} err="failed to get container status \"1ad1940d07e37217089642478ece9ccdaca74784738a400929a9f5359d158c01\": rpc error: code = NotFound desc = could not find container \"1ad1940d07e37217089642478ece9ccdaca74784738a400929a9f5359d158c01\": container with ID starting with 1ad1940d07e37217089642478ece9ccdaca74784738a400929a9f5359d158c01 not found: ID does not exist" Oct 05 06:59:43 crc kubenswrapper[4846]: I1005 06:59:43.384845 4846 scope.go:117] "RemoveContainer" containerID="8da85222775abc7a772f6a8ff8255f1a0d5b15807ba177a7e3e2b2d01a4f24b9" Oct 05 06:59:43 crc kubenswrapper[4846]: I1005 06:59:43.385612 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8da85222775abc7a772f6a8ff8255f1a0d5b15807ba177a7e3e2b2d01a4f24b9"} err="failed to get container status \"8da85222775abc7a772f6a8ff8255f1a0d5b15807ba177a7e3e2b2d01a4f24b9\": rpc error: code = NotFound desc = could not find container \"8da85222775abc7a772f6a8ff8255f1a0d5b15807ba177a7e3e2b2d01a4f24b9\": container with ID starting with 8da85222775abc7a772f6a8ff8255f1a0d5b15807ba177a7e3e2b2d01a4f24b9 not found: ID does not exist" Oct 05 06:59:43 crc kubenswrapper[4846]: I1005 06:59:43.385654 4846 scope.go:117] "RemoveContainer" containerID="1409651de8c897804f87be03d016c138f8acf27e662dce265a93e15c1848adcf" Oct 05 06:59:43 crc kubenswrapper[4846]: I1005 06:59:43.388850 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1409651de8c897804f87be03d016c138f8acf27e662dce265a93e15c1848adcf"} err="failed to get container status \"1409651de8c897804f87be03d016c138f8acf27e662dce265a93e15c1848adcf\": rpc error: code = NotFound desc = could not find container \"1409651de8c897804f87be03d016c138f8acf27e662dce265a93e15c1848adcf\": container with ID starting with 1409651de8c897804f87be03d016c138f8acf27e662dce265a93e15c1848adcf not found: ID does not exist" Oct 05 06:59:43 crc kubenswrapper[4846]: I1005 06:59:43.388903 4846 scope.go:117] "RemoveContainer" containerID="ff3c6545ff3d2436da27bbd223b950b30b0c69a3e7f3034cb0e61380cb38d338" Oct 05 06:59:43 crc kubenswrapper[4846]: I1005 06:59:43.391702 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ff3c6545ff3d2436da27bbd223b950b30b0c69a3e7f3034cb0e61380cb38d338"} err="failed to get container status \"ff3c6545ff3d2436da27bbd223b950b30b0c69a3e7f3034cb0e61380cb38d338\": rpc error: code = NotFound desc = could not find container \"ff3c6545ff3d2436da27bbd223b950b30b0c69a3e7f3034cb0e61380cb38d338\": container with ID starting with ff3c6545ff3d2436da27bbd223b950b30b0c69a3e7f3034cb0e61380cb38d338 not found: ID does not exist" Oct 05 06:59:44 crc kubenswrapper[4846]: I1005 06:59:44.057582 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-7wm5q_9bc7744f-d700-450f-8b8d-98140877fee9/kube-multus/2.log" Oct 05 06:59:44 crc kubenswrapper[4846]: I1005 06:59:44.060076 4846 generic.go:334] "Generic (PLEG): container finished" podID="778242f5-89b8-4f86-9f58-0a81e2d70569" containerID="1e3358e9c18b4506fa76b435540059fa5cbaa4634132262d7d0f86e351815ba5" exitCode=0 Oct 05 06:59:44 crc kubenswrapper[4846]: I1005 06:59:44.060134 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-q224c" event={"ID":"778242f5-89b8-4f86-9f58-0a81e2d70569","Type":"ContainerDied","Data":"1e3358e9c18b4506fa76b435540059fa5cbaa4634132262d7d0f86e351815ba5"} Oct 05 06:59:44 crc kubenswrapper[4846]: I1005 06:59:44.060212 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-q224c" event={"ID":"778242f5-89b8-4f86-9f58-0a81e2d70569","Type":"ContainerStarted","Data":"18d6a129ff6c7803d9184938a9064d0c8aaee60f7ba5adc13a3021eb8c378fcf"} Oct 05 06:59:44 crc kubenswrapper[4846]: I1005 06:59:44.505735 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="93a6ace4-8d64-44d7-9637-457d3af6543c" path="/var/lib/kubelet/pods/93a6ace4-8d64-44d7-9637-457d3af6543c/volumes" Oct 05 06:59:45 crc kubenswrapper[4846]: I1005 06:59:45.079749 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-q224c" event={"ID":"778242f5-89b8-4f86-9f58-0a81e2d70569","Type":"ContainerStarted","Data":"5b07280b8227f49aebcd757a04cad0da90f20be4a6a7d65c4dd6502fb363faca"} Oct 05 06:59:45 crc kubenswrapper[4846]: I1005 06:59:45.080304 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-q224c" event={"ID":"778242f5-89b8-4f86-9f58-0a81e2d70569","Type":"ContainerStarted","Data":"8efb3a4e1e88a3761a7e2876e7162341e477eb2ad4ac8e2d034bc2c7bc397f79"} Oct 05 06:59:45 crc kubenswrapper[4846]: I1005 06:59:45.080329 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-q224c" event={"ID":"778242f5-89b8-4f86-9f58-0a81e2d70569","Type":"ContainerStarted","Data":"2c78d390af0e1d4b73bb47131ccbf1523b62b0d0d3b6d1988c8a7a148d3d8427"} Oct 05 06:59:45 crc kubenswrapper[4846]: I1005 06:59:45.080350 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-q224c" event={"ID":"778242f5-89b8-4f86-9f58-0a81e2d70569","Type":"ContainerStarted","Data":"f9fedc06b1216629f311ab7f71aed5e5b3eb6da9ccd50f558ae4dbf7c668c1f2"} Oct 05 06:59:45 crc kubenswrapper[4846]: I1005 06:59:45.080370 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-q224c" event={"ID":"778242f5-89b8-4f86-9f58-0a81e2d70569","Type":"ContainerStarted","Data":"5905445fa0e1c850f51a762a7b37032e3f4c6b9c601ded1daaecabe850b64193"} Oct 05 06:59:45 crc kubenswrapper[4846]: I1005 06:59:45.080387 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-q224c" event={"ID":"778242f5-89b8-4f86-9f58-0a81e2d70569","Type":"ContainerStarted","Data":"4ed19e4051d16ec6282a3ef6f38f8787b627768572ddb1a1eda7d28c94c32d48"} Oct 05 06:59:48 crc kubenswrapper[4846]: I1005 06:59:48.121612 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-q224c" event={"ID":"778242f5-89b8-4f86-9f58-0a81e2d70569","Type":"ContainerStarted","Data":"ea530054fe3eba981f23d99fa3c24705f026381173320cd149b7da07bd5a525f"} Oct 05 06:59:50 crc kubenswrapper[4846]: I1005 06:59:50.153755 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-q224c" event={"ID":"778242f5-89b8-4f86-9f58-0a81e2d70569","Type":"ContainerStarted","Data":"4e4c230fc1eab628596be4be40a15d3b709f05143c28f16cce6a078bac1168d0"} Oct 05 06:59:50 crc kubenswrapper[4846]: I1005 06:59:50.154151 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-q224c" Oct 05 06:59:50 crc kubenswrapper[4846]: I1005 06:59:50.154168 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-q224c" Oct 05 06:59:50 crc kubenswrapper[4846]: I1005 06:59:50.190382 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-q224c" Oct 05 06:59:50 crc kubenswrapper[4846]: I1005 06:59:50.194472 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-q224c" podStartSLOduration=8.194451617 podStartE2EDuration="8.194451617s" podCreationTimestamp="2025-10-05 06:59:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 06:59:50.191569161 +0000 UTC m=+712.432421946" watchObservedRunningTime="2025-10-05 06:59:50.194451617 +0000 UTC m=+712.435304392" Oct 05 06:59:51 crc kubenswrapper[4846]: I1005 06:59:51.160406 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-q224c" Oct 05 06:59:51 crc kubenswrapper[4846]: I1005 06:59:51.190721 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-q224c" Oct 05 06:59:52 crc kubenswrapper[4846]: I1005 06:59:52.448761 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["crc-storage/crc-storage-crc-dr8j2"] Oct 05 06:59:52 crc kubenswrapper[4846]: I1005 06:59:52.451107 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-dr8j2" Oct 05 06:59:52 crc kubenswrapper[4846]: I1005 06:59:52.455959 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"crc-storage" Oct 05 06:59:52 crc kubenswrapper[4846]: I1005 06:59:52.456335 4846 reflector.go:368] Caches populated for *v1.Secret from object-"crc-storage"/"crc-storage-dockercfg-w2v6h" Oct 05 06:59:52 crc kubenswrapper[4846]: I1005 06:59:52.456621 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"openshift-service-ca.crt" Oct 05 06:59:52 crc kubenswrapper[4846]: I1005 06:59:52.457867 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"kube-root-ca.crt" Oct 05 06:59:52 crc kubenswrapper[4846]: I1005 06:59:52.464308 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["crc-storage/crc-storage-crc-dr8j2"] Oct 05 06:59:52 crc kubenswrapper[4846]: I1005 06:59:52.549998 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5cmdk\" (UniqueName: \"kubernetes.io/projected/a3699d3c-ac3f-4e15-983c-1b6332c9accc-kube-api-access-5cmdk\") pod \"crc-storage-crc-dr8j2\" (UID: \"a3699d3c-ac3f-4e15-983c-1b6332c9accc\") " pod="crc-storage/crc-storage-crc-dr8j2" Oct 05 06:59:52 crc kubenswrapper[4846]: I1005 06:59:52.550092 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/a3699d3c-ac3f-4e15-983c-1b6332c9accc-node-mnt\") pod \"crc-storage-crc-dr8j2\" (UID: \"a3699d3c-ac3f-4e15-983c-1b6332c9accc\") " pod="crc-storage/crc-storage-crc-dr8j2" Oct 05 06:59:52 crc kubenswrapper[4846]: I1005 06:59:52.550160 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/a3699d3c-ac3f-4e15-983c-1b6332c9accc-crc-storage\") pod \"crc-storage-crc-dr8j2\" (UID: \"a3699d3c-ac3f-4e15-983c-1b6332c9accc\") " pod="crc-storage/crc-storage-crc-dr8j2" Oct 05 06:59:52 crc kubenswrapper[4846]: I1005 06:59:52.652408 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5cmdk\" (UniqueName: \"kubernetes.io/projected/a3699d3c-ac3f-4e15-983c-1b6332c9accc-kube-api-access-5cmdk\") pod \"crc-storage-crc-dr8j2\" (UID: \"a3699d3c-ac3f-4e15-983c-1b6332c9accc\") " pod="crc-storage/crc-storage-crc-dr8j2" Oct 05 06:59:52 crc kubenswrapper[4846]: I1005 06:59:52.652566 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/a3699d3c-ac3f-4e15-983c-1b6332c9accc-node-mnt\") pod \"crc-storage-crc-dr8j2\" (UID: \"a3699d3c-ac3f-4e15-983c-1b6332c9accc\") " pod="crc-storage/crc-storage-crc-dr8j2" Oct 05 06:59:52 crc kubenswrapper[4846]: I1005 06:59:52.652624 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/a3699d3c-ac3f-4e15-983c-1b6332c9accc-crc-storage\") pod \"crc-storage-crc-dr8j2\" (UID: \"a3699d3c-ac3f-4e15-983c-1b6332c9accc\") " pod="crc-storage/crc-storage-crc-dr8j2" Oct 05 06:59:52 crc kubenswrapper[4846]: I1005 06:59:52.653210 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/a3699d3c-ac3f-4e15-983c-1b6332c9accc-node-mnt\") pod \"crc-storage-crc-dr8j2\" (UID: \"a3699d3c-ac3f-4e15-983c-1b6332c9accc\") " pod="crc-storage/crc-storage-crc-dr8j2" Oct 05 06:59:52 crc kubenswrapper[4846]: I1005 06:59:52.654014 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/a3699d3c-ac3f-4e15-983c-1b6332c9accc-crc-storage\") pod \"crc-storage-crc-dr8j2\" (UID: \"a3699d3c-ac3f-4e15-983c-1b6332c9accc\") " pod="crc-storage/crc-storage-crc-dr8j2" Oct 05 06:59:52 crc kubenswrapper[4846]: I1005 06:59:52.690620 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5cmdk\" (UniqueName: \"kubernetes.io/projected/a3699d3c-ac3f-4e15-983c-1b6332c9accc-kube-api-access-5cmdk\") pod \"crc-storage-crc-dr8j2\" (UID: \"a3699d3c-ac3f-4e15-983c-1b6332c9accc\") " pod="crc-storage/crc-storage-crc-dr8j2" Oct 05 06:59:52 crc kubenswrapper[4846]: I1005 06:59:52.787303 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-dr8j2" Oct 05 06:59:52 crc kubenswrapper[4846]: E1005 06:59:52.838843 4846 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-dr8j2_crc-storage_a3699d3c-ac3f-4e15-983c-1b6332c9accc_0(d7ed0193628ee7e4f0ed9f1dd256b20275bc9c5b8602f430278c43d7d5ca9bae): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Oct 05 06:59:52 crc kubenswrapper[4846]: E1005 06:59:52.839155 4846 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-dr8j2_crc-storage_a3699d3c-ac3f-4e15-983c-1b6332c9accc_0(d7ed0193628ee7e4f0ed9f1dd256b20275bc9c5b8602f430278c43d7d5ca9bae): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="crc-storage/crc-storage-crc-dr8j2" Oct 05 06:59:52 crc kubenswrapper[4846]: E1005 06:59:52.839247 4846 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-dr8j2_crc-storage_a3699d3c-ac3f-4e15-983c-1b6332c9accc_0(d7ed0193628ee7e4f0ed9f1dd256b20275bc9c5b8602f430278c43d7d5ca9bae): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="crc-storage/crc-storage-crc-dr8j2" Oct 05 06:59:52 crc kubenswrapper[4846]: E1005 06:59:52.839416 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"crc-storage-crc-dr8j2_crc-storage(a3699d3c-ac3f-4e15-983c-1b6332c9accc)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"crc-storage-crc-dr8j2_crc-storage(a3699d3c-ac3f-4e15-983c-1b6332c9accc)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-dr8j2_crc-storage_a3699d3c-ac3f-4e15-983c-1b6332c9accc_0(d7ed0193628ee7e4f0ed9f1dd256b20275bc9c5b8602f430278c43d7d5ca9bae): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="crc-storage/crc-storage-crc-dr8j2" podUID="a3699d3c-ac3f-4e15-983c-1b6332c9accc" Oct 05 06:59:53 crc kubenswrapper[4846]: I1005 06:59:53.173336 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-dr8j2" Oct 05 06:59:53 crc kubenswrapper[4846]: I1005 06:59:53.174653 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-dr8j2" Oct 05 06:59:53 crc kubenswrapper[4846]: E1005 06:59:53.202365 4846 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-dr8j2_crc-storage_a3699d3c-ac3f-4e15-983c-1b6332c9accc_0(37ac6922ec4dd85e8c96987b7bd9d75bea176f8f42de8569e91954d339727a30): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Oct 05 06:59:53 crc kubenswrapper[4846]: E1005 06:59:53.202456 4846 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-dr8j2_crc-storage_a3699d3c-ac3f-4e15-983c-1b6332c9accc_0(37ac6922ec4dd85e8c96987b7bd9d75bea176f8f42de8569e91954d339727a30): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="crc-storage/crc-storage-crc-dr8j2" Oct 05 06:59:53 crc kubenswrapper[4846]: E1005 06:59:53.202489 4846 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-dr8j2_crc-storage_a3699d3c-ac3f-4e15-983c-1b6332c9accc_0(37ac6922ec4dd85e8c96987b7bd9d75bea176f8f42de8569e91954d339727a30): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="crc-storage/crc-storage-crc-dr8j2" Oct 05 06:59:53 crc kubenswrapper[4846]: E1005 06:59:53.202551 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"crc-storage-crc-dr8j2_crc-storage(a3699d3c-ac3f-4e15-983c-1b6332c9accc)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"crc-storage-crc-dr8j2_crc-storage(a3699d3c-ac3f-4e15-983c-1b6332c9accc)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-dr8j2_crc-storage_a3699d3c-ac3f-4e15-983c-1b6332c9accc_0(37ac6922ec4dd85e8c96987b7bd9d75bea176f8f42de8569e91954d339727a30): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="crc-storage/crc-storage-crc-dr8j2" podUID="a3699d3c-ac3f-4e15-983c-1b6332c9accc" Oct 05 06:59:54 crc kubenswrapper[4846]: I1005 06:59:54.498390 4846 scope.go:117] "RemoveContainer" containerID="75ffecdf3834699d0ea47cca69028168c32263e5b5ddc5043cc0c1dc1d7bf74d" Oct 05 06:59:54 crc kubenswrapper[4846]: E1005 06:59:54.505257 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-7wm5q_openshift-multus(9bc7744f-d700-450f-8b8d-98140877fee9)\"" pod="openshift-multus/multus-7wm5q" podUID="9bc7744f-d700-450f-8b8d-98140877fee9" Oct 05 07:00:00 crc kubenswrapper[4846]: I1005 07:00:00.155882 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29327460-x7z7p"] Oct 05 07:00:00 crc kubenswrapper[4846]: I1005 07:00:00.158277 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29327460-x7z7p" Oct 05 07:00:00 crc kubenswrapper[4846]: I1005 07:00:00.161239 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 05 07:00:00 crc kubenswrapper[4846]: I1005 07:00:00.161710 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 05 07:00:00 crc kubenswrapper[4846]: I1005 07:00:00.177088 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29327460-x7z7p"] Oct 05 07:00:00 crc kubenswrapper[4846]: I1005 07:00:00.182558 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6fc79ab2-6284-412f-b641-01e7d1f4b1c6-config-volume\") pod \"collect-profiles-29327460-x7z7p\" (UID: \"6fc79ab2-6284-412f-b641-01e7d1f4b1c6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327460-x7z7p" Oct 05 07:00:00 crc kubenswrapper[4846]: I1005 07:00:00.182651 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/6fc79ab2-6284-412f-b641-01e7d1f4b1c6-secret-volume\") pod \"collect-profiles-29327460-x7z7p\" (UID: \"6fc79ab2-6284-412f-b641-01e7d1f4b1c6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327460-x7z7p" Oct 05 07:00:00 crc kubenswrapper[4846]: I1005 07:00:00.182795 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ksfsl\" (UniqueName: \"kubernetes.io/projected/6fc79ab2-6284-412f-b641-01e7d1f4b1c6-kube-api-access-ksfsl\") pod \"collect-profiles-29327460-x7z7p\" (UID: \"6fc79ab2-6284-412f-b641-01e7d1f4b1c6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327460-x7z7p" Oct 05 07:00:00 crc kubenswrapper[4846]: I1005 07:00:00.284373 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ksfsl\" (UniqueName: \"kubernetes.io/projected/6fc79ab2-6284-412f-b641-01e7d1f4b1c6-kube-api-access-ksfsl\") pod \"collect-profiles-29327460-x7z7p\" (UID: \"6fc79ab2-6284-412f-b641-01e7d1f4b1c6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327460-x7z7p" Oct 05 07:00:00 crc kubenswrapper[4846]: I1005 07:00:00.284547 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6fc79ab2-6284-412f-b641-01e7d1f4b1c6-config-volume\") pod \"collect-profiles-29327460-x7z7p\" (UID: \"6fc79ab2-6284-412f-b641-01e7d1f4b1c6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327460-x7z7p" Oct 05 07:00:00 crc kubenswrapper[4846]: I1005 07:00:00.284619 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/6fc79ab2-6284-412f-b641-01e7d1f4b1c6-secret-volume\") pod \"collect-profiles-29327460-x7z7p\" (UID: \"6fc79ab2-6284-412f-b641-01e7d1f4b1c6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327460-x7z7p" Oct 05 07:00:00 crc kubenswrapper[4846]: I1005 07:00:00.285928 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6fc79ab2-6284-412f-b641-01e7d1f4b1c6-config-volume\") pod \"collect-profiles-29327460-x7z7p\" (UID: \"6fc79ab2-6284-412f-b641-01e7d1f4b1c6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327460-x7z7p" Oct 05 07:00:00 crc kubenswrapper[4846]: I1005 07:00:00.294468 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/6fc79ab2-6284-412f-b641-01e7d1f4b1c6-secret-volume\") pod \"collect-profiles-29327460-x7z7p\" (UID: \"6fc79ab2-6284-412f-b641-01e7d1f4b1c6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327460-x7z7p" Oct 05 07:00:00 crc kubenswrapper[4846]: I1005 07:00:00.316418 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ksfsl\" (UniqueName: \"kubernetes.io/projected/6fc79ab2-6284-412f-b641-01e7d1f4b1c6-kube-api-access-ksfsl\") pod \"collect-profiles-29327460-x7z7p\" (UID: \"6fc79ab2-6284-412f-b641-01e7d1f4b1c6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327460-x7z7p" Oct 05 07:00:00 crc kubenswrapper[4846]: I1005 07:00:00.487865 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29327460-x7z7p" Oct 05 07:00:00 crc kubenswrapper[4846]: E1005 07:00:00.526505 4846 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_collect-profiles-29327460-x7z7p_openshift-operator-lifecycle-manager_6fc79ab2-6284-412f-b641-01e7d1f4b1c6_0(842cabd3f733dfb80005b2f550e5255959f54184840ae0c94f4f5031316703bd): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Oct 05 07:00:00 crc kubenswrapper[4846]: E1005 07:00:00.526626 4846 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_collect-profiles-29327460-x7z7p_openshift-operator-lifecycle-manager_6fc79ab2-6284-412f-b641-01e7d1f4b1c6_0(842cabd3f733dfb80005b2f550e5255959f54184840ae0c94f4f5031316703bd): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operator-lifecycle-manager/collect-profiles-29327460-x7z7p" Oct 05 07:00:00 crc kubenswrapper[4846]: E1005 07:00:00.526665 4846 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_collect-profiles-29327460-x7z7p_openshift-operator-lifecycle-manager_6fc79ab2-6284-412f-b641-01e7d1f4b1c6_0(842cabd3f733dfb80005b2f550e5255959f54184840ae0c94f4f5031316703bd): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operator-lifecycle-manager/collect-profiles-29327460-x7z7p" Oct 05 07:00:00 crc kubenswrapper[4846]: E1005 07:00:00.526753 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"collect-profiles-29327460-x7z7p_openshift-operator-lifecycle-manager(6fc79ab2-6284-412f-b641-01e7d1f4b1c6)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"collect-profiles-29327460-x7z7p_openshift-operator-lifecycle-manager(6fc79ab2-6284-412f-b641-01e7d1f4b1c6)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_collect-profiles-29327460-x7z7p_openshift-operator-lifecycle-manager_6fc79ab2-6284-412f-b641-01e7d1f4b1c6_0(842cabd3f733dfb80005b2f550e5255959f54184840ae0c94f4f5031316703bd): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operator-lifecycle-manager/collect-profiles-29327460-x7z7p" podUID="6fc79ab2-6284-412f-b641-01e7d1f4b1c6" Oct 05 07:00:01 crc kubenswrapper[4846]: I1005 07:00:01.235299 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29327460-x7z7p" Oct 05 07:00:01 crc kubenswrapper[4846]: I1005 07:00:01.236107 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29327460-x7z7p" Oct 05 07:00:01 crc kubenswrapper[4846]: E1005 07:00:01.274227 4846 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_collect-profiles-29327460-x7z7p_openshift-operator-lifecycle-manager_6fc79ab2-6284-412f-b641-01e7d1f4b1c6_0(2e59ce8f729d24b5aa47d662df065344721315a340e4d087c342a5a7186d076d): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Oct 05 07:00:01 crc kubenswrapper[4846]: E1005 07:00:01.274798 4846 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_collect-profiles-29327460-x7z7p_openshift-operator-lifecycle-manager_6fc79ab2-6284-412f-b641-01e7d1f4b1c6_0(2e59ce8f729d24b5aa47d662df065344721315a340e4d087c342a5a7186d076d): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operator-lifecycle-manager/collect-profiles-29327460-x7z7p" Oct 05 07:00:01 crc kubenswrapper[4846]: E1005 07:00:01.274837 4846 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_collect-profiles-29327460-x7z7p_openshift-operator-lifecycle-manager_6fc79ab2-6284-412f-b641-01e7d1f4b1c6_0(2e59ce8f729d24b5aa47d662df065344721315a340e4d087c342a5a7186d076d): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operator-lifecycle-manager/collect-profiles-29327460-x7z7p" Oct 05 07:00:01 crc kubenswrapper[4846]: E1005 07:00:01.274926 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"collect-profiles-29327460-x7z7p_openshift-operator-lifecycle-manager(6fc79ab2-6284-412f-b641-01e7d1f4b1c6)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"collect-profiles-29327460-x7z7p_openshift-operator-lifecycle-manager(6fc79ab2-6284-412f-b641-01e7d1f4b1c6)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_collect-profiles-29327460-x7z7p_openshift-operator-lifecycle-manager_6fc79ab2-6284-412f-b641-01e7d1f4b1c6_0(2e59ce8f729d24b5aa47d662df065344721315a340e4d087c342a5a7186d076d): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operator-lifecycle-manager/collect-profiles-29327460-x7z7p" podUID="6fc79ab2-6284-412f-b641-01e7d1f4b1c6" Oct 05 07:00:06 crc kubenswrapper[4846]: I1005 07:00:06.498165 4846 scope.go:117] "RemoveContainer" containerID="75ffecdf3834699d0ea47cca69028168c32263e5b5ddc5043cc0c1dc1d7bf74d" Oct 05 07:00:07 crc kubenswrapper[4846]: I1005 07:00:07.280353 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-7wm5q_9bc7744f-d700-450f-8b8d-98140877fee9/kube-multus/2.log" Oct 05 07:00:07 crc kubenswrapper[4846]: I1005 07:00:07.281054 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-7wm5q" event={"ID":"9bc7744f-d700-450f-8b8d-98140877fee9","Type":"ContainerStarted","Data":"146b78e679e6d32ba6eb7e43a274adb237e57e8e1a6cacf78423af97f6c84d5f"} Oct 05 07:00:08 crc kubenswrapper[4846]: I1005 07:00:08.497842 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-dr8j2" Oct 05 07:00:08 crc kubenswrapper[4846]: I1005 07:00:08.505444 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-dr8j2" Oct 05 07:00:08 crc kubenswrapper[4846]: I1005 07:00:08.844070 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["crc-storage/crc-storage-crc-dr8j2"] Oct 05 07:00:08 crc kubenswrapper[4846]: I1005 07:00:08.862867 4846 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 05 07:00:09 crc kubenswrapper[4846]: I1005 07:00:09.298985 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-dr8j2" event={"ID":"a3699d3c-ac3f-4e15-983c-1b6332c9accc","Type":"ContainerStarted","Data":"6d6e52e8888bf88da2766a29670ac227ba7afd94ac7af501440a848f6d3c5cbf"} Oct 05 07:00:11 crc kubenswrapper[4846]: I1005 07:00:11.315854 4846 generic.go:334] "Generic (PLEG): container finished" podID="a3699d3c-ac3f-4e15-983c-1b6332c9accc" containerID="c428789eb7e249054d1de1e757e22c2d6318371fc59227c254b993a1c774bb6c" exitCode=0 Oct 05 07:00:11 crc kubenswrapper[4846]: I1005 07:00:11.315972 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-dr8j2" event={"ID":"a3699d3c-ac3f-4e15-983c-1b6332c9accc","Type":"ContainerDied","Data":"c428789eb7e249054d1de1e757e22c2d6318371fc59227c254b993a1c774bb6c"} Oct 05 07:00:12 crc kubenswrapper[4846]: I1005 07:00:12.496969 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29327460-x7z7p" Oct 05 07:00:12 crc kubenswrapper[4846]: I1005 07:00:12.499343 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29327460-x7z7p" Oct 05 07:00:12 crc kubenswrapper[4846]: I1005 07:00:12.662841 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-dr8j2" Oct 05 07:00:12 crc kubenswrapper[4846]: I1005 07:00:12.801362 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29327460-x7z7p"] Oct 05 07:00:12 crc kubenswrapper[4846]: W1005 07:00:12.811828 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6fc79ab2_6284_412f_b641_01e7d1f4b1c6.slice/crio-c1f891c6f51b7f25480a97b3cff95e703ce41cf047db7cd936b8a9e4efe6e44f WatchSource:0}: Error finding container c1f891c6f51b7f25480a97b3cff95e703ce41cf047db7cd936b8a9e4efe6e44f: Status 404 returned error can't find the container with id c1f891c6f51b7f25480a97b3cff95e703ce41cf047db7cd936b8a9e4efe6e44f Oct 05 07:00:12 crc kubenswrapper[4846]: I1005 07:00:12.815881 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/a3699d3c-ac3f-4e15-983c-1b6332c9accc-node-mnt\") pod \"a3699d3c-ac3f-4e15-983c-1b6332c9accc\" (UID: \"a3699d3c-ac3f-4e15-983c-1b6332c9accc\") " Oct 05 07:00:12 crc kubenswrapper[4846]: I1005 07:00:12.815976 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a3699d3c-ac3f-4e15-983c-1b6332c9accc-node-mnt" (OuterVolumeSpecName: "node-mnt") pod "a3699d3c-ac3f-4e15-983c-1b6332c9accc" (UID: "a3699d3c-ac3f-4e15-983c-1b6332c9accc"). InnerVolumeSpecName "node-mnt". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 05 07:00:12 crc kubenswrapper[4846]: I1005 07:00:12.816069 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5cmdk\" (UniqueName: \"kubernetes.io/projected/a3699d3c-ac3f-4e15-983c-1b6332c9accc-kube-api-access-5cmdk\") pod \"a3699d3c-ac3f-4e15-983c-1b6332c9accc\" (UID: \"a3699d3c-ac3f-4e15-983c-1b6332c9accc\") " Oct 05 07:00:12 crc kubenswrapper[4846]: I1005 07:00:12.816142 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/a3699d3c-ac3f-4e15-983c-1b6332c9accc-crc-storage\") pod \"a3699d3c-ac3f-4e15-983c-1b6332c9accc\" (UID: \"a3699d3c-ac3f-4e15-983c-1b6332c9accc\") " Oct 05 07:00:12 crc kubenswrapper[4846]: I1005 07:00:12.816448 4846 reconciler_common.go:293] "Volume detached for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/a3699d3c-ac3f-4e15-983c-1b6332c9accc-node-mnt\") on node \"crc\" DevicePath \"\"" Oct 05 07:00:12 crc kubenswrapper[4846]: I1005 07:00:12.826814 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a3699d3c-ac3f-4e15-983c-1b6332c9accc-kube-api-access-5cmdk" (OuterVolumeSpecName: "kube-api-access-5cmdk") pod "a3699d3c-ac3f-4e15-983c-1b6332c9accc" (UID: "a3699d3c-ac3f-4e15-983c-1b6332c9accc"). InnerVolumeSpecName "kube-api-access-5cmdk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:00:12 crc kubenswrapper[4846]: I1005 07:00:12.846315 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a3699d3c-ac3f-4e15-983c-1b6332c9accc-crc-storage" (OuterVolumeSpecName: "crc-storage") pod "a3699d3c-ac3f-4e15-983c-1b6332c9accc" (UID: "a3699d3c-ac3f-4e15-983c-1b6332c9accc"). InnerVolumeSpecName "crc-storage". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 07:00:12 crc kubenswrapper[4846]: I1005 07:00:12.917704 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5cmdk\" (UniqueName: \"kubernetes.io/projected/a3699d3c-ac3f-4e15-983c-1b6332c9accc-kube-api-access-5cmdk\") on node \"crc\" DevicePath \"\"" Oct 05 07:00:12 crc kubenswrapper[4846]: I1005 07:00:12.917766 4846 reconciler_common.go:293] "Volume detached for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/a3699d3c-ac3f-4e15-983c-1b6332c9accc-crc-storage\") on node \"crc\" DevicePath \"\"" Oct 05 07:00:13 crc kubenswrapper[4846]: I1005 07:00:13.181791 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-q224c" Oct 05 07:00:13 crc kubenswrapper[4846]: I1005 07:00:13.333927 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-dr8j2" Oct 05 07:00:13 crc kubenswrapper[4846]: I1005 07:00:13.333998 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-dr8j2" event={"ID":"a3699d3c-ac3f-4e15-983c-1b6332c9accc","Type":"ContainerDied","Data":"6d6e52e8888bf88da2766a29670ac227ba7afd94ac7af501440a848f6d3c5cbf"} Oct 05 07:00:13 crc kubenswrapper[4846]: I1005 07:00:13.334069 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6d6e52e8888bf88da2766a29670ac227ba7afd94ac7af501440a848f6d3c5cbf" Oct 05 07:00:13 crc kubenswrapper[4846]: I1005 07:00:13.336952 4846 generic.go:334] "Generic (PLEG): container finished" podID="6fc79ab2-6284-412f-b641-01e7d1f4b1c6" containerID="4d929a7b2e9007b4930575af61db05d35381bb42b7507051e5aadda242a701bb" exitCode=0 Oct 05 07:00:13 crc kubenswrapper[4846]: I1005 07:00:13.337124 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29327460-x7z7p" event={"ID":"6fc79ab2-6284-412f-b641-01e7d1f4b1c6","Type":"ContainerDied","Data":"4d929a7b2e9007b4930575af61db05d35381bb42b7507051e5aadda242a701bb"} Oct 05 07:00:13 crc kubenswrapper[4846]: I1005 07:00:13.337888 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29327460-x7z7p" event={"ID":"6fc79ab2-6284-412f-b641-01e7d1f4b1c6","Type":"ContainerStarted","Data":"c1f891c6f51b7f25480a97b3cff95e703ce41cf047db7cd936b8a9e4efe6e44f"} Oct 05 07:00:14 crc kubenswrapper[4846]: I1005 07:00:14.734353 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29327460-x7z7p" Oct 05 07:00:14 crc kubenswrapper[4846]: I1005 07:00:14.848538 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/6fc79ab2-6284-412f-b641-01e7d1f4b1c6-secret-volume\") pod \"6fc79ab2-6284-412f-b641-01e7d1f4b1c6\" (UID: \"6fc79ab2-6284-412f-b641-01e7d1f4b1c6\") " Oct 05 07:00:14 crc kubenswrapper[4846]: I1005 07:00:14.848676 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6fc79ab2-6284-412f-b641-01e7d1f4b1c6-config-volume\") pod \"6fc79ab2-6284-412f-b641-01e7d1f4b1c6\" (UID: \"6fc79ab2-6284-412f-b641-01e7d1f4b1c6\") " Oct 05 07:00:14 crc kubenswrapper[4846]: I1005 07:00:14.848770 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ksfsl\" (UniqueName: \"kubernetes.io/projected/6fc79ab2-6284-412f-b641-01e7d1f4b1c6-kube-api-access-ksfsl\") pod \"6fc79ab2-6284-412f-b641-01e7d1f4b1c6\" (UID: \"6fc79ab2-6284-412f-b641-01e7d1f4b1c6\") " Oct 05 07:00:14 crc kubenswrapper[4846]: I1005 07:00:14.849829 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6fc79ab2-6284-412f-b641-01e7d1f4b1c6-config-volume" (OuterVolumeSpecName: "config-volume") pod "6fc79ab2-6284-412f-b641-01e7d1f4b1c6" (UID: "6fc79ab2-6284-412f-b641-01e7d1f4b1c6"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 07:00:14 crc kubenswrapper[4846]: I1005 07:00:14.858222 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6fc79ab2-6284-412f-b641-01e7d1f4b1c6-kube-api-access-ksfsl" (OuterVolumeSpecName: "kube-api-access-ksfsl") pod "6fc79ab2-6284-412f-b641-01e7d1f4b1c6" (UID: "6fc79ab2-6284-412f-b641-01e7d1f4b1c6"). InnerVolumeSpecName "kube-api-access-ksfsl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:00:14 crc kubenswrapper[4846]: I1005 07:00:14.858358 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6fc79ab2-6284-412f-b641-01e7d1f4b1c6-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "6fc79ab2-6284-412f-b641-01e7d1f4b1c6" (UID: "6fc79ab2-6284-412f-b641-01e7d1f4b1c6"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:00:14 crc kubenswrapper[4846]: I1005 07:00:14.950365 4846 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6fc79ab2-6284-412f-b641-01e7d1f4b1c6-config-volume\") on node \"crc\" DevicePath \"\"" Oct 05 07:00:14 crc kubenswrapper[4846]: I1005 07:00:14.950981 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ksfsl\" (UniqueName: \"kubernetes.io/projected/6fc79ab2-6284-412f-b641-01e7d1f4b1c6-kube-api-access-ksfsl\") on node \"crc\" DevicePath \"\"" Oct 05 07:00:14 crc kubenswrapper[4846]: I1005 07:00:14.951006 4846 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/6fc79ab2-6284-412f-b641-01e7d1f4b1c6-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 05 07:00:15 crc kubenswrapper[4846]: I1005 07:00:15.356961 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29327460-x7z7p" event={"ID":"6fc79ab2-6284-412f-b641-01e7d1f4b1c6","Type":"ContainerDied","Data":"c1f891c6f51b7f25480a97b3cff95e703ce41cf047db7cd936b8a9e4efe6e44f"} Oct 05 07:00:15 crc kubenswrapper[4846]: I1005 07:00:15.357034 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c1f891c6f51b7f25480a97b3cff95e703ce41cf047db7cd936b8a9e4efe6e44f" Oct 05 07:00:15 crc kubenswrapper[4846]: I1005 07:00:15.357378 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29327460-x7z7p" Oct 05 07:00:21 crc kubenswrapper[4846]: I1005 07:00:21.613138 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c65gkl"] Oct 05 07:00:21 crc kubenswrapper[4846]: E1005 07:00:21.614041 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6fc79ab2-6284-412f-b641-01e7d1f4b1c6" containerName="collect-profiles" Oct 05 07:00:21 crc kubenswrapper[4846]: I1005 07:00:21.614056 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="6fc79ab2-6284-412f-b641-01e7d1f4b1c6" containerName="collect-profiles" Oct 05 07:00:21 crc kubenswrapper[4846]: E1005 07:00:21.614070 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a3699d3c-ac3f-4e15-983c-1b6332c9accc" containerName="storage" Oct 05 07:00:21 crc kubenswrapper[4846]: I1005 07:00:21.614077 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="a3699d3c-ac3f-4e15-983c-1b6332c9accc" containerName="storage" Oct 05 07:00:21 crc kubenswrapper[4846]: I1005 07:00:21.614202 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="a3699d3c-ac3f-4e15-983c-1b6332c9accc" containerName="storage" Oct 05 07:00:21 crc kubenswrapper[4846]: I1005 07:00:21.614212 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="6fc79ab2-6284-412f-b641-01e7d1f4b1c6" containerName="collect-profiles" Oct 05 07:00:21 crc kubenswrapper[4846]: I1005 07:00:21.614934 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c65gkl" Oct 05 07:00:21 crc kubenswrapper[4846]: I1005 07:00:21.617576 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Oct 05 07:00:21 crc kubenswrapper[4846]: I1005 07:00:21.628952 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c65gkl"] Oct 05 07:00:21 crc kubenswrapper[4846]: I1005 07:00:21.775033 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/efcc5448-8dd3-4e98-8729-042643de817d-util\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c65gkl\" (UID: \"efcc5448-8dd3-4e98-8729-042643de817d\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c65gkl" Oct 05 07:00:21 crc kubenswrapper[4846]: I1005 07:00:21.775118 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2ntz7\" (UniqueName: \"kubernetes.io/projected/efcc5448-8dd3-4e98-8729-042643de817d-kube-api-access-2ntz7\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c65gkl\" (UID: \"efcc5448-8dd3-4e98-8729-042643de817d\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c65gkl" Oct 05 07:00:21 crc kubenswrapper[4846]: I1005 07:00:21.775218 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/efcc5448-8dd3-4e98-8729-042643de817d-bundle\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c65gkl\" (UID: \"efcc5448-8dd3-4e98-8729-042643de817d\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c65gkl" Oct 05 07:00:21 crc kubenswrapper[4846]: I1005 07:00:21.876673 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/efcc5448-8dd3-4e98-8729-042643de817d-util\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c65gkl\" (UID: \"efcc5448-8dd3-4e98-8729-042643de817d\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c65gkl" Oct 05 07:00:21 crc kubenswrapper[4846]: I1005 07:00:21.876829 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2ntz7\" (UniqueName: \"kubernetes.io/projected/efcc5448-8dd3-4e98-8729-042643de817d-kube-api-access-2ntz7\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c65gkl\" (UID: \"efcc5448-8dd3-4e98-8729-042643de817d\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c65gkl" Oct 05 07:00:21 crc kubenswrapper[4846]: I1005 07:00:21.876944 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/efcc5448-8dd3-4e98-8729-042643de817d-bundle\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c65gkl\" (UID: \"efcc5448-8dd3-4e98-8729-042643de817d\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c65gkl" Oct 05 07:00:21 crc kubenswrapper[4846]: I1005 07:00:21.877685 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/efcc5448-8dd3-4e98-8729-042643de817d-util\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c65gkl\" (UID: \"efcc5448-8dd3-4e98-8729-042643de817d\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c65gkl" Oct 05 07:00:21 crc kubenswrapper[4846]: I1005 07:00:21.877929 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/efcc5448-8dd3-4e98-8729-042643de817d-bundle\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c65gkl\" (UID: \"efcc5448-8dd3-4e98-8729-042643de817d\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c65gkl" Oct 05 07:00:21 crc kubenswrapper[4846]: I1005 07:00:21.913945 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2ntz7\" (UniqueName: \"kubernetes.io/projected/efcc5448-8dd3-4e98-8729-042643de817d-kube-api-access-2ntz7\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c65gkl\" (UID: \"efcc5448-8dd3-4e98-8729-042643de817d\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c65gkl" Oct 05 07:00:21 crc kubenswrapper[4846]: I1005 07:00:21.976838 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c65gkl" Oct 05 07:00:22 crc kubenswrapper[4846]: I1005 07:00:22.316007 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c65gkl"] Oct 05 07:00:22 crc kubenswrapper[4846]: W1005 07:00:22.330131 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podefcc5448_8dd3_4e98_8729_042643de817d.slice/crio-63a3ae149eb74a59ea843cbc613eb3066523cc3e79cf782bccdf49e3c70c0f96 WatchSource:0}: Error finding container 63a3ae149eb74a59ea843cbc613eb3066523cc3e79cf782bccdf49e3c70c0f96: Status 404 returned error can't find the container with id 63a3ae149eb74a59ea843cbc613eb3066523cc3e79cf782bccdf49e3c70c0f96 Oct 05 07:00:22 crc kubenswrapper[4846]: I1005 07:00:22.407493 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c65gkl" event={"ID":"efcc5448-8dd3-4e98-8729-042643de817d","Type":"ContainerStarted","Data":"63a3ae149eb74a59ea843cbc613eb3066523cc3e79cf782bccdf49e3c70c0f96"} Oct 05 07:00:23 crc kubenswrapper[4846]: I1005 07:00:23.326318 4846 patch_prober.go:28] interesting pod/machine-config-daemon-fscvf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 07:00:23 crc kubenswrapper[4846]: I1005 07:00:23.326967 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 07:00:23 crc kubenswrapper[4846]: I1005 07:00:23.416093 4846 generic.go:334] "Generic (PLEG): container finished" podID="efcc5448-8dd3-4e98-8729-042643de817d" containerID="f53b406ea68a486b4a35e4f4c797a5ab71d66bbb0dfe2e2cde731cb069cd8556" exitCode=0 Oct 05 07:00:23 crc kubenswrapper[4846]: I1005 07:00:23.416171 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c65gkl" event={"ID":"efcc5448-8dd3-4e98-8729-042643de817d","Type":"ContainerDied","Data":"f53b406ea68a486b4a35e4f4c797a5ab71d66bbb0dfe2e2cde731cb069cd8556"} Oct 05 07:00:25 crc kubenswrapper[4846]: I1005 07:00:25.431980 4846 generic.go:334] "Generic (PLEG): container finished" podID="efcc5448-8dd3-4e98-8729-042643de817d" containerID="9fd18b119a115a12bcc5bb1c13678de2ded3dfdba9e55edd16341c99f0fb848a" exitCode=0 Oct 05 07:00:25 crc kubenswrapper[4846]: I1005 07:00:25.432087 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c65gkl" event={"ID":"efcc5448-8dd3-4e98-8729-042643de817d","Type":"ContainerDied","Data":"9fd18b119a115a12bcc5bb1c13678de2ded3dfdba9e55edd16341c99f0fb848a"} Oct 05 07:00:26 crc kubenswrapper[4846]: I1005 07:00:26.452300 4846 generic.go:334] "Generic (PLEG): container finished" podID="efcc5448-8dd3-4e98-8729-042643de817d" containerID="bfe36d6bd8a4284869d2b4c98c8c2f59b5b1335c4355e8525a0fa7aea3e7e30c" exitCode=0 Oct 05 07:00:26 crc kubenswrapper[4846]: I1005 07:00:26.452363 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c65gkl" event={"ID":"efcc5448-8dd3-4e98-8729-042643de817d","Type":"ContainerDied","Data":"bfe36d6bd8a4284869d2b4c98c8c2f59b5b1335c4355e8525a0fa7aea3e7e30c"} Oct 05 07:00:26 crc kubenswrapper[4846]: I1005 07:00:26.828660 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-vjmbx"] Oct 05 07:00:26 crc kubenswrapper[4846]: I1005 07:00:26.829021 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-vjmbx" podUID="6be3f74a-b215-4588-8b0f-735eba580bbd" containerName="controller-manager" containerID="cri-o://38cfaabdd0e1e1e2fe9b14b6b4553dc224a05211452f25e0ccce1f49c8e68578" gracePeriod=30 Oct 05 07:00:26 crc kubenswrapper[4846]: I1005 07:00:26.945379 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-vwktr"] Oct 05 07:00:26 crc kubenswrapper[4846]: I1005 07:00:26.945651 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-vwktr" podUID="beec0ea6-0c57-44d6-bfef-d00e5b3724ad" containerName="route-controller-manager" containerID="cri-o://572da8558f11938b557fcd55ba9ad07a9e2c86156f81abfce221a75fa0f165ab" gracePeriod=30 Oct 05 07:00:27 crc kubenswrapper[4846]: I1005 07:00:27.458750 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-vwktr" Oct 05 07:00:27 crc kubenswrapper[4846]: I1005 07:00:27.460718 4846 generic.go:334] "Generic (PLEG): container finished" podID="beec0ea6-0c57-44d6-bfef-d00e5b3724ad" containerID="572da8558f11938b557fcd55ba9ad07a9e2c86156f81abfce221a75fa0f165ab" exitCode=0 Oct 05 07:00:27 crc kubenswrapper[4846]: I1005 07:00:27.460828 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-vwktr" event={"ID":"beec0ea6-0c57-44d6-bfef-d00e5b3724ad","Type":"ContainerDied","Data":"572da8558f11938b557fcd55ba9ad07a9e2c86156f81abfce221a75fa0f165ab"} Oct 05 07:00:27 crc kubenswrapper[4846]: I1005 07:00:27.460923 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-vwktr" event={"ID":"beec0ea6-0c57-44d6-bfef-d00e5b3724ad","Type":"ContainerDied","Data":"d48285c00b668451b292a1fae7b40af066195b9dc894206d63635b4698949e65"} Oct 05 07:00:27 crc kubenswrapper[4846]: I1005 07:00:27.460956 4846 scope.go:117] "RemoveContainer" containerID="572da8558f11938b557fcd55ba9ad07a9e2c86156f81abfce221a75fa0f165ab" Oct 05 07:00:27 crc kubenswrapper[4846]: I1005 07:00:27.463590 4846 generic.go:334] "Generic (PLEG): container finished" podID="6be3f74a-b215-4588-8b0f-735eba580bbd" containerID="38cfaabdd0e1e1e2fe9b14b6b4553dc224a05211452f25e0ccce1f49c8e68578" exitCode=0 Oct 05 07:00:27 crc kubenswrapper[4846]: I1005 07:00:27.463699 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-vjmbx" event={"ID":"6be3f74a-b215-4588-8b0f-735eba580bbd","Type":"ContainerDied","Data":"38cfaabdd0e1e1e2fe9b14b6b4553dc224a05211452f25e0ccce1f49c8e68578"} Oct 05 07:00:27 crc kubenswrapper[4846]: I1005 07:00:27.467150 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/beec0ea6-0c57-44d6-bfef-d00e5b3724ad-config\") pod \"beec0ea6-0c57-44d6-bfef-d00e5b3724ad\" (UID: \"beec0ea6-0c57-44d6-bfef-d00e5b3724ad\") " Oct 05 07:00:27 crc kubenswrapper[4846]: I1005 07:00:27.467249 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pmhpw\" (UniqueName: \"kubernetes.io/projected/beec0ea6-0c57-44d6-bfef-d00e5b3724ad-kube-api-access-pmhpw\") pod \"beec0ea6-0c57-44d6-bfef-d00e5b3724ad\" (UID: \"beec0ea6-0c57-44d6-bfef-d00e5b3724ad\") " Oct 05 07:00:27 crc kubenswrapper[4846]: I1005 07:00:27.467369 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/beec0ea6-0c57-44d6-bfef-d00e5b3724ad-client-ca\") pod \"beec0ea6-0c57-44d6-bfef-d00e5b3724ad\" (UID: \"beec0ea6-0c57-44d6-bfef-d00e5b3724ad\") " Oct 05 07:00:27 crc kubenswrapper[4846]: I1005 07:00:27.467393 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/beec0ea6-0c57-44d6-bfef-d00e5b3724ad-serving-cert\") pod \"beec0ea6-0c57-44d6-bfef-d00e5b3724ad\" (UID: \"beec0ea6-0c57-44d6-bfef-d00e5b3724ad\") " Oct 05 07:00:27 crc kubenswrapper[4846]: I1005 07:00:27.470620 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/beec0ea6-0c57-44d6-bfef-d00e5b3724ad-client-ca" (OuterVolumeSpecName: "client-ca") pod "beec0ea6-0c57-44d6-bfef-d00e5b3724ad" (UID: "beec0ea6-0c57-44d6-bfef-d00e5b3724ad"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 07:00:27 crc kubenswrapper[4846]: I1005 07:00:27.496318 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/beec0ea6-0c57-44d6-bfef-d00e5b3724ad-config" (OuterVolumeSpecName: "config") pod "beec0ea6-0c57-44d6-bfef-d00e5b3724ad" (UID: "beec0ea6-0c57-44d6-bfef-d00e5b3724ad"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 07:00:27 crc kubenswrapper[4846]: I1005 07:00:27.497737 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/beec0ea6-0c57-44d6-bfef-d00e5b3724ad-kube-api-access-pmhpw" (OuterVolumeSpecName: "kube-api-access-pmhpw") pod "beec0ea6-0c57-44d6-bfef-d00e5b3724ad" (UID: "beec0ea6-0c57-44d6-bfef-d00e5b3724ad"). InnerVolumeSpecName "kube-api-access-pmhpw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:00:27 crc kubenswrapper[4846]: I1005 07:00:27.508353 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/beec0ea6-0c57-44d6-bfef-d00e5b3724ad-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "beec0ea6-0c57-44d6-bfef-d00e5b3724ad" (UID: "beec0ea6-0c57-44d6-bfef-d00e5b3724ad"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:00:27 crc kubenswrapper[4846]: I1005 07:00:27.542094 4846 scope.go:117] "RemoveContainer" containerID="572da8558f11938b557fcd55ba9ad07a9e2c86156f81abfce221a75fa0f165ab" Oct 05 07:00:27 crc kubenswrapper[4846]: E1005 07:00:27.544757 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"572da8558f11938b557fcd55ba9ad07a9e2c86156f81abfce221a75fa0f165ab\": container with ID starting with 572da8558f11938b557fcd55ba9ad07a9e2c86156f81abfce221a75fa0f165ab not found: ID does not exist" containerID="572da8558f11938b557fcd55ba9ad07a9e2c86156f81abfce221a75fa0f165ab" Oct 05 07:00:27 crc kubenswrapper[4846]: I1005 07:00:27.544800 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"572da8558f11938b557fcd55ba9ad07a9e2c86156f81abfce221a75fa0f165ab"} err="failed to get container status \"572da8558f11938b557fcd55ba9ad07a9e2c86156f81abfce221a75fa0f165ab\": rpc error: code = NotFound desc = could not find container \"572da8558f11938b557fcd55ba9ad07a9e2c86156f81abfce221a75fa0f165ab\": container with ID starting with 572da8558f11938b557fcd55ba9ad07a9e2c86156f81abfce221a75fa0f165ab not found: ID does not exist" Oct 05 07:00:27 crc kubenswrapper[4846]: I1005 07:00:27.568949 4846 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/beec0ea6-0c57-44d6-bfef-d00e5b3724ad-config\") on node \"crc\" DevicePath \"\"" Oct 05 07:00:27 crc kubenswrapper[4846]: I1005 07:00:27.568989 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pmhpw\" (UniqueName: \"kubernetes.io/projected/beec0ea6-0c57-44d6-bfef-d00e5b3724ad-kube-api-access-pmhpw\") on node \"crc\" DevicePath \"\"" Oct 05 07:00:27 crc kubenswrapper[4846]: I1005 07:00:27.569001 4846 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/beec0ea6-0c57-44d6-bfef-d00e5b3724ad-client-ca\") on node \"crc\" DevicePath \"\"" Oct 05 07:00:27 crc kubenswrapper[4846]: I1005 07:00:27.569011 4846 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/beec0ea6-0c57-44d6-bfef-d00e5b3724ad-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 05 07:00:27 crc kubenswrapper[4846]: I1005 07:00:27.710201 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c65gkl" Oct 05 07:00:27 crc kubenswrapper[4846]: I1005 07:00:27.739824 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-vjmbx" Oct 05 07:00:27 crc kubenswrapper[4846]: I1005 07:00:27.773083 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/efcc5448-8dd3-4e98-8729-042643de817d-util\") pod \"efcc5448-8dd3-4e98-8729-042643de817d\" (UID: \"efcc5448-8dd3-4e98-8729-042643de817d\") " Oct 05 07:00:27 crc kubenswrapper[4846]: I1005 07:00:27.773133 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gbd7g\" (UniqueName: \"kubernetes.io/projected/6be3f74a-b215-4588-8b0f-735eba580bbd-kube-api-access-gbd7g\") pod \"6be3f74a-b215-4588-8b0f-735eba580bbd\" (UID: \"6be3f74a-b215-4588-8b0f-735eba580bbd\") " Oct 05 07:00:27 crc kubenswrapper[4846]: I1005 07:00:27.773168 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6be3f74a-b215-4588-8b0f-735eba580bbd-config\") pod \"6be3f74a-b215-4588-8b0f-735eba580bbd\" (UID: \"6be3f74a-b215-4588-8b0f-735eba580bbd\") " Oct 05 07:00:27 crc kubenswrapper[4846]: I1005 07:00:27.773281 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/6be3f74a-b215-4588-8b0f-735eba580bbd-proxy-ca-bundles\") pod \"6be3f74a-b215-4588-8b0f-735eba580bbd\" (UID: \"6be3f74a-b215-4588-8b0f-735eba580bbd\") " Oct 05 07:00:27 crc kubenswrapper[4846]: I1005 07:00:27.773311 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6be3f74a-b215-4588-8b0f-735eba580bbd-serving-cert\") pod \"6be3f74a-b215-4588-8b0f-735eba580bbd\" (UID: \"6be3f74a-b215-4588-8b0f-735eba580bbd\") " Oct 05 07:00:27 crc kubenswrapper[4846]: I1005 07:00:27.773362 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/efcc5448-8dd3-4e98-8729-042643de817d-bundle\") pod \"efcc5448-8dd3-4e98-8729-042643de817d\" (UID: \"efcc5448-8dd3-4e98-8729-042643de817d\") " Oct 05 07:00:27 crc kubenswrapper[4846]: I1005 07:00:27.773382 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2ntz7\" (UniqueName: \"kubernetes.io/projected/efcc5448-8dd3-4e98-8729-042643de817d-kube-api-access-2ntz7\") pod \"efcc5448-8dd3-4e98-8729-042643de817d\" (UID: \"efcc5448-8dd3-4e98-8729-042643de817d\") " Oct 05 07:00:27 crc kubenswrapper[4846]: I1005 07:00:27.773431 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/6be3f74a-b215-4588-8b0f-735eba580bbd-client-ca\") pod \"6be3f74a-b215-4588-8b0f-735eba580bbd\" (UID: \"6be3f74a-b215-4588-8b0f-735eba580bbd\") " Oct 05 07:00:27 crc kubenswrapper[4846]: I1005 07:00:27.774613 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6be3f74a-b215-4588-8b0f-735eba580bbd-client-ca" (OuterVolumeSpecName: "client-ca") pod "6be3f74a-b215-4588-8b0f-735eba580bbd" (UID: "6be3f74a-b215-4588-8b0f-735eba580bbd"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 07:00:27 crc kubenswrapper[4846]: I1005 07:00:27.775203 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/efcc5448-8dd3-4e98-8729-042643de817d-bundle" (OuterVolumeSpecName: "bundle") pod "efcc5448-8dd3-4e98-8729-042643de817d" (UID: "efcc5448-8dd3-4e98-8729-042643de817d"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 07:00:27 crc kubenswrapper[4846]: I1005 07:00:27.775656 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6be3f74a-b215-4588-8b0f-735eba580bbd-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "6be3f74a-b215-4588-8b0f-735eba580bbd" (UID: "6be3f74a-b215-4588-8b0f-735eba580bbd"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 07:00:27 crc kubenswrapper[4846]: I1005 07:00:27.776490 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6be3f74a-b215-4588-8b0f-735eba580bbd-config" (OuterVolumeSpecName: "config") pod "6be3f74a-b215-4588-8b0f-735eba580bbd" (UID: "6be3f74a-b215-4588-8b0f-735eba580bbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 07:00:27 crc kubenswrapper[4846]: I1005 07:00:27.780285 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efcc5448-8dd3-4e98-8729-042643de817d-kube-api-access-2ntz7" (OuterVolumeSpecName: "kube-api-access-2ntz7") pod "efcc5448-8dd3-4e98-8729-042643de817d" (UID: "efcc5448-8dd3-4e98-8729-042643de817d"). InnerVolumeSpecName "kube-api-access-2ntz7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:00:27 crc kubenswrapper[4846]: I1005 07:00:27.780701 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6be3f74a-b215-4588-8b0f-735eba580bbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6be3f74a-b215-4588-8b0f-735eba580bbd" (UID: "6be3f74a-b215-4588-8b0f-735eba580bbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:00:27 crc kubenswrapper[4846]: I1005 07:00:27.780805 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6be3f74a-b215-4588-8b0f-735eba580bbd-kube-api-access-gbd7g" (OuterVolumeSpecName: "kube-api-access-gbd7g") pod "6be3f74a-b215-4588-8b0f-735eba580bbd" (UID: "6be3f74a-b215-4588-8b0f-735eba580bbd"). InnerVolumeSpecName "kube-api-access-gbd7g". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:00:27 crc kubenswrapper[4846]: I1005 07:00:27.805650 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/efcc5448-8dd3-4e98-8729-042643de817d-util" (OuterVolumeSpecName: "util") pod "efcc5448-8dd3-4e98-8729-042643de817d" (UID: "efcc5448-8dd3-4e98-8729-042643de817d"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 07:00:27 crc kubenswrapper[4846]: I1005 07:00:27.874591 4846 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/6be3f74a-b215-4588-8b0f-735eba580bbd-client-ca\") on node \"crc\" DevicePath \"\"" Oct 05 07:00:27 crc kubenswrapper[4846]: I1005 07:00:27.874650 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gbd7g\" (UniqueName: \"kubernetes.io/projected/6be3f74a-b215-4588-8b0f-735eba580bbd-kube-api-access-gbd7g\") on node \"crc\" DevicePath \"\"" Oct 05 07:00:27 crc kubenswrapper[4846]: I1005 07:00:27.874670 4846 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/efcc5448-8dd3-4e98-8729-042643de817d-util\") on node \"crc\" DevicePath \"\"" Oct 05 07:00:27 crc kubenswrapper[4846]: I1005 07:00:27.874688 4846 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6be3f74a-b215-4588-8b0f-735eba580bbd-config\") on node \"crc\" DevicePath \"\"" Oct 05 07:00:27 crc kubenswrapper[4846]: I1005 07:00:27.874709 4846 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/6be3f74a-b215-4588-8b0f-735eba580bbd-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Oct 05 07:00:27 crc kubenswrapper[4846]: I1005 07:00:27.874728 4846 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6be3f74a-b215-4588-8b0f-735eba580bbd-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 05 07:00:27 crc kubenswrapper[4846]: I1005 07:00:27.874745 4846 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/efcc5448-8dd3-4e98-8729-042643de817d-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 07:00:27 crc kubenswrapper[4846]: I1005 07:00:27.874763 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2ntz7\" (UniqueName: \"kubernetes.io/projected/efcc5448-8dd3-4e98-8729-042643de817d-kube-api-access-2ntz7\") on node \"crc\" DevicePath \"\"" Oct 05 07:00:28 crc kubenswrapper[4846]: I1005 07:00:28.396351 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-84d6646cc4-rv76g"] Oct 05 07:00:28 crc kubenswrapper[4846]: E1005 07:00:28.397439 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="efcc5448-8dd3-4e98-8729-042643de817d" containerName="util" Oct 05 07:00:28 crc kubenswrapper[4846]: I1005 07:00:28.397480 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="efcc5448-8dd3-4e98-8729-042643de817d" containerName="util" Oct 05 07:00:28 crc kubenswrapper[4846]: E1005 07:00:28.397511 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="efcc5448-8dd3-4e98-8729-042643de817d" containerName="extract" Oct 05 07:00:28 crc kubenswrapper[4846]: I1005 07:00:28.397529 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="efcc5448-8dd3-4e98-8729-042643de817d" containerName="extract" Oct 05 07:00:28 crc kubenswrapper[4846]: E1005 07:00:28.397565 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="beec0ea6-0c57-44d6-bfef-d00e5b3724ad" containerName="route-controller-manager" Oct 05 07:00:28 crc kubenswrapper[4846]: I1005 07:00:28.397585 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="beec0ea6-0c57-44d6-bfef-d00e5b3724ad" containerName="route-controller-manager" Oct 05 07:00:28 crc kubenswrapper[4846]: E1005 07:00:28.397621 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6be3f74a-b215-4588-8b0f-735eba580bbd" containerName="controller-manager" Oct 05 07:00:28 crc kubenswrapper[4846]: I1005 07:00:28.397639 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="6be3f74a-b215-4588-8b0f-735eba580bbd" containerName="controller-manager" Oct 05 07:00:28 crc kubenswrapper[4846]: E1005 07:00:28.397661 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="efcc5448-8dd3-4e98-8729-042643de817d" containerName="pull" Oct 05 07:00:28 crc kubenswrapper[4846]: I1005 07:00:28.397676 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="efcc5448-8dd3-4e98-8729-042643de817d" containerName="pull" Oct 05 07:00:28 crc kubenswrapper[4846]: I1005 07:00:28.397913 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="efcc5448-8dd3-4e98-8729-042643de817d" containerName="extract" Oct 05 07:00:28 crc kubenswrapper[4846]: I1005 07:00:28.397961 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="beec0ea6-0c57-44d6-bfef-d00e5b3724ad" containerName="route-controller-manager" Oct 05 07:00:28 crc kubenswrapper[4846]: I1005 07:00:28.397989 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="6be3f74a-b215-4588-8b0f-735eba580bbd" containerName="controller-manager" Oct 05 07:00:28 crc kubenswrapper[4846]: I1005 07:00:28.398958 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-84d6646cc4-rv76g" Oct 05 07:00:28 crc kubenswrapper[4846]: I1005 07:00:28.404524 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-7fb97886bb-n65bb"] Oct 05 07:00:28 crc kubenswrapper[4846]: I1005 07:00:28.405751 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7fb97886bb-n65bb" Oct 05 07:00:28 crc kubenswrapper[4846]: I1005 07:00:28.422051 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-84d6646cc4-rv76g"] Oct 05 07:00:28 crc kubenswrapper[4846]: I1005 07:00:28.429852 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-7fb97886bb-n65bb"] Oct 05 07:00:28 crc kubenswrapper[4846]: I1005 07:00:28.478533 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c65gkl" event={"ID":"efcc5448-8dd3-4e98-8729-042643de817d","Type":"ContainerDied","Data":"63a3ae149eb74a59ea843cbc613eb3066523cc3e79cf782bccdf49e3c70c0f96"} Oct 05 07:00:28 crc kubenswrapper[4846]: I1005 07:00:28.478596 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="63a3ae149eb74a59ea843cbc613eb3066523cc3e79cf782bccdf49e3c70c0f96" Oct 05 07:00:28 crc kubenswrapper[4846]: I1005 07:00:28.480311 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-vwktr" Oct 05 07:00:28 crc kubenswrapper[4846]: I1005 07:00:28.478568 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c65gkl" Oct 05 07:00:28 crc kubenswrapper[4846]: I1005 07:00:28.484635 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-vjmbx" event={"ID":"6be3f74a-b215-4588-8b0f-735eba580bbd","Type":"ContainerDied","Data":"87a40d3be84539a20530df2487213814433c61ad32b51f566c8a99041eb5102f"} Oct 05 07:00:28 crc kubenswrapper[4846]: I1005 07:00:28.484723 4846 scope.go:117] "RemoveContainer" containerID="38cfaabdd0e1e1e2fe9b14b6b4553dc224a05211452f25e0ccce1f49c8e68578" Oct 05 07:00:28 crc kubenswrapper[4846]: I1005 07:00:28.484964 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-vjmbx" Oct 05 07:00:28 crc kubenswrapper[4846]: I1005 07:00:28.485575 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/f9dd0824-e0c1-446f-9786-8b57557415da-proxy-ca-bundles\") pod \"controller-manager-7fb97886bb-n65bb\" (UID: \"f9dd0824-e0c1-446f-9786-8b57557415da\") " pod="openshift-controller-manager/controller-manager-7fb97886bb-n65bb" Oct 05 07:00:28 crc kubenswrapper[4846]: I1005 07:00:28.492001 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f9dd0824-e0c1-446f-9786-8b57557415da-config\") pod \"controller-manager-7fb97886bb-n65bb\" (UID: \"f9dd0824-e0c1-446f-9786-8b57557415da\") " pod="openshift-controller-manager/controller-manager-7fb97886bb-n65bb" Oct 05 07:00:28 crc kubenswrapper[4846]: I1005 07:00:28.492030 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f9dd0824-e0c1-446f-9786-8b57557415da-client-ca\") pod \"controller-manager-7fb97886bb-n65bb\" (UID: \"f9dd0824-e0c1-446f-9786-8b57557415da\") " pod="openshift-controller-manager/controller-manager-7fb97886bb-n65bb" Oct 05 07:00:28 crc kubenswrapper[4846]: I1005 07:00:28.492054 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t7bp4\" (UniqueName: \"kubernetes.io/projected/f9dd0824-e0c1-446f-9786-8b57557415da-kube-api-access-t7bp4\") pod \"controller-manager-7fb97886bb-n65bb\" (UID: \"f9dd0824-e0c1-446f-9786-8b57557415da\") " pod="openshift-controller-manager/controller-manager-7fb97886bb-n65bb" Oct 05 07:00:28 crc kubenswrapper[4846]: I1005 07:00:28.492078 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f9dd0824-e0c1-446f-9786-8b57557415da-serving-cert\") pod \"controller-manager-7fb97886bb-n65bb\" (UID: \"f9dd0824-e0c1-446f-9786-8b57557415da\") " pod="openshift-controller-manager/controller-manager-7fb97886bb-n65bb" Oct 05 07:00:28 crc kubenswrapper[4846]: I1005 07:00:28.492139 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5bb66058-e123-46bb-aede-a02814c3ab5e-config\") pod \"route-controller-manager-84d6646cc4-rv76g\" (UID: \"5bb66058-e123-46bb-aede-a02814c3ab5e\") " pod="openshift-route-controller-manager/route-controller-manager-84d6646cc4-rv76g" Oct 05 07:00:28 crc kubenswrapper[4846]: I1005 07:00:28.492216 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5bb66058-e123-46bb-aede-a02814c3ab5e-client-ca\") pod \"route-controller-manager-84d6646cc4-rv76g\" (UID: \"5bb66058-e123-46bb-aede-a02814c3ab5e\") " pod="openshift-route-controller-manager/route-controller-manager-84d6646cc4-rv76g" Oct 05 07:00:28 crc kubenswrapper[4846]: I1005 07:00:28.492245 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n8jwz\" (UniqueName: \"kubernetes.io/projected/5bb66058-e123-46bb-aede-a02814c3ab5e-kube-api-access-n8jwz\") pod \"route-controller-manager-84d6646cc4-rv76g\" (UID: \"5bb66058-e123-46bb-aede-a02814c3ab5e\") " pod="openshift-route-controller-manager/route-controller-manager-84d6646cc4-rv76g" Oct 05 07:00:28 crc kubenswrapper[4846]: I1005 07:00:28.492282 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5bb66058-e123-46bb-aede-a02814c3ab5e-serving-cert\") pod \"route-controller-manager-84d6646cc4-rv76g\" (UID: \"5bb66058-e123-46bb-aede-a02814c3ab5e\") " pod="openshift-route-controller-manager/route-controller-manager-84d6646cc4-rv76g" Oct 05 07:00:28 crc kubenswrapper[4846]: I1005 07:00:28.539729 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-vwktr"] Oct 05 07:00:28 crc kubenswrapper[4846]: I1005 07:00:28.550901 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-vwktr"] Oct 05 07:00:28 crc kubenswrapper[4846]: I1005 07:00:28.557159 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-vjmbx"] Oct 05 07:00:28 crc kubenswrapper[4846]: I1005 07:00:28.560849 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-vjmbx"] Oct 05 07:00:28 crc kubenswrapper[4846]: I1005 07:00:28.592774 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/f9dd0824-e0c1-446f-9786-8b57557415da-proxy-ca-bundles\") pod \"controller-manager-7fb97886bb-n65bb\" (UID: \"f9dd0824-e0c1-446f-9786-8b57557415da\") " pod="openshift-controller-manager/controller-manager-7fb97886bb-n65bb" Oct 05 07:00:28 crc kubenswrapper[4846]: I1005 07:00:28.592870 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f9dd0824-e0c1-446f-9786-8b57557415da-config\") pod \"controller-manager-7fb97886bb-n65bb\" (UID: \"f9dd0824-e0c1-446f-9786-8b57557415da\") " pod="openshift-controller-manager/controller-manager-7fb97886bb-n65bb" Oct 05 07:00:28 crc kubenswrapper[4846]: I1005 07:00:28.592897 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f9dd0824-e0c1-446f-9786-8b57557415da-client-ca\") pod \"controller-manager-7fb97886bb-n65bb\" (UID: \"f9dd0824-e0c1-446f-9786-8b57557415da\") " pod="openshift-controller-manager/controller-manager-7fb97886bb-n65bb" Oct 05 07:00:28 crc kubenswrapper[4846]: I1005 07:00:28.592922 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t7bp4\" (UniqueName: \"kubernetes.io/projected/f9dd0824-e0c1-446f-9786-8b57557415da-kube-api-access-t7bp4\") pod \"controller-manager-7fb97886bb-n65bb\" (UID: \"f9dd0824-e0c1-446f-9786-8b57557415da\") " pod="openshift-controller-manager/controller-manager-7fb97886bb-n65bb" Oct 05 07:00:28 crc kubenswrapper[4846]: I1005 07:00:28.592949 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f9dd0824-e0c1-446f-9786-8b57557415da-serving-cert\") pod \"controller-manager-7fb97886bb-n65bb\" (UID: \"f9dd0824-e0c1-446f-9786-8b57557415da\") " pod="openshift-controller-manager/controller-manager-7fb97886bb-n65bb" Oct 05 07:00:28 crc kubenswrapper[4846]: I1005 07:00:28.593007 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5bb66058-e123-46bb-aede-a02814c3ab5e-config\") pod \"route-controller-manager-84d6646cc4-rv76g\" (UID: \"5bb66058-e123-46bb-aede-a02814c3ab5e\") " pod="openshift-route-controller-manager/route-controller-manager-84d6646cc4-rv76g" Oct 05 07:00:28 crc kubenswrapper[4846]: I1005 07:00:28.593064 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5bb66058-e123-46bb-aede-a02814c3ab5e-client-ca\") pod \"route-controller-manager-84d6646cc4-rv76g\" (UID: \"5bb66058-e123-46bb-aede-a02814c3ab5e\") " pod="openshift-route-controller-manager/route-controller-manager-84d6646cc4-rv76g" Oct 05 07:00:28 crc kubenswrapper[4846]: I1005 07:00:28.593093 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n8jwz\" (UniqueName: \"kubernetes.io/projected/5bb66058-e123-46bb-aede-a02814c3ab5e-kube-api-access-n8jwz\") pod \"route-controller-manager-84d6646cc4-rv76g\" (UID: \"5bb66058-e123-46bb-aede-a02814c3ab5e\") " pod="openshift-route-controller-manager/route-controller-manager-84d6646cc4-rv76g" Oct 05 07:00:28 crc kubenswrapper[4846]: I1005 07:00:28.593230 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5bb66058-e123-46bb-aede-a02814c3ab5e-serving-cert\") pod \"route-controller-manager-84d6646cc4-rv76g\" (UID: \"5bb66058-e123-46bb-aede-a02814c3ab5e\") " pod="openshift-route-controller-manager/route-controller-manager-84d6646cc4-rv76g" Oct 05 07:00:28 crc kubenswrapper[4846]: I1005 07:00:28.594646 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5bb66058-e123-46bb-aede-a02814c3ab5e-config\") pod \"route-controller-manager-84d6646cc4-rv76g\" (UID: \"5bb66058-e123-46bb-aede-a02814c3ab5e\") " pod="openshift-route-controller-manager/route-controller-manager-84d6646cc4-rv76g" Oct 05 07:00:28 crc kubenswrapper[4846]: I1005 07:00:28.594671 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5bb66058-e123-46bb-aede-a02814c3ab5e-client-ca\") pod \"route-controller-manager-84d6646cc4-rv76g\" (UID: \"5bb66058-e123-46bb-aede-a02814c3ab5e\") " pod="openshift-route-controller-manager/route-controller-manager-84d6646cc4-rv76g" Oct 05 07:00:28 crc kubenswrapper[4846]: I1005 07:00:28.595287 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/f9dd0824-e0c1-446f-9786-8b57557415da-proxy-ca-bundles\") pod \"controller-manager-7fb97886bb-n65bb\" (UID: \"f9dd0824-e0c1-446f-9786-8b57557415da\") " pod="openshift-controller-manager/controller-manager-7fb97886bb-n65bb" Oct 05 07:00:28 crc kubenswrapper[4846]: I1005 07:00:28.595998 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f9dd0824-e0c1-446f-9786-8b57557415da-client-ca\") pod \"controller-manager-7fb97886bb-n65bb\" (UID: \"f9dd0824-e0c1-446f-9786-8b57557415da\") " pod="openshift-controller-manager/controller-manager-7fb97886bb-n65bb" Oct 05 07:00:28 crc kubenswrapper[4846]: I1005 07:00:28.596539 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f9dd0824-e0c1-446f-9786-8b57557415da-config\") pod \"controller-manager-7fb97886bb-n65bb\" (UID: \"f9dd0824-e0c1-446f-9786-8b57557415da\") " pod="openshift-controller-manager/controller-manager-7fb97886bb-n65bb" Oct 05 07:00:28 crc kubenswrapper[4846]: I1005 07:00:28.599963 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5bb66058-e123-46bb-aede-a02814c3ab5e-serving-cert\") pod \"route-controller-manager-84d6646cc4-rv76g\" (UID: \"5bb66058-e123-46bb-aede-a02814c3ab5e\") " pod="openshift-route-controller-manager/route-controller-manager-84d6646cc4-rv76g" Oct 05 07:00:28 crc kubenswrapper[4846]: I1005 07:00:28.611894 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f9dd0824-e0c1-446f-9786-8b57557415da-serving-cert\") pod \"controller-manager-7fb97886bb-n65bb\" (UID: \"f9dd0824-e0c1-446f-9786-8b57557415da\") " pod="openshift-controller-manager/controller-manager-7fb97886bb-n65bb" Oct 05 07:00:28 crc kubenswrapper[4846]: I1005 07:00:28.612603 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n8jwz\" (UniqueName: \"kubernetes.io/projected/5bb66058-e123-46bb-aede-a02814c3ab5e-kube-api-access-n8jwz\") pod \"route-controller-manager-84d6646cc4-rv76g\" (UID: \"5bb66058-e123-46bb-aede-a02814c3ab5e\") " pod="openshift-route-controller-manager/route-controller-manager-84d6646cc4-rv76g" Oct 05 07:00:28 crc kubenswrapper[4846]: I1005 07:00:28.617575 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t7bp4\" (UniqueName: \"kubernetes.io/projected/f9dd0824-e0c1-446f-9786-8b57557415da-kube-api-access-t7bp4\") pod \"controller-manager-7fb97886bb-n65bb\" (UID: \"f9dd0824-e0c1-446f-9786-8b57557415da\") " pod="openshift-controller-manager/controller-manager-7fb97886bb-n65bb" Oct 05 07:00:28 crc kubenswrapper[4846]: I1005 07:00:28.737334 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-84d6646cc4-rv76g" Oct 05 07:00:28 crc kubenswrapper[4846]: I1005 07:00:28.752686 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7fb97886bb-n65bb" Oct 05 07:00:29 crc kubenswrapper[4846]: I1005 07:00:29.244816 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-84d6646cc4-rv76g"] Oct 05 07:00:29 crc kubenswrapper[4846]: I1005 07:00:29.247325 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-7fb97886bb-n65bb"] Oct 05 07:00:29 crc kubenswrapper[4846]: W1005 07:00:29.264327 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5bb66058_e123_46bb_aede_a02814c3ab5e.slice/crio-1a61cefcd1f699de2366d4d7db55e669e3bc20402654bd3001eca6b264393580 WatchSource:0}: Error finding container 1a61cefcd1f699de2366d4d7db55e669e3bc20402654bd3001eca6b264393580: Status 404 returned error can't find the container with id 1a61cefcd1f699de2366d4d7db55e669e3bc20402654bd3001eca6b264393580 Oct 05 07:00:29 crc kubenswrapper[4846]: W1005 07:00:29.264976 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf9dd0824_e0c1_446f_9786_8b57557415da.slice/crio-d939700ff97260bc71b6418790d3e8942811e0f1c698b9154a7dc40217f96a45 WatchSource:0}: Error finding container d939700ff97260bc71b6418790d3e8942811e0f1c698b9154a7dc40217f96a45: Status 404 returned error can't find the container with id d939700ff97260bc71b6418790d3e8942811e0f1c698b9154a7dc40217f96a45 Oct 05 07:00:29 crc kubenswrapper[4846]: I1005 07:00:29.494248 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-84d6646cc4-rv76g" event={"ID":"5bb66058-e123-46bb-aede-a02814c3ab5e","Type":"ContainerStarted","Data":"a191f97455160e34bc5a68e0ae98e179478e422833166e15bf2e12522b567d9c"} Oct 05 07:00:29 crc kubenswrapper[4846]: I1005 07:00:29.494666 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-84d6646cc4-rv76g" event={"ID":"5bb66058-e123-46bb-aede-a02814c3ab5e","Type":"ContainerStarted","Data":"1a61cefcd1f699de2366d4d7db55e669e3bc20402654bd3001eca6b264393580"} Oct 05 07:00:29 crc kubenswrapper[4846]: I1005 07:00:29.501565 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-84d6646cc4-rv76g" Oct 05 07:00:29 crc kubenswrapper[4846]: I1005 07:00:29.505220 4846 patch_prober.go:28] interesting pod/route-controller-manager-84d6646cc4-rv76g container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.38:8443/healthz\": dial tcp 10.217.0.38:8443: connect: connection refused" start-of-body= Oct 05 07:00:29 crc kubenswrapper[4846]: I1005 07:00:29.505353 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7fb97886bb-n65bb" event={"ID":"f9dd0824-e0c1-446f-9786-8b57557415da","Type":"ContainerStarted","Data":"cbf30ab141849da7eb8c183bc1157290dafb528f7f886e3130904a74715bd6a4"} Oct 05 07:00:29 crc kubenswrapper[4846]: I1005 07:00:29.505439 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7fb97886bb-n65bb" event={"ID":"f9dd0824-e0c1-446f-9786-8b57557415da","Type":"ContainerStarted","Data":"d939700ff97260bc71b6418790d3e8942811e0f1c698b9154a7dc40217f96a45"} Oct 05 07:00:29 crc kubenswrapper[4846]: I1005 07:00:29.505380 4846 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-84d6646cc4-rv76g" podUID="5bb66058-e123-46bb-aede-a02814c3ab5e" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.38:8443/healthz\": dial tcp 10.217.0.38:8443: connect: connection refused" Oct 05 07:00:29 crc kubenswrapper[4846]: I1005 07:00:29.506490 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-7fb97886bb-n65bb" Oct 05 07:00:29 crc kubenswrapper[4846]: I1005 07:00:29.507796 4846 patch_prober.go:28] interesting pod/controller-manager-7fb97886bb-n65bb container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.44:8443/healthz\": dial tcp 10.217.0.44:8443: connect: connection refused" start-of-body= Oct 05 07:00:29 crc kubenswrapper[4846]: I1005 07:00:29.507892 4846 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-7fb97886bb-n65bb" podUID="f9dd0824-e0c1-446f-9786-8b57557415da" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.44:8443/healthz\": dial tcp 10.217.0.44:8443: connect: connection refused" Oct 05 07:00:29 crc kubenswrapper[4846]: I1005 07:00:29.544172 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-84d6646cc4-rv76g" podStartSLOduration=2.544130298 podStartE2EDuration="2.544130298s" podCreationTimestamp="2025-10-05 07:00:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 07:00:29.543839391 +0000 UTC m=+751.784692196" watchObservedRunningTime="2025-10-05 07:00:29.544130298 +0000 UTC m=+751.784983123" Oct 05 07:00:29 crc kubenswrapper[4846]: I1005 07:00:29.565490 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-7fb97886bb-n65bb" podStartSLOduration=2.56546554 podStartE2EDuration="2.56546554s" podCreationTimestamp="2025-10-05 07:00:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 07:00:29.564762702 +0000 UTC m=+751.805615497" watchObservedRunningTime="2025-10-05 07:00:29.56546554 +0000 UTC m=+751.806318325" Oct 05 07:00:30 crc kubenswrapper[4846]: I1005 07:00:30.506737 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6be3f74a-b215-4588-8b0f-735eba580bbd" path="/var/lib/kubelet/pods/6be3f74a-b215-4588-8b0f-735eba580bbd/volumes" Oct 05 07:00:30 crc kubenswrapper[4846]: I1005 07:00:30.507922 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="beec0ea6-0c57-44d6-bfef-d00e5b3724ad" path="/var/lib/kubelet/pods/beec0ea6-0c57-44d6-bfef-d00e5b3724ad/volumes" Oct 05 07:00:30 crc kubenswrapper[4846]: I1005 07:00:30.516418 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-7fb97886bb-n65bb" Oct 05 07:00:30 crc kubenswrapper[4846]: I1005 07:00:30.518605 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-84d6646cc4-rv76g" Oct 05 07:00:30 crc kubenswrapper[4846]: I1005 07:00:30.912772 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-operator-858ddd8f98-lwsfd"] Oct 05 07:00:30 crc kubenswrapper[4846]: I1005 07:00:30.913885 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-858ddd8f98-lwsfd" Oct 05 07:00:30 crc kubenswrapper[4846]: I1005 07:00:30.915977 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"openshift-service-ca.crt" Oct 05 07:00:30 crc kubenswrapper[4846]: I1005 07:00:30.916693 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-operator-dockercfg-2bk79" Oct 05 07:00:30 crc kubenswrapper[4846]: I1005 07:00:30.916750 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"kube-root-ca.crt" Oct 05 07:00:30 crc kubenswrapper[4846]: I1005 07:00:30.934018 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-858ddd8f98-lwsfd"] Oct 05 07:00:30 crc kubenswrapper[4846]: I1005 07:00:30.940534 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8q9lq\" (UniqueName: \"kubernetes.io/projected/46fb70d7-c48a-4d57-990f-5e583f96fa16-kube-api-access-8q9lq\") pod \"nmstate-operator-858ddd8f98-lwsfd\" (UID: \"46fb70d7-c48a-4d57-990f-5e583f96fa16\") " pod="openshift-nmstate/nmstate-operator-858ddd8f98-lwsfd" Oct 05 07:00:31 crc kubenswrapper[4846]: I1005 07:00:31.042027 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8q9lq\" (UniqueName: \"kubernetes.io/projected/46fb70d7-c48a-4d57-990f-5e583f96fa16-kube-api-access-8q9lq\") pod \"nmstate-operator-858ddd8f98-lwsfd\" (UID: \"46fb70d7-c48a-4d57-990f-5e583f96fa16\") " pod="openshift-nmstate/nmstate-operator-858ddd8f98-lwsfd" Oct 05 07:00:31 crc kubenswrapper[4846]: I1005 07:00:31.069701 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8q9lq\" (UniqueName: \"kubernetes.io/projected/46fb70d7-c48a-4d57-990f-5e583f96fa16-kube-api-access-8q9lq\") pod \"nmstate-operator-858ddd8f98-lwsfd\" (UID: \"46fb70d7-c48a-4d57-990f-5e583f96fa16\") " pod="openshift-nmstate/nmstate-operator-858ddd8f98-lwsfd" Oct 05 07:00:31 crc kubenswrapper[4846]: I1005 07:00:31.233443 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-858ddd8f98-lwsfd" Oct 05 07:00:31 crc kubenswrapper[4846]: I1005 07:00:31.501921 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-858ddd8f98-lwsfd"] Oct 05 07:00:31 crc kubenswrapper[4846]: W1005 07:00:31.512846 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod46fb70d7_c48a_4d57_990f_5e583f96fa16.slice/crio-9895f2fa34c299ff5105b7367f57945b829b9c7b4710859c502d2b3ad555eb08 WatchSource:0}: Error finding container 9895f2fa34c299ff5105b7367f57945b829b9c7b4710859c502d2b3ad555eb08: Status 404 returned error can't find the container with id 9895f2fa34c299ff5105b7367f57945b829b9c7b4710859c502d2b3ad555eb08 Oct 05 07:00:32 crc kubenswrapper[4846]: I1005 07:00:32.533450 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-858ddd8f98-lwsfd" event={"ID":"46fb70d7-c48a-4d57-990f-5e583f96fa16","Type":"ContainerStarted","Data":"9895f2fa34c299ff5105b7367f57945b829b9c7b4710859c502d2b3ad555eb08"} Oct 05 07:00:32 crc kubenswrapper[4846]: I1005 07:00:32.769633 4846 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Oct 05 07:00:34 crc kubenswrapper[4846]: I1005 07:00:34.547029 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-858ddd8f98-lwsfd" event={"ID":"46fb70d7-c48a-4d57-990f-5e583f96fa16","Type":"ContainerStarted","Data":"ac0d81b618a9789ba0c287b0b6acfd9e7112e6c83534812f7c45ba1bfba1f11a"} Oct 05 07:00:34 crc kubenswrapper[4846]: I1005 07:00:34.580464 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-operator-858ddd8f98-lwsfd" podStartSLOduration=2.463843655 podStartE2EDuration="4.580435602s" podCreationTimestamp="2025-10-05 07:00:30 +0000 UTC" firstStartedPulling="2025-10-05 07:00:31.515090408 +0000 UTC m=+753.755943173" lastFinishedPulling="2025-10-05 07:00:33.631682335 +0000 UTC m=+755.872535120" observedRunningTime="2025-10-05 07:00:34.575922883 +0000 UTC m=+756.816775678" watchObservedRunningTime="2025-10-05 07:00:34.580435602 +0000 UTC m=+756.821288377" Oct 05 07:00:42 crc kubenswrapper[4846]: I1005 07:00:42.437573 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-metrics-fdff9cb8d-dlm67"] Oct 05 07:00:42 crc kubenswrapper[4846]: I1005 07:00:42.439298 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-dlm67" Oct 05 07:00:42 crc kubenswrapper[4846]: I1005 07:00:42.442247 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-handler-dockercfg-r4m58" Oct 05 07:00:42 crc kubenswrapper[4846]: I1005 07:00:42.448970 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-webhook-6cdbc54649-h5krq"] Oct 05 07:00:42 crc kubenswrapper[4846]: I1005 07:00:42.450153 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-h5krq" Oct 05 07:00:42 crc kubenswrapper[4846]: I1005 07:00:42.453349 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"openshift-nmstate-webhook" Oct 05 07:00:42 crc kubenswrapper[4846]: I1005 07:00:42.453441 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-fdff9cb8d-dlm67"] Oct 05 07:00:42 crc kubenswrapper[4846]: I1005 07:00:42.483291 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-6cdbc54649-h5krq"] Oct 05 07:00:42 crc kubenswrapper[4846]: I1005 07:00:42.487634 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-handler-5dcf6"] Oct 05 07:00:42 crc kubenswrapper[4846]: I1005 07:00:42.488474 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-5dcf6" Oct 05 07:00:42 crc kubenswrapper[4846]: I1005 07:00:42.626902 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-console-plugin-6b874cbd85-ltpns"] Oct 05 07:00:42 crc kubenswrapper[4846]: I1005 07:00:42.629325 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-ltpns" Oct 05 07:00:42 crc kubenswrapper[4846]: I1005 07:00:42.632935 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"plugin-serving-cert" Oct 05 07:00:42 crc kubenswrapper[4846]: I1005 07:00:42.633342 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"default-dockercfg-jclr6" Oct 05 07:00:42 crc kubenswrapper[4846]: I1005 07:00:42.633957 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"nginx-conf" Oct 05 07:00:42 crc kubenswrapper[4846]: I1005 07:00:42.634854 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/ceef70f7-1471-4dd1-aabc-85d885a302da-tls-key-pair\") pod \"nmstate-webhook-6cdbc54649-h5krq\" (UID: \"ceef70f7-1471-4dd1-aabc-85d885a302da\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-h5krq" Oct 05 07:00:42 crc kubenswrapper[4846]: I1005 07:00:42.634909 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/8383ac5a-9f8c-4bdc-aded-25b6b9cb837a-ovs-socket\") pod \"nmstate-handler-5dcf6\" (UID: \"8383ac5a-9f8c-4bdc-aded-25b6b9cb837a\") " pod="openshift-nmstate/nmstate-handler-5dcf6" Oct 05 07:00:42 crc kubenswrapper[4846]: I1005 07:00:42.634929 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jgrq2\" (UniqueName: \"kubernetes.io/projected/8383ac5a-9f8c-4bdc-aded-25b6b9cb837a-kube-api-access-jgrq2\") pod \"nmstate-handler-5dcf6\" (UID: \"8383ac5a-9f8c-4bdc-aded-25b6b9cb837a\") " pod="openshift-nmstate/nmstate-handler-5dcf6" Oct 05 07:00:42 crc kubenswrapper[4846]: I1005 07:00:42.634994 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/8383ac5a-9f8c-4bdc-aded-25b6b9cb837a-nmstate-lock\") pod \"nmstate-handler-5dcf6\" (UID: \"8383ac5a-9f8c-4bdc-aded-25b6b9cb837a\") " pod="openshift-nmstate/nmstate-handler-5dcf6" Oct 05 07:00:42 crc kubenswrapper[4846]: I1005 07:00:42.635014 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/8383ac5a-9f8c-4bdc-aded-25b6b9cb837a-dbus-socket\") pod \"nmstate-handler-5dcf6\" (UID: \"8383ac5a-9f8c-4bdc-aded-25b6b9cb837a\") " pod="openshift-nmstate/nmstate-handler-5dcf6" Oct 05 07:00:42 crc kubenswrapper[4846]: I1005 07:00:42.635032 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pwm5q\" (UniqueName: \"kubernetes.io/projected/62d96d39-477a-468c-8b8b-02521146f2ba-kube-api-access-pwm5q\") pod \"nmstate-metrics-fdff9cb8d-dlm67\" (UID: \"62d96d39-477a-468c-8b8b-02521146f2ba\") " pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-dlm67" Oct 05 07:00:42 crc kubenswrapper[4846]: I1005 07:00:42.635050 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gch9m\" (UniqueName: \"kubernetes.io/projected/ceef70f7-1471-4dd1-aabc-85d885a302da-kube-api-access-gch9m\") pod \"nmstate-webhook-6cdbc54649-h5krq\" (UID: \"ceef70f7-1471-4dd1-aabc-85d885a302da\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-h5krq" Oct 05 07:00:42 crc kubenswrapper[4846]: I1005 07:00:42.635682 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-6b874cbd85-ltpns"] Oct 05 07:00:42 crc kubenswrapper[4846]: I1005 07:00:42.736751 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/624c96f8-0f60-48c9-abd3-5c9594a525a7-nginx-conf\") pod \"nmstate-console-plugin-6b874cbd85-ltpns\" (UID: \"624c96f8-0f60-48c9-abd3-5c9594a525a7\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-ltpns" Oct 05 07:00:42 crc kubenswrapper[4846]: I1005 07:00:42.736811 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/8383ac5a-9f8c-4bdc-aded-25b6b9cb837a-nmstate-lock\") pod \"nmstate-handler-5dcf6\" (UID: \"8383ac5a-9f8c-4bdc-aded-25b6b9cb837a\") " pod="openshift-nmstate/nmstate-handler-5dcf6" Oct 05 07:00:42 crc kubenswrapper[4846]: I1005 07:00:42.736834 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/8383ac5a-9f8c-4bdc-aded-25b6b9cb837a-dbus-socket\") pod \"nmstate-handler-5dcf6\" (UID: \"8383ac5a-9f8c-4bdc-aded-25b6b9cb837a\") " pod="openshift-nmstate/nmstate-handler-5dcf6" Oct 05 07:00:42 crc kubenswrapper[4846]: I1005 07:00:42.736852 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pwm5q\" (UniqueName: \"kubernetes.io/projected/62d96d39-477a-468c-8b8b-02521146f2ba-kube-api-access-pwm5q\") pod \"nmstate-metrics-fdff9cb8d-dlm67\" (UID: \"62d96d39-477a-468c-8b8b-02521146f2ba\") " pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-dlm67" Oct 05 07:00:42 crc kubenswrapper[4846]: I1005 07:00:42.736997 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/8383ac5a-9f8c-4bdc-aded-25b6b9cb837a-nmstate-lock\") pod \"nmstate-handler-5dcf6\" (UID: \"8383ac5a-9f8c-4bdc-aded-25b6b9cb837a\") " pod="openshift-nmstate/nmstate-handler-5dcf6" Oct 05 07:00:42 crc kubenswrapper[4846]: I1005 07:00:42.737268 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/8383ac5a-9f8c-4bdc-aded-25b6b9cb837a-dbus-socket\") pod \"nmstate-handler-5dcf6\" (UID: \"8383ac5a-9f8c-4bdc-aded-25b6b9cb837a\") " pod="openshift-nmstate/nmstate-handler-5dcf6" Oct 05 07:00:42 crc kubenswrapper[4846]: I1005 07:00:42.737327 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gch9m\" (UniqueName: \"kubernetes.io/projected/ceef70f7-1471-4dd1-aabc-85d885a302da-kube-api-access-gch9m\") pod \"nmstate-webhook-6cdbc54649-h5krq\" (UID: \"ceef70f7-1471-4dd1-aabc-85d885a302da\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-h5krq" Oct 05 07:00:42 crc kubenswrapper[4846]: I1005 07:00:42.738222 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/624c96f8-0f60-48c9-abd3-5c9594a525a7-plugin-serving-cert\") pod \"nmstate-console-plugin-6b874cbd85-ltpns\" (UID: \"624c96f8-0f60-48c9-abd3-5c9594a525a7\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-ltpns" Oct 05 07:00:42 crc kubenswrapper[4846]: I1005 07:00:42.738280 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/ceef70f7-1471-4dd1-aabc-85d885a302da-tls-key-pair\") pod \"nmstate-webhook-6cdbc54649-h5krq\" (UID: \"ceef70f7-1471-4dd1-aabc-85d885a302da\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-h5krq" Oct 05 07:00:42 crc kubenswrapper[4846]: I1005 07:00:42.738321 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pchqw\" (UniqueName: \"kubernetes.io/projected/624c96f8-0f60-48c9-abd3-5c9594a525a7-kube-api-access-pchqw\") pod \"nmstate-console-plugin-6b874cbd85-ltpns\" (UID: \"624c96f8-0f60-48c9-abd3-5c9594a525a7\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-ltpns" Oct 05 07:00:42 crc kubenswrapper[4846]: I1005 07:00:42.738344 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jgrq2\" (UniqueName: \"kubernetes.io/projected/8383ac5a-9f8c-4bdc-aded-25b6b9cb837a-kube-api-access-jgrq2\") pod \"nmstate-handler-5dcf6\" (UID: \"8383ac5a-9f8c-4bdc-aded-25b6b9cb837a\") " pod="openshift-nmstate/nmstate-handler-5dcf6" Oct 05 07:00:42 crc kubenswrapper[4846]: I1005 07:00:42.738364 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/8383ac5a-9f8c-4bdc-aded-25b6b9cb837a-ovs-socket\") pod \"nmstate-handler-5dcf6\" (UID: \"8383ac5a-9f8c-4bdc-aded-25b6b9cb837a\") " pod="openshift-nmstate/nmstate-handler-5dcf6" Oct 05 07:00:42 crc kubenswrapper[4846]: I1005 07:00:42.738438 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/8383ac5a-9f8c-4bdc-aded-25b6b9cb837a-ovs-socket\") pod \"nmstate-handler-5dcf6\" (UID: \"8383ac5a-9f8c-4bdc-aded-25b6b9cb837a\") " pod="openshift-nmstate/nmstate-handler-5dcf6" Oct 05 07:00:42 crc kubenswrapper[4846]: E1005 07:00:42.738787 4846 secret.go:188] Couldn't get secret openshift-nmstate/openshift-nmstate-webhook: secret "openshift-nmstate-webhook" not found Oct 05 07:00:42 crc kubenswrapper[4846]: E1005 07:00:42.739081 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ceef70f7-1471-4dd1-aabc-85d885a302da-tls-key-pair podName:ceef70f7-1471-4dd1-aabc-85d885a302da nodeName:}" failed. No retries permitted until 2025-10-05 07:00:43.239011289 +0000 UTC m=+765.479864064 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "tls-key-pair" (UniqueName: "kubernetes.io/secret/ceef70f7-1471-4dd1-aabc-85d885a302da-tls-key-pair") pod "nmstate-webhook-6cdbc54649-h5krq" (UID: "ceef70f7-1471-4dd1-aabc-85d885a302da") : secret "openshift-nmstate-webhook" not found Oct 05 07:00:42 crc kubenswrapper[4846]: I1005 07:00:42.761419 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pwm5q\" (UniqueName: \"kubernetes.io/projected/62d96d39-477a-468c-8b8b-02521146f2ba-kube-api-access-pwm5q\") pod \"nmstate-metrics-fdff9cb8d-dlm67\" (UID: \"62d96d39-477a-468c-8b8b-02521146f2ba\") " pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-dlm67" Oct 05 07:00:42 crc kubenswrapper[4846]: I1005 07:00:42.761648 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gch9m\" (UniqueName: \"kubernetes.io/projected/ceef70f7-1471-4dd1-aabc-85d885a302da-kube-api-access-gch9m\") pod \"nmstate-webhook-6cdbc54649-h5krq\" (UID: \"ceef70f7-1471-4dd1-aabc-85d885a302da\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-h5krq" Oct 05 07:00:42 crc kubenswrapper[4846]: I1005 07:00:42.764041 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-dlm67" Oct 05 07:00:42 crc kubenswrapper[4846]: I1005 07:00:42.769226 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jgrq2\" (UniqueName: \"kubernetes.io/projected/8383ac5a-9f8c-4bdc-aded-25b6b9cb837a-kube-api-access-jgrq2\") pod \"nmstate-handler-5dcf6\" (UID: \"8383ac5a-9f8c-4bdc-aded-25b6b9cb837a\") " pod="openshift-nmstate/nmstate-handler-5dcf6" Oct 05 07:00:42 crc kubenswrapper[4846]: I1005 07:00:42.810097 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-8f6745764-bwqzz"] Oct 05 07:00:42 crc kubenswrapper[4846]: I1005 07:00:42.817868 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-5dcf6" Oct 05 07:00:42 crc kubenswrapper[4846]: I1005 07:00:42.818694 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-8f6745764-bwqzz" Oct 05 07:00:42 crc kubenswrapper[4846]: I1005 07:00:42.819071 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-8f6745764-bwqzz"] Oct 05 07:00:42 crc kubenswrapper[4846]: I1005 07:00:42.839950 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pchqw\" (UniqueName: \"kubernetes.io/projected/624c96f8-0f60-48c9-abd3-5c9594a525a7-kube-api-access-pchqw\") pod \"nmstate-console-plugin-6b874cbd85-ltpns\" (UID: \"624c96f8-0f60-48c9-abd3-5c9594a525a7\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-ltpns" Oct 05 07:00:42 crc kubenswrapper[4846]: I1005 07:00:42.840056 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/624c96f8-0f60-48c9-abd3-5c9594a525a7-nginx-conf\") pod \"nmstate-console-plugin-6b874cbd85-ltpns\" (UID: \"624c96f8-0f60-48c9-abd3-5c9594a525a7\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-ltpns" Oct 05 07:00:42 crc kubenswrapper[4846]: I1005 07:00:42.840135 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/624c96f8-0f60-48c9-abd3-5c9594a525a7-plugin-serving-cert\") pod \"nmstate-console-plugin-6b874cbd85-ltpns\" (UID: \"624c96f8-0f60-48c9-abd3-5c9594a525a7\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-ltpns" Oct 05 07:00:42 crc kubenswrapper[4846]: I1005 07:00:42.841748 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/624c96f8-0f60-48c9-abd3-5c9594a525a7-nginx-conf\") pod \"nmstate-console-plugin-6b874cbd85-ltpns\" (UID: \"624c96f8-0f60-48c9-abd3-5c9594a525a7\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-ltpns" Oct 05 07:00:42 crc kubenswrapper[4846]: E1005 07:00:42.842090 4846 secret.go:188] Couldn't get secret openshift-nmstate/plugin-serving-cert: secret "plugin-serving-cert" not found Oct 05 07:00:42 crc kubenswrapper[4846]: E1005 07:00:42.842144 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/624c96f8-0f60-48c9-abd3-5c9594a525a7-plugin-serving-cert podName:624c96f8-0f60-48c9-abd3-5c9594a525a7 nodeName:}" failed. No retries permitted until 2025-10-05 07:00:43.342125275 +0000 UTC m=+765.582978050 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "plugin-serving-cert" (UniqueName: "kubernetes.io/secret/624c96f8-0f60-48c9-abd3-5c9594a525a7-plugin-serving-cert") pod "nmstate-console-plugin-6b874cbd85-ltpns" (UID: "624c96f8-0f60-48c9-abd3-5c9594a525a7") : secret "plugin-serving-cert" not found Oct 05 07:00:42 crc kubenswrapper[4846]: I1005 07:00:42.871922 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pchqw\" (UniqueName: \"kubernetes.io/projected/624c96f8-0f60-48c9-abd3-5c9594a525a7-kube-api-access-pchqw\") pod \"nmstate-console-plugin-6b874cbd85-ltpns\" (UID: \"624c96f8-0f60-48c9-abd3-5c9594a525a7\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-ltpns" Oct 05 07:00:42 crc kubenswrapper[4846]: W1005 07:00:42.872952 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8383ac5a_9f8c_4bdc_aded_25b6b9cb837a.slice/crio-780b7bbd49def19f22b86f22b1227e2ef31357f8788187b6dda5d70493393647 WatchSource:0}: Error finding container 780b7bbd49def19f22b86f22b1227e2ef31357f8788187b6dda5d70493393647: Status 404 returned error can't find the container with id 780b7bbd49def19f22b86f22b1227e2ef31357f8788187b6dda5d70493393647 Oct 05 07:00:42 crc kubenswrapper[4846]: I1005 07:00:42.943101 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/d325fc46-671a-41f9-bf01-b76e427b7c00-console-oauth-config\") pod \"console-8f6745764-bwqzz\" (UID: \"d325fc46-671a-41f9-bf01-b76e427b7c00\") " pod="openshift-console/console-8f6745764-bwqzz" Oct 05 07:00:42 crc kubenswrapper[4846]: I1005 07:00:42.943205 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d325fc46-671a-41f9-bf01-b76e427b7c00-trusted-ca-bundle\") pod \"console-8f6745764-bwqzz\" (UID: \"d325fc46-671a-41f9-bf01-b76e427b7c00\") " pod="openshift-console/console-8f6745764-bwqzz" Oct 05 07:00:42 crc kubenswrapper[4846]: I1005 07:00:42.943256 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/d325fc46-671a-41f9-bf01-b76e427b7c00-service-ca\") pod \"console-8f6745764-bwqzz\" (UID: \"d325fc46-671a-41f9-bf01-b76e427b7c00\") " pod="openshift-console/console-8f6745764-bwqzz" Oct 05 07:00:42 crc kubenswrapper[4846]: I1005 07:00:42.943377 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/d325fc46-671a-41f9-bf01-b76e427b7c00-oauth-serving-cert\") pod \"console-8f6745764-bwqzz\" (UID: \"d325fc46-671a-41f9-bf01-b76e427b7c00\") " pod="openshift-console/console-8f6745764-bwqzz" Oct 05 07:00:42 crc kubenswrapper[4846]: I1005 07:00:42.943438 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/d325fc46-671a-41f9-bf01-b76e427b7c00-console-serving-cert\") pod \"console-8f6745764-bwqzz\" (UID: \"d325fc46-671a-41f9-bf01-b76e427b7c00\") " pod="openshift-console/console-8f6745764-bwqzz" Oct 05 07:00:42 crc kubenswrapper[4846]: I1005 07:00:42.943517 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/d325fc46-671a-41f9-bf01-b76e427b7c00-console-config\") pod \"console-8f6745764-bwqzz\" (UID: \"d325fc46-671a-41f9-bf01-b76e427b7c00\") " pod="openshift-console/console-8f6745764-bwqzz" Oct 05 07:00:42 crc kubenswrapper[4846]: I1005 07:00:42.943713 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z7bn2\" (UniqueName: \"kubernetes.io/projected/d325fc46-671a-41f9-bf01-b76e427b7c00-kube-api-access-z7bn2\") pod \"console-8f6745764-bwqzz\" (UID: \"d325fc46-671a-41f9-bf01-b76e427b7c00\") " pod="openshift-console/console-8f6745764-bwqzz" Oct 05 07:00:43 crc kubenswrapper[4846]: I1005 07:00:43.045008 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z7bn2\" (UniqueName: \"kubernetes.io/projected/d325fc46-671a-41f9-bf01-b76e427b7c00-kube-api-access-z7bn2\") pod \"console-8f6745764-bwqzz\" (UID: \"d325fc46-671a-41f9-bf01-b76e427b7c00\") " pod="openshift-console/console-8f6745764-bwqzz" Oct 05 07:00:43 crc kubenswrapper[4846]: I1005 07:00:43.045096 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/d325fc46-671a-41f9-bf01-b76e427b7c00-console-oauth-config\") pod \"console-8f6745764-bwqzz\" (UID: \"d325fc46-671a-41f9-bf01-b76e427b7c00\") " pod="openshift-console/console-8f6745764-bwqzz" Oct 05 07:00:43 crc kubenswrapper[4846]: I1005 07:00:43.045157 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d325fc46-671a-41f9-bf01-b76e427b7c00-trusted-ca-bundle\") pod \"console-8f6745764-bwqzz\" (UID: \"d325fc46-671a-41f9-bf01-b76e427b7c00\") " pod="openshift-console/console-8f6745764-bwqzz" Oct 05 07:00:43 crc kubenswrapper[4846]: I1005 07:00:43.045225 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/d325fc46-671a-41f9-bf01-b76e427b7c00-service-ca\") pod \"console-8f6745764-bwqzz\" (UID: \"d325fc46-671a-41f9-bf01-b76e427b7c00\") " pod="openshift-console/console-8f6745764-bwqzz" Oct 05 07:00:43 crc kubenswrapper[4846]: I1005 07:00:43.045281 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/d325fc46-671a-41f9-bf01-b76e427b7c00-oauth-serving-cert\") pod \"console-8f6745764-bwqzz\" (UID: \"d325fc46-671a-41f9-bf01-b76e427b7c00\") " pod="openshift-console/console-8f6745764-bwqzz" Oct 05 07:00:43 crc kubenswrapper[4846]: I1005 07:00:43.045341 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/d325fc46-671a-41f9-bf01-b76e427b7c00-console-serving-cert\") pod \"console-8f6745764-bwqzz\" (UID: \"d325fc46-671a-41f9-bf01-b76e427b7c00\") " pod="openshift-console/console-8f6745764-bwqzz" Oct 05 07:00:43 crc kubenswrapper[4846]: I1005 07:00:43.045392 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/d325fc46-671a-41f9-bf01-b76e427b7c00-console-config\") pod \"console-8f6745764-bwqzz\" (UID: \"d325fc46-671a-41f9-bf01-b76e427b7c00\") " pod="openshift-console/console-8f6745764-bwqzz" Oct 05 07:00:43 crc kubenswrapper[4846]: I1005 07:00:43.046751 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/d325fc46-671a-41f9-bf01-b76e427b7c00-oauth-serving-cert\") pod \"console-8f6745764-bwqzz\" (UID: \"d325fc46-671a-41f9-bf01-b76e427b7c00\") " pod="openshift-console/console-8f6745764-bwqzz" Oct 05 07:00:43 crc kubenswrapper[4846]: I1005 07:00:43.047607 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/d325fc46-671a-41f9-bf01-b76e427b7c00-service-ca\") pod \"console-8f6745764-bwqzz\" (UID: \"d325fc46-671a-41f9-bf01-b76e427b7c00\") " pod="openshift-console/console-8f6745764-bwqzz" Oct 05 07:00:43 crc kubenswrapper[4846]: I1005 07:00:43.047672 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d325fc46-671a-41f9-bf01-b76e427b7c00-trusted-ca-bundle\") pod \"console-8f6745764-bwqzz\" (UID: \"d325fc46-671a-41f9-bf01-b76e427b7c00\") " pod="openshift-console/console-8f6745764-bwqzz" Oct 05 07:00:43 crc kubenswrapper[4846]: I1005 07:00:43.048955 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/d325fc46-671a-41f9-bf01-b76e427b7c00-console-config\") pod \"console-8f6745764-bwqzz\" (UID: \"d325fc46-671a-41f9-bf01-b76e427b7c00\") " pod="openshift-console/console-8f6745764-bwqzz" Oct 05 07:00:43 crc kubenswrapper[4846]: I1005 07:00:43.051638 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/d325fc46-671a-41f9-bf01-b76e427b7c00-console-oauth-config\") pod \"console-8f6745764-bwqzz\" (UID: \"d325fc46-671a-41f9-bf01-b76e427b7c00\") " pod="openshift-console/console-8f6745764-bwqzz" Oct 05 07:00:43 crc kubenswrapper[4846]: I1005 07:00:43.060717 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/d325fc46-671a-41f9-bf01-b76e427b7c00-console-serving-cert\") pod \"console-8f6745764-bwqzz\" (UID: \"d325fc46-671a-41f9-bf01-b76e427b7c00\") " pod="openshift-console/console-8f6745764-bwqzz" Oct 05 07:00:43 crc kubenswrapper[4846]: I1005 07:00:43.069627 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z7bn2\" (UniqueName: \"kubernetes.io/projected/d325fc46-671a-41f9-bf01-b76e427b7c00-kube-api-access-z7bn2\") pod \"console-8f6745764-bwqzz\" (UID: \"d325fc46-671a-41f9-bf01-b76e427b7c00\") " pod="openshift-console/console-8f6745764-bwqzz" Oct 05 07:00:43 crc kubenswrapper[4846]: I1005 07:00:43.154357 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-8f6745764-bwqzz" Oct 05 07:00:43 crc kubenswrapper[4846]: I1005 07:00:43.243062 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-fdff9cb8d-dlm67"] Oct 05 07:00:43 crc kubenswrapper[4846]: I1005 07:00:43.249659 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/ceef70f7-1471-4dd1-aabc-85d885a302da-tls-key-pair\") pod \"nmstate-webhook-6cdbc54649-h5krq\" (UID: \"ceef70f7-1471-4dd1-aabc-85d885a302da\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-h5krq" Oct 05 07:00:43 crc kubenswrapper[4846]: I1005 07:00:43.259022 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/ceef70f7-1471-4dd1-aabc-85d885a302da-tls-key-pair\") pod \"nmstate-webhook-6cdbc54649-h5krq\" (UID: \"ceef70f7-1471-4dd1-aabc-85d885a302da\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-h5krq" Oct 05 07:00:43 crc kubenswrapper[4846]: I1005 07:00:43.350852 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/624c96f8-0f60-48c9-abd3-5c9594a525a7-plugin-serving-cert\") pod \"nmstate-console-plugin-6b874cbd85-ltpns\" (UID: \"624c96f8-0f60-48c9-abd3-5c9594a525a7\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-ltpns" Oct 05 07:00:43 crc kubenswrapper[4846]: I1005 07:00:43.355094 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/624c96f8-0f60-48c9-abd3-5c9594a525a7-plugin-serving-cert\") pod \"nmstate-console-plugin-6b874cbd85-ltpns\" (UID: \"624c96f8-0f60-48c9-abd3-5c9594a525a7\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-ltpns" Oct 05 07:00:43 crc kubenswrapper[4846]: I1005 07:00:43.376005 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-h5krq" Oct 05 07:00:43 crc kubenswrapper[4846]: I1005 07:00:43.548860 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-ltpns" Oct 05 07:00:43 crc kubenswrapper[4846]: I1005 07:00:43.627736 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-dlm67" event={"ID":"62d96d39-477a-468c-8b8b-02521146f2ba","Type":"ContainerStarted","Data":"eb82b55e4084e00d3e4dbd1e038db7a2b0d1357a8af9669c199819051d10052b"} Oct 05 07:00:43 crc kubenswrapper[4846]: I1005 07:00:43.630270 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-5dcf6" event={"ID":"8383ac5a-9f8c-4bdc-aded-25b6b9cb837a","Type":"ContainerStarted","Data":"780b7bbd49def19f22b86f22b1227e2ef31357f8788187b6dda5d70493393647"} Oct 05 07:00:43 crc kubenswrapper[4846]: I1005 07:00:43.681227 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-8f6745764-bwqzz"] Oct 05 07:00:43 crc kubenswrapper[4846]: W1005 07:00:43.690822 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd325fc46_671a_41f9_bf01_b76e427b7c00.slice/crio-24e560ea74f7a5baaeb74757d3f6573d1dc05f7a595784433e29b5c484099f6d WatchSource:0}: Error finding container 24e560ea74f7a5baaeb74757d3f6573d1dc05f7a595784433e29b5c484099f6d: Status 404 returned error can't find the container with id 24e560ea74f7a5baaeb74757d3f6573d1dc05f7a595784433e29b5c484099f6d Oct 05 07:00:43 crc kubenswrapper[4846]: I1005 07:00:43.834259 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-6cdbc54649-h5krq"] Oct 05 07:00:44 crc kubenswrapper[4846]: I1005 07:00:44.005214 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-6b874cbd85-ltpns"] Oct 05 07:00:44 crc kubenswrapper[4846]: W1005 07:00:44.012008 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod624c96f8_0f60_48c9_abd3_5c9594a525a7.slice/crio-03118fac39388c0d04611d4fe0bae8e38de6ad95381fa9a17633f33fd74bd88c WatchSource:0}: Error finding container 03118fac39388c0d04611d4fe0bae8e38de6ad95381fa9a17633f33fd74bd88c: Status 404 returned error can't find the container with id 03118fac39388c0d04611d4fe0bae8e38de6ad95381fa9a17633f33fd74bd88c Oct 05 07:00:44 crc kubenswrapper[4846]: I1005 07:00:44.640616 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-ltpns" event={"ID":"624c96f8-0f60-48c9-abd3-5c9594a525a7","Type":"ContainerStarted","Data":"03118fac39388c0d04611d4fe0bae8e38de6ad95381fa9a17633f33fd74bd88c"} Oct 05 07:00:44 crc kubenswrapper[4846]: I1005 07:00:44.642850 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-8f6745764-bwqzz" event={"ID":"d325fc46-671a-41f9-bf01-b76e427b7c00","Type":"ContainerStarted","Data":"eaa68e56fd16b6095bb053f425013e2ddd6c426d932c4dc1f65ee3cc5467c691"} Oct 05 07:00:44 crc kubenswrapper[4846]: I1005 07:00:44.642905 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-8f6745764-bwqzz" event={"ID":"d325fc46-671a-41f9-bf01-b76e427b7c00","Type":"ContainerStarted","Data":"24e560ea74f7a5baaeb74757d3f6573d1dc05f7a595784433e29b5c484099f6d"} Oct 05 07:00:44 crc kubenswrapper[4846]: I1005 07:00:44.644019 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-h5krq" event={"ID":"ceef70f7-1471-4dd1-aabc-85d885a302da","Type":"ContainerStarted","Data":"2bfb9c224d3fdb8e2e07570fa4e050b3bd7e6b4f332cb3fa1e91b60fb483897e"} Oct 05 07:00:44 crc kubenswrapper[4846]: I1005 07:00:44.671047 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-8f6745764-bwqzz" podStartSLOduration=2.6710206640000003 podStartE2EDuration="2.671020664s" podCreationTimestamp="2025-10-05 07:00:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 07:00:44.663231799 +0000 UTC m=+766.904084584" watchObservedRunningTime="2025-10-05 07:00:44.671020664 +0000 UTC m=+766.911873439" Oct 05 07:00:45 crc kubenswrapper[4846]: I1005 07:00:45.652609 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-h5krq" event={"ID":"ceef70f7-1471-4dd1-aabc-85d885a302da","Type":"ContainerStarted","Data":"0efbb5b8a71c2ea39c87d6e48b1e7b09e8cf604843f11e8d5b1e0d82adbef489"} Oct 05 07:00:45 crc kubenswrapper[4846]: I1005 07:00:45.652895 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-h5krq" Oct 05 07:00:45 crc kubenswrapper[4846]: I1005 07:00:45.654498 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-dlm67" event={"ID":"62d96d39-477a-468c-8b8b-02521146f2ba","Type":"ContainerStarted","Data":"b05926b2228e345806b7444dad7eb9f9fa84216e4dc7f6d216a69a608ddde5e0"} Oct 05 07:00:45 crc kubenswrapper[4846]: I1005 07:00:45.655822 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-5dcf6" event={"ID":"8383ac5a-9f8c-4bdc-aded-25b6b9cb837a","Type":"ContainerStarted","Data":"cb3460735e73d49163493ba88169698300a6e3da1c3a8110479226f9c2ae2121"} Oct 05 07:00:45 crc kubenswrapper[4846]: I1005 07:00:45.656051 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-handler-5dcf6" Oct 05 07:00:45 crc kubenswrapper[4846]: I1005 07:00:45.704721 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-h5krq" podStartSLOduration=2.293398849 podStartE2EDuration="3.704688748s" podCreationTimestamp="2025-10-05 07:00:42 +0000 UTC" firstStartedPulling="2025-10-05 07:00:43.849833776 +0000 UTC m=+766.090686551" lastFinishedPulling="2025-10-05 07:00:45.261123665 +0000 UTC m=+767.501976450" observedRunningTime="2025-10-05 07:00:45.679998707 +0000 UTC m=+767.920851502" watchObservedRunningTime="2025-10-05 07:00:45.704688748 +0000 UTC m=+767.945541543" Oct 05 07:00:45 crc kubenswrapper[4846]: I1005 07:00:45.710523 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-handler-5dcf6" podStartSLOduration=1.329033379 podStartE2EDuration="3.710505321s" podCreationTimestamp="2025-10-05 07:00:42 +0000 UTC" firstStartedPulling="2025-10-05 07:00:42.876415588 +0000 UTC m=+765.117268363" lastFinishedPulling="2025-10-05 07:00:45.25788752 +0000 UTC m=+767.498740305" observedRunningTime="2025-10-05 07:00:45.701143794 +0000 UTC m=+767.941996579" watchObservedRunningTime="2025-10-05 07:00:45.710505321 +0000 UTC m=+767.951358106" Oct 05 07:00:46 crc kubenswrapper[4846]: I1005 07:00:46.670566 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-ltpns" event={"ID":"624c96f8-0f60-48c9-abd3-5c9594a525a7","Type":"ContainerStarted","Data":"f960ac1e53ad16d8a9a4b60805cf550b7edcd3e820c6f85045cf154241f080d5"} Oct 05 07:00:46 crc kubenswrapper[4846]: I1005 07:00:46.698714 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-ltpns" podStartSLOduration=2.422400046 podStartE2EDuration="4.698695118s" podCreationTimestamp="2025-10-05 07:00:42 +0000 UTC" firstStartedPulling="2025-10-05 07:00:44.015572391 +0000 UTC m=+766.256425176" lastFinishedPulling="2025-10-05 07:00:46.291867473 +0000 UTC m=+768.532720248" observedRunningTime="2025-10-05 07:00:46.695196096 +0000 UTC m=+768.936048871" watchObservedRunningTime="2025-10-05 07:00:46.698695118 +0000 UTC m=+768.939547893" Oct 05 07:00:48 crc kubenswrapper[4846]: I1005 07:00:48.691041 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-dlm67" event={"ID":"62d96d39-477a-468c-8b8b-02521146f2ba","Type":"ContainerStarted","Data":"03b92eef36e7336db44f4e80602c2971735a27534ec0c3d5ac4ad2ac67b2a548"} Oct 05 07:00:52 crc kubenswrapper[4846]: I1005 07:00:52.866368 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-handler-5dcf6" Oct 05 07:00:52 crc kubenswrapper[4846]: I1005 07:00:52.903848 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-dlm67" podStartSLOduration=6.18870945 podStartE2EDuration="10.903813285s" podCreationTimestamp="2025-10-05 07:00:42 +0000 UTC" firstStartedPulling="2025-10-05 07:00:43.262029574 +0000 UTC m=+765.502882359" lastFinishedPulling="2025-10-05 07:00:47.977133379 +0000 UTC m=+770.217986194" observedRunningTime="2025-10-05 07:00:48.717756434 +0000 UTC m=+770.958609249" watchObservedRunningTime="2025-10-05 07:00:52.903813285 +0000 UTC m=+775.144666100" Oct 05 07:00:53 crc kubenswrapper[4846]: I1005 07:00:53.156150 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-8f6745764-bwqzz" Oct 05 07:00:53 crc kubenswrapper[4846]: I1005 07:00:53.156327 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-8f6745764-bwqzz" Oct 05 07:00:53 crc kubenswrapper[4846]: I1005 07:00:53.171951 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-8f6745764-bwqzz" Oct 05 07:00:53 crc kubenswrapper[4846]: I1005 07:00:53.325434 4846 patch_prober.go:28] interesting pod/machine-config-daemon-fscvf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 07:00:53 crc kubenswrapper[4846]: I1005 07:00:53.325553 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 07:00:53 crc kubenswrapper[4846]: I1005 07:00:53.736042 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-8f6745764-bwqzz" Oct 05 07:00:53 crc kubenswrapper[4846]: I1005 07:00:53.823649 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-9hcbc"] Oct 05 07:01:03 crc kubenswrapper[4846]: I1005 07:01:03.388669 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-h5krq" Oct 05 07:01:18 crc kubenswrapper[4846]: I1005 07:01:18.910456 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-f9d7485db-9hcbc" podUID="74f16025-fe11-4a23-8c1a-ff3da2a3d751" containerName="console" containerID="cri-o://8b88162ddf15a27840637668117eb2e0ff0a349ff0b37277d499b60fec10c880" gracePeriod=15 Oct 05 07:01:19 crc kubenswrapper[4846]: I1005 07:01:19.460164 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-9hcbc_74f16025-fe11-4a23-8c1a-ff3da2a3d751/console/0.log" Oct 05 07:01:19 crc kubenswrapper[4846]: I1005 07:01:19.460652 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-9hcbc" Oct 05 07:01:19 crc kubenswrapper[4846]: I1005 07:01:19.560294 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/74f16025-fe11-4a23-8c1a-ff3da2a3d751-oauth-serving-cert\") pod \"74f16025-fe11-4a23-8c1a-ff3da2a3d751\" (UID: \"74f16025-fe11-4a23-8c1a-ff3da2a3d751\") " Oct 05 07:01:19 crc kubenswrapper[4846]: I1005 07:01:19.560353 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/74f16025-fe11-4a23-8c1a-ff3da2a3d751-trusted-ca-bundle\") pod \"74f16025-fe11-4a23-8c1a-ff3da2a3d751\" (UID: \"74f16025-fe11-4a23-8c1a-ff3da2a3d751\") " Oct 05 07:01:19 crc kubenswrapper[4846]: I1005 07:01:19.560509 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r8vnl\" (UniqueName: \"kubernetes.io/projected/74f16025-fe11-4a23-8c1a-ff3da2a3d751-kube-api-access-r8vnl\") pod \"74f16025-fe11-4a23-8c1a-ff3da2a3d751\" (UID: \"74f16025-fe11-4a23-8c1a-ff3da2a3d751\") " Oct 05 07:01:19 crc kubenswrapper[4846]: I1005 07:01:19.560559 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/74f16025-fe11-4a23-8c1a-ff3da2a3d751-console-config\") pod \"74f16025-fe11-4a23-8c1a-ff3da2a3d751\" (UID: \"74f16025-fe11-4a23-8c1a-ff3da2a3d751\") " Oct 05 07:01:19 crc kubenswrapper[4846]: I1005 07:01:19.560639 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/74f16025-fe11-4a23-8c1a-ff3da2a3d751-console-oauth-config\") pod \"74f16025-fe11-4a23-8c1a-ff3da2a3d751\" (UID: \"74f16025-fe11-4a23-8c1a-ff3da2a3d751\") " Oct 05 07:01:19 crc kubenswrapper[4846]: I1005 07:01:19.560677 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/74f16025-fe11-4a23-8c1a-ff3da2a3d751-service-ca\") pod \"74f16025-fe11-4a23-8c1a-ff3da2a3d751\" (UID: \"74f16025-fe11-4a23-8c1a-ff3da2a3d751\") " Oct 05 07:01:19 crc kubenswrapper[4846]: I1005 07:01:19.560709 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/74f16025-fe11-4a23-8c1a-ff3da2a3d751-console-serving-cert\") pod \"74f16025-fe11-4a23-8c1a-ff3da2a3d751\" (UID: \"74f16025-fe11-4a23-8c1a-ff3da2a3d751\") " Oct 05 07:01:19 crc kubenswrapper[4846]: I1005 07:01:19.561505 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/74f16025-fe11-4a23-8c1a-ff3da2a3d751-service-ca" (OuterVolumeSpecName: "service-ca") pod "74f16025-fe11-4a23-8c1a-ff3da2a3d751" (UID: "74f16025-fe11-4a23-8c1a-ff3da2a3d751"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 07:01:19 crc kubenswrapper[4846]: I1005 07:01:19.561535 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/74f16025-fe11-4a23-8c1a-ff3da2a3d751-console-config" (OuterVolumeSpecName: "console-config") pod "74f16025-fe11-4a23-8c1a-ff3da2a3d751" (UID: "74f16025-fe11-4a23-8c1a-ff3da2a3d751"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 07:01:19 crc kubenswrapper[4846]: I1005 07:01:19.561638 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/74f16025-fe11-4a23-8c1a-ff3da2a3d751-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "74f16025-fe11-4a23-8c1a-ff3da2a3d751" (UID: "74f16025-fe11-4a23-8c1a-ff3da2a3d751"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 07:01:19 crc kubenswrapper[4846]: I1005 07:01:19.562569 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/74f16025-fe11-4a23-8c1a-ff3da2a3d751-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "74f16025-fe11-4a23-8c1a-ff3da2a3d751" (UID: "74f16025-fe11-4a23-8c1a-ff3da2a3d751"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 07:01:19 crc kubenswrapper[4846]: I1005 07:01:19.568326 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/74f16025-fe11-4a23-8c1a-ff3da2a3d751-kube-api-access-r8vnl" (OuterVolumeSpecName: "kube-api-access-r8vnl") pod "74f16025-fe11-4a23-8c1a-ff3da2a3d751" (UID: "74f16025-fe11-4a23-8c1a-ff3da2a3d751"). InnerVolumeSpecName "kube-api-access-r8vnl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:01:19 crc kubenswrapper[4846]: I1005 07:01:19.568646 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/74f16025-fe11-4a23-8c1a-ff3da2a3d751-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "74f16025-fe11-4a23-8c1a-ff3da2a3d751" (UID: "74f16025-fe11-4a23-8c1a-ff3da2a3d751"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:01:19 crc kubenswrapper[4846]: I1005 07:01:19.569166 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/74f16025-fe11-4a23-8c1a-ff3da2a3d751-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "74f16025-fe11-4a23-8c1a-ff3da2a3d751" (UID: "74f16025-fe11-4a23-8c1a-ff3da2a3d751"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:01:19 crc kubenswrapper[4846]: I1005 07:01:19.663126 4846 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/74f16025-fe11-4a23-8c1a-ff3da2a3d751-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 05 07:01:19 crc kubenswrapper[4846]: I1005 07:01:19.663258 4846 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/74f16025-fe11-4a23-8c1a-ff3da2a3d751-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 07:01:19 crc kubenswrapper[4846]: I1005 07:01:19.663287 4846 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/74f16025-fe11-4a23-8c1a-ff3da2a3d751-console-config\") on node \"crc\" DevicePath \"\"" Oct 05 07:01:19 crc kubenswrapper[4846]: I1005 07:01:19.663313 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r8vnl\" (UniqueName: \"kubernetes.io/projected/74f16025-fe11-4a23-8c1a-ff3da2a3d751-kube-api-access-r8vnl\") on node \"crc\" DevicePath \"\"" Oct 05 07:01:19 crc kubenswrapper[4846]: I1005 07:01:19.663343 4846 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/74f16025-fe11-4a23-8c1a-ff3da2a3d751-service-ca\") on node \"crc\" DevicePath \"\"" Oct 05 07:01:19 crc kubenswrapper[4846]: I1005 07:01:19.663370 4846 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/74f16025-fe11-4a23-8c1a-ff3da2a3d751-console-oauth-config\") on node \"crc\" DevicePath \"\"" Oct 05 07:01:19 crc kubenswrapper[4846]: I1005 07:01:19.663396 4846 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/74f16025-fe11-4a23-8c1a-ff3da2a3d751-console-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 05 07:01:19 crc kubenswrapper[4846]: I1005 07:01:19.954787 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-9hcbc_74f16025-fe11-4a23-8c1a-ff3da2a3d751/console/0.log" Oct 05 07:01:19 crc kubenswrapper[4846]: I1005 07:01:19.954863 4846 generic.go:334] "Generic (PLEG): container finished" podID="74f16025-fe11-4a23-8c1a-ff3da2a3d751" containerID="8b88162ddf15a27840637668117eb2e0ff0a349ff0b37277d499b60fec10c880" exitCode=2 Oct 05 07:01:19 crc kubenswrapper[4846]: I1005 07:01:19.954910 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-9hcbc" event={"ID":"74f16025-fe11-4a23-8c1a-ff3da2a3d751","Type":"ContainerDied","Data":"8b88162ddf15a27840637668117eb2e0ff0a349ff0b37277d499b60fec10c880"} Oct 05 07:01:19 crc kubenswrapper[4846]: I1005 07:01:19.954981 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-9hcbc" event={"ID":"74f16025-fe11-4a23-8c1a-ff3da2a3d751","Type":"ContainerDied","Data":"6a2010efd88b09b2d8cf8c289be1ddd4504bd85b79adc3f42557ed0bb350b4f8"} Oct 05 07:01:19 crc kubenswrapper[4846]: I1005 07:01:19.955008 4846 scope.go:117] "RemoveContainer" containerID="8b88162ddf15a27840637668117eb2e0ff0a349ff0b37277d499b60fec10c880" Oct 05 07:01:19 crc kubenswrapper[4846]: I1005 07:01:19.955301 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-9hcbc" Oct 05 07:01:19 crc kubenswrapper[4846]: I1005 07:01:19.980381 4846 scope.go:117] "RemoveContainer" containerID="8b88162ddf15a27840637668117eb2e0ff0a349ff0b37277d499b60fec10c880" Oct 05 07:01:19 crc kubenswrapper[4846]: E1005 07:01:19.984869 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8b88162ddf15a27840637668117eb2e0ff0a349ff0b37277d499b60fec10c880\": container with ID starting with 8b88162ddf15a27840637668117eb2e0ff0a349ff0b37277d499b60fec10c880 not found: ID does not exist" containerID="8b88162ddf15a27840637668117eb2e0ff0a349ff0b37277d499b60fec10c880" Oct 05 07:01:19 crc kubenswrapper[4846]: I1005 07:01:19.984948 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8b88162ddf15a27840637668117eb2e0ff0a349ff0b37277d499b60fec10c880"} err="failed to get container status \"8b88162ddf15a27840637668117eb2e0ff0a349ff0b37277d499b60fec10c880\": rpc error: code = NotFound desc = could not find container \"8b88162ddf15a27840637668117eb2e0ff0a349ff0b37277d499b60fec10c880\": container with ID starting with 8b88162ddf15a27840637668117eb2e0ff0a349ff0b37277d499b60fec10c880 not found: ID does not exist" Oct 05 07:01:20 crc kubenswrapper[4846]: I1005 07:01:20.001408 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2xndcw"] Oct 05 07:01:20 crc kubenswrapper[4846]: E1005 07:01:20.001719 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="74f16025-fe11-4a23-8c1a-ff3da2a3d751" containerName="console" Oct 05 07:01:20 crc kubenswrapper[4846]: I1005 07:01:20.001743 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="74f16025-fe11-4a23-8c1a-ff3da2a3d751" containerName="console" Oct 05 07:01:20 crc kubenswrapper[4846]: I1005 07:01:20.001915 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="74f16025-fe11-4a23-8c1a-ff3da2a3d751" containerName="console" Oct 05 07:01:20 crc kubenswrapper[4846]: I1005 07:01:20.002925 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2xndcw" Oct 05 07:01:20 crc kubenswrapper[4846]: I1005 07:01:20.004773 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Oct 05 07:01:20 crc kubenswrapper[4846]: I1005 07:01:20.018083 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-9hcbc"] Oct 05 07:01:20 crc kubenswrapper[4846]: I1005 07:01:20.023315 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2xndcw"] Oct 05 07:01:20 crc kubenswrapper[4846]: I1005 07:01:20.027740 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-f9d7485db-9hcbc"] Oct 05 07:01:20 crc kubenswrapper[4846]: I1005 07:01:20.070025 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/18f66388-66f4-4457-877f-d47dff48bb98-util\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2xndcw\" (UID: \"18f66388-66f4-4457-877f-d47dff48bb98\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2xndcw" Oct 05 07:01:20 crc kubenswrapper[4846]: I1005 07:01:20.070160 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7vnhn\" (UniqueName: \"kubernetes.io/projected/18f66388-66f4-4457-877f-d47dff48bb98-kube-api-access-7vnhn\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2xndcw\" (UID: \"18f66388-66f4-4457-877f-d47dff48bb98\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2xndcw" Oct 05 07:01:20 crc kubenswrapper[4846]: I1005 07:01:20.070271 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/18f66388-66f4-4457-877f-d47dff48bb98-bundle\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2xndcw\" (UID: \"18f66388-66f4-4457-877f-d47dff48bb98\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2xndcw" Oct 05 07:01:20 crc kubenswrapper[4846]: I1005 07:01:20.172230 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7vnhn\" (UniqueName: \"kubernetes.io/projected/18f66388-66f4-4457-877f-d47dff48bb98-kube-api-access-7vnhn\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2xndcw\" (UID: \"18f66388-66f4-4457-877f-d47dff48bb98\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2xndcw" Oct 05 07:01:20 crc kubenswrapper[4846]: I1005 07:01:20.172315 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/18f66388-66f4-4457-877f-d47dff48bb98-bundle\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2xndcw\" (UID: \"18f66388-66f4-4457-877f-d47dff48bb98\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2xndcw" Oct 05 07:01:20 crc kubenswrapper[4846]: I1005 07:01:20.172359 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/18f66388-66f4-4457-877f-d47dff48bb98-util\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2xndcw\" (UID: \"18f66388-66f4-4457-877f-d47dff48bb98\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2xndcw" Oct 05 07:01:20 crc kubenswrapper[4846]: I1005 07:01:20.173465 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/18f66388-66f4-4457-877f-d47dff48bb98-util\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2xndcw\" (UID: \"18f66388-66f4-4457-877f-d47dff48bb98\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2xndcw" Oct 05 07:01:20 crc kubenswrapper[4846]: I1005 07:01:20.173612 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/18f66388-66f4-4457-877f-d47dff48bb98-bundle\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2xndcw\" (UID: \"18f66388-66f4-4457-877f-d47dff48bb98\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2xndcw" Oct 05 07:01:20 crc kubenswrapper[4846]: I1005 07:01:20.205436 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7vnhn\" (UniqueName: \"kubernetes.io/projected/18f66388-66f4-4457-877f-d47dff48bb98-kube-api-access-7vnhn\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2xndcw\" (UID: \"18f66388-66f4-4457-877f-d47dff48bb98\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2xndcw" Oct 05 07:01:20 crc kubenswrapper[4846]: I1005 07:01:20.322477 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2xndcw" Oct 05 07:01:20 crc kubenswrapper[4846]: I1005 07:01:20.515610 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="74f16025-fe11-4a23-8c1a-ff3da2a3d751" path="/var/lib/kubelet/pods/74f16025-fe11-4a23-8c1a-ff3da2a3d751/volumes" Oct 05 07:01:20 crc kubenswrapper[4846]: I1005 07:01:20.840267 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2xndcw"] Oct 05 07:01:20 crc kubenswrapper[4846]: I1005 07:01:20.976434 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2xndcw" event={"ID":"18f66388-66f4-4457-877f-d47dff48bb98","Type":"ContainerStarted","Data":"93dac73088eed762eb85193d46d51bae777e77e1cedf53afbc3cd2d76f954f58"} Oct 05 07:01:21 crc kubenswrapper[4846]: I1005 07:01:21.994890 4846 generic.go:334] "Generic (PLEG): container finished" podID="18f66388-66f4-4457-877f-d47dff48bb98" containerID="0a26c549b8a64c8e340dc8a146b352cdb061db74a77b5b295e92eb2a69cf6f50" exitCode=0 Oct 05 07:01:21 crc kubenswrapper[4846]: I1005 07:01:21.994987 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2xndcw" event={"ID":"18f66388-66f4-4457-877f-d47dff48bb98","Type":"ContainerDied","Data":"0a26c549b8a64c8e340dc8a146b352cdb061db74a77b5b295e92eb2a69cf6f50"} Oct 05 07:01:22 crc kubenswrapper[4846]: I1005 07:01:22.343512 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-xvxxx"] Oct 05 07:01:22 crc kubenswrapper[4846]: I1005 07:01:22.345490 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xvxxx" Oct 05 07:01:22 crc kubenswrapper[4846]: I1005 07:01:22.354443 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-xvxxx"] Oct 05 07:01:22 crc kubenswrapper[4846]: I1005 07:01:22.409001 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-789l5\" (UniqueName: \"kubernetes.io/projected/40d018c5-6953-435a-8a65-b52745cb42c8-kube-api-access-789l5\") pod \"redhat-operators-xvxxx\" (UID: \"40d018c5-6953-435a-8a65-b52745cb42c8\") " pod="openshift-marketplace/redhat-operators-xvxxx" Oct 05 07:01:22 crc kubenswrapper[4846]: I1005 07:01:22.409072 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/40d018c5-6953-435a-8a65-b52745cb42c8-utilities\") pod \"redhat-operators-xvxxx\" (UID: \"40d018c5-6953-435a-8a65-b52745cb42c8\") " pod="openshift-marketplace/redhat-operators-xvxxx" Oct 05 07:01:22 crc kubenswrapper[4846]: I1005 07:01:22.409670 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/40d018c5-6953-435a-8a65-b52745cb42c8-catalog-content\") pod \"redhat-operators-xvxxx\" (UID: \"40d018c5-6953-435a-8a65-b52745cb42c8\") " pod="openshift-marketplace/redhat-operators-xvxxx" Oct 05 07:01:22 crc kubenswrapper[4846]: I1005 07:01:22.511007 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/40d018c5-6953-435a-8a65-b52745cb42c8-catalog-content\") pod \"redhat-operators-xvxxx\" (UID: \"40d018c5-6953-435a-8a65-b52745cb42c8\") " pod="openshift-marketplace/redhat-operators-xvxxx" Oct 05 07:01:22 crc kubenswrapper[4846]: I1005 07:01:22.511105 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-789l5\" (UniqueName: \"kubernetes.io/projected/40d018c5-6953-435a-8a65-b52745cb42c8-kube-api-access-789l5\") pod \"redhat-operators-xvxxx\" (UID: \"40d018c5-6953-435a-8a65-b52745cb42c8\") " pod="openshift-marketplace/redhat-operators-xvxxx" Oct 05 07:01:22 crc kubenswrapper[4846]: I1005 07:01:22.511155 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/40d018c5-6953-435a-8a65-b52745cb42c8-utilities\") pod \"redhat-operators-xvxxx\" (UID: \"40d018c5-6953-435a-8a65-b52745cb42c8\") " pod="openshift-marketplace/redhat-operators-xvxxx" Oct 05 07:01:22 crc kubenswrapper[4846]: I1005 07:01:22.511664 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/40d018c5-6953-435a-8a65-b52745cb42c8-catalog-content\") pod \"redhat-operators-xvxxx\" (UID: \"40d018c5-6953-435a-8a65-b52745cb42c8\") " pod="openshift-marketplace/redhat-operators-xvxxx" Oct 05 07:01:22 crc kubenswrapper[4846]: I1005 07:01:22.511992 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/40d018c5-6953-435a-8a65-b52745cb42c8-utilities\") pod \"redhat-operators-xvxxx\" (UID: \"40d018c5-6953-435a-8a65-b52745cb42c8\") " pod="openshift-marketplace/redhat-operators-xvxxx" Oct 05 07:01:22 crc kubenswrapper[4846]: I1005 07:01:22.551667 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-789l5\" (UniqueName: \"kubernetes.io/projected/40d018c5-6953-435a-8a65-b52745cb42c8-kube-api-access-789l5\") pod \"redhat-operators-xvxxx\" (UID: \"40d018c5-6953-435a-8a65-b52745cb42c8\") " pod="openshift-marketplace/redhat-operators-xvxxx" Oct 05 07:01:22 crc kubenswrapper[4846]: I1005 07:01:22.683604 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xvxxx" Oct 05 07:01:22 crc kubenswrapper[4846]: I1005 07:01:22.995212 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-xvxxx"] Oct 05 07:01:23 crc kubenswrapper[4846]: W1005 07:01:23.028152 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod40d018c5_6953_435a_8a65_b52745cb42c8.slice/crio-c4d6f5285ed6f8e2d48ec982702f9ea994135810022e6cacc109b6def07b12dc WatchSource:0}: Error finding container c4d6f5285ed6f8e2d48ec982702f9ea994135810022e6cacc109b6def07b12dc: Status 404 returned error can't find the container with id c4d6f5285ed6f8e2d48ec982702f9ea994135810022e6cacc109b6def07b12dc Oct 05 07:01:23 crc kubenswrapper[4846]: I1005 07:01:23.325077 4846 patch_prober.go:28] interesting pod/machine-config-daemon-fscvf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 07:01:23 crc kubenswrapper[4846]: I1005 07:01:23.325655 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 07:01:23 crc kubenswrapper[4846]: I1005 07:01:23.325720 4846 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" Oct 05 07:01:23 crc kubenswrapper[4846]: I1005 07:01:23.326574 4846 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"468d3aa8613beb034776fd6c8b08e9b13031340dca2680a10051a40fbbb92cf2"} pod="openshift-machine-config-operator/machine-config-daemon-fscvf" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 05 07:01:23 crc kubenswrapper[4846]: I1005 07:01:23.326651 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" containerName="machine-config-daemon" containerID="cri-o://468d3aa8613beb034776fd6c8b08e9b13031340dca2680a10051a40fbbb92cf2" gracePeriod=600 Oct 05 07:01:24 crc kubenswrapper[4846]: I1005 07:01:24.017110 4846 generic.go:334] "Generic (PLEG): container finished" podID="40d018c5-6953-435a-8a65-b52745cb42c8" containerID="12744f88d32b07359ec7339b76aacc3e84e6c5126b57ab42b396c583c94579e9" exitCode=0 Oct 05 07:01:24 crc kubenswrapper[4846]: I1005 07:01:24.018310 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xvxxx" event={"ID":"40d018c5-6953-435a-8a65-b52745cb42c8","Type":"ContainerDied","Data":"12744f88d32b07359ec7339b76aacc3e84e6c5126b57ab42b396c583c94579e9"} Oct 05 07:01:24 crc kubenswrapper[4846]: I1005 07:01:24.018452 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xvxxx" event={"ID":"40d018c5-6953-435a-8a65-b52745cb42c8","Type":"ContainerStarted","Data":"c4d6f5285ed6f8e2d48ec982702f9ea994135810022e6cacc109b6def07b12dc"} Oct 05 07:01:24 crc kubenswrapper[4846]: I1005 07:01:24.022344 4846 generic.go:334] "Generic (PLEG): container finished" podID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" containerID="468d3aa8613beb034776fd6c8b08e9b13031340dca2680a10051a40fbbb92cf2" exitCode=0 Oct 05 07:01:24 crc kubenswrapper[4846]: I1005 07:01:24.022437 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" event={"ID":"ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7","Type":"ContainerDied","Data":"468d3aa8613beb034776fd6c8b08e9b13031340dca2680a10051a40fbbb92cf2"} Oct 05 07:01:24 crc kubenswrapper[4846]: I1005 07:01:24.022481 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" event={"ID":"ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7","Type":"ContainerStarted","Data":"ec13db87cb5ffec077d726e058e8274c67177ec66498d56f437d83d9bfbe9292"} Oct 05 07:01:24 crc kubenswrapper[4846]: I1005 07:01:24.022503 4846 scope.go:117] "RemoveContainer" containerID="3b7026879bf093ff49269f1b47a66adf5f31a2fca4f61fcc9b2ba0d01156c507" Oct 05 07:01:24 crc kubenswrapper[4846]: I1005 07:01:24.026738 4846 generic.go:334] "Generic (PLEG): container finished" podID="18f66388-66f4-4457-877f-d47dff48bb98" containerID="fea9f2e7cf9f5fef5d3106b239da3612bc71da234aaa82306792c4d0608ad2ab" exitCode=0 Oct 05 07:01:24 crc kubenswrapper[4846]: I1005 07:01:24.026772 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2xndcw" event={"ID":"18f66388-66f4-4457-877f-d47dff48bb98","Type":"ContainerDied","Data":"fea9f2e7cf9f5fef5d3106b239da3612bc71da234aaa82306792c4d0608ad2ab"} Oct 05 07:01:25 crc kubenswrapper[4846]: I1005 07:01:25.038782 4846 generic.go:334] "Generic (PLEG): container finished" podID="18f66388-66f4-4457-877f-d47dff48bb98" containerID="c9fa0bcfada048633daba827357b974f05b6fc47058d6ad899b31e4ed8d56bdb" exitCode=0 Oct 05 07:01:25 crc kubenswrapper[4846]: I1005 07:01:25.038910 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2xndcw" event={"ID":"18f66388-66f4-4457-877f-d47dff48bb98","Type":"ContainerDied","Data":"c9fa0bcfada048633daba827357b974f05b6fc47058d6ad899b31e4ed8d56bdb"} Oct 05 07:01:25 crc kubenswrapper[4846]: I1005 07:01:25.045890 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xvxxx" event={"ID":"40d018c5-6953-435a-8a65-b52745cb42c8","Type":"ContainerStarted","Data":"7d04353732a6a485d02b925fe8f14f68fb04cfbde98beaf08f5118593e23db27"} Oct 05 07:01:26 crc kubenswrapper[4846]: I1005 07:01:26.057135 4846 generic.go:334] "Generic (PLEG): container finished" podID="40d018c5-6953-435a-8a65-b52745cb42c8" containerID="7d04353732a6a485d02b925fe8f14f68fb04cfbde98beaf08f5118593e23db27" exitCode=0 Oct 05 07:01:26 crc kubenswrapper[4846]: I1005 07:01:26.057361 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xvxxx" event={"ID":"40d018c5-6953-435a-8a65-b52745cb42c8","Type":"ContainerDied","Data":"7d04353732a6a485d02b925fe8f14f68fb04cfbde98beaf08f5118593e23db27"} Oct 05 07:01:26 crc kubenswrapper[4846]: I1005 07:01:26.452072 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2xndcw" Oct 05 07:01:26 crc kubenswrapper[4846]: I1005 07:01:26.587784 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/18f66388-66f4-4457-877f-d47dff48bb98-util\") pod \"18f66388-66f4-4457-877f-d47dff48bb98\" (UID: \"18f66388-66f4-4457-877f-d47dff48bb98\") " Oct 05 07:01:26 crc kubenswrapper[4846]: I1005 07:01:26.587963 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7vnhn\" (UniqueName: \"kubernetes.io/projected/18f66388-66f4-4457-877f-d47dff48bb98-kube-api-access-7vnhn\") pod \"18f66388-66f4-4457-877f-d47dff48bb98\" (UID: \"18f66388-66f4-4457-877f-d47dff48bb98\") " Oct 05 07:01:26 crc kubenswrapper[4846]: I1005 07:01:26.588026 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/18f66388-66f4-4457-877f-d47dff48bb98-bundle\") pod \"18f66388-66f4-4457-877f-d47dff48bb98\" (UID: \"18f66388-66f4-4457-877f-d47dff48bb98\") " Oct 05 07:01:26 crc kubenswrapper[4846]: I1005 07:01:26.589521 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/18f66388-66f4-4457-877f-d47dff48bb98-bundle" (OuterVolumeSpecName: "bundle") pod "18f66388-66f4-4457-877f-d47dff48bb98" (UID: "18f66388-66f4-4457-877f-d47dff48bb98"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 07:01:26 crc kubenswrapper[4846]: I1005 07:01:26.597365 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/18f66388-66f4-4457-877f-d47dff48bb98-kube-api-access-7vnhn" (OuterVolumeSpecName: "kube-api-access-7vnhn") pod "18f66388-66f4-4457-877f-d47dff48bb98" (UID: "18f66388-66f4-4457-877f-d47dff48bb98"). InnerVolumeSpecName "kube-api-access-7vnhn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:01:26 crc kubenswrapper[4846]: I1005 07:01:26.690352 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7vnhn\" (UniqueName: \"kubernetes.io/projected/18f66388-66f4-4457-877f-d47dff48bb98-kube-api-access-7vnhn\") on node \"crc\" DevicePath \"\"" Oct 05 07:01:26 crc kubenswrapper[4846]: I1005 07:01:26.690395 4846 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/18f66388-66f4-4457-877f-d47dff48bb98-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 07:01:26 crc kubenswrapper[4846]: I1005 07:01:26.705816 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/18f66388-66f4-4457-877f-d47dff48bb98-util" (OuterVolumeSpecName: "util") pod "18f66388-66f4-4457-877f-d47dff48bb98" (UID: "18f66388-66f4-4457-877f-d47dff48bb98"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 07:01:26 crc kubenswrapper[4846]: I1005 07:01:26.791577 4846 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/18f66388-66f4-4457-877f-d47dff48bb98-util\") on node \"crc\" DevicePath \"\"" Oct 05 07:01:27 crc kubenswrapper[4846]: I1005 07:01:27.065598 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2xndcw" event={"ID":"18f66388-66f4-4457-877f-d47dff48bb98","Type":"ContainerDied","Data":"93dac73088eed762eb85193d46d51bae777e77e1cedf53afbc3cd2d76f954f58"} Oct 05 07:01:27 crc kubenswrapper[4846]: I1005 07:01:27.066108 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="93dac73088eed762eb85193d46d51bae777e77e1cedf53afbc3cd2d76f954f58" Oct 05 07:01:27 crc kubenswrapper[4846]: I1005 07:01:27.065790 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2xndcw" Oct 05 07:01:27 crc kubenswrapper[4846]: I1005 07:01:27.078270 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xvxxx" event={"ID":"40d018c5-6953-435a-8a65-b52745cb42c8","Type":"ContainerStarted","Data":"e48e513daf030e5c93b40b8a7e1535f60313e848ddde5737bce207fa69ec2bb5"} Oct 05 07:01:27 crc kubenswrapper[4846]: I1005 07:01:27.101377 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-xvxxx" podStartSLOduration=2.625233658 podStartE2EDuration="5.101351578s" podCreationTimestamp="2025-10-05 07:01:22 +0000 UTC" firstStartedPulling="2025-10-05 07:01:24.02086031 +0000 UTC m=+806.261713085" lastFinishedPulling="2025-10-05 07:01:26.49697819 +0000 UTC m=+808.737831005" observedRunningTime="2025-10-05 07:01:27.097770023 +0000 UTC m=+809.338622848" watchObservedRunningTime="2025-10-05 07:01:27.101351578 +0000 UTC m=+809.342204353" Oct 05 07:01:31 crc kubenswrapper[4846]: I1005 07:01:31.747283 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-vfdgn"] Oct 05 07:01:31 crc kubenswrapper[4846]: E1005 07:01:31.748606 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="18f66388-66f4-4457-877f-d47dff48bb98" containerName="pull" Oct 05 07:01:31 crc kubenswrapper[4846]: I1005 07:01:31.748622 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="18f66388-66f4-4457-877f-d47dff48bb98" containerName="pull" Oct 05 07:01:31 crc kubenswrapper[4846]: E1005 07:01:31.748649 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="18f66388-66f4-4457-877f-d47dff48bb98" containerName="extract" Oct 05 07:01:31 crc kubenswrapper[4846]: I1005 07:01:31.748655 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="18f66388-66f4-4457-877f-d47dff48bb98" containerName="extract" Oct 05 07:01:31 crc kubenswrapper[4846]: E1005 07:01:31.748663 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="18f66388-66f4-4457-877f-d47dff48bb98" containerName="util" Oct 05 07:01:31 crc kubenswrapper[4846]: I1005 07:01:31.748670 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="18f66388-66f4-4457-877f-d47dff48bb98" containerName="util" Oct 05 07:01:31 crc kubenswrapper[4846]: I1005 07:01:31.748870 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="18f66388-66f4-4457-877f-d47dff48bb98" containerName="extract" Oct 05 07:01:31 crc kubenswrapper[4846]: I1005 07:01:31.751153 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vfdgn" Oct 05 07:01:31 crc kubenswrapper[4846]: I1005 07:01:31.774452 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/86f56e04-eb7d-4fdd-9174-e2633483e5b0-catalog-content\") pod \"community-operators-vfdgn\" (UID: \"86f56e04-eb7d-4fdd-9174-e2633483e5b0\") " pod="openshift-marketplace/community-operators-vfdgn" Oct 05 07:01:31 crc kubenswrapper[4846]: I1005 07:01:31.774508 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5p6bd\" (UniqueName: \"kubernetes.io/projected/86f56e04-eb7d-4fdd-9174-e2633483e5b0-kube-api-access-5p6bd\") pod \"community-operators-vfdgn\" (UID: \"86f56e04-eb7d-4fdd-9174-e2633483e5b0\") " pod="openshift-marketplace/community-operators-vfdgn" Oct 05 07:01:31 crc kubenswrapper[4846]: I1005 07:01:31.774550 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/86f56e04-eb7d-4fdd-9174-e2633483e5b0-utilities\") pod \"community-operators-vfdgn\" (UID: \"86f56e04-eb7d-4fdd-9174-e2633483e5b0\") " pod="openshift-marketplace/community-operators-vfdgn" Oct 05 07:01:31 crc kubenswrapper[4846]: I1005 07:01:31.797357 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-vfdgn"] Oct 05 07:01:31 crc kubenswrapper[4846]: I1005 07:01:31.876221 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/86f56e04-eb7d-4fdd-9174-e2633483e5b0-catalog-content\") pod \"community-operators-vfdgn\" (UID: \"86f56e04-eb7d-4fdd-9174-e2633483e5b0\") " pod="openshift-marketplace/community-operators-vfdgn" Oct 05 07:01:31 crc kubenswrapper[4846]: I1005 07:01:31.876286 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5p6bd\" (UniqueName: \"kubernetes.io/projected/86f56e04-eb7d-4fdd-9174-e2633483e5b0-kube-api-access-5p6bd\") pod \"community-operators-vfdgn\" (UID: \"86f56e04-eb7d-4fdd-9174-e2633483e5b0\") " pod="openshift-marketplace/community-operators-vfdgn" Oct 05 07:01:31 crc kubenswrapper[4846]: I1005 07:01:31.876330 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/86f56e04-eb7d-4fdd-9174-e2633483e5b0-utilities\") pod \"community-operators-vfdgn\" (UID: \"86f56e04-eb7d-4fdd-9174-e2633483e5b0\") " pod="openshift-marketplace/community-operators-vfdgn" Oct 05 07:01:31 crc kubenswrapper[4846]: I1005 07:01:31.877036 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/86f56e04-eb7d-4fdd-9174-e2633483e5b0-utilities\") pod \"community-operators-vfdgn\" (UID: \"86f56e04-eb7d-4fdd-9174-e2633483e5b0\") " pod="openshift-marketplace/community-operators-vfdgn" Oct 05 07:01:31 crc kubenswrapper[4846]: I1005 07:01:31.877407 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/86f56e04-eb7d-4fdd-9174-e2633483e5b0-catalog-content\") pod \"community-operators-vfdgn\" (UID: \"86f56e04-eb7d-4fdd-9174-e2633483e5b0\") " pod="openshift-marketplace/community-operators-vfdgn" Oct 05 07:01:31 crc kubenswrapper[4846]: I1005 07:01:31.899294 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5p6bd\" (UniqueName: \"kubernetes.io/projected/86f56e04-eb7d-4fdd-9174-e2633483e5b0-kube-api-access-5p6bd\") pod \"community-operators-vfdgn\" (UID: \"86f56e04-eb7d-4fdd-9174-e2633483e5b0\") " pod="openshift-marketplace/community-operators-vfdgn" Oct 05 07:01:32 crc kubenswrapper[4846]: I1005 07:01:32.078319 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vfdgn" Oct 05 07:01:32 crc kubenswrapper[4846]: I1005 07:01:32.537095 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-vfdgn"] Oct 05 07:01:32 crc kubenswrapper[4846]: I1005 07:01:32.684727 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-xvxxx" Oct 05 07:01:32 crc kubenswrapper[4846]: I1005 07:01:32.685265 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-xvxxx" Oct 05 07:01:32 crc kubenswrapper[4846]: I1005 07:01:32.750435 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-xvxxx" Oct 05 07:01:33 crc kubenswrapper[4846]: I1005 07:01:33.119998 4846 generic.go:334] "Generic (PLEG): container finished" podID="86f56e04-eb7d-4fdd-9174-e2633483e5b0" containerID="ba3ceb4071999f854a725c7b1d4d38cb81a40c50eb85165316f07d1476ceeead" exitCode=0 Oct 05 07:01:33 crc kubenswrapper[4846]: I1005 07:01:33.120203 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vfdgn" event={"ID":"86f56e04-eb7d-4fdd-9174-e2633483e5b0","Type":"ContainerDied","Data":"ba3ceb4071999f854a725c7b1d4d38cb81a40c50eb85165316f07d1476ceeead"} Oct 05 07:01:33 crc kubenswrapper[4846]: I1005 07:01:33.120280 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vfdgn" event={"ID":"86f56e04-eb7d-4fdd-9174-e2633483e5b0","Type":"ContainerStarted","Data":"bf65c7b340b933c0d9be843ebcfe4a21c1a60ab31b91b551ab7323d1ffdecd83"} Oct 05 07:01:33 crc kubenswrapper[4846]: I1005 07:01:33.204241 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-xvxxx" Oct 05 07:01:34 crc kubenswrapper[4846]: I1005 07:01:34.128735 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vfdgn" event={"ID":"86f56e04-eb7d-4fdd-9174-e2633483e5b0","Type":"ContainerStarted","Data":"bffddf60c3823ef06c5012f41a1397d317aefec5ba24636261aa2c208512b2cc"} Oct 05 07:01:35 crc kubenswrapper[4846]: I1005 07:01:35.140325 4846 generic.go:334] "Generic (PLEG): container finished" podID="86f56e04-eb7d-4fdd-9174-e2633483e5b0" containerID="bffddf60c3823ef06c5012f41a1397d317aefec5ba24636261aa2c208512b2cc" exitCode=0 Oct 05 07:01:35 crc kubenswrapper[4846]: I1005 07:01:35.140425 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vfdgn" event={"ID":"86f56e04-eb7d-4fdd-9174-e2633483e5b0","Type":"ContainerDied","Data":"bffddf60c3823ef06c5012f41a1397d317aefec5ba24636261aa2c208512b2cc"} Oct 05 07:01:35 crc kubenswrapper[4846]: I1005 07:01:35.835019 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-controller-manager-8fccc95cf-j9fdr"] Oct 05 07:01:35 crc kubenswrapper[4846]: I1005 07:01:35.836097 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-8fccc95cf-j9fdr" Oct 05 07:01:35 crc kubenswrapper[4846]: I1005 07:01:35.839135 4846 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-cert" Oct 05 07:01:35 crc kubenswrapper[4846]: I1005 07:01:35.839833 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"openshift-service-ca.crt" Oct 05 07:01:35 crc kubenswrapper[4846]: I1005 07:01:35.840027 4846 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"manager-account-dockercfg-knxwf" Oct 05 07:01:35 crc kubenswrapper[4846]: I1005 07:01:35.840232 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"kube-root-ca.crt" Oct 05 07:01:35 crc kubenswrapper[4846]: I1005 07:01:35.839839 4846 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-controller-manager-service-cert" Oct 05 07:01:35 crc kubenswrapper[4846]: I1005 07:01:35.862518 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-8fccc95cf-j9fdr"] Oct 05 07:01:35 crc kubenswrapper[4846]: I1005 07:01:35.937959 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6lhx4\" (UniqueName: \"kubernetes.io/projected/cbc67894-51b7-4d7d-92b2-eb8b70a77af5-kube-api-access-6lhx4\") pod \"metallb-operator-controller-manager-8fccc95cf-j9fdr\" (UID: \"cbc67894-51b7-4d7d-92b2-eb8b70a77af5\") " pod="metallb-system/metallb-operator-controller-manager-8fccc95cf-j9fdr" Oct 05 07:01:35 crc kubenswrapper[4846]: I1005 07:01:35.938065 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/cbc67894-51b7-4d7d-92b2-eb8b70a77af5-webhook-cert\") pod \"metallb-operator-controller-manager-8fccc95cf-j9fdr\" (UID: \"cbc67894-51b7-4d7d-92b2-eb8b70a77af5\") " pod="metallb-system/metallb-operator-controller-manager-8fccc95cf-j9fdr" Oct 05 07:01:35 crc kubenswrapper[4846]: I1005 07:01:35.938134 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/cbc67894-51b7-4d7d-92b2-eb8b70a77af5-apiservice-cert\") pod \"metallb-operator-controller-manager-8fccc95cf-j9fdr\" (UID: \"cbc67894-51b7-4d7d-92b2-eb8b70a77af5\") " pod="metallb-system/metallb-operator-controller-manager-8fccc95cf-j9fdr" Oct 05 07:01:36 crc kubenswrapper[4846]: I1005 07:01:36.040018 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6lhx4\" (UniqueName: \"kubernetes.io/projected/cbc67894-51b7-4d7d-92b2-eb8b70a77af5-kube-api-access-6lhx4\") pod \"metallb-operator-controller-manager-8fccc95cf-j9fdr\" (UID: \"cbc67894-51b7-4d7d-92b2-eb8b70a77af5\") " pod="metallb-system/metallb-operator-controller-manager-8fccc95cf-j9fdr" Oct 05 07:01:36 crc kubenswrapper[4846]: I1005 07:01:36.040117 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/cbc67894-51b7-4d7d-92b2-eb8b70a77af5-webhook-cert\") pod \"metallb-operator-controller-manager-8fccc95cf-j9fdr\" (UID: \"cbc67894-51b7-4d7d-92b2-eb8b70a77af5\") " pod="metallb-system/metallb-operator-controller-manager-8fccc95cf-j9fdr" Oct 05 07:01:36 crc kubenswrapper[4846]: I1005 07:01:36.040198 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/cbc67894-51b7-4d7d-92b2-eb8b70a77af5-apiservice-cert\") pod \"metallb-operator-controller-manager-8fccc95cf-j9fdr\" (UID: \"cbc67894-51b7-4d7d-92b2-eb8b70a77af5\") " pod="metallb-system/metallb-operator-controller-manager-8fccc95cf-j9fdr" Oct 05 07:01:36 crc kubenswrapper[4846]: I1005 07:01:36.049312 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/cbc67894-51b7-4d7d-92b2-eb8b70a77af5-webhook-cert\") pod \"metallb-operator-controller-manager-8fccc95cf-j9fdr\" (UID: \"cbc67894-51b7-4d7d-92b2-eb8b70a77af5\") " pod="metallb-system/metallb-operator-controller-manager-8fccc95cf-j9fdr" Oct 05 07:01:36 crc kubenswrapper[4846]: I1005 07:01:36.054606 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/cbc67894-51b7-4d7d-92b2-eb8b70a77af5-apiservice-cert\") pod \"metallb-operator-controller-manager-8fccc95cf-j9fdr\" (UID: \"cbc67894-51b7-4d7d-92b2-eb8b70a77af5\") " pod="metallb-system/metallb-operator-controller-manager-8fccc95cf-j9fdr" Oct 05 07:01:36 crc kubenswrapper[4846]: I1005 07:01:36.060030 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6lhx4\" (UniqueName: \"kubernetes.io/projected/cbc67894-51b7-4d7d-92b2-eb8b70a77af5-kube-api-access-6lhx4\") pod \"metallb-operator-controller-manager-8fccc95cf-j9fdr\" (UID: \"cbc67894-51b7-4d7d-92b2-eb8b70a77af5\") " pod="metallb-system/metallb-operator-controller-manager-8fccc95cf-j9fdr" Oct 05 07:01:36 crc kubenswrapper[4846]: I1005 07:01:36.153889 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-8fccc95cf-j9fdr" Oct 05 07:01:36 crc kubenswrapper[4846]: I1005 07:01:36.183918 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-webhook-server-7bd844db7f-6hpbl"] Oct 05 07:01:36 crc kubenswrapper[4846]: I1005 07:01:36.184980 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-7bd844db7f-6hpbl" Oct 05 07:01:36 crc kubenswrapper[4846]: I1005 07:01:36.192272 4846 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-service-cert" Oct 05 07:01:36 crc kubenswrapper[4846]: I1005 07:01:36.192523 4846 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-dockercfg-xr7fw" Oct 05 07:01:36 crc kubenswrapper[4846]: I1005 07:01:36.192733 4846 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Oct 05 07:01:36 crc kubenswrapper[4846]: I1005 07:01:36.205543 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-7bd844db7f-6hpbl"] Oct 05 07:01:36 crc kubenswrapper[4846]: I1005 07:01:36.246086 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/15c86c60-ec26-4e57-bb15-cd3bc1843e80-webhook-cert\") pod \"metallb-operator-webhook-server-7bd844db7f-6hpbl\" (UID: \"15c86c60-ec26-4e57-bb15-cd3bc1843e80\") " pod="metallb-system/metallb-operator-webhook-server-7bd844db7f-6hpbl" Oct 05 07:01:36 crc kubenswrapper[4846]: I1005 07:01:36.246134 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-92qgl\" (UniqueName: \"kubernetes.io/projected/15c86c60-ec26-4e57-bb15-cd3bc1843e80-kube-api-access-92qgl\") pod \"metallb-operator-webhook-server-7bd844db7f-6hpbl\" (UID: \"15c86c60-ec26-4e57-bb15-cd3bc1843e80\") " pod="metallb-system/metallb-operator-webhook-server-7bd844db7f-6hpbl" Oct 05 07:01:36 crc kubenswrapper[4846]: I1005 07:01:36.246192 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/15c86c60-ec26-4e57-bb15-cd3bc1843e80-apiservice-cert\") pod \"metallb-operator-webhook-server-7bd844db7f-6hpbl\" (UID: \"15c86c60-ec26-4e57-bb15-cd3bc1843e80\") " pod="metallb-system/metallb-operator-webhook-server-7bd844db7f-6hpbl" Oct 05 07:01:36 crc kubenswrapper[4846]: I1005 07:01:36.347157 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/15c86c60-ec26-4e57-bb15-cd3bc1843e80-webhook-cert\") pod \"metallb-operator-webhook-server-7bd844db7f-6hpbl\" (UID: \"15c86c60-ec26-4e57-bb15-cd3bc1843e80\") " pod="metallb-system/metallb-operator-webhook-server-7bd844db7f-6hpbl" Oct 05 07:01:36 crc kubenswrapper[4846]: I1005 07:01:36.348437 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-92qgl\" (UniqueName: \"kubernetes.io/projected/15c86c60-ec26-4e57-bb15-cd3bc1843e80-kube-api-access-92qgl\") pod \"metallb-operator-webhook-server-7bd844db7f-6hpbl\" (UID: \"15c86c60-ec26-4e57-bb15-cd3bc1843e80\") " pod="metallb-system/metallb-operator-webhook-server-7bd844db7f-6hpbl" Oct 05 07:01:36 crc kubenswrapper[4846]: I1005 07:01:36.348579 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/15c86c60-ec26-4e57-bb15-cd3bc1843e80-apiservice-cert\") pod \"metallb-operator-webhook-server-7bd844db7f-6hpbl\" (UID: \"15c86c60-ec26-4e57-bb15-cd3bc1843e80\") " pod="metallb-system/metallb-operator-webhook-server-7bd844db7f-6hpbl" Oct 05 07:01:36 crc kubenswrapper[4846]: I1005 07:01:36.356327 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/15c86c60-ec26-4e57-bb15-cd3bc1843e80-apiservice-cert\") pod \"metallb-operator-webhook-server-7bd844db7f-6hpbl\" (UID: \"15c86c60-ec26-4e57-bb15-cd3bc1843e80\") " pod="metallb-system/metallb-operator-webhook-server-7bd844db7f-6hpbl" Oct 05 07:01:36 crc kubenswrapper[4846]: I1005 07:01:36.363447 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/15c86c60-ec26-4e57-bb15-cd3bc1843e80-webhook-cert\") pod \"metallb-operator-webhook-server-7bd844db7f-6hpbl\" (UID: \"15c86c60-ec26-4e57-bb15-cd3bc1843e80\") " pod="metallb-system/metallb-operator-webhook-server-7bd844db7f-6hpbl" Oct 05 07:01:36 crc kubenswrapper[4846]: I1005 07:01:36.374656 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-92qgl\" (UniqueName: \"kubernetes.io/projected/15c86c60-ec26-4e57-bb15-cd3bc1843e80-kube-api-access-92qgl\") pod \"metallb-operator-webhook-server-7bd844db7f-6hpbl\" (UID: \"15c86c60-ec26-4e57-bb15-cd3bc1843e80\") " pod="metallb-system/metallb-operator-webhook-server-7bd844db7f-6hpbl" Oct 05 07:01:36 crc kubenswrapper[4846]: I1005 07:01:36.541668 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-xvxxx"] Oct 05 07:01:36 crc kubenswrapper[4846]: I1005 07:01:36.541969 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-xvxxx" podUID="40d018c5-6953-435a-8a65-b52745cb42c8" containerName="registry-server" containerID="cri-o://e48e513daf030e5c93b40b8a7e1535f60313e848ddde5737bce207fa69ec2bb5" gracePeriod=2 Oct 05 07:01:36 crc kubenswrapper[4846]: I1005 07:01:36.557884 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-7bd844db7f-6hpbl" Oct 05 07:01:36 crc kubenswrapper[4846]: I1005 07:01:36.569169 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-8fccc95cf-j9fdr"] Oct 05 07:01:36 crc kubenswrapper[4846]: W1005 07:01:36.576533 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcbc67894_51b7_4d7d_92b2_eb8b70a77af5.slice/crio-548d10f6683598de94faf4dcb30e3b040c14c0ded9c048cd5367aaddd396dfc4 WatchSource:0}: Error finding container 548d10f6683598de94faf4dcb30e3b040c14c0ded9c048cd5367aaddd396dfc4: Status 404 returned error can't find the container with id 548d10f6683598de94faf4dcb30e3b040c14c0ded9c048cd5367aaddd396dfc4 Oct 05 07:01:37 crc kubenswrapper[4846]: I1005 07:01:37.038772 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-7bd844db7f-6hpbl"] Oct 05 07:01:37 crc kubenswrapper[4846]: W1005 07:01:37.078209 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod15c86c60_ec26_4e57_bb15_cd3bc1843e80.slice/crio-875a03c9828a52f599302916f72d5e49126676c4ae1764907530deb37f9f5472 WatchSource:0}: Error finding container 875a03c9828a52f599302916f72d5e49126676c4ae1764907530deb37f9f5472: Status 404 returned error can't find the container with id 875a03c9828a52f599302916f72d5e49126676c4ae1764907530deb37f9f5472 Oct 05 07:01:37 crc kubenswrapper[4846]: I1005 07:01:37.156442 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-8fccc95cf-j9fdr" event={"ID":"cbc67894-51b7-4d7d-92b2-eb8b70a77af5","Type":"ContainerStarted","Data":"548d10f6683598de94faf4dcb30e3b040c14c0ded9c048cd5367aaddd396dfc4"} Oct 05 07:01:37 crc kubenswrapper[4846]: I1005 07:01:37.159606 4846 generic.go:334] "Generic (PLEG): container finished" podID="40d018c5-6953-435a-8a65-b52745cb42c8" containerID="e48e513daf030e5c93b40b8a7e1535f60313e848ddde5737bce207fa69ec2bb5" exitCode=0 Oct 05 07:01:37 crc kubenswrapper[4846]: I1005 07:01:37.159643 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xvxxx" event={"ID":"40d018c5-6953-435a-8a65-b52745cb42c8","Type":"ContainerDied","Data":"e48e513daf030e5c93b40b8a7e1535f60313e848ddde5737bce207fa69ec2bb5"} Oct 05 07:01:37 crc kubenswrapper[4846]: I1005 07:01:37.161049 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-7bd844db7f-6hpbl" event={"ID":"15c86c60-ec26-4e57-bb15-cd3bc1843e80","Type":"ContainerStarted","Data":"875a03c9828a52f599302916f72d5e49126676c4ae1764907530deb37f9f5472"} Oct 05 07:01:37 crc kubenswrapper[4846]: I1005 07:01:37.163688 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vfdgn" event={"ID":"86f56e04-eb7d-4fdd-9174-e2633483e5b0","Type":"ContainerStarted","Data":"92b5624cdded2cdaa847fd6f760a07ad503189d53cd57fd7e278d2c15551bf78"} Oct 05 07:01:37 crc kubenswrapper[4846]: I1005 07:01:37.186518 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-vfdgn" podStartSLOduration=3.19944847 podStartE2EDuration="6.18649147s" podCreationTimestamp="2025-10-05 07:01:31 +0000 UTC" firstStartedPulling="2025-10-05 07:01:33.123498519 +0000 UTC m=+815.364351324" lastFinishedPulling="2025-10-05 07:01:36.110541549 +0000 UTC m=+818.351394324" observedRunningTime="2025-10-05 07:01:37.182827142 +0000 UTC m=+819.423679917" watchObservedRunningTime="2025-10-05 07:01:37.18649147 +0000 UTC m=+819.427344255" Oct 05 07:01:37 crc kubenswrapper[4846]: I1005 07:01:37.722156 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xvxxx" Oct 05 07:01:37 crc kubenswrapper[4846]: I1005 07:01:37.780731 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/40d018c5-6953-435a-8a65-b52745cb42c8-catalog-content\") pod \"40d018c5-6953-435a-8a65-b52745cb42c8\" (UID: \"40d018c5-6953-435a-8a65-b52745cb42c8\") " Oct 05 07:01:37 crc kubenswrapper[4846]: I1005 07:01:37.780825 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-789l5\" (UniqueName: \"kubernetes.io/projected/40d018c5-6953-435a-8a65-b52745cb42c8-kube-api-access-789l5\") pod \"40d018c5-6953-435a-8a65-b52745cb42c8\" (UID: \"40d018c5-6953-435a-8a65-b52745cb42c8\") " Oct 05 07:01:37 crc kubenswrapper[4846]: I1005 07:01:37.780860 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/40d018c5-6953-435a-8a65-b52745cb42c8-utilities\") pod \"40d018c5-6953-435a-8a65-b52745cb42c8\" (UID: \"40d018c5-6953-435a-8a65-b52745cb42c8\") " Oct 05 07:01:37 crc kubenswrapper[4846]: I1005 07:01:37.782017 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/40d018c5-6953-435a-8a65-b52745cb42c8-utilities" (OuterVolumeSpecName: "utilities") pod "40d018c5-6953-435a-8a65-b52745cb42c8" (UID: "40d018c5-6953-435a-8a65-b52745cb42c8"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 07:01:37 crc kubenswrapper[4846]: I1005 07:01:37.788543 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/40d018c5-6953-435a-8a65-b52745cb42c8-kube-api-access-789l5" (OuterVolumeSpecName: "kube-api-access-789l5") pod "40d018c5-6953-435a-8a65-b52745cb42c8" (UID: "40d018c5-6953-435a-8a65-b52745cb42c8"). InnerVolumeSpecName "kube-api-access-789l5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:01:37 crc kubenswrapper[4846]: I1005 07:01:37.864726 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/40d018c5-6953-435a-8a65-b52745cb42c8-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "40d018c5-6953-435a-8a65-b52745cb42c8" (UID: "40d018c5-6953-435a-8a65-b52745cb42c8"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 07:01:37 crc kubenswrapper[4846]: I1005 07:01:37.881794 4846 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/40d018c5-6953-435a-8a65-b52745cb42c8-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 05 07:01:37 crc kubenswrapper[4846]: I1005 07:01:37.881837 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-789l5\" (UniqueName: \"kubernetes.io/projected/40d018c5-6953-435a-8a65-b52745cb42c8-kube-api-access-789l5\") on node \"crc\" DevicePath \"\"" Oct 05 07:01:37 crc kubenswrapper[4846]: I1005 07:01:37.881852 4846 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/40d018c5-6953-435a-8a65-b52745cb42c8-utilities\") on node \"crc\" DevicePath \"\"" Oct 05 07:01:38 crc kubenswrapper[4846]: I1005 07:01:38.172959 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xvxxx" Oct 05 07:01:38 crc kubenswrapper[4846]: I1005 07:01:38.172966 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xvxxx" event={"ID":"40d018c5-6953-435a-8a65-b52745cb42c8","Type":"ContainerDied","Data":"c4d6f5285ed6f8e2d48ec982702f9ea994135810022e6cacc109b6def07b12dc"} Oct 05 07:01:38 crc kubenswrapper[4846]: I1005 07:01:38.173067 4846 scope.go:117] "RemoveContainer" containerID="e48e513daf030e5c93b40b8a7e1535f60313e848ddde5737bce207fa69ec2bb5" Oct 05 07:01:38 crc kubenswrapper[4846]: I1005 07:01:38.204029 4846 scope.go:117] "RemoveContainer" containerID="7d04353732a6a485d02b925fe8f14f68fb04cfbde98beaf08f5118593e23db27" Oct 05 07:01:38 crc kubenswrapper[4846]: I1005 07:01:38.205515 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-xvxxx"] Oct 05 07:01:38 crc kubenswrapper[4846]: I1005 07:01:38.215686 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-xvxxx"] Oct 05 07:01:38 crc kubenswrapper[4846]: I1005 07:01:38.232856 4846 scope.go:117] "RemoveContainer" containerID="12744f88d32b07359ec7339b76aacc3e84e6c5126b57ab42b396c583c94579e9" Oct 05 07:01:38 crc kubenswrapper[4846]: I1005 07:01:38.514143 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="40d018c5-6953-435a-8a65-b52745cb42c8" path="/var/lib/kubelet/pods/40d018c5-6953-435a-8a65-b52745cb42c8/volumes" Oct 05 07:01:42 crc kubenswrapper[4846]: I1005 07:01:42.078625 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-vfdgn" Oct 05 07:01:42 crc kubenswrapper[4846]: I1005 07:01:42.079824 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-vfdgn" Oct 05 07:01:42 crc kubenswrapper[4846]: I1005 07:01:42.178587 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-vfdgn" Oct 05 07:01:42 crc kubenswrapper[4846]: I1005 07:01:42.206226 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-8fccc95cf-j9fdr" event={"ID":"cbc67894-51b7-4d7d-92b2-eb8b70a77af5","Type":"ContainerStarted","Data":"1ea67eced024857a5bed7cb6ce94bc43d22c9dc2b6efff98969eff120d311374"} Oct 05 07:01:42 crc kubenswrapper[4846]: I1005 07:01:42.207065 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-8fccc95cf-j9fdr" Oct 05 07:01:42 crc kubenswrapper[4846]: I1005 07:01:42.209776 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-7bd844db7f-6hpbl" event={"ID":"15c86c60-ec26-4e57-bb15-cd3bc1843e80","Type":"ContainerStarted","Data":"741d2769a4119e74b93af64c6fe9b9ad36e7f694aaa1efeb5a966b58682b59bb"} Oct 05 07:01:42 crc kubenswrapper[4846]: I1005 07:01:42.209802 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-7bd844db7f-6hpbl" Oct 05 07:01:42 crc kubenswrapper[4846]: I1005 07:01:42.274942 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-webhook-server-7bd844db7f-6hpbl" podStartSLOduration=1.472340877 podStartE2EDuration="6.274922275s" podCreationTimestamp="2025-10-05 07:01:36 +0000 UTC" firstStartedPulling="2025-10-05 07:01:37.081639718 +0000 UTC m=+819.322492503" lastFinishedPulling="2025-10-05 07:01:41.884221126 +0000 UTC m=+824.125073901" observedRunningTime="2025-10-05 07:01:42.270736094 +0000 UTC m=+824.511588869" watchObservedRunningTime="2025-10-05 07:01:42.274922275 +0000 UTC m=+824.515775050" Oct 05 07:01:42 crc kubenswrapper[4846]: I1005 07:01:42.313384 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-vfdgn" Oct 05 07:01:42 crc kubenswrapper[4846]: I1005 07:01:42.372529 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-controller-manager-8fccc95cf-j9fdr" podStartSLOduration=2.086842926 podStartE2EDuration="7.372503053s" podCreationTimestamp="2025-10-05 07:01:35 +0000 UTC" firstStartedPulling="2025-10-05 07:01:36.582244835 +0000 UTC m=+818.823097600" lastFinishedPulling="2025-10-05 07:01:41.867904912 +0000 UTC m=+824.108757727" observedRunningTime="2025-10-05 07:01:42.328701507 +0000 UTC m=+824.569554292" watchObservedRunningTime="2025-10-05 07:01:42.372503053 +0000 UTC m=+824.613355828" Oct 05 07:01:43 crc kubenswrapper[4846]: I1005 07:01:43.949343 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-wqmsh"] Oct 05 07:01:43 crc kubenswrapper[4846]: E1005 07:01:43.950203 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="40d018c5-6953-435a-8a65-b52745cb42c8" containerName="registry-server" Oct 05 07:01:43 crc kubenswrapper[4846]: I1005 07:01:43.950219 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="40d018c5-6953-435a-8a65-b52745cb42c8" containerName="registry-server" Oct 05 07:01:43 crc kubenswrapper[4846]: E1005 07:01:43.950250 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="40d018c5-6953-435a-8a65-b52745cb42c8" containerName="extract-utilities" Oct 05 07:01:43 crc kubenswrapper[4846]: I1005 07:01:43.950260 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="40d018c5-6953-435a-8a65-b52745cb42c8" containerName="extract-utilities" Oct 05 07:01:43 crc kubenswrapper[4846]: E1005 07:01:43.950277 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="40d018c5-6953-435a-8a65-b52745cb42c8" containerName="extract-content" Oct 05 07:01:43 crc kubenswrapper[4846]: I1005 07:01:43.950285 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="40d018c5-6953-435a-8a65-b52745cb42c8" containerName="extract-content" Oct 05 07:01:43 crc kubenswrapper[4846]: I1005 07:01:43.950431 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="40d018c5-6953-435a-8a65-b52745cb42c8" containerName="registry-server" Oct 05 07:01:43 crc kubenswrapper[4846]: I1005 07:01:43.951573 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wqmsh" Oct 05 07:01:43 crc kubenswrapper[4846]: I1005 07:01:43.978125 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bdw6v\" (UniqueName: \"kubernetes.io/projected/ab709817-df14-4fe9-9c58-ae232db7915e-kube-api-access-bdw6v\") pod \"certified-operators-wqmsh\" (UID: \"ab709817-df14-4fe9-9c58-ae232db7915e\") " pod="openshift-marketplace/certified-operators-wqmsh" Oct 05 07:01:43 crc kubenswrapper[4846]: I1005 07:01:43.978285 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ab709817-df14-4fe9-9c58-ae232db7915e-utilities\") pod \"certified-operators-wqmsh\" (UID: \"ab709817-df14-4fe9-9c58-ae232db7915e\") " pod="openshift-marketplace/certified-operators-wqmsh" Oct 05 07:01:43 crc kubenswrapper[4846]: I1005 07:01:43.978373 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ab709817-df14-4fe9-9c58-ae232db7915e-catalog-content\") pod \"certified-operators-wqmsh\" (UID: \"ab709817-df14-4fe9-9c58-ae232db7915e\") " pod="openshift-marketplace/certified-operators-wqmsh" Oct 05 07:01:43 crc kubenswrapper[4846]: I1005 07:01:43.981100 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-wqmsh"] Oct 05 07:01:44 crc kubenswrapper[4846]: I1005 07:01:44.079125 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bdw6v\" (UniqueName: \"kubernetes.io/projected/ab709817-df14-4fe9-9c58-ae232db7915e-kube-api-access-bdw6v\") pod \"certified-operators-wqmsh\" (UID: \"ab709817-df14-4fe9-9c58-ae232db7915e\") " pod="openshift-marketplace/certified-operators-wqmsh" Oct 05 07:01:44 crc kubenswrapper[4846]: I1005 07:01:44.079224 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ab709817-df14-4fe9-9c58-ae232db7915e-utilities\") pod \"certified-operators-wqmsh\" (UID: \"ab709817-df14-4fe9-9c58-ae232db7915e\") " pod="openshift-marketplace/certified-operators-wqmsh" Oct 05 07:01:44 crc kubenswrapper[4846]: I1005 07:01:44.079286 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ab709817-df14-4fe9-9c58-ae232db7915e-catalog-content\") pod \"certified-operators-wqmsh\" (UID: \"ab709817-df14-4fe9-9c58-ae232db7915e\") " pod="openshift-marketplace/certified-operators-wqmsh" Oct 05 07:01:44 crc kubenswrapper[4846]: I1005 07:01:44.079904 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ab709817-df14-4fe9-9c58-ae232db7915e-catalog-content\") pod \"certified-operators-wqmsh\" (UID: \"ab709817-df14-4fe9-9c58-ae232db7915e\") " pod="openshift-marketplace/certified-operators-wqmsh" Oct 05 07:01:44 crc kubenswrapper[4846]: I1005 07:01:44.080025 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ab709817-df14-4fe9-9c58-ae232db7915e-utilities\") pod \"certified-operators-wqmsh\" (UID: \"ab709817-df14-4fe9-9c58-ae232db7915e\") " pod="openshift-marketplace/certified-operators-wqmsh" Oct 05 07:01:44 crc kubenswrapper[4846]: I1005 07:01:44.101050 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bdw6v\" (UniqueName: \"kubernetes.io/projected/ab709817-df14-4fe9-9c58-ae232db7915e-kube-api-access-bdw6v\") pod \"certified-operators-wqmsh\" (UID: \"ab709817-df14-4fe9-9c58-ae232db7915e\") " pod="openshift-marketplace/certified-operators-wqmsh" Oct 05 07:01:44 crc kubenswrapper[4846]: I1005 07:01:44.274480 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wqmsh" Oct 05 07:01:44 crc kubenswrapper[4846]: I1005 07:01:44.815306 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-wqmsh"] Oct 05 07:01:45 crc kubenswrapper[4846]: I1005 07:01:45.232722 4846 generic.go:334] "Generic (PLEG): container finished" podID="ab709817-df14-4fe9-9c58-ae232db7915e" containerID="77afb2db24398a36b3ea5680fca36296a09cb517a70ccc2ad54547441055d8b5" exitCode=0 Oct 05 07:01:45 crc kubenswrapper[4846]: I1005 07:01:45.232774 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wqmsh" event={"ID":"ab709817-df14-4fe9-9c58-ae232db7915e","Type":"ContainerDied","Data":"77afb2db24398a36b3ea5680fca36296a09cb517a70ccc2ad54547441055d8b5"} Oct 05 07:01:45 crc kubenswrapper[4846]: I1005 07:01:45.233165 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wqmsh" event={"ID":"ab709817-df14-4fe9-9c58-ae232db7915e","Type":"ContainerStarted","Data":"a5e3125f2aa7c011ca25e5cae610eae2b02bc0f501b73d297ceeafc08d529985"} Oct 05 07:01:46 crc kubenswrapper[4846]: I1005 07:01:46.131086 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-vfdgn"] Oct 05 07:01:46 crc kubenswrapper[4846]: I1005 07:01:46.131524 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-vfdgn" podUID="86f56e04-eb7d-4fdd-9174-e2633483e5b0" containerName="registry-server" containerID="cri-o://92b5624cdded2cdaa847fd6f760a07ad503189d53cd57fd7e278d2c15551bf78" gracePeriod=2 Oct 05 07:01:46 crc kubenswrapper[4846]: I1005 07:01:46.246336 4846 generic.go:334] "Generic (PLEG): container finished" podID="ab709817-df14-4fe9-9c58-ae232db7915e" containerID="96a3d3a6032af6874a3958b1d2c71dd57661b3b2aec3a53772007afb6600d948" exitCode=0 Oct 05 07:01:46 crc kubenswrapper[4846]: I1005 07:01:46.246408 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wqmsh" event={"ID":"ab709817-df14-4fe9-9c58-ae232db7915e","Type":"ContainerDied","Data":"96a3d3a6032af6874a3958b1d2c71dd57661b3b2aec3a53772007afb6600d948"} Oct 05 07:01:46 crc kubenswrapper[4846]: I1005 07:01:46.665959 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vfdgn" Oct 05 07:01:46 crc kubenswrapper[4846]: I1005 07:01:46.816590 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/86f56e04-eb7d-4fdd-9174-e2633483e5b0-catalog-content\") pod \"86f56e04-eb7d-4fdd-9174-e2633483e5b0\" (UID: \"86f56e04-eb7d-4fdd-9174-e2633483e5b0\") " Oct 05 07:01:46 crc kubenswrapper[4846]: I1005 07:01:46.816727 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5p6bd\" (UniqueName: \"kubernetes.io/projected/86f56e04-eb7d-4fdd-9174-e2633483e5b0-kube-api-access-5p6bd\") pod \"86f56e04-eb7d-4fdd-9174-e2633483e5b0\" (UID: \"86f56e04-eb7d-4fdd-9174-e2633483e5b0\") " Oct 05 07:01:46 crc kubenswrapper[4846]: I1005 07:01:46.816778 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/86f56e04-eb7d-4fdd-9174-e2633483e5b0-utilities\") pod \"86f56e04-eb7d-4fdd-9174-e2633483e5b0\" (UID: \"86f56e04-eb7d-4fdd-9174-e2633483e5b0\") " Oct 05 07:01:46 crc kubenswrapper[4846]: I1005 07:01:46.817972 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/86f56e04-eb7d-4fdd-9174-e2633483e5b0-utilities" (OuterVolumeSpecName: "utilities") pod "86f56e04-eb7d-4fdd-9174-e2633483e5b0" (UID: "86f56e04-eb7d-4fdd-9174-e2633483e5b0"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 07:01:46 crc kubenswrapper[4846]: I1005 07:01:46.826407 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/86f56e04-eb7d-4fdd-9174-e2633483e5b0-kube-api-access-5p6bd" (OuterVolumeSpecName: "kube-api-access-5p6bd") pod "86f56e04-eb7d-4fdd-9174-e2633483e5b0" (UID: "86f56e04-eb7d-4fdd-9174-e2633483e5b0"). InnerVolumeSpecName "kube-api-access-5p6bd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:01:46 crc kubenswrapper[4846]: I1005 07:01:46.887055 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/86f56e04-eb7d-4fdd-9174-e2633483e5b0-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "86f56e04-eb7d-4fdd-9174-e2633483e5b0" (UID: "86f56e04-eb7d-4fdd-9174-e2633483e5b0"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 07:01:46 crc kubenswrapper[4846]: I1005 07:01:46.918660 4846 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/86f56e04-eb7d-4fdd-9174-e2633483e5b0-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 05 07:01:46 crc kubenswrapper[4846]: I1005 07:01:46.918720 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5p6bd\" (UniqueName: \"kubernetes.io/projected/86f56e04-eb7d-4fdd-9174-e2633483e5b0-kube-api-access-5p6bd\") on node \"crc\" DevicePath \"\"" Oct 05 07:01:46 crc kubenswrapper[4846]: I1005 07:01:46.918738 4846 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/86f56e04-eb7d-4fdd-9174-e2633483e5b0-utilities\") on node \"crc\" DevicePath \"\"" Oct 05 07:01:47 crc kubenswrapper[4846]: I1005 07:01:47.261935 4846 generic.go:334] "Generic (PLEG): container finished" podID="86f56e04-eb7d-4fdd-9174-e2633483e5b0" containerID="92b5624cdded2cdaa847fd6f760a07ad503189d53cd57fd7e278d2c15551bf78" exitCode=0 Oct 05 07:01:47 crc kubenswrapper[4846]: I1005 07:01:47.262075 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vfdgn" Oct 05 07:01:47 crc kubenswrapper[4846]: I1005 07:01:47.262733 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vfdgn" event={"ID":"86f56e04-eb7d-4fdd-9174-e2633483e5b0","Type":"ContainerDied","Data":"92b5624cdded2cdaa847fd6f760a07ad503189d53cd57fd7e278d2c15551bf78"} Oct 05 07:01:47 crc kubenswrapper[4846]: I1005 07:01:47.262810 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vfdgn" event={"ID":"86f56e04-eb7d-4fdd-9174-e2633483e5b0","Type":"ContainerDied","Data":"bf65c7b340b933c0d9be843ebcfe4a21c1a60ab31b91b551ab7323d1ffdecd83"} Oct 05 07:01:47 crc kubenswrapper[4846]: I1005 07:01:47.262836 4846 scope.go:117] "RemoveContainer" containerID="92b5624cdded2cdaa847fd6f760a07ad503189d53cd57fd7e278d2c15551bf78" Oct 05 07:01:47 crc kubenswrapper[4846]: I1005 07:01:47.277053 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wqmsh" event={"ID":"ab709817-df14-4fe9-9c58-ae232db7915e","Type":"ContainerStarted","Data":"4d1924bae279f63f5709332d16e9bdef749b98c781e4b15c7408342b5273ef9b"} Oct 05 07:01:47 crc kubenswrapper[4846]: I1005 07:01:47.291103 4846 scope.go:117] "RemoveContainer" containerID="bffddf60c3823ef06c5012f41a1397d317aefec5ba24636261aa2c208512b2cc" Oct 05 07:01:47 crc kubenswrapper[4846]: I1005 07:01:47.296863 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-wqmsh" podStartSLOduration=2.823840453 podStartE2EDuration="4.296845972s" podCreationTimestamp="2025-10-05 07:01:43 +0000 UTC" firstStartedPulling="2025-10-05 07:01:45.234465145 +0000 UTC m=+827.475317920" lastFinishedPulling="2025-10-05 07:01:46.707470644 +0000 UTC m=+828.948323439" observedRunningTime="2025-10-05 07:01:47.292775244 +0000 UTC m=+829.533628029" watchObservedRunningTime="2025-10-05 07:01:47.296845972 +0000 UTC m=+829.537698747" Oct 05 07:01:47 crc kubenswrapper[4846]: I1005 07:01:47.309040 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-vfdgn"] Oct 05 07:01:47 crc kubenswrapper[4846]: I1005 07:01:47.324511 4846 scope.go:117] "RemoveContainer" containerID="ba3ceb4071999f854a725c7b1d4d38cb81a40c50eb85165316f07d1476ceeead" Oct 05 07:01:47 crc kubenswrapper[4846]: I1005 07:01:47.326222 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-vfdgn"] Oct 05 07:01:47 crc kubenswrapper[4846]: I1005 07:01:47.341888 4846 scope.go:117] "RemoveContainer" containerID="92b5624cdded2cdaa847fd6f760a07ad503189d53cd57fd7e278d2c15551bf78" Oct 05 07:01:47 crc kubenswrapper[4846]: E1005 07:01:47.342505 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"92b5624cdded2cdaa847fd6f760a07ad503189d53cd57fd7e278d2c15551bf78\": container with ID starting with 92b5624cdded2cdaa847fd6f760a07ad503189d53cd57fd7e278d2c15551bf78 not found: ID does not exist" containerID="92b5624cdded2cdaa847fd6f760a07ad503189d53cd57fd7e278d2c15551bf78" Oct 05 07:01:47 crc kubenswrapper[4846]: I1005 07:01:47.342555 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"92b5624cdded2cdaa847fd6f760a07ad503189d53cd57fd7e278d2c15551bf78"} err="failed to get container status \"92b5624cdded2cdaa847fd6f760a07ad503189d53cd57fd7e278d2c15551bf78\": rpc error: code = NotFound desc = could not find container \"92b5624cdded2cdaa847fd6f760a07ad503189d53cd57fd7e278d2c15551bf78\": container with ID starting with 92b5624cdded2cdaa847fd6f760a07ad503189d53cd57fd7e278d2c15551bf78 not found: ID does not exist" Oct 05 07:01:47 crc kubenswrapper[4846]: I1005 07:01:47.342613 4846 scope.go:117] "RemoveContainer" containerID="bffddf60c3823ef06c5012f41a1397d317aefec5ba24636261aa2c208512b2cc" Oct 05 07:01:47 crc kubenswrapper[4846]: E1005 07:01:47.343083 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bffddf60c3823ef06c5012f41a1397d317aefec5ba24636261aa2c208512b2cc\": container with ID starting with bffddf60c3823ef06c5012f41a1397d317aefec5ba24636261aa2c208512b2cc not found: ID does not exist" containerID="bffddf60c3823ef06c5012f41a1397d317aefec5ba24636261aa2c208512b2cc" Oct 05 07:01:47 crc kubenswrapper[4846]: I1005 07:01:47.343107 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bffddf60c3823ef06c5012f41a1397d317aefec5ba24636261aa2c208512b2cc"} err="failed to get container status \"bffddf60c3823ef06c5012f41a1397d317aefec5ba24636261aa2c208512b2cc\": rpc error: code = NotFound desc = could not find container \"bffddf60c3823ef06c5012f41a1397d317aefec5ba24636261aa2c208512b2cc\": container with ID starting with bffddf60c3823ef06c5012f41a1397d317aefec5ba24636261aa2c208512b2cc not found: ID does not exist" Oct 05 07:01:47 crc kubenswrapper[4846]: I1005 07:01:47.343122 4846 scope.go:117] "RemoveContainer" containerID="ba3ceb4071999f854a725c7b1d4d38cb81a40c50eb85165316f07d1476ceeead" Oct 05 07:01:47 crc kubenswrapper[4846]: E1005 07:01:47.343447 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ba3ceb4071999f854a725c7b1d4d38cb81a40c50eb85165316f07d1476ceeead\": container with ID starting with ba3ceb4071999f854a725c7b1d4d38cb81a40c50eb85165316f07d1476ceeead not found: ID does not exist" containerID="ba3ceb4071999f854a725c7b1d4d38cb81a40c50eb85165316f07d1476ceeead" Oct 05 07:01:47 crc kubenswrapper[4846]: I1005 07:01:47.343466 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ba3ceb4071999f854a725c7b1d4d38cb81a40c50eb85165316f07d1476ceeead"} err="failed to get container status \"ba3ceb4071999f854a725c7b1d4d38cb81a40c50eb85165316f07d1476ceeead\": rpc error: code = NotFound desc = could not find container \"ba3ceb4071999f854a725c7b1d4d38cb81a40c50eb85165316f07d1476ceeead\": container with ID starting with ba3ceb4071999f854a725c7b1d4d38cb81a40c50eb85165316f07d1476ceeead not found: ID does not exist" Oct 05 07:01:47 crc kubenswrapper[4846]: I1005 07:01:47.544498 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-98wxg"] Oct 05 07:01:47 crc kubenswrapper[4846]: E1005 07:01:47.544852 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="86f56e04-eb7d-4fdd-9174-e2633483e5b0" containerName="extract-content" Oct 05 07:01:47 crc kubenswrapper[4846]: I1005 07:01:47.544870 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="86f56e04-eb7d-4fdd-9174-e2633483e5b0" containerName="extract-content" Oct 05 07:01:47 crc kubenswrapper[4846]: E1005 07:01:47.544894 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="86f56e04-eb7d-4fdd-9174-e2633483e5b0" containerName="registry-server" Oct 05 07:01:47 crc kubenswrapper[4846]: I1005 07:01:47.544909 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="86f56e04-eb7d-4fdd-9174-e2633483e5b0" containerName="registry-server" Oct 05 07:01:47 crc kubenswrapper[4846]: E1005 07:01:47.544931 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="86f56e04-eb7d-4fdd-9174-e2633483e5b0" containerName="extract-utilities" Oct 05 07:01:47 crc kubenswrapper[4846]: I1005 07:01:47.544945 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="86f56e04-eb7d-4fdd-9174-e2633483e5b0" containerName="extract-utilities" Oct 05 07:01:47 crc kubenswrapper[4846]: I1005 07:01:47.545088 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="86f56e04-eb7d-4fdd-9174-e2633483e5b0" containerName="registry-server" Oct 05 07:01:47 crc kubenswrapper[4846]: I1005 07:01:47.546259 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-98wxg" Oct 05 07:01:47 crc kubenswrapper[4846]: I1005 07:01:47.563701 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-98wxg"] Oct 05 07:01:47 crc kubenswrapper[4846]: I1005 07:01:47.731618 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-855lq\" (UniqueName: \"kubernetes.io/projected/0c327613-e7ab-4d2d-882c-fecaf1590c87-kube-api-access-855lq\") pod \"redhat-marketplace-98wxg\" (UID: \"0c327613-e7ab-4d2d-882c-fecaf1590c87\") " pod="openshift-marketplace/redhat-marketplace-98wxg" Oct 05 07:01:47 crc kubenswrapper[4846]: I1005 07:01:47.731725 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0c327613-e7ab-4d2d-882c-fecaf1590c87-utilities\") pod \"redhat-marketplace-98wxg\" (UID: \"0c327613-e7ab-4d2d-882c-fecaf1590c87\") " pod="openshift-marketplace/redhat-marketplace-98wxg" Oct 05 07:01:47 crc kubenswrapper[4846]: I1005 07:01:47.731750 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0c327613-e7ab-4d2d-882c-fecaf1590c87-catalog-content\") pod \"redhat-marketplace-98wxg\" (UID: \"0c327613-e7ab-4d2d-882c-fecaf1590c87\") " pod="openshift-marketplace/redhat-marketplace-98wxg" Oct 05 07:01:47 crc kubenswrapper[4846]: I1005 07:01:47.833672 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-855lq\" (UniqueName: \"kubernetes.io/projected/0c327613-e7ab-4d2d-882c-fecaf1590c87-kube-api-access-855lq\") pod \"redhat-marketplace-98wxg\" (UID: \"0c327613-e7ab-4d2d-882c-fecaf1590c87\") " pod="openshift-marketplace/redhat-marketplace-98wxg" Oct 05 07:01:47 crc kubenswrapper[4846]: I1005 07:01:47.833792 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0c327613-e7ab-4d2d-882c-fecaf1590c87-utilities\") pod \"redhat-marketplace-98wxg\" (UID: \"0c327613-e7ab-4d2d-882c-fecaf1590c87\") " pod="openshift-marketplace/redhat-marketplace-98wxg" Oct 05 07:01:47 crc kubenswrapper[4846]: I1005 07:01:47.833835 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0c327613-e7ab-4d2d-882c-fecaf1590c87-catalog-content\") pod \"redhat-marketplace-98wxg\" (UID: \"0c327613-e7ab-4d2d-882c-fecaf1590c87\") " pod="openshift-marketplace/redhat-marketplace-98wxg" Oct 05 07:01:47 crc kubenswrapper[4846]: I1005 07:01:47.834419 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0c327613-e7ab-4d2d-882c-fecaf1590c87-utilities\") pod \"redhat-marketplace-98wxg\" (UID: \"0c327613-e7ab-4d2d-882c-fecaf1590c87\") " pod="openshift-marketplace/redhat-marketplace-98wxg" Oct 05 07:01:47 crc kubenswrapper[4846]: I1005 07:01:47.834454 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0c327613-e7ab-4d2d-882c-fecaf1590c87-catalog-content\") pod \"redhat-marketplace-98wxg\" (UID: \"0c327613-e7ab-4d2d-882c-fecaf1590c87\") " pod="openshift-marketplace/redhat-marketplace-98wxg" Oct 05 07:01:47 crc kubenswrapper[4846]: I1005 07:01:47.859147 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-855lq\" (UniqueName: \"kubernetes.io/projected/0c327613-e7ab-4d2d-882c-fecaf1590c87-kube-api-access-855lq\") pod \"redhat-marketplace-98wxg\" (UID: \"0c327613-e7ab-4d2d-882c-fecaf1590c87\") " pod="openshift-marketplace/redhat-marketplace-98wxg" Oct 05 07:01:47 crc kubenswrapper[4846]: I1005 07:01:47.864721 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-98wxg" Oct 05 07:01:48 crc kubenswrapper[4846]: I1005 07:01:48.355491 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-98wxg"] Oct 05 07:01:48 crc kubenswrapper[4846]: I1005 07:01:48.508553 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="86f56e04-eb7d-4fdd-9174-e2633483e5b0" path="/var/lib/kubelet/pods/86f56e04-eb7d-4fdd-9174-e2633483e5b0/volumes" Oct 05 07:01:49 crc kubenswrapper[4846]: I1005 07:01:49.294047 4846 generic.go:334] "Generic (PLEG): container finished" podID="0c327613-e7ab-4d2d-882c-fecaf1590c87" containerID="b77a3358a92b865a06be476f33059afa0a1d84281e9360ebd7c9edc5dd0b9f15" exitCode=0 Oct 05 07:01:49 crc kubenswrapper[4846]: I1005 07:01:49.294150 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-98wxg" event={"ID":"0c327613-e7ab-4d2d-882c-fecaf1590c87","Type":"ContainerDied","Data":"b77a3358a92b865a06be476f33059afa0a1d84281e9360ebd7c9edc5dd0b9f15"} Oct 05 07:01:49 crc kubenswrapper[4846]: I1005 07:01:49.294239 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-98wxg" event={"ID":"0c327613-e7ab-4d2d-882c-fecaf1590c87","Type":"ContainerStarted","Data":"8d4baa880a21087ea285ed0d0e3ea948b5e80a6635ec391580cfea21ffebe9cf"} Oct 05 07:01:50 crc kubenswrapper[4846]: I1005 07:01:50.303668 4846 generic.go:334] "Generic (PLEG): container finished" podID="0c327613-e7ab-4d2d-882c-fecaf1590c87" containerID="7d32d057852ee0fb06261062ba698899df977605fd6c8eae4b1e82369694a7e6" exitCode=0 Oct 05 07:01:50 crc kubenswrapper[4846]: I1005 07:01:50.303791 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-98wxg" event={"ID":"0c327613-e7ab-4d2d-882c-fecaf1590c87","Type":"ContainerDied","Data":"7d32d057852ee0fb06261062ba698899df977605fd6c8eae4b1e82369694a7e6"} Oct 05 07:01:51 crc kubenswrapper[4846]: I1005 07:01:51.314237 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-98wxg" event={"ID":"0c327613-e7ab-4d2d-882c-fecaf1590c87","Type":"ContainerStarted","Data":"842f9f6a9f51f1952e1d4ce07f325b43328999a351180e118b454a0e2477ed14"} Oct 05 07:01:51 crc kubenswrapper[4846]: I1005 07:01:51.341513 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-98wxg" podStartSLOduration=2.94644405 podStartE2EDuration="4.341490775s" podCreationTimestamp="2025-10-05 07:01:47 +0000 UTC" firstStartedPulling="2025-10-05 07:01:49.29688209 +0000 UTC m=+831.537734875" lastFinishedPulling="2025-10-05 07:01:50.691928785 +0000 UTC m=+832.932781600" observedRunningTime="2025-10-05 07:01:51.337366345 +0000 UTC m=+833.578219120" watchObservedRunningTime="2025-10-05 07:01:51.341490775 +0000 UTC m=+833.582343550" Oct 05 07:01:54 crc kubenswrapper[4846]: I1005 07:01:54.274841 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-wqmsh" Oct 05 07:01:54 crc kubenswrapper[4846]: I1005 07:01:54.275348 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-wqmsh" Oct 05 07:01:54 crc kubenswrapper[4846]: I1005 07:01:54.328332 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-wqmsh" Oct 05 07:01:54 crc kubenswrapper[4846]: I1005 07:01:54.391152 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-wqmsh" Oct 05 07:01:56 crc kubenswrapper[4846]: I1005 07:01:56.131083 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-wqmsh"] Oct 05 07:01:56 crc kubenswrapper[4846]: I1005 07:01:56.366039 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-wqmsh" podUID="ab709817-df14-4fe9-9c58-ae232db7915e" containerName="registry-server" containerID="cri-o://4d1924bae279f63f5709332d16e9bdef749b98c781e4b15c7408342b5273ef9b" gracePeriod=2 Oct 05 07:01:56 crc kubenswrapper[4846]: I1005 07:01:56.566592 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-webhook-server-7bd844db7f-6hpbl" Oct 05 07:01:56 crc kubenswrapper[4846]: I1005 07:01:56.892531 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wqmsh" Oct 05 07:01:56 crc kubenswrapper[4846]: I1005 07:01:56.999302 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ab709817-df14-4fe9-9c58-ae232db7915e-catalog-content\") pod \"ab709817-df14-4fe9-9c58-ae232db7915e\" (UID: \"ab709817-df14-4fe9-9c58-ae232db7915e\") " Oct 05 07:01:56 crc kubenswrapper[4846]: I1005 07:01:56.999545 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bdw6v\" (UniqueName: \"kubernetes.io/projected/ab709817-df14-4fe9-9c58-ae232db7915e-kube-api-access-bdw6v\") pod \"ab709817-df14-4fe9-9c58-ae232db7915e\" (UID: \"ab709817-df14-4fe9-9c58-ae232db7915e\") " Oct 05 07:01:56 crc kubenswrapper[4846]: I1005 07:01:56.999628 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ab709817-df14-4fe9-9c58-ae232db7915e-utilities\") pod \"ab709817-df14-4fe9-9c58-ae232db7915e\" (UID: \"ab709817-df14-4fe9-9c58-ae232db7915e\") " Oct 05 07:01:57 crc kubenswrapper[4846]: I1005 07:01:57.000809 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ab709817-df14-4fe9-9c58-ae232db7915e-utilities" (OuterVolumeSpecName: "utilities") pod "ab709817-df14-4fe9-9c58-ae232db7915e" (UID: "ab709817-df14-4fe9-9c58-ae232db7915e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 07:01:57 crc kubenswrapper[4846]: I1005 07:01:57.009208 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ab709817-df14-4fe9-9c58-ae232db7915e-kube-api-access-bdw6v" (OuterVolumeSpecName: "kube-api-access-bdw6v") pod "ab709817-df14-4fe9-9c58-ae232db7915e" (UID: "ab709817-df14-4fe9-9c58-ae232db7915e"). InnerVolumeSpecName "kube-api-access-bdw6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:01:57 crc kubenswrapper[4846]: I1005 07:01:57.051228 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ab709817-df14-4fe9-9c58-ae232db7915e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ab709817-df14-4fe9-9c58-ae232db7915e" (UID: "ab709817-df14-4fe9-9c58-ae232db7915e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 07:01:57 crc kubenswrapper[4846]: I1005 07:01:57.101388 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bdw6v\" (UniqueName: \"kubernetes.io/projected/ab709817-df14-4fe9-9c58-ae232db7915e-kube-api-access-bdw6v\") on node \"crc\" DevicePath \"\"" Oct 05 07:01:57 crc kubenswrapper[4846]: I1005 07:01:57.101438 4846 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ab709817-df14-4fe9-9c58-ae232db7915e-utilities\") on node \"crc\" DevicePath \"\"" Oct 05 07:01:57 crc kubenswrapper[4846]: I1005 07:01:57.101448 4846 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ab709817-df14-4fe9-9c58-ae232db7915e-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 05 07:01:57 crc kubenswrapper[4846]: I1005 07:01:57.380420 4846 generic.go:334] "Generic (PLEG): container finished" podID="ab709817-df14-4fe9-9c58-ae232db7915e" containerID="4d1924bae279f63f5709332d16e9bdef749b98c781e4b15c7408342b5273ef9b" exitCode=0 Oct 05 07:01:57 crc kubenswrapper[4846]: I1005 07:01:57.380486 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wqmsh" event={"ID":"ab709817-df14-4fe9-9c58-ae232db7915e","Type":"ContainerDied","Data":"4d1924bae279f63f5709332d16e9bdef749b98c781e4b15c7408342b5273ef9b"} Oct 05 07:01:57 crc kubenswrapper[4846]: I1005 07:01:57.380559 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wqmsh" Oct 05 07:01:57 crc kubenswrapper[4846]: I1005 07:01:57.380553 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wqmsh" event={"ID":"ab709817-df14-4fe9-9c58-ae232db7915e","Type":"ContainerDied","Data":"a5e3125f2aa7c011ca25e5cae610eae2b02bc0f501b73d297ceeafc08d529985"} Oct 05 07:01:57 crc kubenswrapper[4846]: I1005 07:01:57.380578 4846 scope.go:117] "RemoveContainer" containerID="4d1924bae279f63f5709332d16e9bdef749b98c781e4b15c7408342b5273ef9b" Oct 05 07:01:57 crc kubenswrapper[4846]: I1005 07:01:57.406827 4846 scope.go:117] "RemoveContainer" containerID="96a3d3a6032af6874a3958b1d2c71dd57661b3b2aec3a53772007afb6600d948" Oct 05 07:01:57 crc kubenswrapper[4846]: I1005 07:01:57.426547 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-wqmsh"] Oct 05 07:01:57 crc kubenswrapper[4846]: I1005 07:01:57.430732 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-wqmsh"] Oct 05 07:01:57 crc kubenswrapper[4846]: I1005 07:01:57.447135 4846 scope.go:117] "RemoveContainer" containerID="77afb2db24398a36b3ea5680fca36296a09cb517a70ccc2ad54547441055d8b5" Oct 05 07:01:57 crc kubenswrapper[4846]: I1005 07:01:57.480078 4846 scope.go:117] "RemoveContainer" containerID="4d1924bae279f63f5709332d16e9bdef749b98c781e4b15c7408342b5273ef9b" Oct 05 07:01:57 crc kubenswrapper[4846]: E1005 07:01:57.480731 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4d1924bae279f63f5709332d16e9bdef749b98c781e4b15c7408342b5273ef9b\": container with ID starting with 4d1924bae279f63f5709332d16e9bdef749b98c781e4b15c7408342b5273ef9b not found: ID does not exist" containerID="4d1924bae279f63f5709332d16e9bdef749b98c781e4b15c7408342b5273ef9b" Oct 05 07:01:57 crc kubenswrapper[4846]: I1005 07:01:57.480822 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4d1924bae279f63f5709332d16e9bdef749b98c781e4b15c7408342b5273ef9b"} err="failed to get container status \"4d1924bae279f63f5709332d16e9bdef749b98c781e4b15c7408342b5273ef9b\": rpc error: code = NotFound desc = could not find container \"4d1924bae279f63f5709332d16e9bdef749b98c781e4b15c7408342b5273ef9b\": container with ID starting with 4d1924bae279f63f5709332d16e9bdef749b98c781e4b15c7408342b5273ef9b not found: ID does not exist" Oct 05 07:01:57 crc kubenswrapper[4846]: I1005 07:01:57.480881 4846 scope.go:117] "RemoveContainer" containerID="96a3d3a6032af6874a3958b1d2c71dd57661b3b2aec3a53772007afb6600d948" Oct 05 07:01:57 crc kubenswrapper[4846]: E1005 07:01:57.481898 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"96a3d3a6032af6874a3958b1d2c71dd57661b3b2aec3a53772007afb6600d948\": container with ID starting with 96a3d3a6032af6874a3958b1d2c71dd57661b3b2aec3a53772007afb6600d948 not found: ID does not exist" containerID="96a3d3a6032af6874a3958b1d2c71dd57661b3b2aec3a53772007afb6600d948" Oct 05 07:01:57 crc kubenswrapper[4846]: I1005 07:01:57.481930 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"96a3d3a6032af6874a3958b1d2c71dd57661b3b2aec3a53772007afb6600d948"} err="failed to get container status \"96a3d3a6032af6874a3958b1d2c71dd57661b3b2aec3a53772007afb6600d948\": rpc error: code = NotFound desc = could not find container \"96a3d3a6032af6874a3958b1d2c71dd57661b3b2aec3a53772007afb6600d948\": container with ID starting with 96a3d3a6032af6874a3958b1d2c71dd57661b3b2aec3a53772007afb6600d948 not found: ID does not exist" Oct 05 07:01:57 crc kubenswrapper[4846]: I1005 07:01:57.481949 4846 scope.go:117] "RemoveContainer" containerID="77afb2db24398a36b3ea5680fca36296a09cb517a70ccc2ad54547441055d8b5" Oct 05 07:01:57 crc kubenswrapper[4846]: E1005 07:01:57.482248 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"77afb2db24398a36b3ea5680fca36296a09cb517a70ccc2ad54547441055d8b5\": container with ID starting with 77afb2db24398a36b3ea5680fca36296a09cb517a70ccc2ad54547441055d8b5 not found: ID does not exist" containerID="77afb2db24398a36b3ea5680fca36296a09cb517a70ccc2ad54547441055d8b5" Oct 05 07:01:57 crc kubenswrapper[4846]: I1005 07:01:57.482299 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"77afb2db24398a36b3ea5680fca36296a09cb517a70ccc2ad54547441055d8b5"} err="failed to get container status \"77afb2db24398a36b3ea5680fca36296a09cb517a70ccc2ad54547441055d8b5\": rpc error: code = NotFound desc = could not find container \"77afb2db24398a36b3ea5680fca36296a09cb517a70ccc2ad54547441055d8b5\": container with ID starting with 77afb2db24398a36b3ea5680fca36296a09cb517a70ccc2ad54547441055d8b5 not found: ID does not exist" Oct 05 07:01:57 crc kubenswrapper[4846]: I1005 07:01:57.865342 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-98wxg" Oct 05 07:01:57 crc kubenswrapper[4846]: I1005 07:01:57.865386 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-98wxg" Oct 05 07:01:57 crc kubenswrapper[4846]: I1005 07:01:57.962616 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-98wxg" Oct 05 07:01:58 crc kubenswrapper[4846]: I1005 07:01:58.445359 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-98wxg" Oct 05 07:01:58 crc kubenswrapper[4846]: I1005 07:01:58.505798 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ab709817-df14-4fe9-9c58-ae232db7915e" path="/var/lib/kubelet/pods/ab709817-df14-4fe9-9c58-ae232db7915e/volumes" Oct 05 07:02:00 crc kubenswrapper[4846]: I1005 07:02:00.933311 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-98wxg"] Oct 05 07:02:00 crc kubenswrapper[4846]: I1005 07:02:00.934299 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-98wxg" podUID="0c327613-e7ab-4d2d-882c-fecaf1590c87" containerName="registry-server" containerID="cri-o://842f9f6a9f51f1952e1d4ce07f325b43328999a351180e118b454a0e2477ed14" gracePeriod=2 Oct 05 07:02:01 crc kubenswrapper[4846]: I1005 07:02:01.383054 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-98wxg" Oct 05 07:02:01 crc kubenswrapper[4846]: I1005 07:02:01.421732 4846 generic.go:334] "Generic (PLEG): container finished" podID="0c327613-e7ab-4d2d-882c-fecaf1590c87" containerID="842f9f6a9f51f1952e1d4ce07f325b43328999a351180e118b454a0e2477ed14" exitCode=0 Oct 05 07:02:01 crc kubenswrapper[4846]: I1005 07:02:01.421800 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-98wxg" event={"ID":"0c327613-e7ab-4d2d-882c-fecaf1590c87","Type":"ContainerDied","Data":"842f9f6a9f51f1952e1d4ce07f325b43328999a351180e118b454a0e2477ed14"} Oct 05 07:02:01 crc kubenswrapper[4846]: I1005 07:02:01.421851 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-98wxg" event={"ID":"0c327613-e7ab-4d2d-882c-fecaf1590c87","Type":"ContainerDied","Data":"8d4baa880a21087ea285ed0d0e3ea948b5e80a6635ec391580cfea21ffebe9cf"} Oct 05 07:02:01 crc kubenswrapper[4846]: I1005 07:02:01.421863 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-98wxg" Oct 05 07:02:01 crc kubenswrapper[4846]: I1005 07:02:01.421882 4846 scope.go:117] "RemoveContainer" containerID="842f9f6a9f51f1952e1d4ce07f325b43328999a351180e118b454a0e2477ed14" Oct 05 07:02:01 crc kubenswrapper[4846]: I1005 07:02:01.451812 4846 scope.go:117] "RemoveContainer" containerID="7d32d057852ee0fb06261062ba698899df977605fd6c8eae4b1e82369694a7e6" Oct 05 07:02:01 crc kubenswrapper[4846]: I1005 07:02:01.474764 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0c327613-e7ab-4d2d-882c-fecaf1590c87-utilities\") pod \"0c327613-e7ab-4d2d-882c-fecaf1590c87\" (UID: \"0c327613-e7ab-4d2d-882c-fecaf1590c87\") " Oct 05 07:02:01 crc kubenswrapper[4846]: I1005 07:02:01.474825 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-855lq\" (UniqueName: \"kubernetes.io/projected/0c327613-e7ab-4d2d-882c-fecaf1590c87-kube-api-access-855lq\") pod \"0c327613-e7ab-4d2d-882c-fecaf1590c87\" (UID: \"0c327613-e7ab-4d2d-882c-fecaf1590c87\") " Oct 05 07:02:01 crc kubenswrapper[4846]: I1005 07:02:01.474897 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0c327613-e7ab-4d2d-882c-fecaf1590c87-catalog-content\") pod \"0c327613-e7ab-4d2d-882c-fecaf1590c87\" (UID: \"0c327613-e7ab-4d2d-882c-fecaf1590c87\") " Oct 05 07:02:01 crc kubenswrapper[4846]: I1005 07:02:01.476320 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0c327613-e7ab-4d2d-882c-fecaf1590c87-utilities" (OuterVolumeSpecName: "utilities") pod "0c327613-e7ab-4d2d-882c-fecaf1590c87" (UID: "0c327613-e7ab-4d2d-882c-fecaf1590c87"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 07:02:01 crc kubenswrapper[4846]: I1005 07:02:01.477845 4846 scope.go:117] "RemoveContainer" containerID="b77a3358a92b865a06be476f33059afa0a1d84281e9360ebd7c9edc5dd0b9f15" Oct 05 07:02:01 crc kubenswrapper[4846]: I1005 07:02:01.481860 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0c327613-e7ab-4d2d-882c-fecaf1590c87-kube-api-access-855lq" (OuterVolumeSpecName: "kube-api-access-855lq") pod "0c327613-e7ab-4d2d-882c-fecaf1590c87" (UID: "0c327613-e7ab-4d2d-882c-fecaf1590c87"). InnerVolumeSpecName "kube-api-access-855lq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:02:01 crc kubenswrapper[4846]: I1005 07:02:01.494777 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0c327613-e7ab-4d2d-882c-fecaf1590c87-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0c327613-e7ab-4d2d-882c-fecaf1590c87" (UID: "0c327613-e7ab-4d2d-882c-fecaf1590c87"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 07:02:01 crc kubenswrapper[4846]: I1005 07:02:01.515115 4846 scope.go:117] "RemoveContainer" containerID="842f9f6a9f51f1952e1d4ce07f325b43328999a351180e118b454a0e2477ed14" Oct 05 07:02:01 crc kubenswrapper[4846]: E1005 07:02:01.515635 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"842f9f6a9f51f1952e1d4ce07f325b43328999a351180e118b454a0e2477ed14\": container with ID starting with 842f9f6a9f51f1952e1d4ce07f325b43328999a351180e118b454a0e2477ed14 not found: ID does not exist" containerID="842f9f6a9f51f1952e1d4ce07f325b43328999a351180e118b454a0e2477ed14" Oct 05 07:02:01 crc kubenswrapper[4846]: I1005 07:02:01.515679 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"842f9f6a9f51f1952e1d4ce07f325b43328999a351180e118b454a0e2477ed14"} err="failed to get container status \"842f9f6a9f51f1952e1d4ce07f325b43328999a351180e118b454a0e2477ed14\": rpc error: code = NotFound desc = could not find container \"842f9f6a9f51f1952e1d4ce07f325b43328999a351180e118b454a0e2477ed14\": container with ID starting with 842f9f6a9f51f1952e1d4ce07f325b43328999a351180e118b454a0e2477ed14 not found: ID does not exist" Oct 05 07:02:01 crc kubenswrapper[4846]: I1005 07:02:01.515708 4846 scope.go:117] "RemoveContainer" containerID="7d32d057852ee0fb06261062ba698899df977605fd6c8eae4b1e82369694a7e6" Oct 05 07:02:01 crc kubenswrapper[4846]: E1005 07:02:01.516255 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7d32d057852ee0fb06261062ba698899df977605fd6c8eae4b1e82369694a7e6\": container with ID starting with 7d32d057852ee0fb06261062ba698899df977605fd6c8eae4b1e82369694a7e6 not found: ID does not exist" containerID="7d32d057852ee0fb06261062ba698899df977605fd6c8eae4b1e82369694a7e6" Oct 05 07:02:01 crc kubenswrapper[4846]: I1005 07:02:01.516329 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7d32d057852ee0fb06261062ba698899df977605fd6c8eae4b1e82369694a7e6"} err="failed to get container status \"7d32d057852ee0fb06261062ba698899df977605fd6c8eae4b1e82369694a7e6\": rpc error: code = NotFound desc = could not find container \"7d32d057852ee0fb06261062ba698899df977605fd6c8eae4b1e82369694a7e6\": container with ID starting with 7d32d057852ee0fb06261062ba698899df977605fd6c8eae4b1e82369694a7e6 not found: ID does not exist" Oct 05 07:02:01 crc kubenswrapper[4846]: I1005 07:02:01.516378 4846 scope.go:117] "RemoveContainer" containerID="b77a3358a92b865a06be476f33059afa0a1d84281e9360ebd7c9edc5dd0b9f15" Oct 05 07:02:01 crc kubenswrapper[4846]: E1005 07:02:01.516865 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b77a3358a92b865a06be476f33059afa0a1d84281e9360ebd7c9edc5dd0b9f15\": container with ID starting with b77a3358a92b865a06be476f33059afa0a1d84281e9360ebd7c9edc5dd0b9f15 not found: ID does not exist" containerID="b77a3358a92b865a06be476f33059afa0a1d84281e9360ebd7c9edc5dd0b9f15" Oct 05 07:02:01 crc kubenswrapper[4846]: I1005 07:02:01.516916 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b77a3358a92b865a06be476f33059afa0a1d84281e9360ebd7c9edc5dd0b9f15"} err="failed to get container status \"b77a3358a92b865a06be476f33059afa0a1d84281e9360ebd7c9edc5dd0b9f15\": rpc error: code = NotFound desc = could not find container \"b77a3358a92b865a06be476f33059afa0a1d84281e9360ebd7c9edc5dd0b9f15\": container with ID starting with b77a3358a92b865a06be476f33059afa0a1d84281e9360ebd7c9edc5dd0b9f15 not found: ID does not exist" Oct 05 07:02:01 crc kubenswrapper[4846]: I1005 07:02:01.577027 4846 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0c327613-e7ab-4d2d-882c-fecaf1590c87-utilities\") on node \"crc\" DevicePath \"\"" Oct 05 07:02:01 crc kubenswrapper[4846]: I1005 07:02:01.577073 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-855lq\" (UniqueName: \"kubernetes.io/projected/0c327613-e7ab-4d2d-882c-fecaf1590c87-kube-api-access-855lq\") on node \"crc\" DevicePath \"\"" Oct 05 07:02:01 crc kubenswrapper[4846]: I1005 07:02:01.577089 4846 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0c327613-e7ab-4d2d-882c-fecaf1590c87-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 05 07:02:01 crc kubenswrapper[4846]: I1005 07:02:01.753315 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-98wxg"] Oct 05 07:02:01 crc kubenswrapper[4846]: I1005 07:02:01.757171 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-98wxg"] Oct 05 07:02:02 crc kubenswrapper[4846]: I1005 07:02:02.510297 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0c327613-e7ab-4d2d-882c-fecaf1590c87" path="/var/lib/kubelet/pods/0c327613-e7ab-4d2d-882c-fecaf1590c87/volumes" Oct 05 07:02:16 crc kubenswrapper[4846]: I1005 07:02:16.158820 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-8fccc95cf-j9fdr" Oct 05 07:02:17 crc kubenswrapper[4846]: I1005 07:02:17.012103 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-kz2bc"] Oct 05 07:02:17 crc kubenswrapper[4846]: E1005 07:02:17.013263 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0c327613-e7ab-4d2d-882c-fecaf1590c87" containerName="registry-server" Oct 05 07:02:17 crc kubenswrapper[4846]: I1005 07:02:17.013362 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="0c327613-e7ab-4d2d-882c-fecaf1590c87" containerName="registry-server" Oct 05 07:02:17 crc kubenswrapper[4846]: E1005 07:02:17.013562 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ab709817-df14-4fe9-9c58-ae232db7915e" containerName="registry-server" Oct 05 07:02:17 crc kubenswrapper[4846]: I1005 07:02:17.013637 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="ab709817-df14-4fe9-9c58-ae232db7915e" containerName="registry-server" Oct 05 07:02:17 crc kubenswrapper[4846]: E1005 07:02:17.013716 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ab709817-df14-4fe9-9c58-ae232db7915e" containerName="extract-utilities" Oct 05 07:02:17 crc kubenswrapper[4846]: I1005 07:02:17.013785 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="ab709817-df14-4fe9-9c58-ae232db7915e" containerName="extract-utilities" Oct 05 07:02:17 crc kubenswrapper[4846]: E1005 07:02:17.013860 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0c327613-e7ab-4d2d-882c-fecaf1590c87" containerName="extract-utilities" Oct 05 07:02:17 crc kubenswrapper[4846]: I1005 07:02:17.013933 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="0c327613-e7ab-4d2d-882c-fecaf1590c87" containerName="extract-utilities" Oct 05 07:02:17 crc kubenswrapper[4846]: E1005 07:02:17.014013 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0c327613-e7ab-4d2d-882c-fecaf1590c87" containerName="extract-content" Oct 05 07:02:17 crc kubenswrapper[4846]: I1005 07:02:17.014088 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="0c327613-e7ab-4d2d-882c-fecaf1590c87" containerName="extract-content" Oct 05 07:02:17 crc kubenswrapper[4846]: E1005 07:02:17.014161 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ab709817-df14-4fe9-9c58-ae232db7915e" containerName="extract-content" Oct 05 07:02:17 crc kubenswrapper[4846]: I1005 07:02:17.014254 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="ab709817-df14-4fe9-9c58-ae232db7915e" containerName="extract-content" Oct 05 07:02:17 crc kubenswrapper[4846]: I1005 07:02:17.014498 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="ab709817-df14-4fe9-9c58-ae232db7915e" containerName="registry-server" Oct 05 07:02:17 crc kubenswrapper[4846]: I1005 07:02:17.014587 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="0c327613-e7ab-4d2d-882c-fecaf1590c87" containerName="registry-server" Oct 05 07:02:17 crc kubenswrapper[4846]: I1005 07:02:17.017242 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-kz2bc" Oct 05 07:02:17 crc kubenswrapper[4846]: I1005 07:02:17.019355 4846 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-certs-secret" Oct 05 07:02:17 crc kubenswrapper[4846]: I1005 07:02:17.019655 4846 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-daemon-dockercfg-6b72v" Oct 05 07:02:17 crc kubenswrapper[4846]: I1005 07:02:17.023264 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-webhook-server-64bf5d555-sf9k8"] Oct 05 07:02:17 crc kubenswrapper[4846]: I1005 07:02:17.024138 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-sf9k8" Oct 05 07:02:17 crc kubenswrapper[4846]: I1005 07:02:17.024892 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"frr-startup" Oct 05 07:02:17 crc kubenswrapper[4846]: I1005 07:02:17.027312 4846 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-webhook-server-cert" Oct 05 07:02:17 crc kubenswrapper[4846]: I1005 07:02:17.044503 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-64bf5d555-sf9k8"] Oct 05 07:02:17 crc kubenswrapper[4846]: I1005 07:02:17.113020 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/speaker-czf4m"] Oct 05 07:02:17 crc kubenswrapper[4846]: I1005 07:02:17.114142 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-czf4m" Oct 05 07:02:17 crc kubenswrapper[4846]: I1005 07:02:17.118532 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ab2bc1f2-31f3-4793-926a-e649b556d8de-metrics-certs\") pod \"frr-k8s-kz2bc\" (UID: \"ab2bc1f2-31f3-4793-926a-e649b556d8de\") " pod="metallb-system/frr-k8s-kz2bc" Oct 05 07:02:17 crc kubenswrapper[4846]: I1005 07:02:17.118616 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lcxpc\" (UniqueName: \"kubernetes.io/projected/ab2bc1f2-31f3-4793-926a-e649b556d8de-kube-api-access-lcxpc\") pod \"frr-k8s-kz2bc\" (UID: \"ab2bc1f2-31f3-4793-926a-e649b556d8de\") " pod="metallb-system/frr-k8s-kz2bc" Oct 05 07:02:17 crc kubenswrapper[4846]: I1005 07:02:17.118677 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/ab2bc1f2-31f3-4793-926a-e649b556d8de-metrics\") pod \"frr-k8s-kz2bc\" (UID: \"ab2bc1f2-31f3-4793-926a-e649b556d8de\") " pod="metallb-system/frr-k8s-kz2bc" Oct 05 07:02:17 crc kubenswrapper[4846]: I1005 07:02:17.118723 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/ab2bc1f2-31f3-4793-926a-e649b556d8de-frr-conf\") pod \"frr-k8s-kz2bc\" (UID: \"ab2bc1f2-31f3-4793-926a-e649b556d8de\") " pod="metallb-system/frr-k8s-kz2bc" Oct 05 07:02:17 crc kubenswrapper[4846]: I1005 07:02:17.118774 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/ab2bc1f2-31f3-4793-926a-e649b556d8de-frr-sockets\") pod \"frr-k8s-kz2bc\" (UID: \"ab2bc1f2-31f3-4793-926a-e649b556d8de\") " pod="metallb-system/frr-k8s-kz2bc" Oct 05 07:02:17 crc kubenswrapper[4846]: I1005 07:02:17.118800 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/ab2bc1f2-31f3-4793-926a-e649b556d8de-frr-startup\") pod \"frr-k8s-kz2bc\" (UID: \"ab2bc1f2-31f3-4793-926a-e649b556d8de\") " pod="metallb-system/frr-k8s-kz2bc" Oct 05 07:02:17 crc kubenswrapper[4846]: I1005 07:02:17.118829 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/ab2bc1f2-31f3-4793-926a-e649b556d8de-reloader\") pod \"frr-k8s-kz2bc\" (UID: \"ab2bc1f2-31f3-4793-926a-e649b556d8de\") " pod="metallb-system/frr-k8s-kz2bc" Oct 05 07:02:17 crc kubenswrapper[4846]: I1005 07:02:17.120854 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"metallb-excludel2" Oct 05 07:02:17 crc kubenswrapper[4846]: I1005 07:02:17.121123 4846 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-g2tnd" Oct 05 07:02:17 crc kubenswrapper[4846]: I1005 07:02:17.121287 4846 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-certs-secret" Oct 05 07:02:17 crc kubenswrapper[4846]: I1005 07:02:17.121429 4846 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-memberlist" Oct 05 07:02:17 crc kubenswrapper[4846]: I1005 07:02:17.128477 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/controller-68d546b9d8-vhpgd"] Oct 05 07:02:17 crc kubenswrapper[4846]: I1005 07:02:17.131526 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-68d546b9d8-vhpgd" Oct 05 07:02:17 crc kubenswrapper[4846]: I1005 07:02:17.134849 4846 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-certs-secret" Oct 05 07:02:17 crc kubenswrapper[4846]: I1005 07:02:17.144054 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-68d546b9d8-vhpgd"] Oct 05 07:02:17 crc kubenswrapper[4846]: I1005 07:02:17.220390 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ab2bc1f2-31f3-4793-926a-e649b556d8de-metrics-certs\") pod \"frr-k8s-kz2bc\" (UID: \"ab2bc1f2-31f3-4793-926a-e649b556d8de\") " pod="metallb-system/frr-k8s-kz2bc" Oct 05 07:02:17 crc kubenswrapper[4846]: I1005 07:02:17.221652 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lcxpc\" (UniqueName: \"kubernetes.io/projected/ab2bc1f2-31f3-4793-926a-e649b556d8de-kube-api-access-lcxpc\") pod \"frr-k8s-kz2bc\" (UID: \"ab2bc1f2-31f3-4793-926a-e649b556d8de\") " pod="metallb-system/frr-k8s-kz2bc" Oct 05 07:02:17 crc kubenswrapper[4846]: I1005 07:02:17.221793 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/dad102f2-8887-4a01-98b1-be88b2836eb5-memberlist\") pod \"speaker-czf4m\" (UID: \"dad102f2-8887-4a01-98b1-be88b2836eb5\") " pod="metallb-system/speaker-czf4m" Oct 05 07:02:17 crc kubenswrapper[4846]: I1005 07:02:17.221842 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fj2k2\" (UniqueName: \"kubernetes.io/projected/dad102f2-8887-4a01-98b1-be88b2836eb5-kube-api-access-fj2k2\") pod \"speaker-czf4m\" (UID: \"dad102f2-8887-4a01-98b1-be88b2836eb5\") " pod="metallb-system/speaker-czf4m" Oct 05 07:02:17 crc kubenswrapper[4846]: I1005 07:02:17.221918 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/ab2bc1f2-31f3-4793-926a-e649b556d8de-metrics\") pod \"frr-k8s-kz2bc\" (UID: \"ab2bc1f2-31f3-4793-926a-e649b556d8de\") " pod="metallb-system/frr-k8s-kz2bc" Oct 05 07:02:17 crc kubenswrapper[4846]: I1005 07:02:17.222002 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/dad102f2-8887-4a01-98b1-be88b2836eb5-metallb-excludel2\") pod \"speaker-czf4m\" (UID: \"dad102f2-8887-4a01-98b1-be88b2836eb5\") " pod="metallb-system/speaker-czf4m" Oct 05 07:02:17 crc kubenswrapper[4846]: I1005 07:02:17.222088 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/ab2bc1f2-31f3-4793-926a-e649b556d8de-frr-conf\") pod \"frr-k8s-kz2bc\" (UID: \"ab2bc1f2-31f3-4793-926a-e649b556d8de\") " pod="metallb-system/frr-k8s-kz2bc" Oct 05 07:02:17 crc kubenswrapper[4846]: I1005 07:02:17.222140 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/0a4508c6-6dff-467f-9cac-ef5c7b921984-cert\") pod \"frr-k8s-webhook-server-64bf5d555-sf9k8\" (UID: \"0a4508c6-6dff-467f-9cac-ef5c7b921984\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-sf9k8" Oct 05 07:02:17 crc kubenswrapper[4846]: I1005 07:02:17.222187 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xk6gv\" (UniqueName: \"kubernetes.io/projected/0a4508c6-6dff-467f-9cac-ef5c7b921984-kube-api-access-xk6gv\") pod \"frr-k8s-webhook-server-64bf5d555-sf9k8\" (UID: \"0a4508c6-6dff-467f-9cac-ef5c7b921984\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-sf9k8" Oct 05 07:02:17 crc kubenswrapper[4846]: I1005 07:02:17.222256 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/ab2bc1f2-31f3-4793-926a-e649b556d8de-frr-sockets\") pod \"frr-k8s-kz2bc\" (UID: \"ab2bc1f2-31f3-4793-926a-e649b556d8de\") " pod="metallb-system/frr-k8s-kz2bc" Oct 05 07:02:17 crc kubenswrapper[4846]: I1005 07:02:17.222297 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/ab2bc1f2-31f3-4793-926a-e649b556d8de-frr-startup\") pod \"frr-k8s-kz2bc\" (UID: \"ab2bc1f2-31f3-4793-926a-e649b556d8de\") " pod="metallb-system/frr-k8s-kz2bc" Oct 05 07:02:17 crc kubenswrapper[4846]: I1005 07:02:17.222329 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/ab2bc1f2-31f3-4793-926a-e649b556d8de-reloader\") pod \"frr-k8s-kz2bc\" (UID: \"ab2bc1f2-31f3-4793-926a-e649b556d8de\") " pod="metallb-system/frr-k8s-kz2bc" Oct 05 07:02:17 crc kubenswrapper[4846]: I1005 07:02:17.222353 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/dad102f2-8887-4a01-98b1-be88b2836eb5-metrics-certs\") pod \"speaker-czf4m\" (UID: \"dad102f2-8887-4a01-98b1-be88b2836eb5\") " pod="metallb-system/speaker-czf4m" Oct 05 07:02:17 crc kubenswrapper[4846]: I1005 07:02:17.222655 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/ab2bc1f2-31f3-4793-926a-e649b556d8de-metrics\") pod \"frr-k8s-kz2bc\" (UID: \"ab2bc1f2-31f3-4793-926a-e649b556d8de\") " pod="metallb-system/frr-k8s-kz2bc" Oct 05 07:02:17 crc kubenswrapper[4846]: I1005 07:02:17.222753 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/ab2bc1f2-31f3-4793-926a-e649b556d8de-frr-conf\") pod \"frr-k8s-kz2bc\" (UID: \"ab2bc1f2-31f3-4793-926a-e649b556d8de\") " pod="metallb-system/frr-k8s-kz2bc" Oct 05 07:02:17 crc kubenswrapper[4846]: I1005 07:02:17.222890 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/ab2bc1f2-31f3-4793-926a-e649b556d8de-reloader\") pod \"frr-k8s-kz2bc\" (UID: \"ab2bc1f2-31f3-4793-926a-e649b556d8de\") " pod="metallb-system/frr-k8s-kz2bc" Oct 05 07:02:17 crc kubenswrapper[4846]: I1005 07:02:17.222937 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/ab2bc1f2-31f3-4793-926a-e649b556d8de-frr-sockets\") pod \"frr-k8s-kz2bc\" (UID: \"ab2bc1f2-31f3-4793-926a-e649b556d8de\") " pod="metallb-system/frr-k8s-kz2bc" Oct 05 07:02:17 crc kubenswrapper[4846]: I1005 07:02:17.223279 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/ab2bc1f2-31f3-4793-926a-e649b556d8de-frr-startup\") pod \"frr-k8s-kz2bc\" (UID: \"ab2bc1f2-31f3-4793-926a-e649b556d8de\") " pod="metallb-system/frr-k8s-kz2bc" Oct 05 07:02:17 crc kubenswrapper[4846]: I1005 07:02:17.230265 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ab2bc1f2-31f3-4793-926a-e649b556d8de-metrics-certs\") pod \"frr-k8s-kz2bc\" (UID: \"ab2bc1f2-31f3-4793-926a-e649b556d8de\") " pod="metallb-system/frr-k8s-kz2bc" Oct 05 07:02:17 crc kubenswrapper[4846]: I1005 07:02:17.240138 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lcxpc\" (UniqueName: \"kubernetes.io/projected/ab2bc1f2-31f3-4793-926a-e649b556d8de-kube-api-access-lcxpc\") pod \"frr-k8s-kz2bc\" (UID: \"ab2bc1f2-31f3-4793-926a-e649b556d8de\") " pod="metallb-system/frr-k8s-kz2bc" Oct 05 07:02:17 crc kubenswrapper[4846]: I1005 07:02:17.323875 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/dad102f2-8887-4a01-98b1-be88b2836eb5-memberlist\") pod \"speaker-czf4m\" (UID: \"dad102f2-8887-4a01-98b1-be88b2836eb5\") " pod="metallb-system/speaker-czf4m" Oct 05 07:02:17 crc kubenswrapper[4846]: I1005 07:02:17.323944 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fj2k2\" (UniqueName: \"kubernetes.io/projected/dad102f2-8887-4a01-98b1-be88b2836eb5-kube-api-access-fj2k2\") pod \"speaker-czf4m\" (UID: \"dad102f2-8887-4a01-98b1-be88b2836eb5\") " pod="metallb-system/speaker-czf4m" Oct 05 07:02:17 crc kubenswrapper[4846]: I1005 07:02:17.323980 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/3541826d-961c-4259-be16-51f4dc05aec6-cert\") pod \"controller-68d546b9d8-vhpgd\" (UID: \"3541826d-961c-4259-be16-51f4dc05aec6\") " pod="metallb-system/controller-68d546b9d8-vhpgd" Oct 05 07:02:17 crc kubenswrapper[4846]: I1005 07:02:17.324011 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/3541826d-961c-4259-be16-51f4dc05aec6-metrics-certs\") pod \"controller-68d546b9d8-vhpgd\" (UID: \"3541826d-961c-4259-be16-51f4dc05aec6\") " pod="metallb-system/controller-68d546b9d8-vhpgd" Oct 05 07:02:17 crc kubenswrapper[4846]: I1005 07:02:17.324029 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/dad102f2-8887-4a01-98b1-be88b2836eb5-metallb-excludel2\") pod \"speaker-czf4m\" (UID: \"dad102f2-8887-4a01-98b1-be88b2836eb5\") " pod="metallb-system/speaker-czf4m" Oct 05 07:02:17 crc kubenswrapper[4846]: I1005 07:02:17.324050 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/0a4508c6-6dff-467f-9cac-ef5c7b921984-cert\") pod \"frr-k8s-webhook-server-64bf5d555-sf9k8\" (UID: \"0a4508c6-6dff-467f-9cac-ef5c7b921984\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-sf9k8" Oct 05 07:02:17 crc kubenswrapper[4846]: I1005 07:02:17.324081 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xk6gv\" (UniqueName: \"kubernetes.io/projected/0a4508c6-6dff-467f-9cac-ef5c7b921984-kube-api-access-xk6gv\") pod \"frr-k8s-webhook-server-64bf5d555-sf9k8\" (UID: \"0a4508c6-6dff-467f-9cac-ef5c7b921984\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-sf9k8" Oct 05 07:02:17 crc kubenswrapper[4846]: I1005 07:02:17.324110 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lqtrr\" (UniqueName: \"kubernetes.io/projected/3541826d-961c-4259-be16-51f4dc05aec6-kube-api-access-lqtrr\") pod \"controller-68d546b9d8-vhpgd\" (UID: \"3541826d-961c-4259-be16-51f4dc05aec6\") " pod="metallb-system/controller-68d546b9d8-vhpgd" Oct 05 07:02:17 crc kubenswrapper[4846]: E1005 07:02:17.324310 4846 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Oct 05 07:02:17 crc kubenswrapper[4846]: E1005 07:02:17.324665 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/dad102f2-8887-4a01-98b1-be88b2836eb5-memberlist podName:dad102f2-8887-4a01-98b1-be88b2836eb5 nodeName:}" failed. No retries permitted until 2025-10-05 07:02:17.824615209 +0000 UTC m=+860.065468194 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/dad102f2-8887-4a01-98b1-be88b2836eb5-memberlist") pod "speaker-czf4m" (UID: "dad102f2-8887-4a01-98b1-be88b2836eb5") : secret "metallb-memberlist" not found Oct 05 07:02:17 crc kubenswrapper[4846]: I1005 07:02:17.324749 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/dad102f2-8887-4a01-98b1-be88b2836eb5-metrics-certs\") pod \"speaker-czf4m\" (UID: \"dad102f2-8887-4a01-98b1-be88b2836eb5\") " pod="metallb-system/speaker-czf4m" Oct 05 07:02:17 crc kubenswrapper[4846]: E1005 07:02:17.324762 4846 secret.go:188] Couldn't get secret metallb-system/frr-k8s-webhook-server-cert: secret "frr-k8s-webhook-server-cert" not found Oct 05 07:02:17 crc kubenswrapper[4846]: E1005 07:02:17.324830 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/0a4508c6-6dff-467f-9cac-ef5c7b921984-cert podName:0a4508c6-6dff-467f-9cac-ef5c7b921984 nodeName:}" failed. No retries permitted until 2025-10-05 07:02:17.824810154 +0000 UTC m=+860.065662929 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/0a4508c6-6dff-467f-9cac-ef5c7b921984-cert") pod "frr-k8s-webhook-server-64bf5d555-sf9k8" (UID: "0a4508c6-6dff-467f-9cac-ef5c7b921984") : secret "frr-k8s-webhook-server-cert" not found Oct 05 07:02:17 crc kubenswrapper[4846]: I1005 07:02:17.325369 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/dad102f2-8887-4a01-98b1-be88b2836eb5-metallb-excludel2\") pod \"speaker-czf4m\" (UID: \"dad102f2-8887-4a01-98b1-be88b2836eb5\") " pod="metallb-system/speaker-czf4m" Oct 05 07:02:17 crc kubenswrapper[4846]: I1005 07:02:17.328593 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/dad102f2-8887-4a01-98b1-be88b2836eb5-metrics-certs\") pod \"speaker-czf4m\" (UID: \"dad102f2-8887-4a01-98b1-be88b2836eb5\") " pod="metallb-system/speaker-czf4m" Oct 05 07:02:17 crc kubenswrapper[4846]: I1005 07:02:17.343396 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-kz2bc" Oct 05 07:02:17 crc kubenswrapper[4846]: I1005 07:02:17.343849 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xk6gv\" (UniqueName: \"kubernetes.io/projected/0a4508c6-6dff-467f-9cac-ef5c7b921984-kube-api-access-xk6gv\") pod \"frr-k8s-webhook-server-64bf5d555-sf9k8\" (UID: \"0a4508c6-6dff-467f-9cac-ef5c7b921984\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-sf9k8" Oct 05 07:02:17 crc kubenswrapper[4846]: I1005 07:02:17.363210 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fj2k2\" (UniqueName: \"kubernetes.io/projected/dad102f2-8887-4a01-98b1-be88b2836eb5-kube-api-access-fj2k2\") pod \"speaker-czf4m\" (UID: \"dad102f2-8887-4a01-98b1-be88b2836eb5\") " pod="metallb-system/speaker-czf4m" Oct 05 07:02:17 crc kubenswrapper[4846]: I1005 07:02:17.426936 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/3541826d-961c-4259-be16-51f4dc05aec6-metrics-certs\") pod \"controller-68d546b9d8-vhpgd\" (UID: \"3541826d-961c-4259-be16-51f4dc05aec6\") " pod="metallb-system/controller-68d546b9d8-vhpgd" Oct 05 07:02:17 crc kubenswrapper[4846]: I1005 07:02:17.427038 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lqtrr\" (UniqueName: \"kubernetes.io/projected/3541826d-961c-4259-be16-51f4dc05aec6-kube-api-access-lqtrr\") pod \"controller-68d546b9d8-vhpgd\" (UID: \"3541826d-961c-4259-be16-51f4dc05aec6\") " pod="metallb-system/controller-68d546b9d8-vhpgd" Oct 05 07:02:17 crc kubenswrapper[4846]: I1005 07:02:17.427111 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/3541826d-961c-4259-be16-51f4dc05aec6-cert\") pod \"controller-68d546b9d8-vhpgd\" (UID: \"3541826d-961c-4259-be16-51f4dc05aec6\") " pod="metallb-system/controller-68d546b9d8-vhpgd" Oct 05 07:02:17 crc kubenswrapper[4846]: E1005 07:02:17.427327 4846 secret.go:188] Couldn't get secret metallb-system/controller-certs-secret: secret "controller-certs-secret" not found Oct 05 07:02:17 crc kubenswrapper[4846]: E1005 07:02:17.427395 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/3541826d-961c-4259-be16-51f4dc05aec6-metrics-certs podName:3541826d-961c-4259-be16-51f4dc05aec6 nodeName:}" failed. No retries permitted until 2025-10-05 07:02:17.927370224 +0000 UTC m=+860.168222999 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/3541826d-961c-4259-be16-51f4dc05aec6-metrics-certs") pod "controller-68d546b9d8-vhpgd" (UID: "3541826d-961c-4259-be16-51f4dc05aec6") : secret "controller-certs-secret" not found Oct 05 07:02:17 crc kubenswrapper[4846]: I1005 07:02:17.432032 4846 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Oct 05 07:02:17 crc kubenswrapper[4846]: I1005 07:02:17.442157 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/3541826d-961c-4259-be16-51f4dc05aec6-cert\") pod \"controller-68d546b9d8-vhpgd\" (UID: \"3541826d-961c-4259-be16-51f4dc05aec6\") " pod="metallb-system/controller-68d546b9d8-vhpgd" Oct 05 07:02:17 crc kubenswrapper[4846]: I1005 07:02:17.444940 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lqtrr\" (UniqueName: \"kubernetes.io/projected/3541826d-961c-4259-be16-51f4dc05aec6-kube-api-access-lqtrr\") pod \"controller-68d546b9d8-vhpgd\" (UID: \"3541826d-961c-4259-be16-51f4dc05aec6\") " pod="metallb-system/controller-68d546b9d8-vhpgd" Oct 05 07:02:17 crc kubenswrapper[4846]: I1005 07:02:17.832309 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/0a4508c6-6dff-467f-9cac-ef5c7b921984-cert\") pod \"frr-k8s-webhook-server-64bf5d555-sf9k8\" (UID: \"0a4508c6-6dff-467f-9cac-ef5c7b921984\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-sf9k8" Oct 05 07:02:17 crc kubenswrapper[4846]: I1005 07:02:17.832817 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/dad102f2-8887-4a01-98b1-be88b2836eb5-memberlist\") pod \"speaker-czf4m\" (UID: \"dad102f2-8887-4a01-98b1-be88b2836eb5\") " pod="metallb-system/speaker-czf4m" Oct 05 07:02:17 crc kubenswrapper[4846]: E1005 07:02:17.833106 4846 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Oct 05 07:02:17 crc kubenswrapper[4846]: E1005 07:02:17.833287 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/dad102f2-8887-4a01-98b1-be88b2836eb5-memberlist podName:dad102f2-8887-4a01-98b1-be88b2836eb5 nodeName:}" failed. No retries permitted until 2025-10-05 07:02:18.833254327 +0000 UTC m=+861.074107142 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/dad102f2-8887-4a01-98b1-be88b2836eb5-memberlist") pod "speaker-czf4m" (UID: "dad102f2-8887-4a01-98b1-be88b2836eb5") : secret "metallb-memberlist" not found Oct 05 07:02:17 crc kubenswrapper[4846]: I1005 07:02:17.837827 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/0a4508c6-6dff-467f-9cac-ef5c7b921984-cert\") pod \"frr-k8s-webhook-server-64bf5d555-sf9k8\" (UID: \"0a4508c6-6dff-467f-9cac-ef5c7b921984\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-sf9k8" Oct 05 07:02:17 crc kubenswrapper[4846]: I1005 07:02:17.933980 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/3541826d-961c-4259-be16-51f4dc05aec6-metrics-certs\") pod \"controller-68d546b9d8-vhpgd\" (UID: \"3541826d-961c-4259-be16-51f4dc05aec6\") " pod="metallb-system/controller-68d546b9d8-vhpgd" Oct 05 07:02:17 crc kubenswrapper[4846]: I1005 07:02:17.940047 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/3541826d-961c-4259-be16-51f4dc05aec6-metrics-certs\") pod \"controller-68d546b9d8-vhpgd\" (UID: \"3541826d-961c-4259-be16-51f4dc05aec6\") " pod="metallb-system/controller-68d546b9d8-vhpgd" Oct 05 07:02:17 crc kubenswrapper[4846]: I1005 07:02:17.953298 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-sf9k8" Oct 05 07:02:18 crc kubenswrapper[4846]: I1005 07:02:18.048087 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-68d546b9d8-vhpgd" Oct 05 07:02:18 crc kubenswrapper[4846]: I1005 07:02:18.194597 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-64bf5d555-sf9k8"] Oct 05 07:02:18 crc kubenswrapper[4846]: I1005 07:02:18.476215 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-68d546b9d8-vhpgd"] Oct 05 07:02:18 crc kubenswrapper[4846]: W1005 07:02:18.484945 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3541826d_961c_4259_be16_51f4dc05aec6.slice/crio-76256d2839e1934344c2ba96d505409037de196af0d7ab96baf773c9a3eb34c0 WatchSource:0}: Error finding container 76256d2839e1934344c2ba96d505409037de196af0d7ab96baf773c9a3eb34c0: Status 404 returned error can't find the container with id 76256d2839e1934344c2ba96d505409037de196af0d7ab96baf773c9a3eb34c0 Oct 05 07:02:18 crc kubenswrapper[4846]: I1005 07:02:18.548848 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-kz2bc" event={"ID":"ab2bc1f2-31f3-4793-926a-e649b556d8de","Type":"ContainerStarted","Data":"74f14f5ef00a544f1cbcf40e0b86384a402988d541b205f64c1d62423b5fea9e"} Oct 05 07:02:18 crc kubenswrapper[4846]: I1005 07:02:18.550522 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-sf9k8" event={"ID":"0a4508c6-6dff-467f-9cac-ef5c7b921984","Type":"ContainerStarted","Data":"5893b373ae59c170a57a0477be8c11859c03e48b389a3d816eac1c581245dee2"} Oct 05 07:02:18 crc kubenswrapper[4846]: I1005 07:02:18.553343 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-68d546b9d8-vhpgd" event={"ID":"3541826d-961c-4259-be16-51f4dc05aec6","Type":"ContainerStarted","Data":"76256d2839e1934344c2ba96d505409037de196af0d7ab96baf773c9a3eb34c0"} Oct 05 07:02:18 crc kubenswrapper[4846]: I1005 07:02:18.852948 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/dad102f2-8887-4a01-98b1-be88b2836eb5-memberlist\") pod \"speaker-czf4m\" (UID: \"dad102f2-8887-4a01-98b1-be88b2836eb5\") " pod="metallb-system/speaker-czf4m" Oct 05 07:02:18 crc kubenswrapper[4846]: I1005 07:02:18.861023 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/dad102f2-8887-4a01-98b1-be88b2836eb5-memberlist\") pod \"speaker-czf4m\" (UID: \"dad102f2-8887-4a01-98b1-be88b2836eb5\") " pod="metallb-system/speaker-czf4m" Oct 05 07:02:18 crc kubenswrapper[4846]: I1005 07:02:18.939069 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-czf4m" Oct 05 07:02:19 crc kubenswrapper[4846]: I1005 07:02:19.573134 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-68d546b9d8-vhpgd" event={"ID":"3541826d-961c-4259-be16-51f4dc05aec6","Type":"ContainerStarted","Data":"15e4f35b0de5a93722aeaa87e4e751bfe539372704eb6f3ce9a1ea87e452556c"} Oct 05 07:02:19 crc kubenswrapper[4846]: I1005 07:02:19.573632 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-68d546b9d8-vhpgd" event={"ID":"3541826d-961c-4259-be16-51f4dc05aec6","Type":"ContainerStarted","Data":"37550fbe4c81154e6f6e8503e849b4f34c2c649a1b3ccfd4b8c27f7bb59cded1"} Oct 05 07:02:19 crc kubenswrapper[4846]: I1005 07:02:19.574515 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/controller-68d546b9d8-vhpgd" Oct 05 07:02:19 crc kubenswrapper[4846]: I1005 07:02:19.576496 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-czf4m" event={"ID":"dad102f2-8887-4a01-98b1-be88b2836eb5","Type":"ContainerStarted","Data":"0b210b5da98229af9acd5a1a20376280fa54e4dc962070e963bff87acf912579"} Oct 05 07:02:19 crc kubenswrapper[4846]: I1005 07:02:19.576528 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-czf4m" event={"ID":"dad102f2-8887-4a01-98b1-be88b2836eb5","Type":"ContainerStarted","Data":"f616ded81d1b3e346f3c8c08237b7371f7944ce23d9687150e162674441dd793"} Oct 05 07:02:19 crc kubenswrapper[4846]: I1005 07:02:19.599482 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/controller-68d546b9d8-vhpgd" podStartSLOduration=2.599459811 podStartE2EDuration="2.599459811s" podCreationTimestamp="2025-10-05 07:02:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 07:02:19.597564821 +0000 UTC m=+861.838417596" watchObservedRunningTime="2025-10-05 07:02:19.599459811 +0000 UTC m=+861.840312586" Oct 05 07:02:20 crc kubenswrapper[4846]: I1005 07:02:20.600043 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-czf4m" event={"ID":"dad102f2-8887-4a01-98b1-be88b2836eb5","Type":"ContainerStarted","Data":"862c9eb8cb8e94cff1b32293f5c7c6f992566be6103fcd766a244f862a47a771"} Oct 05 07:02:20 crc kubenswrapper[4846]: I1005 07:02:20.600342 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/speaker-czf4m" Oct 05 07:02:20 crc kubenswrapper[4846]: I1005 07:02:20.625161 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/speaker-czf4m" podStartSLOduration=3.625139354 podStartE2EDuration="3.625139354s" podCreationTimestamp="2025-10-05 07:02:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 07:02:20.622067732 +0000 UTC m=+862.862920527" watchObservedRunningTime="2025-10-05 07:02:20.625139354 +0000 UTC m=+862.865992129" Oct 05 07:02:25 crc kubenswrapper[4846]: I1005 07:02:25.635587 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-sf9k8" event={"ID":"0a4508c6-6dff-467f-9cac-ef5c7b921984","Type":"ContainerStarted","Data":"23ed2139046dc4ec4292cde62b816c4101cffae13068c35cb212ae1c2d07ce1d"} Oct 05 07:02:25 crc kubenswrapper[4846]: I1005 07:02:25.636691 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-sf9k8" Oct 05 07:02:25 crc kubenswrapper[4846]: I1005 07:02:25.640399 4846 generic.go:334] "Generic (PLEG): container finished" podID="ab2bc1f2-31f3-4793-926a-e649b556d8de" containerID="6b359e052db131468beb500f70fcf7028a907fe0b1eabe53db5b937a0284d793" exitCode=0 Oct 05 07:02:25 crc kubenswrapper[4846]: I1005 07:02:25.640476 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-kz2bc" event={"ID":"ab2bc1f2-31f3-4793-926a-e649b556d8de","Type":"ContainerDied","Data":"6b359e052db131468beb500f70fcf7028a907fe0b1eabe53db5b937a0284d793"} Oct 05 07:02:25 crc kubenswrapper[4846]: I1005 07:02:25.661776 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-sf9k8" podStartSLOduration=1.6918938529999998 podStartE2EDuration="8.66174755s" podCreationTimestamp="2025-10-05 07:02:17 +0000 UTC" firstStartedPulling="2025-10-05 07:02:18.203974185 +0000 UTC m=+860.444826960" lastFinishedPulling="2025-10-05 07:02:25.173827842 +0000 UTC m=+867.414680657" observedRunningTime="2025-10-05 07:02:25.660344823 +0000 UTC m=+867.901197638" watchObservedRunningTime="2025-10-05 07:02:25.66174755 +0000 UTC m=+867.902600375" Oct 05 07:02:26 crc kubenswrapper[4846]: I1005 07:02:26.647786 4846 generic.go:334] "Generic (PLEG): container finished" podID="ab2bc1f2-31f3-4793-926a-e649b556d8de" containerID="5ff677841f470f7a18d3d98a2bde62d04ef8d9e374cbcfac0fe9eca609e2e7b3" exitCode=0 Oct 05 07:02:26 crc kubenswrapper[4846]: I1005 07:02:26.648851 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-kz2bc" event={"ID":"ab2bc1f2-31f3-4793-926a-e649b556d8de","Type":"ContainerDied","Data":"5ff677841f470f7a18d3d98a2bde62d04ef8d9e374cbcfac0fe9eca609e2e7b3"} Oct 05 07:02:27 crc kubenswrapper[4846]: I1005 07:02:27.660497 4846 generic.go:334] "Generic (PLEG): container finished" podID="ab2bc1f2-31f3-4793-926a-e649b556d8de" containerID="4527bd1b3d0cd65d1c32f31818d5d1c70c52f14f8703a6b9dc42ab10292f5cd9" exitCode=0 Oct 05 07:02:27 crc kubenswrapper[4846]: I1005 07:02:27.660628 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-kz2bc" event={"ID":"ab2bc1f2-31f3-4793-926a-e649b556d8de","Type":"ContainerDied","Data":"4527bd1b3d0cd65d1c32f31818d5d1c70c52f14f8703a6b9dc42ab10292f5cd9"} Oct 05 07:02:28 crc kubenswrapper[4846]: I1005 07:02:28.054079 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/controller-68d546b9d8-vhpgd" Oct 05 07:02:28 crc kubenswrapper[4846]: I1005 07:02:28.707449 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-kz2bc" event={"ID":"ab2bc1f2-31f3-4793-926a-e649b556d8de","Type":"ContainerStarted","Data":"75c8dbff78c2215b9ea09ac62489b46c5ad4ae08ab3538f774441d9ea621f85f"} Oct 05 07:02:28 crc kubenswrapper[4846]: I1005 07:02:28.708035 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-kz2bc" event={"ID":"ab2bc1f2-31f3-4793-926a-e649b556d8de","Type":"ContainerStarted","Data":"472c2048c30e876bc74b7419bfe050057d37a8b35407d2c902d417a84c6a8018"} Oct 05 07:02:28 crc kubenswrapper[4846]: I1005 07:02:28.708051 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-kz2bc" event={"ID":"ab2bc1f2-31f3-4793-926a-e649b556d8de","Type":"ContainerStarted","Data":"052433fa96d789edd33620352ab5c372242e133c299e447cf1d06d0f87345db2"} Oct 05 07:02:28 crc kubenswrapper[4846]: I1005 07:02:28.708067 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-kz2bc" event={"ID":"ab2bc1f2-31f3-4793-926a-e649b556d8de","Type":"ContainerStarted","Data":"f7e45d27bf2161e7bb21c52f064b7518e9417b690fffe6cd27bc91bb055847e4"} Oct 05 07:02:28 crc kubenswrapper[4846]: I1005 07:02:28.708082 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-kz2bc" event={"ID":"ab2bc1f2-31f3-4793-926a-e649b556d8de","Type":"ContainerStarted","Data":"b6edcaca2d9e2551d788ac3c3c760e185060205ee083c0aeb78bfbc6af476db8"} Oct 05 07:02:29 crc kubenswrapper[4846]: I1005 07:02:29.724885 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-kz2bc" event={"ID":"ab2bc1f2-31f3-4793-926a-e649b556d8de","Type":"ContainerStarted","Data":"6981a6134acc09f7e86e8a1767290fedbda9220a835c0d4e96116020b7c940b5"} Oct 05 07:02:29 crc kubenswrapper[4846]: I1005 07:02:29.725238 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-kz2bc" Oct 05 07:02:29 crc kubenswrapper[4846]: I1005 07:02:29.764657 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-kz2bc" podStartSLOduration=6.165766484 podStartE2EDuration="13.764617133s" podCreationTimestamp="2025-10-05 07:02:16 +0000 UTC" firstStartedPulling="2025-10-05 07:02:17.56547777 +0000 UTC m=+859.806330555" lastFinishedPulling="2025-10-05 07:02:25.164328429 +0000 UTC m=+867.405181204" observedRunningTime="2025-10-05 07:02:29.760241856 +0000 UTC m=+872.001094671" watchObservedRunningTime="2025-10-05 07:02:29.764617133 +0000 UTC m=+872.005469958" Oct 05 07:02:32 crc kubenswrapper[4846]: I1005 07:02:32.344842 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="metallb-system/frr-k8s-kz2bc" Oct 05 07:02:32 crc kubenswrapper[4846]: I1005 07:02:32.415150 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="metallb-system/frr-k8s-kz2bc" Oct 05 07:02:37 crc kubenswrapper[4846]: I1005 07:02:37.349442 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-kz2bc" Oct 05 07:02:37 crc kubenswrapper[4846]: I1005 07:02:37.961713 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-sf9k8" Oct 05 07:02:38 crc kubenswrapper[4846]: I1005 07:02:38.950436 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/speaker-czf4m" Oct 05 07:02:40 crc kubenswrapper[4846]: I1005 07:02:40.623253 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb697lj68"] Oct 05 07:02:40 crc kubenswrapper[4846]: I1005 07:02:40.624586 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb697lj68" Oct 05 07:02:40 crc kubenswrapper[4846]: I1005 07:02:40.626824 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Oct 05 07:02:40 crc kubenswrapper[4846]: I1005 07:02:40.640663 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb697lj68"] Oct 05 07:02:40 crc kubenswrapper[4846]: I1005 07:02:40.750968 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/57b21e86-380b-4403-a4f4-5196748e08e3-util\") pod \"695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb697lj68\" (UID: \"57b21e86-380b-4403-a4f4-5196748e08e3\") " pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb697lj68" Oct 05 07:02:40 crc kubenswrapper[4846]: I1005 07:02:40.751048 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/57b21e86-380b-4403-a4f4-5196748e08e3-bundle\") pod \"695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb697lj68\" (UID: \"57b21e86-380b-4403-a4f4-5196748e08e3\") " pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb697lj68" Oct 05 07:02:40 crc kubenswrapper[4846]: I1005 07:02:40.751090 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k962k\" (UniqueName: \"kubernetes.io/projected/57b21e86-380b-4403-a4f4-5196748e08e3-kube-api-access-k962k\") pod \"695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb697lj68\" (UID: \"57b21e86-380b-4403-a4f4-5196748e08e3\") " pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb697lj68" Oct 05 07:02:40 crc kubenswrapper[4846]: I1005 07:02:40.852269 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/57b21e86-380b-4403-a4f4-5196748e08e3-bundle\") pod \"695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb697lj68\" (UID: \"57b21e86-380b-4403-a4f4-5196748e08e3\") " pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb697lj68" Oct 05 07:02:40 crc kubenswrapper[4846]: I1005 07:02:40.852798 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k962k\" (UniqueName: \"kubernetes.io/projected/57b21e86-380b-4403-a4f4-5196748e08e3-kube-api-access-k962k\") pod \"695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb697lj68\" (UID: \"57b21e86-380b-4403-a4f4-5196748e08e3\") " pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb697lj68" Oct 05 07:02:40 crc kubenswrapper[4846]: I1005 07:02:40.852850 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/57b21e86-380b-4403-a4f4-5196748e08e3-util\") pod \"695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb697lj68\" (UID: \"57b21e86-380b-4403-a4f4-5196748e08e3\") " pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb697lj68" Oct 05 07:02:40 crc kubenswrapper[4846]: I1005 07:02:40.853056 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/57b21e86-380b-4403-a4f4-5196748e08e3-bundle\") pod \"695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb697lj68\" (UID: \"57b21e86-380b-4403-a4f4-5196748e08e3\") " pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb697lj68" Oct 05 07:02:40 crc kubenswrapper[4846]: I1005 07:02:40.853384 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/57b21e86-380b-4403-a4f4-5196748e08e3-util\") pod \"695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb697lj68\" (UID: \"57b21e86-380b-4403-a4f4-5196748e08e3\") " pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb697lj68" Oct 05 07:02:40 crc kubenswrapper[4846]: I1005 07:02:40.879521 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k962k\" (UniqueName: \"kubernetes.io/projected/57b21e86-380b-4403-a4f4-5196748e08e3-kube-api-access-k962k\") pod \"695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb697lj68\" (UID: \"57b21e86-380b-4403-a4f4-5196748e08e3\") " pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb697lj68" Oct 05 07:02:40 crc kubenswrapper[4846]: I1005 07:02:40.950171 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb697lj68" Oct 05 07:02:41 crc kubenswrapper[4846]: I1005 07:02:41.189035 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb697lj68"] Oct 05 07:02:41 crc kubenswrapper[4846]: W1005 07:02:41.197324 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod57b21e86_380b_4403_a4f4_5196748e08e3.slice/crio-fc775fcc877247c6d062cab962becd21ba5f440c458dac7061973b48d1e149b8 WatchSource:0}: Error finding container fc775fcc877247c6d062cab962becd21ba5f440c458dac7061973b48d1e149b8: Status 404 returned error can't find the container with id fc775fcc877247c6d062cab962becd21ba5f440c458dac7061973b48d1e149b8 Oct 05 07:02:41 crc kubenswrapper[4846]: I1005 07:02:41.830683 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb697lj68" event={"ID":"57b21e86-380b-4403-a4f4-5196748e08e3","Type":"ContainerStarted","Data":"fc775fcc877247c6d062cab962becd21ba5f440c458dac7061973b48d1e149b8"} Oct 05 07:02:42 crc kubenswrapper[4846]: I1005 07:02:42.846953 4846 generic.go:334] "Generic (PLEG): container finished" podID="57b21e86-380b-4403-a4f4-5196748e08e3" containerID="8ba511c720902bd0a16c04c91c98c3e993af5bcb7c7488c427c9b008c55b2047" exitCode=0 Oct 05 07:02:42 crc kubenswrapper[4846]: I1005 07:02:42.847522 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb697lj68" event={"ID":"57b21e86-380b-4403-a4f4-5196748e08e3","Type":"ContainerDied","Data":"8ba511c720902bd0a16c04c91c98c3e993af5bcb7c7488c427c9b008c55b2047"} Oct 05 07:02:46 crc kubenswrapper[4846]: I1005 07:02:46.886755 4846 generic.go:334] "Generic (PLEG): container finished" podID="57b21e86-380b-4403-a4f4-5196748e08e3" containerID="9784970031b7410dcc861e327f30dfdbd2be6859e4bb0917e6c46303d0ab45a6" exitCode=0 Oct 05 07:02:46 crc kubenswrapper[4846]: I1005 07:02:46.886868 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb697lj68" event={"ID":"57b21e86-380b-4403-a4f4-5196748e08e3","Type":"ContainerDied","Data":"9784970031b7410dcc861e327f30dfdbd2be6859e4bb0917e6c46303d0ab45a6"} Oct 05 07:02:47 crc kubenswrapper[4846]: I1005 07:02:47.898854 4846 generic.go:334] "Generic (PLEG): container finished" podID="57b21e86-380b-4403-a4f4-5196748e08e3" containerID="66945b28a0c97afb596fb1bb4dcfaf48ea4b0b3dbedc83720baa9de7a62cee92" exitCode=0 Oct 05 07:02:47 crc kubenswrapper[4846]: I1005 07:02:47.898906 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb697lj68" event={"ID":"57b21e86-380b-4403-a4f4-5196748e08e3","Type":"ContainerDied","Data":"66945b28a0c97afb596fb1bb4dcfaf48ea4b0b3dbedc83720baa9de7a62cee92"} Oct 05 07:02:49 crc kubenswrapper[4846]: I1005 07:02:49.201049 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb697lj68" Oct 05 07:02:49 crc kubenswrapper[4846]: I1005 07:02:49.259647 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/57b21e86-380b-4403-a4f4-5196748e08e3-bundle\") pod \"57b21e86-380b-4403-a4f4-5196748e08e3\" (UID: \"57b21e86-380b-4403-a4f4-5196748e08e3\") " Oct 05 07:02:49 crc kubenswrapper[4846]: I1005 07:02:49.259744 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/57b21e86-380b-4403-a4f4-5196748e08e3-util\") pod \"57b21e86-380b-4403-a4f4-5196748e08e3\" (UID: \"57b21e86-380b-4403-a4f4-5196748e08e3\") " Oct 05 07:02:49 crc kubenswrapper[4846]: I1005 07:02:49.259830 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k962k\" (UniqueName: \"kubernetes.io/projected/57b21e86-380b-4403-a4f4-5196748e08e3-kube-api-access-k962k\") pod \"57b21e86-380b-4403-a4f4-5196748e08e3\" (UID: \"57b21e86-380b-4403-a4f4-5196748e08e3\") " Oct 05 07:02:49 crc kubenswrapper[4846]: I1005 07:02:49.261680 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57b21e86-380b-4403-a4f4-5196748e08e3-bundle" (OuterVolumeSpecName: "bundle") pod "57b21e86-380b-4403-a4f4-5196748e08e3" (UID: "57b21e86-380b-4403-a4f4-5196748e08e3"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 07:02:49 crc kubenswrapper[4846]: I1005 07:02:49.267422 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57b21e86-380b-4403-a4f4-5196748e08e3-kube-api-access-k962k" (OuterVolumeSpecName: "kube-api-access-k962k") pod "57b21e86-380b-4403-a4f4-5196748e08e3" (UID: "57b21e86-380b-4403-a4f4-5196748e08e3"). InnerVolumeSpecName "kube-api-access-k962k". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:02:49 crc kubenswrapper[4846]: I1005 07:02:49.270999 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57b21e86-380b-4403-a4f4-5196748e08e3-util" (OuterVolumeSpecName: "util") pod "57b21e86-380b-4403-a4f4-5196748e08e3" (UID: "57b21e86-380b-4403-a4f4-5196748e08e3"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 07:02:49 crc kubenswrapper[4846]: I1005 07:02:49.361463 4846 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/57b21e86-380b-4403-a4f4-5196748e08e3-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 07:02:49 crc kubenswrapper[4846]: I1005 07:02:49.361535 4846 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/57b21e86-380b-4403-a4f4-5196748e08e3-util\") on node \"crc\" DevicePath \"\"" Oct 05 07:02:49 crc kubenswrapper[4846]: I1005 07:02:49.361549 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k962k\" (UniqueName: \"kubernetes.io/projected/57b21e86-380b-4403-a4f4-5196748e08e3-kube-api-access-k962k\") on node \"crc\" DevicePath \"\"" Oct 05 07:02:49 crc kubenswrapper[4846]: I1005 07:02:49.919393 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb697lj68" event={"ID":"57b21e86-380b-4403-a4f4-5196748e08e3","Type":"ContainerDied","Data":"fc775fcc877247c6d062cab962becd21ba5f440c458dac7061973b48d1e149b8"} Oct 05 07:02:49 crc kubenswrapper[4846]: I1005 07:02:49.919808 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb697lj68" Oct 05 07:02:49 crc kubenswrapper[4846]: I1005 07:02:49.919839 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fc775fcc877247c6d062cab962becd21ba5f440c458dac7061973b48d1e149b8" Oct 05 07:02:54 crc kubenswrapper[4846]: I1005 07:02:54.840541 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager-operator/cert-manager-operator-controller-manager-57cd46d6d-knsqr"] Oct 05 07:02:54 crc kubenswrapper[4846]: E1005 07:02:54.841420 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="57b21e86-380b-4403-a4f4-5196748e08e3" containerName="pull" Oct 05 07:02:54 crc kubenswrapper[4846]: I1005 07:02:54.841437 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="57b21e86-380b-4403-a4f4-5196748e08e3" containerName="pull" Oct 05 07:02:54 crc kubenswrapper[4846]: E1005 07:02:54.841448 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="57b21e86-380b-4403-a4f4-5196748e08e3" containerName="extract" Oct 05 07:02:54 crc kubenswrapper[4846]: I1005 07:02:54.841455 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="57b21e86-380b-4403-a4f4-5196748e08e3" containerName="extract" Oct 05 07:02:54 crc kubenswrapper[4846]: E1005 07:02:54.841477 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="57b21e86-380b-4403-a4f4-5196748e08e3" containerName="util" Oct 05 07:02:54 crc kubenswrapper[4846]: I1005 07:02:54.841485 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="57b21e86-380b-4403-a4f4-5196748e08e3" containerName="util" Oct 05 07:02:54 crc kubenswrapper[4846]: I1005 07:02:54.841807 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="57b21e86-380b-4403-a4f4-5196748e08e3" containerName="extract" Oct 05 07:02:54 crc kubenswrapper[4846]: I1005 07:02:54.846718 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager-operator/cert-manager-operator-controller-manager-57cd46d6d-knsqr" Oct 05 07:02:54 crc kubenswrapper[4846]: I1005 07:02:54.854974 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lphpr\" (UniqueName: \"kubernetes.io/projected/19c3377a-4b96-46f3-b85e-eb4ab4f9146f-kube-api-access-lphpr\") pod \"cert-manager-operator-controller-manager-57cd46d6d-knsqr\" (UID: \"19c3377a-4b96-46f3-b85e-eb4ab4f9146f\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-57cd46d6d-knsqr" Oct 05 07:02:54 crc kubenswrapper[4846]: I1005 07:02:54.857936 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager-operator"/"openshift-service-ca.crt" Oct 05 07:02:54 crc kubenswrapper[4846]: I1005 07:02:54.864900 4846 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager-operator"/"cert-manager-operator-controller-manager-dockercfg-tk4bh" Oct 05 07:02:54 crc kubenswrapper[4846]: I1005 07:02:54.865367 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager-operator"/"kube-root-ca.crt" Oct 05 07:02:54 crc kubenswrapper[4846]: I1005 07:02:54.870661 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager-operator/cert-manager-operator-controller-manager-57cd46d6d-knsqr"] Oct 05 07:02:54 crc kubenswrapper[4846]: I1005 07:02:54.956856 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lphpr\" (UniqueName: \"kubernetes.io/projected/19c3377a-4b96-46f3-b85e-eb4ab4f9146f-kube-api-access-lphpr\") pod \"cert-manager-operator-controller-manager-57cd46d6d-knsqr\" (UID: \"19c3377a-4b96-46f3-b85e-eb4ab4f9146f\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-57cd46d6d-knsqr" Oct 05 07:02:54 crc kubenswrapper[4846]: I1005 07:02:54.978533 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lphpr\" (UniqueName: \"kubernetes.io/projected/19c3377a-4b96-46f3-b85e-eb4ab4f9146f-kube-api-access-lphpr\") pod \"cert-manager-operator-controller-manager-57cd46d6d-knsqr\" (UID: \"19c3377a-4b96-46f3-b85e-eb4ab4f9146f\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-57cd46d6d-knsqr" Oct 05 07:02:55 crc kubenswrapper[4846]: I1005 07:02:55.214333 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager-operator/cert-manager-operator-controller-manager-57cd46d6d-knsqr" Oct 05 07:02:55 crc kubenswrapper[4846]: I1005 07:02:55.526723 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager-operator/cert-manager-operator-controller-manager-57cd46d6d-knsqr"] Oct 05 07:02:55 crc kubenswrapper[4846]: I1005 07:02:55.964116 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager-operator/cert-manager-operator-controller-manager-57cd46d6d-knsqr" event={"ID":"19c3377a-4b96-46f3-b85e-eb4ab4f9146f","Type":"ContainerStarted","Data":"2eca4d14707aeac12d5882d47659499e2576be1646d15e401177b34bec19da6e"} Oct 05 07:03:03 crc kubenswrapper[4846]: I1005 07:03:03.023657 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager-operator/cert-manager-operator-controller-manager-57cd46d6d-knsqr" event={"ID":"19c3377a-4b96-46f3-b85e-eb4ab4f9146f","Type":"ContainerStarted","Data":"b1839e69243c8565d4ccd38266bee4e4d95ed42068a2ba16d1d0bb5e396f1167"} Oct 05 07:03:03 crc kubenswrapper[4846]: I1005 07:03:03.046465 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager-operator/cert-manager-operator-controller-manager-57cd46d6d-knsqr" podStartSLOduration=2.152053621 podStartE2EDuration="9.046440601s" podCreationTimestamp="2025-10-05 07:02:54 +0000 UTC" firstStartedPulling="2025-10-05 07:02:55.538410897 +0000 UTC m=+897.779263662" lastFinishedPulling="2025-10-05 07:03:02.432797867 +0000 UTC m=+904.673650642" observedRunningTime="2025-10-05 07:03:03.04377999 +0000 UTC m=+905.284632765" watchObservedRunningTime="2025-10-05 07:03:03.046440601 +0000 UTC m=+905.287293366" Oct 05 07:03:07 crc kubenswrapper[4846]: I1005 07:03:07.122697 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-webhook-d969966f-fb5qg"] Oct 05 07:03:07 crc kubenswrapper[4846]: I1005 07:03:07.124679 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-d969966f-fb5qg" Oct 05 07:03:07 crc kubenswrapper[4846]: I1005 07:03:07.127763 4846 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-webhook-dockercfg-mxqdg" Oct 05 07:03:07 crc kubenswrapper[4846]: I1005 07:03:07.128259 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"kube-root-ca.crt" Oct 05 07:03:07 crc kubenswrapper[4846]: I1005 07:03:07.133635 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-d969966f-fb5qg"] Oct 05 07:03:07 crc kubenswrapper[4846]: I1005 07:03:07.139002 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"openshift-service-ca.crt" Oct 05 07:03:07 crc kubenswrapper[4846]: I1005 07:03:07.269815 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/4823450d-b1c6-4d08-9796-109059819ec0-bound-sa-token\") pod \"cert-manager-webhook-d969966f-fb5qg\" (UID: \"4823450d-b1c6-4d08-9796-109059819ec0\") " pod="cert-manager/cert-manager-webhook-d969966f-fb5qg" Oct 05 07:03:07 crc kubenswrapper[4846]: I1005 07:03:07.270036 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zxrxh\" (UniqueName: \"kubernetes.io/projected/4823450d-b1c6-4d08-9796-109059819ec0-kube-api-access-zxrxh\") pod \"cert-manager-webhook-d969966f-fb5qg\" (UID: \"4823450d-b1c6-4d08-9796-109059819ec0\") " pod="cert-manager/cert-manager-webhook-d969966f-fb5qg" Oct 05 07:03:07 crc kubenswrapper[4846]: I1005 07:03:07.371866 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/4823450d-b1c6-4d08-9796-109059819ec0-bound-sa-token\") pod \"cert-manager-webhook-d969966f-fb5qg\" (UID: \"4823450d-b1c6-4d08-9796-109059819ec0\") " pod="cert-manager/cert-manager-webhook-d969966f-fb5qg" Oct 05 07:03:07 crc kubenswrapper[4846]: I1005 07:03:07.372069 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zxrxh\" (UniqueName: \"kubernetes.io/projected/4823450d-b1c6-4d08-9796-109059819ec0-kube-api-access-zxrxh\") pod \"cert-manager-webhook-d969966f-fb5qg\" (UID: \"4823450d-b1c6-4d08-9796-109059819ec0\") " pod="cert-manager/cert-manager-webhook-d969966f-fb5qg" Oct 05 07:03:07 crc kubenswrapper[4846]: I1005 07:03:07.399574 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/4823450d-b1c6-4d08-9796-109059819ec0-bound-sa-token\") pod \"cert-manager-webhook-d969966f-fb5qg\" (UID: \"4823450d-b1c6-4d08-9796-109059819ec0\") " pod="cert-manager/cert-manager-webhook-d969966f-fb5qg" Oct 05 07:03:07 crc kubenswrapper[4846]: I1005 07:03:07.404606 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zxrxh\" (UniqueName: \"kubernetes.io/projected/4823450d-b1c6-4d08-9796-109059819ec0-kube-api-access-zxrxh\") pod \"cert-manager-webhook-d969966f-fb5qg\" (UID: \"4823450d-b1c6-4d08-9796-109059819ec0\") " pod="cert-manager/cert-manager-webhook-d969966f-fb5qg" Oct 05 07:03:07 crc kubenswrapper[4846]: I1005 07:03:07.454873 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-d969966f-fb5qg" Oct 05 07:03:07 crc kubenswrapper[4846]: I1005 07:03:07.906112 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-d969966f-fb5qg"] Oct 05 07:03:08 crc kubenswrapper[4846]: I1005 07:03:08.058770 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-d969966f-fb5qg" event={"ID":"4823450d-b1c6-4d08-9796-109059819ec0","Type":"ContainerStarted","Data":"a5be30728a26e6ddca1c5f0a6868d01d876be9bea00cd36cd994119b5abb6894"} Oct 05 07:03:08 crc kubenswrapper[4846]: I1005 07:03:08.869753 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-cainjector-7d9f95dbf-ht8qn"] Oct 05 07:03:08 crc kubenswrapper[4846]: I1005 07:03:08.871399 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7d9f95dbf-ht8qn" Oct 05 07:03:08 crc kubenswrapper[4846]: I1005 07:03:08.875689 4846 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-cainjector-dockercfg-pb4vr" Oct 05 07:03:08 crc kubenswrapper[4846]: I1005 07:03:08.893885 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7d9f95dbf-ht8qn"] Oct 05 07:03:09 crc kubenswrapper[4846]: I1005 07:03:09.002678 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/7520b957-0745-432e-a8d5-37e2f9e7405b-bound-sa-token\") pod \"cert-manager-cainjector-7d9f95dbf-ht8qn\" (UID: \"7520b957-0745-432e-a8d5-37e2f9e7405b\") " pod="cert-manager/cert-manager-cainjector-7d9f95dbf-ht8qn" Oct 05 07:03:09 crc kubenswrapper[4846]: I1005 07:03:09.003126 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lft98\" (UniqueName: \"kubernetes.io/projected/7520b957-0745-432e-a8d5-37e2f9e7405b-kube-api-access-lft98\") pod \"cert-manager-cainjector-7d9f95dbf-ht8qn\" (UID: \"7520b957-0745-432e-a8d5-37e2f9e7405b\") " pod="cert-manager/cert-manager-cainjector-7d9f95dbf-ht8qn" Oct 05 07:03:09 crc kubenswrapper[4846]: I1005 07:03:09.105384 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/7520b957-0745-432e-a8d5-37e2f9e7405b-bound-sa-token\") pod \"cert-manager-cainjector-7d9f95dbf-ht8qn\" (UID: \"7520b957-0745-432e-a8d5-37e2f9e7405b\") " pod="cert-manager/cert-manager-cainjector-7d9f95dbf-ht8qn" Oct 05 07:03:09 crc kubenswrapper[4846]: I1005 07:03:09.105527 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lft98\" (UniqueName: \"kubernetes.io/projected/7520b957-0745-432e-a8d5-37e2f9e7405b-kube-api-access-lft98\") pod \"cert-manager-cainjector-7d9f95dbf-ht8qn\" (UID: \"7520b957-0745-432e-a8d5-37e2f9e7405b\") " pod="cert-manager/cert-manager-cainjector-7d9f95dbf-ht8qn" Oct 05 07:03:09 crc kubenswrapper[4846]: I1005 07:03:09.132905 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/7520b957-0745-432e-a8d5-37e2f9e7405b-bound-sa-token\") pod \"cert-manager-cainjector-7d9f95dbf-ht8qn\" (UID: \"7520b957-0745-432e-a8d5-37e2f9e7405b\") " pod="cert-manager/cert-manager-cainjector-7d9f95dbf-ht8qn" Oct 05 07:03:09 crc kubenswrapper[4846]: I1005 07:03:09.133241 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lft98\" (UniqueName: \"kubernetes.io/projected/7520b957-0745-432e-a8d5-37e2f9e7405b-kube-api-access-lft98\") pod \"cert-manager-cainjector-7d9f95dbf-ht8qn\" (UID: \"7520b957-0745-432e-a8d5-37e2f9e7405b\") " pod="cert-manager/cert-manager-cainjector-7d9f95dbf-ht8qn" Oct 05 07:03:09 crc kubenswrapper[4846]: I1005 07:03:09.218945 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7d9f95dbf-ht8qn" Oct 05 07:03:09 crc kubenswrapper[4846]: I1005 07:03:09.687318 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7d9f95dbf-ht8qn"] Oct 05 07:03:10 crc kubenswrapper[4846]: I1005 07:03:10.073689 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7d9f95dbf-ht8qn" event={"ID":"7520b957-0745-432e-a8d5-37e2f9e7405b","Type":"ContainerStarted","Data":"4d692db6dfa1c729e320345c2e92e05dcf45073ec1898606e3355251d831396f"} Oct 05 07:03:14 crc kubenswrapper[4846]: I1005 07:03:14.102645 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7d9f95dbf-ht8qn" event={"ID":"7520b957-0745-432e-a8d5-37e2f9e7405b","Type":"ContainerStarted","Data":"4ddcb39380f0921e500712ffcebd9dd38660456edf9927da8a3cbbe416a6211b"} Oct 05 07:03:14 crc kubenswrapper[4846]: I1005 07:03:14.104217 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-d969966f-fb5qg" event={"ID":"4823450d-b1c6-4d08-9796-109059819ec0","Type":"ContainerStarted","Data":"69b83360d18ed141a048b17bdbcab2e4698517c5904b91bc6f26d6bfd7d54945"} Oct 05 07:03:14 crc kubenswrapper[4846]: I1005 07:03:14.105060 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cert-manager/cert-manager-webhook-d969966f-fb5qg" Oct 05 07:03:14 crc kubenswrapper[4846]: I1005 07:03:14.121228 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-cainjector-7d9f95dbf-ht8qn" podStartSLOduration=2.879581268 podStartE2EDuration="6.121158865s" podCreationTimestamp="2025-10-05 07:03:08 +0000 UTC" firstStartedPulling="2025-10-05 07:03:09.70448669 +0000 UTC m=+911.945339465" lastFinishedPulling="2025-10-05 07:03:12.946064287 +0000 UTC m=+915.186917062" observedRunningTime="2025-10-05 07:03:14.119374997 +0000 UTC m=+916.360227772" watchObservedRunningTime="2025-10-05 07:03:14.121158865 +0000 UTC m=+916.362011630" Oct 05 07:03:14 crc kubenswrapper[4846]: I1005 07:03:14.165016 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-webhook-d969966f-fb5qg" podStartSLOduration=2.159291601 podStartE2EDuration="7.164982463s" podCreationTimestamp="2025-10-05 07:03:07 +0000 UTC" firstStartedPulling="2025-10-05 07:03:07.922433546 +0000 UTC m=+910.163286321" lastFinishedPulling="2025-10-05 07:03:12.928124408 +0000 UTC m=+915.168977183" observedRunningTime="2025-10-05 07:03:14.158929792 +0000 UTC m=+916.399782557" watchObservedRunningTime="2025-10-05 07:03:14.164982463 +0000 UTC m=+916.405835268" Oct 05 07:03:22 crc kubenswrapper[4846]: I1005 07:03:22.458796 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cert-manager/cert-manager-webhook-d969966f-fb5qg" Oct 05 07:03:23 crc kubenswrapper[4846]: I1005 07:03:23.326630 4846 patch_prober.go:28] interesting pod/machine-config-daemon-fscvf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 07:03:23 crc kubenswrapper[4846]: I1005 07:03:23.326738 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 07:03:25 crc kubenswrapper[4846]: I1005 07:03:25.828136 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-7d4cc89fcb-84jrb"] Oct 05 07:03:25 crc kubenswrapper[4846]: I1005 07:03:25.830286 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-7d4cc89fcb-84jrb" Oct 05 07:03:25 crc kubenswrapper[4846]: I1005 07:03:25.832765 4846 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-dockercfg-d4gtw" Oct 05 07:03:25 crc kubenswrapper[4846]: I1005 07:03:25.845522 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-7d4cc89fcb-84jrb"] Oct 05 07:03:25 crc kubenswrapper[4846]: I1005 07:03:25.907550 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a1c9433e-11a8-47fb-91ee-4da4b43facf3-bound-sa-token\") pod \"cert-manager-7d4cc89fcb-84jrb\" (UID: \"a1c9433e-11a8-47fb-91ee-4da4b43facf3\") " pod="cert-manager/cert-manager-7d4cc89fcb-84jrb" Oct 05 07:03:25 crc kubenswrapper[4846]: I1005 07:03:25.907687 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vbcxb\" (UniqueName: \"kubernetes.io/projected/a1c9433e-11a8-47fb-91ee-4da4b43facf3-kube-api-access-vbcxb\") pod \"cert-manager-7d4cc89fcb-84jrb\" (UID: \"a1c9433e-11a8-47fb-91ee-4da4b43facf3\") " pod="cert-manager/cert-manager-7d4cc89fcb-84jrb" Oct 05 07:03:26 crc kubenswrapper[4846]: I1005 07:03:26.008686 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a1c9433e-11a8-47fb-91ee-4da4b43facf3-bound-sa-token\") pod \"cert-manager-7d4cc89fcb-84jrb\" (UID: \"a1c9433e-11a8-47fb-91ee-4da4b43facf3\") " pod="cert-manager/cert-manager-7d4cc89fcb-84jrb" Oct 05 07:03:26 crc kubenswrapper[4846]: I1005 07:03:26.008779 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vbcxb\" (UniqueName: \"kubernetes.io/projected/a1c9433e-11a8-47fb-91ee-4da4b43facf3-kube-api-access-vbcxb\") pod \"cert-manager-7d4cc89fcb-84jrb\" (UID: \"a1c9433e-11a8-47fb-91ee-4da4b43facf3\") " pod="cert-manager/cert-manager-7d4cc89fcb-84jrb" Oct 05 07:03:26 crc kubenswrapper[4846]: I1005 07:03:26.046133 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a1c9433e-11a8-47fb-91ee-4da4b43facf3-bound-sa-token\") pod \"cert-manager-7d4cc89fcb-84jrb\" (UID: \"a1c9433e-11a8-47fb-91ee-4da4b43facf3\") " pod="cert-manager/cert-manager-7d4cc89fcb-84jrb" Oct 05 07:03:26 crc kubenswrapper[4846]: I1005 07:03:26.046405 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vbcxb\" (UniqueName: \"kubernetes.io/projected/a1c9433e-11a8-47fb-91ee-4da4b43facf3-kube-api-access-vbcxb\") pod \"cert-manager-7d4cc89fcb-84jrb\" (UID: \"a1c9433e-11a8-47fb-91ee-4da4b43facf3\") " pod="cert-manager/cert-manager-7d4cc89fcb-84jrb" Oct 05 07:03:26 crc kubenswrapper[4846]: I1005 07:03:26.155279 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-7d4cc89fcb-84jrb" Oct 05 07:03:26 crc kubenswrapper[4846]: I1005 07:03:26.673035 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-7d4cc89fcb-84jrb"] Oct 05 07:03:27 crc kubenswrapper[4846]: I1005 07:03:27.207052 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-7d4cc89fcb-84jrb" event={"ID":"a1c9433e-11a8-47fb-91ee-4da4b43facf3","Type":"ContainerStarted","Data":"44c6962071ac2c7350955e5c7e6d00107faba5fc1793d7c66327cb63f7c48c15"} Oct 05 07:03:27 crc kubenswrapper[4846]: I1005 07:03:27.207707 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-7d4cc89fcb-84jrb" event={"ID":"a1c9433e-11a8-47fb-91ee-4da4b43facf3","Type":"ContainerStarted","Data":"3f3c792bd07ea25ffab9171f3649296509a1b1e6ca814f81e74071a472184f72"} Oct 05 07:03:27 crc kubenswrapper[4846]: I1005 07:03:27.237104 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-7d4cc89fcb-84jrb" podStartSLOduration=2.237069353 podStartE2EDuration="2.237069353s" podCreationTimestamp="2025-10-05 07:03:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 07:03:27.22347206 +0000 UTC m=+929.464324835" watchObservedRunningTime="2025-10-05 07:03:27.237069353 +0000 UTC m=+929.477922128" Oct 05 07:03:37 crc kubenswrapper[4846]: I1005 07:03:37.015013 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-r9gh6"] Oct 05 07:03:37 crc kubenswrapper[4846]: I1005 07:03:37.018233 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-r9gh6" Oct 05 07:03:37 crc kubenswrapper[4846]: I1005 07:03:37.025814 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"kube-root-ca.crt" Oct 05 07:03:37 crc kubenswrapper[4846]: I1005 07:03:37.025831 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"openshift-service-ca.crt" Oct 05 07:03:37 crc kubenswrapper[4846]: I1005 07:03:37.026144 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-index-dockercfg-ljjm8" Oct 05 07:03:37 crc kubenswrapper[4846]: I1005 07:03:37.037605 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-r9gh6"] Oct 05 07:03:37 crc kubenswrapper[4846]: I1005 07:03:37.118736 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mq4c5\" (UniqueName: \"kubernetes.io/projected/166f2bf0-4a96-4ead-a583-78d502a7ab80-kube-api-access-mq4c5\") pod \"openstack-operator-index-r9gh6\" (UID: \"166f2bf0-4a96-4ead-a583-78d502a7ab80\") " pod="openstack-operators/openstack-operator-index-r9gh6" Oct 05 07:03:37 crc kubenswrapper[4846]: I1005 07:03:37.220240 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mq4c5\" (UniqueName: \"kubernetes.io/projected/166f2bf0-4a96-4ead-a583-78d502a7ab80-kube-api-access-mq4c5\") pod \"openstack-operator-index-r9gh6\" (UID: \"166f2bf0-4a96-4ead-a583-78d502a7ab80\") " pod="openstack-operators/openstack-operator-index-r9gh6" Oct 05 07:03:37 crc kubenswrapper[4846]: I1005 07:03:37.242870 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mq4c5\" (UniqueName: \"kubernetes.io/projected/166f2bf0-4a96-4ead-a583-78d502a7ab80-kube-api-access-mq4c5\") pod \"openstack-operator-index-r9gh6\" (UID: \"166f2bf0-4a96-4ead-a583-78d502a7ab80\") " pod="openstack-operators/openstack-operator-index-r9gh6" Oct 05 07:03:37 crc kubenswrapper[4846]: I1005 07:03:37.345269 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-r9gh6" Oct 05 07:03:37 crc kubenswrapper[4846]: I1005 07:03:37.644077 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-r9gh6"] Oct 05 07:03:38 crc kubenswrapper[4846]: I1005 07:03:38.306934 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-r9gh6" event={"ID":"166f2bf0-4a96-4ead-a583-78d502a7ab80","Type":"ContainerStarted","Data":"1fdf40246fc89cd4a2519240454c5d8ba38e1be23d2c0bfa93049fd908693036"} Oct 05 07:03:39 crc kubenswrapper[4846]: I1005 07:03:39.318635 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-r9gh6" event={"ID":"166f2bf0-4a96-4ead-a583-78d502a7ab80","Type":"ContainerStarted","Data":"ce6a7bec3f3045708ca96a0f36af55685fd938692a1e5bf2ecf29850340389fc"} Oct 05 07:03:39 crc kubenswrapper[4846]: I1005 07:03:39.341784 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-r9gh6" podStartSLOduration=2.404193346 podStartE2EDuration="3.341742849s" podCreationTimestamp="2025-10-05 07:03:36 +0000 UTC" firstStartedPulling="2025-10-05 07:03:37.658319405 +0000 UTC m=+939.899172200" lastFinishedPulling="2025-10-05 07:03:38.595868918 +0000 UTC m=+940.836721703" observedRunningTime="2025-10-05 07:03:39.335444131 +0000 UTC m=+941.576296976" watchObservedRunningTime="2025-10-05 07:03:39.341742849 +0000 UTC m=+941.582595674" Oct 05 07:03:40 crc kubenswrapper[4846]: I1005 07:03:40.380171 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-r9gh6"] Oct 05 07:03:40 crc kubenswrapper[4846]: I1005 07:03:40.993395 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-4d9cv"] Oct 05 07:03:40 crc kubenswrapper[4846]: I1005 07:03:40.995063 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-4d9cv" Oct 05 07:03:41 crc kubenswrapper[4846]: I1005 07:03:41.008949 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-4d9cv"] Oct 05 07:03:41 crc kubenswrapper[4846]: I1005 07:03:41.090231 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fqlpk\" (UniqueName: \"kubernetes.io/projected/21f31e13-73d1-4eb0-bd66-5878496bfdb2-kube-api-access-fqlpk\") pod \"openstack-operator-index-4d9cv\" (UID: \"21f31e13-73d1-4eb0-bd66-5878496bfdb2\") " pod="openstack-operators/openstack-operator-index-4d9cv" Oct 05 07:03:41 crc kubenswrapper[4846]: I1005 07:03:41.192380 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fqlpk\" (UniqueName: \"kubernetes.io/projected/21f31e13-73d1-4eb0-bd66-5878496bfdb2-kube-api-access-fqlpk\") pod \"openstack-operator-index-4d9cv\" (UID: \"21f31e13-73d1-4eb0-bd66-5878496bfdb2\") " pod="openstack-operators/openstack-operator-index-4d9cv" Oct 05 07:03:41 crc kubenswrapper[4846]: I1005 07:03:41.222035 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fqlpk\" (UniqueName: \"kubernetes.io/projected/21f31e13-73d1-4eb0-bd66-5878496bfdb2-kube-api-access-fqlpk\") pod \"openstack-operator-index-4d9cv\" (UID: \"21f31e13-73d1-4eb0-bd66-5878496bfdb2\") " pod="openstack-operators/openstack-operator-index-4d9cv" Oct 05 07:03:41 crc kubenswrapper[4846]: I1005 07:03:41.341555 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-4d9cv" Oct 05 07:03:41 crc kubenswrapper[4846]: I1005 07:03:41.345296 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/openstack-operator-index-r9gh6" podUID="166f2bf0-4a96-4ead-a583-78d502a7ab80" containerName="registry-server" containerID="cri-o://ce6a7bec3f3045708ca96a0f36af55685fd938692a1e5bf2ecf29850340389fc" gracePeriod=2 Oct 05 07:03:41 crc kubenswrapper[4846]: I1005 07:03:41.866993 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-r9gh6" Oct 05 07:03:41 crc kubenswrapper[4846]: I1005 07:03:41.906040 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mq4c5\" (UniqueName: \"kubernetes.io/projected/166f2bf0-4a96-4ead-a583-78d502a7ab80-kube-api-access-mq4c5\") pod \"166f2bf0-4a96-4ead-a583-78d502a7ab80\" (UID: \"166f2bf0-4a96-4ead-a583-78d502a7ab80\") " Oct 05 07:03:41 crc kubenswrapper[4846]: W1005 07:03:41.906741 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod21f31e13_73d1_4eb0_bd66_5878496bfdb2.slice/crio-98ad9bd3843cdac3ddfb46ed67b764a63bf573339daf1af98f6f0c8a743dd83b WatchSource:0}: Error finding container 98ad9bd3843cdac3ddfb46ed67b764a63bf573339daf1af98f6f0c8a743dd83b: Status 404 returned error can't find the container with id 98ad9bd3843cdac3ddfb46ed67b764a63bf573339daf1af98f6f0c8a743dd83b Oct 05 07:03:41 crc kubenswrapper[4846]: I1005 07:03:41.911357 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-4d9cv"] Oct 05 07:03:41 crc kubenswrapper[4846]: I1005 07:03:41.911634 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/166f2bf0-4a96-4ead-a583-78d502a7ab80-kube-api-access-mq4c5" (OuterVolumeSpecName: "kube-api-access-mq4c5") pod "166f2bf0-4a96-4ead-a583-78d502a7ab80" (UID: "166f2bf0-4a96-4ead-a583-78d502a7ab80"). InnerVolumeSpecName "kube-api-access-mq4c5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:03:42 crc kubenswrapper[4846]: I1005 07:03:42.008970 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mq4c5\" (UniqueName: \"kubernetes.io/projected/166f2bf0-4a96-4ead-a583-78d502a7ab80-kube-api-access-mq4c5\") on node \"crc\" DevicePath \"\"" Oct 05 07:03:42 crc kubenswrapper[4846]: I1005 07:03:42.355804 4846 generic.go:334] "Generic (PLEG): container finished" podID="166f2bf0-4a96-4ead-a583-78d502a7ab80" containerID="ce6a7bec3f3045708ca96a0f36af55685fd938692a1e5bf2ecf29850340389fc" exitCode=0 Oct 05 07:03:42 crc kubenswrapper[4846]: I1005 07:03:42.355920 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-r9gh6" Oct 05 07:03:42 crc kubenswrapper[4846]: I1005 07:03:42.355930 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-r9gh6" event={"ID":"166f2bf0-4a96-4ead-a583-78d502a7ab80","Type":"ContainerDied","Data":"ce6a7bec3f3045708ca96a0f36af55685fd938692a1e5bf2ecf29850340389fc"} Oct 05 07:03:42 crc kubenswrapper[4846]: I1005 07:03:42.356026 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-r9gh6" event={"ID":"166f2bf0-4a96-4ead-a583-78d502a7ab80","Type":"ContainerDied","Data":"1fdf40246fc89cd4a2519240454c5d8ba38e1be23d2c0bfa93049fd908693036"} Oct 05 07:03:42 crc kubenswrapper[4846]: I1005 07:03:42.356064 4846 scope.go:117] "RemoveContainer" containerID="ce6a7bec3f3045708ca96a0f36af55685fd938692a1e5bf2ecf29850340389fc" Oct 05 07:03:42 crc kubenswrapper[4846]: I1005 07:03:42.358572 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-4d9cv" event={"ID":"21f31e13-73d1-4eb0-bd66-5878496bfdb2","Type":"ContainerStarted","Data":"98ad9bd3843cdac3ddfb46ed67b764a63bf573339daf1af98f6f0c8a743dd83b"} Oct 05 07:03:42 crc kubenswrapper[4846]: I1005 07:03:42.390393 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-r9gh6"] Oct 05 07:03:42 crc kubenswrapper[4846]: I1005 07:03:42.392489 4846 scope.go:117] "RemoveContainer" containerID="ce6a7bec3f3045708ca96a0f36af55685fd938692a1e5bf2ecf29850340389fc" Oct 05 07:03:42 crc kubenswrapper[4846]: E1005 07:03:42.393277 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ce6a7bec3f3045708ca96a0f36af55685fd938692a1e5bf2ecf29850340389fc\": container with ID starting with ce6a7bec3f3045708ca96a0f36af55685fd938692a1e5bf2ecf29850340389fc not found: ID does not exist" containerID="ce6a7bec3f3045708ca96a0f36af55685fd938692a1e5bf2ecf29850340389fc" Oct 05 07:03:42 crc kubenswrapper[4846]: I1005 07:03:42.393330 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ce6a7bec3f3045708ca96a0f36af55685fd938692a1e5bf2ecf29850340389fc"} err="failed to get container status \"ce6a7bec3f3045708ca96a0f36af55685fd938692a1e5bf2ecf29850340389fc\": rpc error: code = NotFound desc = could not find container \"ce6a7bec3f3045708ca96a0f36af55685fd938692a1e5bf2ecf29850340389fc\": container with ID starting with ce6a7bec3f3045708ca96a0f36af55685fd938692a1e5bf2ecf29850340389fc not found: ID does not exist" Oct 05 07:03:42 crc kubenswrapper[4846]: I1005 07:03:42.397337 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/openstack-operator-index-r9gh6"] Oct 05 07:03:42 crc kubenswrapper[4846]: I1005 07:03:42.507621 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="166f2bf0-4a96-4ead-a583-78d502a7ab80" path="/var/lib/kubelet/pods/166f2bf0-4a96-4ead-a583-78d502a7ab80/volumes" Oct 05 07:03:43 crc kubenswrapper[4846]: I1005 07:03:43.369750 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-4d9cv" event={"ID":"21f31e13-73d1-4eb0-bd66-5878496bfdb2","Type":"ContainerStarted","Data":"fe3b53b17bc60cf5c5f00f0ece12f4d308ab2e78a69a7d48cf852fd6abe77906"} Oct 05 07:03:43 crc kubenswrapper[4846]: I1005 07:03:43.394683 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-4d9cv" podStartSLOduration=2.910573914 podStartE2EDuration="3.394652703s" podCreationTimestamp="2025-10-05 07:03:40 +0000 UTC" firstStartedPulling="2025-10-05 07:03:41.911483449 +0000 UTC m=+944.152336264" lastFinishedPulling="2025-10-05 07:03:42.395562268 +0000 UTC m=+944.636415053" observedRunningTime="2025-10-05 07:03:43.392621119 +0000 UTC m=+945.633473944" watchObservedRunningTime="2025-10-05 07:03:43.394652703 +0000 UTC m=+945.635505518" Oct 05 07:03:51 crc kubenswrapper[4846]: I1005 07:03:51.341922 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/openstack-operator-index-4d9cv" Oct 05 07:03:51 crc kubenswrapper[4846]: I1005 07:03:51.343110 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-index-4d9cv" Oct 05 07:03:51 crc kubenswrapper[4846]: I1005 07:03:51.400896 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/openstack-operator-index-4d9cv" Oct 05 07:03:51 crc kubenswrapper[4846]: I1005 07:03:51.495452 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-index-4d9cv" Oct 05 07:03:52 crc kubenswrapper[4846]: I1005 07:03:52.636003 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/0da5b2e9368362304c733e302da704023e73b0b2df8ed109170f4705a8lxrbm"] Oct 05 07:03:52 crc kubenswrapper[4846]: E1005 07:03:52.638296 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="166f2bf0-4a96-4ead-a583-78d502a7ab80" containerName="registry-server" Oct 05 07:03:52 crc kubenswrapper[4846]: I1005 07:03:52.638403 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="166f2bf0-4a96-4ead-a583-78d502a7ab80" containerName="registry-server" Oct 05 07:03:52 crc kubenswrapper[4846]: I1005 07:03:52.638905 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="166f2bf0-4a96-4ead-a583-78d502a7ab80" containerName="registry-server" Oct 05 07:03:52 crc kubenswrapper[4846]: I1005 07:03:52.640267 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/0da5b2e9368362304c733e302da704023e73b0b2df8ed109170f4705a8lxrbm" Oct 05 07:03:52 crc kubenswrapper[4846]: I1005 07:03:52.644507 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-9ddtb" Oct 05 07:03:52 crc kubenswrapper[4846]: I1005 07:03:52.657181 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/0da5b2e9368362304c733e302da704023e73b0b2df8ed109170f4705a8lxrbm"] Oct 05 07:03:52 crc kubenswrapper[4846]: I1005 07:03:52.803419 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/1a853c3b-02bb-4ad2-8f91-2629d9eade31-bundle\") pod \"0da5b2e9368362304c733e302da704023e73b0b2df8ed109170f4705a8lxrbm\" (UID: \"1a853c3b-02bb-4ad2-8f91-2629d9eade31\") " pod="openstack-operators/0da5b2e9368362304c733e302da704023e73b0b2df8ed109170f4705a8lxrbm" Oct 05 07:03:52 crc kubenswrapper[4846]: I1005 07:03:52.803528 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/1a853c3b-02bb-4ad2-8f91-2629d9eade31-util\") pod \"0da5b2e9368362304c733e302da704023e73b0b2df8ed109170f4705a8lxrbm\" (UID: \"1a853c3b-02bb-4ad2-8f91-2629d9eade31\") " pod="openstack-operators/0da5b2e9368362304c733e302da704023e73b0b2df8ed109170f4705a8lxrbm" Oct 05 07:03:52 crc kubenswrapper[4846]: I1005 07:03:52.803665 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xtf72\" (UniqueName: \"kubernetes.io/projected/1a853c3b-02bb-4ad2-8f91-2629d9eade31-kube-api-access-xtf72\") pod \"0da5b2e9368362304c733e302da704023e73b0b2df8ed109170f4705a8lxrbm\" (UID: \"1a853c3b-02bb-4ad2-8f91-2629d9eade31\") " pod="openstack-operators/0da5b2e9368362304c733e302da704023e73b0b2df8ed109170f4705a8lxrbm" Oct 05 07:03:52 crc kubenswrapper[4846]: I1005 07:03:52.906292 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xtf72\" (UniqueName: \"kubernetes.io/projected/1a853c3b-02bb-4ad2-8f91-2629d9eade31-kube-api-access-xtf72\") pod \"0da5b2e9368362304c733e302da704023e73b0b2df8ed109170f4705a8lxrbm\" (UID: \"1a853c3b-02bb-4ad2-8f91-2629d9eade31\") " pod="openstack-operators/0da5b2e9368362304c733e302da704023e73b0b2df8ed109170f4705a8lxrbm" Oct 05 07:03:52 crc kubenswrapper[4846]: I1005 07:03:52.906477 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/1a853c3b-02bb-4ad2-8f91-2629d9eade31-bundle\") pod \"0da5b2e9368362304c733e302da704023e73b0b2df8ed109170f4705a8lxrbm\" (UID: \"1a853c3b-02bb-4ad2-8f91-2629d9eade31\") " pod="openstack-operators/0da5b2e9368362304c733e302da704023e73b0b2df8ed109170f4705a8lxrbm" Oct 05 07:03:52 crc kubenswrapper[4846]: I1005 07:03:52.906534 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/1a853c3b-02bb-4ad2-8f91-2629d9eade31-util\") pod \"0da5b2e9368362304c733e302da704023e73b0b2df8ed109170f4705a8lxrbm\" (UID: \"1a853c3b-02bb-4ad2-8f91-2629d9eade31\") " pod="openstack-operators/0da5b2e9368362304c733e302da704023e73b0b2df8ed109170f4705a8lxrbm" Oct 05 07:03:52 crc kubenswrapper[4846]: I1005 07:03:52.907466 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/1a853c3b-02bb-4ad2-8f91-2629d9eade31-util\") pod \"0da5b2e9368362304c733e302da704023e73b0b2df8ed109170f4705a8lxrbm\" (UID: \"1a853c3b-02bb-4ad2-8f91-2629d9eade31\") " pod="openstack-operators/0da5b2e9368362304c733e302da704023e73b0b2df8ed109170f4705a8lxrbm" Oct 05 07:03:52 crc kubenswrapper[4846]: I1005 07:03:52.907607 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/1a853c3b-02bb-4ad2-8f91-2629d9eade31-bundle\") pod \"0da5b2e9368362304c733e302da704023e73b0b2df8ed109170f4705a8lxrbm\" (UID: \"1a853c3b-02bb-4ad2-8f91-2629d9eade31\") " pod="openstack-operators/0da5b2e9368362304c733e302da704023e73b0b2df8ed109170f4705a8lxrbm" Oct 05 07:03:52 crc kubenswrapper[4846]: I1005 07:03:52.941557 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xtf72\" (UniqueName: \"kubernetes.io/projected/1a853c3b-02bb-4ad2-8f91-2629d9eade31-kube-api-access-xtf72\") pod \"0da5b2e9368362304c733e302da704023e73b0b2df8ed109170f4705a8lxrbm\" (UID: \"1a853c3b-02bb-4ad2-8f91-2629d9eade31\") " pod="openstack-operators/0da5b2e9368362304c733e302da704023e73b0b2df8ed109170f4705a8lxrbm" Oct 05 07:03:52 crc kubenswrapper[4846]: I1005 07:03:52.972266 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/0da5b2e9368362304c733e302da704023e73b0b2df8ed109170f4705a8lxrbm" Oct 05 07:03:53 crc kubenswrapper[4846]: I1005 07:03:53.289484 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/0da5b2e9368362304c733e302da704023e73b0b2df8ed109170f4705a8lxrbm"] Oct 05 07:03:53 crc kubenswrapper[4846]: I1005 07:03:53.325665 4846 patch_prober.go:28] interesting pod/machine-config-daemon-fscvf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 07:03:53 crc kubenswrapper[4846]: I1005 07:03:53.325758 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 07:03:53 crc kubenswrapper[4846]: I1005 07:03:53.461549 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/0da5b2e9368362304c733e302da704023e73b0b2df8ed109170f4705a8lxrbm" event={"ID":"1a853c3b-02bb-4ad2-8f91-2629d9eade31","Type":"ContainerStarted","Data":"ec069d5d8ee9b459d00c749bfc68828572a70f8d9ea1107a4bcd9f6f65d64bc9"} Oct 05 07:03:53 crc kubenswrapper[4846]: I1005 07:03:53.461651 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/0da5b2e9368362304c733e302da704023e73b0b2df8ed109170f4705a8lxrbm" event={"ID":"1a853c3b-02bb-4ad2-8f91-2629d9eade31","Type":"ContainerStarted","Data":"9d86f9f05c0e495cd3cab0bc873fb319d1537442cbf9cf162a2629db7d2234a3"} Oct 05 07:03:54 crc kubenswrapper[4846]: I1005 07:03:54.475347 4846 generic.go:334] "Generic (PLEG): container finished" podID="1a853c3b-02bb-4ad2-8f91-2629d9eade31" containerID="ec069d5d8ee9b459d00c749bfc68828572a70f8d9ea1107a4bcd9f6f65d64bc9" exitCode=0 Oct 05 07:03:54 crc kubenswrapper[4846]: I1005 07:03:54.475645 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/0da5b2e9368362304c733e302da704023e73b0b2df8ed109170f4705a8lxrbm" event={"ID":"1a853c3b-02bb-4ad2-8f91-2629d9eade31","Type":"ContainerDied","Data":"ec069d5d8ee9b459d00c749bfc68828572a70f8d9ea1107a4bcd9f6f65d64bc9"} Oct 05 07:03:55 crc kubenswrapper[4846]: I1005 07:03:55.489599 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/0da5b2e9368362304c733e302da704023e73b0b2df8ed109170f4705a8lxrbm" event={"ID":"1a853c3b-02bb-4ad2-8f91-2629d9eade31","Type":"ContainerStarted","Data":"b3a5e00c8bb0d259b453ea2c7571764840edbadfbda737a265806ee8703e5682"} Oct 05 07:03:56 crc kubenswrapper[4846]: I1005 07:03:56.506266 4846 generic.go:334] "Generic (PLEG): container finished" podID="1a853c3b-02bb-4ad2-8f91-2629d9eade31" containerID="b3a5e00c8bb0d259b453ea2c7571764840edbadfbda737a265806ee8703e5682" exitCode=0 Oct 05 07:03:56 crc kubenswrapper[4846]: I1005 07:03:56.520861 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/0da5b2e9368362304c733e302da704023e73b0b2df8ed109170f4705a8lxrbm" event={"ID":"1a853c3b-02bb-4ad2-8f91-2629d9eade31","Type":"ContainerDied","Data":"b3a5e00c8bb0d259b453ea2c7571764840edbadfbda737a265806ee8703e5682"} Oct 05 07:03:57 crc kubenswrapper[4846]: I1005 07:03:57.520904 4846 generic.go:334] "Generic (PLEG): container finished" podID="1a853c3b-02bb-4ad2-8f91-2629d9eade31" containerID="8beb66e51335e5391535e1153e500948b315451414aa71e6760c5e07ae713230" exitCode=0 Oct 05 07:03:57 crc kubenswrapper[4846]: I1005 07:03:57.521042 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/0da5b2e9368362304c733e302da704023e73b0b2df8ed109170f4705a8lxrbm" event={"ID":"1a853c3b-02bb-4ad2-8f91-2629d9eade31","Type":"ContainerDied","Data":"8beb66e51335e5391535e1153e500948b315451414aa71e6760c5e07ae713230"} Oct 05 07:03:58 crc kubenswrapper[4846]: I1005 07:03:58.875893 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/0da5b2e9368362304c733e302da704023e73b0b2df8ed109170f4705a8lxrbm" Oct 05 07:03:58 crc kubenswrapper[4846]: I1005 07:03:58.912377 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/1a853c3b-02bb-4ad2-8f91-2629d9eade31-util\") pod \"1a853c3b-02bb-4ad2-8f91-2629d9eade31\" (UID: \"1a853c3b-02bb-4ad2-8f91-2629d9eade31\") " Oct 05 07:03:58 crc kubenswrapper[4846]: I1005 07:03:58.912631 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/1a853c3b-02bb-4ad2-8f91-2629d9eade31-bundle\") pod \"1a853c3b-02bb-4ad2-8f91-2629d9eade31\" (UID: \"1a853c3b-02bb-4ad2-8f91-2629d9eade31\") " Oct 05 07:03:58 crc kubenswrapper[4846]: I1005 07:03:58.912732 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xtf72\" (UniqueName: \"kubernetes.io/projected/1a853c3b-02bb-4ad2-8f91-2629d9eade31-kube-api-access-xtf72\") pod \"1a853c3b-02bb-4ad2-8f91-2629d9eade31\" (UID: \"1a853c3b-02bb-4ad2-8f91-2629d9eade31\") " Oct 05 07:03:58 crc kubenswrapper[4846]: I1005 07:03:58.914088 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1a853c3b-02bb-4ad2-8f91-2629d9eade31-bundle" (OuterVolumeSpecName: "bundle") pod "1a853c3b-02bb-4ad2-8f91-2629d9eade31" (UID: "1a853c3b-02bb-4ad2-8f91-2629d9eade31"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 07:03:58 crc kubenswrapper[4846]: I1005 07:03:58.922598 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1a853c3b-02bb-4ad2-8f91-2629d9eade31-kube-api-access-xtf72" (OuterVolumeSpecName: "kube-api-access-xtf72") pod "1a853c3b-02bb-4ad2-8f91-2629d9eade31" (UID: "1a853c3b-02bb-4ad2-8f91-2629d9eade31"). InnerVolumeSpecName "kube-api-access-xtf72". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:03:59 crc kubenswrapper[4846]: I1005 07:03:59.014250 4846 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/1a853c3b-02bb-4ad2-8f91-2629d9eade31-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 07:03:59 crc kubenswrapper[4846]: I1005 07:03:59.014729 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xtf72\" (UniqueName: \"kubernetes.io/projected/1a853c3b-02bb-4ad2-8f91-2629d9eade31-kube-api-access-xtf72\") on node \"crc\" DevicePath \"\"" Oct 05 07:03:59 crc kubenswrapper[4846]: I1005 07:03:59.298030 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1a853c3b-02bb-4ad2-8f91-2629d9eade31-util" (OuterVolumeSpecName: "util") pod "1a853c3b-02bb-4ad2-8f91-2629d9eade31" (UID: "1a853c3b-02bb-4ad2-8f91-2629d9eade31"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 07:03:59 crc kubenswrapper[4846]: I1005 07:03:59.318986 4846 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/1a853c3b-02bb-4ad2-8f91-2629d9eade31-util\") on node \"crc\" DevicePath \"\"" Oct 05 07:03:59 crc kubenswrapper[4846]: I1005 07:03:59.556021 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/0da5b2e9368362304c733e302da704023e73b0b2df8ed109170f4705a8lxrbm" event={"ID":"1a853c3b-02bb-4ad2-8f91-2629d9eade31","Type":"ContainerDied","Data":"9d86f9f05c0e495cd3cab0bc873fb319d1537442cbf9cf162a2629db7d2234a3"} Oct 05 07:03:59 crc kubenswrapper[4846]: I1005 07:03:59.556088 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9d86f9f05c0e495cd3cab0bc873fb319d1537442cbf9cf162a2629db7d2234a3" Oct 05 07:03:59 crc kubenswrapper[4846]: I1005 07:03:59.556247 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/0da5b2e9368362304c733e302da704023e73b0b2df8ed109170f4705a8lxrbm" Oct 05 07:04:05 crc kubenswrapper[4846]: I1005 07:04:05.258116 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-operator-677d5bb784-mzr42"] Oct 05 07:04:05 crc kubenswrapper[4846]: E1005 07:04:05.259399 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1a853c3b-02bb-4ad2-8f91-2629d9eade31" containerName="util" Oct 05 07:04:05 crc kubenswrapper[4846]: I1005 07:04:05.259418 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="1a853c3b-02bb-4ad2-8f91-2629d9eade31" containerName="util" Oct 05 07:04:05 crc kubenswrapper[4846]: E1005 07:04:05.259439 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1a853c3b-02bb-4ad2-8f91-2629d9eade31" containerName="extract" Oct 05 07:04:05 crc kubenswrapper[4846]: I1005 07:04:05.259448 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="1a853c3b-02bb-4ad2-8f91-2629d9eade31" containerName="extract" Oct 05 07:04:05 crc kubenswrapper[4846]: E1005 07:04:05.259465 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1a853c3b-02bb-4ad2-8f91-2629d9eade31" containerName="pull" Oct 05 07:04:05 crc kubenswrapper[4846]: I1005 07:04:05.259476 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="1a853c3b-02bb-4ad2-8f91-2629d9eade31" containerName="pull" Oct 05 07:04:05 crc kubenswrapper[4846]: I1005 07:04:05.259684 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="1a853c3b-02bb-4ad2-8f91-2629d9eade31" containerName="extract" Oct 05 07:04:05 crc kubenswrapper[4846]: I1005 07:04:05.260619 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-677d5bb784-mzr42" Oct 05 07:04:05 crc kubenswrapper[4846]: I1005 07:04:05.263227 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-operator-dockercfg-qd8f4" Oct 05 07:04:05 crc kubenswrapper[4846]: I1005 07:04:05.288760 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-677d5bb784-mzr42"] Oct 05 07:04:05 crc kubenswrapper[4846]: I1005 07:04:05.323568 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9jtcp\" (UniqueName: \"kubernetes.io/projected/af8c317e-4f15-4f06-9a50-198efd07cc4b-kube-api-access-9jtcp\") pod \"openstack-operator-controller-operator-677d5bb784-mzr42\" (UID: \"af8c317e-4f15-4f06-9a50-198efd07cc4b\") " pod="openstack-operators/openstack-operator-controller-operator-677d5bb784-mzr42" Oct 05 07:04:05 crc kubenswrapper[4846]: I1005 07:04:05.424840 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9jtcp\" (UniqueName: \"kubernetes.io/projected/af8c317e-4f15-4f06-9a50-198efd07cc4b-kube-api-access-9jtcp\") pod \"openstack-operator-controller-operator-677d5bb784-mzr42\" (UID: \"af8c317e-4f15-4f06-9a50-198efd07cc4b\") " pod="openstack-operators/openstack-operator-controller-operator-677d5bb784-mzr42" Oct 05 07:04:05 crc kubenswrapper[4846]: I1005 07:04:05.450438 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9jtcp\" (UniqueName: \"kubernetes.io/projected/af8c317e-4f15-4f06-9a50-198efd07cc4b-kube-api-access-9jtcp\") pod \"openstack-operator-controller-operator-677d5bb784-mzr42\" (UID: \"af8c317e-4f15-4f06-9a50-198efd07cc4b\") " pod="openstack-operators/openstack-operator-controller-operator-677d5bb784-mzr42" Oct 05 07:04:05 crc kubenswrapper[4846]: I1005 07:04:05.584149 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-677d5bb784-mzr42" Oct 05 07:04:05 crc kubenswrapper[4846]: I1005 07:04:05.846674 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-677d5bb784-mzr42"] Oct 05 07:04:06 crc kubenswrapper[4846]: I1005 07:04:06.616634 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-677d5bb784-mzr42" event={"ID":"af8c317e-4f15-4f06-9a50-198efd07cc4b","Type":"ContainerStarted","Data":"4fd8dc91905b91861174e16f600aff8b433aa5684a60ee3d8ba9d7dac0df8970"} Oct 05 07:04:10 crc kubenswrapper[4846]: I1005 07:04:10.648721 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-677d5bb784-mzr42" event={"ID":"af8c317e-4f15-4f06-9a50-198efd07cc4b","Type":"ContainerStarted","Data":"6171fe24194aadc3dd998f9237a0d763734a8fd21c95ef151d2eae2b3578af3e"} Oct 05 07:04:13 crc kubenswrapper[4846]: I1005 07:04:13.728480 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-677d5bb784-mzr42" event={"ID":"af8c317e-4f15-4f06-9a50-198efd07cc4b","Type":"ContainerStarted","Data":"d3437f4842229f463c44a5baa74cde485dc7d1b72f8cac47e40e0ee3cdf47e11"} Oct 05 07:04:13 crc kubenswrapper[4846]: I1005 07:04:13.729291 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-operator-677d5bb784-mzr42" Oct 05 07:04:13 crc kubenswrapper[4846]: I1005 07:04:13.791322 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-operator-677d5bb784-mzr42" podStartSLOduration=1.98595688 podStartE2EDuration="8.791291986s" podCreationTimestamp="2025-10-05 07:04:05 +0000 UTC" firstStartedPulling="2025-10-05 07:04:05.861977875 +0000 UTC m=+968.102830680" lastFinishedPulling="2025-10-05 07:04:12.667313001 +0000 UTC m=+974.908165786" observedRunningTime="2025-10-05 07:04:13.783043556 +0000 UTC m=+976.023896411" watchObservedRunningTime="2025-10-05 07:04:13.791291986 +0000 UTC m=+976.032144791" Oct 05 07:04:15 crc kubenswrapper[4846]: I1005 07:04:15.589992 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-operator-677d5bb784-mzr42" Oct 05 07:04:23 crc kubenswrapper[4846]: I1005 07:04:23.325463 4846 patch_prober.go:28] interesting pod/machine-config-daemon-fscvf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 07:04:23 crc kubenswrapper[4846]: I1005 07:04:23.326302 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 07:04:23 crc kubenswrapper[4846]: I1005 07:04:23.326370 4846 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" Oct 05 07:04:23 crc kubenswrapper[4846]: I1005 07:04:23.327269 4846 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"ec13db87cb5ffec077d726e058e8274c67177ec66498d56f437d83d9bfbe9292"} pod="openshift-machine-config-operator/machine-config-daemon-fscvf" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 05 07:04:23 crc kubenswrapper[4846]: I1005 07:04:23.327341 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" containerName="machine-config-daemon" containerID="cri-o://ec13db87cb5ffec077d726e058e8274c67177ec66498d56f437d83d9bfbe9292" gracePeriod=600 Oct 05 07:04:23 crc kubenswrapper[4846]: I1005 07:04:23.804495 4846 generic.go:334] "Generic (PLEG): container finished" podID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" containerID="ec13db87cb5ffec077d726e058e8274c67177ec66498d56f437d83d9bfbe9292" exitCode=0 Oct 05 07:04:23 crc kubenswrapper[4846]: I1005 07:04:23.804535 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" event={"ID":"ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7","Type":"ContainerDied","Data":"ec13db87cb5ffec077d726e058e8274c67177ec66498d56f437d83d9bfbe9292"} Oct 05 07:04:23 crc kubenswrapper[4846]: I1005 07:04:23.804956 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" event={"ID":"ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7","Type":"ContainerStarted","Data":"a25250ef8861b2bca088edb9f5248f80d97de22c7bacf6dfaaf8d42d30c96858"} Oct 05 07:04:23 crc kubenswrapper[4846]: I1005 07:04:23.804989 4846 scope.go:117] "RemoveContainer" containerID="468d3aa8613beb034776fd6c8b08e9b13031340dca2680a10051a40fbbb92cf2" Oct 05 07:04:32 crc kubenswrapper[4846]: I1005 07:04:32.257004 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/barbican-operator-controller-manager-5b974f6766-khk5k"] Oct 05 07:04:32 crc kubenswrapper[4846]: I1005 07:04:32.259601 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-5b974f6766-khk5k" Oct 05 07:04:32 crc kubenswrapper[4846]: I1005 07:04:32.263059 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/cinder-operator-controller-manager-84bd8f6848-b4rhd"] Oct 05 07:04:32 crc kubenswrapper[4846]: I1005 07:04:32.265031 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-84bd8f6848-b4rhd" Oct 05 07:04:32 crc kubenswrapper[4846]: I1005 07:04:32.268314 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"cinder-operator-controller-manager-dockercfg-kdcdw" Oct 05 07:04:32 crc kubenswrapper[4846]: I1005 07:04:32.273469 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"barbican-operator-controller-manager-dockercfg-jqcnh" Oct 05 07:04:32 crc kubenswrapper[4846]: I1005 07:04:32.277528 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-84bd8f6848-b4rhd"] Oct 05 07:04:32 crc kubenswrapper[4846]: I1005 07:04:32.280928 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-5b974f6766-khk5k"] Oct 05 07:04:32 crc kubenswrapper[4846]: I1005 07:04:32.289100 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/designate-operator-controller-manager-58d86cd59d-gnb7x"] Oct 05 07:04:32 crc kubenswrapper[4846]: I1005 07:04:32.290609 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-58d86cd59d-gnb7x" Oct 05 07:04:32 crc kubenswrapper[4846]: I1005 07:04:32.301765 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"designate-operator-controller-manager-dockercfg-9tcsb" Oct 05 07:04:32 crc kubenswrapper[4846]: I1005 07:04:32.303135 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-58d86cd59d-gnb7x"] Oct 05 07:04:32 crc kubenswrapper[4846]: I1005 07:04:32.316153 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-controller-manager-698456cdc6-w4fst"] Oct 05 07:04:32 crc kubenswrapper[4846]: I1005 07:04:32.317278 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-698456cdc6-w4fst" Oct 05 07:04:32 crc kubenswrapper[4846]: I1005 07:04:32.323195 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-dockercfg-8qwfx" Oct 05 07:04:32 crc kubenswrapper[4846]: I1005 07:04:32.332257 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/heat-operator-controller-manager-5c497dbdb-g4bkx"] Oct 05 07:04:32 crc kubenswrapper[4846]: I1005 07:04:32.333943 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-5c497dbdb-g4bkx" Oct 05 07:04:32 crc kubenswrapper[4846]: I1005 07:04:32.335356 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-698456cdc6-w4fst"] Oct 05 07:04:32 crc kubenswrapper[4846]: I1005 07:04:32.339557 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"heat-operator-controller-manager-dockercfg-84mxh" Oct 05 07:04:32 crc kubenswrapper[4846]: I1005 07:04:32.358659 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-5c497dbdb-g4bkx"] Oct 05 07:04:32 crc kubenswrapper[4846]: I1005 07:04:32.366648 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-controller-manager-6675647785-d2n5v"] Oct 05 07:04:32 crc kubenswrapper[4846]: I1005 07:04:32.367967 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-6675647785-d2n5v" Oct 05 07:04:32 crc kubenswrapper[4846]: I1005 07:04:32.375817 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-dockercfg-jdqrc" Oct 05 07:04:32 crc kubenswrapper[4846]: I1005 07:04:32.392474 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-6675647785-d2n5v"] Oct 05 07:04:32 crc kubenswrapper[4846]: I1005 07:04:32.406464 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-controller-manager-84788b6bc5-pg8c6"] Oct 05 07:04:32 crc kubenswrapper[4846]: I1005 07:04:32.407708 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-84788b6bc5-pg8c6" Oct 05 07:04:32 crc kubenswrapper[4846]: I1005 07:04:32.415790 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ironic-operator-controller-manager-6f5894c49f-fkx4r"] Oct 05 07:04:32 crc kubenswrapper[4846]: I1005 07:04:32.417116 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-6f5894c49f-fkx4r" Oct 05 07:04:32 crc kubenswrapper[4846]: I1005 07:04:32.419623 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-webhook-server-cert" Oct 05 07:04:32 crc kubenswrapper[4846]: I1005 07:04:32.419896 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-fddv4" Oct 05 07:04:32 crc kubenswrapper[4846]: I1005 07:04:32.426586 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ironic-operator-controller-manager-dockercfg-dfcx5" Oct 05 07:04:32 crc kubenswrapper[4846]: I1005 07:04:32.429565 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tthlf\" (UniqueName: \"kubernetes.io/projected/c1b6f754-a4fc-4981-a722-415569d3e7a3-kube-api-access-tthlf\") pod \"glance-operator-controller-manager-698456cdc6-w4fst\" (UID: \"c1b6f754-a4fc-4981-a722-415569d3e7a3\") " pod="openstack-operators/glance-operator-controller-manager-698456cdc6-w4fst" Oct 05 07:04:32 crc kubenswrapper[4846]: I1005 07:04:32.429612 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2g7lh\" (UniqueName: \"kubernetes.io/projected/f626116b-4f3a-48cb-ad75-c6550333ee6e-kube-api-access-2g7lh\") pod \"barbican-operator-controller-manager-5b974f6766-khk5k\" (UID: \"f626116b-4f3a-48cb-ad75-c6550333ee6e\") " pod="openstack-operators/barbican-operator-controller-manager-5b974f6766-khk5k" Oct 05 07:04:32 crc kubenswrapper[4846]: I1005 07:04:32.429648 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t2m8s\" (UniqueName: \"kubernetes.io/projected/ac8cdf70-77c0-44c9-b61b-11325d91b698-kube-api-access-t2m8s\") pod \"designate-operator-controller-manager-58d86cd59d-gnb7x\" (UID: \"ac8cdf70-77c0-44c9-b61b-11325d91b698\") " pod="openstack-operators/designate-operator-controller-manager-58d86cd59d-gnb7x" Oct 05 07:04:32 crc kubenswrapper[4846]: I1005 07:04:32.429668 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-whwsg\" (UniqueName: \"kubernetes.io/projected/f0825244-e759-46a4-a16b-4d993dac339d-kube-api-access-whwsg\") pod \"infra-operator-controller-manager-84788b6bc5-pg8c6\" (UID: \"f0825244-e759-46a4-a16b-4d993dac339d\") " pod="openstack-operators/infra-operator-controller-manager-84788b6bc5-pg8c6" Oct 05 07:04:32 crc kubenswrapper[4846]: I1005 07:04:32.429698 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c7dfd\" (UniqueName: \"kubernetes.io/projected/a1899a06-0cb1-44aa-ba1e-d6876051bab1-kube-api-access-c7dfd\") pod \"horizon-operator-controller-manager-6675647785-d2n5v\" (UID: \"a1899a06-0cb1-44aa-ba1e-d6876051bab1\") " pod="openstack-operators/horizon-operator-controller-manager-6675647785-d2n5v" Oct 05 07:04:32 crc kubenswrapper[4846]: I1005 07:04:32.429725 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sp2kl\" (UniqueName: \"kubernetes.io/projected/92c93055-428a-4d41-b034-a805b68f3e19-kube-api-access-sp2kl\") pod \"ironic-operator-controller-manager-6f5894c49f-fkx4r\" (UID: \"92c93055-428a-4d41-b034-a805b68f3e19\") " pod="openstack-operators/ironic-operator-controller-manager-6f5894c49f-fkx4r" Oct 05 07:04:32 crc kubenswrapper[4846]: I1005 07:04:32.429748 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9t5rr\" (UniqueName: \"kubernetes.io/projected/58088fc8-5aaa-422a-84cd-d4ef7d10e989-kube-api-access-9t5rr\") pod \"cinder-operator-controller-manager-84bd8f6848-b4rhd\" (UID: \"58088fc8-5aaa-422a-84cd-d4ef7d10e989\") " pod="openstack-operators/cinder-operator-controller-manager-84bd8f6848-b4rhd" Oct 05 07:04:32 crc kubenswrapper[4846]: I1005 07:04:32.429773 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f0825244-e759-46a4-a16b-4d993dac339d-cert\") pod \"infra-operator-controller-manager-84788b6bc5-pg8c6\" (UID: \"f0825244-e759-46a4-a16b-4d993dac339d\") " pod="openstack-operators/infra-operator-controller-manager-84788b6bc5-pg8c6" Oct 05 07:04:32 crc kubenswrapper[4846]: I1005 07:04:32.429832 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xs8rb\" (UniqueName: \"kubernetes.io/projected/931bcbda-d249-4625-a93c-bfa49dcb38ae-kube-api-access-xs8rb\") pod \"heat-operator-controller-manager-5c497dbdb-g4bkx\" (UID: \"931bcbda-d249-4625-a93c-bfa49dcb38ae\") " pod="openstack-operators/heat-operator-controller-manager-5c497dbdb-g4bkx" Oct 05 07:04:32 crc kubenswrapper[4846]: I1005 07:04:32.450144 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-84788b6bc5-pg8c6"] Oct 05 07:04:32 crc kubenswrapper[4846]: I1005 07:04:32.461505 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-6f5894c49f-fkx4r"] Oct 05 07:04:32 crc kubenswrapper[4846]: I1005 07:04:32.504748 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-controller-manager-57c9cdcf57-j9742"] Oct 05 07:04:32 crc kubenswrapper[4846]: I1005 07:04:32.528753 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-57c9cdcf57-j9742" Oct 05 07:04:32 crc kubenswrapper[4846]: I1005 07:04:32.538159 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tthlf\" (UniqueName: \"kubernetes.io/projected/c1b6f754-a4fc-4981-a722-415569d3e7a3-kube-api-access-tthlf\") pod \"glance-operator-controller-manager-698456cdc6-w4fst\" (UID: \"c1b6f754-a4fc-4981-a722-415569d3e7a3\") " pod="openstack-operators/glance-operator-controller-manager-698456cdc6-w4fst" Oct 05 07:04:32 crc kubenswrapper[4846]: I1005 07:04:32.538239 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2g7lh\" (UniqueName: \"kubernetes.io/projected/f626116b-4f3a-48cb-ad75-c6550333ee6e-kube-api-access-2g7lh\") pod \"barbican-operator-controller-manager-5b974f6766-khk5k\" (UID: \"f626116b-4f3a-48cb-ad75-c6550333ee6e\") " pod="openstack-operators/barbican-operator-controller-manager-5b974f6766-khk5k" Oct 05 07:04:32 crc kubenswrapper[4846]: I1005 07:04:32.538293 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t2m8s\" (UniqueName: \"kubernetes.io/projected/ac8cdf70-77c0-44c9-b61b-11325d91b698-kube-api-access-t2m8s\") pod \"designate-operator-controller-manager-58d86cd59d-gnb7x\" (UID: \"ac8cdf70-77c0-44c9-b61b-11325d91b698\") " pod="openstack-operators/designate-operator-controller-manager-58d86cd59d-gnb7x" Oct 05 07:04:32 crc kubenswrapper[4846]: I1005 07:04:32.538315 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-whwsg\" (UniqueName: \"kubernetes.io/projected/f0825244-e759-46a4-a16b-4d993dac339d-kube-api-access-whwsg\") pod \"infra-operator-controller-manager-84788b6bc5-pg8c6\" (UID: \"f0825244-e759-46a4-a16b-4d993dac339d\") " pod="openstack-operators/infra-operator-controller-manager-84788b6bc5-pg8c6" Oct 05 07:04:32 crc kubenswrapper[4846]: I1005 07:04:32.538344 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c7dfd\" (UniqueName: \"kubernetes.io/projected/a1899a06-0cb1-44aa-ba1e-d6876051bab1-kube-api-access-c7dfd\") pod \"horizon-operator-controller-manager-6675647785-d2n5v\" (UID: \"a1899a06-0cb1-44aa-ba1e-d6876051bab1\") " pod="openstack-operators/horizon-operator-controller-manager-6675647785-d2n5v" Oct 05 07:04:32 crc kubenswrapper[4846]: I1005 07:04:32.538398 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sp2kl\" (UniqueName: \"kubernetes.io/projected/92c93055-428a-4d41-b034-a805b68f3e19-kube-api-access-sp2kl\") pod \"ironic-operator-controller-manager-6f5894c49f-fkx4r\" (UID: \"92c93055-428a-4d41-b034-a805b68f3e19\") " pod="openstack-operators/ironic-operator-controller-manager-6f5894c49f-fkx4r" Oct 05 07:04:32 crc kubenswrapper[4846]: I1005 07:04:32.538424 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9t5rr\" (UniqueName: \"kubernetes.io/projected/58088fc8-5aaa-422a-84cd-d4ef7d10e989-kube-api-access-9t5rr\") pod \"cinder-operator-controller-manager-84bd8f6848-b4rhd\" (UID: \"58088fc8-5aaa-422a-84cd-d4ef7d10e989\") " pod="openstack-operators/cinder-operator-controller-manager-84bd8f6848-b4rhd" Oct 05 07:04:32 crc kubenswrapper[4846]: I1005 07:04:32.538471 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f0825244-e759-46a4-a16b-4d993dac339d-cert\") pod \"infra-operator-controller-manager-84788b6bc5-pg8c6\" (UID: \"f0825244-e759-46a4-a16b-4d993dac339d\") " pod="openstack-operators/infra-operator-controller-manager-84788b6bc5-pg8c6" Oct 05 07:04:32 crc kubenswrapper[4846]: I1005 07:04:32.538580 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xs8rb\" (UniqueName: \"kubernetes.io/projected/931bcbda-d249-4625-a93c-bfa49dcb38ae-kube-api-access-xs8rb\") pod \"heat-operator-controller-manager-5c497dbdb-g4bkx\" (UID: \"931bcbda-d249-4625-a93c-bfa49dcb38ae\") " pod="openstack-operators/heat-operator-controller-manager-5c497dbdb-g4bkx" Oct 05 07:04:32 crc kubenswrapper[4846]: E1005 07:04:32.541208 4846 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Oct 05 07:04:32 crc kubenswrapper[4846]: E1005 07:04:32.541298 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f0825244-e759-46a4-a16b-4d993dac339d-cert podName:f0825244-e759-46a4-a16b-4d993dac339d nodeName:}" failed. No retries permitted until 2025-10-05 07:04:33.041260913 +0000 UTC m=+995.282113688 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/f0825244-e759-46a4-a16b-4d993dac339d-cert") pod "infra-operator-controller-manager-84788b6bc5-pg8c6" (UID: "f0825244-e759-46a4-a16b-4d993dac339d") : secret "infra-operator-webhook-server-cert" not found Oct 05 07:04:32 crc kubenswrapper[4846]: I1005 07:04:32.543442 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-dockercfg-8hfgj" Oct 05 07:04:32 crc kubenswrapper[4846]: I1005 07:04:32.590681 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/manila-operator-controller-manager-7cb48dbc-cvtkp"] Oct 05 07:04:32 crc kubenswrapper[4846]: I1005 07:04:32.600155 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-7cb48dbc-cvtkp" Oct 05 07:04:32 crc kubenswrapper[4846]: I1005 07:04:32.603531 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"manila-operator-controller-manager-dockercfg-6s664" Oct 05 07:04:32 crc kubenswrapper[4846]: I1005 07:04:32.618303 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t2m8s\" (UniqueName: \"kubernetes.io/projected/ac8cdf70-77c0-44c9-b61b-11325d91b698-kube-api-access-t2m8s\") pod \"designate-operator-controller-manager-58d86cd59d-gnb7x\" (UID: \"ac8cdf70-77c0-44c9-b61b-11325d91b698\") " pod="openstack-operators/designate-operator-controller-manager-58d86cd59d-gnb7x" Oct 05 07:04:32 crc kubenswrapper[4846]: I1005 07:04:32.618829 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-58d86cd59d-gnb7x" Oct 05 07:04:32 crc kubenswrapper[4846]: I1005 07:04:32.620777 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xs8rb\" (UniqueName: \"kubernetes.io/projected/931bcbda-d249-4625-a93c-bfa49dcb38ae-kube-api-access-xs8rb\") pod \"heat-operator-controller-manager-5c497dbdb-g4bkx\" (UID: \"931bcbda-d249-4625-a93c-bfa49dcb38ae\") " pod="openstack-operators/heat-operator-controller-manager-5c497dbdb-g4bkx" Oct 05 07:04:32 crc kubenswrapper[4846]: I1005 07:04:32.624649 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-whwsg\" (UniqueName: \"kubernetes.io/projected/f0825244-e759-46a4-a16b-4d993dac339d-kube-api-access-whwsg\") pod \"infra-operator-controller-manager-84788b6bc5-pg8c6\" (UID: \"f0825244-e759-46a4-a16b-4d993dac339d\") " pod="openstack-operators/infra-operator-controller-manager-84788b6bc5-pg8c6" Oct 05 07:04:32 crc kubenswrapper[4846]: I1005 07:04:32.625968 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2g7lh\" (UniqueName: \"kubernetes.io/projected/f626116b-4f3a-48cb-ad75-c6550333ee6e-kube-api-access-2g7lh\") pod \"barbican-operator-controller-manager-5b974f6766-khk5k\" (UID: \"f626116b-4f3a-48cb-ad75-c6550333ee6e\") " pod="openstack-operators/barbican-operator-controller-manager-5b974f6766-khk5k" Oct 05 07:04:32 crc kubenswrapper[4846]: I1005 07:04:32.626694 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9t5rr\" (UniqueName: \"kubernetes.io/projected/58088fc8-5aaa-422a-84cd-d4ef7d10e989-kube-api-access-9t5rr\") pod \"cinder-operator-controller-manager-84bd8f6848-b4rhd\" (UID: \"58088fc8-5aaa-422a-84cd-d4ef7d10e989\") " pod="openstack-operators/cinder-operator-controller-manager-84bd8f6848-b4rhd" Oct 05 07:04:32 crc kubenswrapper[4846]: I1005 07:04:32.645285 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nzjlt\" (UniqueName: \"kubernetes.io/projected/6ff6c74d-2f8b-42d0-af60-88baa4399b32-kube-api-access-nzjlt\") pod \"keystone-operator-controller-manager-57c9cdcf57-j9742\" (UID: \"6ff6c74d-2f8b-42d0-af60-88baa4399b32\") " pod="openstack-operators/keystone-operator-controller-manager-57c9cdcf57-j9742" Oct 05 07:04:32 crc kubenswrapper[4846]: I1005 07:04:32.647001 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sp2kl\" (UniqueName: \"kubernetes.io/projected/92c93055-428a-4d41-b034-a805b68f3e19-kube-api-access-sp2kl\") pod \"ironic-operator-controller-manager-6f5894c49f-fkx4r\" (UID: \"92c93055-428a-4d41-b034-a805b68f3e19\") " pod="openstack-operators/ironic-operator-controller-manager-6f5894c49f-fkx4r" Oct 05 07:04:32 crc kubenswrapper[4846]: I1005 07:04:32.649227 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tthlf\" (UniqueName: \"kubernetes.io/projected/c1b6f754-a4fc-4981-a722-415569d3e7a3-kube-api-access-tthlf\") pod \"glance-operator-controller-manager-698456cdc6-w4fst\" (UID: \"c1b6f754-a4fc-4981-a722-415569d3e7a3\") " pod="openstack-operators/glance-operator-controller-manager-698456cdc6-w4fst" Oct 05 07:04:32 crc kubenswrapper[4846]: I1005 07:04:32.649855 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c7dfd\" (UniqueName: \"kubernetes.io/projected/a1899a06-0cb1-44aa-ba1e-d6876051bab1-kube-api-access-c7dfd\") pod \"horizon-operator-controller-manager-6675647785-d2n5v\" (UID: \"a1899a06-0cb1-44aa-ba1e-d6876051bab1\") " pod="openstack-operators/horizon-operator-controller-manager-6675647785-d2n5v" Oct 05 07:04:32 crc kubenswrapper[4846]: I1005 07:04:32.652400 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-57c9cdcf57-j9742"] Oct 05 07:04:32 crc kubenswrapper[4846]: I1005 07:04:32.665483 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-5c497dbdb-g4bkx" Oct 05 07:04:32 crc kubenswrapper[4846]: I1005 07:04:32.670103 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/neutron-operator-controller-manager-69b956fbf6-xsltv"] Oct 05 07:04:32 crc kubenswrapper[4846]: I1005 07:04:32.671421 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-69b956fbf6-xsltv" Oct 05 07:04:32 crc kubenswrapper[4846]: I1005 07:04:32.674949 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-d6c9dc5bc-2fb7l"] Oct 05 07:04:32 crc kubenswrapper[4846]: I1005 07:04:32.675757 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"neutron-operator-controller-manager-dockercfg-47sqx" Oct 05 07:04:32 crc kubenswrapper[4846]: I1005 07:04:32.677474 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-d6c9dc5bc-2fb7l" Oct 05 07:04:32 crc kubenswrapper[4846]: I1005 07:04:32.690780 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-dockercfg-d2skq" Oct 05 07:04:32 crc kubenswrapper[4846]: I1005 07:04:32.702251 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-6675647785-d2n5v" Oct 05 07:04:32 crc kubenswrapper[4846]: I1005 07:04:32.753536 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rmqdl\" (UniqueName: \"kubernetes.io/projected/e46909e0-1642-436c-b597-167a42e0a72d-kube-api-access-rmqdl\") pod \"manila-operator-controller-manager-7cb48dbc-cvtkp\" (UID: \"e46909e0-1642-436c-b597-167a42e0a72d\") " pod="openstack-operators/manila-operator-controller-manager-7cb48dbc-cvtkp" Oct 05 07:04:32 crc kubenswrapper[4846]: I1005 07:04:32.753624 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nzjlt\" (UniqueName: \"kubernetes.io/projected/6ff6c74d-2f8b-42d0-af60-88baa4399b32-kube-api-access-nzjlt\") pod \"keystone-operator-controller-manager-57c9cdcf57-j9742\" (UID: \"6ff6c74d-2f8b-42d0-af60-88baa4399b32\") " pod="openstack-operators/keystone-operator-controller-manager-57c9cdcf57-j9742" Oct 05 07:04:32 crc kubenswrapper[4846]: I1005 07:04:32.768565 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-d6c9dc5bc-2fb7l"] Oct 05 07:04:32 crc kubenswrapper[4846]: I1005 07:04:32.768829 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-6f5894c49f-fkx4r" Oct 05 07:04:32 crc kubenswrapper[4846]: I1005 07:04:32.778568 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-7cb48dbc-cvtkp"] Oct 05 07:04:32 crc kubenswrapper[4846]: I1005 07:04:32.786714 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nzjlt\" (UniqueName: \"kubernetes.io/projected/6ff6c74d-2f8b-42d0-af60-88baa4399b32-kube-api-access-nzjlt\") pod \"keystone-operator-controller-manager-57c9cdcf57-j9742\" (UID: \"6ff6c74d-2f8b-42d0-af60-88baa4399b32\") " pod="openstack-operators/keystone-operator-controller-manager-57c9cdcf57-j9742" Oct 05 07:04:32 crc kubenswrapper[4846]: I1005 07:04:32.795039 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/nova-operator-controller-manager-6c9b57c67-ggtrm"] Oct 05 07:04:32 crc kubenswrapper[4846]: I1005 07:04:32.799171 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-6c9b57c67-ggtrm" Oct 05 07:04:32 crc kubenswrapper[4846]: I1005 07:04:32.808444 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/octavia-operator-controller-manager-69f59f9d8-s4qf7"] Oct 05 07:04:32 crc kubenswrapper[4846]: I1005 07:04:32.811728 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"nova-operator-controller-manager-dockercfg-95ppw" Oct 05 07:04:32 crc kubenswrapper[4846]: I1005 07:04:32.811754 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-69f59f9d8-s4qf7" Oct 05 07:04:32 crc kubenswrapper[4846]: I1005 07:04:32.824467 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"octavia-operator-controller-manager-dockercfg-6l6ln" Oct 05 07:04:32 crc kubenswrapper[4846]: I1005 07:04:32.829558 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-69f59f9d8-s4qf7"] Oct 05 07:04:32 crc kubenswrapper[4846]: I1005 07:04:32.838148 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-69b956fbf6-xsltv"] Oct 05 07:04:32 crc kubenswrapper[4846]: I1005 07:04:32.855664 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-6c9b57c67-ggtrm"] Oct 05 07:04:32 crc kubenswrapper[4846]: I1005 07:04:32.856986 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ovn-operator-controller-manager-c968bb45-rmwgr"] Oct 05 07:04:32 crc kubenswrapper[4846]: I1005 07:04:32.858535 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-c968bb45-rmwgr" Oct 05 07:04:32 crc kubenswrapper[4846]: I1005 07:04:32.860828 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-57c9cdcf57-j9742" Oct 05 07:04:32 crc kubenswrapper[4846]: I1005 07:04:32.861764 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rzw4b\" (UniqueName: \"kubernetes.io/projected/15b8ca14-e689-430c-84b2-40da6f1e83a5-kube-api-access-rzw4b\") pod \"octavia-operator-controller-manager-69f59f9d8-s4qf7\" (UID: \"15b8ca14-e689-430c-84b2-40da6f1e83a5\") " pod="openstack-operators/octavia-operator-controller-manager-69f59f9d8-s4qf7" Oct 05 07:04:32 crc kubenswrapper[4846]: I1005 07:04:32.861805 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2rrpl\" (UniqueName: \"kubernetes.io/projected/644a3a04-e2d4-4352-9bac-e8d661fd61b0-kube-api-access-2rrpl\") pod \"neutron-operator-controller-manager-69b956fbf6-xsltv\" (UID: \"644a3a04-e2d4-4352-9bac-e8d661fd61b0\") " pod="openstack-operators/neutron-operator-controller-manager-69b956fbf6-xsltv" Oct 05 07:04:32 crc kubenswrapper[4846]: I1005 07:04:32.861870 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-txqrd\" (UniqueName: \"kubernetes.io/projected/71de4f09-b26f-443c-9142-7edbc0d18e38-kube-api-access-txqrd\") pod \"mariadb-operator-controller-manager-d6c9dc5bc-2fb7l\" (UID: \"71de4f09-b26f-443c-9142-7edbc0d18e38\") " pod="openstack-operators/mariadb-operator-controller-manager-d6c9dc5bc-2fb7l" Oct 05 07:04:32 crc kubenswrapper[4846]: I1005 07:04:32.861897 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rmqdl\" (UniqueName: \"kubernetes.io/projected/e46909e0-1642-436c-b597-167a42e0a72d-kube-api-access-rmqdl\") pod \"manila-operator-controller-manager-7cb48dbc-cvtkp\" (UID: \"e46909e0-1642-436c-b597-167a42e0a72d\") " pod="openstack-operators/manila-operator-controller-manager-7cb48dbc-cvtkp" Oct 05 07:04:32 crc kubenswrapper[4846]: I1005 07:04:32.865677 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ovn-operator-controller-manager-dockercfg-nz2zz" Oct 05 07:04:32 crc kubenswrapper[4846]: I1005 07:04:32.882563 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-5b974f6766-khk5k" Oct 05 07:04:32 crc kubenswrapper[4846]: I1005 07:04:32.893170 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-84bd8f6848-b4rhd" Oct 05 07:04:32 crc kubenswrapper[4846]: I1005 07:04:32.907985 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-66cc85b5d5lsczg"] Oct 05 07:04:32 crc kubenswrapper[4846]: I1005 07:04:32.909554 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-66cc85b5d5lsczg" Oct 05 07:04:32 crc kubenswrapper[4846]: I1005 07:04:32.911318 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-webhook-server-cert" Oct 05 07:04:32 crc kubenswrapper[4846]: I1005 07:04:32.912088 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-x4c6p" Oct 05 07:04:32 crc kubenswrapper[4846]: I1005 07:04:32.916949 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rmqdl\" (UniqueName: \"kubernetes.io/projected/e46909e0-1642-436c-b597-167a42e0a72d-kube-api-access-rmqdl\") pod \"manila-operator-controller-manager-7cb48dbc-cvtkp\" (UID: \"e46909e0-1642-436c-b597-167a42e0a72d\") " pod="openstack-operators/manila-operator-controller-manager-7cb48dbc-cvtkp" Oct 05 07:04:32 crc kubenswrapper[4846]: I1005 07:04:32.927797 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-c968bb45-rmwgr"] Oct 05 07:04:32 crc kubenswrapper[4846]: I1005 07:04:32.941095 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-66cc85b5d5lsczg"] Oct 05 07:04:32 crc kubenswrapper[4846]: I1005 07:04:32.941162 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/placement-operator-controller-manager-66f6d6849b-62zd4"] Oct 05 07:04:32 crc kubenswrapper[4846]: I1005 07:04:32.942392 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-66f6d6849b-62zd4" Oct 05 07:04:32 crc kubenswrapper[4846]: I1005 07:04:32.949898 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"placement-operator-controller-manager-dockercfg-v7t6l" Oct 05 07:04:32 crc kubenswrapper[4846]: I1005 07:04:32.951133 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-controller-manager-76d5577b-pkjcv"] Oct 05 07:04:32 crc kubenswrapper[4846]: I1005 07:04:32.952358 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-76d5577b-pkjcv" Oct 05 07:04:32 crc kubenswrapper[4846]: I1005 07:04:32.952386 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-698456cdc6-w4fst" Oct 05 07:04:32 crc kubenswrapper[4846]: I1005 07:04:32.970025 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-txqrd\" (UniqueName: \"kubernetes.io/projected/71de4f09-b26f-443c-9142-7edbc0d18e38-kube-api-access-txqrd\") pod \"mariadb-operator-controller-manager-d6c9dc5bc-2fb7l\" (UID: \"71de4f09-b26f-443c-9142-7edbc0d18e38\") " pod="openstack-operators/mariadb-operator-controller-manager-d6c9dc5bc-2fb7l" Oct 05 07:04:32 crc kubenswrapper[4846]: I1005 07:04:32.970097 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rzw4b\" (UniqueName: \"kubernetes.io/projected/15b8ca14-e689-430c-84b2-40da6f1e83a5-kube-api-access-rzw4b\") pod \"octavia-operator-controller-manager-69f59f9d8-s4qf7\" (UID: \"15b8ca14-e689-430c-84b2-40da6f1e83a5\") " pod="openstack-operators/octavia-operator-controller-manager-69f59f9d8-s4qf7" Oct 05 07:04:32 crc kubenswrapper[4846]: I1005 07:04:32.970159 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-57hdf\" (UniqueName: \"kubernetes.io/projected/49e72a92-e750-4046-9894-f3a0359531b2-kube-api-access-57hdf\") pod \"nova-operator-controller-manager-6c9b57c67-ggtrm\" (UID: \"49e72a92-e750-4046-9894-f3a0359531b2\") " pod="openstack-operators/nova-operator-controller-manager-6c9b57c67-ggtrm" Oct 05 07:04:32 crc kubenswrapper[4846]: I1005 07:04:32.970208 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2rrpl\" (UniqueName: \"kubernetes.io/projected/644a3a04-e2d4-4352-9bac-e8d661fd61b0-kube-api-access-2rrpl\") pod \"neutron-operator-controller-manager-69b956fbf6-xsltv\" (UID: \"644a3a04-e2d4-4352-9bac-e8d661fd61b0\") " pod="openstack-operators/neutron-operator-controller-manager-69b956fbf6-xsltv" Oct 05 07:04:32 crc kubenswrapper[4846]: I1005 07:04:32.984711 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-76d5577b-pkjcv"] Oct 05 07:04:32 crc kubenswrapper[4846]: I1005 07:04:32.991170 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-dockercfg-679lz" Oct 05 07:04:33 crc kubenswrapper[4846]: I1005 07:04:33.035037 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2rrpl\" (UniqueName: \"kubernetes.io/projected/644a3a04-e2d4-4352-9bac-e8d661fd61b0-kube-api-access-2rrpl\") pod \"neutron-operator-controller-manager-69b956fbf6-xsltv\" (UID: \"644a3a04-e2d4-4352-9bac-e8d661fd61b0\") " pod="openstack-operators/neutron-operator-controller-manager-69b956fbf6-xsltv" Oct 05 07:04:33 crc kubenswrapper[4846]: I1005 07:04:33.035527 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rzw4b\" (UniqueName: \"kubernetes.io/projected/15b8ca14-e689-430c-84b2-40da6f1e83a5-kube-api-access-rzw4b\") pod \"octavia-operator-controller-manager-69f59f9d8-s4qf7\" (UID: \"15b8ca14-e689-430c-84b2-40da6f1e83a5\") " pod="openstack-operators/octavia-operator-controller-manager-69f59f9d8-s4qf7" Oct 05 07:04:33 crc kubenswrapper[4846]: I1005 07:04:33.037617 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-7cb48dbc-cvtkp" Oct 05 07:04:33 crc kubenswrapper[4846]: I1005 07:04:33.063906 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-txqrd\" (UniqueName: \"kubernetes.io/projected/71de4f09-b26f-443c-9142-7edbc0d18e38-kube-api-access-txqrd\") pod \"mariadb-operator-controller-manager-d6c9dc5bc-2fb7l\" (UID: \"71de4f09-b26f-443c-9142-7edbc0d18e38\") " pod="openstack-operators/mariadb-operator-controller-manager-d6c9dc5bc-2fb7l" Oct 05 07:04:33 crc kubenswrapper[4846]: I1005 07:04:33.071474 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p697b\" (UniqueName: \"kubernetes.io/projected/95847a57-fbe8-43f2-b4b4-e630665f17e3-kube-api-access-p697b\") pod \"placement-operator-controller-manager-66f6d6849b-62zd4\" (UID: \"95847a57-fbe8-43f2-b4b4-e630665f17e3\") " pod="openstack-operators/placement-operator-controller-manager-66f6d6849b-62zd4" Oct 05 07:04:33 crc kubenswrapper[4846]: I1005 07:04:33.071549 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6kzfn\" (UniqueName: \"kubernetes.io/projected/eb017e3a-1627-4dc7-bda3-1faec8b72739-kube-api-access-6kzfn\") pod \"openstack-baremetal-operator-controller-manager-66cc85b5d5lsczg\" (UID: \"eb017e3a-1627-4dc7-bda3-1faec8b72739\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-66cc85b5d5lsczg" Oct 05 07:04:33 crc kubenswrapper[4846]: I1005 07:04:33.071602 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lmjch\" (UniqueName: \"kubernetes.io/projected/ab1677d6-3cb0-4dce-8e8d-4d2b67025e6d-kube-api-access-lmjch\") pod \"swift-operator-controller-manager-76d5577b-pkjcv\" (UID: \"ab1677d6-3cb0-4dce-8e8d-4d2b67025e6d\") " pod="openstack-operators/swift-operator-controller-manager-76d5577b-pkjcv" Oct 05 07:04:33 crc kubenswrapper[4846]: I1005 07:04:33.079830 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f0825244-e759-46a4-a16b-4d993dac339d-cert\") pod \"infra-operator-controller-manager-84788b6bc5-pg8c6\" (UID: \"f0825244-e759-46a4-a16b-4d993dac339d\") " pod="openstack-operators/infra-operator-controller-manager-84788b6bc5-pg8c6" Oct 05 07:04:33 crc kubenswrapper[4846]: I1005 07:04:33.079976 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v2b4v\" (UniqueName: \"kubernetes.io/projected/945ab775-ffae-41cb-a315-5e05cf65484d-kube-api-access-v2b4v\") pod \"ovn-operator-controller-manager-c968bb45-rmwgr\" (UID: \"945ab775-ffae-41cb-a315-5e05cf65484d\") " pod="openstack-operators/ovn-operator-controller-manager-c968bb45-rmwgr" Oct 05 07:04:33 crc kubenswrapper[4846]: I1005 07:04:33.080088 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-57hdf\" (UniqueName: \"kubernetes.io/projected/49e72a92-e750-4046-9894-f3a0359531b2-kube-api-access-57hdf\") pod \"nova-operator-controller-manager-6c9b57c67-ggtrm\" (UID: \"49e72a92-e750-4046-9894-f3a0359531b2\") " pod="openstack-operators/nova-operator-controller-manager-6c9b57c67-ggtrm" Oct 05 07:04:33 crc kubenswrapper[4846]: I1005 07:04:33.080156 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/eb017e3a-1627-4dc7-bda3-1faec8b72739-cert\") pod \"openstack-baremetal-operator-controller-manager-66cc85b5d5lsczg\" (UID: \"eb017e3a-1627-4dc7-bda3-1faec8b72739\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-66cc85b5d5lsczg" Oct 05 07:04:33 crc kubenswrapper[4846]: I1005 07:04:33.084373 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-66f6d6849b-62zd4"] Oct 05 07:04:33 crc kubenswrapper[4846]: I1005 07:04:33.085346 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-69b956fbf6-xsltv" Oct 05 07:04:33 crc kubenswrapper[4846]: I1005 07:04:33.105901 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f0825244-e759-46a4-a16b-4d993dac339d-cert\") pod \"infra-operator-controller-manager-84788b6bc5-pg8c6\" (UID: \"f0825244-e759-46a4-a16b-4d993dac339d\") " pod="openstack-operators/infra-operator-controller-manager-84788b6bc5-pg8c6" Oct 05 07:04:33 crc kubenswrapper[4846]: I1005 07:04:33.112462 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-f589c7597-jzsnd"] Oct 05 07:04:33 crc kubenswrapper[4846]: I1005 07:04:33.113760 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-f589c7597-jzsnd" Oct 05 07:04:33 crc kubenswrapper[4846]: I1005 07:04:33.120066 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"telemetry-operator-controller-manager-dockercfg-4bsvb" Oct 05 07:04:33 crc kubenswrapper[4846]: I1005 07:04:33.137742 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-d6c9dc5bc-2fb7l" Oct 05 07:04:33 crc kubenswrapper[4846]: I1005 07:04:33.162917 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-57hdf\" (UniqueName: \"kubernetes.io/projected/49e72a92-e750-4046-9894-f3a0359531b2-kube-api-access-57hdf\") pod \"nova-operator-controller-manager-6c9b57c67-ggtrm\" (UID: \"49e72a92-e750-4046-9894-f3a0359531b2\") " pod="openstack-operators/nova-operator-controller-manager-6c9b57c67-ggtrm" Oct 05 07:04:33 crc kubenswrapper[4846]: I1005 07:04:33.184038 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p697b\" (UniqueName: \"kubernetes.io/projected/95847a57-fbe8-43f2-b4b4-e630665f17e3-kube-api-access-p697b\") pod \"placement-operator-controller-manager-66f6d6849b-62zd4\" (UID: \"95847a57-fbe8-43f2-b4b4-e630665f17e3\") " pod="openstack-operators/placement-operator-controller-manager-66f6d6849b-62zd4" Oct 05 07:04:33 crc kubenswrapper[4846]: I1005 07:04:33.184123 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6kzfn\" (UniqueName: \"kubernetes.io/projected/eb017e3a-1627-4dc7-bda3-1faec8b72739-kube-api-access-6kzfn\") pod \"openstack-baremetal-operator-controller-manager-66cc85b5d5lsczg\" (UID: \"eb017e3a-1627-4dc7-bda3-1faec8b72739\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-66cc85b5d5lsczg" Oct 05 07:04:33 crc kubenswrapper[4846]: I1005 07:04:33.184161 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lmjch\" (UniqueName: \"kubernetes.io/projected/ab1677d6-3cb0-4dce-8e8d-4d2b67025e6d-kube-api-access-lmjch\") pod \"swift-operator-controller-manager-76d5577b-pkjcv\" (UID: \"ab1677d6-3cb0-4dce-8e8d-4d2b67025e6d\") " pod="openstack-operators/swift-operator-controller-manager-76d5577b-pkjcv" Oct 05 07:04:33 crc kubenswrapper[4846]: I1005 07:04:33.184231 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v2b4v\" (UniqueName: \"kubernetes.io/projected/945ab775-ffae-41cb-a315-5e05cf65484d-kube-api-access-v2b4v\") pod \"ovn-operator-controller-manager-c968bb45-rmwgr\" (UID: \"945ab775-ffae-41cb-a315-5e05cf65484d\") " pod="openstack-operators/ovn-operator-controller-manager-c968bb45-rmwgr" Oct 05 07:04:33 crc kubenswrapper[4846]: I1005 07:04:33.184284 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/eb017e3a-1627-4dc7-bda3-1faec8b72739-cert\") pod \"openstack-baremetal-operator-controller-manager-66cc85b5d5lsczg\" (UID: \"eb017e3a-1627-4dc7-bda3-1faec8b72739\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-66cc85b5d5lsczg" Oct 05 07:04:33 crc kubenswrapper[4846]: E1005 07:04:33.184453 4846 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Oct 05 07:04:33 crc kubenswrapper[4846]: E1005 07:04:33.184547 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/eb017e3a-1627-4dc7-bda3-1faec8b72739-cert podName:eb017e3a-1627-4dc7-bda3-1faec8b72739 nodeName:}" failed. No retries permitted until 2025-10-05 07:04:33.684516288 +0000 UTC m=+995.925369063 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/eb017e3a-1627-4dc7-bda3-1faec8b72739-cert") pod "openstack-baremetal-operator-controller-manager-66cc85b5d5lsczg" (UID: "eb017e3a-1627-4dc7-bda3-1faec8b72739") : secret "openstack-baremetal-operator-webhook-server-cert" not found Oct 05 07:04:33 crc kubenswrapper[4846]: I1005 07:04:33.184969 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-6c9b57c67-ggtrm" Oct 05 07:04:33 crc kubenswrapper[4846]: I1005 07:04:33.236432 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6kzfn\" (UniqueName: \"kubernetes.io/projected/eb017e3a-1627-4dc7-bda3-1faec8b72739-kube-api-access-6kzfn\") pod \"openstack-baremetal-operator-controller-manager-66cc85b5d5lsczg\" (UID: \"eb017e3a-1627-4dc7-bda3-1faec8b72739\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-66cc85b5d5lsczg" Oct 05 07:04:33 crc kubenswrapper[4846]: I1005 07:04:33.236672 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-69f59f9d8-s4qf7" Oct 05 07:04:33 crc kubenswrapper[4846]: I1005 07:04:33.236900 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-f589c7597-jzsnd"] Oct 05 07:04:33 crc kubenswrapper[4846]: I1005 07:04:33.242714 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lmjch\" (UniqueName: \"kubernetes.io/projected/ab1677d6-3cb0-4dce-8e8d-4d2b67025e6d-kube-api-access-lmjch\") pod \"swift-operator-controller-manager-76d5577b-pkjcv\" (UID: \"ab1677d6-3cb0-4dce-8e8d-4d2b67025e6d\") " pod="openstack-operators/swift-operator-controller-manager-76d5577b-pkjcv" Oct 05 07:04:33 crc kubenswrapper[4846]: I1005 07:04:33.259063 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p697b\" (UniqueName: \"kubernetes.io/projected/95847a57-fbe8-43f2-b4b4-e630665f17e3-kube-api-access-p697b\") pod \"placement-operator-controller-manager-66f6d6849b-62zd4\" (UID: \"95847a57-fbe8-43f2-b4b4-e630665f17e3\") " pod="openstack-operators/placement-operator-controller-manager-66f6d6849b-62zd4" Oct 05 07:04:33 crc kubenswrapper[4846]: I1005 07:04:33.272111 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v2b4v\" (UniqueName: \"kubernetes.io/projected/945ab775-ffae-41cb-a315-5e05cf65484d-kube-api-access-v2b4v\") pod \"ovn-operator-controller-manager-c968bb45-rmwgr\" (UID: \"945ab775-ffae-41cb-a315-5e05cf65484d\") " pod="openstack-operators/ovn-operator-controller-manager-c968bb45-rmwgr" Oct 05 07:04:33 crc kubenswrapper[4846]: I1005 07:04:33.287144 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gx76x\" (UniqueName: \"kubernetes.io/projected/3053ab31-f463-49b7-9b32-d37b0506462b-kube-api-access-gx76x\") pod \"telemetry-operator-controller-manager-f589c7597-jzsnd\" (UID: \"3053ab31-f463-49b7-9b32-d37b0506462b\") " pod="openstack-operators/telemetry-operator-controller-manager-f589c7597-jzsnd" Oct 05 07:04:33 crc kubenswrapper[4846]: I1005 07:04:33.372282 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/test-operator-controller-manager-6bb6dcddc-4krwr"] Oct 05 07:04:33 crc kubenswrapper[4846]: I1005 07:04:33.376368 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-6bb6dcddc-4krwr" Oct 05 07:04:33 crc kubenswrapper[4846]: I1005 07:04:33.382847 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-84788b6bc5-pg8c6" Oct 05 07:04:33 crc kubenswrapper[4846]: I1005 07:04:33.384327 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-66f6d6849b-62zd4" Oct 05 07:04:33 crc kubenswrapper[4846]: I1005 07:04:33.390902 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"test-operator-controller-manager-dockercfg-92nff" Oct 05 07:04:33 crc kubenswrapper[4846]: I1005 07:04:33.409836 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z4c9w\" (UniqueName: \"kubernetes.io/projected/b7a3163e-66e4-406c-9285-e2d125024bc8-kube-api-access-z4c9w\") pod \"test-operator-controller-manager-6bb6dcddc-4krwr\" (UID: \"b7a3163e-66e4-406c-9285-e2d125024bc8\") " pod="openstack-operators/test-operator-controller-manager-6bb6dcddc-4krwr" Oct 05 07:04:33 crc kubenswrapper[4846]: I1005 07:04:33.409905 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gx76x\" (UniqueName: \"kubernetes.io/projected/3053ab31-f463-49b7-9b32-d37b0506462b-kube-api-access-gx76x\") pod \"telemetry-operator-controller-manager-f589c7597-jzsnd\" (UID: \"3053ab31-f463-49b7-9b32-d37b0506462b\") " pod="openstack-operators/telemetry-operator-controller-manager-f589c7597-jzsnd" Oct 05 07:04:33 crc kubenswrapper[4846]: I1005 07:04:33.446093 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-76d5577b-pkjcv" Oct 05 07:04:33 crc kubenswrapper[4846]: I1005 07:04:33.453764 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gx76x\" (UniqueName: \"kubernetes.io/projected/3053ab31-f463-49b7-9b32-d37b0506462b-kube-api-access-gx76x\") pod \"telemetry-operator-controller-manager-f589c7597-jzsnd\" (UID: \"3053ab31-f463-49b7-9b32-d37b0506462b\") " pod="openstack-operators/telemetry-operator-controller-manager-f589c7597-jzsnd" Oct 05 07:04:33 crc kubenswrapper[4846]: I1005 07:04:33.461122 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-6bb6dcddc-4krwr"] Oct 05 07:04:33 crc kubenswrapper[4846]: I1005 07:04:33.492335 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-f589c7597-jzsnd" Oct 05 07:04:33 crc kubenswrapper[4846]: I1005 07:04:33.502165 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/watcher-operator-controller-manager-5d98cc5575-blkpr"] Oct 05 07:04:33 crc kubenswrapper[4846]: I1005 07:04:33.503703 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-5d98cc5575-blkpr" Oct 05 07:04:33 crc kubenswrapper[4846]: I1005 07:04:33.520310 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z4c9w\" (UniqueName: \"kubernetes.io/projected/b7a3163e-66e4-406c-9285-e2d125024bc8-kube-api-access-z4c9w\") pod \"test-operator-controller-manager-6bb6dcddc-4krwr\" (UID: \"b7a3163e-66e4-406c-9285-e2d125024bc8\") " pod="openstack-operators/test-operator-controller-manager-6bb6dcddc-4krwr" Oct 05 07:04:33 crc kubenswrapper[4846]: I1005 07:04:33.544536 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"watcher-operator-controller-manager-dockercfg-lmlf2" Oct 05 07:04:33 crc kubenswrapper[4846]: I1005 07:04:33.562032 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-c968bb45-rmwgr" Oct 05 07:04:33 crc kubenswrapper[4846]: I1005 07:04:33.574400 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z4c9w\" (UniqueName: \"kubernetes.io/projected/b7a3163e-66e4-406c-9285-e2d125024bc8-kube-api-access-z4c9w\") pod \"test-operator-controller-manager-6bb6dcddc-4krwr\" (UID: \"b7a3163e-66e4-406c-9285-e2d125024bc8\") " pod="openstack-operators/test-operator-controller-manager-6bb6dcddc-4krwr" Oct 05 07:04:33 crc kubenswrapper[4846]: I1005 07:04:33.585489 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-5d98cc5575-blkpr"] Oct 05 07:04:33 crc kubenswrapper[4846]: I1005 07:04:33.642120 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-597ff\" (UniqueName: \"kubernetes.io/projected/d4de72d9-7d9c-4d90-ba8a-4179d9028033-kube-api-access-597ff\") pod \"watcher-operator-controller-manager-5d98cc5575-blkpr\" (UID: \"d4de72d9-7d9c-4d90-ba8a-4179d9028033\") " pod="openstack-operators/watcher-operator-controller-manager-5d98cc5575-blkpr" Oct 05 07:04:33 crc kubenswrapper[4846]: I1005 07:04:33.681552 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-manager-7cfc658b9-22b7k"] Oct 05 07:04:33 crc kubenswrapper[4846]: I1005 07:04:33.683335 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-7cfc658b9-22b7k" Oct 05 07:04:33 crc kubenswrapper[4846]: I1005 07:04:33.688107 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"webhook-server-cert" Oct 05 07:04:33 crc kubenswrapper[4846]: I1005 07:04:33.688318 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-hchvw" Oct 05 07:04:33 crc kubenswrapper[4846]: I1005 07:04:33.700686 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-7cfc658b9-22b7k"] Oct 05 07:04:33 crc kubenswrapper[4846]: I1005 07:04:33.710740 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-pdqdx"] Oct 05 07:04:33 crc kubenswrapper[4846]: I1005 07:04:33.721672 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-pdqdx" Oct 05 07:04:33 crc kubenswrapper[4846]: I1005 07:04:33.721541 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-pdqdx"] Oct 05 07:04:33 crc kubenswrapper[4846]: I1005 07:04:33.726515 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-controller-manager-dockercfg-tz74b" Oct 05 07:04:33 crc kubenswrapper[4846]: I1005 07:04:33.743292 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/d7419a91-cc67-4cdb-b8a4-d41480e5854f-cert\") pod \"openstack-operator-controller-manager-7cfc658b9-22b7k\" (UID: \"d7419a91-cc67-4cdb-b8a4-d41480e5854f\") " pod="openstack-operators/openstack-operator-controller-manager-7cfc658b9-22b7k" Oct 05 07:04:33 crc kubenswrapper[4846]: I1005 07:04:33.744281 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/eb017e3a-1627-4dc7-bda3-1faec8b72739-cert\") pod \"openstack-baremetal-operator-controller-manager-66cc85b5d5lsczg\" (UID: \"eb017e3a-1627-4dc7-bda3-1faec8b72739\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-66cc85b5d5lsczg" Oct 05 07:04:33 crc kubenswrapper[4846]: I1005 07:04:33.744396 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-89ptb\" (UniqueName: \"kubernetes.io/projected/d7419a91-cc67-4cdb-b8a4-d41480e5854f-kube-api-access-89ptb\") pod \"openstack-operator-controller-manager-7cfc658b9-22b7k\" (UID: \"d7419a91-cc67-4cdb-b8a4-d41480e5854f\") " pod="openstack-operators/openstack-operator-controller-manager-7cfc658b9-22b7k" Oct 05 07:04:33 crc kubenswrapper[4846]: I1005 07:04:33.744477 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-597ff\" (UniqueName: \"kubernetes.io/projected/d4de72d9-7d9c-4d90-ba8a-4179d9028033-kube-api-access-597ff\") pod \"watcher-operator-controller-manager-5d98cc5575-blkpr\" (UID: \"d4de72d9-7d9c-4d90-ba8a-4179d9028033\") " pod="openstack-operators/watcher-operator-controller-manager-5d98cc5575-blkpr" Oct 05 07:04:33 crc kubenswrapper[4846]: E1005 07:04:33.744836 4846 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Oct 05 07:04:33 crc kubenswrapper[4846]: E1005 07:04:33.744908 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/eb017e3a-1627-4dc7-bda3-1faec8b72739-cert podName:eb017e3a-1627-4dc7-bda3-1faec8b72739 nodeName:}" failed. No retries permitted until 2025-10-05 07:04:34.744881152 +0000 UTC m=+996.985733927 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/eb017e3a-1627-4dc7-bda3-1faec8b72739-cert") pod "openstack-baremetal-operator-controller-manager-66cc85b5d5lsczg" (UID: "eb017e3a-1627-4dc7-bda3-1faec8b72739") : secret "openstack-baremetal-operator-webhook-server-cert" not found Oct 05 07:04:33 crc kubenswrapper[4846]: I1005 07:04:33.751036 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-58d86cd59d-gnb7x"] Oct 05 07:04:33 crc kubenswrapper[4846]: I1005 07:04:33.776570 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-6675647785-d2n5v"] Oct 05 07:04:33 crc kubenswrapper[4846]: I1005 07:04:33.784130 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-597ff\" (UniqueName: \"kubernetes.io/projected/d4de72d9-7d9c-4d90-ba8a-4179d9028033-kube-api-access-597ff\") pod \"watcher-operator-controller-manager-5d98cc5575-blkpr\" (UID: \"d4de72d9-7d9c-4d90-ba8a-4179d9028033\") " pod="openstack-operators/watcher-operator-controller-manager-5d98cc5575-blkpr" Oct 05 07:04:33 crc kubenswrapper[4846]: I1005 07:04:33.845480 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-5c497dbdb-g4bkx"] Oct 05 07:04:33 crc kubenswrapper[4846]: I1005 07:04:33.850812 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/d7419a91-cc67-4cdb-b8a4-d41480e5854f-cert\") pod \"openstack-operator-controller-manager-7cfc658b9-22b7k\" (UID: \"d7419a91-cc67-4cdb-b8a4-d41480e5854f\") " pod="openstack-operators/openstack-operator-controller-manager-7cfc658b9-22b7k" Oct 05 07:04:33 crc kubenswrapper[4846]: I1005 07:04:33.851439 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gqvpp\" (UniqueName: \"kubernetes.io/projected/aae22754-9bda-494a-afb8-42debef021dd-kube-api-access-gqvpp\") pod \"rabbitmq-cluster-operator-manager-5f97d8c699-pdqdx\" (UID: \"aae22754-9bda-494a-afb8-42debef021dd\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-pdqdx" Oct 05 07:04:33 crc kubenswrapper[4846]: I1005 07:04:33.851525 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-89ptb\" (UniqueName: \"kubernetes.io/projected/d7419a91-cc67-4cdb-b8a4-d41480e5854f-kube-api-access-89ptb\") pod \"openstack-operator-controller-manager-7cfc658b9-22b7k\" (UID: \"d7419a91-cc67-4cdb-b8a4-d41480e5854f\") " pod="openstack-operators/openstack-operator-controller-manager-7cfc658b9-22b7k" Oct 05 07:04:33 crc kubenswrapper[4846]: I1005 07:04:33.862412 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/d7419a91-cc67-4cdb-b8a4-d41480e5854f-cert\") pod \"openstack-operator-controller-manager-7cfc658b9-22b7k\" (UID: \"d7419a91-cc67-4cdb-b8a4-d41480e5854f\") " pod="openstack-operators/openstack-operator-controller-manager-7cfc658b9-22b7k" Oct 05 07:04:33 crc kubenswrapper[4846]: I1005 07:04:33.827305 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-6bb6dcddc-4krwr" Oct 05 07:04:33 crc kubenswrapper[4846]: I1005 07:04:33.888916 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-89ptb\" (UniqueName: \"kubernetes.io/projected/d7419a91-cc67-4cdb-b8a4-d41480e5854f-kube-api-access-89ptb\") pod \"openstack-operator-controller-manager-7cfc658b9-22b7k\" (UID: \"d7419a91-cc67-4cdb-b8a4-d41480e5854f\") " pod="openstack-operators/openstack-operator-controller-manager-7cfc658b9-22b7k" Oct 05 07:04:33 crc kubenswrapper[4846]: I1005 07:04:33.897021 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-5d98cc5575-blkpr" Oct 05 07:04:33 crc kubenswrapper[4846]: I1005 07:04:33.933754 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-58d86cd59d-gnb7x" event={"ID":"ac8cdf70-77c0-44c9-b61b-11325d91b698","Type":"ContainerStarted","Data":"ae4679f2fce9e912247b053b00753d713cb98c067696e17f07dbebe239bb2126"} Oct 05 07:04:33 crc kubenswrapper[4846]: I1005 07:04:33.935213 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5c497dbdb-g4bkx" event={"ID":"931bcbda-d249-4625-a93c-bfa49dcb38ae","Type":"ContainerStarted","Data":"62612c2df701a55b762f54ff8c8a57d06a0598403d9daaf93d0703ef50cc49df"} Oct 05 07:04:33 crc kubenswrapper[4846]: I1005 07:04:33.938382 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-6675647785-d2n5v" event={"ID":"a1899a06-0cb1-44aa-ba1e-d6876051bab1","Type":"ContainerStarted","Data":"982e3924aecc06ee902cdf4109419f0fd08dd68dfea87622edd38a9bde4b83fb"} Oct 05 07:04:33 crc kubenswrapper[4846]: I1005 07:04:33.952451 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gqvpp\" (UniqueName: \"kubernetes.io/projected/aae22754-9bda-494a-afb8-42debef021dd-kube-api-access-gqvpp\") pod \"rabbitmq-cluster-operator-manager-5f97d8c699-pdqdx\" (UID: \"aae22754-9bda-494a-afb8-42debef021dd\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-pdqdx" Oct 05 07:04:33 crc kubenswrapper[4846]: I1005 07:04:33.975268 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gqvpp\" (UniqueName: \"kubernetes.io/projected/aae22754-9bda-494a-afb8-42debef021dd-kube-api-access-gqvpp\") pod \"rabbitmq-cluster-operator-manager-5f97d8c699-pdqdx\" (UID: \"aae22754-9bda-494a-afb8-42debef021dd\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-pdqdx" Oct 05 07:04:34 crc kubenswrapper[4846]: I1005 07:04:34.128968 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-6f5894c49f-fkx4r"] Oct 05 07:04:34 crc kubenswrapper[4846]: I1005 07:04:34.151988 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-5b974f6766-khk5k"] Oct 05 07:04:34 crc kubenswrapper[4846]: I1005 07:04:34.159387 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-57c9cdcf57-j9742"] Oct 05 07:04:34 crc kubenswrapper[4846]: I1005 07:04:34.163132 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-7cfc658b9-22b7k" Oct 05 07:04:34 crc kubenswrapper[4846]: W1005 07:04:34.167399 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf626116b_4f3a_48cb_ad75_c6550333ee6e.slice/crio-cdcef1c38503dfd797148d5f77c3df1588c815d43efbab8d5fb7848583611a6c WatchSource:0}: Error finding container cdcef1c38503dfd797148d5f77c3df1588c815d43efbab8d5fb7848583611a6c: Status 404 returned error can't find the container with id cdcef1c38503dfd797148d5f77c3df1588c815d43efbab8d5fb7848583611a6c Oct 05 07:04:34 crc kubenswrapper[4846]: I1005 07:04:34.194534 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-84bd8f6848-b4rhd"] Oct 05 07:04:34 crc kubenswrapper[4846]: I1005 07:04:34.199367 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-pdqdx" Oct 05 07:04:34 crc kubenswrapper[4846]: W1005 07:04:34.202014 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod58088fc8_5aaa_422a_84cd_d4ef7d10e989.slice/crio-113c91a0b627bd6da811d368e9cf26b8cb7fa5ff007d269414a6e8b50220c6aa WatchSource:0}: Error finding container 113c91a0b627bd6da811d368e9cf26b8cb7fa5ff007d269414a6e8b50220c6aa: Status 404 returned error can't find the container with id 113c91a0b627bd6da811d368e9cf26b8cb7fa5ff007d269414a6e8b50220c6aa Oct 05 07:04:34 crc kubenswrapper[4846]: I1005 07:04:34.315937 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-7cb48dbc-cvtkp"] Oct 05 07:04:34 crc kubenswrapper[4846]: I1005 07:04:34.323228 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-69b956fbf6-xsltv"] Oct 05 07:04:34 crc kubenswrapper[4846]: W1005 07:04:34.352412 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod644a3a04_e2d4_4352_9bac_e8d661fd61b0.slice/crio-fd171c107534d9a67a6cb89b23d15ae13e8c8f49b737ebe44c1656ab58a98b72 WatchSource:0}: Error finding container fd171c107534d9a67a6cb89b23d15ae13e8c8f49b737ebe44c1656ab58a98b72: Status 404 returned error can't find the container with id fd171c107534d9a67a6cb89b23d15ae13e8c8f49b737ebe44c1656ab58a98b72 Oct 05 07:04:34 crc kubenswrapper[4846]: I1005 07:04:34.352347 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-76d5577b-pkjcv"] Oct 05 07:04:34 crc kubenswrapper[4846]: I1005 07:04:34.382617 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-698456cdc6-w4fst"] Oct 05 07:04:34 crc kubenswrapper[4846]: I1005 07:04:34.456261 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-6c9b57c67-ggtrm"] Oct 05 07:04:34 crc kubenswrapper[4846]: I1005 07:04:34.460911 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-d6c9dc5bc-2fb7l"] Oct 05 07:04:34 crc kubenswrapper[4846]: W1005 07:04:34.464001 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod71de4f09_b26f_443c_9142_7edbc0d18e38.slice/crio-7a5cebcc9e5e543caba600bb5fe90ec5950520d0ddffc09af9e42bfc9af74770 WatchSource:0}: Error finding container 7a5cebcc9e5e543caba600bb5fe90ec5950520d0ddffc09af9e42bfc9af74770: Status 404 returned error can't find the container with id 7a5cebcc9e5e543caba600bb5fe90ec5950520d0ddffc09af9e42bfc9af74770 Oct 05 07:04:34 crc kubenswrapper[4846]: I1005 07:04:34.465818 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-69f59f9d8-s4qf7"] Oct 05 07:04:34 crc kubenswrapper[4846]: E1005 07:04:34.475398 4846 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/nova-operator@sha256:e4c4ff39c54c0af231fb781759ab50ed86285c74d38bdea43fa75646b762d842,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-57hdf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod nova-operator-controller-manager-6c9b57c67-ggtrm_openstack-operators(49e72a92-e750-4046-9894-f3a0359531b2): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 05 07:04:34 crc kubenswrapper[4846]: I1005 07:04:34.561547 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-f589c7597-jzsnd"] Oct 05 07:04:34 crc kubenswrapper[4846]: I1005 07:04:34.569090 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-6bb6dcddc-4krwr"] Oct 05 07:04:34 crc kubenswrapper[4846]: I1005 07:04:34.575948 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-84788b6bc5-pg8c6"] Oct 05 07:04:34 crc kubenswrapper[4846]: W1005 07:04:34.583936 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb7a3163e_66e4_406c_9285_e2d125024bc8.slice/crio-1905312461cf693b3b2f6911afbb1c3d850f7d01c606e8d6e7b17b2c867022c1 WatchSource:0}: Error finding container 1905312461cf693b3b2f6911afbb1c3d850f7d01c606e8d6e7b17b2c867022c1: Status 404 returned error can't find the container with id 1905312461cf693b3b2f6911afbb1c3d850f7d01c606e8d6e7b17b2c867022c1 Oct 05 07:04:34 crc kubenswrapper[4846]: E1005 07:04:34.587041 4846 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/test-operator@sha256:0daf76cc40ab619ae266b11defcc1b65beb22d859369e7b1b04de9169089a4cb,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-z4c9w,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod test-operator-controller-manager-6bb6dcddc-4krwr_openstack-operators(b7a3163e-66e4-406c-9285-e2d125024bc8): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 05 07:04:34 crc kubenswrapper[4846]: I1005 07:04:34.603521 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-c968bb45-rmwgr"] Oct 05 07:04:34 crc kubenswrapper[4846]: I1005 07:04:34.616537 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-66f6d6849b-62zd4"] Oct 05 07:04:34 crc kubenswrapper[4846]: E1005 07:04:34.628453 4846 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/ovn-operator@sha256:f923b76c1dd8fde02a5faf8a0a433cfacfb7b743f371de64a12e30d6efcde254,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-v2b4v,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovn-operator-controller-manager-c968bb45-rmwgr_openstack-operators(945ab775-ffae-41cb-a315-5e05cf65484d): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 05 07:04:34 crc kubenswrapper[4846]: E1005 07:04:34.628863 4846 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/telemetry-operator@sha256:bf55026ba10b80e1e24733078bd204cef8766d21a305fd000707a1e3b30ff52e,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-gx76x,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod telemetry-operator-controller-manager-f589c7597-jzsnd_openstack-operators(3053ab31-f463-49b7-9b32-d37b0506462b): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 05 07:04:34 crc kubenswrapper[4846]: W1005 07:04:34.641806 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod95847a57_fbe8_43f2_b4b4_e630665f17e3.slice/crio-1576a47f64d0022f3afd05d6eb8cd6044390b0013eb110d31cf8617a6c7c886a WatchSource:0}: Error finding container 1576a47f64d0022f3afd05d6eb8cd6044390b0013eb110d31cf8617a6c7c886a: Status 404 returned error can't find the container with id 1576a47f64d0022f3afd05d6eb8cd6044390b0013eb110d31cf8617a6c7c886a Oct 05 07:04:34 crc kubenswrapper[4846]: E1005 07:04:34.650516 4846 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/placement-operator@sha256:adc23c5fd1aece2b16dc8e22ceed628f9a719455e39d3f98c77544665c6749e1,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-p697b,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-operator-controller-manager-66f6d6849b-62zd4_openstack-operators(95847a57-fbe8-43f2-b4b4-e630665f17e3): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 05 07:04:34 crc kubenswrapper[4846]: E1005 07:04:34.677402 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/nova-operator-controller-manager-6c9b57c67-ggtrm" podUID="49e72a92-e750-4046-9894-f3a0359531b2" Oct 05 07:04:34 crc kubenswrapper[4846]: I1005 07:04:34.720996 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-5d98cc5575-blkpr"] Oct 05 07:04:34 crc kubenswrapper[4846]: W1005 07:04:34.733541 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd4de72d9_7d9c_4d90_ba8a_4179d9028033.slice/crio-497ef1d824e4d2d486c92fb47e68e501ccd24e35a93b3d4245ffbcfdc9b24d6d WatchSource:0}: Error finding container 497ef1d824e4d2d486c92fb47e68e501ccd24e35a93b3d4245ffbcfdc9b24d6d: Status 404 returned error can't find the container with id 497ef1d824e4d2d486c92fb47e68e501ccd24e35a93b3d4245ffbcfdc9b24d6d Oct 05 07:04:34 crc kubenswrapper[4846]: E1005 07:04:34.737955 4846 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/watcher-operator@sha256:64f57b2b59dea2bd9fae91490c5bec2687131884a049e6579819d9f951b877c6,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-597ff,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod watcher-operator-controller-manager-5d98cc5575-blkpr_openstack-operators(d4de72d9-7d9c-4d90-ba8a-4179d9028033): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 05 07:04:34 crc kubenswrapper[4846]: I1005 07:04:34.771202 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/eb017e3a-1627-4dc7-bda3-1faec8b72739-cert\") pod \"openstack-baremetal-operator-controller-manager-66cc85b5d5lsczg\" (UID: \"eb017e3a-1627-4dc7-bda3-1faec8b72739\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-66cc85b5d5lsczg" Oct 05 07:04:34 crc kubenswrapper[4846]: I1005 07:04:34.784558 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/eb017e3a-1627-4dc7-bda3-1faec8b72739-cert\") pod \"openstack-baremetal-operator-controller-manager-66cc85b5d5lsczg\" (UID: \"eb017e3a-1627-4dc7-bda3-1faec8b72739\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-66cc85b5d5lsczg" Oct 05 07:04:34 crc kubenswrapper[4846]: I1005 07:04:34.801350 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-pdqdx"] Oct 05 07:04:34 crc kubenswrapper[4846]: W1005 07:04:34.806156 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podaae22754_9bda_494a_afb8_42debef021dd.slice/crio-cee4fc9a31fb0486f5dc501b1d7e6210664984c5925a5066e7a41b2665d48f7e WatchSource:0}: Error finding container cee4fc9a31fb0486f5dc501b1d7e6210664984c5925a5066e7a41b2665d48f7e: Status 404 returned error can't find the container with id cee4fc9a31fb0486f5dc501b1d7e6210664984c5925a5066e7a41b2665d48f7e Oct 05 07:04:34 crc kubenswrapper[4846]: I1005 07:04:34.808992 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-7cfc658b9-22b7k"] Oct 05 07:04:34 crc kubenswrapper[4846]: I1005 07:04:34.814413 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-66cc85b5d5lsczg" Oct 05 07:04:34 crc kubenswrapper[4846]: W1005 07:04:34.836437 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd7419a91_cc67_4cdb_b8a4_d41480e5854f.slice/crio-106daf22b2c3d4c48d22db2107bb06d8fdb143fa72758205ce2985f630c5ed2e WatchSource:0}: Error finding container 106daf22b2c3d4c48d22db2107bb06d8fdb143fa72758205ce2985f630c5ed2e: Status 404 returned error can't find the container with id 106daf22b2c3d4c48d22db2107bb06d8fdb143fa72758205ce2985f630c5ed2e Oct 05 07:04:34 crc kubenswrapper[4846]: E1005 07:04:34.885260 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/test-operator-controller-manager-6bb6dcddc-4krwr" podUID="b7a3163e-66e4-406c-9285-e2d125024bc8" Oct 05 07:04:34 crc kubenswrapper[4846]: I1005 07:04:34.997861 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-7cfc658b9-22b7k" event={"ID":"d7419a91-cc67-4cdb-b8a4-d41480e5854f","Type":"ContainerStarted","Data":"106daf22b2c3d4c48d22db2107bb06d8fdb143fa72758205ce2985f630c5ed2e"} Oct 05 07:04:35 crc kubenswrapper[4846]: E1005 07:04:35.000488 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/telemetry-operator-controller-manager-f589c7597-jzsnd" podUID="3053ab31-f463-49b7-9b32-d37b0506462b" Oct 05 07:04:35 crc kubenswrapper[4846]: I1005 07:04:35.002500 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-5d98cc5575-blkpr" event={"ID":"d4de72d9-7d9c-4d90-ba8a-4179d9028033","Type":"ContainerStarted","Data":"497ef1d824e4d2d486c92fb47e68e501ccd24e35a93b3d4245ffbcfdc9b24d6d"} Oct 05 07:04:35 crc kubenswrapper[4846]: I1005 07:04:35.005444 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-84788b6bc5-pg8c6" event={"ID":"f0825244-e759-46a4-a16b-4d993dac339d","Type":"ContainerStarted","Data":"8d83f1334d482cacdad07edaa777bc2f86cc7625788b880feabdd4a14644791a"} Oct 05 07:04:35 crc kubenswrapper[4846]: I1005 07:04:35.006721 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-69b956fbf6-xsltv" event={"ID":"644a3a04-e2d4-4352-9bac-e8d661fd61b0","Type":"ContainerStarted","Data":"fd171c107534d9a67a6cb89b23d15ae13e8c8f49b737ebe44c1656ab58a98b72"} Oct 05 07:04:35 crc kubenswrapper[4846]: I1005 07:04:35.007586 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-69f59f9d8-s4qf7" event={"ID":"15b8ca14-e689-430c-84b2-40da6f1e83a5","Type":"ContainerStarted","Data":"ff1a20d0aa08aee3d486909476d49aa627c379c35c13b93d537d9172377671ac"} Oct 05 07:04:35 crc kubenswrapper[4846]: I1005 07:04:35.008385 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-5b974f6766-khk5k" event={"ID":"f626116b-4f3a-48cb-ad75-c6550333ee6e","Type":"ContainerStarted","Data":"cdcef1c38503dfd797148d5f77c3df1588c815d43efbab8d5fb7848583611a6c"} Oct 05 07:04:35 crc kubenswrapper[4846]: I1005 07:04:35.010062 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-d6c9dc5bc-2fb7l" event={"ID":"71de4f09-b26f-443c-9142-7edbc0d18e38","Type":"ContainerStarted","Data":"7a5cebcc9e5e543caba600bb5fe90ec5950520d0ddffc09af9e42bfc9af74770"} Oct 05 07:04:35 crc kubenswrapper[4846]: I1005 07:04:35.011961 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-6c9b57c67-ggtrm" event={"ID":"49e72a92-e750-4046-9894-f3a0359531b2","Type":"ContainerStarted","Data":"293ce5fb54625af99999be11142cc08c55e4a81dab7126ad3e147db7b519e260"} Oct 05 07:04:35 crc kubenswrapper[4846]: I1005 07:04:35.011984 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-6c9b57c67-ggtrm" event={"ID":"49e72a92-e750-4046-9894-f3a0359531b2","Type":"ContainerStarted","Data":"ee422835007bcfdd2f212702ac4b4a02b246da8b7cc1bf891b0d3407e7825f34"} Oct 05 07:04:35 crc kubenswrapper[4846]: E1005 07:04:35.018929 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/nova-operator@sha256:e4c4ff39c54c0af231fb781759ab50ed86285c74d38bdea43fa75646b762d842\\\"\"" pod="openstack-operators/nova-operator-controller-manager-6c9b57c67-ggtrm" podUID="49e72a92-e750-4046-9894-f3a0359531b2" Oct 05 07:04:35 crc kubenswrapper[4846]: I1005 07:04:35.033985 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-pdqdx" event={"ID":"aae22754-9bda-494a-afb8-42debef021dd","Type":"ContainerStarted","Data":"cee4fc9a31fb0486f5dc501b1d7e6210664984c5925a5066e7a41b2665d48f7e"} Oct 05 07:04:35 crc kubenswrapper[4846]: I1005 07:04:35.040068 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-6f5894c49f-fkx4r" event={"ID":"92c93055-428a-4d41-b034-a805b68f3e19","Type":"ContainerStarted","Data":"cad0f2c1d096ac1fa556dc3112fd2fe464d189a4fad060182ae2cc7ce9330ec9"} Oct 05 07:04:35 crc kubenswrapper[4846]: E1005 07:04:35.049775 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/placement-operator-controller-manager-66f6d6849b-62zd4" podUID="95847a57-fbe8-43f2-b4b4-e630665f17e3" Oct 05 07:04:35 crc kubenswrapper[4846]: E1005 07:04:35.050284 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/watcher-operator-controller-manager-5d98cc5575-blkpr" podUID="d4de72d9-7d9c-4d90-ba8a-4179d9028033" Oct 05 07:04:35 crc kubenswrapper[4846]: I1005 07:04:35.051458 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-84bd8f6848-b4rhd" event={"ID":"58088fc8-5aaa-422a-84cd-d4ef7d10e989","Type":"ContainerStarted","Data":"113c91a0b627bd6da811d368e9cf26b8cb7fa5ff007d269414a6e8b50220c6aa"} Oct 05 07:04:35 crc kubenswrapper[4846]: I1005 07:04:35.054920 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-76d5577b-pkjcv" event={"ID":"ab1677d6-3cb0-4dce-8e8d-4d2b67025e6d","Type":"ContainerStarted","Data":"945c7f3c4fe2688a61192b1a3594104d617d9be9b062b66bbfd5aaa60253db73"} Oct 05 07:04:35 crc kubenswrapper[4846]: E1005 07:04:35.059427 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/ovn-operator-controller-manager-c968bb45-rmwgr" podUID="945ab775-ffae-41cb-a315-5e05cf65484d" Oct 05 07:04:35 crc kubenswrapper[4846]: I1005 07:04:35.060147 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-c968bb45-rmwgr" event={"ID":"945ab775-ffae-41cb-a315-5e05cf65484d","Type":"ContainerStarted","Data":"87823aad15b8db319cf1e0c71c146e5514f987ee339670442b955b6690312044"} Oct 05 07:04:35 crc kubenswrapper[4846]: I1005 07:04:35.065339 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-57c9cdcf57-j9742" event={"ID":"6ff6c74d-2f8b-42d0-af60-88baa4399b32","Type":"ContainerStarted","Data":"582ee36fa7f0fe89d8769251afcfa3a96230b6eeff449aa5f18193b1c70d2918"} Oct 05 07:04:35 crc kubenswrapper[4846]: I1005 07:04:35.082925 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-698456cdc6-w4fst" event={"ID":"c1b6f754-a4fc-4981-a722-415569d3e7a3","Type":"ContainerStarted","Data":"d0adcc1fb21aae95797c9668f2723644f3e842aaa0a4d408fc58ef43fe573fa2"} Oct 05 07:04:35 crc kubenswrapper[4846]: I1005 07:04:35.086038 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-6bb6dcddc-4krwr" event={"ID":"b7a3163e-66e4-406c-9285-e2d125024bc8","Type":"ContainerStarted","Data":"e69c0f78f9ff4f72f0a99abbb693400663ab7936204b0d74b30ec70761cd5e64"} Oct 05 07:04:35 crc kubenswrapper[4846]: I1005 07:04:35.086099 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-6bb6dcddc-4krwr" event={"ID":"b7a3163e-66e4-406c-9285-e2d125024bc8","Type":"ContainerStarted","Data":"1905312461cf693b3b2f6911afbb1c3d850f7d01c606e8d6e7b17b2c867022c1"} Oct 05 07:04:35 crc kubenswrapper[4846]: E1005 07:04:35.087866 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:0daf76cc40ab619ae266b11defcc1b65beb22d859369e7b1b04de9169089a4cb\\\"\"" pod="openstack-operators/test-operator-controller-manager-6bb6dcddc-4krwr" podUID="b7a3163e-66e4-406c-9285-e2d125024bc8" Oct 05 07:04:35 crc kubenswrapper[4846]: I1005 07:04:35.089599 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-66f6d6849b-62zd4" event={"ID":"95847a57-fbe8-43f2-b4b4-e630665f17e3","Type":"ContainerStarted","Data":"1576a47f64d0022f3afd05d6eb8cd6044390b0013eb110d31cf8617a6c7c886a"} Oct 05 07:04:35 crc kubenswrapper[4846]: E1005 07:04:35.115331 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/placement-operator@sha256:adc23c5fd1aece2b16dc8e22ceed628f9a719455e39d3f98c77544665c6749e1\\\"\"" pod="openstack-operators/placement-operator-controller-manager-66f6d6849b-62zd4" podUID="95847a57-fbe8-43f2-b4b4-e630665f17e3" Oct 05 07:04:35 crc kubenswrapper[4846]: I1005 07:04:35.122199 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-7cb48dbc-cvtkp" event={"ID":"e46909e0-1642-436c-b597-167a42e0a72d","Type":"ContainerStarted","Data":"0d0d1832842738465dd009dd7d4646e34f1b66732a6f7e7acb1264115204aa4a"} Oct 05 07:04:35 crc kubenswrapper[4846]: I1005 07:04:35.134392 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-f589c7597-jzsnd" event={"ID":"3053ab31-f463-49b7-9b32-d37b0506462b","Type":"ContainerStarted","Data":"f8a48ce67bfac0a3a4b12a9dc9d7c1982d91f0c31e6da841ab96aea25401b424"} Oct 05 07:04:35 crc kubenswrapper[4846]: E1005 07:04:35.142197 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/telemetry-operator@sha256:bf55026ba10b80e1e24733078bd204cef8766d21a305fd000707a1e3b30ff52e\\\"\"" pod="openstack-operators/telemetry-operator-controller-manager-f589c7597-jzsnd" podUID="3053ab31-f463-49b7-9b32-d37b0506462b" Oct 05 07:04:35 crc kubenswrapper[4846]: I1005 07:04:35.461234 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-66cc85b5d5lsczg"] Oct 05 07:04:36 crc kubenswrapper[4846]: I1005 07:04:36.153331 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-66cc85b5d5lsczg" event={"ID":"eb017e3a-1627-4dc7-bda3-1faec8b72739","Type":"ContainerStarted","Data":"2fd13bbdff4d05dc13ae840ffe9ceb7faad5c9c59f780c5cf55bae7dbb918be4"} Oct 05 07:04:36 crc kubenswrapper[4846]: I1005 07:04:36.157967 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-66f6d6849b-62zd4" event={"ID":"95847a57-fbe8-43f2-b4b4-e630665f17e3","Type":"ContainerStarted","Data":"5d94dee89587359d58abf575c679734680e135c924834f6e8b178871ef6acbbd"} Oct 05 07:04:36 crc kubenswrapper[4846]: E1005 07:04:36.167275 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/placement-operator@sha256:adc23c5fd1aece2b16dc8e22ceed628f9a719455e39d3f98c77544665c6749e1\\\"\"" pod="openstack-operators/placement-operator-controller-manager-66f6d6849b-62zd4" podUID="95847a57-fbe8-43f2-b4b4-e630665f17e3" Oct 05 07:04:36 crc kubenswrapper[4846]: I1005 07:04:36.171358 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-f589c7597-jzsnd" event={"ID":"3053ab31-f463-49b7-9b32-d37b0506462b","Type":"ContainerStarted","Data":"0c56954fec9b3d7b87b0e7a1ff29d240a2558c06dbf891e805a06ff8bd064628"} Oct 05 07:04:36 crc kubenswrapper[4846]: E1005 07:04:36.172943 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/telemetry-operator@sha256:bf55026ba10b80e1e24733078bd204cef8766d21a305fd000707a1e3b30ff52e\\\"\"" pod="openstack-operators/telemetry-operator-controller-manager-f589c7597-jzsnd" podUID="3053ab31-f463-49b7-9b32-d37b0506462b" Oct 05 07:04:36 crc kubenswrapper[4846]: I1005 07:04:36.187074 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-7cfc658b9-22b7k" event={"ID":"d7419a91-cc67-4cdb-b8a4-d41480e5854f","Type":"ContainerStarted","Data":"1462049c51cf30683922f73b6e9c3a3bf0b165f41716411e7e2fe01f0493da5d"} Oct 05 07:04:36 crc kubenswrapper[4846]: I1005 07:04:36.187145 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-7cfc658b9-22b7k" event={"ID":"d7419a91-cc67-4cdb-b8a4-d41480e5854f","Type":"ContainerStarted","Data":"fcd1cc021bbc5fa9b9cbfe78c55a24cdfd06022aa7ac538651cd2066389db9b2"} Oct 05 07:04:36 crc kubenswrapper[4846]: I1005 07:04:36.246639 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-manager-7cfc658b9-22b7k" Oct 05 07:04:36 crc kubenswrapper[4846]: I1005 07:04:36.254778 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-5d98cc5575-blkpr" event={"ID":"d4de72d9-7d9c-4d90-ba8a-4179d9028033","Type":"ContainerStarted","Data":"d3e49c97956402a754c38bfeaa7f763b32f1df61c56a82af624815d891c2a829"} Oct 05 07:04:36 crc kubenswrapper[4846]: E1005 07:04:36.261432 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:64f57b2b59dea2bd9fae91490c5bec2687131884a049e6579819d9f951b877c6\\\"\"" pod="openstack-operators/watcher-operator-controller-manager-5d98cc5575-blkpr" podUID="d4de72d9-7d9c-4d90-ba8a-4179d9028033" Oct 05 07:04:36 crc kubenswrapper[4846]: I1005 07:04:36.264436 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-c968bb45-rmwgr" event={"ID":"945ab775-ffae-41cb-a315-5e05cf65484d","Type":"ContainerStarted","Data":"c0b9264875e2b1ea54be470f82e85d178a4bff4b644e9ae75158df9bc253d601"} Oct 05 07:04:36 crc kubenswrapper[4846]: I1005 07:04:36.284008 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-manager-7cfc658b9-22b7k" podStartSLOduration=3.283975215 podStartE2EDuration="3.283975215s" podCreationTimestamp="2025-10-05 07:04:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 07:04:36.277199664 +0000 UTC m=+998.518052439" watchObservedRunningTime="2025-10-05 07:04:36.283975215 +0000 UTC m=+998.524827990" Oct 05 07:04:36 crc kubenswrapper[4846]: E1005 07:04:36.290765 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/ovn-operator@sha256:f923b76c1dd8fde02a5faf8a0a433cfacfb7b743f371de64a12e30d6efcde254\\\"\"" pod="openstack-operators/ovn-operator-controller-manager-c968bb45-rmwgr" podUID="945ab775-ffae-41cb-a315-5e05cf65484d" Oct 05 07:04:36 crc kubenswrapper[4846]: E1005 07:04:36.290780 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:0daf76cc40ab619ae266b11defcc1b65beb22d859369e7b1b04de9169089a4cb\\\"\"" pod="openstack-operators/test-operator-controller-manager-6bb6dcddc-4krwr" podUID="b7a3163e-66e4-406c-9285-e2d125024bc8" Oct 05 07:04:36 crc kubenswrapper[4846]: E1005 07:04:36.291348 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/nova-operator@sha256:e4c4ff39c54c0af231fb781759ab50ed86285c74d38bdea43fa75646b762d842\\\"\"" pod="openstack-operators/nova-operator-controller-manager-6c9b57c67-ggtrm" podUID="49e72a92-e750-4046-9894-f3a0359531b2" Oct 05 07:04:37 crc kubenswrapper[4846]: E1005 07:04:37.277525 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/ovn-operator@sha256:f923b76c1dd8fde02a5faf8a0a433cfacfb7b743f371de64a12e30d6efcde254\\\"\"" pod="openstack-operators/ovn-operator-controller-manager-c968bb45-rmwgr" podUID="945ab775-ffae-41cb-a315-5e05cf65484d" Oct 05 07:04:37 crc kubenswrapper[4846]: E1005 07:04:37.277540 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:64f57b2b59dea2bd9fae91490c5bec2687131884a049e6579819d9f951b877c6\\\"\"" pod="openstack-operators/watcher-operator-controller-manager-5d98cc5575-blkpr" podUID="d4de72d9-7d9c-4d90-ba8a-4179d9028033" Oct 05 07:04:37 crc kubenswrapper[4846]: E1005 07:04:37.278209 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/placement-operator@sha256:adc23c5fd1aece2b16dc8e22ceed628f9a719455e39d3f98c77544665c6749e1\\\"\"" pod="openstack-operators/placement-operator-controller-manager-66f6d6849b-62zd4" podUID="95847a57-fbe8-43f2-b4b4-e630665f17e3" Oct 05 07:04:37 crc kubenswrapper[4846]: E1005 07:04:37.280200 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/telemetry-operator@sha256:bf55026ba10b80e1e24733078bd204cef8766d21a305fd000707a1e3b30ff52e\\\"\"" pod="openstack-operators/telemetry-operator-controller-manager-f589c7597-jzsnd" podUID="3053ab31-f463-49b7-9b32-d37b0506462b" Oct 05 07:04:44 crc kubenswrapper[4846]: I1005 07:04:44.175314 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-manager-7cfc658b9-22b7k" Oct 05 07:04:45 crc kubenswrapper[4846]: I1005 07:04:45.355599 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-6675647785-d2n5v" event={"ID":"a1899a06-0cb1-44aa-ba1e-d6876051bab1","Type":"ContainerStarted","Data":"981ec1e6773ec34a654aca10f27673e1e4f7baccbf0a06bbb35a96c592c32cb6"} Oct 05 07:04:46 crc kubenswrapper[4846]: I1005 07:04:46.466451 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5c497dbdb-g4bkx" event={"ID":"931bcbda-d249-4625-a93c-bfa49dcb38ae","Type":"ContainerStarted","Data":"8447e7c63603b6dc0362054943ca74b8b71160b1ff1ed5f3410b2b74a2235640"} Oct 05 07:04:46 crc kubenswrapper[4846]: I1005 07:04:46.609560 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-84788b6bc5-pg8c6" event={"ID":"f0825244-e759-46a4-a16b-4d993dac339d","Type":"ContainerStarted","Data":"d1371f8fb3fe7821145cd197931708297b631be9461bfa422333b2165a2f01e6"} Oct 05 07:04:46 crc kubenswrapper[4846]: I1005 07:04:46.609598 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-7cb48dbc-cvtkp" event={"ID":"e46909e0-1642-436c-b597-167a42e0a72d","Type":"ContainerStarted","Data":"875c77949a8aba4c13671ac03a9c5c2fab0fa077ceb66a87e6feecc5410c15c1"} Oct 05 07:04:46 crc kubenswrapper[4846]: I1005 07:04:46.609614 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-controller-manager-698456cdc6-w4fst" Oct 05 07:04:46 crc kubenswrapper[4846]: I1005 07:04:46.609628 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/designate-operator-controller-manager-58d86cd59d-gnb7x" Oct 05 07:04:46 crc kubenswrapper[4846]: I1005 07:04:46.609637 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-69f59f9d8-s4qf7" event={"ID":"15b8ca14-e689-430c-84b2-40da6f1e83a5","Type":"ContainerStarted","Data":"cca2eb96b69e88e9fd28dc1f9770bfbb0b1527b18af2ca563809506903e2a632"} Oct 05 07:04:46 crc kubenswrapper[4846]: I1005 07:04:46.609648 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ironic-operator-controller-manager-6f5894c49f-fkx4r" Oct 05 07:04:46 crc kubenswrapper[4846]: I1005 07:04:46.609658 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-698456cdc6-w4fst" event={"ID":"c1b6f754-a4fc-4981-a722-415569d3e7a3","Type":"ContainerStarted","Data":"0ee9f980670bdfa509a454fd7b84ba86d45656fba91d7664601430525da739da"} Oct 05 07:04:46 crc kubenswrapper[4846]: I1005 07:04:46.609667 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-57c9cdcf57-j9742" event={"ID":"6ff6c74d-2f8b-42d0-af60-88baa4399b32","Type":"ContainerStarted","Data":"f9b19bf269e77d93a2a06c191ed8c9759a662cd938c893a2cb8f7286537f31ea"} Oct 05 07:04:46 crc kubenswrapper[4846]: I1005 07:04:46.609676 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-58d86cd59d-gnb7x" event={"ID":"ac8cdf70-77c0-44c9-b61b-11325d91b698","Type":"ContainerStarted","Data":"cc10038da8004dd8b4450aa1747be759c1af2fd43a40dfb219329eaae14c5073"} Oct 05 07:04:46 crc kubenswrapper[4846]: I1005 07:04:46.609688 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-58d86cd59d-gnb7x" event={"ID":"ac8cdf70-77c0-44c9-b61b-11325d91b698","Type":"ContainerStarted","Data":"6b728b12695884d38ca4b4745c9ca46a3d4d8c66dba6387f6db5ef102cfc36a2"} Oct 05 07:04:46 crc kubenswrapper[4846]: I1005 07:04:46.609696 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-6f5894c49f-fkx4r" event={"ID":"92c93055-428a-4d41-b034-a805b68f3e19","Type":"ContainerStarted","Data":"13124e5c16743db2d921b84d877d27a9bc8d2c7738b88121da80ba07f70d910e"} Oct 05 07:04:46 crc kubenswrapper[4846]: I1005 07:04:46.620994 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-controller-manager-698456cdc6-w4fst" podStartSLOduration=4.066672179 podStartE2EDuration="14.620975814s" podCreationTimestamp="2025-10-05 07:04:32 +0000 UTC" firstStartedPulling="2025-10-05 07:04:34.379713461 +0000 UTC m=+996.620566236" lastFinishedPulling="2025-10-05 07:04:44.934017096 +0000 UTC m=+1007.174869871" observedRunningTime="2025-10-05 07:04:46.618741725 +0000 UTC m=+1008.859594500" watchObservedRunningTime="2025-10-05 07:04:46.620975814 +0000 UTC m=+1008.861828589" Oct 05 07:04:46 crc kubenswrapper[4846]: I1005 07:04:46.628659 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-6675647785-d2n5v" event={"ID":"a1899a06-0cb1-44aa-ba1e-d6876051bab1","Type":"ContainerStarted","Data":"da96443d62e4f9212fb6138de3629635fc45b8155fa2615170a005d9cb75770d"} Oct 05 07:04:46 crc kubenswrapper[4846]: I1005 07:04:46.629808 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-6675647785-d2n5v" Oct 05 07:04:46 crc kubenswrapper[4846]: I1005 07:04:46.661603 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-5b974f6766-khk5k" event={"ID":"f626116b-4f3a-48cb-ad75-c6550333ee6e","Type":"ContainerStarted","Data":"916410ab966dd9aad830cab8903e1a48ed334491f86348ac229564547ebd743d"} Oct 05 07:04:46 crc kubenswrapper[4846]: I1005 07:04:46.661654 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-5b974f6766-khk5k" event={"ID":"f626116b-4f3a-48cb-ad75-c6550333ee6e","Type":"ContainerStarted","Data":"5a6950bcde9362be6b507927480ba7d27f599f5849af1da0c7708e0b5152c709"} Oct 05 07:04:46 crc kubenswrapper[4846]: I1005 07:04:46.662635 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/barbican-operator-controller-manager-5b974f6766-khk5k" Oct 05 07:04:46 crc kubenswrapper[4846]: I1005 07:04:46.697396 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/designate-operator-controller-manager-58d86cd59d-gnb7x" podStartSLOduration=3.462117977 podStartE2EDuration="14.697365051s" podCreationTimestamp="2025-10-05 07:04:32 +0000 UTC" firstStartedPulling="2025-10-05 07:04:33.75531346 +0000 UTC m=+995.996166225" lastFinishedPulling="2025-10-05 07:04:44.990560524 +0000 UTC m=+1007.231413299" observedRunningTime="2025-10-05 07:04:46.673386612 +0000 UTC m=+1008.914239387" watchObservedRunningTime="2025-10-05 07:04:46.697365051 +0000 UTC m=+1008.938217826" Oct 05 07:04:46 crc kubenswrapper[4846]: I1005 07:04:46.701111 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-69b956fbf6-xsltv" event={"ID":"644a3a04-e2d4-4352-9bac-e8d661fd61b0","Type":"ContainerStarted","Data":"0768e073ddfec898a4e1d60c2db0d00bfdcb1af6a0ea0cba7de456b5188f3cd2"} Oct 05 07:04:46 crc kubenswrapper[4846]: I1005 07:04:46.701217 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/neutron-operator-controller-manager-69b956fbf6-xsltv" Oct 05 07:04:46 crc kubenswrapper[4846]: I1005 07:04:46.701236 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-69b956fbf6-xsltv" event={"ID":"644a3a04-e2d4-4352-9bac-e8d661fd61b0","Type":"ContainerStarted","Data":"33614cb82fdb3716270860461b16db2e7893428fa228438d26be8ae0fded7c11"} Oct 05 07:04:46 crc kubenswrapper[4846]: I1005 07:04:46.723819 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-66cc85b5d5lsczg" event={"ID":"eb017e3a-1627-4dc7-bda3-1faec8b72739","Type":"ContainerStarted","Data":"0e372c52a0accd114864d187fce165c04beb1daeddde269df438f0dc4aac9562"} Oct 05 07:04:46 crc kubenswrapper[4846]: I1005 07:04:46.739444 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ironic-operator-controller-manager-6f5894c49f-fkx4r" podStartSLOduration=3.91968961 podStartE2EDuration="14.739413913s" podCreationTimestamp="2025-10-05 07:04:32 +0000 UTC" firstStartedPulling="2025-10-05 07:04:34.159629182 +0000 UTC m=+996.400481957" lastFinishedPulling="2025-10-05 07:04:44.979353475 +0000 UTC m=+1007.220206260" observedRunningTime="2025-10-05 07:04:46.724135825 +0000 UTC m=+1008.964988600" watchObservedRunningTime="2025-10-05 07:04:46.739413913 +0000 UTC m=+1008.980266688" Oct 05 07:04:46 crc kubenswrapper[4846]: I1005 07:04:46.758888 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-pdqdx" event={"ID":"aae22754-9bda-494a-afb8-42debef021dd","Type":"ContainerStarted","Data":"ca588833c9d428ee8dd19ee28bfae2f68584f8f270a268b1e98853ba4f293199"} Oct 05 07:04:46 crc kubenswrapper[4846]: I1005 07:04:46.805454 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-84bd8f6848-b4rhd" event={"ID":"58088fc8-5aaa-422a-84cd-d4ef7d10e989","Type":"ContainerStarted","Data":"9cd798cf74b2dda59d8b35f2ad5a5105eef685fd5c816c8c40f2def363848a5d"} Oct 05 07:04:46 crc kubenswrapper[4846]: I1005 07:04:46.812577 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/neutron-operator-controller-manager-69b956fbf6-xsltv" podStartSLOduration=4.245159469 podStartE2EDuration="14.812544093s" podCreationTimestamp="2025-10-05 07:04:32 +0000 UTC" firstStartedPulling="2025-10-05 07:04:34.360112688 +0000 UTC m=+996.600965463" lastFinishedPulling="2025-10-05 07:04:44.927497312 +0000 UTC m=+1007.168350087" observedRunningTime="2025-10-05 07:04:46.793792433 +0000 UTC m=+1009.034645208" watchObservedRunningTime="2025-10-05 07:04:46.812544093 +0000 UTC m=+1009.053396868" Oct 05 07:04:46 crc kubenswrapper[4846]: I1005 07:04:46.813125 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-controller-manager-6675647785-d2n5v" podStartSLOduration=3.843302782 podStartE2EDuration="14.813119358s" podCreationTimestamp="2025-10-05 07:04:32 +0000 UTC" firstStartedPulling="2025-10-05 07:04:33.942729897 +0000 UTC m=+996.183582672" lastFinishedPulling="2025-10-05 07:04:44.912546463 +0000 UTC m=+1007.153399248" observedRunningTime="2025-10-05 07:04:46.760864075 +0000 UTC m=+1009.001716850" watchObservedRunningTime="2025-10-05 07:04:46.813119358 +0000 UTC m=+1009.053972133" Oct 05 07:04:46 crc kubenswrapper[4846]: I1005 07:04:46.821657 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-76d5577b-pkjcv" event={"ID":"ab1677d6-3cb0-4dce-8e8d-4d2b67025e6d","Type":"ContainerStarted","Data":"ff6d6a43b0a9857c66926140d5fc2e5e8d1d496ba113861a6011674f657d442a"} Oct 05 07:04:46 crc kubenswrapper[4846]: I1005 07:04:46.825955 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/barbican-operator-controller-manager-5b974f6766-khk5k" podStartSLOduration=4.01607441 podStartE2EDuration="14.8259335s" podCreationTimestamp="2025-10-05 07:04:32 +0000 UTC" firstStartedPulling="2025-10-05 07:04:34.169849984 +0000 UTC m=+996.410702759" lastFinishedPulling="2025-10-05 07:04:44.979709064 +0000 UTC m=+1007.220561849" observedRunningTime="2025-10-05 07:04:46.824874682 +0000 UTC m=+1009.065727457" watchObservedRunningTime="2025-10-05 07:04:46.8259335 +0000 UTC m=+1009.066786275" Oct 05 07:04:46 crc kubenswrapper[4846]: I1005 07:04:46.841652 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-d6c9dc5bc-2fb7l" event={"ID":"71de4f09-b26f-443c-9142-7edbc0d18e38","Type":"ContainerStarted","Data":"6dab05aab134f4ecad5fb1a8e32e772465b03df0d0bd7405c7c07eb1958978e3"} Oct 05 07:04:46 crc kubenswrapper[4846]: I1005 07:04:46.859014 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-pdqdx" podStartSLOduration=3.674211572 podStartE2EDuration="13.858994912s" podCreationTimestamp="2025-10-05 07:04:33 +0000 UTC" firstStartedPulling="2025-10-05 07:04:34.809460742 +0000 UTC m=+997.050313517" lastFinishedPulling="2025-10-05 07:04:44.994244082 +0000 UTC m=+1007.235096857" observedRunningTime="2025-10-05 07:04:46.852800217 +0000 UTC m=+1009.093652992" watchObservedRunningTime="2025-10-05 07:04:46.858994912 +0000 UTC m=+1009.099847687" Oct 05 07:04:47 crc kubenswrapper[4846]: I1005 07:04:47.855913 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-84bd8f6848-b4rhd" event={"ID":"58088fc8-5aaa-422a-84cd-d4ef7d10e989","Type":"ContainerStarted","Data":"646bc466e21f08f2d6116a962bb060c7a3eefe5a4c0220285281f634a9d2314b"} Oct 05 07:04:47 crc kubenswrapper[4846]: I1005 07:04:47.856008 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/cinder-operator-controller-manager-84bd8f6848-b4rhd" Oct 05 07:04:47 crc kubenswrapper[4846]: I1005 07:04:47.861931 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-d6c9dc5bc-2fb7l" event={"ID":"71de4f09-b26f-443c-9142-7edbc0d18e38","Type":"ContainerStarted","Data":"d0d58c29264b38a76a38cc8ecc160795efa1c6ce48bba90e045825bac6fa64e9"} Oct 05 07:04:47 crc kubenswrapper[4846]: I1005 07:04:47.862806 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-d6c9dc5bc-2fb7l" Oct 05 07:04:47 crc kubenswrapper[4846]: I1005 07:04:47.864998 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-84788b6bc5-pg8c6" event={"ID":"f0825244-e759-46a4-a16b-4d993dac339d","Type":"ContainerStarted","Data":"3ac5d74d1ec00beaab97aa78208911eb7a4bc3179bf24b47609e5a6da308bfc3"} Oct 05 07:04:47 crc kubenswrapper[4846]: I1005 07:04:47.865144 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-84788b6bc5-pg8c6" Oct 05 07:04:47 crc kubenswrapper[4846]: I1005 07:04:47.867564 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-698456cdc6-w4fst" event={"ID":"c1b6f754-a4fc-4981-a722-415569d3e7a3","Type":"ContainerStarted","Data":"0fd550a2a8d0d400b1d91c64513e0c1c542ebe039d828f5d64d60f1a1c052cae"} Oct 05 07:04:47 crc kubenswrapper[4846]: I1005 07:04:47.870141 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-7cb48dbc-cvtkp" event={"ID":"e46909e0-1642-436c-b597-167a42e0a72d","Type":"ContainerStarted","Data":"3175127895f97fb601a8968a934e69a40d709092f0cc8739b0be9cfcff8323be"} Oct 05 07:04:47 crc kubenswrapper[4846]: I1005 07:04:47.870701 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/manila-operator-controller-manager-7cb48dbc-cvtkp" Oct 05 07:04:47 crc kubenswrapper[4846]: I1005 07:04:47.875498 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-66cc85b5d5lsczg" event={"ID":"eb017e3a-1627-4dc7-bda3-1faec8b72739","Type":"ContainerStarted","Data":"17a5356460cd51a78abcd3c18bf93436669aa8b74f5ab545b3dcaa425399eaed"} Oct 05 07:04:47 crc kubenswrapper[4846]: I1005 07:04:47.876048 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-baremetal-operator-controller-manager-66cc85b5d5lsczg" Oct 05 07:04:47 crc kubenswrapper[4846]: I1005 07:04:47.885379 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/cinder-operator-controller-manager-84bd8f6848-b4rhd" podStartSLOduration=5.105553135 podStartE2EDuration="15.885348253s" podCreationTimestamp="2025-10-05 07:04:32 +0000 UTC" firstStartedPulling="2025-10-05 07:04:34.223338201 +0000 UTC m=+996.464190976" lastFinishedPulling="2025-10-05 07:04:45.003133319 +0000 UTC m=+1007.243986094" observedRunningTime="2025-10-05 07:04:47.878583942 +0000 UTC m=+1010.119436727" watchObservedRunningTime="2025-10-05 07:04:47.885348253 +0000 UTC m=+1010.126201038" Oct 05 07:04:47 crc kubenswrapper[4846]: I1005 07:04:47.888585 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-6f5894c49f-fkx4r" event={"ID":"92c93055-428a-4d41-b034-a805b68f3e19","Type":"ContainerStarted","Data":"30103164ed710d0a8a28e8291222c31028008a25789b02eb2c9001c492c9b1ce"} Oct 05 07:04:47 crc kubenswrapper[4846]: I1005 07:04:47.891982 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-76d5577b-pkjcv" event={"ID":"ab1677d6-3cb0-4dce-8e8d-4d2b67025e6d","Type":"ContainerStarted","Data":"bbd45718e0656777d6cf723f0d2d97520ca3c8baebc6d1cf34e360c95605eaa6"} Oct 05 07:04:47 crc kubenswrapper[4846]: I1005 07:04:47.892938 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-76d5577b-pkjcv" Oct 05 07:04:47 crc kubenswrapper[4846]: I1005 07:04:47.898150 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5c497dbdb-g4bkx" event={"ID":"931bcbda-d249-4625-a93c-bfa49dcb38ae","Type":"ContainerStarted","Data":"2db99b08f3f1caab9bb1f2aa4abad18dadf0f667d35fd96ae29a9642a9204a78"} Oct 05 07:04:47 crc kubenswrapper[4846]: I1005 07:04:47.898923 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/heat-operator-controller-manager-5c497dbdb-g4bkx" Oct 05 07:04:47 crc kubenswrapper[4846]: I1005 07:04:47.902444 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-69f59f9d8-s4qf7" event={"ID":"15b8ca14-e689-430c-84b2-40da6f1e83a5","Type":"ContainerStarted","Data":"d4ff02fcd782efdbb6c6efc0add47922524a35b7f12649b32837715cbf1221f8"} Oct 05 07:04:47 crc kubenswrapper[4846]: I1005 07:04:47.903456 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/octavia-operator-controller-manager-69f59f9d8-s4qf7" Oct 05 07:04:47 crc kubenswrapper[4846]: I1005 07:04:47.907948 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-57c9cdcf57-j9742" event={"ID":"6ff6c74d-2f8b-42d0-af60-88baa4399b32","Type":"ContainerStarted","Data":"d9a03979054159ec29d311c945a19aaaa3c7358f60b2bebc15445f52d202966c"} Oct 05 07:04:47 crc kubenswrapper[4846]: I1005 07:04:47.908399 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-57c9cdcf57-j9742" Oct 05 07:04:47 crc kubenswrapper[4846]: I1005 07:04:47.923047 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-baremetal-operator-controller-manager-66cc85b5d5lsczg" podStartSLOduration=6.544294193 podStartE2EDuration="15.923022987s" podCreationTimestamp="2025-10-05 07:04:32 +0000 UTC" firstStartedPulling="2025-10-05 07:04:35.555008564 +0000 UTC m=+997.795861339" lastFinishedPulling="2025-10-05 07:04:44.933737358 +0000 UTC m=+1007.174590133" observedRunningTime="2025-10-05 07:04:47.915552278 +0000 UTC m=+1010.156405093" watchObservedRunningTime="2025-10-05 07:04:47.923022987 +0000 UTC m=+1010.163875762" Oct 05 07:04:47 crc kubenswrapper[4846]: I1005 07:04:47.962593 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-controller-manager-d6c9dc5bc-2fb7l" podStartSLOduration=5.451605663 podStartE2EDuration="15.962559272s" podCreationTimestamp="2025-10-05 07:04:32 +0000 UTC" firstStartedPulling="2025-10-05 07:04:34.468771726 +0000 UTC m=+996.709624501" lastFinishedPulling="2025-10-05 07:04:44.979725335 +0000 UTC m=+1007.220578110" observedRunningTime="2025-10-05 07:04:47.950017067 +0000 UTC m=+1010.190869862" watchObservedRunningTime="2025-10-05 07:04:47.962559272 +0000 UTC m=+1010.203412057" Oct 05 07:04:47 crc kubenswrapper[4846]: I1005 07:04:47.968376 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/manila-operator-controller-manager-7cb48dbc-cvtkp" podStartSLOduration=5.320707762 podStartE2EDuration="15.968344746s" podCreationTimestamp="2025-10-05 07:04:32 +0000 UTC" firstStartedPulling="2025-10-05 07:04:34.351063337 +0000 UTC m=+996.591916112" lastFinishedPulling="2025-10-05 07:04:44.998700321 +0000 UTC m=+1007.239553096" observedRunningTime="2025-10-05 07:04:47.966569739 +0000 UTC m=+1010.207422514" watchObservedRunningTime="2025-10-05 07:04:47.968344746 +0000 UTC m=+1010.209197521" Oct 05 07:04:47 crc kubenswrapper[4846]: I1005 07:04:47.988554 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-controller-manager-84788b6bc5-pg8c6" podStartSLOduration=5.595975863 podStartE2EDuration="15.988529374s" podCreationTimestamp="2025-10-05 07:04:32 +0000 UTC" firstStartedPulling="2025-10-05 07:04:34.583912487 +0000 UTC m=+996.824765262" lastFinishedPulling="2025-10-05 07:04:44.976465978 +0000 UTC m=+1007.217318773" observedRunningTime="2025-10-05 07:04:47.981874137 +0000 UTC m=+1010.222726922" watchObservedRunningTime="2025-10-05 07:04:47.988529374 +0000 UTC m=+1010.229382149" Oct 05 07:04:48 crc kubenswrapper[4846]: I1005 07:04:48.007641 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/heat-operator-controller-manager-5c497dbdb-g4bkx" podStartSLOduration=4.957783233 podStartE2EDuration="16.007608863s" podCreationTimestamp="2025-10-05 07:04:32 +0000 UTC" firstStartedPulling="2025-10-05 07:04:33.924806929 +0000 UTC m=+996.165659704" lastFinishedPulling="2025-10-05 07:04:44.974632539 +0000 UTC m=+1007.215485334" observedRunningTime="2025-10-05 07:04:47.998320346 +0000 UTC m=+1010.239173121" watchObservedRunningTime="2025-10-05 07:04:48.007608863 +0000 UTC m=+1010.248461648" Oct 05 07:04:48 crc kubenswrapper[4846]: I1005 07:04:48.026868 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-controller-manager-76d5577b-pkjcv" podStartSLOduration=5.446059685 podStartE2EDuration="16.026843566s" podCreationTimestamp="2025-10-05 07:04:32 +0000 UTC" firstStartedPulling="2025-10-05 07:04:34.351065197 +0000 UTC m=+996.591917972" lastFinishedPulling="2025-10-05 07:04:44.931849078 +0000 UTC m=+1007.172701853" observedRunningTime="2025-10-05 07:04:48.021161885 +0000 UTC m=+1010.262014670" watchObservedRunningTime="2025-10-05 07:04:48.026843566 +0000 UTC m=+1010.267696331" Oct 05 07:04:48 crc kubenswrapper[4846]: I1005 07:04:48.039501 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-controller-manager-57c9cdcf57-j9742" podStartSLOduration=5.248896827 podStartE2EDuration="16.039476333s" podCreationTimestamp="2025-10-05 07:04:32 +0000 UTC" firstStartedPulling="2025-10-05 07:04:34.188346937 +0000 UTC m=+996.429199712" lastFinishedPulling="2025-10-05 07:04:44.978926433 +0000 UTC m=+1007.219779218" observedRunningTime="2025-10-05 07:04:48.037837609 +0000 UTC m=+1010.278690394" watchObservedRunningTime="2025-10-05 07:04:48.039476333 +0000 UTC m=+1010.280329108" Oct 05 07:04:48 crc kubenswrapper[4846]: I1005 07:04:48.063780 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/octavia-operator-controller-manager-69f59f9d8-s4qf7" podStartSLOduration=5.556684965 podStartE2EDuration="16.06375369s" podCreationTimestamp="2025-10-05 07:04:32 +0000 UTC" firstStartedPulling="2025-10-05 07:04:34.474653893 +0000 UTC m=+996.715506668" lastFinishedPulling="2025-10-05 07:04:44.981722618 +0000 UTC m=+1007.222575393" observedRunningTime="2025-10-05 07:04:48.05923956 +0000 UTC m=+1010.300092335" watchObservedRunningTime="2025-10-05 07:04:48.06375369 +0000 UTC m=+1010.304606465" Oct 05 07:04:52 crc kubenswrapper[4846]: I1005 07:04:52.622026 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/designate-operator-controller-manager-58d86cd59d-gnb7x" Oct 05 07:04:52 crc kubenswrapper[4846]: I1005 07:04:52.668617 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/heat-operator-controller-manager-5c497dbdb-g4bkx" Oct 05 07:04:52 crc kubenswrapper[4846]: I1005 07:04:52.707803 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-controller-manager-6675647785-d2n5v" Oct 05 07:04:52 crc kubenswrapper[4846]: I1005 07:04:52.772300 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ironic-operator-controller-manager-6f5894c49f-fkx4r" Oct 05 07:04:52 crc kubenswrapper[4846]: I1005 07:04:52.867471 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-57c9cdcf57-j9742" Oct 05 07:04:52 crc kubenswrapper[4846]: I1005 07:04:52.888444 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/barbican-operator-controller-manager-5b974f6766-khk5k" Oct 05 07:04:52 crc kubenswrapper[4846]: I1005 07:04:52.896744 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/cinder-operator-controller-manager-84bd8f6848-b4rhd" Oct 05 07:04:52 crc kubenswrapper[4846]: I1005 07:04:52.955983 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-controller-manager-698456cdc6-w4fst" Oct 05 07:04:52 crc kubenswrapper[4846]: I1005 07:04:52.968244 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-6bb6dcddc-4krwr" event={"ID":"b7a3163e-66e4-406c-9285-e2d125024bc8","Type":"ContainerStarted","Data":"8019a80291dc66e84c717a5c681f305aa738603fbf48ac2d1e11d1e6326220e3"} Oct 05 07:04:52 crc kubenswrapper[4846]: I1005 07:04:52.969064 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/test-operator-controller-manager-6bb6dcddc-4krwr" Oct 05 07:04:52 crc kubenswrapper[4846]: I1005 07:04:52.973903 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-66f6d6849b-62zd4" event={"ID":"95847a57-fbe8-43f2-b4b4-e630665f17e3","Type":"ContainerStarted","Data":"6d768dd99d03159d5b68227a6b1b10d0f7e025a90f0be713d8fc5458a5d52e04"} Oct 05 07:04:52 crc kubenswrapper[4846]: I1005 07:04:52.974688 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/placement-operator-controller-manager-66f6d6849b-62zd4" Oct 05 07:04:53 crc kubenswrapper[4846]: I1005 07:04:53.001279 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/test-operator-controller-manager-6bb6dcddc-4krwr" podStartSLOduration=3.8975867600000003 podStartE2EDuration="21.001257004s" podCreationTimestamp="2025-10-05 07:04:32 +0000 UTC" firstStartedPulling="2025-10-05 07:04:34.586875366 +0000 UTC m=+996.827728141" lastFinishedPulling="2025-10-05 07:04:51.69054561 +0000 UTC m=+1013.931398385" observedRunningTime="2025-10-05 07:04:52.991601886 +0000 UTC m=+1015.232454661" watchObservedRunningTime="2025-10-05 07:04:53.001257004 +0000 UTC m=+1015.242109789" Oct 05 07:04:53 crc kubenswrapper[4846]: I1005 07:04:53.014814 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/placement-operator-controller-manager-66f6d6849b-62zd4" podStartSLOduration=3.97861515 podStartE2EDuration="21.014790275s" podCreationTimestamp="2025-10-05 07:04:32 +0000 UTC" firstStartedPulling="2025-10-05 07:04:34.650241455 +0000 UTC m=+996.891094230" lastFinishedPulling="2025-10-05 07:04:51.68641657 +0000 UTC m=+1013.927269355" observedRunningTime="2025-10-05 07:04:53.007370217 +0000 UTC m=+1015.248222992" watchObservedRunningTime="2025-10-05 07:04:53.014790275 +0000 UTC m=+1015.255643050" Oct 05 07:04:53 crc kubenswrapper[4846]: I1005 07:04:53.042845 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/manila-operator-controller-manager-7cb48dbc-cvtkp" Oct 05 07:04:53 crc kubenswrapper[4846]: I1005 07:04:53.092896 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/neutron-operator-controller-manager-69b956fbf6-xsltv" Oct 05 07:04:53 crc kubenswrapper[4846]: I1005 07:04:53.140982 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-d6c9dc5bc-2fb7l" Oct 05 07:04:53 crc kubenswrapper[4846]: I1005 07:04:53.241041 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/octavia-operator-controller-manager-69f59f9d8-s4qf7" Oct 05 07:04:53 crc kubenswrapper[4846]: I1005 07:04:53.390560 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-84788b6bc5-pg8c6" Oct 05 07:04:53 crc kubenswrapper[4846]: I1005 07:04:53.449859 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-controller-manager-76d5577b-pkjcv" Oct 05 07:04:54 crc kubenswrapper[4846]: I1005 07:04:54.821503 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-baremetal-operator-controller-manager-66cc85b5d5lsczg" Oct 05 07:04:56 crc kubenswrapper[4846]: I1005 07:04:56.029590 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-f589c7597-jzsnd" event={"ID":"3053ab31-f463-49b7-9b32-d37b0506462b","Type":"ContainerStarted","Data":"9bcf166051ee70be26e1d5e201fac76cf9ed0a6508bbf33bceea71ede7953b93"} Oct 05 07:04:56 crc kubenswrapper[4846]: I1005 07:04:56.030467 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/telemetry-operator-controller-manager-f589c7597-jzsnd" Oct 05 07:04:56 crc kubenswrapper[4846]: I1005 07:04:56.033551 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-5d98cc5575-blkpr" event={"ID":"d4de72d9-7d9c-4d90-ba8a-4179d9028033","Type":"ContainerStarted","Data":"2b4370a362ed7690613ebcf8501d0d7b42be8b41a8bd060faa531b326711faf5"} Oct 05 07:04:56 crc kubenswrapper[4846]: I1005 07:04:56.034525 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/watcher-operator-controller-manager-5d98cc5575-blkpr" Oct 05 07:04:56 crc kubenswrapper[4846]: I1005 07:04:56.036685 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-c968bb45-rmwgr" event={"ID":"945ab775-ffae-41cb-a315-5e05cf65484d","Type":"ContainerStarted","Data":"5fb60585a3804959802579304368567e29850296cdddf7cbc2f239069a7cea82"} Oct 05 07:04:56 crc kubenswrapper[4846]: I1005 07:04:56.037013 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ovn-operator-controller-manager-c968bb45-rmwgr" Oct 05 07:04:56 crc kubenswrapper[4846]: I1005 07:04:56.038631 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-6c9b57c67-ggtrm" event={"ID":"49e72a92-e750-4046-9894-f3a0359531b2","Type":"ContainerStarted","Data":"e4e6f4fd1e593e84c2dda279e2dae7d822d2fde54fcd26c6c2ebdcf27ee0eab3"} Oct 05 07:04:56 crc kubenswrapper[4846]: I1005 07:04:56.038940 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/nova-operator-controller-manager-6c9b57c67-ggtrm" Oct 05 07:04:56 crc kubenswrapper[4846]: I1005 07:04:56.059169 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/telemetry-operator-controller-manager-f589c7597-jzsnd" podStartSLOduration=3.806899701 podStartE2EDuration="24.059137012s" podCreationTimestamp="2025-10-05 07:04:32 +0000 UTC" firstStartedPulling="2025-10-05 07:04:34.628772913 +0000 UTC m=+996.869625688" lastFinishedPulling="2025-10-05 07:04:54.881010214 +0000 UTC m=+1017.121862999" observedRunningTime="2025-10-05 07:04:56.052115505 +0000 UTC m=+1018.292968290" watchObservedRunningTime="2025-10-05 07:04:56.059137012 +0000 UTC m=+1018.299989797" Oct 05 07:04:56 crc kubenswrapper[4846]: I1005 07:04:56.072729 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ovn-operator-controller-manager-c968bb45-rmwgr" podStartSLOduration=3.81772258 podStartE2EDuration="24.072700994s" podCreationTimestamp="2025-10-05 07:04:32 +0000 UTC" firstStartedPulling="2025-10-05 07:04:34.62829081 +0000 UTC m=+996.869143585" lastFinishedPulling="2025-10-05 07:04:54.883269224 +0000 UTC m=+1017.124121999" observedRunningTime="2025-10-05 07:04:56.069498848 +0000 UTC m=+1018.310351633" watchObservedRunningTime="2025-10-05 07:04:56.072700994 +0000 UTC m=+1018.313553769" Oct 05 07:04:56 crc kubenswrapper[4846]: I1005 07:04:56.096151 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/watcher-operator-controller-manager-5d98cc5575-blkpr" podStartSLOduration=3.884364667 podStartE2EDuration="24.096131329s" podCreationTimestamp="2025-10-05 07:04:32 +0000 UTC" firstStartedPulling="2025-10-05 07:04:34.737734209 +0000 UTC m=+996.978586984" lastFinishedPulling="2025-10-05 07:04:54.949500871 +0000 UTC m=+1017.190353646" observedRunningTime="2025-10-05 07:04:56.088997968 +0000 UTC m=+1018.329850743" watchObservedRunningTime="2025-10-05 07:04:56.096131329 +0000 UTC m=+1018.336984104" Oct 05 07:04:56 crc kubenswrapper[4846]: I1005 07:04:56.115211 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/nova-operator-controller-manager-6c9b57c67-ggtrm" podStartSLOduration=3.700524434 podStartE2EDuration="24.115189927s" podCreationTimestamp="2025-10-05 07:04:32 +0000 UTC" firstStartedPulling="2025-10-05 07:04:34.475218198 +0000 UTC m=+996.716070963" lastFinishedPulling="2025-10-05 07:04:54.889883681 +0000 UTC m=+1017.130736456" observedRunningTime="2025-10-05 07:04:56.113735008 +0000 UTC m=+1018.354587813" watchObservedRunningTime="2025-10-05 07:04:56.115189927 +0000 UTC m=+1018.356042702" Oct 05 07:05:03 crc kubenswrapper[4846]: I1005 07:05:03.192707 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/nova-operator-controller-manager-6c9b57c67-ggtrm" Oct 05 07:05:03 crc kubenswrapper[4846]: I1005 07:05:03.387572 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/placement-operator-controller-manager-66f6d6849b-62zd4" Oct 05 07:05:03 crc kubenswrapper[4846]: I1005 07:05:03.496508 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/telemetry-operator-controller-manager-f589c7597-jzsnd" Oct 05 07:05:03 crc kubenswrapper[4846]: I1005 07:05:03.566538 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ovn-operator-controller-manager-c968bb45-rmwgr" Oct 05 07:05:03 crc kubenswrapper[4846]: I1005 07:05:03.833960 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/test-operator-controller-manager-6bb6dcddc-4krwr" Oct 05 07:05:03 crc kubenswrapper[4846]: I1005 07:05:03.901504 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/watcher-operator-controller-manager-5d98cc5575-blkpr" Oct 05 07:05:20 crc kubenswrapper[4846]: I1005 07:05:20.846795 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-546d69f86c-sp94h"] Oct 05 07:05:20 crc kubenswrapper[4846]: I1005 07:05:20.852226 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-546d69f86c-sp94h" Oct 05 07:05:20 crc kubenswrapper[4846]: I1005 07:05:20.857525 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openshift-service-ca.crt" Oct 05 07:05:20 crc kubenswrapper[4846]: I1005 07:05:20.858088 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"kube-root-ca.crt" Oct 05 07:05:20 crc kubenswrapper[4846]: I1005 07:05:20.858224 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns" Oct 05 07:05:20 crc kubenswrapper[4846]: I1005 07:05:20.858347 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dnsmasq-dns-dockercfg-x44cx" Oct 05 07:05:20 crc kubenswrapper[4846]: I1005 07:05:20.875680 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-546d69f86c-sp94h"] Oct 05 07:05:20 crc kubenswrapper[4846]: I1005 07:05:20.922439 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7f9579fb85-mf94n"] Oct 05 07:05:20 crc kubenswrapper[4846]: I1005 07:05:20.923897 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7f9579fb85-mf94n" Oct 05 07:05:20 crc kubenswrapper[4846]: I1005 07:05:20.929162 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-svc" Oct 05 07:05:20 crc kubenswrapper[4846]: I1005 07:05:20.944873 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7f9579fb85-mf94n"] Oct 05 07:05:21 crc kubenswrapper[4846]: I1005 07:05:21.040318 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6f3484f0-58c1-4b8b-bb79-04bfdbf3b5d9-dns-svc\") pod \"dnsmasq-dns-7f9579fb85-mf94n\" (UID: \"6f3484f0-58c1-4b8b-bb79-04bfdbf3b5d9\") " pod="openstack/dnsmasq-dns-7f9579fb85-mf94n" Oct 05 07:05:21 crc kubenswrapper[4846]: I1005 07:05:21.040381 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b04f83fc-2da1-49f8-a485-a2a437ff8914-config\") pod \"dnsmasq-dns-546d69f86c-sp94h\" (UID: \"b04f83fc-2da1-49f8-a485-a2a437ff8914\") " pod="openstack/dnsmasq-dns-546d69f86c-sp94h" Oct 05 07:05:21 crc kubenswrapper[4846]: I1005 07:05:21.040410 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-29g7r\" (UniqueName: \"kubernetes.io/projected/b04f83fc-2da1-49f8-a485-a2a437ff8914-kube-api-access-29g7r\") pod \"dnsmasq-dns-546d69f86c-sp94h\" (UID: \"b04f83fc-2da1-49f8-a485-a2a437ff8914\") " pod="openstack/dnsmasq-dns-546d69f86c-sp94h" Oct 05 07:05:21 crc kubenswrapper[4846]: I1005 07:05:21.040789 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6f3484f0-58c1-4b8b-bb79-04bfdbf3b5d9-config\") pod \"dnsmasq-dns-7f9579fb85-mf94n\" (UID: \"6f3484f0-58c1-4b8b-bb79-04bfdbf3b5d9\") " pod="openstack/dnsmasq-dns-7f9579fb85-mf94n" Oct 05 07:05:21 crc kubenswrapper[4846]: I1005 07:05:21.040900 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v8xpv\" (UniqueName: \"kubernetes.io/projected/6f3484f0-58c1-4b8b-bb79-04bfdbf3b5d9-kube-api-access-v8xpv\") pod \"dnsmasq-dns-7f9579fb85-mf94n\" (UID: \"6f3484f0-58c1-4b8b-bb79-04bfdbf3b5d9\") " pod="openstack/dnsmasq-dns-7f9579fb85-mf94n" Oct 05 07:05:21 crc kubenswrapper[4846]: I1005 07:05:21.142897 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6f3484f0-58c1-4b8b-bb79-04bfdbf3b5d9-config\") pod \"dnsmasq-dns-7f9579fb85-mf94n\" (UID: \"6f3484f0-58c1-4b8b-bb79-04bfdbf3b5d9\") " pod="openstack/dnsmasq-dns-7f9579fb85-mf94n" Oct 05 07:05:21 crc kubenswrapper[4846]: I1005 07:05:21.142956 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v8xpv\" (UniqueName: \"kubernetes.io/projected/6f3484f0-58c1-4b8b-bb79-04bfdbf3b5d9-kube-api-access-v8xpv\") pod \"dnsmasq-dns-7f9579fb85-mf94n\" (UID: \"6f3484f0-58c1-4b8b-bb79-04bfdbf3b5d9\") " pod="openstack/dnsmasq-dns-7f9579fb85-mf94n" Oct 05 07:05:21 crc kubenswrapper[4846]: I1005 07:05:21.143029 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6f3484f0-58c1-4b8b-bb79-04bfdbf3b5d9-dns-svc\") pod \"dnsmasq-dns-7f9579fb85-mf94n\" (UID: \"6f3484f0-58c1-4b8b-bb79-04bfdbf3b5d9\") " pod="openstack/dnsmasq-dns-7f9579fb85-mf94n" Oct 05 07:05:21 crc kubenswrapper[4846]: I1005 07:05:21.143056 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b04f83fc-2da1-49f8-a485-a2a437ff8914-config\") pod \"dnsmasq-dns-546d69f86c-sp94h\" (UID: \"b04f83fc-2da1-49f8-a485-a2a437ff8914\") " pod="openstack/dnsmasq-dns-546d69f86c-sp94h" Oct 05 07:05:21 crc kubenswrapper[4846]: I1005 07:05:21.143072 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-29g7r\" (UniqueName: \"kubernetes.io/projected/b04f83fc-2da1-49f8-a485-a2a437ff8914-kube-api-access-29g7r\") pod \"dnsmasq-dns-546d69f86c-sp94h\" (UID: \"b04f83fc-2da1-49f8-a485-a2a437ff8914\") " pod="openstack/dnsmasq-dns-546d69f86c-sp94h" Oct 05 07:05:21 crc kubenswrapper[4846]: I1005 07:05:21.144441 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6f3484f0-58c1-4b8b-bb79-04bfdbf3b5d9-config\") pod \"dnsmasq-dns-7f9579fb85-mf94n\" (UID: \"6f3484f0-58c1-4b8b-bb79-04bfdbf3b5d9\") " pod="openstack/dnsmasq-dns-7f9579fb85-mf94n" Oct 05 07:05:21 crc kubenswrapper[4846]: I1005 07:05:21.145314 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6f3484f0-58c1-4b8b-bb79-04bfdbf3b5d9-dns-svc\") pod \"dnsmasq-dns-7f9579fb85-mf94n\" (UID: \"6f3484f0-58c1-4b8b-bb79-04bfdbf3b5d9\") " pod="openstack/dnsmasq-dns-7f9579fb85-mf94n" Oct 05 07:05:21 crc kubenswrapper[4846]: I1005 07:05:21.145595 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b04f83fc-2da1-49f8-a485-a2a437ff8914-config\") pod \"dnsmasq-dns-546d69f86c-sp94h\" (UID: \"b04f83fc-2da1-49f8-a485-a2a437ff8914\") " pod="openstack/dnsmasq-dns-546d69f86c-sp94h" Oct 05 07:05:21 crc kubenswrapper[4846]: I1005 07:05:21.170427 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v8xpv\" (UniqueName: \"kubernetes.io/projected/6f3484f0-58c1-4b8b-bb79-04bfdbf3b5d9-kube-api-access-v8xpv\") pod \"dnsmasq-dns-7f9579fb85-mf94n\" (UID: \"6f3484f0-58c1-4b8b-bb79-04bfdbf3b5d9\") " pod="openstack/dnsmasq-dns-7f9579fb85-mf94n" Oct 05 07:05:21 crc kubenswrapper[4846]: I1005 07:05:21.173341 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-29g7r\" (UniqueName: \"kubernetes.io/projected/b04f83fc-2da1-49f8-a485-a2a437ff8914-kube-api-access-29g7r\") pod \"dnsmasq-dns-546d69f86c-sp94h\" (UID: \"b04f83fc-2da1-49f8-a485-a2a437ff8914\") " pod="openstack/dnsmasq-dns-546d69f86c-sp94h" Oct 05 07:05:21 crc kubenswrapper[4846]: I1005 07:05:21.182062 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-546d69f86c-sp94h" Oct 05 07:05:21 crc kubenswrapper[4846]: I1005 07:05:21.247564 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7f9579fb85-mf94n" Oct 05 07:05:21 crc kubenswrapper[4846]: I1005 07:05:21.591456 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-546d69f86c-sp94h"] Oct 05 07:05:21 crc kubenswrapper[4846]: I1005 07:05:21.600805 4846 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 05 07:05:21 crc kubenswrapper[4846]: I1005 07:05:21.747667 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7f9579fb85-mf94n"] Oct 05 07:05:21 crc kubenswrapper[4846]: W1005 07:05:21.756540 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6f3484f0_58c1_4b8b_bb79_04bfdbf3b5d9.slice/crio-dc41e06be71625cf73c16c4d0d36a79d4562f6dba94e1361e9364894c90013c5 WatchSource:0}: Error finding container dc41e06be71625cf73c16c4d0d36a79d4562f6dba94e1361e9364894c90013c5: Status 404 returned error can't find the container with id dc41e06be71625cf73c16c4d0d36a79d4562f6dba94e1361e9364894c90013c5 Oct 05 07:05:22 crc kubenswrapper[4846]: I1005 07:05:22.331457 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-546d69f86c-sp94h" event={"ID":"b04f83fc-2da1-49f8-a485-a2a437ff8914","Type":"ContainerStarted","Data":"ce8c0bc51284cbbf6261fbe3940924f87c2bc8d10a86a9e35bf19ae08209d0c1"} Oct 05 07:05:22 crc kubenswrapper[4846]: I1005 07:05:22.333367 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7f9579fb85-mf94n" event={"ID":"6f3484f0-58c1-4b8b-bb79-04bfdbf3b5d9","Type":"ContainerStarted","Data":"dc41e06be71625cf73c16c4d0d36a79d4562f6dba94e1361e9364894c90013c5"} Oct 05 07:05:23 crc kubenswrapper[4846]: I1005 07:05:23.749494 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-546d69f86c-sp94h"] Oct 05 07:05:23 crc kubenswrapper[4846]: I1005 07:05:23.783228 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6c64dcbb65-fw9xh"] Oct 05 07:05:23 crc kubenswrapper[4846]: I1005 07:05:23.784631 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6c64dcbb65-fw9xh" Oct 05 07:05:23 crc kubenswrapper[4846]: I1005 07:05:23.797635 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6c64dcbb65-fw9xh"] Oct 05 07:05:23 crc kubenswrapper[4846]: I1005 07:05:23.894988 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s84t5\" (UniqueName: \"kubernetes.io/projected/70abcfd4-3bfd-47c0-bb6a-03883ab21a76-kube-api-access-s84t5\") pod \"dnsmasq-dns-6c64dcbb65-fw9xh\" (UID: \"70abcfd4-3bfd-47c0-bb6a-03883ab21a76\") " pod="openstack/dnsmasq-dns-6c64dcbb65-fw9xh" Oct 05 07:05:23 crc kubenswrapper[4846]: I1005 07:05:23.895118 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/70abcfd4-3bfd-47c0-bb6a-03883ab21a76-dns-svc\") pod \"dnsmasq-dns-6c64dcbb65-fw9xh\" (UID: \"70abcfd4-3bfd-47c0-bb6a-03883ab21a76\") " pod="openstack/dnsmasq-dns-6c64dcbb65-fw9xh" Oct 05 07:05:23 crc kubenswrapper[4846]: I1005 07:05:23.895140 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/70abcfd4-3bfd-47c0-bb6a-03883ab21a76-config\") pod \"dnsmasq-dns-6c64dcbb65-fw9xh\" (UID: \"70abcfd4-3bfd-47c0-bb6a-03883ab21a76\") " pod="openstack/dnsmasq-dns-6c64dcbb65-fw9xh" Oct 05 07:05:23 crc kubenswrapper[4846]: I1005 07:05:23.997070 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s84t5\" (UniqueName: \"kubernetes.io/projected/70abcfd4-3bfd-47c0-bb6a-03883ab21a76-kube-api-access-s84t5\") pod \"dnsmasq-dns-6c64dcbb65-fw9xh\" (UID: \"70abcfd4-3bfd-47c0-bb6a-03883ab21a76\") " pod="openstack/dnsmasq-dns-6c64dcbb65-fw9xh" Oct 05 07:05:23 crc kubenswrapper[4846]: I1005 07:05:23.997229 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/70abcfd4-3bfd-47c0-bb6a-03883ab21a76-dns-svc\") pod \"dnsmasq-dns-6c64dcbb65-fw9xh\" (UID: \"70abcfd4-3bfd-47c0-bb6a-03883ab21a76\") " pod="openstack/dnsmasq-dns-6c64dcbb65-fw9xh" Oct 05 07:05:23 crc kubenswrapper[4846]: I1005 07:05:23.997259 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/70abcfd4-3bfd-47c0-bb6a-03883ab21a76-config\") pod \"dnsmasq-dns-6c64dcbb65-fw9xh\" (UID: \"70abcfd4-3bfd-47c0-bb6a-03883ab21a76\") " pod="openstack/dnsmasq-dns-6c64dcbb65-fw9xh" Oct 05 07:05:23 crc kubenswrapper[4846]: I1005 07:05:23.998519 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/70abcfd4-3bfd-47c0-bb6a-03883ab21a76-config\") pod \"dnsmasq-dns-6c64dcbb65-fw9xh\" (UID: \"70abcfd4-3bfd-47c0-bb6a-03883ab21a76\") " pod="openstack/dnsmasq-dns-6c64dcbb65-fw9xh" Oct 05 07:05:23 crc kubenswrapper[4846]: I1005 07:05:23.998631 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/70abcfd4-3bfd-47c0-bb6a-03883ab21a76-dns-svc\") pod \"dnsmasq-dns-6c64dcbb65-fw9xh\" (UID: \"70abcfd4-3bfd-47c0-bb6a-03883ab21a76\") " pod="openstack/dnsmasq-dns-6c64dcbb65-fw9xh" Oct 05 07:05:24 crc kubenswrapper[4846]: I1005 07:05:24.027904 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s84t5\" (UniqueName: \"kubernetes.io/projected/70abcfd4-3bfd-47c0-bb6a-03883ab21a76-kube-api-access-s84t5\") pod \"dnsmasq-dns-6c64dcbb65-fw9xh\" (UID: \"70abcfd4-3bfd-47c0-bb6a-03883ab21a76\") " pod="openstack/dnsmasq-dns-6c64dcbb65-fw9xh" Oct 05 07:05:24 crc kubenswrapper[4846]: I1005 07:05:24.075435 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7f9579fb85-mf94n"] Oct 05 07:05:24 crc kubenswrapper[4846]: I1005 07:05:24.092779 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5d4d9f7875-29gcr"] Oct 05 07:05:24 crc kubenswrapper[4846]: I1005 07:05:24.097035 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5d4d9f7875-29gcr" Oct 05 07:05:24 crc kubenswrapper[4846]: I1005 07:05:24.109926 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5d4d9f7875-29gcr"] Oct 05 07:05:24 crc kubenswrapper[4846]: I1005 07:05:24.119692 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6c64dcbb65-fw9xh" Oct 05 07:05:24 crc kubenswrapper[4846]: I1005 07:05:24.209675 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7p4vq\" (UniqueName: \"kubernetes.io/projected/5fa7b35e-d3fd-45b6-bb8c-b7599419d53f-kube-api-access-7p4vq\") pod \"dnsmasq-dns-5d4d9f7875-29gcr\" (UID: \"5fa7b35e-d3fd-45b6-bb8c-b7599419d53f\") " pod="openstack/dnsmasq-dns-5d4d9f7875-29gcr" Oct 05 07:05:24 crc kubenswrapper[4846]: I1005 07:05:24.209728 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5fa7b35e-d3fd-45b6-bb8c-b7599419d53f-config\") pod \"dnsmasq-dns-5d4d9f7875-29gcr\" (UID: \"5fa7b35e-d3fd-45b6-bb8c-b7599419d53f\") " pod="openstack/dnsmasq-dns-5d4d9f7875-29gcr" Oct 05 07:05:24 crc kubenswrapper[4846]: I1005 07:05:24.209753 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5fa7b35e-d3fd-45b6-bb8c-b7599419d53f-dns-svc\") pod \"dnsmasq-dns-5d4d9f7875-29gcr\" (UID: \"5fa7b35e-d3fd-45b6-bb8c-b7599419d53f\") " pod="openstack/dnsmasq-dns-5d4d9f7875-29gcr" Oct 05 07:05:24 crc kubenswrapper[4846]: I1005 07:05:24.317745 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7p4vq\" (UniqueName: \"kubernetes.io/projected/5fa7b35e-d3fd-45b6-bb8c-b7599419d53f-kube-api-access-7p4vq\") pod \"dnsmasq-dns-5d4d9f7875-29gcr\" (UID: \"5fa7b35e-d3fd-45b6-bb8c-b7599419d53f\") " pod="openstack/dnsmasq-dns-5d4d9f7875-29gcr" Oct 05 07:05:24 crc kubenswrapper[4846]: I1005 07:05:24.317873 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5fa7b35e-d3fd-45b6-bb8c-b7599419d53f-config\") pod \"dnsmasq-dns-5d4d9f7875-29gcr\" (UID: \"5fa7b35e-d3fd-45b6-bb8c-b7599419d53f\") " pod="openstack/dnsmasq-dns-5d4d9f7875-29gcr" Oct 05 07:05:24 crc kubenswrapper[4846]: I1005 07:05:24.317964 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5fa7b35e-d3fd-45b6-bb8c-b7599419d53f-dns-svc\") pod \"dnsmasq-dns-5d4d9f7875-29gcr\" (UID: \"5fa7b35e-d3fd-45b6-bb8c-b7599419d53f\") " pod="openstack/dnsmasq-dns-5d4d9f7875-29gcr" Oct 05 07:05:24 crc kubenswrapper[4846]: I1005 07:05:24.318979 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5fa7b35e-d3fd-45b6-bb8c-b7599419d53f-config\") pod \"dnsmasq-dns-5d4d9f7875-29gcr\" (UID: \"5fa7b35e-d3fd-45b6-bb8c-b7599419d53f\") " pod="openstack/dnsmasq-dns-5d4d9f7875-29gcr" Oct 05 07:05:24 crc kubenswrapper[4846]: I1005 07:05:24.319298 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5fa7b35e-d3fd-45b6-bb8c-b7599419d53f-dns-svc\") pod \"dnsmasq-dns-5d4d9f7875-29gcr\" (UID: \"5fa7b35e-d3fd-45b6-bb8c-b7599419d53f\") " pod="openstack/dnsmasq-dns-5d4d9f7875-29gcr" Oct 05 07:05:24 crc kubenswrapper[4846]: I1005 07:05:24.368675 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7p4vq\" (UniqueName: \"kubernetes.io/projected/5fa7b35e-d3fd-45b6-bb8c-b7599419d53f-kube-api-access-7p4vq\") pod \"dnsmasq-dns-5d4d9f7875-29gcr\" (UID: \"5fa7b35e-d3fd-45b6-bb8c-b7599419d53f\") " pod="openstack/dnsmasq-dns-5d4d9f7875-29gcr" Oct 05 07:05:24 crc kubenswrapper[4846]: I1005 07:05:24.425784 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5d4d9f7875-29gcr" Oct 05 07:05:24 crc kubenswrapper[4846]: I1005 07:05:24.693455 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6c64dcbb65-fw9xh"] Oct 05 07:05:24 crc kubenswrapper[4846]: I1005 07:05:24.920470 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Oct 05 07:05:24 crc kubenswrapper[4846]: I1005 07:05:24.958077 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 05 07:05:24 crc kubenswrapper[4846]: I1005 07:05:24.965580 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Oct 05 07:05:24 crc kubenswrapper[4846]: I1005 07:05:24.966791 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Oct 05 07:05:24 crc kubenswrapper[4846]: I1005 07:05:24.967136 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Oct 05 07:05:24 crc kubenswrapper[4846]: I1005 07:05:24.967376 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Oct 05 07:05:24 crc kubenswrapper[4846]: I1005 07:05:24.967578 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Oct 05 07:05:24 crc kubenswrapper[4846]: I1005 07:05:24.968851 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Oct 05 07:05:24 crc kubenswrapper[4846]: I1005 07:05:24.971602 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-tm8zp" Oct 05 07:05:24 crc kubenswrapper[4846]: I1005 07:05:24.972652 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 05 07:05:24 crc kubenswrapper[4846]: I1005 07:05:24.998221 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5d4d9f7875-29gcr"] Oct 05 07:05:25 crc kubenswrapper[4846]: I1005 07:05:25.052583 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/35261dbc-cabe-4bf0-88f6-b8613ddec0f1-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"35261dbc-cabe-4bf0-88f6-b8613ddec0f1\") " pod="openstack/rabbitmq-server-0" Oct 05 07:05:25 crc kubenswrapper[4846]: I1005 07:05:25.052672 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/35261dbc-cabe-4bf0-88f6-b8613ddec0f1-server-conf\") pod \"rabbitmq-server-0\" (UID: \"35261dbc-cabe-4bf0-88f6-b8613ddec0f1\") " pod="openstack/rabbitmq-server-0" Oct 05 07:05:25 crc kubenswrapper[4846]: I1005 07:05:25.052737 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/35261dbc-cabe-4bf0-88f6-b8613ddec0f1-config-data\") pod \"rabbitmq-server-0\" (UID: \"35261dbc-cabe-4bf0-88f6-b8613ddec0f1\") " pod="openstack/rabbitmq-server-0" Oct 05 07:05:25 crc kubenswrapper[4846]: I1005 07:05:25.052792 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/35261dbc-cabe-4bf0-88f6-b8613ddec0f1-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"35261dbc-cabe-4bf0-88f6-b8613ddec0f1\") " pod="openstack/rabbitmq-server-0" Oct 05 07:05:25 crc kubenswrapper[4846]: I1005 07:05:25.052831 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j5pb6\" (UniqueName: \"kubernetes.io/projected/35261dbc-cabe-4bf0-88f6-b8613ddec0f1-kube-api-access-j5pb6\") pod \"rabbitmq-server-0\" (UID: \"35261dbc-cabe-4bf0-88f6-b8613ddec0f1\") " pod="openstack/rabbitmq-server-0" Oct 05 07:05:25 crc kubenswrapper[4846]: I1005 07:05:25.052894 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"rabbitmq-server-0\" (UID: \"35261dbc-cabe-4bf0-88f6-b8613ddec0f1\") " pod="openstack/rabbitmq-server-0" Oct 05 07:05:25 crc kubenswrapper[4846]: I1005 07:05:25.052958 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/35261dbc-cabe-4bf0-88f6-b8613ddec0f1-pod-info\") pod \"rabbitmq-server-0\" (UID: \"35261dbc-cabe-4bf0-88f6-b8613ddec0f1\") " pod="openstack/rabbitmq-server-0" Oct 05 07:05:25 crc kubenswrapper[4846]: I1005 07:05:25.052981 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/35261dbc-cabe-4bf0-88f6-b8613ddec0f1-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"35261dbc-cabe-4bf0-88f6-b8613ddec0f1\") " pod="openstack/rabbitmq-server-0" Oct 05 07:05:25 crc kubenswrapper[4846]: I1005 07:05:25.053009 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/35261dbc-cabe-4bf0-88f6-b8613ddec0f1-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"35261dbc-cabe-4bf0-88f6-b8613ddec0f1\") " pod="openstack/rabbitmq-server-0" Oct 05 07:05:25 crc kubenswrapper[4846]: I1005 07:05:25.053026 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/35261dbc-cabe-4bf0-88f6-b8613ddec0f1-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"35261dbc-cabe-4bf0-88f6-b8613ddec0f1\") " pod="openstack/rabbitmq-server-0" Oct 05 07:05:25 crc kubenswrapper[4846]: I1005 07:05:25.053113 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/35261dbc-cabe-4bf0-88f6-b8613ddec0f1-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"35261dbc-cabe-4bf0-88f6-b8613ddec0f1\") " pod="openstack/rabbitmq-server-0" Oct 05 07:05:25 crc kubenswrapper[4846]: I1005 07:05:25.154404 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/35261dbc-cabe-4bf0-88f6-b8613ddec0f1-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"35261dbc-cabe-4bf0-88f6-b8613ddec0f1\") " pod="openstack/rabbitmq-server-0" Oct 05 07:05:25 crc kubenswrapper[4846]: I1005 07:05:25.154501 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j5pb6\" (UniqueName: \"kubernetes.io/projected/35261dbc-cabe-4bf0-88f6-b8613ddec0f1-kube-api-access-j5pb6\") pod \"rabbitmq-server-0\" (UID: \"35261dbc-cabe-4bf0-88f6-b8613ddec0f1\") " pod="openstack/rabbitmq-server-0" Oct 05 07:05:25 crc kubenswrapper[4846]: I1005 07:05:25.154580 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"rabbitmq-server-0\" (UID: \"35261dbc-cabe-4bf0-88f6-b8613ddec0f1\") " pod="openstack/rabbitmq-server-0" Oct 05 07:05:25 crc kubenswrapper[4846]: I1005 07:05:25.154649 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/35261dbc-cabe-4bf0-88f6-b8613ddec0f1-pod-info\") pod \"rabbitmq-server-0\" (UID: \"35261dbc-cabe-4bf0-88f6-b8613ddec0f1\") " pod="openstack/rabbitmq-server-0" Oct 05 07:05:25 crc kubenswrapper[4846]: I1005 07:05:25.154671 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/35261dbc-cabe-4bf0-88f6-b8613ddec0f1-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"35261dbc-cabe-4bf0-88f6-b8613ddec0f1\") " pod="openstack/rabbitmq-server-0" Oct 05 07:05:25 crc kubenswrapper[4846]: I1005 07:05:25.154692 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/35261dbc-cabe-4bf0-88f6-b8613ddec0f1-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"35261dbc-cabe-4bf0-88f6-b8613ddec0f1\") " pod="openstack/rabbitmq-server-0" Oct 05 07:05:25 crc kubenswrapper[4846]: I1005 07:05:25.154714 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/35261dbc-cabe-4bf0-88f6-b8613ddec0f1-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"35261dbc-cabe-4bf0-88f6-b8613ddec0f1\") " pod="openstack/rabbitmq-server-0" Oct 05 07:05:25 crc kubenswrapper[4846]: I1005 07:05:25.154778 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/35261dbc-cabe-4bf0-88f6-b8613ddec0f1-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"35261dbc-cabe-4bf0-88f6-b8613ddec0f1\") " pod="openstack/rabbitmq-server-0" Oct 05 07:05:25 crc kubenswrapper[4846]: I1005 07:05:25.154798 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/35261dbc-cabe-4bf0-88f6-b8613ddec0f1-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"35261dbc-cabe-4bf0-88f6-b8613ddec0f1\") " pod="openstack/rabbitmq-server-0" Oct 05 07:05:25 crc kubenswrapper[4846]: I1005 07:05:25.154834 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/35261dbc-cabe-4bf0-88f6-b8613ddec0f1-server-conf\") pod \"rabbitmq-server-0\" (UID: \"35261dbc-cabe-4bf0-88f6-b8613ddec0f1\") " pod="openstack/rabbitmq-server-0" Oct 05 07:05:25 crc kubenswrapper[4846]: I1005 07:05:25.154880 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/35261dbc-cabe-4bf0-88f6-b8613ddec0f1-config-data\") pod \"rabbitmq-server-0\" (UID: \"35261dbc-cabe-4bf0-88f6-b8613ddec0f1\") " pod="openstack/rabbitmq-server-0" Oct 05 07:05:25 crc kubenswrapper[4846]: I1005 07:05:25.155096 4846 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"rabbitmq-server-0\" (UID: \"35261dbc-cabe-4bf0-88f6-b8613ddec0f1\") device mount path \"/mnt/openstack/pv10\"" pod="openstack/rabbitmq-server-0" Oct 05 07:05:25 crc kubenswrapper[4846]: I1005 07:05:25.156024 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/35261dbc-cabe-4bf0-88f6-b8613ddec0f1-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"35261dbc-cabe-4bf0-88f6-b8613ddec0f1\") " pod="openstack/rabbitmq-server-0" Oct 05 07:05:25 crc kubenswrapper[4846]: I1005 07:05:25.156074 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/35261dbc-cabe-4bf0-88f6-b8613ddec0f1-config-data\") pod \"rabbitmq-server-0\" (UID: \"35261dbc-cabe-4bf0-88f6-b8613ddec0f1\") " pod="openstack/rabbitmq-server-0" Oct 05 07:05:25 crc kubenswrapper[4846]: I1005 07:05:25.156443 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/35261dbc-cabe-4bf0-88f6-b8613ddec0f1-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"35261dbc-cabe-4bf0-88f6-b8613ddec0f1\") " pod="openstack/rabbitmq-server-0" Oct 05 07:05:25 crc kubenswrapper[4846]: I1005 07:05:25.162288 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/35261dbc-cabe-4bf0-88f6-b8613ddec0f1-server-conf\") pod \"rabbitmq-server-0\" (UID: \"35261dbc-cabe-4bf0-88f6-b8613ddec0f1\") " pod="openstack/rabbitmq-server-0" Oct 05 07:05:25 crc kubenswrapper[4846]: I1005 07:05:25.162349 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/35261dbc-cabe-4bf0-88f6-b8613ddec0f1-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"35261dbc-cabe-4bf0-88f6-b8613ddec0f1\") " pod="openstack/rabbitmq-server-0" Oct 05 07:05:25 crc kubenswrapper[4846]: I1005 07:05:25.163334 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/35261dbc-cabe-4bf0-88f6-b8613ddec0f1-pod-info\") pod \"rabbitmq-server-0\" (UID: \"35261dbc-cabe-4bf0-88f6-b8613ddec0f1\") " pod="openstack/rabbitmq-server-0" Oct 05 07:05:25 crc kubenswrapper[4846]: I1005 07:05:25.163621 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/35261dbc-cabe-4bf0-88f6-b8613ddec0f1-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"35261dbc-cabe-4bf0-88f6-b8613ddec0f1\") " pod="openstack/rabbitmq-server-0" Oct 05 07:05:25 crc kubenswrapper[4846]: I1005 07:05:25.164638 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/35261dbc-cabe-4bf0-88f6-b8613ddec0f1-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"35261dbc-cabe-4bf0-88f6-b8613ddec0f1\") " pod="openstack/rabbitmq-server-0" Oct 05 07:05:25 crc kubenswrapper[4846]: I1005 07:05:25.167600 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/35261dbc-cabe-4bf0-88f6-b8613ddec0f1-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"35261dbc-cabe-4bf0-88f6-b8613ddec0f1\") " pod="openstack/rabbitmq-server-0" Oct 05 07:05:25 crc kubenswrapper[4846]: I1005 07:05:25.172643 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j5pb6\" (UniqueName: \"kubernetes.io/projected/35261dbc-cabe-4bf0-88f6-b8613ddec0f1-kube-api-access-j5pb6\") pod \"rabbitmq-server-0\" (UID: \"35261dbc-cabe-4bf0-88f6-b8613ddec0f1\") " pod="openstack/rabbitmq-server-0" Oct 05 07:05:25 crc kubenswrapper[4846]: I1005 07:05:25.184814 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"rabbitmq-server-0\" (UID: \"35261dbc-cabe-4bf0-88f6-b8613ddec0f1\") " pod="openstack/rabbitmq-server-0" Oct 05 07:05:25 crc kubenswrapper[4846]: I1005 07:05:25.269460 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 05 07:05:25 crc kubenswrapper[4846]: I1005 07:05:25.271040 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 05 07:05:25 crc kubenswrapper[4846]: I1005 07:05:25.275534 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Oct 05 07:05:25 crc kubenswrapper[4846]: I1005 07:05:25.276219 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Oct 05 07:05:25 crc kubenswrapper[4846]: I1005 07:05:25.276439 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Oct 05 07:05:25 crc kubenswrapper[4846]: I1005 07:05:25.276888 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-fzrwv" Oct 05 07:05:25 crc kubenswrapper[4846]: I1005 07:05:25.277851 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Oct 05 07:05:25 crc kubenswrapper[4846]: I1005 07:05:25.278061 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Oct 05 07:05:25 crc kubenswrapper[4846]: I1005 07:05:25.278324 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Oct 05 07:05:25 crc kubenswrapper[4846]: I1005 07:05:25.303847 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 05 07:05:25 crc kubenswrapper[4846]: I1005 07:05:25.314370 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 05 07:05:25 crc kubenswrapper[4846]: I1005 07:05:25.382482 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6c64dcbb65-fw9xh" event={"ID":"70abcfd4-3bfd-47c0-bb6a-03883ab21a76","Type":"ContainerStarted","Data":"2b50fdfe935c5c957db115433e7ab1ff933fc94008c7bac99666a100e44b97a7"} Oct 05 07:05:25 crc kubenswrapper[4846]: I1005 07:05:25.384502 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5d4d9f7875-29gcr" event={"ID":"5fa7b35e-d3fd-45b6-bb8c-b7599419d53f","Type":"ContainerStarted","Data":"154248f13f2fb77e6d12d6f625024a46ebd020a7b0eccf5022ac9a11905c36d5"} Oct 05 07:05:25 crc kubenswrapper[4846]: I1005 07:05:25.460100 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/626e6bce-7422-4fbd-b978-dfa5cb93de34-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"626e6bce-7422-4fbd-b978-dfa5cb93de34\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 07:05:25 crc kubenswrapper[4846]: I1005 07:05:25.460234 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/626e6bce-7422-4fbd-b978-dfa5cb93de34-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"626e6bce-7422-4fbd-b978-dfa5cb93de34\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 07:05:25 crc kubenswrapper[4846]: I1005 07:05:25.460526 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/626e6bce-7422-4fbd-b978-dfa5cb93de34-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"626e6bce-7422-4fbd-b978-dfa5cb93de34\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 07:05:25 crc kubenswrapper[4846]: I1005 07:05:25.460678 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"626e6bce-7422-4fbd-b978-dfa5cb93de34\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 07:05:25 crc kubenswrapper[4846]: I1005 07:05:25.460841 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/626e6bce-7422-4fbd-b978-dfa5cb93de34-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"626e6bce-7422-4fbd-b978-dfa5cb93de34\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 07:05:25 crc kubenswrapper[4846]: I1005 07:05:25.460906 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/626e6bce-7422-4fbd-b978-dfa5cb93de34-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"626e6bce-7422-4fbd-b978-dfa5cb93de34\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 07:05:25 crc kubenswrapper[4846]: I1005 07:05:25.460955 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/626e6bce-7422-4fbd-b978-dfa5cb93de34-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"626e6bce-7422-4fbd-b978-dfa5cb93de34\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 07:05:25 crc kubenswrapper[4846]: I1005 07:05:25.461074 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/626e6bce-7422-4fbd-b978-dfa5cb93de34-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"626e6bce-7422-4fbd-b978-dfa5cb93de34\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 07:05:25 crc kubenswrapper[4846]: I1005 07:05:25.461114 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/626e6bce-7422-4fbd-b978-dfa5cb93de34-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"626e6bce-7422-4fbd-b978-dfa5cb93de34\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 07:05:25 crc kubenswrapper[4846]: I1005 07:05:25.461213 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gtvkd\" (UniqueName: \"kubernetes.io/projected/626e6bce-7422-4fbd-b978-dfa5cb93de34-kube-api-access-gtvkd\") pod \"rabbitmq-cell1-server-0\" (UID: \"626e6bce-7422-4fbd-b978-dfa5cb93de34\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 07:05:25 crc kubenswrapper[4846]: I1005 07:05:25.461413 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/626e6bce-7422-4fbd-b978-dfa5cb93de34-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"626e6bce-7422-4fbd-b978-dfa5cb93de34\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 07:05:25 crc kubenswrapper[4846]: I1005 07:05:25.563657 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/626e6bce-7422-4fbd-b978-dfa5cb93de34-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"626e6bce-7422-4fbd-b978-dfa5cb93de34\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 07:05:25 crc kubenswrapper[4846]: I1005 07:05:25.563753 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/626e6bce-7422-4fbd-b978-dfa5cb93de34-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"626e6bce-7422-4fbd-b978-dfa5cb93de34\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 07:05:25 crc kubenswrapper[4846]: I1005 07:05:25.563789 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/626e6bce-7422-4fbd-b978-dfa5cb93de34-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"626e6bce-7422-4fbd-b978-dfa5cb93de34\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 07:05:25 crc kubenswrapper[4846]: I1005 07:05:25.563825 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/626e6bce-7422-4fbd-b978-dfa5cb93de34-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"626e6bce-7422-4fbd-b978-dfa5cb93de34\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 07:05:25 crc kubenswrapper[4846]: I1005 07:05:25.563863 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"626e6bce-7422-4fbd-b978-dfa5cb93de34\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 07:05:25 crc kubenswrapper[4846]: I1005 07:05:25.563920 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/626e6bce-7422-4fbd-b978-dfa5cb93de34-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"626e6bce-7422-4fbd-b978-dfa5cb93de34\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 07:05:25 crc kubenswrapper[4846]: I1005 07:05:25.563944 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/626e6bce-7422-4fbd-b978-dfa5cb93de34-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"626e6bce-7422-4fbd-b978-dfa5cb93de34\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 07:05:25 crc kubenswrapper[4846]: I1005 07:05:25.565442 4846 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"626e6bce-7422-4fbd-b978-dfa5cb93de34\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/rabbitmq-cell1-server-0" Oct 05 07:05:25 crc kubenswrapper[4846]: I1005 07:05:25.565620 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/626e6bce-7422-4fbd-b978-dfa5cb93de34-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"626e6bce-7422-4fbd-b978-dfa5cb93de34\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 07:05:25 crc kubenswrapper[4846]: I1005 07:05:25.565667 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/626e6bce-7422-4fbd-b978-dfa5cb93de34-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"626e6bce-7422-4fbd-b978-dfa5cb93de34\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 07:05:25 crc kubenswrapper[4846]: I1005 07:05:25.565709 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/626e6bce-7422-4fbd-b978-dfa5cb93de34-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"626e6bce-7422-4fbd-b978-dfa5cb93de34\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 07:05:25 crc kubenswrapper[4846]: I1005 07:05:25.565732 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gtvkd\" (UniqueName: \"kubernetes.io/projected/626e6bce-7422-4fbd-b978-dfa5cb93de34-kube-api-access-gtvkd\") pod \"rabbitmq-cell1-server-0\" (UID: \"626e6bce-7422-4fbd-b978-dfa5cb93de34\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 07:05:25 crc kubenswrapper[4846]: I1005 07:05:25.570015 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/626e6bce-7422-4fbd-b978-dfa5cb93de34-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"626e6bce-7422-4fbd-b978-dfa5cb93de34\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 07:05:25 crc kubenswrapper[4846]: I1005 07:05:25.570593 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/626e6bce-7422-4fbd-b978-dfa5cb93de34-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"626e6bce-7422-4fbd-b978-dfa5cb93de34\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 07:05:25 crc kubenswrapper[4846]: I1005 07:05:25.574013 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/626e6bce-7422-4fbd-b978-dfa5cb93de34-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"626e6bce-7422-4fbd-b978-dfa5cb93de34\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 07:05:25 crc kubenswrapper[4846]: I1005 07:05:25.574740 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/626e6bce-7422-4fbd-b978-dfa5cb93de34-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"626e6bce-7422-4fbd-b978-dfa5cb93de34\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 07:05:25 crc kubenswrapper[4846]: I1005 07:05:25.575974 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/626e6bce-7422-4fbd-b978-dfa5cb93de34-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"626e6bce-7422-4fbd-b978-dfa5cb93de34\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 07:05:25 crc kubenswrapper[4846]: I1005 07:05:25.579965 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/626e6bce-7422-4fbd-b978-dfa5cb93de34-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"626e6bce-7422-4fbd-b978-dfa5cb93de34\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 07:05:25 crc kubenswrapper[4846]: I1005 07:05:25.586754 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/626e6bce-7422-4fbd-b978-dfa5cb93de34-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"626e6bce-7422-4fbd-b978-dfa5cb93de34\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 07:05:25 crc kubenswrapper[4846]: I1005 07:05:25.589229 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/626e6bce-7422-4fbd-b978-dfa5cb93de34-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"626e6bce-7422-4fbd-b978-dfa5cb93de34\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 07:05:25 crc kubenswrapper[4846]: I1005 07:05:25.589902 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/626e6bce-7422-4fbd-b978-dfa5cb93de34-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"626e6bce-7422-4fbd-b978-dfa5cb93de34\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 07:05:25 crc kubenswrapper[4846]: I1005 07:05:25.598045 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gtvkd\" (UniqueName: \"kubernetes.io/projected/626e6bce-7422-4fbd-b978-dfa5cb93de34-kube-api-access-gtvkd\") pod \"rabbitmq-cell1-server-0\" (UID: \"626e6bce-7422-4fbd-b978-dfa5cb93de34\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 07:05:25 crc kubenswrapper[4846]: I1005 07:05:25.606939 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"626e6bce-7422-4fbd-b978-dfa5cb93de34\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 07:05:25 crc kubenswrapper[4846]: I1005 07:05:25.900706 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 05 07:05:25 crc kubenswrapper[4846]: I1005 07:05:25.955314 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 05 07:05:26 crc kubenswrapper[4846]: I1005 07:05:26.414213 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"35261dbc-cabe-4bf0-88f6-b8613ddec0f1","Type":"ContainerStarted","Data":"3c185c416d138656ea723c6753d54a6c7cca27516a80c1aaf2874043c2241621"} Oct 05 07:05:26 crc kubenswrapper[4846]: I1005 07:05:26.467576 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 05 07:05:26 crc kubenswrapper[4846]: I1005 07:05:26.610319 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-galera-0"] Oct 05 07:05:26 crc kubenswrapper[4846]: I1005 07:05:26.614537 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Oct 05 07:05:26 crc kubenswrapper[4846]: I1005 07:05:26.618466 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config-data" Oct 05 07:05:26 crc kubenswrapper[4846]: I1005 07:05:26.618763 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-scripts" Oct 05 07:05:26 crc kubenswrapper[4846]: I1005 07:05:26.619333 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-svc" Oct 05 07:05:26 crc kubenswrapper[4846]: I1005 07:05:26.620432 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Oct 05 07:05:26 crc kubenswrapper[4846]: I1005 07:05:26.630595 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"combined-ca-bundle" Oct 05 07:05:26 crc kubenswrapper[4846]: I1005 07:05:26.639000 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-dockercfg-vhfts" Oct 05 07:05:26 crc kubenswrapper[4846]: I1005 07:05:26.642535 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Oct 05 07:05:26 crc kubenswrapper[4846]: I1005 07:05:26.790892 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/d88c48f9-948e-4888-986e-e00163ec93a6-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"d88c48f9-948e-4888-986e-e00163ec93a6\") " pod="openstack/openstack-galera-0" Oct 05 07:05:26 crc kubenswrapper[4846]: I1005 07:05:26.791436 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d88c48f9-948e-4888-986e-e00163ec93a6-operator-scripts\") pod \"openstack-galera-0\" (UID: \"d88c48f9-948e-4888-986e-e00163ec93a6\") " pod="openstack/openstack-galera-0" Oct 05 07:05:26 crc kubenswrapper[4846]: I1005 07:05:26.791461 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"openstack-galera-0\" (UID: \"d88c48f9-948e-4888-986e-e00163ec93a6\") " pod="openstack/openstack-galera-0" Oct 05 07:05:26 crc kubenswrapper[4846]: I1005 07:05:26.791546 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/d88c48f9-948e-4888-986e-e00163ec93a6-config-data-default\") pod \"openstack-galera-0\" (UID: \"d88c48f9-948e-4888-986e-e00163ec93a6\") " pod="openstack/openstack-galera-0" Oct 05 07:05:26 crc kubenswrapper[4846]: I1005 07:05:26.791574 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/d88c48f9-948e-4888-986e-e00163ec93a6-kolla-config\") pod \"openstack-galera-0\" (UID: \"d88c48f9-948e-4888-986e-e00163ec93a6\") " pod="openstack/openstack-galera-0" Oct 05 07:05:26 crc kubenswrapper[4846]: I1005 07:05:26.791592 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/d88c48f9-948e-4888-986e-e00163ec93a6-config-data-generated\") pod \"openstack-galera-0\" (UID: \"d88c48f9-948e-4888-986e-e00163ec93a6\") " pod="openstack/openstack-galera-0" Oct 05 07:05:26 crc kubenswrapper[4846]: I1005 07:05:26.791609 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d88c48f9-948e-4888-986e-e00163ec93a6-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"d88c48f9-948e-4888-986e-e00163ec93a6\") " pod="openstack/openstack-galera-0" Oct 05 07:05:26 crc kubenswrapper[4846]: I1005 07:05:26.791630 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/d88c48f9-948e-4888-986e-e00163ec93a6-secrets\") pod \"openstack-galera-0\" (UID: \"d88c48f9-948e-4888-986e-e00163ec93a6\") " pod="openstack/openstack-galera-0" Oct 05 07:05:26 crc kubenswrapper[4846]: I1005 07:05:26.791646 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pvxnh\" (UniqueName: \"kubernetes.io/projected/d88c48f9-948e-4888-986e-e00163ec93a6-kube-api-access-pvxnh\") pod \"openstack-galera-0\" (UID: \"d88c48f9-948e-4888-986e-e00163ec93a6\") " pod="openstack/openstack-galera-0" Oct 05 07:05:26 crc kubenswrapper[4846]: I1005 07:05:26.894027 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/d88c48f9-948e-4888-986e-e00163ec93a6-config-data-default\") pod \"openstack-galera-0\" (UID: \"d88c48f9-948e-4888-986e-e00163ec93a6\") " pod="openstack/openstack-galera-0" Oct 05 07:05:26 crc kubenswrapper[4846]: I1005 07:05:26.894089 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/d88c48f9-948e-4888-986e-e00163ec93a6-kolla-config\") pod \"openstack-galera-0\" (UID: \"d88c48f9-948e-4888-986e-e00163ec93a6\") " pod="openstack/openstack-galera-0" Oct 05 07:05:26 crc kubenswrapper[4846]: I1005 07:05:26.894121 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/d88c48f9-948e-4888-986e-e00163ec93a6-config-data-generated\") pod \"openstack-galera-0\" (UID: \"d88c48f9-948e-4888-986e-e00163ec93a6\") " pod="openstack/openstack-galera-0" Oct 05 07:05:26 crc kubenswrapper[4846]: I1005 07:05:26.894145 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/d88c48f9-948e-4888-986e-e00163ec93a6-secrets\") pod \"openstack-galera-0\" (UID: \"d88c48f9-948e-4888-986e-e00163ec93a6\") " pod="openstack/openstack-galera-0" Oct 05 07:05:26 crc kubenswrapper[4846]: I1005 07:05:26.894186 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d88c48f9-948e-4888-986e-e00163ec93a6-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"d88c48f9-948e-4888-986e-e00163ec93a6\") " pod="openstack/openstack-galera-0" Oct 05 07:05:26 crc kubenswrapper[4846]: I1005 07:05:26.894208 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pvxnh\" (UniqueName: \"kubernetes.io/projected/d88c48f9-948e-4888-986e-e00163ec93a6-kube-api-access-pvxnh\") pod \"openstack-galera-0\" (UID: \"d88c48f9-948e-4888-986e-e00163ec93a6\") " pod="openstack/openstack-galera-0" Oct 05 07:05:26 crc kubenswrapper[4846]: I1005 07:05:26.894260 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/d88c48f9-948e-4888-986e-e00163ec93a6-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"d88c48f9-948e-4888-986e-e00163ec93a6\") " pod="openstack/openstack-galera-0" Oct 05 07:05:26 crc kubenswrapper[4846]: I1005 07:05:26.894287 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d88c48f9-948e-4888-986e-e00163ec93a6-operator-scripts\") pod \"openstack-galera-0\" (UID: \"d88c48f9-948e-4888-986e-e00163ec93a6\") " pod="openstack/openstack-galera-0" Oct 05 07:05:26 crc kubenswrapper[4846]: I1005 07:05:26.894315 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"openstack-galera-0\" (UID: \"d88c48f9-948e-4888-986e-e00163ec93a6\") " pod="openstack/openstack-galera-0" Oct 05 07:05:26 crc kubenswrapper[4846]: I1005 07:05:26.894766 4846 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"openstack-galera-0\" (UID: \"d88c48f9-948e-4888-986e-e00163ec93a6\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/openstack-galera-0" Oct 05 07:05:26 crc kubenswrapper[4846]: I1005 07:05:26.895010 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/d88c48f9-948e-4888-986e-e00163ec93a6-config-data-default\") pod \"openstack-galera-0\" (UID: \"d88c48f9-948e-4888-986e-e00163ec93a6\") " pod="openstack/openstack-galera-0" Oct 05 07:05:26 crc kubenswrapper[4846]: I1005 07:05:26.895168 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/d88c48f9-948e-4888-986e-e00163ec93a6-kolla-config\") pod \"openstack-galera-0\" (UID: \"d88c48f9-948e-4888-986e-e00163ec93a6\") " pod="openstack/openstack-galera-0" Oct 05 07:05:26 crc kubenswrapper[4846]: I1005 07:05:26.896396 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/d88c48f9-948e-4888-986e-e00163ec93a6-config-data-generated\") pod \"openstack-galera-0\" (UID: \"d88c48f9-948e-4888-986e-e00163ec93a6\") " pod="openstack/openstack-galera-0" Oct 05 07:05:26 crc kubenswrapper[4846]: I1005 07:05:26.897369 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d88c48f9-948e-4888-986e-e00163ec93a6-operator-scripts\") pod \"openstack-galera-0\" (UID: \"d88c48f9-948e-4888-986e-e00163ec93a6\") " pod="openstack/openstack-galera-0" Oct 05 07:05:26 crc kubenswrapper[4846]: I1005 07:05:26.903394 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d88c48f9-948e-4888-986e-e00163ec93a6-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"d88c48f9-948e-4888-986e-e00163ec93a6\") " pod="openstack/openstack-galera-0" Oct 05 07:05:26 crc kubenswrapper[4846]: I1005 07:05:26.903685 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/d88c48f9-948e-4888-986e-e00163ec93a6-secrets\") pod \"openstack-galera-0\" (UID: \"d88c48f9-948e-4888-986e-e00163ec93a6\") " pod="openstack/openstack-galera-0" Oct 05 07:05:26 crc kubenswrapper[4846]: I1005 07:05:26.904941 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/d88c48f9-948e-4888-986e-e00163ec93a6-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"d88c48f9-948e-4888-986e-e00163ec93a6\") " pod="openstack/openstack-galera-0" Oct 05 07:05:26 crc kubenswrapper[4846]: I1005 07:05:26.920781 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pvxnh\" (UniqueName: \"kubernetes.io/projected/d88c48f9-948e-4888-986e-e00163ec93a6-kube-api-access-pvxnh\") pod \"openstack-galera-0\" (UID: \"d88c48f9-948e-4888-986e-e00163ec93a6\") " pod="openstack/openstack-galera-0" Oct 05 07:05:26 crc kubenswrapper[4846]: I1005 07:05:26.943508 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"openstack-galera-0\" (UID: \"d88c48f9-948e-4888-986e-e00163ec93a6\") " pod="openstack/openstack-galera-0" Oct 05 07:05:26 crc kubenswrapper[4846]: I1005 07:05:26.964721 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Oct 05 07:05:27 crc kubenswrapper[4846]: I1005 07:05:27.433826 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"626e6bce-7422-4fbd-b978-dfa5cb93de34","Type":"ContainerStarted","Data":"9a98fff94c37b3930255c3af93b8ce72516be04af351656fe24b90b08afac214"} Oct 05 07:05:27 crc kubenswrapper[4846]: I1005 07:05:27.653815 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Oct 05 07:05:27 crc kubenswrapper[4846]: I1005 07:05:27.701709 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-cell1-galera-0"] Oct 05 07:05:27 crc kubenswrapper[4846]: I1005 07:05:27.712980 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Oct 05 07:05:27 crc kubenswrapper[4846]: I1005 07:05:27.721587 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-config-data" Oct 05 07:05:27 crc kubenswrapper[4846]: I1005 07:05:27.723563 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-scripts" Oct 05 07:05:27 crc kubenswrapper[4846]: I1005 07:05:27.723887 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-cell1-svc" Oct 05 07:05:27 crc kubenswrapper[4846]: I1005 07:05:27.733763 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-cell1-dockercfg-cwsw2" Oct 05 07:05:27 crc kubenswrapper[4846]: I1005 07:05:27.735601 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Oct 05 07:05:27 crc kubenswrapper[4846]: I1005 07:05:27.821125 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/221918b0-eb0e-4a32-8eed-07acd18bd31e-secrets\") pod \"openstack-cell1-galera-0\" (UID: \"221918b0-eb0e-4a32-8eed-07acd18bd31e\") " pod="openstack/openstack-cell1-galera-0" Oct 05 07:05:27 crc kubenswrapper[4846]: I1005 07:05:27.821171 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/221918b0-eb0e-4a32-8eed-07acd18bd31e-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"221918b0-eb0e-4a32-8eed-07acd18bd31e\") " pod="openstack/openstack-cell1-galera-0" Oct 05 07:05:27 crc kubenswrapper[4846]: I1005 07:05:27.821211 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/221918b0-eb0e-4a32-8eed-07acd18bd31e-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"221918b0-eb0e-4a32-8eed-07acd18bd31e\") " pod="openstack/openstack-cell1-galera-0" Oct 05 07:05:27 crc kubenswrapper[4846]: I1005 07:05:27.821231 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/221918b0-eb0e-4a32-8eed-07acd18bd31e-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"221918b0-eb0e-4a32-8eed-07acd18bd31e\") " pod="openstack/openstack-cell1-galera-0" Oct 05 07:05:27 crc kubenswrapper[4846]: I1005 07:05:27.821260 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/221918b0-eb0e-4a32-8eed-07acd18bd31e-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"221918b0-eb0e-4a32-8eed-07acd18bd31e\") " pod="openstack/openstack-cell1-galera-0" Oct 05 07:05:27 crc kubenswrapper[4846]: I1005 07:05:27.821279 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/221918b0-eb0e-4a32-8eed-07acd18bd31e-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"221918b0-eb0e-4a32-8eed-07acd18bd31e\") " pod="openstack/openstack-cell1-galera-0" Oct 05 07:05:27 crc kubenswrapper[4846]: I1005 07:05:27.823474 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"openstack-cell1-galera-0\" (UID: \"221918b0-eb0e-4a32-8eed-07acd18bd31e\") " pod="openstack/openstack-cell1-galera-0" Oct 05 07:05:27 crc kubenswrapper[4846]: I1005 07:05:27.823536 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/221918b0-eb0e-4a32-8eed-07acd18bd31e-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"221918b0-eb0e-4a32-8eed-07acd18bd31e\") " pod="openstack/openstack-cell1-galera-0" Oct 05 07:05:27 crc kubenswrapper[4846]: I1005 07:05:27.823557 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5zl68\" (UniqueName: \"kubernetes.io/projected/221918b0-eb0e-4a32-8eed-07acd18bd31e-kube-api-access-5zl68\") pod \"openstack-cell1-galera-0\" (UID: \"221918b0-eb0e-4a32-8eed-07acd18bd31e\") " pod="openstack/openstack-cell1-galera-0" Oct 05 07:05:27 crc kubenswrapper[4846]: I1005 07:05:27.896476 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/memcached-0"] Oct 05 07:05:27 crc kubenswrapper[4846]: I1005 07:05:27.897634 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Oct 05 07:05:27 crc kubenswrapper[4846]: I1005 07:05:27.905350 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"memcached-config-data" Oct 05 07:05:27 crc kubenswrapper[4846]: I1005 07:05:27.905871 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-memcached-svc" Oct 05 07:05:27 crc kubenswrapper[4846]: I1005 07:05:27.906021 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"memcached-memcached-dockercfg-7rrlj" Oct 05 07:05:27 crc kubenswrapper[4846]: I1005 07:05:27.921236 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Oct 05 07:05:27 crc kubenswrapper[4846]: I1005 07:05:27.925972 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/221918b0-eb0e-4a32-8eed-07acd18bd31e-secrets\") pod \"openstack-cell1-galera-0\" (UID: \"221918b0-eb0e-4a32-8eed-07acd18bd31e\") " pod="openstack/openstack-cell1-galera-0" Oct 05 07:05:27 crc kubenswrapper[4846]: I1005 07:05:27.926014 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/221918b0-eb0e-4a32-8eed-07acd18bd31e-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"221918b0-eb0e-4a32-8eed-07acd18bd31e\") " pod="openstack/openstack-cell1-galera-0" Oct 05 07:05:27 crc kubenswrapper[4846]: I1005 07:05:27.926039 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/221918b0-eb0e-4a32-8eed-07acd18bd31e-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"221918b0-eb0e-4a32-8eed-07acd18bd31e\") " pod="openstack/openstack-cell1-galera-0" Oct 05 07:05:27 crc kubenswrapper[4846]: I1005 07:05:27.926058 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/221918b0-eb0e-4a32-8eed-07acd18bd31e-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"221918b0-eb0e-4a32-8eed-07acd18bd31e\") " pod="openstack/openstack-cell1-galera-0" Oct 05 07:05:27 crc kubenswrapper[4846]: I1005 07:05:27.926091 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/221918b0-eb0e-4a32-8eed-07acd18bd31e-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"221918b0-eb0e-4a32-8eed-07acd18bd31e\") " pod="openstack/openstack-cell1-galera-0" Oct 05 07:05:27 crc kubenswrapper[4846]: I1005 07:05:27.926106 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/221918b0-eb0e-4a32-8eed-07acd18bd31e-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"221918b0-eb0e-4a32-8eed-07acd18bd31e\") " pod="openstack/openstack-cell1-galera-0" Oct 05 07:05:27 crc kubenswrapper[4846]: I1005 07:05:27.928322 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"openstack-cell1-galera-0\" (UID: \"221918b0-eb0e-4a32-8eed-07acd18bd31e\") " pod="openstack/openstack-cell1-galera-0" Oct 05 07:05:27 crc kubenswrapper[4846]: I1005 07:05:27.928359 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/221918b0-eb0e-4a32-8eed-07acd18bd31e-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"221918b0-eb0e-4a32-8eed-07acd18bd31e\") " pod="openstack/openstack-cell1-galera-0" Oct 05 07:05:27 crc kubenswrapper[4846]: I1005 07:05:27.928382 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5zl68\" (UniqueName: \"kubernetes.io/projected/221918b0-eb0e-4a32-8eed-07acd18bd31e-kube-api-access-5zl68\") pod \"openstack-cell1-galera-0\" (UID: \"221918b0-eb0e-4a32-8eed-07acd18bd31e\") " pod="openstack/openstack-cell1-galera-0" Oct 05 07:05:27 crc kubenswrapper[4846]: I1005 07:05:27.931899 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/221918b0-eb0e-4a32-8eed-07acd18bd31e-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"221918b0-eb0e-4a32-8eed-07acd18bd31e\") " pod="openstack/openstack-cell1-galera-0" Oct 05 07:05:27 crc kubenswrapper[4846]: I1005 07:05:27.932733 4846 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"openstack-cell1-galera-0\" (UID: \"221918b0-eb0e-4a32-8eed-07acd18bd31e\") device mount path \"/mnt/openstack/pv05\"" pod="openstack/openstack-cell1-galera-0" Oct 05 07:05:27 crc kubenswrapper[4846]: I1005 07:05:27.934237 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/221918b0-eb0e-4a32-8eed-07acd18bd31e-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"221918b0-eb0e-4a32-8eed-07acd18bd31e\") " pod="openstack/openstack-cell1-galera-0" Oct 05 07:05:27 crc kubenswrapper[4846]: I1005 07:05:27.934503 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/221918b0-eb0e-4a32-8eed-07acd18bd31e-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"221918b0-eb0e-4a32-8eed-07acd18bd31e\") " pod="openstack/openstack-cell1-galera-0" Oct 05 07:05:27 crc kubenswrapper[4846]: I1005 07:05:27.934906 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/221918b0-eb0e-4a32-8eed-07acd18bd31e-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"221918b0-eb0e-4a32-8eed-07acd18bd31e\") " pod="openstack/openstack-cell1-galera-0" Oct 05 07:05:27 crc kubenswrapper[4846]: I1005 07:05:27.943487 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/221918b0-eb0e-4a32-8eed-07acd18bd31e-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"221918b0-eb0e-4a32-8eed-07acd18bd31e\") " pod="openstack/openstack-cell1-galera-0" Oct 05 07:05:27 crc kubenswrapper[4846]: I1005 07:05:27.962344 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/221918b0-eb0e-4a32-8eed-07acd18bd31e-secrets\") pod \"openstack-cell1-galera-0\" (UID: \"221918b0-eb0e-4a32-8eed-07acd18bd31e\") " pod="openstack/openstack-cell1-galera-0" Oct 05 07:05:27 crc kubenswrapper[4846]: I1005 07:05:27.962989 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/221918b0-eb0e-4a32-8eed-07acd18bd31e-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"221918b0-eb0e-4a32-8eed-07acd18bd31e\") " pod="openstack/openstack-cell1-galera-0" Oct 05 07:05:27 crc kubenswrapper[4846]: I1005 07:05:27.972014 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5zl68\" (UniqueName: \"kubernetes.io/projected/221918b0-eb0e-4a32-8eed-07acd18bd31e-kube-api-access-5zl68\") pod \"openstack-cell1-galera-0\" (UID: \"221918b0-eb0e-4a32-8eed-07acd18bd31e\") " pod="openstack/openstack-cell1-galera-0" Oct 05 07:05:28 crc kubenswrapper[4846]: I1005 07:05:28.009497 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"openstack-cell1-galera-0\" (UID: \"221918b0-eb0e-4a32-8eed-07acd18bd31e\") " pod="openstack/openstack-cell1-galera-0" Oct 05 07:05:28 crc kubenswrapper[4846]: I1005 07:05:28.030053 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/751734cb-38c8-46a0-9bac-05bcf24cc589-memcached-tls-certs\") pod \"memcached-0\" (UID: \"751734cb-38c8-46a0-9bac-05bcf24cc589\") " pod="openstack/memcached-0" Oct 05 07:05:28 crc kubenswrapper[4846]: I1005 07:05:28.030170 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/751734cb-38c8-46a0-9bac-05bcf24cc589-config-data\") pod \"memcached-0\" (UID: \"751734cb-38c8-46a0-9bac-05bcf24cc589\") " pod="openstack/memcached-0" Oct 05 07:05:28 crc kubenswrapper[4846]: I1005 07:05:28.030376 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b8cz2\" (UniqueName: \"kubernetes.io/projected/751734cb-38c8-46a0-9bac-05bcf24cc589-kube-api-access-b8cz2\") pod \"memcached-0\" (UID: \"751734cb-38c8-46a0-9bac-05bcf24cc589\") " pod="openstack/memcached-0" Oct 05 07:05:28 crc kubenswrapper[4846]: I1005 07:05:28.030397 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/751734cb-38c8-46a0-9bac-05bcf24cc589-combined-ca-bundle\") pod \"memcached-0\" (UID: \"751734cb-38c8-46a0-9bac-05bcf24cc589\") " pod="openstack/memcached-0" Oct 05 07:05:28 crc kubenswrapper[4846]: I1005 07:05:28.030417 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/751734cb-38c8-46a0-9bac-05bcf24cc589-kolla-config\") pod \"memcached-0\" (UID: \"751734cb-38c8-46a0-9bac-05bcf24cc589\") " pod="openstack/memcached-0" Oct 05 07:05:28 crc kubenswrapper[4846]: I1005 07:05:28.091873 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Oct 05 07:05:28 crc kubenswrapper[4846]: I1005 07:05:28.131787 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/751734cb-38c8-46a0-9bac-05bcf24cc589-memcached-tls-certs\") pod \"memcached-0\" (UID: \"751734cb-38c8-46a0-9bac-05bcf24cc589\") " pod="openstack/memcached-0" Oct 05 07:05:28 crc kubenswrapper[4846]: I1005 07:05:28.131878 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/751734cb-38c8-46a0-9bac-05bcf24cc589-config-data\") pod \"memcached-0\" (UID: \"751734cb-38c8-46a0-9bac-05bcf24cc589\") " pod="openstack/memcached-0" Oct 05 07:05:28 crc kubenswrapper[4846]: I1005 07:05:28.131923 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b8cz2\" (UniqueName: \"kubernetes.io/projected/751734cb-38c8-46a0-9bac-05bcf24cc589-kube-api-access-b8cz2\") pod \"memcached-0\" (UID: \"751734cb-38c8-46a0-9bac-05bcf24cc589\") " pod="openstack/memcached-0" Oct 05 07:05:28 crc kubenswrapper[4846]: I1005 07:05:28.131970 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/751734cb-38c8-46a0-9bac-05bcf24cc589-combined-ca-bundle\") pod \"memcached-0\" (UID: \"751734cb-38c8-46a0-9bac-05bcf24cc589\") " pod="openstack/memcached-0" Oct 05 07:05:28 crc kubenswrapper[4846]: I1005 07:05:28.132000 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/751734cb-38c8-46a0-9bac-05bcf24cc589-kolla-config\") pod \"memcached-0\" (UID: \"751734cb-38c8-46a0-9bac-05bcf24cc589\") " pod="openstack/memcached-0" Oct 05 07:05:28 crc kubenswrapper[4846]: I1005 07:05:28.132930 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/751734cb-38c8-46a0-9bac-05bcf24cc589-kolla-config\") pod \"memcached-0\" (UID: \"751734cb-38c8-46a0-9bac-05bcf24cc589\") " pod="openstack/memcached-0" Oct 05 07:05:28 crc kubenswrapper[4846]: I1005 07:05:28.134697 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/751734cb-38c8-46a0-9bac-05bcf24cc589-config-data\") pod \"memcached-0\" (UID: \"751734cb-38c8-46a0-9bac-05bcf24cc589\") " pod="openstack/memcached-0" Oct 05 07:05:28 crc kubenswrapper[4846]: I1005 07:05:28.144262 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/751734cb-38c8-46a0-9bac-05bcf24cc589-combined-ca-bundle\") pod \"memcached-0\" (UID: \"751734cb-38c8-46a0-9bac-05bcf24cc589\") " pod="openstack/memcached-0" Oct 05 07:05:28 crc kubenswrapper[4846]: I1005 07:05:28.147653 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/751734cb-38c8-46a0-9bac-05bcf24cc589-memcached-tls-certs\") pod \"memcached-0\" (UID: \"751734cb-38c8-46a0-9bac-05bcf24cc589\") " pod="openstack/memcached-0" Oct 05 07:05:28 crc kubenswrapper[4846]: I1005 07:05:28.150527 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b8cz2\" (UniqueName: \"kubernetes.io/projected/751734cb-38c8-46a0-9bac-05bcf24cc589-kube-api-access-b8cz2\") pod \"memcached-0\" (UID: \"751734cb-38c8-46a0-9bac-05bcf24cc589\") " pod="openstack/memcached-0" Oct 05 07:05:28 crc kubenswrapper[4846]: I1005 07:05:28.228691 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Oct 05 07:05:28 crc kubenswrapper[4846]: I1005 07:05:28.457495 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"d88c48f9-948e-4888-986e-e00163ec93a6","Type":"ContainerStarted","Data":"6bc8133968b75fac499facde615345124fd87e14184b7ee6c4a0b3c0f8eb0806"} Oct 05 07:05:28 crc kubenswrapper[4846]: I1005 07:05:28.926979 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Oct 05 07:05:29 crc kubenswrapper[4846]: I1005 07:05:29.945726 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Oct 05 07:05:29 crc kubenswrapper[4846]: I1005 07:05:29.949746 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 05 07:05:29 crc kubenswrapper[4846]: I1005 07:05:29.953036 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-ceilometer-dockercfg-6dsgf" Oct 05 07:05:29 crc kubenswrapper[4846]: I1005 07:05:29.957091 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 05 07:05:30 crc kubenswrapper[4846]: I1005 07:05:30.096603 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7clx2\" (UniqueName: \"kubernetes.io/projected/31c9cd11-358f-4b1e-b661-dfb09f480277-kube-api-access-7clx2\") pod \"kube-state-metrics-0\" (UID: \"31c9cd11-358f-4b1e-b661-dfb09f480277\") " pod="openstack/kube-state-metrics-0" Oct 05 07:05:30 crc kubenswrapper[4846]: I1005 07:05:30.199276 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7clx2\" (UniqueName: \"kubernetes.io/projected/31c9cd11-358f-4b1e-b661-dfb09f480277-kube-api-access-7clx2\") pod \"kube-state-metrics-0\" (UID: \"31c9cd11-358f-4b1e-b661-dfb09f480277\") " pod="openstack/kube-state-metrics-0" Oct 05 07:05:30 crc kubenswrapper[4846]: I1005 07:05:30.240418 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7clx2\" (UniqueName: \"kubernetes.io/projected/31c9cd11-358f-4b1e-b661-dfb09f480277-kube-api-access-7clx2\") pod \"kube-state-metrics-0\" (UID: \"31c9cd11-358f-4b1e-b661-dfb09f480277\") " pod="openstack/kube-state-metrics-0" Oct 05 07:05:30 crc kubenswrapper[4846]: I1005 07:05:30.295965 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 05 07:05:32 crc kubenswrapper[4846]: I1005 07:05:32.535645 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-4kqnw"] Oct 05 07:05:32 crc kubenswrapper[4846]: I1005 07:05:32.538605 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-4kqnw" Oct 05 07:05:32 crc kubenswrapper[4846]: I1005 07:05:32.541492 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovncontroller-ovndbs" Oct 05 07:05:32 crc kubenswrapper[4846]: I1005 07:05:32.541759 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-scripts" Oct 05 07:05:32 crc kubenswrapper[4846]: I1005 07:05:32.541884 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncontroller-ovncontroller-dockercfg-xgkdv" Oct 05 07:05:32 crc kubenswrapper[4846]: I1005 07:05:32.588993 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-4kqnw"] Oct 05 07:05:32 crc kubenswrapper[4846]: I1005 07:05:32.612288 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-ovs-sg4vg"] Oct 05 07:05:32 crc kubenswrapper[4846]: I1005 07:05:32.616984 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-sg4vg" Oct 05 07:05:32 crc kubenswrapper[4846]: I1005 07:05:32.619670 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-sg4vg"] Oct 05 07:05:32 crc kubenswrapper[4846]: I1005 07:05:32.659973 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/169918d8-597e-4ef4-b00c-fb14c4d0347f-var-run\") pod \"ovn-controller-4kqnw\" (UID: \"169918d8-597e-4ef4-b00c-fb14c4d0347f\") " pod="openstack/ovn-controller-4kqnw" Oct 05 07:05:32 crc kubenswrapper[4846]: I1005 07:05:32.660041 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hmtzx\" (UniqueName: \"kubernetes.io/projected/169918d8-597e-4ef4-b00c-fb14c4d0347f-kube-api-access-hmtzx\") pod \"ovn-controller-4kqnw\" (UID: \"169918d8-597e-4ef4-b00c-fb14c4d0347f\") " pod="openstack/ovn-controller-4kqnw" Oct 05 07:05:32 crc kubenswrapper[4846]: I1005 07:05:32.660105 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/169918d8-597e-4ef4-b00c-fb14c4d0347f-scripts\") pod \"ovn-controller-4kqnw\" (UID: \"169918d8-597e-4ef4-b00c-fb14c4d0347f\") " pod="openstack/ovn-controller-4kqnw" Oct 05 07:05:32 crc kubenswrapper[4846]: I1005 07:05:32.660147 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/169918d8-597e-4ef4-b00c-fb14c4d0347f-combined-ca-bundle\") pod \"ovn-controller-4kqnw\" (UID: \"169918d8-597e-4ef4-b00c-fb14c4d0347f\") " pod="openstack/ovn-controller-4kqnw" Oct 05 07:05:32 crc kubenswrapper[4846]: I1005 07:05:32.660267 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/169918d8-597e-4ef4-b00c-fb14c4d0347f-var-run-ovn\") pod \"ovn-controller-4kqnw\" (UID: \"169918d8-597e-4ef4-b00c-fb14c4d0347f\") " pod="openstack/ovn-controller-4kqnw" Oct 05 07:05:32 crc kubenswrapper[4846]: I1005 07:05:32.660287 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/169918d8-597e-4ef4-b00c-fb14c4d0347f-var-log-ovn\") pod \"ovn-controller-4kqnw\" (UID: \"169918d8-597e-4ef4-b00c-fb14c4d0347f\") " pod="openstack/ovn-controller-4kqnw" Oct 05 07:05:32 crc kubenswrapper[4846]: I1005 07:05:32.660306 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/169918d8-597e-4ef4-b00c-fb14c4d0347f-ovn-controller-tls-certs\") pod \"ovn-controller-4kqnw\" (UID: \"169918d8-597e-4ef4-b00c-fb14c4d0347f\") " pod="openstack/ovn-controller-4kqnw" Oct 05 07:05:32 crc kubenswrapper[4846]: I1005 07:05:32.761706 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hmtzx\" (UniqueName: \"kubernetes.io/projected/169918d8-597e-4ef4-b00c-fb14c4d0347f-kube-api-access-hmtzx\") pod \"ovn-controller-4kqnw\" (UID: \"169918d8-597e-4ef4-b00c-fb14c4d0347f\") " pod="openstack/ovn-controller-4kqnw" Oct 05 07:05:32 crc kubenswrapper[4846]: I1005 07:05:32.762128 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/169918d8-597e-4ef4-b00c-fb14c4d0347f-var-run\") pod \"ovn-controller-4kqnw\" (UID: \"169918d8-597e-4ef4-b00c-fb14c4d0347f\") " pod="openstack/ovn-controller-4kqnw" Oct 05 07:05:32 crc kubenswrapper[4846]: I1005 07:05:32.762172 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/29c0f404-fc7d-4909-b6bf-e1826a092a72-etc-ovs\") pod \"ovn-controller-ovs-sg4vg\" (UID: \"29c0f404-fc7d-4909-b6bf-e1826a092a72\") " pod="openstack/ovn-controller-ovs-sg4vg" Oct 05 07:05:32 crc kubenswrapper[4846]: I1005 07:05:32.762227 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/29c0f404-fc7d-4909-b6bf-e1826a092a72-var-lib\") pod \"ovn-controller-ovs-sg4vg\" (UID: \"29c0f404-fc7d-4909-b6bf-e1826a092a72\") " pod="openstack/ovn-controller-ovs-sg4vg" Oct 05 07:05:32 crc kubenswrapper[4846]: I1005 07:05:32.762262 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/169918d8-597e-4ef4-b00c-fb14c4d0347f-scripts\") pod \"ovn-controller-4kqnw\" (UID: \"169918d8-597e-4ef4-b00c-fb14c4d0347f\") " pod="openstack/ovn-controller-4kqnw" Oct 05 07:05:32 crc kubenswrapper[4846]: I1005 07:05:32.762290 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/29c0f404-fc7d-4909-b6bf-e1826a092a72-var-log\") pod \"ovn-controller-ovs-sg4vg\" (UID: \"29c0f404-fc7d-4909-b6bf-e1826a092a72\") " pod="openstack/ovn-controller-ovs-sg4vg" Oct 05 07:05:32 crc kubenswrapper[4846]: I1005 07:05:32.762696 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/169918d8-597e-4ef4-b00c-fb14c4d0347f-combined-ca-bundle\") pod \"ovn-controller-4kqnw\" (UID: \"169918d8-597e-4ef4-b00c-fb14c4d0347f\") " pod="openstack/ovn-controller-4kqnw" Oct 05 07:05:32 crc kubenswrapper[4846]: I1005 07:05:32.762999 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/169918d8-597e-4ef4-b00c-fb14c4d0347f-var-run\") pod \"ovn-controller-4kqnw\" (UID: \"169918d8-597e-4ef4-b00c-fb14c4d0347f\") " pod="openstack/ovn-controller-4kqnw" Oct 05 07:05:32 crc kubenswrapper[4846]: I1005 07:05:32.763130 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x9dgh\" (UniqueName: \"kubernetes.io/projected/29c0f404-fc7d-4909-b6bf-e1826a092a72-kube-api-access-x9dgh\") pod \"ovn-controller-ovs-sg4vg\" (UID: \"29c0f404-fc7d-4909-b6bf-e1826a092a72\") " pod="openstack/ovn-controller-ovs-sg4vg" Oct 05 07:05:32 crc kubenswrapper[4846]: I1005 07:05:32.763209 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/169918d8-597e-4ef4-b00c-fb14c4d0347f-var-run-ovn\") pod \"ovn-controller-4kqnw\" (UID: \"169918d8-597e-4ef4-b00c-fb14c4d0347f\") " pod="openstack/ovn-controller-4kqnw" Oct 05 07:05:32 crc kubenswrapper[4846]: I1005 07:05:32.763235 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/169918d8-597e-4ef4-b00c-fb14c4d0347f-var-log-ovn\") pod \"ovn-controller-4kqnw\" (UID: \"169918d8-597e-4ef4-b00c-fb14c4d0347f\") " pod="openstack/ovn-controller-4kqnw" Oct 05 07:05:32 crc kubenswrapper[4846]: I1005 07:05:32.763258 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/29c0f404-fc7d-4909-b6bf-e1826a092a72-var-run\") pod \"ovn-controller-ovs-sg4vg\" (UID: \"29c0f404-fc7d-4909-b6bf-e1826a092a72\") " pod="openstack/ovn-controller-ovs-sg4vg" Oct 05 07:05:32 crc kubenswrapper[4846]: I1005 07:05:32.763299 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/29c0f404-fc7d-4909-b6bf-e1826a092a72-scripts\") pod \"ovn-controller-ovs-sg4vg\" (UID: \"29c0f404-fc7d-4909-b6bf-e1826a092a72\") " pod="openstack/ovn-controller-ovs-sg4vg" Oct 05 07:05:32 crc kubenswrapper[4846]: I1005 07:05:32.763324 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/169918d8-597e-4ef4-b00c-fb14c4d0347f-ovn-controller-tls-certs\") pod \"ovn-controller-4kqnw\" (UID: \"169918d8-597e-4ef4-b00c-fb14c4d0347f\") " pod="openstack/ovn-controller-4kqnw" Oct 05 07:05:32 crc kubenswrapper[4846]: I1005 07:05:32.763578 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/169918d8-597e-4ef4-b00c-fb14c4d0347f-var-run-ovn\") pod \"ovn-controller-4kqnw\" (UID: \"169918d8-597e-4ef4-b00c-fb14c4d0347f\") " pod="openstack/ovn-controller-4kqnw" Oct 05 07:05:32 crc kubenswrapper[4846]: I1005 07:05:32.763693 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/169918d8-597e-4ef4-b00c-fb14c4d0347f-var-log-ovn\") pod \"ovn-controller-4kqnw\" (UID: \"169918d8-597e-4ef4-b00c-fb14c4d0347f\") " pod="openstack/ovn-controller-4kqnw" Oct 05 07:05:32 crc kubenswrapper[4846]: I1005 07:05:32.764376 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/169918d8-597e-4ef4-b00c-fb14c4d0347f-scripts\") pod \"ovn-controller-4kqnw\" (UID: \"169918d8-597e-4ef4-b00c-fb14c4d0347f\") " pod="openstack/ovn-controller-4kqnw" Oct 05 07:05:32 crc kubenswrapper[4846]: I1005 07:05:32.770072 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/169918d8-597e-4ef4-b00c-fb14c4d0347f-ovn-controller-tls-certs\") pod \"ovn-controller-4kqnw\" (UID: \"169918d8-597e-4ef4-b00c-fb14c4d0347f\") " pod="openstack/ovn-controller-4kqnw" Oct 05 07:05:32 crc kubenswrapper[4846]: I1005 07:05:32.770215 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/169918d8-597e-4ef4-b00c-fb14c4d0347f-combined-ca-bundle\") pod \"ovn-controller-4kqnw\" (UID: \"169918d8-597e-4ef4-b00c-fb14c4d0347f\") " pod="openstack/ovn-controller-4kqnw" Oct 05 07:05:32 crc kubenswrapper[4846]: I1005 07:05:32.786932 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hmtzx\" (UniqueName: \"kubernetes.io/projected/169918d8-597e-4ef4-b00c-fb14c4d0347f-kube-api-access-hmtzx\") pod \"ovn-controller-4kqnw\" (UID: \"169918d8-597e-4ef4-b00c-fb14c4d0347f\") " pod="openstack/ovn-controller-4kqnw" Oct 05 07:05:32 crc kubenswrapper[4846]: I1005 07:05:32.865445 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/29c0f404-fc7d-4909-b6bf-e1826a092a72-var-lib\") pod \"ovn-controller-ovs-sg4vg\" (UID: \"29c0f404-fc7d-4909-b6bf-e1826a092a72\") " pod="openstack/ovn-controller-ovs-sg4vg" Oct 05 07:05:32 crc kubenswrapper[4846]: I1005 07:05:32.865519 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/29c0f404-fc7d-4909-b6bf-e1826a092a72-var-log\") pod \"ovn-controller-ovs-sg4vg\" (UID: \"29c0f404-fc7d-4909-b6bf-e1826a092a72\") " pod="openstack/ovn-controller-ovs-sg4vg" Oct 05 07:05:32 crc kubenswrapper[4846]: I1005 07:05:32.865587 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x9dgh\" (UniqueName: \"kubernetes.io/projected/29c0f404-fc7d-4909-b6bf-e1826a092a72-kube-api-access-x9dgh\") pod \"ovn-controller-ovs-sg4vg\" (UID: \"29c0f404-fc7d-4909-b6bf-e1826a092a72\") " pod="openstack/ovn-controller-ovs-sg4vg" Oct 05 07:05:32 crc kubenswrapper[4846]: I1005 07:05:32.865609 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/29c0f404-fc7d-4909-b6bf-e1826a092a72-var-run\") pod \"ovn-controller-ovs-sg4vg\" (UID: \"29c0f404-fc7d-4909-b6bf-e1826a092a72\") " pod="openstack/ovn-controller-ovs-sg4vg" Oct 05 07:05:32 crc kubenswrapper[4846]: I1005 07:05:32.865647 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/29c0f404-fc7d-4909-b6bf-e1826a092a72-scripts\") pod \"ovn-controller-ovs-sg4vg\" (UID: \"29c0f404-fc7d-4909-b6bf-e1826a092a72\") " pod="openstack/ovn-controller-ovs-sg4vg" Oct 05 07:05:32 crc kubenswrapper[4846]: I1005 07:05:32.865650 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/29c0f404-fc7d-4909-b6bf-e1826a092a72-var-lib\") pod \"ovn-controller-ovs-sg4vg\" (UID: \"29c0f404-fc7d-4909-b6bf-e1826a092a72\") " pod="openstack/ovn-controller-ovs-sg4vg" Oct 05 07:05:32 crc kubenswrapper[4846]: I1005 07:05:32.865696 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/29c0f404-fc7d-4909-b6bf-e1826a092a72-etc-ovs\") pod \"ovn-controller-ovs-sg4vg\" (UID: \"29c0f404-fc7d-4909-b6bf-e1826a092a72\") " pod="openstack/ovn-controller-ovs-sg4vg" Oct 05 07:05:32 crc kubenswrapper[4846]: I1005 07:05:32.865856 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/29c0f404-fc7d-4909-b6bf-e1826a092a72-etc-ovs\") pod \"ovn-controller-ovs-sg4vg\" (UID: \"29c0f404-fc7d-4909-b6bf-e1826a092a72\") " pod="openstack/ovn-controller-ovs-sg4vg" Oct 05 07:05:32 crc kubenswrapper[4846]: I1005 07:05:32.866215 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/29c0f404-fc7d-4909-b6bf-e1826a092a72-var-run\") pod \"ovn-controller-ovs-sg4vg\" (UID: \"29c0f404-fc7d-4909-b6bf-e1826a092a72\") " pod="openstack/ovn-controller-ovs-sg4vg" Oct 05 07:05:32 crc kubenswrapper[4846]: I1005 07:05:32.866388 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/29c0f404-fc7d-4909-b6bf-e1826a092a72-var-log\") pod \"ovn-controller-ovs-sg4vg\" (UID: \"29c0f404-fc7d-4909-b6bf-e1826a092a72\") " pod="openstack/ovn-controller-ovs-sg4vg" Oct 05 07:05:32 crc kubenswrapper[4846]: I1005 07:05:32.870303 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/29c0f404-fc7d-4909-b6bf-e1826a092a72-scripts\") pod \"ovn-controller-ovs-sg4vg\" (UID: \"29c0f404-fc7d-4909-b6bf-e1826a092a72\") " pod="openstack/ovn-controller-ovs-sg4vg" Oct 05 07:05:32 crc kubenswrapper[4846]: I1005 07:05:32.873731 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-4kqnw" Oct 05 07:05:32 crc kubenswrapper[4846]: I1005 07:05:32.882065 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x9dgh\" (UniqueName: \"kubernetes.io/projected/29c0f404-fc7d-4909-b6bf-e1826a092a72-kube-api-access-x9dgh\") pod \"ovn-controller-ovs-sg4vg\" (UID: \"29c0f404-fc7d-4909-b6bf-e1826a092a72\") " pod="openstack/ovn-controller-ovs-sg4vg" Oct 05 07:05:32 crc kubenswrapper[4846]: I1005 07:05:32.941460 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-sg4vg" Oct 05 07:05:33 crc kubenswrapper[4846]: I1005 07:05:33.574289 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"221918b0-eb0e-4a32-8eed-07acd18bd31e","Type":"ContainerStarted","Data":"231cdc556bc3d56e30d9aa386e028a0c6fb06c7db20f716efb83a6c6f9bfc33e"} Oct 05 07:05:35 crc kubenswrapper[4846]: I1005 07:05:35.672877 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-0"] Oct 05 07:05:35 crc kubenswrapper[4846]: I1005 07:05:35.678668 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Oct 05 07:05:35 crc kubenswrapper[4846]: I1005 07:05:35.685815 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-nb-ovndbs" Oct 05 07:05:35 crc kubenswrapper[4846]: I1005 07:05:35.686041 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovn-metrics" Oct 05 07:05:35 crc kubenswrapper[4846]: I1005 07:05:35.688067 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-config" Oct 05 07:05:35 crc kubenswrapper[4846]: I1005 07:05:35.688528 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-nb-dockercfg-mqf5r" Oct 05 07:05:35 crc kubenswrapper[4846]: I1005 07:05:35.688920 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-scripts" Oct 05 07:05:35 crc kubenswrapper[4846]: I1005 07:05:35.700677 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Oct 05 07:05:35 crc kubenswrapper[4846]: I1005 07:05:35.722851 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bdf95824-fa04-4a83-92cc-7e36b4d00289-config\") pod \"ovsdbserver-nb-0\" (UID: \"bdf95824-fa04-4a83-92cc-7e36b4d00289\") " pod="openstack/ovsdbserver-nb-0" Oct 05 07:05:35 crc kubenswrapper[4846]: I1005 07:05:35.722932 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/bdf95824-fa04-4a83-92cc-7e36b4d00289-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"bdf95824-fa04-4a83-92cc-7e36b4d00289\") " pod="openstack/ovsdbserver-nb-0" Oct 05 07:05:35 crc kubenswrapper[4846]: I1005 07:05:35.722960 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b9m6b\" (UniqueName: \"kubernetes.io/projected/bdf95824-fa04-4a83-92cc-7e36b4d00289-kube-api-access-b9m6b\") pod \"ovsdbserver-nb-0\" (UID: \"bdf95824-fa04-4a83-92cc-7e36b4d00289\") " pod="openstack/ovsdbserver-nb-0" Oct 05 07:05:35 crc kubenswrapper[4846]: I1005 07:05:35.722982 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/bdf95824-fa04-4a83-92cc-7e36b4d00289-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"bdf95824-fa04-4a83-92cc-7e36b4d00289\") " pod="openstack/ovsdbserver-nb-0" Oct 05 07:05:35 crc kubenswrapper[4846]: I1005 07:05:35.723009 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/bdf95824-fa04-4a83-92cc-7e36b4d00289-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"bdf95824-fa04-4a83-92cc-7e36b4d00289\") " pod="openstack/ovsdbserver-nb-0" Oct 05 07:05:35 crc kubenswrapper[4846]: I1005 07:05:35.723046 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/bdf95824-fa04-4a83-92cc-7e36b4d00289-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"bdf95824-fa04-4a83-92cc-7e36b4d00289\") " pod="openstack/ovsdbserver-nb-0" Oct 05 07:05:35 crc kubenswrapper[4846]: I1005 07:05:35.723085 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"ovsdbserver-nb-0\" (UID: \"bdf95824-fa04-4a83-92cc-7e36b4d00289\") " pod="openstack/ovsdbserver-nb-0" Oct 05 07:05:35 crc kubenswrapper[4846]: I1005 07:05:35.723116 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bdf95824-fa04-4a83-92cc-7e36b4d00289-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"bdf95824-fa04-4a83-92cc-7e36b4d00289\") " pod="openstack/ovsdbserver-nb-0" Oct 05 07:05:35 crc kubenswrapper[4846]: I1005 07:05:35.829491 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bdf95824-fa04-4a83-92cc-7e36b4d00289-config\") pod \"ovsdbserver-nb-0\" (UID: \"bdf95824-fa04-4a83-92cc-7e36b4d00289\") " pod="openstack/ovsdbserver-nb-0" Oct 05 07:05:35 crc kubenswrapper[4846]: I1005 07:05:35.829583 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/bdf95824-fa04-4a83-92cc-7e36b4d00289-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"bdf95824-fa04-4a83-92cc-7e36b4d00289\") " pod="openstack/ovsdbserver-nb-0" Oct 05 07:05:35 crc kubenswrapper[4846]: I1005 07:05:35.829611 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b9m6b\" (UniqueName: \"kubernetes.io/projected/bdf95824-fa04-4a83-92cc-7e36b4d00289-kube-api-access-b9m6b\") pod \"ovsdbserver-nb-0\" (UID: \"bdf95824-fa04-4a83-92cc-7e36b4d00289\") " pod="openstack/ovsdbserver-nb-0" Oct 05 07:05:35 crc kubenswrapper[4846]: I1005 07:05:35.829637 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/bdf95824-fa04-4a83-92cc-7e36b4d00289-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"bdf95824-fa04-4a83-92cc-7e36b4d00289\") " pod="openstack/ovsdbserver-nb-0" Oct 05 07:05:35 crc kubenswrapper[4846]: I1005 07:05:35.829665 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/bdf95824-fa04-4a83-92cc-7e36b4d00289-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"bdf95824-fa04-4a83-92cc-7e36b4d00289\") " pod="openstack/ovsdbserver-nb-0" Oct 05 07:05:35 crc kubenswrapper[4846]: I1005 07:05:35.829704 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/bdf95824-fa04-4a83-92cc-7e36b4d00289-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"bdf95824-fa04-4a83-92cc-7e36b4d00289\") " pod="openstack/ovsdbserver-nb-0" Oct 05 07:05:35 crc kubenswrapper[4846]: I1005 07:05:35.829741 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"ovsdbserver-nb-0\" (UID: \"bdf95824-fa04-4a83-92cc-7e36b4d00289\") " pod="openstack/ovsdbserver-nb-0" Oct 05 07:05:35 crc kubenswrapper[4846]: I1005 07:05:35.829774 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bdf95824-fa04-4a83-92cc-7e36b4d00289-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"bdf95824-fa04-4a83-92cc-7e36b4d00289\") " pod="openstack/ovsdbserver-nb-0" Oct 05 07:05:35 crc kubenswrapper[4846]: I1005 07:05:35.831737 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/bdf95824-fa04-4a83-92cc-7e36b4d00289-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"bdf95824-fa04-4a83-92cc-7e36b4d00289\") " pod="openstack/ovsdbserver-nb-0" Oct 05 07:05:35 crc kubenswrapper[4846]: I1005 07:05:35.832025 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/bdf95824-fa04-4a83-92cc-7e36b4d00289-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"bdf95824-fa04-4a83-92cc-7e36b4d00289\") " pod="openstack/ovsdbserver-nb-0" Oct 05 07:05:35 crc kubenswrapper[4846]: I1005 07:05:35.832350 4846 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"ovsdbserver-nb-0\" (UID: \"bdf95824-fa04-4a83-92cc-7e36b4d00289\") device mount path \"/mnt/openstack/pv11\"" pod="openstack/ovsdbserver-nb-0" Oct 05 07:05:35 crc kubenswrapper[4846]: I1005 07:05:35.835754 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bdf95824-fa04-4a83-92cc-7e36b4d00289-config\") pod \"ovsdbserver-nb-0\" (UID: \"bdf95824-fa04-4a83-92cc-7e36b4d00289\") " pod="openstack/ovsdbserver-nb-0" Oct 05 07:05:35 crc kubenswrapper[4846]: I1005 07:05:35.837415 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/bdf95824-fa04-4a83-92cc-7e36b4d00289-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"bdf95824-fa04-4a83-92cc-7e36b4d00289\") " pod="openstack/ovsdbserver-nb-0" Oct 05 07:05:35 crc kubenswrapper[4846]: I1005 07:05:35.839092 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bdf95824-fa04-4a83-92cc-7e36b4d00289-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"bdf95824-fa04-4a83-92cc-7e36b4d00289\") " pod="openstack/ovsdbserver-nb-0" Oct 05 07:05:35 crc kubenswrapper[4846]: I1005 07:05:35.843440 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/bdf95824-fa04-4a83-92cc-7e36b4d00289-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"bdf95824-fa04-4a83-92cc-7e36b4d00289\") " pod="openstack/ovsdbserver-nb-0" Oct 05 07:05:35 crc kubenswrapper[4846]: I1005 07:05:35.854202 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b9m6b\" (UniqueName: \"kubernetes.io/projected/bdf95824-fa04-4a83-92cc-7e36b4d00289-kube-api-access-b9m6b\") pod \"ovsdbserver-nb-0\" (UID: \"bdf95824-fa04-4a83-92cc-7e36b4d00289\") " pod="openstack/ovsdbserver-nb-0" Oct 05 07:05:35 crc kubenswrapper[4846]: I1005 07:05:35.864248 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"ovsdbserver-nb-0\" (UID: \"bdf95824-fa04-4a83-92cc-7e36b4d00289\") " pod="openstack/ovsdbserver-nb-0" Oct 05 07:05:36 crc kubenswrapper[4846]: I1005 07:05:36.005435 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Oct 05 07:05:37 crc kubenswrapper[4846]: I1005 07:05:37.082829 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-0"] Oct 05 07:05:37 crc kubenswrapper[4846]: I1005 07:05:37.084451 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Oct 05 07:05:37 crc kubenswrapper[4846]: I1005 07:05:37.094458 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-sb-ovndbs" Oct 05 07:05:37 crc kubenswrapper[4846]: I1005 07:05:37.095565 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-sb-dockercfg-6bdtk" Oct 05 07:05:37 crc kubenswrapper[4846]: I1005 07:05:37.095745 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-config" Oct 05 07:05:37 crc kubenswrapper[4846]: I1005 07:05:37.095919 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-scripts" Oct 05 07:05:37 crc kubenswrapper[4846]: I1005 07:05:37.147391 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Oct 05 07:05:37 crc kubenswrapper[4846]: I1005 07:05:37.154942 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/87e3744a-b8db-4e76-afa9-0e1129410d4c-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"87e3744a-b8db-4e76-afa9-0e1129410d4c\") " pod="openstack/ovsdbserver-sb-0" Oct 05 07:05:37 crc kubenswrapper[4846]: I1005 07:05:37.155006 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/87e3744a-b8db-4e76-afa9-0e1129410d4c-config\") pod \"ovsdbserver-sb-0\" (UID: \"87e3744a-b8db-4e76-afa9-0e1129410d4c\") " pod="openstack/ovsdbserver-sb-0" Oct 05 07:05:37 crc kubenswrapper[4846]: I1005 07:05:37.155038 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qvxqv\" (UniqueName: \"kubernetes.io/projected/87e3744a-b8db-4e76-afa9-0e1129410d4c-kube-api-access-qvxqv\") pod \"ovsdbserver-sb-0\" (UID: \"87e3744a-b8db-4e76-afa9-0e1129410d4c\") " pod="openstack/ovsdbserver-sb-0" Oct 05 07:05:37 crc kubenswrapper[4846]: I1005 07:05:37.155405 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/87e3744a-b8db-4e76-afa9-0e1129410d4c-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"87e3744a-b8db-4e76-afa9-0e1129410d4c\") " pod="openstack/ovsdbserver-sb-0" Oct 05 07:05:37 crc kubenswrapper[4846]: I1005 07:05:37.155580 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/87e3744a-b8db-4e76-afa9-0e1129410d4c-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"87e3744a-b8db-4e76-afa9-0e1129410d4c\") " pod="openstack/ovsdbserver-sb-0" Oct 05 07:05:37 crc kubenswrapper[4846]: I1005 07:05:37.155624 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/87e3744a-b8db-4e76-afa9-0e1129410d4c-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"87e3744a-b8db-4e76-afa9-0e1129410d4c\") " pod="openstack/ovsdbserver-sb-0" Oct 05 07:05:37 crc kubenswrapper[4846]: I1005 07:05:37.155802 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/87e3744a-b8db-4e76-afa9-0e1129410d4c-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"87e3744a-b8db-4e76-afa9-0e1129410d4c\") " pod="openstack/ovsdbserver-sb-0" Oct 05 07:05:37 crc kubenswrapper[4846]: I1005 07:05:37.155841 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"ovsdbserver-sb-0\" (UID: \"87e3744a-b8db-4e76-afa9-0e1129410d4c\") " pod="openstack/ovsdbserver-sb-0" Oct 05 07:05:37 crc kubenswrapper[4846]: I1005 07:05:37.259234 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/87e3744a-b8db-4e76-afa9-0e1129410d4c-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"87e3744a-b8db-4e76-afa9-0e1129410d4c\") " pod="openstack/ovsdbserver-sb-0" Oct 05 07:05:37 crc kubenswrapper[4846]: I1005 07:05:37.259339 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qvxqv\" (UniqueName: \"kubernetes.io/projected/87e3744a-b8db-4e76-afa9-0e1129410d4c-kube-api-access-qvxqv\") pod \"ovsdbserver-sb-0\" (UID: \"87e3744a-b8db-4e76-afa9-0e1129410d4c\") " pod="openstack/ovsdbserver-sb-0" Oct 05 07:05:37 crc kubenswrapper[4846]: I1005 07:05:37.259380 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/87e3744a-b8db-4e76-afa9-0e1129410d4c-config\") pod \"ovsdbserver-sb-0\" (UID: \"87e3744a-b8db-4e76-afa9-0e1129410d4c\") " pod="openstack/ovsdbserver-sb-0" Oct 05 07:05:37 crc kubenswrapper[4846]: I1005 07:05:37.259485 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/87e3744a-b8db-4e76-afa9-0e1129410d4c-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"87e3744a-b8db-4e76-afa9-0e1129410d4c\") " pod="openstack/ovsdbserver-sb-0" Oct 05 07:05:37 crc kubenswrapper[4846]: I1005 07:05:37.259565 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/87e3744a-b8db-4e76-afa9-0e1129410d4c-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"87e3744a-b8db-4e76-afa9-0e1129410d4c\") " pod="openstack/ovsdbserver-sb-0" Oct 05 07:05:37 crc kubenswrapper[4846]: I1005 07:05:37.259600 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/87e3744a-b8db-4e76-afa9-0e1129410d4c-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"87e3744a-b8db-4e76-afa9-0e1129410d4c\") " pod="openstack/ovsdbserver-sb-0" Oct 05 07:05:37 crc kubenswrapper[4846]: I1005 07:05:37.259684 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/87e3744a-b8db-4e76-afa9-0e1129410d4c-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"87e3744a-b8db-4e76-afa9-0e1129410d4c\") " pod="openstack/ovsdbserver-sb-0" Oct 05 07:05:37 crc kubenswrapper[4846]: I1005 07:05:37.259711 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"ovsdbserver-sb-0\" (UID: \"87e3744a-b8db-4e76-afa9-0e1129410d4c\") " pod="openstack/ovsdbserver-sb-0" Oct 05 07:05:37 crc kubenswrapper[4846]: I1005 07:05:37.260065 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/87e3744a-b8db-4e76-afa9-0e1129410d4c-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"87e3744a-b8db-4e76-afa9-0e1129410d4c\") " pod="openstack/ovsdbserver-sb-0" Oct 05 07:05:37 crc kubenswrapper[4846]: I1005 07:05:37.260107 4846 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"ovsdbserver-sb-0\" (UID: \"87e3744a-b8db-4e76-afa9-0e1129410d4c\") device mount path \"/mnt/openstack/pv08\"" pod="openstack/ovsdbserver-sb-0" Oct 05 07:05:37 crc kubenswrapper[4846]: I1005 07:05:37.261618 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/87e3744a-b8db-4e76-afa9-0e1129410d4c-config\") pod \"ovsdbserver-sb-0\" (UID: \"87e3744a-b8db-4e76-afa9-0e1129410d4c\") " pod="openstack/ovsdbserver-sb-0" Oct 05 07:05:37 crc kubenswrapper[4846]: I1005 07:05:37.261691 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/87e3744a-b8db-4e76-afa9-0e1129410d4c-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"87e3744a-b8db-4e76-afa9-0e1129410d4c\") " pod="openstack/ovsdbserver-sb-0" Oct 05 07:05:37 crc kubenswrapper[4846]: I1005 07:05:37.268167 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/87e3744a-b8db-4e76-afa9-0e1129410d4c-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"87e3744a-b8db-4e76-afa9-0e1129410d4c\") " pod="openstack/ovsdbserver-sb-0" Oct 05 07:05:37 crc kubenswrapper[4846]: I1005 07:05:37.271344 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/87e3744a-b8db-4e76-afa9-0e1129410d4c-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"87e3744a-b8db-4e76-afa9-0e1129410d4c\") " pod="openstack/ovsdbserver-sb-0" Oct 05 07:05:37 crc kubenswrapper[4846]: I1005 07:05:37.279669 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qvxqv\" (UniqueName: \"kubernetes.io/projected/87e3744a-b8db-4e76-afa9-0e1129410d4c-kube-api-access-qvxqv\") pod \"ovsdbserver-sb-0\" (UID: \"87e3744a-b8db-4e76-afa9-0e1129410d4c\") " pod="openstack/ovsdbserver-sb-0" Oct 05 07:05:37 crc kubenswrapper[4846]: I1005 07:05:37.280900 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/87e3744a-b8db-4e76-afa9-0e1129410d4c-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"87e3744a-b8db-4e76-afa9-0e1129410d4c\") " pod="openstack/ovsdbserver-sb-0" Oct 05 07:05:37 crc kubenswrapper[4846]: I1005 07:05:37.291999 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"ovsdbserver-sb-0\" (UID: \"87e3744a-b8db-4e76-afa9-0e1129410d4c\") " pod="openstack/ovsdbserver-sb-0" Oct 05 07:05:37 crc kubenswrapper[4846]: I1005 07:05:37.449500 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Oct 05 07:05:42 crc kubenswrapper[4846]: E1005 07:05:42.511749 4846 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-rabbitmq@sha256:d142e6e9123b2a4f16d1d65c5d11a132ae9755c2c8cf429ca7ef8c9cd00f4f42" Oct 05 07:05:42 crc kubenswrapper[4846]: E1005 07:05:42.513119 4846 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:setup-container,Image:quay.io/podified-antelope-centos9/openstack-rabbitmq@sha256:d142e6e9123b2a4f16d1d65c5d11a132ae9755c2c8cf429ca7ef8c9cd00f4f42,Command:[sh -c cp /tmp/erlang-cookie-secret/.erlang.cookie /var/lib/rabbitmq/.erlang.cookie && chmod 600 /var/lib/rabbitmq/.erlang.cookie ; cp /tmp/rabbitmq-plugins/enabled_plugins /operator/enabled_plugins ; echo '[default]' > /var/lib/rabbitmq/.rabbitmqadmin.conf && sed -e 's/default_user/username/' -e 's/default_pass/password/' /tmp/default_user.conf >> /var/lib/rabbitmq/.rabbitmqadmin.conf && chmod 600 /var/lib/rabbitmq/.rabbitmqadmin.conf ; sleep 30],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Requests:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:plugins-conf,ReadOnly:false,MountPath:/tmp/rabbitmq-plugins/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-erlang-cookie,ReadOnly:false,MountPath:/var/lib/rabbitmq/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:erlang-cookie-secret,ReadOnly:false,MountPath:/tmp/erlang-cookie-secret/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-plugins,ReadOnly:false,MountPath:/operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:persistence,ReadOnly:false,MountPath:/var/lib/rabbitmq/mnesia/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-confd,ReadOnly:false,MountPath:/tmp/default_user.conf,SubPath:default_user.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-j5pb6,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-server-0_openstack(35261dbc-cabe-4bf0-88f6-b8613ddec0f1): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 05 07:05:42 crc kubenswrapper[4846]: E1005 07:05:42.514423 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/rabbitmq-server-0" podUID="35261dbc-cabe-4bf0-88f6-b8613ddec0f1" Oct 05 07:05:43 crc kubenswrapper[4846]: E1005 07:05:43.297543 4846 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:053c95cc75e5bc6de83a08f3196125bb5fbbfea1795643daf3f1378cbaad5d26" Oct 05 07:05:43 crc kubenswrapper[4846]: E1005 07:05:43.297831 4846 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:053c95cc75e5bc6de83a08f3196125bb5fbbfea1795643daf3f1378cbaad5d26,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n659h4h664hbh658h587h67ch89h587h8fh679hc6hf9h55fh644h5d5h698h68dh5cdh5ffh669h54ch9h689hb8hd4h5bfhd8h5d7h5fh665h574q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-7p4vq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-5d4d9f7875-29gcr_openstack(5fa7b35e-d3fd-45b6-bb8c-b7599419d53f): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 05 07:05:43 crc kubenswrapper[4846]: E1005 07:05:43.299129 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-5d4d9f7875-29gcr" podUID="5fa7b35e-d3fd-45b6-bb8c-b7599419d53f" Oct 05 07:05:43 crc kubenswrapper[4846]: E1005 07:05:43.680930 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:053c95cc75e5bc6de83a08f3196125bb5fbbfea1795643daf3f1378cbaad5d26\\\"\"" pod="openstack/dnsmasq-dns-5d4d9f7875-29gcr" podUID="5fa7b35e-d3fd-45b6-bb8c-b7599419d53f" Oct 05 07:05:46 crc kubenswrapper[4846]: I1005 07:05:46.030618 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Oct 05 07:05:46 crc kubenswrapper[4846]: I1005 07:05:46.064658 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 05 07:05:46 crc kubenswrapper[4846]: W1005 07:05:46.068017 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod31c9cd11_358f_4b1e_b661_dfb09f480277.slice/crio-ba731ef3d0b7d56aa452c33dbc344f20625eab2da03b9960c336bb4e1c31388a WatchSource:0}: Error finding container ba731ef3d0b7d56aa452c33dbc344f20625eab2da03b9960c336bb4e1c31388a: Status 404 returned error can't find the container with id ba731ef3d0b7d56aa452c33dbc344f20625eab2da03b9960c336bb4e1c31388a Oct 05 07:05:46 crc kubenswrapper[4846]: I1005 07:05:46.103261 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Oct 05 07:05:46 crc kubenswrapper[4846]: I1005 07:05:46.121519 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-4kqnw"] Oct 05 07:05:46 crc kubenswrapper[4846]: I1005 07:05:46.182019 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Oct 05 07:05:46 crc kubenswrapper[4846]: W1005 07:05:46.299651 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod751734cb_38c8_46a0_9bac_05bcf24cc589.slice/crio-cc6b8981a9438d5d7acae93755280feb389f6382aef3df813588262d04e57ddb WatchSource:0}: Error finding container cc6b8981a9438d5d7acae93755280feb389f6382aef3df813588262d04e57ddb: Status 404 returned error can't find the container with id cc6b8981a9438d5d7acae93755280feb389f6382aef3df813588262d04e57ddb Oct 05 07:05:46 crc kubenswrapper[4846]: E1005 07:05:46.368283 4846 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:053c95cc75e5bc6de83a08f3196125bb5fbbfea1795643daf3f1378cbaad5d26" Oct 05 07:05:46 crc kubenswrapper[4846]: E1005 07:05:46.368705 4846 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:053c95cc75e5bc6de83a08f3196125bb5fbbfea1795643daf3f1378cbaad5d26,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:ndfhb5h667h568h584h5f9h58dh565h664h587h597h577h64bh5c4h66fh647hbdh68ch5c5h68dh686h5f7h64hd7hc6h55fh57bh98h57fh87h5fh57fq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-v8xpv,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-7f9579fb85-mf94n_openstack(6f3484f0-58c1-4b8b-bb79-04bfdbf3b5d9): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 05 07:05:46 crc kubenswrapper[4846]: E1005 07:05:46.370737 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-7f9579fb85-mf94n" podUID="6f3484f0-58c1-4b8b-bb79-04bfdbf3b5d9" Oct 05 07:05:46 crc kubenswrapper[4846]: E1005 07:05:46.382584 4846 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:053c95cc75e5bc6de83a08f3196125bb5fbbfea1795643daf3f1378cbaad5d26" Oct 05 07:05:46 crc kubenswrapper[4846]: E1005 07:05:46.382778 4846 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:053c95cc75e5bc6de83a08f3196125bb5fbbfea1795643daf3f1378cbaad5d26,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nffh5bdhf4h5f8h79h55h77h58fh56dh7bh6fh578hbch55dh68h56bhd9h65dh57ch658hc9h566h666h688h58h65dh684h5d7h6ch575h5d6h88q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-29g7r,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-546d69f86c-sp94h_openstack(b04f83fc-2da1-49f8-a485-a2a437ff8914): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 05 07:05:46 crc kubenswrapper[4846]: E1005 07:05:46.383912 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-546d69f86c-sp94h" podUID="b04f83fc-2da1-49f8-a485-a2a437ff8914" Oct 05 07:05:46 crc kubenswrapper[4846]: W1005 07:05:46.401304 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod169918d8_597e_4ef4_b00c_fb14c4d0347f.slice/crio-404742809b09f98d85eb545d8420b8a5096b5cbf77f103b9390023f16993c807 WatchSource:0}: Error finding container 404742809b09f98d85eb545d8420b8a5096b5cbf77f103b9390023f16993c807: Status 404 returned error can't find the container with id 404742809b09f98d85eb545d8420b8a5096b5cbf77f103b9390023f16993c807 Oct 05 07:05:46 crc kubenswrapper[4846]: E1005 07:05:46.473802 4846 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:053c95cc75e5bc6de83a08f3196125bb5fbbfea1795643daf3f1378cbaad5d26" Oct 05 07:05:46 crc kubenswrapper[4846]: E1005 07:05:46.473991 4846 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:053c95cc75e5bc6de83a08f3196125bb5fbbfea1795643daf3f1378cbaad5d26,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n68chd6h679hbfh55fhc6h5ffh5d8h94h56ch589hb4hc5h57bh677hcdh655h8dh667h675h654h66ch567h8fh659h5b4h675h566h55bh54h67dh6dq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-s84t5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-6c64dcbb65-fw9xh_openstack(70abcfd4-3bfd-47c0-bb6a-03883ab21a76): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 05 07:05:46 crc kubenswrapper[4846]: E1005 07:05:46.475398 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-6c64dcbb65-fw9xh" podUID="70abcfd4-3bfd-47c0-bb6a-03883ab21a76" Oct 05 07:05:46 crc kubenswrapper[4846]: I1005 07:05:46.587976 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-sg4vg"] Oct 05 07:05:46 crc kubenswrapper[4846]: I1005 07:05:46.709109 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"87e3744a-b8db-4e76-afa9-0e1129410d4c","Type":"ContainerStarted","Data":"d45d3167f19b6fa1537d87164850bb4f2bddcfe3fcc965f46e24c0042c091067"} Oct 05 07:05:46 crc kubenswrapper[4846]: I1005 07:05:46.710567 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-4kqnw" event={"ID":"169918d8-597e-4ef4-b00c-fb14c4d0347f","Type":"ContainerStarted","Data":"404742809b09f98d85eb545d8420b8a5096b5cbf77f103b9390023f16993c807"} Oct 05 07:05:46 crc kubenswrapper[4846]: I1005 07:05:46.712074 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"31c9cd11-358f-4b1e-b661-dfb09f480277","Type":"ContainerStarted","Data":"ba731ef3d0b7d56aa452c33dbc344f20625eab2da03b9960c336bb4e1c31388a"} Oct 05 07:05:46 crc kubenswrapper[4846]: I1005 07:05:46.713747 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"bdf95824-fa04-4a83-92cc-7e36b4d00289","Type":"ContainerStarted","Data":"241c79da4bec1edce442e764915bccd26a2210f02bcf72b47d133047c73ea3e6"} Oct 05 07:05:46 crc kubenswrapper[4846]: I1005 07:05:46.715205 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"751734cb-38c8-46a0-9bac-05bcf24cc589","Type":"ContainerStarted","Data":"cc6b8981a9438d5d7acae93755280feb389f6382aef3df813588262d04e57ddb"} Oct 05 07:05:46 crc kubenswrapper[4846]: I1005 07:05:46.716438 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-sg4vg" event={"ID":"29c0f404-fc7d-4909-b6bf-e1826a092a72","Type":"ContainerStarted","Data":"3a7e569f8b29ff1301494b1573cbe42db801dc5571d083cb02917ec6fba8c253"} Oct 05 07:05:46 crc kubenswrapper[4846]: E1005 07:05:46.720384 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:053c95cc75e5bc6de83a08f3196125bb5fbbfea1795643daf3f1378cbaad5d26\\\"\"" pod="openstack/dnsmasq-dns-6c64dcbb65-fw9xh" podUID="70abcfd4-3bfd-47c0-bb6a-03883ab21a76" Oct 05 07:05:47 crc kubenswrapper[4846]: I1005 07:05:47.680041 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-546d69f86c-sp94h" Oct 05 07:05:47 crc kubenswrapper[4846]: I1005 07:05:47.692217 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7f9579fb85-mf94n" Oct 05 07:05:47 crc kubenswrapper[4846]: I1005 07:05:47.728647 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6f3484f0-58c1-4b8b-bb79-04bfdbf3b5d9-config\") pod \"6f3484f0-58c1-4b8b-bb79-04bfdbf3b5d9\" (UID: \"6f3484f0-58c1-4b8b-bb79-04bfdbf3b5d9\") " Oct 05 07:05:47 crc kubenswrapper[4846]: I1005 07:05:47.729140 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v8xpv\" (UniqueName: \"kubernetes.io/projected/6f3484f0-58c1-4b8b-bb79-04bfdbf3b5d9-kube-api-access-v8xpv\") pod \"6f3484f0-58c1-4b8b-bb79-04bfdbf3b5d9\" (UID: \"6f3484f0-58c1-4b8b-bb79-04bfdbf3b5d9\") " Oct 05 07:05:47 crc kubenswrapper[4846]: I1005 07:05:47.729385 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6f3484f0-58c1-4b8b-bb79-04bfdbf3b5d9-dns-svc\") pod \"6f3484f0-58c1-4b8b-bb79-04bfdbf3b5d9\" (UID: \"6f3484f0-58c1-4b8b-bb79-04bfdbf3b5d9\") " Oct 05 07:05:47 crc kubenswrapper[4846]: I1005 07:05:47.729424 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b04f83fc-2da1-49f8-a485-a2a437ff8914-config\") pod \"b04f83fc-2da1-49f8-a485-a2a437ff8914\" (UID: \"b04f83fc-2da1-49f8-a485-a2a437ff8914\") " Oct 05 07:05:47 crc kubenswrapper[4846]: I1005 07:05:47.729461 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-29g7r\" (UniqueName: \"kubernetes.io/projected/b04f83fc-2da1-49f8-a485-a2a437ff8914-kube-api-access-29g7r\") pod \"b04f83fc-2da1-49f8-a485-a2a437ff8914\" (UID: \"b04f83fc-2da1-49f8-a485-a2a437ff8914\") " Oct 05 07:05:47 crc kubenswrapper[4846]: I1005 07:05:47.729915 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6f3484f0-58c1-4b8b-bb79-04bfdbf3b5d9-config" (OuterVolumeSpecName: "config") pod "6f3484f0-58c1-4b8b-bb79-04bfdbf3b5d9" (UID: "6f3484f0-58c1-4b8b-bb79-04bfdbf3b5d9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 07:05:47 crc kubenswrapper[4846]: I1005 07:05:47.730317 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6f3484f0-58c1-4b8b-bb79-04bfdbf3b5d9-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "6f3484f0-58c1-4b8b-bb79-04bfdbf3b5d9" (UID: "6f3484f0-58c1-4b8b-bb79-04bfdbf3b5d9"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 07:05:47 crc kubenswrapper[4846]: I1005 07:05:47.731242 4846 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6f3484f0-58c1-4b8b-bb79-04bfdbf3b5d9-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 05 07:05:47 crc kubenswrapper[4846]: I1005 07:05:47.733686 4846 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6f3484f0-58c1-4b8b-bb79-04bfdbf3b5d9-config\") on node \"crc\" DevicePath \"\"" Oct 05 07:05:47 crc kubenswrapper[4846]: I1005 07:05:47.731287 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b04f83fc-2da1-49f8-a485-a2a437ff8914-config" (OuterVolumeSpecName: "config") pod "b04f83fc-2da1-49f8-a485-a2a437ff8914" (UID: "b04f83fc-2da1-49f8-a485-a2a437ff8914"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 07:05:47 crc kubenswrapper[4846]: I1005 07:05:47.736772 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6f3484f0-58c1-4b8b-bb79-04bfdbf3b5d9-kube-api-access-v8xpv" (OuterVolumeSpecName: "kube-api-access-v8xpv") pod "6f3484f0-58c1-4b8b-bb79-04bfdbf3b5d9" (UID: "6f3484f0-58c1-4b8b-bb79-04bfdbf3b5d9"). InnerVolumeSpecName "kube-api-access-v8xpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:05:47 crc kubenswrapper[4846]: I1005 07:05:47.738324 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b04f83fc-2da1-49f8-a485-a2a437ff8914-kube-api-access-29g7r" (OuterVolumeSpecName: "kube-api-access-29g7r") pod "b04f83fc-2da1-49f8-a485-a2a437ff8914" (UID: "b04f83fc-2da1-49f8-a485-a2a437ff8914"). InnerVolumeSpecName "kube-api-access-29g7r". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:05:47 crc kubenswrapper[4846]: I1005 07:05:47.740166 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"35261dbc-cabe-4bf0-88f6-b8613ddec0f1","Type":"ContainerStarted","Data":"9919ee8b928899c8295fa6d8951cffdd5cffd78c7de5c1e7b1a230ab2550fbf6"} Oct 05 07:05:47 crc kubenswrapper[4846]: I1005 07:05:47.743967 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"221918b0-eb0e-4a32-8eed-07acd18bd31e","Type":"ContainerStarted","Data":"48d194c3f1852cef7897b4f3cbb72f3d1aeca5087732bda856574a4d0614661b"} Oct 05 07:05:47 crc kubenswrapper[4846]: I1005 07:05:47.747378 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-546d69f86c-sp94h" Oct 05 07:05:47 crc kubenswrapper[4846]: I1005 07:05:47.747406 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-546d69f86c-sp94h" event={"ID":"b04f83fc-2da1-49f8-a485-a2a437ff8914","Type":"ContainerDied","Data":"ce8c0bc51284cbbf6261fbe3940924f87c2bc8d10a86a9e35bf19ae08209d0c1"} Oct 05 07:05:47 crc kubenswrapper[4846]: I1005 07:05:47.749030 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7f9579fb85-mf94n" event={"ID":"6f3484f0-58c1-4b8b-bb79-04bfdbf3b5d9","Type":"ContainerDied","Data":"dc41e06be71625cf73c16c4d0d36a79d4562f6dba94e1361e9364894c90013c5"} Oct 05 07:05:47 crc kubenswrapper[4846]: I1005 07:05:47.749070 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7f9579fb85-mf94n" Oct 05 07:05:47 crc kubenswrapper[4846]: I1005 07:05:47.751116 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"626e6bce-7422-4fbd-b978-dfa5cb93de34","Type":"ContainerStarted","Data":"516d1ce74e479dfd0f7702334c1c73ac78a55e7ab16e082869f8bbbabc5a6c16"} Oct 05 07:05:47 crc kubenswrapper[4846]: I1005 07:05:47.835416 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v8xpv\" (UniqueName: \"kubernetes.io/projected/6f3484f0-58c1-4b8b-bb79-04bfdbf3b5d9-kube-api-access-v8xpv\") on node \"crc\" DevicePath \"\"" Oct 05 07:05:47 crc kubenswrapper[4846]: I1005 07:05:47.835450 4846 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b04f83fc-2da1-49f8-a485-a2a437ff8914-config\") on node \"crc\" DevicePath \"\"" Oct 05 07:05:47 crc kubenswrapper[4846]: I1005 07:05:47.835460 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-29g7r\" (UniqueName: \"kubernetes.io/projected/b04f83fc-2da1-49f8-a485-a2a437ff8914-kube-api-access-29g7r\") on node \"crc\" DevicePath \"\"" Oct 05 07:05:47 crc kubenswrapper[4846]: I1005 07:05:47.872391 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-546d69f86c-sp94h"] Oct 05 07:05:47 crc kubenswrapper[4846]: I1005 07:05:47.880872 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-546d69f86c-sp94h"] Oct 05 07:05:47 crc kubenswrapper[4846]: I1005 07:05:47.919061 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7f9579fb85-mf94n"] Oct 05 07:05:47 crc kubenswrapper[4846]: I1005 07:05:47.931036 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7f9579fb85-mf94n"] Oct 05 07:05:48 crc kubenswrapper[4846]: I1005 07:05:48.515316 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6f3484f0-58c1-4b8b-bb79-04bfdbf3b5d9" path="/var/lib/kubelet/pods/6f3484f0-58c1-4b8b-bb79-04bfdbf3b5d9/volumes" Oct 05 07:05:48 crc kubenswrapper[4846]: I1005 07:05:48.515855 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b04f83fc-2da1-49f8-a485-a2a437ff8914" path="/var/lib/kubelet/pods/b04f83fc-2da1-49f8-a485-a2a437ff8914/volumes" Oct 05 07:05:48 crc kubenswrapper[4846]: I1005 07:05:48.766696 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"d88c48f9-948e-4888-986e-e00163ec93a6","Type":"ContainerStarted","Data":"bb847394eafe6318ca0e44440040ff96d7a8d7be64f605561ce5e3a0d06b6195"} Oct 05 07:05:50 crc kubenswrapper[4846]: I1005 07:05:50.785376 4846 generic.go:334] "Generic (PLEG): container finished" podID="221918b0-eb0e-4a32-8eed-07acd18bd31e" containerID="48d194c3f1852cef7897b4f3cbb72f3d1aeca5087732bda856574a4d0614661b" exitCode=0 Oct 05 07:05:50 crc kubenswrapper[4846]: I1005 07:05:50.785456 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"221918b0-eb0e-4a32-8eed-07acd18bd31e","Type":"ContainerDied","Data":"48d194c3f1852cef7897b4f3cbb72f3d1aeca5087732bda856574a4d0614661b"} Oct 05 07:05:51 crc kubenswrapper[4846]: I1005 07:05:51.802575 4846 generic.go:334] "Generic (PLEG): container finished" podID="d88c48f9-948e-4888-986e-e00163ec93a6" containerID="bb847394eafe6318ca0e44440040ff96d7a8d7be64f605561ce5e3a0d06b6195" exitCode=0 Oct 05 07:05:51 crc kubenswrapper[4846]: I1005 07:05:51.802684 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"d88c48f9-948e-4888-986e-e00163ec93a6","Type":"ContainerDied","Data":"bb847394eafe6318ca0e44440040ff96d7a8d7be64f605561ce5e3a0d06b6195"} Oct 05 07:05:52 crc kubenswrapper[4846]: I1005 07:05:52.813232 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"751734cb-38c8-46a0-9bac-05bcf24cc589","Type":"ContainerStarted","Data":"ed2b341c0cbc39aa9be3b1731eb5e7d43c786f18e7d5a05613b33a6282bdffcd"} Oct 05 07:05:52 crc kubenswrapper[4846]: I1005 07:05:52.813935 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/memcached-0" Oct 05 07:05:52 crc kubenswrapper[4846]: I1005 07:05:52.815298 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-sg4vg" event={"ID":"29c0f404-fc7d-4909-b6bf-e1826a092a72","Type":"ContainerStarted","Data":"0bb4176d1021478fccab08b6b67c5b5b6ecac6ad30df0ab98cc66c771c4ba7a4"} Oct 05 07:05:52 crc kubenswrapper[4846]: I1005 07:05:52.816859 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"87e3744a-b8db-4e76-afa9-0e1129410d4c","Type":"ContainerStarted","Data":"097a50245fd880a8216f94f3a81c3b749e8486580e37064c166b13f4fa1fc83c"} Oct 05 07:05:52 crc kubenswrapper[4846]: I1005 07:05:52.818720 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-4kqnw" event={"ID":"169918d8-597e-4ef4-b00c-fb14c4d0347f","Type":"ContainerStarted","Data":"d171365315f951cc3fe3c42e399c18ed39c3027267575259a8e68fb2d38283e1"} Oct 05 07:05:52 crc kubenswrapper[4846]: I1005 07:05:52.818902 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-4kqnw" Oct 05 07:05:52 crc kubenswrapper[4846]: I1005 07:05:52.822933 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"221918b0-eb0e-4a32-8eed-07acd18bd31e","Type":"ContainerStarted","Data":"ec314d49982bc70fa3bb12b1a024bc708bb21c05eda9001d00d334940b04d2e6"} Oct 05 07:05:52 crc kubenswrapper[4846]: I1005 07:05:52.825433 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"31c9cd11-358f-4b1e-b661-dfb09f480277","Type":"ContainerStarted","Data":"72b415d01c8c5a8bc6919241cfdcfd6680afd1a0a3a7d66fc08eba56f1c5828d"} Oct 05 07:05:52 crc kubenswrapper[4846]: I1005 07:05:52.825969 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Oct 05 07:05:52 crc kubenswrapper[4846]: I1005 07:05:52.827337 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"bdf95824-fa04-4a83-92cc-7e36b4d00289","Type":"ContainerStarted","Data":"b585eb7d447d9af1ab5bd601fbb7ea5ad51797c817722421db131f2fba3c2373"} Oct 05 07:05:52 crc kubenswrapper[4846]: I1005 07:05:52.829786 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"d88c48f9-948e-4888-986e-e00163ec93a6","Type":"ContainerStarted","Data":"56b8f764086c61683708e28e3f2929710527aada412eda06ddb4b14e487478c3"} Oct 05 07:05:52 crc kubenswrapper[4846]: I1005 07:05:52.868521 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/memcached-0" podStartSLOduration=20.554811798 podStartE2EDuration="25.868488895s" podCreationTimestamp="2025-10-05 07:05:27 +0000 UTC" firstStartedPulling="2025-10-05 07:05:46.302346038 +0000 UTC m=+1068.543198813" lastFinishedPulling="2025-10-05 07:05:51.616023125 +0000 UTC m=+1073.856875910" observedRunningTime="2025-10-05 07:05:52.83527326 +0000 UTC m=+1075.076126075" watchObservedRunningTime="2025-10-05 07:05:52.868488895 +0000 UTC m=+1075.109341670" Oct 05 07:05:52 crc kubenswrapper[4846]: I1005 07:05:52.868686 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-4kqnw" podStartSLOduration=15.156766064 podStartE2EDuration="20.868680131s" podCreationTimestamp="2025-10-05 07:05:32 +0000 UTC" firstStartedPulling="2025-10-05 07:05:46.40365027 +0000 UTC m=+1068.644503045" lastFinishedPulling="2025-10-05 07:05:52.115564337 +0000 UTC m=+1074.356417112" observedRunningTime="2025-10-05 07:05:52.86867134 +0000 UTC m=+1075.109524125" watchObservedRunningTime="2025-10-05 07:05:52.868680131 +0000 UTC m=+1075.109532906" Oct 05 07:05:52 crc kubenswrapper[4846]: I1005 07:05:52.899812 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-cell1-galera-0" podStartSLOduration=13.66192366 podStartE2EDuration="26.89978643s" podCreationTimestamp="2025-10-05 07:05:26 +0000 UTC" firstStartedPulling="2025-10-05 07:05:33.324768069 +0000 UTC m=+1055.565620854" lastFinishedPulling="2025-10-05 07:05:46.562630859 +0000 UTC m=+1068.803483624" observedRunningTime="2025-10-05 07:05:52.893267076 +0000 UTC m=+1075.134119861" watchObservedRunningTime="2025-10-05 07:05:52.89978643 +0000 UTC m=+1075.140639225" Oct 05 07:05:52 crc kubenswrapper[4846]: I1005 07:05:52.924577 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-galera-0" podStartSLOduration=9.077555023 podStartE2EDuration="27.92454313s" podCreationTimestamp="2025-10-05 07:05:25 +0000 UTC" firstStartedPulling="2025-10-05 07:05:27.690266466 +0000 UTC m=+1049.931119241" lastFinishedPulling="2025-10-05 07:05:46.537254573 +0000 UTC m=+1068.778107348" observedRunningTime="2025-10-05 07:05:52.914975585 +0000 UTC m=+1075.155828360" watchObservedRunningTime="2025-10-05 07:05:52.92454313 +0000 UTC m=+1075.165395915" Oct 05 07:05:52 crc kubenswrapper[4846]: I1005 07:05:52.944345 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=17.704017219 podStartE2EDuration="23.944324578s" podCreationTimestamp="2025-10-05 07:05:29 +0000 UTC" firstStartedPulling="2025-10-05 07:05:46.071317927 +0000 UTC m=+1068.312170702" lastFinishedPulling="2025-10-05 07:05:52.311625246 +0000 UTC m=+1074.552478061" observedRunningTime="2025-10-05 07:05:52.940058984 +0000 UTC m=+1075.180911769" watchObservedRunningTime="2025-10-05 07:05:52.944324578 +0000 UTC m=+1075.185177353" Oct 05 07:05:53 crc kubenswrapper[4846]: I1005 07:05:53.847415 4846 generic.go:334] "Generic (PLEG): container finished" podID="29c0f404-fc7d-4909-b6bf-e1826a092a72" containerID="0bb4176d1021478fccab08b6b67c5b5b6ecac6ad30df0ab98cc66c771c4ba7a4" exitCode=0 Oct 05 07:05:53 crc kubenswrapper[4846]: I1005 07:05:53.847479 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-sg4vg" event={"ID":"29c0f404-fc7d-4909-b6bf-e1826a092a72","Type":"ContainerDied","Data":"0bb4176d1021478fccab08b6b67c5b5b6ecac6ad30df0ab98cc66c771c4ba7a4"} Oct 05 07:05:54 crc kubenswrapper[4846]: I1005 07:05:54.866473 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-sg4vg" event={"ID":"29c0f404-fc7d-4909-b6bf-e1826a092a72","Type":"ContainerStarted","Data":"22929dba446bc2516169306e09540fa52839ffc6b8d024f494720cc39ad0ab7d"} Oct 05 07:05:54 crc kubenswrapper[4846]: I1005 07:05:54.866939 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-sg4vg" Oct 05 07:05:54 crc kubenswrapper[4846]: I1005 07:05:54.866957 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-sg4vg" Oct 05 07:05:54 crc kubenswrapper[4846]: I1005 07:05:54.866970 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-sg4vg" event={"ID":"29c0f404-fc7d-4909-b6bf-e1826a092a72","Type":"ContainerStarted","Data":"f4fa8732ebd0803a3d57af2df182505f645e900471768d31fb1d43cb0a83e0cf"} Oct 05 07:05:54 crc kubenswrapper[4846]: I1005 07:05:54.889820 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-ovs-sg4vg" podStartSLOduration=17.378954426 podStartE2EDuration="22.889799611s" podCreationTimestamp="2025-10-05 07:05:32 +0000 UTC" firstStartedPulling="2025-10-05 07:05:46.604748013 +0000 UTC m=+1068.845600788" lastFinishedPulling="2025-10-05 07:05:52.115593198 +0000 UTC m=+1074.356445973" observedRunningTime="2025-10-05 07:05:54.889326068 +0000 UTC m=+1077.130178843" watchObservedRunningTime="2025-10-05 07:05:54.889799611 +0000 UTC m=+1077.130652386" Oct 05 07:05:55 crc kubenswrapper[4846]: I1005 07:05:55.759509 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-metrics-58xdl"] Oct 05 07:05:55 crc kubenswrapper[4846]: I1005 07:05:55.761115 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-58xdl" Oct 05 07:05:55 crc kubenswrapper[4846]: I1005 07:05:55.800833 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-metrics-config" Oct 05 07:05:55 crc kubenswrapper[4846]: I1005 07:05:55.807708 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/19f3a362-4973-4437-ac49-de71508aa4f2-combined-ca-bundle\") pod \"ovn-controller-metrics-58xdl\" (UID: \"19f3a362-4973-4437-ac49-de71508aa4f2\") " pod="openstack/ovn-controller-metrics-58xdl" Oct 05 07:05:55 crc kubenswrapper[4846]: I1005 07:05:55.808047 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/19f3a362-4973-4437-ac49-de71508aa4f2-ovs-rundir\") pod \"ovn-controller-metrics-58xdl\" (UID: \"19f3a362-4973-4437-ac49-de71508aa4f2\") " pod="openstack/ovn-controller-metrics-58xdl" Oct 05 07:05:55 crc kubenswrapper[4846]: I1005 07:05:55.808167 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/19f3a362-4973-4437-ac49-de71508aa4f2-config\") pod \"ovn-controller-metrics-58xdl\" (UID: \"19f3a362-4973-4437-ac49-de71508aa4f2\") " pod="openstack/ovn-controller-metrics-58xdl" Oct 05 07:05:55 crc kubenswrapper[4846]: I1005 07:05:55.808337 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/19f3a362-4973-4437-ac49-de71508aa4f2-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-58xdl\" (UID: \"19f3a362-4973-4437-ac49-de71508aa4f2\") " pod="openstack/ovn-controller-metrics-58xdl" Oct 05 07:05:55 crc kubenswrapper[4846]: I1005 07:05:55.808460 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/19f3a362-4973-4437-ac49-de71508aa4f2-ovn-rundir\") pod \"ovn-controller-metrics-58xdl\" (UID: \"19f3a362-4973-4437-ac49-de71508aa4f2\") " pod="openstack/ovn-controller-metrics-58xdl" Oct 05 07:05:55 crc kubenswrapper[4846]: I1005 07:05:55.808641 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n7jqd\" (UniqueName: \"kubernetes.io/projected/19f3a362-4973-4437-ac49-de71508aa4f2-kube-api-access-n7jqd\") pod \"ovn-controller-metrics-58xdl\" (UID: \"19f3a362-4973-4437-ac49-de71508aa4f2\") " pod="openstack/ovn-controller-metrics-58xdl" Oct 05 07:05:55 crc kubenswrapper[4846]: I1005 07:05:55.816962 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-58xdl"] Oct 05 07:05:55 crc kubenswrapper[4846]: I1005 07:05:55.912523 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/19f3a362-4973-4437-ac49-de71508aa4f2-ovn-rundir\") pod \"ovn-controller-metrics-58xdl\" (UID: \"19f3a362-4973-4437-ac49-de71508aa4f2\") " pod="openstack/ovn-controller-metrics-58xdl" Oct 05 07:05:55 crc kubenswrapper[4846]: I1005 07:05:55.912634 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n7jqd\" (UniqueName: \"kubernetes.io/projected/19f3a362-4973-4437-ac49-de71508aa4f2-kube-api-access-n7jqd\") pod \"ovn-controller-metrics-58xdl\" (UID: \"19f3a362-4973-4437-ac49-de71508aa4f2\") " pod="openstack/ovn-controller-metrics-58xdl" Oct 05 07:05:55 crc kubenswrapper[4846]: I1005 07:05:55.912696 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/19f3a362-4973-4437-ac49-de71508aa4f2-combined-ca-bundle\") pod \"ovn-controller-metrics-58xdl\" (UID: \"19f3a362-4973-4437-ac49-de71508aa4f2\") " pod="openstack/ovn-controller-metrics-58xdl" Oct 05 07:05:55 crc kubenswrapper[4846]: I1005 07:05:55.912724 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/19f3a362-4973-4437-ac49-de71508aa4f2-ovs-rundir\") pod \"ovn-controller-metrics-58xdl\" (UID: \"19f3a362-4973-4437-ac49-de71508aa4f2\") " pod="openstack/ovn-controller-metrics-58xdl" Oct 05 07:05:55 crc kubenswrapper[4846]: I1005 07:05:55.912748 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/19f3a362-4973-4437-ac49-de71508aa4f2-config\") pod \"ovn-controller-metrics-58xdl\" (UID: \"19f3a362-4973-4437-ac49-de71508aa4f2\") " pod="openstack/ovn-controller-metrics-58xdl" Oct 05 07:05:55 crc kubenswrapper[4846]: I1005 07:05:55.912796 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/19f3a362-4973-4437-ac49-de71508aa4f2-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-58xdl\" (UID: \"19f3a362-4973-4437-ac49-de71508aa4f2\") " pod="openstack/ovn-controller-metrics-58xdl" Oct 05 07:05:55 crc kubenswrapper[4846]: I1005 07:05:55.917436 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/19f3a362-4973-4437-ac49-de71508aa4f2-config\") pod \"ovn-controller-metrics-58xdl\" (UID: \"19f3a362-4973-4437-ac49-de71508aa4f2\") " pod="openstack/ovn-controller-metrics-58xdl" Oct 05 07:05:55 crc kubenswrapper[4846]: I1005 07:05:55.917675 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/19f3a362-4973-4437-ac49-de71508aa4f2-ovs-rundir\") pod \"ovn-controller-metrics-58xdl\" (UID: \"19f3a362-4973-4437-ac49-de71508aa4f2\") " pod="openstack/ovn-controller-metrics-58xdl" Oct 05 07:05:55 crc kubenswrapper[4846]: I1005 07:05:55.918564 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/19f3a362-4973-4437-ac49-de71508aa4f2-ovn-rundir\") pod \"ovn-controller-metrics-58xdl\" (UID: \"19f3a362-4973-4437-ac49-de71508aa4f2\") " pod="openstack/ovn-controller-metrics-58xdl" Oct 05 07:05:55 crc kubenswrapper[4846]: I1005 07:05:55.939890 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/19f3a362-4973-4437-ac49-de71508aa4f2-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-58xdl\" (UID: \"19f3a362-4973-4437-ac49-de71508aa4f2\") " pod="openstack/ovn-controller-metrics-58xdl" Oct 05 07:05:55 crc kubenswrapper[4846]: I1005 07:05:55.942399 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/19f3a362-4973-4437-ac49-de71508aa4f2-combined-ca-bundle\") pod \"ovn-controller-metrics-58xdl\" (UID: \"19f3a362-4973-4437-ac49-de71508aa4f2\") " pod="openstack/ovn-controller-metrics-58xdl" Oct 05 07:05:55 crc kubenswrapper[4846]: I1005 07:05:55.947165 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n7jqd\" (UniqueName: \"kubernetes.io/projected/19f3a362-4973-4437-ac49-de71508aa4f2-kube-api-access-n7jqd\") pod \"ovn-controller-metrics-58xdl\" (UID: \"19f3a362-4973-4437-ac49-de71508aa4f2\") " pod="openstack/ovn-controller-metrics-58xdl" Oct 05 07:05:56 crc kubenswrapper[4846]: I1005 07:05:56.016460 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6c64dcbb65-fw9xh"] Oct 05 07:05:56 crc kubenswrapper[4846]: I1005 07:05:56.063832 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7c559c78c7-tkv5q"] Oct 05 07:05:56 crc kubenswrapper[4846]: I1005 07:05:56.065355 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7c559c78c7-tkv5q" Oct 05 07:05:56 crc kubenswrapper[4846]: I1005 07:05:56.078999 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-sb" Oct 05 07:05:56 crc kubenswrapper[4846]: I1005 07:05:56.098131 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7c559c78c7-tkv5q"] Oct 05 07:05:56 crc kubenswrapper[4846]: I1005 07:05:56.114971 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zcsj9\" (UniqueName: \"kubernetes.io/projected/727b9fbc-1de3-4e7e-ac8a-ffbbd76eac0e-kube-api-access-zcsj9\") pod \"dnsmasq-dns-7c559c78c7-tkv5q\" (UID: \"727b9fbc-1de3-4e7e-ac8a-ffbbd76eac0e\") " pod="openstack/dnsmasq-dns-7c559c78c7-tkv5q" Oct 05 07:05:56 crc kubenswrapper[4846]: I1005 07:05:56.115042 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/727b9fbc-1de3-4e7e-ac8a-ffbbd76eac0e-ovsdbserver-sb\") pod \"dnsmasq-dns-7c559c78c7-tkv5q\" (UID: \"727b9fbc-1de3-4e7e-ac8a-ffbbd76eac0e\") " pod="openstack/dnsmasq-dns-7c559c78c7-tkv5q" Oct 05 07:05:56 crc kubenswrapper[4846]: I1005 07:05:56.115092 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/727b9fbc-1de3-4e7e-ac8a-ffbbd76eac0e-dns-svc\") pod \"dnsmasq-dns-7c559c78c7-tkv5q\" (UID: \"727b9fbc-1de3-4e7e-ac8a-ffbbd76eac0e\") " pod="openstack/dnsmasq-dns-7c559c78c7-tkv5q" Oct 05 07:05:56 crc kubenswrapper[4846]: I1005 07:05:56.115112 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/727b9fbc-1de3-4e7e-ac8a-ffbbd76eac0e-config\") pod \"dnsmasq-dns-7c559c78c7-tkv5q\" (UID: \"727b9fbc-1de3-4e7e-ac8a-ffbbd76eac0e\") " pod="openstack/dnsmasq-dns-7c559c78c7-tkv5q" Oct 05 07:05:56 crc kubenswrapper[4846]: I1005 07:05:56.135796 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-58xdl" Oct 05 07:05:56 crc kubenswrapper[4846]: I1005 07:05:56.216715 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/727b9fbc-1de3-4e7e-ac8a-ffbbd76eac0e-dns-svc\") pod \"dnsmasq-dns-7c559c78c7-tkv5q\" (UID: \"727b9fbc-1de3-4e7e-ac8a-ffbbd76eac0e\") " pod="openstack/dnsmasq-dns-7c559c78c7-tkv5q" Oct 05 07:05:56 crc kubenswrapper[4846]: I1005 07:05:56.217089 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/727b9fbc-1de3-4e7e-ac8a-ffbbd76eac0e-config\") pod \"dnsmasq-dns-7c559c78c7-tkv5q\" (UID: \"727b9fbc-1de3-4e7e-ac8a-ffbbd76eac0e\") " pod="openstack/dnsmasq-dns-7c559c78c7-tkv5q" Oct 05 07:05:56 crc kubenswrapper[4846]: I1005 07:05:56.217236 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zcsj9\" (UniqueName: \"kubernetes.io/projected/727b9fbc-1de3-4e7e-ac8a-ffbbd76eac0e-kube-api-access-zcsj9\") pod \"dnsmasq-dns-7c559c78c7-tkv5q\" (UID: \"727b9fbc-1de3-4e7e-ac8a-ffbbd76eac0e\") " pod="openstack/dnsmasq-dns-7c559c78c7-tkv5q" Oct 05 07:05:56 crc kubenswrapper[4846]: I1005 07:05:56.217279 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/727b9fbc-1de3-4e7e-ac8a-ffbbd76eac0e-ovsdbserver-sb\") pod \"dnsmasq-dns-7c559c78c7-tkv5q\" (UID: \"727b9fbc-1de3-4e7e-ac8a-ffbbd76eac0e\") " pod="openstack/dnsmasq-dns-7c559c78c7-tkv5q" Oct 05 07:05:56 crc kubenswrapper[4846]: I1005 07:05:56.217870 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/727b9fbc-1de3-4e7e-ac8a-ffbbd76eac0e-dns-svc\") pod \"dnsmasq-dns-7c559c78c7-tkv5q\" (UID: \"727b9fbc-1de3-4e7e-ac8a-ffbbd76eac0e\") " pod="openstack/dnsmasq-dns-7c559c78c7-tkv5q" Oct 05 07:05:56 crc kubenswrapper[4846]: I1005 07:05:56.218062 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/727b9fbc-1de3-4e7e-ac8a-ffbbd76eac0e-config\") pod \"dnsmasq-dns-7c559c78c7-tkv5q\" (UID: \"727b9fbc-1de3-4e7e-ac8a-ffbbd76eac0e\") " pod="openstack/dnsmasq-dns-7c559c78c7-tkv5q" Oct 05 07:05:56 crc kubenswrapper[4846]: I1005 07:05:56.218151 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/727b9fbc-1de3-4e7e-ac8a-ffbbd76eac0e-ovsdbserver-sb\") pod \"dnsmasq-dns-7c559c78c7-tkv5q\" (UID: \"727b9fbc-1de3-4e7e-ac8a-ffbbd76eac0e\") " pod="openstack/dnsmasq-dns-7c559c78c7-tkv5q" Oct 05 07:05:56 crc kubenswrapper[4846]: I1005 07:05:56.240222 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zcsj9\" (UniqueName: \"kubernetes.io/projected/727b9fbc-1de3-4e7e-ac8a-ffbbd76eac0e-kube-api-access-zcsj9\") pod \"dnsmasq-dns-7c559c78c7-tkv5q\" (UID: \"727b9fbc-1de3-4e7e-ac8a-ffbbd76eac0e\") " pod="openstack/dnsmasq-dns-7c559c78c7-tkv5q" Oct 05 07:05:56 crc kubenswrapper[4846]: I1005 07:05:56.285456 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5d4d9f7875-29gcr"] Oct 05 07:05:56 crc kubenswrapper[4846]: I1005 07:05:56.333932 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-84f55cbc7-bl827"] Oct 05 07:05:56 crc kubenswrapper[4846]: I1005 07:05:56.335413 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-84f55cbc7-bl827" Oct 05 07:05:56 crc kubenswrapper[4846]: I1005 07:05:56.348532 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-nb" Oct 05 07:05:56 crc kubenswrapper[4846]: I1005 07:05:56.396740 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-84f55cbc7-bl827"] Oct 05 07:05:56 crc kubenswrapper[4846]: I1005 07:05:56.440908 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7c559c78c7-tkv5q" Oct 05 07:05:56 crc kubenswrapper[4846]: I1005 07:05:56.531890 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0ca13142-fb38-4f67-88f4-834bb39fcf32-config\") pod \"dnsmasq-dns-84f55cbc7-bl827\" (UID: \"0ca13142-fb38-4f67-88f4-834bb39fcf32\") " pod="openstack/dnsmasq-dns-84f55cbc7-bl827" Oct 05 07:05:56 crc kubenswrapper[4846]: I1005 07:05:56.533568 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0ca13142-fb38-4f67-88f4-834bb39fcf32-ovsdbserver-nb\") pod \"dnsmasq-dns-84f55cbc7-bl827\" (UID: \"0ca13142-fb38-4f67-88f4-834bb39fcf32\") " pod="openstack/dnsmasq-dns-84f55cbc7-bl827" Oct 05 07:05:56 crc kubenswrapper[4846]: I1005 07:05:56.533792 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cxlh2\" (UniqueName: \"kubernetes.io/projected/0ca13142-fb38-4f67-88f4-834bb39fcf32-kube-api-access-cxlh2\") pod \"dnsmasq-dns-84f55cbc7-bl827\" (UID: \"0ca13142-fb38-4f67-88f4-834bb39fcf32\") " pod="openstack/dnsmasq-dns-84f55cbc7-bl827" Oct 05 07:05:56 crc kubenswrapper[4846]: I1005 07:05:56.534356 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0ca13142-fb38-4f67-88f4-834bb39fcf32-ovsdbserver-sb\") pod \"dnsmasq-dns-84f55cbc7-bl827\" (UID: \"0ca13142-fb38-4f67-88f4-834bb39fcf32\") " pod="openstack/dnsmasq-dns-84f55cbc7-bl827" Oct 05 07:05:56 crc kubenswrapper[4846]: I1005 07:05:56.534593 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0ca13142-fb38-4f67-88f4-834bb39fcf32-dns-svc\") pod \"dnsmasq-dns-84f55cbc7-bl827\" (UID: \"0ca13142-fb38-4f67-88f4-834bb39fcf32\") " pod="openstack/dnsmasq-dns-84f55cbc7-bl827" Oct 05 07:05:56 crc kubenswrapper[4846]: I1005 07:05:56.637273 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0ca13142-fb38-4f67-88f4-834bb39fcf32-ovsdbserver-nb\") pod \"dnsmasq-dns-84f55cbc7-bl827\" (UID: \"0ca13142-fb38-4f67-88f4-834bb39fcf32\") " pod="openstack/dnsmasq-dns-84f55cbc7-bl827" Oct 05 07:05:56 crc kubenswrapper[4846]: I1005 07:05:56.637349 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cxlh2\" (UniqueName: \"kubernetes.io/projected/0ca13142-fb38-4f67-88f4-834bb39fcf32-kube-api-access-cxlh2\") pod \"dnsmasq-dns-84f55cbc7-bl827\" (UID: \"0ca13142-fb38-4f67-88f4-834bb39fcf32\") " pod="openstack/dnsmasq-dns-84f55cbc7-bl827" Oct 05 07:05:56 crc kubenswrapper[4846]: I1005 07:05:56.637443 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0ca13142-fb38-4f67-88f4-834bb39fcf32-ovsdbserver-sb\") pod \"dnsmasq-dns-84f55cbc7-bl827\" (UID: \"0ca13142-fb38-4f67-88f4-834bb39fcf32\") " pod="openstack/dnsmasq-dns-84f55cbc7-bl827" Oct 05 07:05:56 crc kubenswrapper[4846]: I1005 07:05:56.637481 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0ca13142-fb38-4f67-88f4-834bb39fcf32-dns-svc\") pod \"dnsmasq-dns-84f55cbc7-bl827\" (UID: \"0ca13142-fb38-4f67-88f4-834bb39fcf32\") " pod="openstack/dnsmasq-dns-84f55cbc7-bl827" Oct 05 07:05:56 crc kubenswrapper[4846]: I1005 07:05:56.637511 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0ca13142-fb38-4f67-88f4-834bb39fcf32-config\") pod \"dnsmasq-dns-84f55cbc7-bl827\" (UID: \"0ca13142-fb38-4f67-88f4-834bb39fcf32\") " pod="openstack/dnsmasq-dns-84f55cbc7-bl827" Oct 05 07:05:56 crc kubenswrapper[4846]: I1005 07:05:56.638569 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0ca13142-fb38-4f67-88f4-834bb39fcf32-config\") pod \"dnsmasq-dns-84f55cbc7-bl827\" (UID: \"0ca13142-fb38-4f67-88f4-834bb39fcf32\") " pod="openstack/dnsmasq-dns-84f55cbc7-bl827" Oct 05 07:05:56 crc kubenswrapper[4846]: I1005 07:05:56.638742 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0ca13142-fb38-4f67-88f4-834bb39fcf32-ovsdbserver-nb\") pod \"dnsmasq-dns-84f55cbc7-bl827\" (UID: \"0ca13142-fb38-4f67-88f4-834bb39fcf32\") " pod="openstack/dnsmasq-dns-84f55cbc7-bl827" Oct 05 07:05:56 crc kubenswrapper[4846]: I1005 07:05:56.639670 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0ca13142-fb38-4f67-88f4-834bb39fcf32-dns-svc\") pod \"dnsmasq-dns-84f55cbc7-bl827\" (UID: \"0ca13142-fb38-4f67-88f4-834bb39fcf32\") " pod="openstack/dnsmasq-dns-84f55cbc7-bl827" Oct 05 07:05:56 crc kubenswrapper[4846]: I1005 07:05:56.640505 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0ca13142-fb38-4f67-88f4-834bb39fcf32-ovsdbserver-sb\") pod \"dnsmasq-dns-84f55cbc7-bl827\" (UID: \"0ca13142-fb38-4f67-88f4-834bb39fcf32\") " pod="openstack/dnsmasq-dns-84f55cbc7-bl827" Oct 05 07:05:56 crc kubenswrapper[4846]: I1005 07:05:56.663084 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cxlh2\" (UniqueName: \"kubernetes.io/projected/0ca13142-fb38-4f67-88f4-834bb39fcf32-kube-api-access-cxlh2\") pod \"dnsmasq-dns-84f55cbc7-bl827\" (UID: \"0ca13142-fb38-4f67-88f4-834bb39fcf32\") " pod="openstack/dnsmasq-dns-84f55cbc7-bl827" Oct 05 07:05:56 crc kubenswrapper[4846]: I1005 07:05:56.674941 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-84f55cbc7-bl827" Oct 05 07:05:56 crc kubenswrapper[4846]: I1005 07:05:56.965171 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-galera-0" Oct 05 07:05:56 crc kubenswrapper[4846]: I1005 07:05:56.965605 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-galera-0" Oct 05 07:05:58 crc kubenswrapper[4846]: I1005 07:05:58.092520 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-cell1-galera-0" Oct 05 07:05:58 crc kubenswrapper[4846]: I1005 07:05:58.092628 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-cell1-galera-0" Oct 05 07:05:58 crc kubenswrapper[4846]: I1005 07:05:58.231456 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/memcached-0" Oct 05 07:06:00 crc kubenswrapper[4846]: I1005 07:06:00.303067 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Oct 05 07:06:00 crc kubenswrapper[4846]: I1005 07:06:00.374673 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7c559c78c7-tkv5q"] Oct 05 07:06:00 crc kubenswrapper[4846]: I1005 07:06:00.435318 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7bc89dc697-k6cm6"] Oct 05 07:06:00 crc kubenswrapper[4846]: I1005 07:06:00.437109 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7bc89dc697-k6cm6" Oct 05 07:06:00 crc kubenswrapper[4846]: I1005 07:06:00.442414 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7bc89dc697-k6cm6"] Oct 05 07:06:00 crc kubenswrapper[4846]: I1005 07:06:00.519324 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/51a08294-d363-4169-8b63-c759c3cc30f0-ovsdbserver-sb\") pod \"dnsmasq-dns-7bc89dc697-k6cm6\" (UID: \"51a08294-d363-4169-8b63-c759c3cc30f0\") " pod="openstack/dnsmasq-dns-7bc89dc697-k6cm6" Oct 05 07:06:00 crc kubenswrapper[4846]: I1005 07:06:00.519391 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gskj2\" (UniqueName: \"kubernetes.io/projected/51a08294-d363-4169-8b63-c759c3cc30f0-kube-api-access-gskj2\") pod \"dnsmasq-dns-7bc89dc697-k6cm6\" (UID: \"51a08294-d363-4169-8b63-c759c3cc30f0\") " pod="openstack/dnsmasq-dns-7bc89dc697-k6cm6" Oct 05 07:06:00 crc kubenswrapper[4846]: I1005 07:06:00.519451 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/51a08294-d363-4169-8b63-c759c3cc30f0-dns-svc\") pod \"dnsmasq-dns-7bc89dc697-k6cm6\" (UID: \"51a08294-d363-4169-8b63-c759c3cc30f0\") " pod="openstack/dnsmasq-dns-7bc89dc697-k6cm6" Oct 05 07:06:00 crc kubenswrapper[4846]: I1005 07:06:00.519489 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/51a08294-d363-4169-8b63-c759c3cc30f0-config\") pod \"dnsmasq-dns-7bc89dc697-k6cm6\" (UID: \"51a08294-d363-4169-8b63-c759c3cc30f0\") " pod="openstack/dnsmasq-dns-7bc89dc697-k6cm6" Oct 05 07:06:00 crc kubenswrapper[4846]: I1005 07:06:00.519524 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/51a08294-d363-4169-8b63-c759c3cc30f0-ovsdbserver-nb\") pod \"dnsmasq-dns-7bc89dc697-k6cm6\" (UID: \"51a08294-d363-4169-8b63-c759c3cc30f0\") " pod="openstack/dnsmasq-dns-7bc89dc697-k6cm6" Oct 05 07:06:00 crc kubenswrapper[4846]: I1005 07:06:00.622658 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/51a08294-d363-4169-8b63-c759c3cc30f0-ovsdbserver-sb\") pod \"dnsmasq-dns-7bc89dc697-k6cm6\" (UID: \"51a08294-d363-4169-8b63-c759c3cc30f0\") " pod="openstack/dnsmasq-dns-7bc89dc697-k6cm6" Oct 05 07:06:00 crc kubenswrapper[4846]: I1005 07:06:00.622755 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gskj2\" (UniqueName: \"kubernetes.io/projected/51a08294-d363-4169-8b63-c759c3cc30f0-kube-api-access-gskj2\") pod \"dnsmasq-dns-7bc89dc697-k6cm6\" (UID: \"51a08294-d363-4169-8b63-c759c3cc30f0\") " pod="openstack/dnsmasq-dns-7bc89dc697-k6cm6" Oct 05 07:06:00 crc kubenswrapper[4846]: I1005 07:06:00.622883 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/51a08294-d363-4169-8b63-c759c3cc30f0-dns-svc\") pod \"dnsmasq-dns-7bc89dc697-k6cm6\" (UID: \"51a08294-d363-4169-8b63-c759c3cc30f0\") " pod="openstack/dnsmasq-dns-7bc89dc697-k6cm6" Oct 05 07:06:00 crc kubenswrapper[4846]: I1005 07:06:00.622935 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/51a08294-d363-4169-8b63-c759c3cc30f0-config\") pod \"dnsmasq-dns-7bc89dc697-k6cm6\" (UID: \"51a08294-d363-4169-8b63-c759c3cc30f0\") " pod="openstack/dnsmasq-dns-7bc89dc697-k6cm6" Oct 05 07:06:00 crc kubenswrapper[4846]: I1005 07:06:00.622991 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/51a08294-d363-4169-8b63-c759c3cc30f0-ovsdbserver-nb\") pod \"dnsmasq-dns-7bc89dc697-k6cm6\" (UID: \"51a08294-d363-4169-8b63-c759c3cc30f0\") " pod="openstack/dnsmasq-dns-7bc89dc697-k6cm6" Oct 05 07:06:00 crc kubenswrapper[4846]: I1005 07:06:00.624616 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/51a08294-d363-4169-8b63-c759c3cc30f0-dns-svc\") pod \"dnsmasq-dns-7bc89dc697-k6cm6\" (UID: \"51a08294-d363-4169-8b63-c759c3cc30f0\") " pod="openstack/dnsmasq-dns-7bc89dc697-k6cm6" Oct 05 07:06:00 crc kubenswrapper[4846]: I1005 07:06:00.624896 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/51a08294-d363-4169-8b63-c759c3cc30f0-ovsdbserver-sb\") pod \"dnsmasq-dns-7bc89dc697-k6cm6\" (UID: \"51a08294-d363-4169-8b63-c759c3cc30f0\") " pod="openstack/dnsmasq-dns-7bc89dc697-k6cm6" Oct 05 07:06:00 crc kubenswrapper[4846]: I1005 07:06:00.626160 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/51a08294-d363-4169-8b63-c759c3cc30f0-config\") pod \"dnsmasq-dns-7bc89dc697-k6cm6\" (UID: \"51a08294-d363-4169-8b63-c759c3cc30f0\") " pod="openstack/dnsmasq-dns-7bc89dc697-k6cm6" Oct 05 07:06:00 crc kubenswrapper[4846]: I1005 07:06:00.647039 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gskj2\" (UniqueName: \"kubernetes.io/projected/51a08294-d363-4169-8b63-c759c3cc30f0-kube-api-access-gskj2\") pod \"dnsmasq-dns-7bc89dc697-k6cm6\" (UID: \"51a08294-d363-4169-8b63-c759c3cc30f0\") " pod="openstack/dnsmasq-dns-7bc89dc697-k6cm6" Oct 05 07:06:00 crc kubenswrapper[4846]: I1005 07:06:00.745732 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/51a08294-d363-4169-8b63-c759c3cc30f0-ovsdbserver-nb\") pod \"dnsmasq-dns-7bc89dc697-k6cm6\" (UID: \"51a08294-d363-4169-8b63-c759c3cc30f0\") " pod="openstack/dnsmasq-dns-7bc89dc697-k6cm6" Oct 05 07:06:00 crc kubenswrapper[4846]: I1005 07:06:00.769957 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7bc89dc697-k6cm6" Oct 05 07:06:01 crc kubenswrapper[4846]: I1005 07:06:01.514850 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-storage-0"] Oct 05 07:06:01 crc kubenswrapper[4846]: I1005 07:06:01.533333 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Oct 05 07:06:01 crc kubenswrapper[4846]: I1005 07:06:01.533542 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Oct 05 07:06:01 crc kubenswrapper[4846]: I1005 07:06:01.536430 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-storage-config-data" Oct 05 07:06:01 crc kubenswrapper[4846]: I1005 07:06:01.537233 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-swift-dockercfg-q8wdn" Oct 05 07:06:01 crc kubenswrapper[4846]: I1005 07:06:01.537944 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-files" Oct 05 07:06:01 crc kubenswrapper[4846]: I1005 07:06:01.538147 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-conf" Oct 05 07:06:01 crc kubenswrapper[4846]: I1005 07:06:01.645124 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dv6nn\" (UniqueName: \"kubernetes.io/projected/62d2ef82-9372-4b55-990e-30eb8e16e439-kube-api-access-dv6nn\") pod \"swift-storage-0\" (UID: \"62d2ef82-9372-4b55-990e-30eb8e16e439\") " pod="openstack/swift-storage-0" Oct 05 07:06:01 crc kubenswrapper[4846]: I1005 07:06:01.645203 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/62d2ef82-9372-4b55-990e-30eb8e16e439-lock\") pod \"swift-storage-0\" (UID: \"62d2ef82-9372-4b55-990e-30eb8e16e439\") " pod="openstack/swift-storage-0" Oct 05 07:06:01 crc kubenswrapper[4846]: I1005 07:06:01.645344 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/62d2ef82-9372-4b55-990e-30eb8e16e439-etc-swift\") pod \"swift-storage-0\" (UID: \"62d2ef82-9372-4b55-990e-30eb8e16e439\") " pod="openstack/swift-storage-0" Oct 05 07:06:01 crc kubenswrapper[4846]: I1005 07:06:01.645541 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"swift-storage-0\" (UID: \"62d2ef82-9372-4b55-990e-30eb8e16e439\") " pod="openstack/swift-storage-0" Oct 05 07:06:01 crc kubenswrapper[4846]: I1005 07:06:01.645770 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/62d2ef82-9372-4b55-990e-30eb8e16e439-cache\") pod \"swift-storage-0\" (UID: \"62d2ef82-9372-4b55-990e-30eb8e16e439\") " pod="openstack/swift-storage-0" Oct 05 07:06:01 crc kubenswrapper[4846]: I1005 07:06:01.747761 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"swift-storage-0\" (UID: \"62d2ef82-9372-4b55-990e-30eb8e16e439\") " pod="openstack/swift-storage-0" Oct 05 07:06:01 crc kubenswrapper[4846]: I1005 07:06:01.747895 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/62d2ef82-9372-4b55-990e-30eb8e16e439-cache\") pod \"swift-storage-0\" (UID: \"62d2ef82-9372-4b55-990e-30eb8e16e439\") " pod="openstack/swift-storage-0" Oct 05 07:06:01 crc kubenswrapper[4846]: I1005 07:06:01.747995 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dv6nn\" (UniqueName: \"kubernetes.io/projected/62d2ef82-9372-4b55-990e-30eb8e16e439-kube-api-access-dv6nn\") pod \"swift-storage-0\" (UID: \"62d2ef82-9372-4b55-990e-30eb8e16e439\") " pod="openstack/swift-storage-0" Oct 05 07:06:01 crc kubenswrapper[4846]: I1005 07:06:01.748044 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/62d2ef82-9372-4b55-990e-30eb8e16e439-lock\") pod \"swift-storage-0\" (UID: \"62d2ef82-9372-4b55-990e-30eb8e16e439\") " pod="openstack/swift-storage-0" Oct 05 07:06:01 crc kubenswrapper[4846]: I1005 07:06:01.748074 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/62d2ef82-9372-4b55-990e-30eb8e16e439-etc-swift\") pod \"swift-storage-0\" (UID: \"62d2ef82-9372-4b55-990e-30eb8e16e439\") " pod="openstack/swift-storage-0" Oct 05 07:06:01 crc kubenswrapper[4846]: E1005 07:06:01.748231 4846 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 05 07:06:01 crc kubenswrapper[4846]: I1005 07:06:01.748238 4846 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"swift-storage-0\" (UID: \"62d2ef82-9372-4b55-990e-30eb8e16e439\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/swift-storage-0" Oct 05 07:06:01 crc kubenswrapper[4846]: E1005 07:06:01.748251 4846 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Oct 05 07:06:01 crc kubenswrapper[4846]: E1005 07:06:01.748496 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/62d2ef82-9372-4b55-990e-30eb8e16e439-etc-swift podName:62d2ef82-9372-4b55-990e-30eb8e16e439 nodeName:}" failed. No retries permitted until 2025-10-05 07:06:02.24847843 +0000 UTC m=+1084.489331205 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/62d2ef82-9372-4b55-990e-30eb8e16e439-etc-swift") pod "swift-storage-0" (UID: "62d2ef82-9372-4b55-990e-30eb8e16e439") : configmap "swift-ring-files" not found Oct 05 07:06:01 crc kubenswrapper[4846]: I1005 07:06:01.748577 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/62d2ef82-9372-4b55-990e-30eb8e16e439-cache\") pod \"swift-storage-0\" (UID: \"62d2ef82-9372-4b55-990e-30eb8e16e439\") " pod="openstack/swift-storage-0" Oct 05 07:06:01 crc kubenswrapper[4846]: I1005 07:06:01.748613 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/62d2ef82-9372-4b55-990e-30eb8e16e439-lock\") pod \"swift-storage-0\" (UID: \"62d2ef82-9372-4b55-990e-30eb8e16e439\") " pod="openstack/swift-storage-0" Oct 05 07:06:01 crc kubenswrapper[4846]: I1005 07:06:01.768417 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dv6nn\" (UniqueName: \"kubernetes.io/projected/62d2ef82-9372-4b55-990e-30eb8e16e439-kube-api-access-dv6nn\") pod \"swift-storage-0\" (UID: \"62d2ef82-9372-4b55-990e-30eb8e16e439\") " pod="openstack/swift-storage-0" Oct 05 07:06:01 crc kubenswrapper[4846]: I1005 07:06:01.772023 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"swift-storage-0\" (UID: \"62d2ef82-9372-4b55-990e-30eb8e16e439\") " pod="openstack/swift-storage-0" Oct 05 07:06:02 crc kubenswrapper[4846]: I1005 07:06:02.050787 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-lrj2x"] Oct 05 07:06:02 crc kubenswrapper[4846]: I1005 07:06:02.055129 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-lrj2x" Oct 05 07:06:02 crc kubenswrapper[4846]: I1005 07:06:02.058602 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-config-data" Oct 05 07:06:02 crc kubenswrapper[4846]: I1005 07:06:02.058635 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Oct 05 07:06:02 crc kubenswrapper[4846]: I1005 07:06:02.058960 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-scripts" Oct 05 07:06:02 crc kubenswrapper[4846]: I1005 07:06:02.094791 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-ring-rebalance-lrj2x"] Oct 05 07:06:02 crc kubenswrapper[4846]: E1005 07:06:02.095786 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[combined-ca-bundle dispersionconf etc-swift kube-api-access-t527f ring-data-devices scripts swiftconf], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openstack/swift-ring-rebalance-lrj2x" podUID="182bb989-7f0e-4aec-8405-9d8e4f73fece" Oct 05 07:06:02 crc kubenswrapper[4846]: I1005 07:06:02.113867 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-b9gxc"] Oct 05 07:06:02 crc kubenswrapper[4846]: I1005 07:06:02.115126 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-b9gxc" Oct 05 07:06:02 crc kubenswrapper[4846]: I1005 07:06:02.131883 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-b9gxc"] Oct 05 07:06:02 crc kubenswrapper[4846]: I1005 07:06:02.146864 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-ring-rebalance-lrj2x"] Oct 05 07:06:02 crc kubenswrapper[4846]: I1005 07:06:02.157615 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/182bb989-7f0e-4aec-8405-9d8e4f73fece-dispersionconf\") pod \"swift-ring-rebalance-lrj2x\" (UID: \"182bb989-7f0e-4aec-8405-9d8e4f73fece\") " pod="openstack/swift-ring-rebalance-lrj2x" Oct 05 07:06:02 crc kubenswrapper[4846]: I1005 07:06:02.157673 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/182bb989-7f0e-4aec-8405-9d8e4f73fece-combined-ca-bundle\") pod \"swift-ring-rebalance-lrj2x\" (UID: \"182bb989-7f0e-4aec-8405-9d8e4f73fece\") " pod="openstack/swift-ring-rebalance-lrj2x" Oct 05 07:06:02 crc kubenswrapper[4846]: I1005 07:06:02.157818 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/182bb989-7f0e-4aec-8405-9d8e4f73fece-ring-data-devices\") pod \"swift-ring-rebalance-lrj2x\" (UID: \"182bb989-7f0e-4aec-8405-9d8e4f73fece\") " pod="openstack/swift-ring-rebalance-lrj2x" Oct 05 07:06:02 crc kubenswrapper[4846]: I1005 07:06:02.158009 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/182bb989-7f0e-4aec-8405-9d8e4f73fece-swiftconf\") pod \"swift-ring-rebalance-lrj2x\" (UID: \"182bb989-7f0e-4aec-8405-9d8e4f73fece\") " pod="openstack/swift-ring-rebalance-lrj2x" Oct 05 07:06:02 crc kubenswrapper[4846]: I1005 07:06:02.158086 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/182bb989-7f0e-4aec-8405-9d8e4f73fece-etc-swift\") pod \"swift-ring-rebalance-lrj2x\" (UID: \"182bb989-7f0e-4aec-8405-9d8e4f73fece\") " pod="openstack/swift-ring-rebalance-lrj2x" Oct 05 07:06:02 crc kubenswrapper[4846]: I1005 07:06:02.158129 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/182bb989-7f0e-4aec-8405-9d8e4f73fece-scripts\") pod \"swift-ring-rebalance-lrj2x\" (UID: \"182bb989-7f0e-4aec-8405-9d8e4f73fece\") " pod="openstack/swift-ring-rebalance-lrj2x" Oct 05 07:06:02 crc kubenswrapper[4846]: I1005 07:06:02.158301 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t527f\" (UniqueName: \"kubernetes.io/projected/182bb989-7f0e-4aec-8405-9d8e4f73fece-kube-api-access-t527f\") pod \"swift-ring-rebalance-lrj2x\" (UID: \"182bb989-7f0e-4aec-8405-9d8e4f73fece\") " pod="openstack/swift-ring-rebalance-lrj2x" Oct 05 07:06:02 crc kubenswrapper[4846]: I1005 07:06:02.259898 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0ffabf45-809d-43f4-902b-755111599aa8-scripts\") pod \"swift-ring-rebalance-b9gxc\" (UID: \"0ffabf45-809d-43f4-902b-755111599aa8\") " pod="openstack/swift-ring-rebalance-b9gxc" Oct 05 07:06:02 crc kubenswrapper[4846]: I1005 07:06:02.260278 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/0ffabf45-809d-43f4-902b-755111599aa8-dispersionconf\") pod \"swift-ring-rebalance-b9gxc\" (UID: \"0ffabf45-809d-43f4-902b-755111599aa8\") " pod="openstack/swift-ring-rebalance-b9gxc" Oct 05 07:06:02 crc kubenswrapper[4846]: I1005 07:06:02.260410 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/62d2ef82-9372-4b55-990e-30eb8e16e439-etc-swift\") pod \"swift-storage-0\" (UID: \"62d2ef82-9372-4b55-990e-30eb8e16e439\") " pod="openstack/swift-storage-0" Oct 05 07:06:02 crc kubenswrapper[4846]: I1005 07:06:02.260502 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t527f\" (UniqueName: \"kubernetes.io/projected/182bb989-7f0e-4aec-8405-9d8e4f73fece-kube-api-access-t527f\") pod \"swift-ring-rebalance-lrj2x\" (UID: \"182bb989-7f0e-4aec-8405-9d8e4f73fece\") " pod="openstack/swift-ring-rebalance-lrj2x" Oct 05 07:06:02 crc kubenswrapper[4846]: I1005 07:06:02.260622 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/182bb989-7f0e-4aec-8405-9d8e4f73fece-dispersionconf\") pod \"swift-ring-rebalance-lrj2x\" (UID: \"182bb989-7f0e-4aec-8405-9d8e4f73fece\") " pod="openstack/swift-ring-rebalance-lrj2x" Oct 05 07:06:02 crc kubenswrapper[4846]: E1005 07:06:02.260674 4846 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 05 07:06:02 crc kubenswrapper[4846]: E1005 07:06:02.260744 4846 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Oct 05 07:06:02 crc kubenswrapper[4846]: I1005 07:06:02.260706 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/182bb989-7f0e-4aec-8405-9d8e4f73fece-combined-ca-bundle\") pod \"swift-ring-rebalance-lrj2x\" (UID: \"182bb989-7f0e-4aec-8405-9d8e4f73fece\") " pod="openstack/swift-ring-rebalance-lrj2x" Oct 05 07:06:02 crc kubenswrapper[4846]: E1005 07:06:02.260830 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/62d2ef82-9372-4b55-990e-30eb8e16e439-etc-swift podName:62d2ef82-9372-4b55-990e-30eb8e16e439 nodeName:}" failed. No retries permitted until 2025-10-05 07:06:03.260790762 +0000 UTC m=+1085.501643547 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/62d2ef82-9372-4b55-990e-30eb8e16e439-etc-swift") pod "swift-storage-0" (UID: "62d2ef82-9372-4b55-990e-30eb8e16e439") : configmap "swift-ring-files" not found Oct 05 07:06:02 crc kubenswrapper[4846]: I1005 07:06:02.260965 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hztth\" (UniqueName: \"kubernetes.io/projected/0ffabf45-809d-43f4-902b-755111599aa8-kube-api-access-hztth\") pod \"swift-ring-rebalance-b9gxc\" (UID: \"0ffabf45-809d-43f4-902b-755111599aa8\") " pod="openstack/swift-ring-rebalance-b9gxc" Oct 05 07:06:02 crc kubenswrapper[4846]: I1005 07:06:02.261060 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/182bb989-7f0e-4aec-8405-9d8e4f73fece-ring-data-devices\") pod \"swift-ring-rebalance-lrj2x\" (UID: \"182bb989-7f0e-4aec-8405-9d8e4f73fece\") " pod="openstack/swift-ring-rebalance-lrj2x" Oct 05 07:06:02 crc kubenswrapper[4846]: I1005 07:06:02.261138 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0ffabf45-809d-43f4-902b-755111599aa8-combined-ca-bundle\") pod \"swift-ring-rebalance-b9gxc\" (UID: \"0ffabf45-809d-43f4-902b-755111599aa8\") " pod="openstack/swift-ring-rebalance-b9gxc" Oct 05 07:06:02 crc kubenswrapper[4846]: I1005 07:06:02.261612 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/0ffabf45-809d-43f4-902b-755111599aa8-swiftconf\") pod \"swift-ring-rebalance-b9gxc\" (UID: \"0ffabf45-809d-43f4-902b-755111599aa8\") " pod="openstack/swift-ring-rebalance-b9gxc" Oct 05 07:06:02 crc kubenswrapper[4846]: I1005 07:06:02.261688 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/182bb989-7f0e-4aec-8405-9d8e4f73fece-swiftconf\") pod \"swift-ring-rebalance-lrj2x\" (UID: \"182bb989-7f0e-4aec-8405-9d8e4f73fece\") " pod="openstack/swift-ring-rebalance-lrj2x" Oct 05 07:06:02 crc kubenswrapper[4846]: I1005 07:06:02.261737 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/182bb989-7f0e-4aec-8405-9d8e4f73fece-etc-swift\") pod \"swift-ring-rebalance-lrj2x\" (UID: \"182bb989-7f0e-4aec-8405-9d8e4f73fece\") " pod="openstack/swift-ring-rebalance-lrj2x" Oct 05 07:06:02 crc kubenswrapper[4846]: I1005 07:06:02.261810 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/182bb989-7f0e-4aec-8405-9d8e4f73fece-scripts\") pod \"swift-ring-rebalance-lrj2x\" (UID: \"182bb989-7f0e-4aec-8405-9d8e4f73fece\") " pod="openstack/swift-ring-rebalance-lrj2x" Oct 05 07:06:02 crc kubenswrapper[4846]: I1005 07:06:02.261860 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/0ffabf45-809d-43f4-902b-755111599aa8-etc-swift\") pod \"swift-ring-rebalance-b9gxc\" (UID: \"0ffabf45-809d-43f4-902b-755111599aa8\") " pod="openstack/swift-ring-rebalance-b9gxc" Oct 05 07:06:02 crc kubenswrapper[4846]: I1005 07:06:02.261898 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/0ffabf45-809d-43f4-902b-755111599aa8-ring-data-devices\") pod \"swift-ring-rebalance-b9gxc\" (UID: \"0ffabf45-809d-43f4-902b-755111599aa8\") " pod="openstack/swift-ring-rebalance-b9gxc" Oct 05 07:06:02 crc kubenswrapper[4846]: I1005 07:06:02.262110 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/182bb989-7f0e-4aec-8405-9d8e4f73fece-ring-data-devices\") pod \"swift-ring-rebalance-lrj2x\" (UID: \"182bb989-7f0e-4aec-8405-9d8e4f73fece\") " pod="openstack/swift-ring-rebalance-lrj2x" Oct 05 07:06:02 crc kubenswrapper[4846]: I1005 07:06:02.262454 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/182bb989-7f0e-4aec-8405-9d8e4f73fece-etc-swift\") pod \"swift-ring-rebalance-lrj2x\" (UID: \"182bb989-7f0e-4aec-8405-9d8e4f73fece\") " pod="openstack/swift-ring-rebalance-lrj2x" Oct 05 07:06:02 crc kubenswrapper[4846]: I1005 07:06:02.262977 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/182bb989-7f0e-4aec-8405-9d8e4f73fece-scripts\") pod \"swift-ring-rebalance-lrj2x\" (UID: \"182bb989-7f0e-4aec-8405-9d8e4f73fece\") " pod="openstack/swift-ring-rebalance-lrj2x" Oct 05 07:06:02 crc kubenswrapper[4846]: I1005 07:06:02.265238 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/182bb989-7f0e-4aec-8405-9d8e4f73fece-dispersionconf\") pod \"swift-ring-rebalance-lrj2x\" (UID: \"182bb989-7f0e-4aec-8405-9d8e4f73fece\") " pod="openstack/swift-ring-rebalance-lrj2x" Oct 05 07:06:02 crc kubenswrapper[4846]: I1005 07:06:02.265279 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/182bb989-7f0e-4aec-8405-9d8e4f73fece-combined-ca-bundle\") pod \"swift-ring-rebalance-lrj2x\" (UID: \"182bb989-7f0e-4aec-8405-9d8e4f73fece\") " pod="openstack/swift-ring-rebalance-lrj2x" Oct 05 07:06:02 crc kubenswrapper[4846]: I1005 07:06:02.266289 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/182bb989-7f0e-4aec-8405-9d8e4f73fece-swiftconf\") pod \"swift-ring-rebalance-lrj2x\" (UID: \"182bb989-7f0e-4aec-8405-9d8e4f73fece\") " pod="openstack/swift-ring-rebalance-lrj2x" Oct 05 07:06:02 crc kubenswrapper[4846]: I1005 07:06:02.283352 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t527f\" (UniqueName: \"kubernetes.io/projected/182bb989-7f0e-4aec-8405-9d8e4f73fece-kube-api-access-t527f\") pod \"swift-ring-rebalance-lrj2x\" (UID: \"182bb989-7f0e-4aec-8405-9d8e4f73fece\") " pod="openstack/swift-ring-rebalance-lrj2x" Oct 05 07:06:02 crc kubenswrapper[4846]: I1005 07:06:02.363774 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/0ffabf45-809d-43f4-902b-755111599aa8-etc-swift\") pod \"swift-ring-rebalance-b9gxc\" (UID: \"0ffabf45-809d-43f4-902b-755111599aa8\") " pod="openstack/swift-ring-rebalance-b9gxc" Oct 05 07:06:02 crc kubenswrapper[4846]: I1005 07:06:02.364065 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/0ffabf45-809d-43f4-902b-755111599aa8-ring-data-devices\") pod \"swift-ring-rebalance-b9gxc\" (UID: \"0ffabf45-809d-43f4-902b-755111599aa8\") " pod="openstack/swift-ring-rebalance-b9gxc" Oct 05 07:06:02 crc kubenswrapper[4846]: I1005 07:06:02.364225 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0ffabf45-809d-43f4-902b-755111599aa8-scripts\") pod \"swift-ring-rebalance-b9gxc\" (UID: \"0ffabf45-809d-43f4-902b-755111599aa8\") " pod="openstack/swift-ring-rebalance-b9gxc" Oct 05 07:06:02 crc kubenswrapper[4846]: I1005 07:06:02.364307 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/0ffabf45-809d-43f4-902b-755111599aa8-dispersionconf\") pod \"swift-ring-rebalance-b9gxc\" (UID: \"0ffabf45-809d-43f4-902b-755111599aa8\") " pod="openstack/swift-ring-rebalance-b9gxc" Oct 05 07:06:02 crc kubenswrapper[4846]: I1005 07:06:02.364472 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hztth\" (UniqueName: \"kubernetes.io/projected/0ffabf45-809d-43f4-902b-755111599aa8-kube-api-access-hztth\") pod \"swift-ring-rebalance-b9gxc\" (UID: \"0ffabf45-809d-43f4-902b-755111599aa8\") " pod="openstack/swift-ring-rebalance-b9gxc" Oct 05 07:06:02 crc kubenswrapper[4846]: I1005 07:06:02.364554 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0ffabf45-809d-43f4-902b-755111599aa8-combined-ca-bundle\") pod \"swift-ring-rebalance-b9gxc\" (UID: \"0ffabf45-809d-43f4-902b-755111599aa8\") " pod="openstack/swift-ring-rebalance-b9gxc" Oct 05 07:06:02 crc kubenswrapper[4846]: I1005 07:06:02.364677 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/0ffabf45-809d-43f4-902b-755111599aa8-swiftconf\") pod \"swift-ring-rebalance-b9gxc\" (UID: \"0ffabf45-809d-43f4-902b-755111599aa8\") " pod="openstack/swift-ring-rebalance-b9gxc" Oct 05 07:06:02 crc kubenswrapper[4846]: I1005 07:06:02.365014 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0ffabf45-809d-43f4-902b-755111599aa8-scripts\") pod \"swift-ring-rebalance-b9gxc\" (UID: \"0ffabf45-809d-43f4-902b-755111599aa8\") " pod="openstack/swift-ring-rebalance-b9gxc" Oct 05 07:06:02 crc kubenswrapper[4846]: I1005 07:06:02.365174 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/0ffabf45-809d-43f4-902b-755111599aa8-ring-data-devices\") pod \"swift-ring-rebalance-b9gxc\" (UID: \"0ffabf45-809d-43f4-902b-755111599aa8\") " pod="openstack/swift-ring-rebalance-b9gxc" Oct 05 07:06:02 crc kubenswrapper[4846]: I1005 07:06:02.365472 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/0ffabf45-809d-43f4-902b-755111599aa8-etc-swift\") pod \"swift-ring-rebalance-b9gxc\" (UID: \"0ffabf45-809d-43f4-902b-755111599aa8\") " pod="openstack/swift-ring-rebalance-b9gxc" Oct 05 07:06:02 crc kubenswrapper[4846]: I1005 07:06:02.370129 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/0ffabf45-809d-43f4-902b-755111599aa8-swiftconf\") pod \"swift-ring-rebalance-b9gxc\" (UID: \"0ffabf45-809d-43f4-902b-755111599aa8\") " pod="openstack/swift-ring-rebalance-b9gxc" Oct 05 07:06:02 crc kubenswrapper[4846]: I1005 07:06:02.370468 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/0ffabf45-809d-43f4-902b-755111599aa8-dispersionconf\") pod \"swift-ring-rebalance-b9gxc\" (UID: \"0ffabf45-809d-43f4-902b-755111599aa8\") " pod="openstack/swift-ring-rebalance-b9gxc" Oct 05 07:06:02 crc kubenswrapper[4846]: I1005 07:06:02.371472 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0ffabf45-809d-43f4-902b-755111599aa8-combined-ca-bundle\") pod \"swift-ring-rebalance-b9gxc\" (UID: \"0ffabf45-809d-43f4-902b-755111599aa8\") " pod="openstack/swift-ring-rebalance-b9gxc" Oct 05 07:06:02 crc kubenswrapper[4846]: I1005 07:06:02.387140 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hztth\" (UniqueName: \"kubernetes.io/projected/0ffabf45-809d-43f4-902b-755111599aa8-kube-api-access-hztth\") pod \"swift-ring-rebalance-b9gxc\" (UID: \"0ffabf45-809d-43f4-902b-755111599aa8\") " pod="openstack/swift-ring-rebalance-b9gxc" Oct 05 07:06:02 crc kubenswrapper[4846]: I1005 07:06:02.431335 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-b9gxc" Oct 05 07:06:02 crc kubenswrapper[4846]: I1005 07:06:02.955668 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-lrj2x" Oct 05 07:06:02 crc kubenswrapper[4846]: I1005 07:06:02.968983 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-lrj2x" Oct 05 07:06:03 crc kubenswrapper[4846]: I1005 07:06:03.080785 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/182bb989-7f0e-4aec-8405-9d8e4f73fece-combined-ca-bundle\") pod \"182bb989-7f0e-4aec-8405-9d8e4f73fece\" (UID: \"182bb989-7f0e-4aec-8405-9d8e4f73fece\") " Oct 05 07:06:03 crc kubenswrapper[4846]: I1005 07:06:03.080979 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/182bb989-7f0e-4aec-8405-9d8e4f73fece-scripts\") pod \"182bb989-7f0e-4aec-8405-9d8e4f73fece\" (UID: \"182bb989-7f0e-4aec-8405-9d8e4f73fece\") " Oct 05 07:06:03 crc kubenswrapper[4846]: I1005 07:06:03.081219 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/182bb989-7f0e-4aec-8405-9d8e4f73fece-etc-swift\") pod \"182bb989-7f0e-4aec-8405-9d8e4f73fece\" (UID: \"182bb989-7f0e-4aec-8405-9d8e4f73fece\") " Oct 05 07:06:03 crc kubenswrapper[4846]: I1005 07:06:03.081265 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/182bb989-7f0e-4aec-8405-9d8e4f73fece-dispersionconf\") pod \"182bb989-7f0e-4aec-8405-9d8e4f73fece\" (UID: \"182bb989-7f0e-4aec-8405-9d8e4f73fece\") " Oct 05 07:06:03 crc kubenswrapper[4846]: I1005 07:06:03.081325 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/182bb989-7f0e-4aec-8405-9d8e4f73fece-ring-data-devices\") pod \"182bb989-7f0e-4aec-8405-9d8e4f73fece\" (UID: \"182bb989-7f0e-4aec-8405-9d8e4f73fece\") " Oct 05 07:06:03 crc kubenswrapper[4846]: I1005 07:06:03.081421 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t527f\" (UniqueName: \"kubernetes.io/projected/182bb989-7f0e-4aec-8405-9d8e4f73fece-kube-api-access-t527f\") pod \"182bb989-7f0e-4aec-8405-9d8e4f73fece\" (UID: \"182bb989-7f0e-4aec-8405-9d8e4f73fece\") " Oct 05 07:06:03 crc kubenswrapper[4846]: I1005 07:06:03.081670 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/182bb989-7f0e-4aec-8405-9d8e4f73fece-scripts" (OuterVolumeSpecName: "scripts") pod "182bb989-7f0e-4aec-8405-9d8e4f73fece" (UID: "182bb989-7f0e-4aec-8405-9d8e4f73fece"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 07:06:03 crc kubenswrapper[4846]: I1005 07:06:03.081980 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/182bb989-7f0e-4aec-8405-9d8e4f73fece-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "182bb989-7f0e-4aec-8405-9d8e4f73fece" (UID: "182bb989-7f0e-4aec-8405-9d8e4f73fece"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 07:06:03 crc kubenswrapper[4846]: I1005 07:06:03.082043 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/182bb989-7f0e-4aec-8405-9d8e4f73fece-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "182bb989-7f0e-4aec-8405-9d8e4f73fece" (UID: "182bb989-7f0e-4aec-8405-9d8e4f73fece"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 07:06:03 crc kubenswrapper[4846]: I1005 07:06:03.081475 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/182bb989-7f0e-4aec-8405-9d8e4f73fece-swiftconf\") pod \"182bb989-7f0e-4aec-8405-9d8e4f73fece\" (UID: \"182bb989-7f0e-4aec-8405-9d8e4f73fece\") " Oct 05 07:06:03 crc kubenswrapper[4846]: I1005 07:06:03.083097 4846 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/182bb989-7f0e-4aec-8405-9d8e4f73fece-scripts\") on node \"crc\" DevicePath \"\"" Oct 05 07:06:03 crc kubenswrapper[4846]: I1005 07:06:03.083137 4846 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/182bb989-7f0e-4aec-8405-9d8e4f73fece-etc-swift\") on node \"crc\" DevicePath \"\"" Oct 05 07:06:03 crc kubenswrapper[4846]: I1005 07:06:03.083164 4846 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/182bb989-7f0e-4aec-8405-9d8e4f73fece-ring-data-devices\") on node \"crc\" DevicePath \"\"" Oct 05 07:06:03 crc kubenswrapper[4846]: I1005 07:06:03.086137 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/182bb989-7f0e-4aec-8405-9d8e4f73fece-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "182bb989-7f0e-4aec-8405-9d8e4f73fece" (UID: "182bb989-7f0e-4aec-8405-9d8e4f73fece"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:06:03 crc kubenswrapper[4846]: I1005 07:06:03.086479 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/182bb989-7f0e-4aec-8405-9d8e4f73fece-kube-api-access-t527f" (OuterVolumeSpecName: "kube-api-access-t527f") pod "182bb989-7f0e-4aec-8405-9d8e4f73fece" (UID: "182bb989-7f0e-4aec-8405-9d8e4f73fece"). InnerVolumeSpecName "kube-api-access-t527f". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:06:03 crc kubenswrapper[4846]: I1005 07:06:03.087231 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/182bb989-7f0e-4aec-8405-9d8e4f73fece-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "182bb989-7f0e-4aec-8405-9d8e4f73fece" (UID: "182bb989-7f0e-4aec-8405-9d8e4f73fece"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:06:03 crc kubenswrapper[4846]: I1005 07:06:03.088809 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/182bb989-7f0e-4aec-8405-9d8e4f73fece-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "182bb989-7f0e-4aec-8405-9d8e4f73fece" (UID: "182bb989-7f0e-4aec-8405-9d8e4f73fece"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:06:03 crc kubenswrapper[4846]: I1005 07:06:03.185888 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t527f\" (UniqueName: \"kubernetes.io/projected/182bb989-7f0e-4aec-8405-9d8e4f73fece-kube-api-access-t527f\") on node \"crc\" DevicePath \"\"" Oct 05 07:06:03 crc kubenswrapper[4846]: I1005 07:06:03.185944 4846 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/182bb989-7f0e-4aec-8405-9d8e4f73fece-swiftconf\") on node \"crc\" DevicePath \"\"" Oct 05 07:06:03 crc kubenswrapper[4846]: I1005 07:06:03.185959 4846 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/182bb989-7f0e-4aec-8405-9d8e4f73fece-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 07:06:03 crc kubenswrapper[4846]: I1005 07:06:03.185972 4846 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/182bb989-7f0e-4aec-8405-9d8e4f73fece-dispersionconf\") on node \"crc\" DevicePath \"\"" Oct 05 07:06:03 crc kubenswrapper[4846]: I1005 07:06:03.287785 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/62d2ef82-9372-4b55-990e-30eb8e16e439-etc-swift\") pod \"swift-storage-0\" (UID: \"62d2ef82-9372-4b55-990e-30eb8e16e439\") " pod="openstack/swift-storage-0" Oct 05 07:06:03 crc kubenswrapper[4846]: E1005 07:06:03.288096 4846 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 05 07:06:03 crc kubenswrapper[4846]: E1005 07:06:03.288144 4846 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Oct 05 07:06:03 crc kubenswrapper[4846]: E1005 07:06:03.288239 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/62d2ef82-9372-4b55-990e-30eb8e16e439-etc-swift podName:62d2ef82-9372-4b55-990e-30eb8e16e439 nodeName:}" failed. No retries permitted until 2025-10-05 07:06:05.288215692 +0000 UTC m=+1087.529068467 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/62d2ef82-9372-4b55-990e-30eb8e16e439-etc-swift") pod "swift-storage-0" (UID: "62d2ef82-9372-4b55-990e-30eb8e16e439") : configmap "swift-ring-files" not found Oct 05 07:06:03 crc kubenswrapper[4846]: I1005 07:06:03.968030 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-lrj2x" Oct 05 07:06:04 crc kubenswrapper[4846]: I1005 07:06:04.044905 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-ring-rebalance-lrj2x"] Oct 05 07:06:04 crc kubenswrapper[4846]: I1005 07:06:04.050860 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/swift-ring-rebalance-lrj2x"] Oct 05 07:06:04 crc kubenswrapper[4846]: I1005 07:06:04.509836 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="182bb989-7f0e-4aec-8405-9d8e4f73fece" path="/var/lib/kubelet/pods/182bb989-7f0e-4aec-8405-9d8e4f73fece/volumes" Oct 05 07:06:05 crc kubenswrapper[4846]: I1005 07:06:05.328257 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/62d2ef82-9372-4b55-990e-30eb8e16e439-etc-swift\") pod \"swift-storage-0\" (UID: \"62d2ef82-9372-4b55-990e-30eb8e16e439\") " pod="openstack/swift-storage-0" Oct 05 07:06:05 crc kubenswrapper[4846]: E1005 07:06:05.328563 4846 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 05 07:06:05 crc kubenswrapper[4846]: E1005 07:06:05.328608 4846 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Oct 05 07:06:05 crc kubenswrapper[4846]: E1005 07:06:05.328694 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/62d2ef82-9372-4b55-990e-30eb8e16e439-etc-swift podName:62d2ef82-9372-4b55-990e-30eb8e16e439 nodeName:}" failed. No retries permitted until 2025-10-05 07:06:09.328670497 +0000 UTC m=+1091.569523272 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/62d2ef82-9372-4b55-990e-30eb8e16e439-etc-swift") pod "swift-storage-0" (UID: "62d2ef82-9372-4b55-990e-30eb8e16e439") : configmap "swift-ring-files" not found Oct 05 07:06:05 crc kubenswrapper[4846]: I1005 07:06:05.971481 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-galera-0" Oct 05 07:06:06 crc kubenswrapper[4846]: I1005 07:06:06.041756 4846 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/openstack-galera-0" podUID="d88c48f9-948e-4888-986e-e00163ec93a6" containerName="galera" probeResult="failure" output=< Oct 05 07:06:06 crc kubenswrapper[4846]: wsrep_local_state_comment (Joined) differs from Synced Oct 05 07:06:06 crc kubenswrapper[4846]: > Oct 05 07:06:07 crc kubenswrapper[4846]: I1005 07:06:07.016845 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-galera-0" Oct 05 07:06:07 crc kubenswrapper[4846]: I1005 07:06:07.921497 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-create-z9b4l"] Oct 05 07:06:07 crc kubenswrapper[4846]: I1005 07:06:07.923020 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-z9b4l" Oct 05 07:06:07 crc kubenswrapper[4846]: I1005 07:06:07.936600 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-z9b4l"] Oct 05 07:06:07 crc kubenswrapper[4846]: I1005 07:06:07.980245 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6llv4\" (UniqueName: \"kubernetes.io/projected/f8c7a428-fb8a-4992-bf65-d603b4cfc323-kube-api-access-6llv4\") pod \"keystone-db-create-z9b4l\" (UID: \"f8c7a428-fb8a-4992-bf65-d603b4cfc323\") " pod="openstack/keystone-db-create-z9b4l" Oct 05 07:06:08 crc kubenswrapper[4846]: I1005 07:06:08.082919 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6llv4\" (UniqueName: \"kubernetes.io/projected/f8c7a428-fb8a-4992-bf65-d603b4cfc323-kube-api-access-6llv4\") pod \"keystone-db-create-z9b4l\" (UID: \"f8c7a428-fb8a-4992-bf65-d603b4cfc323\") " pod="openstack/keystone-db-create-z9b4l" Oct 05 07:06:08 crc kubenswrapper[4846]: I1005 07:06:08.108284 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6llv4\" (UniqueName: \"kubernetes.io/projected/f8c7a428-fb8a-4992-bf65-d603b4cfc323-kube-api-access-6llv4\") pod \"keystone-db-create-z9b4l\" (UID: \"f8c7a428-fb8a-4992-bf65-d603b4cfc323\") " pod="openstack/keystone-db-create-z9b4l" Oct 05 07:06:08 crc kubenswrapper[4846]: I1005 07:06:08.141975 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-create-ll6h2"] Oct 05 07:06:08 crc kubenswrapper[4846]: I1005 07:06:08.143342 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-ll6h2" Oct 05 07:06:08 crc kubenswrapper[4846]: I1005 07:06:08.153282 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-ll6h2"] Oct 05 07:06:08 crc kubenswrapper[4846]: I1005 07:06:08.185232 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hhpvm\" (UniqueName: \"kubernetes.io/projected/d8234451-0d89-47e7-af4c-0263c3b6f58b-kube-api-access-hhpvm\") pod \"placement-db-create-ll6h2\" (UID: \"d8234451-0d89-47e7-af4c-0263c3b6f58b\") " pod="openstack/placement-db-create-ll6h2" Oct 05 07:06:08 crc kubenswrapper[4846]: I1005 07:06:08.247155 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-z9b4l" Oct 05 07:06:08 crc kubenswrapper[4846]: I1005 07:06:08.287542 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hhpvm\" (UniqueName: \"kubernetes.io/projected/d8234451-0d89-47e7-af4c-0263c3b6f58b-kube-api-access-hhpvm\") pod \"placement-db-create-ll6h2\" (UID: \"d8234451-0d89-47e7-af4c-0263c3b6f58b\") " pod="openstack/placement-db-create-ll6h2" Oct 05 07:06:08 crc kubenswrapper[4846]: I1005 07:06:08.306054 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hhpvm\" (UniqueName: \"kubernetes.io/projected/d8234451-0d89-47e7-af4c-0263c3b6f58b-kube-api-access-hhpvm\") pod \"placement-db-create-ll6h2\" (UID: \"d8234451-0d89-47e7-af4c-0263c3b6f58b\") " pod="openstack/placement-db-create-ll6h2" Oct 05 07:06:08 crc kubenswrapper[4846]: I1005 07:06:08.419946 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-create-sqkh4"] Oct 05 07:06:08 crc kubenswrapper[4846]: I1005 07:06:08.421878 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-sqkh4" Oct 05 07:06:08 crc kubenswrapper[4846]: I1005 07:06:08.430683 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-sqkh4"] Oct 05 07:06:08 crc kubenswrapper[4846]: I1005 07:06:08.462971 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-ll6h2" Oct 05 07:06:08 crc kubenswrapper[4846]: I1005 07:06:08.493058 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gkt64\" (UniqueName: \"kubernetes.io/projected/cfa439eb-10b4-4b11-9baa-377a45fe68d0-kube-api-access-gkt64\") pod \"glance-db-create-sqkh4\" (UID: \"cfa439eb-10b4-4b11-9baa-377a45fe68d0\") " pod="openstack/glance-db-create-sqkh4" Oct 05 07:06:08 crc kubenswrapper[4846]: I1005 07:06:08.596045 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gkt64\" (UniqueName: \"kubernetes.io/projected/cfa439eb-10b4-4b11-9baa-377a45fe68d0-kube-api-access-gkt64\") pod \"glance-db-create-sqkh4\" (UID: \"cfa439eb-10b4-4b11-9baa-377a45fe68d0\") " pod="openstack/glance-db-create-sqkh4" Oct 05 07:06:08 crc kubenswrapper[4846]: I1005 07:06:08.620902 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gkt64\" (UniqueName: \"kubernetes.io/projected/cfa439eb-10b4-4b11-9baa-377a45fe68d0-kube-api-access-gkt64\") pod \"glance-db-create-sqkh4\" (UID: \"cfa439eb-10b4-4b11-9baa-377a45fe68d0\") " pod="openstack/glance-db-create-sqkh4" Oct 05 07:06:08 crc kubenswrapper[4846]: I1005 07:06:08.742821 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-sqkh4" Oct 05 07:06:09 crc kubenswrapper[4846]: I1005 07:06:09.397815 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-cell1-galera-0" Oct 05 07:06:09 crc kubenswrapper[4846]: I1005 07:06:09.425980 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/62d2ef82-9372-4b55-990e-30eb8e16e439-etc-swift\") pod \"swift-storage-0\" (UID: \"62d2ef82-9372-4b55-990e-30eb8e16e439\") " pod="openstack/swift-storage-0" Oct 05 07:06:09 crc kubenswrapper[4846]: E1005 07:06:09.426290 4846 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 05 07:06:09 crc kubenswrapper[4846]: E1005 07:06:09.426310 4846 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Oct 05 07:06:09 crc kubenswrapper[4846]: E1005 07:06:09.426392 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/62d2ef82-9372-4b55-990e-30eb8e16e439-etc-swift podName:62d2ef82-9372-4b55-990e-30eb8e16e439 nodeName:}" failed. No retries permitted until 2025-10-05 07:06:17.426349484 +0000 UTC m=+1099.667202259 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/62d2ef82-9372-4b55-990e-30eb8e16e439-etc-swift") pod "swift-storage-0" (UID: "62d2ef82-9372-4b55-990e-30eb8e16e439") : configmap "swift-ring-files" not found Oct 05 07:06:09 crc kubenswrapper[4846]: I1005 07:06:09.511224 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6c64dcbb65-fw9xh" Oct 05 07:06:09 crc kubenswrapper[4846]: I1005 07:06:09.515485 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5d4d9f7875-29gcr" Oct 05 07:06:09 crc kubenswrapper[4846]: I1005 07:06:09.629384 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5fa7b35e-d3fd-45b6-bb8c-b7599419d53f-config\") pod \"5fa7b35e-d3fd-45b6-bb8c-b7599419d53f\" (UID: \"5fa7b35e-d3fd-45b6-bb8c-b7599419d53f\") " Oct 05 07:06:09 crc kubenswrapper[4846]: I1005 07:06:09.630189 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7p4vq\" (UniqueName: \"kubernetes.io/projected/5fa7b35e-d3fd-45b6-bb8c-b7599419d53f-kube-api-access-7p4vq\") pod \"5fa7b35e-d3fd-45b6-bb8c-b7599419d53f\" (UID: \"5fa7b35e-d3fd-45b6-bb8c-b7599419d53f\") " Oct 05 07:06:09 crc kubenswrapper[4846]: I1005 07:06:09.630270 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5fa7b35e-d3fd-45b6-bb8c-b7599419d53f-dns-svc\") pod \"5fa7b35e-d3fd-45b6-bb8c-b7599419d53f\" (UID: \"5fa7b35e-d3fd-45b6-bb8c-b7599419d53f\") " Oct 05 07:06:09 crc kubenswrapper[4846]: I1005 07:06:09.630359 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s84t5\" (UniqueName: \"kubernetes.io/projected/70abcfd4-3bfd-47c0-bb6a-03883ab21a76-kube-api-access-s84t5\") pod \"70abcfd4-3bfd-47c0-bb6a-03883ab21a76\" (UID: \"70abcfd4-3bfd-47c0-bb6a-03883ab21a76\") " Oct 05 07:06:09 crc kubenswrapper[4846]: I1005 07:06:09.630386 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/70abcfd4-3bfd-47c0-bb6a-03883ab21a76-config\") pod \"70abcfd4-3bfd-47c0-bb6a-03883ab21a76\" (UID: \"70abcfd4-3bfd-47c0-bb6a-03883ab21a76\") " Oct 05 07:06:09 crc kubenswrapper[4846]: I1005 07:06:09.630430 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/70abcfd4-3bfd-47c0-bb6a-03883ab21a76-dns-svc\") pod \"70abcfd4-3bfd-47c0-bb6a-03883ab21a76\" (UID: \"70abcfd4-3bfd-47c0-bb6a-03883ab21a76\") " Oct 05 07:06:09 crc kubenswrapper[4846]: I1005 07:06:09.630713 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5fa7b35e-d3fd-45b6-bb8c-b7599419d53f-config" (OuterVolumeSpecName: "config") pod "5fa7b35e-d3fd-45b6-bb8c-b7599419d53f" (UID: "5fa7b35e-d3fd-45b6-bb8c-b7599419d53f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 07:06:09 crc kubenswrapper[4846]: I1005 07:06:09.630980 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5fa7b35e-d3fd-45b6-bb8c-b7599419d53f-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "5fa7b35e-d3fd-45b6-bb8c-b7599419d53f" (UID: "5fa7b35e-d3fd-45b6-bb8c-b7599419d53f"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 07:06:09 crc kubenswrapper[4846]: I1005 07:06:09.631279 4846 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5fa7b35e-d3fd-45b6-bb8c-b7599419d53f-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 05 07:06:09 crc kubenswrapper[4846]: I1005 07:06:09.631299 4846 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5fa7b35e-d3fd-45b6-bb8c-b7599419d53f-config\") on node \"crc\" DevicePath \"\"" Oct 05 07:06:09 crc kubenswrapper[4846]: I1005 07:06:09.633688 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/70abcfd4-3bfd-47c0-bb6a-03883ab21a76-config" (OuterVolumeSpecName: "config") pod "70abcfd4-3bfd-47c0-bb6a-03883ab21a76" (UID: "70abcfd4-3bfd-47c0-bb6a-03883ab21a76"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 07:06:09 crc kubenswrapper[4846]: I1005 07:06:09.634351 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/70abcfd4-3bfd-47c0-bb6a-03883ab21a76-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "70abcfd4-3bfd-47c0-bb6a-03883ab21a76" (UID: "70abcfd4-3bfd-47c0-bb6a-03883ab21a76"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 07:06:09 crc kubenswrapper[4846]: I1005 07:06:09.635761 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fa7b35e-d3fd-45b6-bb8c-b7599419d53f-kube-api-access-7p4vq" (OuterVolumeSpecName: "kube-api-access-7p4vq") pod "5fa7b35e-d3fd-45b6-bb8c-b7599419d53f" (UID: "5fa7b35e-d3fd-45b6-bb8c-b7599419d53f"). InnerVolumeSpecName "kube-api-access-7p4vq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:06:09 crc kubenswrapper[4846]: I1005 07:06:09.643477 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/70abcfd4-3bfd-47c0-bb6a-03883ab21a76-kube-api-access-s84t5" (OuterVolumeSpecName: "kube-api-access-s84t5") pod "70abcfd4-3bfd-47c0-bb6a-03883ab21a76" (UID: "70abcfd4-3bfd-47c0-bb6a-03883ab21a76"). InnerVolumeSpecName "kube-api-access-s84t5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:06:09 crc kubenswrapper[4846]: I1005 07:06:09.727733 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-cell1-galera-0" Oct 05 07:06:09 crc kubenswrapper[4846]: I1005 07:06:09.733481 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s84t5\" (UniqueName: \"kubernetes.io/projected/70abcfd4-3bfd-47c0-bb6a-03883ab21a76-kube-api-access-s84t5\") on node \"crc\" DevicePath \"\"" Oct 05 07:06:09 crc kubenswrapper[4846]: I1005 07:06:09.733525 4846 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/70abcfd4-3bfd-47c0-bb6a-03883ab21a76-config\") on node \"crc\" DevicePath \"\"" Oct 05 07:06:09 crc kubenswrapper[4846]: I1005 07:06:09.733534 4846 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/70abcfd4-3bfd-47c0-bb6a-03883ab21a76-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 05 07:06:09 crc kubenswrapper[4846]: I1005 07:06:09.733548 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7p4vq\" (UniqueName: \"kubernetes.io/projected/5fa7b35e-d3fd-45b6-bb8c-b7599419d53f-kube-api-access-7p4vq\") on node \"crc\" DevicePath \"\"" Oct 05 07:06:10 crc kubenswrapper[4846]: I1005 07:06:10.034617 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5d4d9f7875-29gcr" event={"ID":"5fa7b35e-d3fd-45b6-bb8c-b7599419d53f","Type":"ContainerDied","Data":"154248f13f2fb77e6d12d6f625024a46ebd020a7b0eccf5022ac9a11905c36d5"} Oct 05 07:06:10 crc kubenswrapper[4846]: I1005 07:06:10.034639 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5d4d9f7875-29gcr" Oct 05 07:06:10 crc kubenswrapper[4846]: I1005 07:06:10.037906 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6c64dcbb65-fw9xh" Oct 05 07:06:10 crc kubenswrapper[4846]: I1005 07:06:10.037923 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6c64dcbb65-fw9xh" event={"ID":"70abcfd4-3bfd-47c0-bb6a-03883ab21a76","Type":"ContainerDied","Data":"2b50fdfe935c5c957db115433e7ab1ff933fc94008c7bac99666a100e44b97a7"} Oct 05 07:06:10 crc kubenswrapper[4846]: I1005 07:06:10.143872 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6c64dcbb65-fw9xh"] Oct 05 07:06:10 crc kubenswrapper[4846]: I1005 07:06:10.153756 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6c64dcbb65-fw9xh"] Oct 05 07:06:10 crc kubenswrapper[4846]: I1005 07:06:10.171433 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5d4d9f7875-29gcr"] Oct 05 07:06:10 crc kubenswrapper[4846]: I1005 07:06:10.174754 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5d4d9f7875-29gcr"] Oct 05 07:06:10 crc kubenswrapper[4846]: I1005 07:06:10.511131 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fa7b35e-d3fd-45b6-bb8c-b7599419d53f" path="/var/lib/kubelet/pods/5fa7b35e-d3fd-45b6-bb8c-b7599419d53f/volumes" Oct 05 07:06:10 crc kubenswrapper[4846]: I1005 07:06:10.511519 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="70abcfd4-3bfd-47c0-bb6a-03883ab21a76" path="/var/lib/kubelet/pods/70abcfd4-3bfd-47c0-bb6a-03883ab21a76/volumes" Oct 05 07:06:10 crc kubenswrapper[4846]: E1005 07:06:10.927124 4846 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7" Oct 05 07:06:10 crc kubenswrapper[4846]: E1005 07:06:10.927385 4846 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:openstack-network-exporter,Image:quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7,Command:[/app/openstack-network-exporter],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:OPENSTACK_NETWORK_EXPORTER_YAML,Value:/etc/config/openstack-network-exporter.yaml,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:ovsdb-rundir,ReadOnly:false,MountPath:/tmp,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:metrics-certs-tls-certs,ReadOnly:true,MountPath:/etc/pki/tls/certs/ovnmetrics.crt,SubPath:tls.crt,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:metrics-certs-tls-certs,ReadOnly:true,MountPath:/etc/pki/tls/private/ovnmetrics.key,SubPath:tls.key,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:metrics-certs-tls-certs,ReadOnly:true,MountPath:/etc/pki/tls/certs/ovndbca.crt,SubPath:ca.crt,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-b9m6b,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovsdbserver-nb-0_openstack(bdf95824-fa04-4a83-92cc-7e36b4d00289): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 05 07:06:10 crc kubenswrapper[4846]: E1005 07:06:10.929035 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"openstack-network-exporter\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/ovsdbserver-nb-0" podUID="bdf95824-fa04-4a83-92cc-7e36b4d00289" Oct 05 07:06:11 crc kubenswrapper[4846]: E1005 07:06:11.010188 4846 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7" Oct 05 07:06:11 crc kubenswrapper[4846]: E1005 07:06:11.010696 4846 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:openstack-network-exporter,Image:quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7,Command:[/app/openstack-network-exporter],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:OPENSTACK_NETWORK_EXPORTER_YAML,Value:/etc/config/openstack-network-exporter.yaml,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:ovsdb-rundir,ReadOnly:false,MountPath:/tmp,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:metrics-certs-tls-certs,ReadOnly:true,MountPath:/etc/pki/tls/certs/ovnmetrics.crt,SubPath:tls.crt,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:metrics-certs-tls-certs,ReadOnly:true,MountPath:/etc/pki/tls/private/ovnmetrics.key,SubPath:tls.key,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:metrics-certs-tls-certs,ReadOnly:true,MountPath:/etc/pki/tls/certs/ovndbca.crt,SubPath:ca.crt,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-qvxqv,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovsdbserver-sb-0_openstack(87e3744a-b8db-4e76-afa9-0e1129410d4c): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 05 07:06:11 crc kubenswrapper[4846]: E1005 07:06:11.012187 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"openstack-network-exporter\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/ovsdbserver-sb-0" podUID="87e3744a-b8db-4e76-afa9-0e1129410d4c" Oct 05 07:06:11 crc kubenswrapper[4846]: E1005 07:06:11.063094 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"openstack-network-exporter\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7\\\"\"" pod="openstack/ovsdbserver-sb-0" podUID="87e3744a-b8db-4e76-afa9-0e1129410d4c" Oct 05 07:06:11 crc kubenswrapper[4846]: E1005 07:06:11.064433 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"openstack-network-exporter\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7\\\"\"" pod="openstack/ovsdbserver-nb-0" podUID="bdf95824-fa04-4a83-92cc-7e36b4d00289" Oct 05 07:06:11 crc kubenswrapper[4846]: I1005 07:06:11.224546 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-58xdl"] Oct 05 07:06:11 crc kubenswrapper[4846]: W1005 07:06:11.229542 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod19f3a362_4973_4437_ac49_de71508aa4f2.slice/crio-8c40e8e43da6d24404af9690a38549cd8ea25b81acc50916f08729a15332dc27 WatchSource:0}: Error finding container 8c40e8e43da6d24404af9690a38549cd8ea25b81acc50916f08729a15332dc27: Status 404 returned error can't find the container with id 8c40e8e43da6d24404af9690a38549cd8ea25b81acc50916f08729a15332dc27 Oct 05 07:06:11 crc kubenswrapper[4846]: I1005 07:06:11.372961 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-ll6h2"] Oct 05 07:06:11 crc kubenswrapper[4846]: I1005 07:06:11.421485 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-b9gxc"] Oct 05 07:06:11 crc kubenswrapper[4846]: I1005 07:06:11.453193 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-84f55cbc7-bl827"] Oct 05 07:06:11 crc kubenswrapper[4846]: I1005 07:06:11.460131 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-sqkh4"] Oct 05 07:06:11 crc kubenswrapper[4846]: I1005 07:06:11.465876 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-z9b4l"] Oct 05 07:06:11 crc kubenswrapper[4846]: I1005 07:06:11.487354 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7bc89dc697-k6cm6"] Oct 05 07:06:11 crc kubenswrapper[4846]: I1005 07:06:11.496103 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7c559c78c7-tkv5q"] Oct 05 07:06:11 crc kubenswrapper[4846]: W1005 07:06:11.500121 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod51a08294_d363_4169_8b63_c759c3cc30f0.slice/crio-2183eeb2dac0ba1ef5254851e14b8cfa434cef36ba86d18d83b2c7647451d4db WatchSource:0}: Error finding container 2183eeb2dac0ba1ef5254851e14b8cfa434cef36ba86d18d83b2c7647451d4db: Status 404 returned error can't find the container with id 2183eeb2dac0ba1ef5254851e14b8cfa434cef36ba86d18d83b2c7647451d4db Oct 05 07:06:11 crc kubenswrapper[4846]: W1005 07:06:11.506648 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0ca13142_fb38_4f67_88f4_834bb39fcf32.slice/crio-f2b61f21d3aca7d3ab9a60ea56b98d765c89349f7f7fa0256e65131f62419815 WatchSource:0}: Error finding container f2b61f21d3aca7d3ab9a60ea56b98d765c89349f7f7fa0256e65131f62419815: Status 404 returned error can't find the container with id f2b61f21d3aca7d3ab9a60ea56b98d765c89349f7f7fa0256e65131f62419815 Oct 05 07:06:12 crc kubenswrapper[4846]: I1005 07:06:12.007812 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-0" Oct 05 07:06:12 crc kubenswrapper[4846]: I1005 07:06:12.073799 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-0" Oct 05 07:06:12 crc kubenswrapper[4846]: I1005 07:06:12.086881 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-58xdl" event={"ID":"19f3a362-4973-4437-ac49-de71508aa4f2","Type":"ContainerStarted","Data":"8c40e8e43da6d24404af9690a38549cd8ea25b81acc50916f08729a15332dc27"} Oct 05 07:06:12 crc kubenswrapper[4846]: I1005 07:06:12.089066 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-b9gxc" event={"ID":"0ffabf45-809d-43f4-902b-755111599aa8","Type":"ContainerStarted","Data":"a674a09f7ff0641659924eeebaaa8b2aec3960afb067099baad0cac94674df95"} Oct 05 07:06:12 crc kubenswrapper[4846]: I1005 07:06:12.090548 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-84f55cbc7-bl827" event={"ID":"0ca13142-fb38-4f67-88f4-834bb39fcf32","Type":"ContainerStarted","Data":"f2b61f21d3aca7d3ab9a60ea56b98d765c89349f7f7fa0256e65131f62419815"} Oct 05 07:06:12 crc kubenswrapper[4846]: I1005 07:06:12.092611 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7bc89dc697-k6cm6" event={"ID":"51a08294-d363-4169-8b63-c759c3cc30f0","Type":"ContainerStarted","Data":"2183eeb2dac0ba1ef5254851e14b8cfa434cef36ba86d18d83b2c7647451d4db"} Oct 05 07:06:12 crc kubenswrapper[4846]: I1005 07:06:12.094632 4846 generic.go:334] "Generic (PLEG): container finished" podID="d8234451-0d89-47e7-af4c-0263c3b6f58b" containerID="841de69a750baa7db33c72de16defdadd02908d58b435db7b55539b3d4710699" exitCode=0 Oct 05 07:06:12 crc kubenswrapper[4846]: I1005 07:06:12.094719 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-ll6h2" event={"ID":"d8234451-0d89-47e7-af4c-0263c3b6f58b","Type":"ContainerDied","Data":"841de69a750baa7db33c72de16defdadd02908d58b435db7b55539b3d4710699"} Oct 05 07:06:12 crc kubenswrapper[4846]: I1005 07:06:12.094744 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-ll6h2" event={"ID":"d8234451-0d89-47e7-af4c-0263c3b6f58b","Type":"ContainerStarted","Data":"fe9bf6018402422e25199252b587b8ac0551b194182d5cbfbb29f1b0714daca8"} Oct 05 07:06:12 crc kubenswrapper[4846]: I1005 07:06:12.116466 4846 generic.go:334] "Generic (PLEG): container finished" podID="cfa439eb-10b4-4b11-9baa-377a45fe68d0" containerID="1997b10872b4f4d32c524583e381e6591ecdae66bc71a4d1d18521fe506a1eb7" exitCode=0 Oct 05 07:06:12 crc kubenswrapper[4846]: I1005 07:06:12.116600 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-sqkh4" event={"ID":"cfa439eb-10b4-4b11-9baa-377a45fe68d0","Type":"ContainerDied","Data":"1997b10872b4f4d32c524583e381e6591ecdae66bc71a4d1d18521fe506a1eb7"} Oct 05 07:06:12 crc kubenswrapper[4846]: I1005 07:06:12.116638 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-sqkh4" event={"ID":"cfa439eb-10b4-4b11-9baa-377a45fe68d0","Type":"ContainerStarted","Data":"2733b30bd744b33fd15a806e2e1eab4c653d004ba78d97090111b850ef54d02d"} Oct 05 07:06:12 crc kubenswrapper[4846]: I1005 07:06:12.122370 4846 generic.go:334] "Generic (PLEG): container finished" podID="f8c7a428-fb8a-4992-bf65-d603b4cfc323" containerID="9da68a94c483e18bf05dd62a5687e31c8c93c5ae056bb98a27b899f3ba190d7e" exitCode=0 Oct 05 07:06:12 crc kubenswrapper[4846]: I1005 07:06:12.122512 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-z9b4l" event={"ID":"f8c7a428-fb8a-4992-bf65-d603b4cfc323","Type":"ContainerDied","Data":"9da68a94c483e18bf05dd62a5687e31c8c93c5ae056bb98a27b899f3ba190d7e"} Oct 05 07:06:12 crc kubenswrapper[4846]: I1005 07:06:12.122564 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-z9b4l" event={"ID":"f8c7a428-fb8a-4992-bf65-d603b4cfc323","Type":"ContainerStarted","Data":"c259458af5dbc27b015ba714d8cd86c34c1ae27ad02bac8163744ca8b3aef96c"} Oct 05 07:06:12 crc kubenswrapper[4846]: I1005 07:06:12.127564 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7c559c78c7-tkv5q" event={"ID":"727b9fbc-1de3-4e7e-ac8a-ffbbd76eac0e","Type":"ContainerStarted","Data":"167956f163cc51c70f5ab34e290194a3b2ee971d65901a3fd0cb0df422e30006"} Oct 05 07:06:12 crc kubenswrapper[4846]: I1005 07:06:12.127826 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-0" Oct 05 07:06:12 crc kubenswrapper[4846]: I1005 07:06:12.223309 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-0" Oct 05 07:06:12 crc kubenswrapper[4846]: I1005 07:06:12.450397 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-0" Oct 05 07:06:13 crc kubenswrapper[4846]: I1005 07:06:13.140770 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-58xdl" event={"ID":"19f3a362-4973-4437-ac49-de71508aa4f2","Type":"ContainerStarted","Data":"0e19580f528cd14688ae88ab26841372cd0d6d7f86e25927a1ceae0f4c254a43"} Oct 05 07:06:13 crc kubenswrapper[4846]: I1005 07:06:13.144300 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"bdf95824-fa04-4a83-92cc-7e36b4d00289","Type":"ContainerStarted","Data":"e5574edbd5d5e803159cc31cad023d93fa72c648be7c1d91c85d67e4ae1b4f70"} Oct 05 07:06:13 crc kubenswrapper[4846]: I1005 07:06:13.146340 4846 generic.go:334] "Generic (PLEG): container finished" podID="0ca13142-fb38-4f67-88f4-834bb39fcf32" containerID="aef214901a03384318e4ab0cf42ca86e8ad7c1c83fa23a6b88fc18b42c079e2b" exitCode=0 Oct 05 07:06:13 crc kubenswrapper[4846]: I1005 07:06:13.146410 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-84f55cbc7-bl827" event={"ID":"0ca13142-fb38-4f67-88f4-834bb39fcf32","Type":"ContainerDied","Data":"aef214901a03384318e4ab0cf42ca86e8ad7c1c83fa23a6b88fc18b42c079e2b"} Oct 05 07:06:13 crc kubenswrapper[4846]: I1005 07:06:13.153601 4846 generic.go:334] "Generic (PLEG): container finished" podID="51a08294-d363-4169-8b63-c759c3cc30f0" containerID="95e9c9ce0e28d42ee5f003828a17dd4fa4f47d75bece1d967a29705097f3e344" exitCode=0 Oct 05 07:06:13 crc kubenswrapper[4846]: I1005 07:06:13.153871 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7bc89dc697-k6cm6" event={"ID":"51a08294-d363-4169-8b63-c759c3cc30f0","Type":"ContainerDied","Data":"95e9c9ce0e28d42ee5f003828a17dd4fa4f47d75bece1d967a29705097f3e344"} Oct 05 07:06:13 crc kubenswrapper[4846]: I1005 07:06:13.165986 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"87e3744a-b8db-4e76-afa9-0e1129410d4c","Type":"ContainerStarted","Data":"4110f42bd835d3c1d5633761ee0d01337fb709f2d6a2517116e7ed33818b184d"} Oct 05 07:06:13 crc kubenswrapper[4846]: I1005 07:06:13.166585 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-metrics-58xdl" podStartSLOduration=17.701152157 podStartE2EDuration="18.166550759s" podCreationTimestamp="2025-10-05 07:05:55 +0000 UTC" firstStartedPulling="2025-10-05 07:06:11.231984377 +0000 UTC m=+1093.472837152" lastFinishedPulling="2025-10-05 07:06:11.697382979 +0000 UTC m=+1093.938235754" observedRunningTime="2025-10-05 07:06:13.155927126 +0000 UTC m=+1095.396779901" watchObservedRunningTime="2025-10-05 07:06:13.166550759 +0000 UTC m=+1095.407403524" Oct 05 07:06:13 crc kubenswrapper[4846]: I1005 07:06:13.168891 4846 generic.go:334] "Generic (PLEG): container finished" podID="727b9fbc-1de3-4e7e-ac8a-ffbbd76eac0e" containerID="9f9c6e343853e98f64a82284da8f22a9bf63a1fd07fa477b713b61dcb22e9bc8" exitCode=0 Oct 05 07:06:13 crc kubenswrapper[4846]: I1005 07:06:13.168989 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7c559c78c7-tkv5q" event={"ID":"727b9fbc-1de3-4e7e-ac8a-ffbbd76eac0e","Type":"ContainerDied","Data":"9f9c6e343853e98f64a82284da8f22a9bf63a1fd07fa477b713b61dcb22e9bc8"} Oct 05 07:06:13 crc kubenswrapper[4846]: I1005 07:06:13.273723 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-0" podStartSLOduration=34.087846859 podStartE2EDuration="39.273697157s" podCreationTimestamp="2025-10-05 07:05:34 +0000 UTC" firstStartedPulling="2025-10-05 07:05:46.530510743 +0000 UTC m=+1068.771363518" lastFinishedPulling="2025-10-05 07:05:51.716361041 +0000 UTC m=+1073.957213816" observedRunningTime="2025-10-05 07:06:13.263889085 +0000 UTC m=+1095.504741860" watchObservedRunningTime="2025-10-05 07:06:13.273697157 +0000 UTC m=+1095.514549932" Oct 05 07:06:13 crc kubenswrapper[4846]: I1005 07:06:13.313763 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-0" podStartSLOduration=31.227819554 podStartE2EDuration="37.313737375s" podCreationTimestamp="2025-10-05 07:05:36 +0000 UTC" firstStartedPulling="2025-10-05 07:05:46.03205754 +0000 UTC m=+1068.272910315" lastFinishedPulling="2025-10-05 07:05:52.117975361 +0000 UTC m=+1074.358828136" observedRunningTime="2025-10-05 07:06:13.292520299 +0000 UTC m=+1095.533373074" watchObservedRunningTime="2025-10-05 07:06:13.313737375 +0000 UTC m=+1095.554590140" Oct 05 07:06:13 crc kubenswrapper[4846]: I1005 07:06:13.450633 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-0" Oct 05 07:06:13 crc kubenswrapper[4846]: I1005 07:06:13.513355 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-0" Oct 05 07:06:14 crc kubenswrapper[4846]: I1005 07:06:14.256695 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-0" Oct 05 07:06:14 crc kubenswrapper[4846]: I1005 07:06:14.461823 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-northd-0"] Oct 05 07:06:14 crc kubenswrapper[4846]: I1005 07:06:14.463804 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Oct 05 07:06:14 crc kubenswrapper[4846]: I1005 07:06:14.469454 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-scripts" Oct 05 07:06:14 crc kubenswrapper[4846]: I1005 07:06:14.469751 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovnnorthd-ovnnorthd-dockercfg-z464g" Oct 05 07:06:14 crc kubenswrapper[4846]: I1005 07:06:14.469950 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovnnorthd-ovndbs" Oct 05 07:06:14 crc kubenswrapper[4846]: I1005 07:06:14.471433 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-config" Oct 05 07:06:14 crc kubenswrapper[4846]: I1005 07:06:14.480229 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Oct 05 07:06:14 crc kubenswrapper[4846]: I1005 07:06:14.554006 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/18a7be37-c8dc-4d8a-808f-2a63eb4d9692-scripts\") pod \"ovn-northd-0\" (UID: \"18a7be37-c8dc-4d8a-808f-2a63eb4d9692\") " pod="openstack/ovn-northd-0" Oct 05 07:06:14 crc kubenswrapper[4846]: I1005 07:06:14.554064 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/18a7be37-c8dc-4d8a-808f-2a63eb4d9692-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"18a7be37-c8dc-4d8a-808f-2a63eb4d9692\") " pod="openstack/ovn-northd-0" Oct 05 07:06:14 crc kubenswrapper[4846]: I1005 07:06:14.554116 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/18a7be37-c8dc-4d8a-808f-2a63eb4d9692-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"18a7be37-c8dc-4d8a-808f-2a63eb4d9692\") " pod="openstack/ovn-northd-0" Oct 05 07:06:14 crc kubenswrapper[4846]: I1005 07:06:14.554148 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/18a7be37-c8dc-4d8a-808f-2a63eb4d9692-config\") pod \"ovn-northd-0\" (UID: \"18a7be37-c8dc-4d8a-808f-2a63eb4d9692\") " pod="openstack/ovn-northd-0" Oct 05 07:06:14 crc kubenswrapper[4846]: I1005 07:06:14.554230 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/18a7be37-c8dc-4d8a-808f-2a63eb4d9692-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"18a7be37-c8dc-4d8a-808f-2a63eb4d9692\") " pod="openstack/ovn-northd-0" Oct 05 07:06:14 crc kubenswrapper[4846]: I1005 07:06:14.554264 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p8kvb\" (UniqueName: \"kubernetes.io/projected/18a7be37-c8dc-4d8a-808f-2a63eb4d9692-kube-api-access-p8kvb\") pod \"ovn-northd-0\" (UID: \"18a7be37-c8dc-4d8a-808f-2a63eb4d9692\") " pod="openstack/ovn-northd-0" Oct 05 07:06:14 crc kubenswrapper[4846]: I1005 07:06:14.554279 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/18a7be37-c8dc-4d8a-808f-2a63eb4d9692-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"18a7be37-c8dc-4d8a-808f-2a63eb4d9692\") " pod="openstack/ovn-northd-0" Oct 05 07:06:14 crc kubenswrapper[4846]: I1005 07:06:14.656322 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/18a7be37-c8dc-4d8a-808f-2a63eb4d9692-config\") pod \"ovn-northd-0\" (UID: \"18a7be37-c8dc-4d8a-808f-2a63eb4d9692\") " pod="openstack/ovn-northd-0" Oct 05 07:06:14 crc kubenswrapper[4846]: I1005 07:06:14.656440 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/18a7be37-c8dc-4d8a-808f-2a63eb4d9692-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"18a7be37-c8dc-4d8a-808f-2a63eb4d9692\") " pod="openstack/ovn-northd-0" Oct 05 07:06:14 crc kubenswrapper[4846]: I1005 07:06:14.656496 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p8kvb\" (UniqueName: \"kubernetes.io/projected/18a7be37-c8dc-4d8a-808f-2a63eb4d9692-kube-api-access-p8kvb\") pod \"ovn-northd-0\" (UID: \"18a7be37-c8dc-4d8a-808f-2a63eb4d9692\") " pod="openstack/ovn-northd-0" Oct 05 07:06:14 crc kubenswrapper[4846]: I1005 07:06:14.656522 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/18a7be37-c8dc-4d8a-808f-2a63eb4d9692-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"18a7be37-c8dc-4d8a-808f-2a63eb4d9692\") " pod="openstack/ovn-northd-0" Oct 05 07:06:14 crc kubenswrapper[4846]: I1005 07:06:14.656690 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/18a7be37-c8dc-4d8a-808f-2a63eb4d9692-scripts\") pod \"ovn-northd-0\" (UID: \"18a7be37-c8dc-4d8a-808f-2a63eb4d9692\") " pod="openstack/ovn-northd-0" Oct 05 07:06:14 crc kubenswrapper[4846]: I1005 07:06:14.656714 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/18a7be37-c8dc-4d8a-808f-2a63eb4d9692-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"18a7be37-c8dc-4d8a-808f-2a63eb4d9692\") " pod="openstack/ovn-northd-0" Oct 05 07:06:14 crc kubenswrapper[4846]: I1005 07:06:14.656787 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/18a7be37-c8dc-4d8a-808f-2a63eb4d9692-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"18a7be37-c8dc-4d8a-808f-2a63eb4d9692\") " pod="openstack/ovn-northd-0" Oct 05 07:06:14 crc kubenswrapper[4846]: I1005 07:06:14.657902 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/18a7be37-c8dc-4d8a-808f-2a63eb4d9692-config\") pod \"ovn-northd-0\" (UID: \"18a7be37-c8dc-4d8a-808f-2a63eb4d9692\") " pod="openstack/ovn-northd-0" Oct 05 07:06:14 crc kubenswrapper[4846]: I1005 07:06:14.658020 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/18a7be37-c8dc-4d8a-808f-2a63eb4d9692-scripts\") pod \"ovn-northd-0\" (UID: \"18a7be37-c8dc-4d8a-808f-2a63eb4d9692\") " pod="openstack/ovn-northd-0" Oct 05 07:06:14 crc kubenswrapper[4846]: I1005 07:06:14.658502 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/18a7be37-c8dc-4d8a-808f-2a63eb4d9692-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"18a7be37-c8dc-4d8a-808f-2a63eb4d9692\") " pod="openstack/ovn-northd-0" Oct 05 07:06:14 crc kubenswrapper[4846]: I1005 07:06:14.667696 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/18a7be37-c8dc-4d8a-808f-2a63eb4d9692-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"18a7be37-c8dc-4d8a-808f-2a63eb4d9692\") " pod="openstack/ovn-northd-0" Oct 05 07:06:14 crc kubenswrapper[4846]: I1005 07:06:14.672632 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/18a7be37-c8dc-4d8a-808f-2a63eb4d9692-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"18a7be37-c8dc-4d8a-808f-2a63eb4d9692\") " pod="openstack/ovn-northd-0" Oct 05 07:06:14 crc kubenswrapper[4846]: I1005 07:06:14.684483 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/18a7be37-c8dc-4d8a-808f-2a63eb4d9692-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"18a7be37-c8dc-4d8a-808f-2a63eb4d9692\") " pod="openstack/ovn-northd-0" Oct 05 07:06:14 crc kubenswrapper[4846]: I1005 07:06:14.688340 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p8kvb\" (UniqueName: \"kubernetes.io/projected/18a7be37-c8dc-4d8a-808f-2a63eb4d9692-kube-api-access-p8kvb\") pod \"ovn-northd-0\" (UID: \"18a7be37-c8dc-4d8a-808f-2a63eb4d9692\") " pod="openstack/ovn-northd-0" Oct 05 07:06:14 crc kubenswrapper[4846]: I1005 07:06:14.809021 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Oct 05 07:06:16 crc kubenswrapper[4846]: I1005 07:06:16.917056 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-ll6h2" Oct 05 07:06:16 crc kubenswrapper[4846]: I1005 07:06:16.942903 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7c559c78c7-tkv5q" Oct 05 07:06:16 crc kubenswrapper[4846]: I1005 07:06:16.960683 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-sqkh4" Oct 05 07:06:17 crc kubenswrapper[4846]: I1005 07:06:17.014010 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gkt64\" (UniqueName: \"kubernetes.io/projected/cfa439eb-10b4-4b11-9baa-377a45fe68d0-kube-api-access-gkt64\") pod \"cfa439eb-10b4-4b11-9baa-377a45fe68d0\" (UID: \"cfa439eb-10b4-4b11-9baa-377a45fe68d0\") " Oct 05 07:06:17 crc kubenswrapper[4846]: I1005 07:06:17.014090 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zcsj9\" (UniqueName: \"kubernetes.io/projected/727b9fbc-1de3-4e7e-ac8a-ffbbd76eac0e-kube-api-access-zcsj9\") pod \"727b9fbc-1de3-4e7e-ac8a-ffbbd76eac0e\" (UID: \"727b9fbc-1de3-4e7e-ac8a-ffbbd76eac0e\") " Oct 05 07:06:17 crc kubenswrapper[4846]: I1005 07:06:17.014163 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/727b9fbc-1de3-4e7e-ac8a-ffbbd76eac0e-ovsdbserver-sb\") pod \"727b9fbc-1de3-4e7e-ac8a-ffbbd76eac0e\" (UID: \"727b9fbc-1de3-4e7e-ac8a-ffbbd76eac0e\") " Oct 05 07:06:17 crc kubenswrapper[4846]: I1005 07:06:17.014274 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hhpvm\" (UniqueName: \"kubernetes.io/projected/d8234451-0d89-47e7-af4c-0263c3b6f58b-kube-api-access-hhpvm\") pod \"d8234451-0d89-47e7-af4c-0263c3b6f58b\" (UID: \"d8234451-0d89-47e7-af4c-0263c3b6f58b\") " Oct 05 07:06:17 crc kubenswrapper[4846]: I1005 07:06:17.014399 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/727b9fbc-1de3-4e7e-ac8a-ffbbd76eac0e-dns-svc\") pod \"727b9fbc-1de3-4e7e-ac8a-ffbbd76eac0e\" (UID: \"727b9fbc-1de3-4e7e-ac8a-ffbbd76eac0e\") " Oct 05 07:06:17 crc kubenswrapper[4846]: I1005 07:06:17.014476 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/727b9fbc-1de3-4e7e-ac8a-ffbbd76eac0e-config\") pod \"727b9fbc-1de3-4e7e-ac8a-ffbbd76eac0e\" (UID: \"727b9fbc-1de3-4e7e-ac8a-ffbbd76eac0e\") " Oct 05 07:06:17 crc kubenswrapper[4846]: I1005 07:06:17.029280 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-z9b4l" Oct 05 07:06:17 crc kubenswrapper[4846]: I1005 07:06:17.031774 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d8234451-0d89-47e7-af4c-0263c3b6f58b-kube-api-access-hhpvm" (OuterVolumeSpecName: "kube-api-access-hhpvm") pod "d8234451-0d89-47e7-af4c-0263c3b6f58b" (UID: "d8234451-0d89-47e7-af4c-0263c3b6f58b"). InnerVolumeSpecName "kube-api-access-hhpvm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:06:17 crc kubenswrapper[4846]: I1005 07:06:17.032582 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cfa439eb-10b4-4b11-9baa-377a45fe68d0-kube-api-access-gkt64" (OuterVolumeSpecName: "kube-api-access-gkt64") pod "cfa439eb-10b4-4b11-9baa-377a45fe68d0" (UID: "cfa439eb-10b4-4b11-9baa-377a45fe68d0"). InnerVolumeSpecName "kube-api-access-gkt64". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:06:17 crc kubenswrapper[4846]: I1005 07:06:17.037312 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/727b9fbc-1de3-4e7e-ac8a-ffbbd76eac0e-kube-api-access-zcsj9" (OuterVolumeSpecName: "kube-api-access-zcsj9") pod "727b9fbc-1de3-4e7e-ac8a-ffbbd76eac0e" (UID: "727b9fbc-1de3-4e7e-ac8a-ffbbd76eac0e"). InnerVolumeSpecName "kube-api-access-zcsj9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:06:17 crc kubenswrapper[4846]: I1005 07:06:17.103431 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/727b9fbc-1de3-4e7e-ac8a-ffbbd76eac0e-config" (OuterVolumeSpecName: "config") pod "727b9fbc-1de3-4e7e-ac8a-ffbbd76eac0e" (UID: "727b9fbc-1de3-4e7e-ac8a-ffbbd76eac0e"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 07:06:17 crc kubenswrapper[4846]: I1005 07:06:17.104289 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/727b9fbc-1de3-4e7e-ac8a-ffbbd76eac0e-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "727b9fbc-1de3-4e7e-ac8a-ffbbd76eac0e" (UID: "727b9fbc-1de3-4e7e-ac8a-ffbbd76eac0e"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 07:06:17 crc kubenswrapper[4846]: I1005 07:06:17.109747 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/727b9fbc-1de3-4e7e-ac8a-ffbbd76eac0e-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "727b9fbc-1de3-4e7e-ac8a-ffbbd76eac0e" (UID: "727b9fbc-1de3-4e7e-ac8a-ffbbd76eac0e"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 07:06:17 crc kubenswrapper[4846]: I1005 07:06:17.117212 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6llv4\" (UniqueName: \"kubernetes.io/projected/f8c7a428-fb8a-4992-bf65-d603b4cfc323-kube-api-access-6llv4\") pod \"f8c7a428-fb8a-4992-bf65-d603b4cfc323\" (UID: \"f8c7a428-fb8a-4992-bf65-d603b4cfc323\") " Oct 05 07:06:17 crc kubenswrapper[4846]: I1005 07:06:17.117999 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gkt64\" (UniqueName: \"kubernetes.io/projected/cfa439eb-10b4-4b11-9baa-377a45fe68d0-kube-api-access-gkt64\") on node \"crc\" DevicePath \"\"" Oct 05 07:06:17 crc kubenswrapper[4846]: I1005 07:06:17.118021 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zcsj9\" (UniqueName: \"kubernetes.io/projected/727b9fbc-1de3-4e7e-ac8a-ffbbd76eac0e-kube-api-access-zcsj9\") on node \"crc\" DevicePath \"\"" Oct 05 07:06:17 crc kubenswrapper[4846]: I1005 07:06:17.118034 4846 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/727b9fbc-1de3-4e7e-ac8a-ffbbd76eac0e-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 05 07:06:17 crc kubenswrapper[4846]: I1005 07:06:17.118049 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hhpvm\" (UniqueName: \"kubernetes.io/projected/d8234451-0d89-47e7-af4c-0263c3b6f58b-kube-api-access-hhpvm\") on node \"crc\" DevicePath \"\"" Oct 05 07:06:17 crc kubenswrapper[4846]: I1005 07:06:17.118062 4846 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/727b9fbc-1de3-4e7e-ac8a-ffbbd76eac0e-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 05 07:06:17 crc kubenswrapper[4846]: I1005 07:06:17.118074 4846 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/727b9fbc-1de3-4e7e-ac8a-ffbbd76eac0e-config\") on node \"crc\" DevicePath \"\"" Oct 05 07:06:17 crc kubenswrapper[4846]: I1005 07:06:17.121305 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f8c7a428-fb8a-4992-bf65-d603b4cfc323-kube-api-access-6llv4" (OuterVolumeSpecName: "kube-api-access-6llv4") pod "f8c7a428-fb8a-4992-bf65-d603b4cfc323" (UID: "f8c7a428-fb8a-4992-bf65-d603b4cfc323"). InnerVolumeSpecName "kube-api-access-6llv4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:06:17 crc kubenswrapper[4846]: I1005 07:06:17.221828 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6llv4\" (UniqueName: \"kubernetes.io/projected/f8c7a428-fb8a-4992-bf65-d603b4cfc323-kube-api-access-6llv4\") on node \"crc\" DevicePath \"\"" Oct 05 07:06:17 crc kubenswrapper[4846]: I1005 07:06:17.230645 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-z9b4l" event={"ID":"f8c7a428-fb8a-4992-bf65-d603b4cfc323","Type":"ContainerDied","Data":"c259458af5dbc27b015ba714d8cd86c34c1ae27ad02bac8163744ca8b3aef96c"} Oct 05 07:06:17 crc kubenswrapper[4846]: I1005 07:06:17.230733 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c259458af5dbc27b015ba714d8cd86c34c1ae27ad02bac8163744ca8b3aef96c" Oct 05 07:06:17 crc kubenswrapper[4846]: I1005 07:06:17.231016 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-z9b4l" Oct 05 07:06:17 crc kubenswrapper[4846]: I1005 07:06:17.232851 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7c559c78c7-tkv5q" event={"ID":"727b9fbc-1de3-4e7e-ac8a-ffbbd76eac0e","Type":"ContainerDied","Data":"167956f163cc51c70f5ab34e290194a3b2ee971d65901a3fd0cb0df422e30006"} Oct 05 07:06:17 crc kubenswrapper[4846]: I1005 07:06:17.232914 4846 scope.go:117] "RemoveContainer" containerID="9f9c6e343853e98f64a82284da8f22a9bf63a1fd07fa477b713b61dcb22e9bc8" Oct 05 07:06:17 crc kubenswrapper[4846]: I1005 07:06:17.232974 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7c559c78c7-tkv5q" Oct 05 07:06:17 crc kubenswrapper[4846]: I1005 07:06:17.236773 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-b9gxc" event={"ID":"0ffabf45-809d-43f4-902b-755111599aa8","Type":"ContainerStarted","Data":"1026ae5247bafab34579d4dd5c0cd5d8bd570d34ab240d0ce80bdef49ddaa60a"} Oct 05 07:06:17 crc kubenswrapper[4846]: I1005 07:06:17.241975 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-84f55cbc7-bl827" event={"ID":"0ca13142-fb38-4f67-88f4-834bb39fcf32","Type":"ContainerStarted","Data":"a9c3b90330182e62ece7422bd451742b02796c1046c7782b459456cc7302756b"} Oct 05 07:06:17 crc kubenswrapper[4846]: I1005 07:06:17.242286 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-84f55cbc7-bl827" Oct 05 07:06:17 crc kubenswrapper[4846]: I1005 07:06:17.261772 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7bc89dc697-k6cm6" event={"ID":"51a08294-d363-4169-8b63-c759c3cc30f0","Type":"ContainerStarted","Data":"df24c9bed2030762b841c88e3a16d298b74f8634a1c567708c75f4b65a4066a5"} Oct 05 07:06:17 crc kubenswrapper[4846]: I1005 07:06:17.262971 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7bc89dc697-k6cm6" Oct 05 07:06:17 crc kubenswrapper[4846]: I1005 07:06:17.265307 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-ring-rebalance-b9gxc" podStartSLOduration=9.972460544 podStartE2EDuration="15.265275455s" podCreationTimestamp="2025-10-05 07:06:02 +0000 UTC" firstStartedPulling="2025-10-05 07:06:11.489225367 +0000 UTC m=+1093.730078142" lastFinishedPulling="2025-10-05 07:06:16.782040288 +0000 UTC m=+1099.022893053" observedRunningTime="2025-10-05 07:06:17.257145658 +0000 UTC m=+1099.497998433" watchObservedRunningTime="2025-10-05 07:06:17.265275455 +0000 UTC m=+1099.506128250" Oct 05 07:06:17 crc kubenswrapper[4846]: I1005 07:06:17.272085 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-ll6h2" event={"ID":"d8234451-0d89-47e7-af4c-0263c3b6f58b","Type":"ContainerDied","Data":"fe9bf6018402422e25199252b587b8ac0551b194182d5cbfbb29f1b0714daca8"} Oct 05 07:06:17 crc kubenswrapper[4846]: I1005 07:06:17.272286 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fe9bf6018402422e25199252b587b8ac0551b194182d5cbfbb29f1b0714daca8" Oct 05 07:06:17 crc kubenswrapper[4846]: I1005 07:06:17.272538 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-ll6h2" Oct 05 07:06:17 crc kubenswrapper[4846]: I1005 07:06:17.276627 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-sqkh4" event={"ID":"cfa439eb-10b4-4b11-9baa-377a45fe68d0","Type":"ContainerDied","Data":"2733b30bd744b33fd15a806e2e1eab4c653d004ba78d97090111b850ef54d02d"} Oct 05 07:06:17 crc kubenswrapper[4846]: I1005 07:06:17.276694 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2733b30bd744b33fd15a806e2e1eab4c653d004ba78d97090111b850ef54d02d" Oct 05 07:06:17 crc kubenswrapper[4846]: I1005 07:06:17.276670 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-sqkh4" Oct 05 07:06:17 crc kubenswrapper[4846]: I1005 07:06:17.291222 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-84f55cbc7-bl827" podStartSLOduration=20.873641991 podStartE2EDuration="21.291154655s" podCreationTimestamp="2025-10-05 07:05:56 +0000 UTC" firstStartedPulling="2025-10-05 07:06:11.512010115 +0000 UTC m=+1093.752862890" lastFinishedPulling="2025-10-05 07:06:11.929522759 +0000 UTC m=+1094.170375554" observedRunningTime="2025-10-05 07:06:17.283915382 +0000 UTC m=+1099.524768157" watchObservedRunningTime="2025-10-05 07:06:17.291154655 +0000 UTC m=+1099.532007470" Oct 05 07:06:17 crc kubenswrapper[4846]: I1005 07:06:17.306846 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Oct 05 07:06:17 crc kubenswrapper[4846]: W1005 07:06:17.318950 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod18a7be37_c8dc_4d8a_808f_2a63eb4d9692.slice/crio-4c157e28a99b6e1bcdc1f682bb2c7be41ccc019fcb6d08ba1de9c45e47d39e16 WatchSource:0}: Error finding container 4c157e28a99b6e1bcdc1f682bb2c7be41ccc019fcb6d08ba1de9c45e47d39e16: Status 404 returned error can't find the container with id 4c157e28a99b6e1bcdc1f682bb2c7be41ccc019fcb6d08ba1de9c45e47d39e16 Oct 05 07:06:17 crc kubenswrapper[4846]: I1005 07:06:17.330725 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7c559c78c7-tkv5q"] Oct 05 07:06:17 crc kubenswrapper[4846]: I1005 07:06:17.340047 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7c559c78c7-tkv5q"] Oct 05 07:06:17 crc kubenswrapper[4846]: I1005 07:06:17.363326 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7bc89dc697-k6cm6" podStartSLOduration=16.868889684 podStartE2EDuration="17.363304869s" podCreationTimestamp="2025-10-05 07:06:00 +0000 UTC" firstStartedPulling="2025-10-05 07:06:11.505331617 +0000 UTC m=+1093.746184402" lastFinishedPulling="2025-10-05 07:06:11.999746812 +0000 UTC m=+1094.240599587" observedRunningTime="2025-10-05 07:06:17.359799346 +0000 UTC m=+1099.600652121" watchObservedRunningTime="2025-10-05 07:06:17.363304869 +0000 UTC m=+1099.604157644" Oct 05 07:06:17 crc kubenswrapper[4846]: I1005 07:06:17.431324 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/62d2ef82-9372-4b55-990e-30eb8e16e439-etc-swift\") pod \"swift-storage-0\" (UID: \"62d2ef82-9372-4b55-990e-30eb8e16e439\") " pod="openstack/swift-storage-0" Oct 05 07:06:17 crc kubenswrapper[4846]: E1005 07:06:17.431594 4846 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 05 07:06:17 crc kubenswrapper[4846]: E1005 07:06:17.431640 4846 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Oct 05 07:06:17 crc kubenswrapper[4846]: E1005 07:06:17.431712 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/62d2ef82-9372-4b55-990e-30eb8e16e439-etc-swift podName:62d2ef82-9372-4b55-990e-30eb8e16e439 nodeName:}" failed. No retries permitted until 2025-10-05 07:06:33.431687823 +0000 UTC m=+1115.672540598 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/62d2ef82-9372-4b55-990e-30eb8e16e439-etc-swift") pod "swift-storage-0" (UID: "62d2ef82-9372-4b55-990e-30eb8e16e439") : configmap "swift-ring-files" not found Oct 05 07:06:18 crc kubenswrapper[4846]: I1005 07:06:18.291973 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"18a7be37-c8dc-4d8a-808f-2a63eb4d9692","Type":"ContainerStarted","Data":"4c157e28a99b6e1bcdc1f682bb2c7be41ccc019fcb6d08ba1de9c45e47d39e16"} Oct 05 07:06:18 crc kubenswrapper[4846]: I1005 07:06:18.513502 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="727b9fbc-1de3-4e7e-ac8a-ffbbd76eac0e" path="/var/lib/kubelet/pods/727b9fbc-1de3-4e7e-ac8a-ffbbd76eac0e/volumes" Oct 05 07:06:19 crc kubenswrapper[4846]: I1005 07:06:19.334963 4846 generic.go:334] "Generic (PLEG): container finished" podID="35261dbc-cabe-4bf0-88f6-b8613ddec0f1" containerID="9919ee8b928899c8295fa6d8951cffdd5cffd78c7de5c1e7b1a230ab2550fbf6" exitCode=0 Oct 05 07:06:19 crc kubenswrapper[4846]: I1005 07:06:19.335094 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"35261dbc-cabe-4bf0-88f6-b8613ddec0f1","Type":"ContainerDied","Data":"9919ee8b928899c8295fa6d8951cffdd5cffd78c7de5c1e7b1a230ab2550fbf6"} Oct 05 07:06:19 crc kubenswrapper[4846]: I1005 07:06:19.341771 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"18a7be37-c8dc-4d8a-808f-2a63eb4d9692","Type":"ContainerStarted","Data":"05e6ade463feffe3d2a4a45255b934b5f81dd79a5b4cd4809663d8a570eb260c"} Oct 05 07:06:19 crc kubenswrapper[4846]: I1005 07:06:19.341878 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"18a7be37-c8dc-4d8a-808f-2a63eb4d9692","Type":"ContainerStarted","Data":"4351908cf00eea1071aca5080f9d536de25e56a78e5d0713bec8f51caab84ecb"} Oct 05 07:06:19 crc kubenswrapper[4846]: I1005 07:06:19.416876 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-northd-0" podStartSLOduration=4.306940304 podStartE2EDuration="5.416846183s" podCreationTimestamp="2025-10-05 07:06:14 +0000 UTC" firstStartedPulling="2025-10-05 07:06:17.32884347 +0000 UTC m=+1099.569696245" lastFinishedPulling="2025-10-05 07:06:18.438749349 +0000 UTC m=+1100.679602124" observedRunningTime="2025-10-05 07:06:19.402538532 +0000 UTC m=+1101.643391347" watchObservedRunningTime="2025-10-05 07:06:19.416846183 +0000 UTC m=+1101.657698968" Oct 05 07:06:19 crc kubenswrapper[4846]: I1005 07:06:19.811260 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-northd-0" Oct 05 07:06:20 crc kubenswrapper[4846]: I1005 07:06:20.356567 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"35261dbc-cabe-4bf0-88f6-b8613ddec0f1","Type":"ContainerStarted","Data":"0b472a4ecaa620dd0581763afea4046cc8d15b55265470a90367ed2c9f54e39c"} Oct 05 07:06:20 crc kubenswrapper[4846]: I1005 07:06:20.356847 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Oct 05 07:06:20 crc kubenswrapper[4846]: I1005 07:06:20.358432 4846 generic.go:334] "Generic (PLEG): container finished" podID="626e6bce-7422-4fbd-b978-dfa5cb93de34" containerID="516d1ce74e479dfd0f7702334c1c73ac78a55e7ab16e082869f8bbbabc5a6c16" exitCode=0 Oct 05 07:06:20 crc kubenswrapper[4846]: I1005 07:06:20.358517 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"626e6bce-7422-4fbd-b978-dfa5cb93de34","Type":"ContainerDied","Data":"516d1ce74e479dfd0f7702334c1c73ac78a55e7ab16e082869f8bbbabc5a6c16"} Oct 05 07:06:20 crc kubenswrapper[4846]: I1005 07:06:20.420081 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=-9223371979.434752 podStartE2EDuration="57.420023487s" podCreationTimestamp="2025-10-05 07:05:23 +0000 UTC" firstStartedPulling="2025-10-05 07:05:25.991019272 +0000 UTC m=+1048.231872057" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 07:06:20.401140273 +0000 UTC m=+1102.641993058" watchObservedRunningTime="2025-10-05 07:06:20.420023487 +0000 UTC m=+1102.660876262" Oct 05 07:06:21 crc kubenswrapper[4846]: I1005 07:06:21.371574 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"626e6bce-7422-4fbd-b978-dfa5cb93de34","Type":"ContainerStarted","Data":"c63a03c8ae2116953b605cde504ae09f678d59deda18e41527a386da67ee2ce4"} Oct 05 07:06:21 crc kubenswrapper[4846]: I1005 07:06:21.373778 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Oct 05 07:06:21 crc kubenswrapper[4846]: I1005 07:06:21.415495 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=38.558391118 podStartE2EDuration="57.415463582s" podCreationTimestamp="2025-10-05 07:05:24 +0000 UTC" firstStartedPulling="2025-10-05 07:05:26.498971828 +0000 UTC m=+1048.739824603" lastFinishedPulling="2025-10-05 07:05:45.356044282 +0000 UTC m=+1067.596897067" observedRunningTime="2025-10-05 07:06:21.411665481 +0000 UTC m=+1103.652518256" watchObservedRunningTime="2025-10-05 07:06:21.415463582 +0000 UTC m=+1103.656316357" Oct 05 07:06:22 crc kubenswrapper[4846]: I1005 07:06:22.927543 4846 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-4kqnw" podUID="169918d8-597e-4ef4-b00c-fb14c4d0347f" containerName="ovn-controller" probeResult="failure" output=< Oct 05 07:06:22 crc kubenswrapper[4846]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Oct 05 07:06:22 crc kubenswrapper[4846]: > Oct 05 07:06:23 crc kubenswrapper[4846]: I1005 07:06:23.325353 4846 patch_prober.go:28] interesting pod/machine-config-daemon-fscvf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 07:06:23 crc kubenswrapper[4846]: I1005 07:06:23.325430 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 07:06:25 crc kubenswrapper[4846]: I1005 07:06:25.411581 4846 generic.go:334] "Generic (PLEG): container finished" podID="0ffabf45-809d-43f4-902b-755111599aa8" containerID="1026ae5247bafab34579d4dd5c0cd5d8bd570d34ab240d0ce80bdef49ddaa60a" exitCode=0 Oct 05 07:06:25 crc kubenswrapper[4846]: I1005 07:06:25.411666 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-b9gxc" event={"ID":"0ffabf45-809d-43f4-902b-755111599aa8","Type":"ContainerDied","Data":"1026ae5247bafab34579d4dd5c0cd5d8bd570d34ab240d0ce80bdef49ddaa60a"} Oct 05 07:06:25 crc kubenswrapper[4846]: I1005 07:06:25.772452 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-7bc89dc697-k6cm6" Oct 05 07:06:25 crc kubenswrapper[4846]: I1005 07:06:25.841049 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-84f55cbc7-bl827"] Oct 05 07:06:25 crc kubenswrapper[4846]: I1005 07:06:25.841349 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-84f55cbc7-bl827" podUID="0ca13142-fb38-4f67-88f4-834bb39fcf32" containerName="dnsmasq-dns" containerID="cri-o://a9c3b90330182e62ece7422bd451742b02796c1046c7782b459456cc7302756b" gracePeriod=10 Oct 05 07:06:25 crc kubenswrapper[4846]: I1005 07:06:25.847771 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-84f55cbc7-bl827" Oct 05 07:06:26 crc kubenswrapper[4846]: I1005 07:06:26.344596 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-84f55cbc7-bl827" Oct 05 07:06:26 crc kubenswrapper[4846]: I1005 07:06:26.429032 4846 generic.go:334] "Generic (PLEG): container finished" podID="0ca13142-fb38-4f67-88f4-834bb39fcf32" containerID="a9c3b90330182e62ece7422bd451742b02796c1046c7782b459456cc7302756b" exitCode=0 Oct 05 07:06:26 crc kubenswrapper[4846]: I1005 07:06:26.429404 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-84f55cbc7-bl827" Oct 05 07:06:26 crc kubenswrapper[4846]: I1005 07:06:26.430274 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-84f55cbc7-bl827" event={"ID":"0ca13142-fb38-4f67-88f4-834bb39fcf32","Type":"ContainerDied","Data":"a9c3b90330182e62ece7422bd451742b02796c1046c7782b459456cc7302756b"} Oct 05 07:06:26 crc kubenswrapper[4846]: I1005 07:06:26.430320 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-84f55cbc7-bl827" event={"ID":"0ca13142-fb38-4f67-88f4-834bb39fcf32","Type":"ContainerDied","Data":"f2b61f21d3aca7d3ab9a60ea56b98d765c89349f7f7fa0256e65131f62419815"} Oct 05 07:06:26 crc kubenswrapper[4846]: I1005 07:06:26.430347 4846 scope.go:117] "RemoveContainer" containerID="a9c3b90330182e62ece7422bd451742b02796c1046c7782b459456cc7302756b" Oct 05 07:06:26 crc kubenswrapper[4846]: I1005 07:06:26.455978 4846 scope.go:117] "RemoveContainer" containerID="aef214901a03384318e4ab0cf42ca86e8ad7c1c83fa23a6b88fc18b42c079e2b" Oct 05 07:06:26 crc kubenswrapper[4846]: I1005 07:06:26.497603 4846 scope.go:117] "RemoveContainer" containerID="a9c3b90330182e62ece7422bd451742b02796c1046c7782b459456cc7302756b" Oct 05 07:06:26 crc kubenswrapper[4846]: E1005 07:06:26.498440 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a9c3b90330182e62ece7422bd451742b02796c1046c7782b459456cc7302756b\": container with ID starting with a9c3b90330182e62ece7422bd451742b02796c1046c7782b459456cc7302756b not found: ID does not exist" containerID="a9c3b90330182e62ece7422bd451742b02796c1046c7782b459456cc7302756b" Oct 05 07:06:26 crc kubenswrapper[4846]: I1005 07:06:26.498868 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a9c3b90330182e62ece7422bd451742b02796c1046c7782b459456cc7302756b"} err="failed to get container status \"a9c3b90330182e62ece7422bd451742b02796c1046c7782b459456cc7302756b\": rpc error: code = NotFound desc = could not find container \"a9c3b90330182e62ece7422bd451742b02796c1046c7782b459456cc7302756b\": container with ID starting with a9c3b90330182e62ece7422bd451742b02796c1046c7782b459456cc7302756b not found: ID does not exist" Oct 05 07:06:26 crc kubenswrapper[4846]: I1005 07:06:26.498910 4846 scope.go:117] "RemoveContainer" containerID="aef214901a03384318e4ab0cf42ca86e8ad7c1c83fa23a6b88fc18b42c079e2b" Oct 05 07:06:26 crc kubenswrapper[4846]: E1005 07:06:26.499647 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"aef214901a03384318e4ab0cf42ca86e8ad7c1c83fa23a6b88fc18b42c079e2b\": container with ID starting with aef214901a03384318e4ab0cf42ca86e8ad7c1c83fa23a6b88fc18b42c079e2b not found: ID does not exist" containerID="aef214901a03384318e4ab0cf42ca86e8ad7c1c83fa23a6b88fc18b42c079e2b" Oct 05 07:06:26 crc kubenswrapper[4846]: I1005 07:06:26.499690 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aef214901a03384318e4ab0cf42ca86e8ad7c1c83fa23a6b88fc18b42c079e2b"} err="failed to get container status \"aef214901a03384318e4ab0cf42ca86e8ad7c1c83fa23a6b88fc18b42c079e2b\": rpc error: code = NotFound desc = could not find container \"aef214901a03384318e4ab0cf42ca86e8ad7c1c83fa23a6b88fc18b42c079e2b\": container with ID starting with aef214901a03384318e4ab0cf42ca86e8ad7c1c83fa23a6b88fc18b42c079e2b not found: ID does not exist" Oct 05 07:06:26 crc kubenswrapper[4846]: I1005 07:06:26.534424 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0ca13142-fb38-4f67-88f4-834bb39fcf32-ovsdbserver-nb\") pod \"0ca13142-fb38-4f67-88f4-834bb39fcf32\" (UID: \"0ca13142-fb38-4f67-88f4-834bb39fcf32\") " Oct 05 07:06:26 crc kubenswrapper[4846]: I1005 07:06:26.534575 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0ca13142-fb38-4f67-88f4-834bb39fcf32-ovsdbserver-sb\") pod \"0ca13142-fb38-4f67-88f4-834bb39fcf32\" (UID: \"0ca13142-fb38-4f67-88f4-834bb39fcf32\") " Oct 05 07:06:26 crc kubenswrapper[4846]: I1005 07:06:26.534685 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0ca13142-fb38-4f67-88f4-834bb39fcf32-config\") pod \"0ca13142-fb38-4f67-88f4-834bb39fcf32\" (UID: \"0ca13142-fb38-4f67-88f4-834bb39fcf32\") " Oct 05 07:06:26 crc kubenswrapper[4846]: I1005 07:06:26.534802 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cxlh2\" (UniqueName: \"kubernetes.io/projected/0ca13142-fb38-4f67-88f4-834bb39fcf32-kube-api-access-cxlh2\") pod \"0ca13142-fb38-4f67-88f4-834bb39fcf32\" (UID: \"0ca13142-fb38-4f67-88f4-834bb39fcf32\") " Oct 05 07:06:26 crc kubenswrapper[4846]: I1005 07:06:26.534832 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0ca13142-fb38-4f67-88f4-834bb39fcf32-dns-svc\") pod \"0ca13142-fb38-4f67-88f4-834bb39fcf32\" (UID: \"0ca13142-fb38-4f67-88f4-834bb39fcf32\") " Oct 05 07:06:26 crc kubenswrapper[4846]: I1005 07:06:26.565725 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0ca13142-fb38-4f67-88f4-834bb39fcf32-kube-api-access-cxlh2" (OuterVolumeSpecName: "kube-api-access-cxlh2") pod "0ca13142-fb38-4f67-88f4-834bb39fcf32" (UID: "0ca13142-fb38-4f67-88f4-834bb39fcf32"). InnerVolumeSpecName "kube-api-access-cxlh2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:06:26 crc kubenswrapper[4846]: I1005 07:06:26.632918 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0ca13142-fb38-4f67-88f4-834bb39fcf32-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "0ca13142-fb38-4f67-88f4-834bb39fcf32" (UID: "0ca13142-fb38-4f67-88f4-834bb39fcf32"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 07:06:26 crc kubenswrapper[4846]: I1005 07:06:26.635942 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0ca13142-fb38-4f67-88f4-834bb39fcf32-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "0ca13142-fb38-4f67-88f4-834bb39fcf32" (UID: "0ca13142-fb38-4f67-88f4-834bb39fcf32"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 07:06:26 crc kubenswrapper[4846]: I1005 07:06:26.637701 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cxlh2\" (UniqueName: \"kubernetes.io/projected/0ca13142-fb38-4f67-88f4-834bb39fcf32-kube-api-access-cxlh2\") on node \"crc\" DevicePath \"\"" Oct 05 07:06:26 crc kubenswrapper[4846]: I1005 07:06:26.637746 4846 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0ca13142-fb38-4f67-88f4-834bb39fcf32-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 05 07:06:26 crc kubenswrapper[4846]: I1005 07:06:26.637766 4846 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0ca13142-fb38-4f67-88f4-834bb39fcf32-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 05 07:06:26 crc kubenswrapper[4846]: I1005 07:06:26.639303 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0ca13142-fb38-4f67-88f4-834bb39fcf32-config" (OuterVolumeSpecName: "config") pod "0ca13142-fb38-4f67-88f4-834bb39fcf32" (UID: "0ca13142-fb38-4f67-88f4-834bb39fcf32"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 07:06:26 crc kubenswrapper[4846]: I1005 07:06:26.645098 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0ca13142-fb38-4f67-88f4-834bb39fcf32-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "0ca13142-fb38-4f67-88f4-834bb39fcf32" (UID: "0ca13142-fb38-4f67-88f4-834bb39fcf32"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 07:06:26 crc kubenswrapper[4846]: I1005 07:06:26.740694 4846 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0ca13142-fb38-4f67-88f4-834bb39fcf32-config\") on node \"crc\" DevicePath \"\"" Oct 05 07:06:26 crc kubenswrapper[4846]: I1005 07:06:26.741100 4846 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0ca13142-fb38-4f67-88f4-834bb39fcf32-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 05 07:06:26 crc kubenswrapper[4846]: I1005 07:06:26.814685 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-84f55cbc7-bl827"] Oct 05 07:06:26 crc kubenswrapper[4846]: I1005 07:06:26.857592 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-84f55cbc7-bl827"] Oct 05 07:06:26 crc kubenswrapper[4846]: I1005 07:06:26.888102 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-b9gxc" Oct 05 07:06:26 crc kubenswrapper[4846]: E1005 07:06:26.909504 4846 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0ca13142_fb38_4f67_88f4_834bb39fcf32.slice\": RecentStats: unable to find data in memory cache]" Oct 05 07:06:27 crc kubenswrapper[4846]: I1005 07:06:27.046765 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/0ffabf45-809d-43f4-902b-755111599aa8-etc-swift\") pod \"0ffabf45-809d-43f4-902b-755111599aa8\" (UID: \"0ffabf45-809d-43f4-902b-755111599aa8\") " Oct 05 07:06:27 crc kubenswrapper[4846]: I1005 07:06:27.047107 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hztth\" (UniqueName: \"kubernetes.io/projected/0ffabf45-809d-43f4-902b-755111599aa8-kube-api-access-hztth\") pod \"0ffabf45-809d-43f4-902b-755111599aa8\" (UID: \"0ffabf45-809d-43f4-902b-755111599aa8\") " Oct 05 07:06:27 crc kubenswrapper[4846]: I1005 07:06:27.047256 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/0ffabf45-809d-43f4-902b-755111599aa8-dispersionconf\") pod \"0ffabf45-809d-43f4-902b-755111599aa8\" (UID: \"0ffabf45-809d-43f4-902b-755111599aa8\") " Oct 05 07:06:27 crc kubenswrapper[4846]: I1005 07:06:27.047933 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0ffabf45-809d-43f4-902b-755111599aa8-combined-ca-bundle\") pod \"0ffabf45-809d-43f4-902b-755111599aa8\" (UID: \"0ffabf45-809d-43f4-902b-755111599aa8\") " Oct 05 07:06:27 crc kubenswrapper[4846]: I1005 07:06:27.048078 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/0ffabf45-809d-43f4-902b-755111599aa8-ring-data-devices\") pod \"0ffabf45-809d-43f4-902b-755111599aa8\" (UID: \"0ffabf45-809d-43f4-902b-755111599aa8\") " Oct 05 07:06:27 crc kubenswrapper[4846]: I1005 07:06:27.048227 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/0ffabf45-809d-43f4-902b-755111599aa8-swiftconf\") pod \"0ffabf45-809d-43f4-902b-755111599aa8\" (UID: \"0ffabf45-809d-43f4-902b-755111599aa8\") " Oct 05 07:06:27 crc kubenswrapper[4846]: I1005 07:06:27.048392 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0ffabf45-809d-43f4-902b-755111599aa8-scripts\") pod \"0ffabf45-809d-43f4-902b-755111599aa8\" (UID: \"0ffabf45-809d-43f4-902b-755111599aa8\") " Oct 05 07:06:27 crc kubenswrapper[4846]: I1005 07:06:27.048979 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0ffabf45-809d-43f4-902b-755111599aa8-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "0ffabf45-809d-43f4-902b-755111599aa8" (UID: "0ffabf45-809d-43f4-902b-755111599aa8"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 07:06:27 crc kubenswrapper[4846]: I1005 07:06:27.049455 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0ffabf45-809d-43f4-902b-755111599aa8-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "0ffabf45-809d-43f4-902b-755111599aa8" (UID: "0ffabf45-809d-43f4-902b-755111599aa8"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 07:06:27 crc kubenswrapper[4846]: I1005 07:06:27.049769 4846 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/0ffabf45-809d-43f4-902b-755111599aa8-ring-data-devices\") on node \"crc\" DevicePath \"\"" Oct 05 07:06:27 crc kubenswrapper[4846]: I1005 07:06:27.050255 4846 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/0ffabf45-809d-43f4-902b-755111599aa8-etc-swift\") on node \"crc\" DevicePath \"\"" Oct 05 07:06:27 crc kubenswrapper[4846]: I1005 07:06:27.071064 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0ffabf45-809d-43f4-902b-755111599aa8-kube-api-access-hztth" (OuterVolumeSpecName: "kube-api-access-hztth") pod "0ffabf45-809d-43f4-902b-755111599aa8" (UID: "0ffabf45-809d-43f4-902b-755111599aa8"). InnerVolumeSpecName "kube-api-access-hztth". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:06:27 crc kubenswrapper[4846]: I1005 07:06:27.075283 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0ffabf45-809d-43f4-902b-755111599aa8-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "0ffabf45-809d-43f4-902b-755111599aa8" (UID: "0ffabf45-809d-43f4-902b-755111599aa8"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:06:27 crc kubenswrapper[4846]: I1005 07:06:27.076954 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0ffabf45-809d-43f4-902b-755111599aa8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0ffabf45-809d-43f4-902b-755111599aa8" (UID: "0ffabf45-809d-43f4-902b-755111599aa8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:06:27 crc kubenswrapper[4846]: I1005 07:06:27.085026 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0ffabf45-809d-43f4-902b-755111599aa8-scripts" (OuterVolumeSpecName: "scripts") pod "0ffabf45-809d-43f4-902b-755111599aa8" (UID: "0ffabf45-809d-43f4-902b-755111599aa8"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 07:06:27 crc kubenswrapper[4846]: I1005 07:06:27.108380 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0ffabf45-809d-43f4-902b-755111599aa8-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "0ffabf45-809d-43f4-902b-755111599aa8" (UID: "0ffabf45-809d-43f4-902b-755111599aa8"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:06:27 crc kubenswrapper[4846]: I1005 07:06:27.152396 4846 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0ffabf45-809d-43f4-902b-755111599aa8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 07:06:27 crc kubenswrapper[4846]: I1005 07:06:27.152445 4846 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/0ffabf45-809d-43f4-902b-755111599aa8-swiftconf\") on node \"crc\" DevicePath \"\"" Oct 05 07:06:27 crc kubenswrapper[4846]: I1005 07:06:27.152459 4846 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0ffabf45-809d-43f4-902b-755111599aa8-scripts\") on node \"crc\" DevicePath \"\"" Oct 05 07:06:27 crc kubenswrapper[4846]: I1005 07:06:27.152474 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hztth\" (UniqueName: \"kubernetes.io/projected/0ffabf45-809d-43f4-902b-755111599aa8-kube-api-access-hztth\") on node \"crc\" DevicePath \"\"" Oct 05 07:06:27 crc kubenswrapper[4846]: I1005 07:06:27.152489 4846 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/0ffabf45-809d-43f4-902b-755111599aa8-dispersionconf\") on node \"crc\" DevicePath \"\"" Oct 05 07:06:27 crc kubenswrapper[4846]: I1005 07:06:27.440318 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-b9gxc" event={"ID":"0ffabf45-809d-43f4-902b-755111599aa8","Type":"ContainerDied","Data":"a674a09f7ff0641659924eeebaaa8b2aec3960afb067099baad0cac94674df95"} Oct 05 07:06:27 crc kubenswrapper[4846]: I1005 07:06:27.441988 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a674a09f7ff0641659924eeebaaa8b2aec3960afb067099baad0cac94674df95" Oct 05 07:06:27 crc kubenswrapper[4846]: I1005 07:06:27.440363 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-b9gxc" Oct 05 07:06:27 crc kubenswrapper[4846]: I1005 07:06:27.924706 4846 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-4kqnw" podUID="169918d8-597e-4ef4-b00c-fb14c4d0347f" containerName="ovn-controller" probeResult="failure" output=< Oct 05 07:06:27 crc kubenswrapper[4846]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Oct 05 07:06:27 crc kubenswrapper[4846]: > Oct 05 07:06:27 crc kubenswrapper[4846]: I1005 07:06:27.961424 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-67b1-account-create-lzlh4"] Oct 05 07:06:27 crc kubenswrapper[4846]: E1005 07:06:27.962003 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d8234451-0d89-47e7-af4c-0263c3b6f58b" containerName="mariadb-database-create" Oct 05 07:06:27 crc kubenswrapper[4846]: I1005 07:06:27.962028 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="d8234451-0d89-47e7-af4c-0263c3b6f58b" containerName="mariadb-database-create" Oct 05 07:06:27 crc kubenswrapper[4846]: E1005 07:06:27.962050 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0ca13142-fb38-4f67-88f4-834bb39fcf32" containerName="dnsmasq-dns" Oct 05 07:06:27 crc kubenswrapper[4846]: I1005 07:06:27.962059 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="0ca13142-fb38-4f67-88f4-834bb39fcf32" containerName="dnsmasq-dns" Oct 05 07:06:27 crc kubenswrapper[4846]: E1005 07:06:27.962070 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cfa439eb-10b4-4b11-9baa-377a45fe68d0" containerName="mariadb-database-create" Oct 05 07:06:27 crc kubenswrapper[4846]: I1005 07:06:27.962080 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="cfa439eb-10b4-4b11-9baa-377a45fe68d0" containerName="mariadb-database-create" Oct 05 07:06:27 crc kubenswrapper[4846]: E1005 07:06:27.962105 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f8c7a428-fb8a-4992-bf65-d603b4cfc323" containerName="mariadb-database-create" Oct 05 07:06:27 crc kubenswrapper[4846]: I1005 07:06:27.962112 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="f8c7a428-fb8a-4992-bf65-d603b4cfc323" containerName="mariadb-database-create" Oct 05 07:06:27 crc kubenswrapper[4846]: E1005 07:06:27.962128 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0ca13142-fb38-4f67-88f4-834bb39fcf32" containerName="init" Oct 05 07:06:27 crc kubenswrapper[4846]: I1005 07:06:27.962135 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="0ca13142-fb38-4f67-88f4-834bb39fcf32" containerName="init" Oct 05 07:06:27 crc kubenswrapper[4846]: E1005 07:06:27.962153 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="727b9fbc-1de3-4e7e-ac8a-ffbbd76eac0e" containerName="init" Oct 05 07:06:27 crc kubenswrapper[4846]: I1005 07:06:27.962161 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="727b9fbc-1de3-4e7e-ac8a-ffbbd76eac0e" containerName="init" Oct 05 07:06:27 crc kubenswrapper[4846]: E1005 07:06:27.962195 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0ffabf45-809d-43f4-902b-755111599aa8" containerName="swift-ring-rebalance" Oct 05 07:06:27 crc kubenswrapper[4846]: I1005 07:06:27.962204 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="0ffabf45-809d-43f4-902b-755111599aa8" containerName="swift-ring-rebalance" Oct 05 07:06:27 crc kubenswrapper[4846]: I1005 07:06:27.962426 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="d8234451-0d89-47e7-af4c-0263c3b6f58b" containerName="mariadb-database-create" Oct 05 07:06:27 crc kubenswrapper[4846]: I1005 07:06:27.962446 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="0ffabf45-809d-43f4-902b-755111599aa8" containerName="swift-ring-rebalance" Oct 05 07:06:27 crc kubenswrapper[4846]: I1005 07:06:27.962459 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="727b9fbc-1de3-4e7e-ac8a-ffbbd76eac0e" containerName="init" Oct 05 07:06:27 crc kubenswrapper[4846]: I1005 07:06:27.962482 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="f8c7a428-fb8a-4992-bf65-d603b4cfc323" containerName="mariadb-database-create" Oct 05 07:06:27 crc kubenswrapper[4846]: I1005 07:06:27.962525 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="0ca13142-fb38-4f67-88f4-834bb39fcf32" containerName="dnsmasq-dns" Oct 05 07:06:27 crc kubenswrapper[4846]: I1005 07:06:27.962545 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="cfa439eb-10b4-4b11-9baa-377a45fe68d0" containerName="mariadb-database-create" Oct 05 07:06:27 crc kubenswrapper[4846]: I1005 07:06:27.963348 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-67b1-account-create-lzlh4" Oct 05 07:06:27 crc kubenswrapper[4846]: I1005 07:06:27.966280 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-db-secret" Oct 05 07:06:27 crc kubenswrapper[4846]: I1005 07:06:27.972911 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n64sg\" (UniqueName: \"kubernetes.io/projected/7fd963e7-aa89-4a46-8b1b-ac1bd6bc503d-kube-api-access-n64sg\") pod \"keystone-67b1-account-create-lzlh4\" (UID: \"7fd963e7-aa89-4a46-8b1b-ac1bd6bc503d\") " pod="openstack/keystone-67b1-account-create-lzlh4" Oct 05 07:06:27 crc kubenswrapper[4846]: I1005 07:06:27.976866 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-67b1-account-create-lzlh4"] Oct 05 07:06:28 crc kubenswrapper[4846]: I1005 07:06:28.027608 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-sg4vg" Oct 05 07:06:28 crc kubenswrapper[4846]: I1005 07:06:28.033643 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-sg4vg" Oct 05 07:06:28 crc kubenswrapper[4846]: I1005 07:06:28.074872 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n64sg\" (UniqueName: \"kubernetes.io/projected/7fd963e7-aa89-4a46-8b1b-ac1bd6bc503d-kube-api-access-n64sg\") pod \"keystone-67b1-account-create-lzlh4\" (UID: \"7fd963e7-aa89-4a46-8b1b-ac1bd6bc503d\") " pod="openstack/keystone-67b1-account-create-lzlh4" Oct 05 07:06:28 crc kubenswrapper[4846]: I1005 07:06:28.100372 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n64sg\" (UniqueName: \"kubernetes.io/projected/7fd963e7-aa89-4a46-8b1b-ac1bd6bc503d-kube-api-access-n64sg\") pod \"keystone-67b1-account-create-lzlh4\" (UID: \"7fd963e7-aa89-4a46-8b1b-ac1bd6bc503d\") " pod="openstack/keystone-67b1-account-create-lzlh4" Oct 05 07:06:28 crc kubenswrapper[4846]: I1005 07:06:28.253335 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-2760-account-create-pvm8n"] Oct 05 07:06:28 crc kubenswrapper[4846]: I1005 07:06:28.255669 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-2760-account-create-pvm8n" Oct 05 07:06:28 crc kubenswrapper[4846]: I1005 07:06:28.258126 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-db-secret" Oct 05 07:06:28 crc kubenswrapper[4846]: I1005 07:06:28.262188 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-2760-account-create-pvm8n"] Oct 05 07:06:28 crc kubenswrapper[4846]: I1005 07:06:28.279111 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hj729\" (UniqueName: \"kubernetes.io/projected/68ba5901-a360-4343-8e70-c7e1c9e829f3-kube-api-access-hj729\") pod \"placement-2760-account-create-pvm8n\" (UID: \"68ba5901-a360-4343-8e70-c7e1c9e829f3\") " pod="openstack/placement-2760-account-create-pvm8n" Oct 05 07:06:28 crc kubenswrapper[4846]: I1005 07:06:28.287346 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-67b1-account-create-lzlh4" Oct 05 07:06:28 crc kubenswrapper[4846]: I1005 07:06:28.292555 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-4kqnw-config-dhsjv"] Oct 05 07:06:28 crc kubenswrapper[4846]: I1005 07:06:28.296968 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-4kqnw-config-dhsjv" Oct 05 07:06:28 crc kubenswrapper[4846]: I1005 07:06:28.307989 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Oct 05 07:06:28 crc kubenswrapper[4846]: I1005 07:06:28.330483 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-4kqnw-config-dhsjv"] Oct 05 07:06:28 crc kubenswrapper[4846]: I1005 07:06:28.381497 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hj729\" (UniqueName: \"kubernetes.io/projected/68ba5901-a360-4343-8e70-c7e1c9e829f3-kube-api-access-hj729\") pod \"placement-2760-account-create-pvm8n\" (UID: \"68ba5901-a360-4343-8e70-c7e1c9e829f3\") " pod="openstack/placement-2760-account-create-pvm8n" Oct 05 07:06:28 crc kubenswrapper[4846]: I1005 07:06:28.403234 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hj729\" (UniqueName: \"kubernetes.io/projected/68ba5901-a360-4343-8e70-c7e1c9e829f3-kube-api-access-hj729\") pod \"placement-2760-account-create-pvm8n\" (UID: \"68ba5901-a360-4343-8e70-c7e1c9e829f3\") " pod="openstack/placement-2760-account-create-pvm8n" Oct 05 07:06:28 crc kubenswrapper[4846]: I1005 07:06:28.483850 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/88563126-2bb3-482f-8991-fe6ad17d5e24-var-run-ovn\") pod \"ovn-controller-4kqnw-config-dhsjv\" (UID: \"88563126-2bb3-482f-8991-fe6ad17d5e24\") " pod="openstack/ovn-controller-4kqnw-config-dhsjv" Oct 05 07:06:28 crc kubenswrapper[4846]: I1005 07:06:28.483963 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/88563126-2bb3-482f-8991-fe6ad17d5e24-var-log-ovn\") pod \"ovn-controller-4kqnw-config-dhsjv\" (UID: \"88563126-2bb3-482f-8991-fe6ad17d5e24\") " pod="openstack/ovn-controller-4kqnw-config-dhsjv" Oct 05 07:06:28 crc kubenswrapper[4846]: I1005 07:06:28.484034 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lblrn\" (UniqueName: \"kubernetes.io/projected/88563126-2bb3-482f-8991-fe6ad17d5e24-kube-api-access-lblrn\") pod \"ovn-controller-4kqnw-config-dhsjv\" (UID: \"88563126-2bb3-482f-8991-fe6ad17d5e24\") " pod="openstack/ovn-controller-4kqnw-config-dhsjv" Oct 05 07:06:28 crc kubenswrapper[4846]: I1005 07:06:28.484065 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/88563126-2bb3-482f-8991-fe6ad17d5e24-additional-scripts\") pod \"ovn-controller-4kqnw-config-dhsjv\" (UID: \"88563126-2bb3-482f-8991-fe6ad17d5e24\") " pod="openstack/ovn-controller-4kqnw-config-dhsjv" Oct 05 07:06:28 crc kubenswrapper[4846]: I1005 07:06:28.484084 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/88563126-2bb3-482f-8991-fe6ad17d5e24-var-run\") pod \"ovn-controller-4kqnw-config-dhsjv\" (UID: \"88563126-2bb3-482f-8991-fe6ad17d5e24\") " pod="openstack/ovn-controller-4kqnw-config-dhsjv" Oct 05 07:06:28 crc kubenswrapper[4846]: I1005 07:06:28.484108 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/88563126-2bb3-482f-8991-fe6ad17d5e24-scripts\") pod \"ovn-controller-4kqnw-config-dhsjv\" (UID: \"88563126-2bb3-482f-8991-fe6ad17d5e24\") " pod="openstack/ovn-controller-4kqnw-config-dhsjv" Oct 05 07:06:28 crc kubenswrapper[4846]: I1005 07:06:28.511008 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0ca13142-fb38-4f67-88f4-834bb39fcf32" path="/var/lib/kubelet/pods/0ca13142-fb38-4f67-88f4-834bb39fcf32/volumes" Oct 05 07:06:28 crc kubenswrapper[4846]: I1005 07:06:28.555212 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-bc5f-account-create-spj8j"] Oct 05 07:06:28 crc kubenswrapper[4846]: I1005 07:06:28.556910 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-bc5f-account-create-spj8j" Oct 05 07:06:28 crc kubenswrapper[4846]: I1005 07:06:28.561809 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-db-secret" Oct 05 07:06:28 crc kubenswrapper[4846]: I1005 07:06:28.574352 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-bc5f-account-create-spj8j"] Oct 05 07:06:28 crc kubenswrapper[4846]: I1005 07:06:28.576437 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-2760-account-create-pvm8n" Oct 05 07:06:28 crc kubenswrapper[4846]: I1005 07:06:28.585515 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/88563126-2bb3-482f-8991-fe6ad17d5e24-var-log-ovn\") pod \"ovn-controller-4kqnw-config-dhsjv\" (UID: \"88563126-2bb3-482f-8991-fe6ad17d5e24\") " pod="openstack/ovn-controller-4kqnw-config-dhsjv" Oct 05 07:06:28 crc kubenswrapper[4846]: I1005 07:06:28.585631 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lblrn\" (UniqueName: \"kubernetes.io/projected/88563126-2bb3-482f-8991-fe6ad17d5e24-kube-api-access-lblrn\") pod \"ovn-controller-4kqnw-config-dhsjv\" (UID: \"88563126-2bb3-482f-8991-fe6ad17d5e24\") " pod="openstack/ovn-controller-4kqnw-config-dhsjv" Oct 05 07:06:28 crc kubenswrapper[4846]: I1005 07:06:28.585676 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/88563126-2bb3-482f-8991-fe6ad17d5e24-additional-scripts\") pod \"ovn-controller-4kqnw-config-dhsjv\" (UID: \"88563126-2bb3-482f-8991-fe6ad17d5e24\") " pod="openstack/ovn-controller-4kqnw-config-dhsjv" Oct 05 07:06:28 crc kubenswrapper[4846]: I1005 07:06:28.585694 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/88563126-2bb3-482f-8991-fe6ad17d5e24-var-run\") pod \"ovn-controller-4kqnw-config-dhsjv\" (UID: \"88563126-2bb3-482f-8991-fe6ad17d5e24\") " pod="openstack/ovn-controller-4kqnw-config-dhsjv" Oct 05 07:06:28 crc kubenswrapper[4846]: I1005 07:06:28.585722 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/88563126-2bb3-482f-8991-fe6ad17d5e24-scripts\") pod \"ovn-controller-4kqnw-config-dhsjv\" (UID: \"88563126-2bb3-482f-8991-fe6ad17d5e24\") " pod="openstack/ovn-controller-4kqnw-config-dhsjv" Oct 05 07:06:28 crc kubenswrapper[4846]: I1005 07:06:28.585746 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/88563126-2bb3-482f-8991-fe6ad17d5e24-var-run-ovn\") pod \"ovn-controller-4kqnw-config-dhsjv\" (UID: \"88563126-2bb3-482f-8991-fe6ad17d5e24\") " pod="openstack/ovn-controller-4kqnw-config-dhsjv" Oct 05 07:06:28 crc kubenswrapper[4846]: I1005 07:06:28.586830 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/88563126-2bb3-482f-8991-fe6ad17d5e24-var-run-ovn\") pod \"ovn-controller-4kqnw-config-dhsjv\" (UID: \"88563126-2bb3-482f-8991-fe6ad17d5e24\") " pod="openstack/ovn-controller-4kqnw-config-dhsjv" Oct 05 07:06:28 crc kubenswrapper[4846]: I1005 07:06:28.586831 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/88563126-2bb3-482f-8991-fe6ad17d5e24-var-log-ovn\") pod \"ovn-controller-4kqnw-config-dhsjv\" (UID: \"88563126-2bb3-482f-8991-fe6ad17d5e24\") " pod="openstack/ovn-controller-4kqnw-config-dhsjv" Oct 05 07:06:28 crc kubenswrapper[4846]: I1005 07:06:28.586897 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/88563126-2bb3-482f-8991-fe6ad17d5e24-var-run\") pod \"ovn-controller-4kqnw-config-dhsjv\" (UID: \"88563126-2bb3-482f-8991-fe6ad17d5e24\") " pod="openstack/ovn-controller-4kqnw-config-dhsjv" Oct 05 07:06:28 crc kubenswrapper[4846]: I1005 07:06:28.600127 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/88563126-2bb3-482f-8991-fe6ad17d5e24-scripts\") pod \"ovn-controller-4kqnw-config-dhsjv\" (UID: \"88563126-2bb3-482f-8991-fe6ad17d5e24\") " pod="openstack/ovn-controller-4kqnw-config-dhsjv" Oct 05 07:06:28 crc kubenswrapper[4846]: I1005 07:06:28.603980 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/88563126-2bb3-482f-8991-fe6ad17d5e24-additional-scripts\") pod \"ovn-controller-4kqnw-config-dhsjv\" (UID: \"88563126-2bb3-482f-8991-fe6ad17d5e24\") " pod="openstack/ovn-controller-4kqnw-config-dhsjv" Oct 05 07:06:28 crc kubenswrapper[4846]: I1005 07:06:28.654486 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lblrn\" (UniqueName: \"kubernetes.io/projected/88563126-2bb3-482f-8991-fe6ad17d5e24-kube-api-access-lblrn\") pod \"ovn-controller-4kqnw-config-dhsjv\" (UID: \"88563126-2bb3-482f-8991-fe6ad17d5e24\") " pod="openstack/ovn-controller-4kqnw-config-dhsjv" Oct 05 07:06:28 crc kubenswrapper[4846]: I1005 07:06:28.687765 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qd9q7\" (UniqueName: \"kubernetes.io/projected/bd630f2e-4b0e-4cc7-a0de-4f60c72d941c-kube-api-access-qd9q7\") pod \"glance-bc5f-account-create-spj8j\" (UID: \"bd630f2e-4b0e-4cc7-a0de-4f60c72d941c\") " pod="openstack/glance-bc5f-account-create-spj8j" Oct 05 07:06:28 crc kubenswrapper[4846]: I1005 07:06:28.789488 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qd9q7\" (UniqueName: \"kubernetes.io/projected/bd630f2e-4b0e-4cc7-a0de-4f60c72d941c-kube-api-access-qd9q7\") pod \"glance-bc5f-account-create-spj8j\" (UID: \"bd630f2e-4b0e-4cc7-a0de-4f60c72d941c\") " pod="openstack/glance-bc5f-account-create-spj8j" Oct 05 07:06:28 crc kubenswrapper[4846]: I1005 07:06:28.808279 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qd9q7\" (UniqueName: \"kubernetes.io/projected/bd630f2e-4b0e-4cc7-a0de-4f60c72d941c-kube-api-access-qd9q7\") pod \"glance-bc5f-account-create-spj8j\" (UID: \"bd630f2e-4b0e-4cc7-a0de-4f60c72d941c\") " pod="openstack/glance-bc5f-account-create-spj8j" Oct 05 07:06:28 crc kubenswrapper[4846]: I1005 07:06:28.877670 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-67b1-account-create-lzlh4"] Oct 05 07:06:28 crc kubenswrapper[4846]: I1005 07:06:28.887487 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-2760-account-create-pvm8n"] Oct 05 07:06:28 crc kubenswrapper[4846]: I1005 07:06:28.898086 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-bc5f-account-create-spj8j" Oct 05 07:06:28 crc kubenswrapper[4846]: I1005 07:06:28.916155 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-4kqnw-config-dhsjv" Oct 05 07:06:28 crc kubenswrapper[4846]: W1005 07:06:28.920396 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod68ba5901_a360_4343_8e70_c7e1c9e829f3.slice/crio-274907ece8cd88475f7700ef70429c0079004ed92e023c5b0ce3f380d346de0a WatchSource:0}: Error finding container 274907ece8cd88475f7700ef70429c0079004ed92e023c5b0ce3f380d346de0a: Status 404 returned error can't find the container with id 274907ece8cd88475f7700ef70429c0079004ed92e023c5b0ce3f380d346de0a Oct 05 07:06:29 crc kubenswrapper[4846]: I1005 07:06:29.408323 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-bc5f-account-create-spj8j"] Oct 05 07:06:29 crc kubenswrapper[4846]: W1005 07:06:29.426481 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbd630f2e_4b0e_4cc7_a0de_4f60c72d941c.slice/crio-afe27f5f42bf600f435a71bbeb762ebc05f11aef89f8cc97f3dc276c23c1e440 WatchSource:0}: Error finding container afe27f5f42bf600f435a71bbeb762ebc05f11aef89f8cc97f3dc276c23c1e440: Status 404 returned error can't find the container with id afe27f5f42bf600f435a71bbeb762ebc05f11aef89f8cc97f3dc276c23c1e440 Oct 05 07:06:29 crc kubenswrapper[4846]: I1005 07:06:29.463006 4846 generic.go:334] "Generic (PLEG): container finished" podID="68ba5901-a360-4343-8e70-c7e1c9e829f3" containerID="1b20e0f9240e1bcc74644cccf0d33c9f6e702a6015a35a33c23dbd86eb5b3d91" exitCode=0 Oct 05 07:06:29 crc kubenswrapper[4846]: I1005 07:06:29.463106 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-2760-account-create-pvm8n" event={"ID":"68ba5901-a360-4343-8e70-c7e1c9e829f3","Type":"ContainerDied","Data":"1b20e0f9240e1bcc74644cccf0d33c9f6e702a6015a35a33c23dbd86eb5b3d91"} Oct 05 07:06:29 crc kubenswrapper[4846]: I1005 07:06:29.463172 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-2760-account-create-pvm8n" event={"ID":"68ba5901-a360-4343-8e70-c7e1c9e829f3","Type":"ContainerStarted","Data":"274907ece8cd88475f7700ef70429c0079004ed92e023c5b0ce3f380d346de0a"} Oct 05 07:06:29 crc kubenswrapper[4846]: I1005 07:06:29.465518 4846 generic.go:334] "Generic (PLEG): container finished" podID="7fd963e7-aa89-4a46-8b1b-ac1bd6bc503d" containerID="cf40f2c3e601ab1a12429ff595152653f7656b17ce27280b2f147555aeee04d2" exitCode=0 Oct 05 07:06:29 crc kubenswrapper[4846]: I1005 07:06:29.465647 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-67b1-account-create-lzlh4" event={"ID":"7fd963e7-aa89-4a46-8b1b-ac1bd6bc503d","Type":"ContainerDied","Data":"cf40f2c3e601ab1a12429ff595152653f7656b17ce27280b2f147555aeee04d2"} Oct 05 07:06:29 crc kubenswrapper[4846]: I1005 07:06:29.465702 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-67b1-account-create-lzlh4" event={"ID":"7fd963e7-aa89-4a46-8b1b-ac1bd6bc503d","Type":"ContainerStarted","Data":"052566be39bea655dce8912bf5a5881d5e0edf7f11044cf92807438953702cb3"} Oct 05 07:06:29 crc kubenswrapper[4846]: I1005 07:06:29.467908 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-bc5f-account-create-spj8j" event={"ID":"bd630f2e-4b0e-4cc7-a0de-4f60c72d941c","Type":"ContainerStarted","Data":"afe27f5f42bf600f435a71bbeb762ebc05f11aef89f8cc97f3dc276c23c1e440"} Oct 05 07:06:29 crc kubenswrapper[4846]: I1005 07:06:29.555277 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-4kqnw-config-dhsjv"] Oct 05 07:06:29 crc kubenswrapper[4846]: W1005 07:06:29.560419 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod88563126_2bb3_482f_8991_fe6ad17d5e24.slice/crio-bbefd4f1d3fb183d71bbefd3781d215382045a3ded6df8a5d5a21631e8733e3f WatchSource:0}: Error finding container bbefd4f1d3fb183d71bbefd3781d215382045a3ded6df8a5d5a21631e8733e3f: Status 404 returned error can't find the container with id bbefd4f1d3fb183d71bbefd3781d215382045a3ded6df8a5d5a21631e8733e3f Oct 05 07:06:29 crc kubenswrapper[4846]: I1005 07:06:29.884147 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-northd-0" Oct 05 07:06:30 crc kubenswrapper[4846]: I1005 07:06:30.481015 4846 generic.go:334] "Generic (PLEG): container finished" podID="bd630f2e-4b0e-4cc7-a0de-4f60c72d941c" containerID="f80ec72ff1ba79fd1d723988d040ed000dc6a49b01ce6daeba2f5db0549784a5" exitCode=0 Oct 05 07:06:30 crc kubenswrapper[4846]: I1005 07:06:30.481235 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-bc5f-account-create-spj8j" event={"ID":"bd630f2e-4b0e-4cc7-a0de-4f60c72d941c","Type":"ContainerDied","Data":"f80ec72ff1ba79fd1d723988d040ed000dc6a49b01ce6daeba2f5db0549784a5"} Oct 05 07:06:30 crc kubenswrapper[4846]: I1005 07:06:30.484202 4846 generic.go:334] "Generic (PLEG): container finished" podID="88563126-2bb3-482f-8991-fe6ad17d5e24" containerID="b1b00d22e5b0700b970a8485f74339e8139dbc0cb0edc8276f797b2d2f338545" exitCode=0 Oct 05 07:06:30 crc kubenswrapper[4846]: I1005 07:06:30.484282 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-4kqnw-config-dhsjv" event={"ID":"88563126-2bb3-482f-8991-fe6ad17d5e24","Type":"ContainerDied","Data":"b1b00d22e5b0700b970a8485f74339e8139dbc0cb0edc8276f797b2d2f338545"} Oct 05 07:06:30 crc kubenswrapper[4846]: I1005 07:06:30.484347 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-4kqnw-config-dhsjv" event={"ID":"88563126-2bb3-482f-8991-fe6ad17d5e24","Type":"ContainerStarted","Data":"bbefd4f1d3fb183d71bbefd3781d215382045a3ded6df8a5d5a21631e8733e3f"} Oct 05 07:06:30 crc kubenswrapper[4846]: I1005 07:06:30.991872 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-2760-account-create-pvm8n" Oct 05 07:06:31 crc kubenswrapper[4846]: I1005 07:06:31.002005 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-67b1-account-create-lzlh4" Oct 05 07:06:31 crc kubenswrapper[4846]: I1005 07:06:31.140487 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hj729\" (UniqueName: \"kubernetes.io/projected/68ba5901-a360-4343-8e70-c7e1c9e829f3-kube-api-access-hj729\") pod \"68ba5901-a360-4343-8e70-c7e1c9e829f3\" (UID: \"68ba5901-a360-4343-8e70-c7e1c9e829f3\") " Oct 05 07:06:31 crc kubenswrapper[4846]: I1005 07:06:31.140881 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n64sg\" (UniqueName: \"kubernetes.io/projected/7fd963e7-aa89-4a46-8b1b-ac1bd6bc503d-kube-api-access-n64sg\") pod \"7fd963e7-aa89-4a46-8b1b-ac1bd6bc503d\" (UID: \"7fd963e7-aa89-4a46-8b1b-ac1bd6bc503d\") " Oct 05 07:06:31 crc kubenswrapper[4846]: I1005 07:06:31.148898 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/68ba5901-a360-4343-8e70-c7e1c9e829f3-kube-api-access-hj729" (OuterVolumeSpecName: "kube-api-access-hj729") pod "68ba5901-a360-4343-8e70-c7e1c9e829f3" (UID: "68ba5901-a360-4343-8e70-c7e1c9e829f3"). InnerVolumeSpecName "kube-api-access-hj729". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:06:31 crc kubenswrapper[4846]: I1005 07:06:31.149133 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7fd963e7-aa89-4a46-8b1b-ac1bd6bc503d-kube-api-access-n64sg" (OuterVolumeSpecName: "kube-api-access-n64sg") pod "7fd963e7-aa89-4a46-8b1b-ac1bd6bc503d" (UID: "7fd963e7-aa89-4a46-8b1b-ac1bd6bc503d"). InnerVolumeSpecName "kube-api-access-n64sg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:06:31 crc kubenswrapper[4846]: I1005 07:06:31.245414 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n64sg\" (UniqueName: \"kubernetes.io/projected/7fd963e7-aa89-4a46-8b1b-ac1bd6bc503d-kube-api-access-n64sg\") on node \"crc\" DevicePath \"\"" Oct 05 07:06:31 crc kubenswrapper[4846]: I1005 07:06:31.245476 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hj729\" (UniqueName: \"kubernetes.io/projected/68ba5901-a360-4343-8e70-c7e1c9e829f3-kube-api-access-hj729\") on node \"crc\" DevicePath \"\"" Oct 05 07:06:31 crc kubenswrapper[4846]: I1005 07:06:31.513421 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-2760-account-create-pvm8n" event={"ID":"68ba5901-a360-4343-8e70-c7e1c9e829f3","Type":"ContainerDied","Data":"274907ece8cd88475f7700ef70429c0079004ed92e023c5b0ce3f380d346de0a"} Oct 05 07:06:31 crc kubenswrapper[4846]: I1005 07:06:31.514117 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="274907ece8cd88475f7700ef70429c0079004ed92e023c5b0ce3f380d346de0a" Oct 05 07:06:31 crc kubenswrapper[4846]: I1005 07:06:31.513414 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-2760-account-create-pvm8n" Oct 05 07:06:31 crc kubenswrapper[4846]: I1005 07:06:31.515880 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-67b1-account-create-lzlh4" Oct 05 07:06:31 crc kubenswrapper[4846]: I1005 07:06:31.516783 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-67b1-account-create-lzlh4" event={"ID":"7fd963e7-aa89-4a46-8b1b-ac1bd6bc503d","Type":"ContainerDied","Data":"052566be39bea655dce8912bf5a5881d5e0edf7f11044cf92807438953702cb3"} Oct 05 07:06:31 crc kubenswrapper[4846]: I1005 07:06:31.516834 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="052566be39bea655dce8912bf5a5881d5e0edf7f11044cf92807438953702cb3" Oct 05 07:06:31 crc kubenswrapper[4846]: I1005 07:06:31.970053 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-bc5f-account-create-spj8j" Oct 05 07:06:31 crc kubenswrapper[4846]: I1005 07:06:31.977046 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-4kqnw-config-dhsjv" Oct 05 07:06:32 crc kubenswrapper[4846]: I1005 07:06:32.064318 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/88563126-2bb3-482f-8991-fe6ad17d5e24-additional-scripts\") pod \"88563126-2bb3-482f-8991-fe6ad17d5e24\" (UID: \"88563126-2bb3-482f-8991-fe6ad17d5e24\") " Oct 05 07:06:32 crc kubenswrapper[4846]: I1005 07:06:32.064419 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/88563126-2bb3-482f-8991-fe6ad17d5e24-var-run\") pod \"88563126-2bb3-482f-8991-fe6ad17d5e24\" (UID: \"88563126-2bb3-482f-8991-fe6ad17d5e24\") " Oct 05 07:06:32 crc kubenswrapper[4846]: I1005 07:06:32.064538 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/88563126-2bb3-482f-8991-fe6ad17d5e24-var-log-ovn\") pod \"88563126-2bb3-482f-8991-fe6ad17d5e24\" (UID: \"88563126-2bb3-482f-8991-fe6ad17d5e24\") " Oct 05 07:06:32 crc kubenswrapper[4846]: I1005 07:06:32.064642 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/88563126-2bb3-482f-8991-fe6ad17d5e24-scripts\") pod \"88563126-2bb3-482f-8991-fe6ad17d5e24\" (UID: \"88563126-2bb3-482f-8991-fe6ad17d5e24\") " Oct 05 07:06:32 crc kubenswrapper[4846]: I1005 07:06:32.064760 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lblrn\" (UniqueName: \"kubernetes.io/projected/88563126-2bb3-482f-8991-fe6ad17d5e24-kube-api-access-lblrn\") pod \"88563126-2bb3-482f-8991-fe6ad17d5e24\" (UID: \"88563126-2bb3-482f-8991-fe6ad17d5e24\") " Oct 05 07:06:32 crc kubenswrapper[4846]: I1005 07:06:32.064784 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/88563126-2bb3-482f-8991-fe6ad17d5e24-var-run-ovn\") pod \"88563126-2bb3-482f-8991-fe6ad17d5e24\" (UID: \"88563126-2bb3-482f-8991-fe6ad17d5e24\") " Oct 05 07:06:32 crc kubenswrapper[4846]: I1005 07:06:32.064811 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qd9q7\" (UniqueName: \"kubernetes.io/projected/bd630f2e-4b0e-4cc7-a0de-4f60c72d941c-kube-api-access-qd9q7\") pod \"bd630f2e-4b0e-4cc7-a0de-4f60c72d941c\" (UID: \"bd630f2e-4b0e-4cc7-a0de-4f60c72d941c\") " Oct 05 07:06:32 crc kubenswrapper[4846]: I1005 07:06:32.067314 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/88563126-2bb3-482f-8991-fe6ad17d5e24-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "88563126-2bb3-482f-8991-fe6ad17d5e24" (UID: "88563126-2bb3-482f-8991-fe6ad17d5e24"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 05 07:06:32 crc kubenswrapper[4846]: I1005 07:06:32.067680 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/88563126-2bb3-482f-8991-fe6ad17d5e24-var-run" (OuterVolumeSpecName: "var-run") pod "88563126-2bb3-482f-8991-fe6ad17d5e24" (UID: "88563126-2bb3-482f-8991-fe6ad17d5e24"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 05 07:06:32 crc kubenswrapper[4846]: I1005 07:06:32.068239 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/88563126-2bb3-482f-8991-fe6ad17d5e24-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "88563126-2bb3-482f-8991-fe6ad17d5e24" (UID: "88563126-2bb3-482f-8991-fe6ad17d5e24"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 05 07:06:32 crc kubenswrapper[4846]: I1005 07:06:32.070226 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/88563126-2bb3-482f-8991-fe6ad17d5e24-scripts" (OuterVolumeSpecName: "scripts") pod "88563126-2bb3-482f-8991-fe6ad17d5e24" (UID: "88563126-2bb3-482f-8991-fe6ad17d5e24"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 07:06:32 crc kubenswrapper[4846]: I1005 07:06:32.073538 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/88563126-2bb3-482f-8991-fe6ad17d5e24-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "88563126-2bb3-482f-8991-fe6ad17d5e24" (UID: "88563126-2bb3-482f-8991-fe6ad17d5e24"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 07:06:32 crc kubenswrapper[4846]: I1005 07:06:32.088615 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/88563126-2bb3-482f-8991-fe6ad17d5e24-kube-api-access-lblrn" (OuterVolumeSpecName: "kube-api-access-lblrn") pod "88563126-2bb3-482f-8991-fe6ad17d5e24" (UID: "88563126-2bb3-482f-8991-fe6ad17d5e24"). InnerVolumeSpecName "kube-api-access-lblrn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:06:32 crc kubenswrapper[4846]: I1005 07:06:32.105913 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd630f2e-4b0e-4cc7-a0de-4f60c72d941c-kube-api-access-qd9q7" (OuterVolumeSpecName: "kube-api-access-qd9q7") pod "bd630f2e-4b0e-4cc7-a0de-4f60c72d941c" (UID: "bd630f2e-4b0e-4cc7-a0de-4f60c72d941c"). InnerVolumeSpecName "kube-api-access-qd9q7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:06:32 crc kubenswrapper[4846]: I1005 07:06:32.170575 4846 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/88563126-2bb3-482f-8991-fe6ad17d5e24-var-run\") on node \"crc\" DevicePath \"\"" Oct 05 07:06:32 crc kubenswrapper[4846]: I1005 07:06:32.170611 4846 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/88563126-2bb3-482f-8991-fe6ad17d5e24-var-log-ovn\") on node \"crc\" DevicePath \"\"" Oct 05 07:06:32 crc kubenswrapper[4846]: I1005 07:06:32.170642 4846 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/88563126-2bb3-482f-8991-fe6ad17d5e24-scripts\") on node \"crc\" DevicePath \"\"" Oct 05 07:06:32 crc kubenswrapper[4846]: I1005 07:06:32.170657 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lblrn\" (UniqueName: \"kubernetes.io/projected/88563126-2bb3-482f-8991-fe6ad17d5e24-kube-api-access-lblrn\") on node \"crc\" DevicePath \"\"" Oct 05 07:06:32 crc kubenswrapper[4846]: I1005 07:06:32.170668 4846 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/88563126-2bb3-482f-8991-fe6ad17d5e24-var-run-ovn\") on node \"crc\" DevicePath \"\"" Oct 05 07:06:32 crc kubenswrapper[4846]: I1005 07:06:32.170679 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qd9q7\" (UniqueName: \"kubernetes.io/projected/bd630f2e-4b0e-4cc7-a0de-4f60c72d941c-kube-api-access-qd9q7\") on node \"crc\" DevicePath \"\"" Oct 05 07:06:32 crc kubenswrapper[4846]: I1005 07:06:32.170688 4846 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/88563126-2bb3-482f-8991-fe6ad17d5e24-additional-scripts\") on node \"crc\" DevicePath \"\"" Oct 05 07:06:32 crc kubenswrapper[4846]: I1005 07:06:32.527428 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-4kqnw-config-dhsjv" event={"ID":"88563126-2bb3-482f-8991-fe6ad17d5e24","Type":"ContainerDied","Data":"bbefd4f1d3fb183d71bbefd3781d215382045a3ded6df8a5d5a21631e8733e3f"} Oct 05 07:06:32 crc kubenswrapper[4846]: I1005 07:06:32.527480 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bbefd4f1d3fb183d71bbefd3781d215382045a3ded6df8a5d5a21631e8733e3f" Oct 05 07:06:32 crc kubenswrapper[4846]: I1005 07:06:32.527549 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-4kqnw-config-dhsjv" Oct 05 07:06:32 crc kubenswrapper[4846]: I1005 07:06:32.531863 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-bc5f-account-create-spj8j" event={"ID":"bd630f2e-4b0e-4cc7-a0de-4f60c72d941c","Type":"ContainerDied","Data":"afe27f5f42bf600f435a71bbeb762ebc05f11aef89f8cc97f3dc276c23c1e440"} Oct 05 07:06:32 crc kubenswrapper[4846]: I1005 07:06:32.531920 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="afe27f5f42bf600f435a71bbeb762ebc05f11aef89f8cc97f3dc276c23c1e440" Oct 05 07:06:32 crc kubenswrapper[4846]: I1005 07:06:32.531935 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-bc5f-account-create-spj8j" Oct 05 07:06:32 crc kubenswrapper[4846]: I1005 07:06:32.941085 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-4kqnw" Oct 05 07:06:33 crc kubenswrapper[4846]: I1005 07:06:33.109473 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-4kqnw-config-dhsjv"] Oct 05 07:06:33 crc kubenswrapper[4846]: I1005 07:06:33.114392 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-4kqnw-config-dhsjv"] Oct 05 07:06:33 crc kubenswrapper[4846]: I1005 07:06:33.160329 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-4kqnw-config-mzd79"] Oct 05 07:06:33 crc kubenswrapper[4846]: E1005 07:06:33.160745 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bd630f2e-4b0e-4cc7-a0de-4f60c72d941c" containerName="mariadb-account-create" Oct 05 07:06:33 crc kubenswrapper[4846]: I1005 07:06:33.160764 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="bd630f2e-4b0e-4cc7-a0de-4f60c72d941c" containerName="mariadb-account-create" Oct 05 07:06:33 crc kubenswrapper[4846]: E1005 07:06:33.160788 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="88563126-2bb3-482f-8991-fe6ad17d5e24" containerName="ovn-config" Oct 05 07:06:33 crc kubenswrapper[4846]: I1005 07:06:33.160794 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="88563126-2bb3-482f-8991-fe6ad17d5e24" containerName="ovn-config" Oct 05 07:06:33 crc kubenswrapper[4846]: E1005 07:06:33.160803 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7fd963e7-aa89-4a46-8b1b-ac1bd6bc503d" containerName="mariadb-account-create" Oct 05 07:06:33 crc kubenswrapper[4846]: I1005 07:06:33.160810 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="7fd963e7-aa89-4a46-8b1b-ac1bd6bc503d" containerName="mariadb-account-create" Oct 05 07:06:33 crc kubenswrapper[4846]: E1005 07:06:33.160821 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="68ba5901-a360-4343-8e70-c7e1c9e829f3" containerName="mariadb-account-create" Oct 05 07:06:33 crc kubenswrapper[4846]: I1005 07:06:33.160828 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="68ba5901-a360-4343-8e70-c7e1c9e829f3" containerName="mariadb-account-create" Oct 05 07:06:33 crc kubenswrapper[4846]: I1005 07:06:33.160997 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="68ba5901-a360-4343-8e70-c7e1c9e829f3" containerName="mariadb-account-create" Oct 05 07:06:33 crc kubenswrapper[4846]: I1005 07:06:33.161015 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="88563126-2bb3-482f-8991-fe6ad17d5e24" containerName="ovn-config" Oct 05 07:06:33 crc kubenswrapper[4846]: I1005 07:06:33.161033 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="bd630f2e-4b0e-4cc7-a0de-4f60c72d941c" containerName="mariadb-account-create" Oct 05 07:06:33 crc kubenswrapper[4846]: I1005 07:06:33.161046 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="7fd963e7-aa89-4a46-8b1b-ac1bd6bc503d" containerName="mariadb-account-create" Oct 05 07:06:33 crc kubenswrapper[4846]: I1005 07:06:33.161625 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-4kqnw-config-mzd79" Oct 05 07:06:33 crc kubenswrapper[4846]: I1005 07:06:33.169600 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Oct 05 07:06:33 crc kubenswrapper[4846]: I1005 07:06:33.192711 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b3001e94-7ed4-4f76-a0b5-882965cba150-scripts\") pod \"ovn-controller-4kqnw-config-mzd79\" (UID: \"b3001e94-7ed4-4f76-a0b5-882965cba150\") " pod="openstack/ovn-controller-4kqnw-config-mzd79" Oct 05 07:06:33 crc kubenswrapper[4846]: I1005 07:06:33.192783 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/b3001e94-7ed4-4f76-a0b5-882965cba150-var-run\") pod \"ovn-controller-4kqnw-config-mzd79\" (UID: \"b3001e94-7ed4-4f76-a0b5-882965cba150\") " pod="openstack/ovn-controller-4kqnw-config-mzd79" Oct 05 07:06:33 crc kubenswrapper[4846]: I1005 07:06:33.192812 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/b3001e94-7ed4-4f76-a0b5-882965cba150-var-log-ovn\") pod \"ovn-controller-4kqnw-config-mzd79\" (UID: \"b3001e94-7ed4-4f76-a0b5-882965cba150\") " pod="openstack/ovn-controller-4kqnw-config-mzd79" Oct 05 07:06:33 crc kubenswrapper[4846]: I1005 07:06:33.192838 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/b3001e94-7ed4-4f76-a0b5-882965cba150-additional-scripts\") pod \"ovn-controller-4kqnw-config-mzd79\" (UID: \"b3001e94-7ed4-4f76-a0b5-882965cba150\") " pod="openstack/ovn-controller-4kqnw-config-mzd79" Oct 05 07:06:33 crc kubenswrapper[4846]: I1005 07:06:33.192890 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5kw4q\" (UniqueName: \"kubernetes.io/projected/b3001e94-7ed4-4f76-a0b5-882965cba150-kube-api-access-5kw4q\") pod \"ovn-controller-4kqnw-config-mzd79\" (UID: \"b3001e94-7ed4-4f76-a0b5-882965cba150\") " pod="openstack/ovn-controller-4kqnw-config-mzd79" Oct 05 07:06:33 crc kubenswrapper[4846]: I1005 07:06:33.192943 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/b3001e94-7ed4-4f76-a0b5-882965cba150-var-run-ovn\") pod \"ovn-controller-4kqnw-config-mzd79\" (UID: \"b3001e94-7ed4-4f76-a0b5-882965cba150\") " pod="openstack/ovn-controller-4kqnw-config-mzd79" Oct 05 07:06:33 crc kubenswrapper[4846]: I1005 07:06:33.212426 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-4kqnw-config-mzd79"] Oct 05 07:06:33 crc kubenswrapper[4846]: I1005 07:06:33.295482 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/b3001e94-7ed4-4f76-a0b5-882965cba150-var-run\") pod \"ovn-controller-4kqnw-config-mzd79\" (UID: \"b3001e94-7ed4-4f76-a0b5-882965cba150\") " pod="openstack/ovn-controller-4kqnw-config-mzd79" Oct 05 07:06:33 crc kubenswrapper[4846]: I1005 07:06:33.295556 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/b3001e94-7ed4-4f76-a0b5-882965cba150-var-log-ovn\") pod \"ovn-controller-4kqnw-config-mzd79\" (UID: \"b3001e94-7ed4-4f76-a0b5-882965cba150\") " pod="openstack/ovn-controller-4kqnw-config-mzd79" Oct 05 07:06:33 crc kubenswrapper[4846]: I1005 07:06:33.295594 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/b3001e94-7ed4-4f76-a0b5-882965cba150-additional-scripts\") pod \"ovn-controller-4kqnw-config-mzd79\" (UID: \"b3001e94-7ed4-4f76-a0b5-882965cba150\") " pod="openstack/ovn-controller-4kqnw-config-mzd79" Oct 05 07:06:33 crc kubenswrapper[4846]: I1005 07:06:33.295656 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5kw4q\" (UniqueName: \"kubernetes.io/projected/b3001e94-7ed4-4f76-a0b5-882965cba150-kube-api-access-5kw4q\") pod \"ovn-controller-4kqnw-config-mzd79\" (UID: \"b3001e94-7ed4-4f76-a0b5-882965cba150\") " pod="openstack/ovn-controller-4kqnw-config-mzd79" Oct 05 07:06:33 crc kubenswrapper[4846]: I1005 07:06:33.295707 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/b3001e94-7ed4-4f76-a0b5-882965cba150-var-run-ovn\") pod \"ovn-controller-4kqnw-config-mzd79\" (UID: \"b3001e94-7ed4-4f76-a0b5-882965cba150\") " pod="openstack/ovn-controller-4kqnw-config-mzd79" Oct 05 07:06:33 crc kubenswrapper[4846]: I1005 07:06:33.295729 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b3001e94-7ed4-4f76-a0b5-882965cba150-scripts\") pod \"ovn-controller-4kqnw-config-mzd79\" (UID: \"b3001e94-7ed4-4f76-a0b5-882965cba150\") " pod="openstack/ovn-controller-4kqnw-config-mzd79" Oct 05 07:06:33 crc kubenswrapper[4846]: I1005 07:06:33.296520 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/b3001e94-7ed4-4f76-a0b5-882965cba150-var-log-ovn\") pod \"ovn-controller-4kqnw-config-mzd79\" (UID: \"b3001e94-7ed4-4f76-a0b5-882965cba150\") " pod="openstack/ovn-controller-4kqnw-config-mzd79" Oct 05 07:06:33 crc kubenswrapper[4846]: I1005 07:06:33.296576 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/b3001e94-7ed4-4f76-a0b5-882965cba150-var-run-ovn\") pod \"ovn-controller-4kqnw-config-mzd79\" (UID: \"b3001e94-7ed4-4f76-a0b5-882965cba150\") " pod="openstack/ovn-controller-4kqnw-config-mzd79" Oct 05 07:06:33 crc kubenswrapper[4846]: I1005 07:06:33.296571 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/b3001e94-7ed4-4f76-a0b5-882965cba150-var-run\") pod \"ovn-controller-4kqnw-config-mzd79\" (UID: \"b3001e94-7ed4-4f76-a0b5-882965cba150\") " pod="openstack/ovn-controller-4kqnw-config-mzd79" Oct 05 07:06:33 crc kubenswrapper[4846]: I1005 07:06:33.297228 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/b3001e94-7ed4-4f76-a0b5-882965cba150-additional-scripts\") pod \"ovn-controller-4kqnw-config-mzd79\" (UID: \"b3001e94-7ed4-4f76-a0b5-882965cba150\") " pod="openstack/ovn-controller-4kqnw-config-mzd79" Oct 05 07:06:33 crc kubenswrapper[4846]: I1005 07:06:33.298231 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b3001e94-7ed4-4f76-a0b5-882965cba150-scripts\") pod \"ovn-controller-4kqnw-config-mzd79\" (UID: \"b3001e94-7ed4-4f76-a0b5-882965cba150\") " pod="openstack/ovn-controller-4kqnw-config-mzd79" Oct 05 07:06:33 crc kubenswrapper[4846]: I1005 07:06:33.321068 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5kw4q\" (UniqueName: \"kubernetes.io/projected/b3001e94-7ed4-4f76-a0b5-882965cba150-kube-api-access-5kw4q\") pod \"ovn-controller-4kqnw-config-mzd79\" (UID: \"b3001e94-7ed4-4f76-a0b5-882965cba150\") " pod="openstack/ovn-controller-4kqnw-config-mzd79" Oct 05 07:06:33 crc kubenswrapper[4846]: I1005 07:06:33.499343 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/62d2ef82-9372-4b55-990e-30eb8e16e439-etc-swift\") pod \"swift-storage-0\" (UID: \"62d2ef82-9372-4b55-990e-30eb8e16e439\") " pod="openstack/swift-storage-0" Oct 05 07:06:33 crc kubenswrapper[4846]: I1005 07:06:33.504725 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/62d2ef82-9372-4b55-990e-30eb8e16e439-etc-swift\") pod \"swift-storage-0\" (UID: \"62d2ef82-9372-4b55-990e-30eb8e16e439\") " pod="openstack/swift-storage-0" Oct 05 07:06:33 crc kubenswrapper[4846]: I1005 07:06:33.530011 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-4kqnw-config-mzd79" Oct 05 07:06:33 crc kubenswrapper[4846]: I1005 07:06:33.629202 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-sync-68h9n"] Oct 05 07:06:33 crc kubenswrapper[4846]: I1005 07:06:33.632103 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-68h9n" Oct 05 07:06:33 crc kubenswrapper[4846]: I1005 07:06:33.634445 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-config-data" Oct 05 07:06:33 crc kubenswrapper[4846]: I1005 07:06:33.634571 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-ff9fp" Oct 05 07:06:33 crc kubenswrapper[4846]: I1005 07:06:33.649970 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-68h9n"] Oct 05 07:06:33 crc kubenswrapper[4846]: I1005 07:06:33.665652 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Oct 05 07:06:33 crc kubenswrapper[4846]: I1005 07:06:33.702582 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/8b6863be-f8e1-4a87-93d6-5cf756983747-db-sync-config-data\") pod \"glance-db-sync-68h9n\" (UID: \"8b6863be-f8e1-4a87-93d6-5cf756983747\") " pod="openstack/glance-db-sync-68h9n" Oct 05 07:06:33 crc kubenswrapper[4846]: I1005 07:06:33.702646 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ctmvj\" (UniqueName: \"kubernetes.io/projected/8b6863be-f8e1-4a87-93d6-5cf756983747-kube-api-access-ctmvj\") pod \"glance-db-sync-68h9n\" (UID: \"8b6863be-f8e1-4a87-93d6-5cf756983747\") " pod="openstack/glance-db-sync-68h9n" Oct 05 07:06:33 crc kubenswrapper[4846]: I1005 07:06:33.702693 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8b6863be-f8e1-4a87-93d6-5cf756983747-combined-ca-bundle\") pod \"glance-db-sync-68h9n\" (UID: \"8b6863be-f8e1-4a87-93d6-5cf756983747\") " pod="openstack/glance-db-sync-68h9n" Oct 05 07:06:33 crc kubenswrapper[4846]: I1005 07:06:33.702771 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8b6863be-f8e1-4a87-93d6-5cf756983747-config-data\") pod \"glance-db-sync-68h9n\" (UID: \"8b6863be-f8e1-4a87-93d6-5cf756983747\") " pod="openstack/glance-db-sync-68h9n" Oct 05 07:06:33 crc kubenswrapper[4846]: I1005 07:06:33.807073 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8b6863be-f8e1-4a87-93d6-5cf756983747-combined-ca-bundle\") pod \"glance-db-sync-68h9n\" (UID: \"8b6863be-f8e1-4a87-93d6-5cf756983747\") " pod="openstack/glance-db-sync-68h9n" Oct 05 07:06:33 crc kubenswrapper[4846]: I1005 07:06:33.807220 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8b6863be-f8e1-4a87-93d6-5cf756983747-config-data\") pod \"glance-db-sync-68h9n\" (UID: \"8b6863be-f8e1-4a87-93d6-5cf756983747\") " pod="openstack/glance-db-sync-68h9n" Oct 05 07:06:33 crc kubenswrapper[4846]: I1005 07:06:33.807275 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/8b6863be-f8e1-4a87-93d6-5cf756983747-db-sync-config-data\") pod \"glance-db-sync-68h9n\" (UID: \"8b6863be-f8e1-4a87-93d6-5cf756983747\") " pod="openstack/glance-db-sync-68h9n" Oct 05 07:06:33 crc kubenswrapper[4846]: I1005 07:06:33.807310 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ctmvj\" (UniqueName: \"kubernetes.io/projected/8b6863be-f8e1-4a87-93d6-5cf756983747-kube-api-access-ctmvj\") pod \"glance-db-sync-68h9n\" (UID: \"8b6863be-f8e1-4a87-93d6-5cf756983747\") " pod="openstack/glance-db-sync-68h9n" Oct 05 07:06:33 crc kubenswrapper[4846]: I1005 07:06:33.817949 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/8b6863be-f8e1-4a87-93d6-5cf756983747-db-sync-config-data\") pod \"glance-db-sync-68h9n\" (UID: \"8b6863be-f8e1-4a87-93d6-5cf756983747\") " pod="openstack/glance-db-sync-68h9n" Oct 05 07:06:33 crc kubenswrapper[4846]: I1005 07:06:33.817956 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8b6863be-f8e1-4a87-93d6-5cf756983747-config-data\") pod \"glance-db-sync-68h9n\" (UID: \"8b6863be-f8e1-4a87-93d6-5cf756983747\") " pod="openstack/glance-db-sync-68h9n" Oct 05 07:06:33 crc kubenswrapper[4846]: I1005 07:06:33.818162 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8b6863be-f8e1-4a87-93d6-5cf756983747-combined-ca-bundle\") pod \"glance-db-sync-68h9n\" (UID: \"8b6863be-f8e1-4a87-93d6-5cf756983747\") " pod="openstack/glance-db-sync-68h9n" Oct 05 07:06:33 crc kubenswrapper[4846]: I1005 07:06:33.829240 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ctmvj\" (UniqueName: \"kubernetes.io/projected/8b6863be-f8e1-4a87-93d6-5cf756983747-kube-api-access-ctmvj\") pod \"glance-db-sync-68h9n\" (UID: \"8b6863be-f8e1-4a87-93d6-5cf756983747\") " pod="openstack/glance-db-sync-68h9n" Oct 05 07:06:33 crc kubenswrapper[4846]: I1005 07:06:33.866065 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-4kqnw-config-mzd79"] Oct 05 07:06:33 crc kubenswrapper[4846]: I1005 07:06:33.953098 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-68h9n" Oct 05 07:06:34 crc kubenswrapper[4846]: I1005 07:06:34.348148 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Oct 05 07:06:34 crc kubenswrapper[4846]: W1005 07:06:34.362969 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod62d2ef82_9372_4b55_990e_30eb8e16e439.slice/crio-ea57f2844ffd33c32881942ebca5ce2841d9fc7ee91c27d3b5e9b78505a3f19b WatchSource:0}: Error finding container ea57f2844ffd33c32881942ebca5ce2841d9fc7ee91c27d3b5e9b78505a3f19b: Status 404 returned error can't find the container with id ea57f2844ffd33c32881942ebca5ce2841d9fc7ee91c27d3b5e9b78505a3f19b Oct 05 07:06:34 crc kubenswrapper[4846]: I1005 07:06:34.514594 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="88563126-2bb3-482f-8991-fe6ad17d5e24" path="/var/lib/kubelet/pods/88563126-2bb3-482f-8991-fe6ad17d5e24/volumes" Oct 05 07:06:34 crc kubenswrapper[4846]: I1005 07:06:34.554714 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-68h9n"] Oct 05 07:06:34 crc kubenswrapper[4846]: I1005 07:06:34.561127 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"62d2ef82-9372-4b55-990e-30eb8e16e439","Type":"ContainerStarted","Data":"ea57f2844ffd33c32881942ebca5ce2841d9fc7ee91c27d3b5e9b78505a3f19b"} Oct 05 07:06:34 crc kubenswrapper[4846]: I1005 07:06:34.572157 4846 generic.go:334] "Generic (PLEG): container finished" podID="b3001e94-7ed4-4f76-a0b5-882965cba150" containerID="65cb7c9f9d81cbaf7f1a0cefd3a7c9535986612c6a0a836f3800e04819ebc0dc" exitCode=0 Oct 05 07:06:34 crc kubenswrapper[4846]: I1005 07:06:34.572218 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-4kqnw-config-mzd79" event={"ID":"b3001e94-7ed4-4f76-a0b5-882965cba150","Type":"ContainerDied","Data":"65cb7c9f9d81cbaf7f1a0cefd3a7c9535986612c6a0a836f3800e04819ebc0dc"} Oct 05 07:06:34 crc kubenswrapper[4846]: I1005 07:06:34.572264 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-4kqnw-config-mzd79" event={"ID":"b3001e94-7ed4-4f76-a0b5-882965cba150","Type":"ContainerStarted","Data":"145bde41923926ecebde5be7372dabb86646d244cc5b8164d6c1702f202f2d2b"} Oct 05 07:06:35 crc kubenswrapper[4846]: I1005 07:06:35.307398 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Oct 05 07:06:35 crc kubenswrapper[4846]: I1005 07:06:35.606151 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-68h9n" event={"ID":"8b6863be-f8e1-4a87-93d6-5cf756983747","Type":"ContainerStarted","Data":"c30f24f2f258f70f53877a7c663c371c329d5e9a5c2706e611af8bd107bfcd8f"} Oct 05 07:06:35 crc kubenswrapper[4846]: I1005 07:06:35.687934 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-create-rlqvn"] Oct 05 07:06:35 crc kubenswrapper[4846]: I1005 07:06:35.689354 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-rlqvn" Oct 05 07:06:35 crc kubenswrapper[4846]: I1005 07:06:35.696825 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-rlqvn"] Oct 05 07:06:35 crc kubenswrapper[4846]: I1005 07:06:35.793365 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-create-xp6rt"] Oct 05 07:06:35 crc kubenswrapper[4846]: I1005 07:06:35.794466 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-xp6rt" Oct 05 07:06:35 crc kubenswrapper[4846]: I1005 07:06:35.817250 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-xp6rt"] Oct 05 07:06:35 crc kubenswrapper[4846]: I1005 07:06:35.854623 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8brqm\" (UniqueName: \"kubernetes.io/projected/00746abe-43f9-4275-bf0c-7f095b71e9bc-kube-api-access-8brqm\") pod \"cinder-db-create-rlqvn\" (UID: \"00746abe-43f9-4275-bf0c-7f095b71e9bc\") " pod="openstack/cinder-db-create-rlqvn" Oct 05 07:06:35 crc kubenswrapper[4846]: I1005 07:06:35.911463 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Oct 05 07:06:35 crc kubenswrapper[4846]: I1005 07:06:35.956057 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w4s87\" (UniqueName: \"kubernetes.io/projected/6a5f9a12-316d-42cf-a592-e418d03e0849-kube-api-access-w4s87\") pod \"barbican-db-create-xp6rt\" (UID: \"6a5f9a12-316d-42cf-a592-e418d03e0849\") " pod="openstack/barbican-db-create-xp6rt" Oct 05 07:06:35 crc kubenswrapper[4846]: I1005 07:06:35.956148 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8brqm\" (UniqueName: \"kubernetes.io/projected/00746abe-43f9-4275-bf0c-7f095b71e9bc-kube-api-access-8brqm\") pod \"cinder-db-create-rlqvn\" (UID: \"00746abe-43f9-4275-bf0c-7f095b71e9bc\") " pod="openstack/cinder-db-create-rlqvn" Oct 05 07:06:35 crc kubenswrapper[4846]: I1005 07:06:35.994936 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-sync-t7t67"] Oct 05 07:06:36 crc kubenswrapper[4846]: I1005 07:06:36.001838 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-t7t67" Oct 05 07:06:36 crc kubenswrapper[4846]: I1005 07:06:36.005649 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 05 07:06:36 crc kubenswrapper[4846]: I1005 07:06:36.005846 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 05 07:06:36 crc kubenswrapper[4846]: I1005 07:06:36.005888 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-76d8v" Oct 05 07:06:36 crc kubenswrapper[4846]: I1005 07:06:36.005987 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 05 07:06:36 crc kubenswrapper[4846]: I1005 07:06:36.021347 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-t7t67"] Oct 05 07:06:36 crc kubenswrapper[4846]: I1005 07:06:36.025383 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8brqm\" (UniqueName: \"kubernetes.io/projected/00746abe-43f9-4275-bf0c-7f095b71e9bc-kube-api-access-8brqm\") pod \"cinder-db-create-rlqvn\" (UID: \"00746abe-43f9-4275-bf0c-7f095b71e9bc\") " pod="openstack/cinder-db-create-rlqvn" Oct 05 07:06:36 crc kubenswrapper[4846]: I1005 07:06:36.059008 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w4s87\" (UniqueName: \"kubernetes.io/projected/6a5f9a12-316d-42cf-a592-e418d03e0849-kube-api-access-w4s87\") pod \"barbican-db-create-xp6rt\" (UID: \"6a5f9a12-316d-42cf-a592-e418d03e0849\") " pod="openstack/barbican-db-create-xp6rt" Oct 05 07:06:36 crc kubenswrapper[4846]: I1005 07:06:36.100927 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w4s87\" (UniqueName: \"kubernetes.io/projected/6a5f9a12-316d-42cf-a592-e418d03e0849-kube-api-access-w4s87\") pod \"barbican-db-create-xp6rt\" (UID: \"6a5f9a12-316d-42cf-a592-e418d03e0849\") " pod="openstack/barbican-db-create-xp6rt" Oct 05 07:06:36 crc kubenswrapper[4846]: I1005 07:06:36.107582 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-create-tbdm6"] Oct 05 07:06:36 crc kubenswrapper[4846]: I1005 07:06:36.109094 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-tbdm6" Oct 05 07:06:36 crc kubenswrapper[4846]: I1005 07:06:36.127811 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-tbdm6"] Oct 05 07:06:36 crc kubenswrapper[4846]: I1005 07:06:36.137481 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-xp6rt" Oct 05 07:06:36 crc kubenswrapper[4846]: I1005 07:06:36.161262 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b4b973ee-209f-477b-92ee-3f1c16f17f43-config-data\") pod \"keystone-db-sync-t7t67\" (UID: \"b4b973ee-209f-477b-92ee-3f1c16f17f43\") " pod="openstack/keystone-db-sync-t7t67" Oct 05 07:06:36 crc kubenswrapper[4846]: I1005 07:06:36.161313 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b4b973ee-209f-477b-92ee-3f1c16f17f43-combined-ca-bundle\") pod \"keystone-db-sync-t7t67\" (UID: \"b4b973ee-209f-477b-92ee-3f1c16f17f43\") " pod="openstack/keystone-db-sync-t7t67" Oct 05 07:06:36 crc kubenswrapper[4846]: I1005 07:06:36.161423 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5dsnc\" (UniqueName: \"kubernetes.io/projected/b4b973ee-209f-477b-92ee-3f1c16f17f43-kube-api-access-5dsnc\") pod \"keystone-db-sync-t7t67\" (UID: \"b4b973ee-209f-477b-92ee-3f1c16f17f43\") " pod="openstack/keystone-db-sync-t7t67" Oct 05 07:06:36 crc kubenswrapper[4846]: I1005 07:06:36.263969 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b4b973ee-209f-477b-92ee-3f1c16f17f43-config-data\") pod \"keystone-db-sync-t7t67\" (UID: \"b4b973ee-209f-477b-92ee-3f1c16f17f43\") " pod="openstack/keystone-db-sync-t7t67" Oct 05 07:06:36 crc kubenswrapper[4846]: I1005 07:06:36.264056 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wn5rr\" (UniqueName: \"kubernetes.io/projected/fb35a7c4-9453-4ce8-b0c0-828893100c92-kube-api-access-wn5rr\") pod \"neutron-db-create-tbdm6\" (UID: \"fb35a7c4-9453-4ce8-b0c0-828893100c92\") " pod="openstack/neutron-db-create-tbdm6" Oct 05 07:06:36 crc kubenswrapper[4846]: I1005 07:06:36.264089 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b4b973ee-209f-477b-92ee-3f1c16f17f43-combined-ca-bundle\") pod \"keystone-db-sync-t7t67\" (UID: \"b4b973ee-209f-477b-92ee-3f1c16f17f43\") " pod="openstack/keystone-db-sync-t7t67" Oct 05 07:06:36 crc kubenswrapper[4846]: I1005 07:06:36.264216 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5dsnc\" (UniqueName: \"kubernetes.io/projected/b4b973ee-209f-477b-92ee-3f1c16f17f43-kube-api-access-5dsnc\") pod \"keystone-db-sync-t7t67\" (UID: \"b4b973ee-209f-477b-92ee-3f1c16f17f43\") " pod="openstack/keystone-db-sync-t7t67" Oct 05 07:06:36 crc kubenswrapper[4846]: I1005 07:06:36.270134 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b4b973ee-209f-477b-92ee-3f1c16f17f43-combined-ca-bundle\") pod \"keystone-db-sync-t7t67\" (UID: \"b4b973ee-209f-477b-92ee-3f1c16f17f43\") " pod="openstack/keystone-db-sync-t7t67" Oct 05 07:06:36 crc kubenswrapper[4846]: I1005 07:06:36.270814 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b4b973ee-209f-477b-92ee-3f1c16f17f43-config-data\") pod \"keystone-db-sync-t7t67\" (UID: \"b4b973ee-209f-477b-92ee-3f1c16f17f43\") " pod="openstack/keystone-db-sync-t7t67" Oct 05 07:06:36 crc kubenswrapper[4846]: I1005 07:06:36.294639 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5dsnc\" (UniqueName: \"kubernetes.io/projected/b4b973ee-209f-477b-92ee-3f1c16f17f43-kube-api-access-5dsnc\") pod \"keystone-db-sync-t7t67\" (UID: \"b4b973ee-209f-477b-92ee-3f1c16f17f43\") " pod="openstack/keystone-db-sync-t7t67" Oct 05 07:06:36 crc kubenswrapper[4846]: I1005 07:06:36.309415 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-rlqvn" Oct 05 07:06:36 crc kubenswrapper[4846]: I1005 07:06:36.365587 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-t7t67" Oct 05 07:06:36 crc kubenswrapper[4846]: I1005 07:06:36.366206 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wn5rr\" (UniqueName: \"kubernetes.io/projected/fb35a7c4-9453-4ce8-b0c0-828893100c92-kube-api-access-wn5rr\") pod \"neutron-db-create-tbdm6\" (UID: \"fb35a7c4-9453-4ce8-b0c0-828893100c92\") " pod="openstack/neutron-db-create-tbdm6" Oct 05 07:06:36 crc kubenswrapper[4846]: I1005 07:06:36.395391 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wn5rr\" (UniqueName: \"kubernetes.io/projected/fb35a7c4-9453-4ce8-b0c0-828893100c92-kube-api-access-wn5rr\") pod \"neutron-db-create-tbdm6\" (UID: \"fb35a7c4-9453-4ce8-b0c0-828893100c92\") " pod="openstack/neutron-db-create-tbdm6" Oct 05 07:06:36 crc kubenswrapper[4846]: I1005 07:06:36.435330 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-tbdm6" Oct 05 07:06:36 crc kubenswrapper[4846]: I1005 07:06:36.508332 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-4kqnw-config-mzd79" Oct 05 07:06:36 crc kubenswrapper[4846]: I1005 07:06:36.573172 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/b3001e94-7ed4-4f76-a0b5-882965cba150-var-log-ovn\") pod \"b3001e94-7ed4-4f76-a0b5-882965cba150\" (UID: \"b3001e94-7ed4-4f76-a0b5-882965cba150\") " Oct 05 07:06:36 crc kubenswrapper[4846]: I1005 07:06:36.574455 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b3001e94-7ed4-4f76-a0b5-882965cba150-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "b3001e94-7ed4-4f76-a0b5-882965cba150" (UID: "b3001e94-7ed4-4f76-a0b5-882965cba150"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 05 07:06:36 crc kubenswrapper[4846]: I1005 07:06:36.576597 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b3001e94-7ed4-4f76-a0b5-882965cba150-scripts\") pod \"b3001e94-7ed4-4f76-a0b5-882965cba150\" (UID: \"b3001e94-7ed4-4f76-a0b5-882965cba150\") " Oct 05 07:06:36 crc kubenswrapper[4846]: I1005 07:06:36.576796 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/b3001e94-7ed4-4f76-a0b5-882965cba150-var-run\") pod \"b3001e94-7ed4-4f76-a0b5-882965cba150\" (UID: \"b3001e94-7ed4-4f76-a0b5-882965cba150\") " Oct 05 07:06:36 crc kubenswrapper[4846]: I1005 07:06:36.577478 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/b3001e94-7ed4-4f76-a0b5-882965cba150-var-run-ovn\") pod \"b3001e94-7ed4-4f76-a0b5-882965cba150\" (UID: \"b3001e94-7ed4-4f76-a0b5-882965cba150\") " Oct 05 07:06:36 crc kubenswrapper[4846]: I1005 07:06:36.577664 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/b3001e94-7ed4-4f76-a0b5-882965cba150-additional-scripts\") pod \"b3001e94-7ed4-4f76-a0b5-882965cba150\" (UID: \"b3001e94-7ed4-4f76-a0b5-882965cba150\") " Oct 05 07:06:36 crc kubenswrapper[4846]: I1005 07:06:36.577776 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5kw4q\" (UniqueName: \"kubernetes.io/projected/b3001e94-7ed4-4f76-a0b5-882965cba150-kube-api-access-5kw4q\") pod \"b3001e94-7ed4-4f76-a0b5-882965cba150\" (UID: \"b3001e94-7ed4-4f76-a0b5-882965cba150\") " Oct 05 07:06:36 crc kubenswrapper[4846]: I1005 07:06:36.578824 4846 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/b3001e94-7ed4-4f76-a0b5-882965cba150-var-log-ovn\") on node \"crc\" DevicePath \"\"" Oct 05 07:06:36 crc kubenswrapper[4846]: I1005 07:06:36.577258 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b3001e94-7ed4-4f76-a0b5-882965cba150-var-run" (OuterVolumeSpecName: "var-run") pod "b3001e94-7ed4-4f76-a0b5-882965cba150" (UID: "b3001e94-7ed4-4f76-a0b5-882965cba150"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 05 07:06:36 crc kubenswrapper[4846]: I1005 07:06:36.578901 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b3001e94-7ed4-4f76-a0b5-882965cba150-scripts" (OuterVolumeSpecName: "scripts") pod "b3001e94-7ed4-4f76-a0b5-882965cba150" (UID: "b3001e94-7ed4-4f76-a0b5-882965cba150"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 07:06:36 crc kubenswrapper[4846]: I1005 07:06:36.579469 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b3001e94-7ed4-4f76-a0b5-882965cba150-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "b3001e94-7ed4-4f76-a0b5-882965cba150" (UID: "b3001e94-7ed4-4f76-a0b5-882965cba150"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 05 07:06:36 crc kubenswrapper[4846]: I1005 07:06:36.580032 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b3001e94-7ed4-4f76-a0b5-882965cba150-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "b3001e94-7ed4-4f76-a0b5-882965cba150" (UID: "b3001e94-7ed4-4f76-a0b5-882965cba150"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 07:06:36 crc kubenswrapper[4846]: I1005 07:06:36.583875 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b3001e94-7ed4-4f76-a0b5-882965cba150-kube-api-access-5kw4q" (OuterVolumeSpecName: "kube-api-access-5kw4q") pod "b3001e94-7ed4-4f76-a0b5-882965cba150" (UID: "b3001e94-7ed4-4f76-a0b5-882965cba150"). InnerVolumeSpecName "kube-api-access-5kw4q". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:06:36 crc kubenswrapper[4846]: I1005 07:06:36.621532 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-4kqnw-config-mzd79" event={"ID":"b3001e94-7ed4-4f76-a0b5-882965cba150","Type":"ContainerDied","Data":"145bde41923926ecebde5be7372dabb86646d244cc5b8164d6c1702f202f2d2b"} Oct 05 07:06:36 crc kubenswrapper[4846]: I1005 07:06:36.621589 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-4kqnw-config-mzd79" Oct 05 07:06:36 crc kubenswrapper[4846]: I1005 07:06:36.621601 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="145bde41923926ecebde5be7372dabb86646d244cc5b8164d6c1702f202f2d2b" Oct 05 07:06:36 crc kubenswrapper[4846]: I1005 07:06:36.680898 4846 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b3001e94-7ed4-4f76-a0b5-882965cba150-scripts\") on node \"crc\" DevicePath \"\"" Oct 05 07:06:36 crc kubenswrapper[4846]: I1005 07:06:36.680932 4846 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/b3001e94-7ed4-4f76-a0b5-882965cba150-var-run\") on node \"crc\" DevicePath \"\"" Oct 05 07:06:36 crc kubenswrapper[4846]: I1005 07:06:36.680944 4846 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/b3001e94-7ed4-4f76-a0b5-882965cba150-var-run-ovn\") on node \"crc\" DevicePath \"\"" Oct 05 07:06:36 crc kubenswrapper[4846]: I1005 07:06:36.680957 4846 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/b3001e94-7ed4-4f76-a0b5-882965cba150-additional-scripts\") on node \"crc\" DevicePath \"\"" Oct 05 07:06:36 crc kubenswrapper[4846]: I1005 07:06:36.680970 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5kw4q\" (UniqueName: \"kubernetes.io/projected/b3001e94-7ed4-4f76-a0b5-882965cba150-kube-api-access-5kw4q\") on node \"crc\" DevicePath \"\"" Oct 05 07:06:37 crc kubenswrapper[4846]: I1005 07:06:37.029057 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-xp6rt"] Oct 05 07:06:37 crc kubenswrapper[4846]: I1005 07:06:37.148319 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-tbdm6"] Oct 05 07:06:37 crc kubenswrapper[4846]: I1005 07:06:37.290763 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-rlqvn"] Oct 05 07:06:37 crc kubenswrapper[4846]: W1005 07:06:37.339110 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod00746abe_43f9_4275_bf0c_7f095b71e9bc.slice/crio-2d00432244837f455705b2ad8f4ffb330bdb64892b198f987bba9afcbf88cff5 WatchSource:0}: Error finding container 2d00432244837f455705b2ad8f4ffb330bdb64892b198f987bba9afcbf88cff5: Status 404 returned error can't find the container with id 2d00432244837f455705b2ad8f4ffb330bdb64892b198f987bba9afcbf88cff5 Oct 05 07:06:37 crc kubenswrapper[4846]: I1005 07:06:37.339926 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-t7t67"] Oct 05 07:06:37 crc kubenswrapper[4846]: I1005 07:06:37.663972 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-rlqvn" event={"ID":"00746abe-43f9-4275-bf0c-7f095b71e9bc","Type":"ContainerStarted","Data":"2d00432244837f455705b2ad8f4ffb330bdb64892b198f987bba9afcbf88cff5"} Oct 05 07:06:37 crc kubenswrapper[4846]: I1005 07:06:37.668773 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-xp6rt" event={"ID":"6a5f9a12-316d-42cf-a592-e418d03e0849","Type":"ContainerStarted","Data":"7754148ace05ad06ee097e13afd91ea75505f2f443534d88a22bc6542491436e"} Oct 05 07:06:37 crc kubenswrapper[4846]: I1005 07:06:37.680254 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-4kqnw-config-mzd79"] Oct 05 07:06:37 crc kubenswrapper[4846]: I1005 07:06:37.691388 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-4kqnw-config-mzd79"] Oct 05 07:06:37 crc kubenswrapper[4846]: I1005 07:06:37.698892 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"62d2ef82-9372-4b55-990e-30eb8e16e439","Type":"ContainerStarted","Data":"888a361d7601c9ec3ffa1d314d1a426b937569ba6430ad0b3420461b118db555"} Oct 05 07:06:37 crc kubenswrapper[4846]: I1005 07:06:37.698935 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"62d2ef82-9372-4b55-990e-30eb8e16e439","Type":"ContainerStarted","Data":"a4edeadbef779fa82267d6b817cf8a19cf59e7146c6c1a8f5c786ad4fea85d4d"} Oct 05 07:06:37 crc kubenswrapper[4846]: I1005 07:06:37.707291 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-t7t67" event={"ID":"b4b973ee-209f-477b-92ee-3f1c16f17f43","Type":"ContainerStarted","Data":"a3daa5a064fa48e66f1d8cc155b44ee62baf8fe1a2cd398a5df3a476e3e992c1"} Oct 05 07:06:37 crc kubenswrapper[4846]: I1005 07:06:37.725430 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-tbdm6" event={"ID":"fb35a7c4-9453-4ce8-b0c0-828893100c92","Type":"ContainerStarted","Data":"1b0a90c1a196224fe2bb54e5475dac1694e64355d8a28b1d4ff6cd37263a50e3"} Oct 05 07:06:37 crc kubenswrapper[4846]: I1005 07:06:37.896038 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-4kqnw-config-rm2h8"] Oct 05 07:06:37 crc kubenswrapper[4846]: E1005 07:06:37.896473 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b3001e94-7ed4-4f76-a0b5-882965cba150" containerName="ovn-config" Oct 05 07:06:37 crc kubenswrapper[4846]: I1005 07:06:37.896491 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="b3001e94-7ed4-4f76-a0b5-882965cba150" containerName="ovn-config" Oct 05 07:06:37 crc kubenswrapper[4846]: I1005 07:06:37.896716 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="b3001e94-7ed4-4f76-a0b5-882965cba150" containerName="ovn-config" Oct 05 07:06:37 crc kubenswrapper[4846]: I1005 07:06:37.897382 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-4kqnw-config-rm2h8" Oct 05 07:06:37 crc kubenswrapper[4846]: I1005 07:06:37.900623 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Oct 05 07:06:37 crc kubenswrapper[4846]: I1005 07:06:37.928263 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-4kqnw-config-rm2h8"] Oct 05 07:06:38 crc kubenswrapper[4846]: I1005 07:06:38.015882 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/9ea3c542-09d7-46ee-942f-a0476f7609d1-var-run-ovn\") pod \"ovn-controller-4kqnw-config-rm2h8\" (UID: \"9ea3c542-09d7-46ee-942f-a0476f7609d1\") " pod="openstack/ovn-controller-4kqnw-config-rm2h8" Oct 05 07:06:38 crc kubenswrapper[4846]: I1005 07:06:38.015933 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/9ea3c542-09d7-46ee-942f-a0476f7609d1-additional-scripts\") pod \"ovn-controller-4kqnw-config-rm2h8\" (UID: \"9ea3c542-09d7-46ee-942f-a0476f7609d1\") " pod="openstack/ovn-controller-4kqnw-config-rm2h8" Oct 05 07:06:38 crc kubenswrapper[4846]: I1005 07:06:38.015973 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/9ea3c542-09d7-46ee-942f-a0476f7609d1-var-log-ovn\") pod \"ovn-controller-4kqnw-config-rm2h8\" (UID: \"9ea3c542-09d7-46ee-942f-a0476f7609d1\") " pod="openstack/ovn-controller-4kqnw-config-rm2h8" Oct 05 07:06:38 crc kubenswrapper[4846]: I1005 07:06:38.015995 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/9ea3c542-09d7-46ee-942f-a0476f7609d1-var-run\") pod \"ovn-controller-4kqnw-config-rm2h8\" (UID: \"9ea3c542-09d7-46ee-942f-a0476f7609d1\") " pod="openstack/ovn-controller-4kqnw-config-rm2h8" Oct 05 07:06:38 crc kubenswrapper[4846]: I1005 07:06:38.016014 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qmwv4\" (UniqueName: \"kubernetes.io/projected/9ea3c542-09d7-46ee-942f-a0476f7609d1-kube-api-access-qmwv4\") pod \"ovn-controller-4kqnw-config-rm2h8\" (UID: \"9ea3c542-09d7-46ee-942f-a0476f7609d1\") " pod="openstack/ovn-controller-4kqnw-config-rm2h8" Oct 05 07:06:38 crc kubenswrapper[4846]: I1005 07:06:38.016066 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9ea3c542-09d7-46ee-942f-a0476f7609d1-scripts\") pod \"ovn-controller-4kqnw-config-rm2h8\" (UID: \"9ea3c542-09d7-46ee-942f-a0476f7609d1\") " pod="openstack/ovn-controller-4kqnw-config-rm2h8" Oct 05 07:06:38 crc kubenswrapper[4846]: I1005 07:06:38.117623 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/9ea3c542-09d7-46ee-942f-a0476f7609d1-var-run-ovn\") pod \"ovn-controller-4kqnw-config-rm2h8\" (UID: \"9ea3c542-09d7-46ee-942f-a0476f7609d1\") " pod="openstack/ovn-controller-4kqnw-config-rm2h8" Oct 05 07:06:38 crc kubenswrapper[4846]: I1005 07:06:38.117679 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/9ea3c542-09d7-46ee-942f-a0476f7609d1-additional-scripts\") pod \"ovn-controller-4kqnw-config-rm2h8\" (UID: \"9ea3c542-09d7-46ee-942f-a0476f7609d1\") " pod="openstack/ovn-controller-4kqnw-config-rm2h8" Oct 05 07:06:38 crc kubenswrapper[4846]: I1005 07:06:38.117715 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/9ea3c542-09d7-46ee-942f-a0476f7609d1-var-log-ovn\") pod \"ovn-controller-4kqnw-config-rm2h8\" (UID: \"9ea3c542-09d7-46ee-942f-a0476f7609d1\") " pod="openstack/ovn-controller-4kqnw-config-rm2h8" Oct 05 07:06:38 crc kubenswrapper[4846]: I1005 07:06:38.117735 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qmwv4\" (UniqueName: \"kubernetes.io/projected/9ea3c542-09d7-46ee-942f-a0476f7609d1-kube-api-access-qmwv4\") pod \"ovn-controller-4kqnw-config-rm2h8\" (UID: \"9ea3c542-09d7-46ee-942f-a0476f7609d1\") " pod="openstack/ovn-controller-4kqnw-config-rm2h8" Oct 05 07:06:38 crc kubenswrapper[4846]: I1005 07:06:38.117759 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/9ea3c542-09d7-46ee-942f-a0476f7609d1-var-run\") pod \"ovn-controller-4kqnw-config-rm2h8\" (UID: \"9ea3c542-09d7-46ee-942f-a0476f7609d1\") " pod="openstack/ovn-controller-4kqnw-config-rm2h8" Oct 05 07:06:38 crc kubenswrapper[4846]: I1005 07:06:38.117842 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9ea3c542-09d7-46ee-942f-a0476f7609d1-scripts\") pod \"ovn-controller-4kqnw-config-rm2h8\" (UID: \"9ea3c542-09d7-46ee-942f-a0476f7609d1\") " pod="openstack/ovn-controller-4kqnw-config-rm2h8" Oct 05 07:06:38 crc kubenswrapper[4846]: I1005 07:06:38.118152 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/9ea3c542-09d7-46ee-942f-a0476f7609d1-var-run-ovn\") pod \"ovn-controller-4kqnw-config-rm2h8\" (UID: \"9ea3c542-09d7-46ee-942f-a0476f7609d1\") " pod="openstack/ovn-controller-4kqnw-config-rm2h8" Oct 05 07:06:38 crc kubenswrapper[4846]: I1005 07:06:38.118172 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/9ea3c542-09d7-46ee-942f-a0476f7609d1-var-log-ovn\") pod \"ovn-controller-4kqnw-config-rm2h8\" (UID: \"9ea3c542-09d7-46ee-942f-a0476f7609d1\") " pod="openstack/ovn-controller-4kqnw-config-rm2h8" Oct 05 07:06:38 crc kubenswrapper[4846]: I1005 07:06:38.118310 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/9ea3c542-09d7-46ee-942f-a0476f7609d1-var-run\") pod \"ovn-controller-4kqnw-config-rm2h8\" (UID: \"9ea3c542-09d7-46ee-942f-a0476f7609d1\") " pod="openstack/ovn-controller-4kqnw-config-rm2h8" Oct 05 07:06:38 crc kubenswrapper[4846]: I1005 07:06:38.119015 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/9ea3c542-09d7-46ee-942f-a0476f7609d1-additional-scripts\") pod \"ovn-controller-4kqnw-config-rm2h8\" (UID: \"9ea3c542-09d7-46ee-942f-a0476f7609d1\") " pod="openstack/ovn-controller-4kqnw-config-rm2h8" Oct 05 07:06:38 crc kubenswrapper[4846]: I1005 07:06:38.119914 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9ea3c542-09d7-46ee-942f-a0476f7609d1-scripts\") pod \"ovn-controller-4kqnw-config-rm2h8\" (UID: \"9ea3c542-09d7-46ee-942f-a0476f7609d1\") " pod="openstack/ovn-controller-4kqnw-config-rm2h8" Oct 05 07:06:38 crc kubenswrapper[4846]: I1005 07:06:38.154644 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qmwv4\" (UniqueName: \"kubernetes.io/projected/9ea3c542-09d7-46ee-942f-a0476f7609d1-kube-api-access-qmwv4\") pod \"ovn-controller-4kqnw-config-rm2h8\" (UID: \"9ea3c542-09d7-46ee-942f-a0476f7609d1\") " pod="openstack/ovn-controller-4kqnw-config-rm2h8" Oct 05 07:06:38 crc kubenswrapper[4846]: I1005 07:06:38.330347 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-4kqnw-config-rm2h8" Oct 05 07:06:38 crc kubenswrapper[4846]: I1005 07:06:38.549312 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b3001e94-7ed4-4f76-a0b5-882965cba150" path="/var/lib/kubelet/pods/b3001e94-7ed4-4f76-a0b5-882965cba150/volumes" Oct 05 07:06:38 crc kubenswrapper[4846]: I1005 07:06:38.697526 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-4kqnw-config-rm2h8"] Oct 05 07:06:38 crc kubenswrapper[4846]: I1005 07:06:38.745310 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-4kqnw-config-rm2h8" event={"ID":"9ea3c542-09d7-46ee-942f-a0476f7609d1","Type":"ContainerStarted","Data":"c06434576120f78c1a316dacb16d71aa764f52aecc8978a0157212cbc05ae43d"} Oct 05 07:06:38 crc kubenswrapper[4846]: I1005 07:06:38.748462 4846 generic.go:334] "Generic (PLEG): container finished" podID="fb35a7c4-9453-4ce8-b0c0-828893100c92" containerID="bc74462851218e64b2af029b68e74068a836521458f244ad01e383cfa57ca002" exitCode=0 Oct 05 07:06:38 crc kubenswrapper[4846]: I1005 07:06:38.748553 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-tbdm6" event={"ID":"fb35a7c4-9453-4ce8-b0c0-828893100c92","Type":"ContainerDied","Data":"bc74462851218e64b2af029b68e74068a836521458f244ad01e383cfa57ca002"} Oct 05 07:06:38 crc kubenswrapper[4846]: I1005 07:06:38.751283 4846 generic.go:334] "Generic (PLEG): container finished" podID="00746abe-43f9-4275-bf0c-7f095b71e9bc" containerID="715bb8fb0bc7969f7b9a465afe9f43ee39ae63611715e80bbcb41d0b5b4445de" exitCode=0 Oct 05 07:06:38 crc kubenswrapper[4846]: I1005 07:06:38.751362 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-rlqvn" event={"ID":"00746abe-43f9-4275-bf0c-7f095b71e9bc","Type":"ContainerDied","Data":"715bb8fb0bc7969f7b9a465afe9f43ee39ae63611715e80bbcb41d0b5b4445de"} Oct 05 07:06:38 crc kubenswrapper[4846]: I1005 07:06:38.757241 4846 generic.go:334] "Generic (PLEG): container finished" podID="6a5f9a12-316d-42cf-a592-e418d03e0849" containerID="2cd30fb4a2bf37a8f593325296e9f9f11eaa096979e85337e1f058f23b5bd224" exitCode=0 Oct 05 07:06:38 crc kubenswrapper[4846]: I1005 07:06:38.757284 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-xp6rt" event={"ID":"6a5f9a12-316d-42cf-a592-e418d03e0849","Type":"ContainerDied","Data":"2cd30fb4a2bf37a8f593325296e9f9f11eaa096979e85337e1f058f23b5bd224"} Oct 05 07:06:38 crc kubenswrapper[4846]: I1005 07:06:38.762247 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"62d2ef82-9372-4b55-990e-30eb8e16e439","Type":"ContainerStarted","Data":"d7516230c3ea4b249ca682d8efe6e221f59dc03b58d2094da6e825d330b539b4"} Oct 05 07:06:38 crc kubenswrapper[4846]: I1005 07:06:38.762305 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"62d2ef82-9372-4b55-990e-30eb8e16e439","Type":"ContainerStarted","Data":"cb25bcb083c8fd00a96201f36477176be16589836d2ad4521c750f525555a9af"} Oct 05 07:06:39 crc kubenswrapper[4846]: I1005 07:06:39.794888 4846 generic.go:334] "Generic (PLEG): container finished" podID="9ea3c542-09d7-46ee-942f-a0476f7609d1" containerID="7cd6caeb5a6926e34fdec088400b24c5c5bc48b887c5e275a91ac1896351fdc3" exitCode=0 Oct 05 07:06:39 crc kubenswrapper[4846]: I1005 07:06:39.795383 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-4kqnw-config-rm2h8" event={"ID":"9ea3c542-09d7-46ee-942f-a0476f7609d1","Type":"ContainerDied","Data":"7cd6caeb5a6926e34fdec088400b24c5c5bc48b887c5e275a91ac1896351fdc3"} Oct 05 07:06:40 crc kubenswrapper[4846]: I1005 07:06:40.471912 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-xp6rt" Oct 05 07:06:40 crc kubenswrapper[4846]: I1005 07:06:40.572997 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4s87\" (UniqueName: \"kubernetes.io/projected/6a5f9a12-316d-42cf-a592-e418d03e0849-kube-api-access-w4s87\") pod \"6a5f9a12-316d-42cf-a592-e418d03e0849\" (UID: \"6a5f9a12-316d-42cf-a592-e418d03e0849\") " Oct 05 07:06:40 crc kubenswrapper[4846]: I1005 07:06:40.585816 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6a5f9a12-316d-42cf-a592-e418d03e0849-kube-api-access-w4s87" (OuterVolumeSpecName: "kube-api-access-w4s87") pod "6a5f9a12-316d-42cf-a592-e418d03e0849" (UID: "6a5f9a12-316d-42cf-a592-e418d03e0849"). InnerVolumeSpecName "kube-api-access-w4s87". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:06:40 crc kubenswrapper[4846]: I1005 07:06:40.657129 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-tbdm6" Oct 05 07:06:40 crc kubenswrapper[4846]: I1005 07:06:40.665412 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-rlqvn" Oct 05 07:06:40 crc kubenswrapper[4846]: I1005 07:06:40.676128 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4s87\" (UniqueName: \"kubernetes.io/projected/6a5f9a12-316d-42cf-a592-e418d03e0849-kube-api-access-w4s87\") on node \"crc\" DevicePath \"\"" Oct 05 07:06:40 crc kubenswrapper[4846]: I1005 07:06:40.778170 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8brqm\" (UniqueName: \"kubernetes.io/projected/00746abe-43f9-4275-bf0c-7f095b71e9bc-kube-api-access-8brqm\") pod \"00746abe-43f9-4275-bf0c-7f095b71e9bc\" (UID: \"00746abe-43f9-4275-bf0c-7f095b71e9bc\") " Oct 05 07:06:40 crc kubenswrapper[4846]: I1005 07:06:40.778398 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wn5rr\" (UniqueName: \"kubernetes.io/projected/fb35a7c4-9453-4ce8-b0c0-828893100c92-kube-api-access-wn5rr\") pod \"fb35a7c4-9453-4ce8-b0c0-828893100c92\" (UID: \"fb35a7c4-9453-4ce8-b0c0-828893100c92\") " Oct 05 07:06:40 crc kubenswrapper[4846]: I1005 07:06:40.787437 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/00746abe-43f9-4275-bf0c-7f095b71e9bc-kube-api-access-8brqm" (OuterVolumeSpecName: "kube-api-access-8brqm") pod "00746abe-43f9-4275-bf0c-7f095b71e9bc" (UID: "00746abe-43f9-4275-bf0c-7f095b71e9bc"). InnerVolumeSpecName "kube-api-access-8brqm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:06:40 crc kubenswrapper[4846]: I1005 07:06:40.787522 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fb35a7c4-9453-4ce8-b0c0-828893100c92-kube-api-access-wn5rr" (OuterVolumeSpecName: "kube-api-access-wn5rr") pod "fb35a7c4-9453-4ce8-b0c0-828893100c92" (UID: "fb35a7c4-9453-4ce8-b0c0-828893100c92"). InnerVolumeSpecName "kube-api-access-wn5rr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:06:40 crc kubenswrapper[4846]: I1005 07:06:40.839993 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-rlqvn" event={"ID":"00746abe-43f9-4275-bf0c-7f095b71e9bc","Type":"ContainerDied","Data":"2d00432244837f455705b2ad8f4ffb330bdb64892b198f987bba9afcbf88cff5"} Oct 05 07:06:40 crc kubenswrapper[4846]: I1005 07:06:40.840050 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2d00432244837f455705b2ad8f4ffb330bdb64892b198f987bba9afcbf88cff5" Oct 05 07:06:40 crc kubenswrapper[4846]: I1005 07:06:40.840015 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-rlqvn" Oct 05 07:06:40 crc kubenswrapper[4846]: I1005 07:06:40.843345 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-xp6rt" event={"ID":"6a5f9a12-316d-42cf-a592-e418d03e0849","Type":"ContainerDied","Data":"7754148ace05ad06ee097e13afd91ea75505f2f443534d88a22bc6542491436e"} Oct 05 07:06:40 crc kubenswrapper[4846]: I1005 07:06:40.843388 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7754148ace05ad06ee097e13afd91ea75505f2f443534d88a22bc6542491436e" Oct 05 07:06:40 crc kubenswrapper[4846]: I1005 07:06:40.843432 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-xp6rt" Oct 05 07:06:40 crc kubenswrapper[4846]: I1005 07:06:40.857038 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"62d2ef82-9372-4b55-990e-30eb8e16e439","Type":"ContainerStarted","Data":"1af615c099267db200d6c220550e8454d08ffe8449631c9c707f1cdad15b401e"} Oct 05 07:06:40 crc kubenswrapper[4846]: I1005 07:06:40.857109 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"62d2ef82-9372-4b55-990e-30eb8e16e439","Type":"ContainerStarted","Data":"19432cd2eb09eb116d05c3d630f9414b8755868167c8ad9d50242b823e184fe6"} Oct 05 07:06:40 crc kubenswrapper[4846]: I1005 07:06:40.857128 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"62d2ef82-9372-4b55-990e-30eb8e16e439","Type":"ContainerStarted","Data":"1892fe91962ddb7379e117ea24de43a317054f7a7cd0983a6e06d7d4a4eafef9"} Oct 05 07:06:40 crc kubenswrapper[4846]: I1005 07:06:40.863488 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-tbdm6" Oct 05 07:06:40 crc kubenswrapper[4846]: I1005 07:06:40.865875 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-tbdm6" event={"ID":"fb35a7c4-9453-4ce8-b0c0-828893100c92","Type":"ContainerDied","Data":"1b0a90c1a196224fe2bb54e5475dac1694e64355d8a28b1d4ff6cd37263a50e3"} Oct 05 07:06:40 crc kubenswrapper[4846]: I1005 07:06:40.865969 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1b0a90c1a196224fe2bb54e5475dac1694e64355d8a28b1d4ff6cd37263a50e3" Oct 05 07:06:40 crc kubenswrapper[4846]: I1005 07:06:40.881135 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8brqm\" (UniqueName: \"kubernetes.io/projected/00746abe-43f9-4275-bf0c-7f095b71e9bc-kube-api-access-8brqm\") on node \"crc\" DevicePath \"\"" Oct 05 07:06:40 crc kubenswrapper[4846]: I1005 07:06:40.881174 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wn5rr\" (UniqueName: \"kubernetes.io/projected/fb35a7c4-9453-4ce8-b0c0-828893100c92-kube-api-access-wn5rr\") on node \"crc\" DevicePath \"\"" Oct 05 07:06:41 crc kubenswrapper[4846]: I1005 07:06:41.176519 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-4kqnw-config-rm2h8" Oct 05 07:06:41 crc kubenswrapper[4846]: I1005 07:06:41.300418 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9ea3c542-09d7-46ee-942f-a0476f7609d1-scripts\") pod \"9ea3c542-09d7-46ee-942f-a0476f7609d1\" (UID: \"9ea3c542-09d7-46ee-942f-a0476f7609d1\") " Oct 05 07:06:41 crc kubenswrapper[4846]: I1005 07:06:41.300764 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/9ea3c542-09d7-46ee-942f-a0476f7609d1-var-log-ovn\") pod \"9ea3c542-09d7-46ee-942f-a0476f7609d1\" (UID: \"9ea3c542-09d7-46ee-942f-a0476f7609d1\") " Oct 05 07:06:41 crc kubenswrapper[4846]: I1005 07:06:41.300870 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9ea3c542-09d7-46ee-942f-a0476f7609d1-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "9ea3c542-09d7-46ee-942f-a0476f7609d1" (UID: "9ea3c542-09d7-46ee-942f-a0476f7609d1"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 05 07:06:41 crc kubenswrapper[4846]: I1005 07:06:41.301034 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/9ea3c542-09d7-46ee-942f-a0476f7609d1-var-run-ovn\") pod \"9ea3c542-09d7-46ee-942f-a0476f7609d1\" (UID: \"9ea3c542-09d7-46ee-942f-a0476f7609d1\") " Oct 05 07:06:41 crc kubenswrapper[4846]: I1005 07:06:41.301147 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/9ea3c542-09d7-46ee-942f-a0476f7609d1-additional-scripts\") pod \"9ea3c542-09d7-46ee-942f-a0476f7609d1\" (UID: \"9ea3c542-09d7-46ee-942f-a0476f7609d1\") " Oct 05 07:06:41 crc kubenswrapper[4846]: I1005 07:06:41.301238 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qmwv4\" (UniqueName: \"kubernetes.io/projected/9ea3c542-09d7-46ee-942f-a0476f7609d1-kube-api-access-qmwv4\") pod \"9ea3c542-09d7-46ee-942f-a0476f7609d1\" (UID: \"9ea3c542-09d7-46ee-942f-a0476f7609d1\") " Oct 05 07:06:41 crc kubenswrapper[4846]: I1005 07:06:41.301347 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/9ea3c542-09d7-46ee-942f-a0476f7609d1-var-run\") pod \"9ea3c542-09d7-46ee-942f-a0476f7609d1\" (UID: \"9ea3c542-09d7-46ee-942f-a0476f7609d1\") " Oct 05 07:06:41 crc kubenswrapper[4846]: I1005 07:06:41.301075 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9ea3c542-09d7-46ee-942f-a0476f7609d1-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "9ea3c542-09d7-46ee-942f-a0476f7609d1" (UID: "9ea3c542-09d7-46ee-942f-a0476f7609d1"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 05 07:06:41 crc kubenswrapper[4846]: I1005 07:06:41.301480 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9ea3c542-09d7-46ee-942f-a0476f7609d1-var-run" (OuterVolumeSpecName: "var-run") pod "9ea3c542-09d7-46ee-942f-a0476f7609d1" (UID: "9ea3c542-09d7-46ee-942f-a0476f7609d1"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 05 07:06:41 crc kubenswrapper[4846]: I1005 07:06:41.301771 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9ea3c542-09d7-46ee-942f-a0476f7609d1-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "9ea3c542-09d7-46ee-942f-a0476f7609d1" (UID: "9ea3c542-09d7-46ee-942f-a0476f7609d1"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 07:06:41 crc kubenswrapper[4846]: I1005 07:06:41.301759 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9ea3c542-09d7-46ee-942f-a0476f7609d1-scripts" (OuterVolumeSpecName: "scripts") pod "9ea3c542-09d7-46ee-942f-a0476f7609d1" (UID: "9ea3c542-09d7-46ee-942f-a0476f7609d1"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 07:06:41 crc kubenswrapper[4846]: I1005 07:06:41.301993 4846 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/9ea3c542-09d7-46ee-942f-a0476f7609d1-var-run-ovn\") on node \"crc\" DevicePath \"\"" Oct 05 07:06:41 crc kubenswrapper[4846]: I1005 07:06:41.302164 4846 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/9ea3c542-09d7-46ee-942f-a0476f7609d1-additional-scripts\") on node \"crc\" DevicePath \"\"" Oct 05 07:06:41 crc kubenswrapper[4846]: I1005 07:06:41.302249 4846 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/9ea3c542-09d7-46ee-942f-a0476f7609d1-var-run\") on node \"crc\" DevicePath \"\"" Oct 05 07:06:41 crc kubenswrapper[4846]: I1005 07:06:41.302353 4846 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9ea3c542-09d7-46ee-942f-a0476f7609d1-scripts\") on node \"crc\" DevicePath \"\"" Oct 05 07:06:41 crc kubenswrapper[4846]: I1005 07:06:41.302410 4846 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/9ea3c542-09d7-46ee-942f-a0476f7609d1-var-log-ovn\") on node \"crc\" DevicePath \"\"" Oct 05 07:06:41 crc kubenswrapper[4846]: I1005 07:06:41.319225 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9ea3c542-09d7-46ee-942f-a0476f7609d1-kube-api-access-qmwv4" (OuterVolumeSpecName: "kube-api-access-qmwv4") pod "9ea3c542-09d7-46ee-942f-a0476f7609d1" (UID: "9ea3c542-09d7-46ee-942f-a0476f7609d1"). InnerVolumeSpecName "kube-api-access-qmwv4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:06:41 crc kubenswrapper[4846]: I1005 07:06:41.404201 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qmwv4\" (UniqueName: \"kubernetes.io/projected/9ea3c542-09d7-46ee-942f-a0476f7609d1-kube-api-access-qmwv4\") on node \"crc\" DevicePath \"\"" Oct 05 07:06:41 crc kubenswrapper[4846]: I1005 07:06:41.879508 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"62d2ef82-9372-4b55-990e-30eb8e16e439","Type":"ContainerStarted","Data":"3335b438bf1fc84c97c1fad661d249e0c3d1828dffb17a746efb34add2788d6d"} Oct 05 07:06:41 crc kubenswrapper[4846]: I1005 07:06:41.882278 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-4kqnw-config-rm2h8" event={"ID":"9ea3c542-09d7-46ee-942f-a0476f7609d1","Type":"ContainerDied","Data":"c06434576120f78c1a316dacb16d71aa764f52aecc8978a0157212cbc05ae43d"} Oct 05 07:06:41 crc kubenswrapper[4846]: I1005 07:06:41.882333 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c06434576120f78c1a316dacb16d71aa764f52aecc8978a0157212cbc05ae43d" Oct 05 07:06:41 crc kubenswrapper[4846]: I1005 07:06:41.882367 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-4kqnw-config-rm2h8" Oct 05 07:06:42 crc kubenswrapper[4846]: I1005 07:06:42.284272 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-4kqnw-config-rm2h8"] Oct 05 07:06:42 crc kubenswrapper[4846]: I1005 07:06:42.290784 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-4kqnw-config-rm2h8"] Oct 05 07:06:42 crc kubenswrapper[4846]: I1005 07:06:42.509962 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9ea3c542-09d7-46ee-942f-a0476f7609d1" path="/var/lib/kubelet/pods/9ea3c542-09d7-46ee-942f-a0476f7609d1/volumes" Oct 05 07:06:53 crc kubenswrapper[4846]: I1005 07:06:53.325148 4846 patch_prober.go:28] interesting pod/machine-config-daemon-fscvf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 07:06:53 crc kubenswrapper[4846]: I1005 07:06:53.327798 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 07:06:53 crc kubenswrapper[4846]: E1005 07:06:53.615616 4846 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-glance-api@sha256:b2f6e417260d4ae6bf70c05bf2243a0b6cac555cd28080772546761015674191" Oct 05 07:06:53 crc kubenswrapper[4846]: E1005 07:06:53.615878 4846 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:glance-db-sync,Image:quay.io/podified-antelope-centos9/openstack-glance-api@sha256:b2f6e417260d4ae6bf70c05bf2243a0b6cac555cd28080772546761015674191,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:true,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/glance/glance.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-ctmvj,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42415,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42415,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod glance-db-sync-68h9n_openstack(8b6863be-f8e1-4a87-93d6-5cf756983747): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 05 07:06:53 crc kubenswrapper[4846]: E1005 07:06:53.617084 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"glance-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/glance-db-sync-68h9n" podUID="8b6863be-f8e1-4a87-93d6-5cf756983747" Oct 05 07:06:54 crc kubenswrapper[4846]: E1005 07:06:54.061281 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"glance-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-glance-api@sha256:b2f6e417260d4ae6bf70c05bf2243a0b6cac555cd28080772546761015674191\\\"\"" pod="openstack/glance-db-sync-68h9n" podUID="8b6863be-f8e1-4a87-93d6-5cf756983747" Oct 05 07:06:55 crc kubenswrapper[4846]: I1005 07:06:55.074375 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"62d2ef82-9372-4b55-990e-30eb8e16e439","Type":"ContainerStarted","Data":"9af99544264e591f2b45ec0655947d9163b35da963449be03de565a72c5e8e50"} Oct 05 07:06:55 crc kubenswrapper[4846]: I1005 07:06:55.076547 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"62d2ef82-9372-4b55-990e-30eb8e16e439","Type":"ContainerStarted","Data":"983443d1fc33fbdef8e7ca49faf419ed150fada692ed4b5e60d29ffcb9cbaf6c"} Oct 05 07:06:55 crc kubenswrapper[4846]: I1005 07:06:55.076600 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"62d2ef82-9372-4b55-990e-30eb8e16e439","Type":"ContainerStarted","Data":"0560e729496f16c1c92b49ae2209e5dcfd4cd0655c6017f90693d429daef0694"} Oct 05 07:06:55 crc kubenswrapper[4846]: I1005 07:06:55.076615 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"62d2ef82-9372-4b55-990e-30eb8e16e439","Type":"ContainerStarted","Data":"3e6f56f54a2bfbb877522ef200058490423742afbcf0b93313a239fae7415942"} Oct 05 07:06:55 crc kubenswrapper[4846]: I1005 07:06:55.076632 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"62d2ef82-9372-4b55-990e-30eb8e16e439","Type":"ContainerStarted","Data":"b448ad891d0bdb677c147854fe487e87246e96353b5d9e033e8ba9ab2dfa1486"} Oct 05 07:06:55 crc kubenswrapper[4846]: I1005 07:06:55.077145 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-t7t67" event={"ID":"b4b973ee-209f-477b-92ee-3f1c16f17f43","Type":"ContainerStarted","Data":"bb66ec24b24cb5604d450f4beb64f774dd42cb6a774b6536eca16fcc238ae31a"} Oct 05 07:06:55 crc kubenswrapper[4846]: I1005 07:06:55.106133 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-sync-t7t67" podStartSLOduration=3.5937511779999998 podStartE2EDuration="20.106104441s" podCreationTimestamp="2025-10-05 07:06:35 +0000 UTC" firstStartedPulling="2025-10-05 07:06:37.350568105 +0000 UTC m=+1119.591420880" lastFinishedPulling="2025-10-05 07:06:53.862921358 +0000 UTC m=+1136.103774143" observedRunningTime="2025-10-05 07:06:55.094124942 +0000 UTC m=+1137.334977727" watchObservedRunningTime="2025-10-05 07:06:55.106104441 +0000 UTC m=+1137.346957226" Oct 05 07:06:55 crc kubenswrapper[4846]: I1005 07:06:55.788395 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-cf3d-account-create-zd4h7"] Oct 05 07:06:55 crc kubenswrapper[4846]: E1005 07:06:55.788952 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fb35a7c4-9453-4ce8-b0c0-828893100c92" containerName="mariadb-database-create" Oct 05 07:06:55 crc kubenswrapper[4846]: I1005 07:06:55.788977 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="fb35a7c4-9453-4ce8-b0c0-828893100c92" containerName="mariadb-database-create" Oct 05 07:06:55 crc kubenswrapper[4846]: E1005 07:06:55.788993 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9ea3c542-09d7-46ee-942f-a0476f7609d1" containerName="ovn-config" Oct 05 07:06:55 crc kubenswrapper[4846]: I1005 07:06:55.789002 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="9ea3c542-09d7-46ee-942f-a0476f7609d1" containerName="ovn-config" Oct 05 07:06:55 crc kubenswrapper[4846]: E1005 07:06:55.789027 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="00746abe-43f9-4275-bf0c-7f095b71e9bc" containerName="mariadb-database-create" Oct 05 07:06:55 crc kubenswrapper[4846]: I1005 07:06:55.789037 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="00746abe-43f9-4275-bf0c-7f095b71e9bc" containerName="mariadb-database-create" Oct 05 07:06:55 crc kubenswrapper[4846]: E1005 07:06:55.789057 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6a5f9a12-316d-42cf-a592-e418d03e0849" containerName="mariadb-database-create" Oct 05 07:06:55 crc kubenswrapper[4846]: I1005 07:06:55.789064 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="6a5f9a12-316d-42cf-a592-e418d03e0849" containerName="mariadb-database-create" Oct 05 07:06:55 crc kubenswrapper[4846]: I1005 07:06:55.789344 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="00746abe-43f9-4275-bf0c-7f095b71e9bc" containerName="mariadb-database-create" Oct 05 07:06:55 crc kubenswrapper[4846]: I1005 07:06:55.789378 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="fb35a7c4-9453-4ce8-b0c0-828893100c92" containerName="mariadb-database-create" Oct 05 07:06:55 crc kubenswrapper[4846]: I1005 07:06:55.789420 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="9ea3c542-09d7-46ee-942f-a0476f7609d1" containerName="ovn-config" Oct 05 07:06:55 crc kubenswrapper[4846]: I1005 07:06:55.789446 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="6a5f9a12-316d-42cf-a592-e418d03e0849" containerName="mariadb-database-create" Oct 05 07:06:55 crc kubenswrapper[4846]: I1005 07:06:55.790263 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-cf3d-account-create-zd4h7" Oct 05 07:06:55 crc kubenswrapper[4846]: I1005 07:06:55.793247 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-db-secret" Oct 05 07:06:55 crc kubenswrapper[4846]: I1005 07:06:55.802556 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-cf3d-account-create-zd4h7"] Oct 05 07:06:55 crc kubenswrapper[4846]: I1005 07:06:55.825962 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vl42d\" (UniqueName: \"kubernetes.io/projected/438586d8-d6b1-445f-8460-0e44ac6037c6-kube-api-access-vl42d\") pod \"cinder-cf3d-account-create-zd4h7\" (UID: \"438586d8-d6b1-445f-8460-0e44ac6037c6\") " pod="openstack/cinder-cf3d-account-create-zd4h7" Oct 05 07:06:55 crc kubenswrapper[4846]: I1005 07:06:55.876600 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-8fc5-account-create-fcxfk"] Oct 05 07:06:55 crc kubenswrapper[4846]: I1005 07:06:55.877894 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-8fc5-account-create-fcxfk" Oct 05 07:06:55 crc kubenswrapper[4846]: I1005 07:06:55.882161 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-db-secret" Oct 05 07:06:55 crc kubenswrapper[4846]: I1005 07:06:55.893448 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-8fc5-account-create-fcxfk"] Oct 05 07:06:55 crc kubenswrapper[4846]: I1005 07:06:55.927713 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nbv6b\" (UniqueName: \"kubernetes.io/projected/36efa3ba-3fa1-4ecd-8ad3-9136d31b6070-kube-api-access-nbv6b\") pod \"barbican-8fc5-account-create-fcxfk\" (UID: \"36efa3ba-3fa1-4ecd-8ad3-9136d31b6070\") " pod="openstack/barbican-8fc5-account-create-fcxfk" Oct 05 07:06:55 crc kubenswrapper[4846]: I1005 07:06:55.927902 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vl42d\" (UniqueName: \"kubernetes.io/projected/438586d8-d6b1-445f-8460-0e44ac6037c6-kube-api-access-vl42d\") pod \"cinder-cf3d-account-create-zd4h7\" (UID: \"438586d8-d6b1-445f-8460-0e44ac6037c6\") " pod="openstack/cinder-cf3d-account-create-zd4h7" Oct 05 07:06:55 crc kubenswrapper[4846]: I1005 07:06:55.948744 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vl42d\" (UniqueName: \"kubernetes.io/projected/438586d8-d6b1-445f-8460-0e44ac6037c6-kube-api-access-vl42d\") pod \"cinder-cf3d-account-create-zd4h7\" (UID: \"438586d8-d6b1-445f-8460-0e44ac6037c6\") " pod="openstack/cinder-cf3d-account-create-zd4h7" Oct 05 07:06:56 crc kubenswrapper[4846]: I1005 07:06:56.029582 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nbv6b\" (UniqueName: \"kubernetes.io/projected/36efa3ba-3fa1-4ecd-8ad3-9136d31b6070-kube-api-access-nbv6b\") pod \"barbican-8fc5-account-create-fcxfk\" (UID: \"36efa3ba-3fa1-4ecd-8ad3-9136d31b6070\") " pod="openstack/barbican-8fc5-account-create-fcxfk" Oct 05 07:06:56 crc kubenswrapper[4846]: I1005 07:06:56.057166 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nbv6b\" (UniqueName: \"kubernetes.io/projected/36efa3ba-3fa1-4ecd-8ad3-9136d31b6070-kube-api-access-nbv6b\") pod \"barbican-8fc5-account-create-fcxfk\" (UID: \"36efa3ba-3fa1-4ecd-8ad3-9136d31b6070\") " pod="openstack/barbican-8fc5-account-create-fcxfk" Oct 05 07:06:56 crc kubenswrapper[4846]: I1005 07:06:56.105137 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"62d2ef82-9372-4b55-990e-30eb8e16e439","Type":"ContainerStarted","Data":"ccba487ad88a231e17d0bff3850ebdc3dc9887084fa35706507c9470dedbf5df"} Oct 05 07:06:56 crc kubenswrapper[4846]: I1005 07:06:56.105220 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"62d2ef82-9372-4b55-990e-30eb8e16e439","Type":"ContainerStarted","Data":"8b1645c0bb7485916cb16906718326320b1c0c52c16426fad29ece1124acb1d4"} Oct 05 07:06:56 crc kubenswrapper[4846]: I1005 07:06:56.118603 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-cf3d-account-create-zd4h7" Oct 05 07:06:56 crc kubenswrapper[4846]: I1005 07:06:56.171874 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-storage-0" podStartSLOduration=36.624955714 podStartE2EDuration="56.171842773s" podCreationTimestamp="2025-10-05 07:06:00 +0000 UTC" firstStartedPulling="2025-10-05 07:06:34.365654462 +0000 UTC m=+1116.606507227" lastFinishedPulling="2025-10-05 07:06:53.912541511 +0000 UTC m=+1136.153394286" observedRunningTime="2025-10-05 07:06:56.155461516 +0000 UTC m=+1138.396314291" watchObservedRunningTime="2025-10-05 07:06:56.171842773 +0000 UTC m=+1138.412695558" Oct 05 07:06:56 crc kubenswrapper[4846]: I1005 07:06:56.204569 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-a5c0-account-create-8cbhj"] Oct 05 07:06:56 crc kubenswrapper[4846]: I1005 07:06:56.207751 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-8fc5-account-create-fcxfk" Oct 05 07:06:56 crc kubenswrapper[4846]: I1005 07:06:56.209420 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-a5c0-account-create-8cbhj" Oct 05 07:06:56 crc kubenswrapper[4846]: I1005 07:06:56.218297 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-db-secret" Oct 05 07:06:56 crc kubenswrapper[4846]: I1005 07:06:56.226713 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-a5c0-account-create-8cbhj"] Oct 05 07:06:56 crc kubenswrapper[4846]: I1005 07:06:56.342697 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lvkgg\" (UniqueName: \"kubernetes.io/projected/577b34f0-581f-4fe2-8717-7e3062fd26fc-kube-api-access-lvkgg\") pod \"neutron-a5c0-account-create-8cbhj\" (UID: \"577b34f0-581f-4fe2-8717-7e3062fd26fc\") " pod="openstack/neutron-a5c0-account-create-8cbhj" Oct 05 07:06:57 crc kubenswrapper[4846]: I1005 07:06:56.445047 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lvkgg\" (UniqueName: \"kubernetes.io/projected/577b34f0-581f-4fe2-8717-7e3062fd26fc-kube-api-access-lvkgg\") pod \"neutron-a5c0-account-create-8cbhj\" (UID: \"577b34f0-581f-4fe2-8717-7e3062fd26fc\") " pod="openstack/neutron-a5c0-account-create-8cbhj" Oct 05 07:06:57 crc kubenswrapper[4846]: I1005 07:06:56.467980 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6f9b65c9b5-lqj5b"] Oct 05 07:06:57 crc kubenswrapper[4846]: I1005 07:06:56.469510 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6f9b65c9b5-lqj5b" Oct 05 07:06:57 crc kubenswrapper[4846]: I1005 07:06:56.474109 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-swift-storage-0" Oct 05 07:06:57 crc kubenswrapper[4846]: I1005 07:06:56.486835 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lvkgg\" (UniqueName: \"kubernetes.io/projected/577b34f0-581f-4fe2-8717-7e3062fd26fc-kube-api-access-lvkgg\") pod \"neutron-a5c0-account-create-8cbhj\" (UID: \"577b34f0-581f-4fe2-8717-7e3062fd26fc\") " pod="openstack/neutron-a5c0-account-create-8cbhj" Oct 05 07:06:57 crc kubenswrapper[4846]: I1005 07:06:56.510348 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6f9b65c9b5-lqj5b"] Oct 05 07:06:57 crc kubenswrapper[4846]: I1005 07:06:56.599768 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-a5c0-account-create-8cbhj" Oct 05 07:06:57 crc kubenswrapper[4846]: I1005 07:06:56.645423 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-cf3d-account-create-zd4h7"] Oct 05 07:06:57 crc kubenswrapper[4846]: W1005 07:06:56.648235 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod438586d8_d6b1_445f_8460_0e44ac6037c6.slice/crio-475fcd9fad874dcc007339b5e4f0f0a1c7497f455ef8908b8825a1eaad68ba55 WatchSource:0}: Error finding container 475fcd9fad874dcc007339b5e4f0f0a1c7497f455ef8908b8825a1eaad68ba55: Status 404 returned error can't find the container with id 475fcd9fad874dcc007339b5e4f0f0a1c7497f455ef8908b8825a1eaad68ba55 Oct 05 07:06:57 crc kubenswrapper[4846]: I1005 07:06:56.649751 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8c207acc-07b9-4d9f-8b20-c84c2b67560f-ovsdbserver-nb\") pod \"dnsmasq-dns-6f9b65c9b5-lqj5b\" (UID: \"8c207acc-07b9-4d9f-8b20-c84c2b67560f\") " pod="openstack/dnsmasq-dns-6f9b65c9b5-lqj5b" Oct 05 07:06:57 crc kubenswrapper[4846]: I1005 07:06:56.649828 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8c207acc-07b9-4d9f-8b20-c84c2b67560f-ovsdbserver-sb\") pod \"dnsmasq-dns-6f9b65c9b5-lqj5b\" (UID: \"8c207acc-07b9-4d9f-8b20-c84c2b67560f\") " pod="openstack/dnsmasq-dns-6f9b65c9b5-lqj5b" Oct 05 07:06:57 crc kubenswrapper[4846]: I1005 07:06:56.649871 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8c207acc-07b9-4d9f-8b20-c84c2b67560f-dns-swift-storage-0\") pod \"dnsmasq-dns-6f9b65c9b5-lqj5b\" (UID: \"8c207acc-07b9-4d9f-8b20-c84c2b67560f\") " pod="openstack/dnsmasq-dns-6f9b65c9b5-lqj5b" Oct 05 07:06:57 crc kubenswrapper[4846]: I1005 07:06:56.649950 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8c207acc-07b9-4d9f-8b20-c84c2b67560f-dns-svc\") pod \"dnsmasq-dns-6f9b65c9b5-lqj5b\" (UID: \"8c207acc-07b9-4d9f-8b20-c84c2b67560f\") " pod="openstack/dnsmasq-dns-6f9b65c9b5-lqj5b" Oct 05 07:06:57 crc kubenswrapper[4846]: I1005 07:06:56.649981 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8c207acc-07b9-4d9f-8b20-c84c2b67560f-config\") pod \"dnsmasq-dns-6f9b65c9b5-lqj5b\" (UID: \"8c207acc-07b9-4d9f-8b20-c84c2b67560f\") " pod="openstack/dnsmasq-dns-6f9b65c9b5-lqj5b" Oct 05 07:06:57 crc kubenswrapper[4846]: I1005 07:06:56.650160 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v7mt2\" (UniqueName: \"kubernetes.io/projected/8c207acc-07b9-4d9f-8b20-c84c2b67560f-kube-api-access-v7mt2\") pod \"dnsmasq-dns-6f9b65c9b5-lqj5b\" (UID: \"8c207acc-07b9-4d9f-8b20-c84c2b67560f\") " pod="openstack/dnsmasq-dns-6f9b65c9b5-lqj5b" Oct 05 07:06:57 crc kubenswrapper[4846]: I1005 07:06:56.751671 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8c207acc-07b9-4d9f-8b20-c84c2b67560f-ovsdbserver-sb\") pod \"dnsmasq-dns-6f9b65c9b5-lqj5b\" (UID: \"8c207acc-07b9-4d9f-8b20-c84c2b67560f\") " pod="openstack/dnsmasq-dns-6f9b65c9b5-lqj5b" Oct 05 07:06:57 crc kubenswrapper[4846]: I1005 07:06:56.751737 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8c207acc-07b9-4d9f-8b20-c84c2b67560f-dns-swift-storage-0\") pod \"dnsmasq-dns-6f9b65c9b5-lqj5b\" (UID: \"8c207acc-07b9-4d9f-8b20-c84c2b67560f\") " pod="openstack/dnsmasq-dns-6f9b65c9b5-lqj5b" Oct 05 07:06:57 crc kubenswrapper[4846]: I1005 07:06:56.751788 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8c207acc-07b9-4d9f-8b20-c84c2b67560f-dns-svc\") pod \"dnsmasq-dns-6f9b65c9b5-lqj5b\" (UID: \"8c207acc-07b9-4d9f-8b20-c84c2b67560f\") " pod="openstack/dnsmasq-dns-6f9b65c9b5-lqj5b" Oct 05 07:06:57 crc kubenswrapper[4846]: I1005 07:06:56.751833 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8c207acc-07b9-4d9f-8b20-c84c2b67560f-config\") pod \"dnsmasq-dns-6f9b65c9b5-lqj5b\" (UID: \"8c207acc-07b9-4d9f-8b20-c84c2b67560f\") " pod="openstack/dnsmasq-dns-6f9b65c9b5-lqj5b" Oct 05 07:06:57 crc kubenswrapper[4846]: I1005 07:06:56.752099 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v7mt2\" (UniqueName: \"kubernetes.io/projected/8c207acc-07b9-4d9f-8b20-c84c2b67560f-kube-api-access-v7mt2\") pod \"dnsmasq-dns-6f9b65c9b5-lqj5b\" (UID: \"8c207acc-07b9-4d9f-8b20-c84c2b67560f\") " pod="openstack/dnsmasq-dns-6f9b65c9b5-lqj5b" Oct 05 07:06:57 crc kubenswrapper[4846]: I1005 07:06:56.752167 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8c207acc-07b9-4d9f-8b20-c84c2b67560f-ovsdbserver-nb\") pod \"dnsmasq-dns-6f9b65c9b5-lqj5b\" (UID: \"8c207acc-07b9-4d9f-8b20-c84c2b67560f\") " pod="openstack/dnsmasq-dns-6f9b65c9b5-lqj5b" Oct 05 07:06:57 crc kubenswrapper[4846]: I1005 07:06:56.753804 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8c207acc-07b9-4d9f-8b20-c84c2b67560f-dns-svc\") pod \"dnsmasq-dns-6f9b65c9b5-lqj5b\" (UID: \"8c207acc-07b9-4d9f-8b20-c84c2b67560f\") " pod="openstack/dnsmasq-dns-6f9b65c9b5-lqj5b" Oct 05 07:06:57 crc kubenswrapper[4846]: I1005 07:06:56.753806 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8c207acc-07b9-4d9f-8b20-c84c2b67560f-config\") pod \"dnsmasq-dns-6f9b65c9b5-lqj5b\" (UID: \"8c207acc-07b9-4d9f-8b20-c84c2b67560f\") " pod="openstack/dnsmasq-dns-6f9b65c9b5-lqj5b" Oct 05 07:06:57 crc kubenswrapper[4846]: I1005 07:06:56.754113 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8c207acc-07b9-4d9f-8b20-c84c2b67560f-dns-swift-storage-0\") pod \"dnsmasq-dns-6f9b65c9b5-lqj5b\" (UID: \"8c207acc-07b9-4d9f-8b20-c84c2b67560f\") " pod="openstack/dnsmasq-dns-6f9b65c9b5-lqj5b" Oct 05 07:06:57 crc kubenswrapper[4846]: I1005 07:06:56.754929 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8c207acc-07b9-4d9f-8b20-c84c2b67560f-ovsdbserver-sb\") pod \"dnsmasq-dns-6f9b65c9b5-lqj5b\" (UID: \"8c207acc-07b9-4d9f-8b20-c84c2b67560f\") " pod="openstack/dnsmasq-dns-6f9b65c9b5-lqj5b" Oct 05 07:06:57 crc kubenswrapper[4846]: I1005 07:06:56.755162 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8c207acc-07b9-4d9f-8b20-c84c2b67560f-ovsdbserver-nb\") pod \"dnsmasq-dns-6f9b65c9b5-lqj5b\" (UID: \"8c207acc-07b9-4d9f-8b20-c84c2b67560f\") " pod="openstack/dnsmasq-dns-6f9b65c9b5-lqj5b" Oct 05 07:06:57 crc kubenswrapper[4846]: I1005 07:06:56.775387 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v7mt2\" (UniqueName: \"kubernetes.io/projected/8c207acc-07b9-4d9f-8b20-c84c2b67560f-kube-api-access-v7mt2\") pod \"dnsmasq-dns-6f9b65c9b5-lqj5b\" (UID: \"8c207acc-07b9-4d9f-8b20-c84c2b67560f\") " pod="openstack/dnsmasq-dns-6f9b65c9b5-lqj5b" Oct 05 07:06:57 crc kubenswrapper[4846]: I1005 07:06:56.792074 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6f9b65c9b5-lqj5b" Oct 05 07:06:57 crc kubenswrapper[4846]: I1005 07:06:57.125500 4846 generic.go:334] "Generic (PLEG): container finished" podID="438586d8-d6b1-445f-8460-0e44ac6037c6" containerID="fd96bbff85b7b2a5303006a90a5cd90ed2a0e07acc25eb2d05bccf60b24cfee7" exitCode=0 Oct 05 07:06:57 crc kubenswrapper[4846]: I1005 07:06:57.125626 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-cf3d-account-create-zd4h7" event={"ID":"438586d8-d6b1-445f-8460-0e44ac6037c6","Type":"ContainerDied","Data":"fd96bbff85b7b2a5303006a90a5cd90ed2a0e07acc25eb2d05bccf60b24cfee7"} Oct 05 07:06:57 crc kubenswrapper[4846]: I1005 07:06:57.125917 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-cf3d-account-create-zd4h7" event={"ID":"438586d8-d6b1-445f-8460-0e44ac6037c6","Type":"ContainerStarted","Data":"475fcd9fad874dcc007339b5e4f0f0a1c7497f455ef8908b8825a1eaad68ba55"} Oct 05 07:06:57 crc kubenswrapper[4846]: I1005 07:06:57.206981 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-8fc5-account-create-fcxfk"] Oct 05 07:06:57 crc kubenswrapper[4846]: W1005 07:06:57.221288 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod36efa3ba_3fa1_4ecd_8ad3_9136d31b6070.slice/crio-5e51ebdda6e6ff2c9cf97ff11618c504fd9c5b9c2b44eb4d1bb8eb57f5741561 WatchSource:0}: Error finding container 5e51ebdda6e6ff2c9cf97ff11618c504fd9c5b9c2b44eb4d1bb8eb57f5741561: Status 404 returned error can't find the container with id 5e51ebdda6e6ff2c9cf97ff11618c504fd9c5b9c2b44eb4d1bb8eb57f5741561 Oct 05 07:06:57 crc kubenswrapper[4846]: I1005 07:06:57.461495 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-a5c0-account-create-8cbhj"] Oct 05 07:06:57 crc kubenswrapper[4846]: W1005 07:06:57.472500 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod577b34f0_581f_4fe2_8717_7e3062fd26fc.slice/crio-a5690caa347625c9e565fe502c81e638d024b7e5ef5c011bb82875d888aae201 WatchSource:0}: Error finding container a5690caa347625c9e565fe502c81e638d024b7e5ef5c011bb82875d888aae201: Status 404 returned error can't find the container with id a5690caa347625c9e565fe502c81e638d024b7e5ef5c011bb82875d888aae201 Oct 05 07:06:57 crc kubenswrapper[4846]: I1005 07:06:57.483394 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6f9b65c9b5-lqj5b"] Oct 05 07:06:58 crc kubenswrapper[4846]: I1005 07:06:58.141204 4846 generic.go:334] "Generic (PLEG): container finished" podID="577b34f0-581f-4fe2-8717-7e3062fd26fc" containerID="2fd09fbd97edd50e2b9f40d0f48d643a047010b082f6e418ea82fb818d979cce" exitCode=0 Oct 05 07:06:58 crc kubenswrapper[4846]: I1005 07:06:58.141358 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-a5c0-account-create-8cbhj" event={"ID":"577b34f0-581f-4fe2-8717-7e3062fd26fc","Type":"ContainerDied","Data":"2fd09fbd97edd50e2b9f40d0f48d643a047010b082f6e418ea82fb818d979cce"} Oct 05 07:06:58 crc kubenswrapper[4846]: I1005 07:06:58.141746 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-a5c0-account-create-8cbhj" event={"ID":"577b34f0-581f-4fe2-8717-7e3062fd26fc","Type":"ContainerStarted","Data":"a5690caa347625c9e565fe502c81e638d024b7e5ef5c011bb82875d888aae201"} Oct 05 07:06:58 crc kubenswrapper[4846]: I1005 07:06:58.146980 4846 generic.go:334] "Generic (PLEG): container finished" podID="36efa3ba-3fa1-4ecd-8ad3-9136d31b6070" containerID="a387124df9f8bc90bc9f6818e068798ea86dcd47589df9f1a1df098f8dfdab60" exitCode=0 Oct 05 07:06:58 crc kubenswrapper[4846]: I1005 07:06:58.147054 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-8fc5-account-create-fcxfk" event={"ID":"36efa3ba-3fa1-4ecd-8ad3-9136d31b6070","Type":"ContainerDied","Data":"a387124df9f8bc90bc9f6818e068798ea86dcd47589df9f1a1df098f8dfdab60"} Oct 05 07:06:58 crc kubenswrapper[4846]: I1005 07:06:58.147073 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-8fc5-account-create-fcxfk" event={"ID":"36efa3ba-3fa1-4ecd-8ad3-9136d31b6070","Type":"ContainerStarted","Data":"5e51ebdda6e6ff2c9cf97ff11618c504fd9c5b9c2b44eb4d1bb8eb57f5741561"} Oct 05 07:06:58 crc kubenswrapper[4846]: I1005 07:06:58.149475 4846 generic.go:334] "Generic (PLEG): container finished" podID="b4b973ee-209f-477b-92ee-3f1c16f17f43" containerID="bb66ec24b24cb5604d450f4beb64f774dd42cb6a774b6536eca16fcc238ae31a" exitCode=0 Oct 05 07:06:58 crc kubenswrapper[4846]: I1005 07:06:58.149540 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-t7t67" event={"ID":"b4b973ee-209f-477b-92ee-3f1c16f17f43","Type":"ContainerDied","Data":"bb66ec24b24cb5604d450f4beb64f774dd42cb6a774b6536eca16fcc238ae31a"} Oct 05 07:06:58 crc kubenswrapper[4846]: I1005 07:06:58.151640 4846 generic.go:334] "Generic (PLEG): container finished" podID="8c207acc-07b9-4d9f-8b20-c84c2b67560f" containerID="3014cd247fcb2ed23a0ded0ef8cba21e3b0661b9ee16f6a89e928f7c6d8a5c69" exitCode=0 Oct 05 07:06:58 crc kubenswrapper[4846]: I1005 07:06:58.151779 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6f9b65c9b5-lqj5b" event={"ID":"8c207acc-07b9-4d9f-8b20-c84c2b67560f","Type":"ContainerDied","Data":"3014cd247fcb2ed23a0ded0ef8cba21e3b0661b9ee16f6a89e928f7c6d8a5c69"} Oct 05 07:06:58 crc kubenswrapper[4846]: I1005 07:06:58.151873 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6f9b65c9b5-lqj5b" event={"ID":"8c207acc-07b9-4d9f-8b20-c84c2b67560f","Type":"ContainerStarted","Data":"3c0746cb95a48aefd998651a9fafd0c4d79f70c5eab5dcffdb7c716fc312679e"} Oct 05 07:06:58 crc kubenswrapper[4846]: I1005 07:06:58.384328 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-cf3d-account-create-zd4h7" Oct 05 07:06:58 crc kubenswrapper[4846]: I1005 07:06:58.404453 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vl42d\" (UniqueName: \"kubernetes.io/projected/438586d8-d6b1-445f-8460-0e44ac6037c6-kube-api-access-vl42d\") pod \"438586d8-d6b1-445f-8460-0e44ac6037c6\" (UID: \"438586d8-d6b1-445f-8460-0e44ac6037c6\") " Oct 05 07:06:58 crc kubenswrapper[4846]: I1005 07:06:58.410611 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/438586d8-d6b1-445f-8460-0e44ac6037c6-kube-api-access-vl42d" (OuterVolumeSpecName: "kube-api-access-vl42d") pod "438586d8-d6b1-445f-8460-0e44ac6037c6" (UID: "438586d8-d6b1-445f-8460-0e44ac6037c6"). InnerVolumeSpecName "kube-api-access-vl42d". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:06:58 crc kubenswrapper[4846]: I1005 07:06:58.506238 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vl42d\" (UniqueName: \"kubernetes.io/projected/438586d8-d6b1-445f-8460-0e44ac6037c6-kube-api-access-vl42d\") on node \"crc\" DevicePath \"\"" Oct 05 07:06:59 crc kubenswrapper[4846]: I1005 07:06:59.166547 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6f9b65c9b5-lqj5b" event={"ID":"8c207acc-07b9-4d9f-8b20-c84c2b67560f","Type":"ContainerStarted","Data":"8d03c956f45603fb749baa477ec5fb7e31072932fb32c3bef02bba3de7b751e9"} Oct 05 07:06:59 crc kubenswrapper[4846]: I1005 07:06:59.167287 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6f9b65c9b5-lqj5b" Oct 05 07:06:59 crc kubenswrapper[4846]: I1005 07:06:59.168603 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-cf3d-account-create-zd4h7" event={"ID":"438586d8-d6b1-445f-8460-0e44ac6037c6","Type":"ContainerDied","Data":"475fcd9fad874dcc007339b5e4f0f0a1c7497f455ef8908b8825a1eaad68ba55"} Oct 05 07:06:59 crc kubenswrapper[4846]: I1005 07:06:59.168700 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="475fcd9fad874dcc007339b5e4f0f0a1c7497f455ef8908b8825a1eaad68ba55" Oct 05 07:06:59 crc kubenswrapper[4846]: I1005 07:06:59.168874 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-cf3d-account-create-zd4h7" Oct 05 07:06:59 crc kubenswrapper[4846]: I1005 07:06:59.198755 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6f9b65c9b5-lqj5b" podStartSLOduration=3.198721485 podStartE2EDuration="3.198721485s" podCreationTimestamp="2025-10-05 07:06:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 07:06:59.190648419 +0000 UTC m=+1141.431501194" watchObservedRunningTime="2025-10-05 07:06:59.198721485 +0000 UTC m=+1141.439574260" Oct 05 07:06:59 crc kubenswrapper[4846]: I1005 07:06:59.516398 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-8fc5-account-create-fcxfk" Oct 05 07:06:59 crc kubenswrapper[4846]: I1005 07:06:59.604965 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-t7t67" Oct 05 07:06:59 crc kubenswrapper[4846]: I1005 07:06:59.611596 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-a5c0-account-create-8cbhj" Oct 05 07:06:59 crc kubenswrapper[4846]: I1005 07:06:59.627612 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nbv6b\" (UniqueName: \"kubernetes.io/projected/36efa3ba-3fa1-4ecd-8ad3-9136d31b6070-kube-api-access-nbv6b\") pod \"36efa3ba-3fa1-4ecd-8ad3-9136d31b6070\" (UID: \"36efa3ba-3fa1-4ecd-8ad3-9136d31b6070\") " Oct 05 07:06:59 crc kubenswrapper[4846]: I1005 07:06:59.627756 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b4b973ee-209f-477b-92ee-3f1c16f17f43-config-data\") pod \"b4b973ee-209f-477b-92ee-3f1c16f17f43\" (UID: \"b4b973ee-209f-477b-92ee-3f1c16f17f43\") " Oct 05 07:06:59 crc kubenswrapper[4846]: I1005 07:06:59.627838 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5dsnc\" (UniqueName: \"kubernetes.io/projected/b4b973ee-209f-477b-92ee-3f1c16f17f43-kube-api-access-5dsnc\") pod \"b4b973ee-209f-477b-92ee-3f1c16f17f43\" (UID: \"b4b973ee-209f-477b-92ee-3f1c16f17f43\") " Oct 05 07:06:59 crc kubenswrapper[4846]: I1005 07:06:59.627867 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lvkgg\" (UniqueName: \"kubernetes.io/projected/577b34f0-581f-4fe2-8717-7e3062fd26fc-kube-api-access-lvkgg\") pod \"577b34f0-581f-4fe2-8717-7e3062fd26fc\" (UID: \"577b34f0-581f-4fe2-8717-7e3062fd26fc\") " Oct 05 07:06:59 crc kubenswrapper[4846]: I1005 07:06:59.627904 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b4b973ee-209f-477b-92ee-3f1c16f17f43-combined-ca-bundle\") pod \"b4b973ee-209f-477b-92ee-3f1c16f17f43\" (UID: \"b4b973ee-209f-477b-92ee-3f1c16f17f43\") " Oct 05 07:06:59 crc kubenswrapper[4846]: I1005 07:06:59.669948 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/36efa3ba-3fa1-4ecd-8ad3-9136d31b6070-kube-api-access-nbv6b" (OuterVolumeSpecName: "kube-api-access-nbv6b") pod "36efa3ba-3fa1-4ecd-8ad3-9136d31b6070" (UID: "36efa3ba-3fa1-4ecd-8ad3-9136d31b6070"). InnerVolumeSpecName "kube-api-access-nbv6b". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:06:59 crc kubenswrapper[4846]: I1005 07:06:59.670038 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b4b973ee-209f-477b-92ee-3f1c16f17f43-kube-api-access-5dsnc" (OuterVolumeSpecName: "kube-api-access-5dsnc") pod "b4b973ee-209f-477b-92ee-3f1c16f17f43" (UID: "b4b973ee-209f-477b-92ee-3f1c16f17f43"). InnerVolumeSpecName "kube-api-access-5dsnc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:06:59 crc kubenswrapper[4846]: I1005 07:06:59.670583 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/577b34f0-581f-4fe2-8717-7e3062fd26fc-kube-api-access-lvkgg" (OuterVolumeSpecName: "kube-api-access-lvkgg") pod "577b34f0-581f-4fe2-8717-7e3062fd26fc" (UID: "577b34f0-581f-4fe2-8717-7e3062fd26fc"). InnerVolumeSpecName "kube-api-access-lvkgg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:06:59 crc kubenswrapper[4846]: I1005 07:06:59.675159 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b4b973ee-209f-477b-92ee-3f1c16f17f43-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b4b973ee-209f-477b-92ee-3f1c16f17f43" (UID: "b4b973ee-209f-477b-92ee-3f1c16f17f43"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:06:59 crc kubenswrapper[4846]: I1005 07:06:59.696459 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b4b973ee-209f-477b-92ee-3f1c16f17f43-config-data" (OuterVolumeSpecName: "config-data") pod "b4b973ee-209f-477b-92ee-3f1c16f17f43" (UID: "b4b973ee-209f-477b-92ee-3f1c16f17f43"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:06:59 crc kubenswrapper[4846]: I1005 07:06:59.730226 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nbv6b\" (UniqueName: \"kubernetes.io/projected/36efa3ba-3fa1-4ecd-8ad3-9136d31b6070-kube-api-access-nbv6b\") on node \"crc\" DevicePath \"\"" Oct 05 07:06:59 crc kubenswrapper[4846]: I1005 07:06:59.730297 4846 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b4b973ee-209f-477b-92ee-3f1c16f17f43-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 07:06:59 crc kubenswrapper[4846]: I1005 07:06:59.730320 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5dsnc\" (UniqueName: \"kubernetes.io/projected/b4b973ee-209f-477b-92ee-3f1c16f17f43-kube-api-access-5dsnc\") on node \"crc\" DevicePath \"\"" Oct 05 07:06:59 crc kubenswrapper[4846]: I1005 07:06:59.730335 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lvkgg\" (UniqueName: \"kubernetes.io/projected/577b34f0-581f-4fe2-8717-7e3062fd26fc-kube-api-access-lvkgg\") on node \"crc\" DevicePath \"\"" Oct 05 07:06:59 crc kubenswrapper[4846]: I1005 07:06:59.730347 4846 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b4b973ee-209f-477b-92ee-3f1c16f17f43-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 07:07:00 crc kubenswrapper[4846]: I1005 07:07:00.180039 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-t7t67" event={"ID":"b4b973ee-209f-477b-92ee-3f1c16f17f43","Type":"ContainerDied","Data":"a3daa5a064fa48e66f1d8cc155b44ee62baf8fe1a2cd398a5df3a476e3e992c1"} Oct 05 07:07:00 crc kubenswrapper[4846]: I1005 07:07:00.181399 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a3daa5a064fa48e66f1d8cc155b44ee62baf8fe1a2cd398a5df3a476e3e992c1" Oct 05 07:07:00 crc kubenswrapper[4846]: I1005 07:07:00.180129 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-t7t67" Oct 05 07:07:00 crc kubenswrapper[4846]: I1005 07:07:00.182827 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-a5c0-account-create-8cbhj" Oct 05 07:07:00 crc kubenswrapper[4846]: I1005 07:07:00.182844 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-a5c0-account-create-8cbhj" event={"ID":"577b34f0-581f-4fe2-8717-7e3062fd26fc","Type":"ContainerDied","Data":"a5690caa347625c9e565fe502c81e638d024b7e5ef5c011bb82875d888aae201"} Oct 05 07:07:00 crc kubenswrapper[4846]: I1005 07:07:00.183082 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a5690caa347625c9e565fe502c81e638d024b7e5ef5c011bb82875d888aae201" Oct 05 07:07:00 crc kubenswrapper[4846]: I1005 07:07:00.185882 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-8fc5-account-create-fcxfk" Oct 05 07:07:00 crc kubenswrapper[4846]: I1005 07:07:00.185872 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-8fc5-account-create-fcxfk" event={"ID":"36efa3ba-3fa1-4ecd-8ad3-9136d31b6070","Type":"ContainerDied","Data":"5e51ebdda6e6ff2c9cf97ff11618c504fd9c5b9c2b44eb4d1bb8eb57f5741561"} Oct 05 07:07:00 crc kubenswrapper[4846]: I1005 07:07:00.186036 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5e51ebdda6e6ff2c9cf97ff11618c504fd9c5b9c2b44eb4d1bb8eb57f5741561" Oct 05 07:07:00 crc kubenswrapper[4846]: I1005 07:07:00.516247 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-dswh8"] Oct 05 07:07:00 crc kubenswrapper[4846]: E1005 07:07:00.516834 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b4b973ee-209f-477b-92ee-3f1c16f17f43" containerName="keystone-db-sync" Oct 05 07:07:00 crc kubenswrapper[4846]: I1005 07:07:00.516859 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="b4b973ee-209f-477b-92ee-3f1c16f17f43" containerName="keystone-db-sync" Oct 05 07:07:00 crc kubenswrapper[4846]: E1005 07:07:00.516902 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="438586d8-d6b1-445f-8460-0e44ac6037c6" containerName="mariadb-account-create" Oct 05 07:07:00 crc kubenswrapper[4846]: I1005 07:07:00.516911 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="438586d8-d6b1-445f-8460-0e44ac6037c6" containerName="mariadb-account-create" Oct 05 07:07:00 crc kubenswrapper[4846]: E1005 07:07:00.516932 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="577b34f0-581f-4fe2-8717-7e3062fd26fc" containerName="mariadb-account-create" Oct 05 07:07:00 crc kubenswrapper[4846]: I1005 07:07:00.516941 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="577b34f0-581f-4fe2-8717-7e3062fd26fc" containerName="mariadb-account-create" Oct 05 07:07:00 crc kubenswrapper[4846]: E1005 07:07:00.516962 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="36efa3ba-3fa1-4ecd-8ad3-9136d31b6070" containerName="mariadb-account-create" Oct 05 07:07:00 crc kubenswrapper[4846]: I1005 07:07:00.516972 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="36efa3ba-3fa1-4ecd-8ad3-9136d31b6070" containerName="mariadb-account-create" Oct 05 07:07:00 crc kubenswrapper[4846]: I1005 07:07:00.517243 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="36efa3ba-3fa1-4ecd-8ad3-9136d31b6070" containerName="mariadb-account-create" Oct 05 07:07:00 crc kubenswrapper[4846]: I1005 07:07:00.517278 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="577b34f0-581f-4fe2-8717-7e3062fd26fc" containerName="mariadb-account-create" Oct 05 07:07:00 crc kubenswrapper[4846]: I1005 07:07:00.517295 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="438586d8-d6b1-445f-8460-0e44ac6037c6" containerName="mariadb-account-create" Oct 05 07:07:00 crc kubenswrapper[4846]: I1005 07:07:00.517313 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="b4b973ee-209f-477b-92ee-3f1c16f17f43" containerName="keystone-db-sync" Oct 05 07:07:00 crc kubenswrapper[4846]: I1005 07:07:00.518395 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-dswh8" Oct 05 07:07:00 crc kubenswrapper[4846]: I1005 07:07:00.530881 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 05 07:07:00 crc kubenswrapper[4846]: I1005 07:07:00.531285 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 05 07:07:00 crc kubenswrapper[4846]: I1005 07:07:00.532270 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 05 07:07:00 crc kubenswrapper[4846]: I1005 07:07:00.536636 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-76d8v" Oct 05 07:07:00 crc kubenswrapper[4846]: I1005 07:07:00.550639 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/7760f846-37e0-4bb8-bc2b-2e10d8950022-credential-keys\") pod \"keystone-bootstrap-dswh8\" (UID: \"7760f846-37e0-4bb8-bc2b-2e10d8950022\") " pod="openstack/keystone-bootstrap-dswh8" Oct 05 07:07:00 crc kubenswrapper[4846]: I1005 07:07:00.550720 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/7760f846-37e0-4bb8-bc2b-2e10d8950022-fernet-keys\") pod \"keystone-bootstrap-dswh8\" (UID: \"7760f846-37e0-4bb8-bc2b-2e10d8950022\") " pod="openstack/keystone-bootstrap-dswh8" Oct 05 07:07:00 crc kubenswrapper[4846]: I1005 07:07:00.550751 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7760f846-37e0-4bb8-bc2b-2e10d8950022-scripts\") pod \"keystone-bootstrap-dswh8\" (UID: \"7760f846-37e0-4bb8-bc2b-2e10d8950022\") " pod="openstack/keystone-bootstrap-dswh8" Oct 05 07:07:00 crc kubenswrapper[4846]: I1005 07:07:00.550905 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7760f846-37e0-4bb8-bc2b-2e10d8950022-config-data\") pod \"keystone-bootstrap-dswh8\" (UID: \"7760f846-37e0-4bb8-bc2b-2e10d8950022\") " pod="openstack/keystone-bootstrap-dswh8" Oct 05 07:07:00 crc kubenswrapper[4846]: I1005 07:07:00.551130 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2fhhn\" (UniqueName: \"kubernetes.io/projected/7760f846-37e0-4bb8-bc2b-2e10d8950022-kube-api-access-2fhhn\") pod \"keystone-bootstrap-dswh8\" (UID: \"7760f846-37e0-4bb8-bc2b-2e10d8950022\") " pod="openstack/keystone-bootstrap-dswh8" Oct 05 07:07:00 crc kubenswrapper[4846]: I1005 07:07:00.551253 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7760f846-37e0-4bb8-bc2b-2e10d8950022-combined-ca-bundle\") pod \"keystone-bootstrap-dswh8\" (UID: \"7760f846-37e0-4bb8-bc2b-2e10d8950022\") " pod="openstack/keystone-bootstrap-dswh8" Oct 05 07:07:00 crc kubenswrapper[4846]: I1005 07:07:00.594576 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6f9b65c9b5-lqj5b"] Oct 05 07:07:00 crc kubenswrapper[4846]: I1005 07:07:00.619933 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-dswh8"] Oct 05 07:07:00 crc kubenswrapper[4846]: I1005 07:07:00.654073 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2fhhn\" (UniqueName: \"kubernetes.io/projected/7760f846-37e0-4bb8-bc2b-2e10d8950022-kube-api-access-2fhhn\") pod \"keystone-bootstrap-dswh8\" (UID: \"7760f846-37e0-4bb8-bc2b-2e10d8950022\") " pod="openstack/keystone-bootstrap-dswh8" Oct 05 07:07:00 crc kubenswrapper[4846]: I1005 07:07:00.654205 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7760f846-37e0-4bb8-bc2b-2e10d8950022-combined-ca-bundle\") pod \"keystone-bootstrap-dswh8\" (UID: \"7760f846-37e0-4bb8-bc2b-2e10d8950022\") " pod="openstack/keystone-bootstrap-dswh8" Oct 05 07:07:00 crc kubenswrapper[4846]: I1005 07:07:00.654257 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/7760f846-37e0-4bb8-bc2b-2e10d8950022-credential-keys\") pod \"keystone-bootstrap-dswh8\" (UID: \"7760f846-37e0-4bb8-bc2b-2e10d8950022\") " pod="openstack/keystone-bootstrap-dswh8" Oct 05 07:07:00 crc kubenswrapper[4846]: I1005 07:07:00.654301 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/7760f846-37e0-4bb8-bc2b-2e10d8950022-fernet-keys\") pod \"keystone-bootstrap-dswh8\" (UID: \"7760f846-37e0-4bb8-bc2b-2e10d8950022\") " pod="openstack/keystone-bootstrap-dswh8" Oct 05 07:07:00 crc kubenswrapper[4846]: I1005 07:07:00.654348 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7760f846-37e0-4bb8-bc2b-2e10d8950022-scripts\") pod \"keystone-bootstrap-dswh8\" (UID: \"7760f846-37e0-4bb8-bc2b-2e10d8950022\") " pod="openstack/keystone-bootstrap-dswh8" Oct 05 07:07:00 crc kubenswrapper[4846]: I1005 07:07:00.654396 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7760f846-37e0-4bb8-bc2b-2e10d8950022-config-data\") pod \"keystone-bootstrap-dswh8\" (UID: \"7760f846-37e0-4bb8-bc2b-2e10d8950022\") " pod="openstack/keystone-bootstrap-dswh8" Oct 05 07:07:00 crc kubenswrapper[4846]: I1005 07:07:00.654578 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6d6f4f645f-97gpx"] Oct 05 07:07:00 crc kubenswrapper[4846]: I1005 07:07:00.656357 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6d6f4f645f-97gpx" Oct 05 07:07:00 crc kubenswrapper[4846]: I1005 07:07:00.670864 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7760f846-37e0-4bb8-bc2b-2e10d8950022-scripts\") pod \"keystone-bootstrap-dswh8\" (UID: \"7760f846-37e0-4bb8-bc2b-2e10d8950022\") " pod="openstack/keystone-bootstrap-dswh8" Oct 05 07:07:00 crc kubenswrapper[4846]: I1005 07:07:00.670914 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7760f846-37e0-4bb8-bc2b-2e10d8950022-combined-ca-bundle\") pod \"keystone-bootstrap-dswh8\" (UID: \"7760f846-37e0-4bb8-bc2b-2e10d8950022\") " pod="openstack/keystone-bootstrap-dswh8" Oct 05 07:07:00 crc kubenswrapper[4846]: I1005 07:07:00.671597 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/7760f846-37e0-4bb8-bc2b-2e10d8950022-credential-keys\") pod \"keystone-bootstrap-dswh8\" (UID: \"7760f846-37e0-4bb8-bc2b-2e10d8950022\") " pod="openstack/keystone-bootstrap-dswh8" Oct 05 07:07:00 crc kubenswrapper[4846]: I1005 07:07:00.671690 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7760f846-37e0-4bb8-bc2b-2e10d8950022-config-data\") pod \"keystone-bootstrap-dswh8\" (UID: \"7760f846-37e0-4bb8-bc2b-2e10d8950022\") " pod="openstack/keystone-bootstrap-dswh8" Oct 05 07:07:00 crc kubenswrapper[4846]: I1005 07:07:00.672017 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6d6f4f645f-97gpx"] Oct 05 07:07:00 crc kubenswrapper[4846]: I1005 07:07:00.679817 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/7760f846-37e0-4bb8-bc2b-2e10d8950022-fernet-keys\") pod \"keystone-bootstrap-dswh8\" (UID: \"7760f846-37e0-4bb8-bc2b-2e10d8950022\") " pod="openstack/keystone-bootstrap-dswh8" Oct 05 07:07:00 crc kubenswrapper[4846]: I1005 07:07:00.688770 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2fhhn\" (UniqueName: \"kubernetes.io/projected/7760f846-37e0-4bb8-bc2b-2e10d8950022-kube-api-access-2fhhn\") pod \"keystone-bootstrap-dswh8\" (UID: \"7760f846-37e0-4bb8-bc2b-2e10d8950022\") " pod="openstack/keystone-bootstrap-dswh8" Oct 05 07:07:00 crc kubenswrapper[4846]: I1005 07:07:00.757760 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b5852c48-40ce-4e9f-8cab-d2822471fd26-ovsdbserver-sb\") pod \"dnsmasq-dns-6d6f4f645f-97gpx\" (UID: \"b5852c48-40ce-4e9f-8cab-d2822471fd26\") " pod="openstack/dnsmasq-dns-6d6f4f645f-97gpx" Oct 05 07:07:00 crc kubenswrapper[4846]: I1005 07:07:00.757928 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8z9xq\" (UniqueName: \"kubernetes.io/projected/b5852c48-40ce-4e9f-8cab-d2822471fd26-kube-api-access-8z9xq\") pod \"dnsmasq-dns-6d6f4f645f-97gpx\" (UID: \"b5852c48-40ce-4e9f-8cab-d2822471fd26\") " pod="openstack/dnsmasq-dns-6d6f4f645f-97gpx" Oct 05 07:07:00 crc kubenswrapper[4846]: I1005 07:07:00.758032 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b5852c48-40ce-4e9f-8cab-d2822471fd26-dns-svc\") pod \"dnsmasq-dns-6d6f4f645f-97gpx\" (UID: \"b5852c48-40ce-4e9f-8cab-d2822471fd26\") " pod="openstack/dnsmasq-dns-6d6f4f645f-97gpx" Oct 05 07:07:00 crc kubenswrapper[4846]: I1005 07:07:00.758136 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b5852c48-40ce-4e9f-8cab-d2822471fd26-dns-swift-storage-0\") pod \"dnsmasq-dns-6d6f4f645f-97gpx\" (UID: \"b5852c48-40ce-4e9f-8cab-d2822471fd26\") " pod="openstack/dnsmasq-dns-6d6f4f645f-97gpx" Oct 05 07:07:00 crc kubenswrapper[4846]: I1005 07:07:00.758266 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b5852c48-40ce-4e9f-8cab-d2822471fd26-ovsdbserver-nb\") pod \"dnsmasq-dns-6d6f4f645f-97gpx\" (UID: \"b5852c48-40ce-4e9f-8cab-d2822471fd26\") " pod="openstack/dnsmasq-dns-6d6f4f645f-97gpx" Oct 05 07:07:00 crc kubenswrapper[4846]: I1005 07:07:00.758375 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b5852c48-40ce-4e9f-8cab-d2822471fd26-config\") pod \"dnsmasq-dns-6d6f4f645f-97gpx\" (UID: \"b5852c48-40ce-4e9f-8cab-d2822471fd26\") " pod="openstack/dnsmasq-dns-6d6f4f645f-97gpx" Oct 05 07:07:00 crc kubenswrapper[4846]: I1005 07:07:00.803502 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 05 07:07:00 crc kubenswrapper[4846]: I1005 07:07:00.812285 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 05 07:07:00 crc kubenswrapper[4846]: I1005 07:07:00.817236 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 05 07:07:00 crc kubenswrapper[4846]: I1005 07:07:00.817414 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 05 07:07:00 crc kubenswrapper[4846]: I1005 07:07:00.827236 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 05 07:07:00 crc kubenswrapper[4846]: I1005 07:07:00.844538 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-sync-7f2fl"] Oct 05 07:07:00 crc kubenswrapper[4846]: I1005 07:07:00.845947 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-7f2fl" Oct 05 07:07:00 crc kubenswrapper[4846]: I1005 07:07:00.851068 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-5d5jr" Oct 05 07:07:00 crc kubenswrapper[4846]: I1005 07:07:00.851945 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Oct 05 07:07:00 crc kubenswrapper[4846]: I1005 07:07:00.853425 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Oct 05 07:07:00 crc kubenswrapper[4846]: I1005 07:07:00.866685 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a6ab76e7-ba49-4976-b718-3ea57bb4b53a-combined-ca-bundle\") pod \"placement-db-sync-7f2fl\" (UID: \"a6ab76e7-ba49-4976-b718-3ea57bb4b53a\") " pod="openstack/placement-db-sync-7f2fl" Oct 05 07:07:00 crc kubenswrapper[4846]: I1005 07:07:00.866739 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vmgr5\" (UniqueName: \"kubernetes.io/projected/ae01bc6c-75fb-4958-9a75-6c212fd0ed3a-kube-api-access-vmgr5\") pod \"ceilometer-0\" (UID: \"ae01bc6c-75fb-4958-9a75-6c212fd0ed3a\") " pod="openstack/ceilometer-0" Oct 05 07:07:00 crc kubenswrapper[4846]: I1005 07:07:00.866785 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b5852c48-40ce-4e9f-8cab-d2822471fd26-config\") pod \"dnsmasq-dns-6d6f4f645f-97gpx\" (UID: \"b5852c48-40ce-4e9f-8cab-d2822471fd26\") " pod="openstack/dnsmasq-dns-6d6f4f645f-97gpx" Oct 05 07:07:00 crc kubenswrapper[4846]: I1005 07:07:00.866980 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ae01bc6c-75fb-4958-9a75-6c212fd0ed3a-run-httpd\") pod \"ceilometer-0\" (UID: \"ae01bc6c-75fb-4958-9a75-6c212fd0ed3a\") " pod="openstack/ceilometer-0" Oct 05 07:07:00 crc kubenswrapper[4846]: I1005 07:07:00.867051 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b5852c48-40ce-4e9f-8cab-d2822471fd26-ovsdbserver-sb\") pod \"dnsmasq-dns-6d6f4f645f-97gpx\" (UID: \"b5852c48-40ce-4e9f-8cab-d2822471fd26\") " pod="openstack/dnsmasq-dns-6d6f4f645f-97gpx" Oct 05 07:07:00 crc kubenswrapper[4846]: I1005 07:07:00.867121 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8z9xq\" (UniqueName: \"kubernetes.io/projected/b5852c48-40ce-4e9f-8cab-d2822471fd26-kube-api-access-8z9xq\") pod \"dnsmasq-dns-6d6f4f645f-97gpx\" (UID: \"b5852c48-40ce-4e9f-8cab-d2822471fd26\") " pod="openstack/dnsmasq-dns-6d6f4f645f-97gpx" Oct 05 07:07:00 crc kubenswrapper[4846]: I1005 07:07:00.867157 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a6ab76e7-ba49-4976-b718-3ea57bb4b53a-config-data\") pod \"placement-db-sync-7f2fl\" (UID: \"a6ab76e7-ba49-4976-b718-3ea57bb4b53a\") " pod="openstack/placement-db-sync-7f2fl" Oct 05 07:07:00 crc kubenswrapper[4846]: I1005 07:07:00.867225 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ae01bc6c-75fb-4958-9a75-6c212fd0ed3a-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"ae01bc6c-75fb-4958-9a75-6c212fd0ed3a\") " pod="openstack/ceilometer-0" Oct 05 07:07:00 crc kubenswrapper[4846]: I1005 07:07:00.867242 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ae01bc6c-75fb-4958-9a75-6c212fd0ed3a-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"ae01bc6c-75fb-4958-9a75-6c212fd0ed3a\") " pod="openstack/ceilometer-0" Oct 05 07:07:00 crc kubenswrapper[4846]: I1005 07:07:00.867271 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b5852c48-40ce-4e9f-8cab-d2822471fd26-dns-svc\") pod \"dnsmasq-dns-6d6f4f645f-97gpx\" (UID: \"b5852c48-40ce-4e9f-8cab-d2822471fd26\") " pod="openstack/dnsmasq-dns-6d6f4f645f-97gpx" Oct 05 07:07:00 crc kubenswrapper[4846]: I1005 07:07:00.867309 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a6ab76e7-ba49-4976-b718-3ea57bb4b53a-scripts\") pod \"placement-db-sync-7f2fl\" (UID: \"a6ab76e7-ba49-4976-b718-3ea57bb4b53a\") " pod="openstack/placement-db-sync-7f2fl" Oct 05 07:07:00 crc kubenswrapper[4846]: I1005 07:07:00.867333 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ae01bc6c-75fb-4958-9a75-6c212fd0ed3a-scripts\") pod \"ceilometer-0\" (UID: \"ae01bc6c-75fb-4958-9a75-6c212fd0ed3a\") " pod="openstack/ceilometer-0" Oct 05 07:07:00 crc kubenswrapper[4846]: I1005 07:07:00.867350 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ntr8z\" (UniqueName: \"kubernetes.io/projected/a6ab76e7-ba49-4976-b718-3ea57bb4b53a-kube-api-access-ntr8z\") pod \"placement-db-sync-7f2fl\" (UID: \"a6ab76e7-ba49-4976-b718-3ea57bb4b53a\") " pod="openstack/placement-db-sync-7f2fl" Oct 05 07:07:00 crc kubenswrapper[4846]: I1005 07:07:00.867391 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ae01bc6c-75fb-4958-9a75-6c212fd0ed3a-log-httpd\") pod \"ceilometer-0\" (UID: \"ae01bc6c-75fb-4958-9a75-6c212fd0ed3a\") " pod="openstack/ceilometer-0" Oct 05 07:07:00 crc kubenswrapper[4846]: I1005 07:07:00.867403 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-7f2fl"] Oct 05 07:07:00 crc kubenswrapper[4846]: I1005 07:07:00.867468 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b5852c48-40ce-4e9f-8cab-d2822471fd26-dns-swift-storage-0\") pod \"dnsmasq-dns-6d6f4f645f-97gpx\" (UID: \"b5852c48-40ce-4e9f-8cab-d2822471fd26\") " pod="openstack/dnsmasq-dns-6d6f4f645f-97gpx" Oct 05 07:07:00 crc kubenswrapper[4846]: I1005 07:07:00.867492 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a6ab76e7-ba49-4976-b718-3ea57bb4b53a-logs\") pod \"placement-db-sync-7f2fl\" (UID: \"a6ab76e7-ba49-4976-b718-3ea57bb4b53a\") " pod="openstack/placement-db-sync-7f2fl" Oct 05 07:07:00 crc kubenswrapper[4846]: I1005 07:07:00.867523 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ae01bc6c-75fb-4958-9a75-6c212fd0ed3a-config-data\") pod \"ceilometer-0\" (UID: \"ae01bc6c-75fb-4958-9a75-6c212fd0ed3a\") " pod="openstack/ceilometer-0" Oct 05 07:07:00 crc kubenswrapper[4846]: I1005 07:07:00.867614 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b5852c48-40ce-4e9f-8cab-d2822471fd26-ovsdbserver-nb\") pod \"dnsmasq-dns-6d6f4f645f-97gpx\" (UID: \"b5852c48-40ce-4e9f-8cab-d2822471fd26\") " pod="openstack/dnsmasq-dns-6d6f4f645f-97gpx" Oct 05 07:07:00 crc kubenswrapper[4846]: I1005 07:07:00.867632 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b5852c48-40ce-4e9f-8cab-d2822471fd26-config\") pod \"dnsmasq-dns-6d6f4f645f-97gpx\" (UID: \"b5852c48-40ce-4e9f-8cab-d2822471fd26\") " pod="openstack/dnsmasq-dns-6d6f4f645f-97gpx" Oct 05 07:07:00 crc kubenswrapper[4846]: I1005 07:07:00.868580 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b5852c48-40ce-4e9f-8cab-d2822471fd26-dns-svc\") pod \"dnsmasq-dns-6d6f4f645f-97gpx\" (UID: \"b5852c48-40ce-4e9f-8cab-d2822471fd26\") " pod="openstack/dnsmasq-dns-6d6f4f645f-97gpx" Oct 05 07:07:00 crc kubenswrapper[4846]: I1005 07:07:00.868865 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b5852c48-40ce-4e9f-8cab-d2822471fd26-dns-swift-storage-0\") pod \"dnsmasq-dns-6d6f4f645f-97gpx\" (UID: \"b5852c48-40ce-4e9f-8cab-d2822471fd26\") " pod="openstack/dnsmasq-dns-6d6f4f645f-97gpx" Oct 05 07:07:00 crc kubenswrapper[4846]: I1005 07:07:00.869155 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b5852c48-40ce-4e9f-8cab-d2822471fd26-ovsdbserver-nb\") pod \"dnsmasq-dns-6d6f4f645f-97gpx\" (UID: \"b5852c48-40ce-4e9f-8cab-d2822471fd26\") " pod="openstack/dnsmasq-dns-6d6f4f645f-97gpx" Oct 05 07:07:00 crc kubenswrapper[4846]: I1005 07:07:00.869267 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b5852c48-40ce-4e9f-8cab-d2822471fd26-ovsdbserver-sb\") pod \"dnsmasq-dns-6d6f4f645f-97gpx\" (UID: \"b5852c48-40ce-4e9f-8cab-d2822471fd26\") " pod="openstack/dnsmasq-dns-6d6f4f645f-97gpx" Oct 05 07:07:00 crc kubenswrapper[4846]: I1005 07:07:00.879823 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-dswh8" Oct 05 07:07:00 crc kubenswrapper[4846]: I1005 07:07:00.909842 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6d6f4f645f-97gpx"] Oct 05 07:07:00 crc kubenswrapper[4846]: E1005 07:07:00.911426 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[kube-api-access-8z9xq], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openstack/dnsmasq-dns-6d6f4f645f-97gpx" podUID="b5852c48-40ce-4e9f-8cab-d2822471fd26" Oct 05 07:07:00 crc kubenswrapper[4846]: I1005 07:07:00.926155 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8z9xq\" (UniqueName: \"kubernetes.io/projected/b5852c48-40ce-4e9f-8cab-d2822471fd26-kube-api-access-8z9xq\") pod \"dnsmasq-dns-6d6f4f645f-97gpx\" (UID: \"b5852c48-40ce-4e9f-8cab-d2822471fd26\") " pod="openstack/dnsmasq-dns-6d6f4f645f-97gpx" Oct 05 07:07:00 crc kubenswrapper[4846]: I1005 07:07:00.964334 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7cf74955b9-njz69"] Oct 05 07:07:00 crc kubenswrapper[4846]: I1005 07:07:00.966273 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7cf74955b9-njz69" Oct 05 07:07:00 crc kubenswrapper[4846]: I1005 07:07:00.969421 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a6ab76e7-ba49-4976-b718-3ea57bb4b53a-combined-ca-bundle\") pod \"placement-db-sync-7f2fl\" (UID: \"a6ab76e7-ba49-4976-b718-3ea57bb4b53a\") " pod="openstack/placement-db-sync-7f2fl" Oct 05 07:07:00 crc kubenswrapper[4846]: I1005 07:07:00.969470 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vmgr5\" (UniqueName: \"kubernetes.io/projected/ae01bc6c-75fb-4958-9a75-6c212fd0ed3a-kube-api-access-vmgr5\") pod \"ceilometer-0\" (UID: \"ae01bc6c-75fb-4958-9a75-6c212fd0ed3a\") " pod="openstack/ceilometer-0" Oct 05 07:07:00 crc kubenswrapper[4846]: I1005 07:07:00.969528 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ae01bc6c-75fb-4958-9a75-6c212fd0ed3a-run-httpd\") pod \"ceilometer-0\" (UID: \"ae01bc6c-75fb-4958-9a75-6c212fd0ed3a\") " pod="openstack/ceilometer-0" Oct 05 07:07:00 crc kubenswrapper[4846]: I1005 07:07:00.969576 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a6ab76e7-ba49-4976-b718-3ea57bb4b53a-config-data\") pod \"placement-db-sync-7f2fl\" (UID: \"a6ab76e7-ba49-4976-b718-3ea57bb4b53a\") " pod="openstack/placement-db-sync-7f2fl" Oct 05 07:07:00 crc kubenswrapper[4846]: I1005 07:07:00.969603 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ae01bc6c-75fb-4958-9a75-6c212fd0ed3a-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"ae01bc6c-75fb-4958-9a75-6c212fd0ed3a\") " pod="openstack/ceilometer-0" Oct 05 07:07:00 crc kubenswrapper[4846]: I1005 07:07:00.969622 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ae01bc6c-75fb-4958-9a75-6c212fd0ed3a-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"ae01bc6c-75fb-4958-9a75-6c212fd0ed3a\") " pod="openstack/ceilometer-0" Oct 05 07:07:00 crc kubenswrapper[4846]: I1005 07:07:00.969649 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a6ab76e7-ba49-4976-b718-3ea57bb4b53a-scripts\") pod \"placement-db-sync-7f2fl\" (UID: \"a6ab76e7-ba49-4976-b718-3ea57bb4b53a\") " pod="openstack/placement-db-sync-7f2fl" Oct 05 07:07:00 crc kubenswrapper[4846]: I1005 07:07:00.969672 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ntr8z\" (UniqueName: \"kubernetes.io/projected/a6ab76e7-ba49-4976-b718-3ea57bb4b53a-kube-api-access-ntr8z\") pod \"placement-db-sync-7f2fl\" (UID: \"a6ab76e7-ba49-4976-b718-3ea57bb4b53a\") " pod="openstack/placement-db-sync-7f2fl" Oct 05 07:07:00 crc kubenswrapper[4846]: I1005 07:07:00.969691 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ae01bc6c-75fb-4958-9a75-6c212fd0ed3a-scripts\") pod \"ceilometer-0\" (UID: \"ae01bc6c-75fb-4958-9a75-6c212fd0ed3a\") " pod="openstack/ceilometer-0" Oct 05 07:07:00 crc kubenswrapper[4846]: I1005 07:07:00.969712 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ae01bc6c-75fb-4958-9a75-6c212fd0ed3a-log-httpd\") pod \"ceilometer-0\" (UID: \"ae01bc6c-75fb-4958-9a75-6c212fd0ed3a\") " pod="openstack/ceilometer-0" Oct 05 07:07:00 crc kubenswrapper[4846]: I1005 07:07:00.969744 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a6ab76e7-ba49-4976-b718-3ea57bb4b53a-logs\") pod \"placement-db-sync-7f2fl\" (UID: \"a6ab76e7-ba49-4976-b718-3ea57bb4b53a\") " pod="openstack/placement-db-sync-7f2fl" Oct 05 07:07:00 crc kubenswrapper[4846]: I1005 07:07:00.969765 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ae01bc6c-75fb-4958-9a75-6c212fd0ed3a-config-data\") pod \"ceilometer-0\" (UID: \"ae01bc6c-75fb-4958-9a75-6c212fd0ed3a\") " pod="openstack/ceilometer-0" Oct 05 07:07:00 crc kubenswrapper[4846]: I1005 07:07:00.971293 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ae01bc6c-75fb-4958-9a75-6c212fd0ed3a-run-httpd\") pod \"ceilometer-0\" (UID: \"ae01bc6c-75fb-4958-9a75-6c212fd0ed3a\") " pod="openstack/ceilometer-0" Oct 05 07:07:00 crc kubenswrapper[4846]: I1005 07:07:00.972050 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ae01bc6c-75fb-4958-9a75-6c212fd0ed3a-log-httpd\") pod \"ceilometer-0\" (UID: \"ae01bc6c-75fb-4958-9a75-6c212fd0ed3a\") " pod="openstack/ceilometer-0" Oct 05 07:07:00 crc kubenswrapper[4846]: I1005 07:07:00.972240 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a6ab76e7-ba49-4976-b718-3ea57bb4b53a-logs\") pod \"placement-db-sync-7f2fl\" (UID: \"a6ab76e7-ba49-4976-b718-3ea57bb4b53a\") " pod="openstack/placement-db-sync-7f2fl" Oct 05 07:07:00 crc kubenswrapper[4846]: I1005 07:07:00.974440 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7cf74955b9-njz69"] Oct 05 07:07:00 crc kubenswrapper[4846]: I1005 07:07:00.976214 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ae01bc6c-75fb-4958-9a75-6c212fd0ed3a-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"ae01bc6c-75fb-4958-9a75-6c212fd0ed3a\") " pod="openstack/ceilometer-0" Oct 05 07:07:00 crc kubenswrapper[4846]: I1005 07:07:00.976682 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ae01bc6c-75fb-4958-9a75-6c212fd0ed3a-scripts\") pod \"ceilometer-0\" (UID: \"ae01bc6c-75fb-4958-9a75-6c212fd0ed3a\") " pod="openstack/ceilometer-0" Oct 05 07:07:00 crc kubenswrapper[4846]: I1005 07:07:00.978621 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ae01bc6c-75fb-4958-9a75-6c212fd0ed3a-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"ae01bc6c-75fb-4958-9a75-6c212fd0ed3a\") " pod="openstack/ceilometer-0" Oct 05 07:07:00 crc kubenswrapper[4846]: I1005 07:07:00.979696 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a6ab76e7-ba49-4976-b718-3ea57bb4b53a-config-data\") pod \"placement-db-sync-7f2fl\" (UID: \"a6ab76e7-ba49-4976-b718-3ea57bb4b53a\") " pod="openstack/placement-db-sync-7f2fl" Oct 05 07:07:00 crc kubenswrapper[4846]: I1005 07:07:00.979846 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ae01bc6c-75fb-4958-9a75-6c212fd0ed3a-config-data\") pod \"ceilometer-0\" (UID: \"ae01bc6c-75fb-4958-9a75-6c212fd0ed3a\") " pod="openstack/ceilometer-0" Oct 05 07:07:00 crc kubenswrapper[4846]: I1005 07:07:00.986566 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a6ab76e7-ba49-4976-b718-3ea57bb4b53a-scripts\") pod \"placement-db-sync-7f2fl\" (UID: \"a6ab76e7-ba49-4976-b718-3ea57bb4b53a\") " pod="openstack/placement-db-sync-7f2fl" Oct 05 07:07:00 crc kubenswrapper[4846]: I1005 07:07:00.990279 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a6ab76e7-ba49-4976-b718-3ea57bb4b53a-combined-ca-bundle\") pod \"placement-db-sync-7f2fl\" (UID: \"a6ab76e7-ba49-4976-b718-3ea57bb4b53a\") " pod="openstack/placement-db-sync-7f2fl" Oct 05 07:07:00 crc kubenswrapper[4846]: I1005 07:07:00.991800 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vmgr5\" (UniqueName: \"kubernetes.io/projected/ae01bc6c-75fb-4958-9a75-6c212fd0ed3a-kube-api-access-vmgr5\") pod \"ceilometer-0\" (UID: \"ae01bc6c-75fb-4958-9a75-6c212fd0ed3a\") " pod="openstack/ceilometer-0" Oct 05 07:07:01 crc kubenswrapper[4846]: I1005 07:07:00.999956 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ntr8z\" (UniqueName: \"kubernetes.io/projected/a6ab76e7-ba49-4976-b718-3ea57bb4b53a-kube-api-access-ntr8z\") pod \"placement-db-sync-7f2fl\" (UID: \"a6ab76e7-ba49-4976-b718-3ea57bb4b53a\") " pod="openstack/placement-db-sync-7f2fl" Oct 05 07:07:01 crc kubenswrapper[4846]: I1005 07:07:01.086365 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5514c26f-0fe1-4e09-9292-896f21f90f71-config\") pod \"dnsmasq-dns-7cf74955b9-njz69\" (UID: \"5514c26f-0fe1-4e09-9292-896f21f90f71\") " pod="openstack/dnsmasq-dns-7cf74955b9-njz69" Oct 05 07:07:01 crc kubenswrapper[4846]: I1005 07:07:01.086424 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5514c26f-0fe1-4e09-9292-896f21f90f71-ovsdbserver-nb\") pod \"dnsmasq-dns-7cf74955b9-njz69\" (UID: \"5514c26f-0fe1-4e09-9292-896f21f90f71\") " pod="openstack/dnsmasq-dns-7cf74955b9-njz69" Oct 05 07:07:01 crc kubenswrapper[4846]: I1005 07:07:01.086468 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5514c26f-0fe1-4e09-9292-896f21f90f71-ovsdbserver-sb\") pod \"dnsmasq-dns-7cf74955b9-njz69\" (UID: \"5514c26f-0fe1-4e09-9292-896f21f90f71\") " pod="openstack/dnsmasq-dns-7cf74955b9-njz69" Oct 05 07:07:01 crc kubenswrapper[4846]: I1005 07:07:01.086498 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9jqwj\" (UniqueName: \"kubernetes.io/projected/5514c26f-0fe1-4e09-9292-896f21f90f71-kube-api-access-9jqwj\") pod \"dnsmasq-dns-7cf74955b9-njz69\" (UID: \"5514c26f-0fe1-4e09-9292-896f21f90f71\") " pod="openstack/dnsmasq-dns-7cf74955b9-njz69" Oct 05 07:07:01 crc kubenswrapper[4846]: I1005 07:07:01.086522 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5514c26f-0fe1-4e09-9292-896f21f90f71-dns-swift-storage-0\") pod \"dnsmasq-dns-7cf74955b9-njz69\" (UID: \"5514c26f-0fe1-4e09-9292-896f21f90f71\") " pod="openstack/dnsmasq-dns-7cf74955b9-njz69" Oct 05 07:07:01 crc kubenswrapper[4846]: I1005 07:07:01.086573 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5514c26f-0fe1-4e09-9292-896f21f90f71-dns-svc\") pod \"dnsmasq-dns-7cf74955b9-njz69\" (UID: \"5514c26f-0fe1-4e09-9292-896f21f90f71\") " pod="openstack/dnsmasq-dns-7cf74955b9-njz69" Oct 05 07:07:01 crc kubenswrapper[4846]: I1005 07:07:01.174227 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 05 07:07:01 crc kubenswrapper[4846]: I1005 07:07:01.187359 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-sync-cf6rt"] Oct 05 07:07:01 crc kubenswrapper[4846]: I1005 07:07:01.188102 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5514c26f-0fe1-4e09-9292-896f21f90f71-dns-svc\") pod \"dnsmasq-dns-7cf74955b9-njz69\" (UID: \"5514c26f-0fe1-4e09-9292-896f21f90f71\") " pod="openstack/dnsmasq-dns-7cf74955b9-njz69" Oct 05 07:07:01 crc kubenswrapper[4846]: I1005 07:07:01.188264 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5514c26f-0fe1-4e09-9292-896f21f90f71-config\") pod \"dnsmasq-dns-7cf74955b9-njz69\" (UID: \"5514c26f-0fe1-4e09-9292-896f21f90f71\") " pod="openstack/dnsmasq-dns-7cf74955b9-njz69" Oct 05 07:07:01 crc kubenswrapper[4846]: I1005 07:07:01.188293 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5514c26f-0fe1-4e09-9292-896f21f90f71-ovsdbserver-nb\") pod \"dnsmasq-dns-7cf74955b9-njz69\" (UID: \"5514c26f-0fe1-4e09-9292-896f21f90f71\") " pod="openstack/dnsmasq-dns-7cf74955b9-njz69" Oct 05 07:07:01 crc kubenswrapper[4846]: I1005 07:07:01.188495 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5514c26f-0fe1-4e09-9292-896f21f90f71-ovsdbserver-sb\") pod \"dnsmasq-dns-7cf74955b9-njz69\" (UID: \"5514c26f-0fe1-4e09-9292-896f21f90f71\") " pod="openstack/dnsmasq-dns-7cf74955b9-njz69" Oct 05 07:07:01 crc kubenswrapper[4846]: I1005 07:07:01.188626 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-cf6rt" Oct 05 07:07:01 crc kubenswrapper[4846]: I1005 07:07:01.189238 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5514c26f-0fe1-4e09-9292-896f21f90f71-dns-svc\") pod \"dnsmasq-dns-7cf74955b9-njz69\" (UID: \"5514c26f-0fe1-4e09-9292-896f21f90f71\") " pod="openstack/dnsmasq-dns-7cf74955b9-njz69" Oct 05 07:07:01 crc kubenswrapper[4846]: I1005 07:07:01.189372 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5514c26f-0fe1-4e09-9292-896f21f90f71-config\") pod \"dnsmasq-dns-7cf74955b9-njz69\" (UID: \"5514c26f-0fe1-4e09-9292-896f21f90f71\") " pod="openstack/dnsmasq-dns-7cf74955b9-njz69" Oct 05 07:07:01 crc kubenswrapper[4846]: I1005 07:07:01.190292 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5514c26f-0fe1-4e09-9292-896f21f90f71-ovsdbserver-sb\") pod \"dnsmasq-dns-7cf74955b9-njz69\" (UID: \"5514c26f-0fe1-4e09-9292-896f21f90f71\") " pod="openstack/dnsmasq-dns-7cf74955b9-njz69" Oct 05 07:07:01 crc kubenswrapper[4846]: I1005 07:07:01.190357 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5514c26f-0fe1-4e09-9292-896f21f90f71-ovsdbserver-nb\") pod \"dnsmasq-dns-7cf74955b9-njz69\" (UID: \"5514c26f-0fe1-4e09-9292-896f21f90f71\") " pod="openstack/dnsmasq-dns-7cf74955b9-njz69" Oct 05 07:07:01 crc kubenswrapper[4846]: I1005 07:07:01.191027 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9jqwj\" (UniqueName: \"kubernetes.io/projected/5514c26f-0fe1-4e09-9292-896f21f90f71-kube-api-access-9jqwj\") pod \"dnsmasq-dns-7cf74955b9-njz69\" (UID: \"5514c26f-0fe1-4e09-9292-896f21f90f71\") " pod="openstack/dnsmasq-dns-7cf74955b9-njz69" Oct 05 07:07:01 crc kubenswrapper[4846]: I1005 07:07:01.191087 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5514c26f-0fe1-4e09-9292-896f21f90f71-dns-swift-storage-0\") pod \"dnsmasq-dns-7cf74955b9-njz69\" (UID: \"5514c26f-0fe1-4e09-9292-896f21f90f71\") " pod="openstack/dnsmasq-dns-7cf74955b9-njz69" Oct 05 07:07:01 crc kubenswrapper[4846]: I1005 07:07:01.191219 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-7f2fl" Oct 05 07:07:01 crc kubenswrapper[4846]: I1005 07:07:01.192082 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5514c26f-0fe1-4e09-9292-896f21f90f71-dns-swift-storage-0\") pod \"dnsmasq-dns-7cf74955b9-njz69\" (UID: \"5514c26f-0fe1-4e09-9292-896f21f90f71\") " pod="openstack/dnsmasq-dns-7cf74955b9-njz69" Oct 05 07:07:01 crc kubenswrapper[4846]: I1005 07:07:01.193004 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Oct 05 07:07:01 crc kubenswrapper[4846]: I1005 07:07:01.193495 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-chkrj" Oct 05 07:07:01 crc kubenswrapper[4846]: I1005 07:07:01.193678 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Oct 05 07:07:01 crc kubenswrapper[4846]: I1005 07:07:01.200162 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6f9b65c9b5-lqj5b" podUID="8c207acc-07b9-4d9f-8b20-c84c2b67560f" containerName="dnsmasq-dns" containerID="cri-o://8d03c956f45603fb749baa477ec5fb7e31072932fb32c3bef02bba3de7b751e9" gracePeriod=10 Oct 05 07:07:01 crc kubenswrapper[4846]: I1005 07:07:01.200388 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6d6f4f645f-97gpx" Oct 05 07:07:01 crc kubenswrapper[4846]: I1005 07:07:01.231393 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9jqwj\" (UniqueName: \"kubernetes.io/projected/5514c26f-0fe1-4e09-9292-896f21f90f71-kube-api-access-9jqwj\") pod \"dnsmasq-dns-7cf74955b9-njz69\" (UID: \"5514c26f-0fe1-4e09-9292-896f21f90f71\") " pod="openstack/dnsmasq-dns-7cf74955b9-njz69" Oct 05 07:07:01 crc kubenswrapper[4846]: I1005 07:07:01.234316 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-cf6rt"] Oct 05 07:07:01 crc kubenswrapper[4846]: I1005 07:07:01.297939 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-sync-t5brq"] Oct 05 07:07:01 crc kubenswrapper[4846]: I1005 07:07:01.299644 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-t5brq" Oct 05 07:07:01 crc kubenswrapper[4846]: I1005 07:07:01.300041 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/ee312852-b2eb-43d8-864c-bc553fed7ea5-db-sync-config-data\") pod \"cinder-db-sync-cf6rt\" (UID: \"ee312852-b2eb-43d8-864c-bc553fed7ea5\") " pod="openstack/cinder-db-sync-cf6rt" Oct 05 07:07:01 crc kubenswrapper[4846]: I1005 07:07:01.300130 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j2hlh\" (UniqueName: \"kubernetes.io/projected/ee312852-b2eb-43d8-864c-bc553fed7ea5-kube-api-access-j2hlh\") pod \"cinder-db-sync-cf6rt\" (UID: \"ee312852-b2eb-43d8-864c-bc553fed7ea5\") " pod="openstack/cinder-db-sync-cf6rt" Oct 05 07:07:01 crc kubenswrapper[4846]: I1005 07:07:01.300213 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ee312852-b2eb-43d8-864c-bc553fed7ea5-combined-ca-bundle\") pod \"cinder-db-sync-cf6rt\" (UID: \"ee312852-b2eb-43d8-864c-bc553fed7ea5\") " pod="openstack/cinder-db-sync-cf6rt" Oct 05 07:07:01 crc kubenswrapper[4846]: I1005 07:07:01.300318 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ee312852-b2eb-43d8-864c-bc553fed7ea5-scripts\") pod \"cinder-db-sync-cf6rt\" (UID: \"ee312852-b2eb-43d8-864c-bc553fed7ea5\") " pod="openstack/cinder-db-sync-cf6rt" Oct 05 07:07:01 crc kubenswrapper[4846]: I1005 07:07:01.300360 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ee312852-b2eb-43d8-864c-bc553fed7ea5-etc-machine-id\") pod \"cinder-db-sync-cf6rt\" (UID: \"ee312852-b2eb-43d8-864c-bc553fed7ea5\") " pod="openstack/cinder-db-sync-cf6rt" Oct 05 07:07:01 crc kubenswrapper[4846]: I1005 07:07:01.300395 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ee312852-b2eb-43d8-864c-bc553fed7ea5-config-data\") pod \"cinder-db-sync-cf6rt\" (UID: \"ee312852-b2eb-43d8-864c-bc553fed7ea5\") " pod="openstack/cinder-db-sync-cf6rt" Oct 05 07:07:01 crc kubenswrapper[4846]: I1005 07:07:01.302744 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-2xgzt" Oct 05 07:07:01 crc kubenswrapper[4846]: I1005 07:07:01.303375 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Oct 05 07:07:01 crc kubenswrapper[4846]: I1005 07:07:01.315851 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-t5brq"] Oct 05 07:07:01 crc kubenswrapper[4846]: I1005 07:07:01.401658 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/ee312852-b2eb-43d8-864c-bc553fed7ea5-db-sync-config-data\") pod \"cinder-db-sync-cf6rt\" (UID: \"ee312852-b2eb-43d8-864c-bc553fed7ea5\") " pod="openstack/cinder-db-sync-cf6rt" Oct 05 07:07:01 crc kubenswrapper[4846]: I1005 07:07:01.401724 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j2hlh\" (UniqueName: \"kubernetes.io/projected/ee312852-b2eb-43d8-864c-bc553fed7ea5-kube-api-access-j2hlh\") pod \"cinder-db-sync-cf6rt\" (UID: \"ee312852-b2eb-43d8-864c-bc553fed7ea5\") " pod="openstack/cinder-db-sync-cf6rt" Oct 05 07:07:01 crc kubenswrapper[4846]: I1005 07:07:01.405625 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ee312852-b2eb-43d8-864c-bc553fed7ea5-combined-ca-bundle\") pod \"cinder-db-sync-cf6rt\" (UID: \"ee312852-b2eb-43d8-864c-bc553fed7ea5\") " pod="openstack/cinder-db-sync-cf6rt" Oct 05 07:07:01 crc kubenswrapper[4846]: I1005 07:07:01.405684 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/b7a58faa-0f37-4b0e-a907-6a8bae28cb27-db-sync-config-data\") pod \"barbican-db-sync-t5brq\" (UID: \"b7a58faa-0f37-4b0e-a907-6a8bae28cb27\") " pod="openstack/barbican-db-sync-t5brq" Oct 05 07:07:01 crc kubenswrapper[4846]: I1005 07:07:01.405911 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7cf74955b9-njz69" Oct 05 07:07:01 crc kubenswrapper[4846]: I1005 07:07:01.405939 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b7a58faa-0f37-4b0e-a907-6a8bae28cb27-combined-ca-bundle\") pod \"barbican-db-sync-t5brq\" (UID: \"b7a58faa-0f37-4b0e-a907-6a8bae28cb27\") " pod="openstack/barbican-db-sync-t5brq" Oct 05 07:07:01 crc kubenswrapper[4846]: I1005 07:07:01.405985 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ee312852-b2eb-43d8-864c-bc553fed7ea5-scripts\") pod \"cinder-db-sync-cf6rt\" (UID: \"ee312852-b2eb-43d8-864c-bc553fed7ea5\") " pod="openstack/cinder-db-sync-cf6rt" Oct 05 07:07:01 crc kubenswrapper[4846]: I1005 07:07:01.406040 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ee312852-b2eb-43d8-864c-bc553fed7ea5-etc-machine-id\") pod \"cinder-db-sync-cf6rt\" (UID: \"ee312852-b2eb-43d8-864c-bc553fed7ea5\") " pod="openstack/cinder-db-sync-cf6rt" Oct 05 07:07:01 crc kubenswrapper[4846]: I1005 07:07:01.406099 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cft2p\" (UniqueName: \"kubernetes.io/projected/b7a58faa-0f37-4b0e-a907-6a8bae28cb27-kube-api-access-cft2p\") pod \"barbican-db-sync-t5brq\" (UID: \"b7a58faa-0f37-4b0e-a907-6a8bae28cb27\") " pod="openstack/barbican-db-sync-t5brq" Oct 05 07:07:01 crc kubenswrapper[4846]: I1005 07:07:01.406136 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ee312852-b2eb-43d8-864c-bc553fed7ea5-config-data\") pod \"cinder-db-sync-cf6rt\" (UID: \"ee312852-b2eb-43d8-864c-bc553fed7ea5\") " pod="openstack/cinder-db-sync-cf6rt" Oct 05 07:07:01 crc kubenswrapper[4846]: I1005 07:07:01.406908 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ee312852-b2eb-43d8-864c-bc553fed7ea5-etc-machine-id\") pod \"cinder-db-sync-cf6rt\" (UID: \"ee312852-b2eb-43d8-864c-bc553fed7ea5\") " pod="openstack/cinder-db-sync-cf6rt" Oct 05 07:07:01 crc kubenswrapper[4846]: I1005 07:07:01.415147 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ee312852-b2eb-43d8-864c-bc553fed7ea5-combined-ca-bundle\") pod \"cinder-db-sync-cf6rt\" (UID: \"ee312852-b2eb-43d8-864c-bc553fed7ea5\") " pod="openstack/cinder-db-sync-cf6rt" Oct 05 07:07:01 crc kubenswrapper[4846]: I1005 07:07:01.430785 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/ee312852-b2eb-43d8-864c-bc553fed7ea5-db-sync-config-data\") pod \"cinder-db-sync-cf6rt\" (UID: \"ee312852-b2eb-43d8-864c-bc553fed7ea5\") " pod="openstack/cinder-db-sync-cf6rt" Oct 05 07:07:01 crc kubenswrapper[4846]: I1005 07:07:01.431951 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ee312852-b2eb-43d8-864c-bc553fed7ea5-scripts\") pod \"cinder-db-sync-cf6rt\" (UID: \"ee312852-b2eb-43d8-864c-bc553fed7ea5\") " pod="openstack/cinder-db-sync-cf6rt" Oct 05 07:07:01 crc kubenswrapper[4846]: I1005 07:07:01.432086 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ee312852-b2eb-43d8-864c-bc553fed7ea5-config-data\") pod \"cinder-db-sync-cf6rt\" (UID: \"ee312852-b2eb-43d8-864c-bc553fed7ea5\") " pod="openstack/cinder-db-sync-cf6rt" Oct 05 07:07:01 crc kubenswrapper[4846]: I1005 07:07:01.433395 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6d6f4f645f-97gpx" Oct 05 07:07:01 crc kubenswrapper[4846]: I1005 07:07:01.439119 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j2hlh\" (UniqueName: \"kubernetes.io/projected/ee312852-b2eb-43d8-864c-bc553fed7ea5-kube-api-access-j2hlh\") pod \"cinder-db-sync-cf6rt\" (UID: \"ee312852-b2eb-43d8-864c-bc553fed7ea5\") " pod="openstack/cinder-db-sync-cf6rt" Oct 05 07:07:01 crc kubenswrapper[4846]: I1005 07:07:01.444541 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-dswh8"] Oct 05 07:07:01 crc kubenswrapper[4846]: I1005 07:07:01.507947 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b5852c48-40ce-4e9f-8cab-d2822471fd26-ovsdbserver-sb\") pod \"b5852c48-40ce-4e9f-8cab-d2822471fd26\" (UID: \"b5852c48-40ce-4e9f-8cab-d2822471fd26\") " Oct 05 07:07:01 crc kubenswrapper[4846]: I1005 07:07:01.508639 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b5852c48-40ce-4e9f-8cab-d2822471fd26-dns-svc\") pod \"b5852c48-40ce-4e9f-8cab-d2822471fd26\" (UID: \"b5852c48-40ce-4e9f-8cab-d2822471fd26\") " Oct 05 07:07:01 crc kubenswrapper[4846]: I1005 07:07:01.508773 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8z9xq\" (UniqueName: \"kubernetes.io/projected/b5852c48-40ce-4e9f-8cab-d2822471fd26-kube-api-access-8z9xq\") pod \"b5852c48-40ce-4e9f-8cab-d2822471fd26\" (UID: \"b5852c48-40ce-4e9f-8cab-d2822471fd26\") " Oct 05 07:07:01 crc kubenswrapper[4846]: I1005 07:07:01.508834 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b5852c48-40ce-4e9f-8cab-d2822471fd26-ovsdbserver-nb\") pod \"b5852c48-40ce-4e9f-8cab-d2822471fd26\" (UID: \"b5852c48-40ce-4e9f-8cab-d2822471fd26\") " Oct 05 07:07:01 crc kubenswrapper[4846]: I1005 07:07:01.508875 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b5852c48-40ce-4e9f-8cab-d2822471fd26-dns-swift-storage-0\") pod \"b5852c48-40ce-4e9f-8cab-d2822471fd26\" (UID: \"b5852c48-40ce-4e9f-8cab-d2822471fd26\") " Oct 05 07:07:01 crc kubenswrapper[4846]: I1005 07:07:01.508988 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b5852c48-40ce-4e9f-8cab-d2822471fd26-config\") pod \"b5852c48-40ce-4e9f-8cab-d2822471fd26\" (UID: \"b5852c48-40ce-4e9f-8cab-d2822471fd26\") " Oct 05 07:07:01 crc kubenswrapper[4846]: I1005 07:07:01.514311 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b5852c48-40ce-4e9f-8cab-d2822471fd26-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "b5852c48-40ce-4e9f-8cab-d2822471fd26" (UID: "b5852c48-40ce-4e9f-8cab-d2822471fd26"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 07:07:01 crc kubenswrapper[4846]: I1005 07:07:01.515689 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cft2p\" (UniqueName: \"kubernetes.io/projected/b7a58faa-0f37-4b0e-a907-6a8bae28cb27-kube-api-access-cft2p\") pod \"barbican-db-sync-t5brq\" (UID: \"b7a58faa-0f37-4b0e-a907-6a8bae28cb27\") " pod="openstack/barbican-db-sync-t5brq" Oct 05 07:07:01 crc kubenswrapper[4846]: I1005 07:07:01.516332 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/b7a58faa-0f37-4b0e-a907-6a8bae28cb27-db-sync-config-data\") pod \"barbican-db-sync-t5brq\" (UID: \"b7a58faa-0f37-4b0e-a907-6a8bae28cb27\") " pod="openstack/barbican-db-sync-t5brq" Oct 05 07:07:01 crc kubenswrapper[4846]: I1005 07:07:01.516457 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b5852c48-40ce-4e9f-8cab-d2822471fd26-config" (OuterVolumeSpecName: "config") pod "b5852c48-40ce-4e9f-8cab-d2822471fd26" (UID: "b5852c48-40ce-4e9f-8cab-d2822471fd26"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 07:07:01 crc kubenswrapper[4846]: I1005 07:07:01.516818 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b5852c48-40ce-4e9f-8cab-d2822471fd26-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "b5852c48-40ce-4e9f-8cab-d2822471fd26" (UID: "b5852c48-40ce-4e9f-8cab-d2822471fd26"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 07:07:01 crc kubenswrapper[4846]: I1005 07:07:01.517793 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b5852c48-40ce-4e9f-8cab-d2822471fd26-kube-api-access-8z9xq" (OuterVolumeSpecName: "kube-api-access-8z9xq") pod "b5852c48-40ce-4e9f-8cab-d2822471fd26" (UID: "b5852c48-40ce-4e9f-8cab-d2822471fd26"). InnerVolumeSpecName "kube-api-access-8z9xq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:07:01 crc kubenswrapper[4846]: I1005 07:07:01.518149 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b5852c48-40ce-4e9f-8cab-d2822471fd26-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "b5852c48-40ce-4e9f-8cab-d2822471fd26" (UID: "b5852c48-40ce-4e9f-8cab-d2822471fd26"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 07:07:01 crc kubenswrapper[4846]: I1005 07:07:01.518219 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b5852c48-40ce-4e9f-8cab-d2822471fd26-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "b5852c48-40ce-4e9f-8cab-d2822471fd26" (UID: "b5852c48-40ce-4e9f-8cab-d2822471fd26"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 07:07:01 crc kubenswrapper[4846]: I1005 07:07:01.521295 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b7a58faa-0f37-4b0e-a907-6a8bae28cb27-combined-ca-bundle\") pod \"barbican-db-sync-t5brq\" (UID: \"b7a58faa-0f37-4b0e-a907-6a8bae28cb27\") " pod="openstack/barbican-db-sync-t5brq" Oct 05 07:07:01 crc kubenswrapper[4846]: I1005 07:07:01.521597 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8z9xq\" (UniqueName: \"kubernetes.io/projected/b5852c48-40ce-4e9f-8cab-d2822471fd26-kube-api-access-8z9xq\") on node \"crc\" DevicePath \"\"" Oct 05 07:07:01 crc kubenswrapper[4846]: I1005 07:07:01.521614 4846 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b5852c48-40ce-4e9f-8cab-d2822471fd26-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 05 07:07:01 crc kubenswrapper[4846]: I1005 07:07:01.521623 4846 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b5852c48-40ce-4e9f-8cab-d2822471fd26-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 05 07:07:01 crc kubenswrapper[4846]: I1005 07:07:01.521634 4846 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b5852c48-40ce-4e9f-8cab-d2822471fd26-config\") on node \"crc\" DevicePath \"\"" Oct 05 07:07:01 crc kubenswrapper[4846]: I1005 07:07:01.521643 4846 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b5852c48-40ce-4e9f-8cab-d2822471fd26-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 05 07:07:01 crc kubenswrapper[4846]: I1005 07:07:01.521653 4846 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b5852c48-40ce-4e9f-8cab-d2822471fd26-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 05 07:07:01 crc kubenswrapper[4846]: I1005 07:07:01.528609 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-sync-2l7b4"] Oct 05 07:07:01 crc kubenswrapper[4846]: I1005 07:07:01.529901 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-2l7b4" Oct 05 07:07:01 crc kubenswrapper[4846]: I1005 07:07:01.534235 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-xft85" Oct 05 07:07:01 crc kubenswrapper[4846]: I1005 07:07:01.534469 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Oct 05 07:07:01 crc kubenswrapper[4846]: I1005 07:07:01.534573 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Oct 05 07:07:01 crc kubenswrapper[4846]: I1005 07:07:01.534954 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b7a58faa-0f37-4b0e-a907-6a8bae28cb27-combined-ca-bundle\") pod \"barbican-db-sync-t5brq\" (UID: \"b7a58faa-0f37-4b0e-a907-6a8bae28cb27\") " pod="openstack/barbican-db-sync-t5brq" Oct 05 07:07:01 crc kubenswrapper[4846]: I1005 07:07:01.538807 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/b7a58faa-0f37-4b0e-a907-6a8bae28cb27-db-sync-config-data\") pod \"barbican-db-sync-t5brq\" (UID: \"b7a58faa-0f37-4b0e-a907-6a8bae28cb27\") " pod="openstack/barbican-db-sync-t5brq" Oct 05 07:07:01 crc kubenswrapper[4846]: I1005 07:07:01.547569 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-2l7b4"] Oct 05 07:07:01 crc kubenswrapper[4846]: I1005 07:07:01.548856 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cft2p\" (UniqueName: \"kubernetes.io/projected/b7a58faa-0f37-4b0e-a907-6a8bae28cb27-kube-api-access-cft2p\") pod \"barbican-db-sync-t5brq\" (UID: \"b7a58faa-0f37-4b0e-a907-6a8bae28cb27\") " pod="openstack/barbican-db-sync-t5brq" Oct 05 07:07:01 crc kubenswrapper[4846]: I1005 07:07:01.589354 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 05 07:07:01 crc kubenswrapper[4846]: I1005 07:07:01.626532 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/371a8404-6b77-418f-8fcb-3c72ed092816-combined-ca-bundle\") pod \"neutron-db-sync-2l7b4\" (UID: \"371a8404-6b77-418f-8fcb-3c72ed092816\") " pod="openstack/neutron-db-sync-2l7b4" Oct 05 07:07:01 crc kubenswrapper[4846]: I1005 07:07:01.626634 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9pm8d\" (UniqueName: \"kubernetes.io/projected/371a8404-6b77-418f-8fcb-3c72ed092816-kube-api-access-9pm8d\") pod \"neutron-db-sync-2l7b4\" (UID: \"371a8404-6b77-418f-8fcb-3c72ed092816\") " pod="openstack/neutron-db-sync-2l7b4" Oct 05 07:07:01 crc kubenswrapper[4846]: I1005 07:07:01.626663 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/371a8404-6b77-418f-8fcb-3c72ed092816-config\") pod \"neutron-db-sync-2l7b4\" (UID: \"371a8404-6b77-418f-8fcb-3c72ed092816\") " pod="openstack/neutron-db-sync-2l7b4" Oct 05 07:07:01 crc kubenswrapper[4846]: W1005 07:07:01.640872 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podae01bc6c_75fb_4958_9a75_6c212fd0ed3a.slice/crio-2545b9452eaf83986bb5d51763da560439fe6f42fda5b657635a72196e0d38e2 WatchSource:0}: Error finding container 2545b9452eaf83986bb5d51763da560439fe6f42fda5b657635a72196e0d38e2: Status 404 returned error can't find the container with id 2545b9452eaf83986bb5d51763da560439fe6f42fda5b657635a72196e0d38e2 Oct 05 07:07:01 crc kubenswrapper[4846]: I1005 07:07:01.728802 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/371a8404-6b77-418f-8fcb-3c72ed092816-combined-ca-bundle\") pod \"neutron-db-sync-2l7b4\" (UID: \"371a8404-6b77-418f-8fcb-3c72ed092816\") " pod="openstack/neutron-db-sync-2l7b4" Oct 05 07:07:01 crc kubenswrapper[4846]: I1005 07:07:01.729975 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9pm8d\" (UniqueName: \"kubernetes.io/projected/371a8404-6b77-418f-8fcb-3c72ed092816-kube-api-access-9pm8d\") pod \"neutron-db-sync-2l7b4\" (UID: \"371a8404-6b77-418f-8fcb-3c72ed092816\") " pod="openstack/neutron-db-sync-2l7b4" Oct 05 07:07:01 crc kubenswrapper[4846]: I1005 07:07:01.730004 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/371a8404-6b77-418f-8fcb-3c72ed092816-config\") pod \"neutron-db-sync-2l7b4\" (UID: \"371a8404-6b77-418f-8fcb-3c72ed092816\") " pod="openstack/neutron-db-sync-2l7b4" Oct 05 07:07:01 crc kubenswrapper[4846]: I1005 07:07:01.732602 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-cf6rt" Oct 05 07:07:01 crc kubenswrapper[4846]: I1005 07:07:01.753517 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/371a8404-6b77-418f-8fcb-3c72ed092816-combined-ca-bundle\") pod \"neutron-db-sync-2l7b4\" (UID: \"371a8404-6b77-418f-8fcb-3c72ed092816\") " pod="openstack/neutron-db-sync-2l7b4" Oct 05 07:07:01 crc kubenswrapper[4846]: I1005 07:07:01.754449 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/371a8404-6b77-418f-8fcb-3c72ed092816-config\") pod \"neutron-db-sync-2l7b4\" (UID: \"371a8404-6b77-418f-8fcb-3c72ed092816\") " pod="openstack/neutron-db-sync-2l7b4" Oct 05 07:07:01 crc kubenswrapper[4846]: I1005 07:07:01.760034 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9pm8d\" (UniqueName: \"kubernetes.io/projected/371a8404-6b77-418f-8fcb-3c72ed092816-kube-api-access-9pm8d\") pod \"neutron-db-sync-2l7b4\" (UID: \"371a8404-6b77-418f-8fcb-3c72ed092816\") " pod="openstack/neutron-db-sync-2l7b4" Oct 05 07:07:01 crc kubenswrapper[4846]: I1005 07:07:01.781859 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-t5brq" Oct 05 07:07:01 crc kubenswrapper[4846]: I1005 07:07:01.890034 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-2l7b4" Oct 05 07:07:01 crc kubenswrapper[4846]: I1005 07:07:01.900569 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-7f2fl"] Oct 05 07:07:01 crc kubenswrapper[4846]: I1005 07:07:01.932070 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6f9b65c9b5-lqj5b" Oct 05 07:07:02 crc kubenswrapper[4846]: I1005 07:07:02.041015 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8c207acc-07b9-4d9f-8b20-c84c2b67560f-ovsdbserver-nb\") pod \"8c207acc-07b9-4d9f-8b20-c84c2b67560f\" (UID: \"8c207acc-07b9-4d9f-8b20-c84c2b67560f\") " Oct 05 07:07:02 crc kubenswrapper[4846]: I1005 07:07:02.042213 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8c207acc-07b9-4d9f-8b20-c84c2b67560f-ovsdbserver-sb\") pod \"8c207acc-07b9-4d9f-8b20-c84c2b67560f\" (UID: \"8c207acc-07b9-4d9f-8b20-c84c2b67560f\") " Oct 05 07:07:02 crc kubenswrapper[4846]: I1005 07:07:02.042263 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8c207acc-07b9-4d9f-8b20-c84c2b67560f-dns-svc\") pod \"8c207acc-07b9-4d9f-8b20-c84c2b67560f\" (UID: \"8c207acc-07b9-4d9f-8b20-c84c2b67560f\") " Oct 05 07:07:02 crc kubenswrapper[4846]: I1005 07:07:02.042306 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v7mt2\" (UniqueName: \"kubernetes.io/projected/8c207acc-07b9-4d9f-8b20-c84c2b67560f-kube-api-access-v7mt2\") pod \"8c207acc-07b9-4d9f-8b20-c84c2b67560f\" (UID: \"8c207acc-07b9-4d9f-8b20-c84c2b67560f\") " Oct 05 07:07:02 crc kubenswrapper[4846]: I1005 07:07:02.042474 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8c207acc-07b9-4d9f-8b20-c84c2b67560f-config\") pod \"8c207acc-07b9-4d9f-8b20-c84c2b67560f\" (UID: \"8c207acc-07b9-4d9f-8b20-c84c2b67560f\") " Oct 05 07:07:02 crc kubenswrapper[4846]: I1005 07:07:02.042639 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8c207acc-07b9-4d9f-8b20-c84c2b67560f-dns-swift-storage-0\") pod \"8c207acc-07b9-4d9f-8b20-c84c2b67560f\" (UID: \"8c207acc-07b9-4d9f-8b20-c84c2b67560f\") " Oct 05 07:07:02 crc kubenswrapper[4846]: I1005 07:07:02.054865 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8c207acc-07b9-4d9f-8b20-c84c2b67560f-kube-api-access-v7mt2" (OuterVolumeSpecName: "kube-api-access-v7mt2") pod "8c207acc-07b9-4d9f-8b20-c84c2b67560f" (UID: "8c207acc-07b9-4d9f-8b20-c84c2b67560f"). InnerVolumeSpecName "kube-api-access-v7mt2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:07:02 crc kubenswrapper[4846]: I1005 07:07:02.140448 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8c207acc-07b9-4d9f-8b20-c84c2b67560f-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "8c207acc-07b9-4d9f-8b20-c84c2b67560f" (UID: "8c207acc-07b9-4d9f-8b20-c84c2b67560f"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 07:07:02 crc kubenswrapper[4846]: I1005 07:07:02.145025 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7cf74955b9-njz69"] Oct 05 07:07:02 crc kubenswrapper[4846]: I1005 07:07:02.145369 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v7mt2\" (UniqueName: \"kubernetes.io/projected/8c207acc-07b9-4d9f-8b20-c84c2b67560f-kube-api-access-v7mt2\") on node \"crc\" DevicePath \"\"" Oct 05 07:07:02 crc kubenswrapper[4846]: I1005 07:07:02.145399 4846 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8c207acc-07b9-4d9f-8b20-c84c2b67560f-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 05 07:07:02 crc kubenswrapper[4846]: W1005 07:07:02.161452 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5514c26f_0fe1_4e09_9292_896f21f90f71.slice/crio-6b016854d4d4884a31e2cc57ac01c8207df5d758726b4508fa4bc238478383d3 WatchSource:0}: Error finding container 6b016854d4d4884a31e2cc57ac01c8207df5d758726b4508fa4bc238478383d3: Status 404 returned error can't find the container with id 6b016854d4d4884a31e2cc57ac01c8207df5d758726b4508fa4bc238478383d3 Oct 05 07:07:02 crc kubenswrapper[4846]: I1005 07:07:02.168886 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8c207acc-07b9-4d9f-8b20-c84c2b67560f-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "8c207acc-07b9-4d9f-8b20-c84c2b67560f" (UID: "8c207acc-07b9-4d9f-8b20-c84c2b67560f"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 07:07:02 crc kubenswrapper[4846]: I1005 07:07:02.171077 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8c207acc-07b9-4d9f-8b20-c84c2b67560f-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "8c207acc-07b9-4d9f-8b20-c84c2b67560f" (UID: "8c207acc-07b9-4d9f-8b20-c84c2b67560f"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 07:07:02 crc kubenswrapper[4846]: I1005 07:07:02.191862 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8c207acc-07b9-4d9f-8b20-c84c2b67560f-config" (OuterVolumeSpecName: "config") pod "8c207acc-07b9-4d9f-8b20-c84c2b67560f" (UID: "8c207acc-07b9-4d9f-8b20-c84c2b67560f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 07:07:02 crc kubenswrapper[4846]: I1005 07:07:02.217078 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-dswh8" event={"ID":"7760f846-37e0-4bb8-bc2b-2e10d8950022","Type":"ContainerStarted","Data":"75309f967d618b39cc1b78b8d76b5a811e908820abd07bd78a313ec106dcce6c"} Oct 05 07:07:02 crc kubenswrapper[4846]: I1005 07:07:02.217439 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-dswh8" event={"ID":"7760f846-37e0-4bb8-bc2b-2e10d8950022","Type":"ContainerStarted","Data":"f73f987a9ae0e8f07d440b71e98c4f9bfcd094289b1f749e49f30c052a4b81d5"} Oct 05 07:07:02 crc kubenswrapper[4846]: I1005 07:07:02.221757 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-7f2fl" event={"ID":"a6ab76e7-ba49-4976-b718-3ea57bb4b53a","Type":"ContainerStarted","Data":"6bd20d0b024c20d2d5d45abd3d0560096db7cf87d891316de66e8ad3f5657424"} Oct 05 07:07:02 crc kubenswrapper[4846]: I1005 07:07:02.232087 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7cf74955b9-njz69" event={"ID":"5514c26f-0fe1-4e09-9292-896f21f90f71","Type":"ContainerStarted","Data":"6b016854d4d4884a31e2cc57ac01c8207df5d758726b4508fa4bc238478383d3"} Oct 05 07:07:02 crc kubenswrapper[4846]: I1005 07:07:02.245559 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-dswh8" podStartSLOduration=2.245532958 podStartE2EDuration="2.245532958s" podCreationTimestamp="2025-10-05 07:07:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 07:07:02.235549562 +0000 UTC m=+1144.476402337" watchObservedRunningTime="2025-10-05 07:07:02.245532958 +0000 UTC m=+1144.486385733" Oct 05 07:07:02 crc kubenswrapper[4846]: I1005 07:07:02.247313 4846 generic.go:334] "Generic (PLEG): container finished" podID="8c207acc-07b9-4d9f-8b20-c84c2b67560f" containerID="8d03c956f45603fb749baa477ec5fb7e31072932fb32c3bef02bba3de7b751e9" exitCode=0 Oct 05 07:07:02 crc kubenswrapper[4846]: I1005 07:07:02.247399 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6f9b65c9b5-lqj5b" event={"ID":"8c207acc-07b9-4d9f-8b20-c84c2b67560f","Type":"ContainerDied","Data":"8d03c956f45603fb749baa477ec5fb7e31072932fb32c3bef02bba3de7b751e9"} Oct 05 07:07:02 crc kubenswrapper[4846]: I1005 07:07:02.247434 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6f9b65c9b5-lqj5b" event={"ID":"8c207acc-07b9-4d9f-8b20-c84c2b67560f","Type":"ContainerDied","Data":"3c0746cb95a48aefd998651a9fafd0c4d79f70c5eab5dcffdb7c716fc312679e"} Oct 05 07:07:02 crc kubenswrapper[4846]: I1005 07:07:02.247436 4846 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8c207acc-07b9-4d9f-8b20-c84c2b67560f-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 05 07:07:02 crc kubenswrapper[4846]: I1005 07:07:02.247453 4846 scope.go:117] "RemoveContainer" containerID="8d03c956f45603fb749baa477ec5fb7e31072932fb32c3bef02bba3de7b751e9" Oct 05 07:07:02 crc kubenswrapper[4846]: I1005 07:07:02.247584 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6f9b65c9b5-lqj5b" Oct 05 07:07:02 crc kubenswrapper[4846]: I1005 07:07:02.248056 4846 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8c207acc-07b9-4d9f-8b20-c84c2b67560f-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 05 07:07:02 crc kubenswrapper[4846]: I1005 07:07:02.248163 4846 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8c207acc-07b9-4d9f-8b20-c84c2b67560f-config\") on node \"crc\" DevicePath \"\"" Oct 05 07:07:02 crc kubenswrapper[4846]: I1005 07:07:02.255579 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6d6f4f645f-97gpx" Oct 05 07:07:02 crc kubenswrapper[4846]: I1005 07:07:02.257213 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ae01bc6c-75fb-4958-9a75-6c212fd0ed3a","Type":"ContainerStarted","Data":"2545b9452eaf83986bb5d51763da560439fe6f42fda5b657635a72196e0d38e2"} Oct 05 07:07:02 crc kubenswrapper[4846]: I1005 07:07:02.266713 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8c207acc-07b9-4d9f-8b20-c84c2b67560f-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "8c207acc-07b9-4d9f-8b20-c84c2b67560f" (UID: "8c207acc-07b9-4d9f-8b20-c84c2b67560f"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 07:07:02 crc kubenswrapper[4846]: I1005 07:07:02.338370 4846 scope.go:117] "RemoveContainer" containerID="3014cd247fcb2ed23a0ded0ef8cba21e3b0661b9ee16f6a89e928f7c6d8a5c69" Oct 05 07:07:02 crc kubenswrapper[4846]: I1005 07:07:02.341650 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6d6f4f645f-97gpx"] Oct 05 07:07:02 crc kubenswrapper[4846]: I1005 07:07:02.356135 4846 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8c207acc-07b9-4d9f-8b20-c84c2b67560f-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 05 07:07:02 crc kubenswrapper[4846]: I1005 07:07:02.357080 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6d6f4f645f-97gpx"] Oct 05 07:07:02 crc kubenswrapper[4846]: I1005 07:07:02.411656 4846 scope.go:117] "RemoveContainer" containerID="8d03c956f45603fb749baa477ec5fb7e31072932fb32c3bef02bba3de7b751e9" Oct 05 07:07:02 crc kubenswrapper[4846]: E1005 07:07:02.415058 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8d03c956f45603fb749baa477ec5fb7e31072932fb32c3bef02bba3de7b751e9\": container with ID starting with 8d03c956f45603fb749baa477ec5fb7e31072932fb32c3bef02bba3de7b751e9 not found: ID does not exist" containerID="8d03c956f45603fb749baa477ec5fb7e31072932fb32c3bef02bba3de7b751e9" Oct 05 07:07:02 crc kubenswrapper[4846]: I1005 07:07:02.415117 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8d03c956f45603fb749baa477ec5fb7e31072932fb32c3bef02bba3de7b751e9"} err="failed to get container status \"8d03c956f45603fb749baa477ec5fb7e31072932fb32c3bef02bba3de7b751e9\": rpc error: code = NotFound desc = could not find container \"8d03c956f45603fb749baa477ec5fb7e31072932fb32c3bef02bba3de7b751e9\": container with ID starting with 8d03c956f45603fb749baa477ec5fb7e31072932fb32c3bef02bba3de7b751e9 not found: ID does not exist" Oct 05 07:07:02 crc kubenswrapper[4846]: I1005 07:07:02.415141 4846 scope.go:117] "RemoveContainer" containerID="3014cd247fcb2ed23a0ded0ef8cba21e3b0661b9ee16f6a89e928f7c6d8a5c69" Oct 05 07:07:02 crc kubenswrapper[4846]: E1005 07:07:02.415711 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3014cd247fcb2ed23a0ded0ef8cba21e3b0661b9ee16f6a89e928f7c6d8a5c69\": container with ID starting with 3014cd247fcb2ed23a0ded0ef8cba21e3b0661b9ee16f6a89e928f7c6d8a5c69 not found: ID does not exist" containerID="3014cd247fcb2ed23a0ded0ef8cba21e3b0661b9ee16f6a89e928f7c6d8a5c69" Oct 05 07:07:02 crc kubenswrapper[4846]: I1005 07:07:02.415737 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3014cd247fcb2ed23a0ded0ef8cba21e3b0661b9ee16f6a89e928f7c6d8a5c69"} err="failed to get container status \"3014cd247fcb2ed23a0ded0ef8cba21e3b0661b9ee16f6a89e928f7c6d8a5c69\": rpc error: code = NotFound desc = could not find container \"3014cd247fcb2ed23a0ded0ef8cba21e3b0661b9ee16f6a89e928f7c6d8a5c69\": container with ID starting with 3014cd247fcb2ed23a0ded0ef8cba21e3b0661b9ee16f6a89e928f7c6d8a5c69 not found: ID does not exist" Oct 05 07:07:02 crc kubenswrapper[4846]: I1005 07:07:02.417985 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-t5brq"] Oct 05 07:07:02 crc kubenswrapper[4846]: I1005 07:07:02.429732 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-cf6rt"] Oct 05 07:07:02 crc kubenswrapper[4846]: I1005 07:07:02.518866 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b5852c48-40ce-4e9f-8cab-d2822471fd26" path="/var/lib/kubelet/pods/b5852c48-40ce-4e9f-8cab-d2822471fd26/volumes" Oct 05 07:07:02 crc kubenswrapper[4846]: I1005 07:07:02.590270 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6f9b65c9b5-lqj5b"] Oct 05 07:07:02 crc kubenswrapper[4846]: I1005 07:07:02.613602 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-2l7b4"] Oct 05 07:07:02 crc kubenswrapper[4846]: I1005 07:07:02.630413 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6f9b65c9b5-lqj5b"] Oct 05 07:07:03 crc kubenswrapper[4846]: I1005 07:07:03.064224 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 05 07:07:03 crc kubenswrapper[4846]: I1005 07:07:03.266732 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-t5brq" event={"ID":"b7a58faa-0f37-4b0e-a907-6a8bae28cb27","Type":"ContainerStarted","Data":"67f82731bda75c617e2cda8b0f68588e249bade2906ab46833c9ec7f2f2ddd3a"} Oct 05 07:07:03 crc kubenswrapper[4846]: I1005 07:07:03.269485 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-2l7b4" event={"ID":"371a8404-6b77-418f-8fcb-3c72ed092816","Type":"ContainerStarted","Data":"b902b0193cb363be23ae30173cd0b9b076f21d8ff2cb0927214f0825a53830ed"} Oct 05 07:07:03 crc kubenswrapper[4846]: I1005 07:07:03.269548 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-2l7b4" event={"ID":"371a8404-6b77-418f-8fcb-3c72ed092816","Type":"ContainerStarted","Data":"77393015fdd7b90c5b4b6a40654021fba9dd9be930e8339d558a3e5ff8a1af16"} Oct 05 07:07:03 crc kubenswrapper[4846]: I1005 07:07:03.273273 4846 generic.go:334] "Generic (PLEG): container finished" podID="5514c26f-0fe1-4e09-9292-896f21f90f71" containerID="150067bb99bb3cea426705b9e67dd71568c5d12ab03723bd75aa455857688ad6" exitCode=0 Oct 05 07:07:03 crc kubenswrapper[4846]: I1005 07:07:03.273368 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7cf74955b9-njz69" event={"ID":"5514c26f-0fe1-4e09-9292-896f21f90f71","Type":"ContainerDied","Data":"150067bb99bb3cea426705b9e67dd71568c5d12ab03723bd75aa455857688ad6"} Oct 05 07:07:03 crc kubenswrapper[4846]: I1005 07:07:03.278135 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-cf6rt" event={"ID":"ee312852-b2eb-43d8-864c-bc553fed7ea5","Type":"ContainerStarted","Data":"79b45ea27ca7ea9da92bbd1c6a7b37e0e776b1d5f97e05d8996700255faac79a"} Oct 05 07:07:03 crc kubenswrapper[4846]: I1005 07:07:03.299555 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-sync-2l7b4" podStartSLOduration=2.299533896 podStartE2EDuration="2.299533896s" podCreationTimestamp="2025-10-05 07:07:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 07:07:03.295124119 +0000 UTC m=+1145.535976894" watchObservedRunningTime="2025-10-05 07:07:03.299533896 +0000 UTC m=+1145.540386671" Oct 05 07:07:04 crc kubenswrapper[4846]: I1005 07:07:04.350601 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7cf74955b9-njz69" event={"ID":"5514c26f-0fe1-4e09-9292-896f21f90f71","Type":"ContainerStarted","Data":"1d7f5cfaee91723a32aaf61fa9be5825f7c8e5a2573f16e15c12504bbef939f2"} Oct 05 07:07:04 crc kubenswrapper[4846]: I1005 07:07:04.353372 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7cf74955b9-njz69" Oct 05 07:07:04 crc kubenswrapper[4846]: I1005 07:07:04.374802 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7cf74955b9-njz69" podStartSLOduration=4.374782421 podStartE2EDuration="4.374782421s" podCreationTimestamp="2025-10-05 07:07:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 07:07:04.36951023 +0000 UTC m=+1146.610363005" watchObservedRunningTime="2025-10-05 07:07:04.374782421 +0000 UTC m=+1146.615635196" Oct 05 07:07:04 crc kubenswrapper[4846]: I1005 07:07:04.520403 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8c207acc-07b9-4d9f-8b20-c84c2b67560f" path="/var/lib/kubelet/pods/8c207acc-07b9-4d9f-8b20-c84c2b67560f/volumes" Oct 05 07:07:06 crc kubenswrapper[4846]: I1005 07:07:06.394606 4846 generic.go:334] "Generic (PLEG): container finished" podID="7760f846-37e0-4bb8-bc2b-2e10d8950022" containerID="75309f967d618b39cc1b78b8d76b5a811e908820abd07bd78a313ec106dcce6c" exitCode=0 Oct 05 07:07:06 crc kubenswrapper[4846]: I1005 07:07:06.394720 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-dswh8" event={"ID":"7760f846-37e0-4bb8-bc2b-2e10d8950022","Type":"ContainerDied","Data":"75309f967d618b39cc1b78b8d76b5a811e908820abd07bd78a313ec106dcce6c"} Oct 05 07:07:11 crc kubenswrapper[4846]: I1005 07:07:11.408614 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-7cf74955b9-njz69" Oct 05 07:07:11 crc kubenswrapper[4846]: I1005 07:07:11.551102 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7bc89dc697-k6cm6"] Oct 05 07:07:11 crc kubenswrapper[4846]: I1005 07:07:11.551491 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7bc89dc697-k6cm6" podUID="51a08294-d363-4169-8b63-c759c3cc30f0" containerName="dnsmasq-dns" containerID="cri-o://df24c9bed2030762b841c88e3a16d298b74f8634a1c567708c75f4b65a4066a5" gracePeriod=10 Oct 05 07:07:12 crc kubenswrapper[4846]: I1005 07:07:12.504061 4846 generic.go:334] "Generic (PLEG): container finished" podID="51a08294-d363-4169-8b63-c759c3cc30f0" containerID="df24c9bed2030762b841c88e3a16d298b74f8634a1c567708c75f4b65a4066a5" exitCode=0 Oct 05 07:07:12 crc kubenswrapper[4846]: I1005 07:07:12.511598 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7bc89dc697-k6cm6" event={"ID":"51a08294-d363-4169-8b63-c759c3cc30f0","Type":"ContainerDied","Data":"df24c9bed2030762b841c88e3a16d298b74f8634a1c567708c75f4b65a4066a5"} Oct 05 07:07:13 crc kubenswrapper[4846]: E1005 07:07:13.830738 4846 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-placement-api@sha256:d4062383c3b55d604b842067a258b63caf51dcea7d26447a53cd681105626951" Oct 05 07:07:13 crc kubenswrapper[4846]: E1005 07:07:13.830932 4846 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:placement-db-sync,Image:quay.io/podified-antelope-centos9/openstack-placement-api@sha256:d4062383c3b55d604b842067a258b63caf51dcea7d26447a53cd681105626951,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:true,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/placement,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:false,MountPath:/var/lib/openstack/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:placement-dbsync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-ntr8z,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42482,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-db-sync-7f2fl_openstack(a6ab76e7-ba49-4976-b718-3ea57bb4b53a): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 05 07:07:13 crc kubenswrapper[4846]: E1005 07:07:13.832760 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"placement-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/placement-db-sync-7f2fl" podUID="a6ab76e7-ba49-4976-b718-3ea57bb4b53a" Oct 05 07:07:13 crc kubenswrapper[4846]: I1005 07:07:13.938924 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-dswh8" Oct 05 07:07:14 crc kubenswrapper[4846]: I1005 07:07:14.076873 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2fhhn\" (UniqueName: \"kubernetes.io/projected/7760f846-37e0-4bb8-bc2b-2e10d8950022-kube-api-access-2fhhn\") pod \"7760f846-37e0-4bb8-bc2b-2e10d8950022\" (UID: \"7760f846-37e0-4bb8-bc2b-2e10d8950022\") " Oct 05 07:07:14 crc kubenswrapper[4846]: I1005 07:07:14.076934 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7760f846-37e0-4bb8-bc2b-2e10d8950022-config-data\") pod \"7760f846-37e0-4bb8-bc2b-2e10d8950022\" (UID: \"7760f846-37e0-4bb8-bc2b-2e10d8950022\") " Oct 05 07:07:14 crc kubenswrapper[4846]: I1005 07:07:14.076974 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7760f846-37e0-4bb8-bc2b-2e10d8950022-combined-ca-bundle\") pod \"7760f846-37e0-4bb8-bc2b-2e10d8950022\" (UID: \"7760f846-37e0-4bb8-bc2b-2e10d8950022\") " Oct 05 07:07:14 crc kubenswrapper[4846]: I1005 07:07:14.077042 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/7760f846-37e0-4bb8-bc2b-2e10d8950022-credential-keys\") pod \"7760f846-37e0-4bb8-bc2b-2e10d8950022\" (UID: \"7760f846-37e0-4bb8-bc2b-2e10d8950022\") " Oct 05 07:07:14 crc kubenswrapper[4846]: I1005 07:07:14.077101 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/7760f846-37e0-4bb8-bc2b-2e10d8950022-fernet-keys\") pod \"7760f846-37e0-4bb8-bc2b-2e10d8950022\" (UID: \"7760f846-37e0-4bb8-bc2b-2e10d8950022\") " Oct 05 07:07:14 crc kubenswrapper[4846]: I1005 07:07:14.077261 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7760f846-37e0-4bb8-bc2b-2e10d8950022-scripts\") pod \"7760f846-37e0-4bb8-bc2b-2e10d8950022\" (UID: \"7760f846-37e0-4bb8-bc2b-2e10d8950022\") " Oct 05 07:07:14 crc kubenswrapper[4846]: I1005 07:07:14.090970 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7760f846-37e0-4bb8-bc2b-2e10d8950022-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "7760f846-37e0-4bb8-bc2b-2e10d8950022" (UID: "7760f846-37e0-4bb8-bc2b-2e10d8950022"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:07:14 crc kubenswrapper[4846]: I1005 07:07:14.092045 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7760f846-37e0-4bb8-bc2b-2e10d8950022-scripts" (OuterVolumeSpecName: "scripts") pod "7760f846-37e0-4bb8-bc2b-2e10d8950022" (UID: "7760f846-37e0-4bb8-bc2b-2e10d8950022"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:07:14 crc kubenswrapper[4846]: I1005 07:07:14.092212 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7760f846-37e0-4bb8-bc2b-2e10d8950022-kube-api-access-2fhhn" (OuterVolumeSpecName: "kube-api-access-2fhhn") pod "7760f846-37e0-4bb8-bc2b-2e10d8950022" (UID: "7760f846-37e0-4bb8-bc2b-2e10d8950022"). InnerVolumeSpecName "kube-api-access-2fhhn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:07:14 crc kubenswrapper[4846]: I1005 07:07:14.094191 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7760f846-37e0-4bb8-bc2b-2e10d8950022-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "7760f846-37e0-4bb8-bc2b-2e10d8950022" (UID: "7760f846-37e0-4bb8-bc2b-2e10d8950022"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:07:14 crc kubenswrapper[4846]: I1005 07:07:14.123695 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7760f846-37e0-4bb8-bc2b-2e10d8950022-config-data" (OuterVolumeSpecName: "config-data") pod "7760f846-37e0-4bb8-bc2b-2e10d8950022" (UID: "7760f846-37e0-4bb8-bc2b-2e10d8950022"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:07:14 crc kubenswrapper[4846]: I1005 07:07:14.124236 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7760f846-37e0-4bb8-bc2b-2e10d8950022-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7760f846-37e0-4bb8-bc2b-2e10d8950022" (UID: "7760f846-37e0-4bb8-bc2b-2e10d8950022"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:07:14 crc kubenswrapper[4846]: I1005 07:07:14.180828 4846 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7760f846-37e0-4bb8-bc2b-2e10d8950022-scripts\") on node \"crc\" DevicePath \"\"" Oct 05 07:07:14 crc kubenswrapper[4846]: I1005 07:07:14.180861 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2fhhn\" (UniqueName: \"kubernetes.io/projected/7760f846-37e0-4bb8-bc2b-2e10d8950022-kube-api-access-2fhhn\") on node \"crc\" DevicePath \"\"" Oct 05 07:07:14 crc kubenswrapper[4846]: I1005 07:07:14.180872 4846 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7760f846-37e0-4bb8-bc2b-2e10d8950022-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 07:07:14 crc kubenswrapper[4846]: I1005 07:07:14.180882 4846 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7760f846-37e0-4bb8-bc2b-2e10d8950022-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 07:07:14 crc kubenswrapper[4846]: I1005 07:07:14.180891 4846 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/7760f846-37e0-4bb8-bc2b-2e10d8950022-credential-keys\") on node \"crc\" DevicePath \"\"" Oct 05 07:07:14 crc kubenswrapper[4846]: I1005 07:07:14.180900 4846 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/7760f846-37e0-4bb8-bc2b-2e10d8950022-fernet-keys\") on node \"crc\" DevicePath \"\"" Oct 05 07:07:14 crc kubenswrapper[4846]: I1005 07:07:14.527629 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-dswh8" event={"ID":"7760f846-37e0-4bb8-bc2b-2e10d8950022","Type":"ContainerDied","Data":"f73f987a9ae0e8f07d440b71e98c4f9bfcd094289b1f749e49f30c052a4b81d5"} Oct 05 07:07:14 crc kubenswrapper[4846]: I1005 07:07:14.527697 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f73f987a9ae0e8f07d440b71e98c4f9bfcd094289b1f749e49f30c052a4b81d5" Oct 05 07:07:14 crc kubenswrapper[4846]: I1005 07:07:14.527664 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-dswh8" Oct 05 07:07:14 crc kubenswrapper[4846]: E1005 07:07:14.533157 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"placement-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-placement-api@sha256:d4062383c3b55d604b842067a258b63caf51dcea7d26447a53cd681105626951\\\"\"" pod="openstack/placement-db-sync-7f2fl" podUID="a6ab76e7-ba49-4976-b718-3ea57bb4b53a" Oct 05 07:07:15 crc kubenswrapper[4846]: I1005 07:07:15.093083 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-dswh8"] Oct 05 07:07:15 crc kubenswrapper[4846]: I1005 07:07:15.104913 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-dswh8"] Oct 05 07:07:15 crc kubenswrapper[4846]: I1005 07:07:15.209133 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-vt9sx"] Oct 05 07:07:15 crc kubenswrapper[4846]: E1005 07:07:15.209572 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8c207acc-07b9-4d9f-8b20-c84c2b67560f" containerName="init" Oct 05 07:07:15 crc kubenswrapper[4846]: I1005 07:07:15.209593 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="8c207acc-07b9-4d9f-8b20-c84c2b67560f" containerName="init" Oct 05 07:07:15 crc kubenswrapper[4846]: E1005 07:07:15.209607 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7760f846-37e0-4bb8-bc2b-2e10d8950022" containerName="keystone-bootstrap" Oct 05 07:07:15 crc kubenswrapper[4846]: I1005 07:07:15.209614 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="7760f846-37e0-4bb8-bc2b-2e10d8950022" containerName="keystone-bootstrap" Oct 05 07:07:15 crc kubenswrapper[4846]: E1005 07:07:15.209633 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8c207acc-07b9-4d9f-8b20-c84c2b67560f" containerName="dnsmasq-dns" Oct 05 07:07:15 crc kubenswrapper[4846]: I1005 07:07:15.209640 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="8c207acc-07b9-4d9f-8b20-c84c2b67560f" containerName="dnsmasq-dns" Oct 05 07:07:15 crc kubenswrapper[4846]: I1005 07:07:15.209839 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="8c207acc-07b9-4d9f-8b20-c84c2b67560f" containerName="dnsmasq-dns" Oct 05 07:07:15 crc kubenswrapper[4846]: I1005 07:07:15.209850 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="7760f846-37e0-4bb8-bc2b-2e10d8950022" containerName="keystone-bootstrap" Oct 05 07:07:15 crc kubenswrapper[4846]: I1005 07:07:15.210486 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-vt9sx" Oct 05 07:07:15 crc kubenswrapper[4846]: I1005 07:07:15.217021 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-76d8v" Oct 05 07:07:15 crc kubenswrapper[4846]: I1005 07:07:15.217210 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 05 07:07:15 crc kubenswrapper[4846]: I1005 07:07:15.217320 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 05 07:07:15 crc kubenswrapper[4846]: I1005 07:07:15.217433 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 05 07:07:15 crc kubenswrapper[4846]: I1005 07:07:15.246874 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-vt9sx"] Oct 05 07:07:15 crc kubenswrapper[4846]: I1005 07:07:15.412783 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/979a46b2-96f3-498a-8838-4b73b4c8dc91-credential-keys\") pod \"keystone-bootstrap-vt9sx\" (UID: \"979a46b2-96f3-498a-8838-4b73b4c8dc91\") " pod="openstack/keystone-bootstrap-vt9sx" Oct 05 07:07:15 crc kubenswrapper[4846]: I1005 07:07:15.412913 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/979a46b2-96f3-498a-8838-4b73b4c8dc91-fernet-keys\") pod \"keystone-bootstrap-vt9sx\" (UID: \"979a46b2-96f3-498a-8838-4b73b4c8dc91\") " pod="openstack/keystone-bootstrap-vt9sx" Oct 05 07:07:15 crc kubenswrapper[4846]: I1005 07:07:15.412951 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/979a46b2-96f3-498a-8838-4b73b4c8dc91-config-data\") pod \"keystone-bootstrap-vt9sx\" (UID: \"979a46b2-96f3-498a-8838-4b73b4c8dc91\") " pod="openstack/keystone-bootstrap-vt9sx" Oct 05 07:07:15 crc kubenswrapper[4846]: I1005 07:07:15.412991 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/979a46b2-96f3-498a-8838-4b73b4c8dc91-combined-ca-bundle\") pod \"keystone-bootstrap-vt9sx\" (UID: \"979a46b2-96f3-498a-8838-4b73b4c8dc91\") " pod="openstack/keystone-bootstrap-vt9sx" Oct 05 07:07:15 crc kubenswrapper[4846]: I1005 07:07:15.413064 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/979a46b2-96f3-498a-8838-4b73b4c8dc91-scripts\") pod \"keystone-bootstrap-vt9sx\" (UID: \"979a46b2-96f3-498a-8838-4b73b4c8dc91\") " pod="openstack/keystone-bootstrap-vt9sx" Oct 05 07:07:15 crc kubenswrapper[4846]: I1005 07:07:15.413114 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fp47l\" (UniqueName: \"kubernetes.io/projected/979a46b2-96f3-498a-8838-4b73b4c8dc91-kube-api-access-fp47l\") pod \"keystone-bootstrap-vt9sx\" (UID: \"979a46b2-96f3-498a-8838-4b73b4c8dc91\") " pod="openstack/keystone-bootstrap-vt9sx" Oct 05 07:07:15 crc kubenswrapper[4846]: I1005 07:07:15.514326 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/979a46b2-96f3-498a-8838-4b73b4c8dc91-scripts\") pod \"keystone-bootstrap-vt9sx\" (UID: \"979a46b2-96f3-498a-8838-4b73b4c8dc91\") " pod="openstack/keystone-bootstrap-vt9sx" Oct 05 07:07:15 crc kubenswrapper[4846]: I1005 07:07:15.514395 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fp47l\" (UniqueName: \"kubernetes.io/projected/979a46b2-96f3-498a-8838-4b73b4c8dc91-kube-api-access-fp47l\") pod \"keystone-bootstrap-vt9sx\" (UID: \"979a46b2-96f3-498a-8838-4b73b4c8dc91\") " pod="openstack/keystone-bootstrap-vt9sx" Oct 05 07:07:15 crc kubenswrapper[4846]: I1005 07:07:15.514428 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/979a46b2-96f3-498a-8838-4b73b4c8dc91-credential-keys\") pod \"keystone-bootstrap-vt9sx\" (UID: \"979a46b2-96f3-498a-8838-4b73b4c8dc91\") " pod="openstack/keystone-bootstrap-vt9sx" Oct 05 07:07:15 crc kubenswrapper[4846]: I1005 07:07:15.514477 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/979a46b2-96f3-498a-8838-4b73b4c8dc91-fernet-keys\") pod \"keystone-bootstrap-vt9sx\" (UID: \"979a46b2-96f3-498a-8838-4b73b4c8dc91\") " pod="openstack/keystone-bootstrap-vt9sx" Oct 05 07:07:15 crc kubenswrapper[4846]: I1005 07:07:15.514496 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/979a46b2-96f3-498a-8838-4b73b4c8dc91-config-data\") pod \"keystone-bootstrap-vt9sx\" (UID: \"979a46b2-96f3-498a-8838-4b73b4c8dc91\") " pod="openstack/keystone-bootstrap-vt9sx" Oct 05 07:07:15 crc kubenswrapper[4846]: I1005 07:07:15.514524 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/979a46b2-96f3-498a-8838-4b73b4c8dc91-combined-ca-bundle\") pod \"keystone-bootstrap-vt9sx\" (UID: \"979a46b2-96f3-498a-8838-4b73b4c8dc91\") " pod="openstack/keystone-bootstrap-vt9sx" Oct 05 07:07:15 crc kubenswrapper[4846]: I1005 07:07:15.522107 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/979a46b2-96f3-498a-8838-4b73b4c8dc91-combined-ca-bundle\") pod \"keystone-bootstrap-vt9sx\" (UID: \"979a46b2-96f3-498a-8838-4b73b4c8dc91\") " pod="openstack/keystone-bootstrap-vt9sx" Oct 05 07:07:15 crc kubenswrapper[4846]: I1005 07:07:15.522647 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/979a46b2-96f3-498a-8838-4b73b4c8dc91-credential-keys\") pod \"keystone-bootstrap-vt9sx\" (UID: \"979a46b2-96f3-498a-8838-4b73b4c8dc91\") " pod="openstack/keystone-bootstrap-vt9sx" Oct 05 07:07:15 crc kubenswrapper[4846]: I1005 07:07:15.523852 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/979a46b2-96f3-498a-8838-4b73b4c8dc91-scripts\") pod \"keystone-bootstrap-vt9sx\" (UID: \"979a46b2-96f3-498a-8838-4b73b4c8dc91\") " pod="openstack/keystone-bootstrap-vt9sx" Oct 05 07:07:15 crc kubenswrapper[4846]: I1005 07:07:15.539107 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fp47l\" (UniqueName: \"kubernetes.io/projected/979a46b2-96f3-498a-8838-4b73b4c8dc91-kube-api-access-fp47l\") pod \"keystone-bootstrap-vt9sx\" (UID: \"979a46b2-96f3-498a-8838-4b73b4c8dc91\") " pod="openstack/keystone-bootstrap-vt9sx" Oct 05 07:07:15 crc kubenswrapper[4846]: I1005 07:07:15.543621 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/979a46b2-96f3-498a-8838-4b73b4c8dc91-fernet-keys\") pod \"keystone-bootstrap-vt9sx\" (UID: \"979a46b2-96f3-498a-8838-4b73b4c8dc91\") " pod="openstack/keystone-bootstrap-vt9sx" Oct 05 07:07:15 crc kubenswrapper[4846]: I1005 07:07:15.565758 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/979a46b2-96f3-498a-8838-4b73b4c8dc91-config-data\") pod \"keystone-bootstrap-vt9sx\" (UID: \"979a46b2-96f3-498a-8838-4b73b4c8dc91\") " pod="openstack/keystone-bootstrap-vt9sx" Oct 05 07:07:15 crc kubenswrapper[4846]: I1005 07:07:15.840642 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-vt9sx" Oct 05 07:07:16 crc kubenswrapper[4846]: I1005 07:07:16.509827 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7760f846-37e0-4bb8-bc2b-2e10d8950022" path="/var/lib/kubelet/pods/7760f846-37e0-4bb8-bc2b-2e10d8950022/volumes" Oct 05 07:07:20 crc kubenswrapper[4846]: I1005 07:07:20.771446 4846 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-7bc89dc697-k6cm6" podUID="51a08294-d363-4169-8b63-c759c3cc30f0" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.117:5353: i/o timeout" Oct 05 07:07:22 crc kubenswrapper[4846]: E1005 07:07:22.303271 4846 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-barbican-api@sha256:7db347424a8c5998059c5bf84c86a1ef8d582d1ffe39f4887551f2ac85a4915f" Oct 05 07:07:22 crc kubenswrapper[4846]: E1005 07:07:22.303478 4846 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:barbican-db-sync,Image:quay.io/podified-antelope-centos9/openstack-barbican-api@sha256:7db347424a8c5998059c5bf84c86a1ef8d582d1ffe39f4887551f2ac85a4915f,Command:[/bin/bash],Args:[-c barbican-manage db upgrade],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/barbican/barbican.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-cft2p,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42403,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42403,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod barbican-db-sync-t5brq_openstack(b7a58faa-0f37-4b0e-a907-6a8bae28cb27): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 05 07:07:22 crc kubenswrapper[4846]: E1005 07:07:22.305354 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/barbican-db-sync-t5brq" podUID="b7a58faa-0f37-4b0e-a907-6a8bae28cb27" Oct 05 07:07:22 crc kubenswrapper[4846]: I1005 07:07:22.384485 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7bc89dc697-k6cm6" Oct 05 07:07:22 crc kubenswrapper[4846]: I1005 07:07:22.489305 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/51a08294-d363-4169-8b63-c759c3cc30f0-ovsdbserver-sb\") pod \"51a08294-d363-4169-8b63-c759c3cc30f0\" (UID: \"51a08294-d363-4169-8b63-c759c3cc30f0\") " Oct 05 07:07:22 crc kubenswrapper[4846]: I1005 07:07:22.489441 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/51a08294-d363-4169-8b63-c759c3cc30f0-ovsdbserver-nb\") pod \"51a08294-d363-4169-8b63-c759c3cc30f0\" (UID: \"51a08294-d363-4169-8b63-c759c3cc30f0\") " Oct 05 07:07:22 crc kubenswrapper[4846]: I1005 07:07:22.489481 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gskj2\" (UniqueName: \"kubernetes.io/projected/51a08294-d363-4169-8b63-c759c3cc30f0-kube-api-access-gskj2\") pod \"51a08294-d363-4169-8b63-c759c3cc30f0\" (UID: \"51a08294-d363-4169-8b63-c759c3cc30f0\") " Oct 05 07:07:22 crc kubenswrapper[4846]: I1005 07:07:22.489621 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/51a08294-d363-4169-8b63-c759c3cc30f0-config\") pod \"51a08294-d363-4169-8b63-c759c3cc30f0\" (UID: \"51a08294-d363-4169-8b63-c759c3cc30f0\") " Oct 05 07:07:22 crc kubenswrapper[4846]: I1005 07:07:22.489640 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/51a08294-d363-4169-8b63-c759c3cc30f0-dns-svc\") pod \"51a08294-d363-4169-8b63-c759c3cc30f0\" (UID: \"51a08294-d363-4169-8b63-c759c3cc30f0\") " Oct 05 07:07:22 crc kubenswrapper[4846]: I1005 07:07:22.500703 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/51a08294-d363-4169-8b63-c759c3cc30f0-kube-api-access-gskj2" (OuterVolumeSpecName: "kube-api-access-gskj2") pod "51a08294-d363-4169-8b63-c759c3cc30f0" (UID: "51a08294-d363-4169-8b63-c759c3cc30f0"). InnerVolumeSpecName "kube-api-access-gskj2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:07:22 crc kubenswrapper[4846]: I1005 07:07:22.545052 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/51a08294-d363-4169-8b63-c759c3cc30f0-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "51a08294-d363-4169-8b63-c759c3cc30f0" (UID: "51a08294-d363-4169-8b63-c759c3cc30f0"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 07:07:22 crc kubenswrapper[4846]: I1005 07:07:22.547430 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/51a08294-d363-4169-8b63-c759c3cc30f0-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "51a08294-d363-4169-8b63-c759c3cc30f0" (UID: "51a08294-d363-4169-8b63-c759c3cc30f0"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 07:07:22 crc kubenswrapper[4846]: I1005 07:07:22.549387 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/51a08294-d363-4169-8b63-c759c3cc30f0-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "51a08294-d363-4169-8b63-c759c3cc30f0" (UID: "51a08294-d363-4169-8b63-c759c3cc30f0"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 07:07:22 crc kubenswrapper[4846]: I1005 07:07:22.568714 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/51a08294-d363-4169-8b63-c759c3cc30f0-config" (OuterVolumeSpecName: "config") pod "51a08294-d363-4169-8b63-c759c3cc30f0" (UID: "51a08294-d363-4169-8b63-c759c3cc30f0"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 07:07:22 crc kubenswrapper[4846]: I1005 07:07:22.592717 4846 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/51a08294-d363-4169-8b63-c759c3cc30f0-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 05 07:07:22 crc kubenswrapper[4846]: I1005 07:07:22.592762 4846 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/51a08294-d363-4169-8b63-c759c3cc30f0-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 05 07:07:22 crc kubenswrapper[4846]: I1005 07:07:22.593336 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gskj2\" (UniqueName: \"kubernetes.io/projected/51a08294-d363-4169-8b63-c759c3cc30f0-kube-api-access-gskj2\") on node \"crc\" DevicePath \"\"" Oct 05 07:07:22 crc kubenswrapper[4846]: I1005 07:07:22.593454 4846 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/51a08294-d363-4169-8b63-c759c3cc30f0-config\") on node \"crc\" DevicePath \"\"" Oct 05 07:07:22 crc kubenswrapper[4846]: I1005 07:07:22.593558 4846 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/51a08294-d363-4169-8b63-c759c3cc30f0-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 05 07:07:22 crc kubenswrapper[4846]: I1005 07:07:22.626066 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7bc89dc697-k6cm6" Oct 05 07:07:22 crc kubenswrapper[4846]: I1005 07:07:22.626103 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7bc89dc697-k6cm6" event={"ID":"51a08294-d363-4169-8b63-c759c3cc30f0","Type":"ContainerDied","Data":"2183eeb2dac0ba1ef5254851e14b8cfa434cef36ba86d18d83b2c7647451d4db"} Oct 05 07:07:22 crc kubenswrapper[4846]: I1005 07:07:22.626167 4846 scope.go:117] "RemoveContainer" containerID="df24c9bed2030762b841c88e3a16d298b74f8634a1c567708c75f4b65a4066a5" Oct 05 07:07:22 crc kubenswrapper[4846]: E1005 07:07:22.636596 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-barbican-api@sha256:7db347424a8c5998059c5bf84c86a1ef8d582d1ffe39f4887551f2ac85a4915f\\\"\"" pod="openstack/barbican-db-sync-t5brq" podUID="b7a58faa-0f37-4b0e-a907-6a8bae28cb27" Oct 05 07:07:22 crc kubenswrapper[4846]: I1005 07:07:22.676588 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7bc89dc697-k6cm6"] Oct 05 07:07:22 crc kubenswrapper[4846]: I1005 07:07:22.683909 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7bc89dc697-k6cm6"] Oct 05 07:07:23 crc kubenswrapper[4846]: I1005 07:07:23.325129 4846 patch_prober.go:28] interesting pod/machine-config-daemon-fscvf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 07:07:23 crc kubenswrapper[4846]: I1005 07:07:23.325701 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 07:07:23 crc kubenswrapper[4846]: I1005 07:07:23.325769 4846 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" Oct 05 07:07:23 crc kubenswrapper[4846]: I1005 07:07:23.326697 4846 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"a25250ef8861b2bca088edb9f5248f80d97de22c7bacf6dfaaf8d42d30c96858"} pod="openshift-machine-config-operator/machine-config-daemon-fscvf" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 05 07:07:23 crc kubenswrapper[4846]: I1005 07:07:23.326779 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" containerName="machine-config-daemon" containerID="cri-o://a25250ef8861b2bca088edb9f5248f80d97de22c7bacf6dfaaf8d42d30c96858" gracePeriod=600 Oct 05 07:07:23 crc kubenswrapper[4846]: E1005 07:07:23.552982 4846 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-cinder-api@sha256:b0c0824763cbfd23b836ee4355015d9f94daa115bcc9ef0ea8b8e8980d5a6213" Oct 05 07:07:23 crc kubenswrapper[4846]: E1005 07:07:23.553230 4846 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:cinder-db-sync,Image:quay.io/podified-antelope-centos9/openstack-cinder-api@sha256:b0c0824763cbfd23b836ee4355015d9f94daa115bcc9ef0ea8b8e8980d5a6213,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_set_configs && /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:etc-machine-id,ReadOnly:true,MountPath:/etc/machine-id,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/config-data/merged,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/cinder/cinder.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-j2hlh,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-db-sync-cf6rt_openstack(ee312852-b2eb-43d8-864c-bc553fed7ea5): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 05 07:07:23 crc kubenswrapper[4846]: E1005 07:07:23.554349 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/cinder-db-sync-cf6rt" podUID="ee312852-b2eb-43d8-864c-bc553fed7ea5" Oct 05 07:07:23 crc kubenswrapper[4846]: I1005 07:07:23.569732 4846 scope.go:117] "RemoveContainer" containerID="95e9c9ce0e28d42ee5f003828a17dd4fa4f47d75bece1d967a29705097f3e344" Oct 05 07:07:23 crc kubenswrapper[4846]: I1005 07:07:23.671908 4846 generic.go:334] "Generic (PLEG): container finished" podID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" containerID="a25250ef8861b2bca088edb9f5248f80d97de22c7bacf6dfaaf8d42d30c96858" exitCode=0 Oct 05 07:07:23 crc kubenswrapper[4846]: I1005 07:07:23.672378 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" event={"ID":"ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7","Type":"ContainerDied","Data":"a25250ef8861b2bca088edb9f5248f80d97de22c7bacf6dfaaf8d42d30c96858"} Oct 05 07:07:23 crc kubenswrapper[4846]: E1005 07:07:23.711249 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-cinder-api@sha256:b0c0824763cbfd23b836ee4355015d9f94daa115bcc9ef0ea8b8e8980d5a6213\\\"\"" pod="openstack/cinder-db-sync-cf6rt" podUID="ee312852-b2eb-43d8-864c-bc553fed7ea5" Oct 05 07:07:23 crc kubenswrapper[4846]: I1005 07:07:23.736246 4846 scope.go:117] "RemoveContainer" containerID="ec13db87cb5ffec077d726e058e8274c67177ec66498d56f437d83d9bfbe9292" Oct 05 07:07:24 crc kubenswrapper[4846]: I1005 07:07:24.031305 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-vt9sx"] Oct 05 07:07:24 crc kubenswrapper[4846]: I1005 07:07:24.509422 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="51a08294-d363-4169-8b63-c759c3cc30f0" path="/var/lib/kubelet/pods/51a08294-d363-4169-8b63-c759c3cc30f0/volumes" Oct 05 07:07:24 crc kubenswrapper[4846]: I1005 07:07:24.690019 4846 generic.go:334] "Generic (PLEG): container finished" podID="371a8404-6b77-418f-8fcb-3c72ed092816" containerID="b902b0193cb363be23ae30173cd0b9b076f21d8ff2cb0927214f0825a53830ed" exitCode=0 Oct 05 07:07:24 crc kubenswrapper[4846]: I1005 07:07:24.690116 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-2l7b4" event={"ID":"371a8404-6b77-418f-8fcb-3c72ed092816","Type":"ContainerDied","Data":"b902b0193cb363be23ae30173cd0b9b076f21d8ff2cb0927214f0825a53830ed"} Oct 05 07:07:24 crc kubenswrapper[4846]: I1005 07:07:24.694116 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-vt9sx" event={"ID":"979a46b2-96f3-498a-8838-4b73b4c8dc91","Type":"ContainerStarted","Data":"b9645c4a90f8320793bcfa361c3d8c571e3730e1f40c3ec1b6cd107151e4d1c2"} Oct 05 07:07:24 crc kubenswrapper[4846]: I1005 07:07:24.694170 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-vt9sx" event={"ID":"979a46b2-96f3-498a-8838-4b73b4c8dc91","Type":"ContainerStarted","Data":"9a1146f8c3547b62d691c15b3f03fb61910c1e307ff455b408673ebd5dcb3176"} Oct 05 07:07:24 crc kubenswrapper[4846]: I1005 07:07:24.696977 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ae01bc6c-75fb-4958-9a75-6c212fd0ed3a","Type":"ContainerStarted","Data":"bedf3eddca9636e7febabe72b13554c3e7139ba3edbeca3aa33a3865508515eb"} Oct 05 07:07:24 crc kubenswrapper[4846]: I1005 07:07:24.699701 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" event={"ID":"ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7","Type":"ContainerStarted","Data":"11b9703888d7e63159496cf8c65783a81599439b16d1a1e5677afe95613a82de"} Oct 05 07:07:24 crc kubenswrapper[4846]: I1005 07:07:24.703611 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-68h9n" event={"ID":"8b6863be-f8e1-4a87-93d6-5cf756983747","Type":"ContainerStarted","Data":"6f45c438b8b209cc382e7a14455205149b1abf123df3b2467fe12ebe7d71967f"} Oct 05 07:07:24 crc kubenswrapper[4846]: I1005 07:07:24.734320 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-vt9sx" podStartSLOduration=9.734298718 podStartE2EDuration="9.734298718s" podCreationTimestamp="2025-10-05 07:07:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 07:07:24.729596673 +0000 UTC m=+1166.970449448" watchObservedRunningTime="2025-10-05 07:07:24.734298718 +0000 UTC m=+1166.975151493" Oct 05 07:07:24 crc kubenswrapper[4846]: I1005 07:07:24.757297 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-sync-68h9n" podStartSLOduration=2.71973477 podStartE2EDuration="51.757274931s" podCreationTimestamp="2025-10-05 07:06:33 +0000 UTC" firstStartedPulling="2025-10-05 07:06:34.567091754 +0000 UTC m=+1116.807944529" lastFinishedPulling="2025-10-05 07:07:23.604631915 +0000 UTC m=+1165.845484690" observedRunningTime="2025-10-05 07:07:24.752774331 +0000 UTC m=+1166.993627106" watchObservedRunningTime="2025-10-05 07:07:24.757274931 +0000 UTC m=+1166.998127706" Oct 05 07:07:25 crc kubenswrapper[4846]: I1005 07:07:25.715803 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ae01bc6c-75fb-4958-9a75-6c212fd0ed3a","Type":"ContainerStarted","Data":"f80062678cc982eb6612835cc64c1c9757ef8fbe87a8c39fe4a660630c89a786"} Oct 05 07:07:25 crc kubenswrapper[4846]: I1005 07:07:25.774661 4846 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-7bc89dc697-k6cm6" podUID="51a08294-d363-4169-8b63-c759c3cc30f0" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.117:5353: i/o timeout" Oct 05 07:07:26 crc kubenswrapper[4846]: I1005 07:07:26.094130 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-2l7b4" Oct 05 07:07:26 crc kubenswrapper[4846]: I1005 07:07:26.191430 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/371a8404-6b77-418f-8fcb-3c72ed092816-config\") pod \"371a8404-6b77-418f-8fcb-3c72ed092816\" (UID: \"371a8404-6b77-418f-8fcb-3c72ed092816\") " Oct 05 07:07:26 crc kubenswrapper[4846]: I1005 07:07:26.191593 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9pm8d\" (UniqueName: \"kubernetes.io/projected/371a8404-6b77-418f-8fcb-3c72ed092816-kube-api-access-9pm8d\") pod \"371a8404-6b77-418f-8fcb-3c72ed092816\" (UID: \"371a8404-6b77-418f-8fcb-3c72ed092816\") " Oct 05 07:07:26 crc kubenswrapper[4846]: I1005 07:07:26.191709 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/371a8404-6b77-418f-8fcb-3c72ed092816-combined-ca-bundle\") pod \"371a8404-6b77-418f-8fcb-3c72ed092816\" (UID: \"371a8404-6b77-418f-8fcb-3c72ed092816\") " Oct 05 07:07:26 crc kubenswrapper[4846]: I1005 07:07:26.204056 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/371a8404-6b77-418f-8fcb-3c72ed092816-kube-api-access-9pm8d" (OuterVolumeSpecName: "kube-api-access-9pm8d") pod "371a8404-6b77-418f-8fcb-3c72ed092816" (UID: "371a8404-6b77-418f-8fcb-3c72ed092816"). InnerVolumeSpecName "kube-api-access-9pm8d". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:07:26 crc kubenswrapper[4846]: I1005 07:07:26.230271 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/371a8404-6b77-418f-8fcb-3c72ed092816-config" (OuterVolumeSpecName: "config") pod "371a8404-6b77-418f-8fcb-3c72ed092816" (UID: "371a8404-6b77-418f-8fcb-3c72ed092816"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:07:26 crc kubenswrapper[4846]: I1005 07:07:26.237692 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/371a8404-6b77-418f-8fcb-3c72ed092816-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "371a8404-6b77-418f-8fcb-3c72ed092816" (UID: "371a8404-6b77-418f-8fcb-3c72ed092816"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:07:26 crc kubenswrapper[4846]: I1005 07:07:26.293901 4846 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/371a8404-6b77-418f-8fcb-3c72ed092816-config\") on node \"crc\" DevicePath \"\"" Oct 05 07:07:26 crc kubenswrapper[4846]: I1005 07:07:26.293946 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9pm8d\" (UniqueName: \"kubernetes.io/projected/371a8404-6b77-418f-8fcb-3c72ed092816-kube-api-access-9pm8d\") on node \"crc\" DevicePath \"\"" Oct 05 07:07:26 crc kubenswrapper[4846]: I1005 07:07:26.293961 4846 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/371a8404-6b77-418f-8fcb-3c72ed092816-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 07:07:26 crc kubenswrapper[4846]: I1005 07:07:26.755828 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-2l7b4" event={"ID":"371a8404-6b77-418f-8fcb-3c72ed092816","Type":"ContainerDied","Data":"77393015fdd7b90c5b4b6a40654021fba9dd9be930e8339d558a3e5ff8a1af16"} Oct 05 07:07:26 crc kubenswrapper[4846]: I1005 07:07:26.756213 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-2l7b4" Oct 05 07:07:26 crc kubenswrapper[4846]: I1005 07:07:26.756231 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="77393015fdd7b90c5b4b6a40654021fba9dd9be930e8339d558a3e5ff8a1af16" Oct 05 07:07:26 crc kubenswrapper[4846]: I1005 07:07:26.869357 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-686765ddbc-cnfdk"] Oct 05 07:07:26 crc kubenswrapper[4846]: E1005 07:07:26.869838 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="51a08294-d363-4169-8b63-c759c3cc30f0" containerName="dnsmasq-dns" Oct 05 07:07:26 crc kubenswrapper[4846]: I1005 07:07:26.869867 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="51a08294-d363-4169-8b63-c759c3cc30f0" containerName="dnsmasq-dns" Oct 05 07:07:26 crc kubenswrapper[4846]: E1005 07:07:26.869913 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="51a08294-d363-4169-8b63-c759c3cc30f0" containerName="init" Oct 05 07:07:26 crc kubenswrapper[4846]: I1005 07:07:26.869923 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="51a08294-d363-4169-8b63-c759c3cc30f0" containerName="init" Oct 05 07:07:26 crc kubenswrapper[4846]: E1005 07:07:26.869947 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="371a8404-6b77-418f-8fcb-3c72ed092816" containerName="neutron-db-sync" Oct 05 07:07:26 crc kubenswrapper[4846]: I1005 07:07:26.869955 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="371a8404-6b77-418f-8fcb-3c72ed092816" containerName="neutron-db-sync" Oct 05 07:07:26 crc kubenswrapper[4846]: I1005 07:07:26.870127 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="51a08294-d363-4169-8b63-c759c3cc30f0" containerName="dnsmasq-dns" Oct 05 07:07:26 crc kubenswrapper[4846]: I1005 07:07:26.870148 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="371a8404-6b77-418f-8fcb-3c72ed092816" containerName="neutron-db-sync" Oct 05 07:07:26 crc kubenswrapper[4846]: I1005 07:07:26.875042 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-686765ddbc-cnfdk" Oct 05 07:07:26 crc kubenswrapper[4846]: I1005 07:07:26.889726 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-686765ddbc-cnfdk"] Oct 05 07:07:26 crc kubenswrapper[4846]: I1005 07:07:26.922671 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vl9v9\" (UniqueName: \"kubernetes.io/projected/edc6bfd3-d138-4afc-8cd7-877acd26f22c-kube-api-access-vl9v9\") pod \"dnsmasq-dns-686765ddbc-cnfdk\" (UID: \"edc6bfd3-d138-4afc-8cd7-877acd26f22c\") " pod="openstack/dnsmasq-dns-686765ddbc-cnfdk" Oct 05 07:07:26 crc kubenswrapper[4846]: I1005 07:07:26.922764 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/edc6bfd3-d138-4afc-8cd7-877acd26f22c-config\") pod \"dnsmasq-dns-686765ddbc-cnfdk\" (UID: \"edc6bfd3-d138-4afc-8cd7-877acd26f22c\") " pod="openstack/dnsmasq-dns-686765ddbc-cnfdk" Oct 05 07:07:26 crc kubenswrapper[4846]: I1005 07:07:26.922799 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/edc6bfd3-d138-4afc-8cd7-877acd26f22c-dns-swift-storage-0\") pod \"dnsmasq-dns-686765ddbc-cnfdk\" (UID: \"edc6bfd3-d138-4afc-8cd7-877acd26f22c\") " pod="openstack/dnsmasq-dns-686765ddbc-cnfdk" Oct 05 07:07:26 crc kubenswrapper[4846]: I1005 07:07:26.922831 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/edc6bfd3-d138-4afc-8cd7-877acd26f22c-ovsdbserver-nb\") pod \"dnsmasq-dns-686765ddbc-cnfdk\" (UID: \"edc6bfd3-d138-4afc-8cd7-877acd26f22c\") " pod="openstack/dnsmasq-dns-686765ddbc-cnfdk" Oct 05 07:07:26 crc kubenswrapper[4846]: I1005 07:07:26.923009 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/edc6bfd3-d138-4afc-8cd7-877acd26f22c-ovsdbserver-sb\") pod \"dnsmasq-dns-686765ddbc-cnfdk\" (UID: \"edc6bfd3-d138-4afc-8cd7-877acd26f22c\") " pod="openstack/dnsmasq-dns-686765ddbc-cnfdk" Oct 05 07:07:26 crc kubenswrapper[4846]: I1005 07:07:26.923057 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/edc6bfd3-d138-4afc-8cd7-877acd26f22c-dns-svc\") pod \"dnsmasq-dns-686765ddbc-cnfdk\" (UID: \"edc6bfd3-d138-4afc-8cd7-877acd26f22c\") " pod="openstack/dnsmasq-dns-686765ddbc-cnfdk" Oct 05 07:07:26 crc kubenswrapper[4846]: I1005 07:07:26.979156 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-64c4f98476-pqrx9"] Oct 05 07:07:26 crc kubenswrapper[4846]: I1005 07:07:26.980820 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-64c4f98476-pqrx9" Oct 05 07:07:26 crc kubenswrapper[4846]: I1005 07:07:26.986970 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-ovndbs" Oct 05 07:07:26 crc kubenswrapper[4846]: I1005 07:07:26.987277 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Oct 05 07:07:26 crc kubenswrapper[4846]: I1005 07:07:26.987441 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-xft85" Oct 05 07:07:26 crc kubenswrapper[4846]: I1005 07:07:26.988228 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Oct 05 07:07:27 crc kubenswrapper[4846]: I1005 07:07:27.008150 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-64c4f98476-pqrx9"] Oct 05 07:07:27 crc kubenswrapper[4846]: I1005 07:07:27.027408 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/818246fd-88cc-4c04-8ea4-0e2521a0fc45-httpd-config\") pod \"neutron-64c4f98476-pqrx9\" (UID: \"818246fd-88cc-4c04-8ea4-0e2521a0fc45\") " pod="openstack/neutron-64c4f98476-pqrx9" Oct 05 07:07:27 crc kubenswrapper[4846]: I1005 07:07:27.027504 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vl9v9\" (UniqueName: \"kubernetes.io/projected/edc6bfd3-d138-4afc-8cd7-877acd26f22c-kube-api-access-vl9v9\") pod \"dnsmasq-dns-686765ddbc-cnfdk\" (UID: \"edc6bfd3-d138-4afc-8cd7-877acd26f22c\") " pod="openstack/dnsmasq-dns-686765ddbc-cnfdk" Oct 05 07:07:27 crc kubenswrapper[4846]: I1005 07:07:27.027553 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/818246fd-88cc-4c04-8ea4-0e2521a0fc45-ovndb-tls-certs\") pod \"neutron-64c4f98476-pqrx9\" (UID: \"818246fd-88cc-4c04-8ea4-0e2521a0fc45\") " pod="openstack/neutron-64c4f98476-pqrx9" Oct 05 07:07:27 crc kubenswrapper[4846]: I1005 07:07:27.027585 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/edc6bfd3-d138-4afc-8cd7-877acd26f22c-config\") pod \"dnsmasq-dns-686765ddbc-cnfdk\" (UID: \"edc6bfd3-d138-4afc-8cd7-877acd26f22c\") " pod="openstack/dnsmasq-dns-686765ddbc-cnfdk" Oct 05 07:07:27 crc kubenswrapper[4846]: I1005 07:07:27.027621 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/edc6bfd3-d138-4afc-8cd7-877acd26f22c-dns-swift-storage-0\") pod \"dnsmasq-dns-686765ddbc-cnfdk\" (UID: \"edc6bfd3-d138-4afc-8cd7-877acd26f22c\") " pod="openstack/dnsmasq-dns-686765ddbc-cnfdk" Oct 05 07:07:27 crc kubenswrapper[4846]: I1005 07:07:27.027658 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/edc6bfd3-d138-4afc-8cd7-877acd26f22c-ovsdbserver-nb\") pod \"dnsmasq-dns-686765ddbc-cnfdk\" (UID: \"edc6bfd3-d138-4afc-8cd7-877acd26f22c\") " pod="openstack/dnsmasq-dns-686765ddbc-cnfdk" Oct 05 07:07:27 crc kubenswrapper[4846]: I1005 07:07:27.027697 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g4645\" (UniqueName: \"kubernetes.io/projected/818246fd-88cc-4c04-8ea4-0e2521a0fc45-kube-api-access-g4645\") pod \"neutron-64c4f98476-pqrx9\" (UID: \"818246fd-88cc-4c04-8ea4-0e2521a0fc45\") " pod="openstack/neutron-64c4f98476-pqrx9" Oct 05 07:07:27 crc kubenswrapper[4846]: I1005 07:07:27.027737 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/edc6bfd3-d138-4afc-8cd7-877acd26f22c-ovsdbserver-sb\") pod \"dnsmasq-dns-686765ddbc-cnfdk\" (UID: \"edc6bfd3-d138-4afc-8cd7-877acd26f22c\") " pod="openstack/dnsmasq-dns-686765ddbc-cnfdk" Oct 05 07:07:27 crc kubenswrapper[4846]: I1005 07:07:27.027762 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/edc6bfd3-d138-4afc-8cd7-877acd26f22c-dns-svc\") pod \"dnsmasq-dns-686765ddbc-cnfdk\" (UID: \"edc6bfd3-d138-4afc-8cd7-877acd26f22c\") " pod="openstack/dnsmasq-dns-686765ddbc-cnfdk" Oct 05 07:07:27 crc kubenswrapper[4846]: I1005 07:07:27.027793 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/818246fd-88cc-4c04-8ea4-0e2521a0fc45-config\") pod \"neutron-64c4f98476-pqrx9\" (UID: \"818246fd-88cc-4c04-8ea4-0e2521a0fc45\") " pod="openstack/neutron-64c4f98476-pqrx9" Oct 05 07:07:27 crc kubenswrapper[4846]: I1005 07:07:27.027818 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/818246fd-88cc-4c04-8ea4-0e2521a0fc45-combined-ca-bundle\") pod \"neutron-64c4f98476-pqrx9\" (UID: \"818246fd-88cc-4c04-8ea4-0e2521a0fc45\") " pod="openstack/neutron-64c4f98476-pqrx9" Oct 05 07:07:27 crc kubenswrapper[4846]: I1005 07:07:27.029534 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/edc6bfd3-d138-4afc-8cd7-877acd26f22c-config\") pod \"dnsmasq-dns-686765ddbc-cnfdk\" (UID: \"edc6bfd3-d138-4afc-8cd7-877acd26f22c\") " pod="openstack/dnsmasq-dns-686765ddbc-cnfdk" Oct 05 07:07:27 crc kubenswrapper[4846]: I1005 07:07:27.030070 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/edc6bfd3-d138-4afc-8cd7-877acd26f22c-dns-svc\") pod \"dnsmasq-dns-686765ddbc-cnfdk\" (UID: \"edc6bfd3-d138-4afc-8cd7-877acd26f22c\") " pod="openstack/dnsmasq-dns-686765ddbc-cnfdk" Oct 05 07:07:27 crc kubenswrapper[4846]: I1005 07:07:27.030739 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/edc6bfd3-d138-4afc-8cd7-877acd26f22c-ovsdbserver-nb\") pod \"dnsmasq-dns-686765ddbc-cnfdk\" (UID: \"edc6bfd3-d138-4afc-8cd7-877acd26f22c\") " pod="openstack/dnsmasq-dns-686765ddbc-cnfdk" Oct 05 07:07:27 crc kubenswrapper[4846]: I1005 07:07:27.031313 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/edc6bfd3-d138-4afc-8cd7-877acd26f22c-dns-swift-storage-0\") pod \"dnsmasq-dns-686765ddbc-cnfdk\" (UID: \"edc6bfd3-d138-4afc-8cd7-877acd26f22c\") " pod="openstack/dnsmasq-dns-686765ddbc-cnfdk" Oct 05 07:07:27 crc kubenswrapper[4846]: I1005 07:07:27.031928 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/edc6bfd3-d138-4afc-8cd7-877acd26f22c-ovsdbserver-sb\") pod \"dnsmasq-dns-686765ddbc-cnfdk\" (UID: \"edc6bfd3-d138-4afc-8cd7-877acd26f22c\") " pod="openstack/dnsmasq-dns-686765ddbc-cnfdk" Oct 05 07:07:27 crc kubenswrapper[4846]: I1005 07:07:27.060677 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vl9v9\" (UniqueName: \"kubernetes.io/projected/edc6bfd3-d138-4afc-8cd7-877acd26f22c-kube-api-access-vl9v9\") pod \"dnsmasq-dns-686765ddbc-cnfdk\" (UID: \"edc6bfd3-d138-4afc-8cd7-877acd26f22c\") " pod="openstack/dnsmasq-dns-686765ddbc-cnfdk" Oct 05 07:07:27 crc kubenswrapper[4846]: I1005 07:07:27.130566 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/818246fd-88cc-4c04-8ea4-0e2521a0fc45-config\") pod \"neutron-64c4f98476-pqrx9\" (UID: \"818246fd-88cc-4c04-8ea4-0e2521a0fc45\") " pod="openstack/neutron-64c4f98476-pqrx9" Oct 05 07:07:27 crc kubenswrapper[4846]: I1005 07:07:27.130615 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/818246fd-88cc-4c04-8ea4-0e2521a0fc45-combined-ca-bundle\") pod \"neutron-64c4f98476-pqrx9\" (UID: \"818246fd-88cc-4c04-8ea4-0e2521a0fc45\") " pod="openstack/neutron-64c4f98476-pqrx9" Oct 05 07:07:27 crc kubenswrapper[4846]: I1005 07:07:27.135134 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/818246fd-88cc-4c04-8ea4-0e2521a0fc45-httpd-config\") pod \"neutron-64c4f98476-pqrx9\" (UID: \"818246fd-88cc-4c04-8ea4-0e2521a0fc45\") " pod="openstack/neutron-64c4f98476-pqrx9" Oct 05 07:07:27 crc kubenswrapper[4846]: I1005 07:07:27.135292 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/818246fd-88cc-4c04-8ea4-0e2521a0fc45-ovndb-tls-certs\") pod \"neutron-64c4f98476-pqrx9\" (UID: \"818246fd-88cc-4c04-8ea4-0e2521a0fc45\") " pod="openstack/neutron-64c4f98476-pqrx9" Oct 05 07:07:27 crc kubenswrapper[4846]: I1005 07:07:27.135463 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g4645\" (UniqueName: \"kubernetes.io/projected/818246fd-88cc-4c04-8ea4-0e2521a0fc45-kube-api-access-g4645\") pod \"neutron-64c4f98476-pqrx9\" (UID: \"818246fd-88cc-4c04-8ea4-0e2521a0fc45\") " pod="openstack/neutron-64c4f98476-pqrx9" Oct 05 07:07:27 crc kubenswrapper[4846]: I1005 07:07:27.135896 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/818246fd-88cc-4c04-8ea4-0e2521a0fc45-config\") pod \"neutron-64c4f98476-pqrx9\" (UID: \"818246fd-88cc-4c04-8ea4-0e2521a0fc45\") " pod="openstack/neutron-64c4f98476-pqrx9" Oct 05 07:07:27 crc kubenswrapper[4846]: I1005 07:07:27.139953 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/818246fd-88cc-4c04-8ea4-0e2521a0fc45-httpd-config\") pod \"neutron-64c4f98476-pqrx9\" (UID: \"818246fd-88cc-4c04-8ea4-0e2521a0fc45\") " pod="openstack/neutron-64c4f98476-pqrx9" Oct 05 07:07:27 crc kubenswrapper[4846]: I1005 07:07:27.141708 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/818246fd-88cc-4c04-8ea4-0e2521a0fc45-combined-ca-bundle\") pod \"neutron-64c4f98476-pqrx9\" (UID: \"818246fd-88cc-4c04-8ea4-0e2521a0fc45\") " pod="openstack/neutron-64c4f98476-pqrx9" Oct 05 07:07:27 crc kubenswrapper[4846]: I1005 07:07:27.143921 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/818246fd-88cc-4c04-8ea4-0e2521a0fc45-ovndb-tls-certs\") pod \"neutron-64c4f98476-pqrx9\" (UID: \"818246fd-88cc-4c04-8ea4-0e2521a0fc45\") " pod="openstack/neutron-64c4f98476-pqrx9" Oct 05 07:07:27 crc kubenswrapper[4846]: I1005 07:07:27.154543 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g4645\" (UniqueName: \"kubernetes.io/projected/818246fd-88cc-4c04-8ea4-0e2521a0fc45-kube-api-access-g4645\") pod \"neutron-64c4f98476-pqrx9\" (UID: \"818246fd-88cc-4c04-8ea4-0e2521a0fc45\") " pod="openstack/neutron-64c4f98476-pqrx9" Oct 05 07:07:27 crc kubenswrapper[4846]: I1005 07:07:27.196660 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-686765ddbc-cnfdk" Oct 05 07:07:27 crc kubenswrapper[4846]: I1005 07:07:27.317555 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-64c4f98476-pqrx9" Oct 05 07:07:27 crc kubenswrapper[4846]: I1005 07:07:27.779562 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-686765ddbc-cnfdk"] Oct 05 07:07:28 crc kubenswrapper[4846]: I1005 07:07:28.397847 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-64c4f98476-pqrx9"] Oct 05 07:07:28 crc kubenswrapper[4846]: I1005 07:07:28.785026 4846 generic.go:334] "Generic (PLEG): container finished" podID="979a46b2-96f3-498a-8838-4b73b4c8dc91" containerID="b9645c4a90f8320793bcfa361c3d8c571e3730e1f40c3ec1b6cd107151e4d1c2" exitCode=0 Oct 05 07:07:28 crc kubenswrapper[4846]: I1005 07:07:28.785443 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-vt9sx" event={"ID":"979a46b2-96f3-498a-8838-4b73b4c8dc91","Type":"ContainerDied","Data":"b9645c4a90f8320793bcfa361c3d8c571e3730e1f40c3ec1b6cd107151e4d1c2"} Oct 05 07:07:30 crc kubenswrapper[4846]: I1005 07:07:30.265196 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-85bbbcfb7f-kwfjt"] Oct 05 07:07:30 crc kubenswrapper[4846]: I1005 07:07:30.267852 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-85bbbcfb7f-kwfjt" Oct 05 07:07:30 crc kubenswrapper[4846]: I1005 07:07:30.274202 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-public-svc" Oct 05 07:07:30 crc kubenswrapper[4846]: I1005 07:07:30.274238 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-internal-svc" Oct 05 07:07:30 crc kubenswrapper[4846]: I1005 07:07:30.277745 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-85bbbcfb7f-kwfjt"] Oct 05 07:07:30 crc kubenswrapper[4846]: I1005 07:07:30.313289 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/a72c0cf5-8cd9-4f5f-8a2b-25bf41ad0638-config\") pod \"neutron-85bbbcfb7f-kwfjt\" (UID: \"a72c0cf5-8cd9-4f5f-8a2b-25bf41ad0638\") " pod="openstack/neutron-85bbbcfb7f-kwfjt" Oct 05 07:07:30 crc kubenswrapper[4846]: I1005 07:07:30.313425 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a72c0cf5-8cd9-4f5f-8a2b-25bf41ad0638-public-tls-certs\") pod \"neutron-85bbbcfb7f-kwfjt\" (UID: \"a72c0cf5-8cd9-4f5f-8a2b-25bf41ad0638\") " pod="openstack/neutron-85bbbcfb7f-kwfjt" Oct 05 07:07:30 crc kubenswrapper[4846]: I1005 07:07:30.313448 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8vlxz\" (UniqueName: \"kubernetes.io/projected/a72c0cf5-8cd9-4f5f-8a2b-25bf41ad0638-kube-api-access-8vlxz\") pod \"neutron-85bbbcfb7f-kwfjt\" (UID: \"a72c0cf5-8cd9-4f5f-8a2b-25bf41ad0638\") " pod="openstack/neutron-85bbbcfb7f-kwfjt" Oct 05 07:07:30 crc kubenswrapper[4846]: I1005 07:07:30.313496 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/a72c0cf5-8cd9-4f5f-8a2b-25bf41ad0638-ovndb-tls-certs\") pod \"neutron-85bbbcfb7f-kwfjt\" (UID: \"a72c0cf5-8cd9-4f5f-8a2b-25bf41ad0638\") " pod="openstack/neutron-85bbbcfb7f-kwfjt" Oct 05 07:07:30 crc kubenswrapper[4846]: I1005 07:07:30.313536 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/a72c0cf5-8cd9-4f5f-8a2b-25bf41ad0638-httpd-config\") pod \"neutron-85bbbcfb7f-kwfjt\" (UID: \"a72c0cf5-8cd9-4f5f-8a2b-25bf41ad0638\") " pod="openstack/neutron-85bbbcfb7f-kwfjt" Oct 05 07:07:30 crc kubenswrapper[4846]: I1005 07:07:30.313556 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a72c0cf5-8cd9-4f5f-8a2b-25bf41ad0638-internal-tls-certs\") pod \"neutron-85bbbcfb7f-kwfjt\" (UID: \"a72c0cf5-8cd9-4f5f-8a2b-25bf41ad0638\") " pod="openstack/neutron-85bbbcfb7f-kwfjt" Oct 05 07:07:30 crc kubenswrapper[4846]: I1005 07:07:30.313578 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a72c0cf5-8cd9-4f5f-8a2b-25bf41ad0638-combined-ca-bundle\") pod \"neutron-85bbbcfb7f-kwfjt\" (UID: \"a72c0cf5-8cd9-4f5f-8a2b-25bf41ad0638\") " pod="openstack/neutron-85bbbcfb7f-kwfjt" Oct 05 07:07:30 crc kubenswrapper[4846]: I1005 07:07:30.415268 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/a72c0cf5-8cd9-4f5f-8a2b-25bf41ad0638-ovndb-tls-certs\") pod \"neutron-85bbbcfb7f-kwfjt\" (UID: \"a72c0cf5-8cd9-4f5f-8a2b-25bf41ad0638\") " pod="openstack/neutron-85bbbcfb7f-kwfjt" Oct 05 07:07:30 crc kubenswrapper[4846]: I1005 07:07:30.415337 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/a72c0cf5-8cd9-4f5f-8a2b-25bf41ad0638-httpd-config\") pod \"neutron-85bbbcfb7f-kwfjt\" (UID: \"a72c0cf5-8cd9-4f5f-8a2b-25bf41ad0638\") " pod="openstack/neutron-85bbbcfb7f-kwfjt" Oct 05 07:07:30 crc kubenswrapper[4846]: I1005 07:07:30.415359 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a72c0cf5-8cd9-4f5f-8a2b-25bf41ad0638-internal-tls-certs\") pod \"neutron-85bbbcfb7f-kwfjt\" (UID: \"a72c0cf5-8cd9-4f5f-8a2b-25bf41ad0638\") " pod="openstack/neutron-85bbbcfb7f-kwfjt" Oct 05 07:07:30 crc kubenswrapper[4846]: I1005 07:07:30.416599 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a72c0cf5-8cd9-4f5f-8a2b-25bf41ad0638-combined-ca-bundle\") pod \"neutron-85bbbcfb7f-kwfjt\" (UID: \"a72c0cf5-8cd9-4f5f-8a2b-25bf41ad0638\") " pod="openstack/neutron-85bbbcfb7f-kwfjt" Oct 05 07:07:30 crc kubenswrapper[4846]: I1005 07:07:30.416675 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/a72c0cf5-8cd9-4f5f-8a2b-25bf41ad0638-config\") pod \"neutron-85bbbcfb7f-kwfjt\" (UID: \"a72c0cf5-8cd9-4f5f-8a2b-25bf41ad0638\") " pod="openstack/neutron-85bbbcfb7f-kwfjt" Oct 05 07:07:30 crc kubenswrapper[4846]: I1005 07:07:30.416755 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a72c0cf5-8cd9-4f5f-8a2b-25bf41ad0638-public-tls-certs\") pod \"neutron-85bbbcfb7f-kwfjt\" (UID: \"a72c0cf5-8cd9-4f5f-8a2b-25bf41ad0638\") " pod="openstack/neutron-85bbbcfb7f-kwfjt" Oct 05 07:07:30 crc kubenswrapper[4846]: I1005 07:07:30.416781 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8vlxz\" (UniqueName: \"kubernetes.io/projected/a72c0cf5-8cd9-4f5f-8a2b-25bf41ad0638-kube-api-access-8vlxz\") pod \"neutron-85bbbcfb7f-kwfjt\" (UID: \"a72c0cf5-8cd9-4f5f-8a2b-25bf41ad0638\") " pod="openstack/neutron-85bbbcfb7f-kwfjt" Oct 05 07:07:30 crc kubenswrapper[4846]: I1005 07:07:30.426559 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/a72c0cf5-8cd9-4f5f-8a2b-25bf41ad0638-ovndb-tls-certs\") pod \"neutron-85bbbcfb7f-kwfjt\" (UID: \"a72c0cf5-8cd9-4f5f-8a2b-25bf41ad0638\") " pod="openstack/neutron-85bbbcfb7f-kwfjt" Oct 05 07:07:30 crc kubenswrapper[4846]: I1005 07:07:30.426705 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/a72c0cf5-8cd9-4f5f-8a2b-25bf41ad0638-httpd-config\") pod \"neutron-85bbbcfb7f-kwfjt\" (UID: \"a72c0cf5-8cd9-4f5f-8a2b-25bf41ad0638\") " pod="openstack/neutron-85bbbcfb7f-kwfjt" Oct 05 07:07:30 crc kubenswrapper[4846]: I1005 07:07:30.426755 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a72c0cf5-8cd9-4f5f-8a2b-25bf41ad0638-combined-ca-bundle\") pod \"neutron-85bbbcfb7f-kwfjt\" (UID: \"a72c0cf5-8cd9-4f5f-8a2b-25bf41ad0638\") " pod="openstack/neutron-85bbbcfb7f-kwfjt" Oct 05 07:07:30 crc kubenswrapper[4846]: I1005 07:07:30.426998 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a72c0cf5-8cd9-4f5f-8a2b-25bf41ad0638-public-tls-certs\") pod \"neutron-85bbbcfb7f-kwfjt\" (UID: \"a72c0cf5-8cd9-4f5f-8a2b-25bf41ad0638\") " pod="openstack/neutron-85bbbcfb7f-kwfjt" Oct 05 07:07:30 crc kubenswrapper[4846]: I1005 07:07:30.427138 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/a72c0cf5-8cd9-4f5f-8a2b-25bf41ad0638-config\") pod \"neutron-85bbbcfb7f-kwfjt\" (UID: \"a72c0cf5-8cd9-4f5f-8a2b-25bf41ad0638\") " pod="openstack/neutron-85bbbcfb7f-kwfjt" Oct 05 07:07:30 crc kubenswrapper[4846]: I1005 07:07:30.433925 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a72c0cf5-8cd9-4f5f-8a2b-25bf41ad0638-internal-tls-certs\") pod \"neutron-85bbbcfb7f-kwfjt\" (UID: \"a72c0cf5-8cd9-4f5f-8a2b-25bf41ad0638\") " pod="openstack/neutron-85bbbcfb7f-kwfjt" Oct 05 07:07:30 crc kubenswrapper[4846]: I1005 07:07:30.440219 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8vlxz\" (UniqueName: \"kubernetes.io/projected/a72c0cf5-8cd9-4f5f-8a2b-25bf41ad0638-kube-api-access-8vlxz\") pod \"neutron-85bbbcfb7f-kwfjt\" (UID: \"a72c0cf5-8cd9-4f5f-8a2b-25bf41ad0638\") " pod="openstack/neutron-85bbbcfb7f-kwfjt" Oct 05 07:07:30 crc kubenswrapper[4846]: I1005 07:07:30.594241 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-85bbbcfb7f-kwfjt" Oct 05 07:07:31 crc kubenswrapper[4846]: W1005 07:07:31.531692 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podedc6bfd3_d138_4afc_8cd7_877acd26f22c.slice/crio-7fa0c75b138b59cef0bebd432b7f8562ce86b9e2492eb9e91f7eee0ac5ba726c WatchSource:0}: Error finding container 7fa0c75b138b59cef0bebd432b7f8562ce86b9e2492eb9e91f7eee0ac5ba726c: Status 404 returned error can't find the container with id 7fa0c75b138b59cef0bebd432b7f8562ce86b9e2492eb9e91f7eee0ac5ba726c Oct 05 07:07:31 crc kubenswrapper[4846]: W1005 07:07:31.549987 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod818246fd_88cc_4c04_8ea4_0e2521a0fc45.slice/crio-a3d50da89a9a16db1a951b96ed40e1e10d58c570efdd288e5afcad3c9fd2e14c WatchSource:0}: Error finding container a3d50da89a9a16db1a951b96ed40e1e10d58c570efdd288e5afcad3c9fd2e14c: Status 404 returned error can't find the container with id a3d50da89a9a16db1a951b96ed40e1e10d58c570efdd288e5afcad3c9fd2e14c Oct 05 07:07:31 crc kubenswrapper[4846]: I1005 07:07:31.824839 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-vt9sx" event={"ID":"979a46b2-96f3-498a-8838-4b73b4c8dc91","Type":"ContainerDied","Data":"9a1146f8c3547b62d691c15b3f03fb61910c1e307ff455b408673ebd5dcb3176"} Oct 05 07:07:31 crc kubenswrapper[4846]: I1005 07:07:31.825383 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9a1146f8c3547b62d691c15b3f03fb61910c1e307ff455b408673ebd5dcb3176" Oct 05 07:07:31 crc kubenswrapper[4846]: I1005 07:07:31.834112 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-686765ddbc-cnfdk" event={"ID":"edc6bfd3-d138-4afc-8cd7-877acd26f22c","Type":"ContainerStarted","Data":"7fa0c75b138b59cef0bebd432b7f8562ce86b9e2492eb9e91f7eee0ac5ba726c"} Oct 05 07:07:31 crc kubenswrapper[4846]: I1005 07:07:31.841502 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-64c4f98476-pqrx9" event={"ID":"818246fd-88cc-4c04-8ea4-0e2521a0fc45","Type":"ContainerStarted","Data":"a3d50da89a9a16db1a951b96ed40e1e10d58c570efdd288e5afcad3c9fd2e14c"} Oct 05 07:07:31 crc kubenswrapper[4846]: I1005 07:07:31.850708 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-vt9sx" Oct 05 07:07:31 crc kubenswrapper[4846]: I1005 07:07:31.950125 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/979a46b2-96f3-498a-8838-4b73b4c8dc91-credential-keys\") pod \"979a46b2-96f3-498a-8838-4b73b4c8dc91\" (UID: \"979a46b2-96f3-498a-8838-4b73b4c8dc91\") " Oct 05 07:07:31 crc kubenswrapper[4846]: I1005 07:07:31.950196 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/979a46b2-96f3-498a-8838-4b73b4c8dc91-config-data\") pod \"979a46b2-96f3-498a-8838-4b73b4c8dc91\" (UID: \"979a46b2-96f3-498a-8838-4b73b4c8dc91\") " Oct 05 07:07:31 crc kubenswrapper[4846]: I1005 07:07:31.950412 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/979a46b2-96f3-498a-8838-4b73b4c8dc91-scripts\") pod \"979a46b2-96f3-498a-8838-4b73b4c8dc91\" (UID: \"979a46b2-96f3-498a-8838-4b73b4c8dc91\") " Oct 05 07:07:31 crc kubenswrapper[4846]: I1005 07:07:31.950442 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/979a46b2-96f3-498a-8838-4b73b4c8dc91-combined-ca-bundle\") pod \"979a46b2-96f3-498a-8838-4b73b4c8dc91\" (UID: \"979a46b2-96f3-498a-8838-4b73b4c8dc91\") " Oct 05 07:07:31 crc kubenswrapper[4846]: I1005 07:07:31.950463 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/979a46b2-96f3-498a-8838-4b73b4c8dc91-fernet-keys\") pod \"979a46b2-96f3-498a-8838-4b73b4c8dc91\" (UID: \"979a46b2-96f3-498a-8838-4b73b4c8dc91\") " Oct 05 07:07:31 crc kubenswrapper[4846]: I1005 07:07:31.950543 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fp47l\" (UniqueName: \"kubernetes.io/projected/979a46b2-96f3-498a-8838-4b73b4c8dc91-kube-api-access-fp47l\") pod \"979a46b2-96f3-498a-8838-4b73b4c8dc91\" (UID: \"979a46b2-96f3-498a-8838-4b73b4c8dc91\") " Oct 05 07:07:31 crc kubenswrapper[4846]: I1005 07:07:31.959219 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/979a46b2-96f3-498a-8838-4b73b4c8dc91-scripts" (OuterVolumeSpecName: "scripts") pod "979a46b2-96f3-498a-8838-4b73b4c8dc91" (UID: "979a46b2-96f3-498a-8838-4b73b4c8dc91"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:07:31 crc kubenswrapper[4846]: I1005 07:07:31.959722 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/979a46b2-96f3-498a-8838-4b73b4c8dc91-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "979a46b2-96f3-498a-8838-4b73b4c8dc91" (UID: "979a46b2-96f3-498a-8838-4b73b4c8dc91"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:07:31 crc kubenswrapper[4846]: I1005 07:07:31.959782 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/979a46b2-96f3-498a-8838-4b73b4c8dc91-kube-api-access-fp47l" (OuterVolumeSpecName: "kube-api-access-fp47l") pod "979a46b2-96f3-498a-8838-4b73b4c8dc91" (UID: "979a46b2-96f3-498a-8838-4b73b4c8dc91"). InnerVolumeSpecName "kube-api-access-fp47l". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:07:31 crc kubenswrapper[4846]: I1005 07:07:31.962351 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/979a46b2-96f3-498a-8838-4b73b4c8dc91-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "979a46b2-96f3-498a-8838-4b73b4c8dc91" (UID: "979a46b2-96f3-498a-8838-4b73b4c8dc91"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:07:31 crc kubenswrapper[4846]: I1005 07:07:31.986207 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/979a46b2-96f3-498a-8838-4b73b4c8dc91-config-data" (OuterVolumeSpecName: "config-data") pod "979a46b2-96f3-498a-8838-4b73b4c8dc91" (UID: "979a46b2-96f3-498a-8838-4b73b4c8dc91"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:07:32 crc kubenswrapper[4846]: I1005 07:07:31.999967 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/979a46b2-96f3-498a-8838-4b73b4c8dc91-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "979a46b2-96f3-498a-8838-4b73b4c8dc91" (UID: "979a46b2-96f3-498a-8838-4b73b4c8dc91"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:07:32 crc kubenswrapper[4846]: I1005 07:07:32.053969 4846 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/979a46b2-96f3-498a-8838-4b73b4c8dc91-credential-keys\") on node \"crc\" DevicePath \"\"" Oct 05 07:07:32 crc kubenswrapper[4846]: I1005 07:07:32.054032 4846 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/979a46b2-96f3-498a-8838-4b73b4c8dc91-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 07:07:32 crc kubenswrapper[4846]: I1005 07:07:32.054048 4846 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/979a46b2-96f3-498a-8838-4b73b4c8dc91-scripts\") on node \"crc\" DevicePath \"\"" Oct 05 07:07:32 crc kubenswrapper[4846]: I1005 07:07:32.054060 4846 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/979a46b2-96f3-498a-8838-4b73b4c8dc91-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 07:07:32 crc kubenswrapper[4846]: I1005 07:07:32.054072 4846 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/979a46b2-96f3-498a-8838-4b73b4c8dc91-fernet-keys\") on node \"crc\" DevicePath \"\"" Oct 05 07:07:32 crc kubenswrapper[4846]: I1005 07:07:32.054082 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fp47l\" (UniqueName: \"kubernetes.io/projected/979a46b2-96f3-498a-8838-4b73b4c8dc91-kube-api-access-fp47l\") on node \"crc\" DevicePath \"\"" Oct 05 07:07:32 crc kubenswrapper[4846]: I1005 07:07:32.269357 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-85bbbcfb7f-kwfjt"] Oct 05 07:07:32 crc kubenswrapper[4846]: I1005 07:07:32.859770 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-64c4f98476-pqrx9" event={"ID":"818246fd-88cc-4c04-8ea4-0e2521a0fc45","Type":"ContainerStarted","Data":"1b5541d40b35256864b3d80cd509b1fb37d2f2e7b5097dfa124f6488c42f7d2f"} Oct 05 07:07:32 crc kubenswrapper[4846]: I1005 07:07:32.862703 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-64c4f98476-pqrx9" event={"ID":"818246fd-88cc-4c04-8ea4-0e2521a0fc45","Type":"ContainerStarted","Data":"d98e1b46601ac7b1c3f03252e286d71bbe2836d44edf5d2b5a62a2b0af440f28"} Oct 05 07:07:32 crc kubenswrapper[4846]: I1005 07:07:32.862847 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-64c4f98476-pqrx9" Oct 05 07:07:32 crc kubenswrapper[4846]: I1005 07:07:32.870381 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-7f2fl" event={"ID":"a6ab76e7-ba49-4976-b718-3ea57bb4b53a","Type":"ContainerStarted","Data":"69871ad24cf3a1c323c9675966861d808a81181ea8ee899b1ff38c1e60c644f9"} Oct 05 07:07:32 crc kubenswrapper[4846]: I1005 07:07:32.883953 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-85bbbcfb7f-kwfjt" event={"ID":"a72c0cf5-8cd9-4f5f-8a2b-25bf41ad0638","Type":"ContainerStarted","Data":"1ac7cfeb9e27da195d8d27a5e4c8617a950c62fafc93500f5db134c1b1af30c2"} Oct 05 07:07:32 crc kubenswrapper[4846]: I1005 07:07:32.884066 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-85bbbcfb7f-kwfjt" event={"ID":"a72c0cf5-8cd9-4f5f-8a2b-25bf41ad0638","Type":"ContainerStarted","Data":"57243f61674b9f96faaeba668f666e7518c0f44adbe9b6c573a200820bbb58d3"} Oct 05 07:07:32 crc kubenswrapper[4846]: I1005 07:07:32.884082 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-85bbbcfb7f-kwfjt" event={"ID":"a72c0cf5-8cd9-4f5f-8a2b-25bf41ad0638","Type":"ContainerStarted","Data":"f1cd9e6d2dcee73cc1f74d34d80432272e9c18c2ccc5fa1800a62740e3a25bb5"} Oct 05 07:07:32 crc kubenswrapper[4846]: I1005 07:07:32.884203 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-85bbbcfb7f-kwfjt" Oct 05 07:07:32 crc kubenswrapper[4846]: I1005 07:07:32.907133 4846 generic.go:334] "Generic (PLEG): container finished" podID="edc6bfd3-d138-4afc-8cd7-877acd26f22c" containerID="0eb9e0d175224b1e0ec66202ad4b7b80ef0288e203822839283ed1b14456659b" exitCode=0 Oct 05 07:07:32 crc kubenswrapper[4846]: I1005 07:07:32.907284 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-686765ddbc-cnfdk" event={"ID":"edc6bfd3-d138-4afc-8cd7-877acd26f22c","Type":"ContainerDied","Data":"0eb9e0d175224b1e0ec66202ad4b7b80ef0288e203822839283ed1b14456659b"} Oct 05 07:07:32 crc kubenswrapper[4846]: I1005 07:07:32.915865 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-64c4f98476-pqrx9" podStartSLOduration=6.915810629 podStartE2EDuration="6.915810629s" podCreationTimestamp="2025-10-05 07:07:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 07:07:32.907841326 +0000 UTC m=+1175.148694111" watchObservedRunningTime="2025-10-05 07:07:32.915810629 +0000 UTC m=+1175.156663404" Oct 05 07:07:32 crc kubenswrapper[4846]: I1005 07:07:32.924474 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-vt9sx" Oct 05 07:07:32 crc kubenswrapper[4846]: I1005 07:07:32.925493 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ae01bc6c-75fb-4958-9a75-6c212fd0ed3a","Type":"ContainerStarted","Data":"cd7a0167e5022ef22dc459912b29a30979c920a9d7327e00b83da07039436faa"} Oct 05 07:07:33 crc kubenswrapper[4846]: I1005 07:07:33.053403 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-85bbbcfb7f-kwfjt" podStartSLOduration=3.053377628 podStartE2EDuration="3.053377628s" podCreationTimestamp="2025-10-05 07:07:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 07:07:33.023115121 +0000 UTC m=+1175.263967896" watchObservedRunningTime="2025-10-05 07:07:33.053377628 +0000 UTC m=+1175.294230403" Oct 05 07:07:33 crc kubenswrapper[4846]: I1005 07:07:33.066674 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-79f67599fc-2sk76"] Oct 05 07:07:33 crc kubenswrapper[4846]: E1005 07:07:33.067544 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="979a46b2-96f3-498a-8838-4b73b4c8dc91" containerName="keystone-bootstrap" Oct 05 07:07:33 crc kubenswrapper[4846]: I1005 07:07:33.067581 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="979a46b2-96f3-498a-8838-4b73b4c8dc91" containerName="keystone-bootstrap" Oct 05 07:07:33 crc kubenswrapper[4846]: I1005 07:07:33.069289 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="979a46b2-96f3-498a-8838-4b73b4c8dc91" containerName="keystone-bootstrap" Oct 05 07:07:33 crc kubenswrapper[4846]: I1005 07:07:33.070256 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-79f67599fc-2sk76" Oct 05 07:07:33 crc kubenswrapper[4846]: I1005 07:07:33.080588 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-internal-svc" Oct 05 07:07:33 crc kubenswrapper[4846]: I1005 07:07:33.081030 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-76d8v" Oct 05 07:07:33 crc kubenswrapper[4846]: I1005 07:07:33.081248 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-public-svc" Oct 05 07:07:33 crc kubenswrapper[4846]: I1005 07:07:33.081493 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 05 07:07:33 crc kubenswrapper[4846]: I1005 07:07:33.082433 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 05 07:07:33 crc kubenswrapper[4846]: I1005 07:07:33.082742 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 05 07:07:33 crc kubenswrapper[4846]: I1005 07:07:33.089564 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/1fa90810-c620-48c2-a49c-762341889743-credential-keys\") pod \"keystone-79f67599fc-2sk76\" (UID: \"1fa90810-c620-48c2-a49c-762341889743\") " pod="openstack/keystone-79f67599fc-2sk76" Oct 05 07:07:33 crc kubenswrapper[4846]: I1005 07:07:33.089687 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m8n46\" (UniqueName: \"kubernetes.io/projected/1fa90810-c620-48c2-a49c-762341889743-kube-api-access-m8n46\") pod \"keystone-79f67599fc-2sk76\" (UID: \"1fa90810-c620-48c2-a49c-762341889743\") " pod="openstack/keystone-79f67599fc-2sk76" Oct 05 07:07:33 crc kubenswrapper[4846]: I1005 07:07:33.089817 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/1fa90810-c620-48c2-a49c-762341889743-fernet-keys\") pod \"keystone-79f67599fc-2sk76\" (UID: \"1fa90810-c620-48c2-a49c-762341889743\") " pod="openstack/keystone-79f67599fc-2sk76" Oct 05 07:07:33 crc kubenswrapper[4846]: I1005 07:07:33.089906 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1fa90810-c620-48c2-a49c-762341889743-combined-ca-bundle\") pod \"keystone-79f67599fc-2sk76\" (UID: \"1fa90810-c620-48c2-a49c-762341889743\") " pod="openstack/keystone-79f67599fc-2sk76" Oct 05 07:07:33 crc kubenswrapper[4846]: I1005 07:07:33.090021 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1fa90810-c620-48c2-a49c-762341889743-scripts\") pod \"keystone-79f67599fc-2sk76\" (UID: \"1fa90810-c620-48c2-a49c-762341889743\") " pod="openstack/keystone-79f67599fc-2sk76" Oct 05 07:07:33 crc kubenswrapper[4846]: I1005 07:07:33.090154 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1fa90810-c620-48c2-a49c-762341889743-public-tls-certs\") pod \"keystone-79f67599fc-2sk76\" (UID: \"1fa90810-c620-48c2-a49c-762341889743\") " pod="openstack/keystone-79f67599fc-2sk76" Oct 05 07:07:33 crc kubenswrapper[4846]: I1005 07:07:33.090250 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1fa90810-c620-48c2-a49c-762341889743-config-data\") pod \"keystone-79f67599fc-2sk76\" (UID: \"1fa90810-c620-48c2-a49c-762341889743\") " pod="openstack/keystone-79f67599fc-2sk76" Oct 05 07:07:33 crc kubenswrapper[4846]: I1005 07:07:33.090373 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1fa90810-c620-48c2-a49c-762341889743-internal-tls-certs\") pod \"keystone-79f67599fc-2sk76\" (UID: \"1fa90810-c620-48c2-a49c-762341889743\") " pod="openstack/keystone-79f67599fc-2sk76" Oct 05 07:07:33 crc kubenswrapper[4846]: I1005 07:07:33.107352 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-79f67599fc-2sk76"] Oct 05 07:07:33 crc kubenswrapper[4846]: I1005 07:07:33.110981 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-sync-7f2fl" podStartSLOduration=3.37742383 podStartE2EDuration="33.110953874s" podCreationTimestamp="2025-10-05 07:07:00 +0000 UTC" firstStartedPulling="2025-10-05 07:07:01.961574105 +0000 UTC m=+1144.202426880" lastFinishedPulling="2025-10-05 07:07:31.695104149 +0000 UTC m=+1173.935956924" observedRunningTime="2025-10-05 07:07:33.046931766 +0000 UTC m=+1175.287784541" watchObservedRunningTime="2025-10-05 07:07:33.110953874 +0000 UTC m=+1175.351806649" Oct 05 07:07:33 crc kubenswrapper[4846]: I1005 07:07:33.192953 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1fa90810-c620-48c2-a49c-762341889743-internal-tls-certs\") pod \"keystone-79f67599fc-2sk76\" (UID: \"1fa90810-c620-48c2-a49c-762341889743\") " pod="openstack/keystone-79f67599fc-2sk76" Oct 05 07:07:33 crc kubenswrapper[4846]: I1005 07:07:33.193037 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/1fa90810-c620-48c2-a49c-762341889743-credential-keys\") pod \"keystone-79f67599fc-2sk76\" (UID: \"1fa90810-c620-48c2-a49c-762341889743\") " pod="openstack/keystone-79f67599fc-2sk76" Oct 05 07:07:33 crc kubenswrapper[4846]: I1005 07:07:33.193061 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m8n46\" (UniqueName: \"kubernetes.io/projected/1fa90810-c620-48c2-a49c-762341889743-kube-api-access-m8n46\") pod \"keystone-79f67599fc-2sk76\" (UID: \"1fa90810-c620-48c2-a49c-762341889743\") " pod="openstack/keystone-79f67599fc-2sk76" Oct 05 07:07:33 crc kubenswrapper[4846]: I1005 07:07:33.193095 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/1fa90810-c620-48c2-a49c-762341889743-fernet-keys\") pod \"keystone-79f67599fc-2sk76\" (UID: \"1fa90810-c620-48c2-a49c-762341889743\") " pod="openstack/keystone-79f67599fc-2sk76" Oct 05 07:07:33 crc kubenswrapper[4846]: I1005 07:07:33.193117 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1fa90810-c620-48c2-a49c-762341889743-combined-ca-bundle\") pod \"keystone-79f67599fc-2sk76\" (UID: \"1fa90810-c620-48c2-a49c-762341889743\") " pod="openstack/keystone-79f67599fc-2sk76" Oct 05 07:07:33 crc kubenswrapper[4846]: I1005 07:07:33.193142 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1fa90810-c620-48c2-a49c-762341889743-scripts\") pod \"keystone-79f67599fc-2sk76\" (UID: \"1fa90810-c620-48c2-a49c-762341889743\") " pod="openstack/keystone-79f67599fc-2sk76" Oct 05 07:07:33 crc kubenswrapper[4846]: I1005 07:07:33.193204 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1fa90810-c620-48c2-a49c-762341889743-public-tls-certs\") pod \"keystone-79f67599fc-2sk76\" (UID: \"1fa90810-c620-48c2-a49c-762341889743\") " pod="openstack/keystone-79f67599fc-2sk76" Oct 05 07:07:33 crc kubenswrapper[4846]: I1005 07:07:33.193227 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1fa90810-c620-48c2-a49c-762341889743-config-data\") pod \"keystone-79f67599fc-2sk76\" (UID: \"1fa90810-c620-48c2-a49c-762341889743\") " pod="openstack/keystone-79f67599fc-2sk76" Oct 05 07:07:33 crc kubenswrapper[4846]: I1005 07:07:33.209994 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/1fa90810-c620-48c2-a49c-762341889743-credential-keys\") pod \"keystone-79f67599fc-2sk76\" (UID: \"1fa90810-c620-48c2-a49c-762341889743\") " pod="openstack/keystone-79f67599fc-2sk76" Oct 05 07:07:33 crc kubenswrapper[4846]: I1005 07:07:33.216102 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1fa90810-c620-48c2-a49c-762341889743-scripts\") pod \"keystone-79f67599fc-2sk76\" (UID: \"1fa90810-c620-48c2-a49c-762341889743\") " pod="openstack/keystone-79f67599fc-2sk76" Oct 05 07:07:33 crc kubenswrapper[4846]: I1005 07:07:33.216304 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1fa90810-c620-48c2-a49c-762341889743-internal-tls-certs\") pod \"keystone-79f67599fc-2sk76\" (UID: \"1fa90810-c620-48c2-a49c-762341889743\") " pod="openstack/keystone-79f67599fc-2sk76" Oct 05 07:07:33 crc kubenswrapper[4846]: I1005 07:07:33.218248 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/1fa90810-c620-48c2-a49c-762341889743-fernet-keys\") pod \"keystone-79f67599fc-2sk76\" (UID: \"1fa90810-c620-48c2-a49c-762341889743\") " pod="openstack/keystone-79f67599fc-2sk76" Oct 05 07:07:33 crc kubenswrapper[4846]: I1005 07:07:33.222394 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1fa90810-c620-48c2-a49c-762341889743-combined-ca-bundle\") pod \"keystone-79f67599fc-2sk76\" (UID: \"1fa90810-c620-48c2-a49c-762341889743\") " pod="openstack/keystone-79f67599fc-2sk76" Oct 05 07:07:33 crc kubenswrapper[4846]: I1005 07:07:33.223891 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1fa90810-c620-48c2-a49c-762341889743-public-tls-certs\") pod \"keystone-79f67599fc-2sk76\" (UID: \"1fa90810-c620-48c2-a49c-762341889743\") " pod="openstack/keystone-79f67599fc-2sk76" Oct 05 07:07:33 crc kubenswrapper[4846]: I1005 07:07:33.225912 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1fa90810-c620-48c2-a49c-762341889743-config-data\") pod \"keystone-79f67599fc-2sk76\" (UID: \"1fa90810-c620-48c2-a49c-762341889743\") " pod="openstack/keystone-79f67599fc-2sk76" Oct 05 07:07:33 crc kubenswrapper[4846]: I1005 07:07:33.225922 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m8n46\" (UniqueName: \"kubernetes.io/projected/1fa90810-c620-48c2-a49c-762341889743-kube-api-access-m8n46\") pod \"keystone-79f67599fc-2sk76\" (UID: \"1fa90810-c620-48c2-a49c-762341889743\") " pod="openstack/keystone-79f67599fc-2sk76" Oct 05 07:07:33 crc kubenswrapper[4846]: I1005 07:07:33.412004 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-79f67599fc-2sk76" Oct 05 07:07:33 crc kubenswrapper[4846]: I1005 07:07:33.929258 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-79f67599fc-2sk76"] Oct 05 07:07:33 crc kubenswrapper[4846]: I1005 07:07:33.939512 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-686765ddbc-cnfdk" event={"ID":"edc6bfd3-d138-4afc-8cd7-877acd26f22c","Type":"ContainerStarted","Data":"0f798ea15d74275842b48b01d8f59b6dc8d976659f232892f7029e083e182d9f"} Oct 05 07:07:33 crc kubenswrapper[4846]: I1005 07:07:33.941228 4846 generic.go:334] "Generic (PLEG): container finished" podID="8b6863be-f8e1-4a87-93d6-5cf756983747" containerID="6f45c438b8b209cc382e7a14455205149b1abf123df3b2467fe12ebe7d71967f" exitCode=0 Oct 05 07:07:33 crc kubenswrapper[4846]: I1005 07:07:33.942344 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-68h9n" event={"ID":"8b6863be-f8e1-4a87-93d6-5cf756983747","Type":"ContainerDied","Data":"6f45c438b8b209cc382e7a14455205149b1abf123df3b2467fe12ebe7d71967f"} Oct 05 07:07:34 crc kubenswrapper[4846]: I1005 07:07:34.954067 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-79f67599fc-2sk76" event={"ID":"1fa90810-c620-48c2-a49c-762341889743","Type":"ContainerStarted","Data":"0c1fa59eb2207671e67d590e69cac606e13f97317e06feb44ee17e3ee38b3559"} Oct 05 07:07:35 crc kubenswrapper[4846]: I1005 07:07:35.419389 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-68h9n" Oct 05 07:07:35 crc kubenswrapper[4846]: I1005 07:07:35.555412 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ctmvj\" (UniqueName: \"kubernetes.io/projected/8b6863be-f8e1-4a87-93d6-5cf756983747-kube-api-access-ctmvj\") pod \"8b6863be-f8e1-4a87-93d6-5cf756983747\" (UID: \"8b6863be-f8e1-4a87-93d6-5cf756983747\") " Oct 05 07:07:35 crc kubenswrapper[4846]: I1005 07:07:35.555521 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8b6863be-f8e1-4a87-93d6-5cf756983747-combined-ca-bundle\") pod \"8b6863be-f8e1-4a87-93d6-5cf756983747\" (UID: \"8b6863be-f8e1-4a87-93d6-5cf756983747\") " Oct 05 07:07:35 crc kubenswrapper[4846]: I1005 07:07:35.555654 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/8b6863be-f8e1-4a87-93d6-5cf756983747-db-sync-config-data\") pod \"8b6863be-f8e1-4a87-93d6-5cf756983747\" (UID: \"8b6863be-f8e1-4a87-93d6-5cf756983747\") " Oct 05 07:07:35 crc kubenswrapper[4846]: I1005 07:07:35.555750 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8b6863be-f8e1-4a87-93d6-5cf756983747-config-data\") pod \"8b6863be-f8e1-4a87-93d6-5cf756983747\" (UID: \"8b6863be-f8e1-4a87-93d6-5cf756983747\") " Oct 05 07:07:35 crc kubenswrapper[4846]: I1005 07:07:35.562812 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8b6863be-f8e1-4a87-93d6-5cf756983747-kube-api-access-ctmvj" (OuterVolumeSpecName: "kube-api-access-ctmvj") pod "8b6863be-f8e1-4a87-93d6-5cf756983747" (UID: "8b6863be-f8e1-4a87-93d6-5cf756983747"). InnerVolumeSpecName "kube-api-access-ctmvj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:07:35 crc kubenswrapper[4846]: I1005 07:07:35.563244 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8b6863be-f8e1-4a87-93d6-5cf756983747-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "8b6863be-f8e1-4a87-93d6-5cf756983747" (UID: "8b6863be-f8e1-4a87-93d6-5cf756983747"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:07:35 crc kubenswrapper[4846]: I1005 07:07:35.583677 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8b6863be-f8e1-4a87-93d6-5cf756983747-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8b6863be-f8e1-4a87-93d6-5cf756983747" (UID: "8b6863be-f8e1-4a87-93d6-5cf756983747"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:07:35 crc kubenswrapper[4846]: I1005 07:07:35.627260 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8b6863be-f8e1-4a87-93d6-5cf756983747-config-data" (OuterVolumeSpecName: "config-data") pod "8b6863be-f8e1-4a87-93d6-5cf756983747" (UID: "8b6863be-f8e1-4a87-93d6-5cf756983747"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:07:35 crc kubenswrapper[4846]: I1005 07:07:35.658563 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ctmvj\" (UniqueName: \"kubernetes.io/projected/8b6863be-f8e1-4a87-93d6-5cf756983747-kube-api-access-ctmvj\") on node \"crc\" DevicePath \"\"" Oct 05 07:07:35 crc kubenswrapper[4846]: I1005 07:07:35.658710 4846 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8b6863be-f8e1-4a87-93d6-5cf756983747-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 07:07:35 crc kubenswrapper[4846]: I1005 07:07:35.658772 4846 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/8b6863be-f8e1-4a87-93d6-5cf756983747-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 07:07:35 crc kubenswrapper[4846]: I1005 07:07:35.658846 4846 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8b6863be-f8e1-4a87-93d6-5cf756983747-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 07:07:35 crc kubenswrapper[4846]: I1005 07:07:35.963871 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-68h9n" event={"ID":"8b6863be-f8e1-4a87-93d6-5cf756983747","Type":"ContainerDied","Data":"c30f24f2f258f70f53877a7c663c371c329d5e9a5c2706e611af8bd107bfcd8f"} Oct 05 07:07:35 crc kubenswrapper[4846]: I1005 07:07:35.964328 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c30f24f2f258f70f53877a7c663c371c329d5e9a5c2706e611af8bd107bfcd8f" Oct 05 07:07:35 crc kubenswrapper[4846]: I1005 07:07:35.964364 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-686765ddbc-cnfdk" Oct 05 07:07:35 crc kubenswrapper[4846]: I1005 07:07:35.963918 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-68h9n" Oct 05 07:07:35 crc kubenswrapper[4846]: I1005 07:07:35.994225 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-686765ddbc-cnfdk" podStartSLOduration=9.994155691 podStartE2EDuration="9.994155691s" podCreationTimestamp="2025-10-05 07:07:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 07:07:35.985370716 +0000 UTC m=+1178.226223501" watchObservedRunningTime="2025-10-05 07:07:35.994155691 +0000 UTC m=+1178.235008466" Oct 05 07:07:36 crc kubenswrapper[4846]: I1005 07:07:36.400863 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-686765ddbc-cnfdk"] Oct 05 07:07:36 crc kubenswrapper[4846]: I1005 07:07:36.443105 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7fcd8fd75c-qmzjd"] Oct 05 07:07:36 crc kubenswrapper[4846]: E1005 07:07:36.443606 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8b6863be-f8e1-4a87-93d6-5cf756983747" containerName="glance-db-sync" Oct 05 07:07:36 crc kubenswrapper[4846]: I1005 07:07:36.443629 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="8b6863be-f8e1-4a87-93d6-5cf756983747" containerName="glance-db-sync" Oct 05 07:07:36 crc kubenswrapper[4846]: I1005 07:07:36.443788 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="8b6863be-f8e1-4a87-93d6-5cf756983747" containerName="glance-db-sync" Oct 05 07:07:36 crc kubenswrapper[4846]: I1005 07:07:36.444822 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7fcd8fd75c-qmzjd" Oct 05 07:07:36 crc kubenswrapper[4846]: I1005 07:07:36.470668 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7fcd8fd75c-qmzjd"] Oct 05 07:07:36 crc kubenswrapper[4846]: I1005 07:07:36.575433 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f6b4f351-efb1-4a53-8431-e7058eace718-ovsdbserver-nb\") pod \"dnsmasq-dns-7fcd8fd75c-qmzjd\" (UID: \"f6b4f351-efb1-4a53-8431-e7058eace718\") " pod="openstack/dnsmasq-dns-7fcd8fd75c-qmzjd" Oct 05 07:07:36 crc kubenswrapper[4846]: I1005 07:07:36.575692 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f6b4f351-efb1-4a53-8431-e7058eace718-dns-svc\") pod \"dnsmasq-dns-7fcd8fd75c-qmzjd\" (UID: \"f6b4f351-efb1-4a53-8431-e7058eace718\") " pod="openstack/dnsmasq-dns-7fcd8fd75c-qmzjd" Oct 05 07:07:36 crc kubenswrapper[4846]: I1005 07:07:36.576054 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f6b4f351-efb1-4a53-8431-e7058eace718-dns-swift-storage-0\") pod \"dnsmasq-dns-7fcd8fd75c-qmzjd\" (UID: \"f6b4f351-efb1-4a53-8431-e7058eace718\") " pod="openstack/dnsmasq-dns-7fcd8fd75c-qmzjd" Oct 05 07:07:36 crc kubenswrapper[4846]: I1005 07:07:36.576217 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tsf6z\" (UniqueName: \"kubernetes.io/projected/f6b4f351-efb1-4a53-8431-e7058eace718-kube-api-access-tsf6z\") pod \"dnsmasq-dns-7fcd8fd75c-qmzjd\" (UID: \"f6b4f351-efb1-4a53-8431-e7058eace718\") " pod="openstack/dnsmasq-dns-7fcd8fd75c-qmzjd" Oct 05 07:07:36 crc kubenswrapper[4846]: I1005 07:07:36.576314 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f6b4f351-efb1-4a53-8431-e7058eace718-ovsdbserver-sb\") pod \"dnsmasq-dns-7fcd8fd75c-qmzjd\" (UID: \"f6b4f351-efb1-4a53-8431-e7058eace718\") " pod="openstack/dnsmasq-dns-7fcd8fd75c-qmzjd" Oct 05 07:07:36 crc kubenswrapper[4846]: I1005 07:07:36.576402 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f6b4f351-efb1-4a53-8431-e7058eace718-config\") pod \"dnsmasq-dns-7fcd8fd75c-qmzjd\" (UID: \"f6b4f351-efb1-4a53-8431-e7058eace718\") " pod="openstack/dnsmasq-dns-7fcd8fd75c-qmzjd" Oct 05 07:07:36 crc kubenswrapper[4846]: I1005 07:07:36.678304 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f6b4f351-efb1-4a53-8431-e7058eace718-dns-swift-storage-0\") pod \"dnsmasq-dns-7fcd8fd75c-qmzjd\" (UID: \"f6b4f351-efb1-4a53-8431-e7058eace718\") " pod="openstack/dnsmasq-dns-7fcd8fd75c-qmzjd" Oct 05 07:07:36 crc kubenswrapper[4846]: I1005 07:07:36.678418 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tsf6z\" (UniqueName: \"kubernetes.io/projected/f6b4f351-efb1-4a53-8431-e7058eace718-kube-api-access-tsf6z\") pod \"dnsmasq-dns-7fcd8fd75c-qmzjd\" (UID: \"f6b4f351-efb1-4a53-8431-e7058eace718\") " pod="openstack/dnsmasq-dns-7fcd8fd75c-qmzjd" Oct 05 07:07:36 crc kubenswrapper[4846]: I1005 07:07:36.678469 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f6b4f351-efb1-4a53-8431-e7058eace718-ovsdbserver-sb\") pod \"dnsmasq-dns-7fcd8fd75c-qmzjd\" (UID: \"f6b4f351-efb1-4a53-8431-e7058eace718\") " pod="openstack/dnsmasq-dns-7fcd8fd75c-qmzjd" Oct 05 07:07:36 crc kubenswrapper[4846]: I1005 07:07:36.678517 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f6b4f351-efb1-4a53-8431-e7058eace718-config\") pod \"dnsmasq-dns-7fcd8fd75c-qmzjd\" (UID: \"f6b4f351-efb1-4a53-8431-e7058eace718\") " pod="openstack/dnsmasq-dns-7fcd8fd75c-qmzjd" Oct 05 07:07:36 crc kubenswrapper[4846]: I1005 07:07:36.678544 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f6b4f351-efb1-4a53-8431-e7058eace718-ovsdbserver-nb\") pod \"dnsmasq-dns-7fcd8fd75c-qmzjd\" (UID: \"f6b4f351-efb1-4a53-8431-e7058eace718\") " pod="openstack/dnsmasq-dns-7fcd8fd75c-qmzjd" Oct 05 07:07:36 crc kubenswrapper[4846]: I1005 07:07:36.678646 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f6b4f351-efb1-4a53-8431-e7058eace718-dns-svc\") pod \"dnsmasq-dns-7fcd8fd75c-qmzjd\" (UID: \"f6b4f351-efb1-4a53-8431-e7058eace718\") " pod="openstack/dnsmasq-dns-7fcd8fd75c-qmzjd" Oct 05 07:07:36 crc kubenswrapper[4846]: I1005 07:07:36.679959 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f6b4f351-efb1-4a53-8431-e7058eace718-config\") pod \"dnsmasq-dns-7fcd8fd75c-qmzjd\" (UID: \"f6b4f351-efb1-4a53-8431-e7058eace718\") " pod="openstack/dnsmasq-dns-7fcd8fd75c-qmzjd" Oct 05 07:07:36 crc kubenswrapper[4846]: I1005 07:07:36.680087 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f6b4f351-efb1-4a53-8431-e7058eace718-dns-svc\") pod \"dnsmasq-dns-7fcd8fd75c-qmzjd\" (UID: \"f6b4f351-efb1-4a53-8431-e7058eace718\") " pod="openstack/dnsmasq-dns-7fcd8fd75c-qmzjd" Oct 05 07:07:36 crc kubenswrapper[4846]: I1005 07:07:36.680322 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f6b4f351-efb1-4a53-8431-e7058eace718-ovsdbserver-nb\") pod \"dnsmasq-dns-7fcd8fd75c-qmzjd\" (UID: \"f6b4f351-efb1-4a53-8431-e7058eace718\") " pod="openstack/dnsmasq-dns-7fcd8fd75c-qmzjd" Oct 05 07:07:36 crc kubenswrapper[4846]: I1005 07:07:36.680851 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f6b4f351-efb1-4a53-8431-e7058eace718-ovsdbserver-sb\") pod \"dnsmasq-dns-7fcd8fd75c-qmzjd\" (UID: \"f6b4f351-efb1-4a53-8431-e7058eace718\") " pod="openstack/dnsmasq-dns-7fcd8fd75c-qmzjd" Oct 05 07:07:36 crc kubenswrapper[4846]: I1005 07:07:36.681326 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f6b4f351-efb1-4a53-8431-e7058eace718-dns-swift-storage-0\") pod \"dnsmasq-dns-7fcd8fd75c-qmzjd\" (UID: \"f6b4f351-efb1-4a53-8431-e7058eace718\") " pod="openstack/dnsmasq-dns-7fcd8fd75c-qmzjd" Oct 05 07:07:36 crc kubenswrapper[4846]: I1005 07:07:36.715515 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tsf6z\" (UniqueName: \"kubernetes.io/projected/f6b4f351-efb1-4a53-8431-e7058eace718-kube-api-access-tsf6z\") pod \"dnsmasq-dns-7fcd8fd75c-qmzjd\" (UID: \"f6b4f351-efb1-4a53-8431-e7058eace718\") " pod="openstack/dnsmasq-dns-7fcd8fd75c-qmzjd" Oct 05 07:07:36 crc kubenswrapper[4846]: I1005 07:07:36.767843 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7fcd8fd75c-qmzjd" Oct 05 07:07:37 crc kubenswrapper[4846]: I1005 07:07:37.000696 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-79f67599fc-2sk76" event={"ID":"1fa90810-c620-48c2-a49c-762341889743","Type":"ContainerStarted","Data":"26da8c60efda49d0ad9e38f18f39b3b3c7a75ffb60bb839d759345619a5d337c"} Oct 05 07:07:37 crc kubenswrapper[4846]: I1005 07:07:37.321529 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-79f67599fc-2sk76" podStartSLOduration=5.321503105 podStartE2EDuration="5.321503105s" podCreationTimestamp="2025-10-05 07:07:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 07:07:37.027465562 +0000 UTC m=+1179.268318347" watchObservedRunningTime="2025-10-05 07:07:37.321503105 +0000 UTC m=+1179.562355890" Oct 05 07:07:37 crc kubenswrapper[4846]: I1005 07:07:37.326389 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7fcd8fd75c-qmzjd"] Oct 05 07:07:37 crc kubenswrapper[4846]: I1005 07:07:37.423697 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Oct 05 07:07:37 crc kubenswrapper[4846]: I1005 07:07:37.428514 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 05 07:07:37 crc kubenswrapper[4846]: I1005 07:07:37.433133 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Oct 05 07:07:37 crc kubenswrapper[4846]: I1005 07:07:37.433432 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Oct 05 07:07:37 crc kubenswrapper[4846]: I1005 07:07:37.433683 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-ff9fp" Oct 05 07:07:37 crc kubenswrapper[4846]: I1005 07:07:37.445099 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 05 07:07:37 crc kubenswrapper[4846]: I1005 07:07:37.535888 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 05 07:07:37 crc kubenswrapper[4846]: I1005 07:07:37.538029 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 05 07:07:37 crc kubenswrapper[4846]: I1005 07:07:37.545494 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Oct 05 07:07:37 crc kubenswrapper[4846]: I1005 07:07:37.557080 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 05 07:07:37 crc kubenswrapper[4846]: I1005 07:07:37.609104 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/345b75dc-52c2-4a1b-a5d6-19235d77f106-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"345b75dc-52c2-4a1b-a5d6-19235d77f106\") " pod="openstack/glance-default-external-api-0" Oct 05 07:07:37 crc kubenswrapper[4846]: I1005 07:07:37.609166 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/345b75dc-52c2-4a1b-a5d6-19235d77f106-logs\") pod \"glance-default-external-api-0\" (UID: \"345b75dc-52c2-4a1b-a5d6-19235d77f106\") " pod="openstack/glance-default-external-api-0" Oct 05 07:07:37 crc kubenswrapper[4846]: I1005 07:07:37.609225 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/345b75dc-52c2-4a1b-a5d6-19235d77f106-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"345b75dc-52c2-4a1b-a5d6-19235d77f106\") " pod="openstack/glance-default-external-api-0" Oct 05 07:07:37 crc kubenswrapper[4846]: I1005 07:07:37.609414 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dj7q5\" (UniqueName: \"kubernetes.io/projected/345b75dc-52c2-4a1b-a5d6-19235d77f106-kube-api-access-dj7q5\") pod \"glance-default-external-api-0\" (UID: \"345b75dc-52c2-4a1b-a5d6-19235d77f106\") " pod="openstack/glance-default-external-api-0" Oct 05 07:07:37 crc kubenswrapper[4846]: I1005 07:07:37.609500 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/345b75dc-52c2-4a1b-a5d6-19235d77f106-config-data\") pod \"glance-default-external-api-0\" (UID: \"345b75dc-52c2-4a1b-a5d6-19235d77f106\") " pod="openstack/glance-default-external-api-0" Oct 05 07:07:37 crc kubenswrapper[4846]: I1005 07:07:37.609558 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-external-api-0\" (UID: \"345b75dc-52c2-4a1b-a5d6-19235d77f106\") " pod="openstack/glance-default-external-api-0" Oct 05 07:07:37 crc kubenswrapper[4846]: I1005 07:07:37.609743 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/345b75dc-52c2-4a1b-a5d6-19235d77f106-scripts\") pod \"glance-default-external-api-0\" (UID: \"345b75dc-52c2-4a1b-a5d6-19235d77f106\") " pod="openstack/glance-default-external-api-0" Oct 05 07:07:37 crc kubenswrapper[4846]: I1005 07:07:37.711270 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"f7172e25-4d68-4c6f-ac5a-9721a3446a2c\") " pod="openstack/glance-default-internal-api-0" Oct 05 07:07:37 crc kubenswrapper[4846]: I1005 07:07:37.711323 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f7172e25-4d68-4c6f-ac5a-9721a3446a2c-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"f7172e25-4d68-4c6f-ac5a-9721a3446a2c\") " pod="openstack/glance-default-internal-api-0" Oct 05 07:07:37 crc kubenswrapper[4846]: I1005 07:07:37.711424 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dj7q5\" (UniqueName: \"kubernetes.io/projected/345b75dc-52c2-4a1b-a5d6-19235d77f106-kube-api-access-dj7q5\") pod \"glance-default-external-api-0\" (UID: \"345b75dc-52c2-4a1b-a5d6-19235d77f106\") " pod="openstack/glance-default-external-api-0" Oct 05 07:07:37 crc kubenswrapper[4846]: I1005 07:07:37.711473 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f7172e25-4d68-4c6f-ac5a-9721a3446a2c-scripts\") pod \"glance-default-internal-api-0\" (UID: \"f7172e25-4d68-4c6f-ac5a-9721a3446a2c\") " pod="openstack/glance-default-internal-api-0" Oct 05 07:07:37 crc kubenswrapper[4846]: I1005 07:07:37.711505 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/345b75dc-52c2-4a1b-a5d6-19235d77f106-config-data\") pod \"glance-default-external-api-0\" (UID: \"345b75dc-52c2-4a1b-a5d6-19235d77f106\") " pod="openstack/glance-default-external-api-0" Oct 05 07:07:37 crc kubenswrapper[4846]: I1005 07:07:37.711538 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f7172e25-4d68-4c6f-ac5a-9721a3446a2c-config-data\") pod \"glance-default-internal-api-0\" (UID: \"f7172e25-4d68-4c6f-ac5a-9721a3446a2c\") " pod="openstack/glance-default-internal-api-0" Oct 05 07:07:37 crc kubenswrapper[4846]: I1005 07:07:37.711573 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-external-api-0\" (UID: \"345b75dc-52c2-4a1b-a5d6-19235d77f106\") " pod="openstack/glance-default-external-api-0" Oct 05 07:07:37 crc kubenswrapper[4846]: I1005 07:07:37.711603 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qcl86\" (UniqueName: \"kubernetes.io/projected/f7172e25-4d68-4c6f-ac5a-9721a3446a2c-kube-api-access-qcl86\") pod \"glance-default-internal-api-0\" (UID: \"f7172e25-4d68-4c6f-ac5a-9721a3446a2c\") " pod="openstack/glance-default-internal-api-0" Oct 05 07:07:37 crc kubenswrapper[4846]: I1005 07:07:37.711633 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7172e25-4d68-4c6f-ac5a-9721a3446a2c-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"f7172e25-4d68-4c6f-ac5a-9721a3446a2c\") " pod="openstack/glance-default-internal-api-0" Oct 05 07:07:37 crc kubenswrapper[4846]: I1005 07:07:37.711686 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f7172e25-4d68-4c6f-ac5a-9721a3446a2c-logs\") pod \"glance-default-internal-api-0\" (UID: \"f7172e25-4d68-4c6f-ac5a-9721a3446a2c\") " pod="openstack/glance-default-internal-api-0" Oct 05 07:07:37 crc kubenswrapper[4846]: I1005 07:07:37.711724 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/345b75dc-52c2-4a1b-a5d6-19235d77f106-scripts\") pod \"glance-default-external-api-0\" (UID: \"345b75dc-52c2-4a1b-a5d6-19235d77f106\") " pod="openstack/glance-default-external-api-0" Oct 05 07:07:37 crc kubenswrapper[4846]: I1005 07:07:37.711762 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/345b75dc-52c2-4a1b-a5d6-19235d77f106-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"345b75dc-52c2-4a1b-a5d6-19235d77f106\") " pod="openstack/glance-default-external-api-0" Oct 05 07:07:37 crc kubenswrapper[4846]: I1005 07:07:37.711792 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/345b75dc-52c2-4a1b-a5d6-19235d77f106-logs\") pod \"glance-default-external-api-0\" (UID: \"345b75dc-52c2-4a1b-a5d6-19235d77f106\") " pod="openstack/glance-default-external-api-0" Oct 05 07:07:37 crc kubenswrapper[4846]: I1005 07:07:37.711827 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/345b75dc-52c2-4a1b-a5d6-19235d77f106-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"345b75dc-52c2-4a1b-a5d6-19235d77f106\") " pod="openstack/glance-default-external-api-0" Oct 05 07:07:37 crc kubenswrapper[4846]: I1005 07:07:37.712216 4846 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-external-api-0\" (UID: \"345b75dc-52c2-4a1b-a5d6-19235d77f106\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/glance-default-external-api-0" Oct 05 07:07:37 crc kubenswrapper[4846]: I1005 07:07:37.712405 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/345b75dc-52c2-4a1b-a5d6-19235d77f106-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"345b75dc-52c2-4a1b-a5d6-19235d77f106\") " pod="openstack/glance-default-external-api-0" Oct 05 07:07:37 crc kubenswrapper[4846]: I1005 07:07:37.712720 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/345b75dc-52c2-4a1b-a5d6-19235d77f106-logs\") pod \"glance-default-external-api-0\" (UID: \"345b75dc-52c2-4a1b-a5d6-19235d77f106\") " pod="openstack/glance-default-external-api-0" Oct 05 07:07:37 crc kubenswrapper[4846]: I1005 07:07:37.719915 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/345b75dc-52c2-4a1b-a5d6-19235d77f106-scripts\") pod \"glance-default-external-api-0\" (UID: \"345b75dc-52c2-4a1b-a5d6-19235d77f106\") " pod="openstack/glance-default-external-api-0" Oct 05 07:07:37 crc kubenswrapper[4846]: I1005 07:07:37.720718 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/345b75dc-52c2-4a1b-a5d6-19235d77f106-config-data\") pod \"glance-default-external-api-0\" (UID: \"345b75dc-52c2-4a1b-a5d6-19235d77f106\") " pod="openstack/glance-default-external-api-0" Oct 05 07:07:37 crc kubenswrapper[4846]: I1005 07:07:37.722171 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/345b75dc-52c2-4a1b-a5d6-19235d77f106-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"345b75dc-52c2-4a1b-a5d6-19235d77f106\") " pod="openstack/glance-default-external-api-0" Oct 05 07:07:37 crc kubenswrapper[4846]: I1005 07:07:37.733757 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dj7q5\" (UniqueName: \"kubernetes.io/projected/345b75dc-52c2-4a1b-a5d6-19235d77f106-kube-api-access-dj7q5\") pod \"glance-default-external-api-0\" (UID: \"345b75dc-52c2-4a1b-a5d6-19235d77f106\") " pod="openstack/glance-default-external-api-0" Oct 05 07:07:37 crc kubenswrapper[4846]: I1005 07:07:37.766037 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-external-api-0\" (UID: \"345b75dc-52c2-4a1b-a5d6-19235d77f106\") " pod="openstack/glance-default-external-api-0" Oct 05 07:07:37 crc kubenswrapper[4846]: I1005 07:07:37.813706 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qcl86\" (UniqueName: \"kubernetes.io/projected/f7172e25-4d68-4c6f-ac5a-9721a3446a2c-kube-api-access-qcl86\") pod \"glance-default-internal-api-0\" (UID: \"f7172e25-4d68-4c6f-ac5a-9721a3446a2c\") " pod="openstack/glance-default-internal-api-0" Oct 05 07:07:37 crc kubenswrapper[4846]: I1005 07:07:37.813760 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7172e25-4d68-4c6f-ac5a-9721a3446a2c-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"f7172e25-4d68-4c6f-ac5a-9721a3446a2c\") " pod="openstack/glance-default-internal-api-0" Oct 05 07:07:37 crc kubenswrapper[4846]: I1005 07:07:37.813805 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f7172e25-4d68-4c6f-ac5a-9721a3446a2c-logs\") pod \"glance-default-internal-api-0\" (UID: \"f7172e25-4d68-4c6f-ac5a-9721a3446a2c\") " pod="openstack/glance-default-internal-api-0" Oct 05 07:07:37 crc kubenswrapper[4846]: I1005 07:07:37.813876 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"f7172e25-4d68-4c6f-ac5a-9721a3446a2c\") " pod="openstack/glance-default-internal-api-0" Oct 05 07:07:37 crc kubenswrapper[4846]: I1005 07:07:37.813893 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f7172e25-4d68-4c6f-ac5a-9721a3446a2c-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"f7172e25-4d68-4c6f-ac5a-9721a3446a2c\") " pod="openstack/glance-default-internal-api-0" Oct 05 07:07:37 crc kubenswrapper[4846]: I1005 07:07:37.813952 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f7172e25-4d68-4c6f-ac5a-9721a3446a2c-scripts\") pod \"glance-default-internal-api-0\" (UID: \"f7172e25-4d68-4c6f-ac5a-9721a3446a2c\") " pod="openstack/glance-default-internal-api-0" Oct 05 07:07:37 crc kubenswrapper[4846]: I1005 07:07:37.813977 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f7172e25-4d68-4c6f-ac5a-9721a3446a2c-config-data\") pod \"glance-default-internal-api-0\" (UID: \"f7172e25-4d68-4c6f-ac5a-9721a3446a2c\") " pod="openstack/glance-default-internal-api-0" Oct 05 07:07:37 crc kubenswrapper[4846]: I1005 07:07:37.814927 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f7172e25-4d68-4c6f-ac5a-9721a3446a2c-logs\") pod \"glance-default-internal-api-0\" (UID: \"f7172e25-4d68-4c6f-ac5a-9721a3446a2c\") " pod="openstack/glance-default-internal-api-0" Oct 05 07:07:37 crc kubenswrapper[4846]: I1005 07:07:37.819122 4846 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"f7172e25-4d68-4c6f-ac5a-9721a3446a2c\") device mount path \"/mnt/openstack/pv12\"" pod="openstack/glance-default-internal-api-0" Oct 05 07:07:37 crc kubenswrapper[4846]: I1005 07:07:37.819297 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f7172e25-4d68-4c6f-ac5a-9721a3446a2c-config-data\") pod \"glance-default-internal-api-0\" (UID: \"f7172e25-4d68-4c6f-ac5a-9721a3446a2c\") " pod="openstack/glance-default-internal-api-0" Oct 05 07:07:37 crc kubenswrapper[4846]: I1005 07:07:37.819467 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 05 07:07:37 crc kubenswrapper[4846]: I1005 07:07:37.822660 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f7172e25-4d68-4c6f-ac5a-9721a3446a2c-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"f7172e25-4d68-4c6f-ac5a-9721a3446a2c\") " pod="openstack/glance-default-internal-api-0" Oct 05 07:07:37 crc kubenswrapper[4846]: I1005 07:07:37.833390 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7172e25-4d68-4c6f-ac5a-9721a3446a2c-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"f7172e25-4d68-4c6f-ac5a-9721a3446a2c\") " pod="openstack/glance-default-internal-api-0" Oct 05 07:07:37 crc kubenswrapper[4846]: I1005 07:07:37.835048 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f7172e25-4d68-4c6f-ac5a-9721a3446a2c-scripts\") pod \"glance-default-internal-api-0\" (UID: \"f7172e25-4d68-4c6f-ac5a-9721a3446a2c\") " pod="openstack/glance-default-internal-api-0" Oct 05 07:07:37 crc kubenswrapper[4846]: I1005 07:07:37.838680 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qcl86\" (UniqueName: \"kubernetes.io/projected/f7172e25-4d68-4c6f-ac5a-9721a3446a2c-kube-api-access-qcl86\") pod \"glance-default-internal-api-0\" (UID: \"f7172e25-4d68-4c6f-ac5a-9721a3446a2c\") " pod="openstack/glance-default-internal-api-0" Oct 05 07:07:37 crc kubenswrapper[4846]: I1005 07:07:37.846550 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"f7172e25-4d68-4c6f-ac5a-9721a3446a2c\") " pod="openstack/glance-default-internal-api-0" Oct 05 07:07:37 crc kubenswrapper[4846]: I1005 07:07:37.863992 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 05 07:07:38 crc kubenswrapper[4846]: I1005 07:07:38.008465 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-686765ddbc-cnfdk" podUID="edc6bfd3-d138-4afc-8cd7-877acd26f22c" containerName="dnsmasq-dns" containerID="cri-o://0f798ea15d74275842b48b01d8f59b6dc8d976659f232892f7029e083e182d9f" gracePeriod=10 Oct 05 07:07:38 crc kubenswrapper[4846]: I1005 07:07:38.008952 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/keystone-79f67599fc-2sk76" Oct 05 07:07:38 crc kubenswrapper[4846]: I1005 07:07:38.010424 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-686765ddbc-cnfdk" Oct 05 07:07:39 crc kubenswrapper[4846]: I1005 07:07:39.007226 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 05 07:07:39 crc kubenswrapper[4846]: I1005 07:07:39.036624 4846 generic.go:334] "Generic (PLEG): container finished" podID="edc6bfd3-d138-4afc-8cd7-877acd26f22c" containerID="0f798ea15d74275842b48b01d8f59b6dc8d976659f232892f7029e083e182d9f" exitCode=0 Oct 05 07:07:39 crc kubenswrapper[4846]: I1005 07:07:39.036722 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-686765ddbc-cnfdk" event={"ID":"edc6bfd3-d138-4afc-8cd7-877acd26f22c","Type":"ContainerDied","Data":"0f798ea15d74275842b48b01d8f59b6dc8d976659f232892f7029e083e182d9f"} Oct 05 07:07:39 crc kubenswrapper[4846]: I1005 07:07:39.055021 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7fcd8fd75c-qmzjd" event={"ID":"f6b4f351-efb1-4a53-8431-e7058eace718","Type":"ContainerStarted","Data":"f7228bbaf3f67d3a50ee1739dd126b0d08d75956f0364773df6454eae6d59fb3"} Oct 05 07:07:39 crc kubenswrapper[4846]: I1005 07:07:39.069483 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 05 07:07:39 crc kubenswrapper[4846]: I1005 07:07:39.179660 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-686765ddbc-cnfdk" Oct 05 07:07:39 crc kubenswrapper[4846]: I1005 07:07:39.243693 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/edc6bfd3-d138-4afc-8cd7-877acd26f22c-ovsdbserver-sb\") pod \"edc6bfd3-d138-4afc-8cd7-877acd26f22c\" (UID: \"edc6bfd3-d138-4afc-8cd7-877acd26f22c\") " Oct 05 07:07:39 crc kubenswrapper[4846]: I1005 07:07:39.243794 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/edc6bfd3-d138-4afc-8cd7-877acd26f22c-config\") pod \"edc6bfd3-d138-4afc-8cd7-877acd26f22c\" (UID: \"edc6bfd3-d138-4afc-8cd7-877acd26f22c\") " Oct 05 07:07:39 crc kubenswrapper[4846]: I1005 07:07:39.243911 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/edc6bfd3-d138-4afc-8cd7-877acd26f22c-ovsdbserver-nb\") pod \"edc6bfd3-d138-4afc-8cd7-877acd26f22c\" (UID: \"edc6bfd3-d138-4afc-8cd7-877acd26f22c\") " Oct 05 07:07:39 crc kubenswrapper[4846]: I1005 07:07:39.243938 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/edc6bfd3-d138-4afc-8cd7-877acd26f22c-dns-swift-storage-0\") pod \"edc6bfd3-d138-4afc-8cd7-877acd26f22c\" (UID: \"edc6bfd3-d138-4afc-8cd7-877acd26f22c\") " Oct 05 07:07:39 crc kubenswrapper[4846]: I1005 07:07:39.244016 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/edc6bfd3-d138-4afc-8cd7-877acd26f22c-dns-svc\") pod \"edc6bfd3-d138-4afc-8cd7-877acd26f22c\" (UID: \"edc6bfd3-d138-4afc-8cd7-877acd26f22c\") " Oct 05 07:07:39 crc kubenswrapper[4846]: I1005 07:07:39.244083 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vl9v9\" (UniqueName: \"kubernetes.io/projected/edc6bfd3-d138-4afc-8cd7-877acd26f22c-kube-api-access-vl9v9\") pod \"edc6bfd3-d138-4afc-8cd7-877acd26f22c\" (UID: \"edc6bfd3-d138-4afc-8cd7-877acd26f22c\") " Oct 05 07:07:39 crc kubenswrapper[4846]: I1005 07:07:39.255826 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/edc6bfd3-d138-4afc-8cd7-877acd26f22c-kube-api-access-vl9v9" (OuterVolumeSpecName: "kube-api-access-vl9v9") pod "edc6bfd3-d138-4afc-8cd7-877acd26f22c" (UID: "edc6bfd3-d138-4afc-8cd7-877acd26f22c"). InnerVolumeSpecName "kube-api-access-vl9v9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:07:39 crc kubenswrapper[4846]: I1005 07:07:39.313066 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/edc6bfd3-d138-4afc-8cd7-877acd26f22c-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "edc6bfd3-d138-4afc-8cd7-877acd26f22c" (UID: "edc6bfd3-d138-4afc-8cd7-877acd26f22c"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 07:07:39 crc kubenswrapper[4846]: I1005 07:07:39.321775 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/edc6bfd3-d138-4afc-8cd7-877acd26f22c-config" (OuterVolumeSpecName: "config") pod "edc6bfd3-d138-4afc-8cd7-877acd26f22c" (UID: "edc6bfd3-d138-4afc-8cd7-877acd26f22c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 07:07:39 crc kubenswrapper[4846]: I1005 07:07:39.333891 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/edc6bfd3-d138-4afc-8cd7-877acd26f22c-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "edc6bfd3-d138-4afc-8cd7-877acd26f22c" (UID: "edc6bfd3-d138-4afc-8cd7-877acd26f22c"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 07:07:39 crc kubenswrapper[4846]: I1005 07:07:39.342391 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 05 07:07:39 crc kubenswrapper[4846]: I1005 07:07:39.350649 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/edc6bfd3-d138-4afc-8cd7-877acd26f22c-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "edc6bfd3-d138-4afc-8cd7-877acd26f22c" (UID: "edc6bfd3-d138-4afc-8cd7-877acd26f22c"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 07:07:39 crc kubenswrapper[4846]: I1005 07:07:39.350819 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/edc6bfd3-d138-4afc-8cd7-877acd26f22c-dns-svc\") pod \"edc6bfd3-d138-4afc-8cd7-877acd26f22c\" (UID: \"edc6bfd3-d138-4afc-8cd7-877acd26f22c\") " Oct 05 07:07:39 crc kubenswrapper[4846]: W1005 07:07:39.351363 4846 empty_dir.go:500] Warning: Unmount skipped because path does not exist: /var/lib/kubelet/pods/edc6bfd3-d138-4afc-8cd7-877acd26f22c/volumes/kubernetes.io~configmap/dns-svc Oct 05 07:07:39 crc kubenswrapper[4846]: I1005 07:07:39.351377 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/edc6bfd3-d138-4afc-8cd7-877acd26f22c-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "edc6bfd3-d138-4afc-8cd7-877acd26f22c" (UID: "edc6bfd3-d138-4afc-8cd7-877acd26f22c"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 07:07:39 crc kubenswrapper[4846]: I1005 07:07:39.351800 4846 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/edc6bfd3-d138-4afc-8cd7-877acd26f22c-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 05 07:07:39 crc kubenswrapper[4846]: I1005 07:07:39.351823 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vl9v9\" (UniqueName: \"kubernetes.io/projected/edc6bfd3-d138-4afc-8cd7-877acd26f22c-kube-api-access-vl9v9\") on node \"crc\" DevicePath \"\"" Oct 05 07:07:39 crc kubenswrapper[4846]: I1005 07:07:39.351836 4846 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/edc6bfd3-d138-4afc-8cd7-877acd26f22c-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 05 07:07:39 crc kubenswrapper[4846]: I1005 07:07:39.351847 4846 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/edc6bfd3-d138-4afc-8cd7-877acd26f22c-config\") on node \"crc\" DevicePath \"\"" Oct 05 07:07:39 crc kubenswrapper[4846]: I1005 07:07:39.351857 4846 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/edc6bfd3-d138-4afc-8cd7-877acd26f22c-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 05 07:07:39 crc kubenswrapper[4846]: I1005 07:07:39.381488 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/edc6bfd3-d138-4afc-8cd7-877acd26f22c-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "edc6bfd3-d138-4afc-8cd7-877acd26f22c" (UID: "edc6bfd3-d138-4afc-8cd7-877acd26f22c"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 07:07:39 crc kubenswrapper[4846]: I1005 07:07:39.453857 4846 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/edc6bfd3-d138-4afc-8cd7-877acd26f22c-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 05 07:07:39 crc kubenswrapper[4846]: I1005 07:07:39.871963 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 05 07:07:40 crc kubenswrapper[4846]: I1005 07:07:40.078136 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-686765ddbc-cnfdk" event={"ID":"edc6bfd3-d138-4afc-8cd7-877acd26f22c","Type":"ContainerDied","Data":"7fa0c75b138b59cef0bebd432b7f8562ce86b9e2492eb9e91f7eee0ac5ba726c"} Oct 05 07:07:40 crc kubenswrapper[4846]: I1005 07:07:40.078634 4846 scope.go:117] "RemoveContainer" containerID="0f798ea15d74275842b48b01d8f59b6dc8d976659f232892f7029e083e182d9f" Oct 05 07:07:40 crc kubenswrapper[4846]: I1005 07:07:40.078861 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-686765ddbc-cnfdk" Oct 05 07:07:40 crc kubenswrapper[4846]: I1005 07:07:40.086363 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-cf6rt" event={"ID":"ee312852-b2eb-43d8-864c-bc553fed7ea5","Type":"ContainerStarted","Data":"8adc95b909659a53237badc065088ebc75df6ae5e870528ac833dde781fc0d99"} Oct 05 07:07:40 crc kubenswrapper[4846]: I1005 07:07:40.089593 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"f7172e25-4d68-4c6f-ac5a-9721a3446a2c","Type":"ContainerStarted","Data":"0120b10514b84b8b6d4cd962e7efaaa9c4b9bede73ad7a9d9c8ff9f21143a8c3"} Oct 05 07:07:40 crc kubenswrapper[4846]: I1005 07:07:40.094268 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-t5brq" event={"ID":"b7a58faa-0f37-4b0e-a907-6a8bae28cb27","Type":"ContainerStarted","Data":"27a50aa96a7a8bb54961eddcd8037b7777a781e62487ae4213c4c228650c5b06"} Oct 05 07:07:40 crc kubenswrapper[4846]: I1005 07:07:40.098575 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"345b75dc-52c2-4a1b-a5d6-19235d77f106","Type":"ContainerStarted","Data":"eb41f5a8cdd134b78ec289feb14c1f9ad5bc4591157a3f5a3eeb345c140582d1"} Oct 05 07:07:40 crc kubenswrapper[4846]: I1005 07:07:40.100575 4846 generic.go:334] "Generic (PLEG): container finished" podID="f6b4f351-efb1-4a53-8431-e7058eace718" containerID="155de7225ff92721c9958b6c64a09e10453594fc7642a9165ae9799de8832fde" exitCode=0 Oct 05 07:07:40 crc kubenswrapper[4846]: I1005 07:07:40.100609 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7fcd8fd75c-qmzjd" event={"ID":"f6b4f351-efb1-4a53-8431-e7058eace718","Type":"ContainerDied","Data":"155de7225ff92721c9958b6c64a09e10453594fc7642a9165ae9799de8832fde"} Oct 05 07:07:40 crc kubenswrapper[4846]: I1005 07:07:40.118542 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-sync-cf6rt" podStartSLOduration=3.013221627 podStartE2EDuration="39.118524422s" podCreationTimestamp="2025-10-05 07:07:01 +0000 UTC" firstStartedPulling="2025-10-05 07:07:02.469370837 +0000 UTC m=+1144.710223602" lastFinishedPulling="2025-10-05 07:07:38.574673622 +0000 UTC m=+1180.815526397" observedRunningTime="2025-10-05 07:07:40.117665889 +0000 UTC m=+1182.358518654" watchObservedRunningTime="2025-10-05 07:07:40.118524422 +0000 UTC m=+1182.359377197" Oct 05 07:07:40 crc kubenswrapper[4846]: I1005 07:07:40.147515 4846 scope.go:117] "RemoveContainer" containerID="0eb9e0d175224b1e0ec66202ad4b7b80ef0288e203822839283ed1b14456659b" Oct 05 07:07:40 crc kubenswrapper[4846]: I1005 07:07:40.149585 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-686765ddbc-cnfdk"] Oct 05 07:07:40 crc kubenswrapper[4846]: I1005 07:07:40.156687 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-686765ddbc-cnfdk"] Oct 05 07:07:40 crc kubenswrapper[4846]: I1005 07:07:40.171297 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-sync-t5brq" podStartSLOduration=3.077348618 podStartE2EDuration="39.171225258s" podCreationTimestamp="2025-10-05 07:07:01 +0000 UTC" firstStartedPulling="2025-10-05 07:07:02.469362617 +0000 UTC m=+1144.710215392" lastFinishedPulling="2025-10-05 07:07:38.563239257 +0000 UTC m=+1180.804092032" observedRunningTime="2025-10-05 07:07:40.161013925 +0000 UTC m=+1182.401866700" watchObservedRunningTime="2025-10-05 07:07:40.171225258 +0000 UTC m=+1182.412078033" Oct 05 07:07:40 crc kubenswrapper[4846]: I1005 07:07:40.509328 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="edc6bfd3-d138-4afc-8cd7-877acd26f22c" path="/var/lib/kubelet/pods/edc6bfd3-d138-4afc-8cd7-877acd26f22c/volumes" Oct 05 07:07:41 crc kubenswrapper[4846]: I1005 07:07:41.126952 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"f7172e25-4d68-4c6f-ac5a-9721a3446a2c","Type":"ContainerStarted","Data":"cb195538f98a0c7c4947d1630fe3c094eeab2ece547636b632401d6c5227b1da"} Oct 05 07:07:41 crc kubenswrapper[4846]: I1005 07:07:41.131196 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"345b75dc-52c2-4a1b-a5d6-19235d77f106","Type":"ContainerStarted","Data":"6b926ce432b446bdf98e92991c9970804e43f5e410b70883be51f7275c83a0f7"} Oct 05 07:07:41 crc kubenswrapper[4846]: I1005 07:07:41.131254 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"345b75dc-52c2-4a1b-a5d6-19235d77f106","Type":"ContainerStarted","Data":"a5cfabacef14577b711c23f0a7a8a84f3276e940fbd1566b76d8ae113a7d7c59"} Oct 05 07:07:41 crc kubenswrapper[4846]: I1005 07:07:41.131393 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="345b75dc-52c2-4a1b-a5d6-19235d77f106" containerName="glance-httpd" containerID="cri-o://6b926ce432b446bdf98e92991c9970804e43f5e410b70883be51f7275c83a0f7" gracePeriod=30 Oct 05 07:07:41 crc kubenswrapper[4846]: I1005 07:07:41.131380 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="345b75dc-52c2-4a1b-a5d6-19235d77f106" containerName="glance-log" containerID="cri-o://a5cfabacef14577b711c23f0a7a8a84f3276e940fbd1566b76d8ae113a7d7c59" gracePeriod=30 Oct 05 07:07:41 crc kubenswrapper[4846]: I1005 07:07:41.141794 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7fcd8fd75c-qmzjd" event={"ID":"f6b4f351-efb1-4a53-8431-e7058eace718","Type":"ContainerStarted","Data":"adb9e57c3af0eccf4eb1a4d4fa7671627385b5ebffd25943a05f6247adfca3ce"} Oct 05 07:07:41 crc kubenswrapper[4846]: I1005 07:07:41.141947 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7fcd8fd75c-qmzjd" Oct 05 07:07:41 crc kubenswrapper[4846]: I1005 07:07:41.157685 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=5.15766126 podStartE2EDuration="5.15766126s" podCreationTimestamp="2025-10-05 07:07:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 07:07:41.156645473 +0000 UTC m=+1183.397498268" watchObservedRunningTime="2025-10-05 07:07:41.15766126 +0000 UTC m=+1183.398514035" Oct 05 07:07:41 crc kubenswrapper[4846]: I1005 07:07:41.184694 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7fcd8fd75c-qmzjd" podStartSLOduration=5.18466864 podStartE2EDuration="5.18466864s" podCreationTimestamp="2025-10-05 07:07:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 07:07:41.184079664 +0000 UTC m=+1183.424932439" watchObservedRunningTime="2025-10-05 07:07:41.18466864 +0000 UTC m=+1183.425521415" Oct 05 07:07:42 crc kubenswrapper[4846]: I1005 07:07:42.152358 4846 generic.go:334] "Generic (PLEG): container finished" podID="a6ab76e7-ba49-4976-b718-3ea57bb4b53a" containerID="69871ad24cf3a1c323c9675966861d808a81181ea8ee899b1ff38c1e60c644f9" exitCode=0 Oct 05 07:07:42 crc kubenswrapper[4846]: I1005 07:07:42.152407 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-7f2fl" event={"ID":"a6ab76e7-ba49-4976-b718-3ea57bb4b53a","Type":"ContainerDied","Data":"69871ad24cf3a1c323c9675966861d808a81181ea8ee899b1ff38c1e60c644f9"} Oct 05 07:07:42 crc kubenswrapper[4846]: I1005 07:07:42.157046 4846 generic.go:334] "Generic (PLEG): container finished" podID="345b75dc-52c2-4a1b-a5d6-19235d77f106" containerID="6b926ce432b446bdf98e92991c9970804e43f5e410b70883be51f7275c83a0f7" exitCode=143 Oct 05 07:07:42 crc kubenswrapper[4846]: I1005 07:07:42.157070 4846 generic.go:334] "Generic (PLEG): container finished" podID="345b75dc-52c2-4a1b-a5d6-19235d77f106" containerID="a5cfabacef14577b711c23f0a7a8a84f3276e940fbd1566b76d8ae113a7d7c59" exitCode=143 Oct 05 07:07:42 crc kubenswrapper[4846]: I1005 07:07:42.157412 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"345b75dc-52c2-4a1b-a5d6-19235d77f106","Type":"ContainerDied","Data":"6b926ce432b446bdf98e92991c9970804e43f5e410b70883be51f7275c83a0f7"} Oct 05 07:07:42 crc kubenswrapper[4846]: I1005 07:07:42.157447 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"345b75dc-52c2-4a1b-a5d6-19235d77f106","Type":"ContainerDied","Data":"a5cfabacef14577b711c23f0a7a8a84f3276e940fbd1566b76d8ae113a7d7c59"} Oct 05 07:07:45 crc kubenswrapper[4846]: I1005 07:07:45.776312 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-7f2fl" Oct 05 07:07:45 crc kubenswrapper[4846]: I1005 07:07:45.834671 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a6ab76e7-ba49-4976-b718-3ea57bb4b53a-scripts\") pod \"a6ab76e7-ba49-4976-b718-3ea57bb4b53a\" (UID: \"a6ab76e7-ba49-4976-b718-3ea57bb4b53a\") " Oct 05 07:07:45 crc kubenswrapper[4846]: I1005 07:07:45.834854 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a6ab76e7-ba49-4976-b718-3ea57bb4b53a-combined-ca-bundle\") pod \"a6ab76e7-ba49-4976-b718-3ea57bb4b53a\" (UID: \"a6ab76e7-ba49-4976-b718-3ea57bb4b53a\") " Oct 05 07:07:45 crc kubenswrapper[4846]: I1005 07:07:45.834940 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ntr8z\" (UniqueName: \"kubernetes.io/projected/a6ab76e7-ba49-4976-b718-3ea57bb4b53a-kube-api-access-ntr8z\") pod \"a6ab76e7-ba49-4976-b718-3ea57bb4b53a\" (UID: \"a6ab76e7-ba49-4976-b718-3ea57bb4b53a\") " Oct 05 07:07:45 crc kubenswrapper[4846]: I1005 07:07:45.835111 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a6ab76e7-ba49-4976-b718-3ea57bb4b53a-config-data\") pod \"a6ab76e7-ba49-4976-b718-3ea57bb4b53a\" (UID: \"a6ab76e7-ba49-4976-b718-3ea57bb4b53a\") " Oct 05 07:07:45 crc kubenswrapper[4846]: I1005 07:07:45.835199 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a6ab76e7-ba49-4976-b718-3ea57bb4b53a-logs\") pod \"a6ab76e7-ba49-4976-b718-3ea57bb4b53a\" (UID: \"a6ab76e7-ba49-4976-b718-3ea57bb4b53a\") " Oct 05 07:07:45 crc kubenswrapper[4846]: I1005 07:07:45.836735 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a6ab76e7-ba49-4976-b718-3ea57bb4b53a-logs" (OuterVolumeSpecName: "logs") pod "a6ab76e7-ba49-4976-b718-3ea57bb4b53a" (UID: "a6ab76e7-ba49-4976-b718-3ea57bb4b53a"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 07:07:45 crc kubenswrapper[4846]: I1005 07:07:45.845031 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a6ab76e7-ba49-4976-b718-3ea57bb4b53a-scripts" (OuterVolumeSpecName: "scripts") pod "a6ab76e7-ba49-4976-b718-3ea57bb4b53a" (UID: "a6ab76e7-ba49-4976-b718-3ea57bb4b53a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:07:45 crc kubenswrapper[4846]: I1005 07:07:45.846594 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a6ab76e7-ba49-4976-b718-3ea57bb4b53a-kube-api-access-ntr8z" (OuterVolumeSpecName: "kube-api-access-ntr8z") pod "a6ab76e7-ba49-4976-b718-3ea57bb4b53a" (UID: "a6ab76e7-ba49-4976-b718-3ea57bb4b53a"). InnerVolumeSpecName "kube-api-access-ntr8z". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:07:45 crc kubenswrapper[4846]: I1005 07:07:45.870598 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a6ab76e7-ba49-4976-b718-3ea57bb4b53a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a6ab76e7-ba49-4976-b718-3ea57bb4b53a" (UID: "a6ab76e7-ba49-4976-b718-3ea57bb4b53a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:07:45 crc kubenswrapper[4846]: I1005 07:07:45.885498 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a6ab76e7-ba49-4976-b718-3ea57bb4b53a-config-data" (OuterVolumeSpecName: "config-data") pod "a6ab76e7-ba49-4976-b718-3ea57bb4b53a" (UID: "a6ab76e7-ba49-4976-b718-3ea57bb4b53a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:07:45 crc kubenswrapper[4846]: I1005 07:07:45.937734 4846 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a6ab76e7-ba49-4976-b718-3ea57bb4b53a-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 07:07:45 crc kubenswrapper[4846]: I1005 07:07:45.937774 4846 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a6ab76e7-ba49-4976-b718-3ea57bb4b53a-logs\") on node \"crc\" DevicePath \"\"" Oct 05 07:07:45 crc kubenswrapper[4846]: I1005 07:07:45.937784 4846 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a6ab76e7-ba49-4976-b718-3ea57bb4b53a-scripts\") on node \"crc\" DevicePath \"\"" Oct 05 07:07:45 crc kubenswrapper[4846]: I1005 07:07:45.937794 4846 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a6ab76e7-ba49-4976-b718-3ea57bb4b53a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 07:07:45 crc kubenswrapper[4846]: I1005 07:07:45.937808 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ntr8z\" (UniqueName: \"kubernetes.io/projected/a6ab76e7-ba49-4976-b718-3ea57bb4b53a-kube-api-access-ntr8z\") on node \"crc\" DevicePath \"\"" Oct 05 07:07:46 crc kubenswrapper[4846]: I1005 07:07:46.030332 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 05 07:07:46 crc kubenswrapper[4846]: I1005 07:07:46.141947 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/345b75dc-52c2-4a1b-a5d6-19235d77f106-combined-ca-bundle\") pod \"345b75dc-52c2-4a1b-a5d6-19235d77f106\" (UID: \"345b75dc-52c2-4a1b-a5d6-19235d77f106\") " Oct 05 07:07:46 crc kubenswrapper[4846]: I1005 07:07:46.142474 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/345b75dc-52c2-4a1b-a5d6-19235d77f106-logs\") pod \"345b75dc-52c2-4a1b-a5d6-19235d77f106\" (UID: \"345b75dc-52c2-4a1b-a5d6-19235d77f106\") " Oct 05 07:07:46 crc kubenswrapper[4846]: I1005 07:07:46.142503 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/345b75dc-52c2-4a1b-a5d6-19235d77f106-config-data\") pod \"345b75dc-52c2-4a1b-a5d6-19235d77f106\" (UID: \"345b75dc-52c2-4a1b-a5d6-19235d77f106\") " Oct 05 07:07:46 crc kubenswrapper[4846]: I1005 07:07:46.142618 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"345b75dc-52c2-4a1b-a5d6-19235d77f106\" (UID: \"345b75dc-52c2-4a1b-a5d6-19235d77f106\") " Oct 05 07:07:46 crc kubenswrapper[4846]: I1005 07:07:46.142655 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/345b75dc-52c2-4a1b-a5d6-19235d77f106-httpd-run\") pod \"345b75dc-52c2-4a1b-a5d6-19235d77f106\" (UID: \"345b75dc-52c2-4a1b-a5d6-19235d77f106\") " Oct 05 07:07:46 crc kubenswrapper[4846]: I1005 07:07:46.142685 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dj7q5\" (UniqueName: \"kubernetes.io/projected/345b75dc-52c2-4a1b-a5d6-19235d77f106-kube-api-access-dj7q5\") pod \"345b75dc-52c2-4a1b-a5d6-19235d77f106\" (UID: \"345b75dc-52c2-4a1b-a5d6-19235d77f106\") " Oct 05 07:07:46 crc kubenswrapper[4846]: I1005 07:07:46.142758 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/345b75dc-52c2-4a1b-a5d6-19235d77f106-scripts\") pod \"345b75dc-52c2-4a1b-a5d6-19235d77f106\" (UID: \"345b75dc-52c2-4a1b-a5d6-19235d77f106\") " Oct 05 07:07:46 crc kubenswrapper[4846]: I1005 07:07:46.143024 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/345b75dc-52c2-4a1b-a5d6-19235d77f106-logs" (OuterVolumeSpecName: "logs") pod "345b75dc-52c2-4a1b-a5d6-19235d77f106" (UID: "345b75dc-52c2-4a1b-a5d6-19235d77f106"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 07:07:46 crc kubenswrapper[4846]: I1005 07:07:46.143323 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/345b75dc-52c2-4a1b-a5d6-19235d77f106-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "345b75dc-52c2-4a1b-a5d6-19235d77f106" (UID: "345b75dc-52c2-4a1b-a5d6-19235d77f106"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 07:07:46 crc kubenswrapper[4846]: I1005 07:07:46.143390 4846 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/345b75dc-52c2-4a1b-a5d6-19235d77f106-logs\") on node \"crc\" DevicePath \"\"" Oct 05 07:07:46 crc kubenswrapper[4846]: I1005 07:07:46.147538 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/345b75dc-52c2-4a1b-a5d6-19235d77f106-scripts" (OuterVolumeSpecName: "scripts") pod "345b75dc-52c2-4a1b-a5d6-19235d77f106" (UID: "345b75dc-52c2-4a1b-a5d6-19235d77f106"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:07:46 crc kubenswrapper[4846]: I1005 07:07:46.147591 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/345b75dc-52c2-4a1b-a5d6-19235d77f106-kube-api-access-dj7q5" (OuterVolumeSpecName: "kube-api-access-dj7q5") pod "345b75dc-52c2-4a1b-a5d6-19235d77f106" (UID: "345b75dc-52c2-4a1b-a5d6-19235d77f106"). InnerVolumeSpecName "kube-api-access-dj7q5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:07:46 crc kubenswrapper[4846]: I1005 07:07:46.148037 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage07-crc" (OuterVolumeSpecName: "glance") pod "345b75dc-52c2-4a1b-a5d6-19235d77f106" (UID: "345b75dc-52c2-4a1b-a5d6-19235d77f106"). InnerVolumeSpecName "local-storage07-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 05 07:07:46 crc kubenswrapper[4846]: I1005 07:07:46.205488 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/345b75dc-52c2-4a1b-a5d6-19235d77f106-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "345b75dc-52c2-4a1b-a5d6-19235d77f106" (UID: "345b75dc-52c2-4a1b-a5d6-19235d77f106"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:07:46 crc kubenswrapper[4846]: I1005 07:07:46.209580 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"345b75dc-52c2-4a1b-a5d6-19235d77f106","Type":"ContainerDied","Data":"eb41f5a8cdd134b78ec289feb14c1f9ad5bc4591157a3f5a3eeb345c140582d1"} Oct 05 07:07:46 crc kubenswrapper[4846]: I1005 07:07:46.209666 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 05 07:07:46 crc kubenswrapper[4846]: I1005 07:07:46.209651 4846 scope.go:117] "RemoveContainer" containerID="6b926ce432b446bdf98e92991c9970804e43f5e410b70883be51f7275c83a0f7" Oct 05 07:07:46 crc kubenswrapper[4846]: I1005 07:07:46.212036 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-7f2fl" event={"ID":"a6ab76e7-ba49-4976-b718-3ea57bb4b53a","Type":"ContainerDied","Data":"6bd20d0b024c20d2d5d45abd3d0560096db7cf87d891316de66e8ad3f5657424"} Oct 05 07:07:46 crc kubenswrapper[4846]: I1005 07:07:46.212094 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-7f2fl" Oct 05 07:07:46 crc kubenswrapper[4846]: I1005 07:07:46.212099 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6bd20d0b024c20d2d5d45abd3d0560096db7cf87d891316de66e8ad3f5657424" Oct 05 07:07:46 crc kubenswrapper[4846]: I1005 07:07:46.232584 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/345b75dc-52c2-4a1b-a5d6-19235d77f106-config-data" (OuterVolumeSpecName: "config-data") pod "345b75dc-52c2-4a1b-a5d6-19235d77f106" (UID: "345b75dc-52c2-4a1b-a5d6-19235d77f106"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:07:46 crc kubenswrapper[4846]: I1005 07:07:46.241437 4846 scope.go:117] "RemoveContainer" containerID="a5cfabacef14577b711c23f0a7a8a84f3276e940fbd1566b76d8ae113a7d7c59" Oct 05 07:07:46 crc kubenswrapper[4846]: I1005 07:07:46.245103 4846 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/345b75dc-52c2-4a1b-a5d6-19235d77f106-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 07:07:46 crc kubenswrapper[4846]: I1005 07:07:46.245134 4846 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/345b75dc-52c2-4a1b-a5d6-19235d77f106-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 07:07:46 crc kubenswrapper[4846]: I1005 07:07:46.245193 4846 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" " Oct 05 07:07:46 crc kubenswrapper[4846]: I1005 07:07:46.245208 4846 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/345b75dc-52c2-4a1b-a5d6-19235d77f106-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 05 07:07:46 crc kubenswrapper[4846]: I1005 07:07:46.245218 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dj7q5\" (UniqueName: \"kubernetes.io/projected/345b75dc-52c2-4a1b-a5d6-19235d77f106-kube-api-access-dj7q5\") on node \"crc\" DevicePath \"\"" Oct 05 07:07:46 crc kubenswrapper[4846]: I1005 07:07:46.245229 4846 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/345b75dc-52c2-4a1b-a5d6-19235d77f106-scripts\") on node \"crc\" DevicePath \"\"" Oct 05 07:07:46 crc kubenswrapper[4846]: I1005 07:07:46.264889 4846 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage07-crc" (UniqueName: "kubernetes.io/local-volume/local-storage07-crc") on node "crc" Oct 05 07:07:46 crc kubenswrapper[4846]: I1005 07:07:46.346688 4846 reconciler_common.go:293] "Volume detached for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" DevicePath \"\"" Oct 05 07:07:46 crc kubenswrapper[4846]: I1005 07:07:46.561526 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 05 07:07:46 crc kubenswrapper[4846]: I1005 07:07:46.571801 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 05 07:07:46 crc kubenswrapper[4846]: I1005 07:07:46.623147 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Oct 05 07:07:46 crc kubenswrapper[4846]: E1005 07:07:46.624321 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="edc6bfd3-d138-4afc-8cd7-877acd26f22c" containerName="dnsmasq-dns" Oct 05 07:07:46 crc kubenswrapper[4846]: I1005 07:07:46.624337 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="edc6bfd3-d138-4afc-8cd7-877acd26f22c" containerName="dnsmasq-dns" Oct 05 07:07:46 crc kubenswrapper[4846]: E1005 07:07:46.624373 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="345b75dc-52c2-4a1b-a5d6-19235d77f106" containerName="glance-log" Oct 05 07:07:46 crc kubenswrapper[4846]: I1005 07:07:46.624380 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="345b75dc-52c2-4a1b-a5d6-19235d77f106" containerName="glance-log" Oct 05 07:07:46 crc kubenswrapper[4846]: E1005 07:07:46.624395 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="345b75dc-52c2-4a1b-a5d6-19235d77f106" containerName="glance-httpd" Oct 05 07:07:46 crc kubenswrapper[4846]: I1005 07:07:46.624402 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="345b75dc-52c2-4a1b-a5d6-19235d77f106" containerName="glance-httpd" Oct 05 07:07:46 crc kubenswrapper[4846]: E1005 07:07:46.624450 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="edc6bfd3-d138-4afc-8cd7-877acd26f22c" containerName="init" Oct 05 07:07:46 crc kubenswrapper[4846]: I1005 07:07:46.624458 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="edc6bfd3-d138-4afc-8cd7-877acd26f22c" containerName="init" Oct 05 07:07:46 crc kubenswrapper[4846]: E1005 07:07:46.624473 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a6ab76e7-ba49-4976-b718-3ea57bb4b53a" containerName="placement-db-sync" Oct 05 07:07:46 crc kubenswrapper[4846]: I1005 07:07:46.624481 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="a6ab76e7-ba49-4976-b718-3ea57bb4b53a" containerName="placement-db-sync" Oct 05 07:07:46 crc kubenswrapper[4846]: I1005 07:07:46.624940 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="a6ab76e7-ba49-4976-b718-3ea57bb4b53a" containerName="placement-db-sync" Oct 05 07:07:46 crc kubenswrapper[4846]: I1005 07:07:46.624975 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="edc6bfd3-d138-4afc-8cd7-877acd26f22c" containerName="dnsmasq-dns" Oct 05 07:07:46 crc kubenswrapper[4846]: I1005 07:07:46.625003 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="345b75dc-52c2-4a1b-a5d6-19235d77f106" containerName="glance-log" Oct 05 07:07:46 crc kubenswrapper[4846]: I1005 07:07:46.625026 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="345b75dc-52c2-4a1b-a5d6-19235d77f106" containerName="glance-httpd" Oct 05 07:07:46 crc kubenswrapper[4846]: I1005 07:07:46.640400 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 05 07:07:46 crc kubenswrapper[4846]: I1005 07:07:46.640585 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 05 07:07:46 crc kubenswrapper[4846]: I1005 07:07:46.643907 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Oct 05 07:07:46 crc kubenswrapper[4846]: I1005 07:07:46.643970 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Oct 05 07:07:46 crc kubenswrapper[4846]: I1005 07:07:46.765009 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/47ea4b78-ed6a-48c6-a7d2-0505378a874d-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"47ea4b78-ed6a-48c6-a7d2-0505378a874d\") " pod="openstack/glance-default-external-api-0" Oct 05 07:07:46 crc kubenswrapper[4846]: I1005 07:07:46.765059 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/47ea4b78-ed6a-48c6-a7d2-0505378a874d-config-data\") pod \"glance-default-external-api-0\" (UID: \"47ea4b78-ed6a-48c6-a7d2-0505378a874d\") " pod="openstack/glance-default-external-api-0" Oct 05 07:07:46 crc kubenswrapper[4846]: I1005 07:07:46.765096 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/47ea4b78-ed6a-48c6-a7d2-0505378a874d-logs\") pod \"glance-default-external-api-0\" (UID: \"47ea4b78-ed6a-48c6-a7d2-0505378a874d\") " pod="openstack/glance-default-external-api-0" Oct 05 07:07:46 crc kubenswrapper[4846]: I1005 07:07:46.765342 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/47ea4b78-ed6a-48c6-a7d2-0505378a874d-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"47ea4b78-ed6a-48c6-a7d2-0505378a874d\") " pod="openstack/glance-default-external-api-0" Oct 05 07:07:46 crc kubenswrapper[4846]: I1005 07:07:46.765470 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/47ea4b78-ed6a-48c6-a7d2-0505378a874d-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"47ea4b78-ed6a-48c6-a7d2-0505378a874d\") " pod="openstack/glance-default-external-api-0" Oct 05 07:07:46 crc kubenswrapper[4846]: I1005 07:07:46.765653 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-external-api-0\" (UID: \"47ea4b78-ed6a-48c6-a7d2-0505378a874d\") " pod="openstack/glance-default-external-api-0" Oct 05 07:07:46 crc kubenswrapper[4846]: I1005 07:07:46.765783 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rztq7\" (UniqueName: \"kubernetes.io/projected/47ea4b78-ed6a-48c6-a7d2-0505378a874d-kube-api-access-rztq7\") pod \"glance-default-external-api-0\" (UID: \"47ea4b78-ed6a-48c6-a7d2-0505378a874d\") " pod="openstack/glance-default-external-api-0" Oct 05 07:07:46 crc kubenswrapper[4846]: I1005 07:07:46.766164 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/47ea4b78-ed6a-48c6-a7d2-0505378a874d-scripts\") pod \"glance-default-external-api-0\" (UID: \"47ea4b78-ed6a-48c6-a7d2-0505378a874d\") " pod="openstack/glance-default-external-api-0" Oct 05 07:07:46 crc kubenswrapper[4846]: I1005 07:07:46.770878 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-7fcd8fd75c-qmzjd" Oct 05 07:07:46 crc kubenswrapper[4846]: I1005 07:07:46.858770 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7cf74955b9-njz69"] Oct 05 07:07:46 crc kubenswrapper[4846]: I1005 07:07:46.859599 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7cf74955b9-njz69" podUID="5514c26f-0fe1-4e09-9292-896f21f90f71" containerName="dnsmasq-dns" containerID="cri-o://1d7f5cfaee91723a32aaf61fa9be5825f7c8e5a2573f16e15c12504bbef939f2" gracePeriod=10 Oct 05 07:07:46 crc kubenswrapper[4846]: I1005 07:07:46.867955 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/47ea4b78-ed6a-48c6-a7d2-0505378a874d-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"47ea4b78-ed6a-48c6-a7d2-0505378a874d\") " pod="openstack/glance-default-external-api-0" Oct 05 07:07:46 crc kubenswrapper[4846]: I1005 07:07:46.868001 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/47ea4b78-ed6a-48c6-a7d2-0505378a874d-config-data\") pod \"glance-default-external-api-0\" (UID: \"47ea4b78-ed6a-48c6-a7d2-0505378a874d\") " pod="openstack/glance-default-external-api-0" Oct 05 07:07:46 crc kubenswrapper[4846]: I1005 07:07:46.868054 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/47ea4b78-ed6a-48c6-a7d2-0505378a874d-logs\") pod \"glance-default-external-api-0\" (UID: \"47ea4b78-ed6a-48c6-a7d2-0505378a874d\") " pod="openstack/glance-default-external-api-0" Oct 05 07:07:46 crc kubenswrapper[4846]: I1005 07:07:46.868085 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/47ea4b78-ed6a-48c6-a7d2-0505378a874d-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"47ea4b78-ed6a-48c6-a7d2-0505378a874d\") " pod="openstack/glance-default-external-api-0" Oct 05 07:07:46 crc kubenswrapper[4846]: I1005 07:07:46.868110 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/47ea4b78-ed6a-48c6-a7d2-0505378a874d-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"47ea4b78-ed6a-48c6-a7d2-0505378a874d\") " pod="openstack/glance-default-external-api-0" Oct 05 07:07:46 crc kubenswrapper[4846]: I1005 07:07:46.868159 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-external-api-0\" (UID: \"47ea4b78-ed6a-48c6-a7d2-0505378a874d\") " pod="openstack/glance-default-external-api-0" Oct 05 07:07:46 crc kubenswrapper[4846]: I1005 07:07:46.868198 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rztq7\" (UniqueName: \"kubernetes.io/projected/47ea4b78-ed6a-48c6-a7d2-0505378a874d-kube-api-access-rztq7\") pod \"glance-default-external-api-0\" (UID: \"47ea4b78-ed6a-48c6-a7d2-0505378a874d\") " pod="openstack/glance-default-external-api-0" Oct 05 07:07:46 crc kubenswrapper[4846]: I1005 07:07:46.868256 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/47ea4b78-ed6a-48c6-a7d2-0505378a874d-scripts\") pod \"glance-default-external-api-0\" (UID: \"47ea4b78-ed6a-48c6-a7d2-0505378a874d\") " pod="openstack/glance-default-external-api-0" Oct 05 07:07:46 crc kubenswrapper[4846]: I1005 07:07:46.870067 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/47ea4b78-ed6a-48c6-a7d2-0505378a874d-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"47ea4b78-ed6a-48c6-a7d2-0505378a874d\") " pod="openstack/glance-default-external-api-0" Oct 05 07:07:46 crc kubenswrapper[4846]: I1005 07:07:46.870436 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/47ea4b78-ed6a-48c6-a7d2-0505378a874d-logs\") pod \"glance-default-external-api-0\" (UID: \"47ea4b78-ed6a-48c6-a7d2-0505378a874d\") " pod="openstack/glance-default-external-api-0" Oct 05 07:07:46 crc kubenswrapper[4846]: I1005 07:07:46.871293 4846 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-external-api-0\" (UID: \"47ea4b78-ed6a-48c6-a7d2-0505378a874d\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/glance-default-external-api-0" Oct 05 07:07:46 crc kubenswrapper[4846]: I1005 07:07:46.878618 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/47ea4b78-ed6a-48c6-a7d2-0505378a874d-config-data\") pod \"glance-default-external-api-0\" (UID: \"47ea4b78-ed6a-48c6-a7d2-0505378a874d\") " pod="openstack/glance-default-external-api-0" Oct 05 07:07:46 crc kubenswrapper[4846]: I1005 07:07:46.882557 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/47ea4b78-ed6a-48c6-a7d2-0505378a874d-scripts\") pod \"glance-default-external-api-0\" (UID: \"47ea4b78-ed6a-48c6-a7d2-0505378a874d\") " pod="openstack/glance-default-external-api-0" Oct 05 07:07:46 crc kubenswrapper[4846]: I1005 07:07:46.894360 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/47ea4b78-ed6a-48c6-a7d2-0505378a874d-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"47ea4b78-ed6a-48c6-a7d2-0505378a874d\") " pod="openstack/glance-default-external-api-0" Oct 05 07:07:46 crc kubenswrapper[4846]: I1005 07:07:46.896442 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/47ea4b78-ed6a-48c6-a7d2-0505378a874d-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"47ea4b78-ed6a-48c6-a7d2-0505378a874d\") " pod="openstack/glance-default-external-api-0" Oct 05 07:07:46 crc kubenswrapper[4846]: I1005 07:07:46.900868 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rztq7\" (UniqueName: \"kubernetes.io/projected/47ea4b78-ed6a-48c6-a7d2-0505378a874d-kube-api-access-rztq7\") pod \"glance-default-external-api-0\" (UID: \"47ea4b78-ed6a-48c6-a7d2-0505378a874d\") " pod="openstack/glance-default-external-api-0" Oct 05 07:07:46 crc kubenswrapper[4846]: I1005 07:07:46.935912 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-external-api-0\" (UID: \"47ea4b78-ed6a-48c6-a7d2-0505378a874d\") " pod="openstack/glance-default-external-api-0" Oct 05 07:07:46 crc kubenswrapper[4846]: I1005 07:07:46.962582 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-65884db4f4-fsk6p"] Oct 05 07:07:46 crc kubenswrapper[4846]: I1005 07:07:46.964257 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-65884db4f4-fsk6p" Oct 05 07:07:46 crc kubenswrapper[4846]: I1005 07:07:46.967818 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Oct 05 07:07:46 crc kubenswrapper[4846]: I1005 07:07:46.968065 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-public-svc" Oct 05 07:07:46 crc kubenswrapper[4846]: I1005 07:07:46.968210 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-5d5jr" Oct 05 07:07:46 crc kubenswrapper[4846]: I1005 07:07:46.968468 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Oct 05 07:07:46 crc kubenswrapper[4846]: I1005 07:07:46.968596 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-internal-svc" Oct 05 07:07:46 crc kubenswrapper[4846]: I1005 07:07:46.978368 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 05 07:07:47 crc kubenswrapper[4846]: I1005 07:07:47.005269 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-65884db4f4-fsk6p"] Oct 05 07:07:47 crc kubenswrapper[4846]: I1005 07:07:47.073468 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2mxsf\" (UniqueName: \"kubernetes.io/projected/c35affe0-da4e-448f-aac0-7bba26f89040-kube-api-access-2mxsf\") pod \"placement-65884db4f4-fsk6p\" (UID: \"c35affe0-da4e-448f-aac0-7bba26f89040\") " pod="openstack/placement-65884db4f4-fsk6p" Oct 05 07:07:47 crc kubenswrapper[4846]: I1005 07:07:47.073681 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c35affe0-da4e-448f-aac0-7bba26f89040-config-data\") pod \"placement-65884db4f4-fsk6p\" (UID: \"c35affe0-da4e-448f-aac0-7bba26f89040\") " pod="openstack/placement-65884db4f4-fsk6p" Oct 05 07:07:47 crc kubenswrapper[4846]: I1005 07:07:47.073788 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c35affe0-da4e-448f-aac0-7bba26f89040-internal-tls-certs\") pod \"placement-65884db4f4-fsk6p\" (UID: \"c35affe0-da4e-448f-aac0-7bba26f89040\") " pod="openstack/placement-65884db4f4-fsk6p" Oct 05 07:07:47 crc kubenswrapper[4846]: I1005 07:07:47.073868 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c35affe0-da4e-448f-aac0-7bba26f89040-combined-ca-bundle\") pod \"placement-65884db4f4-fsk6p\" (UID: \"c35affe0-da4e-448f-aac0-7bba26f89040\") " pod="openstack/placement-65884db4f4-fsk6p" Oct 05 07:07:47 crc kubenswrapper[4846]: I1005 07:07:47.074042 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c35affe0-da4e-448f-aac0-7bba26f89040-public-tls-certs\") pod \"placement-65884db4f4-fsk6p\" (UID: \"c35affe0-da4e-448f-aac0-7bba26f89040\") " pod="openstack/placement-65884db4f4-fsk6p" Oct 05 07:07:47 crc kubenswrapper[4846]: I1005 07:07:47.074122 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c35affe0-da4e-448f-aac0-7bba26f89040-logs\") pod \"placement-65884db4f4-fsk6p\" (UID: \"c35affe0-da4e-448f-aac0-7bba26f89040\") " pod="openstack/placement-65884db4f4-fsk6p" Oct 05 07:07:47 crc kubenswrapper[4846]: I1005 07:07:47.074539 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c35affe0-da4e-448f-aac0-7bba26f89040-scripts\") pod \"placement-65884db4f4-fsk6p\" (UID: \"c35affe0-da4e-448f-aac0-7bba26f89040\") " pod="openstack/placement-65884db4f4-fsk6p" Oct 05 07:07:47 crc kubenswrapper[4846]: I1005 07:07:47.177808 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c35affe0-da4e-448f-aac0-7bba26f89040-internal-tls-certs\") pod \"placement-65884db4f4-fsk6p\" (UID: \"c35affe0-da4e-448f-aac0-7bba26f89040\") " pod="openstack/placement-65884db4f4-fsk6p" Oct 05 07:07:47 crc kubenswrapper[4846]: I1005 07:07:47.178347 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c35affe0-da4e-448f-aac0-7bba26f89040-combined-ca-bundle\") pod \"placement-65884db4f4-fsk6p\" (UID: \"c35affe0-da4e-448f-aac0-7bba26f89040\") " pod="openstack/placement-65884db4f4-fsk6p" Oct 05 07:07:47 crc kubenswrapper[4846]: I1005 07:07:47.178436 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c35affe0-da4e-448f-aac0-7bba26f89040-public-tls-certs\") pod \"placement-65884db4f4-fsk6p\" (UID: \"c35affe0-da4e-448f-aac0-7bba26f89040\") " pod="openstack/placement-65884db4f4-fsk6p" Oct 05 07:07:47 crc kubenswrapper[4846]: I1005 07:07:47.178498 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c35affe0-da4e-448f-aac0-7bba26f89040-logs\") pod \"placement-65884db4f4-fsk6p\" (UID: \"c35affe0-da4e-448f-aac0-7bba26f89040\") " pod="openstack/placement-65884db4f4-fsk6p" Oct 05 07:07:47 crc kubenswrapper[4846]: I1005 07:07:47.178554 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c35affe0-da4e-448f-aac0-7bba26f89040-scripts\") pod \"placement-65884db4f4-fsk6p\" (UID: \"c35affe0-da4e-448f-aac0-7bba26f89040\") " pod="openstack/placement-65884db4f4-fsk6p" Oct 05 07:07:47 crc kubenswrapper[4846]: I1005 07:07:47.178603 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2mxsf\" (UniqueName: \"kubernetes.io/projected/c35affe0-da4e-448f-aac0-7bba26f89040-kube-api-access-2mxsf\") pod \"placement-65884db4f4-fsk6p\" (UID: \"c35affe0-da4e-448f-aac0-7bba26f89040\") " pod="openstack/placement-65884db4f4-fsk6p" Oct 05 07:07:47 crc kubenswrapper[4846]: I1005 07:07:47.178654 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c35affe0-da4e-448f-aac0-7bba26f89040-config-data\") pod \"placement-65884db4f4-fsk6p\" (UID: \"c35affe0-da4e-448f-aac0-7bba26f89040\") " pod="openstack/placement-65884db4f4-fsk6p" Oct 05 07:07:47 crc kubenswrapper[4846]: I1005 07:07:47.179597 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c35affe0-da4e-448f-aac0-7bba26f89040-logs\") pod \"placement-65884db4f4-fsk6p\" (UID: \"c35affe0-da4e-448f-aac0-7bba26f89040\") " pod="openstack/placement-65884db4f4-fsk6p" Oct 05 07:07:47 crc kubenswrapper[4846]: I1005 07:07:47.184957 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c35affe0-da4e-448f-aac0-7bba26f89040-internal-tls-certs\") pod \"placement-65884db4f4-fsk6p\" (UID: \"c35affe0-da4e-448f-aac0-7bba26f89040\") " pod="openstack/placement-65884db4f4-fsk6p" Oct 05 07:07:47 crc kubenswrapper[4846]: I1005 07:07:47.185036 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c35affe0-da4e-448f-aac0-7bba26f89040-public-tls-certs\") pod \"placement-65884db4f4-fsk6p\" (UID: \"c35affe0-da4e-448f-aac0-7bba26f89040\") " pod="openstack/placement-65884db4f4-fsk6p" Oct 05 07:07:47 crc kubenswrapper[4846]: I1005 07:07:47.185355 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c35affe0-da4e-448f-aac0-7bba26f89040-combined-ca-bundle\") pod \"placement-65884db4f4-fsk6p\" (UID: \"c35affe0-da4e-448f-aac0-7bba26f89040\") " pod="openstack/placement-65884db4f4-fsk6p" Oct 05 07:07:47 crc kubenswrapper[4846]: I1005 07:07:47.186463 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c35affe0-da4e-448f-aac0-7bba26f89040-config-data\") pod \"placement-65884db4f4-fsk6p\" (UID: \"c35affe0-da4e-448f-aac0-7bba26f89040\") " pod="openstack/placement-65884db4f4-fsk6p" Oct 05 07:07:47 crc kubenswrapper[4846]: I1005 07:07:47.187327 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c35affe0-da4e-448f-aac0-7bba26f89040-scripts\") pod \"placement-65884db4f4-fsk6p\" (UID: \"c35affe0-da4e-448f-aac0-7bba26f89040\") " pod="openstack/placement-65884db4f4-fsk6p" Oct 05 07:07:47 crc kubenswrapper[4846]: I1005 07:07:47.200600 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2mxsf\" (UniqueName: \"kubernetes.io/projected/c35affe0-da4e-448f-aac0-7bba26f89040-kube-api-access-2mxsf\") pod \"placement-65884db4f4-fsk6p\" (UID: \"c35affe0-da4e-448f-aac0-7bba26f89040\") " pod="openstack/placement-65884db4f4-fsk6p" Oct 05 07:07:47 crc kubenswrapper[4846]: I1005 07:07:47.231506 4846 generic.go:334] "Generic (PLEG): container finished" podID="5514c26f-0fe1-4e09-9292-896f21f90f71" containerID="1d7f5cfaee91723a32aaf61fa9be5825f7c8e5a2573f16e15c12504bbef939f2" exitCode=0 Oct 05 07:07:47 crc kubenswrapper[4846]: I1005 07:07:47.231610 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7cf74955b9-njz69" event={"ID":"5514c26f-0fe1-4e09-9292-896f21f90f71","Type":"ContainerDied","Data":"1d7f5cfaee91723a32aaf61fa9be5825f7c8e5a2573f16e15c12504bbef939f2"} Oct 05 07:07:47 crc kubenswrapper[4846]: I1005 07:07:47.236245 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ae01bc6c-75fb-4958-9a75-6c212fd0ed3a","Type":"ContainerStarted","Data":"e6f0c66814a012bcdf9a55c51c7fd4fb06ab5d3c0cf88579133f5a05220e6214"} Oct 05 07:07:47 crc kubenswrapper[4846]: I1005 07:07:47.236540 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 05 07:07:47 crc kubenswrapper[4846]: I1005 07:07:47.236589 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="ae01bc6c-75fb-4958-9a75-6c212fd0ed3a" containerName="proxy-httpd" containerID="cri-o://e6f0c66814a012bcdf9a55c51c7fd4fb06ab5d3c0cf88579133f5a05220e6214" gracePeriod=30 Oct 05 07:07:47 crc kubenswrapper[4846]: I1005 07:07:47.236768 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="ae01bc6c-75fb-4958-9a75-6c212fd0ed3a" containerName="ceilometer-central-agent" containerID="cri-o://bedf3eddca9636e7febabe72b13554c3e7139ba3edbeca3aa33a3865508515eb" gracePeriod=30 Oct 05 07:07:47 crc kubenswrapper[4846]: I1005 07:07:47.236849 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="ae01bc6c-75fb-4958-9a75-6c212fd0ed3a" containerName="sg-core" containerID="cri-o://cd7a0167e5022ef22dc459912b29a30979c920a9d7327e00b83da07039436faa" gracePeriod=30 Oct 05 07:07:47 crc kubenswrapper[4846]: I1005 07:07:47.236870 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="ae01bc6c-75fb-4958-9a75-6c212fd0ed3a" containerName="ceilometer-notification-agent" containerID="cri-o://f80062678cc982eb6612835cc64c1c9757ef8fbe87a8c39fe4a660630c89a786" gracePeriod=30 Oct 05 07:07:47 crc kubenswrapper[4846]: I1005 07:07:47.260979 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"f7172e25-4d68-4c6f-ac5a-9721a3446a2c","Type":"ContainerStarted","Data":"dfb681f3095dc72a51239d91581b46c2f72cd3096542ebb005216651080f1d8d"} Oct 05 07:07:47 crc kubenswrapper[4846]: I1005 07:07:47.261497 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="f7172e25-4d68-4c6f-ac5a-9721a3446a2c" containerName="glance-httpd" containerID="cri-o://dfb681f3095dc72a51239d91581b46c2f72cd3096542ebb005216651080f1d8d" gracePeriod=30 Oct 05 07:07:47 crc kubenswrapper[4846]: I1005 07:07:47.262041 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="f7172e25-4d68-4c6f-ac5a-9721a3446a2c" containerName="glance-log" containerID="cri-o://cb195538f98a0c7c4947d1630fe3c094eeab2ece547636b632401d6c5227b1da" gracePeriod=30 Oct 05 07:07:47 crc kubenswrapper[4846]: I1005 07:07:47.288447 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.951614784 podStartE2EDuration="47.28841928s" podCreationTimestamp="2025-10-05 07:07:00 +0000 UTC" firstStartedPulling="2025-10-05 07:07:01.659612792 +0000 UTC m=+1143.900465557" lastFinishedPulling="2025-10-05 07:07:45.996417278 +0000 UTC m=+1188.237270053" observedRunningTime="2025-10-05 07:07:47.266751842 +0000 UTC m=+1189.507604617" watchObservedRunningTime="2025-10-05 07:07:47.28841928 +0000 UTC m=+1189.529272055" Oct 05 07:07:47 crc kubenswrapper[4846]: I1005 07:07:47.309404 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=11.309376169 podStartE2EDuration="11.309376169s" podCreationTimestamp="2025-10-05 07:07:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 07:07:47.299453494 +0000 UTC m=+1189.540306289" watchObservedRunningTime="2025-10-05 07:07:47.309376169 +0000 UTC m=+1189.550228944" Oct 05 07:07:47 crc kubenswrapper[4846]: I1005 07:07:47.428639 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7cf74955b9-njz69" Oct 05 07:07:47 crc kubenswrapper[4846]: I1005 07:07:47.463087 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-65884db4f4-fsk6p" Oct 05 07:07:47 crc kubenswrapper[4846]: I1005 07:07:47.484643 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9jqwj\" (UniqueName: \"kubernetes.io/projected/5514c26f-0fe1-4e09-9292-896f21f90f71-kube-api-access-9jqwj\") pod \"5514c26f-0fe1-4e09-9292-896f21f90f71\" (UID: \"5514c26f-0fe1-4e09-9292-896f21f90f71\") " Oct 05 07:07:47 crc kubenswrapper[4846]: I1005 07:07:47.489675 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5514c26f-0fe1-4e09-9292-896f21f90f71-dns-swift-storage-0\") pod \"5514c26f-0fe1-4e09-9292-896f21f90f71\" (UID: \"5514c26f-0fe1-4e09-9292-896f21f90f71\") " Oct 05 07:07:47 crc kubenswrapper[4846]: I1005 07:07:47.489811 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5514c26f-0fe1-4e09-9292-896f21f90f71-config\") pod \"5514c26f-0fe1-4e09-9292-896f21f90f71\" (UID: \"5514c26f-0fe1-4e09-9292-896f21f90f71\") " Oct 05 07:07:47 crc kubenswrapper[4846]: I1005 07:07:47.489873 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5514c26f-0fe1-4e09-9292-896f21f90f71-ovsdbserver-nb\") pod \"5514c26f-0fe1-4e09-9292-896f21f90f71\" (UID: \"5514c26f-0fe1-4e09-9292-896f21f90f71\") " Oct 05 07:07:47 crc kubenswrapper[4846]: I1005 07:07:47.489965 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5514c26f-0fe1-4e09-9292-896f21f90f71-ovsdbserver-sb\") pod \"5514c26f-0fe1-4e09-9292-896f21f90f71\" (UID: \"5514c26f-0fe1-4e09-9292-896f21f90f71\") " Oct 05 07:07:47 crc kubenswrapper[4846]: I1005 07:07:47.490034 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5514c26f-0fe1-4e09-9292-896f21f90f71-dns-svc\") pod \"5514c26f-0fe1-4e09-9292-896f21f90f71\" (UID: \"5514c26f-0fe1-4e09-9292-896f21f90f71\") " Oct 05 07:07:47 crc kubenswrapper[4846]: I1005 07:07:47.499607 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5514c26f-0fe1-4e09-9292-896f21f90f71-kube-api-access-9jqwj" (OuterVolumeSpecName: "kube-api-access-9jqwj") pod "5514c26f-0fe1-4e09-9292-896f21f90f71" (UID: "5514c26f-0fe1-4e09-9292-896f21f90f71"). InnerVolumeSpecName "kube-api-access-9jqwj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:07:47 crc kubenswrapper[4846]: I1005 07:07:47.567118 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5514c26f-0fe1-4e09-9292-896f21f90f71-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "5514c26f-0fe1-4e09-9292-896f21f90f71" (UID: "5514c26f-0fe1-4e09-9292-896f21f90f71"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 07:07:47 crc kubenswrapper[4846]: I1005 07:07:47.586511 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5514c26f-0fe1-4e09-9292-896f21f90f71-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "5514c26f-0fe1-4e09-9292-896f21f90f71" (UID: "5514c26f-0fe1-4e09-9292-896f21f90f71"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 07:07:47 crc kubenswrapper[4846]: I1005 07:07:47.597562 4846 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5514c26f-0fe1-4e09-9292-896f21f90f71-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 05 07:07:47 crc kubenswrapper[4846]: I1005 07:07:47.597600 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9jqwj\" (UniqueName: \"kubernetes.io/projected/5514c26f-0fe1-4e09-9292-896f21f90f71-kube-api-access-9jqwj\") on node \"crc\" DevicePath \"\"" Oct 05 07:07:47 crc kubenswrapper[4846]: I1005 07:07:47.597615 4846 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5514c26f-0fe1-4e09-9292-896f21f90f71-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 05 07:07:47 crc kubenswrapper[4846]: I1005 07:07:47.600031 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5514c26f-0fe1-4e09-9292-896f21f90f71-config" (OuterVolumeSpecName: "config") pod "5514c26f-0fe1-4e09-9292-896f21f90f71" (UID: "5514c26f-0fe1-4e09-9292-896f21f90f71"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 07:07:47 crc kubenswrapper[4846]: I1005 07:07:47.600144 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5514c26f-0fe1-4e09-9292-896f21f90f71-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "5514c26f-0fe1-4e09-9292-896f21f90f71" (UID: "5514c26f-0fe1-4e09-9292-896f21f90f71"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 07:07:47 crc kubenswrapper[4846]: I1005 07:07:47.618559 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5514c26f-0fe1-4e09-9292-896f21f90f71-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "5514c26f-0fe1-4e09-9292-896f21f90f71" (UID: "5514c26f-0fe1-4e09-9292-896f21f90f71"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 07:07:47 crc kubenswrapper[4846]: I1005 07:07:47.699687 4846 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5514c26f-0fe1-4e09-9292-896f21f90f71-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 05 07:07:47 crc kubenswrapper[4846]: I1005 07:07:47.699741 4846 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5514c26f-0fe1-4e09-9292-896f21f90f71-config\") on node \"crc\" DevicePath \"\"" Oct 05 07:07:47 crc kubenswrapper[4846]: I1005 07:07:47.699759 4846 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5514c26f-0fe1-4e09-9292-896f21f90f71-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 05 07:07:47 crc kubenswrapper[4846]: I1005 07:07:47.807408 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 05 07:07:48 crc kubenswrapper[4846]: I1005 07:07:48.041677 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-65884db4f4-fsk6p"] Oct 05 07:07:48 crc kubenswrapper[4846]: I1005 07:07:48.051844 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 05 07:07:48 crc kubenswrapper[4846]: I1005 07:07:48.109126 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f7172e25-4d68-4c6f-ac5a-9721a3446a2c-logs\") pod \"f7172e25-4d68-4c6f-ac5a-9721a3446a2c\" (UID: \"f7172e25-4d68-4c6f-ac5a-9721a3446a2c\") " Oct 05 07:07:48 crc kubenswrapper[4846]: I1005 07:07:48.109223 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7172e25-4d68-4c6f-ac5a-9721a3446a2c-combined-ca-bundle\") pod \"f7172e25-4d68-4c6f-ac5a-9721a3446a2c\" (UID: \"f7172e25-4d68-4c6f-ac5a-9721a3446a2c\") " Oct 05 07:07:48 crc kubenswrapper[4846]: I1005 07:07:48.109264 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f7172e25-4d68-4c6f-ac5a-9721a3446a2c-httpd-run\") pod \"f7172e25-4d68-4c6f-ac5a-9721a3446a2c\" (UID: \"f7172e25-4d68-4c6f-ac5a-9721a3446a2c\") " Oct 05 07:07:48 crc kubenswrapper[4846]: I1005 07:07:48.109326 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qcl86\" (UniqueName: \"kubernetes.io/projected/f7172e25-4d68-4c6f-ac5a-9721a3446a2c-kube-api-access-qcl86\") pod \"f7172e25-4d68-4c6f-ac5a-9721a3446a2c\" (UID: \"f7172e25-4d68-4c6f-ac5a-9721a3446a2c\") " Oct 05 07:07:48 crc kubenswrapper[4846]: I1005 07:07:48.109444 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f7172e25-4d68-4c6f-ac5a-9721a3446a2c-config-data\") pod \"f7172e25-4d68-4c6f-ac5a-9721a3446a2c\" (UID: \"f7172e25-4d68-4c6f-ac5a-9721a3446a2c\") " Oct 05 07:07:48 crc kubenswrapper[4846]: I1005 07:07:48.109585 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"f7172e25-4d68-4c6f-ac5a-9721a3446a2c\" (UID: \"f7172e25-4d68-4c6f-ac5a-9721a3446a2c\") " Oct 05 07:07:48 crc kubenswrapper[4846]: I1005 07:07:48.109608 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f7172e25-4d68-4c6f-ac5a-9721a3446a2c-scripts\") pod \"f7172e25-4d68-4c6f-ac5a-9721a3446a2c\" (UID: \"f7172e25-4d68-4c6f-ac5a-9721a3446a2c\") " Oct 05 07:07:48 crc kubenswrapper[4846]: I1005 07:07:48.110564 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f7172e25-4d68-4c6f-ac5a-9721a3446a2c-logs" (OuterVolumeSpecName: "logs") pod "f7172e25-4d68-4c6f-ac5a-9721a3446a2c" (UID: "f7172e25-4d68-4c6f-ac5a-9721a3446a2c"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 07:07:48 crc kubenswrapper[4846]: I1005 07:07:48.110714 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f7172e25-4d68-4c6f-ac5a-9721a3446a2c-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "f7172e25-4d68-4c6f-ac5a-9721a3446a2c" (UID: "f7172e25-4d68-4c6f-ac5a-9721a3446a2c"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 07:07:48 crc kubenswrapper[4846]: I1005 07:07:48.114847 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f7172e25-4d68-4c6f-ac5a-9721a3446a2c-kube-api-access-qcl86" (OuterVolumeSpecName: "kube-api-access-qcl86") pod "f7172e25-4d68-4c6f-ac5a-9721a3446a2c" (UID: "f7172e25-4d68-4c6f-ac5a-9721a3446a2c"). InnerVolumeSpecName "kube-api-access-qcl86". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:07:48 crc kubenswrapper[4846]: I1005 07:07:48.115390 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage12-crc" (OuterVolumeSpecName: "glance") pod "f7172e25-4d68-4c6f-ac5a-9721a3446a2c" (UID: "f7172e25-4d68-4c6f-ac5a-9721a3446a2c"). InnerVolumeSpecName "local-storage12-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 05 07:07:48 crc kubenswrapper[4846]: I1005 07:07:48.116033 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f7172e25-4d68-4c6f-ac5a-9721a3446a2c-scripts" (OuterVolumeSpecName: "scripts") pod "f7172e25-4d68-4c6f-ac5a-9721a3446a2c" (UID: "f7172e25-4d68-4c6f-ac5a-9721a3446a2c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:07:48 crc kubenswrapper[4846]: I1005 07:07:48.182754 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f7172e25-4d68-4c6f-ac5a-9721a3446a2c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f7172e25-4d68-4c6f-ac5a-9721a3446a2c" (UID: "f7172e25-4d68-4c6f-ac5a-9721a3446a2c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:07:48 crc kubenswrapper[4846]: I1005 07:07:48.194291 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f7172e25-4d68-4c6f-ac5a-9721a3446a2c-config-data" (OuterVolumeSpecName: "config-data") pod "f7172e25-4d68-4c6f-ac5a-9721a3446a2c" (UID: "f7172e25-4d68-4c6f-ac5a-9721a3446a2c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:07:48 crc kubenswrapper[4846]: I1005 07:07:48.213468 4846 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f7172e25-4d68-4c6f-ac5a-9721a3446a2c-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 07:07:48 crc kubenswrapper[4846]: I1005 07:07:48.213537 4846 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" " Oct 05 07:07:48 crc kubenswrapper[4846]: I1005 07:07:48.214411 4846 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f7172e25-4d68-4c6f-ac5a-9721a3446a2c-scripts\") on node \"crc\" DevicePath \"\"" Oct 05 07:07:48 crc kubenswrapper[4846]: I1005 07:07:48.214422 4846 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f7172e25-4d68-4c6f-ac5a-9721a3446a2c-logs\") on node \"crc\" DevicePath \"\"" Oct 05 07:07:48 crc kubenswrapper[4846]: I1005 07:07:48.214436 4846 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7172e25-4d68-4c6f-ac5a-9721a3446a2c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 07:07:48 crc kubenswrapper[4846]: I1005 07:07:48.214472 4846 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f7172e25-4d68-4c6f-ac5a-9721a3446a2c-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 05 07:07:48 crc kubenswrapper[4846]: I1005 07:07:48.214483 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qcl86\" (UniqueName: \"kubernetes.io/projected/f7172e25-4d68-4c6f-ac5a-9721a3446a2c-kube-api-access-qcl86\") on node \"crc\" DevicePath \"\"" Oct 05 07:07:48 crc kubenswrapper[4846]: I1005 07:07:48.237004 4846 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage12-crc" (UniqueName: "kubernetes.io/local-volume/local-storage12-crc") on node "crc" Oct 05 07:07:48 crc kubenswrapper[4846]: I1005 07:07:48.295401 4846 generic.go:334] "Generic (PLEG): container finished" podID="f7172e25-4d68-4c6f-ac5a-9721a3446a2c" containerID="dfb681f3095dc72a51239d91581b46c2f72cd3096542ebb005216651080f1d8d" exitCode=0 Oct 05 07:07:48 crc kubenswrapper[4846]: I1005 07:07:48.295440 4846 generic.go:334] "Generic (PLEG): container finished" podID="f7172e25-4d68-4c6f-ac5a-9721a3446a2c" containerID="cb195538f98a0c7c4947d1630fe3c094eeab2ece547636b632401d6c5227b1da" exitCode=143 Oct 05 07:07:48 crc kubenswrapper[4846]: I1005 07:07:48.295472 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"f7172e25-4d68-4c6f-ac5a-9721a3446a2c","Type":"ContainerDied","Data":"dfb681f3095dc72a51239d91581b46c2f72cd3096542ebb005216651080f1d8d"} Oct 05 07:07:48 crc kubenswrapper[4846]: I1005 07:07:48.295548 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"f7172e25-4d68-4c6f-ac5a-9721a3446a2c","Type":"ContainerDied","Data":"cb195538f98a0c7c4947d1630fe3c094eeab2ece547636b632401d6c5227b1da"} Oct 05 07:07:48 crc kubenswrapper[4846]: I1005 07:07:48.295570 4846 scope.go:117] "RemoveContainer" containerID="dfb681f3095dc72a51239d91581b46c2f72cd3096542ebb005216651080f1d8d" Oct 05 07:07:48 crc kubenswrapper[4846]: I1005 07:07:48.295572 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"f7172e25-4d68-4c6f-ac5a-9721a3446a2c","Type":"ContainerDied","Data":"0120b10514b84b8b6d4cd962e7efaaa9c4b9bede73ad7a9d9c8ff9f21143a8c3"} Oct 05 07:07:48 crc kubenswrapper[4846]: I1005 07:07:48.295547 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 05 07:07:48 crc kubenswrapper[4846]: I1005 07:07:48.304023 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"47ea4b78-ed6a-48c6-a7d2-0505378a874d","Type":"ContainerStarted","Data":"2fa57817123fca2619cf1324f4ec1232fd6cbdee734ffdcbdc7b17945b49d572"} Oct 05 07:07:48 crc kubenswrapper[4846]: I1005 07:07:48.305847 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-65884db4f4-fsk6p" event={"ID":"c35affe0-da4e-448f-aac0-7bba26f89040","Type":"ContainerStarted","Data":"ca8f38824786960e56cd2679a8818d0dc234675708e9dcdb847dd2e8bbbdc666"} Oct 05 07:07:48 crc kubenswrapper[4846]: I1005 07:07:48.310007 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7cf74955b9-njz69" event={"ID":"5514c26f-0fe1-4e09-9292-896f21f90f71","Type":"ContainerDied","Data":"6b016854d4d4884a31e2cc57ac01c8207df5d758726b4508fa4bc238478383d3"} Oct 05 07:07:48 crc kubenswrapper[4846]: I1005 07:07:48.310149 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7cf74955b9-njz69" Oct 05 07:07:48 crc kubenswrapper[4846]: I1005 07:07:48.316900 4846 reconciler_common.go:293] "Volume detached for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" DevicePath \"\"" Oct 05 07:07:48 crc kubenswrapper[4846]: I1005 07:07:48.317951 4846 generic.go:334] "Generic (PLEG): container finished" podID="ae01bc6c-75fb-4958-9a75-6c212fd0ed3a" containerID="e6f0c66814a012bcdf9a55c51c7fd4fb06ab5d3c0cf88579133f5a05220e6214" exitCode=0 Oct 05 07:07:48 crc kubenswrapper[4846]: I1005 07:07:48.318658 4846 generic.go:334] "Generic (PLEG): container finished" podID="ae01bc6c-75fb-4958-9a75-6c212fd0ed3a" containerID="cd7a0167e5022ef22dc459912b29a30979c920a9d7327e00b83da07039436faa" exitCode=2 Oct 05 07:07:48 crc kubenswrapper[4846]: I1005 07:07:48.318825 4846 generic.go:334] "Generic (PLEG): container finished" podID="ae01bc6c-75fb-4958-9a75-6c212fd0ed3a" containerID="bedf3eddca9636e7febabe72b13554c3e7139ba3edbeca3aa33a3865508515eb" exitCode=0 Oct 05 07:07:48 crc kubenswrapper[4846]: I1005 07:07:48.319062 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ae01bc6c-75fb-4958-9a75-6c212fd0ed3a","Type":"ContainerDied","Data":"e6f0c66814a012bcdf9a55c51c7fd4fb06ab5d3c0cf88579133f5a05220e6214"} Oct 05 07:07:48 crc kubenswrapper[4846]: I1005 07:07:48.319109 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ae01bc6c-75fb-4958-9a75-6c212fd0ed3a","Type":"ContainerDied","Data":"cd7a0167e5022ef22dc459912b29a30979c920a9d7327e00b83da07039436faa"} Oct 05 07:07:48 crc kubenswrapper[4846]: I1005 07:07:48.319123 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ae01bc6c-75fb-4958-9a75-6c212fd0ed3a","Type":"ContainerDied","Data":"bedf3eddca9636e7febabe72b13554c3e7139ba3edbeca3aa33a3865508515eb"} Oct 05 07:07:48 crc kubenswrapper[4846]: I1005 07:07:48.348830 4846 scope.go:117] "RemoveContainer" containerID="cb195538f98a0c7c4947d1630fe3c094eeab2ece547636b632401d6c5227b1da" Oct 05 07:07:48 crc kubenswrapper[4846]: I1005 07:07:48.354411 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 05 07:07:48 crc kubenswrapper[4846]: I1005 07:07:48.400299 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 05 07:07:48 crc kubenswrapper[4846]: I1005 07:07:48.409805 4846 scope.go:117] "RemoveContainer" containerID="dfb681f3095dc72a51239d91581b46c2f72cd3096542ebb005216651080f1d8d" Oct 05 07:07:48 crc kubenswrapper[4846]: I1005 07:07:48.413698 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 05 07:07:48 crc kubenswrapper[4846]: E1005 07:07:48.414142 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5514c26f-0fe1-4e09-9292-896f21f90f71" containerName="init" Oct 05 07:07:48 crc kubenswrapper[4846]: I1005 07:07:48.414157 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="5514c26f-0fe1-4e09-9292-896f21f90f71" containerName="init" Oct 05 07:07:48 crc kubenswrapper[4846]: E1005 07:07:48.414196 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f7172e25-4d68-4c6f-ac5a-9721a3446a2c" containerName="glance-httpd" Oct 05 07:07:48 crc kubenswrapper[4846]: I1005 07:07:48.414205 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="f7172e25-4d68-4c6f-ac5a-9721a3446a2c" containerName="glance-httpd" Oct 05 07:07:48 crc kubenswrapper[4846]: E1005 07:07:48.414217 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f7172e25-4d68-4c6f-ac5a-9721a3446a2c" containerName="glance-log" Oct 05 07:07:48 crc kubenswrapper[4846]: I1005 07:07:48.414223 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="f7172e25-4d68-4c6f-ac5a-9721a3446a2c" containerName="glance-log" Oct 05 07:07:48 crc kubenswrapper[4846]: E1005 07:07:48.414236 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5514c26f-0fe1-4e09-9292-896f21f90f71" containerName="dnsmasq-dns" Oct 05 07:07:48 crc kubenswrapper[4846]: I1005 07:07:48.414242 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="5514c26f-0fe1-4e09-9292-896f21f90f71" containerName="dnsmasq-dns" Oct 05 07:07:48 crc kubenswrapper[4846]: I1005 07:07:48.414385 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="f7172e25-4d68-4c6f-ac5a-9721a3446a2c" containerName="glance-log" Oct 05 07:07:48 crc kubenswrapper[4846]: I1005 07:07:48.414398 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="f7172e25-4d68-4c6f-ac5a-9721a3446a2c" containerName="glance-httpd" Oct 05 07:07:48 crc kubenswrapper[4846]: I1005 07:07:48.414413 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="5514c26f-0fe1-4e09-9292-896f21f90f71" containerName="dnsmasq-dns" Oct 05 07:07:48 crc kubenswrapper[4846]: I1005 07:07:48.415460 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 05 07:07:48 crc kubenswrapper[4846]: E1005 07:07:48.433511 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dfb681f3095dc72a51239d91581b46c2f72cd3096542ebb005216651080f1d8d\": container with ID starting with dfb681f3095dc72a51239d91581b46c2f72cd3096542ebb005216651080f1d8d not found: ID does not exist" containerID="dfb681f3095dc72a51239d91581b46c2f72cd3096542ebb005216651080f1d8d" Oct 05 07:07:48 crc kubenswrapper[4846]: I1005 07:07:48.433625 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dfb681f3095dc72a51239d91581b46c2f72cd3096542ebb005216651080f1d8d"} err="failed to get container status \"dfb681f3095dc72a51239d91581b46c2f72cd3096542ebb005216651080f1d8d\": rpc error: code = NotFound desc = could not find container \"dfb681f3095dc72a51239d91581b46c2f72cd3096542ebb005216651080f1d8d\": container with ID starting with dfb681f3095dc72a51239d91581b46c2f72cd3096542ebb005216651080f1d8d not found: ID does not exist" Oct 05 07:07:48 crc kubenswrapper[4846]: I1005 07:07:48.433669 4846 scope.go:117] "RemoveContainer" containerID="cb195538f98a0c7c4947d1630fe3c094eeab2ece547636b632401d6c5227b1da" Oct 05 07:07:48 crc kubenswrapper[4846]: I1005 07:07:48.434129 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Oct 05 07:07:48 crc kubenswrapper[4846]: I1005 07:07:48.434530 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Oct 05 07:07:48 crc kubenswrapper[4846]: E1005 07:07:48.434600 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cb195538f98a0c7c4947d1630fe3c094eeab2ece547636b632401d6c5227b1da\": container with ID starting with cb195538f98a0c7c4947d1630fe3c094eeab2ece547636b632401d6c5227b1da not found: ID does not exist" containerID="cb195538f98a0c7c4947d1630fe3c094eeab2ece547636b632401d6c5227b1da" Oct 05 07:07:48 crc kubenswrapper[4846]: I1005 07:07:48.434713 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cb195538f98a0c7c4947d1630fe3c094eeab2ece547636b632401d6c5227b1da"} err="failed to get container status \"cb195538f98a0c7c4947d1630fe3c094eeab2ece547636b632401d6c5227b1da\": rpc error: code = NotFound desc = could not find container \"cb195538f98a0c7c4947d1630fe3c094eeab2ece547636b632401d6c5227b1da\": container with ID starting with cb195538f98a0c7c4947d1630fe3c094eeab2ece547636b632401d6c5227b1da not found: ID does not exist" Oct 05 07:07:48 crc kubenswrapper[4846]: I1005 07:07:48.434759 4846 scope.go:117] "RemoveContainer" containerID="dfb681f3095dc72a51239d91581b46c2f72cd3096542ebb005216651080f1d8d" Oct 05 07:07:48 crc kubenswrapper[4846]: I1005 07:07:48.444000 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dfb681f3095dc72a51239d91581b46c2f72cd3096542ebb005216651080f1d8d"} err="failed to get container status \"dfb681f3095dc72a51239d91581b46c2f72cd3096542ebb005216651080f1d8d\": rpc error: code = NotFound desc = could not find container \"dfb681f3095dc72a51239d91581b46c2f72cd3096542ebb005216651080f1d8d\": container with ID starting with dfb681f3095dc72a51239d91581b46c2f72cd3096542ebb005216651080f1d8d not found: ID does not exist" Oct 05 07:07:48 crc kubenswrapper[4846]: I1005 07:07:48.444054 4846 scope.go:117] "RemoveContainer" containerID="cb195538f98a0c7c4947d1630fe3c094eeab2ece547636b632401d6c5227b1da" Oct 05 07:07:48 crc kubenswrapper[4846]: I1005 07:07:48.444661 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cb195538f98a0c7c4947d1630fe3c094eeab2ece547636b632401d6c5227b1da"} err="failed to get container status \"cb195538f98a0c7c4947d1630fe3c094eeab2ece547636b632401d6c5227b1da\": rpc error: code = NotFound desc = could not find container \"cb195538f98a0c7c4947d1630fe3c094eeab2ece547636b632401d6c5227b1da\": container with ID starting with cb195538f98a0c7c4947d1630fe3c094eeab2ece547636b632401d6c5227b1da not found: ID does not exist" Oct 05 07:07:48 crc kubenswrapper[4846]: I1005 07:07:48.444712 4846 scope.go:117] "RemoveContainer" containerID="1d7f5cfaee91723a32aaf61fa9be5825f7c8e5a2573f16e15c12504bbef939f2" Oct 05 07:07:48 crc kubenswrapper[4846]: I1005 07:07:48.466688 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7cf74955b9-njz69"] Oct 05 07:07:48 crc kubenswrapper[4846]: I1005 07:07:48.483659 4846 scope.go:117] "RemoveContainer" containerID="150067bb99bb3cea426705b9e67dd71568c5d12ab03723bd75aa455857688ad6" Oct 05 07:07:48 crc kubenswrapper[4846]: I1005 07:07:48.491334 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7cf74955b9-njz69"] Oct 05 07:07:48 crc kubenswrapper[4846]: I1005 07:07:48.520425 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/edd6502e-6e77-4ad3-8f1f-b85426188261-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"edd6502e-6e77-4ad3-8f1f-b85426188261\") " pod="openstack/glance-default-internal-api-0" Oct 05 07:07:48 crc kubenswrapper[4846]: I1005 07:07:48.520486 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/edd6502e-6e77-4ad3-8f1f-b85426188261-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"edd6502e-6e77-4ad3-8f1f-b85426188261\") " pod="openstack/glance-default-internal-api-0" Oct 05 07:07:48 crc kubenswrapper[4846]: I1005 07:07:48.520510 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"edd6502e-6e77-4ad3-8f1f-b85426188261\") " pod="openstack/glance-default-internal-api-0" Oct 05 07:07:48 crc kubenswrapper[4846]: I1005 07:07:48.520577 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/edd6502e-6e77-4ad3-8f1f-b85426188261-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"edd6502e-6e77-4ad3-8f1f-b85426188261\") " pod="openstack/glance-default-internal-api-0" Oct 05 07:07:48 crc kubenswrapper[4846]: I1005 07:07:48.520611 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/edd6502e-6e77-4ad3-8f1f-b85426188261-logs\") pod \"glance-default-internal-api-0\" (UID: \"edd6502e-6e77-4ad3-8f1f-b85426188261\") " pod="openstack/glance-default-internal-api-0" Oct 05 07:07:48 crc kubenswrapper[4846]: I1005 07:07:48.520666 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/edd6502e-6e77-4ad3-8f1f-b85426188261-scripts\") pod \"glance-default-internal-api-0\" (UID: \"edd6502e-6e77-4ad3-8f1f-b85426188261\") " pod="openstack/glance-default-internal-api-0" Oct 05 07:07:48 crc kubenswrapper[4846]: I1005 07:07:48.520708 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9xgjg\" (UniqueName: \"kubernetes.io/projected/edd6502e-6e77-4ad3-8f1f-b85426188261-kube-api-access-9xgjg\") pod \"glance-default-internal-api-0\" (UID: \"edd6502e-6e77-4ad3-8f1f-b85426188261\") " pod="openstack/glance-default-internal-api-0" Oct 05 07:07:48 crc kubenswrapper[4846]: I1005 07:07:48.520742 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/edd6502e-6e77-4ad3-8f1f-b85426188261-config-data\") pod \"glance-default-internal-api-0\" (UID: \"edd6502e-6e77-4ad3-8f1f-b85426188261\") " pod="openstack/glance-default-internal-api-0" Oct 05 07:07:48 crc kubenswrapper[4846]: I1005 07:07:48.531872 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="345b75dc-52c2-4a1b-a5d6-19235d77f106" path="/var/lib/kubelet/pods/345b75dc-52c2-4a1b-a5d6-19235d77f106/volumes" Oct 05 07:07:48 crc kubenswrapper[4846]: I1005 07:07:48.532810 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5514c26f-0fe1-4e09-9292-896f21f90f71" path="/var/lib/kubelet/pods/5514c26f-0fe1-4e09-9292-896f21f90f71/volumes" Oct 05 07:07:48 crc kubenswrapper[4846]: I1005 07:07:48.537401 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f7172e25-4d68-4c6f-ac5a-9721a3446a2c" path="/var/lib/kubelet/pods/f7172e25-4d68-4c6f-ac5a-9721a3446a2c/volumes" Oct 05 07:07:48 crc kubenswrapper[4846]: I1005 07:07:48.539136 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 05 07:07:48 crc kubenswrapper[4846]: I1005 07:07:48.623000 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/edd6502e-6e77-4ad3-8f1f-b85426188261-config-data\") pod \"glance-default-internal-api-0\" (UID: \"edd6502e-6e77-4ad3-8f1f-b85426188261\") " pod="openstack/glance-default-internal-api-0" Oct 05 07:07:48 crc kubenswrapper[4846]: I1005 07:07:48.623063 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/edd6502e-6e77-4ad3-8f1f-b85426188261-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"edd6502e-6e77-4ad3-8f1f-b85426188261\") " pod="openstack/glance-default-internal-api-0" Oct 05 07:07:48 crc kubenswrapper[4846]: I1005 07:07:48.623099 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/edd6502e-6e77-4ad3-8f1f-b85426188261-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"edd6502e-6e77-4ad3-8f1f-b85426188261\") " pod="openstack/glance-default-internal-api-0" Oct 05 07:07:48 crc kubenswrapper[4846]: I1005 07:07:48.623124 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"edd6502e-6e77-4ad3-8f1f-b85426188261\") " pod="openstack/glance-default-internal-api-0" Oct 05 07:07:48 crc kubenswrapper[4846]: I1005 07:07:48.623218 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/edd6502e-6e77-4ad3-8f1f-b85426188261-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"edd6502e-6e77-4ad3-8f1f-b85426188261\") " pod="openstack/glance-default-internal-api-0" Oct 05 07:07:48 crc kubenswrapper[4846]: I1005 07:07:48.623250 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/edd6502e-6e77-4ad3-8f1f-b85426188261-logs\") pod \"glance-default-internal-api-0\" (UID: \"edd6502e-6e77-4ad3-8f1f-b85426188261\") " pod="openstack/glance-default-internal-api-0" Oct 05 07:07:48 crc kubenswrapper[4846]: I1005 07:07:48.623329 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/edd6502e-6e77-4ad3-8f1f-b85426188261-scripts\") pod \"glance-default-internal-api-0\" (UID: \"edd6502e-6e77-4ad3-8f1f-b85426188261\") " pod="openstack/glance-default-internal-api-0" Oct 05 07:07:48 crc kubenswrapper[4846]: I1005 07:07:48.623387 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9xgjg\" (UniqueName: \"kubernetes.io/projected/edd6502e-6e77-4ad3-8f1f-b85426188261-kube-api-access-9xgjg\") pod \"glance-default-internal-api-0\" (UID: \"edd6502e-6e77-4ad3-8f1f-b85426188261\") " pod="openstack/glance-default-internal-api-0" Oct 05 07:07:48 crc kubenswrapper[4846]: I1005 07:07:48.623654 4846 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"edd6502e-6e77-4ad3-8f1f-b85426188261\") device mount path \"/mnt/openstack/pv12\"" pod="openstack/glance-default-internal-api-0" Oct 05 07:07:48 crc kubenswrapper[4846]: I1005 07:07:48.623822 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/edd6502e-6e77-4ad3-8f1f-b85426188261-logs\") pod \"glance-default-internal-api-0\" (UID: \"edd6502e-6e77-4ad3-8f1f-b85426188261\") " pod="openstack/glance-default-internal-api-0" Oct 05 07:07:48 crc kubenswrapper[4846]: I1005 07:07:48.624105 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/edd6502e-6e77-4ad3-8f1f-b85426188261-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"edd6502e-6e77-4ad3-8f1f-b85426188261\") " pod="openstack/glance-default-internal-api-0" Oct 05 07:07:48 crc kubenswrapper[4846]: I1005 07:07:48.629284 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/edd6502e-6e77-4ad3-8f1f-b85426188261-config-data\") pod \"glance-default-internal-api-0\" (UID: \"edd6502e-6e77-4ad3-8f1f-b85426188261\") " pod="openstack/glance-default-internal-api-0" Oct 05 07:07:48 crc kubenswrapper[4846]: I1005 07:07:48.629299 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/edd6502e-6e77-4ad3-8f1f-b85426188261-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"edd6502e-6e77-4ad3-8f1f-b85426188261\") " pod="openstack/glance-default-internal-api-0" Oct 05 07:07:48 crc kubenswrapper[4846]: I1005 07:07:48.629425 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/edd6502e-6e77-4ad3-8f1f-b85426188261-scripts\") pod \"glance-default-internal-api-0\" (UID: \"edd6502e-6e77-4ad3-8f1f-b85426188261\") " pod="openstack/glance-default-internal-api-0" Oct 05 07:07:48 crc kubenswrapper[4846]: I1005 07:07:48.633780 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/edd6502e-6e77-4ad3-8f1f-b85426188261-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"edd6502e-6e77-4ad3-8f1f-b85426188261\") " pod="openstack/glance-default-internal-api-0" Oct 05 07:07:48 crc kubenswrapper[4846]: I1005 07:07:48.646647 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9xgjg\" (UniqueName: \"kubernetes.io/projected/edd6502e-6e77-4ad3-8f1f-b85426188261-kube-api-access-9xgjg\") pod \"glance-default-internal-api-0\" (UID: \"edd6502e-6e77-4ad3-8f1f-b85426188261\") " pod="openstack/glance-default-internal-api-0" Oct 05 07:07:48 crc kubenswrapper[4846]: I1005 07:07:48.664120 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"edd6502e-6e77-4ad3-8f1f-b85426188261\") " pod="openstack/glance-default-internal-api-0" Oct 05 07:07:48 crc kubenswrapper[4846]: I1005 07:07:48.785740 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 05 07:07:49 crc kubenswrapper[4846]: I1005 07:07:49.335174 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-65884db4f4-fsk6p" event={"ID":"c35affe0-da4e-448f-aac0-7bba26f89040","Type":"ContainerStarted","Data":"b48698aea06e5e06f2a72f7568ee41c7c5c3fee9c4bdab0f9cb7cebb4f50ea03"} Oct 05 07:07:49 crc kubenswrapper[4846]: I1005 07:07:49.336065 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-65884db4f4-fsk6p" Oct 05 07:07:49 crc kubenswrapper[4846]: I1005 07:07:49.336082 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-65884db4f4-fsk6p" Oct 05 07:07:49 crc kubenswrapper[4846]: I1005 07:07:49.336092 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-65884db4f4-fsk6p" event={"ID":"c35affe0-da4e-448f-aac0-7bba26f89040","Type":"ContainerStarted","Data":"d0a2d9595d1231b4e28c0eff4482c56b148f02336c7fb50f8a4400e74af973ff"} Oct 05 07:07:49 crc kubenswrapper[4846]: I1005 07:07:49.343959 4846 generic.go:334] "Generic (PLEG): container finished" podID="b7a58faa-0f37-4b0e-a907-6a8bae28cb27" containerID="27a50aa96a7a8bb54961eddcd8037b7777a781e62487ae4213c4c228650c5b06" exitCode=0 Oct 05 07:07:49 crc kubenswrapper[4846]: I1005 07:07:49.344028 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-t5brq" event={"ID":"b7a58faa-0f37-4b0e-a907-6a8bae28cb27","Type":"ContainerDied","Data":"27a50aa96a7a8bb54961eddcd8037b7777a781e62487ae4213c4c228650c5b06"} Oct 05 07:07:49 crc kubenswrapper[4846]: I1005 07:07:49.347451 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"47ea4b78-ed6a-48c6-a7d2-0505378a874d","Type":"ContainerStarted","Data":"3747a48030345be1e2d10045a277345f99cf674c32e826c35d08619673de537e"} Oct 05 07:07:49 crc kubenswrapper[4846]: I1005 07:07:49.347489 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"47ea4b78-ed6a-48c6-a7d2-0505378a874d","Type":"ContainerStarted","Data":"4d44457e6e512ba342767ec6d38d8709da21ae7e09f8643e92055bacd72c34da"} Oct 05 07:07:49 crc kubenswrapper[4846]: I1005 07:07:49.362517 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-65884db4f4-fsk6p" podStartSLOduration=3.362493793 podStartE2EDuration="3.362493793s" podCreationTimestamp="2025-10-05 07:07:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 07:07:49.357910041 +0000 UTC m=+1191.598762816" watchObservedRunningTime="2025-10-05 07:07:49.362493793 +0000 UTC m=+1191.603346568" Oct 05 07:07:49 crc kubenswrapper[4846]: I1005 07:07:49.416348 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=3.416325369 podStartE2EDuration="3.416325369s" podCreationTimestamp="2025-10-05 07:07:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 07:07:49.408585243 +0000 UTC m=+1191.649438018" watchObservedRunningTime="2025-10-05 07:07:49.416325369 +0000 UTC m=+1191.657178144" Oct 05 07:07:49 crc kubenswrapper[4846]: I1005 07:07:49.440901 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 05 07:07:49 crc kubenswrapper[4846]: I1005 07:07:49.917499 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 05 07:07:49 crc kubenswrapper[4846]: I1005 07:07:49.971969 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ae01bc6c-75fb-4958-9a75-6c212fd0ed3a-scripts\") pod \"ae01bc6c-75fb-4958-9a75-6c212fd0ed3a\" (UID: \"ae01bc6c-75fb-4958-9a75-6c212fd0ed3a\") " Oct 05 07:07:49 crc kubenswrapper[4846]: I1005 07:07:49.972042 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ae01bc6c-75fb-4958-9a75-6c212fd0ed3a-log-httpd\") pod \"ae01bc6c-75fb-4958-9a75-6c212fd0ed3a\" (UID: \"ae01bc6c-75fb-4958-9a75-6c212fd0ed3a\") " Oct 05 07:07:49 crc kubenswrapper[4846]: I1005 07:07:49.972203 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ae01bc6c-75fb-4958-9a75-6c212fd0ed3a-sg-core-conf-yaml\") pod \"ae01bc6c-75fb-4958-9a75-6c212fd0ed3a\" (UID: \"ae01bc6c-75fb-4958-9a75-6c212fd0ed3a\") " Oct 05 07:07:49 crc kubenswrapper[4846]: I1005 07:07:49.972281 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vmgr5\" (UniqueName: \"kubernetes.io/projected/ae01bc6c-75fb-4958-9a75-6c212fd0ed3a-kube-api-access-vmgr5\") pod \"ae01bc6c-75fb-4958-9a75-6c212fd0ed3a\" (UID: \"ae01bc6c-75fb-4958-9a75-6c212fd0ed3a\") " Oct 05 07:07:49 crc kubenswrapper[4846]: I1005 07:07:49.972462 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ae01bc6c-75fb-4958-9a75-6c212fd0ed3a-run-httpd\") pod \"ae01bc6c-75fb-4958-9a75-6c212fd0ed3a\" (UID: \"ae01bc6c-75fb-4958-9a75-6c212fd0ed3a\") " Oct 05 07:07:49 crc kubenswrapper[4846]: I1005 07:07:49.972530 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ae01bc6c-75fb-4958-9a75-6c212fd0ed3a-combined-ca-bundle\") pod \"ae01bc6c-75fb-4958-9a75-6c212fd0ed3a\" (UID: \"ae01bc6c-75fb-4958-9a75-6c212fd0ed3a\") " Oct 05 07:07:49 crc kubenswrapper[4846]: I1005 07:07:49.972604 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ae01bc6c-75fb-4958-9a75-6c212fd0ed3a-config-data\") pod \"ae01bc6c-75fb-4958-9a75-6c212fd0ed3a\" (UID: \"ae01bc6c-75fb-4958-9a75-6c212fd0ed3a\") " Oct 05 07:07:49 crc kubenswrapper[4846]: I1005 07:07:49.973773 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ae01bc6c-75fb-4958-9a75-6c212fd0ed3a-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "ae01bc6c-75fb-4958-9a75-6c212fd0ed3a" (UID: "ae01bc6c-75fb-4958-9a75-6c212fd0ed3a"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 07:07:49 crc kubenswrapper[4846]: I1005 07:07:49.979149 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ae01bc6c-75fb-4958-9a75-6c212fd0ed3a-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "ae01bc6c-75fb-4958-9a75-6c212fd0ed3a" (UID: "ae01bc6c-75fb-4958-9a75-6c212fd0ed3a"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 07:07:49 crc kubenswrapper[4846]: I1005 07:07:49.988235 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ae01bc6c-75fb-4958-9a75-6c212fd0ed3a-scripts" (OuterVolumeSpecName: "scripts") pod "ae01bc6c-75fb-4958-9a75-6c212fd0ed3a" (UID: "ae01bc6c-75fb-4958-9a75-6c212fd0ed3a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:07:49 crc kubenswrapper[4846]: I1005 07:07:49.988243 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ae01bc6c-75fb-4958-9a75-6c212fd0ed3a-kube-api-access-vmgr5" (OuterVolumeSpecName: "kube-api-access-vmgr5") pod "ae01bc6c-75fb-4958-9a75-6c212fd0ed3a" (UID: "ae01bc6c-75fb-4958-9a75-6c212fd0ed3a"). InnerVolumeSpecName "kube-api-access-vmgr5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:07:50 crc kubenswrapper[4846]: I1005 07:07:50.016036 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ae01bc6c-75fb-4958-9a75-6c212fd0ed3a-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "ae01bc6c-75fb-4958-9a75-6c212fd0ed3a" (UID: "ae01bc6c-75fb-4958-9a75-6c212fd0ed3a"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:07:50 crc kubenswrapper[4846]: I1005 07:07:50.075293 4846 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ae01bc6c-75fb-4958-9a75-6c212fd0ed3a-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 05 07:07:50 crc kubenswrapper[4846]: I1005 07:07:50.075331 4846 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ae01bc6c-75fb-4958-9a75-6c212fd0ed3a-scripts\") on node \"crc\" DevicePath \"\"" Oct 05 07:07:50 crc kubenswrapper[4846]: I1005 07:07:50.075345 4846 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ae01bc6c-75fb-4958-9a75-6c212fd0ed3a-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 05 07:07:50 crc kubenswrapper[4846]: I1005 07:07:50.075357 4846 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ae01bc6c-75fb-4958-9a75-6c212fd0ed3a-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 05 07:07:50 crc kubenswrapper[4846]: I1005 07:07:50.075372 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vmgr5\" (UniqueName: \"kubernetes.io/projected/ae01bc6c-75fb-4958-9a75-6c212fd0ed3a-kube-api-access-vmgr5\") on node \"crc\" DevicePath \"\"" Oct 05 07:07:50 crc kubenswrapper[4846]: I1005 07:07:50.105535 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ae01bc6c-75fb-4958-9a75-6c212fd0ed3a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ae01bc6c-75fb-4958-9a75-6c212fd0ed3a" (UID: "ae01bc6c-75fb-4958-9a75-6c212fd0ed3a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:07:50 crc kubenswrapper[4846]: I1005 07:07:50.151818 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ae01bc6c-75fb-4958-9a75-6c212fd0ed3a-config-data" (OuterVolumeSpecName: "config-data") pod "ae01bc6c-75fb-4958-9a75-6c212fd0ed3a" (UID: "ae01bc6c-75fb-4958-9a75-6c212fd0ed3a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:07:50 crc kubenswrapper[4846]: I1005 07:07:50.177968 4846 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ae01bc6c-75fb-4958-9a75-6c212fd0ed3a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 07:07:50 crc kubenswrapper[4846]: I1005 07:07:50.178024 4846 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ae01bc6c-75fb-4958-9a75-6c212fd0ed3a-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 07:07:50 crc kubenswrapper[4846]: I1005 07:07:50.374077 4846 generic.go:334] "Generic (PLEG): container finished" podID="ae01bc6c-75fb-4958-9a75-6c212fd0ed3a" containerID="f80062678cc982eb6612835cc64c1c9757ef8fbe87a8c39fe4a660630c89a786" exitCode=0 Oct 05 07:07:50 crc kubenswrapper[4846]: I1005 07:07:50.374170 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ae01bc6c-75fb-4958-9a75-6c212fd0ed3a","Type":"ContainerDied","Data":"f80062678cc982eb6612835cc64c1c9757ef8fbe87a8c39fe4a660630c89a786"} Oct 05 07:07:50 crc kubenswrapper[4846]: I1005 07:07:50.374228 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ae01bc6c-75fb-4958-9a75-6c212fd0ed3a","Type":"ContainerDied","Data":"2545b9452eaf83986bb5d51763da560439fe6f42fda5b657635a72196e0d38e2"} Oct 05 07:07:50 crc kubenswrapper[4846]: I1005 07:07:50.374253 4846 scope.go:117] "RemoveContainer" containerID="e6f0c66814a012bcdf9a55c51c7fd4fb06ab5d3c0cf88579133f5a05220e6214" Oct 05 07:07:50 crc kubenswrapper[4846]: I1005 07:07:50.374449 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 05 07:07:50 crc kubenswrapper[4846]: I1005 07:07:50.379629 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"edd6502e-6e77-4ad3-8f1f-b85426188261","Type":"ContainerStarted","Data":"8aa9405d7a8ddd77ef09fe71dd0b5f1c398759b0a55d1ac4fba47c2b46ef37cc"} Oct 05 07:07:50 crc kubenswrapper[4846]: I1005 07:07:50.379695 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"edd6502e-6e77-4ad3-8f1f-b85426188261","Type":"ContainerStarted","Data":"0d610f817a3051e49134c8aafcdac4fd03f563999d934bd1a6787a8227d0042c"} Oct 05 07:07:50 crc kubenswrapper[4846]: I1005 07:07:50.450001 4846 scope.go:117] "RemoveContainer" containerID="cd7a0167e5022ef22dc459912b29a30979c920a9d7327e00b83da07039436faa" Oct 05 07:07:50 crc kubenswrapper[4846]: I1005 07:07:50.467904 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 05 07:07:50 crc kubenswrapper[4846]: I1005 07:07:50.485804 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 05 07:07:50 crc kubenswrapper[4846]: I1005 07:07:50.496354 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 05 07:07:50 crc kubenswrapper[4846]: E1005 07:07:50.496980 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ae01bc6c-75fb-4958-9a75-6c212fd0ed3a" containerName="ceilometer-notification-agent" Oct 05 07:07:50 crc kubenswrapper[4846]: I1005 07:07:50.497115 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="ae01bc6c-75fb-4958-9a75-6c212fd0ed3a" containerName="ceilometer-notification-agent" Oct 05 07:07:50 crc kubenswrapper[4846]: E1005 07:07:50.497136 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ae01bc6c-75fb-4958-9a75-6c212fd0ed3a" containerName="sg-core" Oct 05 07:07:50 crc kubenswrapper[4846]: I1005 07:07:50.497146 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="ae01bc6c-75fb-4958-9a75-6c212fd0ed3a" containerName="sg-core" Oct 05 07:07:50 crc kubenswrapper[4846]: E1005 07:07:50.497250 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ae01bc6c-75fb-4958-9a75-6c212fd0ed3a" containerName="proxy-httpd" Oct 05 07:07:50 crc kubenswrapper[4846]: I1005 07:07:50.497262 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="ae01bc6c-75fb-4958-9a75-6c212fd0ed3a" containerName="proxy-httpd" Oct 05 07:07:50 crc kubenswrapper[4846]: E1005 07:07:50.497304 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ae01bc6c-75fb-4958-9a75-6c212fd0ed3a" containerName="ceilometer-central-agent" Oct 05 07:07:50 crc kubenswrapper[4846]: I1005 07:07:50.497313 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="ae01bc6c-75fb-4958-9a75-6c212fd0ed3a" containerName="ceilometer-central-agent" Oct 05 07:07:50 crc kubenswrapper[4846]: I1005 07:07:50.497540 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="ae01bc6c-75fb-4958-9a75-6c212fd0ed3a" containerName="ceilometer-central-agent" Oct 05 07:07:50 crc kubenswrapper[4846]: I1005 07:07:50.497562 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="ae01bc6c-75fb-4958-9a75-6c212fd0ed3a" containerName="ceilometer-notification-agent" Oct 05 07:07:50 crc kubenswrapper[4846]: I1005 07:07:50.497576 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="ae01bc6c-75fb-4958-9a75-6c212fd0ed3a" containerName="sg-core" Oct 05 07:07:50 crc kubenswrapper[4846]: I1005 07:07:50.497602 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="ae01bc6c-75fb-4958-9a75-6c212fd0ed3a" containerName="proxy-httpd" Oct 05 07:07:50 crc kubenswrapper[4846]: I1005 07:07:50.506278 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 05 07:07:50 crc kubenswrapper[4846]: I1005 07:07:50.510234 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 05 07:07:50 crc kubenswrapper[4846]: I1005 07:07:50.510520 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 05 07:07:50 crc kubenswrapper[4846]: I1005 07:07:50.528304 4846 scope.go:117] "RemoveContainer" containerID="f80062678cc982eb6612835cc64c1c9757ef8fbe87a8c39fe4a660630c89a786" Oct 05 07:07:50 crc kubenswrapper[4846]: I1005 07:07:50.536500 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ae01bc6c-75fb-4958-9a75-6c212fd0ed3a" path="/var/lib/kubelet/pods/ae01bc6c-75fb-4958-9a75-6c212fd0ed3a/volumes" Oct 05 07:07:50 crc kubenswrapper[4846]: I1005 07:07:50.540736 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 05 07:07:50 crc kubenswrapper[4846]: I1005 07:07:50.573255 4846 scope.go:117] "RemoveContainer" containerID="bedf3eddca9636e7febabe72b13554c3e7139ba3edbeca3aa33a3865508515eb" Oct 05 07:07:50 crc kubenswrapper[4846]: I1005 07:07:50.586280 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/39071bf4-214b-4bbb-8b78-dfec6ad3c940-scripts\") pod \"ceilometer-0\" (UID: \"39071bf4-214b-4bbb-8b78-dfec6ad3c940\") " pod="openstack/ceilometer-0" Oct 05 07:07:50 crc kubenswrapper[4846]: I1005 07:07:50.586332 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bcr96\" (UniqueName: \"kubernetes.io/projected/39071bf4-214b-4bbb-8b78-dfec6ad3c940-kube-api-access-bcr96\") pod \"ceilometer-0\" (UID: \"39071bf4-214b-4bbb-8b78-dfec6ad3c940\") " pod="openstack/ceilometer-0" Oct 05 07:07:50 crc kubenswrapper[4846]: I1005 07:07:50.586788 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/39071bf4-214b-4bbb-8b78-dfec6ad3c940-log-httpd\") pod \"ceilometer-0\" (UID: \"39071bf4-214b-4bbb-8b78-dfec6ad3c940\") " pod="openstack/ceilometer-0" Oct 05 07:07:50 crc kubenswrapper[4846]: I1005 07:07:50.587093 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/39071bf4-214b-4bbb-8b78-dfec6ad3c940-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"39071bf4-214b-4bbb-8b78-dfec6ad3c940\") " pod="openstack/ceilometer-0" Oct 05 07:07:50 crc kubenswrapper[4846]: I1005 07:07:50.587262 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/39071bf4-214b-4bbb-8b78-dfec6ad3c940-run-httpd\") pod \"ceilometer-0\" (UID: \"39071bf4-214b-4bbb-8b78-dfec6ad3c940\") " pod="openstack/ceilometer-0" Oct 05 07:07:50 crc kubenswrapper[4846]: I1005 07:07:50.587576 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/39071bf4-214b-4bbb-8b78-dfec6ad3c940-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"39071bf4-214b-4bbb-8b78-dfec6ad3c940\") " pod="openstack/ceilometer-0" Oct 05 07:07:50 crc kubenswrapper[4846]: I1005 07:07:50.587748 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/39071bf4-214b-4bbb-8b78-dfec6ad3c940-config-data\") pod \"ceilometer-0\" (UID: \"39071bf4-214b-4bbb-8b78-dfec6ad3c940\") " pod="openstack/ceilometer-0" Oct 05 07:07:50 crc kubenswrapper[4846]: I1005 07:07:50.685283 4846 scope.go:117] "RemoveContainer" containerID="e6f0c66814a012bcdf9a55c51c7fd4fb06ab5d3c0cf88579133f5a05220e6214" Oct 05 07:07:50 crc kubenswrapper[4846]: E1005 07:07:50.686298 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e6f0c66814a012bcdf9a55c51c7fd4fb06ab5d3c0cf88579133f5a05220e6214\": container with ID starting with e6f0c66814a012bcdf9a55c51c7fd4fb06ab5d3c0cf88579133f5a05220e6214 not found: ID does not exist" containerID="e6f0c66814a012bcdf9a55c51c7fd4fb06ab5d3c0cf88579133f5a05220e6214" Oct 05 07:07:50 crc kubenswrapper[4846]: I1005 07:07:50.686337 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e6f0c66814a012bcdf9a55c51c7fd4fb06ab5d3c0cf88579133f5a05220e6214"} err="failed to get container status \"e6f0c66814a012bcdf9a55c51c7fd4fb06ab5d3c0cf88579133f5a05220e6214\": rpc error: code = NotFound desc = could not find container \"e6f0c66814a012bcdf9a55c51c7fd4fb06ab5d3c0cf88579133f5a05220e6214\": container with ID starting with e6f0c66814a012bcdf9a55c51c7fd4fb06ab5d3c0cf88579133f5a05220e6214 not found: ID does not exist" Oct 05 07:07:50 crc kubenswrapper[4846]: I1005 07:07:50.686362 4846 scope.go:117] "RemoveContainer" containerID="cd7a0167e5022ef22dc459912b29a30979c920a9d7327e00b83da07039436faa" Oct 05 07:07:50 crc kubenswrapper[4846]: E1005 07:07:50.686809 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cd7a0167e5022ef22dc459912b29a30979c920a9d7327e00b83da07039436faa\": container with ID starting with cd7a0167e5022ef22dc459912b29a30979c920a9d7327e00b83da07039436faa not found: ID does not exist" containerID="cd7a0167e5022ef22dc459912b29a30979c920a9d7327e00b83da07039436faa" Oct 05 07:07:50 crc kubenswrapper[4846]: I1005 07:07:50.686856 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cd7a0167e5022ef22dc459912b29a30979c920a9d7327e00b83da07039436faa"} err="failed to get container status \"cd7a0167e5022ef22dc459912b29a30979c920a9d7327e00b83da07039436faa\": rpc error: code = NotFound desc = could not find container \"cd7a0167e5022ef22dc459912b29a30979c920a9d7327e00b83da07039436faa\": container with ID starting with cd7a0167e5022ef22dc459912b29a30979c920a9d7327e00b83da07039436faa not found: ID does not exist" Oct 05 07:07:50 crc kubenswrapper[4846]: I1005 07:07:50.686892 4846 scope.go:117] "RemoveContainer" containerID="f80062678cc982eb6612835cc64c1c9757ef8fbe87a8c39fe4a660630c89a786" Oct 05 07:07:50 crc kubenswrapper[4846]: E1005 07:07:50.687502 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f80062678cc982eb6612835cc64c1c9757ef8fbe87a8c39fe4a660630c89a786\": container with ID starting with f80062678cc982eb6612835cc64c1c9757ef8fbe87a8c39fe4a660630c89a786 not found: ID does not exist" containerID="f80062678cc982eb6612835cc64c1c9757ef8fbe87a8c39fe4a660630c89a786" Oct 05 07:07:50 crc kubenswrapper[4846]: I1005 07:07:50.687532 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f80062678cc982eb6612835cc64c1c9757ef8fbe87a8c39fe4a660630c89a786"} err="failed to get container status \"f80062678cc982eb6612835cc64c1c9757ef8fbe87a8c39fe4a660630c89a786\": rpc error: code = NotFound desc = could not find container \"f80062678cc982eb6612835cc64c1c9757ef8fbe87a8c39fe4a660630c89a786\": container with ID starting with f80062678cc982eb6612835cc64c1c9757ef8fbe87a8c39fe4a660630c89a786 not found: ID does not exist" Oct 05 07:07:50 crc kubenswrapper[4846]: I1005 07:07:50.687548 4846 scope.go:117] "RemoveContainer" containerID="bedf3eddca9636e7febabe72b13554c3e7139ba3edbeca3aa33a3865508515eb" Oct 05 07:07:50 crc kubenswrapper[4846]: E1005 07:07:50.687922 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bedf3eddca9636e7febabe72b13554c3e7139ba3edbeca3aa33a3865508515eb\": container with ID starting with bedf3eddca9636e7febabe72b13554c3e7139ba3edbeca3aa33a3865508515eb not found: ID does not exist" containerID="bedf3eddca9636e7febabe72b13554c3e7139ba3edbeca3aa33a3865508515eb" Oct 05 07:07:50 crc kubenswrapper[4846]: I1005 07:07:50.687943 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bedf3eddca9636e7febabe72b13554c3e7139ba3edbeca3aa33a3865508515eb"} err="failed to get container status \"bedf3eddca9636e7febabe72b13554c3e7139ba3edbeca3aa33a3865508515eb\": rpc error: code = NotFound desc = could not find container \"bedf3eddca9636e7febabe72b13554c3e7139ba3edbeca3aa33a3865508515eb\": container with ID starting with bedf3eddca9636e7febabe72b13554c3e7139ba3edbeca3aa33a3865508515eb not found: ID does not exist" Oct 05 07:07:50 crc kubenswrapper[4846]: I1005 07:07:50.690257 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/39071bf4-214b-4bbb-8b78-dfec6ad3c940-run-httpd\") pod \"ceilometer-0\" (UID: \"39071bf4-214b-4bbb-8b78-dfec6ad3c940\") " pod="openstack/ceilometer-0" Oct 05 07:07:50 crc kubenswrapper[4846]: I1005 07:07:50.690337 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/39071bf4-214b-4bbb-8b78-dfec6ad3c940-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"39071bf4-214b-4bbb-8b78-dfec6ad3c940\") " pod="openstack/ceilometer-0" Oct 05 07:07:50 crc kubenswrapper[4846]: I1005 07:07:50.690374 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/39071bf4-214b-4bbb-8b78-dfec6ad3c940-config-data\") pod \"ceilometer-0\" (UID: \"39071bf4-214b-4bbb-8b78-dfec6ad3c940\") " pod="openstack/ceilometer-0" Oct 05 07:07:50 crc kubenswrapper[4846]: I1005 07:07:50.690400 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/39071bf4-214b-4bbb-8b78-dfec6ad3c940-scripts\") pod \"ceilometer-0\" (UID: \"39071bf4-214b-4bbb-8b78-dfec6ad3c940\") " pod="openstack/ceilometer-0" Oct 05 07:07:50 crc kubenswrapper[4846]: I1005 07:07:50.690417 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bcr96\" (UniqueName: \"kubernetes.io/projected/39071bf4-214b-4bbb-8b78-dfec6ad3c940-kube-api-access-bcr96\") pod \"ceilometer-0\" (UID: \"39071bf4-214b-4bbb-8b78-dfec6ad3c940\") " pod="openstack/ceilometer-0" Oct 05 07:07:50 crc kubenswrapper[4846]: I1005 07:07:50.690472 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/39071bf4-214b-4bbb-8b78-dfec6ad3c940-log-httpd\") pod \"ceilometer-0\" (UID: \"39071bf4-214b-4bbb-8b78-dfec6ad3c940\") " pod="openstack/ceilometer-0" Oct 05 07:07:50 crc kubenswrapper[4846]: I1005 07:07:50.690524 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/39071bf4-214b-4bbb-8b78-dfec6ad3c940-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"39071bf4-214b-4bbb-8b78-dfec6ad3c940\") " pod="openstack/ceilometer-0" Oct 05 07:07:50 crc kubenswrapper[4846]: I1005 07:07:50.690769 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/39071bf4-214b-4bbb-8b78-dfec6ad3c940-run-httpd\") pod \"ceilometer-0\" (UID: \"39071bf4-214b-4bbb-8b78-dfec6ad3c940\") " pod="openstack/ceilometer-0" Oct 05 07:07:50 crc kubenswrapper[4846]: I1005 07:07:50.691419 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/39071bf4-214b-4bbb-8b78-dfec6ad3c940-log-httpd\") pod \"ceilometer-0\" (UID: \"39071bf4-214b-4bbb-8b78-dfec6ad3c940\") " pod="openstack/ceilometer-0" Oct 05 07:07:50 crc kubenswrapper[4846]: I1005 07:07:50.697630 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/39071bf4-214b-4bbb-8b78-dfec6ad3c940-scripts\") pod \"ceilometer-0\" (UID: \"39071bf4-214b-4bbb-8b78-dfec6ad3c940\") " pod="openstack/ceilometer-0" Oct 05 07:07:50 crc kubenswrapper[4846]: I1005 07:07:50.706428 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/39071bf4-214b-4bbb-8b78-dfec6ad3c940-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"39071bf4-214b-4bbb-8b78-dfec6ad3c940\") " pod="openstack/ceilometer-0" Oct 05 07:07:50 crc kubenswrapper[4846]: I1005 07:07:50.707387 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/39071bf4-214b-4bbb-8b78-dfec6ad3c940-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"39071bf4-214b-4bbb-8b78-dfec6ad3c940\") " pod="openstack/ceilometer-0" Oct 05 07:07:50 crc kubenswrapper[4846]: I1005 07:07:50.707720 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/39071bf4-214b-4bbb-8b78-dfec6ad3c940-config-data\") pod \"ceilometer-0\" (UID: \"39071bf4-214b-4bbb-8b78-dfec6ad3c940\") " pod="openstack/ceilometer-0" Oct 05 07:07:50 crc kubenswrapper[4846]: I1005 07:07:50.710298 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bcr96\" (UniqueName: \"kubernetes.io/projected/39071bf4-214b-4bbb-8b78-dfec6ad3c940-kube-api-access-bcr96\") pod \"ceilometer-0\" (UID: \"39071bf4-214b-4bbb-8b78-dfec6ad3c940\") " pod="openstack/ceilometer-0" Oct 05 07:07:50 crc kubenswrapper[4846]: I1005 07:07:50.812999 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-t5brq" Oct 05 07:07:50 crc kubenswrapper[4846]: I1005 07:07:50.838383 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 05 07:07:50 crc kubenswrapper[4846]: I1005 07:07:50.898469 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b7a58faa-0f37-4b0e-a907-6a8bae28cb27-combined-ca-bundle\") pod \"b7a58faa-0f37-4b0e-a907-6a8bae28cb27\" (UID: \"b7a58faa-0f37-4b0e-a907-6a8bae28cb27\") " Oct 05 07:07:50 crc kubenswrapper[4846]: I1005 07:07:50.899582 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/b7a58faa-0f37-4b0e-a907-6a8bae28cb27-db-sync-config-data\") pod \"b7a58faa-0f37-4b0e-a907-6a8bae28cb27\" (UID: \"b7a58faa-0f37-4b0e-a907-6a8bae28cb27\") " Oct 05 07:07:50 crc kubenswrapper[4846]: I1005 07:07:50.899921 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cft2p\" (UniqueName: \"kubernetes.io/projected/b7a58faa-0f37-4b0e-a907-6a8bae28cb27-kube-api-access-cft2p\") pod \"b7a58faa-0f37-4b0e-a907-6a8bae28cb27\" (UID: \"b7a58faa-0f37-4b0e-a907-6a8bae28cb27\") " Oct 05 07:07:50 crc kubenswrapper[4846]: I1005 07:07:50.908368 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b7a58faa-0f37-4b0e-a907-6a8bae28cb27-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "b7a58faa-0f37-4b0e-a907-6a8bae28cb27" (UID: "b7a58faa-0f37-4b0e-a907-6a8bae28cb27"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:07:50 crc kubenswrapper[4846]: I1005 07:07:50.909410 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b7a58faa-0f37-4b0e-a907-6a8bae28cb27-kube-api-access-cft2p" (OuterVolumeSpecName: "kube-api-access-cft2p") pod "b7a58faa-0f37-4b0e-a907-6a8bae28cb27" (UID: "b7a58faa-0f37-4b0e-a907-6a8bae28cb27"). InnerVolumeSpecName "kube-api-access-cft2p". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:07:50 crc kubenswrapper[4846]: I1005 07:07:50.938831 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b7a58faa-0f37-4b0e-a907-6a8bae28cb27-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b7a58faa-0f37-4b0e-a907-6a8bae28cb27" (UID: "b7a58faa-0f37-4b0e-a907-6a8bae28cb27"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:07:51 crc kubenswrapper[4846]: I1005 07:07:51.004558 4846 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b7a58faa-0f37-4b0e-a907-6a8bae28cb27-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 07:07:51 crc kubenswrapper[4846]: I1005 07:07:51.004598 4846 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/b7a58faa-0f37-4b0e-a907-6a8bae28cb27-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 07:07:51 crc kubenswrapper[4846]: I1005 07:07:51.004610 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cft2p\" (UniqueName: \"kubernetes.io/projected/b7a58faa-0f37-4b0e-a907-6a8bae28cb27-kube-api-access-cft2p\") on node \"crc\" DevicePath \"\"" Oct 05 07:07:51 crc kubenswrapper[4846]: I1005 07:07:51.351458 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 05 07:07:51 crc kubenswrapper[4846]: W1005 07:07:51.356567 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod39071bf4_214b_4bbb_8b78_dfec6ad3c940.slice/crio-08c39de0b53585ece279a655d42b2c0a11038c4cd766ef12f56aa73eb55703ba WatchSource:0}: Error finding container 08c39de0b53585ece279a655d42b2c0a11038c4cd766ef12f56aa73eb55703ba: Status 404 returned error can't find the container with id 08c39de0b53585ece279a655d42b2c0a11038c4cd766ef12f56aa73eb55703ba Oct 05 07:07:51 crc kubenswrapper[4846]: I1005 07:07:51.408943 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"edd6502e-6e77-4ad3-8f1f-b85426188261","Type":"ContainerStarted","Data":"fe4a5d7cbeaf3a5245813e826bd7235fb2d424df813e58a0c069e9a6f5b444ca"} Oct 05 07:07:51 crc kubenswrapper[4846]: I1005 07:07:51.415221 4846 generic.go:334] "Generic (PLEG): container finished" podID="ee312852-b2eb-43d8-864c-bc553fed7ea5" containerID="8adc95b909659a53237badc065088ebc75df6ae5e870528ac833dde781fc0d99" exitCode=0 Oct 05 07:07:51 crc kubenswrapper[4846]: I1005 07:07:51.415311 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-cf6rt" event={"ID":"ee312852-b2eb-43d8-864c-bc553fed7ea5","Type":"ContainerDied","Data":"8adc95b909659a53237badc065088ebc75df6ae5e870528ac833dde781fc0d99"} Oct 05 07:07:51 crc kubenswrapper[4846]: I1005 07:07:51.417956 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-t5brq" event={"ID":"b7a58faa-0f37-4b0e-a907-6a8bae28cb27","Type":"ContainerDied","Data":"67f82731bda75c617e2cda8b0f68588e249bade2906ab46833c9ec7f2f2ddd3a"} Oct 05 07:07:51 crc kubenswrapper[4846]: I1005 07:07:51.418214 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="67f82731bda75c617e2cda8b0f68588e249bade2906ab46833c9ec7f2f2ddd3a" Oct 05 07:07:51 crc kubenswrapper[4846]: I1005 07:07:51.418163 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-t5brq" Oct 05 07:07:51 crc kubenswrapper[4846]: I1005 07:07:51.422022 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"39071bf4-214b-4bbb-8b78-dfec6ad3c940","Type":"ContainerStarted","Data":"08c39de0b53585ece279a655d42b2c0a11038c4cd766ef12f56aa73eb55703ba"} Oct 05 07:07:51 crc kubenswrapper[4846]: I1005 07:07:51.447760 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=3.4477284040000002 podStartE2EDuration="3.447728404s" podCreationTimestamp="2025-10-05 07:07:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 07:07:51.438865437 +0000 UTC m=+1193.679718222" watchObservedRunningTime="2025-10-05 07:07:51.447728404 +0000 UTC m=+1193.688581179" Oct 05 07:07:51 crc kubenswrapper[4846]: I1005 07:07:51.673084 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-worker-78995b9f79-nwl7v"] Oct 05 07:07:51 crc kubenswrapper[4846]: E1005 07:07:51.673558 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b7a58faa-0f37-4b0e-a907-6a8bae28cb27" containerName="barbican-db-sync" Oct 05 07:07:51 crc kubenswrapper[4846]: I1005 07:07:51.673579 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="b7a58faa-0f37-4b0e-a907-6a8bae28cb27" containerName="barbican-db-sync" Oct 05 07:07:51 crc kubenswrapper[4846]: I1005 07:07:51.673792 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="b7a58faa-0f37-4b0e-a907-6a8bae28cb27" containerName="barbican-db-sync" Oct 05 07:07:51 crc kubenswrapper[4846]: I1005 07:07:51.674975 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-78995b9f79-nwl7v" Oct 05 07:07:51 crc kubenswrapper[4846]: I1005 07:07:51.677481 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-2xgzt" Oct 05 07:07:51 crc kubenswrapper[4846]: I1005 07:07:51.678002 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Oct 05 07:07:51 crc kubenswrapper[4846]: I1005 07:07:51.681055 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-worker-config-data" Oct 05 07:07:51 crc kubenswrapper[4846]: I1005 07:07:51.699547 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-78995b9f79-nwl7v"] Oct 05 07:07:51 crc kubenswrapper[4846]: I1005 07:07:51.712946 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-keystone-listener-77664b8dfd-5hkqj"] Oct 05 07:07:51 crc kubenswrapper[4846]: I1005 07:07:51.715215 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-77664b8dfd-5hkqj" Oct 05 07:07:51 crc kubenswrapper[4846]: I1005 07:07:51.721859 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-keystone-listener-config-data" Oct 05 07:07:51 crc kubenswrapper[4846]: I1005 07:07:51.724086 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f55kq\" (UniqueName: \"kubernetes.io/projected/3995994c-2a3f-400f-9a64-b564759b20b2-kube-api-access-f55kq\") pod \"barbican-worker-78995b9f79-nwl7v\" (UID: \"3995994c-2a3f-400f-9a64-b564759b20b2\") " pod="openstack/barbican-worker-78995b9f79-nwl7v" Oct 05 07:07:51 crc kubenswrapper[4846]: I1005 07:07:51.724280 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3995994c-2a3f-400f-9a64-b564759b20b2-combined-ca-bundle\") pod \"barbican-worker-78995b9f79-nwl7v\" (UID: \"3995994c-2a3f-400f-9a64-b564759b20b2\") " pod="openstack/barbican-worker-78995b9f79-nwl7v" Oct 05 07:07:51 crc kubenswrapper[4846]: I1005 07:07:51.724396 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3995994c-2a3f-400f-9a64-b564759b20b2-logs\") pod \"barbican-worker-78995b9f79-nwl7v\" (UID: \"3995994c-2a3f-400f-9a64-b564759b20b2\") " pod="openstack/barbican-worker-78995b9f79-nwl7v" Oct 05 07:07:51 crc kubenswrapper[4846]: I1005 07:07:51.725841 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3995994c-2a3f-400f-9a64-b564759b20b2-config-data-custom\") pod \"barbican-worker-78995b9f79-nwl7v\" (UID: \"3995994c-2a3f-400f-9a64-b564759b20b2\") " pod="openstack/barbican-worker-78995b9f79-nwl7v" Oct 05 07:07:51 crc kubenswrapper[4846]: I1005 07:07:51.726039 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3995994c-2a3f-400f-9a64-b564759b20b2-config-data\") pod \"barbican-worker-78995b9f79-nwl7v\" (UID: \"3995994c-2a3f-400f-9a64-b564759b20b2\") " pod="openstack/barbican-worker-78995b9f79-nwl7v" Oct 05 07:07:51 crc kubenswrapper[4846]: I1005 07:07:51.750994 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-77664b8dfd-5hkqj"] Oct 05 07:07:51 crc kubenswrapper[4846]: I1005 07:07:51.827631 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f55kq\" (UniqueName: \"kubernetes.io/projected/3995994c-2a3f-400f-9a64-b564759b20b2-kube-api-access-f55kq\") pod \"barbican-worker-78995b9f79-nwl7v\" (UID: \"3995994c-2a3f-400f-9a64-b564759b20b2\") " pod="openstack/barbican-worker-78995b9f79-nwl7v" Oct 05 07:07:51 crc kubenswrapper[4846]: I1005 07:07:51.827692 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3995994c-2a3f-400f-9a64-b564759b20b2-combined-ca-bundle\") pod \"barbican-worker-78995b9f79-nwl7v\" (UID: \"3995994c-2a3f-400f-9a64-b564759b20b2\") " pod="openstack/barbican-worker-78995b9f79-nwl7v" Oct 05 07:07:51 crc kubenswrapper[4846]: I1005 07:07:51.827750 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8n2gd\" (UniqueName: \"kubernetes.io/projected/8d55f55a-2d79-4c01-8b0f-d67cefb842ff-kube-api-access-8n2gd\") pod \"barbican-keystone-listener-77664b8dfd-5hkqj\" (UID: \"8d55f55a-2d79-4c01-8b0f-d67cefb842ff\") " pod="openstack/barbican-keystone-listener-77664b8dfd-5hkqj" Oct 05 07:07:51 crc kubenswrapper[4846]: I1005 07:07:51.827778 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3995994c-2a3f-400f-9a64-b564759b20b2-logs\") pod \"barbican-worker-78995b9f79-nwl7v\" (UID: \"3995994c-2a3f-400f-9a64-b564759b20b2\") " pod="openstack/barbican-worker-78995b9f79-nwl7v" Oct 05 07:07:51 crc kubenswrapper[4846]: I1005 07:07:51.827819 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8d55f55a-2d79-4c01-8b0f-d67cefb842ff-logs\") pod \"barbican-keystone-listener-77664b8dfd-5hkqj\" (UID: \"8d55f55a-2d79-4c01-8b0f-d67cefb842ff\") " pod="openstack/barbican-keystone-listener-77664b8dfd-5hkqj" Oct 05 07:07:51 crc kubenswrapper[4846]: I1005 07:07:51.827842 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3995994c-2a3f-400f-9a64-b564759b20b2-config-data-custom\") pod \"barbican-worker-78995b9f79-nwl7v\" (UID: \"3995994c-2a3f-400f-9a64-b564759b20b2\") " pod="openstack/barbican-worker-78995b9f79-nwl7v" Oct 05 07:07:51 crc kubenswrapper[4846]: I1005 07:07:51.827888 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8d55f55a-2d79-4c01-8b0f-d67cefb842ff-config-data\") pod \"barbican-keystone-listener-77664b8dfd-5hkqj\" (UID: \"8d55f55a-2d79-4c01-8b0f-d67cefb842ff\") " pod="openstack/barbican-keystone-listener-77664b8dfd-5hkqj" Oct 05 07:07:51 crc kubenswrapper[4846]: I1005 07:07:51.827905 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8d55f55a-2d79-4c01-8b0f-d67cefb842ff-combined-ca-bundle\") pod \"barbican-keystone-listener-77664b8dfd-5hkqj\" (UID: \"8d55f55a-2d79-4c01-8b0f-d67cefb842ff\") " pod="openstack/barbican-keystone-listener-77664b8dfd-5hkqj" Oct 05 07:07:51 crc kubenswrapper[4846]: I1005 07:07:51.827928 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3995994c-2a3f-400f-9a64-b564759b20b2-config-data\") pod \"barbican-worker-78995b9f79-nwl7v\" (UID: \"3995994c-2a3f-400f-9a64-b564759b20b2\") " pod="openstack/barbican-worker-78995b9f79-nwl7v" Oct 05 07:07:51 crc kubenswrapper[4846]: I1005 07:07:51.827947 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8d55f55a-2d79-4c01-8b0f-d67cefb842ff-config-data-custom\") pod \"barbican-keystone-listener-77664b8dfd-5hkqj\" (UID: \"8d55f55a-2d79-4c01-8b0f-d67cefb842ff\") " pod="openstack/barbican-keystone-listener-77664b8dfd-5hkqj" Oct 05 07:07:51 crc kubenswrapper[4846]: I1005 07:07:51.828474 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-544f66dff9-4zksz"] Oct 05 07:07:51 crc kubenswrapper[4846]: I1005 07:07:51.829873 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3995994c-2a3f-400f-9a64-b564759b20b2-logs\") pod \"barbican-worker-78995b9f79-nwl7v\" (UID: \"3995994c-2a3f-400f-9a64-b564759b20b2\") " pod="openstack/barbican-worker-78995b9f79-nwl7v" Oct 05 07:07:51 crc kubenswrapper[4846]: I1005 07:07:51.831709 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-544f66dff9-4zksz" Oct 05 07:07:51 crc kubenswrapper[4846]: I1005 07:07:51.841185 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3995994c-2a3f-400f-9a64-b564759b20b2-config-data\") pod \"barbican-worker-78995b9f79-nwl7v\" (UID: \"3995994c-2a3f-400f-9a64-b564759b20b2\") " pod="openstack/barbican-worker-78995b9f79-nwl7v" Oct 05 07:07:51 crc kubenswrapper[4846]: I1005 07:07:51.844370 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3995994c-2a3f-400f-9a64-b564759b20b2-config-data-custom\") pod \"barbican-worker-78995b9f79-nwl7v\" (UID: \"3995994c-2a3f-400f-9a64-b564759b20b2\") " pod="openstack/barbican-worker-78995b9f79-nwl7v" Oct 05 07:07:51 crc kubenswrapper[4846]: I1005 07:07:51.869338 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3995994c-2a3f-400f-9a64-b564759b20b2-combined-ca-bundle\") pod \"barbican-worker-78995b9f79-nwl7v\" (UID: \"3995994c-2a3f-400f-9a64-b564759b20b2\") " pod="openstack/barbican-worker-78995b9f79-nwl7v" Oct 05 07:07:51 crc kubenswrapper[4846]: I1005 07:07:51.877222 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-544f66dff9-4zksz"] Oct 05 07:07:51 crc kubenswrapper[4846]: I1005 07:07:51.897964 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f55kq\" (UniqueName: \"kubernetes.io/projected/3995994c-2a3f-400f-9a64-b564759b20b2-kube-api-access-f55kq\") pod \"barbican-worker-78995b9f79-nwl7v\" (UID: \"3995994c-2a3f-400f-9a64-b564759b20b2\") " pod="openstack/barbican-worker-78995b9f79-nwl7v" Oct 05 07:07:51 crc kubenswrapper[4846]: I1005 07:07:51.944133 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8d55f55a-2d79-4c01-8b0f-d67cefb842ff-config-data\") pod \"barbican-keystone-listener-77664b8dfd-5hkqj\" (UID: \"8d55f55a-2d79-4c01-8b0f-d67cefb842ff\") " pod="openstack/barbican-keystone-listener-77664b8dfd-5hkqj" Oct 05 07:07:51 crc kubenswrapper[4846]: I1005 07:07:51.944205 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8d55f55a-2d79-4c01-8b0f-d67cefb842ff-combined-ca-bundle\") pod \"barbican-keystone-listener-77664b8dfd-5hkqj\" (UID: \"8d55f55a-2d79-4c01-8b0f-d67cefb842ff\") " pod="openstack/barbican-keystone-listener-77664b8dfd-5hkqj" Oct 05 07:07:51 crc kubenswrapper[4846]: I1005 07:07:51.944261 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8d55f55a-2d79-4c01-8b0f-d67cefb842ff-config-data-custom\") pod \"barbican-keystone-listener-77664b8dfd-5hkqj\" (UID: \"8d55f55a-2d79-4c01-8b0f-d67cefb842ff\") " pod="openstack/barbican-keystone-listener-77664b8dfd-5hkqj" Oct 05 07:07:51 crc kubenswrapper[4846]: I1005 07:07:51.944295 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d5c6818d-305f-4565-a245-672dde7ddfd2-config\") pod \"dnsmasq-dns-544f66dff9-4zksz\" (UID: \"d5c6818d-305f-4565-a245-672dde7ddfd2\") " pod="openstack/dnsmasq-dns-544f66dff9-4zksz" Oct 05 07:07:51 crc kubenswrapper[4846]: I1005 07:07:51.944407 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d5c6818d-305f-4565-a245-672dde7ddfd2-dns-swift-storage-0\") pod \"dnsmasq-dns-544f66dff9-4zksz\" (UID: \"d5c6818d-305f-4565-a245-672dde7ddfd2\") " pod="openstack/dnsmasq-dns-544f66dff9-4zksz" Oct 05 07:07:51 crc kubenswrapper[4846]: I1005 07:07:51.944465 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d5c6818d-305f-4565-a245-672dde7ddfd2-ovsdbserver-sb\") pod \"dnsmasq-dns-544f66dff9-4zksz\" (UID: \"d5c6818d-305f-4565-a245-672dde7ddfd2\") " pod="openstack/dnsmasq-dns-544f66dff9-4zksz" Oct 05 07:07:51 crc kubenswrapper[4846]: I1005 07:07:51.944551 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2t6b8\" (UniqueName: \"kubernetes.io/projected/d5c6818d-305f-4565-a245-672dde7ddfd2-kube-api-access-2t6b8\") pod \"dnsmasq-dns-544f66dff9-4zksz\" (UID: \"d5c6818d-305f-4565-a245-672dde7ddfd2\") " pod="openstack/dnsmasq-dns-544f66dff9-4zksz" Oct 05 07:07:51 crc kubenswrapper[4846]: I1005 07:07:51.944576 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8n2gd\" (UniqueName: \"kubernetes.io/projected/8d55f55a-2d79-4c01-8b0f-d67cefb842ff-kube-api-access-8n2gd\") pod \"barbican-keystone-listener-77664b8dfd-5hkqj\" (UID: \"8d55f55a-2d79-4c01-8b0f-d67cefb842ff\") " pod="openstack/barbican-keystone-listener-77664b8dfd-5hkqj" Oct 05 07:07:51 crc kubenswrapper[4846]: I1005 07:07:51.944665 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d5c6818d-305f-4565-a245-672dde7ddfd2-dns-svc\") pod \"dnsmasq-dns-544f66dff9-4zksz\" (UID: \"d5c6818d-305f-4565-a245-672dde7ddfd2\") " pod="openstack/dnsmasq-dns-544f66dff9-4zksz" Oct 05 07:07:51 crc kubenswrapper[4846]: I1005 07:07:51.944685 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d5c6818d-305f-4565-a245-672dde7ddfd2-ovsdbserver-nb\") pod \"dnsmasq-dns-544f66dff9-4zksz\" (UID: \"d5c6818d-305f-4565-a245-672dde7ddfd2\") " pod="openstack/dnsmasq-dns-544f66dff9-4zksz" Oct 05 07:07:51 crc kubenswrapper[4846]: I1005 07:07:51.944716 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8d55f55a-2d79-4c01-8b0f-d67cefb842ff-logs\") pod \"barbican-keystone-listener-77664b8dfd-5hkqj\" (UID: \"8d55f55a-2d79-4c01-8b0f-d67cefb842ff\") " pod="openstack/barbican-keystone-listener-77664b8dfd-5hkqj" Oct 05 07:07:51 crc kubenswrapper[4846]: I1005 07:07:51.945405 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8d55f55a-2d79-4c01-8b0f-d67cefb842ff-logs\") pod \"barbican-keystone-listener-77664b8dfd-5hkqj\" (UID: \"8d55f55a-2d79-4c01-8b0f-d67cefb842ff\") " pod="openstack/barbican-keystone-listener-77664b8dfd-5hkqj" Oct 05 07:07:51 crc kubenswrapper[4846]: I1005 07:07:51.949089 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8d55f55a-2d79-4c01-8b0f-d67cefb842ff-combined-ca-bundle\") pod \"barbican-keystone-listener-77664b8dfd-5hkqj\" (UID: \"8d55f55a-2d79-4c01-8b0f-d67cefb842ff\") " pod="openstack/barbican-keystone-listener-77664b8dfd-5hkqj" Oct 05 07:07:51 crc kubenswrapper[4846]: I1005 07:07:51.963758 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8d55f55a-2d79-4c01-8b0f-d67cefb842ff-config-data\") pod \"barbican-keystone-listener-77664b8dfd-5hkqj\" (UID: \"8d55f55a-2d79-4c01-8b0f-d67cefb842ff\") " pod="openstack/barbican-keystone-listener-77664b8dfd-5hkqj" Oct 05 07:07:51 crc kubenswrapper[4846]: I1005 07:07:51.963951 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8d55f55a-2d79-4c01-8b0f-d67cefb842ff-config-data-custom\") pod \"barbican-keystone-listener-77664b8dfd-5hkqj\" (UID: \"8d55f55a-2d79-4c01-8b0f-d67cefb842ff\") " pod="openstack/barbican-keystone-listener-77664b8dfd-5hkqj" Oct 05 07:07:51 crc kubenswrapper[4846]: I1005 07:07:51.974941 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8n2gd\" (UniqueName: \"kubernetes.io/projected/8d55f55a-2d79-4c01-8b0f-d67cefb842ff-kube-api-access-8n2gd\") pod \"barbican-keystone-listener-77664b8dfd-5hkqj\" (UID: \"8d55f55a-2d79-4c01-8b0f-d67cefb842ff\") " pod="openstack/barbican-keystone-listener-77664b8dfd-5hkqj" Oct 05 07:07:52 crc kubenswrapper[4846]: I1005 07:07:51.999938 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-78995b9f79-nwl7v" Oct 05 07:07:52 crc kubenswrapper[4846]: I1005 07:07:52.028295 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-6c9ddc8584-bqgtc"] Oct 05 07:07:52 crc kubenswrapper[4846]: I1005 07:07:52.030227 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-6c9ddc8584-bqgtc" Oct 05 07:07:52 crc kubenswrapper[4846]: I1005 07:07:52.033167 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-api-config-data" Oct 05 07:07:52 crc kubenswrapper[4846]: I1005 07:07:52.042704 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-77664b8dfd-5hkqj" Oct 05 07:07:52 crc kubenswrapper[4846]: I1005 07:07:52.051978 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d5c6818d-305f-4565-a245-672dde7ddfd2-config\") pod \"dnsmasq-dns-544f66dff9-4zksz\" (UID: \"d5c6818d-305f-4565-a245-672dde7ddfd2\") " pod="openstack/dnsmasq-dns-544f66dff9-4zksz" Oct 05 07:07:52 crc kubenswrapper[4846]: I1005 07:07:52.052126 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d5c6818d-305f-4565-a245-672dde7ddfd2-dns-swift-storage-0\") pod \"dnsmasq-dns-544f66dff9-4zksz\" (UID: \"d5c6818d-305f-4565-a245-672dde7ddfd2\") " pod="openstack/dnsmasq-dns-544f66dff9-4zksz" Oct 05 07:07:52 crc kubenswrapper[4846]: I1005 07:07:52.052230 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d5c6818d-305f-4565-a245-672dde7ddfd2-ovsdbserver-sb\") pod \"dnsmasq-dns-544f66dff9-4zksz\" (UID: \"d5c6818d-305f-4565-a245-672dde7ddfd2\") " pod="openstack/dnsmasq-dns-544f66dff9-4zksz" Oct 05 07:07:52 crc kubenswrapper[4846]: I1005 07:07:52.052345 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2t6b8\" (UniqueName: \"kubernetes.io/projected/d5c6818d-305f-4565-a245-672dde7ddfd2-kube-api-access-2t6b8\") pod \"dnsmasq-dns-544f66dff9-4zksz\" (UID: \"d5c6818d-305f-4565-a245-672dde7ddfd2\") " pod="openstack/dnsmasq-dns-544f66dff9-4zksz" Oct 05 07:07:52 crc kubenswrapper[4846]: I1005 07:07:52.052471 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d5c6818d-305f-4565-a245-672dde7ddfd2-dns-svc\") pod \"dnsmasq-dns-544f66dff9-4zksz\" (UID: \"d5c6818d-305f-4565-a245-672dde7ddfd2\") " pod="openstack/dnsmasq-dns-544f66dff9-4zksz" Oct 05 07:07:52 crc kubenswrapper[4846]: I1005 07:07:52.052491 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d5c6818d-305f-4565-a245-672dde7ddfd2-ovsdbserver-nb\") pod \"dnsmasq-dns-544f66dff9-4zksz\" (UID: \"d5c6818d-305f-4565-a245-672dde7ddfd2\") " pod="openstack/dnsmasq-dns-544f66dff9-4zksz" Oct 05 07:07:52 crc kubenswrapper[4846]: I1005 07:07:52.053542 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d5c6818d-305f-4565-a245-672dde7ddfd2-ovsdbserver-nb\") pod \"dnsmasq-dns-544f66dff9-4zksz\" (UID: \"d5c6818d-305f-4565-a245-672dde7ddfd2\") " pod="openstack/dnsmasq-dns-544f66dff9-4zksz" Oct 05 07:07:52 crc kubenswrapper[4846]: I1005 07:07:52.054085 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d5c6818d-305f-4565-a245-672dde7ddfd2-ovsdbserver-sb\") pod \"dnsmasq-dns-544f66dff9-4zksz\" (UID: \"d5c6818d-305f-4565-a245-672dde7ddfd2\") " pod="openstack/dnsmasq-dns-544f66dff9-4zksz" Oct 05 07:07:52 crc kubenswrapper[4846]: I1005 07:07:52.054493 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d5c6818d-305f-4565-a245-672dde7ddfd2-dns-svc\") pod \"dnsmasq-dns-544f66dff9-4zksz\" (UID: \"d5c6818d-305f-4565-a245-672dde7ddfd2\") " pod="openstack/dnsmasq-dns-544f66dff9-4zksz" Oct 05 07:07:52 crc kubenswrapper[4846]: I1005 07:07:52.054543 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-6c9ddc8584-bqgtc"] Oct 05 07:07:52 crc kubenswrapper[4846]: I1005 07:07:52.054721 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d5c6818d-305f-4565-a245-672dde7ddfd2-config\") pod \"dnsmasq-dns-544f66dff9-4zksz\" (UID: \"d5c6818d-305f-4565-a245-672dde7ddfd2\") " pod="openstack/dnsmasq-dns-544f66dff9-4zksz" Oct 05 07:07:52 crc kubenswrapper[4846]: I1005 07:07:52.055696 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d5c6818d-305f-4565-a245-672dde7ddfd2-dns-swift-storage-0\") pod \"dnsmasq-dns-544f66dff9-4zksz\" (UID: \"d5c6818d-305f-4565-a245-672dde7ddfd2\") " pod="openstack/dnsmasq-dns-544f66dff9-4zksz" Oct 05 07:07:52 crc kubenswrapper[4846]: I1005 07:07:52.077691 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2t6b8\" (UniqueName: \"kubernetes.io/projected/d5c6818d-305f-4565-a245-672dde7ddfd2-kube-api-access-2t6b8\") pod \"dnsmasq-dns-544f66dff9-4zksz\" (UID: \"d5c6818d-305f-4565-a245-672dde7ddfd2\") " pod="openstack/dnsmasq-dns-544f66dff9-4zksz" Oct 05 07:07:52 crc kubenswrapper[4846]: I1005 07:07:52.154361 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ba9102c0-185f-4f02-8b05-8bb7fde2d573-logs\") pod \"barbican-api-6c9ddc8584-bqgtc\" (UID: \"ba9102c0-185f-4f02-8b05-8bb7fde2d573\") " pod="openstack/barbican-api-6c9ddc8584-bqgtc" Oct 05 07:07:52 crc kubenswrapper[4846]: I1005 07:07:52.154872 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ba9102c0-185f-4f02-8b05-8bb7fde2d573-combined-ca-bundle\") pod \"barbican-api-6c9ddc8584-bqgtc\" (UID: \"ba9102c0-185f-4f02-8b05-8bb7fde2d573\") " pod="openstack/barbican-api-6c9ddc8584-bqgtc" Oct 05 07:07:52 crc kubenswrapper[4846]: I1005 07:07:52.154967 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ba9102c0-185f-4f02-8b05-8bb7fde2d573-config-data-custom\") pod \"barbican-api-6c9ddc8584-bqgtc\" (UID: \"ba9102c0-185f-4f02-8b05-8bb7fde2d573\") " pod="openstack/barbican-api-6c9ddc8584-bqgtc" Oct 05 07:07:52 crc kubenswrapper[4846]: I1005 07:07:52.155044 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ba9102c0-185f-4f02-8b05-8bb7fde2d573-config-data\") pod \"barbican-api-6c9ddc8584-bqgtc\" (UID: \"ba9102c0-185f-4f02-8b05-8bb7fde2d573\") " pod="openstack/barbican-api-6c9ddc8584-bqgtc" Oct 05 07:07:52 crc kubenswrapper[4846]: I1005 07:07:52.155370 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-chqvs\" (UniqueName: \"kubernetes.io/projected/ba9102c0-185f-4f02-8b05-8bb7fde2d573-kube-api-access-chqvs\") pod \"barbican-api-6c9ddc8584-bqgtc\" (UID: \"ba9102c0-185f-4f02-8b05-8bb7fde2d573\") " pod="openstack/barbican-api-6c9ddc8584-bqgtc" Oct 05 07:07:52 crc kubenswrapper[4846]: I1005 07:07:52.258767 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-chqvs\" (UniqueName: \"kubernetes.io/projected/ba9102c0-185f-4f02-8b05-8bb7fde2d573-kube-api-access-chqvs\") pod \"barbican-api-6c9ddc8584-bqgtc\" (UID: \"ba9102c0-185f-4f02-8b05-8bb7fde2d573\") " pod="openstack/barbican-api-6c9ddc8584-bqgtc" Oct 05 07:07:52 crc kubenswrapper[4846]: I1005 07:07:52.258877 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ba9102c0-185f-4f02-8b05-8bb7fde2d573-logs\") pod \"barbican-api-6c9ddc8584-bqgtc\" (UID: \"ba9102c0-185f-4f02-8b05-8bb7fde2d573\") " pod="openstack/barbican-api-6c9ddc8584-bqgtc" Oct 05 07:07:52 crc kubenswrapper[4846]: I1005 07:07:52.258904 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ba9102c0-185f-4f02-8b05-8bb7fde2d573-combined-ca-bundle\") pod \"barbican-api-6c9ddc8584-bqgtc\" (UID: \"ba9102c0-185f-4f02-8b05-8bb7fde2d573\") " pod="openstack/barbican-api-6c9ddc8584-bqgtc" Oct 05 07:07:52 crc kubenswrapper[4846]: I1005 07:07:52.259117 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ba9102c0-185f-4f02-8b05-8bb7fde2d573-config-data-custom\") pod \"barbican-api-6c9ddc8584-bqgtc\" (UID: \"ba9102c0-185f-4f02-8b05-8bb7fde2d573\") " pod="openstack/barbican-api-6c9ddc8584-bqgtc" Oct 05 07:07:52 crc kubenswrapper[4846]: I1005 07:07:52.259164 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ba9102c0-185f-4f02-8b05-8bb7fde2d573-config-data\") pod \"barbican-api-6c9ddc8584-bqgtc\" (UID: \"ba9102c0-185f-4f02-8b05-8bb7fde2d573\") " pod="openstack/barbican-api-6c9ddc8584-bqgtc" Oct 05 07:07:52 crc kubenswrapper[4846]: I1005 07:07:52.259894 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ba9102c0-185f-4f02-8b05-8bb7fde2d573-logs\") pod \"barbican-api-6c9ddc8584-bqgtc\" (UID: \"ba9102c0-185f-4f02-8b05-8bb7fde2d573\") " pod="openstack/barbican-api-6c9ddc8584-bqgtc" Oct 05 07:07:52 crc kubenswrapper[4846]: I1005 07:07:52.271405 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ba9102c0-185f-4f02-8b05-8bb7fde2d573-combined-ca-bundle\") pod \"barbican-api-6c9ddc8584-bqgtc\" (UID: \"ba9102c0-185f-4f02-8b05-8bb7fde2d573\") " pod="openstack/barbican-api-6c9ddc8584-bqgtc" Oct 05 07:07:52 crc kubenswrapper[4846]: I1005 07:07:52.271616 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ba9102c0-185f-4f02-8b05-8bb7fde2d573-config-data\") pod \"barbican-api-6c9ddc8584-bqgtc\" (UID: \"ba9102c0-185f-4f02-8b05-8bb7fde2d573\") " pod="openstack/barbican-api-6c9ddc8584-bqgtc" Oct 05 07:07:52 crc kubenswrapper[4846]: I1005 07:07:52.271989 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ba9102c0-185f-4f02-8b05-8bb7fde2d573-config-data-custom\") pod \"barbican-api-6c9ddc8584-bqgtc\" (UID: \"ba9102c0-185f-4f02-8b05-8bb7fde2d573\") " pod="openstack/barbican-api-6c9ddc8584-bqgtc" Oct 05 07:07:52 crc kubenswrapper[4846]: I1005 07:07:52.277594 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-chqvs\" (UniqueName: \"kubernetes.io/projected/ba9102c0-185f-4f02-8b05-8bb7fde2d573-kube-api-access-chqvs\") pod \"barbican-api-6c9ddc8584-bqgtc\" (UID: \"ba9102c0-185f-4f02-8b05-8bb7fde2d573\") " pod="openstack/barbican-api-6c9ddc8584-bqgtc" Oct 05 07:07:52 crc kubenswrapper[4846]: I1005 07:07:52.325797 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-544f66dff9-4zksz" Oct 05 07:07:52 crc kubenswrapper[4846]: I1005 07:07:52.367964 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-6c9ddc8584-bqgtc" Oct 05 07:07:52 crc kubenswrapper[4846]: I1005 07:07:52.436263 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"39071bf4-214b-4bbb-8b78-dfec6ad3c940","Type":"ContainerStarted","Data":"82b673a1934d34adb80bf16feac56c7fc2158772b630ab377b61d9b3c475954b"} Oct 05 07:07:52 crc kubenswrapper[4846]: I1005 07:07:52.523354 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-78995b9f79-nwl7v"] Oct 05 07:07:52 crc kubenswrapper[4846]: I1005 07:07:52.604371 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-77664b8dfd-5hkqj"] Oct 05 07:07:52 crc kubenswrapper[4846]: I1005 07:07:52.866926 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-544f66dff9-4zksz"] Oct 05 07:07:52 crc kubenswrapper[4846]: W1005 07:07:52.878246 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd5c6818d_305f_4565_a245_672dde7ddfd2.slice/crio-d28c31c68266f89b66543ff28dcc6676976c4a29c1f41d2d842b3be155593cdd WatchSource:0}: Error finding container d28c31c68266f89b66543ff28dcc6676976c4a29c1f41d2d842b3be155593cdd: Status 404 returned error can't find the container with id d28c31c68266f89b66543ff28dcc6676976c4a29c1f41d2d842b3be155593cdd Oct 05 07:07:52 crc kubenswrapper[4846]: I1005 07:07:52.974040 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-cf6rt" Oct 05 07:07:53 crc kubenswrapper[4846]: I1005 07:07:53.088229 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ee312852-b2eb-43d8-864c-bc553fed7ea5-combined-ca-bundle\") pod \"ee312852-b2eb-43d8-864c-bc553fed7ea5\" (UID: \"ee312852-b2eb-43d8-864c-bc553fed7ea5\") " Oct 05 07:07:53 crc kubenswrapper[4846]: I1005 07:07:53.088297 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ee312852-b2eb-43d8-864c-bc553fed7ea5-etc-machine-id\") pod \"ee312852-b2eb-43d8-864c-bc553fed7ea5\" (UID: \"ee312852-b2eb-43d8-864c-bc553fed7ea5\") " Oct 05 07:07:53 crc kubenswrapper[4846]: I1005 07:07:53.088325 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ee312852-b2eb-43d8-864c-bc553fed7ea5-config-data\") pod \"ee312852-b2eb-43d8-864c-bc553fed7ea5\" (UID: \"ee312852-b2eb-43d8-864c-bc553fed7ea5\") " Oct 05 07:07:53 crc kubenswrapper[4846]: I1005 07:07:53.088444 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j2hlh\" (UniqueName: \"kubernetes.io/projected/ee312852-b2eb-43d8-864c-bc553fed7ea5-kube-api-access-j2hlh\") pod \"ee312852-b2eb-43d8-864c-bc553fed7ea5\" (UID: \"ee312852-b2eb-43d8-864c-bc553fed7ea5\") " Oct 05 07:07:53 crc kubenswrapper[4846]: I1005 07:07:53.088694 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ee312852-b2eb-43d8-864c-bc553fed7ea5-scripts\") pod \"ee312852-b2eb-43d8-864c-bc553fed7ea5\" (UID: \"ee312852-b2eb-43d8-864c-bc553fed7ea5\") " Oct 05 07:07:53 crc kubenswrapper[4846]: I1005 07:07:53.088780 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/ee312852-b2eb-43d8-864c-bc553fed7ea5-db-sync-config-data\") pod \"ee312852-b2eb-43d8-864c-bc553fed7ea5\" (UID: \"ee312852-b2eb-43d8-864c-bc553fed7ea5\") " Oct 05 07:07:53 crc kubenswrapper[4846]: I1005 07:07:53.089619 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ee312852-b2eb-43d8-864c-bc553fed7ea5-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "ee312852-b2eb-43d8-864c-bc553fed7ea5" (UID: "ee312852-b2eb-43d8-864c-bc553fed7ea5"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 05 07:07:53 crc kubenswrapper[4846]: I1005 07:07:53.106335 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ee312852-b2eb-43d8-864c-bc553fed7ea5-scripts" (OuterVolumeSpecName: "scripts") pod "ee312852-b2eb-43d8-864c-bc553fed7ea5" (UID: "ee312852-b2eb-43d8-864c-bc553fed7ea5"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:07:53 crc kubenswrapper[4846]: I1005 07:07:53.107302 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ee312852-b2eb-43d8-864c-bc553fed7ea5-kube-api-access-j2hlh" (OuterVolumeSpecName: "kube-api-access-j2hlh") pod "ee312852-b2eb-43d8-864c-bc553fed7ea5" (UID: "ee312852-b2eb-43d8-864c-bc553fed7ea5"). InnerVolumeSpecName "kube-api-access-j2hlh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:07:53 crc kubenswrapper[4846]: I1005 07:07:53.110233 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ee312852-b2eb-43d8-864c-bc553fed7ea5-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "ee312852-b2eb-43d8-864c-bc553fed7ea5" (UID: "ee312852-b2eb-43d8-864c-bc553fed7ea5"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:07:53 crc kubenswrapper[4846]: I1005 07:07:53.136039 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ee312852-b2eb-43d8-864c-bc553fed7ea5-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ee312852-b2eb-43d8-864c-bc553fed7ea5" (UID: "ee312852-b2eb-43d8-864c-bc553fed7ea5"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:07:53 crc kubenswrapper[4846]: I1005 07:07:53.145887 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-6c9ddc8584-bqgtc"] Oct 05 07:07:53 crc kubenswrapper[4846]: I1005 07:07:53.167256 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ee312852-b2eb-43d8-864c-bc553fed7ea5-config-data" (OuterVolumeSpecName: "config-data") pod "ee312852-b2eb-43d8-864c-bc553fed7ea5" (UID: "ee312852-b2eb-43d8-864c-bc553fed7ea5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:07:53 crc kubenswrapper[4846]: I1005 07:07:53.191535 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j2hlh\" (UniqueName: \"kubernetes.io/projected/ee312852-b2eb-43d8-864c-bc553fed7ea5-kube-api-access-j2hlh\") on node \"crc\" DevicePath \"\"" Oct 05 07:07:53 crc kubenswrapper[4846]: I1005 07:07:53.191590 4846 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ee312852-b2eb-43d8-864c-bc553fed7ea5-scripts\") on node \"crc\" DevicePath \"\"" Oct 05 07:07:53 crc kubenswrapper[4846]: I1005 07:07:53.191601 4846 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/ee312852-b2eb-43d8-864c-bc553fed7ea5-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 07:07:53 crc kubenswrapper[4846]: I1005 07:07:53.191610 4846 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ee312852-b2eb-43d8-864c-bc553fed7ea5-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 07:07:53 crc kubenswrapper[4846]: I1005 07:07:53.191619 4846 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ee312852-b2eb-43d8-864c-bc553fed7ea5-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 05 07:07:53 crc kubenswrapper[4846]: I1005 07:07:53.191629 4846 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ee312852-b2eb-43d8-864c-bc553fed7ea5-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 07:07:53 crc kubenswrapper[4846]: I1005 07:07:53.464133 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-cf6rt" event={"ID":"ee312852-b2eb-43d8-864c-bc553fed7ea5","Type":"ContainerDied","Data":"79b45ea27ca7ea9da92bbd1c6a7b37e0e776b1d5f97e05d8996700255faac79a"} Oct 05 07:07:53 crc kubenswrapper[4846]: I1005 07:07:53.464537 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="79b45ea27ca7ea9da92bbd1c6a7b37e0e776b1d5f97e05d8996700255faac79a" Oct 05 07:07:53 crc kubenswrapper[4846]: I1005 07:07:53.464612 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-cf6rt" Oct 05 07:07:53 crc kubenswrapper[4846]: I1005 07:07:53.475172 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-78995b9f79-nwl7v" event={"ID":"3995994c-2a3f-400f-9a64-b564759b20b2","Type":"ContainerStarted","Data":"5f1d95b021777d90a46ec2e78aa381055bb3200c115fbcc6b0575e04c4a2dae3"} Oct 05 07:07:53 crc kubenswrapper[4846]: I1005 07:07:53.478097 4846 generic.go:334] "Generic (PLEG): container finished" podID="d5c6818d-305f-4565-a245-672dde7ddfd2" containerID="b8dce4eabd2404d514a869915acc0811e9f294ad356a82f4ea75e9c5d4ee7ca9" exitCode=0 Oct 05 07:07:53 crc kubenswrapper[4846]: I1005 07:07:53.478304 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-544f66dff9-4zksz" event={"ID":"d5c6818d-305f-4565-a245-672dde7ddfd2","Type":"ContainerDied","Data":"b8dce4eabd2404d514a869915acc0811e9f294ad356a82f4ea75e9c5d4ee7ca9"} Oct 05 07:07:53 crc kubenswrapper[4846]: I1005 07:07:53.478363 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-544f66dff9-4zksz" event={"ID":"d5c6818d-305f-4565-a245-672dde7ddfd2","Type":"ContainerStarted","Data":"d28c31c68266f89b66543ff28dcc6676976c4a29c1f41d2d842b3be155593cdd"} Oct 05 07:07:53 crc kubenswrapper[4846]: I1005 07:07:53.492020 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"39071bf4-214b-4bbb-8b78-dfec6ad3c940","Type":"ContainerStarted","Data":"15e05c1324747c5238b20b4131b788c1adc0d31ec16d39fc7037ec91c627a278"} Oct 05 07:07:53 crc kubenswrapper[4846]: I1005 07:07:53.496076 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6c9ddc8584-bqgtc" event={"ID":"ba9102c0-185f-4f02-8b05-8bb7fde2d573","Type":"ContainerStarted","Data":"ce32162fd7737fb25fa034d9314322010c146074fc19ea1fd62cf8b2dcfcb7ac"} Oct 05 07:07:53 crc kubenswrapper[4846]: I1005 07:07:53.511924 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-77664b8dfd-5hkqj" event={"ID":"8d55f55a-2d79-4c01-8b0f-d67cefb842ff","Type":"ContainerStarted","Data":"3ddbf2a790be1dd9add94290c14ebc38db8686c6dd45f7665a169e191b7fa9a3"} Oct 05 07:07:53 crc kubenswrapper[4846]: I1005 07:07:53.880435 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Oct 05 07:07:53 crc kubenswrapper[4846]: E1005 07:07:53.880928 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ee312852-b2eb-43d8-864c-bc553fed7ea5" containerName="cinder-db-sync" Oct 05 07:07:53 crc kubenswrapper[4846]: I1005 07:07:53.880947 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee312852-b2eb-43d8-864c-bc553fed7ea5" containerName="cinder-db-sync" Oct 05 07:07:53 crc kubenswrapper[4846]: I1005 07:07:53.881165 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="ee312852-b2eb-43d8-864c-bc553fed7ea5" containerName="cinder-db-sync" Oct 05 07:07:53 crc kubenswrapper[4846]: I1005 07:07:53.889939 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 05 07:07:53 crc kubenswrapper[4846]: I1005 07:07:53.901154 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Oct 05 07:07:53 crc kubenswrapper[4846]: I1005 07:07:53.901495 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Oct 05 07:07:53 crc kubenswrapper[4846]: I1005 07:07:53.901613 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-chkrj" Oct 05 07:07:53 crc kubenswrapper[4846]: I1005 07:07:53.908298 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Oct 05 07:07:53 crc kubenswrapper[4846]: I1005 07:07:53.910690 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-544f66dff9-4zksz"] Oct 05 07:07:53 crc kubenswrapper[4846]: I1005 07:07:53.957348 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 05 07:07:54 crc kubenswrapper[4846]: I1005 07:07:54.013015 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-99495cbc7-v87cl"] Oct 05 07:07:54 crc kubenswrapper[4846]: I1005 07:07:54.016882 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-99495cbc7-v87cl" Oct 05 07:07:54 crc kubenswrapper[4846]: I1005 07:07:54.019537 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/40289760-4ac7-42e8-8eab-39f93c1618aa-config-data\") pod \"cinder-scheduler-0\" (UID: \"40289760-4ac7-42e8-8eab-39f93c1618aa\") " pod="openstack/cinder-scheduler-0" Oct 05 07:07:54 crc kubenswrapper[4846]: I1005 07:07:54.019652 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/40289760-4ac7-42e8-8eab-39f93c1618aa-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"40289760-4ac7-42e8-8eab-39f93c1618aa\") " pod="openstack/cinder-scheduler-0" Oct 05 07:07:54 crc kubenswrapper[4846]: I1005 07:07:54.019684 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/40289760-4ac7-42e8-8eab-39f93c1618aa-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"40289760-4ac7-42e8-8eab-39f93c1618aa\") " pod="openstack/cinder-scheduler-0" Oct 05 07:07:54 crc kubenswrapper[4846]: I1005 07:07:54.019715 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/40289760-4ac7-42e8-8eab-39f93c1618aa-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"40289760-4ac7-42e8-8eab-39f93c1618aa\") " pod="openstack/cinder-scheduler-0" Oct 05 07:07:54 crc kubenswrapper[4846]: I1005 07:07:54.019772 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g45vh\" (UniqueName: \"kubernetes.io/projected/40289760-4ac7-42e8-8eab-39f93c1618aa-kube-api-access-g45vh\") pod \"cinder-scheduler-0\" (UID: \"40289760-4ac7-42e8-8eab-39f93c1618aa\") " pod="openstack/cinder-scheduler-0" Oct 05 07:07:54 crc kubenswrapper[4846]: I1005 07:07:54.019799 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/40289760-4ac7-42e8-8eab-39f93c1618aa-scripts\") pod \"cinder-scheduler-0\" (UID: \"40289760-4ac7-42e8-8eab-39f93c1618aa\") " pod="openstack/cinder-scheduler-0" Oct 05 07:07:54 crc kubenswrapper[4846]: I1005 07:07:54.108962 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-99495cbc7-v87cl"] Oct 05 07:07:54 crc kubenswrapper[4846]: I1005 07:07:54.154573 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/40289760-4ac7-42e8-8eab-39f93c1618aa-config-data\") pod \"cinder-scheduler-0\" (UID: \"40289760-4ac7-42e8-8eab-39f93c1618aa\") " pod="openstack/cinder-scheduler-0" Oct 05 07:07:54 crc kubenswrapper[4846]: I1005 07:07:54.154717 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/507e14a3-601e-461b-b8a7-f218b87b0f47-dns-swift-storage-0\") pod \"dnsmasq-dns-99495cbc7-v87cl\" (UID: \"507e14a3-601e-461b-b8a7-f218b87b0f47\") " pod="openstack/dnsmasq-dns-99495cbc7-v87cl" Oct 05 07:07:54 crc kubenswrapper[4846]: I1005 07:07:54.158837 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pvqr5\" (UniqueName: \"kubernetes.io/projected/507e14a3-601e-461b-b8a7-f218b87b0f47-kube-api-access-pvqr5\") pod \"dnsmasq-dns-99495cbc7-v87cl\" (UID: \"507e14a3-601e-461b-b8a7-f218b87b0f47\") " pod="openstack/dnsmasq-dns-99495cbc7-v87cl" Oct 05 07:07:54 crc kubenswrapper[4846]: I1005 07:07:54.158962 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/40289760-4ac7-42e8-8eab-39f93c1618aa-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"40289760-4ac7-42e8-8eab-39f93c1618aa\") " pod="openstack/cinder-scheduler-0" Oct 05 07:07:54 crc kubenswrapper[4846]: I1005 07:07:54.159085 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/40289760-4ac7-42e8-8eab-39f93c1618aa-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"40289760-4ac7-42e8-8eab-39f93c1618aa\") " pod="openstack/cinder-scheduler-0" Oct 05 07:07:54 crc kubenswrapper[4846]: I1005 07:07:54.159158 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/40289760-4ac7-42e8-8eab-39f93c1618aa-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"40289760-4ac7-42e8-8eab-39f93c1618aa\") " pod="openstack/cinder-scheduler-0" Oct 05 07:07:54 crc kubenswrapper[4846]: I1005 07:07:54.159638 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/507e14a3-601e-461b-b8a7-f218b87b0f47-ovsdbserver-sb\") pod \"dnsmasq-dns-99495cbc7-v87cl\" (UID: \"507e14a3-601e-461b-b8a7-f218b87b0f47\") " pod="openstack/dnsmasq-dns-99495cbc7-v87cl" Oct 05 07:07:54 crc kubenswrapper[4846]: I1005 07:07:54.159730 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/507e14a3-601e-461b-b8a7-f218b87b0f47-ovsdbserver-nb\") pod \"dnsmasq-dns-99495cbc7-v87cl\" (UID: \"507e14a3-601e-461b-b8a7-f218b87b0f47\") " pod="openstack/dnsmasq-dns-99495cbc7-v87cl" Oct 05 07:07:54 crc kubenswrapper[4846]: I1005 07:07:54.159814 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/507e14a3-601e-461b-b8a7-f218b87b0f47-dns-svc\") pod \"dnsmasq-dns-99495cbc7-v87cl\" (UID: \"507e14a3-601e-461b-b8a7-f218b87b0f47\") " pod="openstack/dnsmasq-dns-99495cbc7-v87cl" Oct 05 07:07:54 crc kubenswrapper[4846]: I1005 07:07:54.159932 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g45vh\" (UniqueName: \"kubernetes.io/projected/40289760-4ac7-42e8-8eab-39f93c1618aa-kube-api-access-g45vh\") pod \"cinder-scheduler-0\" (UID: \"40289760-4ac7-42e8-8eab-39f93c1618aa\") " pod="openstack/cinder-scheduler-0" Oct 05 07:07:54 crc kubenswrapper[4846]: I1005 07:07:54.160015 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/40289760-4ac7-42e8-8eab-39f93c1618aa-scripts\") pod \"cinder-scheduler-0\" (UID: \"40289760-4ac7-42e8-8eab-39f93c1618aa\") " pod="openstack/cinder-scheduler-0" Oct 05 07:07:54 crc kubenswrapper[4846]: I1005 07:07:54.161113 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/507e14a3-601e-461b-b8a7-f218b87b0f47-config\") pod \"dnsmasq-dns-99495cbc7-v87cl\" (UID: \"507e14a3-601e-461b-b8a7-f218b87b0f47\") " pod="openstack/dnsmasq-dns-99495cbc7-v87cl" Oct 05 07:07:54 crc kubenswrapper[4846]: I1005 07:07:54.164291 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/40289760-4ac7-42e8-8eab-39f93c1618aa-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"40289760-4ac7-42e8-8eab-39f93c1618aa\") " pod="openstack/cinder-scheduler-0" Oct 05 07:07:54 crc kubenswrapper[4846]: I1005 07:07:54.165650 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/40289760-4ac7-42e8-8eab-39f93c1618aa-config-data\") pod \"cinder-scheduler-0\" (UID: \"40289760-4ac7-42e8-8eab-39f93c1618aa\") " pod="openstack/cinder-scheduler-0" Oct 05 07:07:54 crc kubenswrapper[4846]: I1005 07:07:54.165957 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/40289760-4ac7-42e8-8eab-39f93c1618aa-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"40289760-4ac7-42e8-8eab-39f93c1618aa\") " pod="openstack/cinder-scheduler-0" Oct 05 07:07:54 crc kubenswrapper[4846]: I1005 07:07:54.169923 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/40289760-4ac7-42e8-8eab-39f93c1618aa-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"40289760-4ac7-42e8-8eab-39f93c1618aa\") " pod="openstack/cinder-scheduler-0" Oct 05 07:07:54 crc kubenswrapper[4846]: I1005 07:07:54.182878 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/40289760-4ac7-42e8-8eab-39f93c1618aa-scripts\") pod \"cinder-scheduler-0\" (UID: \"40289760-4ac7-42e8-8eab-39f93c1618aa\") " pod="openstack/cinder-scheduler-0" Oct 05 07:07:54 crc kubenswrapper[4846]: I1005 07:07:54.186825 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g45vh\" (UniqueName: \"kubernetes.io/projected/40289760-4ac7-42e8-8eab-39f93c1618aa-kube-api-access-g45vh\") pod \"cinder-scheduler-0\" (UID: \"40289760-4ac7-42e8-8eab-39f93c1618aa\") " pod="openstack/cinder-scheduler-0" Oct 05 07:07:54 crc kubenswrapper[4846]: I1005 07:07:54.216409 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Oct 05 07:07:54 crc kubenswrapper[4846]: I1005 07:07:54.218976 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 05 07:07:54 crc kubenswrapper[4846]: I1005 07:07:54.221616 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Oct 05 07:07:54 crc kubenswrapper[4846]: I1005 07:07:54.230387 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 05 07:07:54 crc kubenswrapper[4846]: I1005 07:07:54.263767 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/507e14a3-601e-461b-b8a7-f218b87b0f47-dns-swift-storage-0\") pod \"dnsmasq-dns-99495cbc7-v87cl\" (UID: \"507e14a3-601e-461b-b8a7-f218b87b0f47\") " pod="openstack/dnsmasq-dns-99495cbc7-v87cl" Oct 05 07:07:54 crc kubenswrapper[4846]: I1005 07:07:54.263870 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pvqr5\" (UniqueName: \"kubernetes.io/projected/507e14a3-601e-461b-b8a7-f218b87b0f47-kube-api-access-pvqr5\") pod \"dnsmasq-dns-99495cbc7-v87cl\" (UID: \"507e14a3-601e-461b-b8a7-f218b87b0f47\") " pod="openstack/dnsmasq-dns-99495cbc7-v87cl" Oct 05 07:07:54 crc kubenswrapper[4846]: I1005 07:07:54.263930 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/507e14a3-601e-461b-b8a7-f218b87b0f47-ovsdbserver-sb\") pod \"dnsmasq-dns-99495cbc7-v87cl\" (UID: \"507e14a3-601e-461b-b8a7-f218b87b0f47\") " pod="openstack/dnsmasq-dns-99495cbc7-v87cl" Oct 05 07:07:54 crc kubenswrapper[4846]: I1005 07:07:54.263956 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/507e14a3-601e-461b-b8a7-f218b87b0f47-ovsdbserver-nb\") pod \"dnsmasq-dns-99495cbc7-v87cl\" (UID: \"507e14a3-601e-461b-b8a7-f218b87b0f47\") " pod="openstack/dnsmasq-dns-99495cbc7-v87cl" Oct 05 07:07:54 crc kubenswrapper[4846]: I1005 07:07:54.263979 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/507e14a3-601e-461b-b8a7-f218b87b0f47-dns-svc\") pod \"dnsmasq-dns-99495cbc7-v87cl\" (UID: \"507e14a3-601e-461b-b8a7-f218b87b0f47\") " pod="openstack/dnsmasq-dns-99495cbc7-v87cl" Oct 05 07:07:54 crc kubenswrapper[4846]: I1005 07:07:54.264026 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/507e14a3-601e-461b-b8a7-f218b87b0f47-config\") pod \"dnsmasq-dns-99495cbc7-v87cl\" (UID: \"507e14a3-601e-461b-b8a7-f218b87b0f47\") " pod="openstack/dnsmasq-dns-99495cbc7-v87cl" Oct 05 07:07:54 crc kubenswrapper[4846]: I1005 07:07:54.264837 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/507e14a3-601e-461b-b8a7-f218b87b0f47-dns-swift-storage-0\") pod \"dnsmasq-dns-99495cbc7-v87cl\" (UID: \"507e14a3-601e-461b-b8a7-f218b87b0f47\") " pod="openstack/dnsmasq-dns-99495cbc7-v87cl" Oct 05 07:07:54 crc kubenswrapper[4846]: I1005 07:07:54.264964 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/507e14a3-601e-461b-b8a7-f218b87b0f47-ovsdbserver-nb\") pod \"dnsmasq-dns-99495cbc7-v87cl\" (UID: \"507e14a3-601e-461b-b8a7-f218b87b0f47\") " pod="openstack/dnsmasq-dns-99495cbc7-v87cl" Oct 05 07:07:54 crc kubenswrapper[4846]: I1005 07:07:54.265005 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/507e14a3-601e-461b-b8a7-f218b87b0f47-config\") pod \"dnsmasq-dns-99495cbc7-v87cl\" (UID: \"507e14a3-601e-461b-b8a7-f218b87b0f47\") " pod="openstack/dnsmasq-dns-99495cbc7-v87cl" Oct 05 07:07:54 crc kubenswrapper[4846]: I1005 07:07:54.265391 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/507e14a3-601e-461b-b8a7-f218b87b0f47-dns-svc\") pod \"dnsmasq-dns-99495cbc7-v87cl\" (UID: \"507e14a3-601e-461b-b8a7-f218b87b0f47\") " pod="openstack/dnsmasq-dns-99495cbc7-v87cl" Oct 05 07:07:54 crc kubenswrapper[4846]: I1005 07:07:54.265521 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/507e14a3-601e-461b-b8a7-f218b87b0f47-ovsdbserver-sb\") pod \"dnsmasq-dns-99495cbc7-v87cl\" (UID: \"507e14a3-601e-461b-b8a7-f218b87b0f47\") " pod="openstack/dnsmasq-dns-99495cbc7-v87cl" Oct 05 07:07:54 crc kubenswrapper[4846]: I1005 07:07:54.283557 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pvqr5\" (UniqueName: \"kubernetes.io/projected/507e14a3-601e-461b-b8a7-f218b87b0f47-kube-api-access-pvqr5\") pod \"dnsmasq-dns-99495cbc7-v87cl\" (UID: \"507e14a3-601e-461b-b8a7-f218b87b0f47\") " pod="openstack/dnsmasq-dns-99495cbc7-v87cl" Oct 05 07:07:54 crc kubenswrapper[4846]: I1005 07:07:54.317088 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 05 07:07:54 crc kubenswrapper[4846]: I1005 07:07:54.366586 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/185f2496-a012-46e6-9387-031b3d92ea5a-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"185f2496-a012-46e6-9387-031b3d92ea5a\") " pod="openstack/cinder-api-0" Oct 05 07:07:54 crc kubenswrapper[4846]: I1005 07:07:54.366647 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/185f2496-a012-46e6-9387-031b3d92ea5a-config-data-custom\") pod \"cinder-api-0\" (UID: \"185f2496-a012-46e6-9387-031b3d92ea5a\") " pod="openstack/cinder-api-0" Oct 05 07:07:54 crc kubenswrapper[4846]: I1005 07:07:54.366925 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/185f2496-a012-46e6-9387-031b3d92ea5a-etc-machine-id\") pod \"cinder-api-0\" (UID: \"185f2496-a012-46e6-9387-031b3d92ea5a\") " pod="openstack/cinder-api-0" Oct 05 07:07:54 crc kubenswrapper[4846]: I1005 07:07:54.367078 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/185f2496-a012-46e6-9387-031b3d92ea5a-logs\") pod \"cinder-api-0\" (UID: \"185f2496-a012-46e6-9387-031b3d92ea5a\") " pod="openstack/cinder-api-0" Oct 05 07:07:54 crc kubenswrapper[4846]: I1005 07:07:54.367118 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/185f2496-a012-46e6-9387-031b3d92ea5a-scripts\") pod \"cinder-api-0\" (UID: \"185f2496-a012-46e6-9387-031b3d92ea5a\") " pod="openstack/cinder-api-0" Oct 05 07:07:54 crc kubenswrapper[4846]: I1005 07:07:54.367326 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/185f2496-a012-46e6-9387-031b3d92ea5a-config-data\") pod \"cinder-api-0\" (UID: \"185f2496-a012-46e6-9387-031b3d92ea5a\") " pod="openstack/cinder-api-0" Oct 05 07:07:54 crc kubenswrapper[4846]: I1005 07:07:54.367593 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bmqpd\" (UniqueName: \"kubernetes.io/projected/185f2496-a012-46e6-9387-031b3d92ea5a-kube-api-access-bmqpd\") pod \"cinder-api-0\" (UID: \"185f2496-a012-46e6-9387-031b3d92ea5a\") " pod="openstack/cinder-api-0" Oct 05 07:07:54 crc kubenswrapper[4846]: I1005 07:07:54.395594 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-99495cbc7-v87cl" Oct 05 07:07:54 crc kubenswrapper[4846]: I1005 07:07:54.474657 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/185f2496-a012-46e6-9387-031b3d92ea5a-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"185f2496-a012-46e6-9387-031b3d92ea5a\") " pod="openstack/cinder-api-0" Oct 05 07:07:54 crc kubenswrapper[4846]: I1005 07:07:54.475113 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/185f2496-a012-46e6-9387-031b3d92ea5a-config-data-custom\") pod \"cinder-api-0\" (UID: \"185f2496-a012-46e6-9387-031b3d92ea5a\") " pod="openstack/cinder-api-0" Oct 05 07:07:54 crc kubenswrapper[4846]: I1005 07:07:54.475170 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/185f2496-a012-46e6-9387-031b3d92ea5a-etc-machine-id\") pod \"cinder-api-0\" (UID: \"185f2496-a012-46e6-9387-031b3d92ea5a\") " pod="openstack/cinder-api-0" Oct 05 07:07:54 crc kubenswrapper[4846]: I1005 07:07:54.475245 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/185f2496-a012-46e6-9387-031b3d92ea5a-logs\") pod \"cinder-api-0\" (UID: \"185f2496-a012-46e6-9387-031b3d92ea5a\") " pod="openstack/cinder-api-0" Oct 05 07:07:54 crc kubenswrapper[4846]: I1005 07:07:54.475267 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/185f2496-a012-46e6-9387-031b3d92ea5a-scripts\") pod \"cinder-api-0\" (UID: \"185f2496-a012-46e6-9387-031b3d92ea5a\") " pod="openstack/cinder-api-0" Oct 05 07:07:54 crc kubenswrapper[4846]: I1005 07:07:54.475315 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/185f2496-a012-46e6-9387-031b3d92ea5a-config-data\") pod \"cinder-api-0\" (UID: \"185f2496-a012-46e6-9387-031b3d92ea5a\") " pod="openstack/cinder-api-0" Oct 05 07:07:54 crc kubenswrapper[4846]: I1005 07:07:54.475363 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bmqpd\" (UniqueName: \"kubernetes.io/projected/185f2496-a012-46e6-9387-031b3d92ea5a-kube-api-access-bmqpd\") pod \"cinder-api-0\" (UID: \"185f2496-a012-46e6-9387-031b3d92ea5a\") " pod="openstack/cinder-api-0" Oct 05 07:07:54 crc kubenswrapper[4846]: I1005 07:07:54.475860 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/185f2496-a012-46e6-9387-031b3d92ea5a-etc-machine-id\") pod \"cinder-api-0\" (UID: \"185f2496-a012-46e6-9387-031b3d92ea5a\") " pod="openstack/cinder-api-0" Oct 05 07:07:54 crc kubenswrapper[4846]: I1005 07:07:54.476375 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/185f2496-a012-46e6-9387-031b3d92ea5a-logs\") pod \"cinder-api-0\" (UID: \"185f2496-a012-46e6-9387-031b3d92ea5a\") " pod="openstack/cinder-api-0" Oct 05 07:07:54 crc kubenswrapper[4846]: I1005 07:07:54.507039 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/185f2496-a012-46e6-9387-031b3d92ea5a-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"185f2496-a012-46e6-9387-031b3d92ea5a\") " pod="openstack/cinder-api-0" Oct 05 07:07:54 crc kubenswrapper[4846]: I1005 07:07:54.509722 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/185f2496-a012-46e6-9387-031b3d92ea5a-scripts\") pod \"cinder-api-0\" (UID: \"185f2496-a012-46e6-9387-031b3d92ea5a\") " pod="openstack/cinder-api-0" Oct 05 07:07:54 crc kubenswrapper[4846]: I1005 07:07:54.510739 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/185f2496-a012-46e6-9387-031b3d92ea5a-config-data\") pod \"cinder-api-0\" (UID: \"185f2496-a012-46e6-9387-031b3d92ea5a\") " pod="openstack/cinder-api-0" Oct 05 07:07:54 crc kubenswrapper[4846]: I1005 07:07:54.512935 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/185f2496-a012-46e6-9387-031b3d92ea5a-config-data-custom\") pod \"cinder-api-0\" (UID: \"185f2496-a012-46e6-9387-031b3d92ea5a\") " pod="openstack/cinder-api-0" Oct 05 07:07:54 crc kubenswrapper[4846]: I1005 07:07:54.513342 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bmqpd\" (UniqueName: \"kubernetes.io/projected/185f2496-a012-46e6-9387-031b3d92ea5a-kube-api-access-bmqpd\") pod \"cinder-api-0\" (UID: \"185f2496-a012-46e6-9387-031b3d92ea5a\") " pod="openstack/cinder-api-0" Oct 05 07:07:54 crc kubenswrapper[4846]: I1005 07:07:54.534947 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-544f66dff9-4zksz" event={"ID":"d5c6818d-305f-4565-a245-672dde7ddfd2","Type":"ContainerStarted","Data":"a8e6157724c484c9ef9627c7510a1d3832e632c213f2f1252b686228eb9d95c2"} Oct 05 07:07:54 crc kubenswrapper[4846]: I1005 07:07:54.535146 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-544f66dff9-4zksz" podUID="d5c6818d-305f-4565-a245-672dde7ddfd2" containerName="dnsmasq-dns" containerID="cri-o://a8e6157724c484c9ef9627c7510a1d3832e632c213f2f1252b686228eb9d95c2" gracePeriod=10 Oct 05 07:07:54 crc kubenswrapper[4846]: I1005 07:07:54.535432 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-544f66dff9-4zksz" Oct 05 07:07:54 crc kubenswrapper[4846]: I1005 07:07:54.549372 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6c9ddc8584-bqgtc" event={"ID":"ba9102c0-185f-4f02-8b05-8bb7fde2d573","Type":"ContainerStarted","Data":"e5bd6644aa631daa23e2517be3eb2c9e0b0f0672ae63eae56fb2b30510a9cb24"} Oct 05 07:07:54 crc kubenswrapper[4846]: I1005 07:07:54.549416 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6c9ddc8584-bqgtc" event={"ID":"ba9102c0-185f-4f02-8b05-8bb7fde2d573","Type":"ContainerStarted","Data":"39111fc9d7fc57d402d3aff9ff623348099e37f99ec5b23eaa70a56e39c540b9"} Oct 05 07:07:54 crc kubenswrapper[4846]: I1005 07:07:54.550356 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-6c9ddc8584-bqgtc" Oct 05 07:07:54 crc kubenswrapper[4846]: I1005 07:07:54.550477 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-6c9ddc8584-bqgtc" Oct 05 07:07:54 crc kubenswrapper[4846]: I1005 07:07:54.551915 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 05 07:07:54 crc kubenswrapper[4846]: I1005 07:07:54.555955 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"39071bf4-214b-4bbb-8b78-dfec6ad3c940","Type":"ContainerStarted","Data":"82e6c4bfad41089daffabb46ada2064751a4f47c40780fda32f8ec7989bd6a30"} Oct 05 07:07:54 crc kubenswrapper[4846]: I1005 07:07:54.569130 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-544f66dff9-4zksz" podStartSLOduration=3.569110812 podStartE2EDuration="3.569110812s" podCreationTimestamp="2025-10-05 07:07:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 07:07:54.565801064 +0000 UTC m=+1196.806653839" watchObservedRunningTime="2025-10-05 07:07:54.569110812 +0000 UTC m=+1196.809963587" Oct 05 07:07:54 crc kubenswrapper[4846]: I1005 07:07:54.627076 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-6c9ddc8584-bqgtc" podStartSLOduration=3.627048138 podStartE2EDuration="3.627048138s" podCreationTimestamp="2025-10-05 07:07:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 07:07:54.616543038 +0000 UTC m=+1196.857395813" watchObservedRunningTime="2025-10-05 07:07:54.627048138 +0000 UTC m=+1196.867900913" Oct 05 07:07:54 crc kubenswrapper[4846]: I1005 07:07:54.930377 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 05 07:07:55 crc kubenswrapper[4846]: I1005 07:07:55.581693 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"40289760-4ac7-42e8-8eab-39f93c1618aa","Type":"ContainerStarted","Data":"9a9217ce018fb0e4213096c3c7fec98ee147392235029f6fd87273173ab45ce6"} Oct 05 07:07:55 crc kubenswrapper[4846]: I1005 07:07:55.584199 4846 generic.go:334] "Generic (PLEG): container finished" podID="d5c6818d-305f-4565-a245-672dde7ddfd2" containerID="a8e6157724c484c9ef9627c7510a1d3832e632c213f2f1252b686228eb9d95c2" exitCode=0 Oct 05 07:07:55 crc kubenswrapper[4846]: I1005 07:07:55.585285 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-544f66dff9-4zksz" event={"ID":"d5c6818d-305f-4565-a245-672dde7ddfd2","Type":"ContainerDied","Data":"a8e6157724c484c9ef9627c7510a1d3832e632c213f2f1252b686228eb9d95c2"} Oct 05 07:07:55 crc kubenswrapper[4846]: I1005 07:07:55.936335 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 05 07:07:55 crc kubenswrapper[4846]: W1005 07:07:55.983192 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod185f2496_a012_46e6_9387_031b3d92ea5a.slice/crio-a7c790abe290362ad8139a98bcdf2a6d418ae615a9b48e5227ffda0cc2ba6295 WatchSource:0}: Error finding container a7c790abe290362ad8139a98bcdf2a6d418ae615a9b48e5227ffda0cc2ba6295: Status 404 returned error can't find the container with id a7c790abe290362ad8139a98bcdf2a6d418ae615a9b48e5227ffda0cc2ba6295 Oct 05 07:07:56 crc kubenswrapper[4846]: I1005 07:07:56.092022 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-544f66dff9-4zksz" Oct 05 07:07:56 crc kubenswrapper[4846]: I1005 07:07:56.236119 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d5c6818d-305f-4565-a245-672dde7ddfd2-ovsdbserver-nb\") pod \"d5c6818d-305f-4565-a245-672dde7ddfd2\" (UID: \"d5c6818d-305f-4565-a245-672dde7ddfd2\") " Oct 05 07:07:56 crc kubenswrapper[4846]: I1005 07:07:56.236362 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2t6b8\" (UniqueName: \"kubernetes.io/projected/d5c6818d-305f-4565-a245-672dde7ddfd2-kube-api-access-2t6b8\") pod \"d5c6818d-305f-4565-a245-672dde7ddfd2\" (UID: \"d5c6818d-305f-4565-a245-672dde7ddfd2\") " Oct 05 07:07:56 crc kubenswrapper[4846]: I1005 07:07:56.237861 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d5c6818d-305f-4565-a245-672dde7ddfd2-dns-swift-storage-0\") pod \"d5c6818d-305f-4565-a245-672dde7ddfd2\" (UID: \"d5c6818d-305f-4565-a245-672dde7ddfd2\") " Oct 05 07:07:56 crc kubenswrapper[4846]: I1005 07:07:56.237930 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d5c6818d-305f-4565-a245-672dde7ddfd2-dns-svc\") pod \"d5c6818d-305f-4565-a245-672dde7ddfd2\" (UID: \"d5c6818d-305f-4565-a245-672dde7ddfd2\") " Oct 05 07:07:56 crc kubenswrapper[4846]: I1005 07:07:56.237952 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d5c6818d-305f-4565-a245-672dde7ddfd2-config\") pod \"d5c6818d-305f-4565-a245-672dde7ddfd2\" (UID: \"d5c6818d-305f-4565-a245-672dde7ddfd2\") " Oct 05 07:07:56 crc kubenswrapper[4846]: I1005 07:07:56.238101 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d5c6818d-305f-4565-a245-672dde7ddfd2-ovsdbserver-sb\") pod \"d5c6818d-305f-4565-a245-672dde7ddfd2\" (UID: \"d5c6818d-305f-4565-a245-672dde7ddfd2\") " Oct 05 07:07:56 crc kubenswrapper[4846]: I1005 07:07:56.254652 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d5c6818d-305f-4565-a245-672dde7ddfd2-kube-api-access-2t6b8" (OuterVolumeSpecName: "kube-api-access-2t6b8") pod "d5c6818d-305f-4565-a245-672dde7ddfd2" (UID: "d5c6818d-305f-4565-a245-672dde7ddfd2"). InnerVolumeSpecName "kube-api-access-2t6b8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:07:56 crc kubenswrapper[4846]: I1005 07:07:56.323531 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d5c6818d-305f-4565-a245-672dde7ddfd2-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "d5c6818d-305f-4565-a245-672dde7ddfd2" (UID: "d5c6818d-305f-4565-a245-672dde7ddfd2"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 07:07:56 crc kubenswrapper[4846]: I1005 07:07:56.324277 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d5c6818d-305f-4565-a245-672dde7ddfd2-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "d5c6818d-305f-4565-a245-672dde7ddfd2" (UID: "d5c6818d-305f-4565-a245-672dde7ddfd2"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 07:07:56 crc kubenswrapper[4846]: I1005 07:07:56.339480 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d5c6818d-305f-4565-a245-672dde7ddfd2-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "d5c6818d-305f-4565-a245-672dde7ddfd2" (UID: "d5c6818d-305f-4565-a245-672dde7ddfd2"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 07:07:56 crc kubenswrapper[4846]: I1005 07:07:56.340079 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d5c6818d-305f-4565-a245-672dde7ddfd2-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "d5c6818d-305f-4565-a245-672dde7ddfd2" (UID: "d5c6818d-305f-4565-a245-672dde7ddfd2"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 07:07:56 crc kubenswrapper[4846]: I1005 07:07:56.341516 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d5c6818d-305f-4565-a245-672dde7ddfd2-ovsdbserver-sb\") pod \"d5c6818d-305f-4565-a245-672dde7ddfd2\" (UID: \"d5c6818d-305f-4565-a245-672dde7ddfd2\") " Oct 05 07:07:56 crc kubenswrapper[4846]: W1005 07:07:56.341630 4846 empty_dir.go:500] Warning: Unmount skipped because path does not exist: /var/lib/kubelet/pods/d5c6818d-305f-4565-a245-672dde7ddfd2/volumes/kubernetes.io~configmap/ovsdbserver-sb Oct 05 07:07:56 crc kubenswrapper[4846]: I1005 07:07:56.341646 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d5c6818d-305f-4565-a245-672dde7ddfd2-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "d5c6818d-305f-4565-a245-672dde7ddfd2" (UID: "d5c6818d-305f-4565-a245-672dde7ddfd2"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 07:07:56 crc kubenswrapper[4846]: I1005 07:07:56.341794 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d5c6818d-305f-4565-a245-672dde7ddfd2-dns-svc\") pod \"d5c6818d-305f-4565-a245-672dde7ddfd2\" (UID: \"d5c6818d-305f-4565-a245-672dde7ddfd2\") " Oct 05 07:07:56 crc kubenswrapper[4846]: W1005 07:07:56.341852 4846 empty_dir.go:500] Warning: Unmount skipped because path does not exist: /var/lib/kubelet/pods/d5c6818d-305f-4565-a245-672dde7ddfd2/volumes/kubernetes.io~configmap/dns-svc Oct 05 07:07:56 crc kubenswrapper[4846]: I1005 07:07:56.341865 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d5c6818d-305f-4565-a245-672dde7ddfd2-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "d5c6818d-305f-4565-a245-672dde7ddfd2" (UID: "d5c6818d-305f-4565-a245-672dde7ddfd2"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 07:07:56 crc kubenswrapper[4846]: I1005 07:07:56.342713 4846 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d5c6818d-305f-4565-a245-672dde7ddfd2-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 05 07:07:56 crc kubenswrapper[4846]: I1005 07:07:56.342734 4846 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d5c6818d-305f-4565-a245-672dde7ddfd2-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 05 07:07:56 crc kubenswrapper[4846]: I1005 07:07:56.342745 4846 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d5c6818d-305f-4565-a245-672dde7ddfd2-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 05 07:07:56 crc kubenswrapper[4846]: I1005 07:07:56.342757 4846 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d5c6818d-305f-4565-a245-672dde7ddfd2-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 05 07:07:56 crc kubenswrapper[4846]: I1005 07:07:56.342766 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2t6b8\" (UniqueName: \"kubernetes.io/projected/d5c6818d-305f-4565-a245-672dde7ddfd2-kube-api-access-2t6b8\") on node \"crc\" DevicePath \"\"" Oct 05 07:07:56 crc kubenswrapper[4846]: I1005 07:07:56.355866 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d5c6818d-305f-4565-a245-672dde7ddfd2-config" (OuterVolumeSpecName: "config") pod "d5c6818d-305f-4565-a245-672dde7ddfd2" (UID: "d5c6818d-305f-4565-a245-672dde7ddfd2"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 07:07:56 crc kubenswrapper[4846]: I1005 07:07:56.432269 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-99495cbc7-v87cl"] Oct 05 07:07:56 crc kubenswrapper[4846]: W1005 07:07:56.444368 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod507e14a3_601e_461b_b8a7_f218b87b0f47.slice/crio-377f7635d785623b7b156209f59b159fad7649c1df29afbf5d93db2d4f758a93 WatchSource:0}: Error finding container 377f7635d785623b7b156209f59b159fad7649c1df29afbf5d93db2d4f758a93: Status 404 returned error can't find the container with id 377f7635d785623b7b156209f59b159fad7649c1df29afbf5d93db2d4f758a93 Oct 05 07:07:56 crc kubenswrapper[4846]: I1005 07:07:56.445699 4846 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d5c6818d-305f-4565-a245-672dde7ddfd2-config\") on node \"crc\" DevicePath \"\"" Oct 05 07:07:56 crc kubenswrapper[4846]: I1005 07:07:56.604166 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-99495cbc7-v87cl" event={"ID":"507e14a3-601e-461b-b8a7-f218b87b0f47","Type":"ContainerStarted","Data":"377f7635d785623b7b156209f59b159fad7649c1df29afbf5d93db2d4f758a93"} Oct 05 07:07:56 crc kubenswrapper[4846]: I1005 07:07:56.615925 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-77664b8dfd-5hkqj" event={"ID":"8d55f55a-2d79-4c01-8b0f-d67cefb842ff","Type":"ContainerStarted","Data":"bca1da26068e641006909835041107dfa8983891c244081dc35a6bf70c412d1d"} Oct 05 07:07:56 crc kubenswrapper[4846]: I1005 07:07:56.618113 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"185f2496-a012-46e6-9387-031b3d92ea5a","Type":"ContainerStarted","Data":"a7c790abe290362ad8139a98bcdf2a6d418ae615a9b48e5227ffda0cc2ba6295"} Oct 05 07:07:56 crc kubenswrapper[4846]: I1005 07:07:56.621610 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-78995b9f79-nwl7v" event={"ID":"3995994c-2a3f-400f-9a64-b564759b20b2","Type":"ContainerStarted","Data":"afd676f2477334d96744cd37d02c69587dea11053d28e2ae166c10deda1b1560"} Oct 05 07:07:56 crc kubenswrapper[4846]: I1005 07:07:56.624102 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-544f66dff9-4zksz" event={"ID":"d5c6818d-305f-4565-a245-672dde7ddfd2","Type":"ContainerDied","Data":"d28c31c68266f89b66543ff28dcc6676976c4a29c1f41d2d842b3be155593cdd"} Oct 05 07:07:56 crc kubenswrapper[4846]: I1005 07:07:56.624141 4846 scope.go:117] "RemoveContainer" containerID="a8e6157724c484c9ef9627c7510a1d3832e632c213f2f1252b686228eb9d95c2" Oct 05 07:07:56 crc kubenswrapper[4846]: I1005 07:07:56.624308 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-544f66dff9-4zksz" Oct 05 07:07:56 crc kubenswrapper[4846]: I1005 07:07:56.657545 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-544f66dff9-4zksz"] Oct 05 07:07:56 crc kubenswrapper[4846]: I1005 07:07:56.665805 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-544f66dff9-4zksz"] Oct 05 07:07:56 crc kubenswrapper[4846]: I1005 07:07:56.679269 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"39071bf4-214b-4bbb-8b78-dfec6ad3c940","Type":"ContainerStarted","Data":"72d54baff0ab5bf146fb7a95307526bd069594408c4da5a11bbeacf29306be6c"} Oct 05 07:07:56 crc kubenswrapper[4846]: I1005 07:07:56.679504 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 05 07:07:56 crc kubenswrapper[4846]: I1005 07:07:56.696439 4846 scope.go:117] "RemoveContainer" containerID="b8dce4eabd2404d514a869915acc0811e9f294ad356a82f4ea75e9c5d4ee7ca9" Oct 05 07:07:56 crc kubenswrapper[4846]: I1005 07:07:56.979291 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Oct 05 07:07:56 crc kubenswrapper[4846]: I1005 07:07:56.979940 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Oct 05 07:07:57 crc kubenswrapper[4846]: I1005 07:07:57.045522 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Oct 05 07:07:57 crc kubenswrapper[4846]: I1005 07:07:57.052534 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Oct 05 07:07:57 crc kubenswrapper[4846]: I1005 07:07:57.084211 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.366318446 podStartE2EDuration="7.084171329s" podCreationTimestamp="2025-10-05 07:07:50 +0000 UTC" firstStartedPulling="2025-10-05 07:07:51.362263954 +0000 UTC m=+1193.603116729" lastFinishedPulling="2025-10-05 07:07:56.080116807 +0000 UTC m=+1198.320969612" observedRunningTime="2025-10-05 07:07:56.718355931 +0000 UTC m=+1198.959208706" watchObservedRunningTime="2025-10-05 07:07:57.084171329 +0000 UTC m=+1199.325024104" Oct 05 07:07:57 crc kubenswrapper[4846]: I1005 07:07:57.339821 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-64c4f98476-pqrx9" Oct 05 07:07:57 crc kubenswrapper[4846]: I1005 07:07:57.696596 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-78995b9f79-nwl7v" event={"ID":"3995994c-2a3f-400f-9a64-b564759b20b2","Type":"ContainerStarted","Data":"5cae68aeaeb33c53f345bf863894fe1192fa80a8ec2b486a6c2d29e83dec72af"} Oct 05 07:07:57 crc kubenswrapper[4846]: I1005 07:07:57.705524 4846 generic.go:334] "Generic (PLEG): container finished" podID="507e14a3-601e-461b-b8a7-f218b87b0f47" containerID="d650d5bbcb222ec9724e83b217d5482ce80a31ef9cbe9e92ec328f485017d0e1" exitCode=0 Oct 05 07:07:57 crc kubenswrapper[4846]: I1005 07:07:57.705592 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-99495cbc7-v87cl" event={"ID":"507e14a3-601e-461b-b8a7-f218b87b0f47","Type":"ContainerDied","Data":"d650d5bbcb222ec9724e83b217d5482ce80a31ef9cbe9e92ec328f485017d0e1"} Oct 05 07:07:57 crc kubenswrapper[4846]: I1005 07:07:57.713432 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"40289760-4ac7-42e8-8eab-39f93c1618aa","Type":"ContainerStarted","Data":"fd9c22e44bb6d2c4d688311f3f8f67ffbf4f78d9f27e0d6baa248cd31e5ddfae"} Oct 05 07:07:57 crc kubenswrapper[4846]: I1005 07:07:57.721898 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-worker-78995b9f79-nwl7v" podStartSLOduration=3.2590028 podStartE2EDuration="6.721861878s" podCreationTimestamp="2025-10-05 07:07:51 +0000 UTC" firstStartedPulling="2025-10-05 07:07:52.53705006 +0000 UTC m=+1194.777902835" lastFinishedPulling="2025-10-05 07:07:55.999909128 +0000 UTC m=+1198.240761913" observedRunningTime="2025-10-05 07:07:57.714925533 +0000 UTC m=+1199.955778298" watchObservedRunningTime="2025-10-05 07:07:57.721861878 +0000 UTC m=+1199.962714643" Oct 05 07:07:57 crc kubenswrapper[4846]: I1005 07:07:57.723029 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-77664b8dfd-5hkqj" event={"ID":"8d55f55a-2d79-4c01-8b0f-d67cefb842ff","Type":"ContainerStarted","Data":"6a990887fb132123ab2aa7e7c062e5fbecb25e29a746e6d59d303ed6848ccde0"} Oct 05 07:07:57 crc kubenswrapper[4846]: I1005 07:07:57.727218 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"185f2496-a012-46e6-9387-031b3d92ea5a","Type":"ContainerStarted","Data":"368a579434676bb95c41afb3688944e9fa1d6b4a41cfb09c70c12f7641134de5"} Oct 05 07:07:57 crc kubenswrapper[4846]: I1005 07:07:57.728899 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Oct 05 07:07:57 crc kubenswrapper[4846]: I1005 07:07:57.728998 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Oct 05 07:07:57 crc kubenswrapper[4846]: I1005 07:07:57.798375 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-keystone-listener-77664b8dfd-5hkqj" podStartSLOduration=3.374885262 podStartE2EDuration="6.798342658s" podCreationTimestamp="2025-10-05 07:07:51 +0000 UTC" firstStartedPulling="2025-10-05 07:07:52.616027237 +0000 UTC m=+1194.856880012" lastFinishedPulling="2025-10-05 07:07:56.039484633 +0000 UTC m=+1198.280337408" observedRunningTime="2025-10-05 07:07:57.765825041 +0000 UTC m=+1200.006677816" watchObservedRunningTime="2025-10-05 07:07:57.798342658 +0000 UTC m=+1200.039195433" Oct 05 07:07:58 crc kubenswrapper[4846]: I1005 07:07:58.261771 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Oct 05 07:07:58 crc kubenswrapper[4846]: I1005 07:07:58.522011 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d5c6818d-305f-4565-a245-672dde7ddfd2" path="/var/lib/kubelet/pods/d5c6818d-305f-4565-a245-672dde7ddfd2/volumes" Oct 05 07:07:58 crc kubenswrapper[4846]: I1005 07:07:58.741320 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-99495cbc7-v87cl" event={"ID":"507e14a3-601e-461b-b8a7-f218b87b0f47","Type":"ContainerStarted","Data":"efc09bde05d0841f423bbc79bab49db3b7705ff627f508b06fdf9dfef64469a0"} Oct 05 07:07:58 crc kubenswrapper[4846]: I1005 07:07:58.743446 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-99495cbc7-v87cl" Oct 05 07:07:58 crc kubenswrapper[4846]: I1005 07:07:58.744423 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"40289760-4ac7-42e8-8eab-39f93c1618aa","Type":"ContainerStarted","Data":"03eaa092d41b9b080f0832b3cfdb8fb94047441e2dd9f8972556e0afa700034e"} Oct 05 07:07:58 crc kubenswrapper[4846]: I1005 07:07:58.749386 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"185f2496-a012-46e6-9387-031b3d92ea5a","Type":"ContainerStarted","Data":"b7c2d5a4f9c46f6b07cf21fc7f29aa99808d5ba4c0a17af8fe48d435039bf13b"} Oct 05 07:07:58 crc kubenswrapper[4846]: I1005 07:07:58.749547 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="185f2496-a012-46e6-9387-031b3d92ea5a" containerName="cinder-api-log" containerID="cri-o://368a579434676bb95c41afb3688944e9fa1d6b4a41cfb09c70c12f7641134de5" gracePeriod=30 Oct 05 07:07:58 crc kubenswrapper[4846]: I1005 07:07:58.749666 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Oct 05 07:07:58 crc kubenswrapper[4846]: I1005 07:07:58.749707 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="185f2496-a012-46e6-9387-031b3d92ea5a" containerName="cinder-api" containerID="cri-o://b7c2d5a4f9c46f6b07cf21fc7f29aa99808d5ba4c0a17af8fe48d435039bf13b" gracePeriod=30 Oct 05 07:07:58 crc kubenswrapper[4846]: I1005 07:07:58.786914 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Oct 05 07:07:58 crc kubenswrapper[4846]: I1005 07:07:58.786978 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Oct 05 07:07:58 crc kubenswrapper[4846]: I1005 07:07:58.792947 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-99495cbc7-v87cl" podStartSLOduration=5.792916766 podStartE2EDuration="5.792916766s" podCreationTimestamp="2025-10-05 07:07:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 07:07:58.773054886 +0000 UTC m=+1201.013907681" watchObservedRunningTime="2025-10-05 07:07:58.792916766 +0000 UTC m=+1201.033769541" Oct 05 07:07:58 crc kubenswrapper[4846]: I1005 07:07:58.813700 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-57674d6c54-zn5nq"] Oct 05 07:07:58 crc kubenswrapper[4846]: E1005 07:07:58.814727 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d5c6818d-305f-4565-a245-672dde7ddfd2" containerName="dnsmasq-dns" Oct 05 07:07:58 crc kubenswrapper[4846]: I1005 07:07:58.814743 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="d5c6818d-305f-4565-a245-672dde7ddfd2" containerName="dnsmasq-dns" Oct 05 07:07:58 crc kubenswrapper[4846]: E1005 07:07:58.814804 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d5c6818d-305f-4565-a245-672dde7ddfd2" containerName="init" Oct 05 07:07:58 crc kubenswrapper[4846]: I1005 07:07:58.814811 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="d5c6818d-305f-4565-a245-672dde7ddfd2" containerName="init" Oct 05 07:07:58 crc kubenswrapper[4846]: I1005 07:07:58.815358 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="d5c6818d-305f-4565-a245-672dde7ddfd2" containerName="dnsmasq-dns" Oct 05 07:07:58 crc kubenswrapper[4846]: I1005 07:07:58.830710 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-57674d6c54-zn5nq" Oct 05 07:07:58 crc kubenswrapper[4846]: I1005 07:07:58.839591 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-internal-svc" Oct 05 07:07:58 crc kubenswrapper[4846]: I1005 07:07:58.839849 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-public-svc" Oct 05 07:07:58 crc kubenswrapper[4846]: I1005 07:07:58.864321 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Oct 05 07:07:58 crc kubenswrapper[4846]: I1005 07:07:58.896251 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-57674d6c54-zn5nq"] Oct 05 07:07:58 crc kubenswrapper[4846]: I1005 07:07:58.896613 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Oct 05 07:07:58 crc kubenswrapper[4846]: I1005 07:07:58.957957 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=4.957922337 podStartE2EDuration="4.957922337s" podCreationTimestamp="2025-10-05 07:07:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 07:07:58.819583827 +0000 UTC m=+1201.060436602" watchObservedRunningTime="2025-10-05 07:07:58.957922337 +0000 UTC m=+1201.198775112" Oct 05 07:07:58 crc kubenswrapper[4846]: I1005 07:07:58.967688 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c31937eb-9836-461b-9230-5e06ce7a99e8-internal-tls-certs\") pod \"barbican-api-57674d6c54-zn5nq\" (UID: \"c31937eb-9836-461b-9230-5e06ce7a99e8\") " pod="openstack/barbican-api-57674d6c54-zn5nq" Oct 05 07:07:58 crc kubenswrapper[4846]: I1005 07:07:58.967785 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c31937eb-9836-461b-9230-5e06ce7a99e8-logs\") pod \"barbican-api-57674d6c54-zn5nq\" (UID: \"c31937eb-9836-461b-9230-5e06ce7a99e8\") " pod="openstack/barbican-api-57674d6c54-zn5nq" Oct 05 07:07:58 crc kubenswrapper[4846]: I1005 07:07:58.967821 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c31937eb-9836-461b-9230-5e06ce7a99e8-config-data\") pod \"barbican-api-57674d6c54-zn5nq\" (UID: \"c31937eb-9836-461b-9230-5e06ce7a99e8\") " pod="openstack/barbican-api-57674d6c54-zn5nq" Oct 05 07:07:58 crc kubenswrapper[4846]: I1005 07:07:58.967837 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c31937eb-9836-461b-9230-5e06ce7a99e8-combined-ca-bundle\") pod \"barbican-api-57674d6c54-zn5nq\" (UID: \"c31937eb-9836-461b-9230-5e06ce7a99e8\") " pod="openstack/barbican-api-57674d6c54-zn5nq" Oct 05 07:07:58 crc kubenswrapper[4846]: I1005 07:07:58.967893 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c31937eb-9836-461b-9230-5e06ce7a99e8-public-tls-certs\") pod \"barbican-api-57674d6c54-zn5nq\" (UID: \"c31937eb-9836-461b-9230-5e06ce7a99e8\") " pod="openstack/barbican-api-57674d6c54-zn5nq" Oct 05 07:07:58 crc kubenswrapper[4846]: I1005 07:07:58.968000 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x6fpj\" (UniqueName: \"kubernetes.io/projected/c31937eb-9836-461b-9230-5e06ce7a99e8-kube-api-access-x6fpj\") pod \"barbican-api-57674d6c54-zn5nq\" (UID: \"c31937eb-9836-461b-9230-5e06ce7a99e8\") " pod="openstack/barbican-api-57674d6c54-zn5nq" Oct 05 07:07:58 crc kubenswrapper[4846]: I1005 07:07:58.968041 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c31937eb-9836-461b-9230-5e06ce7a99e8-config-data-custom\") pod \"barbican-api-57674d6c54-zn5nq\" (UID: \"c31937eb-9836-461b-9230-5e06ce7a99e8\") " pod="openstack/barbican-api-57674d6c54-zn5nq" Oct 05 07:07:58 crc kubenswrapper[4846]: I1005 07:07:58.971114 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=4.9866529 podStartE2EDuration="5.970950345s" podCreationTimestamp="2025-10-05 07:07:53 +0000 UTC" firstStartedPulling="2025-10-05 07:07:55.347110225 +0000 UTC m=+1197.587963000" lastFinishedPulling="2025-10-05 07:07:56.33140767 +0000 UTC m=+1198.572260445" observedRunningTime="2025-10-05 07:07:58.851907559 +0000 UTC m=+1201.092760334" watchObservedRunningTime="2025-10-05 07:07:58.970950345 +0000 UTC m=+1201.211803120" Oct 05 07:07:59 crc kubenswrapper[4846]: I1005 07:07:59.069512 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c31937eb-9836-461b-9230-5e06ce7a99e8-public-tls-certs\") pod \"barbican-api-57674d6c54-zn5nq\" (UID: \"c31937eb-9836-461b-9230-5e06ce7a99e8\") " pod="openstack/barbican-api-57674d6c54-zn5nq" Oct 05 07:07:59 crc kubenswrapper[4846]: I1005 07:07:59.069601 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x6fpj\" (UniqueName: \"kubernetes.io/projected/c31937eb-9836-461b-9230-5e06ce7a99e8-kube-api-access-x6fpj\") pod \"barbican-api-57674d6c54-zn5nq\" (UID: \"c31937eb-9836-461b-9230-5e06ce7a99e8\") " pod="openstack/barbican-api-57674d6c54-zn5nq" Oct 05 07:07:59 crc kubenswrapper[4846]: I1005 07:07:59.069624 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c31937eb-9836-461b-9230-5e06ce7a99e8-config-data-custom\") pod \"barbican-api-57674d6c54-zn5nq\" (UID: \"c31937eb-9836-461b-9230-5e06ce7a99e8\") " pod="openstack/barbican-api-57674d6c54-zn5nq" Oct 05 07:07:59 crc kubenswrapper[4846]: I1005 07:07:59.069700 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c31937eb-9836-461b-9230-5e06ce7a99e8-internal-tls-certs\") pod \"barbican-api-57674d6c54-zn5nq\" (UID: \"c31937eb-9836-461b-9230-5e06ce7a99e8\") " pod="openstack/barbican-api-57674d6c54-zn5nq" Oct 05 07:07:59 crc kubenswrapper[4846]: I1005 07:07:59.069742 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c31937eb-9836-461b-9230-5e06ce7a99e8-logs\") pod \"barbican-api-57674d6c54-zn5nq\" (UID: \"c31937eb-9836-461b-9230-5e06ce7a99e8\") " pod="openstack/barbican-api-57674d6c54-zn5nq" Oct 05 07:07:59 crc kubenswrapper[4846]: I1005 07:07:59.069770 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c31937eb-9836-461b-9230-5e06ce7a99e8-combined-ca-bundle\") pod \"barbican-api-57674d6c54-zn5nq\" (UID: \"c31937eb-9836-461b-9230-5e06ce7a99e8\") " pod="openstack/barbican-api-57674d6c54-zn5nq" Oct 05 07:07:59 crc kubenswrapper[4846]: I1005 07:07:59.069801 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c31937eb-9836-461b-9230-5e06ce7a99e8-config-data\") pod \"barbican-api-57674d6c54-zn5nq\" (UID: \"c31937eb-9836-461b-9230-5e06ce7a99e8\") " pod="openstack/barbican-api-57674d6c54-zn5nq" Oct 05 07:07:59 crc kubenswrapper[4846]: I1005 07:07:59.070912 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c31937eb-9836-461b-9230-5e06ce7a99e8-logs\") pod \"barbican-api-57674d6c54-zn5nq\" (UID: \"c31937eb-9836-461b-9230-5e06ce7a99e8\") " pod="openstack/barbican-api-57674d6c54-zn5nq" Oct 05 07:07:59 crc kubenswrapper[4846]: I1005 07:07:59.078079 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c31937eb-9836-461b-9230-5e06ce7a99e8-internal-tls-certs\") pod \"barbican-api-57674d6c54-zn5nq\" (UID: \"c31937eb-9836-461b-9230-5e06ce7a99e8\") " pod="openstack/barbican-api-57674d6c54-zn5nq" Oct 05 07:07:59 crc kubenswrapper[4846]: I1005 07:07:59.078997 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c31937eb-9836-461b-9230-5e06ce7a99e8-combined-ca-bundle\") pod \"barbican-api-57674d6c54-zn5nq\" (UID: \"c31937eb-9836-461b-9230-5e06ce7a99e8\") " pod="openstack/barbican-api-57674d6c54-zn5nq" Oct 05 07:07:59 crc kubenswrapper[4846]: I1005 07:07:59.080778 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c31937eb-9836-461b-9230-5e06ce7a99e8-config-data\") pod \"barbican-api-57674d6c54-zn5nq\" (UID: \"c31937eb-9836-461b-9230-5e06ce7a99e8\") " pod="openstack/barbican-api-57674d6c54-zn5nq" Oct 05 07:07:59 crc kubenswrapper[4846]: I1005 07:07:59.084022 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c31937eb-9836-461b-9230-5e06ce7a99e8-config-data-custom\") pod \"barbican-api-57674d6c54-zn5nq\" (UID: \"c31937eb-9836-461b-9230-5e06ce7a99e8\") " pod="openstack/barbican-api-57674d6c54-zn5nq" Oct 05 07:07:59 crc kubenswrapper[4846]: I1005 07:07:59.093551 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c31937eb-9836-461b-9230-5e06ce7a99e8-public-tls-certs\") pod \"barbican-api-57674d6c54-zn5nq\" (UID: \"c31937eb-9836-461b-9230-5e06ce7a99e8\") " pod="openstack/barbican-api-57674d6c54-zn5nq" Oct 05 07:07:59 crc kubenswrapper[4846]: I1005 07:07:59.098853 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x6fpj\" (UniqueName: \"kubernetes.io/projected/c31937eb-9836-461b-9230-5e06ce7a99e8-kube-api-access-x6fpj\") pod \"barbican-api-57674d6c54-zn5nq\" (UID: \"c31937eb-9836-461b-9230-5e06ce7a99e8\") " pod="openstack/barbican-api-57674d6c54-zn5nq" Oct 05 07:07:59 crc kubenswrapper[4846]: I1005 07:07:59.284988 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-57674d6c54-zn5nq" Oct 05 07:07:59 crc kubenswrapper[4846]: I1005 07:07:59.328582 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Oct 05 07:07:59 crc kubenswrapper[4846]: I1005 07:07:59.783721 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"185f2496-a012-46e6-9387-031b3d92ea5a","Type":"ContainerDied","Data":"b7c2d5a4f9c46f6b07cf21fc7f29aa99808d5ba4c0a17af8fe48d435039bf13b"} Oct 05 07:07:59 crc kubenswrapper[4846]: I1005 07:07:59.783651 4846 generic.go:334] "Generic (PLEG): container finished" podID="185f2496-a012-46e6-9387-031b3d92ea5a" containerID="b7c2d5a4f9c46f6b07cf21fc7f29aa99808d5ba4c0a17af8fe48d435039bf13b" exitCode=0 Oct 05 07:07:59 crc kubenswrapper[4846]: I1005 07:07:59.783820 4846 generic.go:334] "Generic (PLEG): container finished" podID="185f2496-a012-46e6-9387-031b3d92ea5a" containerID="368a579434676bb95c41afb3688944e9fa1d6b4a41cfb09c70c12f7641134de5" exitCode=143 Oct 05 07:07:59 crc kubenswrapper[4846]: I1005 07:07:59.783922 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"185f2496-a012-46e6-9387-031b3d92ea5a","Type":"ContainerDied","Data":"368a579434676bb95c41afb3688944e9fa1d6b4a41cfb09c70c12f7641134de5"} Oct 05 07:07:59 crc kubenswrapper[4846]: I1005 07:07:59.785579 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Oct 05 07:07:59 crc kubenswrapper[4846]: I1005 07:07:59.785602 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Oct 05 07:07:59 crc kubenswrapper[4846]: I1005 07:07:59.856583 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-57674d6c54-zn5nq"] Oct 05 07:07:59 crc kubenswrapper[4846]: I1005 07:07:59.946517 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 05 07:08:00 crc kubenswrapper[4846]: I1005 07:08:00.001964 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/185f2496-a012-46e6-9387-031b3d92ea5a-config-data\") pod \"185f2496-a012-46e6-9387-031b3d92ea5a\" (UID: \"185f2496-a012-46e6-9387-031b3d92ea5a\") " Oct 05 07:08:00 crc kubenswrapper[4846]: I1005 07:08:00.002060 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/185f2496-a012-46e6-9387-031b3d92ea5a-config-data-custom\") pod \"185f2496-a012-46e6-9387-031b3d92ea5a\" (UID: \"185f2496-a012-46e6-9387-031b3d92ea5a\") " Oct 05 07:08:00 crc kubenswrapper[4846]: I1005 07:08:00.002106 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/185f2496-a012-46e6-9387-031b3d92ea5a-combined-ca-bundle\") pod \"185f2496-a012-46e6-9387-031b3d92ea5a\" (UID: \"185f2496-a012-46e6-9387-031b3d92ea5a\") " Oct 05 07:08:00 crc kubenswrapper[4846]: I1005 07:08:00.002139 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/185f2496-a012-46e6-9387-031b3d92ea5a-logs\") pod \"185f2496-a012-46e6-9387-031b3d92ea5a\" (UID: \"185f2496-a012-46e6-9387-031b3d92ea5a\") " Oct 05 07:08:00 crc kubenswrapper[4846]: I1005 07:08:00.002161 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/185f2496-a012-46e6-9387-031b3d92ea5a-etc-machine-id\") pod \"185f2496-a012-46e6-9387-031b3d92ea5a\" (UID: \"185f2496-a012-46e6-9387-031b3d92ea5a\") " Oct 05 07:08:00 crc kubenswrapper[4846]: I1005 07:08:00.002265 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bmqpd\" (UniqueName: \"kubernetes.io/projected/185f2496-a012-46e6-9387-031b3d92ea5a-kube-api-access-bmqpd\") pod \"185f2496-a012-46e6-9387-031b3d92ea5a\" (UID: \"185f2496-a012-46e6-9387-031b3d92ea5a\") " Oct 05 07:08:00 crc kubenswrapper[4846]: I1005 07:08:00.002478 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/185f2496-a012-46e6-9387-031b3d92ea5a-scripts\") pod \"185f2496-a012-46e6-9387-031b3d92ea5a\" (UID: \"185f2496-a012-46e6-9387-031b3d92ea5a\") " Oct 05 07:08:00 crc kubenswrapper[4846]: I1005 07:08:00.035357 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/185f2496-a012-46e6-9387-031b3d92ea5a-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "185f2496-a012-46e6-9387-031b3d92ea5a" (UID: "185f2496-a012-46e6-9387-031b3d92ea5a"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 05 07:08:00 crc kubenswrapper[4846]: I1005 07:08:00.035661 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/185f2496-a012-46e6-9387-031b3d92ea5a-logs" (OuterVolumeSpecName: "logs") pod "185f2496-a012-46e6-9387-031b3d92ea5a" (UID: "185f2496-a012-46e6-9387-031b3d92ea5a"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 07:08:00 crc kubenswrapper[4846]: I1005 07:08:00.040040 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/185f2496-a012-46e6-9387-031b3d92ea5a-scripts" (OuterVolumeSpecName: "scripts") pod "185f2496-a012-46e6-9387-031b3d92ea5a" (UID: "185f2496-a012-46e6-9387-031b3d92ea5a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:08:00 crc kubenswrapper[4846]: I1005 07:08:00.041969 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/185f2496-a012-46e6-9387-031b3d92ea5a-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "185f2496-a012-46e6-9387-031b3d92ea5a" (UID: "185f2496-a012-46e6-9387-031b3d92ea5a"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:08:00 crc kubenswrapper[4846]: I1005 07:08:00.067656 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/185f2496-a012-46e6-9387-031b3d92ea5a-kube-api-access-bmqpd" (OuterVolumeSpecName: "kube-api-access-bmqpd") pod "185f2496-a012-46e6-9387-031b3d92ea5a" (UID: "185f2496-a012-46e6-9387-031b3d92ea5a"). InnerVolumeSpecName "kube-api-access-bmqpd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:08:00 crc kubenswrapper[4846]: I1005 07:08:00.112426 4846 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/185f2496-a012-46e6-9387-031b3d92ea5a-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 05 07:08:00 crc kubenswrapper[4846]: I1005 07:08:00.112460 4846 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/185f2496-a012-46e6-9387-031b3d92ea5a-logs\") on node \"crc\" DevicePath \"\"" Oct 05 07:08:00 crc kubenswrapper[4846]: I1005 07:08:00.112486 4846 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/185f2496-a012-46e6-9387-031b3d92ea5a-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 05 07:08:00 crc kubenswrapper[4846]: I1005 07:08:00.112496 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bmqpd\" (UniqueName: \"kubernetes.io/projected/185f2496-a012-46e6-9387-031b3d92ea5a-kube-api-access-bmqpd\") on node \"crc\" DevicePath \"\"" Oct 05 07:08:00 crc kubenswrapper[4846]: I1005 07:08:00.112507 4846 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/185f2496-a012-46e6-9387-031b3d92ea5a-scripts\") on node \"crc\" DevicePath \"\"" Oct 05 07:08:00 crc kubenswrapper[4846]: I1005 07:08:00.132911 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/185f2496-a012-46e6-9387-031b3d92ea5a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "185f2496-a012-46e6-9387-031b3d92ea5a" (UID: "185f2496-a012-46e6-9387-031b3d92ea5a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:08:00 crc kubenswrapper[4846]: I1005 07:08:00.153309 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/185f2496-a012-46e6-9387-031b3d92ea5a-config-data" (OuterVolumeSpecName: "config-data") pod "185f2496-a012-46e6-9387-031b3d92ea5a" (UID: "185f2496-a012-46e6-9387-031b3d92ea5a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:08:00 crc kubenswrapper[4846]: I1005 07:08:00.215423 4846 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/185f2496-a012-46e6-9387-031b3d92ea5a-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 07:08:00 crc kubenswrapper[4846]: I1005 07:08:00.215458 4846 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/185f2496-a012-46e6-9387-031b3d92ea5a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 07:08:00 crc kubenswrapper[4846]: I1005 07:08:00.616712 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-85bbbcfb7f-kwfjt" Oct 05 07:08:00 crc kubenswrapper[4846]: I1005 07:08:00.724841 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-64c4f98476-pqrx9"] Oct 05 07:08:00 crc kubenswrapper[4846]: I1005 07:08:00.725133 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-64c4f98476-pqrx9" podUID="818246fd-88cc-4c04-8ea4-0e2521a0fc45" containerName="neutron-api" containerID="cri-o://d98e1b46601ac7b1c3f03252e286d71bbe2836d44edf5d2b5a62a2b0af440f28" gracePeriod=30 Oct 05 07:08:00 crc kubenswrapper[4846]: I1005 07:08:00.725651 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-64c4f98476-pqrx9" podUID="818246fd-88cc-4c04-8ea4-0e2521a0fc45" containerName="neutron-httpd" containerID="cri-o://1b5541d40b35256864b3d80cd509b1fb37d2f2e7b5097dfa124f6488c42f7d2f" gracePeriod=30 Oct 05 07:08:00 crc kubenswrapper[4846]: I1005 07:08:00.934889 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-57674d6c54-zn5nq" event={"ID":"c31937eb-9836-461b-9230-5e06ce7a99e8","Type":"ContainerStarted","Data":"85a65281831348ab23593ec8a61436c9871720d821d057e65d1423a5e8c5e79c"} Oct 05 07:08:00 crc kubenswrapper[4846]: I1005 07:08:00.934958 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-57674d6c54-zn5nq" event={"ID":"c31937eb-9836-461b-9230-5e06ce7a99e8","Type":"ContainerStarted","Data":"8f6e1c6b60e3d1a211e92f9485d583bd98fbc2e6833ffd3020cdfa2f01f8a1da"} Oct 05 07:08:01 crc kubenswrapper[4846]: I1005 07:08:01.003281 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"185f2496-a012-46e6-9387-031b3d92ea5a","Type":"ContainerDied","Data":"a7c790abe290362ad8139a98bcdf2a6d418ae615a9b48e5227ffda0cc2ba6295"} Oct 05 07:08:01 crc kubenswrapper[4846]: I1005 07:08:01.003353 4846 scope.go:117] "RemoveContainer" containerID="b7c2d5a4f9c46f6b07cf21fc7f29aa99808d5ba4c0a17af8fe48d435039bf13b" Oct 05 07:08:01 crc kubenswrapper[4846]: I1005 07:08:01.003374 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 05 07:08:01 crc kubenswrapper[4846]: I1005 07:08:01.039022 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Oct 05 07:08:01 crc kubenswrapper[4846]: I1005 07:08:01.039576 4846 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 05 07:08:01 crc kubenswrapper[4846]: I1005 07:08:01.108911 4846 scope.go:117] "RemoveContainer" containerID="368a579434676bb95c41afb3688944e9fa1d6b4a41cfb09c70c12f7641134de5" Oct 05 07:08:01 crc kubenswrapper[4846]: I1005 07:08:01.146653 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Oct 05 07:08:01 crc kubenswrapper[4846]: I1005 07:08:01.157309 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Oct 05 07:08:01 crc kubenswrapper[4846]: I1005 07:08:01.167880 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Oct 05 07:08:01 crc kubenswrapper[4846]: E1005 07:08:01.168448 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="185f2496-a012-46e6-9387-031b3d92ea5a" containerName="cinder-api" Oct 05 07:08:01 crc kubenswrapper[4846]: I1005 07:08:01.168469 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="185f2496-a012-46e6-9387-031b3d92ea5a" containerName="cinder-api" Oct 05 07:08:01 crc kubenswrapper[4846]: E1005 07:08:01.168488 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="185f2496-a012-46e6-9387-031b3d92ea5a" containerName="cinder-api-log" Oct 05 07:08:01 crc kubenswrapper[4846]: I1005 07:08:01.168495 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="185f2496-a012-46e6-9387-031b3d92ea5a" containerName="cinder-api-log" Oct 05 07:08:01 crc kubenswrapper[4846]: I1005 07:08:01.168751 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="185f2496-a012-46e6-9387-031b3d92ea5a" containerName="cinder-api-log" Oct 05 07:08:01 crc kubenswrapper[4846]: I1005 07:08:01.168770 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="185f2496-a012-46e6-9387-031b3d92ea5a" containerName="cinder-api" Oct 05 07:08:01 crc kubenswrapper[4846]: I1005 07:08:01.170131 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 05 07:08:01 crc kubenswrapper[4846]: I1005 07:08:01.173747 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-internal-svc" Oct 05 07:08:01 crc kubenswrapper[4846]: I1005 07:08:01.174493 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-public-svc" Oct 05 07:08:01 crc kubenswrapper[4846]: I1005 07:08:01.174886 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Oct 05 07:08:01 crc kubenswrapper[4846]: I1005 07:08:01.197443 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 05 07:08:01 crc kubenswrapper[4846]: I1005 07:08:01.206792 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Oct 05 07:08:01 crc kubenswrapper[4846]: I1005 07:08:01.270483 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/36dca705-2f72-411e-8d80-ed31d49989cc-config-data-custom\") pod \"cinder-api-0\" (UID: \"36dca705-2f72-411e-8d80-ed31d49989cc\") " pod="openstack/cinder-api-0" Oct 05 07:08:01 crc kubenswrapper[4846]: I1005 07:08:01.270531 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xld6b\" (UniqueName: \"kubernetes.io/projected/36dca705-2f72-411e-8d80-ed31d49989cc-kube-api-access-xld6b\") pod \"cinder-api-0\" (UID: \"36dca705-2f72-411e-8d80-ed31d49989cc\") " pod="openstack/cinder-api-0" Oct 05 07:08:01 crc kubenswrapper[4846]: I1005 07:08:01.270568 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/36dca705-2f72-411e-8d80-ed31d49989cc-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"36dca705-2f72-411e-8d80-ed31d49989cc\") " pod="openstack/cinder-api-0" Oct 05 07:08:01 crc kubenswrapper[4846]: I1005 07:08:01.270625 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/36dca705-2f72-411e-8d80-ed31d49989cc-public-tls-certs\") pod \"cinder-api-0\" (UID: \"36dca705-2f72-411e-8d80-ed31d49989cc\") " pod="openstack/cinder-api-0" Oct 05 07:08:01 crc kubenswrapper[4846]: I1005 07:08:01.270657 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/36dca705-2f72-411e-8d80-ed31d49989cc-logs\") pod \"cinder-api-0\" (UID: \"36dca705-2f72-411e-8d80-ed31d49989cc\") " pod="openstack/cinder-api-0" Oct 05 07:08:01 crc kubenswrapper[4846]: I1005 07:08:01.270902 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/36dca705-2f72-411e-8d80-ed31d49989cc-scripts\") pod \"cinder-api-0\" (UID: \"36dca705-2f72-411e-8d80-ed31d49989cc\") " pod="openstack/cinder-api-0" Oct 05 07:08:01 crc kubenswrapper[4846]: I1005 07:08:01.271163 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/36dca705-2f72-411e-8d80-ed31d49989cc-etc-machine-id\") pod \"cinder-api-0\" (UID: \"36dca705-2f72-411e-8d80-ed31d49989cc\") " pod="openstack/cinder-api-0" Oct 05 07:08:01 crc kubenswrapper[4846]: I1005 07:08:01.271244 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/36dca705-2f72-411e-8d80-ed31d49989cc-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"36dca705-2f72-411e-8d80-ed31d49989cc\") " pod="openstack/cinder-api-0" Oct 05 07:08:01 crc kubenswrapper[4846]: I1005 07:08:01.271342 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/36dca705-2f72-411e-8d80-ed31d49989cc-config-data\") pod \"cinder-api-0\" (UID: \"36dca705-2f72-411e-8d80-ed31d49989cc\") " pod="openstack/cinder-api-0" Oct 05 07:08:01 crc kubenswrapper[4846]: I1005 07:08:01.376486 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/36dca705-2f72-411e-8d80-ed31d49989cc-public-tls-certs\") pod \"cinder-api-0\" (UID: \"36dca705-2f72-411e-8d80-ed31d49989cc\") " pod="openstack/cinder-api-0" Oct 05 07:08:01 crc kubenswrapper[4846]: I1005 07:08:01.376556 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/36dca705-2f72-411e-8d80-ed31d49989cc-logs\") pod \"cinder-api-0\" (UID: \"36dca705-2f72-411e-8d80-ed31d49989cc\") " pod="openstack/cinder-api-0" Oct 05 07:08:01 crc kubenswrapper[4846]: I1005 07:08:01.376599 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/36dca705-2f72-411e-8d80-ed31d49989cc-scripts\") pod \"cinder-api-0\" (UID: \"36dca705-2f72-411e-8d80-ed31d49989cc\") " pod="openstack/cinder-api-0" Oct 05 07:08:01 crc kubenswrapper[4846]: I1005 07:08:01.376644 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/36dca705-2f72-411e-8d80-ed31d49989cc-etc-machine-id\") pod \"cinder-api-0\" (UID: \"36dca705-2f72-411e-8d80-ed31d49989cc\") " pod="openstack/cinder-api-0" Oct 05 07:08:01 crc kubenswrapper[4846]: I1005 07:08:01.376667 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/36dca705-2f72-411e-8d80-ed31d49989cc-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"36dca705-2f72-411e-8d80-ed31d49989cc\") " pod="openstack/cinder-api-0" Oct 05 07:08:01 crc kubenswrapper[4846]: I1005 07:08:01.376701 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/36dca705-2f72-411e-8d80-ed31d49989cc-config-data\") pod \"cinder-api-0\" (UID: \"36dca705-2f72-411e-8d80-ed31d49989cc\") " pod="openstack/cinder-api-0" Oct 05 07:08:01 crc kubenswrapper[4846]: I1005 07:08:01.376749 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/36dca705-2f72-411e-8d80-ed31d49989cc-config-data-custom\") pod \"cinder-api-0\" (UID: \"36dca705-2f72-411e-8d80-ed31d49989cc\") " pod="openstack/cinder-api-0" Oct 05 07:08:01 crc kubenswrapper[4846]: I1005 07:08:01.376769 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xld6b\" (UniqueName: \"kubernetes.io/projected/36dca705-2f72-411e-8d80-ed31d49989cc-kube-api-access-xld6b\") pod \"cinder-api-0\" (UID: \"36dca705-2f72-411e-8d80-ed31d49989cc\") " pod="openstack/cinder-api-0" Oct 05 07:08:01 crc kubenswrapper[4846]: I1005 07:08:01.376808 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/36dca705-2f72-411e-8d80-ed31d49989cc-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"36dca705-2f72-411e-8d80-ed31d49989cc\") " pod="openstack/cinder-api-0" Oct 05 07:08:01 crc kubenswrapper[4846]: I1005 07:08:01.377780 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/36dca705-2f72-411e-8d80-ed31d49989cc-logs\") pod \"cinder-api-0\" (UID: \"36dca705-2f72-411e-8d80-ed31d49989cc\") " pod="openstack/cinder-api-0" Oct 05 07:08:01 crc kubenswrapper[4846]: I1005 07:08:01.383088 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/36dca705-2f72-411e-8d80-ed31d49989cc-etc-machine-id\") pod \"cinder-api-0\" (UID: \"36dca705-2f72-411e-8d80-ed31d49989cc\") " pod="openstack/cinder-api-0" Oct 05 07:08:01 crc kubenswrapper[4846]: I1005 07:08:01.397704 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/36dca705-2f72-411e-8d80-ed31d49989cc-config-data\") pod \"cinder-api-0\" (UID: \"36dca705-2f72-411e-8d80-ed31d49989cc\") " pod="openstack/cinder-api-0" Oct 05 07:08:01 crc kubenswrapper[4846]: I1005 07:08:01.397923 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/36dca705-2f72-411e-8d80-ed31d49989cc-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"36dca705-2f72-411e-8d80-ed31d49989cc\") " pod="openstack/cinder-api-0" Oct 05 07:08:01 crc kubenswrapper[4846]: I1005 07:08:01.398104 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/36dca705-2f72-411e-8d80-ed31d49989cc-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"36dca705-2f72-411e-8d80-ed31d49989cc\") " pod="openstack/cinder-api-0" Oct 05 07:08:01 crc kubenswrapper[4846]: I1005 07:08:01.399818 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/36dca705-2f72-411e-8d80-ed31d49989cc-scripts\") pod \"cinder-api-0\" (UID: \"36dca705-2f72-411e-8d80-ed31d49989cc\") " pod="openstack/cinder-api-0" Oct 05 07:08:01 crc kubenswrapper[4846]: I1005 07:08:01.402405 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/36dca705-2f72-411e-8d80-ed31d49989cc-public-tls-certs\") pod \"cinder-api-0\" (UID: \"36dca705-2f72-411e-8d80-ed31d49989cc\") " pod="openstack/cinder-api-0" Oct 05 07:08:01 crc kubenswrapper[4846]: I1005 07:08:01.405025 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/36dca705-2f72-411e-8d80-ed31d49989cc-config-data-custom\") pod \"cinder-api-0\" (UID: \"36dca705-2f72-411e-8d80-ed31d49989cc\") " pod="openstack/cinder-api-0" Oct 05 07:08:01 crc kubenswrapper[4846]: I1005 07:08:01.411925 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xld6b\" (UniqueName: \"kubernetes.io/projected/36dca705-2f72-411e-8d80-ed31d49989cc-kube-api-access-xld6b\") pod \"cinder-api-0\" (UID: \"36dca705-2f72-411e-8d80-ed31d49989cc\") " pod="openstack/cinder-api-0" Oct 05 07:08:01 crc kubenswrapper[4846]: I1005 07:08:01.495861 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 05 07:08:02 crc kubenswrapper[4846]: I1005 07:08:02.018716 4846 generic.go:334] "Generic (PLEG): container finished" podID="818246fd-88cc-4c04-8ea4-0e2521a0fc45" containerID="1b5541d40b35256864b3d80cd509b1fb37d2f2e7b5097dfa124f6488c42f7d2f" exitCode=0 Oct 05 07:08:02 crc kubenswrapper[4846]: I1005 07:08:02.018852 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-64c4f98476-pqrx9" event={"ID":"818246fd-88cc-4c04-8ea4-0e2521a0fc45","Type":"ContainerDied","Data":"1b5541d40b35256864b3d80cd509b1fb37d2f2e7b5097dfa124f6488c42f7d2f"} Oct 05 07:08:02 crc kubenswrapper[4846]: I1005 07:08:02.021915 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-57674d6c54-zn5nq" event={"ID":"c31937eb-9836-461b-9230-5e06ce7a99e8","Type":"ContainerStarted","Data":"62af63220e4d608e4b981ffe4309575edbb012337dc1e84d368d62d0b4cab633"} Oct 05 07:08:02 crc kubenswrapper[4846]: I1005 07:08:02.021994 4846 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 05 07:08:02 crc kubenswrapper[4846]: I1005 07:08:02.022399 4846 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 05 07:08:02 crc kubenswrapper[4846]: I1005 07:08:02.046157 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 05 07:08:02 crc kubenswrapper[4846]: I1005 07:08:02.051574 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-57674d6c54-zn5nq" podStartSLOduration=4.051541425 podStartE2EDuration="4.051541425s" podCreationTimestamp="2025-10-05 07:07:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 07:08:02.045401761 +0000 UTC m=+1204.286254546" watchObservedRunningTime="2025-10-05 07:08:02.051541425 +0000 UTC m=+1204.292394200" Oct 05 07:08:02 crc kubenswrapper[4846]: W1005 07:08:02.059328 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod36dca705_2f72_411e_8d80_ed31d49989cc.slice/crio-afaf1e34eaf63fcbc924b759258cd9887dcac6aeb43b21132100b45d8bbe4ea4 WatchSource:0}: Error finding container afaf1e34eaf63fcbc924b759258cd9887dcac6aeb43b21132100b45d8bbe4ea4: Status 404 returned error can't find the container with id afaf1e34eaf63fcbc924b759258cd9887dcac6aeb43b21132100b45d8bbe4ea4 Oct 05 07:08:02 crc kubenswrapper[4846]: I1005 07:08:02.515633 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="185f2496-a012-46e6-9387-031b3d92ea5a" path="/var/lib/kubelet/pods/185f2496-a012-46e6-9387-031b3d92ea5a/volumes" Oct 05 07:08:02 crc kubenswrapper[4846]: I1005 07:08:02.763321 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Oct 05 07:08:03 crc kubenswrapper[4846]: I1005 07:08:03.041118 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"36dca705-2f72-411e-8d80-ed31d49989cc","Type":"ContainerStarted","Data":"735ae29bd9ab70b9f2083caa08160bc161134c908e991ed9c2afaf1f25347aa0"} Oct 05 07:08:03 crc kubenswrapper[4846]: I1005 07:08:03.041547 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-57674d6c54-zn5nq" Oct 05 07:08:03 crc kubenswrapper[4846]: I1005 07:08:03.041565 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-57674d6c54-zn5nq" Oct 05 07:08:03 crc kubenswrapper[4846]: I1005 07:08:03.041577 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"36dca705-2f72-411e-8d80-ed31d49989cc","Type":"ContainerStarted","Data":"afaf1e34eaf63fcbc924b759258cd9887dcac6aeb43b21132100b45d8bbe4ea4"} Oct 05 07:08:03 crc kubenswrapper[4846]: I1005 07:08:03.041273 4846 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 05 07:08:03 crc kubenswrapper[4846]: I1005 07:08:03.433429 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Oct 05 07:08:04 crc kubenswrapper[4846]: I1005 07:08:04.050972 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"36dca705-2f72-411e-8d80-ed31d49989cc","Type":"ContainerStarted","Data":"2dae962829e3caf9333b75e7f59bede4cf4d679ed6def4e5eddee1a34c169b39"} Oct 05 07:08:04 crc kubenswrapper[4846]: I1005 07:08:04.051453 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Oct 05 07:08:04 crc kubenswrapper[4846]: I1005 07:08:04.073211 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=3.07316582 podStartE2EDuration="3.07316582s" podCreationTimestamp="2025-10-05 07:08:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 07:08:04.070756696 +0000 UTC m=+1206.311609491" watchObservedRunningTime="2025-10-05 07:08:04.07316582 +0000 UTC m=+1206.314018595" Oct 05 07:08:04 crc kubenswrapper[4846]: I1005 07:08:04.324887 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-6c9ddc8584-bqgtc" Oct 05 07:08:04 crc kubenswrapper[4846]: I1005 07:08:04.398429 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-99495cbc7-v87cl" Oct 05 07:08:04 crc kubenswrapper[4846]: I1005 07:08:04.481247 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7fcd8fd75c-qmzjd"] Oct 05 07:08:04 crc kubenswrapper[4846]: I1005 07:08:04.481598 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7fcd8fd75c-qmzjd" podUID="f6b4f351-efb1-4a53-8431-e7058eace718" containerName="dnsmasq-dns" containerID="cri-o://adb9e57c3af0eccf4eb1a4d4fa7671627385b5ebffd25943a05f6247adfca3ce" gracePeriod=10 Oct 05 07:08:04 crc kubenswrapper[4846]: I1005 07:08:04.579349 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-6c9ddc8584-bqgtc" Oct 05 07:08:04 crc kubenswrapper[4846]: I1005 07:08:04.697082 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Oct 05 07:08:04 crc kubenswrapper[4846]: I1005 07:08:04.743006 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 05 07:08:05 crc kubenswrapper[4846]: I1005 07:08:05.068430 4846 generic.go:334] "Generic (PLEG): container finished" podID="f6b4f351-efb1-4a53-8431-e7058eace718" containerID="adb9e57c3af0eccf4eb1a4d4fa7671627385b5ebffd25943a05f6247adfca3ce" exitCode=0 Oct 05 07:08:05 crc kubenswrapper[4846]: I1005 07:08:05.068500 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7fcd8fd75c-qmzjd" event={"ID":"f6b4f351-efb1-4a53-8431-e7058eace718","Type":"ContainerDied","Data":"adb9e57c3af0eccf4eb1a4d4fa7671627385b5ebffd25943a05f6247adfca3ce"} Oct 05 07:08:05 crc kubenswrapper[4846]: I1005 07:08:05.069088 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7fcd8fd75c-qmzjd" event={"ID":"f6b4f351-efb1-4a53-8431-e7058eace718","Type":"ContainerDied","Data":"f7228bbaf3f67d3a50ee1739dd126b0d08d75956f0364773df6454eae6d59fb3"} Oct 05 07:08:05 crc kubenswrapper[4846]: I1005 07:08:05.069111 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f7228bbaf3f67d3a50ee1739dd126b0d08d75956f0364773df6454eae6d59fb3" Oct 05 07:08:05 crc kubenswrapper[4846]: I1005 07:08:05.069520 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="40289760-4ac7-42e8-8eab-39f93c1618aa" containerName="cinder-scheduler" containerID="cri-o://fd9c22e44bb6d2c4d688311f3f8f67ffbf4f78d9f27e0d6baa248cd31e5ddfae" gracePeriod=30 Oct 05 07:08:05 crc kubenswrapper[4846]: I1005 07:08:05.069570 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="40289760-4ac7-42e8-8eab-39f93c1618aa" containerName="probe" containerID="cri-o://03eaa092d41b9b080f0832b3cfdb8fb94047441e2dd9f8972556e0afa700034e" gracePeriod=30 Oct 05 07:08:05 crc kubenswrapper[4846]: I1005 07:08:05.168831 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7fcd8fd75c-qmzjd" Oct 05 07:08:05 crc kubenswrapper[4846]: I1005 07:08:05.304418 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f6b4f351-efb1-4a53-8431-e7058eace718-dns-svc\") pod \"f6b4f351-efb1-4a53-8431-e7058eace718\" (UID: \"f6b4f351-efb1-4a53-8431-e7058eace718\") " Oct 05 07:08:05 crc kubenswrapper[4846]: I1005 07:08:05.304490 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f6b4f351-efb1-4a53-8431-e7058eace718-ovsdbserver-nb\") pod \"f6b4f351-efb1-4a53-8431-e7058eace718\" (UID: \"f6b4f351-efb1-4a53-8431-e7058eace718\") " Oct 05 07:08:05 crc kubenswrapper[4846]: I1005 07:08:05.307013 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tsf6z\" (UniqueName: \"kubernetes.io/projected/f6b4f351-efb1-4a53-8431-e7058eace718-kube-api-access-tsf6z\") pod \"f6b4f351-efb1-4a53-8431-e7058eace718\" (UID: \"f6b4f351-efb1-4a53-8431-e7058eace718\") " Oct 05 07:08:05 crc kubenswrapper[4846]: I1005 07:08:05.307237 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f6b4f351-efb1-4a53-8431-e7058eace718-config\") pod \"f6b4f351-efb1-4a53-8431-e7058eace718\" (UID: \"f6b4f351-efb1-4a53-8431-e7058eace718\") " Oct 05 07:08:05 crc kubenswrapper[4846]: I1005 07:08:05.307321 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f6b4f351-efb1-4a53-8431-e7058eace718-ovsdbserver-sb\") pod \"f6b4f351-efb1-4a53-8431-e7058eace718\" (UID: \"f6b4f351-efb1-4a53-8431-e7058eace718\") " Oct 05 07:08:05 crc kubenswrapper[4846]: I1005 07:08:05.307392 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f6b4f351-efb1-4a53-8431-e7058eace718-dns-swift-storage-0\") pod \"f6b4f351-efb1-4a53-8431-e7058eace718\" (UID: \"f6b4f351-efb1-4a53-8431-e7058eace718\") " Oct 05 07:08:05 crc kubenswrapper[4846]: I1005 07:08:05.316206 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f6b4f351-efb1-4a53-8431-e7058eace718-kube-api-access-tsf6z" (OuterVolumeSpecName: "kube-api-access-tsf6z") pod "f6b4f351-efb1-4a53-8431-e7058eace718" (UID: "f6b4f351-efb1-4a53-8431-e7058eace718"). InnerVolumeSpecName "kube-api-access-tsf6z". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:08:05 crc kubenswrapper[4846]: I1005 07:08:05.378937 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f6b4f351-efb1-4a53-8431-e7058eace718-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "f6b4f351-efb1-4a53-8431-e7058eace718" (UID: "f6b4f351-efb1-4a53-8431-e7058eace718"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 07:08:05 crc kubenswrapper[4846]: I1005 07:08:05.384897 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f6b4f351-efb1-4a53-8431-e7058eace718-config" (OuterVolumeSpecName: "config") pod "f6b4f351-efb1-4a53-8431-e7058eace718" (UID: "f6b4f351-efb1-4a53-8431-e7058eace718"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 07:08:05 crc kubenswrapper[4846]: I1005 07:08:05.397956 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f6b4f351-efb1-4a53-8431-e7058eace718-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "f6b4f351-efb1-4a53-8431-e7058eace718" (UID: "f6b4f351-efb1-4a53-8431-e7058eace718"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 07:08:05 crc kubenswrapper[4846]: I1005 07:08:05.401718 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f6b4f351-efb1-4a53-8431-e7058eace718-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "f6b4f351-efb1-4a53-8431-e7058eace718" (UID: "f6b4f351-efb1-4a53-8431-e7058eace718"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 07:08:05 crc kubenswrapper[4846]: I1005 07:08:05.411884 4846 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f6b4f351-efb1-4a53-8431-e7058eace718-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 05 07:08:05 crc kubenswrapper[4846]: I1005 07:08:05.411946 4846 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f6b4f351-efb1-4a53-8431-e7058eace718-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 05 07:08:05 crc kubenswrapper[4846]: I1005 07:08:05.411962 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tsf6z\" (UniqueName: \"kubernetes.io/projected/f6b4f351-efb1-4a53-8431-e7058eace718-kube-api-access-tsf6z\") on node \"crc\" DevicePath \"\"" Oct 05 07:08:05 crc kubenswrapper[4846]: I1005 07:08:05.411980 4846 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f6b4f351-efb1-4a53-8431-e7058eace718-config\") on node \"crc\" DevicePath \"\"" Oct 05 07:08:05 crc kubenswrapper[4846]: I1005 07:08:05.411995 4846 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f6b4f351-efb1-4a53-8431-e7058eace718-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 05 07:08:05 crc kubenswrapper[4846]: I1005 07:08:05.452912 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f6b4f351-efb1-4a53-8431-e7058eace718-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "f6b4f351-efb1-4a53-8431-e7058eace718" (UID: "f6b4f351-efb1-4a53-8431-e7058eace718"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 07:08:05 crc kubenswrapper[4846]: I1005 07:08:05.513647 4846 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f6b4f351-efb1-4a53-8431-e7058eace718-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 05 07:08:05 crc kubenswrapper[4846]: I1005 07:08:05.730341 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/keystone-79f67599fc-2sk76" Oct 05 07:08:06 crc kubenswrapper[4846]: I1005 07:08:06.086100 4846 generic.go:334] "Generic (PLEG): container finished" podID="40289760-4ac7-42e8-8eab-39f93c1618aa" containerID="03eaa092d41b9b080f0832b3cfdb8fb94047441e2dd9f8972556e0afa700034e" exitCode=0 Oct 05 07:08:06 crc kubenswrapper[4846]: I1005 07:08:06.086819 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7fcd8fd75c-qmzjd" Oct 05 07:08:06 crc kubenswrapper[4846]: I1005 07:08:06.086254 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"40289760-4ac7-42e8-8eab-39f93c1618aa","Type":"ContainerDied","Data":"03eaa092d41b9b080f0832b3cfdb8fb94047441e2dd9f8972556e0afa700034e"} Oct 05 07:08:06 crc kubenswrapper[4846]: I1005 07:08:06.123718 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7fcd8fd75c-qmzjd"] Oct 05 07:08:06 crc kubenswrapper[4846]: I1005 07:08:06.130386 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7fcd8fd75c-qmzjd"] Oct 05 07:08:06 crc kubenswrapper[4846]: I1005 07:08:06.515217 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f6b4f351-efb1-4a53-8431-e7058eace718" path="/var/lib/kubelet/pods/f6b4f351-efb1-4a53-8431-e7058eace718/volumes" Oct 05 07:08:06 crc kubenswrapper[4846]: I1005 07:08:06.654893 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 05 07:08:06 crc kubenswrapper[4846]: I1005 07:08:06.750944 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/40289760-4ac7-42e8-8eab-39f93c1618aa-etc-machine-id\") pod \"40289760-4ac7-42e8-8eab-39f93c1618aa\" (UID: \"40289760-4ac7-42e8-8eab-39f93c1618aa\") " Oct 05 07:08:06 crc kubenswrapper[4846]: I1005 07:08:06.751418 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g45vh\" (UniqueName: \"kubernetes.io/projected/40289760-4ac7-42e8-8eab-39f93c1618aa-kube-api-access-g45vh\") pod \"40289760-4ac7-42e8-8eab-39f93c1618aa\" (UID: \"40289760-4ac7-42e8-8eab-39f93c1618aa\") " Oct 05 07:08:06 crc kubenswrapper[4846]: I1005 07:08:06.751472 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/40289760-4ac7-42e8-8eab-39f93c1618aa-combined-ca-bundle\") pod \"40289760-4ac7-42e8-8eab-39f93c1618aa\" (UID: \"40289760-4ac7-42e8-8eab-39f93c1618aa\") " Oct 05 07:08:06 crc kubenswrapper[4846]: I1005 07:08:06.751526 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/40289760-4ac7-42e8-8eab-39f93c1618aa-config-data-custom\") pod \"40289760-4ac7-42e8-8eab-39f93c1618aa\" (UID: \"40289760-4ac7-42e8-8eab-39f93c1618aa\") " Oct 05 07:08:06 crc kubenswrapper[4846]: I1005 07:08:06.751608 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/40289760-4ac7-42e8-8eab-39f93c1618aa-scripts\") pod \"40289760-4ac7-42e8-8eab-39f93c1618aa\" (UID: \"40289760-4ac7-42e8-8eab-39f93c1618aa\") " Oct 05 07:08:06 crc kubenswrapper[4846]: I1005 07:08:06.751721 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/40289760-4ac7-42e8-8eab-39f93c1618aa-config-data\") pod \"40289760-4ac7-42e8-8eab-39f93c1618aa\" (UID: \"40289760-4ac7-42e8-8eab-39f93c1618aa\") " Oct 05 07:08:06 crc kubenswrapper[4846]: I1005 07:08:06.751191 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/40289760-4ac7-42e8-8eab-39f93c1618aa-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "40289760-4ac7-42e8-8eab-39f93c1618aa" (UID: "40289760-4ac7-42e8-8eab-39f93c1618aa"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 05 07:08:06 crc kubenswrapper[4846]: I1005 07:08:06.758478 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/40289760-4ac7-42e8-8eab-39f93c1618aa-scripts" (OuterVolumeSpecName: "scripts") pod "40289760-4ac7-42e8-8eab-39f93c1618aa" (UID: "40289760-4ac7-42e8-8eab-39f93c1618aa"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:08:06 crc kubenswrapper[4846]: I1005 07:08:06.770673 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/40289760-4ac7-42e8-8eab-39f93c1618aa-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "40289760-4ac7-42e8-8eab-39f93c1618aa" (UID: "40289760-4ac7-42e8-8eab-39f93c1618aa"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:08:06 crc kubenswrapper[4846]: I1005 07:08:06.779598 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/40289760-4ac7-42e8-8eab-39f93c1618aa-kube-api-access-g45vh" (OuterVolumeSpecName: "kube-api-access-g45vh") pod "40289760-4ac7-42e8-8eab-39f93c1618aa" (UID: "40289760-4ac7-42e8-8eab-39f93c1618aa"). InnerVolumeSpecName "kube-api-access-g45vh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:08:06 crc kubenswrapper[4846]: I1005 07:08:06.820235 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/40289760-4ac7-42e8-8eab-39f93c1618aa-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "40289760-4ac7-42e8-8eab-39f93c1618aa" (UID: "40289760-4ac7-42e8-8eab-39f93c1618aa"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:08:06 crc kubenswrapper[4846]: I1005 07:08:06.853642 4846 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/40289760-4ac7-42e8-8eab-39f93c1618aa-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 07:08:06 crc kubenswrapper[4846]: I1005 07:08:06.853676 4846 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/40289760-4ac7-42e8-8eab-39f93c1618aa-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 05 07:08:06 crc kubenswrapper[4846]: I1005 07:08:06.853686 4846 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/40289760-4ac7-42e8-8eab-39f93c1618aa-scripts\") on node \"crc\" DevicePath \"\"" Oct 05 07:08:06 crc kubenswrapper[4846]: I1005 07:08:06.853695 4846 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/40289760-4ac7-42e8-8eab-39f93c1618aa-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 05 07:08:06 crc kubenswrapper[4846]: I1005 07:08:06.853704 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g45vh\" (UniqueName: \"kubernetes.io/projected/40289760-4ac7-42e8-8eab-39f93c1618aa-kube-api-access-g45vh\") on node \"crc\" DevicePath \"\"" Oct 05 07:08:06 crc kubenswrapper[4846]: I1005 07:08:06.863519 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/40289760-4ac7-42e8-8eab-39f93c1618aa-config-data" (OuterVolumeSpecName: "config-data") pod "40289760-4ac7-42e8-8eab-39f93c1618aa" (UID: "40289760-4ac7-42e8-8eab-39f93c1618aa"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:08:06 crc kubenswrapper[4846]: I1005 07:08:06.956388 4846 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/40289760-4ac7-42e8-8eab-39f93c1618aa-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 07:08:07 crc kubenswrapper[4846]: I1005 07:08:07.100483 4846 generic.go:334] "Generic (PLEG): container finished" podID="40289760-4ac7-42e8-8eab-39f93c1618aa" containerID="fd9c22e44bb6d2c4d688311f3f8f67ffbf4f78d9f27e0d6baa248cd31e5ddfae" exitCode=0 Oct 05 07:08:07 crc kubenswrapper[4846]: I1005 07:08:07.100571 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 05 07:08:07 crc kubenswrapper[4846]: I1005 07:08:07.100592 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"40289760-4ac7-42e8-8eab-39f93c1618aa","Type":"ContainerDied","Data":"fd9c22e44bb6d2c4d688311f3f8f67ffbf4f78d9f27e0d6baa248cd31e5ddfae"} Oct 05 07:08:07 crc kubenswrapper[4846]: I1005 07:08:07.100667 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"40289760-4ac7-42e8-8eab-39f93c1618aa","Type":"ContainerDied","Data":"9a9217ce018fb0e4213096c3c7fec98ee147392235029f6fd87273173ab45ce6"} Oct 05 07:08:07 crc kubenswrapper[4846]: I1005 07:08:07.100689 4846 scope.go:117] "RemoveContainer" containerID="03eaa092d41b9b080f0832b3cfdb8fb94047441e2dd9f8972556e0afa700034e" Oct 05 07:08:07 crc kubenswrapper[4846]: I1005 07:08:07.161036 4846 scope.go:117] "RemoveContainer" containerID="fd9c22e44bb6d2c4d688311f3f8f67ffbf4f78d9f27e0d6baa248cd31e5ddfae" Oct 05 07:08:07 crc kubenswrapper[4846]: I1005 07:08:07.168778 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 05 07:08:07 crc kubenswrapper[4846]: I1005 07:08:07.184056 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 05 07:08:07 crc kubenswrapper[4846]: I1005 07:08:07.196460 4846 scope.go:117] "RemoveContainer" containerID="03eaa092d41b9b080f0832b3cfdb8fb94047441e2dd9f8972556e0afa700034e" Oct 05 07:08:07 crc kubenswrapper[4846]: E1005 07:08:07.196947 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"03eaa092d41b9b080f0832b3cfdb8fb94047441e2dd9f8972556e0afa700034e\": container with ID starting with 03eaa092d41b9b080f0832b3cfdb8fb94047441e2dd9f8972556e0afa700034e not found: ID does not exist" containerID="03eaa092d41b9b080f0832b3cfdb8fb94047441e2dd9f8972556e0afa700034e" Oct 05 07:08:07 crc kubenswrapper[4846]: I1005 07:08:07.196983 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"03eaa092d41b9b080f0832b3cfdb8fb94047441e2dd9f8972556e0afa700034e"} err="failed to get container status \"03eaa092d41b9b080f0832b3cfdb8fb94047441e2dd9f8972556e0afa700034e\": rpc error: code = NotFound desc = could not find container \"03eaa092d41b9b080f0832b3cfdb8fb94047441e2dd9f8972556e0afa700034e\": container with ID starting with 03eaa092d41b9b080f0832b3cfdb8fb94047441e2dd9f8972556e0afa700034e not found: ID does not exist" Oct 05 07:08:07 crc kubenswrapper[4846]: I1005 07:08:07.197007 4846 scope.go:117] "RemoveContainer" containerID="fd9c22e44bb6d2c4d688311f3f8f67ffbf4f78d9f27e0d6baa248cd31e5ddfae" Oct 05 07:08:07 crc kubenswrapper[4846]: E1005 07:08:07.197382 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fd9c22e44bb6d2c4d688311f3f8f67ffbf4f78d9f27e0d6baa248cd31e5ddfae\": container with ID starting with fd9c22e44bb6d2c4d688311f3f8f67ffbf4f78d9f27e0d6baa248cd31e5ddfae not found: ID does not exist" containerID="fd9c22e44bb6d2c4d688311f3f8f67ffbf4f78d9f27e0d6baa248cd31e5ddfae" Oct 05 07:08:07 crc kubenswrapper[4846]: I1005 07:08:07.197446 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fd9c22e44bb6d2c4d688311f3f8f67ffbf4f78d9f27e0d6baa248cd31e5ddfae"} err="failed to get container status \"fd9c22e44bb6d2c4d688311f3f8f67ffbf4f78d9f27e0d6baa248cd31e5ddfae\": rpc error: code = NotFound desc = could not find container \"fd9c22e44bb6d2c4d688311f3f8f67ffbf4f78d9f27e0d6baa248cd31e5ddfae\": container with ID starting with fd9c22e44bb6d2c4d688311f3f8f67ffbf4f78d9f27e0d6baa248cd31e5ddfae not found: ID does not exist" Oct 05 07:08:07 crc kubenswrapper[4846]: I1005 07:08:07.198095 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Oct 05 07:08:07 crc kubenswrapper[4846]: E1005 07:08:07.198641 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f6b4f351-efb1-4a53-8431-e7058eace718" containerName="dnsmasq-dns" Oct 05 07:08:07 crc kubenswrapper[4846]: I1005 07:08:07.198665 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="f6b4f351-efb1-4a53-8431-e7058eace718" containerName="dnsmasq-dns" Oct 05 07:08:07 crc kubenswrapper[4846]: E1005 07:08:07.198705 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="40289760-4ac7-42e8-8eab-39f93c1618aa" containerName="probe" Oct 05 07:08:07 crc kubenswrapper[4846]: I1005 07:08:07.198714 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="40289760-4ac7-42e8-8eab-39f93c1618aa" containerName="probe" Oct 05 07:08:07 crc kubenswrapper[4846]: E1005 07:08:07.198751 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f6b4f351-efb1-4a53-8431-e7058eace718" containerName="init" Oct 05 07:08:07 crc kubenswrapper[4846]: I1005 07:08:07.198759 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="f6b4f351-efb1-4a53-8431-e7058eace718" containerName="init" Oct 05 07:08:07 crc kubenswrapper[4846]: E1005 07:08:07.198773 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="40289760-4ac7-42e8-8eab-39f93c1618aa" containerName="cinder-scheduler" Oct 05 07:08:07 crc kubenswrapper[4846]: I1005 07:08:07.198782 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="40289760-4ac7-42e8-8eab-39f93c1618aa" containerName="cinder-scheduler" Oct 05 07:08:07 crc kubenswrapper[4846]: I1005 07:08:07.199018 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="40289760-4ac7-42e8-8eab-39f93c1618aa" containerName="cinder-scheduler" Oct 05 07:08:07 crc kubenswrapper[4846]: I1005 07:08:07.199056 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="f6b4f351-efb1-4a53-8431-e7058eace718" containerName="dnsmasq-dns" Oct 05 07:08:07 crc kubenswrapper[4846]: I1005 07:08:07.199077 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="40289760-4ac7-42e8-8eab-39f93c1618aa" containerName="probe" Oct 05 07:08:07 crc kubenswrapper[4846]: I1005 07:08:07.200495 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 05 07:08:07 crc kubenswrapper[4846]: I1005 07:08:07.204716 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Oct 05 07:08:07 crc kubenswrapper[4846]: I1005 07:08:07.209412 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 05 07:08:07 crc kubenswrapper[4846]: I1005 07:08:07.262841 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bfe96e0c-b9fc-438d-9b17-bec246d48446-config-data\") pod \"cinder-scheduler-0\" (UID: \"bfe96e0c-b9fc-438d-9b17-bec246d48446\") " pod="openstack/cinder-scheduler-0" Oct 05 07:08:07 crc kubenswrapper[4846]: I1005 07:08:07.263254 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/bfe96e0c-b9fc-438d-9b17-bec246d48446-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"bfe96e0c-b9fc-438d-9b17-bec246d48446\") " pod="openstack/cinder-scheduler-0" Oct 05 07:08:07 crc kubenswrapper[4846]: I1005 07:08:07.263381 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/bfe96e0c-b9fc-438d-9b17-bec246d48446-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"bfe96e0c-b9fc-438d-9b17-bec246d48446\") " pod="openstack/cinder-scheduler-0" Oct 05 07:08:07 crc kubenswrapper[4846]: I1005 07:08:07.263479 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bfe96e0c-b9fc-438d-9b17-bec246d48446-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"bfe96e0c-b9fc-438d-9b17-bec246d48446\") " pod="openstack/cinder-scheduler-0" Oct 05 07:08:07 crc kubenswrapper[4846]: I1005 07:08:07.263562 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hd6jb\" (UniqueName: \"kubernetes.io/projected/bfe96e0c-b9fc-438d-9b17-bec246d48446-kube-api-access-hd6jb\") pod \"cinder-scheduler-0\" (UID: \"bfe96e0c-b9fc-438d-9b17-bec246d48446\") " pod="openstack/cinder-scheduler-0" Oct 05 07:08:07 crc kubenswrapper[4846]: I1005 07:08:07.263691 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bfe96e0c-b9fc-438d-9b17-bec246d48446-scripts\") pod \"cinder-scheduler-0\" (UID: \"bfe96e0c-b9fc-438d-9b17-bec246d48446\") " pod="openstack/cinder-scheduler-0" Oct 05 07:08:07 crc kubenswrapper[4846]: I1005 07:08:07.365823 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bfe96e0c-b9fc-438d-9b17-bec246d48446-scripts\") pod \"cinder-scheduler-0\" (UID: \"bfe96e0c-b9fc-438d-9b17-bec246d48446\") " pod="openstack/cinder-scheduler-0" Oct 05 07:08:07 crc kubenswrapper[4846]: I1005 07:08:07.365922 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bfe96e0c-b9fc-438d-9b17-bec246d48446-config-data\") pod \"cinder-scheduler-0\" (UID: \"bfe96e0c-b9fc-438d-9b17-bec246d48446\") " pod="openstack/cinder-scheduler-0" Oct 05 07:08:07 crc kubenswrapper[4846]: I1005 07:08:07.366009 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/bfe96e0c-b9fc-438d-9b17-bec246d48446-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"bfe96e0c-b9fc-438d-9b17-bec246d48446\") " pod="openstack/cinder-scheduler-0" Oct 05 07:08:07 crc kubenswrapper[4846]: I1005 07:08:07.366058 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/bfe96e0c-b9fc-438d-9b17-bec246d48446-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"bfe96e0c-b9fc-438d-9b17-bec246d48446\") " pod="openstack/cinder-scheduler-0" Oct 05 07:08:07 crc kubenswrapper[4846]: I1005 07:08:07.366100 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bfe96e0c-b9fc-438d-9b17-bec246d48446-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"bfe96e0c-b9fc-438d-9b17-bec246d48446\") " pod="openstack/cinder-scheduler-0" Oct 05 07:08:07 crc kubenswrapper[4846]: I1005 07:08:07.366122 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hd6jb\" (UniqueName: \"kubernetes.io/projected/bfe96e0c-b9fc-438d-9b17-bec246d48446-kube-api-access-hd6jb\") pod \"cinder-scheduler-0\" (UID: \"bfe96e0c-b9fc-438d-9b17-bec246d48446\") " pod="openstack/cinder-scheduler-0" Oct 05 07:08:07 crc kubenswrapper[4846]: I1005 07:08:07.366662 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/bfe96e0c-b9fc-438d-9b17-bec246d48446-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"bfe96e0c-b9fc-438d-9b17-bec246d48446\") " pod="openstack/cinder-scheduler-0" Oct 05 07:08:07 crc kubenswrapper[4846]: I1005 07:08:07.375089 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bfe96e0c-b9fc-438d-9b17-bec246d48446-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"bfe96e0c-b9fc-438d-9b17-bec246d48446\") " pod="openstack/cinder-scheduler-0" Oct 05 07:08:07 crc kubenswrapper[4846]: I1005 07:08:07.375678 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bfe96e0c-b9fc-438d-9b17-bec246d48446-scripts\") pod \"cinder-scheduler-0\" (UID: \"bfe96e0c-b9fc-438d-9b17-bec246d48446\") " pod="openstack/cinder-scheduler-0" Oct 05 07:08:07 crc kubenswrapper[4846]: I1005 07:08:07.376226 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bfe96e0c-b9fc-438d-9b17-bec246d48446-config-data\") pod \"cinder-scheduler-0\" (UID: \"bfe96e0c-b9fc-438d-9b17-bec246d48446\") " pod="openstack/cinder-scheduler-0" Oct 05 07:08:07 crc kubenswrapper[4846]: I1005 07:08:07.384430 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/bfe96e0c-b9fc-438d-9b17-bec246d48446-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"bfe96e0c-b9fc-438d-9b17-bec246d48446\") " pod="openstack/cinder-scheduler-0" Oct 05 07:08:07 crc kubenswrapper[4846]: I1005 07:08:07.386699 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hd6jb\" (UniqueName: \"kubernetes.io/projected/bfe96e0c-b9fc-438d-9b17-bec246d48446-kube-api-access-hd6jb\") pod \"cinder-scheduler-0\" (UID: \"bfe96e0c-b9fc-438d-9b17-bec246d48446\") " pod="openstack/cinder-scheduler-0" Oct 05 07:08:07 crc kubenswrapper[4846]: I1005 07:08:07.523567 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 05 07:08:08 crc kubenswrapper[4846]: I1005 07:08:08.057641 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 05 07:08:08 crc kubenswrapper[4846]: I1005 07:08:08.118515 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"bfe96e0c-b9fc-438d-9b17-bec246d48446","Type":"ContainerStarted","Data":"69bab1a17f0a227c56ad117e5f561df14c1c0d6f2a4fd25886109fe5afbec8fe"} Oct 05 07:08:08 crc kubenswrapper[4846]: I1005 07:08:08.564900 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="40289760-4ac7-42e8-8eab-39f93c1618aa" path="/var/lib/kubelet/pods/40289760-4ac7-42e8-8eab-39f93c1618aa/volumes" Oct 05 07:08:08 crc kubenswrapper[4846]: I1005 07:08:08.905156 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-64c4f98476-pqrx9" Oct 05 07:08:09 crc kubenswrapper[4846]: I1005 07:08:09.006616 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/818246fd-88cc-4c04-8ea4-0e2521a0fc45-ovndb-tls-certs\") pod \"818246fd-88cc-4c04-8ea4-0e2521a0fc45\" (UID: \"818246fd-88cc-4c04-8ea4-0e2521a0fc45\") " Oct 05 07:08:09 crc kubenswrapper[4846]: I1005 07:08:09.006919 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g4645\" (UniqueName: \"kubernetes.io/projected/818246fd-88cc-4c04-8ea4-0e2521a0fc45-kube-api-access-g4645\") pod \"818246fd-88cc-4c04-8ea4-0e2521a0fc45\" (UID: \"818246fd-88cc-4c04-8ea4-0e2521a0fc45\") " Oct 05 07:08:09 crc kubenswrapper[4846]: I1005 07:08:09.006965 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/818246fd-88cc-4c04-8ea4-0e2521a0fc45-combined-ca-bundle\") pod \"818246fd-88cc-4c04-8ea4-0e2521a0fc45\" (UID: \"818246fd-88cc-4c04-8ea4-0e2521a0fc45\") " Oct 05 07:08:09 crc kubenswrapper[4846]: I1005 07:08:09.007015 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/818246fd-88cc-4c04-8ea4-0e2521a0fc45-config\") pod \"818246fd-88cc-4c04-8ea4-0e2521a0fc45\" (UID: \"818246fd-88cc-4c04-8ea4-0e2521a0fc45\") " Oct 05 07:08:09 crc kubenswrapper[4846]: I1005 07:08:09.007041 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/818246fd-88cc-4c04-8ea4-0e2521a0fc45-httpd-config\") pod \"818246fd-88cc-4c04-8ea4-0e2521a0fc45\" (UID: \"818246fd-88cc-4c04-8ea4-0e2521a0fc45\") " Oct 05 07:08:09 crc kubenswrapper[4846]: I1005 07:08:09.012079 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/818246fd-88cc-4c04-8ea4-0e2521a0fc45-kube-api-access-g4645" (OuterVolumeSpecName: "kube-api-access-g4645") pod "818246fd-88cc-4c04-8ea4-0e2521a0fc45" (UID: "818246fd-88cc-4c04-8ea4-0e2521a0fc45"). InnerVolumeSpecName "kube-api-access-g4645". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:08:09 crc kubenswrapper[4846]: I1005 07:08:09.015209 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/818246fd-88cc-4c04-8ea4-0e2521a0fc45-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "818246fd-88cc-4c04-8ea4-0e2521a0fc45" (UID: "818246fd-88cc-4c04-8ea4-0e2521a0fc45"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:08:09 crc kubenswrapper[4846]: I1005 07:08:09.083492 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/818246fd-88cc-4c04-8ea4-0e2521a0fc45-config" (OuterVolumeSpecName: "config") pod "818246fd-88cc-4c04-8ea4-0e2521a0fc45" (UID: "818246fd-88cc-4c04-8ea4-0e2521a0fc45"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:08:09 crc kubenswrapper[4846]: I1005 07:08:09.100606 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/818246fd-88cc-4c04-8ea4-0e2521a0fc45-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "818246fd-88cc-4c04-8ea4-0e2521a0fc45" (UID: "818246fd-88cc-4c04-8ea4-0e2521a0fc45"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:08:09 crc kubenswrapper[4846]: I1005 07:08:09.109640 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g4645\" (UniqueName: \"kubernetes.io/projected/818246fd-88cc-4c04-8ea4-0e2521a0fc45-kube-api-access-g4645\") on node \"crc\" DevicePath \"\"" Oct 05 07:08:09 crc kubenswrapper[4846]: I1005 07:08:09.109667 4846 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/818246fd-88cc-4c04-8ea4-0e2521a0fc45-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 07:08:09 crc kubenswrapper[4846]: I1005 07:08:09.109678 4846 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/818246fd-88cc-4c04-8ea4-0e2521a0fc45-config\") on node \"crc\" DevicePath \"\"" Oct 05 07:08:09 crc kubenswrapper[4846]: I1005 07:08:09.109689 4846 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/818246fd-88cc-4c04-8ea4-0e2521a0fc45-httpd-config\") on node \"crc\" DevicePath \"\"" Oct 05 07:08:09 crc kubenswrapper[4846]: I1005 07:08:09.128259 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/818246fd-88cc-4c04-8ea4-0e2521a0fc45-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "818246fd-88cc-4c04-8ea4-0e2521a0fc45" (UID: "818246fd-88cc-4c04-8ea4-0e2521a0fc45"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:08:09 crc kubenswrapper[4846]: I1005 07:08:09.153524 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"bfe96e0c-b9fc-438d-9b17-bec246d48446","Type":"ContainerStarted","Data":"1921e872cb6fabf10cd959960895a03e62a0d256a8d1c87f18ab5389a991e8f4"} Oct 05 07:08:09 crc kubenswrapper[4846]: I1005 07:08:09.155815 4846 generic.go:334] "Generic (PLEG): container finished" podID="818246fd-88cc-4c04-8ea4-0e2521a0fc45" containerID="d98e1b46601ac7b1c3f03252e286d71bbe2836d44edf5d2b5a62a2b0af440f28" exitCode=0 Oct 05 07:08:09 crc kubenswrapper[4846]: I1005 07:08:09.155887 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-64c4f98476-pqrx9" Oct 05 07:08:09 crc kubenswrapper[4846]: I1005 07:08:09.156196 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-64c4f98476-pqrx9" event={"ID":"818246fd-88cc-4c04-8ea4-0e2521a0fc45","Type":"ContainerDied","Data":"d98e1b46601ac7b1c3f03252e286d71bbe2836d44edf5d2b5a62a2b0af440f28"} Oct 05 07:08:09 crc kubenswrapper[4846]: I1005 07:08:09.156382 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-64c4f98476-pqrx9" event={"ID":"818246fd-88cc-4c04-8ea4-0e2521a0fc45","Type":"ContainerDied","Data":"a3d50da89a9a16db1a951b96ed40e1e10d58c570efdd288e5afcad3c9fd2e14c"} Oct 05 07:08:09 crc kubenswrapper[4846]: I1005 07:08:09.156435 4846 scope.go:117] "RemoveContainer" containerID="1b5541d40b35256864b3d80cd509b1fb37d2f2e7b5097dfa124f6488c42f7d2f" Oct 05 07:08:09 crc kubenswrapper[4846]: I1005 07:08:09.187897 4846 scope.go:117] "RemoveContainer" containerID="d98e1b46601ac7b1c3f03252e286d71bbe2836d44edf5d2b5a62a2b0af440f28" Oct 05 07:08:09 crc kubenswrapper[4846]: I1005 07:08:09.195982 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-64c4f98476-pqrx9"] Oct 05 07:08:09 crc kubenswrapper[4846]: I1005 07:08:09.211885 4846 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/818246fd-88cc-4c04-8ea4-0e2521a0fc45-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 05 07:08:09 crc kubenswrapper[4846]: I1005 07:08:09.214764 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-64c4f98476-pqrx9"] Oct 05 07:08:09 crc kubenswrapper[4846]: I1005 07:08:09.226294 4846 scope.go:117] "RemoveContainer" containerID="1b5541d40b35256864b3d80cd509b1fb37d2f2e7b5097dfa124f6488c42f7d2f" Oct 05 07:08:09 crc kubenswrapper[4846]: E1005 07:08:09.226744 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1b5541d40b35256864b3d80cd509b1fb37d2f2e7b5097dfa124f6488c42f7d2f\": container with ID starting with 1b5541d40b35256864b3d80cd509b1fb37d2f2e7b5097dfa124f6488c42f7d2f not found: ID does not exist" containerID="1b5541d40b35256864b3d80cd509b1fb37d2f2e7b5097dfa124f6488c42f7d2f" Oct 05 07:08:09 crc kubenswrapper[4846]: I1005 07:08:09.226779 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1b5541d40b35256864b3d80cd509b1fb37d2f2e7b5097dfa124f6488c42f7d2f"} err="failed to get container status \"1b5541d40b35256864b3d80cd509b1fb37d2f2e7b5097dfa124f6488c42f7d2f\": rpc error: code = NotFound desc = could not find container \"1b5541d40b35256864b3d80cd509b1fb37d2f2e7b5097dfa124f6488c42f7d2f\": container with ID starting with 1b5541d40b35256864b3d80cd509b1fb37d2f2e7b5097dfa124f6488c42f7d2f not found: ID does not exist" Oct 05 07:08:09 crc kubenswrapper[4846]: I1005 07:08:09.226801 4846 scope.go:117] "RemoveContainer" containerID="d98e1b46601ac7b1c3f03252e286d71bbe2836d44edf5d2b5a62a2b0af440f28" Oct 05 07:08:09 crc kubenswrapper[4846]: E1005 07:08:09.227014 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d98e1b46601ac7b1c3f03252e286d71bbe2836d44edf5d2b5a62a2b0af440f28\": container with ID starting with d98e1b46601ac7b1c3f03252e286d71bbe2836d44edf5d2b5a62a2b0af440f28 not found: ID does not exist" containerID="d98e1b46601ac7b1c3f03252e286d71bbe2836d44edf5d2b5a62a2b0af440f28" Oct 05 07:08:09 crc kubenswrapper[4846]: I1005 07:08:09.227075 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d98e1b46601ac7b1c3f03252e286d71bbe2836d44edf5d2b5a62a2b0af440f28"} err="failed to get container status \"d98e1b46601ac7b1c3f03252e286d71bbe2836d44edf5d2b5a62a2b0af440f28\": rpc error: code = NotFound desc = could not find container \"d98e1b46601ac7b1c3f03252e286d71bbe2836d44edf5d2b5a62a2b0af440f28\": container with ID starting with d98e1b46601ac7b1c3f03252e286d71bbe2836d44edf5d2b5a62a2b0af440f28 not found: ID does not exist" Oct 05 07:08:09 crc kubenswrapper[4846]: I1005 07:08:09.864649 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Oct 05 07:08:09 crc kubenswrapper[4846]: E1005 07:08:09.865797 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="818246fd-88cc-4c04-8ea4-0e2521a0fc45" containerName="neutron-httpd" Oct 05 07:08:09 crc kubenswrapper[4846]: I1005 07:08:09.865816 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="818246fd-88cc-4c04-8ea4-0e2521a0fc45" containerName="neutron-httpd" Oct 05 07:08:09 crc kubenswrapper[4846]: E1005 07:08:09.865850 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="818246fd-88cc-4c04-8ea4-0e2521a0fc45" containerName="neutron-api" Oct 05 07:08:09 crc kubenswrapper[4846]: I1005 07:08:09.865858 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="818246fd-88cc-4c04-8ea4-0e2521a0fc45" containerName="neutron-api" Oct 05 07:08:09 crc kubenswrapper[4846]: I1005 07:08:09.866122 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="818246fd-88cc-4c04-8ea4-0e2521a0fc45" containerName="neutron-api" Oct 05 07:08:09 crc kubenswrapper[4846]: I1005 07:08:09.866153 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="818246fd-88cc-4c04-8ea4-0e2521a0fc45" containerName="neutron-httpd" Oct 05 07:08:09 crc kubenswrapper[4846]: I1005 07:08:09.867296 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 05 07:08:09 crc kubenswrapper[4846]: I1005 07:08:09.876011 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstackclient-openstackclient-dockercfg-pqvdf" Oct 05 07:08:09 crc kubenswrapper[4846]: I1005 07:08:09.876170 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config" Oct 05 07:08:09 crc kubenswrapper[4846]: I1005 07:08:09.877038 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-config-secret" Oct 05 07:08:09 crc kubenswrapper[4846]: I1005 07:08:09.893052 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Oct 05 07:08:09 crc kubenswrapper[4846]: I1005 07:08:09.928613 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/98fc38b8-a0ee-46c9-9d3f-9c15e8355b9d-openstack-config\") pod \"openstackclient\" (UID: \"98fc38b8-a0ee-46c9-9d3f-9c15e8355b9d\") " pod="openstack/openstackclient" Oct 05 07:08:09 crc kubenswrapper[4846]: I1005 07:08:09.928691 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/98fc38b8-a0ee-46c9-9d3f-9c15e8355b9d-openstack-config-secret\") pod \"openstackclient\" (UID: \"98fc38b8-a0ee-46c9-9d3f-9c15e8355b9d\") " pod="openstack/openstackclient" Oct 05 07:08:09 crc kubenswrapper[4846]: I1005 07:08:09.928781 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qdvx8\" (UniqueName: \"kubernetes.io/projected/98fc38b8-a0ee-46c9-9d3f-9c15e8355b9d-kube-api-access-qdvx8\") pod \"openstackclient\" (UID: \"98fc38b8-a0ee-46c9-9d3f-9c15e8355b9d\") " pod="openstack/openstackclient" Oct 05 07:08:09 crc kubenswrapper[4846]: I1005 07:08:09.928887 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/98fc38b8-a0ee-46c9-9d3f-9c15e8355b9d-combined-ca-bundle\") pod \"openstackclient\" (UID: \"98fc38b8-a0ee-46c9-9d3f-9c15e8355b9d\") " pod="openstack/openstackclient" Oct 05 07:08:10 crc kubenswrapper[4846]: I1005 07:08:10.031058 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qdvx8\" (UniqueName: \"kubernetes.io/projected/98fc38b8-a0ee-46c9-9d3f-9c15e8355b9d-kube-api-access-qdvx8\") pod \"openstackclient\" (UID: \"98fc38b8-a0ee-46c9-9d3f-9c15e8355b9d\") " pod="openstack/openstackclient" Oct 05 07:08:10 crc kubenswrapper[4846]: I1005 07:08:10.031261 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/98fc38b8-a0ee-46c9-9d3f-9c15e8355b9d-combined-ca-bundle\") pod \"openstackclient\" (UID: \"98fc38b8-a0ee-46c9-9d3f-9c15e8355b9d\") " pod="openstack/openstackclient" Oct 05 07:08:10 crc kubenswrapper[4846]: I1005 07:08:10.031341 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/98fc38b8-a0ee-46c9-9d3f-9c15e8355b9d-openstack-config\") pod \"openstackclient\" (UID: \"98fc38b8-a0ee-46c9-9d3f-9c15e8355b9d\") " pod="openstack/openstackclient" Oct 05 07:08:10 crc kubenswrapper[4846]: I1005 07:08:10.031365 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/98fc38b8-a0ee-46c9-9d3f-9c15e8355b9d-openstack-config-secret\") pod \"openstackclient\" (UID: \"98fc38b8-a0ee-46c9-9d3f-9c15e8355b9d\") " pod="openstack/openstackclient" Oct 05 07:08:10 crc kubenswrapper[4846]: I1005 07:08:10.032689 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/98fc38b8-a0ee-46c9-9d3f-9c15e8355b9d-openstack-config\") pod \"openstackclient\" (UID: \"98fc38b8-a0ee-46c9-9d3f-9c15e8355b9d\") " pod="openstack/openstackclient" Oct 05 07:08:10 crc kubenswrapper[4846]: I1005 07:08:10.044022 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/98fc38b8-a0ee-46c9-9d3f-9c15e8355b9d-openstack-config-secret\") pod \"openstackclient\" (UID: \"98fc38b8-a0ee-46c9-9d3f-9c15e8355b9d\") " pod="openstack/openstackclient" Oct 05 07:08:10 crc kubenswrapper[4846]: I1005 07:08:10.048404 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qdvx8\" (UniqueName: \"kubernetes.io/projected/98fc38b8-a0ee-46c9-9d3f-9c15e8355b9d-kube-api-access-qdvx8\") pod \"openstackclient\" (UID: \"98fc38b8-a0ee-46c9-9d3f-9c15e8355b9d\") " pod="openstack/openstackclient" Oct 05 07:08:10 crc kubenswrapper[4846]: I1005 07:08:10.049929 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/98fc38b8-a0ee-46c9-9d3f-9c15e8355b9d-combined-ca-bundle\") pod \"openstackclient\" (UID: \"98fc38b8-a0ee-46c9-9d3f-9c15e8355b9d\") " pod="openstack/openstackclient" Oct 05 07:08:10 crc kubenswrapper[4846]: I1005 07:08:10.167332 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"bfe96e0c-b9fc-438d-9b17-bec246d48446","Type":"ContainerStarted","Data":"470aec4a8b7085d2d3371c5366c57378850aca957d883939f98653f186109a67"} Oct 05 07:08:10 crc kubenswrapper[4846]: I1005 07:08:10.193502 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 05 07:08:10 crc kubenswrapper[4846]: I1005 07:08:10.198233 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=3.198208537 podStartE2EDuration="3.198208537s" podCreationTimestamp="2025-10-05 07:08:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 07:08:10.18894918 +0000 UTC m=+1212.429801955" watchObservedRunningTime="2025-10-05 07:08:10.198208537 +0000 UTC m=+1212.439061312" Oct 05 07:08:10 crc kubenswrapper[4846]: I1005 07:08:10.568630 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="818246fd-88cc-4c04-8ea4-0e2521a0fc45" path="/var/lib/kubelet/pods/818246fd-88cc-4c04-8ea4-0e2521a0fc45/volumes" Oct 05 07:08:10 crc kubenswrapper[4846]: W1005 07:08:10.730423 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod98fc38b8_a0ee_46c9_9d3f_9c15e8355b9d.slice/crio-e466bda1c968cbdc08d15635a847111c1a6de4dc38f77ced845f9d57710ef0da WatchSource:0}: Error finding container e466bda1c968cbdc08d15635a847111c1a6de4dc38f77ced845f9d57710ef0da: Status 404 returned error can't find the container with id e466bda1c968cbdc08d15635a847111c1a6de4dc38f77ced845f9d57710ef0da Oct 05 07:08:10 crc kubenswrapper[4846]: I1005 07:08:10.733696 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Oct 05 07:08:11 crc kubenswrapper[4846]: I1005 07:08:11.007534 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-57674d6c54-zn5nq" Oct 05 07:08:11 crc kubenswrapper[4846]: I1005 07:08:11.183489 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"98fc38b8-a0ee-46c9-9d3f-9c15e8355b9d","Type":"ContainerStarted","Data":"e466bda1c968cbdc08d15635a847111c1a6de4dc38f77ced845f9d57710ef0da"} Oct 05 07:08:11 crc kubenswrapper[4846]: I1005 07:08:11.213930 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-57674d6c54-zn5nq" Oct 05 07:08:11 crc kubenswrapper[4846]: I1005 07:08:11.295319 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-6c9ddc8584-bqgtc"] Oct 05 07:08:11 crc kubenswrapper[4846]: I1005 07:08:11.295683 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-6c9ddc8584-bqgtc" podUID="ba9102c0-185f-4f02-8b05-8bb7fde2d573" containerName="barbican-api-log" containerID="cri-o://39111fc9d7fc57d402d3aff9ff623348099e37f99ec5b23eaa70a56e39c540b9" gracePeriod=30 Oct 05 07:08:11 crc kubenswrapper[4846]: I1005 07:08:11.295856 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-6c9ddc8584-bqgtc" podUID="ba9102c0-185f-4f02-8b05-8bb7fde2d573" containerName="barbican-api" containerID="cri-o://e5bd6644aa631daa23e2517be3eb2c9e0b0f0672ae63eae56fb2b30510a9cb24" gracePeriod=30 Oct 05 07:08:12 crc kubenswrapper[4846]: I1005 07:08:12.195568 4846 generic.go:334] "Generic (PLEG): container finished" podID="ba9102c0-185f-4f02-8b05-8bb7fde2d573" containerID="39111fc9d7fc57d402d3aff9ff623348099e37f99ec5b23eaa70a56e39c540b9" exitCode=143 Oct 05 07:08:12 crc kubenswrapper[4846]: I1005 07:08:12.196401 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6c9ddc8584-bqgtc" event={"ID":"ba9102c0-185f-4f02-8b05-8bb7fde2d573","Type":"ContainerDied","Data":"39111fc9d7fc57d402d3aff9ff623348099e37f99ec5b23eaa70a56e39c540b9"} Oct 05 07:08:12 crc kubenswrapper[4846]: I1005 07:08:12.525394 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Oct 05 07:08:14 crc kubenswrapper[4846]: I1005 07:08:14.376888 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Oct 05 07:08:14 crc kubenswrapper[4846]: I1005 07:08:14.523041 4846 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-6c9ddc8584-bqgtc" podUID="ba9102c0-185f-4f02-8b05-8bb7fde2d573" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.164:9311/healthcheck\": read tcp 10.217.0.2:60164->10.217.0.164:9311: read: connection reset by peer" Oct 05 07:08:14 crc kubenswrapper[4846]: I1005 07:08:14.523472 4846 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-6c9ddc8584-bqgtc" podUID="ba9102c0-185f-4f02-8b05-8bb7fde2d573" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.164:9311/healthcheck\": read tcp 10.217.0.2:60152->10.217.0.164:9311: read: connection reset by peer" Oct 05 07:08:14 crc kubenswrapper[4846]: I1005 07:08:14.528376 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-proxy-984785977-srp2v"] Oct 05 07:08:14 crc kubenswrapper[4846]: I1005 07:08:14.529931 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-984785977-srp2v" Oct 05 07:08:14 crc kubenswrapper[4846]: I1005 07:08:14.533024 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Oct 05 07:08:14 crc kubenswrapper[4846]: I1005 07:08:14.533489 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-internal-svc" Oct 05 07:08:14 crc kubenswrapper[4846]: I1005 07:08:14.533494 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-public-svc" Oct 05 07:08:14 crc kubenswrapper[4846]: I1005 07:08:14.562764 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-984785977-srp2v"] Oct 05 07:08:14 crc kubenswrapper[4846]: I1005 07:08:14.640127 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/d11b6df6-6da2-4cb8-a4f9-d87cfd5d4d26-etc-swift\") pod \"swift-proxy-984785977-srp2v\" (UID: \"d11b6df6-6da2-4cb8-a4f9-d87cfd5d4d26\") " pod="openstack/swift-proxy-984785977-srp2v" Oct 05 07:08:14 crc kubenswrapper[4846]: I1005 07:08:14.640237 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d11b6df6-6da2-4cb8-a4f9-d87cfd5d4d26-internal-tls-certs\") pod \"swift-proxy-984785977-srp2v\" (UID: \"d11b6df6-6da2-4cb8-a4f9-d87cfd5d4d26\") " pod="openstack/swift-proxy-984785977-srp2v" Oct 05 07:08:14 crc kubenswrapper[4846]: I1005 07:08:14.640311 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d11b6df6-6da2-4cb8-a4f9-d87cfd5d4d26-config-data\") pod \"swift-proxy-984785977-srp2v\" (UID: \"d11b6df6-6da2-4cb8-a4f9-d87cfd5d4d26\") " pod="openstack/swift-proxy-984785977-srp2v" Oct 05 07:08:14 crc kubenswrapper[4846]: I1005 07:08:14.640383 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d11b6df6-6da2-4cb8-a4f9-d87cfd5d4d26-run-httpd\") pod \"swift-proxy-984785977-srp2v\" (UID: \"d11b6df6-6da2-4cb8-a4f9-d87cfd5d4d26\") " pod="openstack/swift-proxy-984785977-srp2v" Oct 05 07:08:14 crc kubenswrapper[4846]: I1005 07:08:14.640541 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d11b6df6-6da2-4cb8-a4f9-d87cfd5d4d26-public-tls-certs\") pod \"swift-proxy-984785977-srp2v\" (UID: \"d11b6df6-6da2-4cb8-a4f9-d87cfd5d4d26\") " pod="openstack/swift-proxy-984785977-srp2v" Oct 05 07:08:14 crc kubenswrapper[4846]: I1005 07:08:14.640654 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d11b6df6-6da2-4cb8-a4f9-d87cfd5d4d26-combined-ca-bundle\") pod \"swift-proxy-984785977-srp2v\" (UID: \"d11b6df6-6da2-4cb8-a4f9-d87cfd5d4d26\") " pod="openstack/swift-proxy-984785977-srp2v" Oct 05 07:08:14 crc kubenswrapper[4846]: I1005 07:08:14.640682 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d11b6df6-6da2-4cb8-a4f9-d87cfd5d4d26-log-httpd\") pod \"swift-proxy-984785977-srp2v\" (UID: \"d11b6df6-6da2-4cb8-a4f9-d87cfd5d4d26\") " pod="openstack/swift-proxy-984785977-srp2v" Oct 05 07:08:14 crc kubenswrapper[4846]: I1005 07:08:14.640710 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-74pmj\" (UniqueName: \"kubernetes.io/projected/d11b6df6-6da2-4cb8-a4f9-d87cfd5d4d26-kube-api-access-74pmj\") pod \"swift-proxy-984785977-srp2v\" (UID: \"d11b6df6-6da2-4cb8-a4f9-d87cfd5d4d26\") " pod="openstack/swift-proxy-984785977-srp2v" Oct 05 07:08:14 crc kubenswrapper[4846]: I1005 07:08:14.743880 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d11b6df6-6da2-4cb8-a4f9-d87cfd5d4d26-combined-ca-bundle\") pod \"swift-proxy-984785977-srp2v\" (UID: \"d11b6df6-6da2-4cb8-a4f9-d87cfd5d4d26\") " pod="openstack/swift-proxy-984785977-srp2v" Oct 05 07:08:14 crc kubenswrapper[4846]: I1005 07:08:14.743944 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d11b6df6-6da2-4cb8-a4f9-d87cfd5d4d26-log-httpd\") pod \"swift-proxy-984785977-srp2v\" (UID: \"d11b6df6-6da2-4cb8-a4f9-d87cfd5d4d26\") " pod="openstack/swift-proxy-984785977-srp2v" Oct 05 07:08:14 crc kubenswrapper[4846]: I1005 07:08:14.743969 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-74pmj\" (UniqueName: \"kubernetes.io/projected/d11b6df6-6da2-4cb8-a4f9-d87cfd5d4d26-kube-api-access-74pmj\") pod \"swift-proxy-984785977-srp2v\" (UID: \"d11b6df6-6da2-4cb8-a4f9-d87cfd5d4d26\") " pod="openstack/swift-proxy-984785977-srp2v" Oct 05 07:08:14 crc kubenswrapper[4846]: I1005 07:08:14.744010 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/d11b6df6-6da2-4cb8-a4f9-d87cfd5d4d26-etc-swift\") pod \"swift-proxy-984785977-srp2v\" (UID: \"d11b6df6-6da2-4cb8-a4f9-d87cfd5d4d26\") " pod="openstack/swift-proxy-984785977-srp2v" Oct 05 07:08:14 crc kubenswrapper[4846]: I1005 07:08:14.744031 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d11b6df6-6da2-4cb8-a4f9-d87cfd5d4d26-internal-tls-certs\") pod \"swift-proxy-984785977-srp2v\" (UID: \"d11b6df6-6da2-4cb8-a4f9-d87cfd5d4d26\") " pod="openstack/swift-proxy-984785977-srp2v" Oct 05 07:08:14 crc kubenswrapper[4846]: I1005 07:08:14.744061 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d11b6df6-6da2-4cb8-a4f9-d87cfd5d4d26-config-data\") pod \"swift-proxy-984785977-srp2v\" (UID: \"d11b6df6-6da2-4cb8-a4f9-d87cfd5d4d26\") " pod="openstack/swift-proxy-984785977-srp2v" Oct 05 07:08:14 crc kubenswrapper[4846]: I1005 07:08:14.744095 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d11b6df6-6da2-4cb8-a4f9-d87cfd5d4d26-run-httpd\") pod \"swift-proxy-984785977-srp2v\" (UID: \"d11b6df6-6da2-4cb8-a4f9-d87cfd5d4d26\") " pod="openstack/swift-proxy-984785977-srp2v" Oct 05 07:08:14 crc kubenswrapper[4846]: I1005 07:08:14.744166 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d11b6df6-6da2-4cb8-a4f9-d87cfd5d4d26-public-tls-certs\") pod \"swift-proxy-984785977-srp2v\" (UID: \"d11b6df6-6da2-4cb8-a4f9-d87cfd5d4d26\") " pod="openstack/swift-proxy-984785977-srp2v" Oct 05 07:08:14 crc kubenswrapper[4846]: I1005 07:08:14.746465 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d11b6df6-6da2-4cb8-a4f9-d87cfd5d4d26-run-httpd\") pod \"swift-proxy-984785977-srp2v\" (UID: \"d11b6df6-6da2-4cb8-a4f9-d87cfd5d4d26\") " pod="openstack/swift-proxy-984785977-srp2v" Oct 05 07:08:14 crc kubenswrapper[4846]: I1005 07:08:14.746761 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d11b6df6-6da2-4cb8-a4f9-d87cfd5d4d26-log-httpd\") pod \"swift-proxy-984785977-srp2v\" (UID: \"d11b6df6-6da2-4cb8-a4f9-d87cfd5d4d26\") " pod="openstack/swift-proxy-984785977-srp2v" Oct 05 07:08:14 crc kubenswrapper[4846]: I1005 07:08:14.754771 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/d11b6df6-6da2-4cb8-a4f9-d87cfd5d4d26-etc-swift\") pod \"swift-proxy-984785977-srp2v\" (UID: \"d11b6df6-6da2-4cb8-a4f9-d87cfd5d4d26\") " pod="openstack/swift-proxy-984785977-srp2v" Oct 05 07:08:14 crc kubenswrapper[4846]: I1005 07:08:14.766285 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d11b6df6-6da2-4cb8-a4f9-d87cfd5d4d26-config-data\") pod \"swift-proxy-984785977-srp2v\" (UID: \"d11b6df6-6da2-4cb8-a4f9-d87cfd5d4d26\") " pod="openstack/swift-proxy-984785977-srp2v" Oct 05 07:08:14 crc kubenswrapper[4846]: I1005 07:08:14.777254 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d11b6df6-6da2-4cb8-a4f9-d87cfd5d4d26-combined-ca-bundle\") pod \"swift-proxy-984785977-srp2v\" (UID: \"d11b6df6-6da2-4cb8-a4f9-d87cfd5d4d26\") " pod="openstack/swift-proxy-984785977-srp2v" Oct 05 07:08:14 crc kubenswrapper[4846]: I1005 07:08:14.786064 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d11b6df6-6da2-4cb8-a4f9-d87cfd5d4d26-public-tls-certs\") pod \"swift-proxy-984785977-srp2v\" (UID: \"d11b6df6-6da2-4cb8-a4f9-d87cfd5d4d26\") " pod="openstack/swift-proxy-984785977-srp2v" Oct 05 07:08:14 crc kubenswrapper[4846]: I1005 07:08:14.787585 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-74pmj\" (UniqueName: \"kubernetes.io/projected/d11b6df6-6da2-4cb8-a4f9-d87cfd5d4d26-kube-api-access-74pmj\") pod \"swift-proxy-984785977-srp2v\" (UID: \"d11b6df6-6da2-4cb8-a4f9-d87cfd5d4d26\") " pod="openstack/swift-proxy-984785977-srp2v" Oct 05 07:08:14 crc kubenswrapper[4846]: I1005 07:08:14.791663 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d11b6df6-6da2-4cb8-a4f9-d87cfd5d4d26-internal-tls-certs\") pod \"swift-proxy-984785977-srp2v\" (UID: \"d11b6df6-6da2-4cb8-a4f9-d87cfd5d4d26\") " pod="openstack/swift-proxy-984785977-srp2v" Oct 05 07:08:14 crc kubenswrapper[4846]: I1005 07:08:14.868615 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 05 07:08:14 crc kubenswrapper[4846]: I1005 07:08:14.869252 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="39071bf4-214b-4bbb-8b78-dfec6ad3c940" containerName="ceilometer-central-agent" containerID="cri-o://82b673a1934d34adb80bf16feac56c7fc2158772b630ab377b61d9b3c475954b" gracePeriod=30 Oct 05 07:08:14 crc kubenswrapper[4846]: I1005 07:08:14.869722 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="39071bf4-214b-4bbb-8b78-dfec6ad3c940" containerName="proxy-httpd" containerID="cri-o://72d54baff0ab5bf146fb7a95307526bd069594408c4da5a11bbeacf29306be6c" gracePeriod=30 Oct 05 07:08:14 crc kubenswrapper[4846]: I1005 07:08:14.869791 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="39071bf4-214b-4bbb-8b78-dfec6ad3c940" containerName="sg-core" containerID="cri-o://82e6c4bfad41089daffabb46ada2064751a4f47c40780fda32f8ec7989bd6a30" gracePeriod=30 Oct 05 07:08:14 crc kubenswrapper[4846]: I1005 07:08:14.869846 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="39071bf4-214b-4bbb-8b78-dfec6ad3c940" containerName="ceilometer-notification-agent" containerID="cri-o://15e05c1324747c5238b20b4131b788c1adc0d31ec16d39fc7037ec91c627a278" gracePeriod=30 Oct 05 07:08:14 crc kubenswrapper[4846]: I1005 07:08:14.882385 4846 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="39071bf4-214b-4bbb-8b78-dfec6ad3c940" containerName="proxy-httpd" probeResult="failure" output="Get \"http://10.217.0.160:3000/\": EOF" Oct 05 07:08:14 crc kubenswrapper[4846]: I1005 07:08:14.939028 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-984785977-srp2v" Oct 05 07:08:15 crc kubenswrapper[4846]: I1005 07:08:15.109934 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-6c9ddc8584-bqgtc" Oct 05 07:08:15 crc kubenswrapper[4846]: I1005 07:08:15.159963 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ba9102c0-185f-4f02-8b05-8bb7fde2d573-logs\") pod \"ba9102c0-185f-4f02-8b05-8bb7fde2d573\" (UID: \"ba9102c0-185f-4f02-8b05-8bb7fde2d573\") " Oct 05 07:08:15 crc kubenswrapper[4846]: I1005 07:08:15.160214 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ba9102c0-185f-4f02-8b05-8bb7fde2d573-config-data-custom\") pod \"ba9102c0-185f-4f02-8b05-8bb7fde2d573\" (UID: \"ba9102c0-185f-4f02-8b05-8bb7fde2d573\") " Oct 05 07:08:15 crc kubenswrapper[4846]: I1005 07:08:15.160264 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-chqvs\" (UniqueName: \"kubernetes.io/projected/ba9102c0-185f-4f02-8b05-8bb7fde2d573-kube-api-access-chqvs\") pod \"ba9102c0-185f-4f02-8b05-8bb7fde2d573\" (UID: \"ba9102c0-185f-4f02-8b05-8bb7fde2d573\") " Oct 05 07:08:15 crc kubenswrapper[4846]: I1005 07:08:15.160306 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ba9102c0-185f-4f02-8b05-8bb7fde2d573-config-data\") pod \"ba9102c0-185f-4f02-8b05-8bb7fde2d573\" (UID: \"ba9102c0-185f-4f02-8b05-8bb7fde2d573\") " Oct 05 07:08:15 crc kubenswrapper[4846]: I1005 07:08:15.160405 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ba9102c0-185f-4f02-8b05-8bb7fde2d573-combined-ca-bundle\") pod \"ba9102c0-185f-4f02-8b05-8bb7fde2d573\" (UID: \"ba9102c0-185f-4f02-8b05-8bb7fde2d573\") " Oct 05 07:08:15 crc kubenswrapper[4846]: I1005 07:08:15.162161 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ba9102c0-185f-4f02-8b05-8bb7fde2d573-logs" (OuterVolumeSpecName: "logs") pod "ba9102c0-185f-4f02-8b05-8bb7fde2d573" (UID: "ba9102c0-185f-4f02-8b05-8bb7fde2d573"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 07:08:15 crc kubenswrapper[4846]: I1005 07:08:15.167138 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ba9102c0-185f-4f02-8b05-8bb7fde2d573-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "ba9102c0-185f-4f02-8b05-8bb7fde2d573" (UID: "ba9102c0-185f-4f02-8b05-8bb7fde2d573"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:08:15 crc kubenswrapper[4846]: I1005 07:08:15.167808 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ba9102c0-185f-4f02-8b05-8bb7fde2d573-kube-api-access-chqvs" (OuterVolumeSpecName: "kube-api-access-chqvs") pod "ba9102c0-185f-4f02-8b05-8bb7fde2d573" (UID: "ba9102c0-185f-4f02-8b05-8bb7fde2d573"). InnerVolumeSpecName "kube-api-access-chqvs". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:08:15 crc kubenswrapper[4846]: I1005 07:08:15.218383 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ba9102c0-185f-4f02-8b05-8bb7fde2d573-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ba9102c0-185f-4f02-8b05-8bb7fde2d573" (UID: "ba9102c0-185f-4f02-8b05-8bb7fde2d573"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:08:15 crc kubenswrapper[4846]: I1005 07:08:15.244783 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ba9102c0-185f-4f02-8b05-8bb7fde2d573-config-data" (OuterVolumeSpecName: "config-data") pod "ba9102c0-185f-4f02-8b05-8bb7fde2d573" (UID: "ba9102c0-185f-4f02-8b05-8bb7fde2d573"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:08:15 crc kubenswrapper[4846]: I1005 07:08:15.263923 4846 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ba9102c0-185f-4f02-8b05-8bb7fde2d573-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 07:08:15 crc kubenswrapper[4846]: I1005 07:08:15.263978 4846 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ba9102c0-185f-4f02-8b05-8bb7fde2d573-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 07:08:15 crc kubenswrapper[4846]: I1005 07:08:15.263996 4846 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ba9102c0-185f-4f02-8b05-8bb7fde2d573-logs\") on node \"crc\" DevicePath \"\"" Oct 05 07:08:15 crc kubenswrapper[4846]: I1005 07:08:15.264006 4846 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ba9102c0-185f-4f02-8b05-8bb7fde2d573-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 05 07:08:15 crc kubenswrapper[4846]: I1005 07:08:15.264015 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-chqvs\" (UniqueName: \"kubernetes.io/projected/ba9102c0-185f-4f02-8b05-8bb7fde2d573-kube-api-access-chqvs\") on node \"crc\" DevicePath \"\"" Oct 05 07:08:15 crc kubenswrapper[4846]: I1005 07:08:15.265122 4846 generic.go:334] "Generic (PLEG): container finished" podID="39071bf4-214b-4bbb-8b78-dfec6ad3c940" containerID="72d54baff0ab5bf146fb7a95307526bd069594408c4da5a11bbeacf29306be6c" exitCode=0 Oct 05 07:08:15 crc kubenswrapper[4846]: I1005 07:08:15.265166 4846 generic.go:334] "Generic (PLEG): container finished" podID="39071bf4-214b-4bbb-8b78-dfec6ad3c940" containerID="82e6c4bfad41089daffabb46ada2064751a4f47c40780fda32f8ec7989bd6a30" exitCode=2 Oct 05 07:08:15 crc kubenswrapper[4846]: I1005 07:08:15.265280 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"39071bf4-214b-4bbb-8b78-dfec6ad3c940","Type":"ContainerDied","Data":"72d54baff0ab5bf146fb7a95307526bd069594408c4da5a11bbeacf29306be6c"} Oct 05 07:08:15 crc kubenswrapper[4846]: I1005 07:08:15.265322 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"39071bf4-214b-4bbb-8b78-dfec6ad3c940","Type":"ContainerDied","Data":"82e6c4bfad41089daffabb46ada2064751a4f47c40780fda32f8ec7989bd6a30"} Oct 05 07:08:15 crc kubenswrapper[4846]: I1005 07:08:15.275395 4846 generic.go:334] "Generic (PLEG): container finished" podID="ba9102c0-185f-4f02-8b05-8bb7fde2d573" containerID="e5bd6644aa631daa23e2517be3eb2c9e0b0f0672ae63eae56fb2b30510a9cb24" exitCode=0 Oct 05 07:08:15 crc kubenswrapper[4846]: I1005 07:08:15.275446 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6c9ddc8584-bqgtc" event={"ID":"ba9102c0-185f-4f02-8b05-8bb7fde2d573","Type":"ContainerDied","Data":"e5bd6644aa631daa23e2517be3eb2c9e0b0f0672ae63eae56fb2b30510a9cb24"} Oct 05 07:08:15 crc kubenswrapper[4846]: I1005 07:08:15.275505 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6c9ddc8584-bqgtc" event={"ID":"ba9102c0-185f-4f02-8b05-8bb7fde2d573","Type":"ContainerDied","Data":"ce32162fd7737fb25fa034d9314322010c146074fc19ea1fd62cf8b2dcfcb7ac"} Oct 05 07:08:15 crc kubenswrapper[4846]: I1005 07:08:15.275528 4846 scope.go:117] "RemoveContainer" containerID="e5bd6644aa631daa23e2517be3eb2c9e0b0f0672ae63eae56fb2b30510a9cb24" Oct 05 07:08:15 crc kubenswrapper[4846]: I1005 07:08:15.275545 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-6c9ddc8584-bqgtc" Oct 05 07:08:15 crc kubenswrapper[4846]: I1005 07:08:15.314837 4846 scope.go:117] "RemoveContainer" containerID="39111fc9d7fc57d402d3aff9ff623348099e37f99ec5b23eaa70a56e39c540b9" Oct 05 07:08:15 crc kubenswrapper[4846]: I1005 07:08:15.326557 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-6c9ddc8584-bqgtc"] Oct 05 07:08:15 crc kubenswrapper[4846]: I1005 07:08:15.338053 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-6c9ddc8584-bqgtc"] Oct 05 07:08:15 crc kubenswrapper[4846]: I1005 07:08:15.345237 4846 scope.go:117] "RemoveContainer" containerID="e5bd6644aa631daa23e2517be3eb2c9e0b0f0672ae63eae56fb2b30510a9cb24" Oct 05 07:08:15 crc kubenswrapper[4846]: E1005 07:08:15.348269 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e5bd6644aa631daa23e2517be3eb2c9e0b0f0672ae63eae56fb2b30510a9cb24\": container with ID starting with e5bd6644aa631daa23e2517be3eb2c9e0b0f0672ae63eae56fb2b30510a9cb24 not found: ID does not exist" containerID="e5bd6644aa631daa23e2517be3eb2c9e0b0f0672ae63eae56fb2b30510a9cb24" Oct 05 07:08:15 crc kubenswrapper[4846]: I1005 07:08:15.348559 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e5bd6644aa631daa23e2517be3eb2c9e0b0f0672ae63eae56fb2b30510a9cb24"} err="failed to get container status \"e5bd6644aa631daa23e2517be3eb2c9e0b0f0672ae63eae56fb2b30510a9cb24\": rpc error: code = NotFound desc = could not find container \"e5bd6644aa631daa23e2517be3eb2c9e0b0f0672ae63eae56fb2b30510a9cb24\": container with ID starting with e5bd6644aa631daa23e2517be3eb2c9e0b0f0672ae63eae56fb2b30510a9cb24 not found: ID does not exist" Oct 05 07:08:15 crc kubenswrapper[4846]: I1005 07:08:15.348590 4846 scope.go:117] "RemoveContainer" containerID="39111fc9d7fc57d402d3aff9ff623348099e37f99ec5b23eaa70a56e39c540b9" Oct 05 07:08:15 crc kubenswrapper[4846]: E1005 07:08:15.350531 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"39111fc9d7fc57d402d3aff9ff623348099e37f99ec5b23eaa70a56e39c540b9\": container with ID starting with 39111fc9d7fc57d402d3aff9ff623348099e37f99ec5b23eaa70a56e39c540b9 not found: ID does not exist" containerID="39111fc9d7fc57d402d3aff9ff623348099e37f99ec5b23eaa70a56e39c540b9" Oct 05 07:08:15 crc kubenswrapper[4846]: I1005 07:08:15.350564 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"39111fc9d7fc57d402d3aff9ff623348099e37f99ec5b23eaa70a56e39c540b9"} err="failed to get container status \"39111fc9d7fc57d402d3aff9ff623348099e37f99ec5b23eaa70a56e39c540b9\": rpc error: code = NotFound desc = could not find container \"39111fc9d7fc57d402d3aff9ff623348099e37f99ec5b23eaa70a56e39c540b9\": container with ID starting with 39111fc9d7fc57d402d3aff9ff623348099e37f99ec5b23eaa70a56e39c540b9 not found: ID does not exist" Oct 05 07:08:15 crc kubenswrapper[4846]: I1005 07:08:15.647060 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-984785977-srp2v"] Oct 05 07:08:15 crc kubenswrapper[4846]: W1005 07:08:15.647781 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd11b6df6_6da2_4cb8_a4f9_d87cfd5d4d26.slice/crio-ee3e69e0f367c67afe20a2d1f8600f785c5579c1f001d1de557eba1d560758fa WatchSource:0}: Error finding container ee3e69e0f367c67afe20a2d1f8600f785c5579c1f001d1de557eba1d560758fa: Status 404 returned error can't find the container with id ee3e69e0f367c67afe20a2d1f8600f785c5579c1f001d1de557eba1d560758fa Oct 05 07:08:16 crc kubenswrapper[4846]: I1005 07:08:16.291772 4846 generic.go:334] "Generic (PLEG): container finished" podID="39071bf4-214b-4bbb-8b78-dfec6ad3c940" containerID="82b673a1934d34adb80bf16feac56c7fc2158772b630ab377b61d9b3c475954b" exitCode=0 Oct 05 07:08:16 crc kubenswrapper[4846]: I1005 07:08:16.291849 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"39071bf4-214b-4bbb-8b78-dfec6ad3c940","Type":"ContainerDied","Data":"82b673a1934d34adb80bf16feac56c7fc2158772b630ab377b61d9b3c475954b"} Oct 05 07:08:16 crc kubenswrapper[4846]: I1005 07:08:16.295582 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-984785977-srp2v" event={"ID":"d11b6df6-6da2-4cb8-a4f9-d87cfd5d4d26","Type":"ContainerStarted","Data":"5b9cd4a1e7ff57757729dbae754fc2fd8f2c6d6deb6c33079a5fb777d6c337c4"} Oct 05 07:08:16 crc kubenswrapper[4846]: I1005 07:08:16.295617 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-984785977-srp2v" event={"ID":"d11b6df6-6da2-4cb8-a4f9-d87cfd5d4d26","Type":"ContainerStarted","Data":"eb3d496875131b7ff843f08ff12b7fdc832265502da5423e00550c759c4bc1b1"} Oct 05 07:08:16 crc kubenswrapper[4846]: I1005 07:08:16.295632 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-984785977-srp2v" event={"ID":"d11b6df6-6da2-4cb8-a4f9-d87cfd5d4d26","Type":"ContainerStarted","Data":"ee3e69e0f367c67afe20a2d1f8600f785c5579c1f001d1de557eba1d560758fa"} Oct 05 07:08:16 crc kubenswrapper[4846]: I1005 07:08:16.295904 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-984785977-srp2v" Oct 05 07:08:16 crc kubenswrapper[4846]: I1005 07:08:16.295931 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-984785977-srp2v" Oct 05 07:08:16 crc kubenswrapper[4846]: I1005 07:08:16.339792 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-proxy-984785977-srp2v" podStartSLOduration=2.339759655 podStartE2EDuration="2.339759655s" podCreationTimestamp="2025-10-05 07:08:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 07:08:16.322551226 +0000 UTC m=+1218.563404021" watchObservedRunningTime="2025-10-05 07:08:16.339759655 +0000 UTC m=+1218.580612430" Oct 05 07:08:16 crc kubenswrapper[4846]: I1005 07:08:16.519697 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ba9102c0-185f-4f02-8b05-8bb7fde2d573" path="/var/lib/kubelet/pods/ba9102c0-185f-4f02-8b05-8bb7fde2d573/volumes" Oct 05 07:08:17 crc kubenswrapper[4846]: I1005 07:08:17.793997 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Oct 05 07:08:18 crc kubenswrapper[4846]: I1005 07:08:18.323868 4846 generic.go:334] "Generic (PLEG): container finished" podID="39071bf4-214b-4bbb-8b78-dfec6ad3c940" containerID="15e05c1324747c5238b20b4131b788c1adc0d31ec16d39fc7037ec91c627a278" exitCode=0 Oct 05 07:08:18 crc kubenswrapper[4846]: I1005 07:08:18.323960 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"39071bf4-214b-4bbb-8b78-dfec6ad3c940","Type":"ContainerDied","Data":"15e05c1324747c5238b20b4131b788c1adc0d31ec16d39fc7037ec91c627a278"} Oct 05 07:08:19 crc kubenswrapper[4846]: I1005 07:08:19.047975 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-65884db4f4-fsk6p" Oct 05 07:08:19 crc kubenswrapper[4846]: I1005 07:08:19.049957 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-65884db4f4-fsk6p" Oct 05 07:08:20 crc kubenswrapper[4846]: I1005 07:08:20.624810 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 05 07:08:20 crc kubenswrapper[4846]: I1005 07:08:20.625694 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="edd6502e-6e77-4ad3-8f1f-b85426188261" containerName="glance-log" containerID="cri-o://8aa9405d7a8ddd77ef09fe71dd0b5f1c398759b0a55d1ac4fba47c2b46ef37cc" gracePeriod=30 Oct 05 07:08:20 crc kubenswrapper[4846]: I1005 07:08:20.625973 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="edd6502e-6e77-4ad3-8f1f-b85426188261" containerName="glance-httpd" containerID="cri-o://fe4a5d7cbeaf3a5245813e826bd7235fb2d424df813e58a0c069e9a6f5b444ca" gracePeriod=30 Oct 05 07:08:20 crc kubenswrapper[4846]: I1005 07:08:20.841066 4846 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="39071bf4-214b-4bbb-8b78-dfec6ad3c940" containerName="proxy-httpd" probeResult="failure" output="Get \"http://10.217.0.160:3000/\": dial tcp 10.217.0.160:3000: connect: connection refused" Oct 05 07:08:21 crc kubenswrapper[4846]: I1005 07:08:21.427727 4846 generic.go:334] "Generic (PLEG): container finished" podID="edd6502e-6e77-4ad3-8f1f-b85426188261" containerID="8aa9405d7a8ddd77ef09fe71dd0b5f1c398759b0a55d1ac4fba47c2b46ef37cc" exitCode=143 Oct 05 07:08:21 crc kubenswrapper[4846]: I1005 07:08:21.427904 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"edd6502e-6e77-4ad3-8f1f-b85426188261","Type":"ContainerDied","Data":"8aa9405d7a8ddd77ef09fe71dd0b5f1c398759b0a55d1ac4fba47c2b46ef37cc"} Oct 05 07:08:24 crc kubenswrapper[4846]: I1005 07:08:24.482104 4846 generic.go:334] "Generic (PLEG): container finished" podID="edd6502e-6e77-4ad3-8f1f-b85426188261" containerID="fe4a5d7cbeaf3a5245813e826bd7235fb2d424df813e58a0c069e9a6f5b444ca" exitCode=0 Oct 05 07:08:24 crc kubenswrapper[4846]: I1005 07:08:24.482232 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"edd6502e-6e77-4ad3-8f1f-b85426188261","Type":"ContainerDied","Data":"fe4a5d7cbeaf3a5245813e826bd7235fb2d424df813e58a0c069e9a6f5b444ca"} Oct 05 07:08:24 crc kubenswrapper[4846]: I1005 07:08:24.803763 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 05 07:08:24 crc kubenswrapper[4846]: I1005 07:08:24.945556 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-984785977-srp2v" Oct 05 07:08:24 crc kubenswrapper[4846]: I1005 07:08:24.948044 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-984785977-srp2v" Oct 05 07:08:25 crc kubenswrapper[4846]: I1005 07:08:25.006282 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/39071bf4-214b-4bbb-8b78-dfec6ad3c940-run-httpd\") pod \"39071bf4-214b-4bbb-8b78-dfec6ad3c940\" (UID: \"39071bf4-214b-4bbb-8b78-dfec6ad3c940\") " Oct 05 07:08:25 crc kubenswrapper[4846]: I1005 07:08:25.006399 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/39071bf4-214b-4bbb-8b78-dfec6ad3c940-log-httpd\") pod \"39071bf4-214b-4bbb-8b78-dfec6ad3c940\" (UID: \"39071bf4-214b-4bbb-8b78-dfec6ad3c940\") " Oct 05 07:08:25 crc kubenswrapper[4846]: I1005 07:08:25.006439 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/39071bf4-214b-4bbb-8b78-dfec6ad3c940-scripts\") pod \"39071bf4-214b-4bbb-8b78-dfec6ad3c940\" (UID: \"39071bf4-214b-4bbb-8b78-dfec6ad3c940\") " Oct 05 07:08:25 crc kubenswrapper[4846]: I1005 07:08:25.006475 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/39071bf4-214b-4bbb-8b78-dfec6ad3c940-combined-ca-bundle\") pod \"39071bf4-214b-4bbb-8b78-dfec6ad3c940\" (UID: \"39071bf4-214b-4bbb-8b78-dfec6ad3c940\") " Oct 05 07:08:25 crc kubenswrapper[4846]: I1005 07:08:25.006526 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/39071bf4-214b-4bbb-8b78-dfec6ad3c940-sg-core-conf-yaml\") pod \"39071bf4-214b-4bbb-8b78-dfec6ad3c940\" (UID: \"39071bf4-214b-4bbb-8b78-dfec6ad3c940\") " Oct 05 07:08:25 crc kubenswrapper[4846]: I1005 07:08:25.006585 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bcr96\" (UniqueName: \"kubernetes.io/projected/39071bf4-214b-4bbb-8b78-dfec6ad3c940-kube-api-access-bcr96\") pod \"39071bf4-214b-4bbb-8b78-dfec6ad3c940\" (UID: \"39071bf4-214b-4bbb-8b78-dfec6ad3c940\") " Oct 05 07:08:25 crc kubenswrapper[4846]: I1005 07:08:25.006625 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/39071bf4-214b-4bbb-8b78-dfec6ad3c940-config-data\") pod \"39071bf4-214b-4bbb-8b78-dfec6ad3c940\" (UID: \"39071bf4-214b-4bbb-8b78-dfec6ad3c940\") " Oct 05 07:08:25 crc kubenswrapper[4846]: I1005 07:08:25.006894 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 05 07:08:25 crc kubenswrapper[4846]: I1005 07:08:25.007497 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/39071bf4-214b-4bbb-8b78-dfec6ad3c940-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "39071bf4-214b-4bbb-8b78-dfec6ad3c940" (UID: "39071bf4-214b-4bbb-8b78-dfec6ad3c940"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 07:08:25 crc kubenswrapper[4846]: I1005 07:08:25.007823 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/39071bf4-214b-4bbb-8b78-dfec6ad3c940-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "39071bf4-214b-4bbb-8b78-dfec6ad3c940" (UID: "39071bf4-214b-4bbb-8b78-dfec6ad3c940"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 07:08:25 crc kubenswrapper[4846]: I1005 07:08:25.017336 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/39071bf4-214b-4bbb-8b78-dfec6ad3c940-scripts" (OuterVolumeSpecName: "scripts") pod "39071bf4-214b-4bbb-8b78-dfec6ad3c940" (UID: "39071bf4-214b-4bbb-8b78-dfec6ad3c940"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:08:25 crc kubenswrapper[4846]: I1005 07:08:25.022714 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/39071bf4-214b-4bbb-8b78-dfec6ad3c940-kube-api-access-bcr96" (OuterVolumeSpecName: "kube-api-access-bcr96") pod "39071bf4-214b-4bbb-8b78-dfec6ad3c940" (UID: "39071bf4-214b-4bbb-8b78-dfec6ad3c940"). InnerVolumeSpecName "kube-api-access-bcr96". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:08:25 crc kubenswrapper[4846]: I1005 07:08:25.059519 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/39071bf4-214b-4bbb-8b78-dfec6ad3c940-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "39071bf4-214b-4bbb-8b78-dfec6ad3c940" (UID: "39071bf4-214b-4bbb-8b78-dfec6ad3c940"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:08:25 crc kubenswrapper[4846]: I1005 07:08:25.108113 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/edd6502e-6e77-4ad3-8f1f-b85426188261-httpd-run\") pod \"edd6502e-6e77-4ad3-8f1f-b85426188261\" (UID: \"edd6502e-6e77-4ad3-8f1f-b85426188261\") " Oct 05 07:08:25 crc kubenswrapper[4846]: I1005 07:08:25.108174 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/edd6502e-6e77-4ad3-8f1f-b85426188261-config-data\") pod \"edd6502e-6e77-4ad3-8f1f-b85426188261\" (UID: \"edd6502e-6e77-4ad3-8f1f-b85426188261\") " Oct 05 07:08:25 crc kubenswrapper[4846]: I1005 07:08:25.108410 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xgjg\" (UniqueName: \"kubernetes.io/projected/edd6502e-6e77-4ad3-8f1f-b85426188261-kube-api-access-9xgjg\") pod \"edd6502e-6e77-4ad3-8f1f-b85426188261\" (UID: \"edd6502e-6e77-4ad3-8f1f-b85426188261\") " Oct 05 07:08:25 crc kubenswrapper[4846]: I1005 07:08:25.108444 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"edd6502e-6e77-4ad3-8f1f-b85426188261\" (UID: \"edd6502e-6e77-4ad3-8f1f-b85426188261\") " Oct 05 07:08:25 crc kubenswrapper[4846]: I1005 07:08:25.108467 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/edd6502e-6e77-4ad3-8f1f-b85426188261-logs\") pod \"edd6502e-6e77-4ad3-8f1f-b85426188261\" (UID: \"edd6502e-6e77-4ad3-8f1f-b85426188261\") " Oct 05 07:08:25 crc kubenswrapper[4846]: I1005 07:08:25.108495 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/edd6502e-6e77-4ad3-8f1f-b85426188261-internal-tls-certs\") pod \"edd6502e-6e77-4ad3-8f1f-b85426188261\" (UID: \"edd6502e-6e77-4ad3-8f1f-b85426188261\") " Oct 05 07:08:25 crc kubenswrapper[4846]: I1005 07:08:25.108511 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/edd6502e-6e77-4ad3-8f1f-b85426188261-combined-ca-bundle\") pod \"edd6502e-6e77-4ad3-8f1f-b85426188261\" (UID: \"edd6502e-6e77-4ad3-8f1f-b85426188261\") " Oct 05 07:08:25 crc kubenswrapper[4846]: I1005 07:08:25.108613 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/edd6502e-6e77-4ad3-8f1f-b85426188261-scripts\") pod \"edd6502e-6e77-4ad3-8f1f-b85426188261\" (UID: \"edd6502e-6e77-4ad3-8f1f-b85426188261\") " Oct 05 07:08:25 crc kubenswrapper[4846]: I1005 07:08:25.108782 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/edd6502e-6e77-4ad3-8f1f-b85426188261-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "edd6502e-6e77-4ad3-8f1f-b85426188261" (UID: "edd6502e-6e77-4ad3-8f1f-b85426188261"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 07:08:25 crc kubenswrapper[4846]: I1005 07:08:25.110916 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/edd6502e-6e77-4ad3-8f1f-b85426188261-logs" (OuterVolumeSpecName: "logs") pod "edd6502e-6e77-4ad3-8f1f-b85426188261" (UID: "edd6502e-6e77-4ad3-8f1f-b85426188261"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 07:08:25 crc kubenswrapper[4846]: I1005 07:08:25.112377 4846 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/39071bf4-214b-4bbb-8b78-dfec6ad3c940-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 05 07:08:25 crc kubenswrapper[4846]: I1005 07:08:25.112406 4846 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/edd6502e-6e77-4ad3-8f1f-b85426188261-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 05 07:08:25 crc kubenswrapper[4846]: I1005 07:08:25.112419 4846 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/39071bf4-214b-4bbb-8b78-dfec6ad3c940-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 05 07:08:25 crc kubenswrapper[4846]: I1005 07:08:25.112430 4846 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/39071bf4-214b-4bbb-8b78-dfec6ad3c940-scripts\") on node \"crc\" DevicePath \"\"" Oct 05 07:08:25 crc kubenswrapper[4846]: I1005 07:08:25.112440 4846 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/39071bf4-214b-4bbb-8b78-dfec6ad3c940-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 05 07:08:25 crc kubenswrapper[4846]: I1005 07:08:25.112457 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bcr96\" (UniqueName: \"kubernetes.io/projected/39071bf4-214b-4bbb-8b78-dfec6ad3c940-kube-api-access-bcr96\") on node \"crc\" DevicePath \"\"" Oct 05 07:08:25 crc kubenswrapper[4846]: I1005 07:08:25.114372 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/edd6502e-6e77-4ad3-8f1f-b85426188261-scripts" (OuterVolumeSpecName: "scripts") pod "edd6502e-6e77-4ad3-8f1f-b85426188261" (UID: "edd6502e-6e77-4ad3-8f1f-b85426188261"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:08:25 crc kubenswrapper[4846]: I1005 07:08:25.114906 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage12-crc" (OuterVolumeSpecName: "glance") pod "edd6502e-6e77-4ad3-8f1f-b85426188261" (UID: "edd6502e-6e77-4ad3-8f1f-b85426188261"). InnerVolumeSpecName "local-storage12-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 05 07:08:25 crc kubenswrapper[4846]: I1005 07:08:25.116752 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/edd6502e-6e77-4ad3-8f1f-b85426188261-kube-api-access-9xgjg" (OuterVolumeSpecName: "kube-api-access-9xgjg") pod "edd6502e-6e77-4ad3-8f1f-b85426188261" (UID: "edd6502e-6e77-4ad3-8f1f-b85426188261"). InnerVolumeSpecName "kube-api-access-9xgjg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:08:25 crc kubenswrapper[4846]: I1005 07:08:25.158538 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/edd6502e-6e77-4ad3-8f1f-b85426188261-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "edd6502e-6e77-4ad3-8f1f-b85426188261" (UID: "edd6502e-6e77-4ad3-8f1f-b85426188261"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:08:25 crc kubenswrapper[4846]: I1005 07:08:25.161861 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/39071bf4-214b-4bbb-8b78-dfec6ad3c940-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "39071bf4-214b-4bbb-8b78-dfec6ad3c940" (UID: "39071bf4-214b-4bbb-8b78-dfec6ad3c940"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:08:25 crc kubenswrapper[4846]: I1005 07:08:25.193553 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/edd6502e-6e77-4ad3-8f1f-b85426188261-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "edd6502e-6e77-4ad3-8f1f-b85426188261" (UID: "edd6502e-6e77-4ad3-8f1f-b85426188261"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:08:25 crc kubenswrapper[4846]: I1005 07:08:25.199893 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/39071bf4-214b-4bbb-8b78-dfec6ad3c940-config-data" (OuterVolumeSpecName: "config-data") pod "39071bf4-214b-4bbb-8b78-dfec6ad3c940" (UID: "39071bf4-214b-4bbb-8b78-dfec6ad3c940"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:08:25 crc kubenswrapper[4846]: I1005 07:08:25.200690 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/edd6502e-6e77-4ad3-8f1f-b85426188261-config-data" (OuterVolumeSpecName: "config-data") pod "edd6502e-6e77-4ad3-8f1f-b85426188261" (UID: "edd6502e-6e77-4ad3-8f1f-b85426188261"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:08:25 crc kubenswrapper[4846]: I1005 07:08:25.213469 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xgjg\" (UniqueName: \"kubernetes.io/projected/edd6502e-6e77-4ad3-8f1f-b85426188261-kube-api-access-9xgjg\") on node \"crc\" DevicePath \"\"" Oct 05 07:08:25 crc kubenswrapper[4846]: I1005 07:08:25.213528 4846 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" " Oct 05 07:08:25 crc kubenswrapper[4846]: I1005 07:08:25.213542 4846 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/edd6502e-6e77-4ad3-8f1f-b85426188261-logs\") on node \"crc\" DevicePath \"\"" Oct 05 07:08:25 crc kubenswrapper[4846]: I1005 07:08:25.213555 4846 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/edd6502e-6e77-4ad3-8f1f-b85426188261-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 05 07:08:25 crc kubenswrapper[4846]: I1005 07:08:25.213565 4846 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/edd6502e-6e77-4ad3-8f1f-b85426188261-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 07:08:25 crc kubenswrapper[4846]: I1005 07:08:25.213576 4846 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/39071bf4-214b-4bbb-8b78-dfec6ad3c940-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 07:08:25 crc kubenswrapper[4846]: I1005 07:08:25.213585 4846 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/edd6502e-6e77-4ad3-8f1f-b85426188261-scripts\") on node \"crc\" DevicePath \"\"" Oct 05 07:08:25 crc kubenswrapper[4846]: I1005 07:08:25.213593 4846 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/edd6502e-6e77-4ad3-8f1f-b85426188261-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 07:08:25 crc kubenswrapper[4846]: I1005 07:08:25.213601 4846 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/39071bf4-214b-4bbb-8b78-dfec6ad3c940-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 07:08:25 crc kubenswrapper[4846]: I1005 07:08:25.247486 4846 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage12-crc" (UniqueName: "kubernetes.io/local-volume/local-storage12-crc") on node "crc" Oct 05 07:08:25 crc kubenswrapper[4846]: I1005 07:08:25.316305 4846 reconciler_common.go:293] "Volume detached for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" DevicePath \"\"" Oct 05 07:08:25 crc kubenswrapper[4846]: I1005 07:08:25.497444 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"39071bf4-214b-4bbb-8b78-dfec6ad3c940","Type":"ContainerDied","Data":"08c39de0b53585ece279a655d42b2c0a11038c4cd766ef12f56aa73eb55703ba"} Oct 05 07:08:25 crc kubenswrapper[4846]: I1005 07:08:25.497515 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 05 07:08:25 crc kubenswrapper[4846]: I1005 07:08:25.497522 4846 scope.go:117] "RemoveContainer" containerID="72d54baff0ab5bf146fb7a95307526bd069594408c4da5a11bbeacf29306be6c" Oct 05 07:08:25 crc kubenswrapper[4846]: I1005 07:08:25.507790 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"98fc38b8-a0ee-46c9-9d3f-9c15e8355b9d","Type":"ContainerStarted","Data":"5edadd35f5bb4b18c0ce1cfc15ed5663752e987563a9e6ef8438aa48f0d79ae2"} Oct 05 07:08:25 crc kubenswrapper[4846]: I1005 07:08:25.511832 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"edd6502e-6e77-4ad3-8f1f-b85426188261","Type":"ContainerDied","Data":"0d610f817a3051e49134c8aafcdac4fd03f563999d934bd1a6787a8227d0042c"} Oct 05 07:08:25 crc kubenswrapper[4846]: I1005 07:08:25.512078 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 05 07:08:25 crc kubenswrapper[4846]: I1005 07:08:25.528671 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=2.727646763 podStartE2EDuration="16.528648507s" podCreationTimestamp="2025-10-05 07:08:09 +0000 UTC" firstStartedPulling="2025-10-05 07:08:10.733364602 +0000 UTC m=+1212.974217377" lastFinishedPulling="2025-10-05 07:08:24.534366336 +0000 UTC m=+1226.775219121" observedRunningTime="2025-10-05 07:08:25.524667771 +0000 UTC m=+1227.765520546" watchObservedRunningTime="2025-10-05 07:08:25.528648507 +0000 UTC m=+1227.769501282" Oct 05 07:08:25 crc kubenswrapper[4846]: I1005 07:08:25.536075 4846 scope.go:117] "RemoveContainer" containerID="82e6c4bfad41089daffabb46ada2064751a4f47c40780fda32f8ec7989bd6a30" Oct 05 07:08:25 crc kubenswrapper[4846]: I1005 07:08:25.572054 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 05 07:08:25 crc kubenswrapper[4846]: I1005 07:08:25.575210 4846 scope.go:117] "RemoveContainer" containerID="15e05c1324747c5238b20b4131b788c1adc0d31ec16d39fc7037ec91c627a278" Oct 05 07:08:25 crc kubenswrapper[4846]: I1005 07:08:25.581671 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 05 07:08:25 crc kubenswrapper[4846]: I1005 07:08:25.598391 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 05 07:08:25 crc kubenswrapper[4846]: I1005 07:08:25.606986 4846 scope.go:117] "RemoveContainer" containerID="82b673a1934d34adb80bf16feac56c7fc2158772b630ab377b61d9b3c475954b" Oct 05 07:08:25 crc kubenswrapper[4846]: I1005 07:08:25.616538 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 05 07:08:25 crc kubenswrapper[4846]: I1005 07:08:25.635099 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 05 07:08:25 crc kubenswrapper[4846]: E1005 07:08:25.635710 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ba9102c0-185f-4f02-8b05-8bb7fde2d573" containerName="barbican-api-log" Oct 05 07:08:25 crc kubenswrapper[4846]: I1005 07:08:25.635732 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="ba9102c0-185f-4f02-8b05-8bb7fde2d573" containerName="barbican-api-log" Oct 05 07:08:25 crc kubenswrapper[4846]: E1005 07:08:25.635754 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="39071bf4-214b-4bbb-8b78-dfec6ad3c940" containerName="ceilometer-central-agent" Oct 05 07:08:25 crc kubenswrapper[4846]: I1005 07:08:25.635761 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="39071bf4-214b-4bbb-8b78-dfec6ad3c940" containerName="ceilometer-central-agent" Oct 05 07:08:25 crc kubenswrapper[4846]: E1005 07:08:25.635781 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="39071bf4-214b-4bbb-8b78-dfec6ad3c940" containerName="sg-core" Oct 05 07:08:25 crc kubenswrapper[4846]: I1005 07:08:25.635787 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="39071bf4-214b-4bbb-8b78-dfec6ad3c940" containerName="sg-core" Oct 05 07:08:25 crc kubenswrapper[4846]: E1005 07:08:25.635800 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="edd6502e-6e77-4ad3-8f1f-b85426188261" containerName="glance-httpd" Oct 05 07:08:25 crc kubenswrapper[4846]: I1005 07:08:25.635809 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="edd6502e-6e77-4ad3-8f1f-b85426188261" containerName="glance-httpd" Oct 05 07:08:25 crc kubenswrapper[4846]: E1005 07:08:25.635827 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ba9102c0-185f-4f02-8b05-8bb7fde2d573" containerName="barbican-api" Oct 05 07:08:25 crc kubenswrapper[4846]: I1005 07:08:25.635834 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="ba9102c0-185f-4f02-8b05-8bb7fde2d573" containerName="barbican-api" Oct 05 07:08:25 crc kubenswrapper[4846]: E1005 07:08:25.635859 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="39071bf4-214b-4bbb-8b78-dfec6ad3c940" containerName="proxy-httpd" Oct 05 07:08:25 crc kubenswrapper[4846]: I1005 07:08:25.635866 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="39071bf4-214b-4bbb-8b78-dfec6ad3c940" containerName="proxy-httpd" Oct 05 07:08:25 crc kubenswrapper[4846]: E1005 07:08:25.635881 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="39071bf4-214b-4bbb-8b78-dfec6ad3c940" containerName="ceilometer-notification-agent" Oct 05 07:08:25 crc kubenswrapper[4846]: I1005 07:08:25.635888 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="39071bf4-214b-4bbb-8b78-dfec6ad3c940" containerName="ceilometer-notification-agent" Oct 05 07:08:25 crc kubenswrapper[4846]: E1005 07:08:25.635902 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="edd6502e-6e77-4ad3-8f1f-b85426188261" containerName="glance-log" Oct 05 07:08:25 crc kubenswrapper[4846]: I1005 07:08:25.635908 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="edd6502e-6e77-4ad3-8f1f-b85426188261" containerName="glance-log" Oct 05 07:08:25 crc kubenswrapper[4846]: I1005 07:08:25.636102 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="39071bf4-214b-4bbb-8b78-dfec6ad3c940" containerName="ceilometer-notification-agent" Oct 05 07:08:25 crc kubenswrapper[4846]: I1005 07:08:25.636119 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="edd6502e-6e77-4ad3-8f1f-b85426188261" containerName="glance-log" Oct 05 07:08:25 crc kubenswrapper[4846]: I1005 07:08:25.636133 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="39071bf4-214b-4bbb-8b78-dfec6ad3c940" containerName="proxy-httpd" Oct 05 07:08:25 crc kubenswrapper[4846]: I1005 07:08:25.636149 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="39071bf4-214b-4bbb-8b78-dfec6ad3c940" containerName="ceilometer-central-agent" Oct 05 07:08:25 crc kubenswrapper[4846]: I1005 07:08:25.636161 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="39071bf4-214b-4bbb-8b78-dfec6ad3c940" containerName="sg-core" Oct 05 07:08:25 crc kubenswrapper[4846]: I1005 07:08:25.636171 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="ba9102c0-185f-4f02-8b05-8bb7fde2d573" containerName="barbican-api" Oct 05 07:08:25 crc kubenswrapper[4846]: I1005 07:08:25.636194 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="edd6502e-6e77-4ad3-8f1f-b85426188261" containerName="glance-httpd" Oct 05 07:08:25 crc kubenswrapper[4846]: I1005 07:08:25.636206 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="ba9102c0-185f-4f02-8b05-8bb7fde2d573" containerName="barbican-api-log" Oct 05 07:08:25 crc kubenswrapper[4846]: I1005 07:08:25.638119 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 05 07:08:25 crc kubenswrapper[4846]: I1005 07:08:25.639264 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 05 07:08:25 crc kubenswrapper[4846]: I1005 07:08:25.640620 4846 scope.go:117] "RemoveContainer" containerID="fe4a5d7cbeaf3a5245813e826bd7235fb2d424df813e58a0c069e9a6f5b444ca" Oct 05 07:08:25 crc kubenswrapper[4846]: I1005 07:08:25.643833 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 05 07:08:25 crc kubenswrapper[4846]: I1005 07:08:25.644780 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 05 07:08:25 crc kubenswrapper[4846]: I1005 07:08:25.646372 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 05 07:08:25 crc kubenswrapper[4846]: I1005 07:08:25.652531 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 05 07:08:25 crc kubenswrapper[4846]: I1005 07:08:25.653391 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 05 07:08:25 crc kubenswrapper[4846]: I1005 07:08:25.654949 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Oct 05 07:08:25 crc kubenswrapper[4846]: I1005 07:08:25.655254 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Oct 05 07:08:25 crc kubenswrapper[4846]: I1005 07:08:25.685407 4846 scope.go:117] "RemoveContainer" containerID="8aa9405d7a8ddd77ef09fe71dd0b5f1c398759b0a55d1ac4fba47c2b46ef37cc" Oct 05 07:08:25 crc kubenswrapper[4846]: I1005 07:08:25.829029 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"2adc8ba0-2810-4bf4-bd24-172b97fb9b22\") " pod="openstack/glance-default-internal-api-0" Oct 05 07:08:25 crc kubenswrapper[4846]: I1005 07:08:25.829096 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2adc8ba0-2810-4bf4-bd24-172b97fb9b22-scripts\") pod \"glance-default-internal-api-0\" (UID: \"2adc8ba0-2810-4bf4-bd24-172b97fb9b22\") " pod="openstack/glance-default-internal-api-0" Oct 05 07:08:25 crc kubenswrapper[4846]: I1005 07:08:25.829120 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/2adc8ba0-2810-4bf4-bd24-172b97fb9b22-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"2adc8ba0-2810-4bf4-bd24-172b97fb9b22\") " pod="openstack/glance-default-internal-api-0" Oct 05 07:08:25 crc kubenswrapper[4846]: I1005 07:08:25.829267 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2adc8ba0-2810-4bf4-bd24-172b97fb9b22-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"2adc8ba0-2810-4bf4-bd24-172b97fb9b22\") " pod="openstack/glance-default-internal-api-0" Oct 05 07:08:25 crc kubenswrapper[4846]: I1005 07:08:25.829295 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/671f4c86-d748-474a-b0f0-400b16b28d7f-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"671f4c86-d748-474a-b0f0-400b16b28d7f\") " pod="openstack/ceilometer-0" Oct 05 07:08:25 crc kubenswrapper[4846]: I1005 07:08:25.829484 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qhrh2\" (UniqueName: \"kubernetes.io/projected/671f4c86-d748-474a-b0f0-400b16b28d7f-kube-api-access-qhrh2\") pod \"ceilometer-0\" (UID: \"671f4c86-d748-474a-b0f0-400b16b28d7f\") " pod="openstack/ceilometer-0" Oct 05 07:08:25 crc kubenswrapper[4846]: I1005 07:08:25.829599 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2adc8ba0-2810-4bf4-bd24-172b97fb9b22-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"2adc8ba0-2810-4bf4-bd24-172b97fb9b22\") " pod="openstack/glance-default-internal-api-0" Oct 05 07:08:25 crc kubenswrapper[4846]: I1005 07:08:25.829835 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6dxzx\" (UniqueName: \"kubernetes.io/projected/2adc8ba0-2810-4bf4-bd24-172b97fb9b22-kube-api-access-6dxzx\") pod \"glance-default-internal-api-0\" (UID: \"2adc8ba0-2810-4bf4-bd24-172b97fb9b22\") " pod="openstack/glance-default-internal-api-0" Oct 05 07:08:25 crc kubenswrapper[4846]: I1005 07:08:25.829988 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2adc8ba0-2810-4bf4-bd24-172b97fb9b22-config-data\") pod \"glance-default-internal-api-0\" (UID: \"2adc8ba0-2810-4bf4-bd24-172b97fb9b22\") " pod="openstack/glance-default-internal-api-0" Oct 05 07:08:25 crc kubenswrapper[4846]: I1005 07:08:25.830051 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2adc8ba0-2810-4bf4-bd24-172b97fb9b22-logs\") pod \"glance-default-internal-api-0\" (UID: \"2adc8ba0-2810-4bf4-bd24-172b97fb9b22\") " pod="openstack/glance-default-internal-api-0" Oct 05 07:08:25 crc kubenswrapper[4846]: I1005 07:08:25.830190 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/671f4c86-d748-474a-b0f0-400b16b28d7f-run-httpd\") pod \"ceilometer-0\" (UID: \"671f4c86-d748-474a-b0f0-400b16b28d7f\") " pod="openstack/ceilometer-0" Oct 05 07:08:25 crc kubenswrapper[4846]: I1005 07:08:25.830243 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/671f4c86-d748-474a-b0f0-400b16b28d7f-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"671f4c86-d748-474a-b0f0-400b16b28d7f\") " pod="openstack/ceilometer-0" Oct 05 07:08:25 crc kubenswrapper[4846]: I1005 07:08:25.830443 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/671f4c86-d748-474a-b0f0-400b16b28d7f-config-data\") pod \"ceilometer-0\" (UID: \"671f4c86-d748-474a-b0f0-400b16b28d7f\") " pod="openstack/ceilometer-0" Oct 05 07:08:25 crc kubenswrapper[4846]: I1005 07:08:25.830508 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/671f4c86-d748-474a-b0f0-400b16b28d7f-scripts\") pod \"ceilometer-0\" (UID: \"671f4c86-d748-474a-b0f0-400b16b28d7f\") " pod="openstack/ceilometer-0" Oct 05 07:08:25 crc kubenswrapper[4846]: I1005 07:08:25.830571 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/671f4c86-d748-474a-b0f0-400b16b28d7f-log-httpd\") pod \"ceilometer-0\" (UID: \"671f4c86-d748-474a-b0f0-400b16b28d7f\") " pod="openstack/ceilometer-0" Oct 05 07:08:25 crc kubenswrapper[4846]: I1005 07:08:25.932986 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2adc8ba0-2810-4bf4-bd24-172b97fb9b22-scripts\") pod \"glance-default-internal-api-0\" (UID: \"2adc8ba0-2810-4bf4-bd24-172b97fb9b22\") " pod="openstack/glance-default-internal-api-0" Oct 05 07:08:25 crc kubenswrapper[4846]: I1005 07:08:25.933053 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/2adc8ba0-2810-4bf4-bd24-172b97fb9b22-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"2adc8ba0-2810-4bf4-bd24-172b97fb9b22\") " pod="openstack/glance-default-internal-api-0" Oct 05 07:08:25 crc kubenswrapper[4846]: I1005 07:08:25.933147 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2adc8ba0-2810-4bf4-bd24-172b97fb9b22-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"2adc8ba0-2810-4bf4-bd24-172b97fb9b22\") " pod="openstack/glance-default-internal-api-0" Oct 05 07:08:25 crc kubenswrapper[4846]: I1005 07:08:25.933206 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/671f4c86-d748-474a-b0f0-400b16b28d7f-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"671f4c86-d748-474a-b0f0-400b16b28d7f\") " pod="openstack/ceilometer-0" Oct 05 07:08:25 crc kubenswrapper[4846]: I1005 07:08:25.933256 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qhrh2\" (UniqueName: \"kubernetes.io/projected/671f4c86-d748-474a-b0f0-400b16b28d7f-kube-api-access-qhrh2\") pod \"ceilometer-0\" (UID: \"671f4c86-d748-474a-b0f0-400b16b28d7f\") " pod="openstack/ceilometer-0" Oct 05 07:08:25 crc kubenswrapper[4846]: I1005 07:08:25.933292 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2adc8ba0-2810-4bf4-bd24-172b97fb9b22-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"2adc8ba0-2810-4bf4-bd24-172b97fb9b22\") " pod="openstack/glance-default-internal-api-0" Oct 05 07:08:25 crc kubenswrapper[4846]: I1005 07:08:25.933358 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6dxzx\" (UniqueName: \"kubernetes.io/projected/2adc8ba0-2810-4bf4-bd24-172b97fb9b22-kube-api-access-6dxzx\") pod \"glance-default-internal-api-0\" (UID: \"2adc8ba0-2810-4bf4-bd24-172b97fb9b22\") " pod="openstack/glance-default-internal-api-0" Oct 05 07:08:25 crc kubenswrapper[4846]: I1005 07:08:25.933407 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2adc8ba0-2810-4bf4-bd24-172b97fb9b22-config-data\") pod \"glance-default-internal-api-0\" (UID: \"2adc8ba0-2810-4bf4-bd24-172b97fb9b22\") " pod="openstack/glance-default-internal-api-0" Oct 05 07:08:25 crc kubenswrapper[4846]: I1005 07:08:25.933442 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2adc8ba0-2810-4bf4-bd24-172b97fb9b22-logs\") pod \"glance-default-internal-api-0\" (UID: \"2adc8ba0-2810-4bf4-bd24-172b97fb9b22\") " pod="openstack/glance-default-internal-api-0" Oct 05 07:08:25 crc kubenswrapper[4846]: I1005 07:08:25.933487 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/671f4c86-d748-474a-b0f0-400b16b28d7f-run-httpd\") pod \"ceilometer-0\" (UID: \"671f4c86-d748-474a-b0f0-400b16b28d7f\") " pod="openstack/ceilometer-0" Oct 05 07:08:25 crc kubenswrapper[4846]: I1005 07:08:25.933518 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/671f4c86-d748-474a-b0f0-400b16b28d7f-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"671f4c86-d748-474a-b0f0-400b16b28d7f\") " pod="openstack/ceilometer-0" Oct 05 07:08:25 crc kubenswrapper[4846]: I1005 07:08:25.933555 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/671f4c86-d748-474a-b0f0-400b16b28d7f-config-data\") pod \"ceilometer-0\" (UID: \"671f4c86-d748-474a-b0f0-400b16b28d7f\") " pod="openstack/ceilometer-0" Oct 05 07:08:25 crc kubenswrapper[4846]: I1005 07:08:25.933587 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/671f4c86-d748-474a-b0f0-400b16b28d7f-scripts\") pod \"ceilometer-0\" (UID: \"671f4c86-d748-474a-b0f0-400b16b28d7f\") " pod="openstack/ceilometer-0" Oct 05 07:08:25 crc kubenswrapper[4846]: I1005 07:08:25.933618 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/671f4c86-d748-474a-b0f0-400b16b28d7f-log-httpd\") pod \"ceilometer-0\" (UID: \"671f4c86-d748-474a-b0f0-400b16b28d7f\") " pod="openstack/ceilometer-0" Oct 05 07:08:25 crc kubenswrapper[4846]: I1005 07:08:25.933651 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"2adc8ba0-2810-4bf4-bd24-172b97fb9b22\") " pod="openstack/glance-default-internal-api-0" Oct 05 07:08:25 crc kubenswrapper[4846]: I1005 07:08:25.933824 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/2adc8ba0-2810-4bf4-bd24-172b97fb9b22-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"2adc8ba0-2810-4bf4-bd24-172b97fb9b22\") " pod="openstack/glance-default-internal-api-0" Oct 05 07:08:25 crc kubenswrapper[4846]: I1005 07:08:25.933995 4846 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"2adc8ba0-2810-4bf4-bd24-172b97fb9b22\") device mount path \"/mnt/openstack/pv12\"" pod="openstack/glance-default-internal-api-0" Oct 05 07:08:25 crc kubenswrapper[4846]: I1005 07:08:25.934266 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2adc8ba0-2810-4bf4-bd24-172b97fb9b22-logs\") pod \"glance-default-internal-api-0\" (UID: \"2adc8ba0-2810-4bf4-bd24-172b97fb9b22\") " pod="openstack/glance-default-internal-api-0" Oct 05 07:08:25 crc kubenswrapper[4846]: I1005 07:08:25.934281 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/671f4c86-d748-474a-b0f0-400b16b28d7f-run-httpd\") pod \"ceilometer-0\" (UID: \"671f4c86-d748-474a-b0f0-400b16b28d7f\") " pod="openstack/ceilometer-0" Oct 05 07:08:25 crc kubenswrapper[4846]: I1005 07:08:25.934316 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/671f4c86-d748-474a-b0f0-400b16b28d7f-log-httpd\") pod \"ceilometer-0\" (UID: \"671f4c86-d748-474a-b0f0-400b16b28d7f\") " pod="openstack/ceilometer-0" Oct 05 07:08:25 crc kubenswrapper[4846]: I1005 07:08:25.940228 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/671f4c86-d748-474a-b0f0-400b16b28d7f-scripts\") pod \"ceilometer-0\" (UID: \"671f4c86-d748-474a-b0f0-400b16b28d7f\") " pod="openstack/ceilometer-0" Oct 05 07:08:25 crc kubenswrapper[4846]: I1005 07:08:25.942459 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/671f4c86-d748-474a-b0f0-400b16b28d7f-config-data\") pod \"ceilometer-0\" (UID: \"671f4c86-d748-474a-b0f0-400b16b28d7f\") " pod="openstack/ceilometer-0" Oct 05 07:08:25 crc kubenswrapper[4846]: I1005 07:08:25.944973 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2adc8ba0-2810-4bf4-bd24-172b97fb9b22-scripts\") pod \"glance-default-internal-api-0\" (UID: \"2adc8ba0-2810-4bf4-bd24-172b97fb9b22\") " pod="openstack/glance-default-internal-api-0" Oct 05 07:08:25 crc kubenswrapper[4846]: I1005 07:08:25.945399 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/671f4c86-d748-474a-b0f0-400b16b28d7f-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"671f4c86-d748-474a-b0f0-400b16b28d7f\") " pod="openstack/ceilometer-0" Oct 05 07:08:25 crc kubenswrapper[4846]: I1005 07:08:25.945520 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2adc8ba0-2810-4bf4-bd24-172b97fb9b22-config-data\") pod \"glance-default-internal-api-0\" (UID: \"2adc8ba0-2810-4bf4-bd24-172b97fb9b22\") " pod="openstack/glance-default-internal-api-0" Oct 05 07:08:25 crc kubenswrapper[4846]: I1005 07:08:25.953541 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2adc8ba0-2810-4bf4-bd24-172b97fb9b22-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"2adc8ba0-2810-4bf4-bd24-172b97fb9b22\") " pod="openstack/glance-default-internal-api-0" Oct 05 07:08:25 crc kubenswrapper[4846]: I1005 07:08:25.953928 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/671f4c86-d748-474a-b0f0-400b16b28d7f-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"671f4c86-d748-474a-b0f0-400b16b28d7f\") " pod="openstack/ceilometer-0" Oct 05 07:08:25 crc kubenswrapper[4846]: I1005 07:08:25.957171 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qhrh2\" (UniqueName: \"kubernetes.io/projected/671f4c86-d748-474a-b0f0-400b16b28d7f-kube-api-access-qhrh2\") pod \"ceilometer-0\" (UID: \"671f4c86-d748-474a-b0f0-400b16b28d7f\") " pod="openstack/ceilometer-0" Oct 05 07:08:25 crc kubenswrapper[4846]: I1005 07:08:25.957906 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2adc8ba0-2810-4bf4-bd24-172b97fb9b22-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"2adc8ba0-2810-4bf4-bd24-172b97fb9b22\") " pod="openstack/glance-default-internal-api-0" Oct 05 07:08:25 crc kubenswrapper[4846]: I1005 07:08:25.960919 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6dxzx\" (UniqueName: \"kubernetes.io/projected/2adc8ba0-2810-4bf4-bd24-172b97fb9b22-kube-api-access-6dxzx\") pod \"glance-default-internal-api-0\" (UID: \"2adc8ba0-2810-4bf4-bd24-172b97fb9b22\") " pod="openstack/glance-default-internal-api-0" Oct 05 07:08:25 crc kubenswrapper[4846]: I1005 07:08:25.975342 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-internal-api-0\" (UID: \"2adc8ba0-2810-4bf4-bd24-172b97fb9b22\") " pod="openstack/glance-default-internal-api-0" Oct 05 07:08:25 crc kubenswrapper[4846]: I1005 07:08:25.992805 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 05 07:08:26 crc kubenswrapper[4846]: I1005 07:08:26.004727 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 05 07:08:26 crc kubenswrapper[4846]: I1005 07:08:26.512158 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="39071bf4-214b-4bbb-8b78-dfec6ad3c940" path="/var/lib/kubelet/pods/39071bf4-214b-4bbb-8b78-dfec6ad3c940/volumes" Oct 05 07:08:26 crc kubenswrapper[4846]: I1005 07:08:26.513499 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="edd6502e-6e77-4ad3-8f1f-b85426188261" path="/var/lib/kubelet/pods/edd6502e-6e77-4ad3-8f1f-b85426188261/volumes" Oct 05 07:08:26 crc kubenswrapper[4846]: I1005 07:08:26.714205 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 05 07:08:26 crc kubenswrapper[4846]: W1005 07:08:26.721429 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2adc8ba0_2810_4bf4_bd24_172b97fb9b22.slice/crio-9ae6aaa0c860e07f083871e9cd161f5f2319bfe8b6db28f062aac92ee397f992 WatchSource:0}: Error finding container 9ae6aaa0c860e07f083871e9cd161f5f2319bfe8b6db28f062aac92ee397f992: Status 404 returned error can't find the container with id 9ae6aaa0c860e07f083871e9cd161f5f2319bfe8b6db28f062aac92ee397f992 Oct 05 07:08:26 crc kubenswrapper[4846]: I1005 07:08:26.726063 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 05 07:08:26 crc kubenswrapper[4846]: I1005 07:08:26.762638 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 05 07:08:26 crc kubenswrapper[4846]: W1005 07:08:26.770939 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod671f4c86_d748_474a_b0f0_400b16b28d7f.slice/crio-b9aaf8c473dc83d993ffcfc71be15375921b9f36516973cae72e19f38ef3a48c WatchSource:0}: Error finding container b9aaf8c473dc83d993ffcfc71be15375921b9f36516973cae72e19f38ef3a48c: Status 404 returned error can't find the container with id b9aaf8c473dc83d993ffcfc71be15375921b9f36516973cae72e19f38ef3a48c Oct 05 07:08:27 crc kubenswrapper[4846]: I1005 07:08:27.591380 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"671f4c86-d748-474a-b0f0-400b16b28d7f","Type":"ContainerStarted","Data":"f9d7094326fecb094c30501623057ad8a358a8b75efbde65e9c1837200d81a04"} Oct 05 07:08:27 crc kubenswrapper[4846]: I1005 07:08:27.591836 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"671f4c86-d748-474a-b0f0-400b16b28d7f","Type":"ContainerStarted","Data":"b9aaf8c473dc83d993ffcfc71be15375921b9f36516973cae72e19f38ef3a48c"} Oct 05 07:08:27 crc kubenswrapper[4846]: I1005 07:08:27.609464 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"2adc8ba0-2810-4bf4-bd24-172b97fb9b22","Type":"ContainerStarted","Data":"3be1afa599376331b916c6589c79dd8bdcf3fefa4275b988917d64c058731c7a"} Oct 05 07:08:27 crc kubenswrapper[4846]: I1005 07:08:27.609516 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"2adc8ba0-2810-4bf4-bd24-172b97fb9b22","Type":"ContainerStarted","Data":"9ae6aaa0c860e07f083871e9cd161f5f2319bfe8b6db28f062aac92ee397f992"} Oct 05 07:08:28 crc kubenswrapper[4846]: I1005 07:08:28.627118 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"671f4c86-d748-474a-b0f0-400b16b28d7f","Type":"ContainerStarted","Data":"0a62f53c440c738a9134baa4ae47b0ee5c491ebfe690b79742628905992fcc5f"} Oct 05 07:08:28 crc kubenswrapper[4846]: I1005 07:08:28.629868 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"2adc8ba0-2810-4bf4-bd24-172b97fb9b22","Type":"ContainerStarted","Data":"9e54d631fb403eb9146c131948244ad644f213facfb43b90944167c8827ee59f"} Oct 05 07:08:28 crc kubenswrapper[4846]: I1005 07:08:28.673524 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=3.673497092 podStartE2EDuration="3.673497092s" podCreationTimestamp="2025-10-05 07:08:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 07:08:28.66030043 +0000 UTC m=+1230.901153225" watchObservedRunningTime="2025-10-05 07:08:28.673497092 +0000 UTC m=+1230.914349887" Oct 05 07:08:30 crc kubenswrapper[4846]: I1005 07:08:30.642658 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 05 07:08:30 crc kubenswrapper[4846]: I1005 07:08:30.645900 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="47ea4b78-ed6a-48c6-a7d2-0505378a874d" containerName="glance-log" containerID="cri-o://4d44457e6e512ba342767ec6d38d8709da21ae7e09f8643e92055bacd72c34da" gracePeriod=30 Oct 05 07:08:30 crc kubenswrapper[4846]: I1005 07:08:30.646122 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="47ea4b78-ed6a-48c6-a7d2-0505378a874d" containerName="glance-httpd" containerID="cri-o://3747a48030345be1e2d10045a277345f99cf674c32e826c35d08619673de537e" gracePeriod=30 Oct 05 07:08:31 crc kubenswrapper[4846]: I1005 07:08:31.687558 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"671f4c86-d748-474a-b0f0-400b16b28d7f","Type":"ContainerStarted","Data":"65af09f075e262e3ee4b22ef2e398bc14640a56118adbc1e85a3b56072b378e7"} Oct 05 07:08:31 crc kubenswrapper[4846]: I1005 07:08:31.689553 4846 generic.go:334] "Generic (PLEG): container finished" podID="47ea4b78-ed6a-48c6-a7d2-0505378a874d" containerID="4d44457e6e512ba342767ec6d38d8709da21ae7e09f8643e92055bacd72c34da" exitCode=143 Oct 05 07:08:31 crc kubenswrapper[4846]: I1005 07:08:31.689583 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"47ea4b78-ed6a-48c6-a7d2-0505378a874d","Type":"ContainerDied","Data":"4d44457e6e512ba342767ec6d38d8709da21ae7e09f8643e92055bacd72c34da"} Oct 05 07:08:32 crc kubenswrapper[4846]: I1005 07:08:32.703419 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"671f4c86-d748-474a-b0f0-400b16b28d7f","Type":"ContainerStarted","Data":"dd2b7dd7d4644b6434c27817ea994d710a784654f202d6bb8622b4eacd71e7a7"} Oct 05 07:08:32 crc kubenswrapper[4846]: I1005 07:08:32.703696 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="671f4c86-d748-474a-b0f0-400b16b28d7f" containerName="ceilometer-central-agent" containerID="cri-o://f9d7094326fecb094c30501623057ad8a358a8b75efbde65e9c1837200d81a04" gracePeriod=30 Oct 05 07:08:32 crc kubenswrapper[4846]: I1005 07:08:32.703723 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="671f4c86-d748-474a-b0f0-400b16b28d7f" containerName="proxy-httpd" containerID="cri-o://dd2b7dd7d4644b6434c27817ea994d710a784654f202d6bb8622b4eacd71e7a7" gracePeriod=30 Oct 05 07:08:32 crc kubenswrapper[4846]: I1005 07:08:32.703791 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="671f4c86-d748-474a-b0f0-400b16b28d7f" containerName="ceilometer-notification-agent" containerID="cri-o://0a62f53c440c738a9134baa4ae47b0ee5c491ebfe690b79742628905992fcc5f" gracePeriod=30 Oct 05 07:08:32 crc kubenswrapper[4846]: I1005 07:08:32.703791 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="671f4c86-d748-474a-b0f0-400b16b28d7f" containerName="sg-core" containerID="cri-o://65af09f075e262e3ee4b22ef2e398bc14640a56118adbc1e85a3b56072b378e7" gracePeriod=30 Oct 05 07:08:32 crc kubenswrapper[4846]: I1005 07:08:32.703992 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 05 07:08:32 crc kubenswrapper[4846]: I1005 07:08:32.736083 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.596111493 podStartE2EDuration="7.736053774s" podCreationTimestamp="2025-10-05 07:08:25 +0000 UTC" firstStartedPulling="2025-10-05 07:08:26.77420096 +0000 UTC m=+1229.015053725" lastFinishedPulling="2025-10-05 07:08:31.914143231 +0000 UTC m=+1234.154996006" observedRunningTime="2025-10-05 07:08:32.726643443 +0000 UTC m=+1234.967496218" watchObservedRunningTime="2025-10-05 07:08:32.736053774 +0000 UTC m=+1234.976906549" Oct 05 07:08:33 crc kubenswrapper[4846]: I1005 07:08:33.716558 4846 generic.go:334] "Generic (PLEG): container finished" podID="671f4c86-d748-474a-b0f0-400b16b28d7f" containerID="dd2b7dd7d4644b6434c27817ea994d710a784654f202d6bb8622b4eacd71e7a7" exitCode=0 Oct 05 07:08:33 crc kubenswrapper[4846]: I1005 07:08:33.717049 4846 generic.go:334] "Generic (PLEG): container finished" podID="671f4c86-d748-474a-b0f0-400b16b28d7f" containerID="65af09f075e262e3ee4b22ef2e398bc14640a56118adbc1e85a3b56072b378e7" exitCode=2 Oct 05 07:08:33 crc kubenswrapper[4846]: I1005 07:08:33.717062 4846 generic.go:334] "Generic (PLEG): container finished" podID="671f4c86-d748-474a-b0f0-400b16b28d7f" containerID="0a62f53c440c738a9134baa4ae47b0ee5c491ebfe690b79742628905992fcc5f" exitCode=0 Oct 05 07:08:33 crc kubenswrapper[4846]: I1005 07:08:33.716672 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"671f4c86-d748-474a-b0f0-400b16b28d7f","Type":"ContainerDied","Data":"dd2b7dd7d4644b6434c27817ea994d710a784654f202d6bb8622b4eacd71e7a7"} Oct 05 07:08:33 crc kubenswrapper[4846]: I1005 07:08:33.717107 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"671f4c86-d748-474a-b0f0-400b16b28d7f","Type":"ContainerDied","Data":"65af09f075e262e3ee4b22ef2e398bc14640a56118adbc1e85a3b56072b378e7"} Oct 05 07:08:33 crc kubenswrapper[4846]: I1005 07:08:33.717124 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"671f4c86-d748-474a-b0f0-400b16b28d7f","Type":"ContainerDied","Data":"0a62f53c440c738a9134baa4ae47b0ee5c491ebfe690b79742628905992fcc5f"} Oct 05 07:08:34 crc kubenswrapper[4846]: I1005 07:08:34.381685 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 05 07:08:34 crc kubenswrapper[4846]: I1005 07:08:34.556757 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/47ea4b78-ed6a-48c6-a7d2-0505378a874d-config-data\") pod \"47ea4b78-ed6a-48c6-a7d2-0505378a874d\" (UID: \"47ea4b78-ed6a-48c6-a7d2-0505378a874d\") " Oct 05 07:08:34 crc kubenswrapper[4846]: I1005 07:08:34.556805 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/47ea4b78-ed6a-48c6-a7d2-0505378a874d-combined-ca-bundle\") pod \"47ea4b78-ed6a-48c6-a7d2-0505378a874d\" (UID: \"47ea4b78-ed6a-48c6-a7d2-0505378a874d\") " Oct 05 07:08:34 crc kubenswrapper[4846]: I1005 07:08:34.556886 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rztq7\" (UniqueName: \"kubernetes.io/projected/47ea4b78-ed6a-48c6-a7d2-0505378a874d-kube-api-access-rztq7\") pod \"47ea4b78-ed6a-48c6-a7d2-0505378a874d\" (UID: \"47ea4b78-ed6a-48c6-a7d2-0505378a874d\") " Oct 05 07:08:34 crc kubenswrapper[4846]: I1005 07:08:34.556953 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/47ea4b78-ed6a-48c6-a7d2-0505378a874d-scripts\") pod \"47ea4b78-ed6a-48c6-a7d2-0505378a874d\" (UID: \"47ea4b78-ed6a-48c6-a7d2-0505378a874d\") " Oct 05 07:08:34 crc kubenswrapper[4846]: I1005 07:08:34.557005 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/47ea4b78-ed6a-48c6-a7d2-0505378a874d-httpd-run\") pod \"47ea4b78-ed6a-48c6-a7d2-0505378a874d\" (UID: \"47ea4b78-ed6a-48c6-a7d2-0505378a874d\") " Oct 05 07:08:34 crc kubenswrapper[4846]: I1005 07:08:34.557045 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"47ea4b78-ed6a-48c6-a7d2-0505378a874d\" (UID: \"47ea4b78-ed6a-48c6-a7d2-0505378a874d\") " Oct 05 07:08:34 crc kubenswrapper[4846]: I1005 07:08:34.557100 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/47ea4b78-ed6a-48c6-a7d2-0505378a874d-public-tls-certs\") pod \"47ea4b78-ed6a-48c6-a7d2-0505378a874d\" (UID: \"47ea4b78-ed6a-48c6-a7d2-0505378a874d\") " Oct 05 07:08:34 crc kubenswrapper[4846]: I1005 07:08:34.557155 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/47ea4b78-ed6a-48c6-a7d2-0505378a874d-logs\") pod \"47ea4b78-ed6a-48c6-a7d2-0505378a874d\" (UID: \"47ea4b78-ed6a-48c6-a7d2-0505378a874d\") " Oct 05 07:08:34 crc kubenswrapper[4846]: I1005 07:08:34.557975 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/47ea4b78-ed6a-48c6-a7d2-0505378a874d-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "47ea4b78-ed6a-48c6-a7d2-0505378a874d" (UID: "47ea4b78-ed6a-48c6-a7d2-0505378a874d"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 07:08:34 crc kubenswrapper[4846]: I1005 07:08:34.558758 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/47ea4b78-ed6a-48c6-a7d2-0505378a874d-logs" (OuterVolumeSpecName: "logs") pod "47ea4b78-ed6a-48c6-a7d2-0505378a874d" (UID: "47ea4b78-ed6a-48c6-a7d2-0505378a874d"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 07:08:34 crc kubenswrapper[4846]: I1005 07:08:34.559597 4846 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/47ea4b78-ed6a-48c6-a7d2-0505378a874d-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 05 07:08:34 crc kubenswrapper[4846]: I1005 07:08:34.565577 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage07-crc" (OuterVolumeSpecName: "glance") pod "47ea4b78-ed6a-48c6-a7d2-0505378a874d" (UID: "47ea4b78-ed6a-48c6-a7d2-0505378a874d"). InnerVolumeSpecName "local-storage07-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 05 07:08:34 crc kubenswrapper[4846]: I1005 07:08:34.565880 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/47ea4b78-ed6a-48c6-a7d2-0505378a874d-scripts" (OuterVolumeSpecName: "scripts") pod "47ea4b78-ed6a-48c6-a7d2-0505378a874d" (UID: "47ea4b78-ed6a-48c6-a7d2-0505378a874d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:08:34 crc kubenswrapper[4846]: I1005 07:08:34.574293 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/47ea4b78-ed6a-48c6-a7d2-0505378a874d-kube-api-access-rztq7" (OuterVolumeSpecName: "kube-api-access-rztq7") pod "47ea4b78-ed6a-48c6-a7d2-0505378a874d" (UID: "47ea4b78-ed6a-48c6-a7d2-0505378a874d"). InnerVolumeSpecName "kube-api-access-rztq7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:08:34 crc kubenswrapper[4846]: I1005 07:08:34.596061 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/47ea4b78-ed6a-48c6-a7d2-0505378a874d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "47ea4b78-ed6a-48c6-a7d2-0505378a874d" (UID: "47ea4b78-ed6a-48c6-a7d2-0505378a874d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:08:34 crc kubenswrapper[4846]: I1005 07:08:34.629520 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/47ea4b78-ed6a-48c6-a7d2-0505378a874d-config-data" (OuterVolumeSpecName: "config-data") pod "47ea4b78-ed6a-48c6-a7d2-0505378a874d" (UID: "47ea4b78-ed6a-48c6-a7d2-0505378a874d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:08:34 crc kubenswrapper[4846]: I1005 07:08:34.631799 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/47ea4b78-ed6a-48c6-a7d2-0505378a874d-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "47ea4b78-ed6a-48c6-a7d2-0505378a874d" (UID: "47ea4b78-ed6a-48c6-a7d2-0505378a874d"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:08:34 crc kubenswrapper[4846]: I1005 07:08:34.661652 4846 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/47ea4b78-ed6a-48c6-a7d2-0505378a874d-logs\") on node \"crc\" DevicePath \"\"" Oct 05 07:08:34 crc kubenswrapper[4846]: I1005 07:08:34.661694 4846 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/47ea4b78-ed6a-48c6-a7d2-0505378a874d-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 07:08:34 crc kubenswrapper[4846]: I1005 07:08:34.661705 4846 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/47ea4b78-ed6a-48c6-a7d2-0505378a874d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 07:08:34 crc kubenswrapper[4846]: I1005 07:08:34.661731 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rztq7\" (UniqueName: \"kubernetes.io/projected/47ea4b78-ed6a-48c6-a7d2-0505378a874d-kube-api-access-rztq7\") on node \"crc\" DevicePath \"\"" Oct 05 07:08:34 crc kubenswrapper[4846]: I1005 07:08:34.661743 4846 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/47ea4b78-ed6a-48c6-a7d2-0505378a874d-scripts\") on node \"crc\" DevicePath \"\"" Oct 05 07:08:34 crc kubenswrapper[4846]: I1005 07:08:34.661770 4846 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" " Oct 05 07:08:34 crc kubenswrapper[4846]: I1005 07:08:34.661780 4846 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/47ea4b78-ed6a-48c6-a7d2-0505378a874d-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 05 07:08:34 crc kubenswrapper[4846]: I1005 07:08:34.682859 4846 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage07-crc" (UniqueName: "kubernetes.io/local-volume/local-storage07-crc") on node "crc" Oct 05 07:08:34 crc kubenswrapper[4846]: I1005 07:08:34.728807 4846 generic.go:334] "Generic (PLEG): container finished" podID="47ea4b78-ed6a-48c6-a7d2-0505378a874d" containerID="3747a48030345be1e2d10045a277345f99cf674c32e826c35d08619673de537e" exitCode=0 Oct 05 07:08:34 crc kubenswrapper[4846]: I1005 07:08:34.728853 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"47ea4b78-ed6a-48c6-a7d2-0505378a874d","Type":"ContainerDied","Data":"3747a48030345be1e2d10045a277345f99cf674c32e826c35d08619673de537e"} Oct 05 07:08:34 crc kubenswrapper[4846]: I1005 07:08:34.728884 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"47ea4b78-ed6a-48c6-a7d2-0505378a874d","Type":"ContainerDied","Data":"2fa57817123fca2619cf1324f4ec1232fd6cbdee734ffdcbdc7b17945b49d572"} Oct 05 07:08:34 crc kubenswrapper[4846]: I1005 07:08:34.728903 4846 scope.go:117] "RemoveContainer" containerID="3747a48030345be1e2d10045a277345f99cf674c32e826c35d08619673de537e" Oct 05 07:08:34 crc kubenswrapper[4846]: I1005 07:08:34.729027 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 05 07:08:34 crc kubenswrapper[4846]: I1005 07:08:34.768095 4846 reconciler_common.go:293] "Volume detached for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" DevicePath \"\"" Oct 05 07:08:34 crc kubenswrapper[4846]: I1005 07:08:34.777931 4846 scope.go:117] "RemoveContainer" containerID="4d44457e6e512ba342767ec6d38d8709da21ae7e09f8643e92055bacd72c34da" Oct 05 07:08:34 crc kubenswrapper[4846]: I1005 07:08:34.781335 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 05 07:08:34 crc kubenswrapper[4846]: I1005 07:08:34.794037 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 05 07:08:34 crc kubenswrapper[4846]: I1005 07:08:34.813682 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Oct 05 07:08:34 crc kubenswrapper[4846]: E1005 07:08:34.822146 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="47ea4b78-ed6a-48c6-a7d2-0505378a874d" containerName="glance-log" Oct 05 07:08:34 crc kubenswrapper[4846]: I1005 07:08:34.822192 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="47ea4b78-ed6a-48c6-a7d2-0505378a874d" containerName="glance-log" Oct 05 07:08:34 crc kubenswrapper[4846]: E1005 07:08:34.822273 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="47ea4b78-ed6a-48c6-a7d2-0505378a874d" containerName="glance-httpd" Oct 05 07:08:34 crc kubenswrapper[4846]: I1005 07:08:34.822279 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="47ea4b78-ed6a-48c6-a7d2-0505378a874d" containerName="glance-httpd" Oct 05 07:08:34 crc kubenswrapper[4846]: I1005 07:08:34.824163 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="47ea4b78-ed6a-48c6-a7d2-0505378a874d" containerName="glance-log" Oct 05 07:08:34 crc kubenswrapper[4846]: I1005 07:08:34.824233 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="47ea4b78-ed6a-48c6-a7d2-0505378a874d" containerName="glance-httpd" Oct 05 07:08:34 crc kubenswrapper[4846]: I1005 07:08:34.827405 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 05 07:08:34 crc kubenswrapper[4846]: I1005 07:08:34.831698 4846 scope.go:117] "RemoveContainer" containerID="3747a48030345be1e2d10045a277345f99cf674c32e826c35d08619673de537e" Oct 05 07:08:34 crc kubenswrapper[4846]: I1005 07:08:34.832026 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Oct 05 07:08:34 crc kubenswrapper[4846]: I1005 07:08:34.832045 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Oct 05 07:08:34 crc kubenswrapper[4846]: E1005 07:08:34.833029 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3747a48030345be1e2d10045a277345f99cf674c32e826c35d08619673de537e\": container with ID starting with 3747a48030345be1e2d10045a277345f99cf674c32e826c35d08619673de537e not found: ID does not exist" containerID="3747a48030345be1e2d10045a277345f99cf674c32e826c35d08619673de537e" Oct 05 07:08:34 crc kubenswrapper[4846]: I1005 07:08:34.833053 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3747a48030345be1e2d10045a277345f99cf674c32e826c35d08619673de537e"} err="failed to get container status \"3747a48030345be1e2d10045a277345f99cf674c32e826c35d08619673de537e\": rpc error: code = NotFound desc = could not find container \"3747a48030345be1e2d10045a277345f99cf674c32e826c35d08619673de537e\": container with ID starting with 3747a48030345be1e2d10045a277345f99cf674c32e826c35d08619673de537e not found: ID does not exist" Oct 05 07:08:34 crc kubenswrapper[4846]: I1005 07:08:34.833072 4846 scope.go:117] "RemoveContainer" containerID="4d44457e6e512ba342767ec6d38d8709da21ae7e09f8643e92055bacd72c34da" Oct 05 07:08:34 crc kubenswrapper[4846]: E1005 07:08:34.835088 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4d44457e6e512ba342767ec6d38d8709da21ae7e09f8643e92055bacd72c34da\": container with ID starting with 4d44457e6e512ba342767ec6d38d8709da21ae7e09f8643e92055bacd72c34da not found: ID does not exist" containerID="4d44457e6e512ba342767ec6d38d8709da21ae7e09f8643e92055bacd72c34da" Oct 05 07:08:34 crc kubenswrapper[4846]: I1005 07:08:34.835132 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4d44457e6e512ba342767ec6d38d8709da21ae7e09f8643e92055bacd72c34da"} err="failed to get container status \"4d44457e6e512ba342767ec6d38d8709da21ae7e09f8643e92055bacd72c34da\": rpc error: code = NotFound desc = could not find container \"4d44457e6e512ba342767ec6d38d8709da21ae7e09f8643e92055bacd72c34da\": container with ID starting with 4d44457e6e512ba342767ec6d38d8709da21ae7e09f8643e92055bacd72c34da not found: ID does not exist" Oct 05 07:08:34 crc kubenswrapper[4846]: I1005 07:08:34.859704 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 05 07:08:34 crc kubenswrapper[4846]: I1005 07:08:34.978761 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0c0f8308-455e-4544-9f53-0c06a54c6e22-logs\") pod \"glance-default-external-api-0\" (UID: \"0c0f8308-455e-4544-9f53-0c06a54c6e22\") " pod="openstack/glance-default-external-api-0" Oct 05 07:08:34 crc kubenswrapper[4846]: I1005 07:08:34.979041 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-external-api-0\" (UID: \"0c0f8308-455e-4544-9f53-0c06a54c6e22\") " pod="openstack/glance-default-external-api-0" Oct 05 07:08:34 crc kubenswrapper[4846]: I1005 07:08:34.979118 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0c0f8308-455e-4544-9f53-0c06a54c6e22-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"0c0f8308-455e-4544-9f53-0c06a54c6e22\") " pod="openstack/glance-default-external-api-0" Oct 05 07:08:34 crc kubenswrapper[4846]: I1005 07:08:34.979266 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0c0f8308-455e-4544-9f53-0c06a54c6e22-scripts\") pod \"glance-default-external-api-0\" (UID: \"0c0f8308-455e-4544-9f53-0c06a54c6e22\") " pod="openstack/glance-default-external-api-0" Oct 05 07:08:34 crc kubenswrapper[4846]: I1005 07:08:34.979375 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pjd99\" (UniqueName: \"kubernetes.io/projected/0c0f8308-455e-4544-9f53-0c06a54c6e22-kube-api-access-pjd99\") pod \"glance-default-external-api-0\" (UID: \"0c0f8308-455e-4544-9f53-0c06a54c6e22\") " pod="openstack/glance-default-external-api-0" Oct 05 07:08:34 crc kubenswrapper[4846]: I1005 07:08:34.979407 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/0c0f8308-455e-4544-9f53-0c06a54c6e22-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"0c0f8308-455e-4544-9f53-0c06a54c6e22\") " pod="openstack/glance-default-external-api-0" Oct 05 07:08:34 crc kubenswrapper[4846]: I1005 07:08:34.979465 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0c0f8308-455e-4544-9f53-0c06a54c6e22-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"0c0f8308-455e-4544-9f53-0c06a54c6e22\") " pod="openstack/glance-default-external-api-0" Oct 05 07:08:34 crc kubenswrapper[4846]: I1005 07:08:34.979583 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0c0f8308-455e-4544-9f53-0c06a54c6e22-config-data\") pod \"glance-default-external-api-0\" (UID: \"0c0f8308-455e-4544-9f53-0c06a54c6e22\") " pod="openstack/glance-default-external-api-0" Oct 05 07:08:35 crc kubenswrapper[4846]: I1005 07:08:35.081845 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0c0f8308-455e-4544-9f53-0c06a54c6e22-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"0c0f8308-455e-4544-9f53-0c06a54c6e22\") " pod="openstack/glance-default-external-api-0" Oct 05 07:08:35 crc kubenswrapper[4846]: I1005 07:08:35.081926 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0c0f8308-455e-4544-9f53-0c06a54c6e22-config-data\") pod \"glance-default-external-api-0\" (UID: \"0c0f8308-455e-4544-9f53-0c06a54c6e22\") " pod="openstack/glance-default-external-api-0" Oct 05 07:08:35 crc kubenswrapper[4846]: I1005 07:08:35.082739 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0c0f8308-455e-4544-9f53-0c06a54c6e22-logs\") pod \"glance-default-external-api-0\" (UID: \"0c0f8308-455e-4544-9f53-0c06a54c6e22\") " pod="openstack/glance-default-external-api-0" Oct 05 07:08:35 crc kubenswrapper[4846]: I1005 07:08:35.082805 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-external-api-0\" (UID: \"0c0f8308-455e-4544-9f53-0c06a54c6e22\") " pod="openstack/glance-default-external-api-0" Oct 05 07:08:35 crc kubenswrapper[4846]: I1005 07:08:35.082829 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0c0f8308-455e-4544-9f53-0c06a54c6e22-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"0c0f8308-455e-4544-9f53-0c06a54c6e22\") " pod="openstack/glance-default-external-api-0" Oct 05 07:08:35 crc kubenswrapper[4846]: I1005 07:08:35.082861 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0c0f8308-455e-4544-9f53-0c06a54c6e22-scripts\") pod \"glance-default-external-api-0\" (UID: \"0c0f8308-455e-4544-9f53-0c06a54c6e22\") " pod="openstack/glance-default-external-api-0" Oct 05 07:08:35 crc kubenswrapper[4846]: I1005 07:08:35.082886 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pjd99\" (UniqueName: \"kubernetes.io/projected/0c0f8308-455e-4544-9f53-0c06a54c6e22-kube-api-access-pjd99\") pod \"glance-default-external-api-0\" (UID: \"0c0f8308-455e-4544-9f53-0c06a54c6e22\") " pod="openstack/glance-default-external-api-0" Oct 05 07:08:35 crc kubenswrapper[4846]: I1005 07:08:35.082903 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/0c0f8308-455e-4544-9f53-0c06a54c6e22-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"0c0f8308-455e-4544-9f53-0c06a54c6e22\") " pod="openstack/glance-default-external-api-0" Oct 05 07:08:35 crc kubenswrapper[4846]: I1005 07:08:35.083052 4846 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-external-api-0\" (UID: \"0c0f8308-455e-4544-9f53-0c06a54c6e22\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/glance-default-external-api-0" Oct 05 07:08:35 crc kubenswrapper[4846]: I1005 07:08:35.083130 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0c0f8308-455e-4544-9f53-0c06a54c6e22-logs\") pod \"glance-default-external-api-0\" (UID: \"0c0f8308-455e-4544-9f53-0c06a54c6e22\") " pod="openstack/glance-default-external-api-0" Oct 05 07:08:35 crc kubenswrapper[4846]: I1005 07:08:35.083280 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/0c0f8308-455e-4544-9f53-0c06a54c6e22-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"0c0f8308-455e-4544-9f53-0c06a54c6e22\") " pod="openstack/glance-default-external-api-0" Oct 05 07:08:35 crc kubenswrapper[4846]: I1005 07:08:35.090624 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0c0f8308-455e-4544-9f53-0c06a54c6e22-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"0c0f8308-455e-4544-9f53-0c06a54c6e22\") " pod="openstack/glance-default-external-api-0" Oct 05 07:08:35 crc kubenswrapper[4846]: I1005 07:08:35.090699 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0c0f8308-455e-4544-9f53-0c06a54c6e22-scripts\") pod \"glance-default-external-api-0\" (UID: \"0c0f8308-455e-4544-9f53-0c06a54c6e22\") " pod="openstack/glance-default-external-api-0" Oct 05 07:08:35 crc kubenswrapper[4846]: I1005 07:08:35.093441 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0c0f8308-455e-4544-9f53-0c06a54c6e22-config-data\") pod \"glance-default-external-api-0\" (UID: \"0c0f8308-455e-4544-9f53-0c06a54c6e22\") " pod="openstack/glance-default-external-api-0" Oct 05 07:08:35 crc kubenswrapper[4846]: I1005 07:08:35.098156 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0c0f8308-455e-4544-9f53-0c06a54c6e22-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"0c0f8308-455e-4544-9f53-0c06a54c6e22\") " pod="openstack/glance-default-external-api-0" Oct 05 07:08:35 crc kubenswrapper[4846]: I1005 07:08:35.104501 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pjd99\" (UniqueName: \"kubernetes.io/projected/0c0f8308-455e-4544-9f53-0c06a54c6e22-kube-api-access-pjd99\") pod \"glance-default-external-api-0\" (UID: \"0c0f8308-455e-4544-9f53-0c06a54c6e22\") " pod="openstack/glance-default-external-api-0" Oct 05 07:08:35 crc kubenswrapper[4846]: I1005 07:08:35.121652 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-external-api-0\" (UID: \"0c0f8308-455e-4544-9f53-0c06a54c6e22\") " pod="openstack/glance-default-external-api-0" Oct 05 07:08:35 crc kubenswrapper[4846]: I1005 07:08:35.161519 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 05 07:08:35 crc kubenswrapper[4846]: I1005 07:08:35.800055 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 05 07:08:35 crc kubenswrapper[4846]: W1005 07:08:35.815767 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0c0f8308_455e_4544_9f53_0c06a54c6e22.slice/crio-69378334ebffca563252133a91b7f973164e307ee0c19be4b2afa5c70cd826c5 WatchSource:0}: Error finding container 69378334ebffca563252133a91b7f973164e307ee0c19be4b2afa5c70cd826c5: Status 404 returned error can't find the container with id 69378334ebffca563252133a91b7f973164e307ee0c19be4b2afa5c70cd826c5 Oct 05 07:08:36 crc kubenswrapper[4846]: I1005 07:08:36.006297 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Oct 05 07:08:36 crc kubenswrapper[4846]: I1005 07:08:36.006360 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Oct 05 07:08:36 crc kubenswrapper[4846]: I1005 07:08:36.044003 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Oct 05 07:08:36 crc kubenswrapper[4846]: I1005 07:08:36.076388 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Oct 05 07:08:36 crc kubenswrapper[4846]: I1005 07:08:36.527540 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="47ea4b78-ed6a-48c6-a7d2-0505378a874d" path="/var/lib/kubelet/pods/47ea4b78-ed6a-48c6-a7d2-0505378a874d/volumes" Oct 05 07:08:36 crc kubenswrapper[4846]: I1005 07:08:36.756506 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"0c0f8308-455e-4544-9f53-0c06a54c6e22","Type":"ContainerStarted","Data":"623a1c303003463ea428d6d40cb44811fb558226463a79ed21399df208bb4505"} Oct 05 07:08:36 crc kubenswrapper[4846]: I1005 07:08:36.756588 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"0c0f8308-455e-4544-9f53-0c06a54c6e22","Type":"ContainerStarted","Data":"69378334ebffca563252133a91b7f973164e307ee0c19be4b2afa5c70cd826c5"} Oct 05 07:08:36 crc kubenswrapper[4846]: I1005 07:08:36.756891 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Oct 05 07:08:36 crc kubenswrapper[4846]: I1005 07:08:36.756961 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Oct 05 07:08:37 crc kubenswrapper[4846]: I1005 07:08:37.559492 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 05 07:08:37 crc kubenswrapper[4846]: I1005 07:08:37.746000 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qhrh2\" (UniqueName: \"kubernetes.io/projected/671f4c86-d748-474a-b0f0-400b16b28d7f-kube-api-access-qhrh2\") pod \"671f4c86-d748-474a-b0f0-400b16b28d7f\" (UID: \"671f4c86-d748-474a-b0f0-400b16b28d7f\") " Oct 05 07:08:37 crc kubenswrapper[4846]: I1005 07:08:37.746102 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/671f4c86-d748-474a-b0f0-400b16b28d7f-scripts\") pod \"671f4c86-d748-474a-b0f0-400b16b28d7f\" (UID: \"671f4c86-d748-474a-b0f0-400b16b28d7f\") " Oct 05 07:08:37 crc kubenswrapper[4846]: I1005 07:08:37.746251 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/671f4c86-d748-474a-b0f0-400b16b28d7f-combined-ca-bundle\") pod \"671f4c86-d748-474a-b0f0-400b16b28d7f\" (UID: \"671f4c86-d748-474a-b0f0-400b16b28d7f\") " Oct 05 07:08:37 crc kubenswrapper[4846]: I1005 07:08:37.746288 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/671f4c86-d748-474a-b0f0-400b16b28d7f-run-httpd\") pod \"671f4c86-d748-474a-b0f0-400b16b28d7f\" (UID: \"671f4c86-d748-474a-b0f0-400b16b28d7f\") " Oct 05 07:08:37 crc kubenswrapper[4846]: I1005 07:08:37.746381 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/671f4c86-d748-474a-b0f0-400b16b28d7f-log-httpd\") pod \"671f4c86-d748-474a-b0f0-400b16b28d7f\" (UID: \"671f4c86-d748-474a-b0f0-400b16b28d7f\") " Oct 05 07:08:37 crc kubenswrapper[4846]: I1005 07:08:37.746424 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/671f4c86-d748-474a-b0f0-400b16b28d7f-config-data\") pod \"671f4c86-d748-474a-b0f0-400b16b28d7f\" (UID: \"671f4c86-d748-474a-b0f0-400b16b28d7f\") " Oct 05 07:08:37 crc kubenswrapper[4846]: I1005 07:08:37.746515 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/671f4c86-d748-474a-b0f0-400b16b28d7f-sg-core-conf-yaml\") pod \"671f4c86-d748-474a-b0f0-400b16b28d7f\" (UID: \"671f4c86-d748-474a-b0f0-400b16b28d7f\") " Oct 05 07:08:37 crc kubenswrapper[4846]: I1005 07:08:37.746932 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/671f4c86-d748-474a-b0f0-400b16b28d7f-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "671f4c86-d748-474a-b0f0-400b16b28d7f" (UID: "671f4c86-d748-474a-b0f0-400b16b28d7f"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 07:08:37 crc kubenswrapper[4846]: I1005 07:08:37.747154 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/671f4c86-d748-474a-b0f0-400b16b28d7f-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "671f4c86-d748-474a-b0f0-400b16b28d7f" (UID: "671f4c86-d748-474a-b0f0-400b16b28d7f"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 07:08:37 crc kubenswrapper[4846]: I1005 07:08:37.754976 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/671f4c86-d748-474a-b0f0-400b16b28d7f-kube-api-access-qhrh2" (OuterVolumeSpecName: "kube-api-access-qhrh2") pod "671f4c86-d748-474a-b0f0-400b16b28d7f" (UID: "671f4c86-d748-474a-b0f0-400b16b28d7f"). InnerVolumeSpecName "kube-api-access-qhrh2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:08:37 crc kubenswrapper[4846]: I1005 07:08:37.771868 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/671f4c86-d748-474a-b0f0-400b16b28d7f-scripts" (OuterVolumeSpecName: "scripts") pod "671f4c86-d748-474a-b0f0-400b16b28d7f" (UID: "671f4c86-d748-474a-b0f0-400b16b28d7f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:08:37 crc kubenswrapper[4846]: I1005 07:08:37.784263 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/671f4c86-d748-474a-b0f0-400b16b28d7f-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "671f4c86-d748-474a-b0f0-400b16b28d7f" (UID: "671f4c86-d748-474a-b0f0-400b16b28d7f"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:08:37 crc kubenswrapper[4846]: I1005 07:08:37.792708 4846 generic.go:334] "Generic (PLEG): container finished" podID="671f4c86-d748-474a-b0f0-400b16b28d7f" containerID="f9d7094326fecb094c30501623057ad8a358a8b75efbde65e9c1837200d81a04" exitCode=0 Oct 05 07:08:37 crc kubenswrapper[4846]: I1005 07:08:37.793011 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 05 07:08:37 crc kubenswrapper[4846]: I1005 07:08:37.793121 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"671f4c86-d748-474a-b0f0-400b16b28d7f","Type":"ContainerDied","Data":"f9d7094326fecb094c30501623057ad8a358a8b75efbde65e9c1837200d81a04"} Oct 05 07:08:37 crc kubenswrapper[4846]: I1005 07:08:37.793244 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"671f4c86-d748-474a-b0f0-400b16b28d7f","Type":"ContainerDied","Data":"b9aaf8c473dc83d993ffcfc71be15375921b9f36516973cae72e19f38ef3a48c"} Oct 05 07:08:37 crc kubenswrapper[4846]: I1005 07:08:37.793271 4846 scope.go:117] "RemoveContainer" containerID="dd2b7dd7d4644b6434c27817ea994d710a784654f202d6bb8622b4eacd71e7a7" Oct 05 07:08:37 crc kubenswrapper[4846]: I1005 07:08:37.800657 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"0c0f8308-455e-4544-9f53-0c06a54c6e22","Type":"ContainerStarted","Data":"708780fcc4181d993d6e7d44ef9444fa8e85b47f3a59d10a6fa23aa84fa714b0"} Oct 05 07:08:37 crc kubenswrapper[4846]: I1005 07:08:37.837322 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=3.837299243 podStartE2EDuration="3.837299243s" podCreationTimestamp="2025-10-05 07:08:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 07:08:37.826192927 +0000 UTC m=+1240.067045712" watchObservedRunningTime="2025-10-05 07:08:37.837299243 +0000 UTC m=+1240.078152018" Oct 05 07:08:37 crc kubenswrapper[4846]: I1005 07:08:37.849680 4846 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/671f4c86-d748-474a-b0f0-400b16b28d7f-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 05 07:08:37 crc kubenswrapper[4846]: I1005 07:08:37.849707 4846 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/671f4c86-d748-474a-b0f0-400b16b28d7f-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 05 07:08:37 crc kubenswrapper[4846]: I1005 07:08:37.849716 4846 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/671f4c86-d748-474a-b0f0-400b16b28d7f-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 05 07:08:37 crc kubenswrapper[4846]: I1005 07:08:37.849729 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qhrh2\" (UniqueName: \"kubernetes.io/projected/671f4c86-d748-474a-b0f0-400b16b28d7f-kube-api-access-qhrh2\") on node \"crc\" DevicePath \"\"" Oct 05 07:08:37 crc kubenswrapper[4846]: I1005 07:08:37.849738 4846 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/671f4c86-d748-474a-b0f0-400b16b28d7f-scripts\") on node \"crc\" DevicePath \"\"" Oct 05 07:08:37 crc kubenswrapper[4846]: I1005 07:08:37.861770 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/671f4c86-d748-474a-b0f0-400b16b28d7f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "671f4c86-d748-474a-b0f0-400b16b28d7f" (UID: "671f4c86-d748-474a-b0f0-400b16b28d7f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:08:37 crc kubenswrapper[4846]: I1005 07:08:37.905974 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/671f4c86-d748-474a-b0f0-400b16b28d7f-config-data" (OuterVolumeSpecName: "config-data") pod "671f4c86-d748-474a-b0f0-400b16b28d7f" (UID: "671f4c86-d748-474a-b0f0-400b16b28d7f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:08:37 crc kubenswrapper[4846]: I1005 07:08:37.937356 4846 scope.go:117] "RemoveContainer" containerID="65af09f075e262e3ee4b22ef2e398bc14640a56118adbc1e85a3b56072b378e7" Oct 05 07:08:37 crc kubenswrapper[4846]: I1005 07:08:37.952469 4846 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/671f4c86-d748-474a-b0f0-400b16b28d7f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 07:08:37 crc kubenswrapper[4846]: I1005 07:08:37.952513 4846 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/671f4c86-d748-474a-b0f0-400b16b28d7f-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 07:08:37 crc kubenswrapper[4846]: I1005 07:08:37.964272 4846 scope.go:117] "RemoveContainer" containerID="0a62f53c440c738a9134baa4ae47b0ee5c491ebfe690b79742628905992fcc5f" Oct 05 07:08:37 crc kubenswrapper[4846]: I1005 07:08:37.993873 4846 scope.go:117] "RemoveContainer" containerID="f9d7094326fecb094c30501623057ad8a358a8b75efbde65e9c1837200d81a04" Oct 05 07:08:38 crc kubenswrapper[4846]: I1005 07:08:38.019678 4846 scope.go:117] "RemoveContainer" containerID="dd2b7dd7d4644b6434c27817ea994d710a784654f202d6bb8622b4eacd71e7a7" Oct 05 07:08:38 crc kubenswrapper[4846]: E1005 07:08:38.020879 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dd2b7dd7d4644b6434c27817ea994d710a784654f202d6bb8622b4eacd71e7a7\": container with ID starting with dd2b7dd7d4644b6434c27817ea994d710a784654f202d6bb8622b4eacd71e7a7 not found: ID does not exist" containerID="dd2b7dd7d4644b6434c27817ea994d710a784654f202d6bb8622b4eacd71e7a7" Oct 05 07:08:38 crc kubenswrapper[4846]: I1005 07:08:38.020950 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dd2b7dd7d4644b6434c27817ea994d710a784654f202d6bb8622b4eacd71e7a7"} err="failed to get container status \"dd2b7dd7d4644b6434c27817ea994d710a784654f202d6bb8622b4eacd71e7a7\": rpc error: code = NotFound desc = could not find container \"dd2b7dd7d4644b6434c27817ea994d710a784654f202d6bb8622b4eacd71e7a7\": container with ID starting with dd2b7dd7d4644b6434c27817ea994d710a784654f202d6bb8622b4eacd71e7a7 not found: ID does not exist" Oct 05 07:08:38 crc kubenswrapper[4846]: I1005 07:08:38.020998 4846 scope.go:117] "RemoveContainer" containerID="65af09f075e262e3ee4b22ef2e398bc14640a56118adbc1e85a3b56072b378e7" Oct 05 07:08:38 crc kubenswrapper[4846]: E1005 07:08:38.021635 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"65af09f075e262e3ee4b22ef2e398bc14640a56118adbc1e85a3b56072b378e7\": container with ID starting with 65af09f075e262e3ee4b22ef2e398bc14640a56118adbc1e85a3b56072b378e7 not found: ID does not exist" containerID="65af09f075e262e3ee4b22ef2e398bc14640a56118adbc1e85a3b56072b378e7" Oct 05 07:08:38 crc kubenswrapper[4846]: I1005 07:08:38.021664 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"65af09f075e262e3ee4b22ef2e398bc14640a56118adbc1e85a3b56072b378e7"} err="failed to get container status \"65af09f075e262e3ee4b22ef2e398bc14640a56118adbc1e85a3b56072b378e7\": rpc error: code = NotFound desc = could not find container \"65af09f075e262e3ee4b22ef2e398bc14640a56118adbc1e85a3b56072b378e7\": container with ID starting with 65af09f075e262e3ee4b22ef2e398bc14640a56118adbc1e85a3b56072b378e7 not found: ID does not exist" Oct 05 07:08:38 crc kubenswrapper[4846]: I1005 07:08:38.021679 4846 scope.go:117] "RemoveContainer" containerID="0a62f53c440c738a9134baa4ae47b0ee5c491ebfe690b79742628905992fcc5f" Oct 05 07:08:38 crc kubenswrapper[4846]: E1005 07:08:38.022142 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0a62f53c440c738a9134baa4ae47b0ee5c491ebfe690b79742628905992fcc5f\": container with ID starting with 0a62f53c440c738a9134baa4ae47b0ee5c491ebfe690b79742628905992fcc5f not found: ID does not exist" containerID="0a62f53c440c738a9134baa4ae47b0ee5c491ebfe690b79742628905992fcc5f" Oct 05 07:08:38 crc kubenswrapper[4846]: I1005 07:08:38.022164 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0a62f53c440c738a9134baa4ae47b0ee5c491ebfe690b79742628905992fcc5f"} err="failed to get container status \"0a62f53c440c738a9134baa4ae47b0ee5c491ebfe690b79742628905992fcc5f\": rpc error: code = NotFound desc = could not find container \"0a62f53c440c738a9134baa4ae47b0ee5c491ebfe690b79742628905992fcc5f\": container with ID starting with 0a62f53c440c738a9134baa4ae47b0ee5c491ebfe690b79742628905992fcc5f not found: ID does not exist" Oct 05 07:08:38 crc kubenswrapper[4846]: I1005 07:08:38.022262 4846 scope.go:117] "RemoveContainer" containerID="f9d7094326fecb094c30501623057ad8a358a8b75efbde65e9c1837200d81a04" Oct 05 07:08:38 crc kubenswrapper[4846]: E1005 07:08:38.022704 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f9d7094326fecb094c30501623057ad8a358a8b75efbde65e9c1837200d81a04\": container with ID starting with f9d7094326fecb094c30501623057ad8a358a8b75efbde65e9c1837200d81a04 not found: ID does not exist" containerID="f9d7094326fecb094c30501623057ad8a358a8b75efbde65e9c1837200d81a04" Oct 05 07:08:38 crc kubenswrapper[4846]: I1005 07:08:38.022732 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f9d7094326fecb094c30501623057ad8a358a8b75efbde65e9c1837200d81a04"} err="failed to get container status \"f9d7094326fecb094c30501623057ad8a358a8b75efbde65e9c1837200d81a04\": rpc error: code = NotFound desc = could not find container \"f9d7094326fecb094c30501623057ad8a358a8b75efbde65e9c1837200d81a04\": container with ID starting with f9d7094326fecb094c30501623057ad8a358a8b75efbde65e9c1837200d81a04 not found: ID does not exist" Oct 05 07:08:38 crc kubenswrapper[4846]: I1005 07:08:38.136331 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 05 07:08:38 crc kubenswrapper[4846]: I1005 07:08:38.165230 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 05 07:08:38 crc kubenswrapper[4846]: I1005 07:08:38.178703 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 05 07:08:38 crc kubenswrapper[4846]: E1005 07:08:38.179235 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="671f4c86-d748-474a-b0f0-400b16b28d7f" containerName="ceilometer-notification-agent" Oct 05 07:08:38 crc kubenswrapper[4846]: I1005 07:08:38.179259 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="671f4c86-d748-474a-b0f0-400b16b28d7f" containerName="ceilometer-notification-agent" Oct 05 07:08:38 crc kubenswrapper[4846]: E1005 07:08:38.179280 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="671f4c86-d748-474a-b0f0-400b16b28d7f" containerName="sg-core" Oct 05 07:08:38 crc kubenswrapper[4846]: I1005 07:08:38.179288 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="671f4c86-d748-474a-b0f0-400b16b28d7f" containerName="sg-core" Oct 05 07:08:38 crc kubenswrapper[4846]: E1005 07:08:38.179314 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="671f4c86-d748-474a-b0f0-400b16b28d7f" containerName="ceilometer-central-agent" Oct 05 07:08:38 crc kubenswrapper[4846]: I1005 07:08:38.179322 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="671f4c86-d748-474a-b0f0-400b16b28d7f" containerName="ceilometer-central-agent" Oct 05 07:08:38 crc kubenswrapper[4846]: E1005 07:08:38.179337 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="671f4c86-d748-474a-b0f0-400b16b28d7f" containerName="proxy-httpd" Oct 05 07:08:38 crc kubenswrapper[4846]: I1005 07:08:38.179343 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="671f4c86-d748-474a-b0f0-400b16b28d7f" containerName="proxy-httpd" Oct 05 07:08:38 crc kubenswrapper[4846]: I1005 07:08:38.179549 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="671f4c86-d748-474a-b0f0-400b16b28d7f" containerName="ceilometer-central-agent" Oct 05 07:08:38 crc kubenswrapper[4846]: I1005 07:08:38.179565 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="671f4c86-d748-474a-b0f0-400b16b28d7f" containerName="ceilometer-notification-agent" Oct 05 07:08:38 crc kubenswrapper[4846]: I1005 07:08:38.179577 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="671f4c86-d748-474a-b0f0-400b16b28d7f" containerName="proxy-httpd" Oct 05 07:08:38 crc kubenswrapper[4846]: I1005 07:08:38.179593 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="671f4c86-d748-474a-b0f0-400b16b28d7f" containerName="sg-core" Oct 05 07:08:38 crc kubenswrapper[4846]: I1005 07:08:38.188796 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 05 07:08:38 crc kubenswrapper[4846]: I1005 07:08:38.196010 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 05 07:08:38 crc kubenswrapper[4846]: I1005 07:08:38.196303 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 05 07:08:38 crc kubenswrapper[4846]: I1005 07:08:38.216496 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 05 07:08:38 crc kubenswrapper[4846]: I1005 07:08:38.363065 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nlbtk\" (UniqueName: \"kubernetes.io/projected/0c1e73f6-0b4e-46ee-9ea1-1b7e52226cf3-kube-api-access-nlbtk\") pod \"ceilometer-0\" (UID: \"0c1e73f6-0b4e-46ee-9ea1-1b7e52226cf3\") " pod="openstack/ceilometer-0" Oct 05 07:08:38 crc kubenswrapper[4846]: I1005 07:08:38.363142 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0c1e73f6-0b4e-46ee-9ea1-1b7e52226cf3-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"0c1e73f6-0b4e-46ee-9ea1-1b7e52226cf3\") " pod="openstack/ceilometer-0" Oct 05 07:08:38 crc kubenswrapper[4846]: I1005 07:08:38.363198 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0c1e73f6-0b4e-46ee-9ea1-1b7e52226cf3-log-httpd\") pod \"ceilometer-0\" (UID: \"0c1e73f6-0b4e-46ee-9ea1-1b7e52226cf3\") " pod="openstack/ceilometer-0" Oct 05 07:08:38 crc kubenswrapper[4846]: I1005 07:08:38.363251 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0c1e73f6-0b4e-46ee-9ea1-1b7e52226cf3-run-httpd\") pod \"ceilometer-0\" (UID: \"0c1e73f6-0b4e-46ee-9ea1-1b7e52226cf3\") " pod="openstack/ceilometer-0" Oct 05 07:08:38 crc kubenswrapper[4846]: I1005 07:08:38.363280 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0c1e73f6-0b4e-46ee-9ea1-1b7e52226cf3-config-data\") pod \"ceilometer-0\" (UID: \"0c1e73f6-0b4e-46ee-9ea1-1b7e52226cf3\") " pod="openstack/ceilometer-0" Oct 05 07:08:38 crc kubenswrapper[4846]: I1005 07:08:38.363309 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0c1e73f6-0b4e-46ee-9ea1-1b7e52226cf3-scripts\") pod \"ceilometer-0\" (UID: \"0c1e73f6-0b4e-46ee-9ea1-1b7e52226cf3\") " pod="openstack/ceilometer-0" Oct 05 07:08:38 crc kubenswrapper[4846]: I1005 07:08:38.363357 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0c1e73f6-0b4e-46ee-9ea1-1b7e52226cf3-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"0c1e73f6-0b4e-46ee-9ea1-1b7e52226cf3\") " pod="openstack/ceilometer-0" Oct 05 07:08:38 crc kubenswrapper[4846]: I1005 07:08:38.466238 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0c1e73f6-0b4e-46ee-9ea1-1b7e52226cf3-scripts\") pod \"ceilometer-0\" (UID: \"0c1e73f6-0b4e-46ee-9ea1-1b7e52226cf3\") " pod="openstack/ceilometer-0" Oct 05 07:08:38 crc kubenswrapper[4846]: I1005 07:08:38.466390 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0c1e73f6-0b4e-46ee-9ea1-1b7e52226cf3-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"0c1e73f6-0b4e-46ee-9ea1-1b7e52226cf3\") " pod="openstack/ceilometer-0" Oct 05 07:08:38 crc kubenswrapper[4846]: I1005 07:08:38.466497 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nlbtk\" (UniqueName: \"kubernetes.io/projected/0c1e73f6-0b4e-46ee-9ea1-1b7e52226cf3-kube-api-access-nlbtk\") pod \"ceilometer-0\" (UID: \"0c1e73f6-0b4e-46ee-9ea1-1b7e52226cf3\") " pod="openstack/ceilometer-0" Oct 05 07:08:38 crc kubenswrapper[4846]: I1005 07:08:38.466533 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0c1e73f6-0b4e-46ee-9ea1-1b7e52226cf3-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"0c1e73f6-0b4e-46ee-9ea1-1b7e52226cf3\") " pod="openstack/ceilometer-0" Oct 05 07:08:38 crc kubenswrapper[4846]: I1005 07:08:38.466564 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0c1e73f6-0b4e-46ee-9ea1-1b7e52226cf3-log-httpd\") pod \"ceilometer-0\" (UID: \"0c1e73f6-0b4e-46ee-9ea1-1b7e52226cf3\") " pod="openstack/ceilometer-0" Oct 05 07:08:38 crc kubenswrapper[4846]: I1005 07:08:38.466613 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0c1e73f6-0b4e-46ee-9ea1-1b7e52226cf3-run-httpd\") pod \"ceilometer-0\" (UID: \"0c1e73f6-0b4e-46ee-9ea1-1b7e52226cf3\") " pod="openstack/ceilometer-0" Oct 05 07:08:38 crc kubenswrapper[4846]: I1005 07:08:38.466635 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0c1e73f6-0b4e-46ee-9ea1-1b7e52226cf3-config-data\") pod \"ceilometer-0\" (UID: \"0c1e73f6-0b4e-46ee-9ea1-1b7e52226cf3\") " pod="openstack/ceilometer-0" Oct 05 07:08:38 crc kubenswrapper[4846]: I1005 07:08:38.468605 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0c1e73f6-0b4e-46ee-9ea1-1b7e52226cf3-run-httpd\") pod \"ceilometer-0\" (UID: \"0c1e73f6-0b4e-46ee-9ea1-1b7e52226cf3\") " pod="openstack/ceilometer-0" Oct 05 07:08:38 crc kubenswrapper[4846]: I1005 07:08:38.469090 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0c1e73f6-0b4e-46ee-9ea1-1b7e52226cf3-log-httpd\") pod \"ceilometer-0\" (UID: \"0c1e73f6-0b4e-46ee-9ea1-1b7e52226cf3\") " pod="openstack/ceilometer-0" Oct 05 07:08:38 crc kubenswrapper[4846]: I1005 07:08:38.474968 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0c1e73f6-0b4e-46ee-9ea1-1b7e52226cf3-scripts\") pod \"ceilometer-0\" (UID: \"0c1e73f6-0b4e-46ee-9ea1-1b7e52226cf3\") " pod="openstack/ceilometer-0" Oct 05 07:08:38 crc kubenswrapper[4846]: I1005 07:08:38.483894 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0c1e73f6-0b4e-46ee-9ea1-1b7e52226cf3-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"0c1e73f6-0b4e-46ee-9ea1-1b7e52226cf3\") " pod="openstack/ceilometer-0" Oct 05 07:08:38 crc kubenswrapper[4846]: I1005 07:08:38.484591 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0c1e73f6-0b4e-46ee-9ea1-1b7e52226cf3-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"0c1e73f6-0b4e-46ee-9ea1-1b7e52226cf3\") " pod="openstack/ceilometer-0" Oct 05 07:08:38 crc kubenswrapper[4846]: I1005 07:08:38.484971 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0c1e73f6-0b4e-46ee-9ea1-1b7e52226cf3-config-data\") pod \"ceilometer-0\" (UID: \"0c1e73f6-0b4e-46ee-9ea1-1b7e52226cf3\") " pod="openstack/ceilometer-0" Oct 05 07:08:38 crc kubenswrapper[4846]: I1005 07:08:38.488903 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nlbtk\" (UniqueName: \"kubernetes.io/projected/0c1e73f6-0b4e-46ee-9ea1-1b7e52226cf3-kube-api-access-nlbtk\") pod \"ceilometer-0\" (UID: \"0c1e73f6-0b4e-46ee-9ea1-1b7e52226cf3\") " pod="openstack/ceilometer-0" Oct 05 07:08:38 crc kubenswrapper[4846]: I1005 07:08:38.518340 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="671f4c86-d748-474a-b0f0-400b16b28d7f" path="/var/lib/kubelet/pods/671f4c86-d748-474a-b0f0-400b16b28d7f/volumes" Oct 05 07:08:38 crc kubenswrapper[4846]: I1005 07:08:38.526111 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 05 07:08:39 crc kubenswrapper[4846]: I1005 07:08:39.042747 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 05 07:08:39 crc kubenswrapper[4846]: I1005 07:08:39.091363 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Oct 05 07:08:39 crc kubenswrapper[4846]: I1005 07:08:39.091553 4846 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 05 07:08:39 crc kubenswrapper[4846]: I1005 07:08:39.164876 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Oct 05 07:08:39 crc kubenswrapper[4846]: I1005 07:08:39.862773 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0c1e73f6-0b4e-46ee-9ea1-1b7e52226cf3","Type":"ContainerStarted","Data":"d55b6e1ab7f73560911aa0d1f991fead109d800cc85fcf5ec3cd1ac2f93c693c"} Oct 05 07:08:39 crc kubenswrapper[4846]: I1005 07:08:39.965603 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-db-create-4mqhm"] Oct 05 07:08:39 crc kubenswrapper[4846]: I1005 07:08:39.967518 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-4mqhm" Oct 05 07:08:39 crc kubenswrapper[4846]: I1005 07:08:39.994145 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-4mqhm"] Oct 05 07:08:40 crc kubenswrapper[4846]: I1005 07:08:40.076521 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-db-create-hpmx9"] Oct 05 07:08:40 crc kubenswrapper[4846]: I1005 07:08:40.080371 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-hpmx9" Oct 05 07:08:40 crc kubenswrapper[4846]: I1005 07:08:40.088255 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-hpmx9"] Oct 05 07:08:40 crc kubenswrapper[4846]: I1005 07:08:40.114599 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-54g2d\" (UniqueName: \"kubernetes.io/projected/514465f9-8432-469d-9260-0958f48f774f-kube-api-access-54g2d\") pod \"nova-api-db-create-4mqhm\" (UID: \"514465f9-8432-469d-9260-0958f48f774f\") " pod="openstack/nova-api-db-create-4mqhm" Oct 05 07:08:40 crc kubenswrapper[4846]: I1005 07:08:40.160061 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-db-create-8s7qn"] Oct 05 07:08:40 crc kubenswrapper[4846]: I1005 07:08:40.163740 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-8s7qn" Oct 05 07:08:40 crc kubenswrapper[4846]: I1005 07:08:40.170421 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-8s7qn"] Oct 05 07:08:40 crc kubenswrapper[4846]: I1005 07:08:40.216536 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vn7r5\" (UniqueName: \"kubernetes.io/projected/d0ccbe35-d3da-4f0d-8c3f-66c016c4604a-kube-api-access-vn7r5\") pod \"nova-cell0-db-create-hpmx9\" (UID: \"d0ccbe35-d3da-4f0d-8c3f-66c016c4604a\") " pod="openstack/nova-cell0-db-create-hpmx9" Oct 05 07:08:40 crc kubenswrapper[4846]: I1005 07:08:40.216677 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qlprc\" (UniqueName: \"kubernetes.io/projected/483b7e9a-7db4-4137-b505-d2b6c404ba06-kube-api-access-qlprc\") pod \"nova-cell1-db-create-8s7qn\" (UID: \"483b7e9a-7db4-4137-b505-d2b6c404ba06\") " pod="openstack/nova-cell1-db-create-8s7qn" Oct 05 07:08:40 crc kubenswrapper[4846]: I1005 07:08:40.216721 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-54g2d\" (UniqueName: \"kubernetes.io/projected/514465f9-8432-469d-9260-0958f48f774f-kube-api-access-54g2d\") pod \"nova-api-db-create-4mqhm\" (UID: \"514465f9-8432-469d-9260-0958f48f774f\") " pod="openstack/nova-api-db-create-4mqhm" Oct 05 07:08:40 crc kubenswrapper[4846]: I1005 07:08:40.246074 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-54g2d\" (UniqueName: \"kubernetes.io/projected/514465f9-8432-469d-9260-0958f48f774f-kube-api-access-54g2d\") pod \"nova-api-db-create-4mqhm\" (UID: \"514465f9-8432-469d-9260-0958f48f774f\") " pod="openstack/nova-api-db-create-4mqhm" Oct 05 07:08:40 crc kubenswrapper[4846]: I1005 07:08:40.296749 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-4mqhm" Oct 05 07:08:40 crc kubenswrapper[4846]: I1005 07:08:40.319274 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vn7r5\" (UniqueName: \"kubernetes.io/projected/d0ccbe35-d3da-4f0d-8c3f-66c016c4604a-kube-api-access-vn7r5\") pod \"nova-cell0-db-create-hpmx9\" (UID: \"d0ccbe35-d3da-4f0d-8c3f-66c016c4604a\") " pod="openstack/nova-cell0-db-create-hpmx9" Oct 05 07:08:40 crc kubenswrapper[4846]: I1005 07:08:40.320096 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qlprc\" (UniqueName: \"kubernetes.io/projected/483b7e9a-7db4-4137-b505-d2b6c404ba06-kube-api-access-qlprc\") pod \"nova-cell1-db-create-8s7qn\" (UID: \"483b7e9a-7db4-4137-b505-d2b6c404ba06\") " pod="openstack/nova-cell1-db-create-8s7qn" Oct 05 07:08:40 crc kubenswrapper[4846]: I1005 07:08:40.353926 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qlprc\" (UniqueName: \"kubernetes.io/projected/483b7e9a-7db4-4137-b505-d2b6c404ba06-kube-api-access-qlprc\") pod \"nova-cell1-db-create-8s7qn\" (UID: \"483b7e9a-7db4-4137-b505-d2b6c404ba06\") " pod="openstack/nova-cell1-db-create-8s7qn" Oct 05 07:08:40 crc kubenswrapper[4846]: I1005 07:08:40.384254 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vn7r5\" (UniqueName: \"kubernetes.io/projected/d0ccbe35-d3da-4f0d-8c3f-66c016c4604a-kube-api-access-vn7r5\") pod \"nova-cell0-db-create-hpmx9\" (UID: \"d0ccbe35-d3da-4f0d-8c3f-66c016c4604a\") " pod="openstack/nova-cell0-db-create-hpmx9" Oct 05 07:08:40 crc kubenswrapper[4846]: I1005 07:08:40.406637 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-hpmx9" Oct 05 07:08:40 crc kubenswrapper[4846]: I1005 07:08:40.486749 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-8s7qn" Oct 05 07:08:40 crc kubenswrapper[4846]: W1005 07:08:40.863013 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod514465f9_8432_469d_9260_0958f48f774f.slice/crio-1b900da68268b1f424cd42b3bc4ad4839452b4934487f798b36da98f112ff12a WatchSource:0}: Error finding container 1b900da68268b1f424cd42b3bc4ad4839452b4934487f798b36da98f112ff12a: Status 404 returned error can't find the container with id 1b900da68268b1f424cd42b3bc4ad4839452b4934487f798b36da98f112ff12a Oct 05 07:08:40 crc kubenswrapper[4846]: I1005 07:08:40.883222 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-4mqhm"] Oct 05 07:08:40 crc kubenswrapper[4846]: I1005 07:08:40.884782 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0c1e73f6-0b4e-46ee-9ea1-1b7e52226cf3","Type":"ContainerStarted","Data":"f0ab84da10dbec024f726d79920ee70f2c8e4645eb5c9333660cdff7a7e06945"} Oct 05 07:08:41 crc kubenswrapper[4846]: I1005 07:08:41.047219 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-hpmx9"] Oct 05 07:08:41 crc kubenswrapper[4846]: I1005 07:08:41.079429 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-8s7qn"] Oct 05 07:08:41 crc kubenswrapper[4846]: I1005 07:08:41.898857 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0c1e73f6-0b4e-46ee-9ea1-1b7e52226cf3","Type":"ContainerStarted","Data":"58064da9228d387abae4a368984ba205dc2577f396cc68d7b4ae2a3f8b25bc4a"} Oct 05 07:08:41 crc kubenswrapper[4846]: I1005 07:08:41.899424 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0c1e73f6-0b4e-46ee-9ea1-1b7e52226cf3","Type":"ContainerStarted","Data":"2011a6bd3f315187706c9865c5eff9573c35a07876057f68548f0732a4adbcdb"} Oct 05 07:08:41 crc kubenswrapper[4846]: I1005 07:08:41.907532 4846 generic.go:334] "Generic (PLEG): container finished" podID="d0ccbe35-d3da-4f0d-8c3f-66c016c4604a" containerID="49603cc789b1bc6ca79b06b68c268619d72a726745c51b51cb8b353e41fcf6b7" exitCode=0 Oct 05 07:08:41 crc kubenswrapper[4846]: I1005 07:08:41.907614 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-hpmx9" event={"ID":"d0ccbe35-d3da-4f0d-8c3f-66c016c4604a","Type":"ContainerDied","Data":"49603cc789b1bc6ca79b06b68c268619d72a726745c51b51cb8b353e41fcf6b7"} Oct 05 07:08:41 crc kubenswrapper[4846]: I1005 07:08:41.907644 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-hpmx9" event={"ID":"d0ccbe35-d3da-4f0d-8c3f-66c016c4604a","Type":"ContainerStarted","Data":"e135acd284e07c4374b23dbe8a2dc5664c74f712b08385077e987d369fd47bed"} Oct 05 07:08:41 crc kubenswrapper[4846]: I1005 07:08:41.912980 4846 generic.go:334] "Generic (PLEG): container finished" podID="514465f9-8432-469d-9260-0958f48f774f" containerID="3cf10e313f2fcf0243e3ffd0ebcf570c10c8a178135588f4b07bf18a38f012e1" exitCode=0 Oct 05 07:08:41 crc kubenswrapper[4846]: I1005 07:08:41.913044 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-4mqhm" event={"ID":"514465f9-8432-469d-9260-0958f48f774f","Type":"ContainerDied","Data":"3cf10e313f2fcf0243e3ffd0ebcf570c10c8a178135588f4b07bf18a38f012e1"} Oct 05 07:08:41 crc kubenswrapper[4846]: I1005 07:08:41.913070 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-4mqhm" event={"ID":"514465f9-8432-469d-9260-0958f48f774f","Type":"ContainerStarted","Data":"1b900da68268b1f424cd42b3bc4ad4839452b4934487f798b36da98f112ff12a"} Oct 05 07:08:41 crc kubenswrapper[4846]: I1005 07:08:41.916028 4846 generic.go:334] "Generic (PLEG): container finished" podID="483b7e9a-7db4-4137-b505-d2b6c404ba06" containerID="ffa470fa54fc1c858cb186b1545ff7805051d894cde1fb62d09980a0f72da492" exitCode=0 Oct 05 07:08:41 crc kubenswrapper[4846]: I1005 07:08:41.916086 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-8s7qn" event={"ID":"483b7e9a-7db4-4137-b505-d2b6c404ba06","Type":"ContainerDied","Data":"ffa470fa54fc1c858cb186b1545ff7805051d894cde1fb62d09980a0f72da492"} Oct 05 07:08:41 crc kubenswrapper[4846]: I1005 07:08:41.916122 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-8s7qn" event={"ID":"483b7e9a-7db4-4137-b505-d2b6c404ba06","Type":"ContainerStarted","Data":"78cb1df8dc79bfbc3431ac02e902f7f63c53917842c1c7fc85ce3cd294da5180"} Oct 05 07:08:42 crc kubenswrapper[4846]: I1005 07:08:42.931989 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0c1e73f6-0b4e-46ee-9ea1-1b7e52226cf3","Type":"ContainerStarted","Data":"4ecc42b3ffd32b9aafddde8e57e6913de0371f1206218ee0aabf2c1631bf739e"} Oct 05 07:08:42 crc kubenswrapper[4846]: I1005 07:08:42.970867 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.420954805 podStartE2EDuration="4.970837844s" podCreationTimestamp="2025-10-05 07:08:38 +0000 UTC" firstStartedPulling="2025-10-05 07:08:39.061427165 +0000 UTC m=+1241.302279940" lastFinishedPulling="2025-10-05 07:08:42.611310204 +0000 UTC m=+1244.852162979" observedRunningTime="2025-10-05 07:08:42.961460944 +0000 UTC m=+1245.202313749" watchObservedRunningTime="2025-10-05 07:08:42.970837844 +0000 UTC m=+1245.211690619" Oct 05 07:08:43 crc kubenswrapper[4846]: I1005 07:08:43.416031 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-hpmx9" Oct 05 07:08:43 crc kubenswrapper[4846]: I1005 07:08:43.492110 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vn7r5\" (UniqueName: \"kubernetes.io/projected/d0ccbe35-d3da-4f0d-8c3f-66c016c4604a-kube-api-access-vn7r5\") pod \"d0ccbe35-d3da-4f0d-8c3f-66c016c4604a\" (UID: \"d0ccbe35-d3da-4f0d-8c3f-66c016c4604a\") " Oct 05 07:08:43 crc kubenswrapper[4846]: I1005 07:08:43.593466 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d0ccbe35-d3da-4f0d-8c3f-66c016c4604a-kube-api-access-vn7r5" (OuterVolumeSpecName: "kube-api-access-vn7r5") pod "d0ccbe35-d3da-4f0d-8c3f-66c016c4604a" (UID: "d0ccbe35-d3da-4f0d-8c3f-66c016c4604a"). InnerVolumeSpecName "kube-api-access-vn7r5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:08:43 crc kubenswrapper[4846]: I1005 07:08:43.635353 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vn7r5\" (UniqueName: \"kubernetes.io/projected/d0ccbe35-d3da-4f0d-8c3f-66c016c4604a-kube-api-access-vn7r5\") on node \"crc\" DevicePath \"\"" Oct 05 07:08:43 crc kubenswrapper[4846]: I1005 07:08:43.716335 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-8s7qn" Oct 05 07:08:43 crc kubenswrapper[4846]: I1005 07:08:43.725634 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-4mqhm" Oct 05 07:08:43 crc kubenswrapper[4846]: I1005 07:08:43.839762 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-54g2d\" (UniqueName: \"kubernetes.io/projected/514465f9-8432-469d-9260-0958f48f774f-kube-api-access-54g2d\") pod \"514465f9-8432-469d-9260-0958f48f774f\" (UID: \"514465f9-8432-469d-9260-0958f48f774f\") " Oct 05 07:08:43 crc kubenswrapper[4846]: I1005 07:08:43.840188 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qlprc\" (UniqueName: \"kubernetes.io/projected/483b7e9a-7db4-4137-b505-d2b6c404ba06-kube-api-access-qlprc\") pod \"483b7e9a-7db4-4137-b505-d2b6c404ba06\" (UID: \"483b7e9a-7db4-4137-b505-d2b6c404ba06\") " Oct 05 07:08:43 crc kubenswrapper[4846]: I1005 07:08:43.844318 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/514465f9-8432-469d-9260-0958f48f774f-kube-api-access-54g2d" (OuterVolumeSpecName: "kube-api-access-54g2d") pod "514465f9-8432-469d-9260-0958f48f774f" (UID: "514465f9-8432-469d-9260-0958f48f774f"). InnerVolumeSpecName "kube-api-access-54g2d". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:08:43 crc kubenswrapper[4846]: I1005 07:08:43.845106 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/483b7e9a-7db4-4137-b505-d2b6c404ba06-kube-api-access-qlprc" (OuterVolumeSpecName: "kube-api-access-qlprc") pod "483b7e9a-7db4-4137-b505-d2b6c404ba06" (UID: "483b7e9a-7db4-4137-b505-d2b6c404ba06"). InnerVolumeSpecName "kube-api-access-qlprc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:08:43 crc kubenswrapper[4846]: I1005 07:08:43.943490 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-54g2d\" (UniqueName: \"kubernetes.io/projected/514465f9-8432-469d-9260-0958f48f774f-kube-api-access-54g2d\") on node \"crc\" DevicePath \"\"" Oct 05 07:08:43 crc kubenswrapper[4846]: I1005 07:08:43.943524 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qlprc\" (UniqueName: \"kubernetes.io/projected/483b7e9a-7db4-4137-b505-d2b6c404ba06-kube-api-access-qlprc\") on node \"crc\" DevicePath \"\"" Oct 05 07:08:43 crc kubenswrapper[4846]: I1005 07:08:43.945980 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-8s7qn" event={"ID":"483b7e9a-7db4-4137-b505-d2b6c404ba06","Type":"ContainerDied","Data":"78cb1df8dc79bfbc3431ac02e902f7f63c53917842c1c7fc85ce3cd294da5180"} Oct 05 07:08:43 crc kubenswrapper[4846]: I1005 07:08:43.946164 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="78cb1df8dc79bfbc3431ac02e902f7f63c53917842c1c7fc85ce3cd294da5180" Oct 05 07:08:43 crc kubenswrapper[4846]: I1005 07:08:43.946492 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-8s7qn" Oct 05 07:08:43 crc kubenswrapper[4846]: I1005 07:08:43.948360 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-hpmx9" event={"ID":"d0ccbe35-d3da-4f0d-8c3f-66c016c4604a","Type":"ContainerDied","Data":"e135acd284e07c4374b23dbe8a2dc5664c74f712b08385077e987d369fd47bed"} Oct 05 07:08:43 crc kubenswrapper[4846]: I1005 07:08:43.948475 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e135acd284e07c4374b23dbe8a2dc5664c74f712b08385077e987d369fd47bed" Oct 05 07:08:43 crc kubenswrapper[4846]: I1005 07:08:43.948394 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-hpmx9" Oct 05 07:08:43 crc kubenswrapper[4846]: I1005 07:08:43.950418 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-4mqhm" event={"ID":"514465f9-8432-469d-9260-0958f48f774f","Type":"ContainerDied","Data":"1b900da68268b1f424cd42b3bc4ad4839452b4934487f798b36da98f112ff12a"} Oct 05 07:08:43 crc kubenswrapper[4846]: I1005 07:08:43.950540 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1b900da68268b1f424cd42b3bc4ad4839452b4934487f798b36da98f112ff12a" Oct 05 07:08:43 crc kubenswrapper[4846]: I1005 07:08:43.950430 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-4mqhm" Oct 05 07:08:43 crc kubenswrapper[4846]: I1005 07:08:43.958819 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 05 07:08:45 crc kubenswrapper[4846]: I1005 07:08:45.161966 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Oct 05 07:08:45 crc kubenswrapper[4846]: I1005 07:08:45.163496 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Oct 05 07:08:45 crc kubenswrapper[4846]: I1005 07:08:45.201243 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Oct 05 07:08:45 crc kubenswrapper[4846]: I1005 07:08:45.207920 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Oct 05 07:08:45 crc kubenswrapper[4846]: I1005 07:08:45.968742 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Oct 05 07:08:45 crc kubenswrapper[4846]: I1005 07:08:45.969098 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Oct 05 07:08:47 crc kubenswrapper[4846]: I1005 07:08:47.917023 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Oct 05 07:08:47 crc kubenswrapper[4846]: I1005 07:08:47.924995 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Oct 05 07:08:50 crc kubenswrapper[4846]: I1005 07:08:50.039071 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-7d67-account-create-v5k6p"] Oct 05 07:08:50 crc kubenswrapper[4846]: E1005 07:08:50.039912 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="483b7e9a-7db4-4137-b505-d2b6c404ba06" containerName="mariadb-database-create" Oct 05 07:08:50 crc kubenswrapper[4846]: I1005 07:08:50.039927 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="483b7e9a-7db4-4137-b505-d2b6c404ba06" containerName="mariadb-database-create" Oct 05 07:08:50 crc kubenswrapper[4846]: E1005 07:08:50.039950 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d0ccbe35-d3da-4f0d-8c3f-66c016c4604a" containerName="mariadb-database-create" Oct 05 07:08:50 crc kubenswrapper[4846]: I1005 07:08:50.039956 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="d0ccbe35-d3da-4f0d-8c3f-66c016c4604a" containerName="mariadb-database-create" Oct 05 07:08:50 crc kubenswrapper[4846]: E1005 07:08:50.039977 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="514465f9-8432-469d-9260-0958f48f774f" containerName="mariadb-database-create" Oct 05 07:08:50 crc kubenswrapper[4846]: I1005 07:08:50.039983 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="514465f9-8432-469d-9260-0958f48f774f" containerName="mariadb-database-create" Oct 05 07:08:50 crc kubenswrapper[4846]: I1005 07:08:50.040209 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="514465f9-8432-469d-9260-0958f48f774f" containerName="mariadb-database-create" Oct 05 07:08:50 crc kubenswrapper[4846]: I1005 07:08:50.040229 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="d0ccbe35-d3da-4f0d-8c3f-66c016c4604a" containerName="mariadb-database-create" Oct 05 07:08:50 crc kubenswrapper[4846]: I1005 07:08:50.040243 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="483b7e9a-7db4-4137-b505-d2b6c404ba06" containerName="mariadb-database-create" Oct 05 07:08:50 crc kubenswrapper[4846]: I1005 07:08:50.040956 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-7d67-account-create-v5k6p" Oct 05 07:08:50 crc kubenswrapper[4846]: I1005 07:08:50.044040 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-db-secret" Oct 05 07:08:50 crc kubenswrapper[4846]: I1005 07:08:50.056496 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-7d67-account-create-v5k6p"] Oct 05 07:08:50 crc kubenswrapper[4846]: I1005 07:08:50.090065 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mr7hj\" (UniqueName: \"kubernetes.io/projected/04602b97-f7db-4bd6-98ce-bd066dbb21f5-kube-api-access-mr7hj\") pod \"nova-api-7d67-account-create-v5k6p\" (UID: \"04602b97-f7db-4bd6-98ce-bd066dbb21f5\") " pod="openstack/nova-api-7d67-account-create-v5k6p" Oct 05 07:08:50 crc kubenswrapper[4846]: I1005 07:08:50.192849 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mr7hj\" (UniqueName: \"kubernetes.io/projected/04602b97-f7db-4bd6-98ce-bd066dbb21f5-kube-api-access-mr7hj\") pod \"nova-api-7d67-account-create-v5k6p\" (UID: \"04602b97-f7db-4bd6-98ce-bd066dbb21f5\") " pod="openstack/nova-api-7d67-account-create-v5k6p" Oct 05 07:08:50 crc kubenswrapper[4846]: I1005 07:08:50.220743 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-7ebc-account-create-k72ft"] Oct 05 07:08:50 crc kubenswrapper[4846]: I1005 07:08:50.223009 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-7ebc-account-create-k72ft" Oct 05 07:08:50 crc kubenswrapper[4846]: I1005 07:08:50.228105 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-db-secret" Oct 05 07:08:50 crc kubenswrapper[4846]: I1005 07:08:50.233155 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-7ebc-account-create-k72ft"] Oct 05 07:08:50 crc kubenswrapper[4846]: I1005 07:08:50.252648 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mr7hj\" (UniqueName: \"kubernetes.io/projected/04602b97-f7db-4bd6-98ce-bd066dbb21f5-kube-api-access-mr7hj\") pod \"nova-api-7d67-account-create-v5k6p\" (UID: \"04602b97-f7db-4bd6-98ce-bd066dbb21f5\") " pod="openstack/nova-api-7d67-account-create-v5k6p" Oct 05 07:08:50 crc kubenswrapper[4846]: I1005 07:08:50.297132 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6rxmj\" (UniqueName: \"kubernetes.io/projected/a63045aa-bdd8-4b7b-954e-73e7799b5859-kube-api-access-6rxmj\") pod \"nova-cell0-7ebc-account-create-k72ft\" (UID: \"a63045aa-bdd8-4b7b-954e-73e7799b5859\") " pod="openstack/nova-cell0-7ebc-account-create-k72ft" Oct 05 07:08:50 crc kubenswrapper[4846]: I1005 07:08:50.321812 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-7e7d-account-create-bcwnf"] Oct 05 07:08:50 crc kubenswrapper[4846]: I1005 07:08:50.323250 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-7e7d-account-create-bcwnf" Oct 05 07:08:50 crc kubenswrapper[4846]: I1005 07:08:50.330885 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-db-secret" Oct 05 07:08:50 crc kubenswrapper[4846]: I1005 07:08:50.358092 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-7e7d-account-create-bcwnf"] Oct 05 07:08:50 crc kubenswrapper[4846]: I1005 07:08:50.362800 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-7d67-account-create-v5k6p" Oct 05 07:08:50 crc kubenswrapper[4846]: I1005 07:08:50.399652 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6rxmj\" (UniqueName: \"kubernetes.io/projected/a63045aa-bdd8-4b7b-954e-73e7799b5859-kube-api-access-6rxmj\") pod \"nova-cell0-7ebc-account-create-k72ft\" (UID: \"a63045aa-bdd8-4b7b-954e-73e7799b5859\") " pod="openstack/nova-cell0-7ebc-account-create-k72ft" Oct 05 07:08:50 crc kubenswrapper[4846]: I1005 07:08:50.399776 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fzrdf\" (UniqueName: \"kubernetes.io/projected/fb30bf11-366c-45af-87fd-254f8eca710d-kube-api-access-fzrdf\") pod \"nova-cell1-7e7d-account-create-bcwnf\" (UID: \"fb30bf11-366c-45af-87fd-254f8eca710d\") " pod="openstack/nova-cell1-7e7d-account-create-bcwnf" Oct 05 07:08:50 crc kubenswrapper[4846]: I1005 07:08:50.416489 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6rxmj\" (UniqueName: \"kubernetes.io/projected/a63045aa-bdd8-4b7b-954e-73e7799b5859-kube-api-access-6rxmj\") pod \"nova-cell0-7ebc-account-create-k72ft\" (UID: \"a63045aa-bdd8-4b7b-954e-73e7799b5859\") " pod="openstack/nova-cell0-7ebc-account-create-k72ft" Oct 05 07:08:50 crc kubenswrapper[4846]: I1005 07:08:50.501674 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fzrdf\" (UniqueName: \"kubernetes.io/projected/fb30bf11-366c-45af-87fd-254f8eca710d-kube-api-access-fzrdf\") pod \"nova-cell1-7e7d-account-create-bcwnf\" (UID: \"fb30bf11-366c-45af-87fd-254f8eca710d\") " pod="openstack/nova-cell1-7e7d-account-create-bcwnf" Oct 05 07:08:50 crc kubenswrapper[4846]: I1005 07:08:50.521347 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fzrdf\" (UniqueName: \"kubernetes.io/projected/fb30bf11-366c-45af-87fd-254f8eca710d-kube-api-access-fzrdf\") pod \"nova-cell1-7e7d-account-create-bcwnf\" (UID: \"fb30bf11-366c-45af-87fd-254f8eca710d\") " pod="openstack/nova-cell1-7e7d-account-create-bcwnf" Oct 05 07:08:50 crc kubenswrapper[4846]: I1005 07:08:50.547375 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-7ebc-account-create-k72ft" Oct 05 07:08:50 crc kubenswrapper[4846]: I1005 07:08:50.675077 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-7e7d-account-create-bcwnf" Oct 05 07:08:50 crc kubenswrapper[4846]: I1005 07:08:50.851424 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-7d67-account-create-v5k6p"] Oct 05 07:08:51 crc kubenswrapper[4846]: I1005 07:08:51.029885 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-7ebc-account-create-k72ft"] Oct 05 07:08:51 crc kubenswrapper[4846]: W1005 07:08:51.035472 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda63045aa_bdd8_4b7b_954e_73e7799b5859.slice/crio-2dd4de9c6ded148a3daad9cfe8bb72bda3136d929c1fa731b435d28650ff6f9d WatchSource:0}: Error finding container 2dd4de9c6ded148a3daad9cfe8bb72bda3136d929c1fa731b435d28650ff6f9d: Status 404 returned error can't find the container with id 2dd4de9c6ded148a3daad9cfe8bb72bda3136d929c1fa731b435d28650ff6f9d Oct 05 07:08:51 crc kubenswrapper[4846]: I1005 07:08:51.037094 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-7d67-account-create-v5k6p" event={"ID":"04602b97-f7db-4bd6-98ce-bd066dbb21f5","Type":"ContainerStarted","Data":"2c91a216d7eedcbfbeedde232a7a0a69fd4e0a7bc8920e6800382b25de1f46ec"} Oct 05 07:08:51 crc kubenswrapper[4846]: I1005 07:08:51.144137 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-7e7d-account-create-bcwnf"] Oct 05 07:08:51 crc kubenswrapper[4846]: W1005 07:08:51.165200 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfb30bf11_366c_45af_87fd_254f8eca710d.slice/crio-5cd1637e254c31aeacd3ab8157451076a0ead83bffaa8e931cdc48a80063bc09 WatchSource:0}: Error finding container 5cd1637e254c31aeacd3ab8157451076a0ead83bffaa8e931cdc48a80063bc09: Status 404 returned error can't find the container with id 5cd1637e254c31aeacd3ab8157451076a0ead83bffaa8e931cdc48a80063bc09 Oct 05 07:08:52 crc kubenswrapper[4846]: I1005 07:08:52.053228 4846 generic.go:334] "Generic (PLEG): container finished" podID="fb30bf11-366c-45af-87fd-254f8eca710d" containerID="ec596b676d1e2029a427c619a2c09007c32cd26a7cd5b08b7f797a076c0a061e" exitCode=0 Oct 05 07:08:52 crc kubenswrapper[4846]: I1005 07:08:52.053476 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-7e7d-account-create-bcwnf" event={"ID":"fb30bf11-366c-45af-87fd-254f8eca710d","Type":"ContainerDied","Data":"ec596b676d1e2029a427c619a2c09007c32cd26a7cd5b08b7f797a076c0a061e"} Oct 05 07:08:52 crc kubenswrapper[4846]: I1005 07:08:52.053797 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-7e7d-account-create-bcwnf" event={"ID":"fb30bf11-366c-45af-87fd-254f8eca710d","Type":"ContainerStarted","Data":"5cd1637e254c31aeacd3ab8157451076a0ead83bffaa8e931cdc48a80063bc09"} Oct 05 07:08:52 crc kubenswrapper[4846]: I1005 07:08:52.057780 4846 generic.go:334] "Generic (PLEG): container finished" podID="04602b97-f7db-4bd6-98ce-bd066dbb21f5" containerID="39e9c49a310f7904a188de2c7d263bafd273efb07adc6a0435e84e760f4fde4c" exitCode=0 Oct 05 07:08:52 crc kubenswrapper[4846]: I1005 07:08:52.057866 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-7d67-account-create-v5k6p" event={"ID":"04602b97-f7db-4bd6-98ce-bd066dbb21f5","Type":"ContainerDied","Data":"39e9c49a310f7904a188de2c7d263bafd273efb07adc6a0435e84e760f4fde4c"} Oct 05 07:08:52 crc kubenswrapper[4846]: I1005 07:08:52.062126 4846 generic.go:334] "Generic (PLEG): container finished" podID="a63045aa-bdd8-4b7b-954e-73e7799b5859" containerID="14b045f6d72991175d4f2fa50e9298251749662cf7376add05f11dc4152f31ce" exitCode=0 Oct 05 07:08:52 crc kubenswrapper[4846]: I1005 07:08:52.062246 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-7ebc-account-create-k72ft" event={"ID":"a63045aa-bdd8-4b7b-954e-73e7799b5859","Type":"ContainerDied","Data":"14b045f6d72991175d4f2fa50e9298251749662cf7376add05f11dc4152f31ce"} Oct 05 07:08:52 crc kubenswrapper[4846]: I1005 07:08:52.062377 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-7ebc-account-create-k72ft" event={"ID":"a63045aa-bdd8-4b7b-954e-73e7799b5859","Type":"ContainerStarted","Data":"2dd4de9c6ded148a3daad9cfe8bb72bda3136d929c1fa731b435d28650ff6f9d"} Oct 05 07:08:53 crc kubenswrapper[4846]: I1005 07:08:53.672881 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-7e7d-account-create-bcwnf" Oct 05 07:08:53 crc kubenswrapper[4846]: I1005 07:08:53.700622 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-7ebc-account-create-k72ft" Oct 05 07:08:53 crc kubenswrapper[4846]: I1005 07:08:53.708851 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-7d67-account-create-v5k6p" Oct 05 07:08:53 crc kubenswrapper[4846]: I1005 07:08:53.787720 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mr7hj\" (UniqueName: \"kubernetes.io/projected/04602b97-f7db-4bd6-98ce-bd066dbb21f5-kube-api-access-mr7hj\") pod \"04602b97-f7db-4bd6-98ce-bd066dbb21f5\" (UID: \"04602b97-f7db-4bd6-98ce-bd066dbb21f5\") " Oct 05 07:08:53 crc kubenswrapper[4846]: I1005 07:08:53.787921 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6rxmj\" (UniqueName: \"kubernetes.io/projected/a63045aa-bdd8-4b7b-954e-73e7799b5859-kube-api-access-6rxmj\") pod \"a63045aa-bdd8-4b7b-954e-73e7799b5859\" (UID: \"a63045aa-bdd8-4b7b-954e-73e7799b5859\") " Oct 05 07:08:53 crc kubenswrapper[4846]: I1005 07:08:53.788105 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fzrdf\" (UniqueName: \"kubernetes.io/projected/fb30bf11-366c-45af-87fd-254f8eca710d-kube-api-access-fzrdf\") pod \"fb30bf11-366c-45af-87fd-254f8eca710d\" (UID: \"fb30bf11-366c-45af-87fd-254f8eca710d\") " Oct 05 07:08:53 crc kubenswrapper[4846]: I1005 07:08:53.795875 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/04602b97-f7db-4bd6-98ce-bd066dbb21f5-kube-api-access-mr7hj" (OuterVolumeSpecName: "kube-api-access-mr7hj") pod "04602b97-f7db-4bd6-98ce-bd066dbb21f5" (UID: "04602b97-f7db-4bd6-98ce-bd066dbb21f5"). InnerVolumeSpecName "kube-api-access-mr7hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:08:53 crc kubenswrapper[4846]: I1005 07:08:53.795935 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a63045aa-bdd8-4b7b-954e-73e7799b5859-kube-api-access-6rxmj" (OuterVolumeSpecName: "kube-api-access-6rxmj") pod "a63045aa-bdd8-4b7b-954e-73e7799b5859" (UID: "a63045aa-bdd8-4b7b-954e-73e7799b5859"). InnerVolumeSpecName "kube-api-access-6rxmj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:08:53 crc kubenswrapper[4846]: I1005 07:08:53.795983 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fb30bf11-366c-45af-87fd-254f8eca710d-kube-api-access-fzrdf" (OuterVolumeSpecName: "kube-api-access-fzrdf") pod "fb30bf11-366c-45af-87fd-254f8eca710d" (UID: "fb30bf11-366c-45af-87fd-254f8eca710d"). InnerVolumeSpecName "kube-api-access-fzrdf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:08:53 crc kubenswrapper[4846]: I1005 07:08:53.890812 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mr7hj\" (UniqueName: \"kubernetes.io/projected/04602b97-f7db-4bd6-98ce-bd066dbb21f5-kube-api-access-mr7hj\") on node \"crc\" DevicePath \"\"" Oct 05 07:08:53 crc kubenswrapper[4846]: I1005 07:08:53.890859 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6rxmj\" (UniqueName: \"kubernetes.io/projected/a63045aa-bdd8-4b7b-954e-73e7799b5859-kube-api-access-6rxmj\") on node \"crc\" DevicePath \"\"" Oct 05 07:08:53 crc kubenswrapper[4846]: I1005 07:08:53.890869 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fzrdf\" (UniqueName: \"kubernetes.io/projected/fb30bf11-366c-45af-87fd-254f8eca710d-kube-api-access-fzrdf\") on node \"crc\" DevicePath \"\"" Oct 05 07:08:54 crc kubenswrapper[4846]: I1005 07:08:54.085675 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-7e7d-account-create-bcwnf" event={"ID":"fb30bf11-366c-45af-87fd-254f8eca710d","Type":"ContainerDied","Data":"5cd1637e254c31aeacd3ab8157451076a0ead83bffaa8e931cdc48a80063bc09"} Oct 05 07:08:54 crc kubenswrapper[4846]: I1005 07:08:54.085735 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5cd1637e254c31aeacd3ab8157451076a0ead83bffaa8e931cdc48a80063bc09" Oct 05 07:08:54 crc kubenswrapper[4846]: I1005 07:08:54.085811 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-7e7d-account-create-bcwnf" Oct 05 07:08:54 crc kubenswrapper[4846]: I1005 07:08:54.088519 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-7d67-account-create-v5k6p" event={"ID":"04602b97-f7db-4bd6-98ce-bd066dbb21f5","Type":"ContainerDied","Data":"2c91a216d7eedcbfbeedde232a7a0a69fd4e0a7bc8920e6800382b25de1f46ec"} Oct 05 07:08:54 crc kubenswrapper[4846]: I1005 07:08:54.088580 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2c91a216d7eedcbfbeedde232a7a0a69fd4e0a7bc8920e6800382b25de1f46ec" Oct 05 07:08:54 crc kubenswrapper[4846]: I1005 07:08:54.088599 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-7d67-account-create-v5k6p" Oct 05 07:08:54 crc kubenswrapper[4846]: I1005 07:08:54.090059 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-7ebc-account-create-k72ft" event={"ID":"a63045aa-bdd8-4b7b-954e-73e7799b5859","Type":"ContainerDied","Data":"2dd4de9c6ded148a3daad9cfe8bb72bda3136d929c1fa731b435d28650ff6f9d"} Oct 05 07:08:54 crc kubenswrapper[4846]: I1005 07:08:54.090094 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2dd4de9c6ded148a3daad9cfe8bb72bda3136d929c1fa731b435d28650ff6f9d" Oct 05 07:08:54 crc kubenswrapper[4846]: I1005 07:08:54.090166 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-7ebc-account-create-k72ft" Oct 05 07:08:55 crc kubenswrapper[4846]: I1005 07:08:55.473476 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-db-sync-88lpl"] Oct 05 07:08:55 crc kubenswrapper[4846]: E1005 07:08:55.474337 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a63045aa-bdd8-4b7b-954e-73e7799b5859" containerName="mariadb-account-create" Oct 05 07:08:55 crc kubenswrapper[4846]: I1005 07:08:55.474351 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="a63045aa-bdd8-4b7b-954e-73e7799b5859" containerName="mariadb-account-create" Oct 05 07:08:55 crc kubenswrapper[4846]: E1005 07:08:55.474369 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="04602b97-f7db-4bd6-98ce-bd066dbb21f5" containerName="mariadb-account-create" Oct 05 07:08:55 crc kubenswrapper[4846]: I1005 07:08:55.474375 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="04602b97-f7db-4bd6-98ce-bd066dbb21f5" containerName="mariadb-account-create" Oct 05 07:08:55 crc kubenswrapper[4846]: E1005 07:08:55.474391 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fb30bf11-366c-45af-87fd-254f8eca710d" containerName="mariadb-account-create" Oct 05 07:08:55 crc kubenswrapper[4846]: I1005 07:08:55.474427 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="fb30bf11-366c-45af-87fd-254f8eca710d" containerName="mariadb-account-create" Oct 05 07:08:55 crc kubenswrapper[4846]: I1005 07:08:55.474595 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="a63045aa-bdd8-4b7b-954e-73e7799b5859" containerName="mariadb-account-create" Oct 05 07:08:55 crc kubenswrapper[4846]: I1005 07:08:55.474612 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="04602b97-f7db-4bd6-98ce-bd066dbb21f5" containerName="mariadb-account-create" Oct 05 07:08:55 crc kubenswrapper[4846]: I1005 07:08:55.474632 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="fb30bf11-366c-45af-87fd-254f8eca710d" containerName="mariadb-account-create" Oct 05 07:08:55 crc kubenswrapper[4846]: I1005 07:08:55.475324 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-88lpl" Oct 05 07:08:55 crc kubenswrapper[4846]: I1005 07:08:55.484331 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-scripts" Oct 05 07:08:55 crc kubenswrapper[4846]: I1005 07:08:55.484593 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Oct 05 07:08:55 crc kubenswrapper[4846]: I1005 07:08:55.484799 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-4c97w" Oct 05 07:08:55 crc kubenswrapper[4846]: I1005 07:08:55.507755 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-88lpl"] Oct 05 07:08:55 crc kubenswrapper[4846]: I1005 07:08:55.529785 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7ac2edb5-3ecf-4162-be56-43585390796a-config-data\") pod \"nova-cell0-conductor-db-sync-88lpl\" (UID: \"7ac2edb5-3ecf-4162-be56-43585390796a\") " pod="openstack/nova-cell0-conductor-db-sync-88lpl" Oct 05 07:08:55 crc kubenswrapper[4846]: I1005 07:08:55.529940 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pq68w\" (UniqueName: \"kubernetes.io/projected/7ac2edb5-3ecf-4162-be56-43585390796a-kube-api-access-pq68w\") pod \"nova-cell0-conductor-db-sync-88lpl\" (UID: \"7ac2edb5-3ecf-4162-be56-43585390796a\") " pod="openstack/nova-cell0-conductor-db-sync-88lpl" Oct 05 07:08:55 crc kubenswrapper[4846]: I1005 07:08:55.529970 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7ac2edb5-3ecf-4162-be56-43585390796a-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-88lpl\" (UID: \"7ac2edb5-3ecf-4162-be56-43585390796a\") " pod="openstack/nova-cell0-conductor-db-sync-88lpl" Oct 05 07:08:55 crc kubenswrapper[4846]: I1005 07:08:55.530001 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7ac2edb5-3ecf-4162-be56-43585390796a-scripts\") pod \"nova-cell0-conductor-db-sync-88lpl\" (UID: \"7ac2edb5-3ecf-4162-be56-43585390796a\") " pod="openstack/nova-cell0-conductor-db-sync-88lpl" Oct 05 07:08:55 crc kubenswrapper[4846]: I1005 07:08:55.631328 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7ac2edb5-3ecf-4162-be56-43585390796a-config-data\") pod \"nova-cell0-conductor-db-sync-88lpl\" (UID: \"7ac2edb5-3ecf-4162-be56-43585390796a\") " pod="openstack/nova-cell0-conductor-db-sync-88lpl" Oct 05 07:08:55 crc kubenswrapper[4846]: I1005 07:08:55.631506 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pq68w\" (UniqueName: \"kubernetes.io/projected/7ac2edb5-3ecf-4162-be56-43585390796a-kube-api-access-pq68w\") pod \"nova-cell0-conductor-db-sync-88lpl\" (UID: \"7ac2edb5-3ecf-4162-be56-43585390796a\") " pod="openstack/nova-cell0-conductor-db-sync-88lpl" Oct 05 07:08:55 crc kubenswrapper[4846]: I1005 07:08:55.631542 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7ac2edb5-3ecf-4162-be56-43585390796a-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-88lpl\" (UID: \"7ac2edb5-3ecf-4162-be56-43585390796a\") " pod="openstack/nova-cell0-conductor-db-sync-88lpl" Oct 05 07:08:55 crc kubenswrapper[4846]: I1005 07:08:55.631571 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7ac2edb5-3ecf-4162-be56-43585390796a-scripts\") pod \"nova-cell0-conductor-db-sync-88lpl\" (UID: \"7ac2edb5-3ecf-4162-be56-43585390796a\") " pod="openstack/nova-cell0-conductor-db-sync-88lpl" Oct 05 07:08:55 crc kubenswrapper[4846]: I1005 07:08:55.639194 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7ac2edb5-3ecf-4162-be56-43585390796a-config-data\") pod \"nova-cell0-conductor-db-sync-88lpl\" (UID: \"7ac2edb5-3ecf-4162-be56-43585390796a\") " pod="openstack/nova-cell0-conductor-db-sync-88lpl" Oct 05 07:08:55 crc kubenswrapper[4846]: I1005 07:08:55.641024 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7ac2edb5-3ecf-4162-be56-43585390796a-scripts\") pod \"nova-cell0-conductor-db-sync-88lpl\" (UID: \"7ac2edb5-3ecf-4162-be56-43585390796a\") " pod="openstack/nova-cell0-conductor-db-sync-88lpl" Oct 05 07:08:55 crc kubenswrapper[4846]: I1005 07:08:55.653907 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7ac2edb5-3ecf-4162-be56-43585390796a-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-88lpl\" (UID: \"7ac2edb5-3ecf-4162-be56-43585390796a\") " pod="openstack/nova-cell0-conductor-db-sync-88lpl" Oct 05 07:08:55 crc kubenswrapper[4846]: I1005 07:08:55.655047 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pq68w\" (UniqueName: \"kubernetes.io/projected/7ac2edb5-3ecf-4162-be56-43585390796a-kube-api-access-pq68w\") pod \"nova-cell0-conductor-db-sync-88lpl\" (UID: \"7ac2edb5-3ecf-4162-be56-43585390796a\") " pod="openstack/nova-cell0-conductor-db-sync-88lpl" Oct 05 07:08:55 crc kubenswrapper[4846]: I1005 07:08:55.813865 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-88lpl" Oct 05 07:08:56 crc kubenswrapper[4846]: I1005 07:08:56.354407 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-88lpl"] Oct 05 07:08:56 crc kubenswrapper[4846]: W1005 07:08:56.358921 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7ac2edb5_3ecf_4162_be56_43585390796a.slice/crio-aa3c919abf97da616f69fc85126ad627d72e32b1d7900ee5364241cb5a84442e WatchSource:0}: Error finding container aa3c919abf97da616f69fc85126ad627d72e32b1d7900ee5364241cb5a84442e: Status 404 returned error can't find the container with id aa3c919abf97da616f69fc85126ad627d72e32b1d7900ee5364241cb5a84442e Oct 05 07:08:57 crc kubenswrapper[4846]: I1005 07:08:57.123840 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-88lpl" event={"ID":"7ac2edb5-3ecf-4162-be56-43585390796a","Type":"ContainerStarted","Data":"aa3c919abf97da616f69fc85126ad627d72e32b1d7900ee5364241cb5a84442e"} Oct 05 07:08:57 crc kubenswrapper[4846]: I1005 07:08:57.372636 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 05 07:08:57 crc kubenswrapper[4846]: I1005 07:08:57.373004 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="0c1e73f6-0b4e-46ee-9ea1-1b7e52226cf3" containerName="ceilometer-central-agent" containerID="cri-o://f0ab84da10dbec024f726d79920ee70f2c8e4645eb5c9333660cdff7a7e06945" gracePeriod=30 Oct 05 07:08:57 crc kubenswrapper[4846]: I1005 07:08:57.373617 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="0c1e73f6-0b4e-46ee-9ea1-1b7e52226cf3" containerName="proxy-httpd" containerID="cri-o://4ecc42b3ffd32b9aafddde8e57e6913de0371f1206218ee0aabf2c1631bf739e" gracePeriod=30 Oct 05 07:08:57 crc kubenswrapper[4846]: I1005 07:08:57.373698 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="0c1e73f6-0b4e-46ee-9ea1-1b7e52226cf3" containerName="ceilometer-notification-agent" containerID="cri-o://2011a6bd3f315187706c9865c5eff9573c35a07876057f68548f0732a4adbcdb" gracePeriod=30 Oct 05 07:08:57 crc kubenswrapper[4846]: I1005 07:08:57.373870 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="0c1e73f6-0b4e-46ee-9ea1-1b7e52226cf3" containerName="sg-core" containerID="cri-o://58064da9228d387abae4a368984ba205dc2577f396cc68d7b4ae2a3f8b25bc4a" gracePeriod=30 Oct 05 07:08:57 crc kubenswrapper[4846]: I1005 07:08:57.496372 4846 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="0c1e73f6-0b4e-46ee-9ea1-1b7e52226cf3" containerName="proxy-httpd" probeResult="failure" output="Get \"http://10.217.0.176:3000/\": read tcp 10.217.0.2:40282->10.217.0.176:3000: read: connection reset by peer" Oct 05 07:08:58 crc kubenswrapper[4846]: I1005 07:08:58.138653 4846 generic.go:334] "Generic (PLEG): container finished" podID="0c1e73f6-0b4e-46ee-9ea1-1b7e52226cf3" containerID="4ecc42b3ffd32b9aafddde8e57e6913de0371f1206218ee0aabf2c1631bf739e" exitCode=0 Oct 05 07:08:58 crc kubenswrapper[4846]: I1005 07:08:58.138702 4846 generic.go:334] "Generic (PLEG): container finished" podID="0c1e73f6-0b4e-46ee-9ea1-1b7e52226cf3" containerID="58064da9228d387abae4a368984ba205dc2577f396cc68d7b4ae2a3f8b25bc4a" exitCode=2 Oct 05 07:08:58 crc kubenswrapper[4846]: I1005 07:08:58.138715 4846 generic.go:334] "Generic (PLEG): container finished" podID="0c1e73f6-0b4e-46ee-9ea1-1b7e52226cf3" containerID="f0ab84da10dbec024f726d79920ee70f2c8e4645eb5c9333660cdff7a7e06945" exitCode=0 Oct 05 07:08:58 crc kubenswrapper[4846]: I1005 07:08:58.138722 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0c1e73f6-0b4e-46ee-9ea1-1b7e52226cf3","Type":"ContainerDied","Data":"4ecc42b3ffd32b9aafddde8e57e6913de0371f1206218ee0aabf2c1631bf739e"} Oct 05 07:08:58 crc kubenswrapper[4846]: I1005 07:08:58.139006 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0c1e73f6-0b4e-46ee-9ea1-1b7e52226cf3","Type":"ContainerDied","Data":"58064da9228d387abae4a368984ba205dc2577f396cc68d7b4ae2a3f8b25bc4a"} Oct 05 07:08:58 crc kubenswrapper[4846]: I1005 07:08:58.139022 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0c1e73f6-0b4e-46ee-9ea1-1b7e52226cf3","Type":"ContainerDied","Data":"f0ab84da10dbec024f726d79920ee70f2c8e4645eb5c9333660cdff7a7e06945"} Oct 05 07:09:03 crc kubenswrapper[4846]: I1005 07:09:03.208449 4846 generic.go:334] "Generic (PLEG): container finished" podID="0c1e73f6-0b4e-46ee-9ea1-1b7e52226cf3" containerID="2011a6bd3f315187706c9865c5eff9573c35a07876057f68548f0732a4adbcdb" exitCode=0 Oct 05 07:09:03 crc kubenswrapper[4846]: I1005 07:09:03.208511 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0c1e73f6-0b4e-46ee-9ea1-1b7e52226cf3","Type":"ContainerDied","Data":"2011a6bd3f315187706c9865c5eff9573c35a07876057f68548f0732a4adbcdb"} Oct 05 07:09:03 crc kubenswrapper[4846]: I1005 07:09:03.685152 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 05 07:09:03 crc kubenswrapper[4846]: I1005 07:09:03.743827 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nlbtk\" (UniqueName: \"kubernetes.io/projected/0c1e73f6-0b4e-46ee-9ea1-1b7e52226cf3-kube-api-access-nlbtk\") pod \"0c1e73f6-0b4e-46ee-9ea1-1b7e52226cf3\" (UID: \"0c1e73f6-0b4e-46ee-9ea1-1b7e52226cf3\") " Oct 05 07:09:03 crc kubenswrapper[4846]: I1005 07:09:03.743927 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0c1e73f6-0b4e-46ee-9ea1-1b7e52226cf3-combined-ca-bundle\") pod \"0c1e73f6-0b4e-46ee-9ea1-1b7e52226cf3\" (UID: \"0c1e73f6-0b4e-46ee-9ea1-1b7e52226cf3\") " Oct 05 07:09:03 crc kubenswrapper[4846]: I1005 07:09:03.744133 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0c1e73f6-0b4e-46ee-9ea1-1b7e52226cf3-scripts\") pod \"0c1e73f6-0b4e-46ee-9ea1-1b7e52226cf3\" (UID: \"0c1e73f6-0b4e-46ee-9ea1-1b7e52226cf3\") " Oct 05 07:09:03 crc kubenswrapper[4846]: I1005 07:09:03.744335 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0c1e73f6-0b4e-46ee-9ea1-1b7e52226cf3-sg-core-conf-yaml\") pod \"0c1e73f6-0b4e-46ee-9ea1-1b7e52226cf3\" (UID: \"0c1e73f6-0b4e-46ee-9ea1-1b7e52226cf3\") " Oct 05 07:09:03 crc kubenswrapper[4846]: I1005 07:09:03.744378 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0c1e73f6-0b4e-46ee-9ea1-1b7e52226cf3-run-httpd\") pod \"0c1e73f6-0b4e-46ee-9ea1-1b7e52226cf3\" (UID: \"0c1e73f6-0b4e-46ee-9ea1-1b7e52226cf3\") " Oct 05 07:09:03 crc kubenswrapper[4846]: I1005 07:09:03.744435 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0c1e73f6-0b4e-46ee-9ea1-1b7e52226cf3-log-httpd\") pod \"0c1e73f6-0b4e-46ee-9ea1-1b7e52226cf3\" (UID: \"0c1e73f6-0b4e-46ee-9ea1-1b7e52226cf3\") " Oct 05 07:09:03 crc kubenswrapper[4846]: I1005 07:09:03.744482 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0c1e73f6-0b4e-46ee-9ea1-1b7e52226cf3-config-data\") pod \"0c1e73f6-0b4e-46ee-9ea1-1b7e52226cf3\" (UID: \"0c1e73f6-0b4e-46ee-9ea1-1b7e52226cf3\") " Oct 05 07:09:03 crc kubenswrapper[4846]: I1005 07:09:03.744803 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0c1e73f6-0b4e-46ee-9ea1-1b7e52226cf3-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "0c1e73f6-0b4e-46ee-9ea1-1b7e52226cf3" (UID: "0c1e73f6-0b4e-46ee-9ea1-1b7e52226cf3"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 07:09:03 crc kubenswrapper[4846]: I1005 07:09:03.745273 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0c1e73f6-0b4e-46ee-9ea1-1b7e52226cf3-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "0c1e73f6-0b4e-46ee-9ea1-1b7e52226cf3" (UID: "0c1e73f6-0b4e-46ee-9ea1-1b7e52226cf3"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 07:09:03 crc kubenswrapper[4846]: I1005 07:09:03.745423 4846 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0c1e73f6-0b4e-46ee-9ea1-1b7e52226cf3-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 05 07:09:03 crc kubenswrapper[4846]: I1005 07:09:03.750414 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0c1e73f6-0b4e-46ee-9ea1-1b7e52226cf3-kube-api-access-nlbtk" (OuterVolumeSpecName: "kube-api-access-nlbtk") pod "0c1e73f6-0b4e-46ee-9ea1-1b7e52226cf3" (UID: "0c1e73f6-0b4e-46ee-9ea1-1b7e52226cf3"). InnerVolumeSpecName "kube-api-access-nlbtk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:09:03 crc kubenswrapper[4846]: I1005 07:09:03.751437 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0c1e73f6-0b4e-46ee-9ea1-1b7e52226cf3-scripts" (OuterVolumeSpecName: "scripts") pod "0c1e73f6-0b4e-46ee-9ea1-1b7e52226cf3" (UID: "0c1e73f6-0b4e-46ee-9ea1-1b7e52226cf3"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:09:03 crc kubenswrapper[4846]: I1005 07:09:03.773302 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0c1e73f6-0b4e-46ee-9ea1-1b7e52226cf3-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "0c1e73f6-0b4e-46ee-9ea1-1b7e52226cf3" (UID: "0c1e73f6-0b4e-46ee-9ea1-1b7e52226cf3"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:09:03 crc kubenswrapper[4846]: I1005 07:09:03.830934 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0c1e73f6-0b4e-46ee-9ea1-1b7e52226cf3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0c1e73f6-0b4e-46ee-9ea1-1b7e52226cf3" (UID: "0c1e73f6-0b4e-46ee-9ea1-1b7e52226cf3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:09:03 crc kubenswrapper[4846]: I1005 07:09:03.847624 4846 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0c1e73f6-0b4e-46ee-9ea1-1b7e52226cf3-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 05 07:09:03 crc kubenswrapper[4846]: I1005 07:09:03.847659 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nlbtk\" (UniqueName: \"kubernetes.io/projected/0c1e73f6-0b4e-46ee-9ea1-1b7e52226cf3-kube-api-access-nlbtk\") on node \"crc\" DevicePath \"\"" Oct 05 07:09:03 crc kubenswrapper[4846]: I1005 07:09:03.847677 4846 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0c1e73f6-0b4e-46ee-9ea1-1b7e52226cf3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 07:09:03 crc kubenswrapper[4846]: I1005 07:09:03.847688 4846 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0c1e73f6-0b4e-46ee-9ea1-1b7e52226cf3-scripts\") on node \"crc\" DevicePath \"\"" Oct 05 07:09:03 crc kubenswrapper[4846]: I1005 07:09:03.847697 4846 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0c1e73f6-0b4e-46ee-9ea1-1b7e52226cf3-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 05 07:09:03 crc kubenswrapper[4846]: I1005 07:09:03.859029 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0c1e73f6-0b4e-46ee-9ea1-1b7e52226cf3-config-data" (OuterVolumeSpecName: "config-data") pod "0c1e73f6-0b4e-46ee-9ea1-1b7e52226cf3" (UID: "0c1e73f6-0b4e-46ee-9ea1-1b7e52226cf3"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:09:03 crc kubenswrapper[4846]: I1005 07:09:03.950624 4846 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0c1e73f6-0b4e-46ee-9ea1-1b7e52226cf3-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 07:09:04 crc kubenswrapper[4846]: I1005 07:09:04.231234 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0c1e73f6-0b4e-46ee-9ea1-1b7e52226cf3","Type":"ContainerDied","Data":"d55b6e1ab7f73560911aa0d1f991fead109d800cc85fcf5ec3cd1ac2f93c693c"} Oct 05 07:09:04 crc kubenswrapper[4846]: I1005 07:09:04.231364 4846 scope.go:117] "RemoveContainer" containerID="4ecc42b3ffd32b9aafddde8e57e6913de0371f1206218ee0aabf2c1631bf739e" Oct 05 07:09:04 crc kubenswrapper[4846]: I1005 07:09:04.231290 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 05 07:09:04 crc kubenswrapper[4846]: I1005 07:09:04.237243 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-88lpl" event={"ID":"7ac2edb5-3ecf-4162-be56-43585390796a","Type":"ContainerStarted","Data":"787d7008e975b8e9978d9240b7d5756af79da41e854c43ea23614d9e93e503b5"} Oct 05 07:09:04 crc kubenswrapper[4846]: I1005 07:09:04.271044 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-db-sync-88lpl" podStartSLOduration=2.23455284 podStartE2EDuration="9.271010929s" podCreationTimestamp="2025-10-05 07:08:55 +0000 UTC" firstStartedPulling="2025-10-05 07:08:56.362257442 +0000 UTC m=+1258.603110217" lastFinishedPulling="2025-10-05 07:09:03.398715531 +0000 UTC m=+1265.639568306" observedRunningTime="2025-10-05 07:09:04.263514449 +0000 UTC m=+1266.504367244" watchObservedRunningTime="2025-10-05 07:09:04.271010929 +0000 UTC m=+1266.511863744" Oct 05 07:09:04 crc kubenswrapper[4846]: I1005 07:09:04.280615 4846 scope.go:117] "RemoveContainer" containerID="58064da9228d387abae4a368984ba205dc2577f396cc68d7b4ae2a3f8b25bc4a" Oct 05 07:09:04 crc kubenswrapper[4846]: I1005 07:09:04.313869 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 05 07:09:04 crc kubenswrapper[4846]: I1005 07:09:04.327686 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 05 07:09:04 crc kubenswrapper[4846]: I1005 07:09:04.331352 4846 scope.go:117] "RemoveContainer" containerID="2011a6bd3f315187706c9865c5eff9573c35a07876057f68548f0732a4adbcdb" Oct 05 07:09:04 crc kubenswrapper[4846]: I1005 07:09:04.339869 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 05 07:09:04 crc kubenswrapper[4846]: E1005 07:09:04.340462 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0c1e73f6-0b4e-46ee-9ea1-1b7e52226cf3" containerName="proxy-httpd" Oct 05 07:09:04 crc kubenswrapper[4846]: I1005 07:09:04.340489 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="0c1e73f6-0b4e-46ee-9ea1-1b7e52226cf3" containerName="proxy-httpd" Oct 05 07:09:04 crc kubenswrapper[4846]: E1005 07:09:04.340509 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0c1e73f6-0b4e-46ee-9ea1-1b7e52226cf3" containerName="ceilometer-notification-agent" Oct 05 07:09:04 crc kubenswrapper[4846]: I1005 07:09:04.340520 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="0c1e73f6-0b4e-46ee-9ea1-1b7e52226cf3" containerName="ceilometer-notification-agent" Oct 05 07:09:04 crc kubenswrapper[4846]: E1005 07:09:04.340941 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0c1e73f6-0b4e-46ee-9ea1-1b7e52226cf3" containerName="ceilometer-central-agent" Oct 05 07:09:04 crc kubenswrapper[4846]: I1005 07:09:04.340968 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="0c1e73f6-0b4e-46ee-9ea1-1b7e52226cf3" containerName="ceilometer-central-agent" Oct 05 07:09:04 crc kubenswrapper[4846]: E1005 07:09:04.340985 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0c1e73f6-0b4e-46ee-9ea1-1b7e52226cf3" containerName="sg-core" Oct 05 07:09:04 crc kubenswrapper[4846]: I1005 07:09:04.340994 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="0c1e73f6-0b4e-46ee-9ea1-1b7e52226cf3" containerName="sg-core" Oct 05 07:09:04 crc kubenswrapper[4846]: I1005 07:09:04.341262 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="0c1e73f6-0b4e-46ee-9ea1-1b7e52226cf3" containerName="proxy-httpd" Oct 05 07:09:04 crc kubenswrapper[4846]: I1005 07:09:04.341297 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="0c1e73f6-0b4e-46ee-9ea1-1b7e52226cf3" containerName="ceilometer-central-agent" Oct 05 07:09:04 crc kubenswrapper[4846]: I1005 07:09:04.341315 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="0c1e73f6-0b4e-46ee-9ea1-1b7e52226cf3" containerName="ceilometer-notification-agent" Oct 05 07:09:04 crc kubenswrapper[4846]: I1005 07:09:04.341326 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="0c1e73f6-0b4e-46ee-9ea1-1b7e52226cf3" containerName="sg-core" Oct 05 07:09:04 crc kubenswrapper[4846]: I1005 07:09:04.343534 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 05 07:09:04 crc kubenswrapper[4846]: I1005 07:09:04.354306 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 05 07:09:04 crc kubenswrapper[4846]: I1005 07:09:04.354786 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 05 07:09:04 crc kubenswrapper[4846]: I1005 07:09:04.381279 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 05 07:09:04 crc kubenswrapper[4846]: I1005 07:09:04.397657 4846 scope.go:117] "RemoveContainer" containerID="f0ab84da10dbec024f726d79920ee70f2c8e4645eb5c9333660cdff7a7e06945" Oct 05 07:09:04 crc kubenswrapper[4846]: I1005 07:09:04.488469 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/06e92a1c-15bf-4cdb-a284-0b5bee2a75c9-config-data\") pod \"ceilometer-0\" (UID: \"06e92a1c-15bf-4cdb-a284-0b5bee2a75c9\") " pod="openstack/ceilometer-0" Oct 05 07:09:04 crc kubenswrapper[4846]: I1005 07:09:04.488571 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/06e92a1c-15bf-4cdb-a284-0b5bee2a75c9-log-httpd\") pod \"ceilometer-0\" (UID: \"06e92a1c-15bf-4cdb-a284-0b5bee2a75c9\") " pod="openstack/ceilometer-0" Oct 05 07:09:04 crc kubenswrapper[4846]: I1005 07:09:04.488681 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/06e92a1c-15bf-4cdb-a284-0b5bee2a75c9-scripts\") pod \"ceilometer-0\" (UID: \"06e92a1c-15bf-4cdb-a284-0b5bee2a75c9\") " pod="openstack/ceilometer-0" Oct 05 07:09:04 crc kubenswrapper[4846]: I1005 07:09:04.488856 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2m5sx\" (UniqueName: \"kubernetes.io/projected/06e92a1c-15bf-4cdb-a284-0b5bee2a75c9-kube-api-access-2m5sx\") pod \"ceilometer-0\" (UID: \"06e92a1c-15bf-4cdb-a284-0b5bee2a75c9\") " pod="openstack/ceilometer-0" Oct 05 07:09:04 crc kubenswrapper[4846]: I1005 07:09:04.488892 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/06e92a1c-15bf-4cdb-a284-0b5bee2a75c9-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"06e92a1c-15bf-4cdb-a284-0b5bee2a75c9\") " pod="openstack/ceilometer-0" Oct 05 07:09:04 crc kubenswrapper[4846]: I1005 07:09:04.488915 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/06e92a1c-15bf-4cdb-a284-0b5bee2a75c9-run-httpd\") pod \"ceilometer-0\" (UID: \"06e92a1c-15bf-4cdb-a284-0b5bee2a75c9\") " pod="openstack/ceilometer-0" Oct 05 07:09:04 crc kubenswrapper[4846]: I1005 07:09:04.488985 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/06e92a1c-15bf-4cdb-a284-0b5bee2a75c9-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"06e92a1c-15bf-4cdb-a284-0b5bee2a75c9\") " pod="openstack/ceilometer-0" Oct 05 07:09:04 crc kubenswrapper[4846]: I1005 07:09:04.521124 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0c1e73f6-0b4e-46ee-9ea1-1b7e52226cf3" path="/var/lib/kubelet/pods/0c1e73f6-0b4e-46ee-9ea1-1b7e52226cf3/volumes" Oct 05 07:09:04 crc kubenswrapper[4846]: I1005 07:09:04.591615 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2m5sx\" (UniqueName: \"kubernetes.io/projected/06e92a1c-15bf-4cdb-a284-0b5bee2a75c9-kube-api-access-2m5sx\") pod \"ceilometer-0\" (UID: \"06e92a1c-15bf-4cdb-a284-0b5bee2a75c9\") " pod="openstack/ceilometer-0" Oct 05 07:09:04 crc kubenswrapper[4846]: I1005 07:09:04.592045 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/06e92a1c-15bf-4cdb-a284-0b5bee2a75c9-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"06e92a1c-15bf-4cdb-a284-0b5bee2a75c9\") " pod="openstack/ceilometer-0" Oct 05 07:09:04 crc kubenswrapper[4846]: I1005 07:09:04.592255 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/06e92a1c-15bf-4cdb-a284-0b5bee2a75c9-run-httpd\") pod \"ceilometer-0\" (UID: \"06e92a1c-15bf-4cdb-a284-0b5bee2a75c9\") " pod="openstack/ceilometer-0" Oct 05 07:09:04 crc kubenswrapper[4846]: I1005 07:09:04.592452 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/06e92a1c-15bf-4cdb-a284-0b5bee2a75c9-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"06e92a1c-15bf-4cdb-a284-0b5bee2a75c9\") " pod="openstack/ceilometer-0" Oct 05 07:09:04 crc kubenswrapper[4846]: I1005 07:09:04.592907 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/06e92a1c-15bf-4cdb-a284-0b5bee2a75c9-config-data\") pod \"ceilometer-0\" (UID: \"06e92a1c-15bf-4cdb-a284-0b5bee2a75c9\") " pod="openstack/ceilometer-0" Oct 05 07:09:04 crc kubenswrapper[4846]: I1005 07:09:04.593236 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/06e92a1c-15bf-4cdb-a284-0b5bee2a75c9-log-httpd\") pod \"ceilometer-0\" (UID: \"06e92a1c-15bf-4cdb-a284-0b5bee2a75c9\") " pod="openstack/ceilometer-0" Oct 05 07:09:04 crc kubenswrapper[4846]: I1005 07:09:04.593246 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/06e92a1c-15bf-4cdb-a284-0b5bee2a75c9-run-httpd\") pod \"ceilometer-0\" (UID: \"06e92a1c-15bf-4cdb-a284-0b5bee2a75c9\") " pod="openstack/ceilometer-0" Oct 05 07:09:04 crc kubenswrapper[4846]: I1005 07:09:04.593565 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/06e92a1c-15bf-4cdb-a284-0b5bee2a75c9-log-httpd\") pod \"ceilometer-0\" (UID: \"06e92a1c-15bf-4cdb-a284-0b5bee2a75c9\") " pod="openstack/ceilometer-0" Oct 05 07:09:04 crc kubenswrapper[4846]: I1005 07:09:04.593486 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/06e92a1c-15bf-4cdb-a284-0b5bee2a75c9-scripts\") pod \"ceilometer-0\" (UID: \"06e92a1c-15bf-4cdb-a284-0b5bee2a75c9\") " pod="openstack/ceilometer-0" Oct 05 07:09:04 crc kubenswrapper[4846]: I1005 07:09:04.602757 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/06e92a1c-15bf-4cdb-a284-0b5bee2a75c9-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"06e92a1c-15bf-4cdb-a284-0b5bee2a75c9\") " pod="openstack/ceilometer-0" Oct 05 07:09:04 crc kubenswrapper[4846]: I1005 07:09:04.603489 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/06e92a1c-15bf-4cdb-a284-0b5bee2a75c9-scripts\") pod \"ceilometer-0\" (UID: \"06e92a1c-15bf-4cdb-a284-0b5bee2a75c9\") " pod="openstack/ceilometer-0" Oct 05 07:09:04 crc kubenswrapper[4846]: I1005 07:09:04.612625 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/06e92a1c-15bf-4cdb-a284-0b5bee2a75c9-config-data\") pod \"ceilometer-0\" (UID: \"06e92a1c-15bf-4cdb-a284-0b5bee2a75c9\") " pod="openstack/ceilometer-0" Oct 05 07:09:04 crc kubenswrapper[4846]: I1005 07:09:04.618029 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/06e92a1c-15bf-4cdb-a284-0b5bee2a75c9-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"06e92a1c-15bf-4cdb-a284-0b5bee2a75c9\") " pod="openstack/ceilometer-0" Oct 05 07:09:04 crc kubenswrapper[4846]: I1005 07:09:04.620604 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2m5sx\" (UniqueName: \"kubernetes.io/projected/06e92a1c-15bf-4cdb-a284-0b5bee2a75c9-kube-api-access-2m5sx\") pod \"ceilometer-0\" (UID: \"06e92a1c-15bf-4cdb-a284-0b5bee2a75c9\") " pod="openstack/ceilometer-0" Oct 05 07:09:04 crc kubenswrapper[4846]: I1005 07:09:04.698125 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 05 07:09:05 crc kubenswrapper[4846]: I1005 07:09:05.175057 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 05 07:09:05 crc kubenswrapper[4846]: W1005 07:09:05.189220 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod06e92a1c_15bf_4cdb_a284_0b5bee2a75c9.slice/crio-0ab3345f5cb08c7a58ce1742d4a3f8bba41ae9e9b2f12e7723b4ec05e1f6da29 WatchSource:0}: Error finding container 0ab3345f5cb08c7a58ce1742d4a3f8bba41ae9e9b2f12e7723b4ec05e1f6da29: Status 404 returned error can't find the container with id 0ab3345f5cb08c7a58ce1742d4a3f8bba41ae9e9b2f12e7723b4ec05e1f6da29 Oct 05 07:09:05 crc kubenswrapper[4846]: I1005 07:09:05.253055 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"06e92a1c-15bf-4cdb-a284-0b5bee2a75c9","Type":"ContainerStarted","Data":"0ab3345f5cb08c7a58ce1742d4a3f8bba41ae9e9b2f12e7723b4ec05e1f6da29"} Oct 05 07:09:06 crc kubenswrapper[4846]: I1005 07:09:06.267609 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"06e92a1c-15bf-4cdb-a284-0b5bee2a75c9","Type":"ContainerStarted","Data":"0a02b9d96618505f33de0057a4a6c5bb3f7d078b7697e171d7309d0fb4459179"} Oct 05 07:09:08 crc kubenswrapper[4846]: I1005 07:09:08.301297 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"06e92a1c-15bf-4cdb-a284-0b5bee2a75c9","Type":"ContainerStarted","Data":"8a4dc29d9a82877eba8bfc9d6ac07fb48d1a0e553bef761e5a88bb04c88cbe6b"} Oct 05 07:09:10 crc kubenswrapper[4846]: I1005 07:09:10.326075 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"06e92a1c-15bf-4cdb-a284-0b5bee2a75c9","Type":"ContainerStarted","Data":"e8da228ce53f0f9b4a512fcd59c901ab9bd35018ddae6a4efd69d1cb97fabdfd"} Oct 05 07:09:11 crc kubenswrapper[4846]: I1005 07:09:11.340468 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"06e92a1c-15bf-4cdb-a284-0b5bee2a75c9","Type":"ContainerStarted","Data":"1366799532bcfbd4e3db914bbc7d301c1fffe5b0fe8364f08939c0180dd80492"} Oct 05 07:09:11 crc kubenswrapper[4846]: I1005 07:09:11.341071 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 05 07:09:19 crc kubenswrapper[4846]: I1005 07:09:19.460522 4846 generic.go:334] "Generic (PLEG): container finished" podID="7ac2edb5-3ecf-4162-be56-43585390796a" containerID="787d7008e975b8e9978d9240b7d5756af79da41e854c43ea23614d9e93e503b5" exitCode=0 Oct 05 07:09:19 crc kubenswrapper[4846]: I1005 07:09:19.460636 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-88lpl" event={"ID":"7ac2edb5-3ecf-4162-be56-43585390796a","Type":"ContainerDied","Data":"787d7008e975b8e9978d9240b7d5756af79da41e854c43ea23614d9e93e503b5"} Oct 05 07:09:19 crc kubenswrapper[4846]: I1005 07:09:19.492746 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=9.878946789 podStartE2EDuration="15.492720228s" podCreationTimestamp="2025-10-05 07:09:04 +0000 UTC" firstStartedPulling="2025-10-05 07:09:05.195381035 +0000 UTC m=+1267.436233810" lastFinishedPulling="2025-10-05 07:09:10.809154434 +0000 UTC m=+1273.050007249" observedRunningTime="2025-10-05 07:09:11.397619601 +0000 UTC m=+1273.638472386" watchObservedRunningTime="2025-10-05 07:09:19.492720228 +0000 UTC m=+1281.733573003" Oct 05 07:09:20 crc kubenswrapper[4846]: I1005 07:09:20.882475 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-88lpl" Oct 05 07:09:20 crc kubenswrapper[4846]: I1005 07:09:20.918196 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7ac2edb5-3ecf-4162-be56-43585390796a-scripts\") pod \"7ac2edb5-3ecf-4162-be56-43585390796a\" (UID: \"7ac2edb5-3ecf-4162-be56-43585390796a\") " Oct 05 07:09:20 crc kubenswrapper[4846]: I1005 07:09:20.918275 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pq68w\" (UniqueName: \"kubernetes.io/projected/7ac2edb5-3ecf-4162-be56-43585390796a-kube-api-access-pq68w\") pod \"7ac2edb5-3ecf-4162-be56-43585390796a\" (UID: \"7ac2edb5-3ecf-4162-be56-43585390796a\") " Oct 05 07:09:20 crc kubenswrapper[4846]: I1005 07:09:20.918305 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7ac2edb5-3ecf-4162-be56-43585390796a-config-data\") pod \"7ac2edb5-3ecf-4162-be56-43585390796a\" (UID: \"7ac2edb5-3ecf-4162-be56-43585390796a\") " Oct 05 07:09:20 crc kubenswrapper[4846]: I1005 07:09:20.918333 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7ac2edb5-3ecf-4162-be56-43585390796a-combined-ca-bundle\") pod \"7ac2edb5-3ecf-4162-be56-43585390796a\" (UID: \"7ac2edb5-3ecf-4162-be56-43585390796a\") " Oct 05 07:09:20 crc kubenswrapper[4846]: I1005 07:09:20.926647 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7ac2edb5-3ecf-4162-be56-43585390796a-kube-api-access-pq68w" (OuterVolumeSpecName: "kube-api-access-pq68w") pod "7ac2edb5-3ecf-4162-be56-43585390796a" (UID: "7ac2edb5-3ecf-4162-be56-43585390796a"). InnerVolumeSpecName "kube-api-access-pq68w". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:09:20 crc kubenswrapper[4846]: I1005 07:09:20.927239 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7ac2edb5-3ecf-4162-be56-43585390796a-scripts" (OuterVolumeSpecName: "scripts") pod "7ac2edb5-3ecf-4162-be56-43585390796a" (UID: "7ac2edb5-3ecf-4162-be56-43585390796a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:09:20 crc kubenswrapper[4846]: I1005 07:09:20.952920 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7ac2edb5-3ecf-4162-be56-43585390796a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7ac2edb5-3ecf-4162-be56-43585390796a" (UID: "7ac2edb5-3ecf-4162-be56-43585390796a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:09:20 crc kubenswrapper[4846]: I1005 07:09:20.964541 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7ac2edb5-3ecf-4162-be56-43585390796a-config-data" (OuterVolumeSpecName: "config-data") pod "7ac2edb5-3ecf-4162-be56-43585390796a" (UID: "7ac2edb5-3ecf-4162-be56-43585390796a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:09:21 crc kubenswrapper[4846]: I1005 07:09:21.020008 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pq68w\" (UniqueName: \"kubernetes.io/projected/7ac2edb5-3ecf-4162-be56-43585390796a-kube-api-access-pq68w\") on node \"crc\" DevicePath \"\"" Oct 05 07:09:21 crc kubenswrapper[4846]: I1005 07:09:21.020282 4846 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7ac2edb5-3ecf-4162-be56-43585390796a-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 07:09:21 crc kubenswrapper[4846]: I1005 07:09:21.020375 4846 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7ac2edb5-3ecf-4162-be56-43585390796a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 07:09:21 crc kubenswrapper[4846]: I1005 07:09:21.020444 4846 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7ac2edb5-3ecf-4162-be56-43585390796a-scripts\") on node \"crc\" DevicePath \"\"" Oct 05 07:09:21 crc kubenswrapper[4846]: I1005 07:09:21.485747 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-88lpl" event={"ID":"7ac2edb5-3ecf-4162-be56-43585390796a","Type":"ContainerDied","Data":"aa3c919abf97da616f69fc85126ad627d72e32b1d7900ee5364241cb5a84442e"} Oct 05 07:09:21 crc kubenswrapper[4846]: I1005 07:09:21.485802 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-88lpl" Oct 05 07:09:21 crc kubenswrapper[4846]: I1005 07:09:21.485804 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="aa3c919abf97da616f69fc85126ad627d72e32b1d7900ee5364241cb5a84442e" Oct 05 07:09:21 crc kubenswrapper[4846]: I1005 07:09:21.619887 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 05 07:09:21 crc kubenswrapper[4846]: E1005 07:09:21.620461 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7ac2edb5-3ecf-4162-be56-43585390796a" containerName="nova-cell0-conductor-db-sync" Oct 05 07:09:21 crc kubenswrapper[4846]: I1005 07:09:21.620488 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="7ac2edb5-3ecf-4162-be56-43585390796a" containerName="nova-cell0-conductor-db-sync" Oct 05 07:09:21 crc kubenswrapper[4846]: I1005 07:09:21.620774 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="7ac2edb5-3ecf-4162-be56-43585390796a" containerName="nova-cell0-conductor-db-sync" Oct 05 07:09:21 crc kubenswrapper[4846]: I1005 07:09:21.621676 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Oct 05 07:09:21 crc kubenswrapper[4846]: I1005 07:09:21.624869 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Oct 05 07:09:21 crc kubenswrapper[4846]: I1005 07:09:21.627189 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-4c97w" Oct 05 07:09:21 crc kubenswrapper[4846]: I1005 07:09:21.645244 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 05 07:09:21 crc kubenswrapper[4846]: I1005 07:09:21.740966 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/98b50269-c6d4-42f2-8cef-afc127bfccb1-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"98b50269-c6d4-42f2-8cef-afc127bfccb1\") " pod="openstack/nova-cell0-conductor-0" Oct 05 07:09:21 crc kubenswrapper[4846]: I1005 07:09:21.741491 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hfbjg\" (UniqueName: \"kubernetes.io/projected/98b50269-c6d4-42f2-8cef-afc127bfccb1-kube-api-access-hfbjg\") pod \"nova-cell0-conductor-0\" (UID: \"98b50269-c6d4-42f2-8cef-afc127bfccb1\") " pod="openstack/nova-cell0-conductor-0" Oct 05 07:09:21 crc kubenswrapper[4846]: I1005 07:09:21.741559 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/98b50269-c6d4-42f2-8cef-afc127bfccb1-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"98b50269-c6d4-42f2-8cef-afc127bfccb1\") " pod="openstack/nova-cell0-conductor-0" Oct 05 07:09:21 crc kubenswrapper[4846]: E1005 07:09:21.742349 4846 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7ac2edb5_3ecf_4162_be56_43585390796a.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7ac2edb5_3ecf_4162_be56_43585390796a.slice/crio-aa3c919abf97da616f69fc85126ad627d72e32b1d7900ee5364241cb5a84442e\": RecentStats: unable to find data in memory cache]" Oct 05 07:09:21 crc kubenswrapper[4846]: I1005 07:09:21.843917 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hfbjg\" (UniqueName: \"kubernetes.io/projected/98b50269-c6d4-42f2-8cef-afc127bfccb1-kube-api-access-hfbjg\") pod \"nova-cell0-conductor-0\" (UID: \"98b50269-c6d4-42f2-8cef-afc127bfccb1\") " pod="openstack/nova-cell0-conductor-0" Oct 05 07:09:21 crc kubenswrapper[4846]: I1005 07:09:21.844024 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/98b50269-c6d4-42f2-8cef-afc127bfccb1-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"98b50269-c6d4-42f2-8cef-afc127bfccb1\") " pod="openstack/nova-cell0-conductor-0" Oct 05 07:09:21 crc kubenswrapper[4846]: I1005 07:09:21.844171 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/98b50269-c6d4-42f2-8cef-afc127bfccb1-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"98b50269-c6d4-42f2-8cef-afc127bfccb1\") " pod="openstack/nova-cell0-conductor-0" Oct 05 07:09:21 crc kubenswrapper[4846]: I1005 07:09:21.850340 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/98b50269-c6d4-42f2-8cef-afc127bfccb1-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"98b50269-c6d4-42f2-8cef-afc127bfccb1\") " pod="openstack/nova-cell0-conductor-0" Oct 05 07:09:21 crc kubenswrapper[4846]: I1005 07:09:21.853641 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/98b50269-c6d4-42f2-8cef-afc127bfccb1-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"98b50269-c6d4-42f2-8cef-afc127bfccb1\") " pod="openstack/nova-cell0-conductor-0" Oct 05 07:09:21 crc kubenswrapper[4846]: I1005 07:09:21.868163 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hfbjg\" (UniqueName: \"kubernetes.io/projected/98b50269-c6d4-42f2-8cef-afc127bfccb1-kube-api-access-hfbjg\") pod \"nova-cell0-conductor-0\" (UID: \"98b50269-c6d4-42f2-8cef-afc127bfccb1\") " pod="openstack/nova-cell0-conductor-0" Oct 05 07:09:21 crc kubenswrapper[4846]: I1005 07:09:21.946409 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Oct 05 07:09:22 crc kubenswrapper[4846]: W1005 07:09:22.468639 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod98b50269_c6d4_42f2_8cef_afc127bfccb1.slice/crio-9f3658c37c9a27d088645083e4f9c2797423c225b6855b8581b1eff8ce69200e WatchSource:0}: Error finding container 9f3658c37c9a27d088645083e4f9c2797423c225b6855b8581b1eff8ce69200e: Status 404 returned error can't find the container with id 9f3658c37c9a27d088645083e4f9c2797423c225b6855b8581b1eff8ce69200e Oct 05 07:09:22 crc kubenswrapper[4846]: I1005 07:09:22.473596 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 05 07:09:22 crc kubenswrapper[4846]: I1005 07:09:22.523356 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"98b50269-c6d4-42f2-8cef-afc127bfccb1","Type":"ContainerStarted","Data":"9f3658c37c9a27d088645083e4f9c2797423c225b6855b8581b1eff8ce69200e"} Oct 05 07:09:23 crc kubenswrapper[4846]: I1005 07:09:23.324557 4846 patch_prober.go:28] interesting pod/machine-config-daemon-fscvf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 07:09:23 crc kubenswrapper[4846]: I1005 07:09:23.325489 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 07:09:23 crc kubenswrapper[4846]: I1005 07:09:23.522983 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"98b50269-c6d4-42f2-8cef-afc127bfccb1","Type":"ContainerStarted","Data":"055332eb57f5e6e889d548a1b4717caea296e84dc70616d5bdb53dd34aaa0697"} Oct 05 07:09:23 crc kubenswrapper[4846]: I1005 07:09:23.523232 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Oct 05 07:09:32 crc kubenswrapper[4846]: I1005 07:09:32.000445 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Oct 05 07:09:32 crc kubenswrapper[4846]: I1005 07:09:32.026105 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=11.026081403 podStartE2EDuration="11.026081403s" podCreationTimestamp="2025-10-05 07:09:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 07:09:23.554263494 +0000 UTC m=+1285.795116279" watchObservedRunningTime="2025-10-05 07:09:32.026081403 +0000 UTC m=+1294.266934198" Oct 05 07:09:32 crc kubenswrapper[4846]: I1005 07:09:32.515309 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-cell-mapping-ptdtq"] Oct 05 07:09:32 crc kubenswrapper[4846]: I1005 07:09:32.519238 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-ptdtq" Oct 05 07:09:32 crc kubenswrapper[4846]: I1005 07:09:32.522605 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-config-data" Oct 05 07:09:32 crc kubenswrapper[4846]: I1005 07:09:32.543365 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-scripts" Oct 05 07:09:32 crc kubenswrapper[4846]: I1005 07:09:32.574966 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-ptdtq"] Oct 05 07:09:32 crc kubenswrapper[4846]: I1005 07:09:32.638431 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tj7rk\" (UniqueName: \"kubernetes.io/projected/a368d12f-6383-4a8e-8d71-23a675946027-kube-api-access-tj7rk\") pod \"nova-cell0-cell-mapping-ptdtq\" (UID: \"a368d12f-6383-4a8e-8d71-23a675946027\") " pod="openstack/nova-cell0-cell-mapping-ptdtq" Oct 05 07:09:32 crc kubenswrapper[4846]: I1005 07:09:32.638535 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a368d12f-6383-4a8e-8d71-23a675946027-scripts\") pod \"nova-cell0-cell-mapping-ptdtq\" (UID: \"a368d12f-6383-4a8e-8d71-23a675946027\") " pod="openstack/nova-cell0-cell-mapping-ptdtq" Oct 05 07:09:32 crc kubenswrapper[4846]: I1005 07:09:32.638570 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a368d12f-6383-4a8e-8d71-23a675946027-config-data\") pod \"nova-cell0-cell-mapping-ptdtq\" (UID: \"a368d12f-6383-4a8e-8d71-23a675946027\") " pod="openstack/nova-cell0-cell-mapping-ptdtq" Oct 05 07:09:32 crc kubenswrapper[4846]: I1005 07:09:32.638713 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a368d12f-6383-4a8e-8d71-23a675946027-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-ptdtq\" (UID: \"a368d12f-6383-4a8e-8d71-23a675946027\") " pod="openstack/nova-cell0-cell-mapping-ptdtq" Oct 05 07:09:32 crc kubenswrapper[4846]: I1005 07:09:32.699713 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 05 07:09:32 crc kubenswrapper[4846]: I1005 07:09:32.716949 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 05 07:09:32 crc kubenswrapper[4846]: I1005 07:09:32.718289 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 05 07:09:32 crc kubenswrapper[4846]: I1005 07:09:32.722759 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 05 07:09:32 crc kubenswrapper[4846]: I1005 07:09:32.747200 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a368d12f-6383-4a8e-8d71-23a675946027-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-ptdtq\" (UID: \"a368d12f-6383-4a8e-8d71-23a675946027\") " pod="openstack/nova-cell0-cell-mapping-ptdtq" Oct 05 07:09:32 crc kubenswrapper[4846]: I1005 07:09:32.747393 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tj7rk\" (UniqueName: \"kubernetes.io/projected/a368d12f-6383-4a8e-8d71-23a675946027-kube-api-access-tj7rk\") pod \"nova-cell0-cell-mapping-ptdtq\" (UID: \"a368d12f-6383-4a8e-8d71-23a675946027\") " pod="openstack/nova-cell0-cell-mapping-ptdtq" Oct 05 07:09:32 crc kubenswrapper[4846]: I1005 07:09:32.747444 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a368d12f-6383-4a8e-8d71-23a675946027-scripts\") pod \"nova-cell0-cell-mapping-ptdtq\" (UID: \"a368d12f-6383-4a8e-8d71-23a675946027\") " pod="openstack/nova-cell0-cell-mapping-ptdtq" Oct 05 07:09:32 crc kubenswrapper[4846]: I1005 07:09:32.747475 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a368d12f-6383-4a8e-8d71-23a675946027-config-data\") pod \"nova-cell0-cell-mapping-ptdtq\" (UID: \"a368d12f-6383-4a8e-8d71-23a675946027\") " pod="openstack/nova-cell0-cell-mapping-ptdtq" Oct 05 07:09:32 crc kubenswrapper[4846]: I1005 07:09:32.763490 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a368d12f-6383-4a8e-8d71-23a675946027-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-ptdtq\" (UID: \"a368d12f-6383-4a8e-8d71-23a675946027\") " pod="openstack/nova-cell0-cell-mapping-ptdtq" Oct 05 07:09:32 crc kubenswrapper[4846]: I1005 07:09:32.766513 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a368d12f-6383-4a8e-8d71-23a675946027-config-data\") pod \"nova-cell0-cell-mapping-ptdtq\" (UID: \"a368d12f-6383-4a8e-8d71-23a675946027\") " pod="openstack/nova-cell0-cell-mapping-ptdtq" Oct 05 07:09:32 crc kubenswrapper[4846]: I1005 07:09:32.777853 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a368d12f-6383-4a8e-8d71-23a675946027-scripts\") pod \"nova-cell0-cell-mapping-ptdtq\" (UID: \"a368d12f-6383-4a8e-8d71-23a675946027\") " pod="openstack/nova-cell0-cell-mapping-ptdtq" Oct 05 07:09:32 crc kubenswrapper[4846]: I1005 07:09:32.802363 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 05 07:09:32 crc kubenswrapper[4846]: I1005 07:09:32.804405 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 05 07:09:32 crc kubenswrapper[4846]: I1005 07:09:32.809344 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tj7rk\" (UniqueName: \"kubernetes.io/projected/a368d12f-6383-4a8e-8d71-23a675946027-kube-api-access-tj7rk\") pod \"nova-cell0-cell-mapping-ptdtq\" (UID: \"a368d12f-6383-4a8e-8d71-23a675946027\") " pod="openstack/nova-cell0-cell-mapping-ptdtq" Oct 05 07:09:32 crc kubenswrapper[4846]: I1005 07:09:32.809816 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 05 07:09:32 crc kubenswrapper[4846]: I1005 07:09:32.857251 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9gb6s\" (UniqueName: \"kubernetes.io/projected/095eba43-2814-4476-b482-efdf6b903de4-kube-api-access-9gb6s\") pod \"nova-api-0\" (UID: \"095eba43-2814-4476-b482-efdf6b903de4\") " pod="openstack/nova-api-0" Oct 05 07:09:32 crc kubenswrapper[4846]: I1005 07:09:32.857326 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/095eba43-2814-4476-b482-efdf6b903de4-logs\") pod \"nova-api-0\" (UID: \"095eba43-2814-4476-b482-efdf6b903de4\") " pod="openstack/nova-api-0" Oct 05 07:09:32 crc kubenswrapper[4846]: I1005 07:09:32.857358 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/095eba43-2814-4476-b482-efdf6b903de4-config-data\") pod \"nova-api-0\" (UID: \"095eba43-2814-4476-b482-efdf6b903de4\") " pod="openstack/nova-api-0" Oct 05 07:09:32 crc kubenswrapper[4846]: I1005 07:09:32.857382 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/095eba43-2814-4476-b482-efdf6b903de4-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"095eba43-2814-4476-b482-efdf6b903de4\") " pod="openstack/nova-api-0" Oct 05 07:09:32 crc kubenswrapper[4846]: I1005 07:09:32.858973 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 05 07:09:32 crc kubenswrapper[4846]: I1005 07:09:32.872463 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-ptdtq" Oct 05 07:09:32 crc kubenswrapper[4846]: I1005 07:09:32.947413 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Oct 05 07:09:32 crc kubenswrapper[4846]: I1005 07:09:32.949319 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 05 07:09:32 crc kubenswrapper[4846]: I1005 07:09:32.958622 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Oct 05 07:09:32 crc kubenswrapper[4846]: I1005 07:09:32.958999 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/819bc991-94dc-4fd3-984e-2f3f711f5bf5-config-data\") pod \"nova-metadata-0\" (UID: \"819bc991-94dc-4fd3-984e-2f3f711f5bf5\") " pod="openstack/nova-metadata-0" Oct 05 07:09:32 crc kubenswrapper[4846]: I1005 07:09:32.959049 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9gb6s\" (UniqueName: \"kubernetes.io/projected/095eba43-2814-4476-b482-efdf6b903de4-kube-api-access-9gb6s\") pod \"nova-api-0\" (UID: \"095eba43-2814-4476-b482-efdf6b903de4\") " pod="openstack/nova-api-0" Oct 05 07:09:32 crc kubenswrapper[4846]: I1005 07:09:32.959077 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/095eba43-2814-4476-b482-efdf6b903de4-logs\") pod \"nova-api-0\" (UID: \"095eba43-2814-4476-b482-efdf6b903de4\") " pod="openstack/nova-api-0" Oct 05 07:09:32 crc kubenswrapper[4846]: I1005 07:09:32.959104 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/095eba43-2814-4476-b482-efdf6b903de4-config-data\") pod \"nova-api-0\" (UID: \"095eba43-2814-4476-b482-efdf6b903de4\") " pod="openstack/nova-api-0" Oct 05 07:09:32 crc kubenswrapper[4846]: I1005 07:09:32.959128 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/095eba43-2814-4476-b482-efdf6b903de4-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"095eba43-2814-4476-b482-efdf6b903de4\") " pod="openstack/nova-api-0" Oct 05 07:09:32 crc kubenswrapper[4846]: I1005 07:09:32.959167 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/819bc991-94dc-4fd3-984e-2f3f711f5bf5-logs\") pod \"nova-metadata-0\" (UID: \"819bc991-94dc-4fd3-984e-2f3f711f5bf5\") " pod="openstack/nova-metadata-0" Oct 05 07:09:32 crc kubenswrapper[4846]: I1005 07:09:32.959204 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/819bc991-94dc-4fd3-984e-2f3f711f5bf5-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"819bc991-94dc-4fd3-984e-2f3f711f5bf5\") " pod="openstack/nova-metadata-0" Oct 05 07:09:32 crc kubenswrapper[4846]: I1005 07:09:32.959236 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mvtzk\" (UniqueName: \"kubernetes.io/projected/819bc991-94dc-4fd3-984e-2f3f711f5bf5-kube-api-access-mvtzk\") pod \"nova-metadata-0\" (UID: \"819bc991-94dc-4fd3-984e-2f3f711f5bf5\") " pod="openstack/nova-metadata-0" Oct 05 07:09:32 crc kubenswrapper[4846]: I1005 07:09:32.959670 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/095eba43-2814-4476-b482-efdf6b903de4-logs\") pod \"nova-api-0\" (UID: \"095eba43-2814-4476-b482-efdf6b903de4\") " pod="openstack/nova-api-0" Oct 05 07:09:32 crc kubenswrapper[4846]: I1005 07:09:32.970274 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/095eba43-2814-4476-b482-efdf6b903de4-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"095eba43-2814-4476-b482-efdf6b903de4\") " pod="openstack/nova-api-0" Oct 05 07:09:32 crc kubenswrapper[4846]: I1005 07:09:32.975047 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/095eba43-2814-4476-b482-efdf6b903de4-config-data\") pod \"nova-api-0\" (UID: \"095eba43-2814-4476-b482-efdf6b903de4\") " pod="openstack/nova-api-0" Oct 05 07:09:32 crc kubenswrapper[4846]: I1005 07:09:32.975134 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 05 07:09:33 crc kubenswrapper[4846]: I1005 07:09:33.000894 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9gb6s\" (UniqueName: \"kubernetes.io/projected/095eba43-2814-4476-b482-efdf6b903de4-kube-api-access-9gb6s\") pod \"nova-api-0\" (UID: \"095eba43-2814-4476-b482-efdf6b903de4\") " pod="openstack/nova-api-0" Oct 05 07:09:33 crc kubenswrapper[4846]: I1005 07:09:33.027283 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-78565f64c9-vkrdf"] Oct 05 07:09:33 crc kubenswrapper[4846]: I1005 07:09:33.029095 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78565f64c9-vkrdf" Oct 05 07:09:33 crc kubenswrapper[4846]: I1005 07:09:33.062953 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 05 07:09:33 crc kubenswrapper[4846]: I1005 07:09:33.066099 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 05 07:09:33 crc kubenswrapper[4846]: I1005 07:09:33.068385 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tnx6l\" (UniqueName: \"kubernetes.io/projected/ef292d07-1248-4edd-a182-9e9d572513ca-kube-api-access-tnx6l\") pod \"nova-scheduler-0\" (UID: \"ef292d07-1248-4edd-a182-9e9d572513ca\") " pod="openstack/nova-scheduler-0" Oct 05 07:09:33 crc kubenswrapper[4846]: I1005 07:09:33.068460 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/819bc991-94dc-4fd3-984e-2f3f711f5bf5-config-data\") pod \"nova-metadata-0\" (UID: \"819bc991-94dc-4fd3-984e-2f3f711f5bf5\") " pod="openstack/nova-metadata-0" Oct 05 07:09:33 crc kubenswrapper[4846]: I1005 07:09:33.068487 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/68a854c2-afda-4dbb-9ac7-06eec2c98f14-ovsdbserver-sb\") pod \"dnsmasq-dns-78565f64c9-vkrdf\" (UID: \"68a854c2-afda-4dbb-9ac7-06eec2c98f14\") " pod="openstack/dnsmasq-dns-78565f64c9-vkrdf" Oct 05 07:09:33 crc kubenswrapper[4846]: I1005 07:09:33.068514 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ef292d07-1248-4edd-a182-9e9d572513ca-config-data\") pod \"nova-scheduler-0\" (UID: \"ef292d07-1248-4edd-a182-9e9d572513ca\") " pod="openstack/nova-scheduler-0" Oct 05 07:09:33 crc kubenswrapper[4846]: I1005 07:09:33.068566 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5spwx\" (UniqueName: \"kubernetes.io/projected/a0844571-8880-4fa4-af42-2da48aa8bdec-kube-api-access-5spwx\") pod \"nova-cell1-novncproxy-0\" (UID: \"a0844571-8880-4fa4-af42-2da48aa8bdec\") " pod="openstack/nova-cell1-novncproxy-0" Oct 05 07:09:33 crc kubenswrapper[4846]: I1005 07:09:33.068596 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/68a854c2-afda-4dbb-9ac7-06eec2c98f14-config\") pod \"dnsmasq-dns-78565f64c9-vkrdf\" (UID: \"68a854c2-afda-4dbb-9ac7-06eec2c98f14\") " pod="openstack/dnsmasq-dns-78565f64c9-vkrdf" Oct 05 07:09:33 crc kubenswrapper[4846]: I1005 07:09:33.068620 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/819bc991-94dc-4fd3-984e-2f3f711f5bf5-logs\") pod \"nova-metadata-0\" (UID: \"819bc991-94dc-4fd3-984e-2f3f711f5bf5\") " pod="openstack/nova-metadata-0" Oct 05 07:09:33 crc kubenswrapper[4846]: I1005 07:09:33.068644 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/819bc991-94dc-4fd3-984e-2f3f711f5bf5-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"819bc991-94dc-4fd3-984e-2f3f711f5bf5\") " pod="openstack/nova-metadata-0" Oct 05 07:09:33 crc kubenswrapper[4846]: I1005 07:09:33.068668 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a0844571-8880-4fa4-af42-2da48aa8bdec-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"a0844571-8880-4fa4-af42-2da48aa8bdec\") " pod="openstack/nova-cell1-novncproxy-0" Oct 05 07:09:33 crc kubenswrapper[4846]: I1005 07:09:33.068694 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a0844571-8880-4fa4-af42-2da48aa8bdec-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"a0844571-8880-4fa4-af42-2da48aa8bdec\") " pod="openstack/nova-cell1-novncproxy-0" Oct 05 07:09:33 crc kubenswrapper[4846]: I1005 07:09:33.068715 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mvtzk\" (UniqueName: \"kubernetes.io/projected/819bc991-94dc-4fd3-984e-2f3f711f5bf5-kube-api-access-mvtzk\") pod \"nova-metadata-0\" (UID: \"819bc991-94dc-4fd3-984e-2f3f711f5bf5\") " pod="openstack/nova-metadata-0" Oct 05 07:09:33 crc kubenswrapper[4846]: I1005 07:09:33.068733 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef292d07-1248-4edd-a182-9e9d572513ca-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"ef292d07-1248-4edd-a182-9e9d572513ca\") " pod="openstack/nova-scheduler-0" Oct 05 07:09:33 crc kubenswrapper[4846]: I1005 07:09:33.068757 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/68a854c2-afda-4dbb-9ac7-06eec2c98f14-dns-svc\") pod \"dnsmasq-dns-78565f64c9-vkrdf\" (UID: \"68a854c2-afda-4dbb-9ac7-06eec2c98f14\") " pod="openstack/dnsmasq-dns-78565f64c9-vkrdf" Oct 05 07:09:33 crc kubenswrapper[4846]: I1005 07:09:33.068776 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qdvf6\" (UniqueName: \"kubernetes.io/projected/68a854c2-afda-4dbb-9ac7-06eec2c98f14-kube-api-access-qdvf6\") pod \"dnsmasq-dns-78565f64c9-vkrdf\" (UID: \"68a854c2-afda-4dbb-9ac7-06eec2c98f14\") " pod="openstack/dnsmasq-dns-78565f64c9-vkrdf" Oct 05 07:09:33 crc kubenswrapper[4846]: I1005 07:09:33.068804 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/68a854c2-afda-4dbb-9ac7-06eec2c98f14-ovsdbserver-nb\") pod \"dnsmasq-dns-78565f64c9-vkrdf\" (UID: \"68a854c2-afda-4dbb-9ac7-06eec2c98f14\") " pod="openstack/dnsmasq-dns-78565f64c9-vkrdf" Oct 05 07:09:33 crc kubenswrapper[4846]: I1005 07:09:33.068831 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/68a854c2-afda-4dbb-9ac7-06eec2c98f14-dns-swift-storage-0\") pod \"dnsmasq-dns-78565f64c9-vkrdf\" (UID: \"68a854c2-afda-4dbb-9ac7-06eec2c98f14\") " pod="openstack/dnsmasq-dns-78565f64c9-vkrdf" Oct 05 07:09:33 crc kubenswrapper[4846]: I1005 07:09:33.074847 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/819bc991-94dc-4fd3-984e-2f3f711f5bf5-logs\") pod \"nova-metadata-0\" (UID: \"819bc991-94dc-4fd3-984e-2f3f711f5bf5\") " pod="openstack/nova-metadata-0" Oct 05 07:09:33 crc kubenswrapper[4846]: I1005 07:09:33.076017 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Oct 05 07:09:33 crc kubenswrapper[4846]: I1005 07:09:33.078792 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/819bc991-94dc-4fd3-984e-2f3f711f5bf5-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"819bc991-94dc-4fd3-984e-2f3f711f5bf5\") " pod="openstack/nova-metadata-0" Oct 05 07:09:33 crc kubenswrapper[4846]: I1005 07:09:33.080833 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/819bc991-94dc-4fd3-984e-2f3f711f5bf5-config-data\") pod \"nova-metadata-0\" (UID: \"819bc991-94dc-4fd3-984e-2f3f711f5bf5\") " pod="openstack/nova-metadata-0" Oct 05 07:09:33 crc kubenswrapper[4846]: I1005 07:09:33.098152 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mvtzk\" (UniqueName: \"kubernetes.io/projected/819bc991-94dc-4fd3-984e-2f3f711f5bf5-kube-api-access-mvtzk\") pod \"nova-metadata-0\" (UID: \"819bc991-94dc-4fd3-984e-2f3f711f5bf5\") " pod="openstack/nova-metadata-0" Oct 05 07:09:33 crc kubenswrapper[4846]: I1005 07:09:33.103770 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 05 07:09:33 crc kubenswrapper[4846]: I1005 07:09:33.112619 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78565f64c9-vkrdf"] Oct 05 07:09:33 crc kubenswrapper[4846]: I1005 07:09:33.170619 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/68a854c2-afda-4dbb-9ac7-06eec2c98f14-dns-swift-storage-0\") pod \"dnsmasq-dns-78565f64c9-vkrdf\" (UID: \"68a854c2-afda-4dbb-9ac7-06eec2c98f14\") " pod="openstack/dnsmasq-dns-78565f64c9-vkrdf" Oct 05 07:09:33 crc kubenswrapper[4846]: I1005 07:09:33.171096 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tnx6l\" (UniqueName: \"kubernetes.io/projected/ef292d07-1248-4edd-a182-9e9d572513ca-kube-api-access-tnx6l\") pod \"nova-scheduler-0\" (UID: \"ef292d07-1248-4edd-a182-9e9d572513ca\") " pod="openstack/nova-scheduler-0" Oct 05 07:09:33 crc kubenswrapper[4846]: I1005 07:09:33.171145 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/68a854c2-afda-4dbb-9ac7-06eec2c98f14-ovsdbserver-sb\") pod \"dnsmasq-dns-78565f64c9-vkrdf\" (UID: \"68a854c2-afda-4dbb-9ac7-06eec2c98f14\") " pod="openstack/dnsmasq-dns-78565f64c9-vkrdf" Oct 05 07:09:33 crc kubenswrapper[4846]: I1005 07:09:33.171190 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ef292d07-1248-4edd-a182-9e9d572513ca-config-data\") pod \"nova-scheduler-0\" (UID: \"ef292d07-1248-4edd-a182-9e9d572513ca\") " pod="openstack/nova-scheduler-0" Oct 05 07:09:33 crc kubenswrapper[4846]: I1005 07:09:33.171268 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5spwx\" (UniqueName: \"kubernetes.io/projected/a0844571-8880-4fa4-af42-2da48aa8bdec-kube-api-access-5spwx\") pod \"nova-cell1-novncproxy-0\" (UID: \"a0844571-8880-4fa4-af42-2da48aa8bdec\") " pod="openstack/nova-cell1-novncproxy-0" Oct 05 07:09:33 crc kubenswrapper[4846]: I1005 07:09:33.171313 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/68a854c2-afda-4dbb-9ac7-06eec2c98f14-config\") pod \"dnsmasq-dns-78565f64c9-vkrdf\" (UID: \"68a854c2-afda-4dbb-9ac7-06eec2c98f14\") " pod="openstack/dnsmasq-dns-78565f64c9-vkrdf" Oct 05 07:09:33 crc kubenswrapper[4846]: I1005 07:09:33.171350 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a0844571-8880-4fa4-af42-2da48aa8bdec-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"a0844571-8880-4fa4-af42-2da48aa8bdec\") " pod="openstack/nova-cell1-novncproxy-0" Oct 05 07:09:33 crc kubenswrapper[4846]: I1005 07:09:33.171379 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a0844571-8880-4fa4-af42-2da48aa8bdec-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"a0844571-8880-4fa4-af42-2da48aa8bdec\") " pod="openstack/nova-cell1-novncproxy-0" Oct 05 07:09:33 crc kubenswrapper[4846]: I1005 07:09:33.171401 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef292d07-1248-4edd-a182-9e9d572513ca-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"ef292d07-1248-4edd-a182-9e9d572513ca\") " pod="openstack/nova-scheduler-0" Oct 05 07:09:33 crc kubenswrapper[4846]: I1005 07:09:33.171426 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/68a854c2-afda-4dbb-9ac7-06eec2c98f14-dns-svc\") pod \"dnsmasq-dns-78565f64c9-vkrdf\" (UID: \"68a854c2-afda-4dbb-9ac7-06eec2c98f14\") " pod="openstack/dnsmasq-dns-78565f64c9-vkrdf" Oct 05 07:09:33 crc kubenswrapper[4846]: I1005 07:09:33.171447 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qdvf6\" (UniqueName: \"kubernetes.io/projected/68a854c2-afda-4dbb-9ac7-06eec2c98f14-kube-api-access-qdvf6\") pod \"dnsmasq-dns-78565f64c9-vkrdf\" (UID: \"68a854c2-afda-4dbb-9ac7-06eec2c98f14\") " pod="openstack/dnsmasq-dns-78565f64c9-vkrdf" Oct 05 07:09:33 crc kubenswrapper[4846]: I1005 07:09:33.171477 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/68a854c2-afda-4dbb-9ac7-06eec2c98f14-ovsdbserver-nb\") pod \"dnsmasq-dns-78565f64c9-vkrdf\" (UID: \"68a854c2-afda-4dbb-9ac7-06eec2c98f14\") " pod="openstack/dnsmasq-dns-78565f64c9-vkrdf" Oct 05 07:09:33 crc kubenswrapper[4846]: I1005 07:09:33.171816 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/68a854c2-afda-4dbb-9ac7-06eec2c98f14-dns-swift-storage-0\") pod \"dnsmasq-dns-78565f64c9-vkrdf\" (UID: \"68a854c2-afda-4dbb-9ac7-06eec2c98f14\") " pod="openstack/dnsmasq-dns-78565f64c9-vkrdf" Oct 05 07:09:33 crc kubenswrapper[4846]: I1005 07:09:33.173283 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/68a854c2-afda-4dbb-9ac7-06eec2c98f14-ovsdbserver-sb\") pod \"dnsmasq-dns-78565f64c9-vkrdf\" (UID: \"68a854c2-afda-4dbb-9ac7-06eec2c98f14\") " pod="openstack/dnsmasq-dns-78565f64c9-vkrdf" Oct 05 07:09:33 crc kubenswrapper[4846]: I1005 07:09:33.175015 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/68a854c2-afda-4dbb-9ac7-06eec2c98f14-config\") pod \"dnsmasq-dns-78565f64c9-vkrdf\" (UID: \"68a854c2-afda-4dbb-9ac7-06eec2c98f14\") " pod="openstack/dnsmasq-dns-78565f64c9-vkrdf" Oct 05 07:09:33 crc kubenswrapper[4846]: I1005 07:09:33.175075 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/68a854c2-afda-4dbb-9ac7-06eec2c98f14-dns-svc\") pod \"dnsmasq-dns-78565f64c9-vkrdf\" (UID: \"68a854c2-afda-4dbb-9ac7-06eec2c98f14\") " pod="openstack/dnsmasq-dns-78565f64c9-vkrdf" Oct 05 07:09:33 crc kubenswrapper[4846]: I1005 07:09:33.175631 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/68a854c2-afda-4dbb-9ac7-06eec2c98f14-ovsdbserver-nb\") pod \"dnsmasq-dns-78565f64c9-vkrdf\" (UID: \"68a854c2-afda-4dbb-9ac7-06eec2c98f14\") " pod="openstack/dnsmasq-dns-78565f64c9-vkrdf" Oct 05 07:09:33 crc kubenswrapper[4846]: I1005 07:09:33.179521 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef292d07-1248-4edd-a182-9e9d572513ca-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"ef292d07-1248-4edd-a182-9e9d572513ca\") " pod="openstack/nova-scheduler-0" Oct 05 07:09:33 crc kubenswrapper[4846]: I1005 07:09:33.180337 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a0844571-8880-4fa4-af42-2da48aa8bdec-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"a0844571-8880-4fa4-af42-2da48aa8bdec\") " pod="openstack/nova-cell1-novncproxy-0" Oct 05 07:09:33 crc kubenswrapper[4846]: I1005 07:09:33.181803 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ef292d07-1248-4edd-a182-9e9d572513ca-config-data\") pod \"nova-scheduler-0\" (UID: \"ef292d07-1248-4edd-a182-9e9d572513ca\") " pod="openstack/nova-scheduler-0" Oct 05 07:09:33 crc kubenswrapper[4846]: I1005 07:09:33.188403 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a0844571-8880-4fa4-af42-2da48aa8bdec-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"a0844571-8880-4fa4-af42-2da48aa8bdec\") " pod="openstack/nova-cell1-novncproxy-0" Oct 05 07:09:33 crc kubenswrapper[4846]: I1005 07:09:33.193563 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5spwx\" (UniqueName: \"kubernetes.io/projected/a0844571-8880-4fa4-af42-2da48aa8bdec-kube-api-access-5spwx\") pod \"nova-cell1-novncproxy-0\" (UID: \"a0844571-8880-4fa4-af42-2da48aa8bdec\") " pod="openstack/nova-cell1-novncproxy-0" Oct 05 07:09:33 crc kubenswrapper[4846]: I1005 07:09:33.194503 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tnx6l\" (UniqueName: \"kubernetes.io/projected/ef292d07-1248-4edd-a182-9e9d572513ca-kube-api-access-tnx6l\") pod \"nova-scheduler-0\" (UID: \"ef292d07-1248-4edd-a182-9e9d572513ca\") " pod="openstack/nova-scheduler-0" Oct 05 07:09:33 crc kubenswrapper[4846]: I1005 07:09:33.197981 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qdvf6\" (UniqueName: \"kubernetes.io/projected/68a854c2-afda-4dbb-9ac7-06eec2c98f14-kube-api-access-qdvf6\") pod \"dnsmasq-dns-78565f64c9-vkrdf\" (UID: \"68a854c2-afda-4dbb-9ac7-06eec2c98f14\") " pod="openstack/dnsmasq-dns-78565f64c9-vkrdf" Oct 05 07:09:33 crc kubenswrapper[4846]: I1005 07:09:33.218165 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 05 07:09:33 crc kubenswrapper[4846]: I1005 07:09:33.229025 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 05 07:09:33 crc kubenswrapper[4846]: I1005 07:09:33.379450 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 05 07:09:33 crc kubenswrapper[4846]: I1005 07:09:33.394485 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78565f64c9-vkrdf" Oct 05 07:09:33 crc kubenswrapper[4846]: I1005 07:09:33.406100 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 05 07:09:33 crc kubenswrapper[4846]: I1005 07:09:33.551720 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-ptdtq"] Oct 05 07:09:33 crc kubenswrapper[4846]: I1005 07:09:33.661642 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-ptdtq" event={"ID":"a368d12f-6383-4a8e-8d71-23a675946027","Type":"ContainerStarted","Data":"7df79d3abb6a623d10ef018a06e397d24e14c8f848beb54e7a08bbcfda71c2f6"} Oct 05 07:09:33 crc kubenswrapper[4846]: I1005 07:09:33.731362 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-db-sync-s6xg9"] Oct 05 07:09:33 crc kubenswrapper[4846]: I1005 07:09:33.732860 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-s6xg9" Oct 05 07:09:33 crc kubenswrapper[4846]: I1005 07:09:33.735592 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-s6xg9"] Oct 05 07:09:33 crc kubenswrapper[4846]: I1005 07:09:33.740838 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Oct 05 07:09:33 crc kubenswrapper[4846]: I1005 07:09:33.741313 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-scripts" Oct 05 07:09:33 crc kubenswrapper[4846]: I1005 07:09:33.809931 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 05 07:09:33 crc kubenswrapper[4846]: I1005 07:09:33.892221 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d9adf2fe-f612-44f5-a7eb-e137909be76a-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-s6xg9\" (UID: \"d9adf2fe-f612-44f5-a7eb-e137909be76a\") " pod="openstack/nova-cell1-conductor-db-sync-s6xg9" Oct 05 07:09:33 crc kubenswrapper[4846]: I1005 07:09:33.892710 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d9adf2fe-f612-44f5-a7eb-e137909be76a-scripts\") pod \"nova-cell1-conductor-db-sync-s6xg9\" (UID: \"d9adf2fe-f612-44f5-a7eb-e137909be76a\") " pod="openstack/nova-cell1-conductor-db-sync-s6xg9" Oct 05 07:09:33 crc kubenswrapper[4846]: I1005 07:09:33.892796 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d9adf2fe-f612-44f5-a7eb-e137909be76a-config-data\") pod \"nova-cell1-conductor-db-sync-s6xg9\" (UID: \"d9adf2fe-f612-44f5-a7eb-e137909be76a\") " pod="openstack/nova-cell1-conductor-db-sync-s6xg9" Oct 05 07:09:33 crc kubenswrapper[4846]: I1005 07:09:33.893975 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l7vjn\" (UniqueName: \"kubernetes.io/projected/d9adf2fe-f612-44f5-a7eb-e137909be76a-kube-api-access-l7vjn\") pod \"nova-cell1-conductor-db-sync-s6xg9\" (UID: \"d9adf2fe-f612-44f5-a7eb-e137909be76a\") " pod="openstack/nova-cell1-conductor-db-sync-s6xg9" Oct 05 07:09:33 crc kubenswrapper[4846]: I1005 07:09:33.899357 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 05 07:09:33 crc kubenswrapper[4846]: I1005 07:09:33.996235 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l7vjn\" (UniqueName: \"kubernetes.io/projected/d9adf2fe-f612-44f5-a7eb-e137909be76a-kube-api-access-l7vjn\") pod \"nova-cell1-conductor-db-sync-s6xg9\" (UID: \"d9adf2fe-f612-44f5-a7eb-e137909be76a\") " pod="openstack/nova-cell1-conductor-db-sync-s6xg9" Oct 05 07:09:33 crc kubenswrapper[4846]: I1005 07:09:33.996300 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d9adf2fe-f612-44f5-a7eb-e137909be76a-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-s6xg9\" (UID: \"d9adf2fe-f612-44f5-a7eb-e137909be76a\") " pod="openstack/nova-cell1-conductor-db-sync-s6xg9" Oct 05 07:09:33 crc kubenswrapper[4846]: I1005 07:09:33.996352 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d9adf2fe-f612-44f5-a7eb-e137909be76a-scripts\") pod \"nova-cell1-conductor-db-sync-s6xg9\" (UID: \"d9adf2fe-f612-44f5-a7eb-e137909be76a\") " pod="openstack/nova-cell1-conductor-db-sync-s6xg9" Oct 05 07:09:33 crc kubenswrapper[4846]: I1005 07:09:33.996415 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d9adf2fe-f612-44f5-a7eb-e137909be76a-config-data\") pod \"nova-cell1-conductor-db-sync-s6xg9\" (UID: \"d9adf2fe-f612-44f5-a7eb-e137909be76a\") " pod="openstack/nova-cell1-conductor-db-sync-s6xg9" Oct 05 07:09:34 crc kubenswrapper[4846]: I1005 07:09:34.003485 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d9adf2fe-f612-44f5-a7eb-e137909be76a-scripts\") pod \"nova-cell1-conductor-db-sync-s6xg9\" (UID: \"d9adf2fe-f612-44f5-a7eb-e137909be76a\") " pod="openstack/nova-cell1-conductor-db-sync-s6xg9" Oct 05 07:09:34 crc kubenswrapper[4846]: I1005 07:09:34.004312 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d9adf2fe-f612-44f5-a7eb-e137909be76a-config-data\") pod \"nova-cell1-conductor-db-sync-s6xg9\" (UID: \"d9adf2fe-f612-44f5-a7eb-e137909be76a\") " pod="openstack/nova-cell1-conductor-db-sync-s6xg9" Oct 05 07:09:34 crc kubenswrapper[4846]: I1005 07:09:34.004438 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d9adf2fe-f612-44f5-a7eb-e137909be76a-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-s6xg9\" (UID: \"d9adf2fe-f612-44f5-a7eb-e137909be76a\") " pod="openstack/nova-cell1-conductor-db-sync-s6xg9" Oct 05 07:09:34 crc kubenswrapper[4846]: I1005 07:09:34.015947 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l7vjn\" (UniqueName: \"kubernetes.io/projected/d9adf2fe-f612-44f5-a7eb-e137909be76a-kube-api-access-l7vjn\") pod \"nova-cell1-conductor-db-sync-s6xg9\" (UID: \"d9adf2fe-f612-44f5-a7eb-e137909be76a\") " pod="openstack/nova-cell1-conductor-db-sync-s6xg9" Oct 05 07:09:34 crc kubenswrapper[4846]: I1005 07:09:34.165838 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 05 07:09:34 crc kubenswrapper[4846]: I1005 07:09:34.177680 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 05 07:09:34 crc kubenswrapper[4846]: I1005 07:09:34.229271 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78565f64c9-vkrdf"] Oct 05 07:09:34 crc kubenswrapper[4846]: I1005 07:09:34.289857 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-s6xg9" Oct 05 07:09:34 crc kubenswrapper[4846]: I1005 07:09:34.673898 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"ef292d07-1248-4edd-a182-9e9d572513ca","Type":"ContainerStarted","Data":"1be7426753a42cbaec966d06366b74abb18d77021d5fcffa3a4b8826192f3744"} Oct 05 07:09:34 crc kubenswrapper[4846]: I1005 07:09:34.677373 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"819bc991-94dc-4fd3-984e-2f3f711f5bf5","Type":"ContainerStarted","Data":"1555a9919a8b6582ec36e5b99d1003d080e634c3fe3debd2f75422e910944b33"} Oct 05 07:09:34 crc kubenswrapper[4846]: I1005 07:09:34.683298 4846 generic.go:334] "Generic (PLEG): container finished" podID="68a854c2-afda-4dbb-9ac7-06eec2c98f14" containerID="21988ef9f5e0ab34b6030d3b8b9a19ca844adb411f5ca6ecbb5d5f7fb49d21a6" exitCode=0 Oct 05 07:09:34 crc kubenswrapper[4846]: I1005 07:09:34.683399 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78565f64c9-vkrdf" event={"ID":"68a854c2-afda-4dbb-9ac7-06eec2c98f14","Type":"ContainerDied","Data":"21988ef9f5e0ab34b6030d3b8b9a19ca844adb411f5ca6ecbb5d5f7fb49d21a6"} Oct 05 07:09:34 crc kubenswrapper[4846]: I1005 07:09:34.683512 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78565f64c9-vkrdf" event={"ID":"68a854c2-afda-4dbb-9ac7-06eec2c98f14","Type":"ContainerStarted","Data":"76e87ad58f355f456114724a6744749e735d7ff76f24ef36aaf392dcddaa11e8"} Oct 05 07:09:34 crc kubenswrapper[4846]: I1005 07:09:34.684838 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"a0844571-8880-4fa4-af42-2da48aa8bdec","Type":"ContainerStarted","Data":"b5192909be2c780d9316ded24f975196712cb7511b0a99cabaa714a897f59917"} Oct 05 07:09:34 crc kubenswrapper[4846]: I1005 07:09:34.688695 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-ptdtq" event={"ID":"a368d12f-6383-4a8e-8d71-23a675946027","Type":"ContainerStarted","Data":"b3e205f106c607cc68f15ea01b24c2b915400791a93310bfef25ced3582cdd85"} Oct 05 07:09:34 crc kubenswrapper[4846]: I1005 07:09:34.693423 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"095eba43-2814-4476-b482-efdf6b903de4","Type":"ContainerStarted","Data":"08caf19eb603b238be913cdd839af47c24c884cc7134c517e2b4c45423bfd385"} Oct 05 07:09:34 crc kubenswrapper[4846]: I1005 07:09:34.721380 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-cell-mapping-ptdtq" podStartSLOduration=2.721359726 podStartE2EDuration="2.721359726s" podCreationTimestamp="2025-10-05 07:09:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 07:09:34.718724626 +0000 UTC m=+1296.959577411" watchObservedRunningTime="2025-10-05 07:09:34.721359726 +0000 UTC m=+1296.962212491" Oct 05 07:09:34 crc kubenswrapper[4846]: I1005 07:09:34.728893 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Oct 05 07:09:34 crc kubenswrapper[4846]: I1005 07:09:34.821854 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-s6xg9"] Oct 05 07:09:34 crc kubenswrapper[4846]: W1005 07:09:34.844322 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd9adf2fe_f612_44f5_a7eb_e137909be76a.slice/crio-e6140a414dd8024e567280f760b95bebc0041fbbbac9769da674d592e2febd92 WatchSource:0}: Error finding container e6140a414dd8024e567280f760b95bebc0041fbbbac9769da674d592e2febd92: Status 404 returned error can't find the container with id e6140a414dd8024e567280f760b95bebc0041fbbbac9769da674d592e2febd92 Oct 05 07:09:35 crc kubenswrapper[4846]: I1005 07:09:35.721323 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78565f64c9-vkrdf" event={"ID":"68a854c2-afda-4dbb-9ac7-06eec2c98f14","Type":"ContainerStarted","Data":"96643ce5b5a2f18ab40b26a3bea0d27ff26ce9e8cfedae27efdb2d0b75872aa3"} Oct 05 07:09:35 crc kubenswrapper[4846]: I1005 07:09:35.721988 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-78565f64c9-vkrdf" Oct 05 07:09:35 crc kubenswrapper[4846]: I1005 07:09:35.726653 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-s6xg9" event={"ID":"d9adf2fe-f612-44f5-a7eb-e137909be76a","Type":"ContainerStarted","Data":"f0993f040dcd6d3145256a44cdc38fc8db9c44f334ee6dc7c33895edf97422b9"} Oct 05 07:09:35 crc kubenswrapper[4846]: I1005 07:09:35.726703 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-s6xg9" event={"ID":"d9adf2fe-f612-44f5-a7eb-e137909be76a","Type":"ContainerStarted","Data":"e6140a414dd8024e567280f760b95bebc0041fbbbac9769da674d592e2febd92"} Oct 05 07:09:35 crc kubenswrapper[4846]: I1005 07:09:35.751298 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-78565f64c9-vkrdf" podStartSLOduration=3.751281037 podStartE2EDuration="3.751281037s" podCreationTimestamp="2025-10-05 07:09:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 07:09:35.748722749 +0000 UTC m=+1297.989575524" watchObservedRunningTime="2025-10-05 07:09:35.751281037 +0000 UTC m=+1297.992133812" Oct 05 07:09:35 crc kubenswrapper[4846]: I1005 07:09:35.771677 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-db-sync-s6xg9" podStartSLOduration=2.771656701 podStartE2EDuration="2.771656701s" podCreationTimestamp="2025-10-05 07:09:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 07:09:35.763886184 +0000 UTC m=+1298.004738959" watchObservedRunningTime="2025-10-05 07:09:35.771656701 +0000 UTC m=+1298.012509476" Oct 05 07:09:36 crc kubenswrapper[4846]: I1005 07:09:36.269493 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 05 07:09:36 crc kubenswrapper[4846]: I1005 07:09:36.292952 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 05 07:09:38 crc kubenswrapper[4846]: I1005 07:09:38.767924 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"819bc991-94dc-4fd3-984e-2f3f711f5bf5","Type":"ContainerStarted","Data":"2260d1fd93897c54f1c80e3d8c3d8a24b2c4ec4f8e5fa9495a23a14fa6d6ceef"} Oct 05 07:09:38 crc kubenswrapper[4846]: I1005 07:09:38.768774 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"819bc991-94dc-4fd3-984e-2f3f711f5bf5","Type":"ContainerStarted","Data":"723bbe425d7fae5972a2ced59b306e826e92dc8947d430b29ecddcc70824fa7f"} Oct 05 07:09:38 crc kubenswrapper[4846]: I1005 07:09:38.768927 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="819bc991-94dc-4fd3-984e-2f3f711f5bf5" containerName="nova-metadata-log" containerID="cri-o://723bbe425d7fae5972a2ced59b306e826e92dc8947d430b29ecddcc70824fa7f" gracePeriod=30 Oct 05 07:09:38 crc kubenswrapper[4846]: I1005 07:09:38.769880 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="819bc991-94dc-4fd3-984e-2f3f711f5bf5" containerName="nova-metadata-metadata" containerID="cri-o://2260d1fd93897c54f1c80e3d8c3d8a24b2c4ec4f8e5fa9495a23a14fa6d6ceef" gracePeriod=30 Oct 05 07:09:38 crc kubenswrapper[4846]: I1005 07:09:38.772724 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"a0844571-8880-4fa4-af42-2da48aa8bdec","Type":"ContainerStarted","Data":"cb184b8b7a1c9f825f5c48f4b882e586772ff863671c32a461519f9e717aab6e"} Oct 05 07:09:38 crc kubenswrapper[4846]: I1005 07:09:38.772836 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="a0844571-8880-4fa4-af42-2da48aa8bdec" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://cb184b8b7a1c9f825f5c48f4b882e586772ff863671c32a461519f9e717aab6e" gracePeriod=30 Oct 05 07:09:38 crc kubenswrapper[4846]: I1005 07:09:38.778319 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"095eba43-2814-4476-b482-efdf6b903de4","Type":"ContainerStarted","Data":"adcdee70fed38ca32d9ea08aa0872b7e277f0811e3da49432a923b205c6732a0"} Oct 05 07:09:38 crc kubenswrapper[4846]: I1005 07:09:38.778376 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"095eba43-2814-4476-b482-efdf6b903de4","Type":"ContainerStarted","Data":"a5ed027b4a2ba75551446c6b871445f74de3480c300dba8c804dc3238212ddc8"} Oct 05 07:09:38 crc kubenswrapper[4846]: I1005 07:09:38.780910 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"ef292d07-1248-4edd-a182-9e9d572513ca","Type":"ContainerStarted","Data":"606907b830a2f5a35bf2b24719a4cd8e8b2a45dfc0e282ccce0bede035c53184"} Oct 05 07:09:38 crc kubenswrapper[4846]: I1005 07:09:38.800447 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=3.122901735 podStartE2EDuration="6.800425407s" podCreationTimestamp="2025-10-05 07:09:32 +0000 UTC" firstStartedPulling="2025-10-05 07:09:33.904476002 +0000 UTC m=+1296.145328777" lastFinishedPulling="2025-10-05 07:09:37.581999644 +0000 UTC m=+1299.822852449" observedRunningTime="2025-10-05 07:09:38.794989382 +0000 UTC m=+1301.035842147" watchObservedRunningTime="2025-10-05 07:09:38.800425407 +0000 UTC m=+1301.041278182" Oct 05 07:09:38 crc kubenswrapper[4846]: I1005 07:09:38.825401 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=3.066883179 podStartE2EDuration="6.825366232s" podCreationTimestamp="2025-10-05 07:09:32 +0000 UTC" firstStartedPulling="2025-10-05 07:09:33.821075406 +0000 UTC m=+1296.061928181" lastFinishedPulling="2025-10-05 07:09:37.579558459 +0000 UTC m=+1299.820411234" observedRunningTime="2025-10-05 07:09:38.816352822 +0000 UTC m=+1301.057205607" watchObservedRunningTime="2025-10-05 07:09:38.825366232 +0000 UTC m=+1301.066219027" Oct 05 07:09:38 crc kubenswrapper[4846]: I1005 07:09:38.845500 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=3.435690413 podStartE2EDuration="6.845471559s" podCreationTimestamp="2025-10-05 07:09:32 +0000 UTC" firstStartedPulling="2025-10-05 07:09:34.172333742 +0000 UTC m=+1296.413186517" lastFinishedPulling="2025-10-05 07:09:37.582114888 +0000 UTC m=+1299.822967663" observedRunningTime="2025-10-05 07:09:38.839729216 +0000 UTC m=+1301.080581991" watchObservedRunningTime="2025-10-05 07:09:38.845471559 +0000 UTC m=+1301.086324334" Oct 05 07:09:38 crc kubenswrapper[4846]: I1005 07:09:38.867656 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=3.398487 podStartE2EDuration="6.86762487s" podCreationTimestamp="2025-10-05 07:09:32 +0000 UTC" firstStartedPulling="2025-10-05 07:09:34.135891099 +0000 UTC m=+1296.376743874" lastFinishedPulling="2025-10-05 07:09:37.605028969 +0000 UTC m=+1299.845881744" observedRunningTime="2025-10-05 07:09:38.861807905 +0000 UTC m=+1301.102660700" watchObservedRunningTime="2025-10-05 07:09:38.86762487 +0000 UTC m=+1301.108477655" Oct 05 07:09:39 crc kubenswrapper[4846]: I1005 07:09:39.529559 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 05 07:09:39 crc kubenswrapper[4846]: I1005 07:09:39.530212 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="31c9cd11-358f-4b1e-b661-dfb09f480277" containerName="kube-state-metrics" containerID="cri-o://72b415d01c8c5a8bc6919241cfdcfd6680afd1a0a3a7d66fc08eba56f1c5828d" gracePeriod=30 Oct 05 07:09:39 crc kubenswrapper[4846]: I1005 07:09:39.809307 4846 generic.go:334] "Generic (PLEG): container finished" podID="819bc991-94dc-4fd3-984e-2f3f711f5bf5" containerID="2260d1fd93897c54f1c80e3d8c3d8a24b2c4ec4f8e5fa9495a23a14fa6d6ceef" exitCode=0 Oct 05 07:09:39 crc kubenswrapper[4846]: I1005 07:09:39.809355 4846 generic.go:334] "Generic (PLEG): container finished" podID="819bc991-94dc-4fd3-984e-2f3f711f5bf5" containerID="723bbe425d7fae5972a2ced59b306e826e92dc8947d430b29ecddcc70824fa7f" exitCode=143 Oct 05 07:09:39 crc kubenswrapper[4846]: I1005 07:09:39.809415 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"819bc991-94dc-4fd3-984e-2f3f711f5bf5","Type":"ContainerDied","Data":"2260d1fd93897c54f1c80e3d8c3d8a24b2c4ec4f8e5fa9495a23a14fa6d6ceef"} Oct 05 07:09:39 crc kubenswrapper[4846]: I1005 07:09:39.809454 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"819bc991-94dc-4fd3-984e-2f3f711f5bf5","Type":"ContainerDied","Data":"723bbe425d7fae5972a2ced59b306e826e92dc8947d430b29ecddcc70824fa7f"} Oct 05 07:09:39 crc kubenswrapper[4846]: I1005 07:09:39.812318 4846 generic.go:334] "Generic (PLEG): container finished" podID="31c9cd11-358f-4b1e-b661-dfb09f480277" containerID="72b415d01c8c5a8bc6919241cfdcfd6680afd1a0a3a7d66fc08eba56f1c5828d" exitCode=2 Oct 05 07:09:39 crc kubenswrapper[4846]: I1005 07:09:39.812911 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"31c9cd11-358f-4b1e-b661-dfb09f480277","Type":"ContainerDied","Data":"72b415d01c8c5a8bc6919241cfdcfd6680afd1a0a3a7d66fc08eba56f1c5828d"} Oct 05 07:09:39 crc kubenswrapper[4846]: I1005 07:09:39.982495 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 05 07:09:40 crc kubenswrapper[4846]: I1005 07:09:40.131602 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 05 07:09:40 crc kubenswrapper[4846]: I1005 07:09:40.173139 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mvtzk\" (UniqueName: \"kubernetes.io/projected/819bc991-94dc-4fd3-984e-2f3f711f5bf5-kube-api-access-mvtzk\") pod \"819bc991-94dc-4fd3-984e-2f3f711f5bf5\" (UID: \"819bc991-94dc-4fd3-984e-2f3f711f5bf5\") " Oct 05 07:09:40 crc kubenswrapper[4846]: I1005 07:09:40.173318 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/819bc991-94dc-4fd3-984e-2f3f711f5bf5-config-data\") pod \"819bc991-94dc-4fd3-984e-2f3f711f5bf5\" (UID: \"819bc991-94dc-4fd3-984e-2f3f711f5bf5\") " Oct 05 07:09:40 crc kubenswrapper[4846]: I1005 07:09:40.173442 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/819bc991-94dc-4fd3-984e-2f3f711f5bf5-combined-ca-bundle\") pod \"819bc991-94dc-4fd3-984e-2f3f711f5bf5\" (UID: \"819bc991-94dc-4fd3-984e-2f3f711f5bf5\") " Oct 05 07:09:40 crc kubenswrapper[4846]: I1005 07:09:40.173494 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/819bc991-94dc-4fd3-984e-2f3f711f5bf5-logs\") pod \"819bc991-94dc-4fd3-984e-2f3f711f5bf5\" (UID: \"819bc991-94dc-4fd3-984e-2f3f711f5bf5\") " Oct 05 07:09:40 crc kubenswrapper[4846]: I1005 07:09:40.174597 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/819bc991-94dc-4fd3-984e-2f3f711f5bf5-logs" (OuterVolumeSpecName: "logs") pod "819bc991-94dc-4fd3-984e-2f3f711f5bf5" (UID: "819bc991-94dc-4fd3-984e-2f3f711f5bf5"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 07:09:40 crc kubenswrapper[4846]: I1005 07:09:40.181639 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/819bc991-94dc-4fd3-984e-2f3f711f5bf5-kube-api-access-mvtzk" (OuterVolumeSpecName: "kube-api-access-mvtzk") pod "819bc991-94dc-4fd3-984e-2f3f711f5bf5" (UID: "819bc991-94dc-4fd3-984e-2f3f711f5bf5"). InnerVolumeSpecName "kube-api-access-mvtzk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:09:40 crc kubenswrapper[4846]: I1005 07:09:40.213314 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/819bc991-94dc-4fd3-984e-2f3f711f5bf5-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "819bc991-94dc-4fd3-984e-2f3f711f5bf5" (UID: "819bc991-94dc-4fd3-984e-2f3f711f5bf5"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:09:40 crc kubenswrapper[4846]: I1005 07:09:40.218281 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/819bc991-94dc-4fd3-984e-2f3f711f5bf5-config-data" (OuterVolumeSpecName: "config-data") pod "819bc991-94dc-4fd3-984e-2f3f711f5bf5" (UID: "819bc991-94dc-4fd3-984e-2f3f711f5bf5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:09:40 crc kubenswrapper[4846]: I1005 07:09:40.276008 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7clx2\" (UniqueName: \"kubernetes.io/projected/31c9cd11-358f-4b1e-b661-dfb09f480277-kube-api-access-7clx2\") pod \"31c9cd11-358f-4b1e-b661-dfb09f480277\" (UID: \"31c9cd11-358f-4b1e-b661-dfb09f480277\") " Oct 05 07:09:40 crc kubenswrapper[4846]: I1005 07:09:40.277076 4846 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/819bc991-94dc-4fd3-984e-2f3f711f5bf5-logs\") on node \"crc\" DevicePath \"\"" Oct 05 07:09:40 crc kubenswrapper[4846]: I1005 07:09:40.277118 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mvtzk\" (UniqueName: \"kubernetes.io/projected/819bc991-94dc-4fd3-984e-2f3f711f5bf5-kube-api-access-mvtzk\") on node \"crc\" DevicePath \"\"" Oct 05 07:09:40 crc kubenswrapper[4846]: I1005 07:09:40.277138 4846 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/819bc991-94dc-4fd3-984e-2f3f711f5bf5-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 07:09:40 crc kubenswrapper[4846]: I1005 07:09:40.277156 4846 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/819bc991-94dc-4fd3-984e-2f3f711f5bf5-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 07:09:40 crc kubenswrapper[4846]: I1005 07:09:40.282512 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31c9cd11-358f-4b1e-b661-dfb09f480277-kube-api-access-7clx2" (OuterVolumeSpecName: "kube-api-access-7clx2") pod "31c9cd11-358f-4b1e-b661-dfb09f480277" (UID: "31c9cd11-358f-4b1e-b661-dfb09f480277"). InnerVolumeSpecName "kube-api-access-7clx2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:09:40 crc kubenswrapper[4846]: I1005 07:09:40.379876 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7clx2\" (UniqueName: \"kubernetes.io/projected/31c9cd11-358f-4b1e-b661-dfb09f480277-kube-api-access-7clx2\") on node \"crc\" DevicePath \"\"" Oct 05 07:09:40 crc kubenswrapper[4846]: I1005 07:09:40.836164 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"31c9cd11-358f-4b1e-b661-dfb09f480277","Type":"ContainerDied","Data":"ba731ef3d0b7d56aa452c33dbc344f20625eab2da03b9960c336bb4e1c31388a"} Oct 05 07:09:40 crc kubenswrapper[4846]: I1005 07:09:40.836698 4846 scope.go:117] "RemoveContainer" containerID="72b415d01c8c5a8bc6919241cfdcfd6680afd1a0a3a7d66fc08eba56f1c5828d" Oct 05 07:09:40 crc kubenswrapper[4846]: I1005 07:09:40.837042 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 05 07:09:40 crc kubenswrapper[4846]: I1005 07:09:40.845973 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"819bc991-94dc-4fd3-984e-2f3f711f5bf5","Type":"ContainerDied","Data":"1555a9919a8b6582ec36e5b99d1003d080e634c3fe3debd2f75422e910944b33"} Oct 05 07:09:40 crc kubenswrapper[4846]: I1005 07:09:40.846149 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 05 07:09:40 crc kubenswrapper[4846]: I1005 07:09:40.875248 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 05 07:09:40 crc kubenswrapper[4846]: I1005 07:09:40.881540 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 05 07:09:40 crc kubenswrapper[4846]: I1005 07:09:40.889163 4846 scope.go:117] "RemoveContainer" containerID="2260d1fd93897c54f1c80e3d8c3d8a24b2c4ec4f8e5fa9495a23a14fa6d6ceef" Oct 05 07:09:40 crc kubenswrapper[4846]: I1005 07:09:40.926951 4846 scope.go:117] "RemoveContainer" containerID="723bbe425d7fae5972a2ced59b306e826e92dc8947d430b29ecddcc70824fa7f" Oct 05 07:09:40 crc kubenswrapper[4846]: I1005 07:09:40.947341 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Oct 05 07:09:40 crc kubenswrapper[4846]: E1005 07:09:40.947909 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="819bc991-94dc-4fd3-984e-2f3f711f5bf5" containerName="nova-metadata-log" Oct 05 07:09:40 crc kubenswrapper[4846]: I1005 07:09:40.947928 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="819bc991-94dc-4fd3-984e-2f3f711f5bf5" containerName="nova-metadata-log" Oct 05 07:09:40 crc kubenswrapper[4846]: E1005 07:09:40.947944 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="819bc991-94dc-4fd3-984e-2f3f711f5bf5" containerName="nova-metadata-metadata" Oct 05 07:09:40 crc kubenswrapper[4846]: I1005 07:09:40.947950 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="819bc991-94dc-4fd3-984e-2f3f711f5bf5" containerName="nova-metadata-metadata" Oct 05 07:09:40 crc kubenswrapper[4846]: E1005 07:09:40.947977 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="31c9cd11-358f-4b1e-b661-dfb09f480277" containerName="kube-state-metrics" Oct 05 07:09:40 crc kubenswrapper[4846]: I1005 07:09:40.947984 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="31c9cd11-358f-4b1e-b661-dfb09f480277" containerName="kube-state-metrics" Oct 05 07:09:40 crc kubenswrapper[4846]: I1005 07:09:40.948200 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="819bc991-94dc-4fd3-984e-2f3f711f5bf5" containerName="nova-metadata-log" Oct 05 07:09:40 crc kubenswrapper[4846]: I1005 07:09:40.948215 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="31c9cd11-358f-4b1e-b661-dfb09f480277" containerName="kube-state-metrics" Oct 05 07:09:40 crc kubenswrapper[4846]: I1005 07:09:40.948234 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="819bc991-94dc-4fd3-984e-2f3f711f5bf5" containerName="nova-metadata-metadata" Oct 05 07:09:40 crc kubenswrapper[4846]: I1005 07:09:40.949011 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 05 07:09:40 crc kubenswrapper[4846]: I1005 07:09:40.953147 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"kube-state-metrics-tls-config" Oct 05 07:09:40 crc kubenswrapper[4846]: I1005 07:09:40.953726 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-kube-state-metrics-svc" Oct 05 07:09:40 crc kubenswrapper[4846]: I1005 07:09:40.972542 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 05 07:09:40 crc kubenswrapper[4846]: I1005 07:09:40.987292 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Oct 05 07:09:40 crc kubenswrapper[4846]: I1005 07:09:40.996332 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 05 07:09:41 crc kubenswrapper[4846]: I1005 07:09:41.005024 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 05 07:09:41 crc kubenswrapper[4846]: I1005 07:09:41.007524 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 05 07:09:41 crc kubenswrapper[4846]: I1005 07:09:41.010219 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Oct 05 07:09:41 crc kubenswrapper[4846]: I1005 07:09:41.010432 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 05 07:09:41 crc kubenswrapper[4846]: I1005 07:09:41.014500 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 05 07:09:41 crc kubenswrapper[4846]: I1005 07:09:41.095777 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/57b09608-08e5-4998-a3b3-e88b13587f58-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"57b09608-08e5-4998-a3b3-e88b13587f58\") " pod="openstack/kube-state-metrics-0" Oct 05 07:09:41 crc kubenswrapper[4846]: I1005 07:09:41.095842 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5lbbc\" (UniqueName: \"kubernetes.io/projected/57b09608-08e5-4998-a3b3-e88b13587f58-kube-api-access-5lbbc\") pod \"kube-state-metrics-0\" (UID: \"57b09608-08e5-4998-a3b3-e88b13587f58\") " pod="openstack/kube-state-metrics-0" Oct 05 07:09:41 crc kubenswrapper[4846]: I1005 07:09:41.095924 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/57b09608-08e5-4998-a3b3-e88b13587f58-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"57b09608-08e5-4998-a3b3-e88b13587f58\") " pod="openstack/kube-state-metrics-0" Oct 05 07:09:41 crc kubenswrapper[4846]: I1005 07:09:41.096114 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/57b09608-08e5-4998-a3b3-e88b13587f58-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"57b09608-08e5-4998-a3b3-e88b13587f58\") " pod="openstack/kube-state-metrics-0" Oct 05 07:09:41 crc kubenswrapper[4846]: I1005 07:09:41.199001 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/2682f5d9-2987-4dad-86e2-750b48a2e697-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"2682f5d9-2987-4dad-86e2-750b48a2e697\") " pod="openstack/nova-metadata-0" Oct 05 07:09:41 crc kubenswrapper[4846]: I1005 07:09:41.199137 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2682f5d9-2987-4dad-86e2-750b48a2e697-config-data\") pod \"nova-metadata-0\" (UID: \"2682f5d9-2987-4dad-86e2-750b48a2e697\") " pod="openstack/nova-metadata-0" Oct 05 07:09:41 crc kubenswrapper[4846]: I1005 07:09:41.199333 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/57b09608-08e5-4998-a3b3-e88b13587f58-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"57b09608-08e5-4998-a3b3-e88b13587f58\") " pod="openstack/kube-state-metrics-0" Oct 05 07:09:41 crc kubenswrapper[4846]: I1005 07:09:41.199382 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/57b09608-08e5-4998-a3b3-e88b13587f58-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"57b09608-08e5-4998-a3b3-e88b13587f58\") " pod="openstack/kube-state-metrics-0" Oct 05 07:09:41 crc kubenswrapper[4846]: I1005 07:09:41.199415 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2682f5d9-2987-4dad-86e2-750b48a2e697-logs\") pod \"nova-metadata-0\" (UID: \"2682f5d9-2987-4dad-86e2-750b48a2e697\") " pod="openstack/nova-metadata-0" Oct 05 07:09:41 crc kubenswrapper[4846]: I1005 07:09:41.200669 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zr6b4\" (UniqueName: \"kubernetes.io/projected/2682f5d9-2987-4dad-86e2-750b48a2e697-kube-api-access-zr6b4\") pod \"nova-metadata-0\" (UID: \"2682f5d9-2987-4dad-86e2-750b48a2e697\") " pod="openstack/nova-metadata-0" Oct 05 07:09:41 crc kubenswrapper[4846]: I1005 07:09:41.200933 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/57b09608-08e5-4998-a3b3-e88b13587f58-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"57b09608-08e5-4998-a3b3-e88b13587f58\") " pod="openstack/kube-state-metrics-0" Oct 05 07:09:41 crc kubenswrapper[4846]: I1005 07:09:41.200990 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2682f5d9-2987-4dad-86e2-750b48a2e697-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"2682f5d9-2987-4dad-86e2-750b48a2e697\") " pod="openstack/nova-metadata-0" Oct 05 07:09:41 crc kubenswrapper[4846]: I1005 07:09:41.201022 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5lbbc\" (UniqueName: \"kubernetes.io/projected/57b09608-08e5-4998-a3b3-e88b13587f58-kube-api-access-5lbbc\") pod \"kube-state-metrics-0\" (UID: \"57b09608-08e5-4998-a3b3-e88b13587f58\") " pod="openstack/kube-state-metrics-0" Oct 05 07:09:41 crc kubenswrapper[4846]: I1005 07:09:41.206453 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/57b09608-08e5-4998-a3b3-e88b13587f58-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"57b09608-08e5-4998-a3b3-e88b13587f58\") " pod="openstack/kube-state-metrics-0" Oct 05 07:09:41 crc kubenswrapper[4846]: I1005 07:09:41.206767 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/57b09608-08e5-4998-a3b3-e88b13587f58-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"57b09608-08e5-4998-a3b3-e88b13587f58\") " pod="openstack/kube-state-metrics-0" Oct 05 07:09:41 crc kubenswrapper[4846]: I1005 07:09:41.207643 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/57b09608-08e5-4998-a3b3-e88b13587f58-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"57b09608-08e5-4998-a3b3-e88b13587f58\") " pod="openstack/kube-state-metrics-0" Oct 05 07:09:41 crc kubenswrapper[4846]: I1005 07:09:41.224051 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5lbbc\" (UniqueName: \"kubernetes.io/projected/57b09608-08e5-4998-a3b3-e88b13587f58-kube-api-access-5lbbc\") pod \"kube-state-metrics-0\" (UID: \"57b09608-08e5-4998-a3b3-e88b13587f58\") " pod="openstack/kube-state-metrics-0" Oct 05 07:09:41 crc kubenswrapper[4846]: I1005 07:09:41.278218 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 05 07:09:41 crc kubenswrapper[4846]: I1005 07:09:41.305348 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/2682f5d9-2987-4dad-86e2-750b48a2e697-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"2682f5d9-2987-4dad-86e2-750b48a2e697\") " pod="openstack/nova-metadata-0" Oct 05 07:09:41 crc kubenswrapper[4846]: I1005 07:09:41.305440 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2682f5d9-2987-4dad-86e2-750b48a2e697-config-data\") pod \"nova-metadata-0\" (UID: \"2682f5d9-2987-4dad-86e2-750b48a2e697\") " pod="openstack/nova-metadata-0" Oct 05 07:09:41 crc kubenswrapper[4846]: I1005 07:09:41.305552 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2682f5d9-2987-4dad-86e2-750b48a2e697-logs\") pod \"nova-metadata-0\" (UID: \"2682f5d9-2987-4dad-86e2-750b48a2e697\") " pod="openstack/nova-metadata-0" Oct 05 07:09:41 crc kubenswrapper[4846]: I1005 07:09:41.305615 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zr6b4\" (UniqueName: \"kubernetes.io/projected/2682f5d9-2987-4dad-86e2-750b48a2e697-kube-api-access-zr6b4\") pod \"nova-metadata-0\" (UID: \"2682f5d9-2987-4dad-86e2-750b48a2e697\") " pod="openstack/nova-metadata-0" Oct 05 07:09:41 crc kubenswrapper[4846]: I1005 07:09:41.305780 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2682f5d9-2987-4dad-86e2-750b48a2e697-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"2682f5d9-2987-4dad-86e2-750b48a2e697\") " pod="openstack/nova-metadata-0" Oct 05 07:09:41 crc kubenswrapper[4846]: I1005 07:09:41.310011 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2682f5d9-2987-4dad-86e2-750b48a2e697-logs\") pod \"nova-metadata-0\" (UID: \"2682f5d9-2987-4dad-86e2-750b48a2e697\") " pod="openstack/nova-metadata-0" Oct 05 07:09:41 crc kubenswrapper[4846]: I1005 07:09:41.310659 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2682f5d9-2987-4dad-86e2-750b48a2e697-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"2682f5d9-2987-4dad-86e2-750b48a2e697\") " pod="openstack/nova-metadata-0" Oct 05 07:09:41 crc kubenswrapper[4846]: I1005 07:09:41.311830 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2682f5d9-2987-4dad-86e2-750b48a2e697-config-data\") pod \"nova-metadata-0\" (UID: \"2682f5d9-2987-4dad-86e2-750b48a2e697\") " pod="openstack/nova-metadata-0" Oct 05 07:09:41 crc kubenswrapper[4846]: I1005 07:09:41.314837 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/2682f5d9-2987-4dad-86e2-750b48a2e697-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"2682f5d9-2987-4dad-86e2-750b48a2e697\") " pod="openstack/nova-metadata-0" Oct 05 07:09:41 crc kubenswrapper[4846]: I1005 07:09:41.331093 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zr6b4\" (UniqueName: \"kubernetes.io/projected/2682f5d9-2987-4dad-86e2-750b48a2e697-kube-api-access-zr6b4\") pod \"nova-metadata-0\" (UID: \"2682f5d9-2987-4dad-86e2-750b48a2e697\") " pod="openstack/nova-metadata-0" Oct 05 07:09:41 crc kubenswrapper[4846]: I1005 07:09:41.517902 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 05 07:09:41 crc kubenswrapper[4846]: I1005 07:09:41.518636 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="06e92a1c-15bf-4cdb-a284-0b5bee2a75c9" containerName="ceilometer-central-agent" containerID="cri-o://0a02b9d96618505f33de0057a4a6c5bb3f7d078b7697e171d7309d0fb4459179" gracePeriod=30 Oct 05 07:09:41 crc kubenswrapper[4846]: I1005 07:09:41.519313 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="06e92a1c-15bf-4cdb-a284-0b5bee2a75c9" containerName="proxy-httpd" containerID="cri-o://1366799532bcfbd4e3db914bbc7d301c1fffe5b0fe8364f08939c0180dd80492" gracePeriod=30 Oct 05 07:09:41 crc kubenswrapper[4846]: I1005 07:09:41.519423 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="06e92a1c-15bf-4cdb-a284-0b5bee2a75c9" containerName="sg-core" containerID="cri-o://e8da228ce53f0f9b4a512fcd59c901ab9bd35018ddae6a4efd69d1cb97fabdfd" gracePeriod=30 Oct 05 07:09:41 crc kubenswrapper[4846]: I1005 07:09:41.519460 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="06e92a1c-15bf-4cdb-a284-0b5bee2a75c9" containerName="ceilometer-notification-agent" containerID="cri-o://8a4dc29d9a82877eba8bfc9d6ac07fb48d1a0e553bef761e5a88bb04c88cbe6b" gracePeriod=30 Oct 05 07:09:41 crc kubenswrapper[4846]: I1005 07:09:41.629653 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 05 07:09:41 crc kubenswrapper[4846]: I1005 07:09:41.769760 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 05 07:09:41 crc kubenswrapper[4846]: I1005 07:09:41.874778 4846 generic.go:334] "Generic (PLEG): container finished" podID="06e92a1c-15bf-4cdb-a284-0b5bee2a75c9" containerID="1366799532bcfbd4e3db914bbc7d301c1fffe5b0fe8364f08939c0180dd80492" exitCode=0 Oct 05 07:09:41 crc kubenswrapper[4846]: I1005 07:09:41.875244 4846 generic.go:334] "Generic (PLEG): container finished" podID="06e92a1c-15bf-4cdb-a284-0b5bee2a75c9" containerID="e8da228ce53f0f9b4a512fcd59c901ab9bd35018ddae6a4efd69d1cb97fabdfd" exitCode=2 Oct 05 07:09:41 crc kubenswrapper[4846]: I1005 07:09:41.874887 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"06e92a1c-15bf-4cdb-a284-0b5bee2a75c9","Type":"ContainerDied","Data":"1366799532bcfbd4e3db914bbc7d301c1fffe5b0fe8364f08939c0180dd80492"} Oct 05 07:09:41 crc kubenswrapper[4846]: I1005 07:09:41.875331 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"06e92a1c-15bf-4cdb-a284-0b5bee2a75c9","Type":"ContainerDied","Data":"e8da228ce53f0f9b4a512fcd59c901ab9bd35018ddae6a4efd69d1cb97fabdfd"} Oct 05 07:09:41 crc kubenswrapper[4846]: I1005 07:09:41.878152 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"57b09608-08e5-4998-a3b3-e88b13587f58","Type":"ContainerStarted","Data":"56ffad5ab9e3eaf08b4a3a4ff530c7b1dd85702163911520bca3b753e9901952"} Oct 05 07:09:42 crc kubenswrapper[4846]: I1005 07:09:42.192829 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 05 07:09:42 crc kubenswrapper[4846]: I1005 07:09:42.514064 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31c9cd11-358f-4b1e-b661-dfb09f480277" path="/var/lib/kubelet/pods/31c9cd11-358f-4b1e-b661-dfb09f480277/volumes" Oct 05 07:09:42 crc kubenswrapper[4846]: I1005 07:09:42.514690 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="819bc991-94dc-4fd3-984e-2f3f711f5bf5" path="/var/lib/kubelet/pods/819bc991-94dc-4fd3-984e-2f3f711f5bf5/volumes" Oct 05 07:09:42 crc kubenswrapper[4846]: I1005 07:09:42.895646 4846 generic.go:334] "Generic (PLEG): container finished" podID="06e92a1c-15bf-4cdb-a284-0b5bee2a75c9" containerID="0a02b9d96618505f33de0057a4a6c5bb3f7d078b7697e171d7309d0fb4459179" exitCode=0 Oct 05 07:09:42 crc kubenswrapper[4846]: I1005 07:09:42.895726 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"06e92a1c-15bf-4cdb-a284-0b5bee2a75c9","Type":"ContainerDied","Data":"0a02b9d96618505f33de0057a4a6c5bb3f7d078b7697e171d7309d0fb4459179"} Oct 05 07:09:42 crc kubenswrapper[4846]: I1005 07:09:42.899138 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"57b09608-08e5-4998-a3b3-e88b13587f58","Type":"ContainerStarted","Data":"0f36596d6da69b2a28102d8a73a67b3c631900ebfdb71663f52974e9dd99b1d1"} Oct 05 07:09:42 crc kubenswrapper[4846]: I1005 07:09:42.905066 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"2682f5d9-2987-4dad-86e2-750b48a2e697","Type":"ContainerStarted","Data":"d7c56637eb3aea87669012b06d0890cc0603d7e208e71bb55c05dbd4e172feed"} Oct 05 07:09:42 crc kubenswrapper[4846]: I1005 07:09:42.905124 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"2682f5d9-2987-4dad-86e2-750b48a2e697","Type":"ContainerStarted","Data":"0c53ca400292db7449e73c76ec3ad77058c533d8fe6bf07b8509da0f332598cc"} Oct 05 07:09:42 crc kubenswrapper[4846]: I1005 07:09:42.905139 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"2682f5d9-2987-4dad-86e2-750b48a2e697","Type":"ContainerStarted","Data":"6f90f20ac50ffa971df116110da7c9581e170038ad7c8f1902aa1de25c3bf3ae"} Oct 05 07:09:42 crc kubenswrapper[4846]: I1005 07:09:42.925385 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=2.576580092 podStartE2EDuration="2.925348331s" podCreationTimestamp="2025-10-05 07:09:40 +0000 UTC" firstStartedPulling="2025-10-05 07:09:41.794488155 +0000 UTC m=+1304.035340931" lastFinishedPulling="2025-10-05 07:09:42.143256395 +0000 UTC m=+1304.384109170" observedRunningTime="2025-10-05 07:09:42.917848391 +0000 UTC m=+1305.158701226" watchObservedRunningTime="2025-10-05 07:09:42.925348331 +0000 UTC m=+1305.166201106" Oct 05 07:09:42 crc kubenswrapper[4846]: I1005 07:09:42.962604 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.962583505 podStartE2EDuration="2.962583505s" podCreationTimestamp="2025-10-05 07:09:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 07:09:42.950098242 +0000 UTC m=+1305.190951027" watchObservedRunningTime="2025-10-05 07:09:42.962583505 +0000 UTC m=+1305.203436290" Oct 05 07:09:43 crc kubenswrapper[4846]: I1005 07:09:43.219105 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 05 07:09:43 crc kubenswrapper[4846]: I1005 07:09:43.219281 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 05 07:09:43 crc kubenswrapper[4846]: I1005 07:09:43.379835 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Oct 05 07:09:43 crc kubenswrapper[4846]: I1005 07:09:43.379892 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Oct 05 07:09:43 crc kubenswrapper[4846]: I1005 07:09:43.397385 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-78565f64c9-vkrdf" Oct 05 07:09:43 crc kubenswrapper[4846]: I1005 07:09:43.407126 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Oct 05 07:09:43 crc kubenswrapper[4846]: I1005 07:09:43.432668 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Oct 05 07:09:43 crc kubenswrapper[4846]: I1005 07:09:43.484307 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-99495cbc7-v87cl"] Oct 05 07:09:43 crc kubenswrapper[4846]: I1005 07:09:43.484653 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-99495cbc7-v87cl" podUID="507e14a3-601e-461b-b8a7-f218b87b0f47" containerName="dnsmasq-dns" containerID="cri-o://efc09bde05d0841f423bbc79bab49db3b7705ff627f508b06fdf9dfef64469a0" gracePeriod=10 Oct 05 07:09:43 crc kubenswrapper[4846]: I1005 07:09:43.927586 4846 generic.go:334] "Generic (PLEG): container finished" podID="507e14a3-601e-461b-b8a7-f218b87b0f47" containerID="efc09bde05d0841f423bbc79bab49db3b7705ff627f508b06fdf9dfef64469a0" exitCode=0 Oct 05 07:09:43 crc kubenswrapper[4846]: I1005 07:09:43.927778 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-99495cbc7-v87cl" event={"ID":"507e14a3-601e-461b-b8a7-f218b87b0f47","Type":"ContainerDied","Data":"efc09bde05d0841f423bbc79bab49db3b7705ff627f508b06fdf9dfef64469a0"} Oct 05 07:09:43 crc kubenswrapper[4846]: I1005 07:09:43.932546 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Oct 05 07:09:43 crc kubenswrapper[4846]: I1005 07:09:43.963597 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Oct 05 07:09:44 crc kubenswrapper[4846]: I1005 07:09:44.210222 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-99495cbc7-v87cl" Oct 05 07:09:44 crc kubenswrapper[4846]: I1005 07:09:44.304575 4846 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="095eba43-2814-4476-b482-efdf6b903de4" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.187:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 05 07:09:44 crc kubenswrapper[4846]: I1005 07:09:44.304627 4846 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="095eba43-2814-4476-b482-efdf6b903de4" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.187:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 05 07:09:44 crc kubenswrapper[4846]: I1005 07:09:44.312371 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/507e14a3-601e-461b-b8a7-f218b87b0f47-dns-swift-storage-0\") pod \"507e14a3-601e-461b-b8a7-f218b87b0f47\" (UID: \"507e14a3-601e-461b-b8a7-f218b87b0f47\") " Oct 05 07:09:44 crc kubenswrapper[4846]: I1005 07:09:44.312481 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/507e14a3-601e-461b-b8a7-f218b87b0f47-config\") pod \"507e14a3-601e-461b-b8a7-f218b87b0f47\" (UID: \"507e14a3-601e-461b-b8a7-f218b87b0f47\") " Oct 05 07:09:44 crc kubenswrapper[4846]: I1005 07:09:44.312529 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pvqr5\" (UniqueName: \"kubernetes.io/projected/507e14a3-601e-461b-b8a7-f218b87b0f47-kube-api-access-pvqr5\") pod \"507e14a3-601e-461b-b8a7-f218b87b0f47\" (UID: \"507e14a3-601e-461b-b8a7-f218b87b0f47\") " Oct 05 07:09:44 crc kubenswrapper[4846]: I1005 07:09:44.312568 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/507e14a3-601e-461b-b8a7-f218b87b0f47-ovsdbserver-nb\") pod \"507e14a3-601e-461b-b8a7-f218b87b0f47\" (UID: \"507e14a3-601e-461b-b8a7-f218b87b0f47\") " Oct 05 07:09:44 crc kubenswrapper[4846]: I1005 07:09:44.312600 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/507e14a3-601e-461b-b8a7-f218b87b0f47-dns-svc\") pod \"507e14a3-601e-461b-b8a7-f218b87b0f47\" (UID: \"507e14a3-601e-461b-b8a7-f218b87b0f47\") " Oct 05 07:09:44 crc kubenswrapper[4846]: I1005 07:09:44.312621 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/507e14a3-601e-461b-b8a7-f218b87b0f47-ovsdbserver-sb\") pod \"507e14a3-601e-461b-b8a7-f218b87b0f47\" (UID: \"507e14a3-601e-461b-b8a7-f218b87b0f47\") " Oct 05 07:09:44 crc kubenswrapper[4846]: I1005 07:09:44.321555 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/507e14a3-601e-461b-b8a7-f218b87b0f47-kube-api-access-pvqr5" (OuterVolumeSpecName: "kube-api-access-pvqr5") pod "507e14a3-601e-461b-b8a7-f218b87b0f47" (UID: "507e14a3-601e-461b-b8a7-f218b87b0f47"). InnerVolumeSpecName "kube-api-access-pvqr5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:09:44 crc kubenswrapper[4846]: I1005 07:09:44.374588 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/507e14a3-601e-461b-b8a7-f218b87b0f47-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "507e14a3-601e-461b-b8a7-f218b87b0f47" (UID: "507e14a3-601e-461b-b8a7-f218b87b0f47"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 07:09:44 crc kubenswrapper[4846]: I1005 07:09:44.375966 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/507e14a3-601e-461b-b8a7-f218b87b0f47-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "507e14a3-601e-461b-b8a7-f218b87b0f47" (UID: "507e14a3-601e-461b-b8a7-f218b87b0f47"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 07:09:44 crc kubenswrapper[4846]: I1005 07:09:44.376682 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/507e14a3-601e-461b-b8a7-f218b87b0f47-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "507e14a3-601e-461b-b8a7-f218b87b0f47" (UID: "507e14a3-601e-461b-b8a7-f218b87b0f47"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 07:09:44 crc kubenswrapper[4846]: I1005 07:09:44.393685 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/507e14a3-601e-461b-b8a7-f218b87b0f47-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "507e14a3-601e-461b-b8a7-f218b87b0f47" (UID: "507e14a3-601e-461b-b8a7-f218b87b0f47"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 07:09:44 crc kubenswrapper[4846]: I1005 07:09:44.399676 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/507e14a3-601e-461b-b8a7-f218b87b0f47-config" (OuterVolumeSpecName: "config") pod "507e14a3-601e-461b-b8a7-f218b87b0f47" (UID: "507e14a3-601e-461b-b8a7-f218b87b0f47"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 07:09:44 crc kubenswrapper[4846]: I1005 07:09:44.414084 4846 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/507e14a3-601e-461b-b8a7-f218b87b0f47-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 05 07:09:44 crc kubenswrapper[4846]: I1005 07:09:44.414128 4846 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/507e14a3-601e-461b-b8a7-f218b87b0f47-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 05 07:09:44 crc kubenswrapper[4846]: I1005 07:09:44.414140 4846 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/507e14a3-601e-461b-b8a7-f218b87b0f47-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 05 07:09:44 crc kubenswrapper[4846]: I1005 07:09:44.414150 4846 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/507e14a3-601e-461b-b8a7-f218b87b0f47-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 05 07:09:44 crc kubenswrapper[4846]: I1005 07:09:44.414160 4846 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/507e14a3-601e-461b-b8a7-f218b87b0f47-config\") on node \"crc\" DevicePath \"\"" Oct 05 07:09:44 crc kubenswrapper[4846]: I1005 07:09:44.414170 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pvqr5\" (UniqueName: \"kubernetes.io/projected/507e14a3-601e-461b-b8a7-f218b87b0f47-kube-api-access-pvqr5\") on node \"crc\" DevicePath \"\"" Oct 05 07:09:44 crc kubenswrapper[4846]: I1005 07:09:44.650426 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 05 07:09:44 crc kubenswrapper[4846]: I1005 07:09:44.826553 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/06e92a1c-15bf-4cdb-a284-0b5bee2a75c9-scripts\") pod \"06e92a1c-15bf-4cdb-a284-0b5bee2a75c9\" (UID: \"06e92a1c-15bf-4cdb-a284-0b5bee2a75c9\") " Oct 05 07:09:44 crc kubenswrapper[4846]: I1005 07:09:44.826651 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/06e92a1c-15bf-4cdb-a284-0b5bee2a75c9-combined-ca-bundle\") pod \"06e92a1c-15bf-4cdb-a284-0b5bee2a75c9\" (UID: \"06e92a1c-15bf-4cdb-a284-0b5bee2a75c9\") " Oct 05 07:09:44 crc kubenswrapper[4846]: I1005 07:09:44.826782 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2m5sx\" (UniqueName: \"kubernetes.io/projected/06e92a1c-15bf-4cdb-a284-0b5bee2a75c9-kube-api-access-2m5sx\") pod \"06e92a1c-15bf-4cdb-a284-0b5bee2a75c9\" (UID: \"06e92a1c-15bf-4cdb-a284-0b5bee2a75c9\") " Oct 05 07:09:44 crc kubenswrapper[4846]: I1005 07:09:44.826977 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/06e92a1c-15bf-4cdb-a284-0b5bee2a75c9-sg-core-conf-yaml\") pod \"06e92a1c-15bf-4cdb-a284-0b5bee2a75c9\" (UID: \"06e92a1c-15bf-4cdb-a284-0b5bee2a75c9\") " Oct 05 07:09:44 crc kubenswrapper[4846]: I1005 07:09:44.827065 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/06e92a1c-15bf-4cdb-a284-0b5bee2a75c9-log-httpd\") pod \"06e92a1c-15bf-4cdb-a284-0b5bee2a75c9\" (UID: \"06e92a1c-15bf-4cdb-a284-0b5bee2a75c9\") " Oct 05 07:09:44 crc kubenswrapper[4846]: I1005 07:09:44.827096 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/06e92a1c-15bf-4cdb-a284-0b5bee2a75c9-run-httpd\") pod \"06e92a1c-15bf-4cdb-a284-0b5bee2a75c9\" (UID: \"06e92a1c-15bf-4cdb-a284-0b5bee2a75c9\") " Oct 05 07:09:44 crc kubenswrapper[4846]: I1005 07:09:44.827115 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/06e92a1c-15bf-4cdb-a284-0b5bee2a75c9-config-data\") pod \"06e92a1c-15bf-4cdb-a284-0b5bee2a75c9\" (UID: \"06e92a1c-15bf-4cdb-a284-0b5bee2a75c9\") " Oct 05 07:09:44 crc kubenswrapper[4846]: I1005 07:09:44.828554 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/06e92a1c-15bf-4cdb-a284-0b5bee2a75c9-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "06e92a1c-15bf-4cdb-a284-0b5bee2a75c9" (UID: "06e92a1c-15bf-4cdb-a284-0b5bee2a75c9"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 07:09:44 crc kubenswrapper[4846]: I1005 07:09:44.828902 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/06e92a1c-15bf-4cdb-a284-0b5bee2a75c9-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "06e92a1c-15bf-4cdb-a284-0b5bee2a75c9" (UID: "06e92a1c-15bf-4cdb-a284-0b5bee2a75c9"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 07:09:44 crc kubenswrapper[4846]: I1005 07:09:44.833374 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/06e92a1c-15bf-4cdb-a284-0b5bee2a75c9-kube-api-access-2m5sx" (OuterVolumeSpecName: "kube-api-access-2m5sx") pod "06e92a1c-15bf-4cdb-a284-0b5bee2a75c9" (UID: "06e92a1c-15bf-4cdb-a284-0b5bee2a75c9"). InnerVolumeSpecName "kube-api-access-2m5sx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:09:44 crc kubenswrapper[4846]: I1005 07:09:44.838552 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/06e92a1c-15bf-4cdb-a284-0b5bee2a75c9-scripts" (OuterVolumeSpecName: "scripts") pod "06e92a1c-15bf-4cdb-a284-0b5bee2a75c9" (UID: "06e92a1c-15bf-4cdb-a284-0b5bee2a75c9"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:09:44 crc kubenswrapper[4846]: I1005 07:09:44.869901 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/06e92a1c-15bf-4cdb-a284-0b5bee2a75c9-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "06e92a1c-15bf-4cdb-a284-0b5bee2a75c9" (UID: "06e92a1c-15bf-4cdb-a284-0b5bee2a75c9"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:09:44 crc kubenswrapper[4846]: I1005 07:09:44.917292 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/06e92a1c-15bf-4cdb-a284-0b5bee2a75c9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "06e92a1c-15bf-4cdb-a284-0b5bee2a75c9" (UID: "06e92a1c-15bf-4cdb-a284-0b5bee2a75c9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:09:44 crc kubenswrapper[4846]: I1005 07:09:44.929192 4846 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/06e92a1c-15bf-4cdb-a284-0b5bee2a75c9-scripts\") on node \"crc\" DevicePath \"\"" Oct 05 07:09:44 crc kubenswrapper[4846]: I1005 07:09:44.929226 4846 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/06e92a1c-15bf-4cdb-a284-0b5bee2a75c9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 07:09:44 crc kubenswrapper[4846]: I1005 07:09:44.929239 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2m5sx\" (UniqueName: \"kubernetes.io/projected/06e92a1c-15bf-4cdb-a284-0b5bee2a75c9-kube-api-access-2m5sx\") on node \"crc\" DevicePath \"\"" Oct 05 07:09:44 crc kubenswrapper[4846]: I1005 07:09:44.929249 4846 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/06e92a1c-15bf-4cdb-a284-0b5bee2a75c9-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 05 07:09:44 crc kubenswrapper[4846]: I1005 07:09:44.929258 4846 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/06e92a1c-15bf-4cdb-a284-0b5bee2a75c9-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 05 07:09:44 crc kubenswrapper[4846]: I1005 07:09:44.929268 4846 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/06e92a1c-15bf-4cdb-a284-0b5bee2a75c9-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 05 07:09:44 crc kubenswrapper[4846]: I1005 07:09:44.947308 4846 generic.go:334] "Generic (PLEG): container finished" podID="d9adf2fe-f612-44f5-a7eb-e137909be76a" containerID="f0993f040dcd6d3145256a44cdc38fc8db9c44f334ee6dc7c33895edf97422b9" exitCode=0 Oct 05 07:09:44 crc kubenswrapper[4846]: I1005 07:09:44.947418 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-s6xg9" event={"ID":"d9adf2fe-f612-44f5-a7eb-e137909be76a","Type":"ContainerDied","Data":"f0993f040dcd6d3145256a44cdc38fc8db9c44f334ee6dc7c33895edf97422b9"} Oct 05 07:09:44 crc kubenswrapper[4846]: I1005 07:09:44.951785 4846 generic.go:334] "Generic (PLEG): container finished" podID="06e92a1c-15bf-4cdb-a284-0b5bee2a75c9" containerID="8a4dc29d9a82877eba8bfc9d6ac07fb48d1a0e553bef761e5a88bb04c88cbe6b" exitCode=0 Oct 05 07:09:44 crc kubenswrapper[4846]: I1005 07:09:44.951891 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"06e92a1c-15bf-4cdb-a284-0b5bee2a75c9","Type":"ContainerDied","Data":"8a4dc29d9a82877eba8bfc9d6ac07fb48d1a0e553bef761e5a88bb04c88cbe6b"} Oct 05 07:09:44 crc kubenswrapper[4846]: I1005 07:09:44.951917 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 05 07:09:44 crc kubenswrapper[4846]: I1005 07:09:44.951936 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"06e92a1c-15bf-4cdb-a284-0b5bee2a75c9","Type":"ContainerDied","Data":"0ab3345f5cb08c7a58ce1742d4a3f8bba41ae9e9b2f12e7723b4ec05e1f6da29"} Oct 05 07:09:44 crc kubenswrapper[4846]: I1005 07:09:44.951985 4846 scope.go:117] "RemoveContainer" containerID="1366799532bcfbd4e3db914bbc7d301c1fffe5b0fe8364f08939c0180dd80492" Oct 05 07:09:44 crc kubenswrapper[4846]: I1005 07:09:44.955944 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-99495cbc7-v87cl" event={"ID":"507e14a3-601e-461b-b8a7-f218b87b0f47","Type":"ContainerDied","Data":"377f7635d785623b7b156209f59b159fad7649c1df29afbf5d93db2d4f758a93"} Oct 05 07:09:44 crc kubenswrapper[4846]: I1005 07:09:44.956210 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-99495cbc7-v87cl" Oct 05 07:09:44 crc kubenswrapper[4846]: I1005 07:09:44.960813 4846 generic.go:334] "Generic (PLEG): container finished" podID="a368d12f-6383-4a8e-8d71-23a675946027" containerID="b3e205f106c607cc68f15ea01b24c2b915400791a93310bfef25ced3582cdd85" exitCode=0 Oct 05 07:09:44 crc kubenswrapper[4846]: I1005 07:09:44.961310 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-ptdtq" event={"ID":"a368d12f-6383-4a8e-8d71-23a675946027","Type":"ContainerDied","Data":"b3e205f106c607cc68f15ea01b24c2b915400791a93310bfef25ced3582cdd85"} Oct 05 07:09:44 crc kubenswrapper[4846]: I1005 07:09:44.971820 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/06e92a1c-15bf-4cdb-a284-0b5bee2a75c9-config-data" (OuterVolumeSpecName: "config-data") pod "06e92a1c-15bf-4cdb-a284-0b5bee2a75c9" (UID: "06e92a1c-15bf-4cdb-a284-0b5bee2a75c9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:09:44 crc kubenswrapper[4846]: I1005 07:09:44.993171 4846 scope.go:117] "RemoveContainer" containerID="e8da228ce53f0f9b4a512fcd59c901ab9bd35018ddae6a4efd69d1cb97fabdfd" Oct 05 07:09:45 crc kubenswrapper[4846]: I1005 07:09:45.024299 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-99495cbc7-v87cl"] Oct 05 07:09:45 crc kubenswrapper[4846]: I1005 07:09:45.031526 4846 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/06e92a1c-15bf-4cdb-a284-0b5bee2a75c9-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 07:09:45 crc kubenswrapper[4846]: I1005 07:09:45.031575 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-99495cbc7-v87cl"] Oct 05 07:09:45 crc kubenswrapper[4846]: I1005 07:09:45.051925 4846 scope.go:117] "RemoveContainer" containerID="8a4dc29d9a82877eba8bfc9d6ac07fb48d1a0e553bef761e5a88bb04c88cbe6b" Oct 05 07:09:45 crc kubenswrapper[4846]: I1005 07:09:45.089489 4846 scope.go:117] "RemoveContainer" containerID="0a02b9d96618505f33de0057a4a6c5bb3f7d078b7697e171d7309d0fb4459179" Oct 05 07:09:45 crc kubenswrapper[4846]: I1005 07:09:45.122499 4846 scope.go:117] "RemoveContainer" containerID="1366799532bcfbd4e3db914bbc7d301c1fffe5b0fe8364f08939c0180dd80492" Oct 05 07:09:45 crc kubenswrapper[4846]: E1005 07:09:45.123296 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1366799532bcfbd4e3db914bbc7d301c1fffe5b0fe8364f08939c0180dd80492\": container with ID starting with 1366799532bcfbd4e3db914bbc7d301c1fffe5b0fe8364f08939c0180dd80492 not found: ID does not exist" containerID="1366799532bcfbd4e3db914bbc7d301c1fffe5b0fe8364f08939c0180dd80492" Oct 05 07:09:45 crc kubenswrapper[4846]: I1005 07:09:45.123381 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1366799532bcfbd4e3db914bbc7d301c1fffe5b0fe8364f08939c0180dd80492"} err="failed to get container status \"1366799532bcfbd4e3db914bbc7d301c1fffe5b0fe8364f08939c0180dd80492\": rpc error: code = NotFound desc = could not find container \"1366799532bcfbd4e3db914bbc7d301c1fffe5b0fe8364f08939c0180dd80492\": container with ID starting with 1366799532bcfbd4e3db914bbc7d301c1fffe5b0fe8364f08939c0180dd80492 not found: ID does not exist" Oct 05 07:09:45 crc kubenswrapper[4846]: I1005 07:09:45.123415 4846 scope.go:117] "RemoveContainer" containerID="e8da228ce53f0f9b4a512fcd59c901ab9bd35018ddae6a4efd69d1cb97fabdfd" Oct 05 07:09:45 crc kubenswrapper[4846]: E1005 07:09:45.123978 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e8da228ce53f0f9b4a512fcd59c901ab9bd35018ddae6a4efd69d1cb97fabdfd\": container with ID starting with e8da228ce53f0f9b4a512fcd59c901ab9bd35018ddae6a4efd69d1cb97fabdfd not found: ID does not exist" containerID="e8da228ce53f0f9b4a512fcd59c901ab9bd35018ddae6a4efd69d1cb97fabdfd" Oct 05 07:09:45 crc kubenswrapper[4846]: I1005 07:09:45.124022 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e8da228ce53f0f9b4a512fcd59c901ab9bd35018ddae6a4efd69d1cb97fabdfd"} err="failed to get container status \"e8da228ce53f0f9b4a512fcd59c901ab9bd35018ddae6a4efd69d1cb97fabdfd\": rpc error: code = NotFound desc = could not find container \"e8da228ce53f0f9b4a512fcd59c901ab9bd35018ddae6a4efd69d1cb97fabdfd\": container with ID starting with e8da228ce53f0f9b4a512fcd59c901ab9bd35018ddae6a4efd69d1cb97fabdfd not found: ID does not exist" Oct 05 07:09:45 crc kubenswrapper[4846]: I1005 07:09:45.124045 4846 scope.go:117] "RemoveContainer" containerID="8a4dc29d9a82877eba8bfc9d6ac07fb48d1a0e553bef761e5a88bb04c88cbe6b" Oct 05 07:09:45 crc kubenswrapper[4846]: E1005 07:09:45.124523 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8a4dc29d9a82877eba8bfc9d6ac07fb48d1a0e553bef761e5a88bb04c88cbe6b\": container with ID starting with 8a4dc29d9a82877eba8bfc9d6ac07fb48d1a0e553bef761e5a88bb04c88cbe6b not found: ID does not exist" containerID="8a4dc29d9a82877eba8bfc9d6ac07fb48d1a0e553bef761e5a88bb04c88cbe6b" Oct 05 07:09:45 crc kubenswrapper[4846]: I1005 07:09:45.124587 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8a4dc29d9a82877eba8bfc9d6ac07fb48d1a0e553bef761e5a88bb04c88cbe6b"} err="failed to get container status \"8a4dc29d9a82877eba8bfc9d6ac07fb48d1a0e553bef761e5a88bb04c88cbe6b\": rpc error: code = NotFound desc = could not find container \"8a4dc29d9a82877eba8bfc9d6ac07fb48d1a0e553bef761e5a88bb04c88cbe6b\": container with ID starting with 8a4dc29d9a82877eba8bfc9d6ac07fb48d1a0e553bef761e5a88bb04c88cbe6b not found: ID does not exist" Oct 05 07:09:45 crc kubenswrapper[4846]: I1005 07:09:45.124633 4846 scope.go:117] "RemoveContainer" containerID="0a02b9d96618505f33de0057a4a6c5bb3f7d078b7697e171d7309d0fb4459179" Oct 05 07:09:45 crc kubenswrapper[4846]: E1005 07:09:45.125019 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0a02b9d96618505f33de0057a4a6c5bb3f7d078b7697e171d7309d0fb4459179\": container with ID starting with 0a02b9d96618505f33de0057a4a6c5bb3f7d078b7697e171d7309d0fb4459179 not found: ID does not exist" containerID="0a02b9d96618505f33de0057a4a6c5bb3f7d078b7697e171d7309d0fb4459179" Oct 05 07:09:45 crc kubenswrapper[4846]: I1005 07:09:45.125068 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0a02b9d96618505f33de0057a4a6c5bb3f7d078b7697e171d7309d0fb4459179"} err="failed to get container status \"0a02b9d96618505f33de0057a4a6c5bb3f7d078b7697e171d7309d0fb4459179\": rpc error: code = NotFound desc = could not find container \"0a02b9d96618505f33de0057a4a6c5bb3f7d078b7697e171d7309d0fb4459179\": container with ID starting with 0a02b9d96618505f33de0057a4a6c5bb3f7d078b7697e171d7309d0fb4459179 not found: ID does not exist" Oct 05 07:09:45 crc kubenswrapper[4846]: I1005 07:09:45.125086 4846 scope.go:117] "RemoveContainer" containerID="efc09bde05d0841f423bbc79bab49db3b7705ff627f508b06fdf9dfef64469a0" Oct 05 07:09:45 crc kubenswrapper[4846]: I1005 07:09:45.146843 4846 scope.go:117] "RemoveContainer" containerID="d650d5bbcb222ec9724e83b217d5482ce80a31ef9cbe9e92ec328f485017d0e1" Oct 05 07:09:45 crc kubenswrapper[4846]: I1005 07:09:45.301230 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 05 07:09:45 crc kubenswrapper[4846]: I1005 07:09:45.315648 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 05 07:09:45 crc kubenswrapper[4846]: I1005 07:09:45.340302 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 05 07:09:45 crc kubenswrapper[4846]: E1005 07:09:45.340839 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="06e92a1c-15bf-4cdb-a284-0b5bee2a75c9" containerName="ceilometer-central-agent" Oct 05 07:09:45 crc kubenswrapper[4846]: I1005 07:09:45.340862 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="06e92a1c-15bf-4cdb-a284-0b5bee2a75c9" containerName="ceilometer-central-agent" Oct 05 07:09:45 crc kubenswrapper[4846]: E1005 07:09:45.340886 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="507e14a3-601e-461b-b8a7-f218b87b0f47" containerName="init" Oct 05 07:09:45 crc kubenswrapper[4846]: I1005 07:09:45.340894 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="507e14a3-601e-461b-b8a7-f218b87b0f47" containerName="init" Oct 05 07:09:45 crc kubenswrapper[4846]: E1005 07:09:45.340914 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="06e92a1c-15bf-4cdb-a284-0b5bee2a75c9" containerName="ceilometer-notification-agent" Oct 05 07:09:45 crc kubenswrapper[4846]: I1005 07:09:45.340922 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="06e92a1c-15bf-4cdb-a284-0b5bee2a75c9" containerName="ceilometer-notification-agent" Oct 05 07:09:45 crc kubenswrapper[4846]: E1005 07:09:45.340942 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="507e14a3-601e-461b-b8a7-f218b87b0f47" containerName="dnsmasq-dns" Oct 05 07:09:45 crc kubenswrapper[4846]: I1005 07:09:45.340954 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="507e14a3-601e-461b-b8a7-f218b87b0f47" containerName="dnsmasq-dns" Oct 05 07:09:45 crc kubenswrapper[4846]: E1005 07:09:45.341002 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="06e92a1c-15bf-4cdb-a284-0b5bee2a75c9" containerName="sg-core" Oct 05 07:09:45 crc kubenswrapper[4846]: I1005 07:09:45.341010 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="06e92a1c-15bf-4cdb-a284-0b5bee2a75c9" containerName="sg-core" Oct 05 07:09:45 crc kubenswrapper[4846]: E1005 07:09:45.342940 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="06e92a1c-15bf-4cdb-a284-0b5bee2a75c9" containerName="proxy-httpd" Oct 05 07:09:45 crc kubenswrapper[4846]: I1005 07:09:45.342961 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="06e92a1c-15bf-4cdb-a284-0b5bee2a75c9" containerName="proxy-httpd" Oct 05 07:09:45 crc kubenswrapper[4846]: I1005 07:09:45.343362 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="06e92a1c-15bf-4cdb-a284-0b5bee2a75c9" containerName="ceilometer-central-agent" Oct 05 07:09:45 crc kubenswrapper[4846]: I1005 07:09:45.343389 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="06e92a1c-15bf-4cdb-a284-0b5bee2a75c9" containerName="proxy-httpd" Oct 05 07:09:45 crc kubenswrapper[4846]: I1005 07:09:45.343404 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="06e92a1c-15bf-4cdb-a284-0b5bee2a75c9" containerName="ceilometer-notification-agent" Oct 05 07:09:45 crc kubenswrapper[4846]: I1005 07:09:45.343427 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="06e92a1c-15bf-4cdb-a284-0b5bee2a75c9" containerName="sg-core" Oct 05 07:09:45 crc kubenswrapper[4846]: I1005 07:09:45.343449 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="507e14a3-601e-461b-b8a7-f218b87b0f47" containerName="dnsmasq-dns" Oct 05 07:09:45 crc kubenswrapper[4846]: I1005 07:09:45.345923 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 05 07:09:45 crc kubenswrapper[4846]: I1005 07:09:45.347965 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 05 07:09:45 crc kubenswrapper[4846]: I1005 07:09:45.352991 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 05 07:09:45 crc kubenswrapper[4846]: I1005 07:09:45.355766 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 05 07:09:45 crc kubenswrapper[4846]: I1005 07:09:45.356440 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Oct 05 07:09:45 crc kubenswrapper[4846]: I1005 07:09:45.541537 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/eb03eca3-f45d-4361-ab00-9ce16bb2c96e-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"eb03eca3-f45d-4361-ab00-9ce16bb2c96e\") " pod="openstack/ceilometer-0" Oct 05 07:09:45 crc kubenswrapper[4846]: I1005 07:09:45.541626 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eb03eca3-f45d-4361-ab00-9ce16bb2c96e-config-data\") pod \"ceilometer-0\" (UID: \"eb03eca3-f45d-4361-ab00-9ce16bb2c96e\") " pod="openstack/ceilometer-0" Oct 05 07:09:45 crc kubenswrapper[4846]: I1005 07:09:45.542039 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/eb03eca3-f45d-4361-ab00-9ce16bb2c96e-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"eb03eca3-f45d-4361-ab00-9ce16bb2c96e\") " pod="openstack/ceilometer-0" Oct 05 07:09:45 crc kubenswrapper[4846]: I1005 07:09:45.542080 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/eb03eca3-f45d-4361-ab00-9ce16bb2c96e-log-httpd\") pod \"ceilometer-0\" (UID: \"eb03eca3-f45d-4361-ab00-9ce16bb2c96e\") " pod="openstack/ceilometer-0" Oct 05 07:09:45 crc kubenswrapper[4846]: I1005 07:09:45.542139 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eb03eca3-f45d-4361-ab00-9ce16bb2c96e-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"eb03eca3-f45d-4361-ab00-9ce16bb2c96e\") " pod="openstack/ceilometer-0" Oct 05 07:09:45 crc kubenswrapper[4846]: I1005 07:09:45.542197 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/eb03eca3-f45d-4361-ab00-9ce16bb2c96e-scripts\") pod \"ceilometer-0\" (UID: \"eb03eca3-f45d-4361-ab00-9ce16bb2c96e\") " pod="openstack/ceilometer-0" Oct 05 07:09:45 crc kubenswrapper[4846]: I1005 07:09:45.542244 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/eb03eca3-f45d-4361-ab00-9ce16bb2c96e-run-httpd\") pod \"ceilometer-0\" (UID: \"eb03eca3-f45d-4361-ab00-9ce16bb2c96e\") " pod="openstack/ceilometer-0" Oct 05 07:09:45 crc kubenswrapper[4846]: I1005 07:09:45.542296 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9hpsb\" (UniqueName: \"kubernetes.io/projected/eb03eca3-f45d-4361-ab00-9ce16bb2c96e-kube-api-access-9hpsb\") pod \"ceilometer-0\" (UID: \"eb03eca3-f45d-4361-ab00-9ce16bb2c96e\") " pod="openstack/ceilometer-0" Oct 05 07:09:45 crc kubenswrapper[4846]: I1005 07:09:45.643833 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eb03eca3-f45d-4361-ab00-9ce16bb2c96e-config-data\") pod \"ceilometer-0\" (UID: \"eb03eca3-f45d-4361-ab00-9ce16bb2c96e\") " pod="openstack/ceilometer-0" Oct 05 07:09:45 crc kubenswrapper[4846]: I1005 07:09:45.643919 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/eb03eca3-f45d-4361-ab00-9ce16bb2c96e-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"eb03eca3-f45d-4361-ab00-9ce16bb2c96e\") " pod="openstack/ceilometer-0" Oct 05 07:09:45 crc kubenswrapper[4846]: I1005 07:09:45.643951 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/eb03eca3-f45d-4361-ab00-9ce16bb2c96e-log-httpd\") pod \"ceilometer-0\" (UID: \"eb03eca3-f45d-4361-ab00-9ce16bb2c96e\") " pod="openstack/ceilometer-0" Oct 05 07:09:45 crc kubenswrapper[4846]: I1005 07:09:45.643994 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eb03eca3-f45d-4361-ab00-9ce16bb2c96e-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"eb03eca3-f45d-4361-ab00-9ce16bb2c96e\") " pod="openstack/ceilometer-0" Oct 05 07:09:45 crc kubenswrapper[4846]: I1005 07:09:45.644034 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/eb03eca3-f45d-4361-ab00-9ce16bb2c96e-scripts\") pod \"ceilometer-0\" (UID: \"eb03eca3-f45d-4361-ab00-9ce16bb2c96e\") " pod="openstack/ceilometer-0" Oct 05 07:09:45 crc kubenswrapper[4846]: I1005 07:09:45.644061 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/eb03eca3-f45d-4361-ab00-9ce16bb2c96e-run-httpd\") pod \"ceilometer-0\" (UID: \"eb03eca3-f45d-4361-ab00-9ce16bb2c96e\") " pod="openstack/ceilometer-0" Oct 05 07:09:45 crc kubenswrapper[4846]: I1005 07:09:45.644111 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9hpsb\" (UniqueName: \"kubernetes.io/projected/eb03eca3-f45d-4361-ab00-9ce16bb2c96e-kube-api-access-9hpsb\") pod \"ceilometer-0\" (UID: \"eb03eca3-f45d-4361-ab00-9ce16bb2c96e\") " pod="openstack/ceilometer-0" Oct 05 07:09:45 crc kubenswrapper[4846]: I1005 07:09:45.644214 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/eb03eca3-f45d-4361-ab00-9ce16bb2c96e-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"eb03eca3-f45d-4361-ab00-9ce16bb2c96e\") " pod="openstack/ceilometer-0" Oct 05 07:09:45 crc kubenswrapper[4846]: I1005 07:09:45.644567 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/eb03eca3-f45d-4361-ab00-9ce16bb2c96e-log-httpd\") pod \"ceilometer-0\" (UID: \"eb03eca3-f45d-4361-ab00-9ce16bb2c96e\") " pod="openstack/ceilometer-0" Oct 05 07:09:45 crc kubenswrapper[4846]: I1005 07:09:45.645030 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/eb03eca3-f45d-4361-ab00-9ce16bb2c96e-run-httpd\") pod \"ceilometer-0\" (UID: \"eb03eca3-f45d-4361-ab00-9ce16bb2c96e\") " pod="openstack/ceilometer-0" Oct 05 07:09:45 crc kubenswrapper[4846]: I1005 07:09:45.648354 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eb03eca3-f45d-4361-ab00-9ce16bb2c96e-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"eb03eca3-f45d-4361-ab00-9ce16bb2c96e\") " pod="openstack/ceilometer-0" Oct 05 07:09:45 crc kubenswrapper[4846]: I1005 07:09:45.648732 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/eb03eca3-f45d-4361-ab00-9ce16bb2c96e-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"eb03eca3-f45d-4361-ab00-9ce16bb2c96e\") " pod="openstack/ceilometer-0" Oct 05 07:09:45 crc kubenswrapper[4846]: I1005 07:09:45.649597 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/eb03eca3-f45d-4361-ab00-9ce16bb2c96e-scripts\") pod \"ceilometer-0\" (UID: \"eb03eca3-f45d-4361-ab00-9ce16bb2c96e\") " pod="openstack/ceilometer-0" Oct 05 07:09:45 crc kubenswrapper[4846]: I1005 07:09:45.650770 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eb03eca3-f45d-4361-ab00-9ce16bb2c96e-config-data\") pod \"ceilometer-0\" (UID: \"eb03eca3-f45d-4361-ab00-9ce16bb2c96e\") " pod="openstack/ceilometer-0" Oct 05 07:09:45 crc kubenswrapper[4846]: I1005 07:09:45.669712 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/eb03eca3-f45d-4361-ab00-9ce16bb2c96e-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"eb03eca3-f45d-4361-ab00-9ce16bb2c96e\") " pod="openstack/ceilometer-0" Oct 05 07:09:45 crc kubenswrapper[4846]: I1005 07:09:45.671987 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9hpsb\" (UniqueName: \"kubernetes.io/projected/eb03eca3-f45d-4361-ab00-9ce16bb2c96e-kube-api-access-9hpsb\") pod \"ceilometer-0\" (UID: \"eb03eca3-f45d-4361-ab00-9ce16bb2c96e\") " pod="openstack/ceilometer-0" Oct 05 07:09:45 crc kubenswrapper[4846]: I1005 07:09:45.968654 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 05 07:09:46 crc kubenswrapper[4846]: I1005 07:09:46.518810 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="06e92a1c-15bf-4cdb-a284-0b5bee2a75c9" path="/var/lib/kubelet/pods/06e92a1c-15bf-4cdb-a284-0b5bee2a75c9/volumes" Oct 05 07:09:46 crc kubenswrapper[4846]: I1005 07:09:46.521687 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="507e14a3-601e-461b-b8a7-f218b87b0f47" path="/var/lib/kubelet/pods/507e14a3-601e-461b-b8a7-f218b87b0f47/volumes" Oct 05 07:09:46 crc kubenswrapper[4846]: I1005 07:09:46.528474 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-ptdtq" Oct 05 07:09:46 crc kubenswrapper[4846]: I1005 07:09:46.534811 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-s6xg9" Oct 05 07:09:46 crc kubenswrapper[4846]: I1005 07:09:46.568596 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d9adf2fe-f612-44f5-a7eb-e137909be76a-config-data\") pod \"d9adf2fe-f612-44f5-a7eb-e137909be76a\" (UID: \"d9adf2fe-f612-44f5-a7eb-e137909be76a\") " Oct 05 07:09:46 crc kubenswrapper[4846]: I1005 07:09:46.568655 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l7vjn\" (UniqueName: \"kubernetes.io/projected/d9adf2fe-f612-44f5-a7eb-e137909be76a-kube-api-access-l7vjn\") pod \"d9adf2fe-f612-44f5-a7eb-e137909be76a\" (UID: \"d9adf2fe-f612-44f5-a7eb-e137909be76a\") " Oct 05 07:09:46 crc kubenswrapper[4846]: I1005 07:09:46.568707 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a368d12f-6383-4a8e-8d71-23a675946027-combined-ca-bundle\") pod \"a368d12f-6383-4a8e-8d71-23a675946027\" (UID: \"a368d12f-6383-4a8e-8d71-23a675946027\") " Oct 05 07:09:46 crc kubenswrapper[4846]: I1005 07:09:46.568742 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tj7rk\" (UniqueName: \"kubernetes.io/projected/a368d12f-6383-4a8e-8d71-23a675946027-kube-api-access-tj7rk\") pod \"a368d12f-6383-4a8e-8d71-23a675946027\" (UID: \"a368d12f-6383-4a8e-8d71-23a675946027\") " Oct 05 07:09:46 crc kubenswrapper[4846]: I1005 07:09:46.588366 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d9adf2fe-f612-44f5-a7eb-e137909be76a-kube-api-access-l7vjn" (OuterVolumeSpecName: "kube-api-access-l7vjn") pod "d9adf2fe-f612-44f5-a7eb-e137909be76a" (UID: "d9adf2fe-f612-44f5-a7eb-e137909be76a"). InnerVolumeSpecName "kube-api-access-l7vjn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:09:46 crc kubenswrapper[4846]: I1005 07:09:46.589450 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a368d12f-6383-4a8e-8d71-23a675946027-kube-api-access-tj7rk" (OuterVolumeSpecName: "kube-api-access-tj7rk") pod "a368d12f-6383-4a8e-8d71-23a675946027" (UID: "a368d12f-6383-4a8e-8d71-23a675946027"). InnerVolumeSpecName "kube-api-access-tj7rk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:09:46 crc kubenswrapper[4846]: I1005 07:09:46.605975 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d9adf2fe-f612-44f5-a7eb-e137909be76a-config-data" (OuterVolumeSpecName: "config-data") pod "d9adf2fe-f612-44f5-a7eb-e137909be76a" (UID: "d9adf2fe-f612-44f5-a7eb-e137909be76a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:09:46 crc kubenswrapper[4846]: I1005 07:09:46.616734 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a368d12f-6383-4a8e-8d71-23a675946027-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a368d12f-6383-4a8e-8d71-23a675946027" (UID: "a368d12f-6383-4a8e-8d71-23a675946027"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:09:46 crc kubenswrapper[4846]: I1005 07:09:46.620776 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 05 07:09:46 crc kubenswrapper[4846]: I1005 07:09:46.630659 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 05 07:09:46 crc kubenswrapper[4846]: I1005 07:09:46.630707 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 05 07:09:46 crc kubenswrapper[4846]: I1005 07:09:46.670829 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d9adf2fe-f612-44f5-a7eb-e137909be76a-scripts\") pod \"d9adf2fe-f612-44f5-a7eb-e137909be76a\" (UID: \"d9adf2fe-f612-44f5-a7eb-e137909be76a\") " Oct 05 07:09:46 crc kubenswrapper[4846]: I1005 07:09:46.670909 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a368d12f-6383-4a8e-8d71-23a675946027-scripts\") pod \"a368d12f-6383-4a8e-8d71-23a675946027\" (UID: \"a368d12f-6383-4a8e-8d71-23a675946027\") " Oct 05 07:09:46 crc kubenswrapper[4846]: I1005 07:09:46.670986 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d9adf2fe-f612-44f5-a7eb-e137909be76a-combined-ca-bundle\") pod \"d9adf2fe-f612-44f5-a7eb-e137909be76a\" (UID: \"d9adf2fe-f612-44f5-a7eb-e137909be76a\") " Oct 05 07:09:46 crc kubenswrapper[4846]: I1005 07:09:46.671031 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a368d12f-6383-4a8e-8d71-23a675946027-config-data\") pod \"a368d12f-6383-4a8e-8d71-23a675946027\" (UID: \"a368d12f-6383-4a8e-8d71-23a675946027\") " Oct 05 07:09:46 crc kubenswrapper[4846]: I1005 07:09:46.671762 4846 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d9adf2fe-f612-44f5-a7eb-e137909be76a-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 07:09:46 crc kubenswrapper[4846]: I1005 07:09:46.671781 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l7vjn\" (UniqueName: \"kubernetes.io/projected/d9adf2fe-f612-44f5-a7eb-e137909be76a-kube-api-access-l7vjn\") on node \"crc\" DevicePath \"\"" Oct 05 07:09:46 crc kubenswrapper[4846]: I1005 07:09:46.671792 4846 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a368d12f-6383-4a8e-8d71-23a675946027-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 07:09:46 crc kubenswrapper[4846]: I1005 07:09:46.671805 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tj7rk\" (UniqueName: \"kubernetes.io/projected/a368d12f-6383-4a8e-8d71-23a675946027-kube-api-access-tj7rk\") on node \"crc\" DevicePath \"\"" Oct 05 07:09:46 crc kubenswrapper[4846]: I1005 07:09:46.674432 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d9adf2fe-f612-44f5-a7eb-e137909be76a-scripts" (OuterVolumeSpecName: "scripts") pod "d9adf2fe-f612-44f5-a7eb-e137909be76a" (UID: "d9adf2fe-f612-44f5-a7eb-e137909be76a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:09:46 crc kubenswrapper[4846]: I1005 07:09:46.675915 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a368d12f-6383-4a8e-8d71-23a675946027-scripts" (OuterVolumeSpecName: "scripts") pod "a368d12f-6383-4a8e-8d71-23a675946027" (UID: "a368d12f-6383-4a8e-8d71-23a675946027"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:09:46 crc kubenswrapper[4846]: I1005 07:09:46.703897 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a368d12f-6383-4a8e-8d71-23a675946027-config-data" (OuterVolumeSpecName: "config-data") pod "a368d12f-6383-4a8e-8d71-23a675946027" (UID: "a368d12f-6383-4a8e-8d71-23a675946027"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:09:46 crc kubenswrapper[4846]: I1005 07:09:46.704686 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d9adf2fe-f612-44f5-a7eb-e137909be76a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d9adf2fe-f612-44f5-a7eb-e137909be76a" (UID: "d9adf2fe-f612-44f5-a7eb-e137909be76a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:09:46 crc kubenswrapper[4846]: I1005 07:09:46.775003 4846 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d9adf2fe-f612-44f5-a7eb-e137909be76a-scripts\") on node \"crc\" DevicePath \"\"" Oct 05 07:09:46 crc kubenswrapper[4846]: I1005 07:09:46.775064 4846 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a368d12f-6383-4a8e-8d71-23a675946027-scripts\") on node \"crc\" DevicePath \"\"" Oct 05 07:09:46 crc kubenswrapper[4846]: I1005 07:09:46.775088 4846 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d9adf2fe-f612-44f5-a7eb-e137909be76a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 07:09:46 crc kubenswrapper[4846]: I1005 07:09:46.775108 4846 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a368d12f-6383-4a8e-8d71-23a675946027-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 07:09:47 crc kubenswrapper[4846]: I1005 07:09:47.035016 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-ptdtq" event={"ID":"a368d12f-6383-4a8e-8d71-23a675946027","Type":"ContainerDied","Data":"7df79d3abb6a623d10ef018a06e397d24e14c8f848beb54e7a08bbcfda71c2f6"} Oct 05 07:09:47 crc kubenswrapper[4846]: I1005 07:09:47.035078 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7df79d3abb6a623d10ef018a06e397d24e14c8f848beb54e7a08bbcfda71c2f6" Oct 05 07:09:47 crc kubenswrapper[4846]: I1005 07:09:47.035158 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-ptdtq" Oct 05 07:09:47 crc kubenswrapper[4846]: I1005 07:09:47.041077 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-s6xg9" event={"ID":"d9adf2fe-f612-44f5-a7eb-e137909be76a","Type":"ContainerDied","Data":"e6140a414dd8024e567280f760b95bebc0041fbbbac9769da674d592e2febd92"} Oct 05 07:09:47 crc kubenswrapper[4846]: I1005 07:09:47.041144 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e6140a414dd8024e567280f760b95bebc0041fbbbac9769da674d592e2febd92" Oct 05 07:09:47 crc kubenswrapper[4846]: I1005 07:09:47.041265 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-s6xg9" Oct 05 07:09:47 crc kubenswrapper[4846]: I1005 07:09:47.046171 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"eb03eca3-f45d-4361-ab00-9ce16bb2c96e","Type":"ContainerStarted","Data":"fc3cd487d783fbbd3afb2b971190c71056b88902b2807e17d5a1faf694fe9480"} Oct 05 07:09:47 crc kubenswrapper[4846]: I1005 07:09:47.096247 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 05 07:09:47 crc kubenswrapper[4846]: E1005 07:09:47.097142 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d9adf2fe-f612-44f5-a7eb-e137909be76a" containerName="nova-cell1-conductor-db-sync" Oct 05 07:09:47 crc kubenswrapper[4846]: I1005 07:09:47.097164 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="d9adf2fe-f612-44f5-a7eb-e137909be76a" containerName="nova-cell1-conductor-db-sync" Oct 05 07:09:47 crc kubenswrapper[4846]: E1005 07:09:47.097229 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a368d12f-6383-4a8e-8d71-23a675946027" containerName="nova-manage" Oct 05 07:09:47 crc kubenswrapper[4846]: I1005 07:09:47.097242 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="a368d12f-6383-4a8e-8d71-23a675946027" containerName="nova-manage" Oct 05 07:09:47 crc kubenswrapper[4846]: I1005 07:09:47.097463 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="d9adf2fe-f612-44f5-a7eb-e137909be76a" containerName="nova-cell1-conductor-db-sync" Oct 05 07:09:47 crc kubenswrapper[4846]: I1005 07:09:47.097507 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="a368d12f-6383-4a8e-8d71-23a675946027" containerName="nova-manage" Oct 05 07:09:47 crc kubenswrapper[4846]: I1005 07:09:47.098696 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Oct 05 07:09:47 crc kubenswrapper[4846]: I1005 07:09:47.108723 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Oct 05 07:09:47 crc kubenswrapper[4846]: I1005 07:09:47.132916 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 05 07:09:47 crc kubenswrapper[4846]: I1005 07:09:47.188486 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/33193805-7aa6-406e-8cad-0ebbb1228ef2-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"33193805-7aa6-406e-8cad-0ebbb1228ef2\") " pod="openstack/nova-cell1-conductor-0" Oct 05 07:09:47 crc kubenswrapper[4846]: I1005 07:09:47.188662 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/33193805-7aa6-406e-8cad-0ebbb1228ef2-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"33193805-7aa6-406e-8cad-0ebbb1228ef2\") " pod="openstack/nova-cell1-conductor-0" Oct 05 07:09:47 crc kubenswrapper[4846]: I1005 07:09:47.188809 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2nd7f\" (UniqueName: \"kubernetes.io/projected/33193805-7aa6-406e-8cad-0ebbb1228ef2-kube-api-access-2nd7f\") pod \"nova-cell1-conductor-0\" (UID: \"33193805-7aa6-406e-8cad-0ebbb1228ef2\") " pod="openstack/nova-cell1-conductor-0" Oct 05 07:09:47 crc kubenswrapper[4846]: I1005 07:09:47.275160 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 05 07:09:47 crc kubenswrapper[4846]: I1005 07:09:47.275558 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="095eba43-2814-4476-b482-efdf6b903de4" containerName="nova-api-log" containerID="cri-o://a5ed027b4a2ba75551446c6b871445f74de3480c300dba8c804dc3238212ddc8" gracePeriod=30 Oct 05 07:09:47 crc kubenswrapper[4846]: I1005 07:09:47.275692 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="095eba43-2814-4476-b482-efdf6b903de4" containerName="nova-api-api" containerID="cri-o://adcdee70fed38ca32d9ea08aa0872b7e277f0811e3da49432a923b205c6732a0" gracePeriod=30 Oct 05 07:09:47 crc kubenswrapper[4846]: I1005 07:09:47.290049 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 05 07:09:47 crc kubenswrapper[4846]: I1005 07:09:47.290310 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="ef292d07-1248-4edd-a182-9e9d572513ca" containerName="nova-scheduler-scheduler" containerID="cri-o://606907b830a2f5a35bf2b24719a4cd8e8b2a45dfc0e282ccce0bede035c53184" gracePeriod=30 Oct 05 07:09:47 crc kubenswrapper[4846]: I1005 07:09:47.290895 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/33193805-7aa6-406e-8cad-0ebbb1228ef2-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"33193805-7aa6-406e-8cad-0ebbb1228ef2\") " pod="openstack/nova-cell1-conductor-0" Oct 05 07:09:47 crc kubenswrapper[4846]: I1005 07:09:47.290958 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/33193805-7aa6-406e-8cad-0ebbb1228ef2-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"33193805-7aa6-406e-8cad-0ebbb1228ef2\") " pod="openstack/nova-cell1-conductor-0" Oct 05 07:09:47 crc kubenswrapper[4846]: I1005 07:09:47.291005 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2nd7f\" (UniqueName: \"kubernetes.io/projected/33193805-7aa6-406e-8cad-0ebbb1228ef2-kube-api-access-2nd7f\") pod \"nova-cell1-conductor-0\" (UID: \"33193805-7aa6-406e-8cad-0ebbb1228ef2\") " pod="openstack/nova-cell1-conductor-0" Oct 05 07:09:47 crc kubenswrapper[4846]: I1005 07:09:47.307239 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/33193805-7aa6-406e-8cad-0ebbb1228ef2-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"33193805-7aa6-406e-8cad-0ebbb1228ef2\") " pod="openstack/nova-cell1-conductor-0" Oct 05 07:09:47 crc kubenswrapper[4846]: I1005 07:09:47.308246 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/33193805-7aa6-406e-8cad-0ebbb1228ef2-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"33193805-7aa6-406e-8cad-0ebbb1228ef2\") " pod="openstack/nova-cell1-conductor-0" Oct 05 07:09:47 crc kubenswrapper[4846]: I1005 07:09:47.314903 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2nd7f\" (UniqueName: \"kubernetes.io/projected/33193805-7aa6-406e-8cad-0ebbb1228ef2-kube-api-access-2nd7f\") pod \"nova-cell1-conductor-0\" (UID: \"33193805-7aa6-406e-8cad-0ebbb1228ef2\") " pod="openstack/nova-cell1-conductor-0" Oct 05 07:09:47 crc kubenswrapper[4846]: I1005 07:09:47.342375 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 05 07:09:47 crc kubenswrapper[4846]: I1005 07:09:47.342719 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="2682f5d9-2987-4dad-86e2-750b48a2e697" containerName="nova-metadata-log" containerID="cri-o://0c53ca400292db7449e73c76ec3ad77058c533d8fe6bf07b8509da0f332598cc" gracePeriod=30 Oct 05 07:09:47 crc kubenswrapper[4846]: I1005 07:09:47.342922 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="2682f5d9-2987-4dad-86e2-750b48a2e697" containerName="nova-metadata-metadata" containerID="cri-o://d7c56637eb3aea87669012b06d0890cc0603d7e208e71bb55c05dbd4e172feed" gracePeriod=30 Oct 05 07:09:47 crc kubenswrapper[4846]: I1005 07:09:47.564284 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Oct 05 07:09:48 crc kubenswrapper[4846]: I1005 07:09:48.102618 4846 generic.go:334] "Generic (PLEG): container finished" podID="2682f5d9-2987-4dad-86e2-750b48a2e697" containerID="d7c56637eb3aea87669012b06d0890cc0603d7e208e71bb55c05dbd4e172feed" exitCode=0 Oct 05 07:09:48 crc kubenswrapper[4846]: I1005 07:09:48.103335 4846 generic.go:334] "Generic (PLEG): container finished" podID="2682f5d9-2987-4dad-86e2-750b48a2e697" containerID="0c53ca400292db7449e73c76ec3ad77058c533d8fe6bf07b8509da0f332598cc" exitCode=143 Oct 05 07:09:48 crc kubenswrapper[4846]: I1005 07:09:48.103417 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"2682f5d9-2987-4dad-86e2-750b48a2e697","Type":"ContainerDied","Data":"d7c56637eb3aea87669012b06d0890cc0603d7e208e71bb55c05dbd4e172feed"} Oct 05 07:09:48 crc kubenswrapper[4846]: I1005 07:09:48.103453 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"2682f5d9-2987-4dad-86e2-750b48a2e697","Type":"ContainerDied","Data":"0c53ca400292db7449e73c76ec3ad77058c533d8fe6bf07b8509da0f332598cc"} Oct 05 07:09:48 crc kubenswrapper[4846]: I1005 07:09:48.111869 4846 generic.go:334] "Generic (PLEG): container finished" podID="095eba43-2814-4476-b482-efdf6b903de4" containerID="a5ed027b4a2ba75551446c6b871445f74de3480c300dba8c804dc3238212ddc8" exitCode=143 Oct 05 07:09:48 crc kubenswrapper[4846]: I1005 07:09:48.111944 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"095eba43-2814-4476-b482-efdf6b903de4","Type":"ContainerDied","Data":"a5ed027b4a2ba75551446c6b871445f74de3480c300dba8c804dc3238212ddc8"} Oct 05 07:09:48 crc kubenswrapper[4846]: I1005 07:09:48.130618 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"eb03eca3-f45d-4361-ab00-9ce16bb2c96e","Type":"ContainerStarted","Data":"74b48fcfb6476893634deecb8bf996c2f28f54df151237a948e6426854564819"} Oct 05 07:09:48 crc kubenswrapper[4846]: I1005 07:09:48.257047 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 05 07:09:48 crc kubenswrapper[4846]: I1005 07:09:48.281656 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 05 07:09:48 crc kubenswrapper[4846]: W1005 07:09:48.288309 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod33193805_7aa6_406e_8cad_0ebbb1228ef2.slice/crio-470889702a1e2e05941ace0261d00f6838c4d71988b1b343ca1f365453bbe0b2 WatchSource:0}: Error finding container 470889702a1e2e05941ace0261d00f6838c4d71988b1b343ca1f365453bbe0b2: Status 404 returned error can't find the container with id 470889702a1e2e05941ace0261d00f6838c4d71988b1b343ca1f365453bbe0b2 Oct 05 07:09:48 crc kubenswrapper[4846]: I1005 07:09:48.336385 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2682f5d9-2987-4dad-86e2-750b48a2e697-config-data\") pod \"2682f5d9-2987-4dad-86e2-750b48a2e697\" (UID: \"2682f5d9-2987-4dad-86e2-750b48a2e697\") " Oct 05 07:09:48 crc kubenswrapper[4846]: I1005 07:09:48.337481 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zr6b4\" (UniqueName: \"kubernetes.io/projected/2682f5d9-2987-4dad-86e2-750b48a2e697-kube-api-access-zr6b4\") pod \"2682f5d9-2987-4dad-86e2-750b48a2e697\" (UID: \"2682f5d9-2987-4dad-86e2-750b48a2e697\") " Oct 05 07:09:48 crc kubenswrapper[4846]: I1005 07:09:48.337555 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2682f5d9-2987-4dad-86e2-750b48a2e697-combined-ca-bundle\") pod \"2682f5d9-2987-4dad-86e2-750b48a2e697\" (UID: \"2682f5d9-2987-4dad-86e2-750b48a2e697\") " Oct 05 07:09:48 crc kubenswrapper[4846]: I1005 07:09:48.337644 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2682f5d9-2987-4dad-86e2-750b48a2e697-logs\") pod \"2682f5d9-2987-4dad-86e2-750b48a2e697\" (UID: \"2682f5d9-2987-4dad-86e2-750b48a2e697\") " Oct 05 07:09:48 crc kubenswrapper[4846]: I1005 07:09:48.337704 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/2682f5d9-2987-4dad-86e2-750b48a2e697-nova-metadata-tls-certs\") pod \"2682f5d9-2987-4dad-86e2-750b48a2e697\" (UID: \"2682f5d9-2987-4dad-86e2-750b48a2e697\") " Oct 05 07:09:48 crc kubenswrapper[4846]: I1005 07:09:48.338315 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2682f5d9-2987-4dad-86e2-750b48a2e697-logs" (OuterVolumeSpecName: "logs") pod "2682f5d9-2987-4dad-86e2-750b48a2e697" (UID: "2682f5d9-2987-4dad-86e2-750b48a2e697"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 07:09:48 crc kubenswrapper[4846]: I1005 07:09:48.338813 4846 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2682f5d9-2987-4dad-86e2-750b48a2e697-logs\") on node \"crc\" DevicePath \"\"" Oct 05 07:09:48 crc kubenswrapper[4846]: I1005 07:09:48.356046 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2682f5d9-2987-4dad-86e2-750b48a2e697-kube-api-access-zr6b4" (OuterVolumeSpecName: "kube-api-access-zr6b4") pod "2682f5d9-2987-4dad-86e2-750b48a2e697" (UID: "2682f5d9-2987-4dad-86e2-750b48a2e697"). InnerVolumeSpecName "kube-api-access-zr6b4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:09:48 crc kubenswrapper[4846]: I1005 07:09:48.369124 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2682f5d9-2987-4dad-86e2-750b48a2e697-config-data" (OuterVolumeSpecName: "config-data") pod "2682f5d9-2987-4dad-86e2-750b48a2e697" (UID: "2682f5d9-2987-4dad-86e2-750b48a2e697"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:09:48 crc kubenswrapper[4846]: I1005 07:09:48.381812 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2682f5d9-2987-4dad-86e2-750b48a2e697-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2682f5d9-2987-4dad-86e2-750b48a2e697" (UID: "2682f5d9-2987-4dad-86e2-750b48a2e697"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:09:48 crc kubenswrapper[4846]: E1005 07:09:48.384381 4846 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="606907b830a2f5a35bf2b24719a4cd8e8b2a45dfc0e282ccce0bede035c53184" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 05 07:09:48 crc kubenswrapper[4846]: E1005 07:09:48.386479 4846 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="606907b830a2f5a35bf2b24719a4cd8e8b2a45dfc0e282ccce0bede035c53184" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 05 07:09:48 crc kubenswrapper[4846]: E1005 07:09:48.397427 4846 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="606907b830a2f5a35bf2b24719a4cd8e8b2a45dfc0e282ccce0bede035c53184" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 05 07:09:48 crc kubenswrapper[4846]: E1005 07:09:48.397496 4846 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="ef292d07-1248-4edd-a182-9e9d572513ca" containerName="nova-scheduler-scheduler" Oct 05 07:09:48 crc kubenswrapper[4846]: I1005 07:09:48.415026 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2682f5d9-2987-4dad-86e2-750b48a2e697-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "2682f5d9-2987-4dad-86e2-750b48a2e697" (UID: "2682f5d9-2987-4dad-86e2-750b48a2e697"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:09:48 crc kubenswrapper[4846]: I1005 07:09:48.441059 4846 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/2682f5d9-2987-4dad-86e2-750b48a2e697-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 05 07:09:48 crc kubenswrapper[4846]: I1005 07:09:48.441100 4846 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2682f5d9-2987-4dad-86e2-750b48a2e697-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 07:09:48 crc kubenswrapper[4846]: I1005 07:09:48.441115 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zr6b4\" (UniqueName: \"kubernetes.io/projected/2682f5d9-2987-4dad-86e2-750b48a2e697-kube-api-access-zr6b4\") on node \"crc\" DevicePath \"\"" Oct 05 07:09:48 crc kubenswrapper[4846]: I1005 07:09:48.441124 4846 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2682f5d9-2987-4dad-86e2-750b48a2e697-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 07:09:49 crc kubenswrapper[4846]: I1005 07:09:49.148561 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"2682f5d9-2987-4dad-86e2-750b48a2e697","Type":"ContainerDied","Data":"6f90f20ac50ffa971df116110da7c9581e170038ad7c8f1902aa1de25c3bf3ae"} Oct 05 07:09:49 crc kubenswrapper[4846]: I1005 07:09:49.149061 4846 scope.go:117] "RemoveContainer" containerID="d7c56637eb3aea87669012b06d0890cc0603d7e208e71bb55c05dbd4e172feed" Oct 05 07:09:49 crc kubenswrapper[4846]: I1005 07:09:49.149214 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 05 07:09:49 crc kubenswrapper[4846]: I1005 07:09:49.159905 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"33193805-7aa6-406e-8cad-0ebbb1228ef2","Type":"ContainerStarted","Data":"a03ec35e533cac1f50168f28070b82ea570152d6d222e9cf097e6f662b34b418"} Oct 05 07:09:49 crc kubenswrapper[4846]: I1005 07:09:49.159948 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"33193805-7aa6-406e-8cad-0ebbb1228ef2","Type":"ContainerStarted","Data":"470889702a1e2e05941ace0261d00f6838c4d71988b1b343ca1f365453bbe0b2"} Oct 05 07:09:49 crc kubenswrapper[4846]: I1005 07:09:49.160460 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Oct 05 07:09:49 crc kubenswrapper[4846]: I1005 07:09:49.174173 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"eb03eca3-f45d-4361-ab00-9ce16bb2c96e","Type":"ContainerStarted","Data":"fafa6b60a9e821e9bc3c916554dcc88b5f10181956baed8a807149ed1db4333e"} Oct 05 07:09:49 crc kubenswrapper[4846]: I1005 07:09:49.177862 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=2.177841695 podStartE2EDuration="2.177841695s" podCreationTimestamp="2025-10-05 07:09:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 07:09:49.176120789 +0000 UTC m=+1311.416973564" watchObservedRunningTime="2025-10-05 07:09:49.177841695 +0000 UTC m=+1311.418694470" Oct 05 07:09:49 crc kubenswrapper[4846]: I1005 07:09:49.309394 4846 scope.go:117] "RemoveContainer" containerID="0c53ca400292db7449e73c76ec3ad77058c533d8fe6bf07b8509da0f332598cc" Oct 05 07:09:49 crc kubenswrapper[4846]: I1005 07:09:49.333962 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 05 07:09:49 crc kubenswrapper[4846]: I1005 07:09:49.353777 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Oct 05 07:09:49 crc kubenswrapper[4846]: I1005 07:09:49.365661 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 05 07:09:49 crc kubenswrapper[4846]: E1005 07:09:49.366601 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2682f5d9-2987-4dad-86e2-750b48a2e697" containerName="nova-metadata-log" Oct 05 07:09:49 crc kubenswrapper[4846]: I1005 07:09:49.366718 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="2682f5d9-2987-4dad-86e2-750b48a2e697" containerName="nova-metadata-log" Oct 05 07:09:49 crc kubenswrapper[4846]: E1005 07:09:49.366811 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2682f5d9-2987-4dad-86e2-750b48a2e697" containerName="nova-metadata-metadata" Oct 05 07:09:49 crc kubenswrapper[4846]: I1005 07:09:49.366898 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="2682f5d9-2987-4dad-86e2-750b48a2e697" containerName="nova-metadata-metadata" Oct 05 07:09:49 crc kubenswrapper[4846]: I1005 07:09:49.367235 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="2682f5d9-2987-4dad-86e2-750b48a2e697" containerName="nova-metadata-metadata" Oct 05 07:09:49 crc kubenswrapper[4846]: I1005 07:09:49.367346 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="2682f5d9-2987-4dad-86e2-750b48a2e697" containerName="nova-metadata-log" Oct 05 07:09:49 crc kubenswrapper[4846]: I1005 07:09:49.369571 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 05 07:09:49 crc kubenswrapper[4846]: I1005 07:09:49.375666 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Oct 05 07:09:49 crc kubenswrapper[4846]: I1005 07:09:49.375885 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 05 07:09:49 crc kubenswrapper[4846]: I1005 07:09:49.381435 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 05 07:09:49 crc kubenswrapper[4846]: I1005 07:09:49.468307 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fe9fdc00-614b-4207-80f5-05a634a497a7-config-data\") pod \"nova-metadata-0\" (UID: \"fe9fdc00-614b-4207-80f5-05a634a497a7\") " pod="openstack/nova-metadata-0" Oct 05 07:09:49 crc kubenswrapper[4846]: I1005 07:09:49.468380 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fe9fdc00-614b-4207-80f5-05a634a497a7-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"fe9fdc00-614b-4207-80f5-05a634a497a7\") " pod="openstack/nova-metadata-0" Oct 05 07:09:49 crc kubenswrapper[4846]: I1005 07:09:49.468454 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/fe9fdc00-614b-4207-80f5-05a634a497a7-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"fe9fdc00-614b-4207-80f5-05a634a497a7\") " pod="openstack/nova-metadata-0" Oct 05 07:09:49 crc kubenswrapper[4846]: I1005 07:09:49.468588 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fe9fdc00-614b-4207-80f5-05a634a497a7-logs\") pod \"nova-metadata-0\" (UID: \"fe9fdc00-614b-4207-80f5-05a634a497a7\") " pod="openstack/nova-metadata-0" Oct 05 07:09:49 crc kubenswrapper[4846]: I1005 07:09:49.468754 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sp2jf\" (UniqueName: \"kubernetes.io/projected/fe9fdc00-614b-4207-80f5-05a634a497a7-kube-api-access-sp2jf\") pod \"nova-metadata-0\" (UID: \"fe9fdc00-614b-4207-80f5-05a634a497a7\") " pod="openstack/nova-metadata-0" Oct 05 07:09:49 crc kubenswrapper[4846]: I1005 07:09:49.569914 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fe9fdc00-614b-4207-80f5-05a634a497a7-logs\") pod \"nova-metadata-0\" (UID: \"fe9fdc00-614b-4207-80f5-05a634a497a7\") " pod="openstack/nova-metadata-0" Oct 05 07:09:49 crc kubenswrapper[4846]: I1005 07:09:49.570024 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sp2jf\" (UniqueName: \"kubernetes.io/projected/fe9fdc00-614b-4207-80f5-05a634a497a7-kube-api-access-sp2jf\") pod \"nova-metadata-0\" (UID: \"fe9fdc00-614b-4207-80f5-05a634a497a7\") " pod="openstack/nova-metadata-0" Oct 05 07:09:49 crc kubenswrapper[4846]: I1005 07:09:49.570068 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fe9fdc00-614b-4207-80f5-05a634a497a7-config-data\") pod \"nova-metadata-0\" (UID: \"fe9fdc00-614b-4207-80f5-05a634a497a7\") " pod="openstack/nova-metadata-0" Oct 05 07:09:49 crc kubenswrapper[4846]: I1005 07:09:49.570095 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fe9fdc00-614b-4207-80f5-05a634a497a7-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"fe9fdc00-614b-4207-80f5-05a634a497a7\") " pod="openstack/nova-metadata-0" Oct 05 07:09:49 crc kubenswrapper[4846]: I1005 07:09:49.570143 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/fe9fdc00-614b-4207-80f5-05a634a497a7-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"fe9fdc00-614b-4207-80f5-05a634a497a7\") " pod="openstack/nova-metadata-0" Oct 05 07:09:49 crc kubenswrapper[4846]: I1005 07:09:49.571078 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fe9fdc00-614b-4207-80f5-05a634a497a7-logs\") pod \"nova-metadata-0\" (UID: \"fe9fdc00-614b-4207-80f5-05a634a497a7\") " pod="openstack/nova-metadata-0" Oct 05 07:09:49 crc kubenswrapper[4846]: I1005 07:09:49.584120 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/fe9fdc00-614b-4207-80f5-05a634a497a7-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"fe9fdc00-614b-4207-80f5-05a634a497a7\") " pod="openstack/nova-metadata-0" Oct 05 07:09:49 crc kubenswrapper[4846]: I1005 07:09:49.584233 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fe9fdc00-614b-4207-80f5-05a634a497a7-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"fe9fdc00-614b-4207-80f5-05a634a497a7\") " pod="openstack/nova-metadata-0" Oct 05 07:09:49 crc kubenswrapper[4846]: I1005 07:09:49.589689 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sp2jf\" (UniqueName: \"kubernetes.io/projected/fe9fdc00-614b-4207-80f5-05a634a497a7-kube-api-access-sp2jf\") pod \"nova-metadata-0\" (UID: \"fe9fdc00-614b-4207-80f5-05a634a497a7\") " pod="openstack/nova-metadata-0" Oct 05 07:09:49 crc kubenswrapper[4846]: I1005 07:09:49.590566 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fe9fdc00-614b-4207-80f5-05a634a497a7-config-data\") pod \"nova-metadata-0\" (UID: \"fe9fdc00-614b-4207-80f5-05a634a497a7\") " pod="openstack/nova-metadata-0" Oct 05 07:09:49 crc kubenswrapper[4846]: I1005 07:09:49.691898 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 05 07:09:50 crc kubenswrapper[4846]: I1005 07:09:50.188583 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"eb03eca3-f45d-4361-ab00-9ce16bb2c96e","Type":"ContainerStarted","Data":"6f1f221eac0cd0e08e90b60a2002245bc6a6ed3b4c9a396d48f47cca0039feb9"} Oct 05 07:09:50 crc kubenswrapper[4846]: I1005 07:09:50.209076 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 05 07:09:50 crc kubenswrapper[4846]: W1005 07:09:50.481526 4846 container.go:586] Failed to update stats for container "/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod095eba43_2814_4476_b482_efdf6b903de4.slice/crio-08caf19eb603b238be913cdd839af47c24c884cc7134c517e2b4c45423bfd385": error while statting cgroup v2: [unable to parse /sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod095eba43_2814_4476_b482_efdf6b903de4.slice/crio-08caf19eb603b238be913cdd839af47c24c884cc7134c517e2b4c45423bfd385/memory.stat: read /sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod095eba43_2814_4476_b482_efdf6b903de4.slice/crio-08caf19eb603b238be913cdd839af47c24c884cc7134c517e2b4c45423bfd385/memory.stat: no such device], continuing to push stats Oct 05 07:09:50 crc kubenswrapper[4846]: I1005 07:09:50.511235 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2682f5d9-2987-4dad-86e2-750b48a2e697" path="/var/lib/kubelet/pods/2682f5d9-2987-4dad-86e2-750b48a2e697/volumes" Oct 05 07:09:50 crc kubenswrapper[4846]: I1005 07:09:50.857385 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 05 07:09:50 crc kubenswrapper[4846]: I1005 07:09:50.906572 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/095eba43-2814-4476-b482-efdf6b903de4-logs\") pod \"095eba43-2814-4476-b482-efdf6b903de4\" (UID: \"095eba43-2814-4476-b482-efdf6b903de4\") " Oct 05 07:09:50 crc kubenswrapper[4846]: I1005 07:09:50.906681 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9gb6s\" (UniqueName: \"kubernetes.io/projected/095eba43-2814-4476-b482-efdf6b903de4-kube-api-access-9gb6s\") pod \"095eba43-2814-4476-b482-efdf6b903de4\" (UID: \"095eba43-2814-4476-b482-efdf6b903de4\") " Oct 05 07:09:50 crc kubenswrapper[4846]: I1005 07:09:50.906756 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/095eba43-2814-4476-b482-efdf6b903de4-combined-ca-bundle\") pod \"095eba43-2814-4476-b482-efdf6b903de4\" (UID: \"095eba43-2814-4476-b482-efdf6b903de4\") " Oct 05 07:09:50 crc kubenswrapper[4846]: I1005 07:09:50.906782 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/095eba43-2814-4476-b482-efdf6b903de4-config-data\") pod \"095eba43-2814-4476-b482-efdf6b903de4\" (UID: \"095eba43-2814-4476-b482-efdf6b903de4\") " Oct 05 07:09:50 crc kubenswrapper[4846]: I1005 07:09:50.907218 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/095eba43-2814-4476-b482-efdf6b903de4-logs" (OuterVolumeSpecName: "logs") pod "095eba43-2814-4476-b482-efdf6b903de4" (UID: "095eba43-2814-4476-b482-efdf6b903de4"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 07:09:50 crc kubenswrapper[4846]: I1005 07:09:50.914495 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/095eba43-2814-4476-b482-efdf6b903de4-kube-api-access-9gb6s" (OuterVolumeSpecName: "kube-api-access-9gb6s") pod "095eba43-2814-4476-b482-efdf6b903de4" (UID: "095eba43-2814-4476-b482-efdf6b903de4"). InnerVolumeSpecName "kube-api-access-9gb6s". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:09:50 crc kubenswrapper[4846]: I1005 07:09:50.940921 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/095eba43-2814-4476-b482-efdf6b903de4-config-data" (OuterVolumeSpecName: "config-data") pod "095eba43-2814-4476-b482-efdf6b903de4" (UID: "095eba43-2814-4476-b482-efdf6b903de4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:09:50 crc kubenswrapper[4846]: I1005 07:09:50.963928 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/095eba43-2814-4476-b482-efdf6b903de4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "095eba43-2814-4476-b482-efdf6b903de4" (UID: "095eba43-2814-4476-b482-efdf6b903de4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:09:51 crc kubenswrapper[4846]: I1005 07:09:51.010236 4846 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/095eba43-2814-4476-b482-efdf6b903de4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 07:09:51 crc kubenswrapper[4846]: I1005 07:09:51.010294 4846 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/095eba43-2814-4476-b482-efdf6b903de4-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 07:09:51 crc kubenswrapper[4846]: I1005 07:09:51.010308 4846 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/095eba43-2814-4476-b482-efdf6b903de4-logs\") on node \"crc\" DevicePath \"\"" Oct 05 07:09:51 crc kubenswrapper[4846]: I1005 07:09:51.010321 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9gb6s\" (UniqueName: \"kubernetes.io/projected/095eba43-2814-4476-b482-efdf6b903de4-kube-api-access-9gb6s\") on node \"crc\" DevicePath \"\"" Oct 05 07:09:51 crc kubenswrapper[4846]: I1005 07:09:51.057876 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 05 07:09:51 crc kubenswrapper[4846]: I1005 07:09:51.110735 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef292d07-1248-4edd-a182-9e9d572513ca-combined-ca-bundle\") pod \"ef292d07-1248-4edd-a182-9e9d572513ca\" (UID: \"ef292d07-1248-4edd-a182-9e9d572513ca\") " Oct 05 07:09:51 crc kubenswrapper[4846]: I1005 07:09:51.110797 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tnx6l\" (UniqueName: \"kubernetes.io/projected/ef292d07-1248-4edd-a182-9e9d572513ca-kube-api-access-tnx6l\") pod \"ef292d07-1248-4edd-a182-9e9d572513ca\" (UID: \"ef292d07-1248-4edd-a182-9e9d572513ca\") " Oct 05 07:09:51 crc kubenswrapper[4846]: I1005 07:09:51.110873 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ef292d07-1248-4edd-a182-9e9d572513ca-config-data\") pod \"ef292d07-1248-4edd-a182-9e9d572513ca\" (UID: \"ef292d07-1248-4edd-a182-9e9d572513ca\") " Oct 05 07:09:51 crc kubenswrapper[4846]: I1005 07:09:51.117557 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ef292d07-1248-4edd-a182-9e9d572513ca-kube-api-access-tnx6l" (OuterVolumeSpecName: "kube-api-access-tnx6l") pod "ef292d07-1248-4edd-a182-9e9d572513ca" (UID: "ef292d07-1248-4edd-a182-9e9d572513ca"). InnerVolumeSpecName "kube-api-access-tnx6l". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:09:51 crc kubenswrapper[4846]: I1005 07:09:51.145410 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ef292d07-1248-4edd-a182-9e9d572513ca-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ef292d07-1248-4edd-a182-9e9d572513ca" (UID: "ef292d07-1248-4edd-a182-9e9d572513ca"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:09:51 crc kubenswrapper[4846]: I1005 07:09:51.151937 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ef292d07-1248-4edd-a182-9e9d572513ca-config-data" (OuterVolumeSpecName: "config-data") pod "ef292d07-1248-4edd-a182-9e9d572513ca" (UID: "ef292d07-1248-4edd-a182-9e9d572513ca"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:09:51 crc kubenswrapper[4846]: I1005 07:09:51.202338 4846 generic.go:334] "Generic (PLEG): container finished" podID="ef292d07-1248-4edd-a182-9e9d572513ca" containerID="606907b830a2f5a35bf2b24719a4cd8e8b2a45dfc0e282ccce0bede035c53184" exitCode=0 Oct 05 07:09:51 crc kubenswrapper[4846]: I1005 07:09:51.202397 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 05 07:09:51 crc kubenswrapper[4846]: I1005 07:09:51.202428 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"ef292d07-1248-4edd-a182-9e9d572513ca","Type":"ContainerDied","Data":"606907b830a2f5a35bf2b24719a4cd8e8b2a45dfc0e282ccce0bede035c53184"} Oct 05 07:09:51 crc kubenswrapper[4846]: I1005 07:09:51.202460 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"ef292d07-1248-4edd-a182-9e9d572513ca","Type":"ContainerDied","Data":"1be7426753a42cbaec966d06366b74abb18d77021d5fcffa3a4b8826192f3744"} Oct 05 07:09:51 crc kubenswrapper[4846]: I1005 07:09:51.202479 4846 scope.go:117] "RemoveContainer" containerID="606907b830a2f5a35bf2b24719a4cd8e8b2a45dfc0e282ccce0bede035c53184" Oct 05 07:09:51 crc kubenswrapper[4846]: I1005 07:09:51.208351 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"eb03eca3-f45d-4361-ab00-9ce16bb2c96e","Type":"ContainerStarted","Data":"32020fcfa11e9aff11fbbf804e18681fcebe28c354eef5dfd4c7dc9be93e5965"} Oct 05 07:09:51 crc kubenswrapper[4846]: I1005 07:09:51.209307 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 05 07:09:51 crc kubenswrapper[4846]: I1005 07:09:51.216206 4846 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef292d07-1248-4edd-a182-9e9d572513ca-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 07:09:51 crc kubenswrapper[4846]: I1005 07:09:51.216244 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tnx6l\" (UniqueName: \"kubernetes.io/projected/ef292d07-1248-4edd-a182-9e9d572513ca-kube-api-access-tnx6l\") on node \"crc\" DevicePath \"\"" Oct 05 07:09:51 crc kubenswrapper[4846]: I1005 07:09:51.216261 4846 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ef292d07-1248-4edd-a182-9e9d572513ca-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 07:09:51 crc kubenswrapper[4846]: I1005 07:09:51.218358 4846 generic.go:334] "Generic (PLEG): container finished" podID="095eba43-2814-4476-b482-efdf6b903de4" containerID="adcdee70fed38ca32d9ea08aa0872b7e277f0811e3da49432a923b205c6732a0" exitCode=0 Oct 05 07:09:51 crc kubenswrapper[4846]: I1005 07:09:51.218426 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"095eba43-2814-4476-b482-efdf6b903de4","Type":"ContainerDied","Data":"adcdee70fed38ca32d9ea08aa0872b7e277f0811e3da49432a923b205c6732a0"} Oct 05 07:09:51 crc kubenswrapper[4846]: I1005 07:09:51.218456 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"095eba43-2814-4476-b482-efdf6b903de4","Type":"ContainerDied","Data":"08caf19eb603b238be913cdd839af47c24c884cc7134c517e2b4c45423bfd385"} Oct 05 07:09:51 crc kubenswrapper[4846]: I1005 07:09:51.218526 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 05 07:09:51 crc kubenswrapper[4846]: I1005 07:09:51.225951 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"fe9fdc00-614b-4207-80f5-05a634a497a7","Type":"ContainerStarted","Data":"6000ba86c9c9f2027bd31401f3a69a05f4e695ece27d93cfabaa5898532368be"} Oct 05 07:09:51 crc kubenswrapper[4846]: I1005 07:09:51.226017 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"fe9fdc00-614b-4207-80f5-05a634a497a7","Type":"ContainerStarted","Data":"a7ae54073f10d9d332e2d936a42d58cf1af318a6d1f5ced07d8420dcfa60102c"} Oct 05 07:09:51 crc kubenswrapper[4846]: I1005 07:09:51.226030 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"fe9fdc00-614b-4207-80f5-05a634a497a7","Type":"ContainerStarted","Data":"fc1182f4fca3742cd41a53703e689fed3143be3c9136c4cd44ab6b84054af6b1"} Oct 05 07:09:51 crc kubenswrapper[4846]: I1005 07:09:51.252163 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.363311242 podStartE2EDuration="6.252133404s" podCreationTimestamp="2025-10-05 07:09:45 +0000 UTC" firstStartedPulling="2025-10-05 07:09:46.628769395 +0000 UTC m=+1308.869622170" lastFinishedPulling="2025-10-05 07:09:50.517591547 +0000 UTC m=+1312.758444332" observedRunningTime="2025-10-05 07:09:51.246639427 +0000 UTC m=+1313.487492212" watchObservedRunningTime="2025-10-05 07:09:51.252133404 +0000 UTC m=+1313.492986189" Oct 05 07:09:51 crc kubenswrapper[4846]: I1005 07:09:51.274752 4846 scope.go:117] "RemoveContainer" containerID="606907b830a2f5a35bf2b24719a4cd8e8b2a45dfc0e282ccce0bede035c53184" Oct 05 07:09:51 crc kubenswrapper[4846]: E1005 07:09:51.277303 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"606907b830a2f5a35bf2b24719a4cd8e8b2a45dfc0e282ccce0bede035c53184\": container with ID starting with 606907b830a2f5a35bf2b24719a4cd8e8b2a45dfc0e282ccce0bede035c53184 not found: ID does not exist" containerID="606907b830a2f5a35bf2b24719a4cd8e8b2a45dfc0e282ccce0bede035c53184" Oct 05 07:09:51 crc kubenswrapper[4846]: I1005 07:09:51.277774 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"606907b830a2f5a35bf2b24719a4cd8e8b2a45dfc0e282ccce0bede035c53184"} err="failed to get container status \"606907b830a2f5a35bf2b24719a4cd8e8b2a45dfc0e282ccce0bede035c53184\": rpc error: code = NotFound desc = could not find container \"606907b830a2f5a35bf2b24719a4cd8e8b2a45dfc0e282ccce0bede035c53184\": container with ID starting with 606907b830a2f5a35bf2b24719a4cd8e8b2a45dfc0e282ccce0bede035c53184 not found: ID does not exist" Oct 05 07:09:51 crc kubenswrapper[4846]: I1005 07:09:51.277836 4846 scope.go:117] "RemoveContainer" containerID="adcdee70fed38ca32d9ea08aa0872b7e277f0811e3da49432a923b205c6732a0" Oct 05 07:09:51 crc kubenswrapper[4846]: I1005 07:09:51.299566 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 05 07:09:51 crc kubenswrapper[4846]: I1005 07:09:51.310814 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Oct 05 07:09:51 crc kubenswrapper[4846]: I1005 07:09:51.313143 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Oct 05 07:09:51 crc kubenswrapper[4846]: I1005 07:09:51.319468 4846 scope.go:117] "RemoveContainer" containerID="a5ed027b4a2ba75551446c6b871445f74de3480c300dba8c804dc3238212ddc8" Oct 05 07:09:51 crc kubenswrapper[4846]: I1005 07:09:51.374965 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 05 07:09:51 crc kubenswrapper[4846]: I1005 07:09:51.405997 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Oct 05 07:09:51 crc kubenswrapper[4846]: I1005 07:09:51.406140 4846 scope.go:117] "RemoveContainer" containerID="adcdee70fed38ca32d9ea08aa0872b7e277f0811e3da49432a923b205c6732a0" Oct 05 07:09:51 crc kubenswrapper[4846]: E1005 07:09:51.406560 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ef292d07-1248-4edd-a182-9e9d572513ca" containerName="nova-scheduler-scheduler" Oct 05 07:09:51 crc kubenswrapper[4846]: I1005 07:09:51.406575 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="ef292d07-1248-4edd-a182-9e9d572513ca" containerName="nova-scheduler-scheduler" Oct 05 07:09:51 crc kubenswrapper[4846]: E1005 07:09:51.406589 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="095eba43-2814-4476-b482-efdf6b903de4" containerName="nova-api-log" Oct 05 07:09:51 crc kubenswrapper[4846]: I1005 07:09:51.406595 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="095eba43-2814-4476-b482-efdf6b903de4" containerName="nova-api-log" Oct 05 07:09:51 crc kubenswrapper[4846]: E1005 07:09:51.406621 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="095eba43-2814-4476-b482-efdf6b903de4" containerName="nova-api-api" Oct 05 07:09:51 crc kubenswrapper[4846]: I1005 07:09:51.406627 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="095eba43-2814-4476-b482-efdf6b903de4" containerName="nova-api-api" Oct 05 07:09:51 crc kubenswrapper[4846]: I1005 07:09:51.406840 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="095eba43-2814-4476-b482-efdf6b903de4" containerName="nova-api-log" Oct 05 07:09:51 crc kubenswrapper[4846]: I1005 07:09:51.406867 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="ef292d07-1248-4edd-a182-9e9d572513ca" containerName="nova-scheduler-scheduler" Oct 05 07:09:51 crc kubenswrapper[4846]: I1005 07:09:51.406879 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="095eba43-2814-4476-b482-efdf6b903de4" containerName="nova-api-api" Oct 05 07:09:51 crc kubenswrapper[4846]: E1005 07:09:51.407421 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"adcdee70fed38ca32d9ea08aa0872b7e277f0811e3da49432a923b205c6732a0\": container with ID starting with adcdee70fed38ca32d9ea08aa0872b7e277f0811e3da49432a923b205c6732a0 not found: ID does not exist" containerID="adcdee70fed38ca32d9ea08aa0872b7e277f0811e3da49432a923b205c6732a0" Oct 05 07:09:51 crc kubenswrapper[4846]: I1005 07:09:51.407456 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"adcdee70fed38ca32d9ea08aa0872b7e277f0811e3da49432a923b205c6732a0"} err="failed to get container status \"adcdee70fed38ca32d9ea08aa0872b7e277f0811e3da49432a923b205c6732a0\": rpc error: code = NotFound desc = could not find container \"adcdee70fed38ca32d9ea08aa0872b7e277f0811e3da49432a923b205c6732a0\": container with ID starting with adcdee70fed38ca32d9ea08aa0872b7e277f0811e3da49432a923b205c6732a0 not found: ID does not exist" Oct 05 07:09:51 crc kubenswrapper[4846]: I1005 07:09:51.407486 4846 scope.go:117] "RemoveContainer" containerID="a5ed027b4a2ba75551446c6b871445f74de3480c300dba8c804dc3238212ddc8" Oct 05 07:09:51 crc kubenswrapper[4846]: I1005 07:09:51.407678 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 05 07:09:51 crc kubenswrapper[4846]: E1005 07:09:51.407949 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a5ed027b4a2ba75551446c6b871445f74de3480c300dba8c804dc3238212ddc8\": container with ID starting with a5ed027b4a2ba75551446c6b871445f74de3480c300dba8c804dc3238212ddc8 not found: ID does not exist" containerID="a5ed027b4a2ba75551446c6b871445f74de3480c300dba8c804dc3238212ddc8" Oct 05 07:09:51 crc kubenswrapper[4846]: I1005 07:09:51.407988 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a5ed027b4a2ba75551446c6b871445f74de3480c300dba8c804dc3238212ddc8"} err="failed to get container status \"a5ed027b4a2ba75551446c6b871445f74de3480c300dba8c804dc3238212ddc8\": rpc error: code = NotFound desc = could not find container \"a5ed027b4a2ba75551446c6b871445f74de3480c300dba8c804dc3238212ddc8\": container with ID starting with a5ed027b4a2ba75551446c6b871445f74de3480c300dba8c804dc3238212ddc8 not found: ID does not exist" Oct 05 07:09:51 crc kubenswrapper[4846]: I1005 07:09:51.410738 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Oct 05 07:09:51 crc kubenswrapper[4846]: I1005 07:09:51.424228 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ce94efea-dc0a-4d7a-9340-9217ead55549-config-data\") pod \"nova-scheduler-0\" (UID: \"ce94efea-dc0a-4d7a-9340-9217ead55549\") " pod="openstack/nova-scheduler-0" Oct 05 07:09:51 crc kubenswrapper[4846]: I1005 07:09:51.426362 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ce94efea-dc0a-4d7a-9340-9217ead55549-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"ce94efea-dc0a-4d7a-9340-9217ead55549\") " pod="openstack/nova-scheduler-0" Oct 05 07:09:51 crc kubenswrapper[4846]: I1005 07:09:51.426438 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-86fq9\" (UniqueName: \"kubernetes.io/projected/ce94efea-dc0a-4d7a-9340-9217ead55549-kube-api-access-86fq9\") pod \"nova-scheduler-0\" (UID: \"ce94efea-dc0a-4d7a-9340-9217ead55549\") " pod="openstack/nova-scheduler-0" Oct 05 07:09:51 crc kubenswrapper[4846]: I1005 07:09:51.431738 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Oct 05 07:09:51 crc kubenswrapper[4846]: I1005 07:09:51.455253 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 05 07:09:51 crc kubenswrapper[4846]: I1005 07:09:51.461061 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.46104151 podStartE2EDuration="2.46104151s" podCreationTimestamp="2025-10-05 07:09:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 07:09:51.320062867 +0000 UTC m=+1313.560915642" watchObservedRunningTime="2025-10-05 07:09:51.46104151 +0000 UTC m=+1313.701894295" Oct 05 07:09:51 crc kubenswrapper[4846]: I1005 07:09:51.473716 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 05 07:09:51 crc kubenswrapper[4846]: I1005 07:09:51.476015 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 05 07:09:51 crc kubenswrapper[4846]: I1005 07:09:51.479083 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 05 07:09:51 crc kubenswrapper[4846]: I1005 07:09:51.482809 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 05 07:09:51 crc kubenswrapper[4846]: I1005 07:09:51.528749 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f1e8f9a4-b671-4417-99c9-90bda3c17339-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"f1e8f9a4-b671-4417-99c9-90bda3c17339\") " pod="openstack/nova-api-0" Oct 05 07:09:51 crc kubenswrapper[4846]: I1005 07:09:51.528807 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f1e8f9a4-b671-4417-99c9-90bda3c17339-config-data\") pod \"nova-api-0\" (UID: \"f1e8f9a4-b671-4417-99c9-90bda3c17339\") " pod="openstack/nova-api-0" Oct 05 07:09:51 crc kubenswrapper[4846]: I1005 07:09:51.528845 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gc265\" (UniqueName: \"kubernetes.io/projected/f1e8f9a4-b671-4417-99c9-90bda3c17339-kube-api-access-gc265\") pod \"nova-api-0\" (UID: \"f1e8f9a4-b671-4417-99c9-90bda3c17339\") " pod="openstack/nova-api-0" Oct 05 07:09:51 crc kubenswrapper[4846]: I1005 07:09:51.529062 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ce94efea-dc0a-4d7a-9340-9217ead55549-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"ce94efea-dc0a-4d7a-9340-9217ead55549\") " pod="openstack/nova-scheduler-0" Oct 05 07:09:51 crc kubenswrapper[4846]: I1005 07:09:51.529129 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f1e8f9a4-b671-4417-99c9-90bda3c17339-logs\") pod \"nova-api-0\" (UID: \"f1e8f9a4-b671-4417-99c9-90bda3c17339\") " pod="openstack/nova-api-0" Oct 05 07:09:51 crc kubenswrapper[4846]: I1005 07:09:51.529323 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-86fq9\" (UniqueName: \"kubernetes.io/projected/ce94efea-dc0a-4d7a-9340-9217ead55549-kube-api-access-86fq9\") pod \"nova-scheduler-0\" (UID: \"ce94efea-dc0a-4d7a-9340-9217ead55549\") " pod="openstack/nova-scheduler-0" Oct 05 07:09:51 crc kubenswrapper[4846]: I1005 07:09:51.529478 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ce94efea-dc0a-4d7a-9340-9217ead55549-config-data\") pod \"nova-scheduler-0\" (UID: \"ce94efea-dc0a-4d7a-9340-9217ead55549\") " pod="openstack/nova-scheduler-0" Oct 05 07:09:51 crc kubenswrapper[4846]: I1005 07:09:51.540275 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ce94efea-dc0a-4d7a-9340-9217ead55549-config-data\") pod \"nova-scheduler-0\" (UID: \"ce94efea-dc0a-4d7a-9340-9217ead55549\") " pod="openstack/nova-scheduler-0" Oct 05 07:09:51 crc kubenswrapper[4846]: I1005 07:09:51.540466 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ce94efea-dc0a-4d7a-9340-9217ead55549-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"ce94efea-dc0a-4d7a-9340-9217ead55549\") " pod="openstack/nova-scheduler-0" Oct 05 07:09:51 crc kubenswrapper[4846]: I1005 07:09:51.553254 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-86fq9\" (UniqueName: \"kubernetes.io/projected/ce94efea-dc0a-4d7a-9340-9217ead55549-kube-api-access-86fq9\") pod \"nova-scheduler-0\" (UID: \"ce94efea-dc0a-4d7a-9340-9217ead55549\") " pod="openstack/nova-scheduler-0" Oct 05 07:09:51 crc kubenswrapper[4846]: I1005 07:09:51.631845 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f1e8f9a4-b671-4417-99c9-90bda3c17339-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"f1e8f9a4-b671-4417-99c9-90bda3c17339\") " pod="openstack/nova-api-0" Oct 05 07:09:51 crc kubenswrapper[4846]: I1005 07:09:51.631932 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f1e8f9a4-b671-4417-99c9-90bda3c17339-config-data\") pod \"nova-api-0\" (UID: \"f1e8f9a4-b671-4417-99c9-90bda3c17339\") " pod="openstack/nova-api-0" Oct 05 07:09:51 crc kubenswrapper[4846]: I1005 07:09:51.631977 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gc265\" (UniqueName: \"kubernetes.io/projected/f1e8f9a4-b671-4417-99c9-90bda3c17339-kube-api-access-gc265\") pod \"nova-api-0\" (UID: \"f1e8f9a4-b671-4417-99c9-90bda3c17339\") " pod="openstack/nova-api-0" Oct 05 07:09:51 crc kubenswrapper[4846]: I1005 07:09:51.632070 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f1e8f9a4-b671-4417-99c9-90bda3c17339-logs\") pod \"nova-api-0\" (UID: \"f1e8f9a4-b671-4417-99c9-90bda3c17339\") " pod="openstack/nova-api-0" Oct 05 07:09:51 crc kubenswrapper[4846]: I1005 07:09:51.632663 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f1e8f9a4-b671-4417-99c9-90bda3c17339-logs\") pod \"nova-api-0\" (UID: \"f1e8f9a4-b671-4417-99c9-90bda3c17339\") " pod="openstack/nova-api-0" Oct 05 07:09:51 crc kubenswrapper[4846]: I1005 07:09:51.636803 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f1e8f9a4-b671-4417-99c9-90bda3c17339-config-data\") pod \"nova-api-0\" (UID: \"f1e8f9a4-b671-4417-99c9-90bda3c17339\") " pod="openstack/nova-api-0" Oct 05 07:09:51 crc kubenswrapper[4846]: I1005 07:09:51.637078 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f1e8f9a4-b671-4417-99c9-90bda3c17339-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"f1e8f9a4-b671-4417-99c9-90bda3c17339\") " pod="openstack/nova-api-0" Oct 05 07:09:51 crc kubenswrapper[4846]: I1005 07:09:51.654154 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gc265\" (UniqueName: \"kubernetes.io/projected/f1e8f9a4-b671-4417-99c9-90bda3c17339-kube-api-access-gc265\") pod \"nova-api-0\" (UID: \"f1e8f9a4-b671-4417-99c9-90bda3c17339\") " pod="openstack/nova-api-0" Oct 05 07:09:51 crc kubenswrapper[4846]: I1005 07:09:51.738620 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 05 07:09:51 crc kubenswrapper[4846]: I1005 07:09:51.797846 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 05 07:09:52 crc kubenswrapper[4846]: I1005 07:09:52.336051 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 05 07:09:52 crc kubenswrapper[4846]: I1005 07:09:52.432129 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 05 07:09:52 crc kubenswrapper[4846]: W1005 07:09:52.435511 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podce94efea_dc0a_4d7a_9340_9217ead55549.slice/crio-ecde9fcad845c81d90b7d0ebb4aa5515cb50bc02a9ba8b568a05fa84ec21b854 WatchSource:0}: Error finding container ecde9fcad845c81d90b7d0ebb4aa5515cb50bc02a9ba8b568a05fa84ec21b854: Status 404 returned error can't find the container with id ecde9fcad845c81d90b7d0ebb4aa5515cb50bc02a9ba8b568a05fa84ec21b854 Oct 05 07:09:52 crc kubenswrapper[4846]: I1005 07:09:52.509077 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="095eba43-2814-4476-b482-efdf6b903de4" path="/var/lib/kubelet/pods/095eba43-2814-4476-b482-efdf6b903de4/volumes" Oct 05 07:09:52 crc kubenswrapper[4846]: I1005 07:09:52.509792 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ef292d07-1248-4edd-a182-9e9d572513ca" path="/var/lib/kubelet/pods/ef292d07-1248-4edd-a182-9e9d572513ca/volumes" Oct 05 07:09:53 crc kubenswrapper[4846]: I1005 07:09:53.255803 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"f1e8f9a4-b671-4417-99c9-90bda3c17339","Type":"ContainerStarted","Data":"53602a509bd81a4226a9664cce98e43c4223e89f4b02f63f45ef63bdac84c69e"} Oct 05 07:09:53 crc kubenswrapper[4846]: I1005 07:09:53.256413 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"f1e8f9a4-b671-4417-99c9-90bda3c17339","Type":"ContainerStarted","Data":"7ec1da9974ee7548066528f3fd8df37c374953980dee2632e3db26ed5b6f0b8f"} Oct 05 07:09:53 crc kubenswrapper[4846]: I1005 07:09:53.256455 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"f1e8f9a4-b671-4417-99c9-90bda3c17339","Type":"ContainerStarted","Data":"40a7a409bbe0ba6f0b9e9714102ae06da4cc85a991723f874c127b4180ebe5cf"} Oct 05 07:09:53 crc kubenswrapper[4846]: I1005 07:09:53.257807 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"ce94efea-dc0a-4d7a-9340-9217ead55549","Type":"ContainerStarted","Data":"f8ef556e3cb59bcf872a739907d82bb2061305bc039de4bbd142d1bef463ec56"} Oct 05 07:09:53 crc kubenswrapper[4846]: I1005 07:09:53.257881 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"ce94efea-dc0a-4d7a-9340-9217ead55549","Type":"ContainerStarted","Data":"ecde9fcad845c81d90b7d0ebb4aa5515cb50bc02a9ba8b568a05fa84ec21b854"} Oct 05 07:09:53 crc kubenswrapper[4846]: I1005 07:09:53.286769 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.286745152 podStartE2EDuration="2.286745152s" podCreationTimestamp="2025-10-05 07:09:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 07:09:53.278651446 +0000 UTC m=+1315.519504231" watchObservedRunningTime="2025-10-05 07:09:53.286745152 +0000 UTC m=+1315.527597927" Oct 05 07:09:53 crc kubenswrapper[4846]: I1005 07:09:53.320954 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.320925944 podStartE2EDuration="2.320925944s" podCreationTimestamp="2025-10-05 07:09:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 07:09:53.312005496 +0000 UTC m=+1315.552858281" watchObservedRunningTime="2025-10-05 07:09:53.320925944 +0000 UTC m=+1315.561778729" Oct 05 07:09:53 crc kubenswrapper[4846]: I1005 07:09:53.325396 4846 patch_prober.go:28] interesting pod/machine-config-daemon-fscvf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 07:09:53 crc kubenswrapper[4846]: I1005 07:09:53.325487 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 07:09:54 crc kubenswrapper[4846]: I1005 07:09:54.692078 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 05 07:09:54 crc kubenswrapper[4846]: I1005 07:09:54.692139 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 05 07:09:56 crc kubenswrapper[4846]: I1005 07:09:56.739562 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Oct 05 07:09:57 crc kubenswrapper[4846]: I1005 07:09:57.611510 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Oct 05 07:09:59 crc kubenswrapper[4846]: I1005 07:09:59.692074 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 05 07:09:59 crc kubenswrapper[4846]: I1005 07:09:59.693100 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 05 07:10:00 crc kubenswrapper[4846]: I1005 07:10:00.707329 4846 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="fe9fdc00-614b-4207-80f5-05a634a497a7" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.197:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 05 07:10:00 crc kubenswrapper[4846]: I1005 07:10:00.707329 4846 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="fe9fdc00-614b-4207-80f5-05a634a497a7" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.197:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 05 07:10:01 crc kubenswrapper[4846]: I1005 07:10:01.740024 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Oct 05 07:10:01 crc kubenswrapper[4846]: I1005 07:10:01.773262 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Oct 05 07:10:01 crc kubenswrapper[4846]: I1005 07:10:01.799094 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 05 07:10:01 crc kubenswrapper[4846]: I1005 07:10:01.799171 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 05 07:10:02 crc kubenswrapper[4846]: I1005 07:10:02.428940 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Oct 05 07:10:02 crc kubenswrapper[4846]: I1005 07:10:02.882802 4846 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="f1e8f9a4-b671-4417-99c9-90bda3c17339" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.199:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 05 07:10:02 crc kubenswrapper[4846]: I1005 07:10:02.882819 4846 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="f1e8f9a4-b671-4417-99c9-90bda3c17339" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.199:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 05 07:10:09 crc kubenswrapper[4846]: I1005 07:10:09.225715 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 05 07:10:09 crc kubenswrapper[4846]: I1005 07:10:09.385821 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a0844571-8880-4fa4-af42-2da48aa8bdec-combined-ca-bundle\") pod \"a0844571-8880-4fa4-af42-2da48aa8bdec\" (UID: \"a0844571-8880-4fa4-af42-2da48aa8bdec\") " Oct 05 07:10:09 crc kubenswrapper[4846]: I1005 07:10:09.385924 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a0844571-8880-4fa4-af42-2da48aa8bdec-config-data\") pod \"a0844571-8880-4fa4-af42-2da48aa8bdec\" (UID: \"a0844571-8880-4fa4-af42-2da48aa8bdec\") " Oct 05 07:10:09 crc kubenswrapper[4846]: I1005 07:10:09.386002 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5spwx\" (UniqueName: \"kubernetes.io/projected/a0844571-8880-4fa4-af42-2da48aa8bdec-kube-api-access-5spwx\") pod \"a0844571-8880-4fa4-af42-2da48aa8bdec\" (UID: \"a0844571-8880-4fa4-af42-2da48aa8bdec\") " Oct 05 07:10:09 crc kubenswrapper[4846]: I1005 07:10:09.394475 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0844571-8880-4fa4-af42-2da48aa8bdec-kube-api-access-5spwx" (OuterVolumeSpecName: "kube-api-access-5spwx") pod "a0844571-8880-4fa4-af42-2da48aa8bdec" (UID: "a0844571-8880-4fa4-af42-2da48aa8bdec"). InnerVolumeSpecName "kube-api-access-5spwx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:10:09 crc kubenswrapper[4846]: I1005 07:10:09.422539 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0844571-8880-4fa4-af42-2da48aa8bdec-config-data" (OuterVolumeSpecName: "config-data") pod "a0844571-8880-4fa4-af42-2da48aa8bdec" (UID: "a0844571-8880-4fa4-af42-2da48aa8bdec"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:10:09 crc kubenswrapper[4846]: I1005 07:10:09.435833 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0844571-8880-4fa4-af42-2da48aa8bdec-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a0844571-8880-4fa4-af42-2da48aa8bdec" (UID: "a0844571-8880-4fa4-af42-2da48aa8bdec"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:10:09 crc kubenswrapper[4846]: I1005 07:10:09.478841 4846 generic.go:334] "Generic (PLEG): container finished" podID="a0844571-8880-4fa4-af42-2da48aa8bdec" containerID="cb184b8b7a1c9f825f5c48f4b882e586772ff863671c32a461519f9e717aab6e" exitCode=137 Oct 05 07:10:09 crc kubenswrapper[4846]: I1005 07:10:09.478904 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"a0844571-8880-4fa4-af42-2da48aa8bdec","Type":"ContainerDied","Data":"cb184b8b7a1c9f825f5c48f4b882e586772ff863671c32a461519f9e717aab6e"} Oct 05 07:10:09 crc kubenswrapper[4846]: I1005 07:10:09.478951 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"a0844571-8880-4fa4-af42-2da48aa8bdec","Type":"ContainerDied","Data":"b5192909be2c780d9316ded24f975196712cb7511b0a99cabaa714a897f59917"} Oct 05 07:10:09 crc kubenswrapper[4846]: I1005 07:10:09.478982 4846 scope.go:117] "RemoveContainer" containerID="cb184b8b7a1c9f825f5c48f4b882e586772ff863671c32a461519f9e717aab6e" Oct 05 07:10:09 crc kubenswrapper[4846]: I1005 07:10:09.481297 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 05 07:10:09 crc kubenswrapper[4846]: I1005 07:10:09.488808 4846 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a0844571-8880-4fa4-af42-2da48aa8bdec-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 07:10:09 crc kubenswrapper[4846]: I1005 07:10:09.488834 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5spwx\" (UniqueName: \"kubernetes.io/projected/a0844571-8880-4fa4-af42-2da48aa8bdec-kube-api-access-5spwx\") on node \"crc\" DevicePath \"\"" Oct 05 07:10:09 crc kubenswrapper[4846]: I1005 07:10:09.488845 4846 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a0844571-8880-4fa4-af42-2da48aa8bdec-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 07:10:09 crc kubenswrapper[4846]: I1005 07:10:09.512592 4846 scope.go:117] "RemoveContainer" containerID="cb184b8b7a1c9f825f5c48f4b882e586772ff863671c32a461519f9e717aab6e" Oct 05 07:10:09 crc kubenswrapper[4846]: E1005 07:10:09.517385 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cb184b8b7a1c9f825f5c48f4b882e586772ff863671c32a461519f9e717aab6e\": container with ID starting with cb184b8b7a1c9f825f5c48f4b882e586772ff863671c32a461519f9e717aab6e not found: ID does not exist" containerID="cb184b8b7a1c9f825f5c48f4b882e586772ff863671c32a461519f9e717aab6e" Oct 05 07:10:09 crc kubenswrapper[4846]: I1005 07:10:09.517468 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cb184b8b7a1c9f825f5c48f4b882e586772ff863671c32a461519f9e717aab6e"} err="failed to get container status \"cb184b8b7a1c9f825f5c48f4b882e586772ff863671c32a461519f9e717aab6e\": rpc error: code = NotFound desc = could not find container \"cb184b8b7a1c9f825f5c48f4b882e586772ff863671c32a461519f9e717aab6e\": container with ID starting with cb184b8b7a1c9f825f5c48f4b882e586772ff863671c32a461519f9e717aab6e not found: ID does not exist" Oct 05 07:10:09 crc kubenswrapper[4846]: I1005 07:10:09.531409 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 05 07:10:09 crc kubenswrapper[4846]: I1005 07:10:09.560295 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 05 07:10:09 crc kubenswrapper[4846]: I1005 07:10:09.575889 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 05 07:10:09 crc kubenswrapper[4846]: E1005 07:10:09.576687 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a0844571-8880-4fa4-af42-2da48aa8bdec" containerName="nova-cell1-novncproxy-novncproxy" Oct 05 07:10:09 crc kubenswrapper[4846]: I1005 07:10:09.576717 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="a0844571-8880-4fa4-af42-2da48aa8bdec" containerName="nova-cell1-novncproxy-novncproxy" Oct 05 07:10:09 crc kubenswrapper[4846]: I1005 07:10:09.577007 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="a0844571-8880-4fa4-af42-2da48aa8bdec" containerName="nova-cell1-novncproxy-novncproxy" Oct 05 07:10:09 crc kubenswrapper[4846]: I1005 07:10:09.578045 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 05 07:10:09 crc kubenswrapper[4846]: I1005 07:10:09.583696 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Oct 05 07:10:09 crc kubenswrapper[4846]: I1005 07:10:09.583739 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-public-svc" Oct 05 07:10:09 crc kubenswrapper[4846]: I1005 07:10:09.585490 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-vencrypt" Oct 05 07:10:09 crc kubenswrapper[4846]: I1005 07:10:09.586019 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 05 07:10:09 crc kubenswrapper[4846]: I1005 07:10:09.697841 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/6660b9ff-df9c-478e-b852-82cb5fb2c60c-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"6660b9ff-df9c-478e-b852-82cb5fb2c60c\") " pod="openstack/nova-cell1-novncproxy-0" Oct 05 07:10:09 crc kubenswrapper[4846]: I1005 07:10:09.698436 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6660b9ff-df9c-478e-b852-82cb5fb2c60c-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"6660b9ff-df9c-478e-b852-82cb5fb2c60c\") " pod="openstack/nova-cell1-novncproxy-0" Oct 05 07:10:09 crc kubenswrapper[4846]: I1005 07:10:09.698602 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/6660b9ff-df9c-478e-b852-82cb5fb2c60c-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"6660b9ff-df9c-478e-b852-82cb5fb2c60c\") " pod="openstack/nova-cell1-novncproxy-0" Oct 05 07:10:09 crc kubenswrapper[4846]: I1005 07:10:09.698779 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6660b9ff-df9c-478e-b852-82cb5fb2c60c-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"6660b9ff-df9c-478e-b852-82cb5fb2c60c\") " pod="openstack/nova-cell1-novncproxy-0" Oct 05 07:10:09 crc kubenswrapper[4846]: I1005 07:10:09.699254 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jwnfj\" (UniqueName: \"kubernetes.io/projected/6660b9ff-df9c-478e-b852-82cb5fb2c60c-kube-api-access-jwnfj\") pod \"nova-cell1-novncproxy-0\" (UID: \"6660b9ff-df9c-478e-b852-82cb5fb2c60c\") " pod="openstack/nova-cell1-novncproxy-0" Oct 05 07:10:09 crc kubenswrapper[4846]: I1005 07:10:09.705138 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 05 07:10:09 crc kubenswrapper[4846]: I1005 07:10:09.707265 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 05 07:10:09 crc kubenswrapper[4846]: I1005 07:10:09.710968 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 05 07:10:09 crc kubenswrapper[4846]: I1005 07:10:09.802285 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6660b9ff-df9c-478e-b852-82cb5fb2c60c-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"6660b9ff-df9c-478e-b852-82cb5fb2c60c\") " pod="openstack/nova-cell1-novncproxy-0" Oct 05 07:10:09 crc kubenswrapper[4846]: I1005 07:10:09.802378 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/6660b9ff-df9c-478e-b852-82cb5fb2c60c-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"6660b9ff-df9c-478e-b852-82cb5fb2c60c\") " pod="openstack/nova-cell1-novncproxy-0" Oct 05 07:10:09 crc kubenswrapper[4846]: I1005 07:10:09.802472 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6660b9ff-df9c-478e-b852-82cb5fb2c60c-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"6660b9ff-df9c-478e-b852-82cb5fb2c60c\") " pod="openstack/nova-cell1-novncproxy-0" Oct 05 07:10:09 crc kubenswrapper[4846]: I1005 07:10:09.802581 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jwnfj\" (UniqueName: \"kubernetes.io/projected/6660b9ff-df9c-478e-b852-82cb5fb2c60c-kube-api-access-jwnfj\") pod \"nova-cell1-novncproxy-0\" (UID: \"6660b9ff-df9c-478e-b852-82cb5fb2c60c\") " pod="openstack/nova-cell1-novncproxy-0" Oct 05 07:10:09 crc kubenswrapper[4846]: I1005 07:10:09.802735 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/6660b9ff-df9c-478e-b852-82cb5fb2c60c-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"6660b9ff-df9c-478e-b852-82cb5fb2c60c\") " pod="openstack/nova-cell1-novncproxy-0" Oct 05 07:10:09 crc kubenswrapper[4846]: I1005 07:10:09.808050 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/6660b9ff-df9c-478e-b852-82cb5fb2c60c-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"6660b9ff-df9c-478e-b852-82cb5fb2c60c\") " pod="openstack/nova-cell1-novncproxy-0" Oct 05 07:10:09 crc kubenswrapper[4846]: I1005 07:10:09.809035 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/6660b9ff-df9c-478e-b852-82cb5fb2c60c-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"6660b9ff-df9c-478e-b852-82cb5fb2c60c\") " pod="openstack/nova-cell1-novncproxy-0" Oct 05 07:10:09 crc kubenswrapper[4846]: I1005 07:10:09.809812 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6660b9ff-df9c-478e-b852-82cb5fb2c60c-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"6660b9ff-df9c-478e-b852-82cb5fb2c60c\") " pod="openstack/nova-cell1-novncproxy-0" Oct 05 07:10:09 crc kubenswrapper[4846]: I1005 07:10:09.812856 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6660b9ff-df9c-478e-b852-82cb5fb2c60c-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"6660b9ff-df9c-478e-b852-82cb5fb2c60c\") " pod="openstack/nova-cell1-novncproxy-0" Oct 05 07:10:09 crc kubenswrapper[4846]: I1005 07:10:09.837591 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jwnfj\" (UniqueName: \"kubernetes.io/projected/6660b9ff-df9c-478e-b852-82cb5fb2c60c-kube-api-access-jwnfj\") pod \"nova-cell1-novncproxy-0\" (UID: \"6660b9ff-df9c-478e-b852-82cb5fb2c60c\") " pod="openstack/nova-cell1-novncproxy-0" Oct 05 07:10:09 crc kubenswrapper[4846]: I1005 07:10:09.905921 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 05 07:10:10 crc kubenswrapper[4846]: I1005 07:10:10.211350 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 05 07:10:10 crc kubenswrapper[4846]: W1005 07:10:10.214691 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6660b9ff_df9c_478e_b852_82cb5fb2c60c.slice/crio-975d98535a2e45846601a9e1e1224188cc0dd1e3bbd28cd7323573e631782916 WatchSource:0}: Error finding container 975d98535a2e45846601a9e1e1224188cc0dd1e3bbd28cd7323573e631782916: Status 404 returned error can't find the container with id 975d98535a2e45846601a9e1e1224188cc0dd1e3bbd28cd7323573e631782916 Oct 05 07:10:10 crc kubenswrapper[4846]: I1005 07:10:10.532652 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0844571-8880-4fa4-af42-2da48aa8bdec" path="/var/lib/kubelet/pods/a0844571-8880-4fa4-af42-2da48aa8bdec/volumes" Oct 05 07:10:10 crc kubenswrapper[4846]: I1005 07:10:10.534379 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"6660b9ff-df9c-478e-b852-82cb5fb2c60c","Type":"ContainerStarted","Data":"975d98535a2e45846601a9e1e1224188cc0dd1e3bbd28cd7323573e631782916"} Oct 05 07:10:10 crc kubenswrapper[4846]: I1005 07:10:10.534477 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 05 07:10:11 crc kubenswrapper[4846]: I1005 07:10:11.521530 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"6660b9ff-df9c-478e-b852-82cb5fb2c60c","Type":"ContainerStarted","Data":"2a96112701d2d7042bce02efb3a776df18e837a8d8981829c635370de3016348"} Oct 05 07:10:11 crc kubenswrapper[4846]: I1005 07:10:11.544306 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.544288852 podStartE2EDuration="2.544288852s" podCreationTimestamp="2025-10-05 07:10:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 07:10:11.539851884 +0000 UTC m=+1333.780704659" watchObservedRunningTime="2025-10-05 07:10:11.544288852 +0000 UTC m=+1333.785141627" Oct 05 07:10:11 crc kubenswrapper[4846]: I1005 07:10:11.804403 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 05 07:10:11 crc kubenswrapper[4846]: I1005 07:10:11.805005 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 05 07:10:11 crc kubenswrapper[4846]: I1005 07:10:11.805545 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 05 07:10:11 crc kubenswrapper[4846]: I1005 07:10:11.823878 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 05 07:10:12 crc kubenswrapper[4846]: I1005 07:10:12.535135 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 05 07:10:12 crc kubenswrapper[4846]: I1005 07:10:12.541313 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 05 07:10:12 crc kubenswrapper[4846]: I1005 07:10:12.809838 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-c7d6584f5-dxcgg"] Oct 05 07:10:12 crc kubenswrapper[4846]: I1005 07:10:12.813647 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-c7d6584f5-dxcgg" Oct 05 07:10:12 crc kubenswrapper[4846]: I1005 07:10:12.839135 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-c7d6584f5-dxcgg"] Oct 05 07:10:12 crc kubenswrapper[4846]: I1005 07:10:12.987725 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7a5c4c82-4b3f-4314-a32d-db5398ca8a9e-ovsdbserver-nb\") pod \"dnsmasq-dns-c7d6584f5-dxcgg\" (UID: \"7a5c4c82-4b3f-4314-a32d-db5398ca8a9e\") " pod="openstack/dnsmasq-dns-c7d6584f5-dxcgg" Oct 05 07:10:12 crc kubenswrapper[4846]: I1005 07:10:12.988006 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7a5c4c82-4b3f-4314-a32d-db5398ca8a9e-ovsdbserver-sb\") pod \"dnsmasq-dns-c7d6584f5-dxcgg\" (UID: \"7a5c4c82-4b3f-4314-a32d-db5398ca8a9e\") " pod="openstack/dnsmasq-dns-c7d6584f5-dxcgg" Oct 05 07:10:12 crc kubenswrapper[4846]: I1005 07:10:12.988125 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7a5c4c82-4b3f-4314-a32d-db5398ca8a9e-config\") pod \"dnsmasq-dns-c7d6584f5-dxcgg\" (UID: \"7a5c4c82-4b3f-4314-a32d-db5398ca8a9e\") " pod="openstack/dnsmasq-dns-c7d6584f5-dxcgg" Oct 05 07:10:12 crc kubenswrapper[4846]: I1005 07:10:12.988241 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7a5c4c82-4b3f-4314-a32d-db5398ca8a9e-dns-svc\") pod \"dnsmasq-dns-c7d6584f5-dxcgg\" (UID: \"7a5c4c82-4b3f-4314-a32d-db5398ca8a9e\") " pod="openstack/dnsmasq-dns-c7d6584f5-dxcgg" Oct 05 07:10:12 crc kubenswrapper[4846]: I1005 07:10:12.988367 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8wmgl\" (UniqueName: \"kubernetes.io/projected/7a5c4c82-4b3f-4314-a32d-db5398ca8a9e-kube-api-access-8wmgl\") pod \"dnsmasq-dns-c7d6584f5-dxcgg\" (UID: \"7a5c4c82-4b3f-4314-a32d-db5398ca8a9e\") " pod="openstack/dnsmasq-dns-c7d6584f5-dxcgg" Oct 05 07:10:12 crc kubenswrapper[4846]: I1005 07:10:12.988478 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7a5c4c82-4b3f-4314-a32d-db5398ca8a9e-dns-swift-storage-0\") pod \"dnsmasq-dns-c7d6584f5-dxcgg\" (UID: \"7a5c4c82-4b3f-4314-a32d-db5398ca8a9e\") " pod="openstack/dnsmasq-dns-c7d6584f5-dxcgg" Oct 05 07:10:13 crc kubenswrapper[4846]: I1005 07:10:13.101799 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7a5c4c82-4b3f-4314-a32d-db5398ca8a9e-ovsdbserver-sb\") pod \"dnsmasq-dns-c7d6584f5-dxcgg\" (UID: \"7a5c4c82-4b3f-4314-a32d-db5398ca8a9e\") " pod="openstack/dnsmasq-dns-c7d6584f5-dxcgg" Oct 05 07:10:13 crc kubenswrapper[4846]: I1005 07:10:13.101941 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7a5c4c82-4b3f-4314-a32d-db5398ca8a9e-config\") pod \"dnsmasq-dns-c7d6584f5-dxcgg\" (UID: \"7a5c4c82-4b3f-4314-a32d-db5398ca8a9e\") " pod="openstack/dnsmasq-dns-c7d6584f5-dxcgg" Oct 05 07:10:13 crc kubenswrapper[4846]: I1005 07:10:13.101968 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7a5c4c82-4b3f-4314-a32d-db5398ca8a9e-dns-svc\") pod \"dnsmasq-dns-c7d6584f5-dxcgg\" (UID: \"7a5c4c82-4b3f-4314-a32d-db5398ca8a9e\") " pod="openstack/dnsmasq-dns-c7d6584f5-dxcgg" Oct 05 07:10:13 crc kubenswrapper[4846]: I1005 07:10:13.102074 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8wmgl\" (UniqueName: \"kubernetes.io/projected/7a5c4c82-4b3f-4314-a32d-db5398ca8a9e-kube-api-access-8wmgl\") pod \"dnsmasq-dns-c7d6584f5-dxcgg\" (UID: \"7a5c4c82-4b3f-4314-a32d-db5398ca8a9e\") " pod="openstack/dnsmasq-dns-c7d6584f5-dxcgg" Oct 05 07:10:13 crc kubenswrapper[4846]: I1005 07:10:13.102154 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7a5c4c82-4b3f-4314-a32d-db5398ca8a9e-dns-swift-storage-0\") pod \"dnsmasq-dns-c7d6584f5-dxcgg\" (UID: \"7a5c4c82-4b3f-4314-a32d-db5398ca8a9e\") " pod="openstack/dnsmasq-dns-c7d6584f5-dxcgg" Oct 05 07:10:13 crc kubenswrapper[4846]: I1005 07:10:13.102332 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7a5c4c82-4b3f-4314-a32d-db5398ca8a9e-ovsdbserver-nb\") pod \"dnsmasq-dns-c7d6584f5-dxcgg\" (UID: \"7a5c4c82-4b3f-4314-a32d-db5398ca8a9e\") " pod="openstack/dnsmasq-dns-c7d6584f5-dxcgg" Oct 05 07:10:13 crc kubenswrapper[4846]: I1005 07:10:13.103497 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7a5c4c82-4b3f-4314-a32d-db5398ca8a9e-config\") pod \"dnsmasq-dns-c7d6584f5-dxcgg\" (UID: \"7a5c4c82-4b3f-4314-a32d-db5398ca8a9e\") " pod="openstack/dnsmasq-dns-c7d6584f5-dxcgg" Oct 05 07:10:13 crc kubenswrapper[4846]: I1005 07:10:13.103595 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7a5c4c82-4b3f-4314-a32d-db5398ca8a9e-ovsdbserver-sb\") pod \"dnsmasq-dns-c7d6584f5-dxcgg\" (UID: \"7a5c4c82-4b3f-4314-a32d-db5398ca8a9e\") " pod="openstack/dnsmasq-dns-c7d6584f5-dxcgg" Oct 05 07:10:13 crc kubenswrapper[4846]: I1005 07:10:13.106085 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7a5c4c82-4b3f-4314-a32d-db5398ca8a9e-ovsdbserver-nb\") pod \"dnsmasq-dns-c7d6584f5-dxcgg\" (UID: \"7a5c4c82-4b3f-4314-a32d-db5398ca8a9e\") " pod="openstack/dnsmasq-dns-c7d6584f5-dxcgg" Oct 05 07:10:13 crc kubenswrapper[4846]: I1005 07:10:13.106153 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7a5c4c82-4b3f-4314-a32d-db5398ca8a9e-dns-swift-storage-0\") pod \"dnsmasq-dns-c7d6584f5-dxcgg\" (UID: \"7a5c4c82-4b3f-4314-a32d-db5398ca8a9e\") " pod="openstack/dnsmasq-dns-c7d6584f5-dxcgg" Oct 05 07:10:13 crc kubenswrapper[4846]: I1005 07:10:13.106415 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7a5c4c82-4b3f-4314-a32d-db5398ca8a9e-dns-svc\") pod \"dnsmasq-dns-c7d6584f5-dxcgg\" (UID: \"7a5c4c82-4b3f-4314-a32d-db5398ca8a9e\") " pod="openstack/dnsmasq-dns-c7d6584f5-dxcgg" Oct 05 07:10:13 crc kubenswrapper[4846]: I1005 07:10:13.131101 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8wmgl\" (UniqueName: \"kubernetes.io/projected/7a5c4c82-4b3f-4314-a32d-db5398ca8a9e-kube-api-access-8wmgl\") pod \"dnsmasq-dns-c7d6584f5-dxcgg\" (UID: \"7a5c4c82-4b3f-4314-a32d-db5398ca8a9e\") " pod="openstack/dnsmasq-dns-c7d6584f5-dxcgg" Oct 05 07:10:13 crc kubenswrapper[4846]: I1005 07:10:13.145718 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-c7d6584f5-dxcgg" Oct 05 07:10:13 crc kubenswrapper[4846]: I1005 07:10:13.711601 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-c7d6584f5-dxcgg"] Oct 05 07:10:14 crc kubenswrapper[4846]: I1005 07:10:14.589118 4846 generic.go:334] "Generic (PLEG): container finished" podID="7a5c4c82-4b3f-4314-a32d-db5398ca8a9e" containerID="c51d875ba49113376b3d8a083c86581b5406413bc84095d5348f99c0a3959cc5" exitCode=0 Oct 05 07:10:14 crc kubenswrapper[4846]: I1005 07:10:14.590064 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-c7d6584f5-dxcgg" event={"ID":"7a5c4c82-4b3f-4314-a32d-db5398ca8a9e","Type":"ContainerDied","Data":"c51d875ba49113376b3d8a083c86581b5406413bc84095d5348f99c0a3959cc5"} Oct 05 07:10:14 crc kubenswrapper[4846]: I1005 07:10:14.590098 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-c7d6584f5-dxcgg" event={"ID":"7a5c4c82-4b3f-4314-a32d-db5398ca8a9e","Type":"ContainerStarted","Data":"ec97db54b2342d26de8e49361527359256837278cf11d3cc5355609f7e73865b"} Oct 05 07:10:14 crc kubenswrapper[4846]: I1005 07:10:14.919774 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Oct 05 07:10:15 crc kubenswrapper[4846]: I1005 07:10:15.083858 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 05 07:10:15 crc kubenswrapper[4846]: I1005 07:10:15.084203 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="eb03eca3-f45d-4361-ab00-9ce16bb2c96e" containerName="ceilometer-central-agent" containerID="cri-o://74b48fcfb6476893634deecb8bf996c2f28f54df151237a948e6426854564819" gracePeriod=30 Oct 05 07:10:15 crc kubenswrapper[4846]: I1005 07:10:15.084285 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="eb03eca3-f45d-4361-ab00-9ce16bb2c96e" containerName="sg-core" containerID="cri-o://6f1f221eac0cd0e08e90b60a2002245bc6a6ed3b4c9a396d48f47cca0039feb9" gracePeriod=30 Oct 05 07:10:15 crc kubenswrapper[4846]: I1005 07:10:15.084300 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="eb03eca3-f45d-4361-ab00-9ce16bb2c96e" containerName="ceilometer-notification-agent" containerID="cri-o://fafa6b60a9e821e9bc3c916554dcc88b5f10181956baed8a807149ed1db4333e" gracePeriod=30 Oct 05 07:10:15 crc kubenswrapper[4846]: I1005 07:10:15.084305 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="eb03eca3-f45d-4361-ab00-9ce16bb2c96e" containerName="proxy-httpd" containerID="cri-o://32020fcfa11e9aff11fbbf804e18681fcebe28c354eef5dfd4c7dc9be93e5965" gracePeriod=30 Oct 05 07:10:15 crc kubenswrapper[4846]: I1005 07:10:15.091466 4846 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="eb03eca3-f45d-4361-ab00-9ce16bb2c96e" containerName="proxy-httpd" probeResult="failure" output="Get \"https://10.217.0.195:3000/\": EOF" Oct 05 07:10:15 crc kubenswrapper[4846]: I1005 07:10:15.581673 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 05 07:10:15 crc kubenswrapper[4846]: I1005 07:10:15.604480 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-c7d6584f5-dxcgg" event={"ID":"7a5c4c82-4b3f-4314-a32d-db5398ca8a9e","Type":"ContainerStarted","Data":"283ce773f8b442a1a87a4848af02010620aa0074924483f0a608d4c3f15d5131"} Oct 05 07:10:15 crc kubenswrapper[4846]: I1005 07:10:15.606018 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-c7d6584f5-dxcgg" Oct 05 07:10:15 crc kubenswrapper[4846]: I1005 07:10:15.611524 4846 generic.go:334] "Generic (PLEG): container finished" podID="eb03eca3-f45d-4361-ab00-9ce16bb2c96e" containerID="32020fcfa11e9aff11fbbf804e18681fcebe28c354eef5dfd4c7dc9be93e5965" exitCode=0 Oct 05 07:10:15 crc kubenswrapper[4846]: I1005 07:10:15.611558 4846 generic.go:334] "Generic (PLEG): container finished" podID="eb03eca3-f45d-4361-ab00-9ce16bb2c96e" containerID="6f1f221eac0cd0e08e90b60a2002245bc6a6ed3b4c9a396d48f47cca0039feb9" exitCode=2 Oct 05 07:10:15 crc kubenswrapper[4846]: I1005 07:10:15.611569 4846 generic.go:334] "Generic (PLEG): container finished" podID="eb03eca3-f45d-4361-ab00-9ce16bb2c96e" containerID="74b48fcfb6476893634deecb8bf996c2f28f54df151237a948e6426854564819" exitCode=0 Oct 05 07:10:15 crc kubenswrapper[4846]: I1005 07:10:15.611746 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="f1e8f9a4-b671-4417-99c9-90bda3c17339" containerName="nova-api-log" containerID="cri-o://7ec1da9974ee7548066528f3fd8df37c374953980dee2632e3db26ed5b6f0b8f" gracePeriod=30 Oct 05 07:10:15 crc kubenswrapper[4846]: I1005 07:10:15.611827 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"eb03eca3-f45d-4361-ab00-9ce16bb2c96e","Type":"ContainerDied","Data":"32020fcfa11e9aff11fbbf804e18681fcebe28c354eef5dfd4c7dc9be93e5965"} Oct 05 07:10:15 crc kubenswrapper[4846]: I1005 07:10:15.611853 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"eb03eca3-f45d-4361-ab00-9ce16bb2c96e","Type":"ContainerDied","Data":"6f1f221eac0cd0e08e90b60a2002245bc6a6ed3b4c9a396d48f47cca0039feb9"} Oct 05 07:10:15 crc kubenswrapper[4846]: I1005 07:10:15.611869 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"eb03eca3-f45d-4361-ab00-9ce16bb2c96e","Type":"ContainerDied","Data":"74b48fcfb6476893634deecb8bf996c2f28f54df151237a948e6426854564819"} Oct 05 07:10:15 crc kubenswrapper[4846]: I1005 07:10:15.611968 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="f1e8f9a4-b671-4417-99c9-90bda3c17339" containerName="nova-api-api" containerID="cri-o://53602a509bd81a4226a9664cce98e43c4223e89f4b02f63f45ef63bdac84c69e" gracePeriod=30 Oct 05 07:10:15 crc kubenswrapper[4846]: I1005 07:10:15.635772 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-c7d6584f5-dxcgg" podStartSLOduration=3.635748744 podStartE2EDuration="3.635748744s" podCreationTimestamp="2025-10-05 07:10:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 07:10:15.622637504 +0000 UTC m=+1337.863490279" watchObservedRunningTime="2025-10-05 07:10:15.635748744 +0000 UTC m=+1337.876601519" Oct 05 07:10:15 crc kubenswrapper[4846]: I1005 07:10:15.969383 4846 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="eb03eca3-f45d-4361-ab00-9ce16bb2c96e" containerName="proxy-httpd" probeResult="failure" output="Get \"https://10.217.0.195:3000/\": dial tcp 10.217.0.195:3000: connect: connection refused" Oct 05 07:10:16 crc kubenswrapper[4846]: I1005 07:10:16.624805 4846 generic.go:334] "Generic (PLEG): container finished" podID="f1e8f9a4-b671-4417-99c9-90bda3c17339" containerID="7ec1da9974ee7548066528f3fd8df37c374953980dee2632e3db26ed5b6f0b8f" exitCode=143 Oct 05 07:10:16 crc kubenswrapper[4846]: I1005 07:10:16.624900 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"f1e8f9a4-b671-4417-99c9-90bda3c17339","Type":"ContainerDied","Data":"7ec1da9974ee7548066528f3fd8df37c374953980dee2632e3db26ed5b6f0b8f"} Oct 05 07:10:19 crc kubenswrapper[4846]: I1005 07:10:19.259625 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 05 07:10:19 crc kubenswrapper[4846]: I1005 07:10:19.366358 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f1e8f9a4-b671-4417-99c9-90bda3c17339-config-data\") pod \"f1e8f9a4-b671-4417-99c9-90bda3c17339\" (UID: \"f1e8f9a4-b671-4417-99c9-90bda3c17339\") " Oct 05 07:10:19 crc kubenswrapper[4846]: I1005 07:10:19.366428 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gc265\" (UniqueName: \"kubernetes.io/projected/f1e8f9a4-b671-4417-99c9-90bda3c17339-kube-api-access-gc265\") pod \"f1e8f9a4-b671-4417-99c9-90bda3c17339\" (UID: \"f1e8f9a4-b671-4417-99c9-90bda3c17339\") " Oct 05 07:10:19 crc kubenswrapper[4846]: I1005 07:10:19.366484 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f1e8f9a4-b671-4417-99c9-90bda3c17339-combined-ca-bundle\") pod \"f1e8f9a4-b671-4417-99c9-90bda3c17339\" (UID: \"f1e8f9a4-b671-4417-99c9-90bda3c17339\") " Oct 05 07:10:19 crc kubenswrapper[4846]: I1005 07:10:19.366551 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f1e8f9a4-b671-4417-99c9-90bda3c17339-logs\") pod \"f1e8f9a4-b671-4417-99c9-90bda3c17339\" (UID: \"f1e8f9a4-b671-4417-99c9-90bda3c17339\") " Oct 05 07:10:19 crc kubenswrapper[4846]: I1005 07:10:19.368300 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f1e8f9a4-b671-4417-99c9-90bda3c17339-logs" (OuterVolumeSpecName: "logs") pod "f1e8f9a4-b671-4417-99c9-90bda3c17339" (UID: "f1e8f9a4-b671-4417-99c9-90bda3c17339"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 07:10:19 crc kubenswrapper[4846]: I1005 07:10:19.378981 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f1e8f9a4-b671-4417-99c9-90bda3c17339-kube-api-access-gc265" (OuterVolumeSpecName: "kube-api-access-gc265") pod "f1e8f9a4-b671-4417-99c9-90bda3c17339" (UID: "f1e8f9a4-b671-4417-99c9-90bda3c17339"). InnerVolumeSpecName "kube-api-access-gc265". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:10:19 crc kubenswrapper[4846]: I1005 07:10:19.400833 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f1e8f9a4-b671-4417-99c9-90bda3c17339-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f1e8f9a4-b671-4417-99c9-90bda3c17339" (UID: "f1e8f9a4-b671-4417-99c9-90bda3c17339"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:10:19 crc kubenswrapper[4846]: I1005 07:10:19.412372 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f1e8f9a4-b671-4417-99c9-90bda3c17339-config-data" (OuterVolumeSpecName: "config-data") pod "f1e8f9a4-b671-4417-99c9-90bda3c17339" (UID: "f1e8f9a4-b671-4417-99c9-90bda3c17339"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:10:19 crc kubenswrapper[4846]: I1005 07:10:19.469417 4846 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f1e8f9a4-b671-4417-99c9-90bda3c17339-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 07:10:19 crc kubenswrapper[4846]: I1005 07:10:19.469544 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gc265\" (UniqueName: \"kubernetes.io/projected/f1e8f9a4-b671-4417-99c9-90bda3c17339-kube-api-access-gc265\") on node \"crc\" DevicePath \"\"" Oct 05 07:10:19 crc kubenswrapper[4846]: I1005 07:10:19.469557 4846 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f1e8f9a4-b671-4417-99c9-90bda3c17339-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 07:10:19 crc kubenswrapper[4846]: I1005 07:10:19.469568 4846 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f1e8f9a4-b671-4417-99c9-90bda3c17339-logs\") on node \"crc\" DevicePath \"\"" Oct 05 07:10:19 crc kubenswrapper[4846]: I1005 07:10:19.659785 4846 generic.go:334] "Generic (PLEG): container finished" podID="f1e8f9a4-b671-4417-99c9-90bda3c17339" containerID="53602a509bd81a4226a9664cce98e43c4223e89f4b02f63f45ef63bdac84c69e" exitCode=0 Oct 05 07:10:19 crc kubenswrapper[4846]: I1005 07:10:19.660147 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"f1e8f9a4-b671-4417-99c9-90bda3c17339","Type":"ContainerDied","Data":"53602a509bd81a4226a9664cce98e43c4223e89f4b02f63f45ef63bdac84c69e"} Oct 05 07:10:19 crc kubenswrapper[4846]: I1005 07:10:19.660292 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"f1e8f9a4-b671-4417-99c9-90bda3c17339","Type":"ContainerDied","Data":"40a7a409bbe0ba6f0b9e9714102ae06da4cc85a991723f874c127b4180ebe5cf"} Oct 05 07:10:19 crc kubenswrapper[4846]: I1005 07:10:19.660326 4846 scope.go:117] "RemoveContainer" containerID="53602a509bd81a4226a9664cce98e43c4223e89f4b02f63f45ef63bdac84c69e" Oct 05 07:10:19 crc kubenswrapper[4846]: I1005 07:10:19.660512 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 05 07:10:19 crc kubenswrapper[4846]: I1005 07:10:19.706260 4846 scope.go:117] "RemoveContainer" containerID="7ec1da9974ee7548066528f3fd8df37c374953980dee2632e3db26ed5b6f0b8f" Oct 05 07:10:19 crc kubenswrapper[4846]: I1005 07:10:19.731841 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 05 07:10:19 crc kubenswrapper[4846]: I1005 07:10:19.741426 4846 scope.go:117] "RemoveContainer" containerID="53602a509bd81a4226a9664cce98e43c4223e89f4b02f63f45ef63bdac84c69e" Oct 05 07:10:19 crc kubenswrapper[4846]: I1005 07:10:19.748434 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Oct 05 07:10:19 crc kubenswrapper[4846]: E1005 07:10:19.749228 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"53602a509bd81a4226a9664cce98e43c4223e89f4b02f63f45ef63bdac84c69e\": container with ID starting with 53602a509bd81a4226a9664cce98e43c4223e89f4b02f63f45ef63bdac84c69e not found: ID does not exist" containerID="53602a509bd81a4226a9664cce98e43c4223e89f4b02f63f45ef63bdac84c69e" Oct 05 07:10:19 crc kubenswrapper[4846]: I1005 07:10:19.749290 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"53602a509bd81a4226a9664cce98e43c4223e89f4b02f63f45ef63bdac84c69e"} err="failed to get container status \"53602a509bd81a4226a9664cce98e43c4223e89f4b02f63f45ef63bdac84c69e\": rpc error: code = NotFound desc = could not find container \"53602a509bd81a4226a9664cce98e43c4223e89f4b02f63f45ef63bdac84c69e\": container with ID starting with 53602a509bd81a4226a9664cce98e43c4223e89f4b02f63f45ef63bdac84c69e not found: ID does not exist" Oct 05 07:10:19 crc kubenswrapper[4846]: I1005 07:10:19.749331 4846 scope.go:117] "RemoveContainer" containerID="7ec1da9974ee7548066528f3fd8df37c374953980dee2632e3db26ed5b6f0b8f" Oct 05 07:10:19 crc kubenswrapper[4846]: E1005 07:10:19.749834 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7ec1da9974ee7548066528f3fd8df37c374953980dee2632e3db26ed5b6f0b8f\": container with ID starting with 7ec1da9974ee7548066528f3fd8df37c374953980dee2632e3db26ed5b6f0b8f not found: ID does not exist" containerID="7ec1da9974ee7548066528f3fd8df37c374953980dee2632e3db26ed5b6f0b8f" Oct 05 07:10:19 crc kubenswrapper[4846]: I1005 07:10:19.749886 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7ec1da9974ee7548066528f3fd8df37c374953980dee2632e3db26ed5b6f0b8f"} err="failed to get container status \"7ec1da9974ee7548066528f3fd8df37c374953980dee2632e3db26ed5b6f0b8f\": rpc error: code = NotFound desc = could not find container \"7ec1da9974ee7548066528f3fd8df37c374953980dee2632e3db26ed5b6f0b8f\": container with ID starting with 7ec1da9974ee7548066528f3fd8df37c374953980dee2632e3db26ed5b6f0b8f not found: ID does not exist" Oct 05 07:10:19 crc kubenswrapper[4846]: I1005 07:10:19.760764 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 05 07:10:19 crc kubenswrapper[4846]: E1005 07:10:19.763992 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f1e8f9a4-b671-4417-99c9-90bda3c17339" containerName="nova-api-log" Oct 05 07:10:19 crc kubenswrapper[4846]: I1005 07:10:19.764022 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="f1e8f9a4-b671-4417-99c9-90bda3c17339" containerName="nova-api-log" Oct 05 07:10:19 crc kubenswrapper[4846]: E1005 07:10:19.764055 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f1e8f9a4-b671-4417-99c9-90bda3c17339" containerName="nova-api-api" Oct 05 07:10:19 crc kubenswrapper[4846]: I1005 07:10:19.764064 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="f1e8f9a4-b671-4417-99c9-90bda3c17339" containerName="nova-api-api" Oct 05 07:10:19 crc kubenswrapper[4846]: I1005 07:10:19.764470 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="f1e8f9a4-b671-4417-99c9-90bda3c17339" containerName="nova-api-api" Oct 05 07:10:19 crc kubenswrapper[4846]: I1005 07:10:19.764582 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="f1e8f9a4-b671-4417-99c9-90bda3c17339" containerName="nova-api-log" Oct 05 07:10:19 crc kubenswrapper[4846]: I1005 07:10:19.766952 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 05 07:10:19 crc kubenswrapper[4846]: I1005 07:10:19.769891 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Oct 05 07:10:19 crc kubenswrapper[4846]: I1005 07:10:19.770243 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 05 07:10:19 crc kubenswrapper[4846]: I1005 07:10:19.770651 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Oct 05 07:10:19 crc kubenswrapper[4846]: I1005 07:10:19.772566 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 05 07:10:19 crc kubenswrapper[4846]: I1005 07:10:19.878806 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/65fd80a1-d4ce-42f9-b1cf-e379333dfdba-internal-tls-certs\") pod \"nova-api-0\" (UID: \"65fd80a1-d4ce-42f9-b1cf-e379333dfdba\") " pod="openstack/nova-api-0" Oct 05 07:10:19 crc kubenswrapper[4846]: I1005 07:10:19.878908 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/65fd80a1-d4ce-42f9-b1cf-e379333dfdba-logs\") pod \"nova-api-0\" (UID: \"65fd80a1-d4ce-42f9-b1cf-e379333dfdba\") " pod="openstack/nova-api-0" Oct 05 07:10:19 crc kubenswrapper[4846]: I1005 07:10:19.879011 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jv99j\" (UniqueName: \"kubernetes.io/projected/65fd80a1-d4ce-42f9-b1cf-e379333dfdba-kube-api-access-jv99j\") pod \"nova-api-0\" (UID: \"65fd80a1-d4ce-42f9-b1cf-e379333dfdba\") " pod="openstack/nova-api-0" Oct 05 07:10:19 crc kubenswrapper[4846]: I1005 07:10:19.879336 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/65fd80a1-d4ce-42f9-b1cf-e379333dfdba-config-data\") pod \"nova-api-0\" (UID: \"65fd80a1-d4ce-42f9-b1cf-e379333dfdba\") " pod="openstack/nova-api-0" Oct 05 07:10:19 crc kubenswrapper[4846]: I1005 07:10:19.879431 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/65fd80a1-d4ce-42f9-b1cf-e379333dfdba-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"65fd80a1-d4ce-42f9-b1cf-e379333dfdba\") " pod="openstack/nova-api-0" Oct 05 07:10:19 crc kubenswrapper[4846]: I1005 07:10:19.879711 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/65fd80a1-d4ce-42f9-b1cf-e379333dfdba-public-tls-certs\") pod \"nova-api-0\" (UID: \"65fd80a1-d4ce-42f9-b1cf-e379333dfdba\") " pod="openstack/nova-api-0" Oct 05 07:10:19 crc kubenswrapper[4846]: I1005 07:10:19.906363 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Oct 05 07:10:19 crc kubenswrapper[4846]: I1005 07:10:19.925985 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Oct 05 07:10:19 crc kubenswrapper[4846]: I1005 07:10:19.981489 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/65fd80a1-d4ce-42f9-b1cf-e379333dfdba-public-tls-certs\") pod \"nova-api-0\" (UID: \"65fd80a1-d4ce-42f9-b1cf-e379333dfdba\") " pod="openstack/nova-api-0" Oct 05 07:10:19 crc kubenswrapper[4846]: I1005 07:10:19.981577 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/65fd80a1-d4ce-42f9-b1cf-e379333dfdba-internal-tls-certs\") pod \"nova-api-0\" (UID: \"65fd80a1-d4ce-42f9-b1cf-e379333dfdba\") " pod="openstack/nova-api-0" Oct 05 07:10:19 crc kubenswrapper[4846]: I1005 07:10:19.981620 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/65fd80a1-d4ce-42f9-b1cf-e379333dfdba-logs\") pod \"nova-api-0\" (UID: \"65fd80a1-d4ce-42f9-b1cf-e379333dfdba\") " pod="openstack/nova-api-0" Oct 05 07:10:19 crc kubenswrapper[4846]: I1005 07:10:19.981710 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jv99j\" (UniqueName: \"kubernetes.io/projected/65fd80a1-d4ce-42f9-b1cf-e379333dfdba-kube-api-access-jv99j\") pod \"nova-api-0\" (UID: \"65fd80a1-d4ce-42f9-b1cf-e379333dfdba\") " pod="openstack/nova-api-0" Oct 05 07:10:19 crc kubenswrapper[4846]: I1005 07:10:19.981761 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/65fd80a1-d4ce-42f9-b1cf-e379333dfdba-config-data\") pod \"nova-api-0\" (UID: \"65fd80a1-d4ce-42f9-b1cf-e379333dfdba\") " pod="openstack/nova-api-0" Oct 05 07:10:19 crc kubenswrapper[4846]: I1005 07:10:19.981787 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/65fd80a1-d4ce-42f9-b1cf-e379333dfdba-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"65fd80a1-d4ce-42f9-b1cf-e379333dfdba\") " pod="openstack/nova-api-0" Oct 05 07:10:19 crc kubenswrapper[4846]: I1005 07:10:19.983019 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/65fd80a1-d4ce-42f9-b1cf-e379333dfdba-logs\") pod \"nova-api-0\" (UID: \"65fd80a1-d4ce-42f9-b1cf-e379333dfdba\") " pod="openstack/nova-api-0" Oct 05 07:10:19 crc kubenswrapper[4846]: I1005 07:10:19.986960 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/65fd80a1-d4ce-42f9-b1cf-e379333dfdba-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"65fd80a1-d4ce-42f9-b1cf-e379333dfdba\") " pod="openstack/nova-api-0" Oct 05 07:10:19 crc kubenswrapper[4846]: I1005 07:10:19.993811 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/65fd80a1-d4ce-42f9-b1cf-e379333dfdba-internal-tls-certs\") pod \"nova-api-0\" (UID: \"65fd80a1-d4ce-42f9-b1cf-e379333dfdba\") " pod="openstack/nova-api-0" Oct 05 07:10:19 crc kubenswrapper[4846]: I1005 07:10:19.994031 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/65fd80a1-d4ce-42f9-b1cf-e379333dfdba-config-data\") pod \"nova-api-0\" (UID: \"65fd80a1-d4ce-42f9-b1cf-e379333dfdba\") " pod="openstack/nova-api-0" Oct 05 07:10:19 crc kubenswrapper[4846]: I1005 07:10:19.994145 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/65fd80a1-d4ce-42f9-b1cf-e379333dfdba-public-tls-certs\") pod \"nova-api-0\" (UID: \"65fd80a1-d4ce-42f9-b1cf-e379333dfdba\") " pod="openstack/nova-api-0" Oct 05 07:10:20 crc kubenswrapper[4846]: I1005 07:10:20.004895 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jv99j\" (UniqueName: \"kubernetes.io/projected/65fd80a1-d4ce-42f9-b1cf-e379333dfdba-kube-api-access-jv99j\") pod \"nova-api-0\" (UID: \"65fd80a1-d4ce-42f9-b1cf-e379333dfdba\") " pod="openstack/nova-api-0" Oct 05 07:10:20 crc kubenswrapper[4846]: I1005 07:10:20.084477 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 05 07:10:20 crc kubenswrapper[4846]: I1005 07:10:20.398203 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 05 07:10:20 crc kubenswrapper[4846]: I1005 07:10:20.496319 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/eb03eca3-f45d-4361-ab00-9ce16bb2c96e-log-httpd\") pod \"eb03eca3-f45d-4361-ab00-9ce16bb2c96e\" (UID: \"eb03eca3-f45d-4361-ab00-9ce16bb2c96e\") " Oct 05 07:10:20 crc kubenswrapper[4846]: I1005 07:10:20.496393 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/eb03eca3-f45d-4361-ab00-9ce16bb2c96e-run-httpd\") pod \"eb03eca3-f45d-4361-ab00-9ce16bb2c96e\" (UID: \"eb03eca3-f45d-4361-ab00-9ce16bb2c96e\") " Oct 05 07:10:20 crc kubenswrapper[4846]: I1005 07:10:20.496467 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/eb03eca3-f45d-4361-ab00-9ce16bb2c96e-scripts\") pod \"eb03eca3-f45d-4361-ab00-9ce16bb2c96e\" (UID: \"eb03eca3-f45d-4361-ab00-9ce16bb2c96e\") " Oct 05 07:10:20 crc kubenswrapper[4846]: I1005 07:10:20.496499 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/eb03eca3-f45d-4361-ab00-9ce16bb2c96e-ceilometer-tls-certs\") pod \"eb03eca3-f45d-4361-ab00-9ce16bb2c96e\" (UID: \"eb03eca3-f45d-4361-ab00-9ce16bb2c96e\") " Oct 05 07:10:20 crc kubenswrapper[4846]: I1005 07:10:20.496622 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/eb03eca3-f45d-4361-ab00-9ce16bb2c96e-sg-core-conf-yaml\") pod \"eb03eca3-f45d-4361-ab00-9ce16bb2c96e\" (UID: \"eb03eca3-f45d-4361-ab00-9ce16bb2c96e\") " Oct 05 07:10:20 crc kubenswrapper[4846]: I1005 07:10:20.496651 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eb03eca3-f45d-4361-ab00-9ce16bb2c96e-combined-ca-bundle\") pod \"eb03eca3-f45d-4361-ab00-9ce16bb2c96e\" (UID: \"eb03eca3-f45d-4361-ab00-9ce16bb2c96e\") " Oct 05 07:10:20 crc kubenswrapper[4846]: I1005 07:10:20.496717 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eb03eca3-f45d-4361-ab00-9ce16bb2c96e-config-data\") pod \"eb03eca3-f45d-4361-ab00-9ce16bb2c96e\" (UID: \"eb03eca3-f45d-4361-ab00-9ce16bb2c96e\") " Oct 05 07:10:20 crc kubenswrapper[4846]: I1005 07:10:20.496864 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/eb03eca3-f45d-4361-ab00-9ce16bb2c96e-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "eb03eca3-f45d-4361-ab00-9ce16bb2c96e" (UID: "eb03eca3-f45d-4361-ab00-9ce16bb2c96e"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 07:10:20 crc kubenswrapper[4846]: I1005 07:10:20.496930 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9hpsb\" (UniqueName: \"kubernetes.io/projected/eb03eca3-f45d-4361-ab00-9ce16bb2c96e-kube-api-access-9hpsb\") pod \"eb03eca3-f45d-4361-ab00-9ce16bb2c96e\" (UID: \"eb03eca3-f45d-4361-ab00-9ce16bb2c96e\") " Oct 05 07:10:20 crc kubenswrapper[4846]: I1005 07:10:20.497545 4846 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/eb03eca3-f45d-4361-ab00-9ce16bb2c96e-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 05 07:10:20 crc kubenswrapper[4846]: I1005 07:10:20.497717 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/eb03eca3-f45d-4361-ab00-9ce16bb2c96e-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "eb03eca3-f45d-4361-ab00-9ce16bb2c96e" (UID: "eb03eca3-f45d-4361-ab00-9ce16bb2c96e"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 07:10:20 crc kubenswrapper[4846]: I1005 07:10:20.502787 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eb03eca3-f45d-4361-ab00-9ce16bb2c96e-kube-api-access-9hpsb" (OuterVolumeSpecName: "kube-api-access-9hpsb") pod "eb03eca3-f45d-4361-ab00-9ce16bb2c96e" (UID: "eb03eca3-f45d-4361-ab00-9ce16bb2c96e"). InnerVolumeSpecName "kube-api-access-9hpsb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:10:20 crc kubenswrapper[4846]: I1005 07:10:20.503612 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eb03eca3-f45d-4361-ab00-9ce16bb2c96e-scripts" (OuterVolumeSpecName: "scripts") pod "eb03eca3-f45d-4361-ab00-9ce16bb2c96e" (UID: "eb03eca3-f45d-4361-ab00-9ce16bb2c96e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:10:20 crc kubenswrapper[4846]: I1005 07:10:20.530863 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f1e8f9a4-b671-4417-99c9-90bda3c17339" path="/var/lib/kubelet/pods/f1e8f9a4-b671-4417-99c9-90bda3c17339/volumes" Oct 05 07:10:20 crc kubenswrapper[4846]: I1005 07:10:20.547598 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eb03eca3-f45d-4361-ab00-9ce16bb2c96e-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "eb03eca3-f45d-4361-ab00-9ce16bb2c96e" (UID: "eb03eca3-f45d-4361-ab00-9ce16bb2c96e"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:10:20 crc kubenswrapper[4846]: I1005 07:10:20.620963 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9hpsb\" (UniqueName: \"kubernetes.io/projected/eb03eca3-f45d-4361-ab00-9ce16bb2c96e-kube-api-access-9hpsb\") on node \"crc\" DevicePath \"\"" Oct 05 07:10:20 crc kubenswrapper[4846]: I1005 07:10:20.621003 4846 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/eb03eca3-f45d-4361-ab00-9ce16bb2c96e-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 05 07:10:20 crc kubenswrapper[4846]: I1005 07:10:20.621013 4846 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/eb03eca3-f45d-4361-ab00-9ce16bb2c96e-scripts\") on node \"crc\" DevicePath \"\"" Oct 05 07:10:20 crc kubenswrapper[4846]: I1005 07:10:20.621021 4846 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/eb03eca3-f45d-4361-ab00-9ce16bb2c96e-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 05 07:10:20 crc kubenswrapper[4846]: I1005 07:10:20.672376 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 05 07:10:20 crc kubenswrapper[4846]: I1005 07:10:20.697383 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eb03eca3-f45d-4361-ab00-9ce16bb2c96e-config-data" (OuterVolumeSpecName: "config-data") pod "eb03eca3-f45d-4361-ab00-9ce16bb2c96e" (UID: "eb03eca3-f45d-4361-ab00-9ce16bb2c96e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:10:20 crc kubenswrapper[4846]: I1005 07:10:20.700301 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"65fd80a1-d4ce-42f9-b1cf-e379333dfdba","Type":"ContainerStarted","Data":"12adb4adca58353410b4aba7e75c9560dd580019484783b4e35d98acb7df5496"} Oct 05 07:10:20 crc kubenswrapper[4846]: I1005 07:10:20.716466 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eb03eca3-f45d-4361-ab00-9ce16bb2c96e-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "eb03eca3-f45d-4361-ab00-9ce16bb2c96e" (UID: "eb03eca3-f45d-4361-ab00-9ce16bb2c96e"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:10:20 crc kubenswrapper[4846]: I1005 07:10:20.722965 4846 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/eb03eca3-f45d-4361-ab00-9ce16bb2c96e-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 05 07:10:20 crc kubenswrapper[4846]: I1005 07:10:20.722994 4846 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eb03eca3-f45d-4361-ab00-9ce16bb2c96e-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 07:10:20 crc kubenswrapper[4846]: I1005 07:10:20.754851 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eb03eca3-f45d-4361-ab00-9ce16bb2c96e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "eb03eca3-f45d-4361-ab00-9ce16bb2c96e" (UID: "eb03eca3-f45d-4361-ab00-9ce16bb2c96e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:10:20 crc kubenswrapper[4846]: I1005 07:10:20.755493 4846 generic.go:334] "Generic (PLEG): container finished" podID="eb03eca3-f45d-4361-ab00-9ce16bb2c96e" containerID="fafa6b60a9e821e9bc3c916554dcc88b5f10181956baed8a807149ed1db4333e" exitCode=0 Oct 05 07:10:20 crc kubenswrapper[4846]: I1005 07:10:20.755562 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"eb03eca3-f45d-4361-ab00-9ce16bb2c96e","Type":"ContainerDied","Data":"fafa6b60a9e821e9bc3c916554dcc88b5f10181956baed8a807149ed1db4333e"} Oct 05 07:10:20 crc kubenswrapper[4846]: I1005 07:10:20.755600 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"eb03eca3-f45d-4361-ab00-9ce16bb2c96e","Type":"ContainerDied","Data":"fc3cd487d783fbbd3afb2b971190c71056b88902b2807e17d5a1faf694fe9480"} Oct 05 07:10:20 crc kubenswrapper[4846]: I1005 07:10:20.755619 4846 scope.go:117] "RemoveContainer" containerID="32020fcfa11e9aff11fbbf804e18681fcebe28c354eef5dfd4c7dc9be93e5965" Oct 05 07:10:20 crc kubenswrapper[4846]: I1005 07:10:20.755812 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 05 07:10:20 crc kubenswrapper[4846]: I1005 07:10:20.825294 4846 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eb03eca3-f45d-4361-ab00-9ce16bb2c96e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 07:10:20 crc kubenswrapper[4846]: I1005 07:10:20.826401 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 05 07:10:20 crc kubenswrapper[4846]: I1005 07:10:20.830378 4846 scope.go:117] "RemoveContainer" containerID="6f1f221eac0cd0e08e90b60a2002245bc6a6ed3b4c9a396d48f47cca0039feb9" Oct 05 07:10:20 crc kubenswrapper[4846]: I1005 07:10:20.836819 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 05 07:10:20 crc kubenswrapper[4846]: I1005 07:10:20.838624 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Oct 05 07:10:20 crc kubenswrapper[4846]: I1005 07:10:20.866258 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 05 07:10:20 crc kubenswrapper[4846]: E1005 07:10:20.866820 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eb03eca3-f45d-4361-ab00-9ce16bb2c96e" containerName="sg-core" Oct 05 07:10:20 crc kubenswrapper[4846]: I1005 07:10:20.866841 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="eb03eca3-f45d-4361-ab00-9ce16bb2c96e" containerName="sg-core" Oct 05 07:10:20 crc kubenswrapper[4846]: E1005 07:10:20.866864 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eb03eca3-f45d-4361-ab00-9ce16bb2c96e" containerName="ceilometer-central-agent" Oct 05 07:10:20 crc kubenswrapper[4846]: I1005 07:10:20.866872 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="eb03eca3-f45d-4361-ab00-9ce16bb2c96e" containerName="ceilometer-central-agent" Oct 05 07:10:20 crc kubenswrapper[4846]: E1005 07:10:20.866891 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eb03eca3-f45d-4361-ab00-9ce16bb2c96e" containerName="ceilometer-notification-agent" Oct 05 07:10:20 crc kubenswrapper[4846]: I1005 07:10:20.866897 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="eb03eca3-f45d-4361-ab00-9ce16bb2c96e" containerName="ceilometer-notification-agent" Oct 05 07:10:20 crc kubenswrapper[4846]: E1005 07:10:20.866920 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eb03eca3-f45d-4361-ab00-9ce16bb2c96e" containerName="proxy-httpd" Oct 05 07:10:20 crc kubenswrapper[4846]: I1005 07:10:20.866926 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="eb03eca3-f45d-4361-ab00-9ce16bb2c96e" containerName="proxy-httpd" Oct 05 07:10:20 crc kubenswrapper[4846]: I1005 07:10:20.867129 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="eb03eca3-f45d-4361-ab00-9ce16bb2c96e" containerName="sg-core" Oct 05 07:10:20 crc kubenswrapper[4846]: I1005 07:10:20.867160 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="eb03eca3-f45d-4361-ab00-9ce16bb2c96e" containerName="ceilometer-central-agent" Oct 05 07:10:20 crc kubenswrapper[4846]: I1005 07:10:20.867169 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="eb03eca3-f45d-4361-ab00-9ce16bb2c96e" containerName="ceilometer-notification-agent" Oct 05 07:10:20 crc kubenswrapper[4846]: I1005 07:10:20.867196 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="eb03eca3-f45d-4361-ab00-9ce16bb2c96e" containerName="proxy-httpd" Oct 05 07:10:20 crc kubenswrapper[4846]: I1005 07:10:20.869021 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 05 07:10:20 crc kubenswrapper[4846]: I1005 07:10:20.872326 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 05 07:10:20 crc kubenswrapper[4846]: I1005 07:10:20.872872 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Oct 05 07:10:20 crc kubenswrapper[4846]: I1005 07:10:20.874401 4846 scope.go:117] "RemoveContainer" containerID="fafa6b60a9e821e9bc3c916554dcc88b5f10181956baed8a807149ed1db4333e" Oct 05 07:10:20 crc kubenswrapper[4846]: I1005 07:10:20.878433 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 05 07:10:20 crc kubenswrapper[4846]: I1005 07:10:20.884644 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 05 07:10:20 crc kubenswrapper[4846]: I1005 07:10:20.911448 4846 scope.go:117] "RemoveContainer" containerID="74b48fcfb6476893634deecb8bf996c2f28f54df151237a948e6426854564819" Oct 05 07:10:20 crc kubenswrapper[4846]: I1005 07:10:20.948834 4846 scope.go:117] "RemoveContainer" containerID="32020fcfa11e9aff11fbbf804e18681fcebe28c354eef5dfd4c7dc9be93e5965" Oct 05 07:10:20 crc kubenswrapper[4846]: E1005 07:10:20.950105 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"32020fcfa11e9aff11fbbf804e18681fcebe28c354eef5dfd4c7dc9be93e5965\": container with ID starting with 32020fcfa11e9aff11fbbf804e18681fcebe28c354eef5dfd4c7dc9be93e5965 not found: ID does not exist" containerID="32020fcfa11e9aff11fbbf804e18681fcebe28c354eef5dfd4c7dc9be93e5965" Oct 05 07:10:20 crc kubenswrapper[4846]: I1005 07:10:20.950139 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"32020fcfa11e9aff11fbbf804e18681fcebe28c354eef5dfd4c7dc9be93e5965"} err="failed to get container status \"32020fcfa11e9aff11fbbf804e18681fcebe28c354eef5dfd4c7dc9be93e5965\": rpc error: code = NotFound desc = could not find container \"32020fcfa11e9aff11fbbf804e18681fcebe28c354eef5dfd4c7dc9be93e5965\": container with ID starting with 32020fcfa11e9aff11fbbf804e18681fcebe28c354eef5dfd4c7dc9be93e5965 not found: ID does not exist" Oct 05 07:10:20 crc kubenswrapper[4846]: I1005 07:10:20.950160 4846 scope.go:117] "RemoveContainer" containerID="6f1f221eac0cd0e08e90b60a2002245bc6a6ed3b4c9a396d48f47cca0039feb9" Oct 05 07:10:20 crc kubenswrapper[4846]: E1005 07:10:20.951383 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6f1f221eac0cd0e08e90b60a2002245bc6a6ed3b4c9a396d48f47cca0039feb9\": container with ID starting with 6f1f221eac0cd0e08e90b60a2002245bc6a6ed3b4c9a396d48f47cca0039feb9 not found: ID does not exist" containerID="6f1f221eac0cd0e08e90b60a2002245bc6a6ed3b4c9a396d48f47cca0039feb9" Oct 05 07:10:20 crc kubenswrapper[4846]: I1005 07:10:20.951405 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6f1f221eac0cd0e08e90b60a2002245bc6a6ed3b4c9a396d48f47cca0039feb9"} err="failed to get container status \"6f1f221eac0cd0e08e90b60a2002245bc6a6ed3b4c9a396d48f47cca0039feb9\": rpc error: code = NotFound desc = could not find container \"6f1f221eac0cd0e08e90b60a2002245bc6a6ed3b4c9a396d48f47cca0039feb9\": container with ID starting with 6f1f221eac0cd0e08e90b60a2002245bc6a6ed3b4c9a396d48f47cca0039feb9 not found: ID does not exist" Oct 05 07:10:20 crc kubenswrapper[4846]: I1005 07:10:20.951419 4846 scope.go:117] "RemoveContainer" containerID="fafa6b60a9e821e9bc3c916554dcc88b5f10181956baed8a807149ed1db4333e" Oct 05 07:10:20 crc kubenswrapper[4846]: E1005 07:10:20.951813 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fafa6b60a9e821e9bc3c916554dcc88b5f10181956baed8a807149ed1db4333e\": container with ID starting with fafa6b60a9e821e9bc3c916554dcc88b5f10181956baed8a807149ed1db4333e not found: ID does not exist" containerID="fafa6b60a9e821e9bc3c916554dcc88b5f10181956baed8a807149ed1db4333e" Oct 05 07:10:20 crc kubenswrapper[4846]: I1005 07:10:20.951831 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fafa6b60a9e821e9bc3c916554dcc88b5f10181956baed8a807149ed1db4333e"} err="failed to get container status \"fafa6b60a9e821e9bc3c916554dcc88b5f10181956baed8a807149ed1db4333e\": rpc error: code = NotFound desc = could not find container \"fafa6b60a9e821e9bc3c916554dcc88b5f10181956baed8a807149ed1db4333e\": container with ID starting with fafa6b60a9e821e9bc3c916554dcc88b5f10181956baed8a807149ed1db4333e not found: ID does not exist" Oct 05 07:10:20 crc kubenswrapper[4846]: I1005 07:10:20.951846 4846 scope.go:117] "RemoveContainer" containerID="74b48fcfb6476893634deecb8bf996c2f28f54df151237a948e6426854564819" Oct 05 07:10:20 crc kubenswrapper[4846]: E1005 07:10:20.952606 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"74b48fcfb6476893634deecb8bf996c2f28f54df151237a948e6426854564819\": container with ID starting with 74b48fcfb6476893634deecb8bf996c2f28f54df151237a948e6426854564819 not found: ID does not exist" containerID="74b48fcfb6476893634deecb8bf996c2f28f54df151237a948e6426854564819" Oct 05 07:10:20 crc kubenswrapper[4846]: I1005 07:10:20.952630 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"74b48fcfb6476893634deecb8bf996c2f28f54df151237a948e6426854564819"} err="failed to get container status \"74b48fcfb6476893634deecb8bf996c2f28f54df151237a948e6426854564819\": rpc error: code = NotFound desc = could not find container \"74b48fcfb6476893634deecb8bf996c2f28f54df151237a948e6426854564819\": container with ID starting with 74b48fcfb6476893634deecb8bf996c2f28f54df151237a948e6426854564819 not found: ID does not exist" Oct 05 07:10:21 crc kubenswrapper[4846]: I1005 07:10:21.033106 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lggkw\" (UniqueName: \"kubernetes.io/projected/b1e394dd-f55b-4af1-af0b-52cc7d09afa1-kube-api-access-lggkw\") pod \"ceilometer-0\" (UID: \"b1e394dd-f55b-4af1-af0b-52cc7d09afa1\") " pod="openstack/ceilometer-0" Oct 05 07:10:21 crc kubenswrapper[4846]: I1005 07:10:21.033163 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b1e394dd-f55b-4af1-af0b-52cc7d09afa1-scripts\") pod \"ceilometer-0\" (UID: \"b1e394dd-f55b-4af1-af0b-52cc7d09afa1\") " pod="openstack/ceilometer-0" Oct 05 07:10:21 crc kubenswrapper[4846]: I1005 07:10:21.033271 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b1e394dd-f55b-4af1-af0b-52cc7d09afa1-run-httpd\") pod \"ceilometer-0\" (UID: \"b1e394dd-f55b-4af1-af0b-52cc7d09afa1\") " pod="openstack/ceilometer-0" Oct 05 07:10:21 crc kubenswrapper[4846]: I1005 07:10:21.033316 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b1e394dd-f55b-4af1-af0b-52cc7d09afa1-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"b1e394dd-f55b-4af1-af0b-52cc7d09afa1\") " pod="openstack/ceilometer-0" Oct 05 07:10:21 crc kubenswrapper[4846]: I1005 07:10:21.033391 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/b1e394dd-f55b-4af1-af0b-52cc7d09afa1-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"b1e394dd-f55b-4af1-af0b-52cc7d09afa1\") " pod="openstack/ceilometer-0" Oct 05 07:10:21 crc kubenswrapper[4846]: I1005 07:10:21.033417 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b1e394dd-f55b-4af1-af0b-52cc7d09afa1-log-httpd\") pod \"ceilometer-0\" (UID: \"b1e394dd-f55b-4af1-af0b-52cc7d09afa1\") " pod="openstack/ceilometer-0" Oct 05 07:10:21 crc kubenswrapper[4846]: I1005 07:10:21.033459 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b1e394dd-f55b-4af1-af0b-52cc7d09afa1-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"b1e394dd-f55b-4af1-af0b-52cc7d09afa1\") " pod="openstack/ceilometer-0" Oct 05 07:10:21 crc kubenswrapper[4846]: I1005 07:10:21.033487 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b1e394dd-f55b-4af1-af0b-52cc7d09afa1-config-data\") pod \"ceilometer-0\" (UID: \"b1e394dd-f55b-4af1-af0b-52cc7d09afa1\") " pod="openstack/ceilometer-0" Oct 05 07:10:21 crc kubenswrapper[4846]: I1005 07:10:21.049053 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-cell-mapping-t7fdh"] Oct 05 07:10:21 crc kubenswrapper[4846]: I1005 07:10:21.052822 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-t7fdh" Oct 05 07:10:21 crc kubenswrapper[4846]: I1005 07:10:21.057991 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-config-data" Oct 05 07:10:21 crc kubenswrapper[4846]: I1005 07:10:21.058399 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-scripts" Oct 05 07:10:21 crc kubenswrapper[4846]: I1005 07:10:21.065735 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-t7fdh"] Oct 05 07:10:21 crc kubenswrapper[4846]: I1005 07:10:21.135229 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/b1e394dd-f55b-4af1-af0b-52cc7d09afa1-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"b1e394dd-f55b-4af1-af0b-52cc7d09afa1\") " pod="openstack/ceilometer-0" Oct 05 07:10:21 crc kubenswrapper[4846]: I1005 07:10:21.135511 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b1e394dd-f55b-4af1-af0b-52cc7d09afa1-log-httpd\") pod \"ceilometer-0\" (UID: \"b1e394dd-f55b-4af1-af0b-52cc7d09afa1\") " pod="openstack/ceilometer-0" Oct 05 07:10:21 crc kubenswrapper[4846]: I1005 07:10:21.135637 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8fa2fcc5-ba88-4a8c-88ec-a8d1882b8c7b-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-t7fdh\" (UID: \"8fa2fcc5-ba88-4a8c-88ec-a8d1882b8c7b\") " pod="openstack/nova-cell1-cell-mapping-t7fdh" Oct 05 07:10:21 crc kubenswrapper[4846]: I1005 07:10:21.135745 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b1e394dd-f55b-4af1-af0b-52cc7d09afa1-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"b1e394dd-f55b-4af1-af0b-52cc7d09afa1\") " pod="openstack/ceilometer-0" Oct 05 07:10:21 crc kubenswrapper[4846]: I1005 07:10:21.135836 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b1e394dd-f55b-4af1-af0b-52cc7d09afa1-config-data\") pod \"ceilometer-0\" (UID: \"b1e394dd-f55b-4af1-af0b-52cc7d09afa1\") " pod="openstack/ceilometer-0" Oct 05 07:10:21 crc kubenswrapper[4846]: I1005 07:10:21.135941 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lggkw\" (UniqueName: \"kubernetes.io/projected/b1e394dd-f55b-4af1-af0b-52cc7d09afa1-kube-api-access-lggkw\") pod \"ceilometer-0\" (UID: \"b1e394dd-f55b-4af1-af0b-52cc7d09afa1\") " pod="openstack/ceilometer-0" Oct 05 07:10:21 crc kubenswrapper[4846]: I1005 07:10:21.136037 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b1e394dd-f55b-4af1-af0b-52cc7d09afa1-scripts\") pod \"ceilometer-0\" (UID: \"b1e394dd-f55b-4af1-af0b-52cc7d09afa1\") " pod="openstack/ceilometer-0" Oct 05 07:10:21 crc kubenswrapper[4846]: I1005 07:10:21.136171 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8fa2fcc5-ba88-4a8c-88ec-a8d1882b8c7b-scripts\") pod \"nova-cell1-cell-mapping-t7fdh\" (UID: \"8fa2fcc5-ba88-4a8c-88ec-a8d1882b8c7b\") " pod="openstack/nova-cell1-cell-mapping-t7fdh" Oct 05 07:10:21 crc kubenswrapper[4846]: I1005 07:10:21.136311 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b1e394dd-f55b-4af1-af0b-52cc7d09afa1-run-httpd\") pod \"ceilometer-0\" (UID: \"b1e394dd-f55b-4af1-af0b-52cc7d09afa1\") " pod="openstack/ceilometer-0" Oct 05 07:10:21 crc kubenswrapper[4846]: I1005 07:10:21.136422 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8fa2fcc5-ba88-4a8c-88ec-a8d1882b8c7b-config-data\") pod \"nova-cell1-cell-mapping-t7fdh\" (UID: \"8fa2fcc5-ba88-4a8c-88ec-a8d1882b8c7b\") " pod="openstack/nova-cell1-cell-mapping-t7fdh" Oct 05 07:10:21 crc kubenswrapper[4846]: I1005 07:10:21.136518 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2k2x4\" (UniqueName: \"kubernetes.io/projected/8fa2fcc5-ba88-4a8c-88ec-a8d1882b8c7b-kube-api-access-2k2x4\") pod \"nova-cell1-cell-mapping-t7fdh\" (UID: \"8fa2fcc5-ba88-4a8c-88ec-a8d1882b8c7b\") " pod="openstack/nova-cell1-cell-mapping-t7fdh" Oct 05 07:10:21 crc kubenswrapper[4846]: I1005 07:10:21.136616 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b1e394dd-f55b-4af1-af0b-52cc7d09afa1-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"b1e394dd-f55b-4af1-af0b-52cc7d09afa1\") " pod="openstack/ceilometer-0" Oct 05 07:10:21 crc kubenswrapper[4846]: I1005 07:10:21.135993 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b1e394dd-f55b-4af1-af0b-52cc7d09afa1-log-httpd\") pod \"ceilometer-0\" (UID: \"b1e394dd-f55b-4af1-af0b-52cc7d09afa1\") " pod="openstack/ceilometer-0" Oct 05 07:10:21 crc kubenswrapper[4846]: I1005 07:10:21.137781 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b1e394dd-f55b-4af1-af0b-52cc7d09afa1-run-httpd\") pod \"ceilometer-0\" (UID: \"b1e394dd-f55b-4af1-af0b-52cc7d09afa1\") " pod="openstack/ceilometer-0" Oct 05 07:10:21 crc kubenswrapper[4846]: I1005 07:10:21.139933 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/b1e394dd-f55b-4af1-af0b-52cc7d09afa1-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"b1e394dd-f55b-4af1-af0b-52cc7d09afa1\") " pod="openstack/ceilometer-0" Oct 05 07:10:21 crc kubenswrapper[4846]: I1005 07:10:21.141241 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b1e394dd-f55b-4af1-af0b-52cc7d09afa1-scripts\") pod \"ceilometer-0\" (UID: \"b1e394dd-f55b-4af1-af0b-52cc7d09afa1\") " pod="openstack/ceilometer-0" Oct 05 07:10:21 crc kubenswrapper[4846]: I1005 07:10:21.141272 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b1e394dd-f55b-4af1-af0b-52cc7d09afa1-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"b1e394dd-f55b-4af1-af0b-52cc7d09afa1\") " pod="openstack/ceilometer-0" Oct 05 07:10:21 crc kubenswrapper[4846]: I1005 07:10:21.141734 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b1e394dd-f55b-4af1-af0b-52cc7d09afa1-config-data\") pod \"ceilometer-0\" (UID: \"b1e394dd-f55b-4af1-af0b-52cc7d09afa1\") " pod="openstack/ceilometer-0" Oct 05 07:10:21 crc kubenswrapper[4846]: I1005 07:10:21.144103 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b1e394dd-f55b-4af1-af0b-52cc7d09afa1-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"b1e394dd-f55b-4af1-af0b-52cc7d09afa1\") " pod="openstack/ceilometer-0" Oct 05 07:10:21 crc kubenswrapper[4846]: I1005 07:10:21.155734 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lggkw\" (UniqueName: \"kubernetes.io/projected/b1e394dd-f55b-4af1-af0b-52cc7d09afa1-kube-api-access-lggkw\") pod \"ceilometer-0\" (UID: \"b1e394dd-f55b-4af1-af0b-52cc7d09afa1\") " pod="openstack/ceilometer-0" Oct 05 07:10:21 crc kubenswrapper[4846]: I1005 07:10:21.193247 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 05 07:10:21 crc kubenswrapper[4846]: I1005 07:10:21.238923 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8fa2fcc5-ba88-4a8c-88ec-a8d1882b8c7b-config-data\") pod \"nova-cell1-cell-mapping-t7fdh\" (UID: \"8fa2fcc5-ba88-4a8c-88ec-a8d1882b8c7b\") " pod="openstack/nova-cell1-cell-mapping-t7fdh" Oct 05 07:10:21 crc kubenswrapper[4846]: I1005 07:10:21.239351 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2k2x4\" (UniqueName: \"kubernetes.io/projected/8fa2fcc5-ba88-4a8c-88ec-a8d1882b8c7b-kube-api-access-2k2x4\") pod \"nova-cell1-cell-mapping-t7fdh\" (UID: \"8fa2fcc5-ba88-4a8c-88ec-a8d1882b8c7b\") " pod="openstack/nova-cell1-cell-mapping-t7fdh" Oct 05 07:10:21 crc kubenswrapper[4846]: I1005 07:10:21.239460 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8fa2fcc5-ba88-4a8c-88ec-a8d1882b8c7b-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-t7fdh\" (UID: \"8fa2fcc5-ba88-4a8c-88ec-a8d1882b8c7b\") " pod="openstack/nova-cell1-cell-mapping-t7fdh" Oct 05 07:10:21 crc kubenswrapper[4846]: I1005 07:10:21.239557 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8fa2fcc5-ba88-4a8c-88ec-a8d1882b8c7b-scripts\") pod \"nova-cell1-cell-mapping-t7fdh\" (UID: \"8fa2fcc5-ba88-4a8c-88ec-a8d1882b8c7b\") " pod="openstack/nova-cell1-cell-mapping-t7fdh" Oct 05 07:10:21 crc kubenswrapper[4846]: I1005 07:10:21.248034 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8fa2fcc5-ba88-4a8c-88ec-a8d1882b8c7b-config-data\") pod \"nova-cell1-cell-mapping-t7fdh\" (UID: \"8fa2fcc5-ba88-4a8c-88ec-a8d1882b8c7b\") " pod="openstack/nova-cell1-cell-mapping-t7fdh" Oct 05 07:10:21 crc kubenswrapper[4846]: I1005 07:10:21.256005 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8fa2fcc5-ba88-4a8c-88ec-a8d1882b8c7b-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-t7fdh\" (UID: \"8fa2fcc5-ba88-4a8c-88ec-a8d1882b8c7b\") " pod="openstack/nova-cell1-cell-mapping-t7fdh" Oct 05 07:10:21 crc kubenswrapper[4846]: I1005 07:10:21.257197 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8fa2fcc5-ba88-4a8c-88ec-a8d1882b8c7b-scripts\") pod \"nova-cell1-cell-mapping-t7fdh\" (UID: \"8fa2fcc5-ba88-4a8c-88ec-a8d1882b8c7b\") " pod="openstack/nova-cell1-cell-mapping-t7fdh" Oct 05 07:10:21 crc kubenswrapper[4846]: I1005 07:10:21.262892 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2k2x4\" (UniqueName: \"kubernetes.io/projected/8fa2fcc5-ba88-4a8c-88ec-a8d1882b8c7b-kube-api-access-2k2x4\") pod \"nova-cell1-cell-mapping-t7fdh\" (UID: \"8fa2fcc5-ba88-4a8c-88ec-a8d1882b8c7b\") " pod="openstack/nova-cell1-cell-mapping-t7fdh" Oct 05 07:10:21 crc kubenswrapper[4846]: I1005 07:10:21.291307 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-t7fdh" Oct 05 07:10:21 crc kubenswrapper[4846]: I1005 07:10:21.644347 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 05 07:10:21 crc kubenswrapper[4846]: I1005 07:10:21.649020 4846 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 05 07:10:21 crc kubenswrapper[4846]: I1005 07:10:21.810146 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b1e394dd-f55b-4af1-af0b-52cc7d09afa1","Type":"ContainerStarted","Data":"d8f6a2cda488a503a4cdbec1246f11f58542b60bc44c2b7a0b7ba804f3e98a70"} Oct 05 07:10:21 crc kubenswrapper[4846]: I1005 07:10:21.815356 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"65fd80a1-d4ce-42f9-b1cf-e379333dfdba","Type":"ContainerStarted","Data":"0bf6cd09e5de807030743657cdd80e0b641715af3513167ff53800b7f6c565be"} Oct 05 07:10:21 crc kubenswrapper[4846]: I1005 07:10:21.815405 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"65fd80a1-d4ce-42f9-b1cf-e379333dfdba","Type":"ContainerStarted","Data":"9a4be5b54dc458b481c513b652f2f0da4ceaaa036f7b850ef512811ccd3355ff"} Oct 05 07:10:21 crc kubenswrapper[4846]: I1005 07:10:21.850388 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.850360057 podStartE2EDuration="2.850360057s" podCreationTimestamp="2025-10-05 07:10:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 07:10:21.835926272 +0000 UTC m=+1344.076779047" watchObservedRunningTime="2025-10-05 07:10:21.850360057 +0000 UTC m=+1344.091212832" Oct 05 07:10:21 crc kubenswrapper[4846]: W1005 07:10:21.853595 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8fa2fcc5_ba88_4a8c_88ec_a8d1882b8c7b.slice/crio-42915f74331bbfe756ff193d37f0ca9231e63a3cd41f0ab7fa3e7af088da7b2c WatchSource:0}: Error finding container 42915f74331bbfe756ff193d37f0ca9231e63a3cd41f0ab7fa3e7af088da7b2c: Status 404 returned error can't find the container with id 42915f74331bbfe756ff193d37f0ca9231e63a3cd41f0ab7fa3e7af088da7b2c Oct 05 07:10:21 crc kubenswrapper[4846]: I1005 07:10:21.860775 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-t7fdh"] Oct 05 07:10:22 crc kubenswrapper[4846]: I1005 07:10:22.513138 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="eb03eca3-f45d-4361-ab00-9ce16bb2c96e" path="/var/lib/kubelet/pods/eb03eca3-f45d-4361-ab00-9ce16bb2c96e/volumes" Oct 05 07:10:22 crc kubenswrapper[4846]: I1005 07:10:22.845424 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-t7fdh" event={"ID":"8fa2fcc5-ba88-4a8c-88ec-a8d1882b8c7b","Type":"ContainerStarted","Data":"df3b6f1e9e992a4de6fcc136d2a372cb06b4dc91089da63e8a9ba4a7378ed72d"} Oct 05 07:10:22 crc kubenswrapper[4846]: I1005 07:10:22.845491 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-t7fdh" event={"ID":"8fa2fcc5-ba88-4a8c-88ec-a8d1882b8c7b","Type":"ContainerStarted","Data":"42915f74331bbfe756ff193d37f0ca9231e63a3cd41f0ab7fa3e7af088da7b2c"} Oct 05 07:10:22 crc kubenswrapper[4846]: I1005 07:10:22.850860 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b1e394dd-f55b-4af1-af0b-52cc7d09afa1","Type":"ContainerStarted","Data":"6682a3a723ef5336f655521943bfbd504709a37185a094744e9c63a0c851b903"} Oct 05 07:10:22 crc kubenswrapper[4846]: I1005 07:10:22.878601 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-cell-mapping-t7fdh" podStartSLOduration=1.8785771329999998 podStartE2EDuration="1.878577133s" podCreationTimestamp="2025-10-05 07:10:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 07:10:22.8679925 +0000 UTC m=+1345.108845285" watchObservedRunningTime="2025-10-05 07:10:22.878577133 +0000 UTC m=+1345.119429928" Oct 05 07:10:23 crc kubenswrapper[4846]: I1005 07:10:23.148375 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-c7d6584f5-dxcgg" Oct 05 07:10:23 crc kubenswrapper[4846]: I1005 07:10:23.219334 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78565f64c9-vkrdf"] Oct 05 07:10:23 crc kubenswrapper[4846]: I1005 07:10:23.219598 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-78565f64c9-vkrdf" podUID="68a854c2-afda-4dbb-9ac7-06eec2c98f14" containerName="dnsmasq-dns" containerID="cri-o://96643ce5b5a2f18ab40b26a3bea0d27ff26ce9e8cfedae27efdb2d0b75872aa3" gracePeriod=10 Oct 05 07:10:23 crc kubenswrapper[4846]: I1005 07:10:23.326839 4846 patch_prober.go:28] interesting pod/machine-config-daemon-fscvf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 07:10:23 crc kubenswrapper[4846]: I1005 07:10:23.327317 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 07:10:23 crc kubenswrapper[4846]: I1005 07:10:23.327394 4846 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" Oct 05 07:10:23 crc kubenswrapper[4846]: I1005 07:10:23.328352 4846 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"11b9703888d7e63159496cf8c65783a81599439b16d1a1e5677afe95613a82de"} pod="openshift-machine-config-operator/machine-config-daemon-fscvf" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 05 07:10:23 crc kubenswrapper[4846]: I1005 07:10:23.328413 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" containerName="machine-config-daemon" containerID="cri-o://11b9703888d7e63159496cf8c65783a81599439b16d1a1e5677afe95613a82de" gracePeriod=600 Oct 05 07:10:23 crc kubenswrapper[4846]: I1005 07:10:23.398960 4846 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-78565f64c9-vkrdf" podUID="68a854c2-afda-4dbb-9ac7-06eec2c98f14" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.191:5353: connect: connection refused" Oct 05 07:10:23 crc kubenswrapper[4846]: I1005 07:10:23.844942 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78565f64c9-vkrdf" Oct 05 07:10:23 crc kubenswrapper[4846]: I1005 07:10:23.883146 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b1e394dd-f55b-4af1-af0b-52cc7d09afa1","Type":"ContainerStarted","Data":"d0d0db8b415aac7ced1865679a2764d980d7574715bf4b8a924f0edb74f9f2fc"} Oct 05 07:10:23 crc kubenswrapper[4846]: I1005 07:10:23.904661 4846 generic.go:334] "Generic (PLEG): container finished" podID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" containerID="11b9703888d7e63159496cf8c65783a81599439b16d1a1e5677afe95613a82de" exitCode=0 Oct 05 07:10:23 crc kubenswrapper[4846]: I1005 07:10:23.904735 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" event={"ID":"ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7","Type":"ContainerDied","Data":"11b9703888d7e63159496cf8c65783a81599439b16d1a1e5677afe95613a82de"} Oct 05 07:10:23 crc kubenswrapper[4846]: I1005 07:10:23.904775 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" event={"ID":"ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7","Type":"ContainerStarted","Data":"7d363dbd53ee92c0974009df77db8574365e53ef73df754e1c6fabb86787d68a"} Oct 05 07:10:23 crc kubenswrapper[4846]: I1005 07:10:23.904808 4846 scope.go:117] "RemoveContainer" containerID="a25250ef8861b2bca088edb9f5248f80d97de22c7bacf6dfaaf8d42d30c96858" Oct 05 07:10:23 crc kubenswrapper[4846]: I1005 07:10:23.915742 4846 generic.go:334] "Generic (PLEG): container finished" podID="68a854c2-afda-4dbb-9ac7-06eec2c98f14" containerID="96643ce5b5a2f18ab40b26a3bea0d27ff26ce9e8cfedae27efdb2d0b75872aa3" exitCode=0 Oct 05 07:10:23 crc kubenswrapper[4846]: I1005 07:10:23.915865 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78565f64c9-vkrdf" Oct 05 07:10:23 crc kubenswrapper[4846]: I1005 07:10:23.916422 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78565f64c9-vkrdf" event={"ID":"68a854c2-afda-4dbb-9ac7-06eec2c98f14","Type":"ContainerDied","Data":"96643ce5b5a2f18ab40b26a3bea0d27ff26ce9e8cfedae27efdb2d0b75872aa3"} Oct 05 07:10:23 crc kubenswrapper[4846]: I1005 07:10:23.916453 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78565f64c9-vkrdf" event={"ID":"68a854c2-afda-4dbb-9ac7-06eec2c98f14","Type":"ContainerDied","Data":"76e87ad58f355f456114724a6744749e735d7ff76f24ef36aaf392dcddaa11e8"} Oct 05 07:10:23 crc kubenswrapper[4846]: I1005 07:10:23.948162 4846 scope.go:117] "RemoveContainer" containerID="96643ce5b5a2f18ab40b26a3bea0d27ff26ce9e8cfedae27efdb2d0b75872aa3" Oct 05 07:10:23 crc kubenswrapper[4846]: I1005 07:10:23.958963 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/68a854c2-afda-4dbb-9ac7-06eec2c98f14-dns-swift-storage-0\") pod \"68a854c2-afda-4dbb-9ac7-06eec2c98f14\" (UID: \"68a854c2-afda-4dbb-9ac7-06eec2c98f14\") " Oct 05 07:10:23 crc kubenswrapper[4846]: I1005 07:10:23.959040 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/68a854c2-afda-4dbb-9ac7-06eec2c98f14-ovsdbserver-sb\") pod \"68a854c2-afda-4dbb-9ac7-06eec2c98f14\" (UID: \"68a854c2-afda-4dbb-9ac7-06eec2c98f14\") " Oct 05 07:10:23 crc kubenswrapper[4846]: I1005 07:10:23.959086 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/68a854c2-afda-4dbb-9ac7-06eec2c98f14-dns-svc\") pod \"68a854c2-afda-4dbb-9ac7-06eec2c98f14\" (UID: \"68a854c2-afda-4dbb-9ac7-06eec2c98f14\") " Oct 05 07:10:23 crc kubenswrapper[4846]: I1005 07:10:23.959141 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/68a854c2-afda-4dbb-9ac7-06eec2c98f14-config\") pod \"68a854c2-afda-4dbb-9ac7-06eec2c98f14\" (UID: \"68a854c2-afda-4dbb-9ac7-06eec2c98f14\") " Oct 05 07:10:23 crc kubenswrapper[4846]: I1005 07:10:23.959207 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/68a854c2-afda-4dbb-9ac7-06eec2c98f14-ovsdbserver-nb\") pod \"68a854c2-afda-4dbb-9ac7-06eec2c98f14\" (UID: \"68a854c2-afda-4dbb-9ac7-06eec2c98f14\") " Oct 05 07:10:23 crc kubenswrapper[4846]: I1005 07:10:23.959429 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qdvf6\" (UniqueName: \"kubernetes.io/projected/68a854c2-afda-4dbb-9ac7-06eec2c98f14-kube-api-access-qdvf6\") pod \"68a854c2-afda-4dbb-9ac7-06eec2c98f14\" (UID: \"68a854c2-afda-4dbb-9ac7-06eec2c98f14\") " Oct 05 07:10:23 crc kubenswrapper[4846]: I1005 07:10:23.970681 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/68a854c2-afda-4dbb-9ac7-06eec2c98f14-kube-api-access-qdvf6" (OuterVolumeSpecName: "kube-api-access-qdvf6") pod "68a854c2-afda-4dbb-9ac7-06eec2c98f14" (UID: "68a854c2-afda-4dbb-9ac7-06eec2c98f14"). InnerVolumeSpecName "kube-api-access-qdvf6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:10:23 crc kubenswrapper[4846]: I1005 07:10:23.995079 4846 scope.go:117] "RemoveContainer" containerID="21988ef9f5e0ab34b6030d3b8b9a19ca844adb411f5ca6ecbb5d5f7fb49d21a6" Oct 05 07:10:24 crc kubenswrapper[4846]: I1005 07:10:24.062476 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qdvf6\" (UniqueName: \"kubernetes.io/projected/68a854c2-afda-4dbb-9ac7-06eec2c98f14-kube-api-access-qdvf6\") on node \"crc\" DevicePath \"\"" Oct 05 07:10:24 crc kubenswrapper[4846]: I1005 07:10:24.062991 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/68a854c2-afda-4dbb-9ac7-06eec2c98f14-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "68a854c2-afda-4dbb-9ac7-06eec2c98f14" (UID: "68a854c2-afda-4dbb-9ac7-06eec2c98f14"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 07:10:24 crc kubenswrapper[4846]: I1005 07:10:24.105402 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/68a854c2-afda-4dbb-9ac7-06eec2c98f14-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "68a854c2-afda-4dbb-9ac7-06eec2c98f14" (UID: "68a854c2-afda-4dbb-9ac7-06eec2c98f14"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 07:10:24 crc kubenswrapper[4846]: I1005 07:10:24.120647 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/68a854c2-afda-4dbb-9ac7-06eec2c98f14-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "68a854c2-afda-4dbb-9ac7-06eec2c98f14" (UID: "68a854c2-afda-4dbb-9ac7-06eec2c98f14"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 07:10:24 crc kubenswrapper[4846]: I1005 07:10:24.137507 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/68a854c2-afda-4dbb-9ac7-06eec2c98f14-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "68a854c2-afda-4dbb-9ac7-06eec2c98f14" (UID: "68a854c2-afda-4dbb-9ac7-06eec2c98f14"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 07:10:24 crc kubenswrapper[4846]: I1005 07:10:24.137894 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/68a854c2-afda-4dbb-9ac7-06eec2c98f14-config" (OuterVolumeSpecName: "config") pod "68a854c2-afda-4dbb-9ac7-06eec2c98f14" (UID: "68a854c2-afda-4dbb-9ac7-06eec2c98f14"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 07:10:24 crc kubenswrapper[4846]: I1005 07:10:24.164773 4846 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/68a854c2-afda-4dbb-9ac7-06eec2c98f14-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 05 07:10:24 crc kubenswrapper[4846]: I1005 07:10:24.164827 4846 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/68a854c2-afda-4dbb-9ac7-06eec2c98f14-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 05 07:10:24 crc kubenswrapper[4846]: I1005 07:10:24.164838 4846 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/68a854c2-afda-4dbb-9ac7-06eec2c98f14-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 05 07:10:24 crc kubenswrapper[4846]: I1005 07:10:24.164848 4846 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/68a854c2-afda-4dbb-9ac7-06eec2c98f14-config\") on node \"crc\" DevicePath \"\"" Oct 05 07:10:24 crc kubenswrapper[4846]: I1005 07:10:24.164863 4846 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/68a854c2-afda-4dbb-9ac7-06eec2c98f14-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 05 07:10:24 crc kubenswrapper[4846]: I1005 07:10:24.213135 4846 scope.go:117] "RemoveContainer" containerID="96643ce5b5a2f18ab40b26a3bea0d27ff26ce9e8cfedae27efdb2d0b75872aa3" Oct 05 07:10:24 crc kubenswrapper[4846]: E1005 07:10:24.213671 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"96643ce5b5a2f18ab40b26a3bea0d27ff26ce9e8cfedae27efdb2d0b75872aa3\": container with ID starting with 96643ce5b5a2f18ab40b26a3bea0d27ff26ce9e8cfedae27efdb2d0b75872aa3 not found: ID does not exist" containerID="96643ce5b5a2f18ab40b26a3bea0d27ff26ce9e8cfedae27efdb2d0b75872aa3" Oct 05 07:10:24 crc kubenswrapper[4846]: I1005 07:10:24.213713 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"96643ce5b5a2f18ab40b26a3bea0d27ff26ce9e8cfedae27efdb2d0b75872aa3"} err="failed to get container status \"96643ce5b5a2f18ab40b26a3bea0d27ff26ce9e8cfedae27efdb2d0b75872aa3\": rpc error: code = NotFound desc = could not find container \"96643ce5b5a2f18ab40b26a3bea0d27ff26ce9e8cfedae27efdb2d0b75872aa3\": container with ID starting with 96643ce5b5a2f18ab40b26a3bea0d27ff26ce9e8cfedae27efdb2d0b75872aa3 not found: ID does not exist" Oct 05 07:10:24 crc kubenswrapper[4846]: I1005 07:10:24.213742 4846 scope.go:117] "RemoveContainer" containerID="21988ef9f5e0ab34b6030d3b8b9a19ca844adb411f5ca6ecbb5d5f7fb49d21a6" Oct 05 07:10:24 crc kubenswrapper[4846]: E1005 07:10:24.214056 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"21988ef9f5e0ab34b6030d3b8b9a19ca844adb411f5ca6ecbb5d5f7fb49d21a6\": container with ID starting with 21988ef9f5e0ab34b6030d3b8b9a19ca844adb411f5ca6ecbb5d5f7fb49d21a6 not found: ID does not exist" containerID="21988ef9f5e0ab34b6030d3b8b9a19ca844adb411f5ca6ecbb5d5f7fb49d21a6" Oct 05 07:10:24 crc kubenswrapper[4846]: I1005 07:10:24.214077 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"21988ef9f5e0ab34b6030d3b8b9a19ca844adb411f5ca6ecbb5d5f7fb49d21a6"} err="failed to get container status \"21988ef9f5e0ab34b6030d3b8b9a19ca844adb411f5ca6ecbb5d5f7fb49d21a6\": rpc error: code = NotFound desc = could not find container \"21988ef9f5e0ab34b6030d3b8b9a19ca844adb411f5ca6ecbb5d5f7fb49d21a6\": container with ID starting with 21988ef9f5e0ab34b6030d3b8b9a19ca844adb411f5ca6ecbb5d5f7fb49d21a6 not found: ID does not exist" Oct 05 07:10:24 crc kubenswrapper[4846]: I1005 07:10:24.279091 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78565f64c9-vkrdf"] Oct 05 07:10:24 crc kubenswrapper[4846]: I1005 07:10:24.296558 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-78565f64c9-vkrdf"] Oct 05 07:10:24 crc kubenswrapper[4846]: I1005 07:10:24.509983 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="68a854c2-afda-4dbb-9ac7-06eec2c98f14" path="/var/lib/kubelet/pods/68a854c2-afda-4dbb-9ac7-06eec2c98f14/volumes" Oct 05 07:10:24 crc kubenswrapper[4846]: I1005 07:10:24.929308 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b1e394dd-f55b-4af1-af0b-52cc7d09afa1","Type":"ContainerStarted","Data":"6db54d2ca22110333154adb6a23a0ec27e54e4f7118d44ea9636189fffd9e0ae"} Oct 05 07:10:26 crc kubenswrapper[4846]: I1005 07:10:26.963686 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b1e394dd-f55b-4af1-af0b-52cc7d09afa1","Type":"ContainerStarted","Data":"e92d479cbafa7786e2514b2d1e99c120ef33197a849bb0c13e2362354bcebd01"} Oct 05 07:10:26 crc kubenswrapper[4846]: I1005 07:10:26.964577 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 05 07:10:26 crc kubenswrapper[4846]: I1005 07:10:26.997607 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.84024226 podStartE2EDuration="6.99758277s" podCreationTimestamp="2025-10-05 07:10:20 +0000 UTC" firstStartedPulling="2025-10-05 07:10:21.648793107 +0000 UTC m=+1343.889645882" lastFinishedPulling="2025-10-05 07:10:25.806133607 +0000 UTC m=+1348.046986392" observedRunningTime="2025-10-05 07:10:26.992636948 +0000 UTC m=+1349.233489733" watchObservedRunningTime="2025-10-05 07:10:26.99758277 +0000 UTC m=+1349.238435545" Oct 05 07:10:30 crc kubenswrapper[4846]: I1005 07:10:30.000318 4846 generic.go:334] "Generic (PLEG): container finished" podID="8fa2fcc5-ba88-4a8c-88ec-a8d1882b8c7b" containerID="df3b6f1e9e992a4de6fcc136d2a372cb06b4dc91089da63e8a9ba4a7378ed72d" exitCode=0 Oct 05 07:10:30 crc kubenswrapper[4846]: I1005 07:10:30.001028 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-t7fdh" event={"ID":"8fa2fcc5-ba88-4a8c-88ec-a8d1882b8c7b","Type":"ContainerDied","Data":"df3b6f1e9e992a4de6fcc136d2a372cb06b4dc91089da63e8a9ba4a7378ed72d"} Oct 05 07:10:30 crc kubenswrapper[4846]: I1005 07:10:30.086054 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 05 07:10:30 crc kubenswrapper[4846]: I1005 07:10:30.086126 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 05 07:10:31 crc kubenswrapper[4846]: I1005 07:10:31.100440 4846 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="65fd80a1-d4ce-42f9-b1cf-e379333dfdba" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.202:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 05 07:10:31 crc kubenswrapper[4846]: I1005 07:10:31.100413 4846 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="65fd80a1-d4ce-42f9-b1cf-e379333dfdba" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.202:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 05 07:10:31 crc kubenswrapper[4846]: I1005 07:10:31.453352 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-t7fdh" Oct 05 07:10:31 crc kubenswrapper[4846]: I1005 07:10:31.533055 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8fa2fcc5-ba88-4a8c-88ec-a8d1882b8c7b-config-data\") pod \"8fa2fcc5-ba88-4a8c-88ec-a8d1882b8c7b\" (UID: \"8fa2fcc5-ba88-4a8c-88ec-a8d1882b8c7b\") " Oct 05 07:10:31 crc kubenswrapper[4846]: I1005 07:10:31.533235 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2k2x4\" (UniqueName: \"kubernetes.io/projected/8fa2fcc5-ba88-4a8c-88ec-a8d1882b8c7b-kube-api-access-2k2x4\") pod \"8fa2fcc5-ba88-4a8c-88ec-a8d1882b8c7b\" (UID: \"8fa2fcc5-ba88-4a8c-88ec-a8d1882b8c7b\") " Oct 05 07:10:31 crc kubenswrapper[4846]: I1005 07:10:31.533347 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8fa2fcc5-ba88-4a8c-88ec-a8d1882b8c7b-scripts\") pod \"8fa2fcc5-ba88-4a8c-88ec-a8d1882b8c7b\" (UID: \"8fa2fcc5-ba88-4a8c-88ec-a8d1882b8c7b\") " Oct 05 07:10:31 crc kubenswrapper[4846]: I1005 07:10:31.533369 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8fa2fcc5-ba88-4a8c-88ec-a8d1882b8c7b-combined-ca-bundle\") pod \"8fa2fcc5-ba88-4a8c-88ec-a8d1882b8c7b\" (UID: \"8fa2fcc5-ba88-4a8c-88ec-a8d1882b8c7b\") " Oct 05 07:10:31 crc kubenswrapper[4846]: I1005 07:10:31.541366 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8fa2fcc5-ba88-4a8c-88ec-a8d1882b8c7b-kube-api-access-2k2x4" (OuterVolumeSpecName: "kube-api-access-2k2x4") pod "8fa2fcc5-ba88-4a8c-88ec-a8d1882b8c7b" (UID: "8fa2fcc5-ba88-4a8c-88ec-a8d1882b8c7b"). InnerVolumeSpecName "kube-api-access-2k2x4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:10:31 crc kubenswrapper[4846]: I1005 07:10:31.542289 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8fa2fcc5-ba88-4a8c-88ec-a8d1882b8c7b-scripts" (OuterVolumeSpecName: "scripts") pod "8fa2fcc5-ba88-4a8c-88ec-a8d1882b8c7b" (UID: "8fa2fcc5-ba88-4a8c-88ec-a8d1882b8c7b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:10:31 crc kubenswrapper[4846]: I1005 07:10:31.569299 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8fa2fcc5-ba88-4a8c-88ec-a8d1882b8c7b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8fa2fcc5-ba88-4a8c-88ec-a8d1882b8c7b" (UID: "8fa2fcc5-ba88-4a8c-88ec-a8d1882b8c7b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:10:31 crc kubenswrapper[4846]: I1005 07:10:31.588597 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8fa2fcc5-ba88-4a8c-88ec-a8d1882b8c7b-config-data" (OuterVolumeSpecName: "config-data") pod "8fa2fcc5-ba88-4a8c-88ec-a8d1882b8c7b" (UID: "8fa2fcc5-ba88-4a8c-88ec-a8d1882b8c7b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:10:31 crc kubenswrapper[4846]: I1005 07:10:31.636869 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2k2x4\" (UniqueName: \"kubernetes.io/projected/8fa2fcc5-ba88-4a8c-88ec-a8d1882b8c7b-kube-api-access-2k2x4\") on node \"crc\" DevicePath \"\"" Oct 05 07:10:31 crc kubenswrapper[4846]: I1005 07:10:31.637298 4846 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8fa2fcc5-ba88-4a8c-88ec-a8d1882b8c7b-scripts\") on node \"crc\" DevicePath \"\"" Oct 05 07:10:31 crc kubenswrapper[4846]: I1005 07:10:31.637339 4846 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8fa2fcc5-ba88-4a8c-88ec-a8d1882b8c7b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 07:10:31 crc kubenswrapper[4846]: I1005 07:10:31.637369 4846 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8fa2fcc5-ba88-4a8c-88ec-a8d1882b8c7b-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 07:10:32 crc kubenswrapper[4846]: I1005 07:10:32.030593 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-t7fdh" event={"ID":"8fa2fcc5-ba88-4a8c-88ec-a8d1882b8c7b","Type":"ContainerDied","Data":"42915f74331bbfe756ff193d37f0ca9231e63a3cd41f0ab7fa3e7af088da7b2c"} Oct 05 07:10:32 crc kubenswrapper[4846]: I1005 07:10:32.030982 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="42915f74331bbfe756ff193d37f0ca9231e63a3cd41f0ab7fa3e7af088da7b2c" Oct 05 07:10:32 crc kubenswrapper[4846]: I1005 07:10:32.030664 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-t7fdh" Oct 05 07:10:32 crc kubenswrapper[4846]: I1005 07:10:32.226075 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 05 07:10:32 crc kubenswrapper[4846]: I1005 07:10:32.226468 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="65fd80a1-d4ce-42f9-b1cf-e379333dfdba" containerName="nova-api-log" containerID="cri-o://9a4be5b54dc458b481c513b652f2f0da4ceaaa036f7b850ef512811ccd3355ff" gracePeriod=30 Oct 05 07:10:32 crc kubenswrapper[4846]: I1005 07:10:32.227007 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="65fd80a1-d4ce-42f9-b1cf-e379333dfdba" containerName="nova-api-api" containerID="cri-o://0bf6cd09e5de807030743657cdd80e0b641715af3513167ff53800b7f6c565be" gracePeriod=30 Oct 05 07:10:32 crc kubenswrapper[4846]: I1005 07:10:32.242591 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 05 07:10:32 crc kubenswrapper[4846]: I1005 07:10:32.242871 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="ce94efea-dc0a-4d7a-9340-9217ead55549" containerName="nova-scheduler-scheduler" containerID="cri-o://f8ef556e3cb59bcf872a739907d82bb2061305bc039de4bbd142d1bef463ec56" gracePeriod=30 Oct 05 07:10:32 crc kubenswrapper[4846]: I1005 07:10:32.267187 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 05 07:10:32 crc kubenswrapper[4846]: I1005 07:10:32.267475 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="fe9fdc00-614b-4207-80f5-05a634a497a7" containerName="nova-metadata-log" containerID="cri-o://a7ae54073f10d9d332e2d936a42d58cf1af318a6d1f5ced07d8420dcfa60102c" gracePeriod=30 Oct 05 07:10:32 crc kubenswrapper[4846]: I1005 07:10:32.267651 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="fe9fdc00-614b-4207-80f5-05a634a497a7" containerName="nova-metadata-metadata" containerID="cri-o://6000ba86c9c9f2027bd31401f3a69a05f4e695ece27d93cfabaa5898532368be" gracePeriod=30 Oct 05 07:10:33 crc kubenswrapper[4846]: I1005 07:10:33.044163 4846 generic.go:334] "Generic (PLEG): container finished" podID="65fd80a1-d4ce-42f9-b1cf-e379333dfdba" containerID="9a4be5b54dc458b481c513b652f2f0da4ceaaa036f7b850ef512811ccd3355ff" exitCode=143 Oct 05 07:10:33 crc kubenswrapper[4846]: I1005 07:10:33.044496 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"65fd80a1-d4ce-42f9-b1cf-e379333dfdba","Type":"ContainerDied","Data":"9a4be5b54dc458b481c513b652f2f0da4ceaaa036f7b850ef512811ccd3355ff"} Oct 05 07:10:33 crc kubenswrapper[4846]: I1005 07:10:33.050392 4846 generic.go:334] "Generic (PLEG): container finished" podID="fe9fdc00-614b-4207-80f5-05a634a497a7" containerID="a7ae54073f10d9d332e2d936a42d58cf1af318a6d1f5ced07d8420dcfa60102c" exitCode=143 Oct 05 07:10:33 crc kubenswrapper[4846]: I1005 07:10:33.050484 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"fe9fdc00-614b-4207-80f5-05a634a497a7","Type":"ContainerDied","Data":"a7ae54073f10d9d332e2d936a42d58cf1af318a6d1f5ced07d8420dcfa60102c"} Oct 05 07:10:35 crc kubenswrapper[4846]: I1005 07:10:35.403464 4846 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="fe9fdc00-614b-4207-80f5-05a634a497a7" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.197:8775/\": read tcp 10.217.0.2:50732->10.217.0.197:8775: read: connection reset by peer" Oct 05 07:10:35 crc kubenswrapper[4846]: I1005 07:10:35.404687 4846 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="fe9fdc00-614b-4207-80f5-05a634a497a7" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.197:8775/\": read tcp 10.217.0.2:50730->10.217.0.197:8775: read: connection reset by peer" Oct 05 07:10:36 crc kubenswrapper[4846]: I1005 07:10:36.049707 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 05 07:10:36 crc kubenswrapper[4846]: I1005 07:10:36.101451 4846 generic.go:334] "Generic (PLEG): container finished" podID="fe9fdc00-614b-4207-80f5-05a634a497a7" containerID="6000ba86c9c9f2027bd31401f3a69a05f4e695ece27d93cfabaa5898532368be" exitCode=0 Oct 05 07:10:36 crc kubenswrapper[4846]: I1005 07:10:36.101528 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 05 07:10:36 crc kubenswrapper[4846]: I1005 07:10:36.101604 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"fe9fdc00-614b-4207-80f5-05a634a497a7","Type":"ContainerDied","Data":"6000ba86c9c9f2027bd31401f3a69a05f4e695ece27d93cfabaa5898532368be"} Oct 05 07:10:36 crc kubenswrapper[4846]: I1005 07:10:36.101647 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"fe9fdc00-614b-4207-80f5-05a634a497a7","Type":"ContainerDied","Data":"fc1182f4fca3742cd41a53703e689fed3143be3c9136c4cd44ab6b84054af6b1"} Oct 05 07:10:36 crc kubenswrapper[4846]: I1005 07:10:36.101673 4846 scope.go:117] "RemoveContainer" containerID="6000ba86c9c9f2027bd31401f3a69a05f4e695ece27d93cfabaa5898532368be" Oct 05 07:10:36 crc kubenswrapper[4846]: I1005 07:10:36.106315 4846 generic.go:334] "Generic (PLEG): container finished" podID="ce94efea-dc0a-4d7a-9340-9217ead55549" containerID="f8ef556e3cb59bcf872a739907d82bb2061305bc039de4bbd142d1bef463ec56" exitCode=0 Oct 05 07:10:36 crc kubenswrapper[4846]: I1005 07:10:36.106605 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"ce94efea-dc0a-4d7a-9340-9217ead55549","Type":"ContainerDied","Data":"f8ef556e3cb59bcf872a739907d82bb2061305bc039de4bbd142d1bef463ec56"} Oct 05 07:10:36 crc kubenswrapper[4846]: I1005 07:10:36.146233 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fe9fdc00-614b-4207-80f5-05a634a497a7-config-data\") pod \"fe9fdc00-614b-4207-80f5-05a634a497a7\" (UID: \"fe9fdc00-614b-4207-80f5-05a634a497a7\") " Oct 05 07:10:36 crc kubenswrapper[4846]: I1005 07:10:36.146294 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/fe9fdc00-614b-4207-80f5-05a634a497a7-nova-metadata-tls-certs\") pod \"fe9fdc00-614b-4207-80f5-05a634a497a7\" (UID: \"fe9fdc00-614b-4207-80f5-05a634a497a7\") " Oct 05 07:10:36 crc kubenswrapper[4846]: I1005 07:10:36.146404 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fe9fdc00-614b-4207-80f5-05a634a497a7-combined-ca-bundle\") pod \"fe9fdc00-614b-4207-80f5-05a634a497a7\" (UID: \"fe9fdc00-614b-4207-80f5-05a634a497a7\") " Oct 05 07:10:36 crc kubenswrapper[4846]: I1005 07:10:36.146567 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sp2jf\" (UniqueName: \"kubernetes.io/projected/fe9fdc00-614b-4207-80f5-05a634a497a7-kube-api-access-sp2jf\") pod \"fe9fdc00-614b-4207-80f5-05a634a497a7\" (UID: \"fe9fdc00-614b-4207-80f5-05a634a497a7\") " Oct 05 07:10:36 crc kubenswrapper[4846]: I1005 07:10:36.146594 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fe9fdc00-614b-4207-80f5-05a634a497a7-logs\") pod \"fe9fdc00-614b-4207-80f5-05a634a497a7\" (UID: \"fe9fdc00-614b-4207-80f5-05a634a497a7\") " Oct 05 07:10:36 crc kubenswrapper[4846]: I1005 07:10:36.146875 4846 scope.go:117] "RemoveContainer" containerID="a7ae54073f10d9d332e2d936a42d58cf1af318a6d1f5ced07d8420dcfa60102c" Oct 05 07:10:36 crc kubenswrapper[4846]: I1005 07:10:36.148257 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fe9fdc00-614b-4207-80f5-05a634a497a7-logs" (OuterVolumeSpecName: "logs") pod "fe9fdc00-614b-4207-80f5-05a634a497a7" (UID: "fe9fdc00-614b-4207-80f5-05a634a497a7"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 07:10:36 crc kubenswrapper[4846]: I1005 07:10:36.166821 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fe9fdc00-614b-4207-80f5-05a634a497a7-kube-api-access-sp2jf" (OuterVolumeSpecName: "kube-api-access-sp2jf") pod "fe9fdc00-614b-4207-80f5-05a634a497a7" (UID: "fe9fdc00-614b-4207-80f5-05a634a497a7"). InnerVolumeSpecName "kube-api-access-sp2jf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:10:36 crc kubenswrapper[4846]: I1005 07:10:36.204492 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fe9fdc00-614b-4207-80f5-05a634a497a7-config-data" (OuterVolumeSpecName: "config-data") pod "fe9fdc00-614b-4207-80f5-05a634a497a7" (UID: "fe9fdc00-614b-4207-80f5-05a634a497a7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:10:36 crc kubenswrapper[4846]: I1005 07:10:36.227190 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fe9fdc00-614b-4207-80f5-05a634a497a7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "fe9fdc00-614b-4207-80f5-05a634a497a7" (UID: "fe9fdc00-614b-4207-80f5-05a634a497a7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:10:36 crc kubenswrapper[4846]: I1005 07:10:36.244429 4846 scope.go:117] "RemoveContainer" containerID="6000ba86c9c9f2027bd31401f3a69a05f4e695ece27d93cfabaa5898532368be" Oct 05 07:10:36 crc kubenswrapper[4846]: E1005 07:10:36.244850 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6000ba86c9c9f2027bd31401f3a69a05f4e695ece27d93cfabaa5898532368be\": container with ID starting with 6000ba86c9c9f2027bd31401f3a69a05f4e695ece27d93cfabaa5898532368be not found: ID does not exist" containerID="6000ba86c9c9f2027bd31401f3a69a05f4e695ece27d93cfabaa5898532368be" Oct 05 07:10:36 crc kubenswrapper[4846]: I1005 07:10:36.244904 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6000ba86c9c9f2027bd31401f3a69a05f4e695ece27d93cfabaa5898532368be"} err="failed to get container status \"6000ba86c9c9f2027bd31401f3a69a05f4e695ece27d93cfabaa5898532368be\": rpc error: code = NotFound desc = could not find container \"6000ba86c9c9f2027bd31401f3a69a05f4e695ece27d93cfabaa5898532368be\": container with ID starting with 6000ba86c9c9f2027bd31401f3a69a05f4e695ece27d93cfabaa5898532368be not found: ID does not exist" Oct 05 07:10:36 crc kubenswrapper[4846]: I1005 07:10:36.244936 4846 scope.go:117] "RemoveContainer" containerID="a7ae54073f10d9d332e2d936a42d58cf1af318a6d1f5ced07d8420dcfa60102c" Oct 05 07:10:36 crc kubenswrapper[4846]: E1005 07:10:36.245218 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a7ae54073f10d9d332e2d936a42d58cf1af318a6d1f5ced07d8420dcfa60102c\": container with ID starting with a7ae54073f10d9d332e2d936a42d58cf1af318a6d1f5ced07d8420dcfa60102c not found: ID does not exist" containerID="a7ae54073f10d9d332e2d936a42d58cf1af318a6d1f5ced07d8420dcfa60102c" Oct 05 07:10:36 crc kubenswrapper[4846]: I1005 07:10:36.245237 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a7ae54073f10d9d332e2d936a42d58cf1af318a6d1f5ced07d8420dcfa60102c"} err="failed to get container status \"a7ae54073f10d9d332e2d936a42d58cf1af318a6d1f5ced07d8420dcfa60102c\": rpc error: code = NotFound desc = could not find container \"a7ae54073f10d9d332e2d936a42d58cf1af318a6d1f5ced07d8420dcfa60102c\": container with ID starting with a7ae54073f10d9d332e2d936a42d58cf1af318a6d1f5ced07d8420dcfa60102c not found: ID does not exist" Oct 05 07:10:36 crc kubenswrapper[4846]: I1005 07:10:36.256311 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sp2jf\" (UniqueName: \"kubernetes.io/projected/fe9fdc00-614b-4207-80f5-05a634a497a7-kube-api-access-sp2jf\") on node \"crc\" DevicePath \"\"" Oct 05 07:10:36 crc kubenswrapper[4846]: I1005 07:10:36.257053 4846 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fe9fdc00-614b-4207-80f5-05a634a497a7-logs\") on node \"crc\" DevicePath \"\"" Oct 05 07:10:36 crc kubenswrapper[4846]: I1005 07:10:36.257093 4846 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fe9fdc00-614b-4207-80f5-05a634a497a7-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 07:10:36 crc kubenswrapper[4846]: I1005 07:10:36.257110 4846 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fe9fdc00-614b-4207-80f5-05a634a497a7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 07:10:36 crc kubenswrapper[4846]: I1005 07:10:36.262478 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fe9fdc00-614b-4207-80f5-05a634a497a7-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "fe9fdc00-614b-4207-80f5-05a634a497a7" (UID: "fe9fdc00-614b-4207-80f5-05a634a497a7"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:10:36 crc kubenswrapper[4846]: I1005 07:10:36.297808 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 05 07:10:36 crc kubenswrapper[4846]: I1005 07:10:36.358380 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ce94efea-dc0a-4d7a-9340-9217ead55549-config-data\") pod \"ce94efea-dc0a-4d7a-9340-9217ead55549\" (UID: \"ce94efea-dc0a-4d7a-9340-9217ead55549\") " Oct 05 07:10:36 crc kubenswrapper[4846]: I1005 07:10:36.358484 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ce94efea-dc0a-4d7a-9340-9217ead55549-combined-ca-bundle\") pod \"ce94efea-dc0a-4d7a-9340-9217ead55549\" (UID: \"ce94efea-dc0a-4d7a-9340-9217ead55549\") " Oct 05 07:10:36 crc kubenswrapper[4846]: I1005 07:10:36.358567 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-86fq9\" (UniqueName: \"kubernetes.io/projected/ce94efea-dc0a-4d7a-9340-9217ead55549-kube-api-access-86fq9\") pod \"ce94efea-dc0a-4d7a-9340-9217ead55549\" (UID: \"ce94efea-dc0a-4d7a-9340-9217ead55549\") " Oct 05 07:10:36 crc kubenswrapper[4846]: I1005 07:10:36.359114 4846 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/fe9fdc00-614b-4207-80f5-05a634a497a7-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 05 07:10:36 crc kubenswrapper[4846]: I1005 07:10:36.363345 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ce94efea-dc0a-4d7a-9340-9217ead55549-kube-api-access-86fq9" (OuterVolumeSpecName: "kube-api-access-86fq9") pod "ce94efea-dc0a-4d7a-9340-9217ead55549" (UID: "ce94efea-dc0a-4d7a-9340-9217ead55549"). InnerVolumeSpecName "kube-api-access-86fq9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:10:36 crc kubenswrapper[4846]: I1005 07:10:36.392613 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ce94efea-dc0a-4d7a-9340-9217ead55549-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ce94efea-dc0a-4d7a-9340-9217ead55549" (UID: "ce94efea-dc0a-4d7a-9340-9217ead55549"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:10:36 crc kubenswrapper[4846]: I1005 07:10:36.397572 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ce94efea-dc0a-4d7a-9340-9217ead55549-config-data" (OuterVolumeSpecName: "config-data") pod "ce94efea-dc0a-4d7a-9340-9217ead55549" (UID: "ce94efea-dc0a-4d7a-9340-9217ead55549"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:10:36 crc kubenswrapper[4846]: I1005 07:10:36.450662 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 05 07:10:36 crc kubenswrapper[4846]: I1005 07:10:36.461615 4846 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ce94efea-dc0a-4d7a-9340-9217ead55549-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 07:10:36 crc kubenswrapper[4846]: I1005 07:10:36.461664 4846 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ce94efea-dc0a-4d7a-9340-9217ead55549-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 07:10:36 crc kubenswrapper[4846]: I1005 07:10:36.461683 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-86fq9\" (UniqueName: \"kubernetes.io/projected/ce94efea-dc0a-4d7a-9340-9217ead55549-kube-api-access-86fq9\") on node \"crc\" DevicePath \"\"" Oct 05 07:10:36 crc kubenswrapper[4846]: I1005 07:10:36.466012 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Oct 05 07:10:36 crc kubenswrapper[4846]: I1005 07:10:36.479520 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 05 07:10:36 crc kubenswrapper[4846]: E1005 07:10:36.480172 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="68a854c2-afda-4dbb-9ac7-06eec2c98f14" containerName="init" Oct 05 07:10:36 crc kubenswrapper[4846]: I1005 07:10:36.480213 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="68a854c2-afda-4dbb-9ac7-06eec2c98f14" containerName="init" Oct 05 07:10:36 crc kubenswrapper[4846]: E1005 07:10:36.480224 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="68a854c2-afda-4dbb-9ac7-06eec2c98f14" containerName="dnsmasq-dns" Oct 05 07:10:36 crc kubenswrapper[4846]: I1005 07:10:36.480234 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="68a854c2-afda-4dbb-9ac7-06eec2c98f14" containerName="dnsmasq-dns" Oct 05 07:10:36 crc kubenswrapper[4846]: E1005 07:10:36.480261 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ce94efea-dc0a-4d7a-9340-9217ead55549" containerName="nova-scheduler-scheduler" Oct 05 07:10:36 crc kubenswrapper[4846]: I1005 07:10:36.480283 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="ce94efea-dc0a-4d7a-9340-9217ead55549" containerName="nova-scheduler-scheduler" Oct 05 07:10:36 crc kubenswrapper[4846]: E1005 07:10:36.480301 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fe9fdc00-614b-4207-80f5-05a634a497a7" containerName="nova-metadata-metadata" Oct 05 07:10:36 crc kubenswrapper[4846]: I1005 07:10:36.480311 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="fe9fdc00-614b-4207-80f5-05a634a497a7" containerName="nova-metadata-metadata" Oct 05 07:10:36 crc kubenswrapper[4846]: E1005 07:10:36.480346 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fe9fdc00-614b-4207-80f5-05a634a497a7" containerName="nova-metadata-log" Oct 05 07:10:36 crc kubenswrapper[4846]: I1005 07:10:36.480355 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="fe9fdc00-614b-4207-80f5-05a634a497a7" containerName="nova-metadata-log" Oct 05 07:10:36 crc kubenswrapper[4846]: E1005 07:10:36.480383 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8fa2fcc5-ba88-4a8c-88ec-a8d1882b8c7b" containerName="nova-manage" Oct 05 07:10:36 crc kubenswrapper[4846]: I1005 07:10:36.480392 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="8fa2fcc5-ba88-4a8c-88ec-a8d1882b8c7b" containerName="nova-manage" Oct 05 07:10:36 crc kubenswrapper[4846]: I1005 07:10:36.480634 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="fe9fdc00-614b-4207-80f5-05a634a497a7" containerName="nova-metadata-log" Oct 05 07:10:36 crc kubenswrapper[4846]: I1005 07:10:36.480653 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="fe9fdc00-614b-4207-80f5-05a634a497a7" containerName="nova-metadata-metadata" Oct 05 07:10:36 crc kubenswrapper[4846]: I1005 07:10:36.480670 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="8fa2fcc5-ba88-4a8c-88ec-a8d1882b8c7b" containerName="nova-manage" Oct 05 07:10:36 crc kubenswrapper[4846]: I1005 07:10:36.480688 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="68a854c2-afda-4dbb-9ac7-06eec2c98f14" containerName="dnsmasq-dns" Oct 05 07:10:36 crc kubenswrapper[4846]: I1005 07:10:36.480711 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="ce94efea-dc0a-4d7a-9340-9217ead55549" containerName="nova-scheduler-scheduler" Oct 05 07:10:36 crc kubenswrapper[4846]: I1005 07:10:36.482144 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 05 07:10:36 crc kubenswrapper[4846]: I1005 07:10:36.486374 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Oct 05 07:10:36 crc kubenswrapper[4846]: I1005 07:10:36.486594 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 05 07:10:36 crc kubenswrapper[4846]: I1005 07:10:36.513813 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fe9fdc00-614b-4207-80f5-05a634a497a7" path="/var/lib/kubelet/pods/fe9fdc00-614b-4207-80f5-05a634a497a7/volumes" Oct 05 07:10:36 crc kubenswrapper[4846]: I1005 07:10:36.514786 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 05 07:10:36 crc kubenswrapper[4846]: I1005 07:10:36.563268 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e6e86b1c-87ac-4594-aeba-943947065374-config-data\") pod \"nova-metadata-0\" (UID: \"e6e86b1c-87ac-4594-aeba-943947065374\") " pod="openstack/nova-metadata-0" Oct 05 07:10:36 crc kubenswrapper[4846]: I1005 07:10:36.563354 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/e6e86b1c-87ac-4594-aeba-943947065374-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"e6e86b1c-87ac-4594-aeba-943947065374\") " pod="openstack/nova-metadata-0" Oct 05 07:10:36 crc kubenswrapper[4846]: I1005 07:10:36.563463 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dbvks\" (UniqueName: \"kubernetes.io/projected/e6e86b1c-87ac-4594-aeba-943947065374-kube-api-access-dbvks\") pod \"nova-metadata-0\" (UID: \"e6e86b1c-87ac-4594-aeba-943947065374\") " pod="openstack/nova-metadata-0" Oct 05 07:10:36 crc kubenswrapper[4846]: I1005 07:10:36.563500 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e6e86b1c-87ac-4594-aeba-943947065374-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"e6e86b1c-87ac-4594-aeba-943947065374\") " pod="openstack/nova-metadata-0" Oct 05 07:10:36 crc kubenswrapper[4846]: I1005 07:10:36.563539 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e6e86b1c-87ac-4594-aeba-943947065374-logs\") pod \"nova-metadata-0\" (UID: \"e6e86b1c-87ac-4594-aeba-943947065374\") " pod="openstack/nova-metadata-0" Oct 05 07:10:36 crc kubenswrapper[4846]: I1005 07:10:36.666022 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e6e86b1c-87ac-4594-aeba-943947065374-config-data\") pod \"nova-metadata-0\" (UID: \"e6e86b1c-87ac-4594-aeba-943947065374\") " pod="openstack/nova-metadata-0" Oct 05 07:10:36 crc kubenswrapper[4846]: I1005 07:10:36.666636 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/e6e86b1c-87ac-4594-aeba-943947065374-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"e6e86b1c-87ac-4594-aeba-943947065374\") " pod="openstack/nova-metadata-0" Oct 05 07:10:36 crc kubenswrapper[4846]: I1005 07:10:36.666753 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dbvks\" (UniqueName: \"kubernetes.io/projected/e6e86b1c-87ac-4594-aeba-943947065374-kube-api-access-dbvks\") pod \"nova-metadata-0\" (UID: \"e6e86b1c-87ac-4594-aeba-943947065374\") " pod="openstack/nova-metadata-0" Oct 05 07:10:36 crc kubenswrapper[4846]: I1005 07:10:36.666823 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e6e86b1c-87ac-4594-aeba-943947065374-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"e6e86b1c-87ac-4594-aeba-943947065374\") " pod="openstack/nova-metadata-0" Oct 05 07:10:36 crc kubenswrapper[4846]: I1005 07:10:36.666899 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e6e86b1c-87ac-4594-aeba-943947065374-logs\") pod \"nova-metadata-0\" (UID: \"e6e86b1c-87ac-4594-aeba-943947065374\") " pod="openstack/nova-metadata-0" Oct 05 07:10:36 crc kubenswrapper[4846]: I1005 07:10:36.667650 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e6e86b1c-87ac-4594-aeba-943947065374-logs\") pod \"nova-metadata-0\" (UID: \"e6e86b1c-87ac-4594-aeba-943947065374\") " pod="openstack/nova-metadata-0" Oct 05 07:10:36 crc kubenswrapper[4846]: I1005 07:10:36.673352 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/e6e86b1c-87ac-4594-aeba-943947065374-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"e6e86b1c-87ac-4594-aeba-943947065374\") " pod="openstack/nova-metadata-0" Oct 05 07:10:36 crc kubenswrapper[4846]: I1005 07:10:36.684544 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e6e86b1c-87ac-4594-aeba-943947065374-config-data\") pod \"nova-metadata-0\" (UID: \"e6e86b1c-87ac-4594-aeba-943947065374\") " pod="openstack/nova-metadata-0" Oct 05 07:10:36 crc kubenswrapper[4846]: I1005 07:10:36.684851 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e6e86b1c-87ac-4594-aeba-943947065374-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"e6e86b1c-87ac-4594-aeba-943947065374\") " pod="openstack/nova-metadata-0" Oct 05 07:10:36 crc kubenswrapper[4846]: I1005 07:10:36.688092 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dbvks\" (UniqueName: \"kubernetes.io/projected/e6e86b1c-87ac-4594-aeba-943947065374-kube-api-access-dbvks\") pod \"nova-metadata-0\" (UID: \"e6e86b1c-87ac-4594-aeba-943947065374\") " pod="openstack/nova-metadata-0" Oct 05 07:10:36 crc kubenswrapper[4846]: I1005 07:10:36.896780 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 05 07:10:37 crc kubenswrapper[4846]: I1005 07:10:37.068665 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 05 07:10:37 crc kubenswrapper[4846]: I1005 07:10:37.118964 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"ce94efea-dc0a-4d7a-9340-9217ead55549","Type":"ContainerDied","Data":"ecde9fcad845c81d90b7d0ebb4aa5515cb50bc02a9ba8b568a05fa84ec21b854"} Oct 05 07:10:37 crc kubenswrapper[4846]: I1005 07:10:37.119015 4846 scope.go:117] "RemoveContainer" containerID="f8ef556e3cb59bcf872a739907d82bb2061305bc039de4bbd142d1bef463ec56" Oct 05 07:10:37 crc kubenswrapper[4846]: I1005 07:10:37.119083 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 05 07:10:37 crc kubenswrapper[4846]: I1005 07:10:37.121681 4846 generic.go:334] "Generic (PLEG): container finished" podID="65fd80a1-d4ce-42f9-b1cf-e379333dfdba" containerID="0bf6cd09e5de807030743657cdd80e0b641715af3513167ff53800b7f6c565be" exitCode=0 Oct 05 07:10:37 crc kubenswrapper[4846]: I1005 07:10:37.121752 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"65fd80a1-d4ce-42f9-b1cf-e379333dfdba","Type":"ContainerDied","Data":"0bf6cd09e5de807030743657cdd80e0b641715af3513167ff53800b7f6c565be"} Oct 05 07:10:37 crc kubenswrapper[4846]: I1005 07:10:37.121786 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"65fd80a1-d4ce-42f9-b1cf-e379333dfdba","Type":"ContainerDied","Data":"12adb4adca58353410b4aba7e75c9560dd580019484783b4e35d98acb7df5496"} Oct 05 07:10:37 crc kubenswrapper[4846]: I1005 07:10:37.121842 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 05 07:10:37 crc kubenswrapper[4846]: I1005 07:10:37.152208 4846 scope.go:117] "RemoveContainer" containerID="0bf6cd09e5de807030743657cdd80e0b641715af3513167ff53800b7f6c565be" Oct 05 07:10:37 crc kubenswrapper[4846]: I1005 07:10:37.152926 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 05 07:10:37 crc kubenswrapper[4846]: I1005 07:10:37.166899 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Oct 05 07:10:37 crc kubenswrapper[4846]: I1005 07:10:37.177170 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/65fd80a1-d4ce-42f9-b1cf-e379333dfdba-combined-ca-bundle\") pod \"65fd80a1-d4ce-42f9-b1cf-e379333dfdba\" (UID: \"65fd80a1-d4ce-42f9-b1cf-e379333dfdba\") " Oct 05 07:10:37 crc kubenswrapper[4846]: I1005 07:10:37.177225 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/65fd80a1-d4ce-42f9-b1cf-e379333dfdba-config-data\") pod \"65fd80a1-d4ce-42f9-b1cf-e379333dfdba\" (UID: \"65fd80a1-d4ce-42f9-b1cf-e379333dfdba\") " Oct 05 07:10:37 crc kubenswrapper[4846]: I1005 07:10:37.177322 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/65fd80a1-d4ce-42f9-b1cf-e379333dfdba-internal-tls-certs\") pod \"65fd80a1-d4ce-42f9-b1cf-e379333dfdba\" (UID: \"65fd80a1-d4ce-42f9-b1cf-e379333dfdba\") " Oct 05 07:10:37 crc kubenswrapper[4846]: I1005 07:10:37.177390 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jv99j\" (UniqueName: \"kubernetes.io/projected/65fd80a1-d4ce-42f9-b1cf-e379333dfdba-kube-api-access-jv99j\") pod \"65fd80a1-d4ce-42f9-b1cf-e379333dfdba\" (UID: \"65fd80a1-d4ce-42f9-b1cf-e379333dfdba\") " Oct 05 07:10:37 crc kubenswrapper[4846]: I1005 07:10:37.177441 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/65fd80a1-d4ce-42f9-b1cf-e379333dfdba-public-tls-certs\") pod \"65fd80a1-d4ce-42f9-b1cf-e379333dfdba\" (UID: \"65fd80a1-d4ce-42f9-b1cf-e379333dfdba\") " Oct 05 07:10:37 crc kubenswrapper[4846]: I1005 07:10:37.177530 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/65fd80a1-d4ce-42f9-b1cf-e379333dfdba-logs\") pod \"65fd80a1-d4ce-42f9-b1cf-e379333dfdba\" (UID: \"65fd80a1-d4ce-42f9-b1cf-e379333dfdba\") " Oct 05 07:10:37 crc kubenswrapper[4846]: I1005 07:10:37.178592 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/65fd80a1-d4ce-42f9-b1cf-e379333dfdba-logs" (OuterVolumeSpecName: "logs") pod "65fd80a1-d4ce-42f9-b1cf-e379333dfdba" (UID: "65fd80a1-d4ce-42f9-b1cf-e379333dfdba"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 07:10:37 crc kubenswrapper[4846]: I1005 07:10:37.181036 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Oct 05 07:10:37 crc kubenswrapper[4846]: E1005 07:10:37.181586 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="65fd80a1-d4ce-42f9-b1cf-e379333dfdba" containerName="nova-api-log" Oct 05 07:10:37 crc kubenswrapper[4846]: I1005 07:10:37.181608 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="65fd80a1-d4ce-42f9-b1cf-e379333dfdba" containerName="nova-api-log" Oct 05 07:10:37 crc kubenswrapper[4846]: E1005 07:10:37.181672 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="65fd80a1-d4ce-42f9-b1cf-e379333dfdba" containerName="nova-api-api" Oct 05 07:10:37 crc kubenswrapper[4846]: I1005 07:10:37.181685 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="65fd80a1-d4ce-42f9-b1cf-e379333dfdba" containerName="nova-api-api" Oct 05 07:10:37 crc kubenswrapper[4846]: I1005 07:10:37.181678 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/65fd80a1-d4ce-42f9-b1cf-e379333dfdba-kube-api-access-jv99j" (OuterVolumeSpecName: "kube-api-access-jv99j") pod "65fd80a1-d4ce-42f9-b1cf-e379333dfdba" (UID: "65fd80a1-d4ce-42f9-b1cf-e379333dfdba"). InnerVolumeSpecName "kube-api-access-jv99j". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:10:37 crc kubenswrapper[4846]: I1005 07:10:37.182109 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="65fd80a1-d4ce-42f9-b1cf-e379333dfdba" containerName="nova-api-log" Oct 05 07:10:37 crc kubenswrapper[4846]: I1005 07:10:37.182132 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="65fd80a1-d4ce-42f9-b1cf-e379333dfdba" containerName="nova-api-api" Oct 05 07:10:37 crc kubenswrapper[4846]: I1005 07:10:37.183101 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 05 07:10:37 crc kubenswrapper[4846]: I1005 07:10:37.186658 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Oct 05 07:10:37 crc kubenswrapper[4846]: I1005 07:10:37.193001 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 05 07:10:37 crc kubenswrapper[4846]: I1005 07:10:37.195317 4846 scope.go:117] "RemoveContainer" containerID="9a4be5b54dc458b481c513b652f2f0da4ceaaa036f7b850ef512811ccd3355ff" Oct 05 07:10:37 crc kubenswrapper[4846]: I1005 07:10:37.228743 4846 scope.go:117] "RemoveContainer" containerID="0bf6cd09e5de807030743657cdd80e0b641715af3513167ff53800b7f6c565be" Oct 05 07:10:37 crc kubenswrapper[4846]: E1005 07:10:37.229657 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0bf6cd09e5de807030743657cdd80e0b641715af3513167ff53800b7f6c565be\": container with ID starting with 0bf6cd09e5de807030743657cdd80e0b641715af3513167ff53800b7f6c565be not found: ID does not exist" containerID="0bf6cd09e5de807030743657cdd80e0b641715af3513167ff53800b7f6c565be" Oct 05 07:10:37 crc kubenswrapper[4846]: I1005 07:10:37.229713 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0bf6cd09e5de807030743657cdd80e0b641715af3513167ff53800b7f6c565be"} err="failed to get container status \"0bf6cd09e5de807030743657cdd80e0b641715af3513167ff53800b7f6c565be\": rpc error: code = NotFound desc = could not find container \"0bf6cd09e5de807030743657cdd80e0b641715af3513167ff53800b7f6c565be\": container with ID starting with 0bf6cd09e5de807030743657cdd80e0b641715af3513167ff53800b7f6c565be not found: ID does not exist" Oct 05 07:10:37 crc kubenswrapper[4846]: I1005 07:10:37.229744 4846 scope.go:117] "RemoveContainer" containerID="9a4be5b54dc458b481c513b652f2f0da4ceaaa036f7b850ef512811ccd3355ff" Oct 05 07:10:37 crc kubenswrapper[4846]: E1005 07:10:37.230221 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9a4be5b54dc458b481c513b652f2f0da4ceaaa036f7b850ef512811ccd3355ff\": container with ID starting with 9a4be5b54dc458b481c513b652f2f0da4ceaaa036f7b850ef512811ccd3355ff not found: ID does not exist" containerID="9a4be5b54dc458b481c513b652f2f0da4ceaaa036f7b850ef512811ccd3355ff" Oct 05 07:10:37 crc kubenswrapper[4846]: I1005 07:10:37.230256 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9a4be5b54dc458b481c513b652f2f0da4ceaaa036f7b850ef512811ccd3355ff"} err="failed to get container status \"9a4be5b54dc458b481c513b652f2f0da4ceaaa036f7b850ef512811ccd3355ff\": rpc error: code = NotFound desc = could not find container \"9a4be5b54dc458b481c513b652f2f0da4ceaaa036f7b850ef512811ccd3355ff\": container with ID starting with 9a4be5b54dc458b481c513b652f2f0da4ceaaa036f7b850ef512811ccd3355ff not found: ID does not exist" Oct 05 07:10:37 crc kubenswrapper[4846]: I1005 07:10:37.242031 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/65fd80a1-d4ce-42f9-b1cf-e379333dfdba-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "65fd80a1-d4ce-42f9-b1cf-e379333dfdba" (UID: "65fd80a1-d4ce-42f9-b1cf-e379333dfdba"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:10:37 crc kubenswrapper[4846]: I1005 07:10:37.243089 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/65fd80a1-d4ce-42f9-b1cf-e379333dfdba-config-data" (OuterVolumeSpecName: "config-data") pod "65fd80a1-d4ce-42f9-b1cf-e379333dfdba" (UID: "65fd80a1-d4ce-42f9-b1cf-e379333dfdba"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:10:37 crc kubenswrapper[4846]: I1005 07:10:37.258641 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/65fd80a1-d4ce-42f9-b1cf-e379333dfdba-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "65fd80a1-d4ce-42f9-b1cf-e379333dfdba" (UID: "65fd80a1-d4ce-42f9-b1cf-e379333dfdba"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:10:37 crc kubenswrapper[4846]: I1005 07:10:37.262294 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/65fd80a1-d4ce-42f9-b1cf-e379333dfdba-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "65fd80a1-d4ce-42f9-b1cf-e379333dfdba" (UID: "65fd80a1-d4ce-42f9-b1cf-e379333dfdba"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:10:37 crc kubenswrapper[4846]: I1005 07:10:37.280454 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b7de2524-bf5c-4668-af74-7584edf72c8b-config-data\") pod \"nova-scheduler-0\" (UID: \"b7de2524-bf5c-4668-af74-7584edf72c8b\") " pod="openstack/nova-scheduler-0" Oct 05 07:10:37 crc kubenswrapper[4846]: I1005 07:10:37.280537 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4lm46\" (UniqueName: \"kubernetes.io/projected/b7de2524-bf5c-4668-af74-7584edf72c8b-kube-api-access-4lm46\") pod \"nova-scheduler-0\" (UID: \"b7de2524-bf5c-4668-af74-7584edf72c8b\") " pod="openstack/nova-scheduler-0" Oct 05 07:10:37 crc kubenswrapper[4846]: I1005 07:10:37.280618 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b7de2524-bf5c-4668-af74-7584edf72c8b-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"b7de2524-bf5c-4668-af74-7584edf72c8b\") " pod="openstack/nova-scheduler-0" Oct 05 07:10:37 crc kubenswrapper[4846]: I1005 07:10:37.280733 4846 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/65fd80a1-d4ce-42f9-b1cf-e379333dfdba-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 05 07:10:37 crc kubenswrapper[4846]: I1005 07:10:37.280747 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jv99j\" (UniqueName: \"kubernetes.io/projected/65fd80a1-d4ce-42f9-b1cf-e379333dfdba-kube-api-access-jv99j\") on node \"crc\" DevicePath \"\"" Oct 05 07:10:37 crc kubenswrapper[4846]: I1005 07:10:37.280758 4846 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/65fd80a1-d4ce-42f9-b1cf-e379333dfdba-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 05 07:10:37 crc kubenswrapper[4846]: I1005 07:10:37.280766 4846 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/65fd80a1-d4ce-42f9-b1cf-e379333dfdba-logs\") on node \"crc\" DevicePath \"\"" Oct 05 07:10:37 crc kubenswrapper[4846]: I1005 07:10:37.280776 4846 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/65fd80a1-d4ce-42f9-b1cf-e379333dfdba-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 07:10:37 crc kubenswrapper[4846]: I1005 07:10:37.280784 4846 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/65fd80a1-d4ce-42f9-b1cf-e379333dfdba-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 07:10:37 crc kubenswrapper[4846]: I1005 07:10:37.383143 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b7de2524-bf5c-4668-af74-7584edf72c8b-config-data\") pod \"nova-scheduler-0\" (UID: \"b7de2524-bf5c-4668-af74-7584edf72c8b\") " pod="openstack/nova-scheduler-0" Oct 05 07:10:37 crc kubenswrapper[4846]: I1005 07:10:37.383262 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4lm46\" (UniqueName: \"kubernetes.io/projected/b7de2524-bf5c-4668-af74-7584edf72c8b-kube-api-access-4lm46\") pod \"nova-scheduler-0\" (UID: \"b7de2524-bf5c-4668-af74-7584edf72c8b\") " pod="openstack/nova-scheduler-0" Oct 05 07:10:37 crc kubenswrapper[4846]: I1005 07:10:37.383426 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b7de2524-bf5c-4668-af74-7584edf72c8b-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"b7de2524-bf5c-4668-af74-7584edf72c8b\") " pod="openstack/nova-scheduler-0" Oct 05 07:10:37 crc kubenswrapper[4846]: I1005 07:10:37.390371 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b7de2524-bf5c-4668-af74-7584edf72c8b-config-data\") pod \"nova-scheduler-0\" (UID: \"b7de2524-bf5c-4668-af74-7584edf72c8b\") " pod="openstack/nova-scheduler-0" Oct 05 07:10:37 crc kubenswrapper[4846]: I1005 07:10:37.391037 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b7de2524-bf5c-4668-af74-7584edf72c8b-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"b7de2524-bf5c-4668-af74-7584edf72c8b\") " pod="openstack/nova-scheduler-0" Oct 05 07:10:37 crc kubenswrapper[4846]: I1005 07:10:37.406970 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4lm46\" (UniqueName: \"kubernetes.io/projected/b7de2524-bf5c-4668-af74-7584edf72c8b-kube-api-access-4lm46\") pod \"nova-scheduler-0\" (UID: \"b7de2524-bf5c-4668-af74-7584edf72c8b\") " pod="openstack/nova-scheduler-0" Oct 05 07:10:37 crc kubenswrapper[4846]: I1005 07:10:37.461876 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 05 07:10:37 crc kubenswrapper[4846]: I1005 07:10:37.477158 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Oct 05 07:10:37 crc kubenswrapper[4846]: I1005 07:10:37.494911 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 05 07:10:37 crc kubenswrapper[4846]: I1005 07:10:37.498060 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 05 07:10:37 crc kubenswrapper[4846]: I1005 07:10:37.505189 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 05 07:10:37 crc kubenswrapper[4846]: I1005 07:10:37.505398 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Oct 05 07:10:37 crc kubenswrapper[4846]: I1005 07:10:37.505541 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Oct 05 07:10:37 crc kubenswrapper[4846]: I1005 07:10:37.507778 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 05 07:10:37 crc kubenswrapper[4846]: I1005 07:10:37.525495 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 05 07:10:37 crc kubenswrapper[4846]: I1005 07:10:37.588020 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6e9e53c6-4cf2-4988-89b4-c69471f616aa-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"6e9e53c6-4cf2-4988-89b4-c69471f616aa\") " pod="openstack/nova-api-0" Oct 05 07:10:37 crc kubenswrapper[4846]: I1005 07:10:37.588086 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6e9e53c6-4cf2-4988-89b4-c69471f616aa-config-data\") pod \"nova-api-0\" (UID: \"6e9e53c6-4cf2-4988-89b4-c69471f616aa\") " pod="openstack/nova-api-0" Oct 05 07:10:37 crc kubenswrapper[4846]: I1005 07:10:37.588198 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6e9e53c6-4cf2-4988-89b4-c69471f616aa-logs\") pod \"nova-api-0\" (UID: \"6e9e53c6-4cf2-4988-89b4-c69471f616aa\") " pod="openstack/nova-api-0" Oct 05 07:10:37 crc kubenswrapper[4846]: I1005 07:10:37.588239 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6e9e53c6-4cf2-4988-89b4-c69471f616aa-internal-tls-certs\") pod \"nova-api-0\" (UID: \"6e9e53c6-4cf2-4988-89b4-c69471f616aa\") " pod="openstack/nova-api-0" Oct 05 07:10:37 crc kubenswrapper[4846]: I1005 07:10:37.588303 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mc992\" (UniqueName: \"kubernetes.io/projected/6e9e53c6-4cf2-4988-89b4-c69471f616aa-kube-api-access-mc992\") pod \"nova-api-0\" (UID: \"6e9e53c6-4cf2-4988-89b4-c69471f616aa\") " pod="openstack/nova-api-0" Oct 05 07:10:37 crc kubenswrapper[4846]: I1005 07:10:37.588415 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6e9e53c6-4cf2-4988-89b4-c69471f616aa-public-tls-certs\") pod \"nova-api-0\" (UID: \"6e9e53c6-4cf2-4988-89b4-c69471f616aa\") " pod="openstack/nova-api-0" Oct 05 07:10:37 crc kubenswrapper[4846]: I1005 07:10:37.690524 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6e9e53c6-4cf2-4988-89b4-c69471f616aa-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"6e9e53c6-4cf2-4988-89b4-c69471f616aa\") " pod="openstack/nova-api-0" Oct 05 07:10:37 crc kubenswrapper[4846]: I1005 07:10:37.690568 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6e9e53c6-4cf2-4988-89b4-c69471f616aa-config-data\") pod \"nova-api-0\" (UID: \"6e9e53c6-4cf2-4988-89b4-c69471f616aa\") " pod="openstack/nova-api-0" Oct 05 07:10:37 crc kubenswrapper[4846]: I1005 07:10:37.690628 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6e9e53c6-4cf2-4988-89b4-c69471f616aa-logs\") pod \"nova-api-0\" (UID: \"6e9e53c6-4cf2-4988-89b4-c69471f616aa\") " pod="openstack/nova-api-0" Oct 05 07:10:37 crc kubenswrapper[4846]: I1005 07:10:37.690675 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6e9e53c6-4cf2-4988-89b4-c69471f616aa-internal-tls-certs\") pod \"nova-api-0\" (UID: \"6e9e53c6-4cf2-4988-89b4-c69471f616aa\") " pod="openstack/nova-api-0" Oct 05 07:10:37 crc kubenswrapper[4846]: I1005 07:10:37.691158 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6e9e53c6-4cf2-4988-89b4-c69471f616aa-logs\") pod \"nova-api-0\" (UID: \"6e9e53c6-4cf2-4988-89b4-c69471f616aa\") " pod="openstack/nova-api-0" Oct 05 07:10:37 crc kubenswrapper[4846]: I1005 07:10:37.691247 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mc992\" (UniqueName: \"kubernetes.io/projected/6e9e53c6-4cf2-4988-89b4-c69471f616aa-kube-api-access-mc992\") pod \"nova-api-0\" (UID: \"6e9e53c6-4cf2-4988-89b4-c69471f616aa\") " pod="openstack/nova-api-0" Oct 05 07:10:37 crc kubenswrapper[4846]: I1005 07:10:37.691626 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6e9e53c6-4cf2-4988-89b4-c69471f616aa-public-tls-certs\") pod \"nova-api-0\" (UID: \"6e9e53c6-4cf2-4988-89b4-c69471f616aa\") " pod="openstack/nova-api-0" Oct 05 07:10:37 crc kubenswrapper[4846]: I1005 07:10:37.695731 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6e9e53c6-4cf2-4988-89b4-c69471f616aa-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"6e9e53c6-4cf2-4988-89b4-c69471f616aa\") " pod="openstack/nova-api-0" Oct 05 07:10:37 crc kubenswrapper[4846]: I1005 07:10:37.695997 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6e9e53c6-4cf2-4988-89b4-c69471f616aa-public-tls-certs\") pod \"nova-api-0\" (UID: \"6e9e53c6-4cf2-4988-89b4-c69471f616aa\") " pod="openstack/nova-api-0" Oct 05 07:10:37 crc kubenswrapper[4846]: I1005 07:10:37.696551 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6e9e53c6-4cf2-4988-89b4-c69471f616aa-internal-tls-certs\") pod \"nova-api-0\" (UID: \"6e9e53c6-4cf2-4988-89b4-c69471f616aa\") " pod="openstack/nova-api-0" Oct 05 07:10:37 crc kubenswrapper[4846]: I1005 07:10:37.697990 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6e9e53c6-4cf2-4988-89b4-c69471f616aa-config-data\") pod \"nova-api-0\" (UID: \"6e9e53c6-4cf2-4988-89b4-c69471f616aa\") " pod="openstack/nova-api-0" Oct 05 07:10:37 crc kubenswrapper[4846]: I1005 07:10:37.708566 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mc992\" (UniqueName: \"kubernetes.io/projected/6e9e53c6-4cf2-4988-89b4-c69471f616aa-kube-api-access-mc992\") pod \"nova-api-0\" (UID: \"6e9e53c6-4cf2-4988-89b4-c69471f616aa\") " pod="openstack/nova-api-0" Oct 05 07:10:37 crc kubenswrapper[4846]: I1005 07:10:37.942989 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 05 07:10:38 crc kubenswrapper[4846]: I1005 07:10:38.146635 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 05 07:10:38 crc kubenswrapper[4846]: I1005 07:10:38.156470 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 05 07:10:38 crc kubenswrapper[4846]: W1005 07:10:38.164507 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode6e86b1c_87ac_4594_aeba_943947065374.slice/crio-648dd0ba4f64806143815b8d201f6f6ff3ed3d36886724d53eedbf8ae1fc835e WatchSource:0}: Error finding container 648dd0ba4f64806143815b8d201f6f6ff3ed3d36886724d53eedbf8ae1fc835e: Status 404 returned error can't find the container with id 648dd0ba4f64806143815b8d201f6f6ff3ed3d36886724d53eedbf8ae1fc835e Oct 05 07:10:38 crc kubenswrapper[4846]: I1005 07:10:38.457673 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 05 07:10:38 crc kubenswrapper[4846]: W1005 07:10:38.459165 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6e9e53c6_4cf2_4988_89b4_c69471f616aa.slice/crio-1ccdf63d0267e37d0a39dfc09521b36fac3246c3415087fde77f78ac72e311ae WatchSource:0}: Error finding container 1ccdf63d0267e37d0a39dfc09521b36fac3246c3415087fde77f78ac72e311ae: Status 404 returned error can't find the container with id 1ccdf63d0267e37d0a39dfc09521b36fac3246c3415087fde77f78ac72e311ae Oct 05 07:10:38 crc kubenswrapper[4846]: I1005 07:10:38.514023 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="65fd80a1-d4ce-42f9-b1cf-e379333dfdba" path="/var/lib/kubelet/pods/65fd80a1-d4ce-42f9-b1cf-e379333dfdba/volumes" Oct 05 07:10:38 crc kubenswrapper[4846]: I1005 07:10:38.523245 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ce94efea-dc0a-4d7a-9340-9217ead55549" path="/var/lib/kubelet/pods/ce94efea-dc0a-4d7a-9340-9217ead55549/volumes" Oct 05 07:10:39 crc kubenswrapper[4846]: I1005 07:10:39.176133 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"e6e86b1c-87ac-4594-aeba-943947065374","Type":"ContainerStarted","Data":"7408c3d57efa73ce926d92ff91954d484124d71bc39b46bd428ba78c4ef03075"} Oct 05 07:10:39 crc kubenswrapper[4846]: I1005 07:10:39.176465 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"e6e86b1c-87ac-4594-aeba-943947065374","Type":"ContainerStarted","Data":"71ab8b0934bfb90a07c9012633af71114aed35423acecf39d3a2e32cd390b303"} Oct 05 07:10:39 crc kubenswrapper[4846]: I1005 07:10:39.176476 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"e6e86b1c-87ac-4594-aeba-943947065374","Type":"ContainerStarted","Data":"648dd0ba4f64806143815b8d201f6f6ff3ed3d36886724d53eedbf8ae1fc835e"} Oct 05 07:10:39 crc kubenswrapper[4846]: I1005 07:10:39.179772 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"b7de2524-bf5c-4668-af74-7584edf72c8b","Type":"ContainerStarted","Data":"316083fa9d64f1f90ee62226b3c28e38700ed5382c669302dda9655852943dab"} Oct 05 07:10:39 crc kubenswrapper[4846]: I1005 07:10:39.179819 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"b7de2524-bf5c-4668-af74-7584edf72c8b","Type":"ContainerStarted","Data":"e82e7415d67ac703880356b4cae71e6bc9b26a12df9bb1e074b952bab4350036"} Oct 05 07:10:39 crc kubenswrapper[4846]: I1005 07:10:39.185639 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"6e9e53c6-4cf2-4988-89b4-c69471f616aa","Type":"ContainerStarted","Data":"7b4d83183acff2c37efb3d06dc23fc65569d8eb1f78f7305d68857a97455040e"} Oct 05 07:10:39 crc kubenswrapper[4846]: I1005 07:10:39.185691 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"6e9e53c6-4cf2-4988-89b4-c69471f616aa","Type":"ContainerStarted","Data":"a2a13c5c05a60506397f6b2d96fab2cded887ebfc95de2c51ee9dd249fae3050"} Oct 05 07:10:39 crc kubenswrapper[4846]: I1005 07:10:39.185705 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"6e9e53c6-4cf2-4988-89b4-c69471f616aa","Type":"ContainerStarted","Data":"1ccdf63d0267e37d0a39dfc09521b36fac3246c3415087fde77f78ac72e311ae"} Oct 05 07:10:39 crc kubenswrapper[4846]: I1005 07:10:39.204707 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=3.204668787 podStartE2EDuration="3.204668787s" podCreationTimestamp="2025-10-05 07:10:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 07:10:39.195906993 +0000 UTC m=+1361.436759808" watchObservedRunningTime="2025-10-05 07:10:39.204668787 +0000 UTC m=+1361.445521562" Oct 05 07:10:39 crc kubenswrapper[4846]: I1005 07:10:39.232376 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.232345926 podStartE2EDuration="2.232345926s" podCreationTimestamp="2025-10-05 07:10:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 07:10:39.221031814 +0000 UTC m=+1361.461884629" watchObservedRunningTime="2025-10-05 07:10:39.232345926 +0000 UTC m=+1361.473198701" Oct 05 07:10:39 crc kubenswrapper[4846]: I1005 07:10:39.248658 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.24861922 podStartE2EDuration="2.24861922s" podCreationTimestamp="2025-10-05 07:10:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 07:10:39.241393407 +0000 UTC m=+1361.482246222" watchObservedRunningTime="2025-10-05 07:10:39.24861922 +0000 UTC m=+1361.489472045" Oct 05 07:10:41 crc kubenswrapper[4846]: I1005 07:10:41.897156 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 05 07:10:41 crc kubenswrapper[4846]: I1005 07:10:41.897316 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 05 07:10:42 crc kubenswrapper[4846]: I1005 07:10:42.518593 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Oct 05 07:10:46 crc kubenswrapper[4846]: I1005 07:10:46.897923 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 05 07:10:46 crc kubenswrapper[4846]: I1005 07:10:46.898543 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 05 07:10:47 crc kubenswrapper[4846]: I1005 07:10:47.508466 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Oct 05 07:10:47 crc kubenswrapper[4846]: I1005 07:10:47.544129 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Oct 05 07:10:47 crc kubenswrapper[4846]: I1005 07:10:47.918365 4846 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="e6e86b1c-87ac-4594-aeba-943947065374" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.205:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 05 07:10:47 crc kubenswrapper[4846]: I1005 07:10:47.918365 4846 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="e6e86b1c-87ac-4594-aeba-943947065374" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.205:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 05 07:10:47 crc kubenswrapper[4846]: I1005 07:10:47.943687 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 05 07:10:47 crc kubenswrapper[4846]: I1005 07:10:47.943742 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 05 07:10:48 crc kubenswrapper[4846]: I1005 07:10:48.339684 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Oct 05 07:10:48 crc kubenswrapper[4846]: I1005 07:10:48.951446 4846 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="6e9e53c6-4cf2-4988-89b4-c69471f616aa" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.207:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 05 07:10:48 crc kubenswrapper[4846]: I1005 07:10:48.959645 4846 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="6e9e53c6-4cf2-4988-89b4-c69471f616aa" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.207:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 05 07:10:51 crc kubenswrapper[4846]: I1005 07:10:51.204782 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Oct 05 07:10:56 crc kubenswrapper[4846]: I1005 07:10:56.905650 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 05 07:10:56 crc kubenswrapper[4846]: I1005 07:10:56.910679 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 05 07:10:56 crc kubenswrapper[4846]: I1005 07:10:56.918111 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 05 07:10:57 crc kubenswrapper[4846]: I1005 07:10:57.426345 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 05 07:10:58 crc kubenswrapper[4846]: I1005 07:10:58.049539 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 05 07:10:58 crc kubenswrapper[4846]: I1005 07:10:58.050950 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 05 07:10:58 crc kubenswrapper[4846]: I1005 07:10:58.075448 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 05 07:10:58 crc kubenswrapper[4846]: I1005 07:10:58.150655 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 05 07:10:58 crc kubenswrapper[4846]: I1005 07:10:58.432039 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 05 07:10:58 crc kubenswrapper[4846]: I1005 07:10:58.440058 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 05 07:11:18 crc kubenswrapper[4846]: I1005 07:11:18.247468 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstackclient"] Oct 05 07:11:18 crc kubenswrapper[4846]: I1005 07:11:18.248234 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/openstackclient" podUID="98fc38b8-a0ee-46c9-9d3f-9c15e8355b9d" containerName="openstackclient" containerID="cri-o://5edadd35f5bb4b18c0ce1cfc15ed5663752e987563a9e6ef8438aa48f0d79ae2" gracePeriod=2 Oct 05 07:11:18 crc kubenswrapper[4846]: I1005 07:11:18.267276 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstackclient"] Oct 05 07:11:18 crc kubenswrapper[4846]: I1005 07:11:18.297365 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 05 07:11:18 crc kubenswrapper[4846]: I1005 07:11:18.323784 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glancebc5f-account-delete-hkzcl"] Oct 05 07:11:18 crc kubenswrapper[4846]: E1005 07:11:18.324289 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="98fc38b8-a0ee-46c9-9d3f-9c15e8355b9d" containerName="openstackclient" Oct 05 07:11:18 crc kubenswrapper[4846]: I1005 07:11:18.324307 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="98fc38b8-a0ee-46c9-9d3f-9c15e8355b9d" containerName="openstackclient" Oct 05 07:11:18 crc kubenswrapper[4846]: I1005 07:11:18.324536 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="98fc38b8-a0ee-46c9-9d3f-9c15e8355b9d" containerName="openstackclient" Oct 05 07:11:18 crc kubenswrapper[4846]: I1005 07:11:18.325247 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glancebc5f-account-delete-hkzcl" Oct 05 07:11:18 crc kubenswrapper[4846]: I1005 07:11:18.339956 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glancebc5f-account-delete-hkzcl"] Oct 05 07:11:18 crc kubenswrapper[4846]: I1005 07:11:18.408272 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-ovs-sg4vg"] Oct 05 07:11:18 crc kubenswrapper[4846]: I1005 07:11:18.416253 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovsdbserver-nb-0"] Oct 05 07:11:18 crc kubenswrapper[4846]: I1005 07:11:18.416681 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovsdbserver-nb-0" podUID="bdf95824-fa04-4a83-92cc-7e36b4d00289" containerName="openstack-network-exporter" containerID="cri-o://e5574edbd5d5e803159cc31cad023d93fa72c648be7c1d91c85d67e4ae1b4f70" gracePeriod=300 Oct 05 07:11:18 crc kubenswrapper[4846]: I1005 07:11:18.428696 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-4kqnw"] Oct 05 07:11:18 crc kubenswrapper[4846]: I1005 07:11:18.460538 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-metrics-58xdl"] Oct 05 07:11:18 crc kubenswrapper[4846]: I1005 07:11:18.460735 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-controller-metrics-58xdl" podUID="19f3a362-4973-4437-ac49-de71508aa4f2" containerName="openstack-network-exporter" containerID="cri-o://0e19580f528cd14688ae88ab26841372cd0d6d7f86e25927a1ceae0f4c254a43" gracePeriod=30 Oct 05 07:11:18 crc kubenswrapper[4846]: I1005 07:11:18.476443 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sszh5\" (UniqueName: \"kubernetes.io/projected/4d3616b4-9f2c-4c62-9760-0300b02d0dbb-kube-api-access-sszh5\") pod \"glancebc5f-account-delete-hkzcl\" (UID: \"4d3616b4-9f2c-4c62-9760-0300b02d0dbb\") " pod="openstack/glancebc5f-account-delete-hkzcl" Oct 05 07:11:18 crc kubenswrapper[4846]: I1005 07:11:18.476596 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovsdbserver-sb-0"] Oct 05 07:11:18 crc kubenswrapper[4846]: I1005 07:11:18.477012 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovsdbserver-sb-0" podUID="87e3744a-b8db-4e76-afa9-0e1129410d4c" containerName="openstack-network-exporter" containerID="cri-o://4110f42bd835d3c1d5633761ee0d01337fb709f2d6a2517116e7ed33818b184d" gracePeriod=300 Oct 05 07:11:18 crc kubenswrapper[4846]: E1005 07:11:18.477200 4846 configmap.go:193] Couldn't get configMap openstack/rabbitmq-config-data: configmap "rabbitmq-config-data" not found Oct 05 07:11:18 crc kubenswrapper[4846]: E1005 07:11:18.477248 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/35261dbc-cabe-4bf0-88f6-b8613ddec0f1-config-data podName:35261dbc-cabe-4bf0-88f6-b8613ddec0f1 nodeName:}" failed. No retries permitted until 2025-10-05 07:11:18.977231839 +0000 UTC m=+1401.218084614 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/35261dbc-cabe-4bf0-88f6-b8613ddec0f1-config-data") pod "rabbitmq-server-0" (UID: "35261dbc-cabe-4bf0-88f6-b8613ddec0f1") : configmap "rabbitmq-config-data" not found Oct 05 07:11:18 crc kubenswrapper[4846]: I1005 07:11:18.495458 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutrona5c0-account-delete-q5wg4"] Oct 05 07:11:18 crc kubenswrapper[4846]: I1005 07:11:18.526749 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutrona5c0-account-delete-q5wg4" Oct 05 07:11:18 crc kubenswrapper[4846]: I1005 07:11:18.602498 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovsdbserver-nb-0" podUID="bdf95824-fa04-4a83-92cc-7e36b4d00289" containerName="ovsdbserver-nb" containerID="cri-o://b585eb7d447d9af1ab5bd601fbb7ea5ad51797c817722421db131f2fba3c2373" gracePeriod=300 Oct 05 07:11:18 crc kubenswrapper[4846]: I1005 07:11:18.615432 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sszh5\" (UniqueName: \"kubernetes.io/projected/4d3616b4-9f2c-4c62-9760-0300b02d0dbb-kube-api-access-sszh5\") pod \"glancebc5f-account-delete-hkzcl\" (UID: \"4d3616b4-9f2c-4c62-9760-0300b02d0dbb\") " pod="openstack/glancebc5f-account-delete-hkzcl" Oct 05 07:11:18 crc kubenswrapper[4846]: I1005 07:11:18.691090 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutrona5c0-account-delete-q5wg4"] Oct 05 07:11:18 crc kubenswrapper[4846]: I1005 07:11:18.691204 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-northd-0"] Oct 05 07:11:18 crc kubenswrapper[4846]: I1005 07:11:18.691613 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-sync-68h9n"] Oct 05 07:11:18 crc kubenswrapper[4846]: I1005 07:11:18.691674 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-sync-68h9n"] Oct 05 07:11:18 crc kubenswrapper[4846]: I1005 07:11:18.691737 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-sync-2l7b4"] Oct 05 07:11:18 crc kubenswrapper[4846]: I1005 07:11:18.706538 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-northd-0" podUID="18a7be37-c8dc-4d8a-808f-2a63eb4d9692" containerName="ovn-northd" containerID="cri-o://4351908cf00eea1071aca5080f9d536de25e56a78e5d0713bec8f51caab84ecb" gracePeriod=30 Oct 05 07:11:18 crc kubenswrapper[4846]: I1005 07:11:18.741949 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sszh5\" (UniqueName: \"kubernetes.io/projected/4d3616b4-9f2c-4c62-9760-0300b02d0dbb-kube-api-access-sszh5\") pod \"glancebc5f-account-delete-hkzcl\" (UID: \"4d3616b4-9f2c-4c62-9760-0300b02d0dbb\") " pod="openstack/glancebc5f-account-delete-hkzcl" Oct 05 07:11:18 crc kubenswrapper[4846]: I1005 07:11:18.742373 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-northd-0" podUID="18a7be37-c8dc-4d8a-808f-2a63eb4d9692" containerName="openstack-network-exporter" containerID="cri-o://05e6ade463feffe3d2a4a45255b934b5f81dd79a5b4cd4809663d8a570eb260c" gracePeriod=30 Oct 05 07:11:18 crc kubenswrapper[4846]: I1005 07:11:18.746515 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vhc7k\" (UniqueName: \"kubernetes.io/projected/e2f9f9a8-a25f-4a99-b425-4799964d4e24-kube-api-access-vhc7k\") pod \"neutrona5c0-account-delete-q5wg4\" (UID: \"e2f9f9a8-a25f-4a99-b425-4799964d4e24\") " pod="openstack/neutrona5c0-account-delete-q5wg4" Oct 05 07:11:18 crc kubenswrapper[4846]: I1005 07:11:18.753355 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovsdbserver-sb-0" podUID="87e3744a-b8db-4e76-afa9-0e1129410d4c" containerName="ovsdbserver-sb" containerID="cri-o://097a50245fd880a8216f94f3a81c3b749e8486580e37064c166b13f4fa1fc83c" gracePeriod=300 Oct 05 07:11:18 crc kubenswrapper[4846]: I1005 07:11:18.773749 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-sync-2l7b4"] Oct 05 07:11:18 crc kubenswrapper[4846]: I1005 07:11:18.789430 4846 generic.go:334] "Generic (PLEG): container finished" podID="87e3744a-b8db-4e76-afa9-0e1129410d4c" containerID="4110f42bd835d3c1d5633761ee0d01337fb709f2d6a2517116e7ed33818b184d" exitCode=2 Oct 05 07:11:18 crc kubenswrapper[4846]: I1005 07:11:18.789499 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"87e3744a-b8db-4e76-afa9-0e1129410d4c","Type":"ContainerDied","Data":"4110f42bd835d3c1d5633761ee0d01337fb709f2d6a2517116e7ed33818b184d"} Oct 05 07:11:18 crc kubenswrapper[4846]: I1005 07:11:18.794151 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-58xdl_19f3a362-4973-4437-ac49-de71508aa4f2/openstack-network-exporter/0.log" Oct 05 07:11:18 crc kubenswrapper[4846]: I1005 07:11:18.794209 4846 generic.go:334] "Generic (PLEG): container finished" podID="19f3a362-4973-4437-ac49-de71508aa4f2" containerID="0e19580f528cd14688ae88ab26841372cd0d6d7f86e25927a1ceae0f4c254a43" exitCode=2 Oct 05 07:11:18 crc kubenswrapper[4846]: I1005 07:11:18.794273 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-58xdl" event={"ID":"19f3a362-4973-4437-ac49-de71508aa4f2","Type":"ContainerDied","Data":"0e19580f528cd14688ae88ab26841372cd0d6d7f86e25927a1ceae0f4c254a43"} Oct 05 07:11:18 crc kubenswrapper[4846]: I1005 07:11:18.799679 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cindercf3d-account-delete-jw5tr"] Oct 05 07:11:18 crc kubenswrapper[4846]: I1005 07:11:18.801157 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cindercf3d-account-delete-jw5tr" Oct 05 07:11:18 crc kubenswrapper[4846]: I1005 07:11:18.809629 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_bdf95824-fa04-4a83-92cc-7e36b4d00289/ovsdbserver-nb/0.log" Oct 05 07:11:18 crc kubenswrapper[4846]: I1005 07:11:18.809680 4846 generic.go:334] "Generic (PLEG): container finished" podID="bdf95824-fa04-4a83-92cc-7e36b4d00289" containerID="e5574edbd5d5e803159cc31cad023d93fa72c648be7c1d91c85d67e4ae1b4f70" exitCode=2 Oct 05 07:11:18 crc kubenswrapper[4846]: I1005 07:11:18.809892 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cindercf3d-account-delete-jw5tr"] Oct 05 07:11:18 crc kubenswrapper[4846]: I1005 07:11:18.809942 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"bdf95824-fa04-4a83-92cc-7e36b4d00289","Type":"ContainerDied","Data":"e5574edbd5d5e803159cc31cad023d93fa72c648be7c1d91c85d67e4ae1b4f70"} Oct 05 07:11:18 crc kubenswrapper[4846]: I1005 07:11:18.821658 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-sync-cf6rt"] Oct 05 07:11:18 crc kubenswrapper[4846]: I1005 07:11:18.849004 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ds9zd\" (UniqueName: \"kubernetes.io/projected/a32cdd03-1398-4055-b051-62adde35200f-kube-api-access-ds9zd\") pod \"cindercf3d-account-delete-jw5tr\" (UID: \"a32cdd03-1398-4055-b051-62adde35200f\") " pod="openstack/cindercf3d-account-delete-jw5tr" Oct 05 07:11:18 crc kubenswrapper[4846]: I1005 07:11:18.849171 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vhc7k\" (UniqueName: \"kubernetes.io/projected/e2f9f9a8-a25f-4a99-b425-4799964d4e24-kube-api-access-vhc7k\") pod \"neutrona5c0-account-delete-q5wg4\" (UID: \"e2f9f9a8-a25f-4a99-b425-4799964d4e24\") " pod="openstack/neutrona5c0-account-delete-q5wg4" Oct 05 07:11:18 crc kubenswrapper[4846]: I1005 07:11:18.850382 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-sync-cf6rt"] Oct 05 07:11:18 crc kubenswrapper[4846]: I1005 07:11:18.867854 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement2760-account-delete-bvfdp"] Oct 05 07:11:18 crc kubenswrapper[4846]: I1005 07:11:18.869316 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement2760-account-delete-bvfdp" Oct 05 07:11:18 crc kubenswrapper[4846]: I1005 07:11:18.903474 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vhc7k\" (UniqueName: \"kubernetes.io/projected/e2f9f9a8-a25f-4a99-b425-4799964d4e24-kube-api-access-vhc7k\") pod \"neutrona5c0-account-delete-q5wg4\" (UID: \"e2f9f9a8-a25f-4a99-b425-4799964d4e24\") " pod="openstack/neutrona5c0-account-delete-q5wg4" Oct 05 07:11:18 crc kubenswrapper[4846]: I1005 07:11:18.910777 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-sync-7f2fl"] Oct 05 07:11:18 crc kubenswrapper[4846]: I1005 07:11:18.920731 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-sync-7f2fl"] Oct 05 07:11:18 crc kubenswrapper[4846]: I1005 07:11:18.929646 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement2760-account-delete-bvfdp"] Oct 05 07:11:18 crc kubenswrapper[4846]: I1005 07:11:18.944515 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-sync-t5brq"] Oct 05 07:11:18 crc kubenswrapper[4846]: I1005 07:11:18.950956 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ds9zd\" (UniqueName: \"kubernetes.io/projected/a32cdd03-1398-4055-b051-62adde35200f-kube-api-access-ds9zd\") pod \"cindercf3d-account-delete-jw5tr\" (UID: \"a32cdd03-1398-4055-b051-62adde35200f\") " pod="openstack/cindercf3d-account-delete-jw5tr" Oct 05 07:11:18 crc kubenswrapper[4846]: I1005 07:11:18.964558 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-sync-t5brq"] Oct 05 07:11:18 crc kubenswrapper[4846]: I1005 07:11:18.968024 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glancebc5f-account-delete-hkzcl" Oct 05 07:11:19 crc kubenswrapper[4846]: I1005 07:11:19.011028 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ds9zd\" (UniqueName: \"kubernetes.io/projected/a32cdd03-1398-4055-b051-62adde35200f-kube-api-access-ds9zd\") pod \"cindercf3d-account-delete-jw5tr\" (UID: \"a32cdd03-1398-4055-b051-62adde35200f\") " pod="openstack/cindercf3d-account-delete-jw5tr" Oct 05 07:11:19 crc kubenswrapper[4846]: I1005 07:11:19.052654 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gdftk\" (UniqueName: \"kubernetes.io/projected/fc35db15-65d4-4cd6-9d09-059822904757-kube-api-access-gdftk\") pod \"placement2760-account-delete-bvfdp\" (UID: \"fc35db15-65d4-4cd6-9d09-059822904757\") " pod="openstack/placement2760-account-delete-bvfdp" Oct 05 07:11:19 crc kubenswrapper[4846]: E1005 07:11:19.052891 4846 configmap.go:193] Couldn't get configMap openstack/rabbitmq-config-data: configmap "rabbitmq-config-data" not found Oct 05 07:11:19 crc kubenswrapper[4846]: E1005 07:11:19.052952 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/35261dbc-cabe-4bf0-88f6-b8613ddec0f1-config-data podName:35261dbc-cabe-4bf0-88f6-b8613ddec0f1 nodeName:}" failed. No retries permitted until 2025-10-05 07:11:20.052932786 +0000 UTC m=+1402.293785561 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/35261dbc-cabe-4bf0-88f6-b8613ddec0f1-config-data") pod "rabbitmq-server-0" (UID: "35261dbc-cabe-4bf0-88f6-b8613ddec0f1") : configmap "rabbitmq-config-data" not found Oct 05 07:11:19 crc kubenswrapper[4846]: I1005 07:11:19.061875 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican8fc5-account-delete-gjp47"] Oct 05 07:11:19 crc kubenswrapper[4846]: I1005 07:11:19.063267 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican8fc5-account-delete-gjp47" Oct 05 07:11:19 crc kubenswrapper[4846]: I1005 07:11:19.069199 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican8fc5-account-delete-gjp47"] Oct 05 07:11:19 crc kubenswrapper[4846]: I1005 07:11:19.084075 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 05 07:11:19 crc kubenswrapper[4846]: I1005 07:11:19.084359 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="2adc8ba0-2810-4bf4-bd24-172b97fb9b22" containerName="glance-log" containerID="cri-o://3be1afa599376331b916c6589c79dd8bdcf3fefa4275b988917d64c058731c7a" gracePeriod=30 Oct 05 07:11:19 crc kubenswrapper[4846]: I1005 07:11:19.084431 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="2adc8ba0-2810-4bf4-bd24-172b97fb9b22" containerName="glance-httpd" containerID="cri-o://9e54d631fb403eb9146c131948244ad644f213facfb43b90944167c8827ee59f" gracePeriod=30 Oct 05 07:11:19 crc kubenswrapper[4846]: I1005 07:11:19.114453 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 05 07:11:19 crc kubenswrapper[4846]: I1005 07:11:19.148320 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-ring-rebalance-b9gxc"] Oct 05 07:11:19 crc kubenswrapper[4846]: I1005 07:11:19.154600 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pvrgj\" (UniqueName: \"kubernetes.io/projected/086cf913-a4a2-41ee-910b-a83be95c69d6-kube-api-access-pvrgj\") pod \"barbican8fc5-account-delete-gjp47\" (UID: \"086cf913-a4a2-41ee-910b-a83be95c69d6\") " pod="openstack/barbican8fc5-account-delete-gjp47" Oct 05 07:11:19 crc kubenswrapper[4846]: I1005 07:11:19.154665 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gdftk\" (UniqueName: \"kubernetes.io/projected/fc35db15-65d4-4cd6-9d09-059822904757-kube-api-access-gdftk\") pod \"placement2760-account-delete-bvfdp\" (UID: \"fc35db15-65d4-4cd6-9d09-059822904757\") " pod="openstack/placement2760-account-delete-bvfdp" Oct 05 07:11:19 crc kubenswrapper[4846]: I1005 07:11:19.181009 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutrona5c0-account-delete-q5wg4" Oct 05 07:11:19 crc kubenswrapper[4846]: I1005 07:11:19.195275 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gdftk\" (UniqueName: \"kubernetes.io/projected/fc35db15-65d4-4cd6-9d09-059822904757-kube-api-access-gdftk\") pod \"placement2760-account-delete-bvfdp\" (UID: \"fc35db15-65d4-4cd6-9d09-059822904757\") " pod="openstack/placement2760-account-delete-bvfdp" Oct 05 07:11:19 crc kubenswrapper[4846]: I1005 07:11:19.212330 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/swift-ring-rebalance-b9gxc"] Oct 05 07:11:19 crc kubenswrapper[4846]: I1005 07:11:19.232085 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-c7d6584f5-dxcgg"] Oct 05 07:11:19 crc kubenswrapper[4846]: I1005 07:11:19.232390 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-c7d6584f5-dxcgg" podUID="7a5c4c82-4b3f-4314-a32d-db5398ca8a9e" containerName="dnsmasq-dns" containerID="cri-o://283ce773f8b442a1a87a4848af02010620aa0074924483f0a608d4c3f15d5131" gracePeriod=10 Oct 05 07:11:19 crc kubenswrapper[4846]: I1005 07:11:19.235924 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cindercf3d-account-delete-jw5tr" Oct 05 07:11:19 crc kubenswrapper[4846]: I1005 07:11:19.256945 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pvrgj\" (UniqueName: \"kubernetes.io/projected/086cf913-a4a2-41ee-910b-a83be95c69d6-kube-api-access-pvrgj\") pod \"barbican8fc5-account-delete-gjp47\" (UID: \"086cf913-a4a2-41ee-910b-a83be95c69d6\") " pod="openstack/barbican8fc5-account-delete-gjp47" Oct 05 07:11:19 crc kubenswrapper[4846]: E1005 07:11:19.257941 4846 configmap.go:193] Couldn't get configMap openstack/rabbitmq-cell1-config-data: configmap "rabbitmq-cell1-config-data" not found Oct 05 07:11:19 crc kubenswrapper[4846]: E1005 07:11:19.257998 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/626e6bce-7422-4fbd-b978-dfa5cb93de34-config-data podName:626e6bce-7422-4fbd-b978-dfa5cb93de34 nodeName:}" failed. No retries permitted until 2025-10-05 07:11:19.757979489 +0000 UTC m=+1401.998832264 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/626e6bce-7422-4fbd-b978-dfa5cb93de34-config-data") pod "rabbitmq-cell1-server-0" (UID: "626e6bce-7422-4fbd-b978-dfa5cb93de34") : configmap "rabbitmq-cell1-config-data" not found Oct 05 07:11:19 crc kubenswrapper[4846]: I1005 07:11:19.272773 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement2760-account-delete-bvfdp" Oct 05 07:11:19 crc kubenswrapper[4846]: I1005 07:11:19.281687 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-85bbbcfb7f-kwfjt"] Oct 05 07:11:19 crc kubenswrapper[4846]: I1005 07:11:19.281941 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-85bbbcfb7f-kwfjt" podUID="a72c0cf5-8cd9-4f5f-8a2b-25bf41ad0638" containerName="neutron-api" containerID="cri-o://57243f61674b9f96faaeba668f666e7518c0f44adbe9b6c573a200820bbb58d3" gracePeriod=30 Oct 05 07:11:19 crc kubenswrapper[4846]: I1005 07:11:19.282073 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-85bbbcfb7f-kwfjt" podUID="a72c0cf5-8cd9-4f5f-8a2b-25bf41ad0638" containerName="neutron-httpd" containerID="cri-o://1ac7cfeb9e27da195d8d27a5e4c8617a950c62fafc93500f5db134c1b1af30c2" gracePeriod=30 Oct 05 07:11:19 crc kubenswrapper[4846]: I1005 07:11:19.325182 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pvrgj\" (UniqueName: \"kubernetes.io/projected/086cf913-a4a2-41ee-910b-a83be95c69d6-kube-api-access-pvrgj\") pod \"barbican8fc5-account-delete-gjp47\" (UID: \"086cf913-a4a2-41ee-910b-a83be95c69d6\") " pod="openstack/barbican8fc5-account-delete-gjp47" Oct 05 07:11:19 crc kubenswrapper[4846]: I1005 07:11:19.341757 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 05 07:11:19 crc kubenswrapper[4846]: I1005 07:11:19.342604 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="0c0f8308-455e-4544-9f53-0c06a54c6e22" containerName="glance-log" containerID="cri-o://623a1c303003463ea428d6d40cb44811fb558226463a79ed21399df208bb4505" gracePeriod=30 Oct 05 07:11:19 crc kubenswrapper[4846]: I1005 07:11:19.343070 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="0c0f8308-455e-4544-9f53-0c06a54c6e22" containerName="glance-httpd" containerID="cri-o://708780fcc4181d993d6e7d44ef9444fa8e85b47f3a59d10a6fa23aa84fa714b0" gracePeriod=30 Oct 05 07:11:19 crc kubenswrapper[4846]: I1005 07:11:19.375619 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-storage-0"] Oct 05 07:11:19 crc kubenswrapper[4846]: I1005 07:11:19.376145 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="62d2ef82-9372-4b55-990e-30eb8e16e439" containerName="object-server" containerID="cri-o://b448ad891d0bdb677c147854fe487e87246e96353b5d9e033e8ba9ab2dfa1486" gracePeriod=30 Oct 05 07:11:19 crc kubenswrapper[4846]: I1005 07:11:19.378521 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="62d2ef82-9372-4b55-990e-30eb8e16e439" containerName="container-server" containerID="cri-o://1892fe91962ddb7379e117ea24de43a317054f7a7cd0983a6e06d7d4a4eafef9" gracePeriod=30 Oct 05 07:11:19 crc kubenswrapper[4846]: I1005 07:11:19.378711 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="62d2ef82-9372-4b55-990e-30eb8e16e439" containerName="container-updater" containerID="cri-o://3335b438bf1fc84c97c1fad661d249e0c3d1828dffb17a746efb34add2788d6d" gracePeriod=30 Oct 05 07:11:19 crc kubenswrapper[4846]: I1005 07:11:19.378759 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="62d2ef82-9372-4b55-990e-30eb8e16e439" containerName="container-auditor" containerID="cri-o://1af615c099267db200d6c220550e8454d08ffe8449631c9c707f1cdad15b401e" gracePeriod=30 Oct 05 07:11:19 crc kubenswrapper[4846]: I1005 07:11:19.378791 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="62d2ef82-9372-4b55-990e-30eb8e16e439" containerName="container-replicator" containerID="cri-o://19432cd2eb09eb116d05c3d630f9414b8755868167c8ad9d50242b823e184fe6" gracePeriod=30 Oct 05 07:11:19 crc kubenswrapper[4846]: I1005 07:11:19.378847 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="62d2ef82-9372-4b55-990e-30eb8e16e439" containerName="account-auditor" containerID="cri-o://cb25bcb083c8fd00a96201f36477176be16589836d2ad4521c750f525555a9af" gracePeriod=30 Oct 05 07:11:19 crc kubenswrapper[4846]: I1005 07:11:19.382253 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="62d2ef82-9372-4b55-990e-30eb8e16e439" containerName="account-reaper" containerID="cri-o://d7516230c3ea4b249ca682d8efe6e221f59dc03b58d2094da6e825d330b539b4" gracePeriod=30 Oct 05 07:11:19 crc kubenswrapper[4846]: I1005 07:11:19.382500 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="62d2ef82-9372-4b55-990e-30eb8e16e439" containerName="account-replicator" containerID="cri-o://888a361d7601c9ec3ffa1d314d1a426b937569ba6430ad0b3420461b118db555" gracePeriod=30 Oct 05 07:11:19 crc kubenswrapper[4846]: I1005 07:11:19.382612 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="62d2ef82-9372-4b55-990e-30eb8e16e439" containerName="object-expirer" containerID="cri-o://9af99544264e591f2b45ec0655947d9163b35da963449be03de565a72c5e8e50" gracePeriod=30 Oct 05 07:11:19 crc kubenswrapper[4846]: I1005 07:11:19.382733 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="62d2ef82-9372-4b55-990e-30eb8e16e439" containerName="object-auditor" containerID="cri-o://0560e729496f16c1c92b49ae2209e5dcfd4cd0655c6017f90693d429daef0694" gracePeriod=30 Oct 05 07:11:19 crc kubenswrapper[4846]: I1005 07:11:19.382794 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="62d2ef82-9372-4b55-990e-30eb8e16e439" containerName="object-updater" containerID="cri-o://983443d1fc33fbdef8e7ca49faf419ed150fada692ed4b5e60d29ffcb9cbaf6c" gracePeriod=30 Oct 05 07:11:19 crc kubenswrapper[4846]: I1005 07:11:19.382837 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="62d2ef82-9372-4b55-990e-30eb8e16e439" containerName="rsync" containerID="cri-o://8b1645c0bb7485916cb16906718326320b1c0c52c16426fad29ece1124acb1d4" gracePeriod=30 Oct 05 07:11:19 crc kubenswrapper[4846]: I1005 07:11:19.382871 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="62d2ef82-9372-4b55-990e-30eb8e16e439" containerName="swift-recon-cron" containerID="cri-o://ccba487ad88a231e17d0bff3850ebdc3dc9887084fa35706507c9470dedbf5df" gracePeriod=30 Oct 05 07:11:19 crc kubenswrapper[4846]: I1005 07:11:19.382920 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="62d2ef82-9372-4b55-990e-30eb8e16e439" containerName="object-replicator" containerID="cri-o://3e6f56f54a2bfbb877522ef200058490423742afbcf0b93313a239fae7415942" gracePeriod=30 Oct 05 07:11:19 crc kubenswrapper[4846]: I1005 07:11:19.376129 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="62d2ef82-9372-4b55-990e-30eb8e16e439" containerName="account-server" containerID="cri-o://a4edeadbef779fa82267d6b817cf8a19cf59e7146c6c1a8f5c786ad4fea85d4d" gracePeriod=30 Oct 05 07:11:19 crc kubenswrapper[4846]: I1005 07:11:19.413851 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 05 07:11:19 crc kubenswrapper[4846]: I1005 07:11:19.414116 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="bfe96e0c-b9fc-438d-9b17-bec246d48446" containerName="cinder-scheduler" containerID="cri-o://1921e872cb6fabf10cd959960895a03e62a0d256a8d1c87f18ab5389a991e8f4" gracePeriod=30 Oct 05 07:11:19 crc kubenswrapper[4846]: I1005 07:11:19.414285 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="bfe96e0c-b9fc-438d-9b17-bec246d48446" containerName="probe" containerID="cri-o://470aec4a8b7085d2d3371c5366c57378850aca957d883939f98653f186109a67" gracePeriod=30 Oct 05 07:11:19 crc kubenswrapper[4846]: I1005 07:11:19.427525 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican8fc5-account-delete-gjp47" Oct 05 07:11:19 crc kubenswrapper[4846]: I1005 07:11:19.448778 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-65884db4f4-fsk6p"] Oct 05 07:11:19 crc kubenswrapper[4846]: I1005 07:11:19.449053 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/placement-65884db4f4-fsk6p" podUID="c35affe0-da4e-448f-aac0-7bba26f89040" containerName="placement-log" containerID="cri-o://d0a2d9595d1231b4e28c0eff4482c56b148f02336c7fb50f8a4400e74af973ff" gracePeriod=30 Oct 05 07:11:19 crc kubenswrapper[4846]: I1005 07:11:19.449401 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/placement-65884db4f4-fsk6p" podUID="c35affe0-da4e-448f-aac0-7bba26f89040" containerName="placement-api" containerID="cri-o://b48698aea06e5e06f2a72f7568ee41c7c5c3fee9c4bdab0f9cb7cebb4f50ea03" gracePeriod=30 Oct 05 07:11:19 crc kubenswrapper[4846]: I1005 07:11:19.509306 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-cell-mapping-ptdtq"] Oct 05 07:11:19 crc kubenswrapper[4846]: I1005 07:11:19.557436 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-cell-mapping-ptdtq"] Oct 05 07:11:19 crc kubenswrapper[4846]: I1005 07:11:19.602284 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-cell-mapping-t7fdh"] Oct 05 07:11:19 crc kubenswrapper[4846]: I1005 07:11:19.637718 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-cell-mapping-t7fdh"] Oct 05 07:11:19 crc kubenswrapper[4846]: I1005 07:11:19.714519 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Oct 05 07:11:19 crc kubenswrapper[4846]: I1005 07:11:19.714783 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="36dca705-2f72-411e-8d80-ed31d49989cc" containerName="cinder-api-log" containerID="cri-o://735ae29bd9ab70b9f2083caa08160bc161134c908e991ed9c2afaf1f25347aa0" gracePeriod=30 Oct 05 07:11:19 crc kubenswrapper[4846]: I1005 07:11:19.715219 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="36dca705-2f72-411e-8d80-ed31d49989cc" containerName="cinder-api" containerID="cri-o://2dae962829e3caf9333b75e7f59bede4cf4d679ed6def4e5eddee1a34c169b39" gracePeriod=30 Oct 05 07:11:19 crc kubenswrapper[4846]: I1005 07:11:19.830240 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/cinder-api-0" podUID="36dca705-2f72-411e-8d80-ed31d49989cc" containerName="cinder-api" probeResult="failure" output="Get \"https://10.217.0.169:8776/healthcheck\": EOF" Oct 05 07:11:19 crc kubenswrapper[4846]: E1005 07:11:19.884715 4846 configmap.go:193] Couldn't get configMap openstack/rabbitmq-cell1-config-data: configmap "rabbitmq-cell1-config-data" not found Oct 05 07:11:19 crc kubenswrapper[4846]: E1005 07:11:19.884789 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/626e6bce-7422-4fbd-b978-dfa5cb93de34-config-data podName:626e6bce-7422-4fbd-b978-dfa5cb93de34 nodeName:}" failed. No retries permitted until 2025-10-05 07:11:20.88477182 +0000 UTC m=+1403.125624595 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/626e6bce-7422-4fbd-b978-dfa5cb93de34-config-data") pod "rabbitmq-cell1-server-0" (UID: "626e6bce-7422-4fbd-b978-dfa5cb93de34") : configmap "rabbitmq-cell1-config-data" not found Oct 05 07:11:19 crc kubenswrapper[4846]: E1005 07:11:19.900835 4846 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="4351908cf00eea1071aca5080f9d536de25e56a78e5d0713bec8f51caab84ecb" cmd=["/usr/local/bin/container-scripts/status_check.sh"] Oct 05 07:11:19 crc kubenswrapper[4846]: I1005 07:11:19.939910 4846 generic.go:334] "Generic (PLEG): container finished" podID="18a7be37-c8dc-4d8a-808f-2a63eb4d9692" containerID="05e6ade463feffe3d2a4a45255b934b5f81dd79a5b4cd4809663d8a570eb260c" exitCode=2 Oct 05 07:11:19 crc kubenswrapper[4846]: I1005 07:11:19.939996 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"18a7be37-c8dc-4d8a-808f-2a63eb4d9692","Type":"ContainerDied","Data":"05e6ade463feffe3d2a4a45255b934b5f81dd79a5b4cd4809663d8a570eb260c"} Oct 05 07:11:19 crc kubenswrapper[4846]: E1005 07:11:19.964487 4846 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="4351908cf00eea1071aca5080f9d536de25e56a78e5d0713bec8f51caab84ecb" cmd=["/usr/local/bin/container-scripts/status_check.sh"] Oct 05 07:11:20 crc kubenswrapper[4846]: I1005 07:11:19.998223 4846 generic.go:334] "Generic (PLEG): container finished" podID="0c0f8308-455e-4544-9f53-0c06a54c6e22" containerID="623a1c303003463ea428d6d40cb44811fb558226463a79ed21399df208bb4505" exitCode=143 Oct 05 07:11:20 crc kubenswrapper[4846]: I1005 07:11:19.999317 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"0c0f8308-455e-4544-9f53-0c06a54c6e22","Type":"ContainerDied","Data":"623a1c303003463ea428d6d40cb44811fb558226463a79ed21399df208bb4505"} Oct 05 07:11:20 crc kubenswrapper[4846]: E1005 07:11:20.022038 4846 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="4351908cf00eea1071aca5080f9d536de25e56a78e5d0713bec8f51caab84ecb" cmd=["/usr/local/bin/container-scripts/status_check.sh"] Oct 05 07:11:20 crc kubenswrapper[4846]: E1005 07:11:20.022312 4846 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-northd-0" podUID="18a7be37-c8dc-4d8a-808f-2a63eb4d9692" containerName="ovn-northd" Oct 05 07:11:20 crc kubenswrapper[4846]: I1005 07:11:20.029815 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-a5c0-account-create-8cbhj"] Oct 05 07:11:20 crc kubenswrapper[4846]: I1005 07:11:20.037065 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_87e3744a-b8db-4e76-afa9-0e1129410d4c/ovsdbserver-sb/0.log" Oct 05 07:11:20 crc kubenswrapper[4846]: I1005 07:11:20.037117 4846 generic.go:334] "Generic (PLEG): container finished" podID="87e3744a-b8db-4e76-afa9-0e1129410d4c" containerID="097a50245fd880a8216f94f3a81c3b749e8486580e37064c166b13f4fa1fc83c" exitCode=143 Oct 05 07:11:20 crc kubenswrapper[4846]: I1005 07:11:20.037228 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"87e3744a-b8db-4e76-afa9-0e1129410d4c","Type":"ContainerDied","Data":"097a50245fd880a8216f94f3a81c3b749e8486580e37064c166b13f4fa1fc83c"} Oct 05 07:11:20 crc kubenswrapper[4846]: I1005 07:11:20.040287 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-create-tbdm6"] Oct 05 07:11:20 crc kubenswrapper[4846]: I1005 07:11:20.124890 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-a5c0-account-create-8cbhj"] Oct 05 07:11:20 crc kubenswrapper[4846]: E1005 07:11:20.125601 4846 configmap.go:193] Couldn't get configMap openstack/rabbitmq-config-data: configmap "rabbitmq-config-data" not found Oct 05 07:11:20 crc kubenswrapper[4846]: E1005 07:11:20.125651 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/35261dbc-cabe-4bf0-88f6-b8613ddec0f1-config-data podName:35261dbc-cabe-4bf0-88f6-b8613ddec0f1 nodeName:}" failed. No retries permitted until 2025-10-05 07:11:22.12563731 +0000 UTC m=+1404.366490085 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/35261dbc-cabe-4bf0-88f6-b8613ddec0f1-config-data") pod "rabbitmq-server-0" (UID: "35261dbc-cabe-4bf0-88f6-b8613ddec0f1") : configmap "rabbitmq-config-data" not found Oct 05 07:11:20 crc kubenswrapper[4846]: I1005 07:11:20.143330 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-create-tbdm6"] Oct 05 07:11:20 crc kubenswrapper[4846]: I1005 07:11:20.151864 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutrona5c0-account-delete-q5wg4"] Oct 05 07:11:20 crc kubenswrapper[4846]: I1005 07:11:20.153640 4846 generic.go:334] "Generic (PLEG): container finished" podID="7a5c4c82-4b3f-4314-a32d-db5398ca8a9e" containerID="283ce773f8b442a1a87a4848af02010620aa0074924483f0a608d4c3f15d5131" exitCode=0 Oct 05 07:11:20 crc kubenswrapper[4846]: I1005 07:11:20.153767 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-c7d6584f5-dxcgg" event={"ID":"7a5c4c82-4b3f-4314-a32d-db5398ca8a9e","Type":"ContainerDied","Data":"283ce773f8b442a1a87a4848af02010620aa0074924483f0a608d4c3f15d5131"} Oct 05 07:11:20 crc kubenswrapper[4846]: I1005 07:11:20.191604 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 05 07:11:20 crc kubenswrapper[4846]: I1005 07:11:20.192138 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="e6e86b1c-87ac-4594-aeba-943947065374" containerName="nova-metadata-log" containerID="cri-o://71ab8b0934bfb90a07c9012633af71114aed35423acecf39d3a2e32cd390b303" gracePeriod=30 Oct 05 07:11:20 crc kubenswrapper[4846]: I1005 07:11:20.192501 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="e6e86b1c-87ac-4594-aeba-943947065374" containerName="nova-metadata-metadata" containerID="cri-o://7408c3d57efa73ce926d92ff91954d484124d71bc39b46bd428ba78c4ef03075" gracePeriod=30 Oct 05 07:11:20 crc kubenswrapper[4846]: I1005 07:11:20.251404 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 05 07:11:20 crc kubenswrapper[4846]: I1005 07:11:20.261952 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-create-rlqvn"] Oct 05 07:11:20 crc kubenswrapper[4846]: I1005 07:11:20.273715 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cindercf3d-account-delete-jw5tr"] Oct 05 07:11:20 crc kubenswrapper[4846]: I1005 07:11:20.284644 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-create-rlqvn"] Oct 05 07:11:20 crc kubenswrapper[4846]: I1005 07:11:20.294083 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstack-cell1-galera-0"] Oct 05 07:11:20 crc kubenswrapper[4846]: I1005 07:11:20.305652 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-cf3d-account-create-zd4h7"] Oct 05 07:11:20 crc kubenswrapper[4846]: I1005 07:11:20.306260 4846 generic.go:334] "Generic (PLEG): container finished" podID="62d2ef82-9372-4b55-990e-30eb8e16e439" containerID="9af99544264e591f2b45ec0655947d9163b35da963449be03de565a72c5e8e50" exitCode=0 Oct 05 07:11:20 crc kubenswrapper[4846]: I1005 07:11:20.306282 4846 generic.go:334] "Generic (PLEG): container finished" podID="62d2ef82-9372-4b55-990e-30eb8e16e439" containerID="983443d1fc33fbdef8e7ca49faf419ed150fada692ed4b5e60d29ffcb9cbaf6c" exitCode=0 Oct 05 07:11:20 crc kubenswrapper[4846]: I1005 07:11:20.306290 4846 generic.go:334] "Generic (PLEG): container finished" podID="62d2ef82-9372-4b55-990e-30eb8e16e439" containerID="0560e729496f16c1c92b49ae2209e5dcfd4cd0655c6017f90693d429daef0694" exitCode=0 Oct 05 07:11:20 crc kubenswrapper[4846]: I1005 07:11:20.306297 4846 generic.go:334] "Generic (PLEG): container finished" podID="62d2ef82-9372-4b55-990e-30eb8e16e439" containerID="3e6f56f54a2bfbb877522ef200058490423742afbcf0b93313a239fae7415942" exitCode=0 Oct 05 07:11:20 crc kubenswrapper[4846]: I1005 07:11:20.306307 4846 generic.go:334] "Generic (PLEG): container finished" podID="62d2ef82-9372-4b55-990e-30eb8e16e439" containerID="3335b438bf1fc84c97c1fad661d249e0c3d1828dffb17a746efb34add2788d6d" exitCode=0 Oct 05 07:11:20 crc kubenswrapper[4846]: I1005 07:11:20.306322 4846 generic.go:334] "Generic (PLEG): container finished" podID="62d2ef82-9372-4b55-990e-30eb8e16e439" containerID="19432cd2eb09eb116d05c3d630f9414b8755868167c8ad9d50242b823e184fe6" exitCode=0 Oct 05 07:11:20 crc kubenswrapper[4846]: I1005 07:11:20.306328 4846 generic.go:334] "Generic (PLEG): container finished" podID="62d2ef82-9372-4b55-990e-30eb8e16e439" containerID="d7516230c3ea4b249ca682d8efe6e221f59dc03b58d2094da6e825d330b539b4" exitCode=0 Oct 05 07:11:20 crc kubenswrapper[4846]: I1005 07:11:20.306334 4846 generic.go:334] "Generic (PLEG): container finished" podID="62d2ef82-9372-4b55-990e-30eb8e16e439" containerID="cb25bcb083c8fd00a96201f36477176be16589836d2ad4521c750f525555a9af" exitCode=0 Oct 05 07:11:20 crc kubenswrapper[4846]: I1005 07:11:20.306342 4846 generic.go:334] "Generic (PLEG): container finished" podID="62d2ef82-9372-4b55-990e-30eb8e16e439" containerID="888a361d7601c9ec3ffa1d314d1a426b937569ba6430ad0b3420461b118db555" exitCode=0 Oct 05 07:11:20 crc kubenswrapper[4846]: I1005 07:11:20.306376 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"62d2ef82-9372-4b55-990e-30eb8e16e439","Type":"ContainerDied","Data":"9af99544264e591f2b45ec0655947d9163b35da963449be03de565a72c5e8e50"} Oct 05 07:11:20 crc kubenswrapper[4846]: I1005 07:11:20.306395 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"62d2ef82-9372-4b55-990e-30eb8e16e439","Type":"ContainerDied","Data":"983443d1fc33fbdef8e7ca49faf419ed150fada692ed4b5e60d29ffcb9cbaf6c"} Oct 05 07:11:20 crc kubenswrapper[4846]: I1005 07:11:20.306404 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"62d2ef82-9372-4b55-990e-30eb8e16e439","Type":"ContainerDied","Data":"0560e729496f16c1c92b49ae2209e5dcfd4cd0655c6017f90693d429daef0694"} Oct 05 07:11:20 crc kubenswrapper[4846]: I1005 07:11:20.306412 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"62d2ef82-9372-4b55-990e-30eb8e16e439","Type":"ContainerDied","Data":"3e6f56f54a2bfbb877522ef200058490423742afbcf0b93313a239fae7415942"} Oct 05 07:11:20 crc kubenswrapper[4846]: I1005 07:11:20.306421 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"62d2ef82-9372-4b55-990e-30eb8e16e439","Type":"ContainerDied","Data":"3335b438bf1fc84c97c1fad661d249e0c3d1828dffb17a746efb34add2788d6d"} Oct 05 07:11:20 crc kubenswrapper[4846]: I1005 07:11:20.306430 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"62d2ef82-9372-4b55-990e-30eb8e16e439","Type":"ContainerDied","Data":"19432cd2eb09eb116d05c3d630f9414b8755868167c8ad9d50242b823e184fe6"} Oct 05 07:11:20 crc kubenswrapper[4846]: I1005 07:11:20.306439 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"62d2ef82-9372-4b55-990e-30eb8e16e439","Type":"ContainerDied","Data":"d7516230c3ea4b249ca682d8efe6e221f59dc03b58d2094da6e825d330b539b4"} Oct 05 07:11:20 crc kubenswrapper[4846]: I1005 07:11:20.306448 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"62d2ef82-9372-4b55-990e-30eb8e16e439","Type":"ContainerDied","Data":"cb25bcb083c8fd00a96201f36477176be16589836d2ad4521c750f525555a9af"} Oct 05 07:11:20 crc kubenswrapper[4846]: I1005 07:11:20.306456 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"62d2ef82-9372-4b55-990e-30eb8e16e439","Type":"ContainerDied","Data":"888a361d7601c9ec3ffa1d314d1a426b937569ba6430ad0b3420461b118db555"} Oct 05 07:11:20 crc kubenswrapper[4846]: I1005 07:11:20.313865 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-keystone-listener-77664b8dfd-5hkqj"] Oct 05 07:11:20 crc kubenswrapper[4846]: I1005 07:11:20.314150 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-keystone-listener-77664b8dfd-5hkqj" podUID="8d55f55a-2d79-4c01-8b0f-d67cefb842ff" containerName="barbican-keystone-listener-log" containerID="cri-o://bca1da26068e641006909835041107dfa8983891c244081dc35a6bf70c412d1d" gracePeriod=30 Oct 05 07:11:20 crc kubenswrapper[4846]: I1005 07:11:20.314637 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-keystone-listener-77664b8dfd-5hkqj" podUID="8d55f55a-2d79-4c01-8b0f-d67cefb842ff" containerName="barbican-keystone-listener" containerID="cri-o://6a990887fb132123ab2aa7e7c062e5fbecb25e29a746e6d59d303ed6848ccde0" gracePeriod=30 Oct 05 07:11:20 crc kubenswrapper[4846]: I1005 07:11:20.323598 4846 generic.go:334] "Generic (PLEG): container finished" podID="2adc8ba0-2810-4bf4-bd24-172b97fb9b22" containerID="3be1afa599376331b916c6589c79dd8bdcf3fefa4275b988917d64c058731c7a" exitCode=143 Oct 05 07:11:20 crc kubenswrapper[4846]: I1005 07:11:20.323653 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"2adc8ba0-2810-4bf4-bd24-172b97fb9b22","Type":"ContainerDied","Data":"3be1afa599376331b916c6589c79dd8bdcf3fefa4275b988917d64c058731c7a"} Oct 05 07:11:20 crc kubenswrapper[4846]: I1005 07:11:20.333712 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_bdf95824-fa04-4a83-92cc-7e36b4d00289/ovsdbserver-nb/0.log" Oct 05 07:11:20 crc kubenswrapper[4846]: I1005 07:11:20.333762 4846 generic.go:334] "Generic (PLEG): container finished" podID="bdf95824-fa04-4a83-92cc-7e36b4d00289" containerID="b585eb7d447d9af1ab5bd601fbb7ea5ad51797c817722421db131f2fba3c2373" exitCode=143 Oct 05 07:11:20 crc kubenswrapper[4846]: I1005 07:11:20.333790 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"bdf95824-fa04-4a83-92cc-7e36b4d00289","Type":"ContainerDied","Data":"b585eb7d447d9af1ab5bd601fbb7ea5ad51797c817722421db131f2fba3c2373"} Oct 05 07:11:20 crc kubenswrapper[4846]: I1005 07:11:20.339034 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-cf3d-account-create-zd4h7"] Oct 05 07:11:20 crc kubenswrapper[4846]: I1005 07:11:20.348534 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 05 07:11:20 crc kubenswrapper[4846]: I1005 07:11:20.349562 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="6e9e53c6-4cf2-4988-89b4-c69471f616aa" containerName="nova-api-api" containerID="cri-o://7b4d83183acff2c37efb3d06dc23fc65569d8eb1f78f7305d68857a97455040e" gracePeriod=30 Oct 05 07:11:20 crc kubenswrapper[4846]: I1005 07:11:20.349698 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="6e9e53c6-4cf2-4988-89b4-c69471f616aa" containerName="nova-api-log" containerID="cri-o://a2a13c5c05a60506397f6b2d96fab2cded887ebfc95de2c51ee9dd249fae3050" gracePeriod=30 Oct 05 07:11:20 crc kubenswrapper[4846]: I1005 07:11:20.359041 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement2760-account-delete-bvfdp"] Oct 05 07:11:20 crc kubenswrapper[4846]: I1005 07:11:20.368831 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-57674d6c54-zn5nq"] Oct 05 07:11:20 crc kubenswrapper[4846]: I1005 07:11:20.369075 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-57674d6c54-zn5nq" podUID="c31937eb-9836-461b-9230-5e06ce7a99e8" containerName="barbican-api-log" containerID="cri-o://85a65281831348ab23593ec8a61436c9871720d821d057e65d1423a5e8c5e79c" gracePeriod=30 Oct 05 07:11:20 crc kubenswrapper[4846]: I1005 07:11:20.369547 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-57674d6c54-zn5nq" podUID="c31937eb-9836-461b-9230-5e06ce7a99e8" containerName="barbican-api" containerID="cri-o://62af63220e4d608e4b981ffe4309575edbb012337dc1e84d368d62d0b4cab633" gracePeriod=30 Oct 05 07:11:20 crc kubenswrapper[4846]: I1005 07:11:20.379053 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-db-create-8s7qn"] Oct 05 07:11:20 crc kubenswrapper[4846]: I1005 07:11:20.392149 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-db-create-8s7qn"] Oct 05 07:11:20 crc kubenswrapper[4846]: I1005 07:11:20.408106 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-58xdl_19f3a362-4973-4437-ac49-de71508aa4f2/openstack-network-exporter/0.log" Oct 05 07:11:20 crc kubenswrapper[4846]: I1005 07:11:20.408171 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-58xdl" Oct 05 07:11:20 crc kubenswrapper[4846]: I1005 07:11:20.426600 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-create-ll6h2"] Oct 05 07:11:20 crc kubenswrapper[4846]: I1005 07:11:20.486387 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_87e3744a-b8db-4e76-afa9-0e1129410d4c/ovsdbserver-sb/0.log" Oct 05 07:11:20 crc kubenswrapper[4846]: I1005 07:11:20.486469 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Oct 05 07:11:20 crc kubenswrapper[4846]: I1005 07:11:20.487280 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-controller-ovs-sg4vg" podUID="29c0f404-fc7d-4909-b6bf-e1826a092a72" containerName="ovs-vswitchd" containerID="cri-o://22929dba446bc2516169306e09540fa52839ffc6b8d024f494720cc39ad0ab7d" gracePeriod=28 Oct 05 07:11:20 crc kubenswrapper[4846]: I1005 07:11:20.516060 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-server-0" podUID="35261dbc-cabe-4bf0-88f6-b8613ddec0f1" containerName="rabbitmq" containerID="cri-o://0b472a4ecaa620dd0581763afea4046cc8d15b55265470a90367ed2c9f54e39c" gracePeriod=604800 Oct 05 07:11:20 crc kubenswrapper[4846]: I1005 07:11:20.516748 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="00746abe-43f9-4275-bf0c-7f095b71e9bc" path="/var/lib/kubelet/pods/00746abe-43f9-4275-bf0c-7f095b71e9bc/volumes" Oct 05 07:11:20 crc kubenswrapper[4846]: I1005 07:11:20.517449 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0ffabf45-809d-43f4-902b-755111599aa8" path="/var/lib/kubelet/pods/0ffabf45-809d-43f4-902b-755111599aa8/volumes" Oct 05 07:11:20 crc kubenswrapper[4846]: I1005 07:11:20.517984 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="371a8404-6b77-418f-8fcb-3c72ed092816" path="/var/lib/kubelet/pods/371a8404-6b77-418f-8fcb-3c72ed092816/volumes" Oct 05 07:11:20 crc kubenswrapper[4846]: I1005 07:11:20.518570 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="438586d8-d6b1-445f-8460-0e44ac6037c6" path="/var/lib/kubelet/pods/438586d8-d6b1-445f-8460-0e44ac6037c6/volumes" Oct 05 07:11:20 crc kubenswrapper[4846]: I1005 07:11:20.519916 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="483b7e9a-7db4-4137-b505-d2b6c404ba06" path="/var/lib/kubelet/pods/483b7e9a-7db4-4137-b505-d2b6c404ba06/volumes" Oct 05 07:11:20 crc kubenswrapper[4846]: I1005 07:11:20.520400 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="577b34f0-581f-4fe2-8717-7e3062fd26fc" path="/var/lib/kubelet/pods/577b34f0-581f-4fe2-8717-7e3062fd26fc/volumes" Oct 05 07:11:20 crc kubenswrapper[4846]: I1005 07:11:20.520894 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8b6863be-f8e1-4a87-93d6-5cf756983747" path="/var/lib/kubelet/pods/8b6863be-f8e1-4a87-93d6-5cf756983747/volumes" Oct 05 07:11:20 crc kubenswrapper[4846]: I1005 07:11:20.522645 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8fa2fcc5-ba88-4a8c-88ec-a8d1882b8c7b" path="/var/lib/kubelet/pods/8fa2fcc5-ba88-4a8c-88ec-a8d1882b8c7b/volumes" Oct 05 07:11:20 crc kubenswrapper[4846]: I1005 07:11:20.524024 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a368d12f-6383-4a8e-8d71-23a675946027" path="/var/lib/kubelet/pods/a368d12f-6383-4a8e-8d71-23a675946027/volumes" Oct 05 07:11:20 crc kubenswrapper[4846]: I1005 07:11:20.524840 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a6ab76e7-ba49-4976-b718-3ea57bb4b53a" path="/var/lib/kubelet/pods/a6ab76e7-ba49-4976-b718-3ea57bb4b53a/volumes" Oct 05 07:11:20 crc kubenswrapper[4846]: I1005 07:11:20.526693 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b7a58faa-0f37-4b0e-a907-6a8bae28cb27" path="/var/lib/kubelet/pods/b7a58faa-0f37-4b0e-a907-6a8bae28cb27/volumes" Oct 05 07:11:20 crc kubenswrapper[4846]: I1005 07:11:20.527419 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ee312852-b2eb-43d8-864c-bc553fed7ea5" path="/var/lib/kubelet/pods/ee312852-b2eb-43d8-864c-bc553fed7ea5/volumes" Oct 05 07:11:20 crc kubenswrapper[4846]: I1005 07:11:20.528119 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fb35a7c4-9453-4ce8-b0c0-828893100c92" path="/var/lib/kubelet/pods/fb35a7c4-9453-4ce8-b0c0-828893100c92/volumes" Oct 05 07:11:20 crc kubenswrapper[4846]: I1005 07:11:20.540121 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n7jqd\" (UniqueName: \"kubernetes.io/projected/19f3a362-4973-4437-ac49-de71508aa4f2-kube-api-access-n7jqd\") pod \"19f3a362-4973-4437-ac49-de71508aa4f2\" (UID: \"19f3a362-4973-4437-ac49-de71508aa4f2\") " Oct 05 07:11:20 crc kubenswrapper[4846]: I1005 07:11:20.540288 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/19f3a362-4973-4437-ac49-de71508aa4f2-combined-ca-bundle\") pod \"19f3a362-4973-4437-ac49-de71508aa4f2\" (UID: \"19f3a362-4973-4437-ac49-de71508aa4f2\") " Oct 05 07:11:20 crc kubenswrapper[4846]: I1005 07:11:20.540324 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/19f3a362-4973-4437-ac49-de71508aa4f2-config\") pod \"19f3a362-4973-4437-ac49-de71508aa4f2\" (UID: \"19f3a362-4973-4437-ac49-de71508aa4f2\") " Oct 05 07:11:20 crc kubenswrapper[4846]: I1005 07:11:20.540421 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/19f3a362-4973-4437-ac49-de71508aa4f2-ovn-rundir\") pod \"19f3a362-4973-4437-ac49-de71508aa4f2\" (UID: \"19f3a362-4973-4437-ac49-de71508aa4f2\") " Oct 05 07:11:20 crc kubenswrapper[4846]: I1005 07:11:20.540488 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/19f3a362-4973-4437-ac49-de71508aa4f2-metrics-certs-tls-certs\") pod \"19f3a362-4973-4437-ac49-de71508aa4f2\" (UID: \"19f3a362-4973-4437-ac49-de71508aa4f2\") " Oct 05 07:11:20 crc kubenswrapper[4846]: I1005 07:11:20.540508 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/19f3a362-4973-4437-ac49-de71508aa4f2-ovs-rundir\") pod \"19f3a362-4973-4437-ac49-de71508aa4f2\" (UID: \"19f3a362-4973-4437-ac49-de71508aa4f2\") " Oct 05 07:11:20 crc kubenswrapper[4846]: I1005 07:11:20.540996 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/19f3a362-4973-4437-ac49-de71508aa4f2-ovs-rundir" (OuterVolumeSpecName: "ovs-rundir") pod "19f3a362-4973-4437-ac49-de71508aa4f2" (UID: "19f3a362-4973-4437-ac49-de71508aa4f2"). InnerVolumeSpecName "ovs-rundir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 05 07:11:20 crc kubenswrapper[4846]: I1005 07:11:20.546481 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/19f3a362-4973-4437-ac49-de71508aa4f2-ovn-rundir" (OuterVolumeSpecName: "ovn-rundir") pod "19f3a362-4973-4437-ac49-de71508aa4f2" (UID: "19f3a362-4973-4437-ac49-de71508aa4f2"). InnerVolumeSpecName "ovn-rundir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 05 07:11:20 crc kubenswrapper[4846]: I1005 07:11:20.546648 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/19f3a362-4973-4437-ac49-de71508aa4f2-config" (OuterVolumeSpecName: "config") pod "19f3a362-4973-4437-ac49-de71508aa4f2" (UID: "19f3a362-4973-4437-ac49-de71508aa4f2"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 07:11:20 crc kubenswrapper[4846]: I1005 07:11:20.553581 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-create-ll6h2"] Oct 05 07:11:20 crc kubenswrapper[4846]: I1005 07:11:20.563736 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-2760-account-create-pvm8n"] Oct 05 07:11:20 crc kubenswrapper[4846]: I1005 07:11:20.566682 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/19f3a362-4973-4437-ac49-de71508aa4f2-kube-api-access-n7jqd" (OuterVolumeSpecName: "kube-api-access-n7jqd") pod "19f3a362-4973-4437-ac49-de71508aa4f2" (UID: "19f3a362-4973-4437-ac49-de71508aa4f2"). InnerVolumeSpecName "kube-api-access-n7jqd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:11:20 crc kubenswrapper[4846]: I1005 07:11:20.580884 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-2760-account-create-pvm8n"] Oct 05 07:11:20 crc kubenswrapper[4846]: I1005 07:11:20.590253 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-db-create-4mqhm"] Oct 05 07:11:20 crc kubenswrapper[4846]: I1005 07:11:20.601485 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-db-create-4mqhm"] Oct 05 07:11:20 crc kubenswrapper[4846]: I1005 07:11:20.609887 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-8fc5-account-create-fcxfk"] Oct 05 07:11:20 crc kubenswrapper[4846]: I1005 07:11:20.616578 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-7ebc-account-create-k72ft"] Oct 05 07:11:20 crc kubenswrapper[4846]: I1005 07:11:20.623444 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-worker-78995b9f79-nwl7v"] Oct 05 07:11:20 crc kubenswrapper[4846]: I1005 07:11:20.623717 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-worker-78995b9f79-nwl7v" podUID="3995994c-2a3f-400f-9a64-b564759b20b2" containerName="barbican-worker-log" containerID="cri-o://afd676f2477334d96744cd37d02c69587dea11053d28e2ae166c10deda1b1560" gracePeriod=30 Oct 05 07:11:20 crc kubenswrapper[4846]: I1005 07:11:20.623941 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-worker-78995b9f79-nwl7v" podUID="3995994c-2a3f-400f-9a64-b564759b20b2" containerName="barbican-worker" containerID="cri-o://5cae68aeaeb33c53f345bf863894fe1192fa80a8ec2b486a6c2d29e83dec72af" gracePeriod=30 Oct 05 07:11:20 crc kubenswrapper[4846]: I1005 07:11:20.630236 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-7e7d-account-create-bcwnf"] Oct 05 07:11:20 crc kubenswrapper[4846]: I1005 07:11:20.640517 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-7d67-account-create-v5k6p"] Oct 05 07:11:20 crc kubenswrapper[4846]: I1005 07:11:20.641635 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/87e3744a-b8db-4e76-afa9-0e1129410d4c-config\") pod \"87e3744a-b8db-4e76-afa9-0e1129410d4c\" (UID: \"87e3744a-b8db-4e76-afa9-0e1129410d4c\") " Oct 05 07:11:20 crc kubenswrapper[4846]: I1005 07:11:20.641828 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/87e3744a-b8db-4e76-afa9-0e1129410d4c-combined-ca-bundle\") pod \"87e3744a-b8db-4e76-afa9-0e1129410d4c\" (UID: \"87e3744a-b8db-4e76-afa9-0e1129410d4c\") " Oct 05 07:11:20 crc kubenswrapper[4846]: I1005 07:11:20.641943 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/87e3744a-b8db-4e76-afa9-0e1129410d4c-scripts\") pod \"87e3744a-b8db-4e76-afa9-0e1129410d4c\" (UID: \"87e3744a-b8db-4e76-afa9-0e1129410d4c\") " Oct 05 07:11:20 crc kubenswrapper[4846]: I1005 07:11:20.642039 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndbcluster-sb-etc-ovn\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"87e3744a-b8db-4e76-afa9-0e1129410d4c\" (UID: \"87e3744a-b8db-4e76-afa9-0e1129410d4c\") " Oct 05 07:11:20 crc kubenswrapper[4846]: I1005 07:11:20.642132 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/87e3744a-b8db-4e76-afa9-0e1129410d4c-metrics-certs-tls-certs\") pod \"87e3744a-b8db-4e76-afa9-0e1129410d4c\" (UID: \"87e3744a-b8db-4e76-afa9-0e1129410d4c\") " Oct 05 07:11:20 crc kubenswrapper[4846]: I1005 07:11:20.642255 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/87e3744a-b8db-4e76-afa9-0e1129410d4c-ovsdb-rundir\") pod \"87e3744a-b8db-4e76-afa9-0e1129410d4c\" (UID: \"87e3744a-b8db-4e76-afa9-0e1129410d4c\") " Oct 05 07:11:20 crc kubenswrapper[4846]: I1005 07:11:20.642321 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qvxqv\" (UniqueName: \"kubernetes.io/projected/87e3744a-b8db-4e76-afa9-0e1129410d4c-kube-api-access-qvxqv\") pod \"87e3744a-b8db-4e76-afa9-0e1129410d4c\" (UID: \"87e3744a-b8db-4e76-afa9-0e1129410d4c\") " Oct 05 07:11:20 crc kubenswrapper[4846]: I1005 07:11:20.642408 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/87e3744a-b8db-4e76-afa9-0e1129410d4c-ovsdbserver-sb-tls-certs\") pod \"87e3744a-b8db-4e76-afa9-0e1129410d4c\" (UID: \"87e3744a-b8db-4e76-afa9-0e1129410d4c\") " Oct 05 07:11:20 crc kubenswrapper[4846]: I1005 07:11:20.642928 4846 reconciler_common.go:293] "Volume detached for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/19f3a362-4973-4437-ac49-de71508aa4f2-ovs-rundir\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:20 crc kubenswrapper[4846]: I1005 07:11:20.648948 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n7jqd\" (UniqueName: \"kubernetes.io/projected/19f3a362-4973-4437-ac49-de71508aa4f2-kube-api-access-n7jqd\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:20 crc kubenswrapper[4846]: I1005 07:11:20.649039 4846 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/19f3a362-4973-4437-ac49-de71508aa4f2-config\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:20 crc kubenswrapper[4846]: I1005 07:11:20.649096 4846 reconciler_common.go:293] "Volume detached for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/19f3a362-4973-4437-ac49-de71508aa4f2-ovn-rundir\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:20 crc kubenswrapper[4846]: I1005 07:11:20.645897 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87e3744a-b8db-4e76-afa9-0e1129410d4c-scripts" (OuterVolumeSpecName: "scripts") pod "87e3744a-b8db-4e76-afa9-0e1129410d4c" (UID: "87e3744a-b8db-4e76-afa9-0e1129410d4c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 07:11:20 crc kubenswrapper[4846]: I1005 07:11:20.646354 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87e3744a-b8db-4e76-afa9-0e1129410d4c-config" (OuterVolumeSpecName: "config") pod "87e3744a-b8db-4e76-afa9-0e1129410d4c" (UID: "87e3744a-b8db-4e76-afa9-0e1129410d4c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 07:11:20 crc kubenswrapper[4846]: I1005 07:11:20.648476 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87e3744a-b8db-4e76-afa9-0e1129410d4c-kube-api-access-qvxqv" (OuterVolumeSpecName: "kube-api-access-qvxqv") pod "87e3744a-b8db-4e76-afa9-0e1129410d4c" (UID: "87e3744a-b8db-4e76-afa9-0e1129410d4c"). InnerVolumeSpecName "kube-api-access-qvxqv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:11:20 crc kubenswrapper[4846]: I1005 07:11:20.662632 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/87e3744a-b8db-4e76-afa9-0e1129410d4c-ovsdb-rundir" (OuterVolumeSpecName: "ovsdb-rundir") pod "87e3744a-b8db-4e76-afa9-0e1129410d4c" (UID: "87e3744a-b8db-4e76-afa9-0e1129410d4c"). InnerVolumeSpecName "ovsdb-rundir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 07:11:20 crc kubenswrapper[4846]: I1005 07:11:20.664203 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage08-crc" (OuterVolumeSpecName: "ovndbcluster-sb-etc-ovn") pod "87e3744a-b8db-4e76-afa9-0e1129410d4c" (UID: "87e3744a-b8db-4e76-afa9-0e1129410d4c"). InnerVolumeSpecName "local-storage08-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 05 07:11:20 crc kubenswrapper[4846]: I1005 07:11:20.695220 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-create-xp6rt"] Oct 05 07:11:20 crc kubenswrapper[4846]: I1005 07:11:20.697295 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican8fc5-account-delete-gjp47"] Oct 05 07:11:20 crc kubenswrapper[4846]: I1005 07:11:20.704320 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-db-create-hpmx9"] Oct 05 07:11:20 crc kubenswrapper[4846]: I1005 07:11:20.708755 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-8fc5-account-create-fcxfk"] Oct 05 07:11:20 crc kubenswrapper[4846]: I1005 07:11:20.713448 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-7ebc-account-create-k72ft"] Oct 05 07:11:20 crc kubenswrapper[4846]: I1005 07:11:20.719905 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 05 07:11:20 crc kubenswrapper[4846]: I1005 07:11:20.720128 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="6660b9ff-df9c-478e-b852-82cb5fb2c60c" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://2a96112701d2d7042bce02efb3a776df18e837a8d8981829c635370de3016348" gracePeriod=30 Oct 05 07:11:20 crc kubenswrapper[4846]: I1005 07:11:20.727785 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-db-create-hpmx9"] Oct 05 07:11:20 crc kubenswrapper[4846]: I1005 07:11:20.740380 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-7d67-account-create-v5k6p"] Oct 05 07:11:20 crc kubenswrapper[4846]: I1005 07:11:20.745675 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/openstack-cell1-galera-0" podUID="221918b0-eb0e-4a32-8eed-07acd18bd31e" containerName="galera" containerID="cri-o://ec314d49982bc70fa3bb12b1a024bc708bb21c05eda9001d00d334940b04d2e6" gracePeriod=30 Oct 05 07:11:20 crc kubenswrapper[4846]: I1005 07:11:20.745915 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-7e7d-account-create-bcwnf"] Oct 05 07:11:20 crc kubenswrapper[4846]: I1005 07:11:20.750970 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-create-xp6rt"] Oct 05 07:11:20 crc kubenswrapper[4846]: I1005 07:11:20.758265 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 05 07:11:20 crc kubenswrapper[4846]: I1005 07:11:20.766633 4846 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/87e3744a-b8db-4e76-afa9-0e1129410d4c-scripts\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:20 crc kubenswrapper[4846]: I1005 07:11:20.767116 4846 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" " Oct 05 07:11:20 crc kubenswrapper[4846]: I1005 07:11:20.767883 4846 reconciler_common.go:293] "Volume detached for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/87e3744a-b8db-4e76-afa9-0e1129410d4c-ovsdb-rundir\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:20 crc kubenswrapper[4846]: I1005 07:11:20.767934 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qvxqv\" (UniqueName: \"kubernetes.io/projected/87e3744a-b8db-4e76-afa9-0e1129410d4c-kube-api-access-qvxqv\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:20 crc kubenswrapper[4846]: I1005 07:11:20.767950 4846 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/87e3744a-b8db-4e76-afa9-0e1129410d4c-config\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:20 crc kubenswrapper[4846]: I1005 07:11:20.792022 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87e3744a-b8db-4e76-afa9-0e1129410d4c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "87e3744a-b8db-4e76-afa9-0e1129410d4c" (UID: "87e3744a-b8db-4e76-afa9-0e1129410d4c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:11:20 crc kubenswrapper[4846]: I1005 07:11:20.809879 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/19f3a362-4973-4437-ac49-de71508aa4f2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "19f3a362-4973-4437-ac49-de71508aa4f2" (UID: "19f3a362-4973-4437-ac49-de71508aa4f2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:11:20 crc kubenswrapper[4846]: E1005 07:11:20.810408 4846 handlers.go:78] "Exec lifecycle hook for Container in Pod failed" err=< Oct 05 07:11:20 crc kubenswrapper[4846]: command '/usr/local/bin/container-scripts/stop-ovsdb-server.sh' exited with 137: ++ dirname /usr/local/bin/container-scripts/stop-ovsdb-server.sh Oct 05 07:11:20 crc kubenswrapper[4846]: + source /usr/local/bin/container-scripts/functions Oct 05 07:11:20 crc kubenswrapper[4846]: ++ OVNBridge=br-int Oct 05 07:11:20 crc kubenswrapper[4846]: ++ OVNRemote=tcp:localhost:6642 Oct 05 07:11:20 crc kubenswrapper[4846]: ++ OVNEncapType=geneve Oct 05 07:11:20 crc kubenswrapper[4846]: ++ OVNAvailabilityZones= Oct 05 07:11:20 crc kubenswrapper[4846]: ++ EnableChassisAsGateway=true Oct 05 07:11:20 crc kubenswrapper[4846]: ++ PhysicalNetworks= Oct 05 07:11:20 crc kubenswrapper[4846]: ++ OVNHostName= Oct 05 07:11:20 crc kubenswrapper[4846]: ++ DB_FILE=/etc/openvswitch/conf.db Oct 05 07:11:20 crc kubenswrapper[4846]: ++ ovs_dir=/var/lib/openvswitch Oct 05 07:11:20 crc kubenswrapper[4846]: ++ FLOWS_RESTORE_SCRIPT=/var/lib/openvswitch/flows-script Oct 05 07:11:20 crc kubenswrapper[4846]: ++ FLOWS_RESTORE_DIR=/var/lib/openvswitch/saved-flows Oct 05 07:11:20 crc kubenswrapper[4846]: ++ SAFE_TO_STOP_OVSDB_SERVER_SEMAPHORE=/var/lib/openvswitch/is_safe_to_stop_ovsdb_server Oct 05 07:11:20 crc kubenswrapper[4846]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Oct 05 07:11:20 crc kubenswrapper[4846]: + sleep 0.5 Oct 05 07:11:20 crc kubenswrapper[4846]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Oct 05 07:11:20 crc kubenswrapper[4846]: + sleep 0.5 Oct 05 07:11:20 crc kubenswrapper[4846]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Oct 05 07:11:20 crc kubenswrapper[4846]: + sleep 0.5 Oct 05 07:11:20 crc kubenswrapper[4846]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Oct 05 07:11:20 crc kubenswrapper[4846]: + cleanup_ovsdb_server_semaphore Oct 05 07:11:20 crc kubenswrapper[4846]: + rm -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server Oct 05 07:11:20 crc kubenswrapper[4846]: + /usr/share/openvswitch/scripts/ovs-ctl stop --no-ovs-vswitchd Oct 05 07:11:20 crc kubenswrapper[4846]: > execCommand=["/usr/local/bin/container-scripts/stop-ovsdb-server.sh"] containerName="ovsdb-server" pod="openstack/ovn-controller-ovs-sg4vg" message=< Oct 05 07:11:20 crc kubenswrapper[4846]: Exiting ovsdb-server (5) [ OK ] Oct 05 07:11:20 crc kubenswrapper[4846]: ++ dirname /usr/local/bin/container-scripts/stop-ovsdb-server.sh Oct 05 07:11:20 crc kubenswrapper[4846]: + source /usr/local/bin/container-scripts/functions Oct 05 07:11:20 crc kubenswrapper[4846]: ++ OVNBridge=br-int Oct 05 07:11:20 crc kubenswrapper[4846]: ++ OVNRemote=tcp:localhost:6642 Oct 05 07:11:20 crc kubenswrapper[4846]: ++ OVNEncapType=geneve Oct 05 07:11:20 crc kubenswrapper[4846]: ++ OVNAvailabilityZones= Oct 05 07:11:20 crc kubenswrapper[4846]: ++ EnableChassisAsGateway=true Oct 05 07:11:20 crc kubenswrapper[4846]: ++ PhysicalNetworks= Oct 05 07:11:20 crc kubenswrapper[4846]: ++ OVNHostName= Oct 05 07:11:20 crc kubenswrapper[4846]: ++ DB_FILE=/etc/openvswitch/conf.db Oct 05 07:11:20 crc kubenswrapper[4846]: ++ ovs_dir=/var/lib/openvswitch Oct 05 07:11:20 crc kubenswrapper[4846]: ++ FLOWS_RESTORE_SCRIPT=/var/lib/openvswitch/flows-script Oct 05 07:11:20 crc kubenswrapper[4846]: ++ FLOWS_RESTORE_DIR=/var/lib/openvswitch/saved-flows Oct 05 07:11:20 crc kubenswrapper[4846]: ++ SAFE_TO_STOP_OVSDB_SERVER_SEMAPHORE=/var/lib/openvswitch/is_safe_to_stop_ovsdb_server Oct 05 07:11:20 crc kubenswrapper[4846]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Oct 05 07:11:20 crc kubenswrapper[4846]: + sleep 0.5 Oct 05 07:11:20 crc kubenswrapper[4846]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Oct 05 07:11:20 crc kubenswrapper[4846]: + sleep 0.5 Oct 05 07:11:20 crc kubenswrapper[4846]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Oct 05 07:11:20 crc kubenswrapper[4846]: + sleep 0.5 Oct 05 07:11:20 crc kubenswrapper[4846]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Oct 05 07:11:20 crc kubenswrapper[4846]: + cleanup_ovsdb_server_semaphore Oct 05 07:11:20 crc kubenswrapper[4846]: + rm -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server Oct 05 07:11:20 crc kubenswrapper[4846]: + /usr/share/openvswitch/scripts/ovs-ctl stop --no-ovs-vswitchd Oct 05 07:11:20 crc kubenswrapper[4846]: > Oct 05 07:11:20 crc kubenswrapper[4846]: E1005 07:11:20.810458 4846 kuberuntime_container.go:691] "PreStop hook failed" err=< Oct 05 07:11:20 crc kubenswrapper[4846]: command '/usr/local/bin/container-scripts/stop-ovsdb-server.sh' exited with 137: ++ dirname /usr/local/bin/container-scripts/stop-ovsdb-server.sh Oct 05 07:11:20 crc kubenswrapper[4846]: + source /usr/local/bin/container-scripts/functions Oct 05 07:11:20 crc kubenswrapper[4846]: ++ OVNBridge=br-int Oct 05 07:11:20 crc kubenswrapper[4846]: ++ OVNRemote=tcp:localhost:6642 Oct 05 07:11:20 crc kubenswrapper[4846]: ++ OVNEncapType=geneve Oct 05 07:11:20 crc kubenswrapper[4846]: ++ OVNAvailabilityZones= Oct 05 07:11:20 crc kubenswrapper[4846]: ++ EnableChassisAsGateway=true Oct 05 07:11:20 crc kubenswrapper[4846]: ++ PhysicalNetworks= Oct 05 07:11:20 crc kubenswrapper[4846]: ++ OVNHostName= Oct 05 07:11:20 crc kubenswrapper[4846]: ++ DB_FILE=/etc/openvswitch/conf.db Oct 05 07:11:20 crc kubenswrapper[4846]: ++ ovs_dir=/var/lib/openvswitch Oct 05 07:11:20 crc kubenswrapper[4846]: ++ FLOWS_RESTORE_SCRIPT=/var/lib/openvswitch/flows-script Oct 05 07:11:20 crc kubenswrapper[4846]: ++ FLOWS_RESTORE_DIR=/var/lib/openvswitch/saved-flows Oct 05 07:11:20 crc kubenswrapper[4846]: ++ SAFE_TO_STOP_OVSDB_SERVER_SEMAPHORE=/var/lib/openvswitch/is_safe_to_stop_ovsdb_server Oct 05 07:11:20 crc kubenswrapper[4846]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Oct 05 07:11:20 crc kubenswrapper[4846]: + sleep 0.5 Oct 05 07:11:20 crc kubenswrapper[4846]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Oct 05 07:11:20 crc kubenswrapper[4846]: + sleep 0.5 Oct 05 07:11:20 crc kubenswrapper[4846]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Oct 05 07:11:20 crc kubenswrapper[4846]: + sleep 0.5 Oct 05 07:11:20 crc kubenswrapper[4846]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Oct 05 07:11:20 crc kubenswrapper[4846]: + cleanup_ovsdb_server_semaphore Oct 05 07:11:20 crc kubenswrapper[4846]: + rm -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server Oct 05 07:11:20 crc kubenswrapper[4846]: + /usr/share/openvswitch/scripts/ovs-ctl stop --no-ovs-vswitchd Oct 05 07:11:20 crc kubenswrapper[4846]: > pod="openstack/ovn-controller-ovs-sg4vg" podUID="29c0f404-fc7d-4909-b6bf-e1826a092a72" containerName="ovsdb-server" containerID="cri-o://f4fa8732ebd0803a3d57af2df182505f645e900471768d31fb1d43cb0a83e0cf" Oct 05 07:11:20 crc kubenswrapper[4846]: I1005 07:11:20.810503 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-controller-ovs-sg4vg" podUID="29c0f404-fc7d-4909-b6bf-e1826a092a72" containerName="ovsdb-server" containerID="cri-o://f4fa8732ebd0803a3d57af2df182505f645e900471768d31fb1d43cb0a83e0cf" gracePeriod=28 Oct 05 07:11:20 crc kubenswrapper[4846]: I1005 07:11:20.864644 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-cell1-server-0" podUID="626e6bce-7422-4fbd-b978-dfa5cb93de34" containerName="rabbitmq" containerID="cri-o://c63a03c8ae2116953b605cde504ae09f678d59deda18e41527a386da67ee2ce4" gracePeriod=604800 Oct 05 07:11:20 crc kubenswrapper[4846]: I1005 07:11:20.898605 4846 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage08-crc" (UniqueName: "kubernetes.io/local-volume/local-storage08-crc") on node "crc" Oct 05 07:11:20 crc kubenswrapper[4846]: I1005 07:11:20.951500 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87e3744a-b8db-4e76-afa9-0e1129410d4c-ovsdbserver-sb-tls-certs" (OuterVolumeSpecName: "ovsdbserver-sb-tls-certs") pod "87e3744a-b8db-4e76-afa9-0e1129410d4c" (UID: "87e3744a-b8db-4e76-afa9-0e1129410d4c"). InnerVolumeSpecName "ovsdbserver-sb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:11:20 crc kubenswrapper[4846]: I1005 07:11:20.958910 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/87e3744a-b8db-4e76-afa9-0e1129410d4c-ovsdbserver-sb-tls-certs\") pod \"87e3744a-b8db-4e76-afa9-0e1129410d4c\" (UID: \"87e3744a-b8db-4e76-afa9-0e1129410d4c\") " Oct 05 07:11:20 crc kubenswrapper[4846]: W1005 07:11:20.959678 4846 empty_dir.go:500] Warning: Unmount skipped because path does not exist: /var/lib/kubelet/pods/87e3744a-b8db-4e76-afa9-0e1129410d4c/volumes/kubernetes.io~secret/ovsdbserver-sb-tls-certs Oct 05 07:11:20 crc kubenswrapper[4846]: I1005 07:11:20.959712 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87e3744a-b8db-4e76-afa9-0e1129410d4c-ovsdbserver-sb-tls-certs" (OuterVolumeSpecName: "ovsdbserver-sb-tls-certs") pod "87e3744a-b8db-4e76-afa9-0e1129410d4c" (UID: "87e3744a-b8db-4e76-afa9-0e1129410d4c"). InnerVolumeSpecName "ovsdbserver-sb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:11:20 crc kubenswrapper[4846]: I1005 07:11:20.960091 4846 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/87e3744a-b8db-4e76-afa9-0e1129410d4c-ovsdbserver-sb-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:20 crc kubenswrapper[4846]: I1005 07:11:20.960119 4846 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/19f3a362-4973-4437-ac49-de71508aa4f2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:20 crc kubenswrapper[4846]: I1005 07:11:20.960135 4846 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/87e3744a-b8db-4e76-afa9-0e1129410d4c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:20 crc kubenswrapper[4846]: I1005 07:11:20.960145 4846 reconciler_common.go:293] "Volume detached for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:20 crc kubenswrapper[4846]: E1005 07:11:20.960233 4846 configmap.go:193] Couldn't get configMap openstack/rabbitmq-cell1-config-data: configmap "rabbitmq-cell1-config-data" not found Oct 05 07:11:20 crc kubenswrapper[4846]: E1005 07:11:20.960284 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/626e6bce-7422-4fbd-b978-dfa5cb93de34-config-data podName:626e6bce-7422-4fbd-b978-dfa5cb93de34 nodeName:}" failed. No retries permitted until 2025-10-05 07:11:22.960267218 +0000 UTC m=+1405.201119993 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/626e6bce-7422-4fbd-b978-dfa5cb93de34-config-data") pod "rabbitmq-cell1-server-0" (UID: "626e6bce-7422-4fbd-b978-dfa5cb93de34") : configmap "rabbitmq-cell1-config-data" not found Oct 05 07:11:21 crc kubenswrapper[4846]: E1005 07:11:21.011275 4846 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of b585eb7d447d9af1ab5bd601fbb7ea5ad51797c817722421db131f2fba3c2373 is running failed: container process not found" containerID="b585eb7d447d9af1ab5bd601fbb7ea5ad51797c817722421db131f2fba3c2373" cmd=["/usr/bin/pidof","ovsdb-server"] Oct 05 07:11:21 crc kubenswrapper[4846]: E1005 07:11:21.013609 4846 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of b585eb7d447d9af1ab5bd601fbb7ea5ad51797c817722421db131f2fba3c2373 is running failed: container process not found" containerID="b585eb7d447d9af1ab5bd601fbb7ea5ad51797c817722421db131f2fba3c2373" cmd=["/usr/bin/pidof","ovsdb-server"] Oct 05 07:11:21 crc kubenswrapper[4846]: E1005 07:11:21.013960 4846 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of b585eb7d447d9af1ab5bd601fbb7ea5ad51797c817722421db131f2fba3c2373 is running failed: container process not found" containerID="b585eb7d447d9af1ab5bd601fbb7ea5ad51797c817722421db131f2fba3c2373" cmd=["/usr/bin/pidof","ovsdb-server"] Oct 05 07:11:21 crc kubenswrapper[4846]: E1005 07:11:21.014003 4846 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of b585eb7d447d9af1ab5bd601fbb7ea5ad51797c817722421db131f2fba3c2373 is running failed: container process not found" probeType="Readiness" pod="openstack/ovsdbserver-nb-0" podUID="bdf95824-fa04-4a83-92cc-7e36b4d00289" containerName="ovsdbserver-nb" Oct 05 07:11:21 crc kubenswrapper[4846]: I1005 07:11:21.021589 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-c7d6584f5-dxcgg" Oct 05 07:11:21 crc kubenswrapper[4846]: I1005 07:11:21.032932 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/19f3a362-4973-4437-ac49-de71508aa4f2-metrics-certs-tls-certs" (OuterVolumeSpecName: "metrics-certs-tls-certs") pod "19f3a362-4973-4437-ac49-de71508aa4f2" (UID: "19f3a362-4973-4437-ac49-de71508aa4f2"). InnerVolumeSpecName "metrics-certs-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:11:21 crc kubenswrapper[4846]: I1005 07:11:21.035592 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_bdf95824-fa04-4a83-92cc-7e36b4d00289/ovsdbserver-nb/0.log" Oct 05 07:11:21 crc kubenswrapper[4846]: I1005 07:11:21.035710 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Oct 05 07:11:21 crc kubenswrapper[4846]: I1005 07:11:21.083495 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndbcluster-nb-etc-ovn\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"bdf95824-fa04-4a83-92cc-7e36b4d00289\" (UID: \"bdf95824-fa04-4a83-92cc-7e36b4d00289\") " Oct 05 07:11:21 crc kubenswrapper[4846]: I1005 07:11:21.083897 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7a5c4c82-4b3f-4314-a32d-db5398ca8a9e-config\") pod \"7a5c4c82-4b3f-4314-a32d-db5398ca8a9e\" (UID: \"7a5c4c82-4b3f-4314-a32d-db5398ca8a9e\") " Oct 05 07:11:21 crc kubenswrapper[4846]: I1005 07:11:21.083930 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/bdf95824-fa04-4a83-92cc-7e36b4d00289-metrics-certs-tls-certs\") pod \"bdf95824-fa04-4a83-92cc-7e36b4d00289\" (UID: \"bdf95824-fa04-4a83-92cc-7e36b4d00289\") " Oct 05 07:11:21 crc kubenswrapper[4846]: I1005 07:11:21.083955 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/bdf95824-fa04-4a83-92cc-7e36b4d00289-ovsdb-rundir\") pod \"bdf95824-fa04-4a83-92cc-7e36b4d00289\" (UID: \"bdf95824-fa04-4a83-92cc-7e36b4d00289\") " Oct 05 07:11:21 crc kubenswrapper[4846]: I1005 07:11:21.084022 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/bdf95824-fa04-4a83-92cc-7e36b4d00289-ovsdbserver-nb-tls-certs\") pod \"bdf95824-fa04-4a83-92cc-7e36b4d00289\" (UID: \"bdf95824-fa04-4a83-92cc-7e36b4d00289\") " Oct 05 07:11:21 crc kubenswrapper[4846]: I1005 07:11:21.084099 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7a5c4c82-4b3f-4314-a32d-db5398ca8a9e-dns-svc\") pod \"7a5c4c82-4b3f-4314-a32d-db5398ca8a9e\" (UID: \"7a5c4c82-4b3f-4314-a32d-db5398ca8a9e\") " Oct 05 07:11:21 crc kubenswrapper[4846]: I1005 07:11:21.084159 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8wmgl\" (UniqueName: \"kubernetes.io/projected/7a5c4c82-4b3f-4314-a32d-db5398ca8a9e-kube-api-access-8wmgl\") pod \"7a5c4c82-4b3f-4314-a32d-db5398ca8a9e\" (UID: \"7a5c4c82-4b3f-4314-a32d-db5398ca8a9e\") " Oct 05 07:11:21 crc kubenswrapper[4846]: I1005 07:11:21.088323 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7a5c4c82-4b3f-4314-a32d-db5398ca8a9e-ovsdbserver-sb\") pod \"7a5c4c82-4b3f-4314-a32d-db5398ca8a9e\" (UID: \"7a5c4c82-4b3f-4314-a32d-db5398ca8a9e\") " Oct 05 07:11:21 crc kubenswrapper[4846]: I1005 07:11:21.088371 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bdf95824-fa04-4a83-92cc-7e36b4d00289-combined-ca-bundle\") pod \"bdf95824-fa04-4a83-92cc-7e36b4d00289\" (UID: \"bdf95824-fa04-4a83-92cc-7e36b4d00289\") " Oct 05 07:11:21 crc kubenswrapper[4846]: I1005 07:11:21.088397 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7a5c4c82-4b3f-4314-a32d-db5398ca8a9e-ovsdbserver-nb\") pod \"7a5c4c82-4b3f-4314-a32d-db5398ca8a9e\" (UID: \"7a5c4c82-4b3f-4314-a32d-db5398ca8a9e\") " Oct 05 07:11:21 crc kubenswrapper[4846]: I1005 07:11:21.088438 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7a5c4c82-4b3f-4314-a32d-db5398ca8a9e-dns-swift-storage-0\") pod \"7a5c4c82-4b3f-4314-a32d-db5398ca8a9e\" (UID: \"7a5c4c82-4b3f-4314-a32d-db5398ca8a9e\") " Oct 05 07:11:21 crc kubenswrapper[4846]: I1005 07:11:21.088461 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b9m6b\" (UniqueName: \"kubernetes.io/projected/bdf95824-fa04-4a83-92cc-7e36b4d00289-kube-api-access-b9m6b\") pod \"bdf95824-fa04-4a83-92cc-7e36b4d00289\" (UID: \"bdf95824-fa04-4a83-92cc-7e36b4d00289\") " Oct 05 07:11:21 crc kubenswrapper[4846]: I1005 07:11:21.088531 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/bdf95824-fa04-4a83-92cc-7e36b4d00289-scripts\") pod \"bdf95824-fa04-4a83-92cc-7e36b4d00289\" (UID: \"bdf95824-fa04-4a83-92cc-7e36b4d00289\") " Oct 05 07:11:21 crc kubenswrapper[4846]: I1005 07:11:21.088606 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bdf95824-fa04-4a83-92cc-7e36b4d00289-config\") pod \"bdf95824-fa04-4a83-92cc-7e36b4d00289\" (UID: \"bdf95824-fa04-4a83-92cc-7e36b4d00289\") " Oct 05 07:11:21 crc kubenswrapper[4846]: I1005 07:11:21.102627 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bdf95824-fa04-4a83-92cc-7e36b4d00289-ovsdb-rundir" (OuterVolumeSpecName: "ovsdb-rundir") pod "bdf95824-fa04-4a83-92cc-7e36b4d00289" (UID: "bdf95824-fa04-4a83-92cc-7e36b4d00289"). InnerVolumeSpecName "ovsdb-rundir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 07:11:21 crc kubenswrapper[4846]: I1005 07:11:21.103072 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bdf95824-fa04-4a83-92cc-7e36b4d00289-config" (OuterVolumeSpecName: "config") pod "bdf95824-fa04-4a83-92cc-7e36b4d00289" (UID: "bdf95824-fa04-4a83-92cc-7e36b4d00289"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 07:11:21 crc kubenswrapper[4846]: I1005 07:11:21.103182 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bdf95824-fa04-4a83-92cc-7e36b4d00289-scripts" (OuterVolumeSpecName: "scripts") pod "bdf95824-fa04-4a83-92cc-7e36b4d00289" (UID: "bdf95824-fa04-4a83-92cc-7e36b4d00289"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 07:11:21 crc kubenswrapper[4846]: I1005 07:11:21.104197 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage11-crc" (OuterVolumeSpecName: "ovndbcluster-nb-etc-ovn") pod "bdf95824-fa04-4a83-92cc-7e36b4d00289" (UID: "bdf95824-fa04-4a83-92cc-7e36b4d00289"). InnerVolumeSpecName "local-storage11-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 05 07:11:21 crc kubenswrapper[4846]: I1005 07:11:21.105008 4846 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bdf95824-fa04-4a83-92cc-7e36b4d00289-config\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:21 crc kubenswrapper[4846]: I1005 07:11:21.105033 4846 reconciler_common.go:293] "Volume detached for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/19f3a362-4973-4437-ac49-de71508aa4f2-metrics-certs-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:21 crc kubenswrapper[4846]: I1005 07:11:21.105048 4846 reconciler_common.go:293] "Volume detached for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/bdf95824-fa04-4a83-92cc-7e36b4d00289-ovsdb-rundir\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:21 crc kubenswrapper[4846]: I1005 07:11:21.159392 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bdf95824-fa04-4a83-92cc-7e36b4d00289-kube-api-access-b9m6b" (OuterVolumeSpecName: "kube-api-access-b9m6b") pod "bdf95824-fa04-4a83-92cc-7e36b4d00289" (UID: "bdf95824-fa04-4a83-92cc-7e36b4d00289"). InnerVolumeSpecName "kube-api-access-b9m6b". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:11:21 crc kubenswrapper[4846]: I1005 07:11:21.173445 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 05 07:11:21 crc kubenswrapper[4846]: I1005 07:11:21.184250 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 05 07:11:21 crc kubenswrapper[4846]: I1005 07:11:21.184535 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="b7de2524-bf5c-4668-af74-7584edf72c8b" containerName="nova-scheduler-scheduler" containerID="cri-o://316083fa9d64f1f90ee62226b3c28e38700ed5382c669302dda9655852943dab" gracePeriod=30 Oct 05 07:11:21 crc kubenswrapper[4846]: I1005 07:11:21.206444 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/98fc38b8-a0ee-46c9-9d3f-9c15e8355b9d-openstack-config\") pod \"98fc38b8-a0ee-46c9-9d3f-9c15e8355b9d\" (UID: \"98fc38b8-a0ee-46c9-9d3f-9c15e8355b9d\") " Oct 05 07:11:21 crc kubenswrapper[4846]: I1005 07:11:21.206539 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/98fc38b8-a0ee-46c9-9d3f-9c15e8355b9d-openstack-config-secret\") pod \"98fc38b8-a0ee-46c9-9d3f-9c15e8355b9d\" (UID: \"98fc38b8-a0ee-46c9-9d3f-9c15e8355b9d\") " Oct 05 07:11:21 crc kubenswrapper[4846]: I1005 07:11:21.206573 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/98fc38b8-a0ee-46c9-9d3f-9c15e8355b9d-combined-ca-bundle\") pod \"98fc38b8-a0ee-46c9-9d3f-9c15e8355b9d\" (UID: \"98fc38b8-a0ee-46c9-9d3f-9c15e8355b9d\") " Oct 05 07:11:21 crc kubenswrapper[4846]: I1005 07:11:21.206787 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qdvx8\" (UniqueName: \"kubernetes.io/projected/98fc38b8-a0ee-46c9-9d3f-9c15e8355b9d-kube-api-access-qdvx8\") pod \"98fc38b8-a0ee-46c9-9d3f-9c15e8355b9d\" (UID: \"98fc38b8-a0ee-46c9-9d3f-9c15e8355b9d\") " Oct 05 07:11:21 crc kubenswrapper[4846]: I1005 07:11:21.207600 4846 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" " Oct 05 07:11:21 crc kubenswrapper[4846]: I1005 07:11:21.207618 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b9m6b\" (UniqueName: \"kubernetes.io/projected/bdf95824-fa04-4a83-92cc-7e36b4d00289-kube-api-access-b9m6b\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:21 crc kubenswrapper[4846]: I1005 07:11:21.207632 4846 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/bdf95824-fa04-4a83-92cc-7e36b4d00289-scripts\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:21 crc kubenswrapper[4846]: I1005 07:11:21.209087 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 05 07:11:21 crc kubenswrapper[4846]: I1005 07:11:21.209317 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-conductor-0" podUID="33193805-7aa6-406e-8cad-0ebbb1228ef2" containerName="nova-cell1-conductor-conductor" containerID="cri-o://a03ec35e533cac1f50168f28070b82ea570152d6d222e9cf097e6f662b34b418" gracePeriod=30 Oct 05 07:11:21 crc kubenswrapper[4846]: I1005 07:11:21.220518 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7a5c4c82-4b3f-4314-a32d-db5398ca8a9e-kube-api-access-8wmgl" (OuterVolumeSpecName: "kube-api-access-8wmgl") pod "7a5c4c82-4b3f-4314-a32d-db5398ca8a9e" (UID: "7a5c4c82-4b3f-4314-a32d-db5398ca8a9e"). InnerVolumeSpecName "kube-api-access-8wmgl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:11:21 crc kubenswrapper[4846]: I1005 07:11:21.230410 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/98fc38b8-a0ee-46c9-9d3f-9c15e8355b9d-kube-api-access-qdvx8" (OuterVolumeSpecName: "kube-api-access-qdvx8") pod "98fc38b8-a0ee-46c9-9d3f-9c15e8355b9d" (UID: "98fc38b8-a0ee-46c9-9d3f-9c15e8355b9d"). InnerVolumeSpecName "kube-api-access-qdvx8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:11:21 crc kubenswrapper[4846]: I1005 07:11:21.233735 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-s6xg9"] Oct 05 07:11:21 crc kubenswrapper[4846]: I1005 07:11:21.270470 4846 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage11-crc" (UniqueName: "kubernetes.io/local-volume/local-storage11-crc") on node "crc" Oct 05 07:11:21 crc kubenswrapper[4846]: I1005 07:11:21.273828 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-s6xg9"] Oct 05 07:11:21 crc kubenswrapper[4846]: I1005 07:11:21.276164 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bdf95824-fa04-4a83-92cc-7e36b4d00289-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "bdf95824-fa04-4a83-92cc-7e36b4d00289" (UID: "bdf95824-fa04-4a83-92cc-7e36b4d00289"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:11:21 crc kubenswrapper[4846]: I1005 07:11:21.292327 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 05 07:11:21 crc kubenswrapper[4846]: I1005 07:11:21.292560 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell0-conductor-0" podUID="98b50269-c6d4-42f2-8cef-afc127bfccb1" containerName="nova-cell0-conductor-conductor" containerID="cri-o://055332eb57f5e6e889d548a1b4717caea296e84dc70616d5bdb53dd34aaa0697" gracePeriod=30 Oct 05 07:11:21 crc kubenswrapper[4846]: I1005 07:11:21.301710 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/98fc38b8-a0ee-46c9-9d3f-9c15e8355b9d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "98fc38b8-a0ee-46c9-9d3f-9c15e8355b9d" (UID: "98fc38b8-a0ee-46c9-9d3f-9c15e8355b9d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:11:21 crc kubenswrapper[4846]: I1005 07:11:21.308249 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-88lpl"] Oct 05 07:11:21 crc kubenswrapper[4846]: I1005 07:11:21.310431 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8wmgl\" (UniqueName: \"kubernetes.io/projected/7a5c4c82-4b3f-4314-a32d-db5398ca8a9e-kube-api-access-8wmgl\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:21 crc kubenswrapper[4846]: I1005 07:11:21.310457 4846 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bdf95824-fa04-4a83-92cc-7e36b4d00289-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:21 crc kubenswrapper[4846]: I1005 07:11:21.310466 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qdvx8\" (UniqueName: \"kubernetes.io/projected/98fc38b8-a0ee-46c9-9d3f-9c15e8355b9d-kube-api-access-qdvx8\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:21 crc kubenswrapper[4846]: I1005 07:11:21.310474 4846 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/98fc38b8-a0ee-46c9-9d3f-9c15e8355b9d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:21 crc kubenswrapper[4846]: I1005 07:11:21.310486 4846 reconciler_common.go:293] "Volume detached for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:21 crc kubenswrapper[4846]: I1005 07:11:21.315905 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-88lpl"] Oct 05 07:11:21 crc kubenswrapper[4846]: I1005 07:11:21.329280 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/98fc38b8-a0ee-46c9-9d3f-9c15e8355b9d-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "98fc38b8-a0ee-46c9-9d3f-9c15e8355b9d" (UID: "98fc38b8-a0ee-46c9-9d3f-9c15e8355b9d"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 07:11:21 crc kubenswrapper[4846]: I1005 07:11:21.353978 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87e3744a-b8db-4e76-afa9-0e1129410d4c-metrics-certs-tls-certs" (OuterVolumeSpecName: "metrics-certs-tls-certs") pod "87e3744a-b8db-4e76-afa9-0e1129410d4c" (UID: "87e3744a-b8db-4e76-afa9-0e1129410d4c"). InnerVolumeSpecName "metrics-certs-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:11:21 crc kubenswrapper[4846]: I1005 07:11:21.358458 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bdf95824-fa04-4a83-92cc-7e36b4d00289-ovsdbserver-nb-tls-certs" (OuterVolumeSpecName: "ovsdbserver-nb-tls-certs") pod "bdf95824-fa04-4a83-92cc-7e36b4d00289" (UID: "bdf95824-fa04-4a83-92cc-7e36b4d00289"). InnerVolumeSpecName "ovsdbserver-nb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:11:21 crc kubenswrapper[4846]: I1005 07:11:21.358806 4846 generic.go:334] "Generic (PLEG): container finished" podID="c31937eb-9836-461b-9230-5e06ce7a99e8" containerID="85a65281831348ab23593ec8a61436c9871720d821d057e65d1423a5e8c5e79c" exitCode=143 Oct 05 07:11:21 crc kubenswrapper[4846]: I1005 07:11:21.358875 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-57674d6c54-zn5nq" event={"ID":"c31937eb-9836-461b-9230-5e06ce7a99e8","Type":"ContainerDied","Data":"85a65281831348ab23593ec8a61436c9871720d821d057e65d1423a5e8c5e79c"} Oct 05 07:11:21 crc kubenswrapper[4846]: I1005 07:11:21.361139 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7a5c4c82-4b3f-4314-a32d-db5398ca8a9e-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "7a5c4c82-4b3f-4314-a32d-db5398ca8a9e" (UID: "7a5c4c82-4b3f-4314-a32d-db5398ca8a9e"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 07:11:21 crc kubenswrapper[4846]: I1005 07:11:21.363863 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_87e3744a-b8db-4e76-afa9-0e1129410d4c/ovsdbserver-sb/0.log" Oct 05 07:11:21 crc kubenswrapper[4846]: I1005 07:11:21.363965 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"87e3744a-b8db-4e76-afa9-0e1129410d4c","Type":"ContainerDied","Data":"d45d3167f19b6fa1537d87164850bb4f2bddcfe3fcc965f46e24c0042c091067"} Oct 05 07:11:21 crc kubenswrapper[4846]: I1005 07:11:21.364004 4846 scope.go:117] "RemoveContainer" containerID="4110f42bd835d3c1d5633761ee0d01337fb709f2d6a2517116e7ed33818b184d" Oct 05 07:11:21 crc kubenswrapper[4846]: I1005 07:11:21.364013 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Oct 05 07:11:21 crc kubenswrapper[4846]: I1005 07:11:21.367363 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/98fc38b8-a0ee-46c9-9d3f-9c15e8355b9d-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "98fc38b8-a0ee-46c9-9d3f-9c15e8355b9d" (UID: "98fc38b8-a0ee-46c9-9d3f-9c15e8355b9d"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:11:21 crc kubenswrapper[4846]: I1005 07:11:21.368365 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7a5c4c82-4b3f-4314-a32d-db5398ca8a9e-config" (OuterVolumeSpecName: "config") pod "7a5c4c82-4b3f-4314-a32d-db5398ca8a9e" (UID: "7a5c4c82-4b3f-4314-a32d-db5398ca8a9e"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 07:11:21 crc kubenswrapper[4846]: I1005 07:11:21.397943 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7a5c4c82-4b3f-4314-a32d-db5398ca8a9e-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "7a5c4c82-4b3f-4314-a32d-db5398ca8a9e" (UID: "7a5c4c82-4b3f-4314-a32d-db5398ca8a9e"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 07:11:21 crc kubenswrapper[4846]: I1005 07:11:21.410704 4846 generic.go:334] "Generic (PLEG): container finished" podID="bfe96e0c-b9fc-438d-9b17-bec246d48446" containerID="470aec4a8b7085d2d3371c5366c57378850aca957d883939f98653f186109a67" exitCode=0 Oct 05 07:11:21 crc kubenswrapper[4846]: I1005 07:11:21.410778 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"bfe96e0c-b9fc-438d-9b17-bec246d48446","Type":"ContainerDied","Data":"470aec4a8b7085d2d3371c5366c57378850aca957d883939f98653f186109a67"} Oct 05 07:11:21 crc kubenswrapper[4846]: I1005 07:11:21.412636 4846 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/98fc38b8-a0ee-46c9-9d3f-9c15e8355b9d-openstack-config\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:21 crc kubenswrapper[4846]: I1005 07:11:21.412672 4846 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/98fc38b8-a0ee-46c9-9d3f-9c15e8355b9d-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:21 crc kubenswrapper[4846]: I1005 07:11:21.412689 4846 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7a5c4c82-4b3f-4314-a32d-db5398ca8a9e-config\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:21 crc kubenswrapper[4846]: I1005 07:11:21.412702 4846 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/bdf95824-fa04-4a83-92cc-7e36b4d00289-ovsdbserver-nb-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:21 crc kubenswrapper[4846]: I1005 07:11:21.412712 4846 reconciler_common.go:293] "Volume detached for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/87e3744a-b8db-4e76-afa9-0e1129410d4c-metrics-certs-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:21 crc kubenswrapper[4846]: I1005 07:11:21.412721 4846 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7a5c4c82-4b3f-4314-a32d-db5398ca8a9e-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:21 crc kubenswrapper[4846]: I1005 07:11:21.412732 4846 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7a5c4c82-4b3f-4314-a32d-db5398ca8a9e-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:21 crc kubenswrapper[4846]: I1005 07:11:21.415083 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_bdf95824-fa04-4a83-92cc-7e36b4d00289/ovsdbserver-nb/0.log" Oct 05 07:11:21 crc kubenswrapper[4846]: I1005 07:11:21.418744 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"bdf95824-fa04-4a83-92cc-7e36b4d00289","Type":"ContainerDied","Data":"241c79da4bec1edce442e764915bccd26a2210f02bcf72b47d133047c73ea3e6"} Oct 05 07:11:21 crc kubenswrapper[4846]: I1005 07:11:21.424260 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Oct 05 07:11:21 crc kubenswrapper[4846]: I1005 07:11:21.425829 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement2760-account-delete-bvfdp"] Oct 05 07:11:21 crc kubenswrapper[4846]: I1005 07:11:21.426805 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-c7d6584f5-dxcgg" event={"ID":"7a5c4c82-4b3f-4314-a32d-db5398ca8a9e","Type":"ContainerDied","Data":"ec97db54b2342d26de8e49361527359256837278cf11d3cc5355609f7e73865b"} Oct 05 07:11:21 crc kubenswrapper[4846]: I1005 07:11:21.426934 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-c7d6584f5-dxcgg" Oct 05 07:11:21 crc kubenswrapper[4846]: I1005 07:11:21.427963 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7a5c4c82-4b3f-4314-a32d-db5398ca8a9e-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "7a5c4c82-4b3f-4314-a32d-db5398ca8a9e" (UID: "7a5c4c82-4b3f-4314-a32d-db5398ca8a9e"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 07:11:21 crc kubenswrapper[4846]: I1005 07:11:21.428447 4846 generic.go:334] "Generic (PLEG): container finished" podID="a72c0cf5-8cd9-4f5f-8a2b-25bf41ad0638" containerID="1ac7cfeb9e27da195d8d27a5e4c8617a950c62fafc93500f5db134c1b1af30c2" exitCode=0 Oct 05 07:11:21 crc kubenswrapper[4846]: I1005 07:11:21.428507 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-85bbbcfb7f-kwfjt" event={"ID":"a72c0cf5-8cd9-4f5f-8a2b-25bf41ad0638","Type":"ContainerDied","Data":"1ac7cfeb9e27da195d8d27a5e4c8617a950c62fafc93500f5db134c1b1af30c2"} Oct 05 07:11:21 crc kubenswrapper[4846]: I1005 07:11:21.429850 4846 generic.go:334] "Generic (PLEG): container finished" podID="6e9e53c6-4cf2-4988-89b4-c69471f616aa" containerID="a2a13c5c05a60506397f6b2d96fab2cded887ebfc95de2c51ee9dd249fae3050" exitCode=143 Oct 05 07:11:21 crc kubenswrapper[4846]: I1005 07:11:21.429894 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"6e9e53c6-4cf2-4988-89b4-c69471f616aa","Type":"ContainerDied","Data":"a2a13c5c05a60506397f6b2d96fab2cded887ebfc95de2c51ee9dd249fae3050"} Oct 05 07:11:21 crc kubenswrapper[4846]: I1005 07:11:21.443481 4846 generic.go:334] "Generic (PLEG): container finished" podID="8d55f55a-2d79-4c01-8b0f-d67cefb842ff" containerID="bca1da26068e641006909835041107dfa8983891c244081dc35a6bf70c412d1d" exitCode=143 Oct 05 07:11:21 crc kubenswrapper[4846]: I1005 07:11:21.443564 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-77664b8dfd-5hkqj" event={"ID":"8d55f55a-2d79-4c01-8b0f-d67cefb842ff","Type":"ContainerDied","Data":"bca1da26068e641006909835041107dfa8983891c244081dc35a6bf70c412d1d"} Oct 05 07:11:21 crc kubenswrapper[4846]: I1005 07:11:21.445495 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cindercf3d-account-delete-jw5tr"] Oct 05 07:11:21 crc kubenswrapper[4846]: I1005 07:11:21.451419 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7a5c4c82-4b3f-4314-a32d-db5398ca8a9e-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "7a5c4c82-4b3f-4314-a32d-db5398ca8a9e" (UID: "7a5c4c82-4b3f-4314-a32d-db5398ca8a9e"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 07:11:21 crc kubenswrapper[4846]: I1005 07:11:21.468592 4846 scope.go:117] "RemoveContainer" containerID="097a50245fd880a8216f94f3a81c3b749e8486580e37064c166b13f4fa1fc83c" Oct 05 07:11:21 crc kubenswrapper[4846]: I1005 07:11:21.474294 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovsdbserver-sb-0"] Oct 05 07:11:21 crc kubenswrapper[4846]: I1005 07:11:21.480274 4846 generic.go:334] "Generic (PLEG): container finished" podID="98fc38b8-a0ee-46c9-9d3f-9c15e8355b9d" containerID="5edadd35f5bb4b18c0ce1cfc15ed5663752e987563a9e6ef8438aa48f0d79ae2" exitCode=137 Oct 05 07:11:21 crc kubenswrapper[4846]: I1005 07:11:21.480419 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 05 07:11:21 crc kubenswrapper[4846]: W1005 07:11:21.490383 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda32cdd03_1398_4055_b051_62adde35200f.slice/crio-63fc4076208dfa8214e57a8f6a726a5fbe145fe441f981f1f2d4a2f7ce1f81bc WatchSource:0}: Error finding container 63fc4076208dfa8214e57a8f6a726a5fbe145fe441f981f1f2d4a2f7ce1f81bc: Status 404 returned error can't find the container with id 63fc4076208dfa8214e57a8f6a726a5fbe145fe441f981f1f2d4a2f7ce1f81bc Oct 05 07:11:21 crc kubenswrapper[4846]: I1005 07:11:21.506523 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovsdbserver-sb-0"] Oct 05 07:11:21 crc kubenswrapper[4846]: I1005 07:11:21.518700 4846 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7a5c4c82-4b3f-4314-a32d-db5398ca8a9e-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:21 crc kubenswrapper[4846]: I1005 07:11:21.518724 4846 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7a5c4c82-4b3f-4314-a32d-db5398ca8a9e-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:21 crc kubenswrapper[4846]: W1005 07:11:21.519420 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4d3616b4_9f2c_4c62_9760_0300b02d0dbb.slice/crio-1da9c21171f99446de2b2aa2c3001e53829eca259615385092f0f013986911bf WatchSource:0}: Error finding container 1da9c21171f99446de2b2aa2c3001e53829eca259615385092f0f013986911bf: Status 404 returned error can't find the container with id 1da9c21171f99446de2b2aa2c3001e53829eca259615385092f0f013986911bf Oct 05 07:11:21 crc kubenswrapper[4846]: I1005 07:11:21.524319 4846 generic.go:334] "Generic (PLEG): container finished" podID="3995994c-2a3f-400f-9a64-b564759b20b2" containerID="afd676f2477334d96744cd37d02c69587dea11053d28e2ae166c10deda1b1560" exitCode=143 Oct 05 07:11:21 crc kubenswrapper[4846]: I1005 07:11:21.524368 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-78995b9f79-nwl7v" event={"ID":"3995994c-2a3f-400f-9a64-b564759b20b2","Type":"ContainerDied","Data":"afd676f2477334d96744cd37d02c69587dea11053d28e2ae166c10deda1b1560"} Oct 05 07:11:21 crc kubenswrapper[4846]: I1005 07:11:21.538925 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glancebc5f-account-delete-hkzcl"] Oct 05 07:11:21 crc kubenswrapper[4846]: I1005 07:11:21.583484 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican8fc5-account-delete-gjp47"] Oct 05 07:11:21 crc kubenswrapper[4846]: I1005 07:11:21.606105 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutrona5c0-account-delete-q5wg4"] Oct 05 07:11:21 crc kubenswrapper[4846]: I1005 07:11:21.606450 4846 generic.go:334] "Generic (PLEG): container finished" podID="29c0f404-fc7d-4909-b6bf-e1826a092a72" containerID="f4fa8732ebd0803a3d57af2df182505f645e900471768d31fb1d43cb0a83e0cf" exitCode=0 Oct 05 07:11:21 crc kubenswrapper[4846]: I1005 07:11:21.606531 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-sg4vg" event={"ID":"29c0f404-fc7d-4909-b6bf-e1826a092a72","Type":"ContainerDied","Data":"f4fa8732ebd0803a3d57af2df182505f645e900471768d31fb1d43cb0a83e0cf"} Oct 05 07:11:21 crc kubenswrapper[4846]: I1005 07:11:21.626582 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bdf95824-fa04-4a83-92cc-7e36b4d00289-metrics-certs-tls-certs" (OuterVolumeSpecName: "metrics-certs-tls-certs") pod "bdf95824-fa04-4a83-92cc-7e36b4d00289" (UID: "bdf95824-fa04-4a83-92cc-7e36b4d00289"). InnerVolumeSpecName "metrics-certs-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:11:21 crc kubenswrapper[4846]: I1005 07:11:21.629126 4846 reconciler_common.go:293] "Volume detached for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/bdf95824-fa04-4a83-92cc-7e36b4d00289-metrics-certs-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:21 crc kubenswrapper[4846]: I1005 07:11:21.656070 4846 generic.go:334] "Generic (PLEG): container finished" podID="36dca705-2f72-411e-8d80-ed31d49989cc" containerID="735ae29bd9ab70b9f2083caa08160bc161134c908e991ed9c2afaf1f25347aa0" exitCode=143 Oct 05 07:11:21 crc kubenswrapper[4846]: I1005 07:11:21.656179 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"36dca705-2f72-411e-8d80-ed31d49989cc","Type":"ContainerDied","Data":"735ae29bd9ab70b9f2083caa08160bc161134c908e991ed9c2afaf1f25347aa0"} Oct 05 07:11:21 crc kubenswrapper[4846]: I1005 07:11:21.704078 4846 generic.go:334] "Generic (PLEG): container finished" podID="c35affe0-da4e-448f-aac0-7bba26f89040" containerID="d0a2d9595d1231b4e28c0eff4482c56b148f02336c7fb50f8a4400e74af973ff" exitCode=143 Oct 05 07:11:21 crc kubenswrapper[4846]: I1005 07:11:21.704150 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-65884db4f4-fsk6p" event={"ID":"c35affe0-da4e-448f-aac0-7bba26f89040","Type":"ContainerDied","Data":"d0a2d9595d1231b4e28c0eff4482c56b148f02336c7fb50f8a4400e74af973ff"} Oct 05 07:11:21 crc kubenswrapper[4846]: I1005 07:11:21.796965 4846 generic.go:334] "Generic (PLEG): container finished" podID="62d2ef82-9372-4b55-990e-30eb8e16e439" containerID="8b1645c0bb7485916cb16906718326320b1c0c52c16426fad29ece1124acb1d4" exitCode=0 Oct 05 07:11:21 crc kubenswrapper[4846]: I1005 07:11:21.796991 4846 generic.go:334] "Generic (PLEG): container finished" podID="62d2ef82-9372-4b55-990e-30eb8e16e439" containerID="b448ad891d0bdb677c147854fe487e87246e96353b5d9e033e8ba9ab2dfa1486" exitCode=0 Oct 05 07:11:21 crc kubenswrapper[4846]: I1005 07:11:21.797000 4846 generic.go:334] "Generic (PLEG): container finished" podID="62d2ef82-9372-4b55-990e-30eb8e16e439" containerID="1af615c099267db200d6c220550e8454d08ffe8449631c9c707f1cdad15b401e" exitCode=0 Oct 05 07:11:21 crc kubenswrapper[4846]: I1005 07:11:21.797008 4846 generic.go:334] "Generic (PLEG): container finished" podID="62d2ef82-9372-4b55-990e-30eb8e16e439" containerID="1892fe91962ddb7379e117ea24de43a317054f7a7cd0983a6e06d7d4a4eafef9" exitCode=0 Oct 05 07:11:21 crc kubenswrapper[4846]: I1005 07:11:21.797015 4846 generic.go:334] "Generic (PLEG): container finished" podID="62d2ef82-9372-4b55-990e-30eb8e16e439" containerID="a4edeadbef779fa82267d6b817cf8a19cf59e7146c6c1a8f5c786ad4fea85d4d" exitCode=0 Oct 05 07:11:21 crc kubenswrapper[4846]: I1005 07:11:21.797057 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"62d2ef82-9372-4b55-990e-30eb8e16e439","Type":"ContainerDied","Data":"8b1645c0bb7485916cb16906718326320b1c0c52c16426fad29ece1124acb1d4"} Oct 05 07:11:21 crc kubenswrapper[4846]: I1005 07:11:21.797083 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"62d2ef82-9372-4b55-990e-30eb8e16e439","Type":"ContainerDied","Data":"b448ad891d0bdb677c147854fe487e87246e96353b5d9e033e8ba9ab2dfa1486"} Oct 05 07:11:21 crc kubenswrapper[4846]: I1005 07:11:21.797095 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"62d2ef82-9372-4b55-990e-30eb8e16e439","Type":"ContainerDied","Data":"1af615c099267db200d6c220550e8454d08ffe8449631c9c707f1cdad15b401e"} Oct 05 07:11:21 crc kubenswrapper[4846]: I1005 07:11:21.797104 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"62d2ef82-9372-4b55-990e-30eb8e16e439","Type":"ContainerDied","Data":"1892fe91962ddb7379e117ea24de43a317054f7a7cd0983a6e06d7d4a4eafef9"} Oct 05 07:11:21 crc kubenswrapper[4846]: I1005 07:11:21.797113 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"62d2ef82-9372-4b55-990e-30eb8e16e439","Type":"ContainerDied","Data":"a4edeadbef779fa82267d6b817cf8a19cf59e7146c6c1a8f5c786ad4fea85d4d"} Oct 05 07:11:21 crc kubenswrapper[4846]: I1005 07:11:21.814222 4846 scope.go:117] "RemoveContainer" containerID="e5574edbd5d5e803159cc31cad023d93fa72c648be7c1d91c85d67e4ae1b4f70" Oct 05 07:11:21 crc kubenswrapper[4846]: I1005 07:11:21.830141 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-58xdl_19f3a362-4973-4437-ac49-de71508aa4f2/openstack-network-exporter/0.log" Oct 05 07:11:21 crc kubenswrapper[4846]: I1005 07:11:21.830249 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-58xdl" event={"ID":"19f3a362-4973-4437-ac49-de71508aa4f2","Type":"ContainerDied","Data":"8c40e8e43da6d24404af9690a38549cd8ea25b81acc50916f08729a15332dc27"} Oct 05 07:11:21 crc kubenswrapper[4846]: I1005 07:11:21.830354 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-58xdl" Oct 05 07:11:21 crc kubenswrapper[4846]: I1005 07:11:21.844057 4846 generic.go:334] "Generic (PLEG): container finished" podID="e6e86b1c-87ac-4594-aeba-943947065374" containerID="71ab8b0934bfb90a07c9012633af71114aed35423acecf39d3a2e32cd390b303" exitCode=143 Oct 05 07:11:21 crc kubenswrapper[4846]: I1005 07:11:21.844099 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"e6e86b1c-87ac-4594-aeba-943947065374","Type":"ContainerDied","Data":"71ab8b0934bfb90a07c9012633af71114aed35423acecf39d3a2e32cd390b303"} Oct 05 07:11:21 crc kubenswrapper[4846]: I1005 07:11:21.872304 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-c7d6584f5-dxcgg"] Oct 05 07:11:21 crc kubenswrapper[4846]: I1005 07:11:21.902143 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-c7d6584f5-dxcgg"] Oct 05 07:11:21 crc kubenswrapper[4846]: I1005 07:11:21.935552 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovsdbserver-nb-0"] Oct 05 07:11:21 crc kubenswrapper[4846]: E1005 07:11:21.959975 4846 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="055332eb57f5e6e889d548a1b4717caea296e84dc70616d5bdb53dd34aaa0697" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Oct 05 07:11:21 crc kubenswrapper[4846]: E1005 07:11:21.963369 4846 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="055332eb57f5e6e889d548a1b4717caea296e84dc70616d5bdb53dd34aaa0697" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Oct 05 07:11:21 crc kubenswrapper[4846]: I1005 07:11:21.971041 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovsdbserver-nb-0"] Oct 05 07:11:21 crc kubenswrapper[4846]: E1005 07:11:21.971200 4846 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="055332eb57f5e6e889d548a1b4717caea296e84dc70616d5bdb53dd34aaa0697" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Oct 05 07:11:21 crc kubenswrapper[4846]: E1005 07:11:21.971265 4846 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-cell0-conductor-0" podUID="98b50269-c6d4-42f2-8cef-afc127bfccb1" containerName="nova-cell0-conductor-conductor" Oct 05 07:11:22 crc kubenswrapper[4846]: I1005 07:11:22.093266 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-metrics-58xdl"] Oct 05 07:11:22 crc kubenswrapper[4846]: I1005 07:11:22.121087 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-metrics-58xdl"] Oct 05 07:11:22 crc kubenswrapper[4846]: E1005 07:11:22.147795 4846 configmap.go:193] Couldn't get configMap openstack/rabbitmq-config-data: configmap "rabbitmq-config-data" not found Oct 05 07:11:22 crc kubenswrapper[4846]: E1005 07:11:22.147860 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/35261dbc-cabe-4bf0-88f6-b8613ddec0f1-config-data podName:35261dbc-cabe-4bf0-88f6-b8613ddec0f1 nodeName:}" failed. No retries permitted until 2025-10-05 07:11:26.147845057 +0000 UTC m=+1408.388697832 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/35261dbc-cabe-4bf0-88f6-b8613ddec0f1-config-data") pod "rabbitmq-server-0" (UID: "35261dbc-cabe-4bf0-88f6-b8613ddec0f1") : configmap "rabbitmq-config-data" not found Oct 05 07:11:22 crc kubenswrapper[4846]: I1005 07:11:22.179625 4846 scope.go:117] "RemoveContainer" containerID="b585eb7d447d9af1ab5bd601fbb7ea5ad51797c817722421db131f2fba3c2373" Oct 05 07:11:22 crc kubenswrapper[4846]: I1005 07:11:22.253743 4846 scope.go:117] "RemoveContainer" containerID="283ce773f8b442a1a87a4848af02010620aa0074924483f0a608d4c3f15d5131" Oct 05 07:11:22 crc kubenswrapper[4846]: I1005 07:11:22.297242 4846 scope.go:117] "RemoveContainer" containerID="c51d875ba49113376b3d8a083c86581b5406413bc84095d5348f99c0a3959cc5" Oct 05 07:11:22 crc kubenswrapper[4846]: I1005 07:11:22.361441 4846 scope.go:117] "RemoveContainer" containerID="5edadd35f5bb4b18c0ce1cfc15ed5663752e987563a9e6ef8438aa48f0d79ae2" Oct 05 07:11:22 crc kubenswrapper[4846]: I1005 07:11:22.458449 4846 scope.go:117] "RemoveContainer" containerID="5edadd35f5bb4b18c0ce1cfc15ed5663752e987563a9e6ef8438aa48f0d79ae2" Oct 05 07:11:22 crc kubenswrapper[4846]: E1005 07:11:22.466197 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5edadd35f5bb4b18c0ce1cfc15ed5663752e987563a9e6ef8438aa48f0d79ae2\": container with ID starting with 5edadd35f5bb4b18c0ce1cfc15ed5663752e987563a9e6ef8438aa48f0d79ae2 not found: ID does not exist" containerID="5edadd35f5bb4b18c0ce1cfc15ed5663752e987563a9e6ef8438aa48f0d79ae2" Oct 05 07:11:22 crc kubenswrapper[4846]: I1005 07:11:22.466332 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5edadd35f5bb4b18c0ce1cfc15ed5663752e987563a9e6ef8438aa48f0d79ae2"} err="failed to get container status \"5edadd35f5bb4b18c0ce1cfc15ed5663752e987563a9e6ef8438aa48f0d79ae2\": rpc error: code = NotFound desc = could not find container \"5edadd35f5bb4b18c0ce1cfc15ed5663752e987563a9e6ef8438aa48f0d79ae2\": container with ID starting with 5edadd35f5bb4b18c0ce1cfc15ed5663752e987563a9e6ef8438aa48f0d79ae2 not found: ID does not exist" Oct 05 07:11:22 crc kubenswrapper[4846]: I1005 07:11:22.466411 4846 scope.go:117] "RemoveContainer" containerID="0e19580f528cd14688ae88ab26841372cd0d6d7f86e25927a1ceae0f4c254a43" Oct 05 07:11:22 crc kubenswrapper[4846]: I1005 07:11:22.474639 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Oct 05 07:11:22 crc kubenswrapper[4846]: I1005 07:11:22.500562 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-proxy-984785977-srp2v"] Oct 05 07:11:22 crc kubenswrapper[4846]: I1005 07:11:22.500825 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-proxy-984785977-srp2v" podUID="d11b6df6-6da2-4cb8-a4f9-d87cfd5d4d26" containerName="proxy-httpd" containerID="cri-o://eb3d496875131b7ff843f08ff12b7fdc832265502da5423e00550c759c4bc1b1" gracePeriod=30 Oct 05 07:11:22 crc kubenswrapper[4846]: I1005 07:11:22.500969 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-proxy-984785977-srp2v" podUID="d11b6df6-6da2-4cb8-a4f9-d87cfd5d4d26" containerName="proxy-server" containerID="cri-o://5b9cd4a1e7ff57757729dbae754fc2fd8f2c6d6deb6c33079a5fb777d6c337c4" gracePeriod=30 Oct 05 07:11:22 crc kubenswrapper[4846]: E1005 07:11:22.521746 4846 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 316083fa9d64f1f90ee62226b3c28e38700ed5382c669302dda9655852943dab is running failed: container process not found" containerID="316083fa9d64f1f90ee62226b3c28e38700ed5382c669302dda9655852943dab" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 05 07:11:22 crc kubenswrapper[4846]: E1005 07:11:22.535083 4846 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 316083fa9d64f1f90ee62226b3c28e38700ed5382c669302dda9655852943dab is running failed: container process not found" containerID="316083fa9d64f1f90ee62226b3c28e38700ed5382c669302dda9655852943dab" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 05 07:11:22 crc kubenswrapper[4846]: E1005 07:11:22.540292 4846 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 316083fa9d64f1f90ee62226b3c28e38700ed5382c669302dda9655852943dab is running failed: container process not found" containerID="316083fa9d64f1f90ee62226b3c28e38700ed5382c669302dda9655852943dab" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 05 07:11:22 crc kubenswrapper[4846]: E1005 07:11:22.540356 4846 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 316083fa9d64f1f90ee62226b3c28e38700ed5382c669302dda9655852943dab is running failed: container process not found" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="b7de2524-bf5c-4668-af74-7584edf72c8b" containerName="nova-scheduler-scheduler" Oct 05 07:11:22 crc kubenswrapper[4846]: I1005 07:11:22.541544 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="04602b97-f7db-4bd6-98ce-bd066dbb21f5" path="/var/lib/kubelet/pods/04602b97-f7db-4bd6-98ce-bd066dbb21f5/volumes" Oct 05 07:11:22 crc kubenswrapper[4846]: I1005 07:11:22.542105 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="19f3a362-4973-4437-ac49-de71508aa4f2" path="/var/lib/kubelet/pods/19f3a362-4973-4437-ac49-de71508aa4f2/volumes" Oct 05 07:11:22 crc kubenswrapper[4846]: I1005 07:11:22.542802 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="36efa3ba-3fa1-4ecd-8ad3-9136d31b6070" path="/var/lib/kubelet/pods/36efa3ba-3fa1-4ecd-8ad3-9136d31b6070/volumes" Oct 05 07:11:22 crc kubenswrapper[4846]: I1005 07:11:22.547821 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="514465f9-8432-469d-9260-0958f48f774f" path="/var/lib/kubelet/pods/514465f9-8432-469d-9260-0958f48f774f/volumes" Oct 05 07:11:22 crc kubenswrapper[4846]: I1005 07:11:22.557294 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="68ba5901-a360-4343-8e70-c7e1c9e829f3" path="/var/lib/kubelet/pods/68ba5901-a360-4343-8e70-c7e1c9e829f3/volumes" Oct 05 07:11:22 crc kubenswrapper[4846]: I1005 07:11:22.560815 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6a5f9a12-316d-42cf-a592-e418d03e0849" path="/var/lib/kubelet/pods/6a5f9a12-316d-42cf-a592-e418d03e0849/volumes" Oct 05 07:11:22 crc kubenswrapper[4846]: I1005 07:11:22.561588 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7a5c4c82-4b3f-4314-a32d-db5398ca8a9e" path="/var/lib/kubelet/pods/7a5c4c82-4b3f-4314-a32d-db5398ca8a9e/volumes" Oct 05 07:11:22 crc kubenswrapper[4846]: I1005 07:11:22.562185 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7ac2edb5-3ecf-4162-be56-43585390796a" path="/var/lib/kubelet/pods/7ac2edb5-3ecf-4162-be56-43585390796a/volumes" Oct 05 07:11:22 crc kubenswrapper[4846]: I1005 07:11:22.563373 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87e3744a-b8db-4e76-afa9-0e1129410d4c" path="/var/lib/kubelet/pods/87e3744a-b8db-4e76-afa9-0e1129410d4c/volumes" Oct 05 07:11:22 crc kubenswrapper[4846]: I1005 07:11:22.574763 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/221918b0-eb0e-4a32-8eed-07acd18bd31e-secrets\") pod \"221918b0-eb0e-4a32-8eed-07acd18bd31e\" (UID: \"221918b0-eb0e-4a32-8eed-07acd18bd31e\") " Oct 05 07:11:22 crc kubenswrapper[4846]: I1005 07:11:22.574878 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/221918b0-eb0e-4a32-8eed-07acd18bd31e-config-data-default\") pod \"221918b0-eb0e-4a32-8eed-07acd18bd31e\" (UID: \"221918b0-eb0e-4a32-8eed-07acd18bd31e\") " Oct 05 07:11:22 crc kubenswrapper[4846]: I1005 07:11:22.575014 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/221918b0-eb0e-4a32-8eed-07acd18bd31e-galera-tls-certs\") pod \"221918b0-eb0e-4a32-8eed-07acd18bd31e\" (UID: \"221918b0-eb0e-4a32-8eed-07acd18bd31e\") " Oct 05 07:11:22 crc kubenswrapper[4846]: I1005 07:11:22.575072 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5zl68\" (UniqueName: \"kubernetes.io/projected/221918b0-eb0e-4a32-8eed-07acd18bd31e-kube-api-access-5zl68\") pod \"221918b0-eb0e-4a32-8eed-07acd18bd31e\" (UID: \"221918b0-eb0e-4a32-8eed-07acd18bd31e\") " Oct 05 07:11:22 crc kubenswrapper[4846]: I1005 07:11:22.575152 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/221918b0-eb0e-4a32-8eed-07acd18bd31e-config-data-generated\") pod \"221918b0-eb0e-4a32-8eed-07acd18bd31e\" (UID: \"221918b0-eb0e-4a32-8eed-07acd18bd31e\") " Oct 05 07:11:22 crc kubenswrapper[4846]: I1005 07:11:22.575202 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/221918b0-eb0e-4a32-8eed-07acd18bd31e-operator-scripts\") pod \"221918b0-eb0e-4a32-8eed-07acd18bd31e\" (UID: \"221918b0-eb0e-4a32-8eed-07acd18bd31e\") " Oct 05 07:11:22 crc kubenswrapper[4846]: I1005 07:11:22.575290 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/221918b0-eb0e-4a32-8eed-07acd18bd31e-kolla-config\") pod \"221918b0-eb0e-4a32-8eed-07acd18bd31e\" (UID: \"221918b0-eb0e-4a32-8eed-07acd18bd31e\") " Oct 05 07:11:22 crc kubenswrapper[4846]: I1005 07:11:22.575311 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mysql-db\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"221918b0-eb0e-4a32-8eed-07acd18bd31e\" (UID: \"221918b0-eb0e-4a32-8eed-07acd18bd31e\") " Oct 05 07:11:22 crc kubenswrapper[4846]: I1005 07:11:22.576634 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/221918b0-eb0e-4a32-8eed-07acd18bd31e-combined-ca-bundle\") pod \"221918b0-eb0e-4a32-8eed-07acd18bd31e\" (UID: \"221918b0-eb0e-4a32-8eed-07acd18bd31e\") " Oct 05 07:11:22 crc kubenswrapper[4846]: E1005 07:11:22.581734 4846 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="a03ec35e533cac1f50168f28070b82ea570152d6d222e9cf097e6f662b34b418" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Oct 05 07:11:22 crc kubenswrapper[4846]: I1005 07:11:22.581847 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/221918b0-eb0e-4a32-8eed-07acd18bd31e-config-data-default" (OuterVolumeSpecName: "config-data-default") pod "221918b0-eb0e-4a32-8eed-07acd18bd31e" (UID: "221918b0-eb0e-4a32-8eed-07acd18bd31e"). InnerVolumeSpecName "config-data-default". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 07:11:22 crc kubenswrapper[4846]: I1005 07:11:22.582781 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/221918b0-eb0e-4a32-8eed-07acd18bd31e-config-data-generated" (OuterVolumeSpecName: "config-data-generated") pod "221918b0-eb0e-4a32-8eed-07acd18bd31e" (UID: "221918b0-eb0e-4a32-8eed-07acd18bd31e"). InnerVolumeSpecName "config-data-generated". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 07:11:22 crc kubenswrapper[4846]: I1005 07:11:22.583352 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/221918b0-eb0e-4a32-8eed-07acd18bd31e-kolla-config" (OuterVolumeSpecName: "kolla-config") pod "221918b0-eb0e-4a32-8eed-07acd18bd31e" (UID: "221918b0-eb0e-4a32-8eed-07acd18bd31e"). InnerVolumeSpecName "kolla-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 07:11:22 crc kubenswrapper[4846]: I1005 07:11:22.583560 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/221918b0-eb0e-4a32-8eed-07acd18bd31e-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "221918b0-eb0e-4a32-8eed-07acd18bd31e" (UID: "221918b0-eb0e-4a32-8eed-07acd18bd31e"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 07:11:22 crc kubenswrapper[4846]: I1005 07:11:22.588230 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="98fc38b8-a0ee-46c9-9d3f-9c15e8355b9d" path="/var/lib/kubelet/pods/98fc38b8-a0ee-46c9-9d3f-9c15e8355b9d/volumes" Oct 05 07:11:22 crc kubenswrapper[4846]: I1005 07:11:22.595400 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a63045aa-bdd8-4b7b-954e-73e7799b5859" path="/var/lib/kubelet/pods/a63045aa-bdd8-4b7b-954e-73e7799b5859/volumes" Oct 05 07:11:22 crc kubenswrapper[4846]: E1005 07:11:22.637501 4846 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="a03ec35e533cac1f50168f28070b82ea570152d6d222e9cf097e6f662b34b418" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Oct 05 07:11:22 crc kubenswrapper[4846]: I1005 07:11:22.642198 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/221918b0-eb0e-4a32-8eed-07acd18bd31e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "221918b0-eb0e-4a32-8eed-07acd18bd31e" (UID: "221918b0-eb0e-4a32-8eed-07acd18bd31e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:11:22 crc kubenswrapper[4846]: E1005 07:11:22.643246 4846 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="a03ec35e533cac1f50168f28070b82ea570152d6d222e9cf097e6f662b34b418" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Oct 05 07:11:22 crc kubenswrapper[4846]: E1005 07:11:22.643317 4846 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-cell1-conductor-0" podUID="33193805-7aa6-406e-8cad-0ebbb1228ef2" containerName="nova-cell1-conductor-conductor" Oct 05 07:11:22 crc kubenswrapper[4846]: I1005 07:11:22.675063 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/221918b0-eb0e-4a32-8eed-07acd18bd31e-kube-api-access-5zl68" (OuterVolumeSpecName: "kube-api-access-5zl68") pod "221918b0-eb0e-4a32-8eed-07acd18bd31e" (UID: "221918b0-eb0e-4a32-8eed-07acd18bd31e"). InnerVolumeSpecName "kube-api-access-5zl68". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:11:22 crc kubenswrapper[4846]: I1005 07:11:22.677103 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/221918b0-eb0e-4a32-8eed-07acd18bd31e-secrets" (OuterVolumeSpecName: "secrets") pod "221918b0-eb0e-4a32-8eed-07acd18bd31e" (UID: "221918b0-eb0e-4a32-8eed-07acd18bd31e"). InnerVolumeSpecName "secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:11:22 crc kubenswrapper[4846]: I1005 07:11:22.677216 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bdf95824-fa04-4a83-92cc-7e36b4d00289" path="/var/lib/kubelet/pods/bdf95824-fa04-4a83-92cc-7e36b4d00289/volumes" Oct 05 07:11:22 crc kubenswrapper[4846]: I1005 07:11:22.683371 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d0ccbe35-d3da-4f0d-8c3f-66c016c4604a" path="/var/lib/kubelet/pods/d0ccbe35-d3da-4f0d-8c3f-66c016c4604a/volumes" Oct 05 07:11:22 crc kubenswrapper[4846]: I1005 07:11:22.686021 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d8234451-0d89-47e7-af4c-0263c3b6f58b" path="/var/lib/kubelet/pods/d8234451-0d89-47e7-af4c-0263c3b6f58b/volumes" Oct 05 07:11:22 crc kubenswrapper[4846]: I1005 07:11:22.686982 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d9adf2fe-f612-44f5-a7eb-e137909be76a" path="/var/lib/kubelet/pods/d9adf2fe-f612-44f5-a7eb-e137909be76a/volumes" Oct 05 07:11:22 crc kubenswrapper[4846]: I1005 07:11:22.687631 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fb30bf11-366c-45af-87fd-254f8eca710d" path="/var/lib/kubelet/pods/fb30bf11-366c-45af-87fd-254f8eca710d/volumes" Oct 05 07:11:22 crc kubenswrapper[4846]: I1005 07:11:22.698210 4846 reconciler_common.go:293] "Volume detached for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/221918b0-eb0e-4a32-8eed-07acd18bd31e-config-data-default\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:22 crc kubenswrapper[4846]: I1005 07:11:22.698247 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5zl68\" (UniqueName: \"kubernetes.io/projected/221918b0-eb0e-4a32-8eed-07acd18bd31e-kube-api-access-5zl68\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:22 crc kubenswrapper[4846]: I1005 07:11:22.698262 4846 reconciler_common.go:293] "Volume detached for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/221918b0-eb0e-4a32-8eed-07acd18bd31e-config-data-generated\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:22 crc kubenswrapper[4846]: I1005 07:11:22.698276 4846 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/221918b0-eb0e-4a32-8eed-07acd18bd31e-operator-scripts\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:22 crc kubenswrapper[4846]: I1005 07:11:22.698286 4846 reconciler_common.go:293] "Volume detached for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/221918b0-eb0e-4a32-8eed-07acd18bd31e-kolla-config\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:22 crc kubenswrapper[4846]: I1005 07:11:22.698297 4846 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/221918b0-eb0e-4a32-8eed-07acd18bd31e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:22 crc kubenswrapper[4846]: I1005 07:11:22.698307 4846 reconciler_common.go:293] "Volume detached for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/221918b0-eb0e-4a32-8eed-07acd18bd31e-secrets\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:22 crc kubenswrapper[4846]: I1005 07:11:22.731388 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/221918b0-eb0e-4a32-8eed-07acd18bd31e-galera-tls-certs" (OuterVolumeSpecName: "galera-tls-certs") pod "221918b0-eb0e-4a32-8eed-07acd18bd31e" (UID: "221918b0-eb0e-4a32-8eed-07acd18bd31e"). InnerVolumeSpecName "galera-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:11:22 crc kubenswrapper[4846]: I1005 07:11:22.748553 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage05-crc" (OuterVolumeSpecName: "mysql-db") pod "221918b0-eb0e-4a32-8eed-07acd18bd31e" (UID: "221918b0-eb0e-4a32-8eed-07acd18bd31e"). InnerVolumeSpecName "local-storage05-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 05 07:11:22 crc kubenswrapper[4846]: I1005 07:11:22.804988 4846 reconciler_common.go:293] "Volume detached for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/221918b0-eb0e-4a32-8eed-07acd18bd31e-galera-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:22 crc kubenswrapper[4846]: I1005 07:11:22.805401 4846 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" " Oct 05 07:11:22 crc kubenswrapper[4846]: I1005 07:11:22.847672 4846 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage05-crc" (UniqueName: "kubernetes.io/local-volume/local-storage05-crc") on node "crc" Oct 05 07:11:22 crc kubenswrapper[4846]: I1005 07:11:22.896420 4846 generic.go:334] "Generic (PLEG): container finished" podID="0c0f8308-455e-4544-9f53-0c06a54c6e22" containerID="708780fcc4181d993d6e7d44ef9444fa8e85b47f3a59d10a6fa23aa84fa714b0" exitCode=0 Oct 05 07:11:22 crc kubenswrapper[4846]: I1005 07:11:22.901480 4846 generic.go:334] "Generic (PLEG): container finished" podID="b7de2524-bf5c-4668-af74-7584edf72c8b" containerID="316083fa9d64f1f90ee62226b3c28e38700ed5382c669302dda9655852943dab" exitCode=0 Oct 05 07:11:22 crc kubenswrapper[4846]: I1005 07:11:22.910401 4846 generic.go:334] "Generic (PLEG): container finished" podID="221918b0-eb0e-4a32-8eed-07acd18bd31e" containerID="ec314d49982bc70fa3bb12b1a024bc708bb21c05eda9001d00d334940b04d2e6" exitCode=0 Oct 05 07:11:22 crc kubenswrapper[4846]: I1005 07:11:22.910528 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Oct 05 07:11:22 crc kubenswrapper[4846]: I1005 07:11:22.918915 4846 generic.go:334] "Generic (PLEG): container finished" podID="d11b6df6-6da2-4cb8-a4f9-d87cfd5d4d26" containerID="eb3d496875131b7ff843f08ff12b7fdc832265502da5423e00550c759c4bc1b1" exitCode=0 Oct 05 07:11:22 crc kubenswrapper[4846]: I1005 07:11:22.923580 4846 generic.go:334] "Generic (PLEG): container finished" podID="2adc8ba0-2810-4bf4-bd24-172b97fb9b22" containerID="9e54d631fb403eb9146c131948244ad644f213facfb43b90944167c8827ee59f" exitCode=0 Oct 05 07:11:22 crc kubenswrapper[4846]: I1005 07:11:22.924357 4846 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-4kqnw" podUID="169918d8-597e-4ef4-b00c-fb14c4d0347f" containerName="ovn-controller" probeResult="failure" output=< Oct 05 07:11:22 crc kubenswrapper[4846]: ERROR - Failed to get connection status from ovn-controller, ovn-appctl exit status: 0 Oct 05 07:11:22 crc kubenswrapper[4846]: > Oct 05 07:11:22 crc kubenswrapper[4846]: I1005 07:11:22.925227 4846 reconciler_common.go:293] "Volume detached for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:22 crc kubenswrapper[4846]: I1005 07:11:22.929450 4846 generic.go:334] "Generic (PLEG): container finished" podID="e2f9f9a8-a25f-4a99-b425-4799964d4e24" containerID="ba19323ff93845912587e75b8b10dfd5b4a0513fb4e38a07aab30ad816bdfa3c" exitCode=0 Oct 05 07:11:22 crc kubenswrapper[4846]: E1005 07:11:22.943088 4846 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of f4fa8732ebd0803a3d57af2df182505f645e900471768d31fb1d43cb0a83e0cf is running failed: container process not found" containerID="f4fa8732ebd0803a3d57af2df182505f645e900471768d31fb1d43cb0a83e0cf" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 05 07:11:22 crc kubenswrapper[4846]: I1005 07:11:22.943347 4846 generic.go:334] "Generic (PLEG): container finished" podID="fc35db15-65d4-4cd6-9d09-059822904757" containerID="8b92f361cea3763ba14f3598387de3c1d364a1f14851d48c0df9f7aa1af31db1" exitCode=0 Oct 05 07:11:22 crc kubenswrapper[4846]: E1005 07:11:22.945649 4846 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="22929dba446bc2516169306e09540fa52839ffc6b8d024f494720cc39ad0ab7d" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 05 07:11:23 crc kubenswrapper[4846]: E1005 07:11:22.998523 4846 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of f4fa8732ebd0803a3d57af2df182505f645e900471768d31fb1d43cb0a83e0cf is running failed: container process not found" containerID="f4fa8732ebd0803a3d57af2df182505f645e900471768d31fb1d43cb0a83e0cf" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 05 07:11:23 crc kubenswrapper[4846]: E1005 07:11:23.006487 4846 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of f4fa8732ebd0803a3d57af2df182505f645e900471768d31fb1d43cb0a83e0cf is running failed: container process not found" containerID="f4fa8732ebd0803a3d57af2df182505f645e900471768d31fb1d43cb0a83e0cf" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 05 07:11:23 crc kubenswrapper[4846]: E1005 07:11:23.006587 4846 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of f4fa8732ebd0803a3d57af2df182505f645e900471768d31fb1d43cb0a83e0cf is running failed: container process not found" probeType="Readiness" pod="openstack/ovn-controller-ovs-sg4vg" podUID="29c0f404-fc7d-4909-b6bf-e1826a092a72" containerName="ovsdb-server" Oct 05 07:11:23 crc kubenswrapper[4846]: E1005 07:11:23.030594 4846 configmap.go:193] Couldn't get configMap openstack/rabbitmq-cell1-config-data: configmap "rabbitmq-cell1-config-data" not found Oct 05 07:11:23 crc kubenswrapper[4846]: E1005 07:11:23.030766 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/626e6bce-7422-4fbd-b978-dfa5cb93de34-config-data podName:626e6bce-7422-4fbd-b978-dfa5cb93de34 nodeName:}" failed. No retries permitted until 2025-10-05 07:11:27.030749014 +0000 UTC m=+1409.271601779 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/626e6bce-7422-4fbd-b978-dfa5cb93de34-config-data") pod "rabbitmq-cell1-server-0" (UID: "626e6bce-7422-4fbd-b978-dfa5cb93de34") : configmap "rabbitmq-cell1-config-data" not found Oct 05 07:11:23 crc kubenswrapper[4846]: E1005 07:11:23.033660 4846 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="22929dba446bc2516169306e09540fa52839ffc6b8d024f494720cc39ad0ab7d" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 05 07:11:23 crc kubenswrapper[4846]: E1005 07:11:23.070484 4846 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="22929dba446bc2516169306e09540fa52839ffc6b8d024f494720cc39ad0ab7d" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 05 07:11:23 crc kubenswrapper[4846]: E1005 07:11:23.070558 4846 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-controller-ovs-sg4vg" podUID="29c0f404-fc7d-4909-b6bf-e1826a092a72" containerName="ovs-vswitchd" Oct 05 07:11:23 crc kubenswrapper[4846]: I1005 07:11:23.103834 4846 generic.go:334] "Generic (PLEG): container finished" podID="33193805-7aa6-406e-8cad-0ebbb1228ef2" containerID="a03ec35e533cac1f50168f28070b82ea570152d6d222e9cf097e6f662b34b418" exitCode=0 Oct 05 07:11:23 crc kubenswrapper[4846]: I1005 07:11:23.121311 4846 generic.go:334] "Generic (PLEG): container finished" podID="086cf913-a4a2-41ee-910b-a83be95c69d6" containerID="18de9a748414f93e117e641968d42b9db1d5720a2c5bf9d8643daeb9b2524940" exitCode=0 Oct 05 07:11:23 crc kubenswrapper[4846]: I1005 07:11:23.123868 4846 generic.go:334] "Generic (PLEG): container finished" podID="a32cdd03-1398-4055-b051-62adde35200f" containerID="6ef0fb95fbc564da5b7b63871e6cea78667f2f37a2822564371b3f419ac22f47" exitCode=0 Oct 05 07:11:23 crc kubenswrapper[4846]: I1005 07:11:23.139310 4846 generic.go:334] "Generic (PLEG): container finished" podID="bfe96e0c-b9fc-438d-9b17-bec246d48446" containerID="1921e872cb6fabf10cd959960895a03e62a0d256a8d1c87f18ab5389a991e8f4" exitCode=0 Oct 05 07:11:23 crc kubenswrapper[4846]: I1005 07:11:23.155509 4846 generic.go:334] "Generic (PLEG): container finished" podID="6660b9ff-df9c-478e-b852-82cb5fb2c60c" containerID="2a96112701d2d7042bce02efb3a776df18e837a8d8981829c635370de3016348" exitCode=0 Oct 05 07:11:23 crc kubenswrapper[4846]: I1005 07:11:23.172934 4846 generic.go:334] "Generic (PLEG): container finished" podID="4d3616b4-9f2c-4c62-9760-0300b02d0dbb" containerID="b30b71de14fefc7cbe669f3c30ee9c9b2aa5f68d63305a95ba9232ac4a893ad2" exitCode=0 Oct 05 07:11:23 crc kubenswrapper[4846]: I1005 07:11:23.198054 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"0c0f8308-455e-4544-9f53-0c06a54c6e22","Type":"ContainerDied","Data":"708780fcc4181d993d6e7d44ef9444fa8e85b47f3a59d10a6fa23aa84fa714b0"} Oct 05 07:11:23 crc kubenswrapper[4846]: I1005 07:11:23.198123 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"b7de2524-bf5c-4668-af74-7584edf72c8b","Type":"ContainerDied","Data":"316083fa9d64f1f90ee62226b3c28e38700ed5382c669302dda9655852943dab"} Oct 05 07:11:23 crc kubenswrapper[4846]: I1005 07:11:23.198142 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"221918b0-eb0e-4a32-8eed-07acd18bd31e","Type":"ContainerDied","Data":"ec314d49982bc70fa3bb12b1a024bc708bb21c05eda9001d00d334940b04d2e6"} Oct 05 07:11:23 crc kubenswrapper[4846]: I1005 07:11:23.198164 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"221918b0-eb0e-4a32-8eed-07acd18bd31e","Type":"ContainerDied","Data":"231cdc556bc3d56e30d9aa386e028a0c6fb06c7db20f716efb83a6c6f9bfc33e"} Oct 05 07:11:23 crc kubenswrapper[4846]: I1005 07:11:23.198178 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-984785977-srp2v" event={"ID":"d11b6df6-6da2-4cb8-a4f9-d87cfd5d4d26","Type":"ContainerDied","Data":"eb3d496875131b7ff843f08ff12b7fdc832265502da5423e00550c759c4bc1b1"} Oct 05 07:11:23 crc kubenswrapper[4846]: I1005 07:11:23.198249 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"2adc8ba0-2810-4bf4-bd24-172b97fb9b22","Type":"ContainerDied","Data":"9e54d631fb403eb9146c131948244ad644f213facfb43b90944167c8827ee59f"} Oct 05 07:11:23 crc kubenswrapper[4846]: I1005 07:11:23.198263 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutrona5c0-account-delete-q5wg4" event={"ID":"e2f9f9a8-a25f-4a99-b425-4799964d4e24","Type":"ContainerDied","Data":"ba19323ff93845912587e75b8b10dfd5b4a0513fb4e38a07aab30ad816bdfa3c"} Oct 05 07:11:23 crc kubenswrapper[4846]: I1005 07:11:23.198288 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutrona5c0-account-delete-q5wg4" event={"ID":"e2f9f9a8-a25f-4a99-b425-4799964d4e24","Type":"ContainerStarted","Data":"3df79f72c856d902d61af470174e9eb4cda409026d12528c0aece2ce522096e9"} Oct 05 07:11:23 crc kubenswrapper[4846]: I1005 07:11:23.198304 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement2760-account-delete-bvfdp" event={"ID":"fc35db15-65d4-4cd6-9d09-059822904757","Type":"ContainerDied","Data":"8b92f361cea3763ba14f3598387de3c1d364a1f14851d48c0df9f7aa1af31db1"} Oct 05 07:11:23 crc kubenswrapper[4846]: I1005 07:11:23.198316 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement2760-account-delete-bvfdp" event={"ID":"fc35db15-65d4-4cd6-9d09-059822904757","Type":"ContainerStarted","Data":"f31bd69595aa75eeda462ba2192f10b248f98c93fa238550ad43636812bfecb1"} Oct 05 07:11:23 crc kubenswrapper[4846]: I1005 07:11:23.198332 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"33193805-7aa6-406e-8cad-0ebbb1228ef2","Type":"ContainerDied","Data":"a03ec35e533cac1f50168f28070b82ea570152d6d222e9cf097e6f662b34b418"} Oct 05 07:11:23 crc kubenswrapper[4846]: I1005 07:11:23.198346 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican8fc5-account-delete-gjp47" event={"ID":"086cf913-a4a2-41ee-910b-a83be95c69d6","Type":"ContainerDied","Data":"18de9a748414f93e117e641968d42b9db1d5720a2c5bf9d8643daeb9b2524940"} Oct 05 07:11:23 crc kubenswrapper[4846]: I1005 07:11:23.198358 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican8fc5-account-delete-gjp47" event={"ID":"086cf913-a4a2-41ee-910b-a83be95c69d6","Type":"ContainerStarted","Data":"525ee924ca58c84f2ca050e2fe5ed97086aca63ac6a15850a1eb3d12e6423876"} Oct 05 07:11:23 crc kubenswrapper[4846]: I1005 07:11:23.198368 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cindercf3d-account-delete-jw5tr" event={"ID":"a32cdd03-1398-4055-b051-62adde35200f","Type":"ContainerDied","Data":"6ef0fb95fbc564da5b7b63871e6cea78667f2f37a2822564371b3f419ac22f47"} Oct 05 07:11:23 crc kubenswrapper[4846]: I1005 07:11:23.198380 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cindercf3d-account-delete-jw5tr" event={"ID":"a32cdd03-1398-4055-b051-62adde35200f","Type":"ContainerStarted","Data":"63fc4076208dfa8214e57a8f6a726a5fbe145fe441f981f1f2d4a2f7ce1f81bc"} Oct 05 07:11:23 crc kubenswrapper[4846]: I1005 07:11:23.198392 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"bfe96e0c-b9fc-438d-9b17-bec246d48446","Type":"ContainerDied","Data":"1921e872cb6fabf10cd959960895a03e62a0d256a8d1c87f18ab5389a991e8f4"} Oct 05 07:11:23 crc kubenswrapper[4846]: I1005 07:11:23.198408 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"6660b9ff-df9c-478e-b852-82cb5fb2c60c","Type":"ContainerDied","Data":"2a96112701d2d7042bce02efb3a776df18e837a8d8981829c635370de3016348"} Oct 05 07:11:23 crc kubenswrapper[4846]: I1005 07:11:23.198422 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"6660b9ff-df9c-478e-b852-82cb5fb2c60c","Type":"ContainerDied","Data":"975d98535a2e45846601a9e1e1224188cc0dd1e3bbd28cd7323573e631782916"} Oct 05 07:11:23 crc kubenswrapper[4846]: I1005 07:11:23.198434 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="975d98535a2e45846601a9e1e1224188cc0dd1e3bbd28cd7323573e631782916" Oct 05 07:11:23 crc kubenswrapper[4846]: I1005 07:11:23.198446 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glancebc5f-account-delete-hkzcl" event={"ID":"4d3616b4-9f2c-4c62-9760-0300b02d0dbb","Type":"ContainerDied","Data":"b30b71de14fefc7cbe669f3c30ee9c9b2aa5f68d63305a95ba9232ac4a893ad2"} Oct 05 07:11:23 crc kubenswrapper[4846]: I1005 07:11:23.198458 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glancebc5f-account-delete-hkzcl" event={"ID":"4d3616b4-9f2c-4c62-9760-0300b02d0dbb","Type":"ContainerStarted","Data":"1da9c21171f99446de2b2aa2c3001e53829eca259615385092f0f013986911bf"} Oct 05 07:11:23 crc kubenswrapper[4846]: I1005 07:11:23.198480 4846 scope.go:117] "RemoveContainer" containerID="ec314d49982bc70fa3bb12b1a024bc708bb21c05eda9001d00d334940b04d2e6" Oct 05 07:11:23 crc kubenswrapper[4846]: I1005 07:11:23.214247 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 05 07:11:23 crc kubenswrapper[4846]: I1005 07:11:23.240661 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstack-cell1-galera-0"] Oct 05 07:11:23 crc kubenswrapper[4846]: I1005 07:11:23.249397 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstack-cell1-galera-0"] Oct 05 07:11:23 crc kubenswrapper[4846]: I1005 07:11:23.289276 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-create-sqkh4"] Oct 05 07:11:23 crc kubenswrapper[4846]: I1005 07:11:23.305584 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-create-sqkh4"] Oct 05 07:11:23 crc kubenswrapper[4846]: I1005 07:11:23.313098 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-bc5f-account-create-spj8j"] Oct 05 07:11:23 crc kubenswrapper[4846]: I1005 07:11:23.317196 4846 scope.go:117] "RemoveContainer" containerID="48d194c3f1852cef7897b4f3cbb72f3d1aeca5087732bda856574a4d0614661b" Oct 05 07:11:23 crc kubenswrapper[4846]: I1005 07:11:23.321243 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glancebc5f-account-delete-hkzcl"] Oct 05 07:11:23 crc kubenswrapper[4846]: I1005 07:11:23.328910 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-bc5f-account-create-spj8j"] Oct 05 07:11:23 crc kubenswrapper[4846]: I1005 07:11:23.336256 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6660b9ff-df9c-478e-b852-82cb5fb2c60c-combined-ca-bundle\") pod \"6660b9ff-df9c-478e-b852-82cb5fb2c60c\" (UID: \"6660b9ff-df9c-478e-b852-82cb5fb2c60c\") " Oct 05 07:11:23 crc kubenswrapper[4846]: I1005 07:11:23.336397 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/6660b9ff-df9c-478e-b852-82cb5fb2c60c-vencrypt-tls-certs\") pod \"6660b9ff-df9c-478e-b852-82cb5fb2c60c\" (UID: \"6660b9ff-df9c-478e-b852-82cb5fb2c60c\") " Oct 05 07:11:23 crc kubenswrapper[4846]: I1005 07:11:23.336451 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jwnfj\" (UniqueName: \"kubernetes.io/projected/6660b9ff-df9c-478e-b852-82cb5fb2c60c-kube-api-access-jwnfj\") pod \"6660b9ff-df9c-478e-b852-82cb5fb2c60c\" (UID: \"6660b9ff-df9c-478e-b852-82cb5fb2c60c\") " Oct 05 07:11:23 crc kubenswrapper[4846]: I1005 07:11:23.336479 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/6660b9ff-df9c-478e-b852-82cb5fb2c60c-nova-novncproxy-tls-certs\") pod \"6660b9ff-df9c-478e-b852-82cb5fb2c60c\" (UID: \"6660b9ff-df9c-478e-b852-82cb5fb2c60c\") " Oct 05 07:11:23 crc kubenswrapper[4846]: I1005 07:11:23.336569 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6660b9ff-df9c-478e-b852-82cb5fb2c60c-config-data\") pod \"6660b9ff-df9c-478e-b852-82cb5fb2c60c\" (UID: \"6660b9ff-df9c-478e-b852-82cb5fb2c60c\") " Oct 05 07:11:23 crc kubenswrapper[4846]: I1005 07:11:23.345691 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6660b9ff-df9c-478e-b852-82cb5fb2c60c-kube-api-access-jwnfj" (OuterVolumeSpecName: "kube-api-access-jwnfj") pod "6660b9ff-df9c-478e-b852-82cb5fb2c60c" (UID: "6660b9ff-df9c-478e-b852-82cb5fb2c60c"). InnerVolumeSpecName "kube-api-access-jwnfj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:11:23 crc kubenswrapper[4846]: I1005 07:11:23.350498 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 05 07:11:23 crc kubenswrapper[4846]: I1005 07:11:23.386091 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6660b9ff-df9c-478e-b852-82cb5fb2c60c-config-data" (OuterVolumeSpecName: "config-data") pod "6660b9ff-df9c-478e-b852-82cb5fb2c60c" (UID: "6660b9ff-df9c-478e-b852-82cb5fb2c60c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:11:23 crc kubenswrapper[4846]: I1005 07:11:23.388787 4846 scope.go:117] "RemoveContainer" containerID="ec314d49982bc70fa3bb12b1a024bc708bb21c05eda9001d00d334940b04d2e6" Oct 05 07:11:23 crc kubenswrapper[4846]: E1005 07:11:23.389414 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ec314d49982bc70fa3bb12b1a024bc708bb21c05eda9001d00d334940b04d2e6\": container with ID starting with ec314d49982bc70fa3bb12b1a024bc708bb21c05eda9001d00d334940b04d2e6 not found: ID does not exist" containerID="ec314d49982bc70fa3bb12b1a024bc708bb21c05eda9001d00d334940b04d2e6" Oct 05 07:11:23 crc kubenswrapper[4846]: I1005 07:11:23.389445 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ec314d49982bc70fa3bb12b1a024bc708bb21c05eda9001d00d334940b04d2e6"} err="failed to get container status \"ec314d49982bc70fa3bb12b1a024bc708bb21c05eda9001d00d334940b04d2e6\": rpc error: code = NotFound desc = could not find container \"ec314d49982bc70fa3bb12b1a024bc708bb21c05eda9001d00d334940b04d2e6\": container with ID starting with ec314d49982bc70fa3bb12b1a024bc708bb21c05eda9001d00d334940b04d2e6 not found: ID does not exist" Oct 05 07:11:23 crc kubenswrapper[4846]: I1005 07:11:23.389467 4846 scope.go:117] "RemoveContainer" containerID="48d194c3f1852cef7897b4f3cbb72f3d1aeca5087732bda856574a4d0614661b" Oct 05 07:11:23 crc kubenswrapper[4846]: E1005 07:11:23.389720 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"48d194c3f1852cef7897b4f3cbb72f3d1aeca5087732bda856574a4d0614661b\": container with ID starting with 48d194c3f1852cef7897b4f3cbb72f3d1aeca5087732bda856574a4d0614661b not found: ID does not exist" containerID="48d194c3f1852cef7897b4f3cbb72f3d1aeca5087732bda856574a4d0614661b" Oct 05 07:11:23 crc kubenswrapper[4846]: I1005 07:11:23.389739 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"48d194c3f1852cef7897b4f3cbb72f3d1aeca5087732bda856574a4d0614661b"} err="failed to get container status \"48d194c3f1852cef7897b4f3cbb72f3d1aeca5087732bda856574a4d0614661b\": rpc error: code = NotFound desc = could not find container \"48d194c3f1852cef7897b4f3cbb72f3d1aeca5087732bda856574a4d0614661b\": container with ID starting with 48d194c3f1852cef7897b4f3cbb72f3d1aeca5087732bda856574a4d0614661b not found: ID does not exist" Oct 05 07:11:23 crc kubenswrapper[4846]: I1005 07:11:23.395538 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6660b9ff-df9c-478e-b852-82cb5fb2c60c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6660b9ff-df9c-478e-b852-82cb5fb2c60c" (UID: "6660b9ff-df9c-478e-b852-82cb5fb2c60c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:11:23 crc kubenswrapper[4846]: I1005 07:11:23.406974 4846 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/cinder-api-0" podUID="36dca705-2f72-411e-8d80-ed31d49989cc" containerName="cinder-api" probeResult="failure" output="Get \"https://10.217.0.169:8776/healthcheck\": read tcp 10.217.0.2:43848->10.217.0.169:8776: read: connection reset by peer" Oct 05 07:11:23 crc kubenswrapper[4846]: I1005 07:11:23.409974 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6660b9ff-df9c-478e-b852-82cb5fb2c60c-vencrypt-tls-certs" (OuterVolumeSpecName: "vencrypt-tls-certs") pod "6660b9ff-df9c-478e-b852-82cb5fb2c60c" (UID: "6660b9ff-df9c-478e-b852-82cb5fb2c60c"). InnerVolumeSpecName "vencrypt-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:11:23 crc kubenswrapper[4846]: I1005 07:11:23.426428 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6660b9ff-df9c-478e-b852-82cb5fb2c60c-nova-novncproxy-tls-certs" (OuterVolumeSpecName: "nova-novncproxy-tls-certs") pod "6660b9ff-df9c-478e-b852-82cb5fb2c60c" (UID: "6660b9ff-df9c-478e-b852-82cb5fb2c60c"). InnerVolumeSpecName "nova-novncproxy-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:11:23 crc kubenswrapper[4846]: I1005 07:11:23.440628 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b7de2524-bf5c-4668-af74-7584edf72c8b-combined-ca-bundle\") pod \"b7de2524-bf5c-4668-af74-7584edf72c8b\" (UID: \"b7de2524-bf5c-4668-af74-7584edf72c8b\") " Oct 05 07:11:23 crc kubenswrapper[4846]: I1005 07:11:23.440728 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4lm46\" (UniqueName: \"kubernetes.io/projected/b7de2524-bf5c-4668-af74-7584edf72c8b-kube-api-access-4lm46\") pod \"b7de2524-bf5c-4668-af74-7584edf72c8b\" (UID: \"b7de2524-bf5c-4668-af74-7584edf72c8b\") " Oct 05 07:11:23 crc kubenswrapper[4846]: I1005 07:11:23.440827 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b7de2524-bf5c-4668-af74-7584edf72c8b-config-data\") pod \"b7de2524-bf5c-4668-af74-7584edf72c8b\" (UID: \"b7de2524-bf5c-4668-af74-7584edf72c8b\") " Oct 05 07:11:23 crc kubenswrapper[4846]: I1005 07:11:23.441312 4846 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6660b9ff-df9c-478e-b852-82cb5fb2c60c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:23 crc kubenswrapper[4846]: I1005 07:11:23.441332 4846 reconciler_common.go:293] "Volume detached for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/6660b9ff-df9c-478e-b852-82cb5fb2c60c-vencrypt-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:23 crc kubenswrapper[4846]: I1005 07:11:23.441341 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jwnfj\" (UniqueName: \"kubernetes.io/projected/6660b9ff-df9c-478e-b852-82cb5fb2c60c-kube-api-access-jwnfj\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:23 crc kubenswrapper[4846]: I1005 07:11:23.441356 4846 reconciler_common.go:293] "Volume detached for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/6660b9ff-df9c-478e-b852-82cb5fb2c60c-nova-novncproxy-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:23 crc kubenswrapper[4846]: I1005 07:11:23.441367 4846 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6660b9ff-df9c-478e-b852-82cb5fb2c60c-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:23 crc kubenswrapper[4846]: I1005 07:11:23.445541 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b7de2524-bf5c-4668-af74-7584edf72c8b-kube-api-access-4lm46" (OuterVolumeSpecName: "kube-api-access-4lm46") pod "b7de2524-bf5c-4668-af74-7584edf72c8b" (UID: "b7de2524-bf5c-4668-af74-7584edf72c8b"). InnerVolumeSpecName "kube-api-access-4lm46". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:11:23 crc kubenswrapper[4846]: I1005 07:11:23.470291 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b7de2524-bf5c-4668-af74-7584edf72c8b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b7de2524-bf5c-4668-af74-7584edf72c8b" (UID: "b7de2524-bf5c-4668-af74-7584edf72c8b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:11:23 crc kubenswrapper[4846]: I1005 07:11:23.480757 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b7de2524-bf5c-4668-af74-7584edf72c8b-config-data" (OuterVolumeSpecName: "config-data") pod "b7de2524-bf5c-4668-af74-7584edf72c8b" (UID: "b7de2524-bf5c-4668-af74-7584edf72c8b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:11:23 crc kubenswrapper[4846]: I1005 07:11:23.543641 4846 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b7de2524-bf5c-4668-af74-7584edf72c8b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:23 crc kubenswrapper[4846]: I1005 07:11:23.543666 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4lm46\" (UniqueName: \"kubernetes.io/projected/b7de2524-bf5c-4668-af74-7584edf72c8b-kube-api-access-4lm46\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:23 crc kubenswrapper[4846]: I1005 07:11:23.543674 4846 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b7de2524-bf5c-4668-af74-7584edf72c8b-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:23 crc kubenswrapper[4846]: I1005 07:11:23.881809 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 05 07:11:23 crc kubenswrapper[4846]: I1005 07:11:23.898830 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Oct 05 07:11:23 crc kubenswrapper[4846]: I1005 07:11:23.921800 4846 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="e6e86b1c-87ac-4594-aeba-943947065374" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.205:8775/\": read tcp 10.217.0.2:33862->10.217.0.205:8775: read: connection reset by peer" Oct 05 07:11:23 crc kubenswrapper[4846]: I1005 07:11:23.922202 4846 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="e6e86b1c-87ac-4594-aeba-943947065374" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.205:8775/\": read tcp 10.217.0.2:33858->10.217.0.205:8775: read: connection reset by peer" Oct 05 07:11:23 crc kubenswrapper[4846]: I1005 07:11:23.924771 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement2760-account-delete-bvfdp" Oct 05 07:11:23 crc kubenswrapper[4846]: I1005 07:11:23.928917 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutrona5c0-account-delete-q5wg4" Oct 05 07:11:23 crc kubenswrapper[4846]: I1005 07:11:23.969376 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/bfe96e0c-b9fc-438d-9b17-bec246d48446-config-data-custom\") pod \"bfe96e0c-b9fc-438d-9b17-bec246d48446\" (UID: \"bfe96e0c-b9fc-438d-9b17-bec246d48446\") " Oct 05 07:11:23 crc kubenswrapper[4846]: I1005 07:11:23.969430 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/33193805-7aa6-406e-8cad-0ebbb1228ef2-config-data\") pod \"33193805-7aa6-406e-8cad-0ebbb1228ef2\" (UID: \"33193805-7aa6-406e-8cad-0ebbb1228ef2\") " Oct 05 07:11:23 crc kubenswrapper[4846]: I1005 07:11:23.969460 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vhc7k\" (UniqueName: \"kubernetes.io/projected/e2f9f9a8-a25f-4a99-b425-4799964d4e24-kube-api-access-vhc7k\") pod \"e2f9f9a8-a25f-4a99-b425-4799964d4e24\" (UID: \"e2f9f9a8-a25f-4a99-b425-4799964d4e24\") " Oct 05 07:11:23 crc kubenswrapper[4846]: I1005 07:11:23.969576 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/33193805-7aa6-406e-8cad-0ebbb1228ef2-combined-ca-bundle\") pod \"33193805-7aa6-406e-8cad-0ebbb1228ef2\" (UID: \"33193805-7aa6-406e-8cad-0ebbb1228ef2\") " Oct 05 07:11:23 crc kubenswrapper[4846]: I1005 07:11:23.969667 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hd6jb\" (UniqueName: \"kubernetes.io/projected/bfe96e0c-b9fc-438d-9b17-bec246d48446-kube-api-access-hd6jb\") pod \"bfe96e0c-b9fc-438d-9b17-bec246d48446\" (UID: \"bfe96e0c-b9fc-438d-9b17-bec246d48446\") " Oct 05 07:11:23 crc kubenswrapper[4846]: I1005 07:11:23.969691 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bfe96e0c-b9fc-438d-9b17-bec246d48446-combined-ca-bundle\") pod \"bfe96e0c-b9fc-438d-9b17-bec246d48446\" (UID: \"bfe96e0c-b9fc-438d-9b17-bec246d48446\") " Oct 05 07:11:23 crc kubenswrapper[4846]: I1005 07:11:23.969716 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bfe96e0c-b9fc-438d-9b17-bec246d48446-scripts\") pod \"bfe96e0c-b9fc-438d-9b17-bec246d48446\" (UID: \"bfe96e0c-b9fc-438d-9b17-bec246d48446\") " Oct 05 07:11:23 crc kubenswrapper[4846]: I1005 07:11:23.969770 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bfe96e0c-b9fc-438d-9b17-bec246d48446-config-data\") pod \"bfe96e0c-b9fc-438d-9b17-bec246d48446\" (UID: \"bfe96e0c-b9fc-438d-9b17-bec246d48446\") " Oct 05 07:11:23 crc kubenswrapper[4846]: I1005 07:11:23.969805 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gdftk\" (UniqueName: \"kubernetes.io/projected/fc35db15-65d4-4cd6-9d09-059822904757-kube-api-access-gdftk\") pod \"fc35db15-65d4-4cd6-9d09-059822904757\" (UID: \"fc35db15-65d4-4cd6-9d09-059822904757\") " Oct 05 07:11:23 crc kubenswrapper[4846]: I1005 07:11:23.969831 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/bfe96e0c-b9fc-438d-9b17-bec246d48446-etc-machine-id\") pod \"bfe96e0c-b9fc-438d-9b17-bec246d48446\" (UID: \"bfe96e0c-b9fc-438d-9b17-bec246d48446\") " Oct 05 07:11:23 crc kubenswrapper[4846]: I1005 07:11:23.969861 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2nd7f\" (UniqueName: \"kubernetes.io/projected/33193805-7aa6-406e-8cad-0ebbb1228ef2-kube-api-access-2nd7f\") pod \"33193805-7aa6-406e-8cad-0ebbb1228ef2\" (UID: \"33193805-7aa6-406e-8cad-0ebbb1228ef2\") " Oct 05 07:11:23 crc kubenswrapper[4846]: I1005 07:11:23.981535 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/33193805-7aa6-406e-8cad-0ebbb1228ef2-kube-api-access-2nd7f" (OuterVolumeSpecName: "kube-api-access-2nd7f") pod "33193805-7aa6-406e-8cad-0ebbb1228ef2" (UID: "33193805-7aa6-406e-8cad-0ebbb1228ef2"). InnerVolumeSpecName "kube-api-access-2nd7f". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:11:23 crc kubenswrapper[4846]: I1005 07:11:23.983135 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/bfe96e0c-b9fc-438d-9b17-bec246d48446-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "bfe96e0c-b9fc-438d-9b17-bec246d48446" (UID: "bfe96e0c-b9fc-438d-9b17-bec246d48446"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 05 07:11:24 crc kubenswrapper[4846]: I1005 07:11:24.000967 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bfe96e0c-b9fc-438d-9b17-bec246d48446-kube-api-access-hd6jb" (OuterVolumeSpecName: "kube-api-access-hd6jb") pod "bfe96e0c-b9fc-438d-9b17-bec246d48446" (UID: "bfe96e0c-b9fc-438d-9b17-bec246d48446"). InnerVolumeSpecName "kube-api-access-hd6jb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:11:24 crc kubenswrapper[4846]: I1005 07:11:24.006605 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fc35db15-65d4-4cd6-9d09-059822904757-kube-api-access-gdftk" (OuterVolumeSpecName: "kube-api-access-gdftk") pod "fc35db15-65d4-4cd6-9d09-059822904757" (UID: "fc35db15-65d4-4cd6-9d09-059822904757"). InnerVolumeSpecName "kube-api-access-gdftk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:11:24 crc kubenswrapper[4846]: I1005 07:11:24.010628 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e2f9f9a8-a25f-4a99-b425-4799964d4e24-kube-api-access-vhc7k" (OuterVolumeSpecName: "kube-api-access-vhc7k") pod "e2f9f9a8-a25f-4a99-b425-4799964d4e24" (UID: "e2f9f9a8-a25f-4a99-b425-4799964d4e24"). InnerVolumeSpecName "kube-api-access-vhc7k". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:11:24 crc kubenswrapper[4846]: I1005 07:11:24.011750 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bfe96e0c-b9fc-438d-9b17-bec246d48446-scripts" (OuterVolumeSpecName: "scripts") pod "bfe96e0c-b9fc-438d-9b17-bec246d48446" (UID: "bfe96e0c-b9fc-438d-9b17-bec246d48446"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:11:24 crc kubenswrapper[4846]: I1005 07:11:24.011870 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bfe96e0c-b9fc-438d-9b17-bec246d48446-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "bfe96e0c-b9fc-438d-9b17-bec246d48446" (UID: "bfe96e0c-b9fc-438d-9b17-bec246d48446"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:11:24 crc kubenswrapper[4846]: I1005 07:11:24.079981 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vhc7k\" (UniqueName: \"kubernetes.io/projected/e2f9f9a8-a25f-4a99-b425-4799964d4e24-kube-api-access-vhc7k\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:24 crc kubenswrapper[4846]: I1005 07:11:24.080008 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hd6jb\" (UniqueName: \"kubernetes.io/projected/bfe96e0c-b9fc-438d-9b17-bec246d48446-kube-api-access-hd6jb\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:24 crc kubenswrapper[4846]: I1005 07:11:24.080017 4846 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bfe96e0c-b9fc-438d-9b17-bec246d48446-scripts\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:24 crc kubenswrapper[4846]: I1005 07:11:24.080029 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gdftk\" (UniqueName: \"kubernetes.io/projected/fc35db15-65d4-4cd6-9d09-059822904757-kube-api-access-gdftk\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:24 crc kubenswrapper[4846]: I1005 07:11:24.080040 4846 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/bfe96e0c-b9fc-438d-9b17-bec246d48446-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:24 crc kubenswrapper[4846]: I1005 07:11:24.080050 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2nd7f\" (UniqueName: \"kubernetes.io/projected/33193805-7aa6-406e-8cad-0ebbb1228ef2-kube-api-access-2nd7f\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:24 crc kubenswrapper[4846]: I1005 07:11:24.080058 4846 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/bfe96e0c-b9fc-438d-9b17-bec246d48446-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:24 crc kubenswrapper[4846]: I1005 07:11:24.096661 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/33193805-7aa6-406e-8cad-0ebbb1228ef2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "33193805-7aa6-406e-8cad-0ebbb1228ef2" (UID: "33193805-7aa6-406e-8cad-0ebbb1228ef2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:11:24 crc kubenswrapper[4846]: I1005 07:11:24.148999 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bfe96e0c-b9fc-438d-9b17-bec246d48446-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "bfe96e0c-b9fc-438d-9b17-bec246d48446" (UID: "bfe96e0c-b9fc-438d-9b17-bec246d48446"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:11:24 crc kubenswrapper[4846]: I1005 07:11:24.150195 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/33193805-7aa6-406e-8cad-0ebbb1228ef2-config-data" (OuterVolumeSpecName: "config-data") pod "33193805-7aa6-406e-8cad-0ebbb1228ef2" (UID: "33193805-7aa6-406e-8cad-0ebbb1228ef2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:11:24 crc kubenswrapper[4846]: I1005 07:11:24.174924 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 05 07:11:24 crc kubenswrapper[4846]: I1005 07:11:24.181766 4846 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/33193805-7aa6-406e-8cad-0ebbb1228ef2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:24 crc kubenswrapper[4846]: I1005 07:11:24.181789 4846 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bfe96e0c-b9fc-438d-9b17-bec246d48446-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:24 crc kubenswrapper[4846]: I1005 07:11:24.181798 4846 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/33193805-7aa6-406e-8cad-0ebbb1228ef2-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:24 crc kubenswrapper[4846]: I1005 07:11:24.189334 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement2760-account-delete-bvfdp" event={"ID":"fc35db15-65d4-4cd6-9d09-059822904757","Type":"ContainerDied","Data":"f31bd69595aa75eeda462ba2192f10b248f98c93fa238550ad43636812bfecb1"} Oct 05 07:11:24 crc kubenswrapper[4846]: I1005 07:11:24.189386 4846 scope.go:117] "RemoveContainer" containerID="8b92f361cea3763ba14f3598387de3c1d364a1f14851d48c0df9f7aa1af31db1" Oct 05 07:11:24 crc kubenswrapper[4846]: I1005 07:11:24.189398 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement2760-account-delete-bvfdp" Oct 05 07:11:24 crc kubenswrapper[4846]: I1005 07:11:24.197993 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"0c0f8308-455e-4544-9f53-0c06a54c6e22","Type":"ContainerDied","Data":"69378334ebffca563252133a91b7f973164e307ee0c19be4b2afa5c70cd826c5"} Oct 05 07:11:24 crc kubenswrapper[4846]: I1005 07:11:24.198086 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 05 07:11:24 crc kubenswrapper[4846]: I1005 07:11:24.214713 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"b7de2524-bf5c-4668-af74-7584edf72c8b","Type":"ContainerDied","Data":"e82e7415d67ac703880356b4cae71e6bc9b26a12df9bb1e074b952bab4350036"} Oct 05 07:11:24 crc kubenswrapper[4846]: I1005 07:11:24.214803 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 05 07:11:24 crc kubenswrapper[4846]: I1005 07:11:24.230154 4846 generic.go:334] "Generic (PLEG): container finished" podID="e6e86b1c-87ac-4594-aeba-943947065374" containerID="7408c3d57efa73ce926d92ff91954d484124d71bc39b46bd428ba78c4ef03075" exitCode=0 Oct 05 07:11:24 crc kubenswrapper[4846]: I1005 07:11:24.230255 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"e6e86b1c-87ac-4594-aeba-943947065374","Type":"ContainerDied","Data":"7408c3d57efa73ce926d92ff91954d484124d71bc39b46bd428ba78c4ef03075"} Oct 05 07:11:24 crc kubenswrapper[4846]: I1005 07:11:24.231002 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bfe96e0c-b9fc-438d-9b17-bec246d48446-config-data" (OuterVolumeSpecName: "config-data") pod "bfe96e0c-b9fc-438d-9b17-bec246d48446" (UID: "bfe96e0c-b9fc-438d-9b17-bec246d48446"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:11:24 crc kubenswrapper[4846]: I1005 07:11:24.248387 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement2760-account-delete-bvfdp"] Oct 05 07:11:24 crc kubenswrapper[4846]: I1005 07:11:24.251935 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"33193805-7aa6-406e-8cad-0ebbb1228ef2","Type":"ContainerDied","Data":"470889702a1e2e05941ace0261d00f6838c4d71988b1b343ca1f365453bbe0b2"} Oct 05 07:11:24 crc kubenswrapper[4846]: I1005 07:11:24.252015 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Oct 05 07:11:24 crc kubenswrapper[4846]: I1005 07:11:24.257164 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement2760-account-delete-bvfdp"] Oct 05 07:11:24 crc kubenswrapper[4846]: I1005 07:11:24.272062 4846 generic.go:334] "Generic (PLEG): container finished" podID="c31937eb-9836-461b-9230-5e06ce7a99e8" containerID="62af63220e4d608e4b981ffe4309575edbb012337dc1e84d368d62d0b4cab633" exitCode=0 Oct 05 07:11:24 crc kubenswrapper[4846]: I1005 07:11:24.272126 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-57674d6c54-zn5nq" event={"ID":"c31937eb-9836-461b-9230-5e06ce7a99e8","Type":"ContainerDied","Data":"62af63220e4d608e4b981ffe4309575edbb012337dc1e84d368d62d0b4cab633"} Oct 05 07:11:24 crc kubenswrapper[4846]: I1005 07:11:24.286463 4846 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-57674d6c54-zn5nq" podUID="c31937eb-9836-461b-9230-5e06ce7a99e8" containerName="barbican-api-log" probeResult="failure" output="Get \"https://10.217.0.168:9311/healthcheck\": dial tcp 10.217.0.168:9311: connect: connection refused" Oct 05 07:11:24 crc kubenswrapper[4846]: I1005 07:11:24.286977 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0c0f8308-455e-4544-9f53-0c06a54c6e22-logs\") pod \"0c0f8308-455e-4544-9f53-0c06a54c6e22\" (UID: \"0c0f8308-455e-4544-9f53-0c06a54c6e22\") " Oct 05 07:11:24 crc kubenswrapper[4846]: I1005 07:11:24.287052 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0c0f8308-455e-4544-9f53-0c06a54c6e22-public-tls-certs\") pod \"0c0f8308-455e-4544-9f53-0c06a54c6e22\" (UID: \"0c0f8308-455e-4544-9f53-0c06a54c6e22\") " Oct 05 07:11:24 crc kubenswrapper[4846]: I1005 07:11:24.287127 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjd99\" (UniqueName: \"kubernetes.io/projected/0c0f8308-455e-4544-9f53-0c06a54c6e22-kube-api-access-pjd99\") pod \"0c0f8308-455e-4544-9f53-0c06a54c6e22\" (UID: \"0c0f8308-455e-4544-9f53-0c06a54c6e22\") " Oct 05 07:11:24 crc kubenswrapper[4846]: I1005 07:11:24.287290 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/0c0f8308-455e-4544-9f53-0c06a54c6e22-httpd-run\") pod \"0c0f8308-455e-4544-9f53-0c06a54c6e22\" (UID: \"0c0f8308-455e-4544-9f53-0c06a54c6e22\") " Oct 05 07:11:24 crc kubenswrapper[4846]: I1005 07:11:24.287359 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0c0f8308-455e-4544-9f53-0c06a54c6e22-combined-ca-bundle\") pod \"0c0f8308-455e-4544-9f53-0c06a54c6e22\" (UID: \"0c0f8308-455e-4544-9f53-0c06a54c6e22\") " Oct 05 07:11:24 crc kubenswrapper[4846]: I1005 07:11:24.287469 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"0c0f8308-455e-4544-9f53-0c06a54c6e22\" (UID: \"0c0f8308-455e-4544-9f53-0c06a54c6e22\") " Oct 05 07:11:24 crc kubenswrapper[4846]: I1005 07:11:24.287532 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0c0f8308-455e-4544-9f53-0c06a54c6e22-config-data\") pod \"0c0f8308-455e-4544-9f53-0c06a54c6e22\" (UID: \"0c0f8308-455e-4544-9f53-0c06a54c6e22\") " Oct 05 07:11:24 crc kubenswrapper[4846]: I1005 07:11:24.287633 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0c0f8308-455e-4544-9f53-0c06a54c6e22-scripts\") pod \"0c0f8308-455e-4544-9f53-0c06a54c6e22\" (UID: \"0c0f8308-455e-4544-9f53-0c06a54c6e22\") " Oct 05 07:11:24 crc kubenswrapper[4846]: I1005 07:11:24.288228 4846 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bfe96e0c-b9fc-438d-9b17-bec246d48446-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:24 crc kubenswrapper[4846]: I1005 07:11:24.288793 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0c0f8308-455e-4544-9f53-0c06a54c6e22-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "0c0f8308-455e-4544-9f53-0c06a54c6e22" (UID: "0c0f8308-455e-4544-9f53-0c06a54c6e22"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 07:11:24 crc kubenswrapper[4846]: I1005 07:11:24.289045 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0c0f8308-455e-4544-9f53-0c06a54c6e22-logs" (OuterVolumeSpecName: "logs") pod "0c0f8308-455e-4544-9f53-0c06a54c6e22" (UID: "0c0f8308-455e-4544-9f53-0c06a54c6e22"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 07:11:24 crc kubenswrapper[4846]: I1005 07:11:24.292476 4846 generic.go:334] "Generic (PLEG): container finished" podID="36dca705-2f72-411e-8d80-ed31d49989cc" containerID="2dae962829e3caf9333b75e7f59bede4cf4d679ed6def4e5eddee1a34c169b39" exitCode=0 Oct 05 07:11:24 crc kubenswrapper[4846]: I1005 07:11:24.292533 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"36dca705-2f72-411e-8d80-ed31d49989cc","Type":"ContainerDied","Data":"2dae962829e3caf9333b75e7f59bede4cf4d679ed6def4e5eddee1a34c169b39"} Oct 05 07:11:24 crc kubenswrapper[4846]: I1005 07:11:24.297647 4846 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-57674d6c54-zn5nq" podUID="c31937eb-9836-461b-9230-5e06ce7a99e8" containerName="barbican-api" probeResult="failure" output="Get \"https://10.217.0.168:9311/healthcheck\": dial tcp 10.217.0.168:9311: connect: connection refused" Oct 05 07:11:24 crc kubenswrapper[4846]: I1005 07:11:24.298377 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cindercf3d-account-delete-jw5tr" Oct 05 07:11:24 crc kubenswrapper[4846]: I1005 07:11:24.301041 4846 generic.go:334] "Generic (PLEG): container finished" podID="c35affe0-da4e-448f-aac0-7bba26f89040" containerID="b48698aea06e5e06f2a72f7568ee41c7c5c3fee9c4bdab0f9cb7cebb4f50ea03" exitCode=0 Oct 05 07:11:24 crc kubenswrapper[4846]: I1005 07:11:24.301103 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-65884db4f4-fsk6p" event={"ID":"c35affe0-da4e-448f-aac0-7bba26f89040","Type":"ContainerDied","Data":"b48698aea06e5e06f2a72f7568ee41c7c5c3fee9c4bdab0f9cb7cebb4f50ea03"} Oct 05 07:11:24 crc kubenswrapper[4846]: I1005 07:11:24.303308 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 05 07:11:24 crc kubenswrapper[4846]: I1005 07:11:24.311135 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage07-crc" (OuterVolumeSpecName: "glance") pod "0c0f8308-455e-4544-9f53-0c06a54c6e22" (UID: "0c0f8308-455e-4544-9f53-0c06a54c6e22"). InnerVolumeSpecName "local-storage07-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 05 07:11:24 crc kubenswrapper[4846]: I1005 07:11:24.317325 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 05 07:11:24 crc kubenswrapper[4846]: I1005 07:11:24.317583 4846 generic.go:334] "Generic (PLEG): container finished" podID="6e9e53c6-4cf2-4988-89b4-c69471f616aa" containerID="7b4d83183acff2c37efb3d06dc23fc65569d8eb1f78f7305d68857a97455040e" exitCode=0 Oct 05 07:11:24 crc kubenswrapper[4846]: I1005 07:11:24.317637 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"6e9e53c6-4cf2-4988-89b4-c69471f616aa","Type":"ContainerDied","Data":"7b4d83183acff2c37efb3d06dc23fc65569d8eb1f78f7305d68857a97455040e"} Oct 05 07:11:24 crc kubenswrapper[4846]: I1005 07:11:24.325787 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0c0f8308-455e-4544-9f53-0c06a54c6e22-kube-api-access-pjd99" (OuterVolumeSpecName: "kube-api-access-pjd99") pod "0c0f8308-455e-4544-9f53-0c06a54c6e22" (UID: "0c0f8308-455e-4544-9f53-0c06a54c6e22"). InnerVolumeSpecName "kube-api-access-pjd99". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:11:24 crc kubenswrapper[4846]: I1005 07:11:24.330495 4846 generic.go:334] "Generic (PLEG): container finished" podID="d11b6df6-6da2-4cb8-a4f9-d87cfd5d4d26" containerID="5b9cd4a1e7ff57757729dbae754fc2fd8f2c6d6deb6c33079a5fb777d6c337c4" exitCode=0 Oct 05 07:11:24 crc kubenswrapper[4846]: I1005 07:11:24.330564 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Oct 05 07:11:24 crc kubenswrapper[4846]: I1005 07:11:24.330592 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-984785977-srp2v" event={"ID":"d11b6df6-6da2-4cb8-a4f9-d87cfd5d4d26","Type":"ContainerDied","Data":"5b9cd4a1e7ff57757729dbae754fc2fd8f2c6d6deb6c33079a5fb777d6c337c4"} Oct 05 07:11:24 crc kubenswrapper[4846]: I1005 07:11:24.335674 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0c0f8308-455e-4544-9f53-0c06a54c6e22-scripts" (OuterVolumeSpecName: "scripts") pod "0c0f8308-455e-4544-9f53-0c06a54c6e22" (UID: "0c0f8308-455e-4544-9f53-0c06a54c6e22"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:11:24 crc kubenswrapper[4846]: I1005 07:11:24.341389 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutrona5c0-account-delete-q5wg4" event={"ID":"e2f9f9a8-a25f-4a99-b425-4799964d4e24","Type":"ContainerDied","Data":"3df79f72c856d902d61af470174e9eb4cda409026d12528c0aece2ce522096e9"} Oct 05 07:11:24 crc kubenswrapper[4846]: I1005 07:11:24.341477 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutrona5c0-account-delete-q5wg4" Oct 05 07:11:24 crc kubenswrapper[4846]: I1005 07:11:24.345951 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican8fc5-account-delete-gjp47" Oct 05 07:11:24 crc kubenswrapper[4846]: I1005 07:11:24.363013 4846 scope.go:117] "RemoveContainer" containerID="708780fcc4181d993d6e7d44ef9444fa8e85b47f3a59d10a6fa23aa84fa714b0" Oct 05 07:11:24 crc kubenswrapper[4846]: I1005 07:11:24.364482 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"bfe96e0c-b9fc-438d-9b17-bec246d48446","Type":"ContainerDied","Data":"69bab1a17f0a227c56ad117e5f561df14c1c0d6f2a4fd25886109fe5afbec8fe"} Oct 05 07:11:24 crc kubenswrapper[4846]: I1005 07:11:24.364770 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 05 07:11:24 crc kubenswrapper[4846]: I1005 07:11:24.365302 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 05 07:11:24 crc kubenswrapper[4846]: I1005 07:11:24.370863 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-984785977-srp2v" Oct 05 07:11:24 crc kubenswrapper[4846]: I1005 07:11:24.390718 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6dxzx\" (UniqueName: \"kubernetes.io/projected/2adc8ba0-2810-4bf4-bd24-172b97fb9b22-kube-api-access-6dxzx\") pod \"2adc8ba0-2810-4bf4-bd24-172b97fb9b22\" (UID: \"2adc8ba0-2810-4bf4-bd24-172b97fb9b22\") " Oct 05 07:11:24 crc kubenswrapper[4846]: I1005 07:11:24.390783 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2adc8ba0-2810-4bf4-bd24-172b97fb9b22-internal-tls-certs\") pod \"2adc8ba0-2810-4bf4-bd24-172b97fb9b22\" (UID: \"2adc8ba0-2810-4bf4-bd24-172b97fb9b22\") " Oct 05 07:11:24 crc kubenswrapper[4846]: I1005 07:11:24.390846 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2adc8ba0-2810-4bf4-bd24-172b97fb9b22-config-data\") pod \"2adc8ba0-2810-4bf4-bd24-172b97fb9b22\" (UID: \"2adc8ba0-2810-4bf4-bd24-172b97fb9b22\") " Oct 05 07:11:24 crc kubenswrapper[4846]: I1005 07:11:24.390975 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"2adc8ba0-2810-4bf4-bd24-172b97fb9b22\" (UID: \"2adc8ba0-2810-4bf4-bd24-172b97fb9b22\") " Oct 05 07:11:24 crc kubenswrapper[4846]: I1005 07:11:24.391011 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2adc8ba0-2810-4bf4-bd24-172b97fb9b22-scripts\") pod \"2adc8ba0-2810-4bf4-bd24-172b97fb9b22\" (UID: \"2adc8ba0-2810-4bf4-bd24-172b97fb9b22\") " Oct 05 07:11:24 crc kubenswrapper[4846]: I1005 07:11:24.391102 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2adc8ba0-2810-4bf4-bd24-172b97fb9b22-logs\") pod \"2adc8ba0-2810-4bf4-bd24-172b97fb9b22\" (UID: \"2adc8ba0-2810-4bf4-bd24-172b97fb9b22\") " Oct 05 07:11:24 crc kubenswrapper[4846]: I1005 07:11:24.391227 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/2adc8ba0-2810-4bf4-bd24-172b97fb9b22-httpd-run\") pod \"2adc8ba0-2810-4bf4-bd24-172b97fb9b22\" (UID: \"2adc8ba0-2810-4bf4-bd24-172b97fb9b22\") " Oct 05 07:11:24 crc kubenswrapper[4846]: I1005 07:11:24.391293 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ds9zd\" (UniqueName: \"kubernetes.io/projected/a32cdd03-1398-4055-b051-62adde35200f-kube-api-access-ds9zd\") pod \"a32cdd03-1398-4055-b051-62adde35200f\" (UID: \"a32cdd03-1398-4055-b051-62adde35200f\") " Oct 05 07:11:24 crc kubenswrapper[4846]: I1005 07:11:24.391348 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2adc8ba0-2810-4bf4-bd24-172b97fb9b22-combined-ca-bundle\") pod \"2adc8ba0-2810-4bf4-bd24-172b97fb9b22\" (UID: \"2adc8ba0-2810-4bf4-bd24-172b97fb9b22\") " Oct 05 07:11:24 crc kubenswrapper[4846]: I1005 07:11:24.392372 4846 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/0c0f8308-455e-4544-9f53-0c06a54c6e22-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:24 crc kubenswrapper[4846]: I1005 07:11:24.392435 4846 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" " Oct 05 07:11:24 crc kubenswrapper[4846]: I1005 07:11:24.392479 4846 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0c0f8308-455e-4544-9f53-0c06a54c6e22-scripts\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:24 crc kubenswrapper[4846]: I1005 07:11:24.392501 4846 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0c0f8308-455e-4544-9f53-0c06a54c6e22-logs\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:24 crc kubenswrapper[4846]: I1005 07:11:24.392514 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjd99\" (UniqueName: \"kubernetes.io/projected/0c0f8308-455e-4544-9f53-0c06a54c6e22-kube-api-access-pjd99\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:24 crc kubenswrapper[4846]: I1005 07:11:24.393063 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2adc8ba0-2810-4bf4-bd24-172b97fb9b22-logs" (OuterVolumeSpecName: "logs") pod "2adc8ba0-2810-4bf4-bd24-172b97fb9b22" (UID: "2adc8ba0-2810-4bf4-bd24-172b97fb9b22"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 07:11:24 crc kubenswrapper[4846]: I1005 07:11:24.393664 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2adc8ba0-2810-4bf4-bd24-172b97fb9b22-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "2adc8ba0-2810-4bf4-bd24-172b97fb9b22" (UID: "2adc8ba0-2810-4bf4-bd24-172b97fb9b22"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 07:11:24 crc kubenswrapper[4846]: I1005 07:11:24.410327 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 05 07:11:24 crc kubenswrapper[4846]: I1005 07:11:24.432125 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage12-crc" (OuterVolumeSpecName: "glance") pod "2adc8ba0-2810-4bf4-bd24-172b97fb9b22" (UID: "2adc8ba0-2810-4bf4-bd24-172b97fb9b22"). InnerVolumeSpecName "local-storage12-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 05 07:11:24 crc kubenswrapper[4846]: I1005 07:11:24.437127 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2adc8ba0-2810-4bf4-bd24-172b97fb9b22-kube-api-access-6dxzx" (OuterVolumeSpecName: "kube-api-access-6dxzx") pod "2adc8ba0-2810-4bf4-bd24-172b97fb9b22" (UID: "2adc8ba0-2810-4bf4-bd24-172b97fb9b22"). InnerVolumeSpecName "kube-api-access-6dxzx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:11:24 crc kubenswrapper[4846]: I1005 07:11:24.437961 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 05 07:11:24 crc kubenswrapper[4846]: I1005 07:11:24.447012 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0c0f8308-455e-4544-9f53-0c06a54c6e22-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0c0f8308-455e-4544-9f53-0c06a54c6e22" (UID: "0c0f8308-455e-4544-9f53-0c06a54c6e22"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:11:24 crc kubenswrapper[4846]: I1005 07:11:24.453016 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a32cdd03-1398-4055-b051-62adde35200f-kube-api-access-ds9zd" (OuterVolumeSpecName: "kube-api-access-ds9zd") pod "a32cdd03-1398-4055-b051-62adde35200f" (UID: "a32cdd03-1398-4055-b051-62adde35200f"). InnerVolumeSpecName "kube-api-access-ds9zd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:11:24 crc kubenswrapper[4846]: I1005 07:11:24.470433 4846 scope.go:117] "RemoveContainer" containerID="623a1c303003463ea428d6d40cb44811fb558226463a79ed21399df208bb4505" Oct 05 07:11:24 crc kubenswrapper[4846]: I1005 07:11:24.496872 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-74pmj\" (UniqueName: \"kubernetes.io/projected/d11b6df6-6da2-4cb8-a4f9-d87cfd5d4d26-kube-api-access-74pmj\") pod \"d11b6df6-6da2-4cb8-a4f9-d87cfd5d4d26\" (UID: \"d11b6df6-6da2-4cb8-a4f9-d87cfd5d4d26\") " Oct 05 07:11:24 crc kubenswrapper[4846]: I1005 07:11:24.497132 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d11b6df6-6da2-4cb8-a4f9-d87cfd5d4d26-combined-ca-bundle\") pod \"d11b6df6-6da2-4cb8-a4f9-d87cfd5d4d26\" (UID: \"d11b6df6-6da2-4cb8-a4f9-d87cfd5d4d26\") " Oct 05 07:11:24 crc kubenswrapper[4846]: I1005 07:11:24.497241 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pvrgj\" (UniqueName: \"kubernetes.io/projected/086cf913-a4a2-41ee-910b-a83be95c69d6-kube-api-access-pvrgj\") pod \"086cf913-a4a2-41ee-910b-a83be95c69d6\" (UID: \"086cf913-a4a2-41ee-910b-a83be95c69d6\") " Oct 05 07:11:24 crc kubenswrapper[4846]: I1005 07:11:24.497349 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d11b6df6-6da2-4cb8-a4f9-d87cfd5d4d26-run-httpd\") pod \"d11b6df6-6da2-4cb8-a4f9-d87cfd5d4d26\" (UID: \"d11b6df6-6da2-4cb8-a4f9-d87cfd5d4d26\") " Oct 05 07:11:24 crc kubenswrapper[4846]: I1005 07:11:24.497443 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d11b6df6-6da2-4cb8-a4f9-d87cfd5d4d26-public-tls-certs\") pod \"d11b6df6-6da2-4cb8-a4f9-d87cfd5d4d26\" (UID: \"d11b6df6-6da2-4cb8-a4f9-d87cfd5d4d26\") " Oct 05 07:11:24 crc kubenswrapper[4846]: I1005 07:11:24.497600 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/d11b6df6-6da2-4cb8-a4f9-d87cfd5d4d26-etc-swift\") pod \"d11b6df6-6da2-4cb8-a4f9-d87cfd5d4d26\" (UID: \"d11b6df6-6da2-4cb8-a4f9-d87cfd5d4d26\") " Oct 05 07:11:24 crc kubenswrapper[4846]: I1005 07:11:24.497675 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d11b6df6-6da2-4cb8-a4f9-d87cfd5d4d26-internal-tls-certs\") pod \"d11b6df6-6da2-4cb8-a4f9-d87cfd5d4d26\" (UID: \"d11b6df6-6da2-4cb8-a4f9-d87cfd5d4d26\") " Oct 05 07:11:24 crc kubenswrapper[4846]: I1005 07:11:24.497784 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d11b6df6-6da2-4cb8-a4f9-d87cfd5d4d26-config-data\") pod \"d11b6df6-6da2-4cb8-a4f9-d87cfd5d4d26\" (UID: \"d11b6df6-6da2-4cb8-a4f9-d87cfd5d4d26\") " Oct 05 07:11:24 crc kubenswrapper[4846]: I1005 07:11:24.497874 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d11b6df6-6da2-4cb8-a4f9-d87cfd5d4d26-log-httpd\") pod \"d11b6df6-6da2-4cb8-a4f9-d87cfd5d4d26\" (UID: \"d11b6df6-6da2-4cb8-a4f9-d87cfd5d4d26\") " Oct 05 07:11:24 crc kubenswrapper[4846]: I1005 07:11:24.498412 4846 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2adc8ba0-2810-4bf4-bd24-172b97fb9b22-logs\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:24 crc kubenswrapper[4846]: I1005 07:11:24.498452 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2adc8ba0-2810-4bf4-bd24-172b97fb9b22-scripts" (OuterVolumeSpecName: "scripts") pod "2adc8ba0-2810-4bf4-bd24-172b97fb9b22" (UID: "2adc8ba0-2810-4bf4-bd24-172b97fb9b22"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:11:24 crc kubenswrapper[4846]: I1005 07:11:24.498481 4846 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/2adc8ba0-2810-4bf4-bd24-172b97fb9b22-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:24 crc kubenswrapper[4846]: I1005 07:11:24.498553 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ds9zd\" (UniqueName: \"kubernetes.io/projected/a32cdd03-1398-4055-b051-62adde35200f-kube-api-access-ds9zd\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:24 crc kubenswrapper[4846]: I1005 07:11:24.498576 4846 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0c0f8308-455e-4544-9f53-0c06a54c6e22-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:24 crc kubenswrapper[4846]: I1005 07:11:24.498591 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6dxzx\" (UniqueName: \"kubernetes.io/projected/2adc8ba0-2810-4bf4-bd24-172b97fb9b22-kube-api-access-6dxzx\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:24 crc kubenswrapper[4846]: I1005 07:11:24.498632 4846 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" " Oct 05 07:11:24 crc kubenswrapper[4846]: I1005 07:11:24.499149 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d11b6df6-6da2-4cb8-a4f9-d87cfd5d4d26-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "d11b6df6-6da2-4cb8-a4f9-d87cfd5d4d26" (UID: "d11b6df6-6da2-4cb8-a4f9-d87cfd5d4d26"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 07:11:24 crc kubenswrapper[4846]: I1005 07:11:24.504957 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2adc8ba0-2810-4bf4-bd24-172b97fb9b22-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2adc8ba0-2810-4bf4-bd24-172b97fb9b22" (UID: "2adc8ba0-2810-4bf4-bd24-172b97fb9b22"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:11:24 crc kubenswrapper[4846]: I1005 07:11:24.511234 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d11b6df6-6da2-4cb8-a4f9-d87cfd5d4d26-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "d11b6df6-6da2-4cb8-a4f9-d87cfd5d4d26" (UID: "d11b6df6-6da2-4cb8-a4f9-d87cfd5d4d26"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 07:11:24 crc kubenswrapper[4846]: I1005 07:11:24.518651 4846 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage07-crc" (UniqueName: "kubernetes.io/local-volume/local-storage07-crc") on node "crc" Oct 05 07:11:24 crc kubenswrapper[4846]: I1005 07:11:24.519829 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d11b6df6-6da2-4cb8-a4f9-d87cfd5d4d26-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "d11b6df6-6da2-4cb8-a4f9-d87cfd5d4d26" (UID: "d11b6df6-6da2-4cb8-a4f9-d87cfd5d4d26"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:11:24 crc kubenswrapper[4846]: I1005 07:11:24.526278 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 05 07:11:24 crc kubenswrapper[4846]: I1005 07:11:24.535629 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/086cf913-a4a2-41ee-910b-a83be95c69d6-kube-api-access-pvrgj" (OuterVolumeSpecName: "kube-api-access-pvrgj") pod "086cf913-a4a2-41ee-910b-a83be95c69d6" (UID: "086cf913-a4a2-41ee-910b-a83be95c69d6"). InnerVolumeSpecName "kube-api-access-pvrgj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:11:24 crc kubenswrapper[4846]: I1005 07:11:24.540823 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="221918b0-eb0e-4a32-8eed-07acd18bd31e" path="/var/lib/kubelet/pods/221918b0-eb0e-4a32-8eed-07acd18bd31e/volumes" Oct 05 07:11:24 crc kubenswrapper[4846]: I1005 07:11:24.546282 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="33193805-7aa6-406e-8cad-0ebbb1228ef2" path="/var/lib/kubelet/pods/33193805-7aa6-406e-8cad-0ebbb1228ef2/volumes" Oct 05 07:11:24 crc kubenswrapper[4846]: I1005 07:11:24.546529 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d11b6df6-6da2-4cb8-a4f9-d87cfd5d4d26-kube-api-access-74pmj" (OuterVolumeSpecName: "kube-api-access-74pmj") pod "d11b6df6-6da2-4cb8-a4f9-d87cfd5d4d26" (UID: "d11b6df6-6da2-4cb8-a4f9-d87cfd5d4d26"). InnerVolumeSpecName "kube-api-access-74pmj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:11:24 crc kubenswrapper[4846]: I1005 07:11:24.546902 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0c0f8308-455e-4544-9f53-0c06a54c6e22-config-data" (OuterVolumeSpecName: "config-data") pod "0c0f8308-455e-4544-9f53-0c06a54c6e22" (UID: "0c0f8308-455e-4544-9f53-0c06a54c6e22"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:11:24 crc kubenswrapper[4846]: I1005 07:11:24.548666 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b7de2524-bf5c-4668-af74-7584edf72c8b" path="/var/lib/kubelet/pods/b7de2524-bf5c-4668-af74-7584edf72c8b/volumes" Oct 05 07:11:24 crc kubenswrapper[4846]: I1005 07:11:24.550970 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd630f2e-4b0e-4cc7-a0de-4f60c72d941c" path="/var/lib/kubelet/pods/bd630f2e-4b0e-4cc7-a0de-4f60c72d941c/volumes" Oct 05 07:11:24 crc kubenswrapper[4846]: I1005 07:11:24.552144 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cfa439eb-10b4-4b11-9baa-377a45fe68d0" path="/var/lib/kubelet/pods/cfa439eb-10b4-4b11-9baa-377a45fe68d0/volumes" Oct 05 07:11:24 crc kubenswrapper[4846]: I1005 07:11:24.560482 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0c0f8308-455e-4544-9f53-0c06a54c6e22-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "0c0f8308-455e-4544-9f53-0c06a54c6e22" (UID: "0c0f8308-455e-4544-9f53-0c06a54c6e22"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:11:24 crc kubenswrapper[4846]: I1005 07:11:24.564172 4846 scope.go:117] "RemoveContainer" containerID="316083fa9d64f1f90ee62226b3c28e38700ed5382c669302dda9655852943dab" Oct 05 07:11:24 crc kubenswrapper[4846]: I1005 07:11:24.567137 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fc35db15-65d4-4cd6-9d09-059822904757" path="/var/lib/kubelet/pods/fc35db15-65d4-4cd6-9d09-059822904757/volumes" Oct 05 07:11:24 crc kubenswrapper[4846]: I1005 07:11:24.600178 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/36dca705-2f72-411e-8d80-ed31d49989cc-public-tls-certs\") pod \"36dca705-2f72-411e-8d80-ed31d49989cc\" (UID: \"36dca705-2f72-411e-8d80-ed31d49989cc\") " Oct 05 07:11:24 crc kubenswrapper[4846]: I1005 07:11:24.600256 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/36dca705-2f72-411e-8d80-ed31d49989cc-logs\") pod \"36dca705-2f72-411e-8d80-ed31d49989cc\" (UID: \"36dca705-2f72-411e-8d80-ed31d49989cc\") " Oct 05 07:11:24 crc kubenswrapper[4846]: I1005 07:11:24.600325 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/36dca705-2f72-411e-8d80-ed31d49989cc-internal-tls-certs\") pod \"36dca705-2f72-411e-8d80-ed31d49989cc\" (UID: \"36dca705-2f72-411e-8d80-ed31d49989cc\") " Oct 05 07:11:24 crc kubenswrapper[4846]: I1005 07:11:24.600405 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/36dca705-2f72-411e-8d80-ed31d49989cc-config-data-custom\") pod \"36dca705-2f72-411e-8d80-ed31d49989cc\" (UID: \"36dca705-2f72-411e-8d80-ed31d49989cc\") " Oct 05 07:11:24 crc kubenswrapper[4846]: I1005 07:11:24.600522 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/36dca705-2f72-411e-8d80-ed31d49989cc-config-data\") pod \"36dca705-2f72-411e-8d80-ed31d49989cc\" (UID: \"36dca705-2f72-411e-8d80-ed31d49989cc\") " Oct 05 07:11:24 crc kubenswrapper[4846]: I1005 07:11:24.600576 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xld6b\" (UniqueName: \"kubernetes.io/projected/36dca705-2f72-411e-8d80-ed31d49989cc-kube-api-access-xld6b\") pod \"36dca705-2f72-411e-8d80-ed31d49989cc\" (UID: \"36dca705-2f72-411e-8d80-ed31d49989cc\") " Oct 05 07:11:24 crc kubenswrapper[4846]: I1005 07:11:24.600615 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/36dca705-2f72-411e-8d80-ed31d49989cc-scripts\") pod \"36dca705-2f72-411e-8d80-ed31d49989cc\" (UID: \"36dca705-2f72-411e-8d80-ed31d49989cc\") " Oct 05 07:11:24 crc kubenswrapper[4846]: I1005 07:11:24.600664 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/36dca705-2f72-411e-8d80-ed31d49989cc-combined-ca-bundle\") pod \"36dca705-2f72-411e-8d80-ed31d49989cc\" (UID: \"36dca705-2f72-411e-8d80-ed31d49989cc\") " Oct 05 07:11:24 crc kubenswrapper[4846]: I1005 07:11:24.600700 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/36dca705-2f72-411e-8d80-ed31d49989cc-etc-machine-id\") pod \"36dca705-2f72-411e-8d80-ed31d49989cc\" (UID: \"36dca705-2f72-411e-8d80-ed31d49989cc\") " Oct 05 07:11:24 crc kubenswrapper[4846]: I1005 07:11:24.601591 4846 reconciler_common.go:293] "Volume detached for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:24 crc kubenswrapper[4846]: I1005 07:11:24.601619 4846 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2adc8ba0-2810-4bf4-bd24-172b97fb9b22-scripts\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:24 crc kubenswrapper[4846]: I1005 07:11:24.601634 4846 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/d11b6df6-6da2-4cb8-a4f9-d87cfd5d4d26-etc-swift\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:24 crc kubenswrapper[4846]: I1005 07:11:24.601644 4846 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0c0f8308-455e-4544-9f53-0c06a54c6e22-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:24 crc kubenswrapper[4846]: I1005 07:11:24.601655 4846 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d11b6df6-6da2-4cb8-a4f9-d87cfd5d4d26-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:24 crc kubenswrapper[4846]: I1005 07:11:24.601664 4846 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0c0f8308-455e-4544-9f53-0c06a54c6e22-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:24 crc kubenswrapper[4846]: I1005 07:11:24.601672 4846 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2adc8ba0-2810-4bf4-bd24-172b97fb9b22-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:24 crc kubenswrapper[4846]: I1005 07:11:24.601682 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-74pmj\" (UniqueName: \"kubernetes.io/projected/d11b6df6-6da2-4cb8-a4f9-d87cfd5d4d26-kube-api-access-74pmj\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:24 crc kubenswrapper[4846]: I1005 07:11:24.601694 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pvrgj\" (UniqueName: \"kubernetes.io/projected/086cf913-a4a2-41ee-910b-a83be95c69d6-kube-api-access-pvrgj\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:24 crc kubenswrapper[4846]: I1005 07:11:24.601702 4846 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d11b6df6-6da2-4cb8-a4f9-d87cfd5d4d26-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:24 crc kubenswrapper[4846]: I1005 07:11:24.603152 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/36dca705-2f72-411e-8d80-ed31d49989cc-logs" (OuterVolumeSpecName: "logs") pod "36dca705-2f72-411e-8d80-ed31d49989cc" (UID: "36dca705-2f72-411e-8d80-ed31d49989cc"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 07:11:24 crc kubenswrapper[4846]: I1005 07:11:24.603575 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/36dca705-2f72-411e-8d80-ed31d49989cc-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "36dca705-2f72-411e-8d80-ed31d49989cc" (UID: "36dca705-2f72-411e-8d80-ed31d49989cc"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 05 07:11:24 crc kubenswrapper[4846]: I1005 07:11:24.618625 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/36dca705-2f72-411e-8d80-ed31d49989cc-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "36dca705-2f72-411e-8d80-ed31d49989cc" (UID: "36dca705-2f72-411e-8d80-ed31d49989cc"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:11:24 crc kubenswrapper[4846]: I1005 07:11:24.618807 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/36dca705-2f72-411e-8d80-ed31d49989cc-kube-api-access-xld6b" (OuterVolumeSpecName: "kube-api-access-xld6b") pod "36dca705-2f72-411e-8d80-ed31d49989cc" (UID: "36dca705-2f72-411e-8d80-ed31d49989cc"). InnerVolumeSpecName "kube-api-access-xld6b". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:11:24 crc kubenswrapper[4846]: I1005 07:11:24.619083 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2adc8ba0-2810-4bf4-bd24-172b97fb9b22-config-data" (OuterVolumeSpecName: "config-data") pod "2adc8ba0-2810-4bf4-bd24-172b97fb9b22" (UID: "2adc8ba0-2810-4bf4-bd24-172b97fb9b22"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:11:24 crc kubenswrapper[4846]: I1005 07:11:24.622073 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/36dca705-2f72-411e-8d80-ed31d49989cc-scripts" (OuterVolumeSpecName: "scripts") pod "36dca705-2f72-411e-8d80-ed31d49989cc" (UID: "36dca705-2f72-411e-8d80-ed31d49989cc"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:11:24 crc kubenswrapper[4846]: I1005 07:11:24.643812 4846 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage12-crc" (UniqueName: "kubernetes.io/local-volume/local-storage12-crc") on node "crc" Oct 05 07:11:24 crc kubenswrapper[4846]: I1005 07:11:24.660666 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2adc8ba0-2810-4bf4-bd24-172b97fb9b22-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "2adc8ba0-2810-4bf4-bd24-172b97fb9b22" (UID: "2adc8ba0-2810-4bf4-bd24-172b97fb9b22"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:11:24 crc kubenswrapper[4846]: I1005 07:11:24.660679 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d11b6df6-6da2-4cb8-a4f9-d87cfd5d4d26-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "d11b6df6-6da2-4cb8-a4f9-d87cfd5d4d26" (UID: "d11b6df6-6da2-4cb8-a4f9-d87cfd5d4d26"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:11:24 crc kubenswrapper[4846]: I1005 07:11:24.675960 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/36dca705-2f72-411e-8d80-ed31d49989cc-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "36dca705-2f72-411e-8d80-ed31d49989cc" (UID: "36dca705-2f72-411e-8d80-ed31d49989cc"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:11:24 crc kubenswrapper[4846]: I1005 07:11:24.685875 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d11b6df6-6da2-4cb8-a4f9-d87cfd5d4d26-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "d11b6df6-6da2-4cb8-a4f9-d87cfd5d4d26" (UID: "d11b6df6-6da2-4cb8-a4f9-d87cfd5d4d26"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:11:24 crc kubenswrapper[4846]: I1005 07:11:24.707568 4846 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/36dca705-2f72-411e-8d80-ed31d49989cc-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:24 crc kubenswrapper[4846]: I1005 07:11:24.707616 4846 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/36dca705-2f72-411e-8d80-ed31d49989cc-logs\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:24 crc kubenswrapper[4846]: I1005 07:11:24.707630 4846 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d11b6df6-6da2-4cb8-a4f9-d87cfd5d4d26-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:24 crc kubenswrapper[4846]: I1005 07:11:24.707643 4846 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2adc8ba0-2810-4bf4-bd24-172b97fb9b22-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:24 crc kubenswrapper[4846]: I1005 07:11:24.707655 4846 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2adc8ba0-2810-4bf4-bd24-172b97fb9b22-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:24 crc kubenswrapper[4846]: I1005 07:11:24.707666 4846 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/36dca705-2f72-411e-8d80-ed31d49989cc-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:24 crc kubenswrapper[4846]: I1005 07:11:24.707679 4846 reconciler_common.go:293] "Volume detached for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:24 crc kubenswrapper[4846]: I1005 07:11:24.707691 4846 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d11b6df6-6da2-4cb8-a4f9-d87cfd5d4d26-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:24 crc kubenswrapper[4846]: I1005 07:11:24.707702 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xld6b\" (UniqueName: \"kubernetes.io/projected/36dca705-2f72-411e-8d80-ed31d49989cc-kube-api-access-xld6b\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:24 crc kubenswrapper[4846]: I1005 07:11:24.707715 4846 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/36dca705-2f72-411e-8d80-ed31d49989cc-scripts\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:24 crc kubenswrapper[4846]: I1005 07:11:24.707726 4846 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/36dca705-2f72-411e-8d80-ed31d49989cc-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:24 crc kubenswrapper[4846]: I1005 07:11:24.777386 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d11b6df6-6da2-4cb8-a4f9-d87cfd5d4d26-config-data" (OuterVolumeSpecName: "config-data") pod "d11b6df6-6da2-4cb8-a4f9-d87cfd5d4d26" (UID: "d11b6df6-6da2-4cb8-a4f9-d87cfd5d4d26"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:11:24 crc kubenswrapper[4846]: I1005 07:11:24.810454 4846 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d11b6df6-6da2-4cb8-a4f9-d87cfd5d4d26-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:24 crc kubenswrapper[4846]: E1005 07:11:24.842713 4846 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="4351908cf00eea1071aca5080f9d536de25e56a78e5d0713bec8f51caab84ecb" cmd=["/usr/local/bin/container-scripts/status_check.sh"] Oct 05 07:11:24 crc kubenswrapper[4846]: E1005 07:11:24.846654 4846 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="4351908cf00eea1071aca5080f9d536de25e56a78e5d0713bec8f51caab84ecb" cmd=["/usr/local/bin/container-scripts/status_check.sh"] Oct 05 07:11:24 crc kubenswrapper[4846]: I1005 07:11:24.847402 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/36dca705-2f72-411e-8d80-ed31d49989cc-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "36dca705-2f72-411e-8d80-ed31d49989cc" (UID: "36dca705-2f72-411e-8d80-ed31d49989cc"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:11:24 crc kubenswrapper[4846]: E1005 07:11:24.849906 4846 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="4351908cf00eea1071aca5080f9d536de25e56a78e5d0713bec8f51caab84ecb" cmd=["/usr/local/bin/container-scripts/status_check.sh"] Oct 05 07:11:24 crc kubenswrapper[4846]: E1005 07:11:24.850062 4846 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-northd-0" podUID="18a7be37-c8dc-4d8a-808f-2a63eb4d9692" containerName="ovn-northd" Oct 05 07:11:24 crc kubenswrapper[4846]: I1005 07:11:24.904684 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/36dca705-2f72-411e-8d80-ed31d49989cc-config-data" (OuterVolumeSpecName: "config-data") pod "36dca705-2f72-411e-8d80-ed31d49989cc" (UID: "36dca705-2f72-411e-8d80-ed31d49989cc"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:11:24 crc kubenswrapper[4846]: I1005 07:11:24.911931 4846 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/36dca705-2f72-411e-8d80-ed31d49989cc-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:24 crc kubenswrapper[4846]: I1005 07:11:24.911956 4846 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/36dca705-2f72-411e-8d80-ed31d49989cc-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:24 crc kubenswrapper[4846]: I1005 07:11:24.936226 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 05 07:11:24 crc kubenswrapper[4846]: I1005 07:11:24.936629 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 05 07:11:24 crc kubenswrapper[4846]: I1005 07:11:24.936647 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutrona5c0-account-delete-q5wg4"] Oct 05 07:11:24 crc kubenswrapper[4846]: I1005 07:11:24.936683 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutrona5c0-account-delete-q5wg4"] Oct 05 07:11:24 crc kubenswrapper[4846]: I1005 07:11:24.936696 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 05 07:11:24 crc kubenswrapper[4846]: I1005 07:11:24.936709 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 05 07:11:24 crc kubenswrapper[4846]: I1005 07:11:24.936724 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 05 07:11:24 crc kubenswrapper[4846]: I1005 07:11:24.936739 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 05 07:11:24 crc kubenswrapper[4846]: I1005 07:11:24.936931 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="57b09608-08e5-4998-a3b3-e88b13587f58" containerName="kube-state-metrics" containerID="cri-o://0f36596d6da69b2a28102d8a73a67b3c631900ebfdb71663f52974e9dd99b1d1" gracePeriod=30 Oct 05 07:11:24 crc kubenswrapper[4846]: I1005 07:11:24.951020 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="b1e394dd-f55b-4af1-af0b-52cc7d09afa1" containerName="ceilometer-central-agent" containerID="cri-o://6682a3a723ef5336f655521943bfbd504709a37185a094744e9c63a0c851b903" gracePeriod=30 Oct 05 07:11:24 crc kubenswrapper[4846]: I1005 07:11:24.951339 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="b1e394dd-f55b-4af1-af0b-52cc7d09afa1" containerName="proxy-httpd" containerID="cri-o://e92d479cbafa7786e2514b2d1e99c120ef33197a849bb0c13e2362354bcebd01" gracePeriod=30 Oct 05 07:11:24 crc kubenswrapper[4846]: I1005 07:11:24.951384 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="b1e394dd-f55b-4af1-af0b-52cc7d09afa1" containerName="sg-core" containerID="cri-o://6db54d2ca22110333154adb6a23a0ec27e54e4f7118d44ea9636189fffd9e0ae" gracePeriod=30 Oct 05 07:11:24 crc kubenswrapper[4846]: I1005 07:11:24.951589 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="b1e394dd-f55b-4af1-af0b-52cc7d09afa1" containerName="ceilometer-notification-agent" containerID="cri-o://d0d0db8b415aac7ced1865679a2764d980d7574715bf4b8a924f0edb74f9f2fc" gracePeriod=30 Oct 05 07:11:25 crc kubenswrapper[4846]: I1005 07:11:25.006992 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 05 07:11:25 crc kubenswrapper[4846]: I1005 07:11:25.010500 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-65884db4f4-fsk6p" Oct 05 07:11:25 crc kubenswrapper[4846]: I1005 07:11:25.017014 4846 scope.go:117] "RemoveContainer" containerID="a03ec35e533cac1f50168f28070b82ea570152d6d222e9cf097e6f662b34b418" Oct 05 07:11:25 crc kubenswrapper[4846]: I1005 07:11:25.017164 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-57674d6c54-zn5nq" Oct 05 07:11:25 crc kubenswrapper[4846]: I1005 07:11:25.018747 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 05 07:11:25 crc kubenswrapper[4846]: I1005 07:11:25.035252 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/36dca705-2f72-411e-8d80-ed31d49989cc-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "36dca705-2f72-411e-8d80-ed31d49989cc" (UID: "36dca705-2f72-411e-8d80-ed31d49989cc"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:11:25 crc kubenswrapper[4846]: I1005 07:11:25.035741 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/memcached-0"] Oct 05 07:11:25 crc kubenswrapper[4846]: I1005 07:11:25.039345 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/memcached-0" podUID="751734cb-38c8-46a0-9bac-05bcf24cc589" containerName="memcached" containerID="cri-o://ed2b341c0cbc39aa9be3b1731eb5e7d43c786f18e7d5a05613b33a6282bdffcd" gracePeriod=30 Oct 05 07:11:25 crc kubenswrapper[4846]: I1005 07:11:25.067575 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 05 07:11:25 crc kubenswrapper[4846]: I1005 07:11:25.107681 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d11b6df6-6da2-4cb8-a4f9-d87cfd5d4d26-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d11b6df6-6da2-4cb8-a4f9-d87cfd5d4d26" (UID: "d11b6df6-6da2-4cb8-a4f9-d87cfd5d4d26"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:11:25 crc kubenswrapper[4846]: I1005 07:11:25.122712 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c31937eb-9836-461b-9230-5e06ce7a99e8-config-data-custom\") pod \"c31937eb-9836-461b-9230-5e06ce7a99e8\" (UID: \"c31937eb-9836-461b-9230-5e06ce7a99e8\") " Oct 05 07:11:25 crc kubenswrapper[4846]: I1005 07:11:25.122775 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6e9e53c6-4cf2-4988-89b4-c69471f616aa-public-tls-certs\") pod \"6e9e53c6-4cf2-4988-89b4-c69471f616aa\" (UID: \"6e9e53c6-4cf2-4988-89b4-c69471f616aa\") " Oct 05 07:11:25 crc kubenswrapper[4846]: I1005 07:11:25.122825 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c35affe0-da4e-448f-aac0-7bba26f89040-config-data\") pod \"c35affe0-da4e-448f-aac0-7bba26f89040\" (UID: \"c35affe0-da4e-448f-aac0-7bba26f89040\") " Oct 05 07:11:25 crc kubenswrapper[4846]: I1005 07:11:25.122877 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c31937eb-9836-461b-9230-5e06ce7a99e8-combined-ca-bundle\") pod \"c31937eb-9836-461b-9230-5e06ce7a99e8\" (UID: \"c31937eb-9836-461b-9230-5e06ce7a99e8\") " Oct 05 07:11:25 crc kubenswrapper[4846]: I1005 07:11:25.122916 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c31937eb-9836-461b-9230-5e06ce7a99e8-logs\") pod \"c31937eb-9836-461b-9230-5e06ce7a99e8\" (UID: \"c31937eb-9836-461b-9230-5e06ce7a99e8\") " Oct 05 07:11:25 crc kubenswrapper[4846]: I1005 07:11:25.122942 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c31937eb-9836-461b-9230-5e06ce7a99e8-public-tls-certs\") pod \"c31937eb-9836-461b-9230-5e06ce7a99e8\" (UID: \"c31937eb-9836-461b-9230-5e06ce7a99e8\") " Oct 05 07:11:25 crc kubenswrapper[4846]: I1005 07:11:25.122981 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c35affe0-da4e-448f-aac0-7bba26f89040-internal-tls-certs\") pod \"c35affe0-da4e-448f-aac0-7bba26f89040\" (UID: \"c35affe0-da4e-448f-aac0-7bba26f89040\") " Oct 05 07:11:25 crc kubenswrapper[4846]: I1005 07:11:25.123009 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c35affe0-da4e-448f-aac0-7bba26f89040-public-tls-certs\") pod \"c35affe0-da4e-448f-aac0-7bba26f89040\" (UID: \"c35affe0-da4e-448f-aac0-7bba26f89040\") " Oct 05 07:11:25 crc kubenswrapper[4846]: I1005 07:11:25.123041 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2mxsf\" (UniqueName: \"kubernetes.io/projected/c35affe0-da4e-448f-aac0-7bba26f89040-kube-api-access-2mxsf\") pod \"c35affe0-da4e-448f-aac0-7bba26f89040\" (UID: \"c35affe0-da4e-448f-aac0-7bba26f89040\") " Oct 05 07:11:25 crc kubenswrapper[4846]: I1005 07:11:25.123108 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x6fpj\" (UniqueName: \"kubernetes.io/projected/c31937eb-9836-461b-9230-5e06ce7a99e8-kube-api-access-x6fpj\") pod \"c31937eb-9836-461b-9230-5e06ce7a99e8\" (UID: \"c31937eb-9836-461b-9230-5e06ce7a99e8\") " Oct 05 07:11:25 crc kubenswrapper[4846]: I1005 07:11:25.123141 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c31937eb-9836-461b-9230-5e06ce7a99e8-config-data\") pod \"c31937eb-9836-461b-9230-5e06ce7a99e8\" (UID: \"c31937eb-9836-461b-9230-5e06ce7a99e8\") " Oct 05 07:11:25 crc kubenswrapper[4846]: I1005 07:11:25.123171 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c31937eb-9836-461b-9230-5e06ce7a99e8-internal-tls-certs\") pod \"c31937eb-9836-461b-9230-5e06ce7a99e8\" (UID: \"c31937eb-9836-461b-9230-5e06ce7a99e8\") " Oct 05 07:11:25 crc kubenswrapper[4846]: I1005 07:11:25.123247 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6e9e53c6-4cf2-4988-89b4-c69471f616aa-config-data\") pod \"6e9e53c6-4cf2-4988-89b4-c69471f616aa\" (UID: \"6e9e53c6-4cf2-4988-89b4-c69471f616aa\") " Oct 05 07:11:25 crc kubenswrapper[4846]: I1005 07:11:25.123285 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6e9e53c6-4cf2-4988-89b4-c69471f616aa-internal-tls-certs\") pod \"6e9e53c6-4cf2-4988-89b4-c69471f616aa\" (UID: \"6e9e53c6-4cf2-4988-89b4-c69471f616aa\") " Oct 05 07:11:25 crc kubenswrapper[4846]: I1005 07:11:25.123326 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mc992\" (UniqueName: \"kubernetes.io/projected/6e9e53c6-4cf2-4988-89b4-c69471f616aa-kube-api-access-mc992\") pod \"6e9e53c6-4cf2-4988-89b4-c69471f616aa\" (UID: \"6e9e53c6-4cf2-4988-89b4-c69471f616aa\") " Oct 05 07:11:25 crc kubenswrapper[4846]: I1005 07:11:25.123355 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6e9e53c6-4cf2-4988-89b4-c69471f616aa-logs\") pod \"6e9e53c6-4cf2-4988-89b4-c69471f616aa\" (UID: \"6e9e53c6-4cf2-4988-89b4-c69471f616aa\") " Oct 05 07:11:25 crc kubenswrapper[4846]: I1005 07:11:25.123377 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c35affe0-da4e-448f-aac0-7bba26f89040-logs\") pod \"c35affe0-da4e-448f-aac0-7bba26f89040\" (UID: \"c35affe0-da4e-448f-aac0-7bba26f89040\") " Oct 05 07:11:25 crc kubenswrapper[4846]: I1005 07:11:25.123417 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c35affe0-da4e-448f-aac0-7bba26f89040-scripts\") pod \"c35affe0-da4e-448f-aac0-7bba26f89040\" (UID: \"c35affe0-da4e-448f-aac0-7bba26f89040\") " Oct 05 07:11:25 crc kubenswrapper[4846]: I1005 07:11:25.123457 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c35affe0-da4e-448f-aac0-7bba26f89040-combined-ca-bundle\") pod \"c35affe0-da4e-448f-aac0-7bba26f89040\" (UID: \"c35affe0-da4e-448f-aac0-7bba26f89040\") " Oct 05 07:11:25 crc kubenswrapper[4846]: I1005 07:11:25.123480 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6e9e53c6-4cf2-4988-89b4-c69471f616aa-combined-ca-bundle\") pod \"6e9e53c6-4cf2-4988-89b4-c69471f616aa\" (UID: \"6e9e53c6-4cf2-4988-89b4-c69471f616aa\") " Oct 05 07:11:25 crc kubenswrapper[4846]: I1005 07:11:25.135918 4846 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d11b6df6-6da2-4cb8-a4f9-d87cfd5d4d26-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:25 crc kubenswrapper[4846]: I1005 07:11:25.135963 4846 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/36dca705-2f72-411e-8d80-ed31d49989cc-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:25 crc kubenswrapper[4846]: I1005 07:11:25.152532 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-sync-t7t67"] Oct 05 07:11:25 crc kubenswrapper[4846]: I1005 07:11:25.163206 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c31937eb-9836-461b-9230-5e06ce7a99e8-logs" (OuterVolumeSpecName: "logs") pod "c31937eb-9836-461b-9230-5e06ce7a99e8" (UID: "c31937eb-9836-461b-9230-5e06ce7a99e8"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 07:11:25 crc kubenswrapper[4846]: I1005 07:11:25.184455 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 05 07:11:25 crc kubenswrapper[4846]: I1005 07:11:25.184987 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone67b1-account-delete-7lt6j"] Oct 05 07:11:25 crc kubenswrapper[4846]: I1005 07:11:25.185431 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c35affe0-da4e-448f-aac0-7bba26f89040-logs" (OuterVolumeSpecName: "logs") pod "c35affe0-da4e-448f-aac0-7bba26f89040" (UID: "c35affe0-da4e-448f-aac0-7bba26f89040"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 07:11:25 crc kubenswrapper[4846]: E1005 07:11:25.185525 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e6e86b1c-87ac-4594-aeba-943947065374" containerName="nova-metadata-metadata" Oct 05 07:11:25 crc kubenswrapper[4846]: I1005 07:11:25.185582 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="e6e86b1c-87ac-4594-aeba-943947065374" containerName="nova-metadata-metadata" Oct 05 07:11:25 crc kubenswrapper[4846]: E1005 07:11:25.185606 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6e9e53c6-4cf2-4988-89b4-c69471f616aa" containerName="nova-api-api" Oct 05 07:11:25 crc kubenswrapper[4846]: I1005 07:11:25.185636 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="6e9e53c6-4cf2-4988-89b4-c69471f616aa" containerName="nova-api-api" Oct 05 07:11:25 crc kubenswrapper[4846]: E1005 07:11:25.185679 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="36dca705-2f72-411e-8d80-ed31d49989cc" containerName="cinder-api" Oct 05 07:11:25 crc kubenswrapper[4846]: I1005 07:11:25.185792 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="36dca705-2f72-411e-8d80-ed31d49989cc" containerName="cinder-api" Oct 05 07:11:25 crc kubenswrapper[4846]: E1005 07:11:25.185813 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a32cdd03-1398-4055-b051-62adde35200f" containerName="mariadb-account-delete" Oct 05 07:11:25 crc kubenswrapper[4846]: I1005 07:11:25.185820 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="a32cdd03-1398-4055-b051-62adde35200f" containerName="mariadb-account-delete" Oct 05 07:11:25 crc kubenswrapper[4846]: E1005 07:11:25.185829 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="87e3744a-b8db-4e76-afa9-0e1129410d4c" containerName="openstack-network-exporter" Oct 05 07:11:25 crc kubenswrapper[4846]: I1005 07:11:25.185838 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="87e3744a-b8db-4e76-afa9-0e1129410d4c" containerName="openstack-network-exporter" Oct 05 07:11:25 crc kubenswrapper[4846]: E1005 07:11:25.185886 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c31937eb-9836-461b-9230-5e06ce7a99e8" containerName="barbican-api" Oct 05 07:11:25 crc kubenswrapper[4846]: I1005 07:11:25.185894 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="c31937eb-9836-461b-9230-5e06ce7a99e8" containerName="barbican-api" Oct 05 07:11:25 crc kubenswrapper[4846]: E1005 07:11:25.185903 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c35affe0-da4e-448f-aac0-7bba26f89040" containerName="placement-log" Oct 05 07:11:25 crc kubenswrapper[4846]: I1005 07:11:25.185910 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="c35affe0-da4e-448f-aac0-7bba26f89040" containerName="placement-log" Oct 05 07:11:25 crc kubenswrapper[4846]: E1005 07:11:25.185920 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="33193805-7aa6-406e-8cad-0ebbb1228ef2" containerName="nova-cell1-conductor-conductor" Oct 05 07:11:25 crc kubenswrapper[4846]: I1005 07:11:25.185927 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="33193805-7aa6-406e-8cad-0ebbb1228ef2" containerName="nova-cell1-conductor-conductor" Oct 05 07:11:25 crc kubenswrapper[4846]: E1005 07:11:25.185969 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="19f3a362-4973-4437-ac49-de71508aa4f2" containerName="openstack-network-exporter" Oct 05 07:11:25 crc kubenswrapper[4846]: I1005 07:11:25.185978 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="19f3a362-4973-4437-ac49-de71508aa4f2" containerName="openstack-network-exporter" Oct 05 07:11:25 crc kubenswrapper[4846]: E1005 07:11:25.185995 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="221918b0-eb0e-4a32-8eed-07acd18bd31e" containerName="galera" Oct 05 07:11:25 crc kubenswrapper[4846]: I1005 07:11:25.186003 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="221918b0-eb0e-4a32-8eed-07acd18bd31e" containerName="galera" Oct 05 07:11:25 crc kubenswrapper[4846]: E1005 07:11:25.186045 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6660b9ff-df9c-478e-b852-82cb5fb2c60c" containerName="nova-cell1-novncproxy-novncproxy" Oct 05 07:11:25 crc kubenswrapper[4846]: I1005 07:11:25.186055 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="6660b9ff-df9c-478e-b852-82cb5fb2c60c" containerName="nova-cell1-novncproxy-novncproxy" Oct 05 07:11:25 crc kubenswrapper[4846]: E1005 07:11:25.186069 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e2f9f9a8-a25f-4a99-b425-4799964d4e24" containerName="mariadb-account-delete" Oct 05 07:11:25 crc kubenswrapper[4846]: I1005 07:11:25.186076 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="e2f9f9a8-a25f-4a99-b425-4799964d4e24" containerName="mariadb-account-delete" Oct 05 07:11:25 crc kubenswrapper[4846]: E1005 07:11:25.186087 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7a5c4c82-4b3f-4314-a32d-db5398ca8a9e" containerName="init" Oct 05 07:11:25 crc kubenswrapper[4846]: I1005 07:11:25.186095 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="7a5c4c82-4b3f-4314-a32d-db5398ca8a9e" containerName="init" Oct 05 07:11:25 crc kubenswrapper[4846]: E1005 07:11:25.186318 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e6e86b1c-87ac-4594-aeba-943947065374" containerName="nova-metadata-log" Oct 05 07:11:25 crc kubenswrapper[4846]: I1005 07:11:25.186330 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="e6e86b1c-87ac-4594-aeba-943947065374" containerName="nova-metadata-log" Oct 05 07:11:25 crc kubenswrapper[4846]: E1005 07:11:25.186344 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2adc8ba0-2810-4bf4-bd24-172b97fb9b22" containerName="glance-log" Oct 05 07:11:25 crc kubenswrapper[4846]: I1005 07:11:25.186352 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="2adc8ba0-2810-4bf4-bd24-172b97fb9b22" containerName="glance-log" Oct 05 07:11:25 crc kubenswrapper[4846]: E1005 07:11:25.186398 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c31937eb-9836-461b-9230-5e06ce7a99e8" containerName="barbican-api-log" Oct 05 07:11:25 crc kubenswrapper[4846]: I1005 07:11:25.186407 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="c31937eb-9836-461b-9230-5e06ce7a99e8" containerName="barbican-api-log" Oct 05 07:11:25 crc kubenswrapper[4846]: E1005 07:11:25.186416 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bfe96e0c-b9fc-438d-9b17-bec246d48446" containerName="probe" Oct 05 07:11:25 crc kubenswrapper[4846]: I1005 07:11:25.186422 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="bfe96e0c-b9fc-438d-9b17-bec246d48446" containerName="probe" Oct 05 07:11:25 crc kubenswrapper[4846]: E1005 07:11:25.186430 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0c0f8308-455e-4544-9f53-0c06a54c6e22" containerName="glance-log" Oct 05 07:11:25 crc kubenswrapper[4846]: I1005 07:11:25.186435 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="0c0f8308-455e-4544-9f53-0c06a54c6e22" containerName="glance-log" Oct 05 07:11:25 crc kubenswrapper[4846]: E1005 07:11:25.186447 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6e9e53c6-4cf2-4988-89b4-c69471f616aa" containerName="nova-api-log" Oct 05 07:11:25 crc kubenswrapper[4846]: I1005 07:11:25.186452 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="6e9e53c6-4cf2-4988-89b4-c69471f616aa" containerName="nova-api-log" Oct 05 07:11:25 crc kubenswrapper[4846]: E1005 07:11:25.186490 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="221918b0-eb0e-4a32-8eed-07acd18bd31e" containerName="mysql-bootstrap" Oct 05 07:11:25 crc kubenswrapper[4846]: I1005 07:11:25.186496 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="221918b0-eb0e-4a32-8eed-07acd18bd31e" containerName="mysql-bootstrap" Oct 05 07:11:25 crc kubenswrapper[4846]: E1005 07:11:25.186505 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fc35db15-65d4-4cd6-9d09-059822904757" containerName="mariadb-account-delete" Oct 05 07:11:25 crc kubenswrapper[4846]: I1005 07:11:25.186511 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="fc35db15-65d4-4cd6-9d09-059822904757" containerName="mariadb-account-delete" Oct 05 07:11:25 crc kubenswrapper[4846]: E1005 07:11:25.186521 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d11b6df6-6da2-4cb8-a4f9-d87cfd5d4d26" containerName="proxy-httpd" Oct 05 07:11:25 crc kubenswrapper[4846]: I1005 07:11:25.186528 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="d11b6df6-6da2-4cb8-a4f9-d87cfd5d4d26" containerName="proxy-httpd" Oct 05 07:11:25 crc kubenswrapper[4846]: E1005 07:11:25.186536 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="36dca705-2f72-411e-8d80-ed31d49989cc" containerName="cinder-api-log" Oct 05 07:11:25 crc kubenswrapper[4846]: I1005 07:11:25.186574 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="36dca705-2f72-411e-8d80-ed31d49989cc" containerName="cinder-api-log" Oct 05 07:11:25 crc kubenswrapper[4846]: E1005 07:11:25.186584 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="086cf913-a4a2-41ee-910b-a83be95c69d6" containerName="mariadb-account-delete" Oct 05 07:11:25 crc kubenswrapper[4846]: I1005 07:11:25.186590 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="086cf913-a4a2-41ee-910b-a83be95c69d6" containerName="mariadb-account-delete" Oct 05 07:11:25 crc kubenswrapper[4846]: E1005 07:11:25.186602 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bfe96e0c-b9fc-438d-9b17-bec246d48446" containerName="cinder-scheduler" Oct 05 07:11:25 crc kubenswrapper[4846]: I1005 07:11:25.186609 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="bfe96e0c-b9fc-438d-9b17-bec246d48446" containerName="cinder-scheduler" Oct 05 07:11:25 crc kubenswrapper[4846]: E1005 07:11:25.186618 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="87e3744a-b8db-4e76-afa9-0e1129410d4c" containerName="ovsdbserver-sb" Oct 05 07:11:25 crc kubenswrapper[4846]: I1005 07:11:25.186628 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="87e3744a-b8db-4e76-afa9-0e1129410d4c" containerName="ovsdbserver-sb" Oct 05 07:11:25 crc kubenswrapper[4846]: E1005 07:11:25.186665 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bdf95824-fa04-4a83-92cc-7e36b4d00289" containerName="openstack-network-exporter" Oct 05 07:11:25 crc kubenswrapper[4846]: I1005 07:11:25.186673 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="bdf95824-fa04-4a83-92cc-7e36b4d00289" containerName="openstack-network-exporter" Oct 05 07:11:25 crc kubenswrapper[4846]: E1005 07:11:25.186685 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d11b6df6-6da2-4cb8-a4f9-d87cfd5d4d26" containerName="proxy-server" Oct 05 07:11:25 crc kubenswrapper[4846]: I1005 07:11:25.186691 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="d11b6df6-6da2-4cb8-a4f9-d87cfd5d4d26" containerName="proxy-server" Oct 05 07:11:25 crc kubenswrapper[4846]: E1005 07:11:25.186699 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7a5c4c82-4b3f-4314-a32d-db5398ca8a9e" containerName="dnsmasq-dns" Oct 05 07:11:25 crc kubenswrapper[4846]: I1005 07:11:25.186705 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="7a5c4c82-4b3f-4314-a32d-db5398ca8a9e" containerName="dnsmasq-dns" Oct 05 07:11:25 crc kubenswrapper[4846]: E1005 07:11:25.186745 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b7de2524-bf5c-4668-af74-7584edf72c8b" containerName="nova-scheduler-scheduler" Oct 05 07:11:25 crc kubenswrapper[4846]: I1005 07:11:25.186752 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="b7de2524-bf5c-4668-af74-7584edf72c8b" containerName="nova-scheduler-scheduler" Oct 05 07:11:25 crc kubenswrapper[4846]: E1005 07:11:25.186765 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c35affe0-da4e-448f-aac0-7bba26f89040" containerName="placement-api" Oct 05 07:11:25 crc kubenswrapper[4846]: I1005 07:11:25.186771 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="c35affe0-da4e-448f-aac0-7bba26f89040" containerName="placement-api" Oct 05 07:11:25 crc kubenswrapper[4846]: E1005 07:11:25.186781 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2adc8ba0-2810-4bf4-bd24-172b97fb9b22" containerName="glance-httpd" Oct 05 07:11:25 crc kubenswrapper[4846]: I1005 07:11:25.186787 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="2adc8ba0-2810-4bf4-bd24-172b97fb9b22" containerName="glance-httpd" Oct 05 07:11:25 crc kubenswrapper[4846]: E1005 07:11:25.186794 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bdf95824-fa04-4a83-92cc-7e36b4d00289" containerName="ovsdbserver-nb" Oct 05 07:11:25 crc kubenswrapper[4846]: I1005 07:11:25.186827 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="bdf95824-fa04-4a83-92cc-7e36b4d00289" containerName="ovsdbserver-nb" Oct 05 07:11:25 crc kubenswrapper[4846]: E1005 07:11:25.186838 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0c0f8308-455e-4544-9f53-0c06a54c6e22" containerName="glance-httpd" Oct 05 07:11:25 crc kubenswrapper[4846]: I1005 07:11:25.186845 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="0c0f8308-455e-4544-9f53-0c06a54c6e22" containerName="glance-httpd" Oct 05 07:11:25 crc kubenswrapper[4846]: I1005 07:11:25.187123 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="c35affe0-da4e-448f-aac0-7bba26f89040" containerName="placement-log" Oct 05 07:11:25 crc kubenswrapper[4846]: I1005 07:11:25.187138 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="086cf913-a4a2-41ee-910b-a83be95c69d6" containerName="mariadb-account-delete" Oct 05 07:11:25 crc kubenswrapper[4846]: I1005 07:11:25.187174 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="33193805-7aa6-406e-8cad-0ebbb1228ef2" containerName="nova-cell1-conductor-conductor" Oct 05 07:11:25 crc kubenswrapper[4846]: I1005 07:11:25.187191 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="2adc8ba0-2810-4bf4-bd24-172b97fb9b22" containerName="glance-httpd" Oct 05 07:11:25 crc kubenswrapper[4846]: I1005 07:11:25.187203 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="fc35db15-65d4-4cd6-9d09-059822904757" containerName="mariadb-account-delete" Oct 05 07:11:25 crc kubenswrapper[4846]: I1005 07:11:25.187210 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="36dca705-2f72-411e-8d80-ed31d49989cc" containerName="cinder-api-log" Oct 05 07:11:25 crc kubenswrapper[4846]: I1005 07:11:25.187218 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="6e9e53c6-4cf2-4988-89b4-c69471f616aa" containerName="nova-api-api" Oct 05 07:11:25 crc kubenswrapper[4846]: I1005 07:11:25.187226 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="19f3a362-4973-4437-ac49-de71508aa4f2" containerName="openstack-network-exporter" Oct 05 07:11:25 crc kubenswrapper[4846]: I1005 07:11:25.187259 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="bfe96e0c-b9fc-438d-9b17-bec246d48446" containerName="probe" Oct 05 07:11:25 crc kubenswrapper[4846]: I1005 07:11:25.187273 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="2adc8ba0-2810-4bf4-bd24-172b97fb9b22" containerName="glance-log" Oct 05 07:11:25 crc kubenswrapper[4846]: I1005 07:11:25.187283 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="36dca705-2f72-411e-8d80-ed31d49989cc" containerName="cinder-api" Oct 05 07:11:25 crc kubenswrapper[4846]: I1005 07:11:25.187292 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="d11b6df6-6da2-4cb8-a4f9-d87cfd5d4d26" containerName="proxy-httpd" Oct 05 07:11:25 crc kubenswrapper[4846]: I1005 07:11:25.187299 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="b7de2524-bf5c-4668-af74-7584edf72c8b" containerName="nova-scheduler-scheduler" Oct 05 07:11:25 crc kubenswrapper[4846]: I1005 07:11:25.187309 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="bdf95824-fa04-4a83-92cc-7e36b4d00289" containerName="openstack-network-exporter" Oct 05 07:11:25 crc kubenswrapper[4846]: I1005 07:11:25.187357 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="e2f9f9a8-a25f-4a99-b425-4799964d4e24" containerName="mariadb-account-delete" Oct 05 07:11:25 crc kubenswrapper[4846]: I1005 07:11:25.187366 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="bdf95824-fa04-4a83-92cc-7e36b4d00289" containerName="ovsdbserver-nb" Oct 05 07:11:25 crc kubenswrapper[4846]: I1005 07:11:25.187376 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="c31937eb-9836-461b-9230-5e06ce7a99e8" containerName="barbican-api" Oct 05 07:11:25 crc kubenswrapper[4846]: I1005 07:11:25.187383 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="e6e86b1c-87ac-4594-aeba-943947065374" containerName="nova-metadata-metadata" Oct 05 07:11:25 crc kubenswrapper[4846]: I1005 07:11:25.187392 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="bfe96e0c-b9fc-438d-9b17-bec246d48446" containerName="cinder-scheduler" Oct 05 07:11:25 crc kubenswrapper[4846]: I1005 07:11:25.187398 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="6660b9ff-df9c-478e-b852-82cb5fb2c60c" containerName="nova-cell1-novncproxy-novncproxy" Oct 05 07:11:25 crc kubenswrapper[4846]: I1005 07:11:25.187408 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="7a5c4c82-4b3f-4314-a32d-db5398ca8a9e" containerName="dnsmasq-dns" Oct 05 07:11:25 crc kubenswrapper[4846]: I1005 07:11:25.187448 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="e6e86b1c-87ac-4594-aeba-943947065374" containerName="nova-metadata-log" Oct 05 07:11:25 crc kubenswrapper[4846]: I1005 07:11:25.187455 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="221918b0-eb0e-4a32-8eed-07acd18bd31e" containerName="galera" Oct 05 07:11:25 crc kubenswrapper[4846]: I1005 07:11:25.187465 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="d11b6df6-6da2-4cb8-a4f9-d87cfd5d4d26" containerName="proxy-server" Oct 05 07:11:25 crc kubenswrapper[4846]: I1005 07:11:25.187475 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="a32cdd03-1398-4055-b051-62adde35200f" containerName="mariadb-account-delete" Oct 05 07:11:25 crc kubenswrapper[4846]: I1005 07:11:25.187481 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="0c0f8308-455e-4544-9f53-0c06a54c6e22" containerName="glance-httpd" Oct 05 07:11:25 crc kubenswrapper[4846]: I1005 07:11:25.187491 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="87e3744a-b8db-4e76-afa9-0e1129410d4c" containerName="ovsdbserver-sb" Oct 05 07:11:25 crc kubenswrapper[4846]: I1005 07:11:25.187526 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="c35affe0-da4e-448f-aac0-7bba26f89040" containerName="placement-api" Oct 05 07:11:25 crc kubenswrapper[4846]: I1005 07:11:25.187536 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="0c0f8308-455e-4544-9f53-0c06a54c6e22" containerName="glance-log" Oct 05 07:11:25 crc kubenswrapper[4846]: I1005 07:11:25.187544 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="87e3744a-b8db-4e76-afa9-0e1129410d4c" containerName="openstack-network-exporter" Oct 05 07:11:25 crc kubenswrapper[4846]: I1005 07:11:25.187555 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="c31937eb-9836-461b-9230-5e06ce7a99e8" containerName="barbican-api-log" Oct 05 07:11:25 crc kubenswrapper[4846]: I1005 07:11:25.187563 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="6e9e53c6-4cf2-4988-89b4-c69471f616aa" containerName="nova-api-log" Oct 05 07:11:25 crc kubenswrapper[4846]: I1005 07:11:25.188681 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone67b1-account-delete-7lt6j" Oct 05 07:11:25 crc kubenswrapper[4846]: I1005 07:11:25.190453 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6e9e53c6-4cf2-4988-89b4-c69471f616aa-logs" (OuterVolumeSpecName: "logs") pod "6e9e53c6-4cf2-4988-89b4-c69471f616aa" (UID: "6e9e53c6-4cf2-4988-89b4-c69471f616aa"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 07:11:25 crc kubenswrapper[4846]: I1005 07:11:25.197619 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-sync-t7t67"] Oct 05 07:11:25 crc kubenswrapper[4846]: I1005 07:11:25.199698 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6e9e53c6-4cf2-4988-89b4-c69471f616aa-kube-api-access-mc992" (OuterVolumeSpecName: "kube-api-access-mc992") pod "6e9e53c6-4cf2-4988-89b4-c69471f616aa" (UID: "6e9e53c6-4cf2-4988-89b4-c69471f616aa"). InnerVolumeSpecName "kube-api-access-mc992". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:11:25 crc kubenswrapper[4846]: I1005 07:11:25.231974 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c35affe0-da4e-448f-aac0-7bba26f89040-scripts" (OuterVolumeSpecName: "scripts") pod "c35affe0-da4e-448f-aac0-7bba26f89040" (UID: "c35affe0-da4e-448f-aac0-7bba26f89040"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:11:25 crc kubenswrapper[4846]: I1005 07:11:25.239227 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9dmcg\" (UniqueName: \"kubernetes.io/projected/c3e9c65e-42bf-4cd6-a47d-b6ef0db5160e-kube-api-access-9dmcg\") pod \"keystone67b1-account-delete-7lt6j\" (UID: \"c3e9c65e-42bf-4cd6-a47d-b6ef0db5160e\") " pod="openstack/keystone67b1-account-delete-7lt6j" Oct 05 07:11:25 crc kubenswrapper[4846]: I1005 07:11:25.239389 4846 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c31937eb-9836-461b-9230-5e06ce7a99e8-logs\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:25 crc kubenswrapper[4846]: I1005 07:11:25.239402 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mc992\" (UniqueName: \"kubernetes.io/projected/6e9e53c6-4cf2-4988-89b4-c69471f616aa-kube-api-access-mc992\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:25 crc kubenswrapper[4846]: I1005 07:11:25.239413 4846 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6e9e53c6-4cf2-4988-89b4-c69471f616aa-logs\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:25 crc kubenswrapper[4846]: I1005 07:11:25.239421 4846 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c35affe0-da4e-448f-aac0-7bba26f89040-logs\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:25 crc kubenswrapper[4846]: I1005 07:11:25.239430 4846 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c35affe0-da4e-448f-aac0-7bba26f89040-scripts\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:25 crc kubenswrapper[4846]: I1005 07:11:25.242483 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone67b1-account-delete-7lt6j"] Oct 05 07:11:25 crc kubenswrapper[4846]: I1005 07:11:25.253500 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c35affe0-da4e-448f-aac0-7bba26f89040-kube-api-access-2mxsf" (OuterVolumeSpecName: "kube-api-access-2mxsf") pod "c35affe0-da4e-448f-aac0-7bba26f89040" (UID: "c35affe0-da4e-448f-aac0-7bba26f89040"). InnerVolumeSpecName "kube-api-access-2mxsf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:11:25 crc kubenswrapper[4846]: I1005 07:11:25.253973 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c31937eb-9836-461b-9230-5e06ce7a99e8-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "c31937eb-9836-461b-9230-5e06ce7a99e8" (UID: "c31937eb-9836-461b-9230-5e06ce7a99e8"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:11:25 crc kubenswrapper[4846]: I1005 07:11:25.261175 4846 scope.go:117] "RemoveContainer" containerID="ba19323ff93845912587e75b8b10dfd5b4a0513fb4e38a07aab30ad816bdfa3c" Oct 05 07:11:25 crc kubenswrapper[4846]: I1005 07:11:25.262071 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-vt9sx"] Oct 05 07:11:25 crc kubenswrapper[4846]: I1005 07:11:25.280132 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c31937eb-9836-461b-9230-5e06ce7a99e8-kube-api-access-x6fpj" (OuterVolumeSpecName: "kube-api-access-x6fpj") pod "c31937eb-9836-461b-9230-5e06ce7a99e8" (UID: "c31937eb-9836-461b-9230-5e06ce7a99e8"). InnerVolumeSpecName "kube-api-access-x6fpj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:11:25 crc kubenswrapper[4846]: I1005 07:11:25.306591 4846 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-0" podUID="35261dbc-cabe-4bf0-88f6-b8613ddec0f1" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.104:5671: connect: connection refused" Oct 05 07:11:25 crc kubenswrapper[4846]: I1005 07:11:25.322395 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-79f67599fc-2sk76"] Oct 05 07:11:25 crc kubenswrapper[4846]: I1005 07:11:25.322464 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c31937eb-9836-461b-9230-5e06ce7a99e8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c31937eb-9836-461b-9230-5e06ce7a99e8" (UID: "c31937eb-9836-461b-9230-5e06ce7a99e8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:11:25 crc kubenswrapper[4846]: I1005 07:11:25.322764 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/keystone-79f67599fc-2sk76" podUID="1fa90810-c620-48c2-a49c-762341889743" containerName="keystone-api" containerID="cri-o://26da8c60efda49d0ad9e38f18f39b3b3c7a75ffb60bb839d759345619a5d337c" gracePeriod=30 Oct 05 07:11:25 crc kubenswrapper[4846]: I1005 07:11:25.340463 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e6e86b1c-87ac-4594-aeba-943947065374-config-data\") pod \"e6e86b1c-87ac-4594-aeba-943947065374\" (UID: \"e6e86b1c-87ac-4594-aeba-943947065374\") " Oct 05 07:11:25 crc kubenswrapper[4846]: I1005 07:11:25.340798 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/e6e86b1c-87ac-4594-aeba-943947065374-nova-metadata-tls-certs\") pod \"e6e86b1c-87ac-4594-aeba-943947065374\" (UID: \"e6e86b1c-87ac-4594-aeba-943947065374\") " Oct 05 07:11:25 crc kubenswrapper[4846]: I1005 07:11:25.340931 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e6e86b1c-87ac-4594-aeba-943947065374-logs\") pod \"e6e86b1c-87ac-4594-aeba-943947065374\" (UID: \"e6e86b1c-87ac-4594-aeba-943947065374\") " Oct 05 07:11:25 crc kubenswrapper[4846]: I1005 07:11:25.341099 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e6e86b1c-87ac-4594-aeba-943947065374-combined-ca-bundle\") pod \"e6e86b1c-87ac-4594-aeba-943947065374\" (UID: \"e6e86b1c-87ac-4594-aeba-943947065374\") " Oct 05 07:11:25 crc kubenswrapper[4846]: I1005 07:11:25.341436 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbvks\" (UniqueName: \"kubernetes.io/projected/e6e86b1c-87ac-4594-aeba-943947065374-kube-api-access-dbvks\") pod \"e6e86b1c-87ac-4594-aeba-943947065374\" (UID: \"e6e86b1c-87ac-4594-aeba-943947065374\") " Oct 05 07:11:25 crc kubenswrapper[4846]: I1005 07:11:25.341925 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9dmcg\" (UniqueName: \"kubernetes.io/projected/c3e9c65e-42bf-4cd6-a47d-b6ef0db5160e-kube-api-access-9dmcg\") pod \"keystone67b1-account-delete-7lt6j\" (UID: \"c3e9c65e-42bf-4cd6-a47d-b6ef0db5160e\") " pod="openstack/keystone67b1-account-delete-7lt6j" Oct 05 07:11:25 crc kubenswrapper[4846]: I1005 07:11:25.342449 4846 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c31937eb-9836-461b-9230-5e06ce7a99e8-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:25 crc kubenswrapper[4846]: I1005 07:11:25.342544 4846 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c31937eb-9836-461b-9230-5e06ce7a99e8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:25 crc kubenswrapper[4846]: I1005 07:11:25.342634 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2mxsf\" (UniqueName: \"kubernetes.io/projected/c35affe0-da4e-448f-aac0-7bba26f89040-kube-api-access-2mxsf\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:25 crc kubenswrapper[4846]: I1005 07:11:25.342721 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x6fpj\" (UniqueName: \"kubernetes.io/projected/c31937eb-9836-461b-9230-5e06ce7a99e8-kube-api-access-x6fpj\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:25 crc kubenswrapper[4846]: I1005 07:11:25.343870 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e6e86b1c-87ac-4594-aeba-943947065374-logs" (OuterVolumeSpecName: "logs") pod "e6e86b1c-87ac-4594-aeba-943947065374" (UID: "e6e86b1c-87ac-4594-aeba-943947065374"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 07:11:25 crc kubenswrapper[4846]: E1005 07:11:25.352519 4846 projected.go:194] Error preparing data for projected volume kube-api-access-9dmcg for pod openstack/keystone67b1-account-delete-7lt6j: failed to fetch token: serviceaccounts "galera-openstack" not found Oct 05 07:11:25 crc kubenswrapper[4846]: E1005 07:11:25.352608 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/c3e9c65e-42bf-4cd6-a47d-b6ef0db5160e-kube-api-access-9dmcg podName:c3e9c65e-42bf-4cd6-a47d-b6ef0db5160e nodeName:}" failed. No retries permitted until 2025-10-05 07:11:25.852581011 +0000 UTC m=+1408.093433786 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-9dmcg" (UniqueName: "kubernetes.io/projected/c3e9c65e-42bf-4cd6-a47d-b6ef0db5160e-kube-api-access-9dmcg") pod "keystone67b1-account-delete-7lt6j" (UID: "c3e9c65e-42bf-4cd6-a47d-b6ef0db5160e") : failed to fetch token: serviceaccounts "galera-openstack" not found Oct 05 07:11:25 crc kubenswrapper[4846]: I1005 07:11:25.357048 4846 scope.go:117] "RemoveContainer" containerID="470aec4a8b7085d2d3371c5366c57378850aca957d883939f98653f186109a67" Oct 05 07:11:25 crc kubenswrapper[4846]: I1005 07:11:25.369098 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e6e86b1c-87ac-4594-aeba-943947065374-kube-api-access-dbvks" (OuterVolumeSpecName: "kube-api-access-dbvks") pod "e6e86b1c-87ac-4594-aeba-943947065374" (UID: "e6e86b1c-87ac-4594-aeba-943947065374"). InnerVolumeSpecName "kube-api-access-dbvks". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:11:25 crc kubenswrapper[4846]: I1005 07:11:25.381700 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6e9e53c6-4cf2-4988-89b4-c69471f616aa-config-data" (OuterVolumeSpecName: "config-data") pod "6e9e53c6-4cf2-4988-89b4-c69471f616aa" (UID: "6e9e53c6-4cf2-4988-89b4-c69471f616aa"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:11:25 crc kubenswrapper[4846]: I1005 07:11:25.393273 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"2adc8ba0-2810-4bf4-bd24-172b97fb9b22","Type":"ContainerDied","Data":"9ae6aaa0c860e07f083871e9cd161f5f2319bfe8b6db28f062aac92ee397f992"} Oct 05 07:11:25 crc kubenswrapper[4846]: I1005 07:11:25.393394 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 05 07:11:25 crc kubenswrapper[4846]: I1005 07:11:25.395711 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-67b1-account-create-lzlh4"] Oct 05 07:11:25 crc kubenswrapper[4846]: I1005 07:11:25.404926 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 05 07:11:25 crc kubenswrapper[4846]: I1005 07:11:25.406133 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"e6e86b1c-87ac-4594-aeba-943947065374","Type":"ContainerDied","Data":"648dd0ba4f64806143815b8d201f6f6ff3ed3d36886724d53eedbf8ae1fc835e"} Oct 05 07:11:25 crc kubenswrapper[4846]: I1005 07:11:25.408724 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-vt9sx"] Oct 05 07:11:25 crc kubenswrapper[4846]: I1005 07:11:25.429293 4846 generic.go:334] "Generic (PLEG): container finished" podID="b1e394dd-f55b-4af1-af0b-52cc7d09afa1" containerID="e92d479cbafa7786e2514b2d1e99c120ef33197a849bb0c13e2362354bcebd01" exitCode=0 Oct 05 07:11:25 crc kubenswrapper[4846]: I1005 07:11:25.429330 4846 generic.go:334] "Generic (PLEG): container finished" podID="b1e394dd-f55b-4af1-af0b-52cc7d09afa1" containerID="6db54d2ca22110333154adb6a23a0ec27e54e4f7118d44ea9636189fffd9e0ae" exitCode=2 Oct 05 07:11:25 crc kubenswrapper[4846]: I1005 07:11:25.429630 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b1e394dd-f55b-4af1-af0b-52cc7d09afa1","Type":"ContainerDied","Data":"e92d479cbafa7786e2514b2d1e99c120ef33197a849bb0c13e2362354bcebd01"} Oct 05 07:11:25 crc kubenswrapper[4846]: I1005 07:11:25.429764 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b1e394dd-f55b-4af1-af0b-52cc7d09afa1","Type":"ContainerDied","Data":"6db54d2ca22110333154adb6a23a0ec27e54e4f7118d44ea9636189fffd9e0ae"} Oct 05 07:11:25 crc kubenswrapper[4846]: I1005 07:11:25.444038 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6e9e53c6-4cf2-4988-89b4-c69471f616aa-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "6e9e53c6-4cf2-4988-89b4-c69471f616aa" (UID: "6e9e53c6-4cf2-4988-89b4-c69471f616aa"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:11:25 crc kubenswrapper[4846]: I1005 07:11:25.449769 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cindercf3d-account-delete-jw5tr" event={"ID":"a32cdd03-1398-4055-b051-62adde35200f","Type":"ContainerDied","Data":"63fc4076208dfa8214e57a8f6a726a5fbe145fe441f981f1f2d4a2f7ce1f81bc"} Oct 05 07:11:25 crc kubenswrapper[4846]: I1005 07:11:25.450883 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cindercf3d-account-delete-jw5tr" Oct 05 07:11:25 crc kubenswrapper[4846]: I1005 07:11:25.456804 4846 generic.go:334] "Generic (PLEG): container finished" podID="57b09608-08e5-4998-a3b3-e88b13587f58" containerID="0f36596d6da69b2a28102d8a73a67b3c631900ebfdb71663f52974e9dd99b1d1" exitCode=2 Oct 05 07:11:25 crc kubenswrapper[4846]: I1005 07:11:25.457015 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"57b09608-08e5-4998-a3b3-e88b13587f58","Type":"ContainerDied","Data":"0f36596d6da69b2a28102d8a73a67b3c631900ebfdb71663f52974e9dd99b1d1"} Oct 05 07:11:25 crc kubenswrapper[4846]: I1005 07:11:25.464463 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-984785977-srp2v" event={"ID":"d11b6df6-6da2-4cb8-a4f9-d87cfd5d4d26","Type":"ContainerDied","Data":"ee3e69e0f367c67afe20a2d1f8600f785c5579c1f001d1de557eba1d560758fa"} Oct 05 07:11:25 crc kubenswrapper[4846]: I1005 07:11:25.464622 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-984785977-srp2v" Oct 05 07:11:25 crc kubenswrapper[4846]: I1005 07:11:25.470357 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican8fc5-account-delete-gjp47" Oct 05 07:11:25 crc kubenswrapper[4846]: I1005 07:11:25.470474 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstack-galera-0"] Oct 05 07:11:25 crc kubenswrapper[4846]: I1005 07:11:25.470513 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican8fc5-account-delete-gjp47" event={"ID":"086cf913-a4a2-41ee-910b-a83be95c69d6","Type":"ContainerDied","Data":"525ee924ca58c84f2ca050e2fe5ed97086aca63ac6a15850a1eb3d12e6423876"} Oct 05 07:11:25 crc kubenswrapper[4846]: I1005 07:11:25.482793 4846 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6e9e53c6-4cf2-4988-89b4-c69471f616aa-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:25 crc kubenswrapper[4846]: I1005 07:11:25.485212 4846 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e6e86b1c-87ac-4594-aeba-943947065374-logs\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:25 crc kubenswrapper[4846]: I1005 07:11:25.485523 4846 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6e9e53c6-4cf2-4988-89b4-c69471f616aa-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:25 crc kubenswrapper[4846]: I1005 07:11:25.485824 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbvks\" (UniqueName: \"kubernetes.io/projected/e6e86b1c-87ac-4594-aeba-943947065374-kube-api-access-dbvks\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:25 crc kubenswrapper[4846]: I1005 07:11:25.503822 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone67b1-account-delete-7lt6j"] Oct 05 07:11:25 crc kubenswrapper[4846]: I1005 07:11:25.507458 4846 scope.go:117] "RemoveContainer" containerID="1921e872cb6fabf10cd959960895a03e62a0d256a8d1c87f18ab5389a991e8f4" Oct 05 07:11:25 crc kubenswrapper[4846]: I1005 07:11:25.510096 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-57674d6c54-zn5nq" event={"ID":"c31937eb-9836-461b-9230-5e06ce7a99e8","Type":"ContainerDied","Data":"8f6e1c6b60e3d1a211e92f9485d583bd98fbc2e6833ffd3020cdfa2f01f8a1da"} Oct 05 07:11:25 crc kubenswrapper[4846]: I1005 07:11:25.510285 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-57674d6c54-zn5nq" Oct 05 07:11:25 crc kubenswrapper[4846]: I1005 07:11:25.512521 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c35affe0-da4e-448f-aac0-7bba26f89040-config-data" (OuterVolumeSpecName: "config-data") pod "c35affe0-da4e-448f-aac0-7bba26f89040" (UID: "c35affe0-da4e-448f-aac0-7bba26f89040"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:11:25 crc kubenswrapper[4846]: E1005 07:11:25.518969 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[kube-api-access-9dmcg], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openstack/keystone67b1-account-delete-7lt6j" podUID="c3e9c65e-42bf-4cd6-a47d-b6ef0db5160e" Oct 05 07:11:25 crc kubenswrapper[4846]: I1005 07:11:25.524308 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-65884db4f4-fsk6p" event={"ID":"c35affe0-da4e-448f-aac0-7bba26f89040","Type":"ContainerDied","Data":"ca8f38824786960e56cd2679a8818d0dc234675708e9dcdb847dd2e8bbbdc666"} Oct 05 07:11:25 crc kubenswrapper[4846]: I1005 07:11:25.524558 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-65884db4f4-fsk6p" Oct 05 07:11:25 crc kubenswrapper[4846]: I1005 07:11:25.532638 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6e9e53c6-4cf2-4988-89b4-c69471f616aa-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6e9e53c6-4cf2-4988-89b4-c69471f616aa" (UID: "6e9e53c6-4cf2-4988-89b4-c69471f616aa"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:11:25 crc kubenswrapper[4846]: I1005 07:11:25.539023 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"36dca705-2f72-411e-8d80-ed31d49989cc","Type":"ContainerDied","Data":"afaf1e34eaf63fcbc924b759258cd9887dcac6aeb43b21132100b45d8bbe4ea4"} Oct 05 07:11:25 crc kubenswrapper[4846]: I1005 07:11:25.539151 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 05 07:11:25 crc kubenswrapper[4846]: I1005 07:11:25.550333 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"6e9e53c6-4cf2-4988-89b4-c69471f616aa","Type":"ContainerDied","Data":"1ccdf63d0267e37d0a39dfc09521b36fac3246c3415087fde77f78ac72e311ae"} Oct 05 07:11:25 crc kubenswrapper[4846]: I1005 07:11:25.550469 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 05 07:11:25 crc kubenswrapper[4846]: I1005 07:11:25.571207 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e6e86b1c-87ac-4594-aeba-943947065374-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e6e86b1c-87ac-4594-aeba-943947065374" (UID: "e6e86b1c-87ac-4594-aeba-943947065374"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:11:25 crc kubenswrapper[4846]: I1005 07:11:25.584761 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-67b1-account-create-lzlh4"] Oct 05 07:11:25 crc kubenswrapper[4846]: I1005 07:11:25.587604 4846 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6e9e53c6-4cf2-4988-89b4-c69471f616aa-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:25 crc kubenswrapper[4846]: I1005 07:11:25.587637 4846 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c35affe0-da4e-448f-aac0-7bba26f89040-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:25 crc kubenswrapper[4846]: I1005 07:11:25.587652 4846 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e6e86b1c-87ac-4594-aeba-943947065374-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:25 crc kubenswrapper[4846]: I1005 07:11:25.603436 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glancebc5f-account-delete-hkzcl" Oct 05 07:11:25 crc kubenswrapper[4846]: I1005 07:11:25.605116 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-create-z9b4l"] Oct 05 07:11:25 crc kubenswrapper[4846]: I1005 07:11:25.613311 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-create-z9b4l"] Oct 05 07:11:25 crc kubenswrapper[4846]: I1005 07:11:25.641996 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 05 07:11:25 crc kubenswrapper[4846]: I1005 07:11:25.646435 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e6e86b1c-87ac-4594-aeba-943947065374-config-data" (OuterVolumeSpecName: "config-data") pod "e6e86b1c-87ac-4594-aeba-943947065374" (UID: "e6e86b1c-87ac-4594-aeba-943947065374"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:11:25 crc kubenswrapper[4846]: I1005 07:11:25.648756 4846 scope.go:117] "RemoveContainer" containerID="9e54d631fb403eb9146c131948244ad644f213facfb43b90944167c8827ee59f" Oct 05 07:11:25 crc kubenswrapper[4846]: I1005 07:11:25.649147 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6e9e53c6-4cf2-4988-89b4-c69471f616aa-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "6e9e53c6-4cf2-4988-89b4-c69471f616aa" (UID: "6e9e53c6-4cf2-4988-89b4-c69471f616aa"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:11:25 crc kubenswrapper[4846]: I1005 07:11:25.649548 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 05 07:11:25 crc kubenswrapper[4846]: I1005 07:11:25.662208 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c35affe0-da4e-448f-aac0-7bba26f89040-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c35affe0-da4e-448f-aac0-7bba26f89040" (UID: "c35affe0-da4e-448f-aac0-7bba26f89040"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:11:25 crc kubenswrapper[4846]: I1005 07:11:25.662796 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican8fc5-account-delete-gjp47"] Oct 05 07:11:25 crc kubenswrapper[4846]: I1005 07:11:25.667770 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c31937eb-9836-461b-9230-5e06ce7a99e8-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "c31937eb-9836-461b-9230-5e06ce7a99e8" (UID: "c31937eb-9836-461b-9230-5e06ce7a99e8"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:11:25 crc kubenswrapper[4846]: I1005 07:11:25.668896 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican8fc5-account-delete-gjp47"] Oct 05 07:11:25 crc kubenswrapper[4846]: I1005 07:11:25.676847 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cindercf3d-account-delete-jw5tr"] Oct 05 07:11:25 crc kubenswrapper[4846]: I1005 07:11:25.690959 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cindercf3d-account-delete-jw5tr"] Oct 05 07:11:25 crc kubenswrapper[4846]: I1005 07:11:25.692241 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Oct 05 07:11:25 crc kubenswrapper[4846]: I1005 07:11:25.693322 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sszh5\" (UniqueName: \"kubernetes.io/projected/4d3616b4-9f2c-4c62-9760-0300b02d0dbb-kube-api-access-sszh5\") pod \"4d3616b4-9f2c-4c62-9760-0300b02d0dbb\" (UID: \"4d3616b4-9f2c-4c62-9760-0300b02d0dbb\") " Oct 05 07:11:25 crc kubenswrapper[4846]: I1005 07:11:25.693901 4846 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c35affe0-da4e-448f-aac0-7bba26f89040-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:25 crc kubenswrapper[4846]: I1005 07:11:25.693918 4846 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c31937eb-9836-461b-9230-5e06ce7a99e8-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:25 crc kubenswrapper[4846]: I1005 07:11:25.693929 4846 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e6e86b1c-87ac-4594-aeba-943947065374-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:25 crc kubenswrapper[4846]: I1005 07:11:25.693938 4846 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6e9e53c6-4cf2-4988-89b4-c69471f616aa-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:25 crc kubenswrapper[4846]: I1005 07:11:25.700567 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4d3616b4-9f2c-4c62-9760-0300b02d0dbb-kube-api-access-sszh5" (OuterVolumeSpecName: "kube-api-access-sszh5") pod "4d3616b4-9f2c-4c62-9760-0300b02d0dbb" (UID: "4d3616b4-9f2c-4c62-9760-0300b02d0dbb"). InnerVolumeSpecName "kube-api-access-sszh5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:11:25 crc kubenswrapper[4846]: I1005 07:11:25.711760 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c31937eb-9836-461b-9230-5e06ce7a99e8-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "c31937eb-9836-461b-9230-5e06ce7a99e8" (UID: "c31937eb-9836-461b-9230-5e06ce7a99e8"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:11:25 crc kubenswrapper[4846]: I1005 07:11:25.713790 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Oct 05 07:11:25 crc kubenswrapper[4846]: I1005 07:11:25.718710 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e6e86b1c-87ac-4594-aeba-943947065374-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "e6e86b1c-87ac-4594-aeba-943947065374" (UID: "e6e86b1c-87ac-4594-aeba-943947065374"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:11:25 crc kubenswrapper[4846]: I1005 07:11:25.732069 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c35affe0-da4e-448f-aac0-7bba26f89040-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "c35affe0-da4e-448f-aac0-7bba26f89040" (UID: "c35affe0-da4e-448f-aac0-7bba26f89040"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:11:25 crc kubenswrapper[4846]: I1005 07:11:25.753741 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/openstack-galera-0" podUID="d88c48f9-948e-4888-986e-e00163ec93a6" containerName="galera" containerID="cri-o://56b8f764086c61683708e28e3f2929710527aada412eda06ddb4b14e487478c3" gracePeriod=30 Oct 05 07:11:25 crc kubenswrapper[4846]: E1005 07:11:25.776321 4846 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod36dca705_2f72_411e_8d80_ed31d49989cc.slice/crio-afaf1e34eaf63fcbc924b759258cd9887dcac6aeb43b21132100b45d8bbe4ea4\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2adc8ba0_2810_4bf4_bd24_172b97fb9b22.slice/crio-9ae6aaa0c860e07f083871e9cd161f5f2319bfe8b6db28f062aac92ee397f992\": RecentStats: unable to find data in memory cache]" Oct 05 07:11:25 crc kubenswrapper[4846]: I1005 07:11:25.791573 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c35affe0-da4e-448f-aac0-7bba26f89040-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "c35affe0-da4e-448f-aac0-7bba26f89040" (UID: "c35affe0-da4e-448f-aac0-7bba26f89040"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:11:25 crc kubenswrapper[4846]: I1005 07:11:25.804264 4846 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c31937eb-9836-461b-9230-5e06ce7a99e8-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:25 crc kubenswrapper[4846]: I1005 07:11:25.804297 4846 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c35affe0-da4e-448f-aac0-7bba26f89040-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:25 crc kubenswrapper[4846]: I1005 07:11:25.804306 4846 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c35affe0-da4e-448f-aac0-7bba26f89040-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:25 crc kubenswrapper[4846]: I1005 07:11:25.804318 4846 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/e6e86b1c-87ac-4594-aeba-943947065374-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:25 crc kubenswrapper[4846]: I1005 07:11:25.804329 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sszh5\" (UniqueName: \"kubernetes.io/projected/4d3616b4-9f2c-4c62-9760-0300b02d0dbb-kube-api-access-sszh5\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:25 crc kubenswrapper[4846]: I1005 07:11:25.816281 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c31937eb-9836-461b-9230-5e06ce7a99e8-config-data" (OuterVolumeSpecName: "config-data") pod "c31937eb-9836-461b-9230-5e06ce7a99e8" (UID: "c31937eb-9836-461b-9230-5e06ce7a99e8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:11:25 crc kubenswrapper[4846]: I1005 07:11:25.905888 4846 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-cell1-server-0" podUID="626e6bce-7422-4fbd-b978-dfa5cb93de34" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.105:5671: connect: connection refused" Oct 05 07:11:25 crc kubenswrapper[4846]: I1005 07:11:25.907395 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9dmcg\" (UniqueName: \"kubernetes.io/projected/c3e9c65e-42bf-4cd6-a47d-b6ef0db5160e-kube-api-access-9dmcg\") pod \"keystone67b1-account-delete-7lt6j\" (UID: \"c3e9c65e-42bf-4cd6-a47d-b6ef0db5160e\") " pod="openstack/keystone67b1-account-delete-7lt6j" Oct 05 07:11:25 crc kubenswrapper[4846]: I1005 07:11:25.907537 4846 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c31937eb-9836-461b-9230-5e06ce7a99e8-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:25 crc kubenswrapper[4846]: E1005 07:11:25.910767 4846 projected.go:194] Error preparing data for projected volume kube-api-access-9dmcg for pod openstack/keystone67b1-account-delete-7lt6j: failed to fetch token: serviceaccounts "galera-openstack" not found Oct 05 07:11:25 crc kubenswrapper[4846]: E1005 07:11:25.910852 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/c3e9c65e-42bf-4cd6-a47d-b6ef0db5160e-kube-api-access-9dmcg podName:c3e9c65e-42bf-4cd6-a47d-b6ef0db5160e nodeName:}" failed. No retries permitted until 2025-10-05 07:11:26.910830321 +0000 UTC m=+1409.151683096 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-9dmcg" (UniqueName: "kubernetes.io/projected/c3e9c65e-42bf-4cd6-a47d-b6ef0db5160e-kube-api-access-9dmcg") pod "keystone67b1-account-delete-7lt6j" (UID: "c3e9c65e-42bf-4cd6-a47d-b6ef0db5160e") : failed to fetch token: serviceaccounts "galera-openstack" not found Oct 05 07:11:25 crc kubenswrapper[4846]: I1005 07:11:25.957412 4846 scope.go:117] "RemoveContainer" containerID="3be1afa599376331b916c6589c79dd8bdcf3fefa4275b988917d64c058731c7a" Oct 05 07:11:25 crc kubenswrapper[4846]: I1005 07:11:25.976604 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 05 07:11:26 crc kubenswrapper[4846]: I1005 07:11:26.005912 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-65884db4f4-fsk6p"] Oct 05 07:11:26 crc kubenswrapper[4846]: I1005 07:11:26.026036 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-65884db4f4-fsk6p"] Oct 05 07:11:26 crc kubenswrapper[4846]: I1005 07:11:26.028076 4846 scope.go:117] "RemoveContainer" containerID="7408c3d57efa73ce926d92ff91954d484124d71bc39b46bd428ba78c4ef03075" Oct 05 07:11:26 crc kubenswrapper[4846]: I1005 07:11:26.046853 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-proxy-984785977-srp2v"] Oct 05 07:11:26 crc kubenswrapper[4846]: I1005 07:11:26.065831 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/swift-proxy-984785977-srp2v"] Oct 05 07:11:26 crc kubenswrapper[4846]: I1005 07:11:26.074404 4846 scope.go:117] "RemoveContainer" containerID="71ab8b0934bfb90a07c9012633af71114aed35423acecf39d3a2e32cd390b303" Oct 05 07:11:26 crc kubenswrapper[4846]: I1005 07:11:26.104492 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 05 07:11:26 crc kubenswrapper[4846]: I1005 07:11:26.111132 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/57b09608-08e5-4998-a3b3-e88b13587f58-kube-state-metrics-tls-certs\") pod \"57b09608-08e5-4998-a3b3-e88b13587f58\" (UID: \"57b09608-08e5-4998-a3b3-e88b13587f58\") " Oct 05 07:11:26 crc kubenswrapper[4846]: I1005 07:11:26.111383 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/57b09608-08e5-4998-a3b3-e88b13587f58-combined-ca-bundle\") pod \"57b09608-08e5-4998-a3b3-e88b13587f58\" (UID: \"57b09608-08e5-4998-a3b3-e88b13587f58\") " Oct 05 07:11:26 crc kubenswrapper[4846]: I1005 07:11:26.111484 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5lbbc\" (UniqueName: \"kubernetes.io/projected/57b09608-08e5-4998-a3b3-e88b13587f58-kube-api-access-5lbbc\") pod \"57b09608-08e5-4998-a3b3-e88b13587f58\" (UID: \"57b09608-08e5-4998-a3b3-e88b13587f58\") " Oct 05 07:11:26 crc kubenswrapper[4846]: I1005 07:11:26.111507 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/57b09608-08e5-4998-a3b3-e88b13587f58-kube-state-metrics-tls-config\") pod \"57b09608-08e5-4998-a3b3-e88b13587f58\" (UID: \"57b09608-08e5-4998-a3b3-e88b13587f58\") " Oct 05 07:11:26 crc kubenswrapper[4846]: I1005 07:11:26.116813 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57b09608-08e5-4998-a3b3-e88b13587f58-kube-api-access-5lbbc" (OuterVolumeSpecName: "kube-api-access-5lbbc") pod "57b09608-08e5-4998-a3b3-e88b13587f58" (UID: "57b09608-08e5-4998-a3b3-e88b13587f58"). InnerVolumeSpecName "kube-api-access-5lbbc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:11:26 crc kubenswrapper[4846]: I1005 07:11:26.138437 4846 scope.go:117] "RemoveContainer" containerID="6ef0fb95fbc564da5b7b63871e6cea78667f2f37a2822564371b3f419ac22f47" Oct 05 07:11:26 crc kubenswrapper[4846]: I1005 07:11:26.151932 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/57b09608-08e5-4998-a3b3-e88b13587f58-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "57b09608-08e5-4998-a3b3-e88b13587f58" (UID: "57b09608-08e5-4998-a3b3-e88b13587f58"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:11:26 crc kubenswrapper[4846]: I1005 07:11:26.167275 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Oct 05 07:11:26 crc kubenswrapper[4846]: I1005 07:11:26.175461 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/57b09608-08e5-4998-a3b3-e88b13587f58-kube-state-metrics-tls-config" (OuterVolumeSpecName: "kube-state-metrics-tls-config") pod "57b09608-08e5-4998-a3b3-e88b13587f58" (UID: "57b09608-08e5-4998-a3b3-e88b13587f58"). InnerVolumeSpecName "kube-state-metrics-tls-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:11:26 crc kubenswrapper[4846]: I1005 07:11:26.181261 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 05 07:11:26 crc kubenswrapper[4846]: I1005 07:11:26.190472 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Oct 05 07:11:26 crc kubenswrapper[4846]: I1005 07:11:26.196691 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-57674d6c54-zn5nq"] Oct 05 07:11:26 crc kubenswrapper[4846]: I1005 07:11:26.204459 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-57674d6c54-zn5nq"] Oct 05 07:11:26 crc kubenswrapper[4846]: I1005 07:11:26.207330 4846 scope.go:117] "RemoveContainer" containerID="5b9cd4a1e7ff57757729dbae754fc2fd8f2c6d6deb6c33079a5fb777d6c337c4" Oct 05 07:11:26 crc kubenswrapper[4846]: I1005 07:11:26.212456 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/57b09608-08e5-4998-a3b3-e88b13587f58-kube-state-metrics-tls-certs" (OuterVolumeSpecName: "kube-state-metrics-tls-certs") pod "57b09608-08e5-4998-a3b3-e88b13587f58" (UID: "57b09608-08e5-4998-a3b3-e88b13587f58"). InnerVolumeSpecName "kube-state-metrics-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:11:26 crc kubenswrapper[4846]: I1005 07:11:26.214764 4846 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/57b09608-08e5-4998-a3b3-e88b13587f58-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:26 crc kubenswrapper[4846]: I1005 07:11:26.214801 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5lbbc\" (UniqueName: \"kubernetes.io/projected/57b09608-08e5-4998-a3b3-e88b13587f58-kube-api-access-5lbbc\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:26 crc kubenswrapper[4846]: I1005 07:11:26.214822 4846 reconciler_common.go:293] "Volume detached for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/57b09608-08e5-4998-a3b3-e88b13587f58-kube-state-metrics-tls-config\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:26 crc kubenswrapper[4846]: I1005 07:11:26.214837 4846 reconciler_common.go:293] "Volume detached for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/57b09608-08e5-4998-a3b3-e88b13587f58-kube-state-metrics-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:26 crc kubenswrapper[4846]: E1005 07:11:26.215892 4846 configmap.go:193] Couldn't get configMap openstack/rabbitmq-config-data: configmap "rabbitmq-config-data" not found Oct 05 07:11:26 crc kubenswrapper[4846]: E1005 07:11:26.215982 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/35261dbc-cabe-4bf0-88f6-b8613ddec0f1-config-data podName:35261dbc-cabe-4bf0-88f6-b8613ddec0f1 nodeName:}" failed. No retries permitted until 2025-10-05 07:11:34.215941795 +0000 UTC m=+1416.456794570 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/35261dbc-cabe-4bf0-88f6-b8613ddec0f1-config-data") pod "rabbitmq-server-0" (UID: "35261dbc-cabe-4bf0-88f6-b8613ddec0f1") : configmap "rabbitmq-config-data" not found Oct 05 07:11:26 crc kubenswrapper[4846]: I1005 07:11:26.246813 4846 scope.go:117] "RemoveContainer" containerID="eb3d496875131b7ff843f08ff12b7fdc832265502da5423e00550c759c4bc1b1" Oct 05 07:11:26 crc kubenswrapper[4846]: I1005 07:11:26.275323 4846 scope.go:117] "RemoveContainer" containerID="18de9a748414f93e117e641968d42b9db1d5720a2c5bf9d8643daeb9b2524940" Oct 05 07:11:26 crc kubenswrapper[4846]: I1005 07:11:26.306142 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Oct 05 07:11:26 crc kubenswrapper[4846]: I1005 07:11:26.322004 4846 scope.go:117] "RemoveContainer" containerID="62af63220e4d608e4b981ffe4309575edbb012337dc1e84d368d62d0b4cab633" Oct 05 07:11:26 crc kubenswrapper[4846]: I1005 07:11:26.366543 4846 scope.go:117] "RemoveContainer" containerID="85a65281831348ab23593ec8a61436c9871720d821d057e65d1423a5e8c5e79c" Oct 05 07:11:26 crc kubenswrapper[4846]: I1005 07:11:26.394483 4846 scope.go:117] "RemoveContainer" containerID="b48698aea06e5e06f2a72f7568ee41c7c5c3fee9c4bdab0f9cb7cebb4f50ea03" Oct 05 07:11:26 crc kubenswrapper[4846]: I1005 07:11:26.418493 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/751734cb-38c8-46a0-9bac-05bcf24cc589-memcached-tls-certs\") pod \"751734cb-38c8-46a0-9bac-05bcf24cc589\" (UID: \"751734cb-38c8-46a0-9bac-05bcf24cc589\") " Oct 05 07:11:26 crc kubenswrapper[4846]: I1005 07:11:26.418599 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/751734cb-38c8-46a0-9bac-05bcf24cc589-combined-ca-bundle\") pod \"751734cb-38c8-46a0-9bac-05bcf24cc589\" (UID: \"751734cb-38c8-46a0-9bac-05bcf24cc589\") " Oct 05 07:11:26 crc kubenswrapper[4846]: I1005 07:11:26.418667 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/751734cb-38c8-46a0-9bac-05bcf24cc589-kolla-config\") pod \"751734cb-38c8-46a0-9bac-05bcf24cc589\" (UID: \"751734cb-38c8-46a0-9bac-05bcf24cc589\") " Oct 05 07:11:26 crc kubenswrapper[4846]: I1005 07:11:26.418756 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b8cz2\" (UniqueName: \"kubernetes.io/projected/751734cb-38c8-46a0-9bac-05bcf24cc589-kube-api-access-b8cz2\") pod \"751734cb-38c8-46a0-9bac-05bcf24cc589\" (UID: \"751734cb-38c8-46a0-9bac-05bcf24cc589\") " Oct 05 07:11:26 crc kubenswrapper[4846]: I1005 07:11:26.418798 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/751734cb-38c8-46a0-9bac-05bcf24cc589-config-data\") pod \"751734cb-38c8-46a0-9bac-05bcf24cc589\" (UID: \"751734cb-38c8-46a0-9bac-05bcf24cc589\") " Oct 05 07:11:26 crc kubenswrapper[4846]: I1005 07:11:26.420822 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/751734cb-38c8-46a0-9bac-05bcf24cc589-config-data" (OuterVolumeSpecName: "config-data") pod "751734cb-38c8-46a0-9bac-05bcf24cc589" (UID: "751734cb-38c8-46a0-9bac-05bcf24cc589"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 07:11:26 crc kubenswrapper[4846]: I1005 07:11:26.421253 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/751734cb-38c8-46a0-9bac-05bcf24cc589-kolla-config" (OuterVolumeSpecName: "kolla-config") pod "751734cb-38c8-46a0-9bac-05bcf24cc589" (UID: "751734cb-38c8-46a0-9bac-05bcf24cc589"). InnerVolumeSpecName "kolla-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 07:11:26 crc kubenswrapper[4846]: I1005 07:11:26.423085 4846 scope.go:117] "RemoveContainer" containerID="d0a2d9595d1231b4e28c0eff4482c56b148f02336c7fb50f8a4400e74af973ff" Oct 05 07:11:26 crc kubenswrapper[4846]: I1005 07:11:26.425716 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/751734cb-38c8-46a0-9bac-05bcf24cc589-kube-api-access-b8cz2" (OuterVolumeSpecName: "kube-api-access-b8cz2") pod "751734cb-38c8-46a0-9bac-05bcf24cc589" (UID: "751734cb-38c8-46a0-9bac-05bcf24cc589"). InnerVolumeSpecName "kube-api-access-b8cz2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:11:26 crc kubenswrapper[4846]: I1005 07:11:26.449433 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/751734cb-38c8-46a0-9bac-05bcf24cc589-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "751734cb-38c8-46a0-9bac-05bcf24cc589" (UID: "751734cb-38c8-46a0-9bac-05bcf24cc589"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:11:26 crc kubenswrapper[4846]: I1005 07:11:26.449840 4846 scope.go:117] "RemoveContainer" containerID="2dae962829e3caf9333b75e7f59bede4cf4d679ed6def4e5eddee1a34c169b39" Oct 05 07:11:26 crc kubenswrapper[4846]: I1005 07:11:26.481453 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/751734cb-38c8-46a0-9bac-05bcf24cc589-memcached-tls-certs" (OuterVolumeSpecName: "memcached-tls-certs") pod "751734cb-38c8-46a0-9bac-05bcf24cc589" (UID: "751734cb-38c8-46a0-9bac-05bcf24cc589"). InnerVolumeSpecName "memcached-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:11:26 crc kubenswrapper[4846]: I1005 07:11:26.485897 4846 scope.go:117] "RemoveContainer" containerID="735ae29bd9ab70b9f2083caa08160bc161134c908e991ed9c2afaf1f25347aa0" Oct 05 07:11:26 crc kubenswrapper[4846]: I1005 07:11:26.520700 4846 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/751734cb-38c8-46a0-9bac-05bcf24cc589-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:26 crc kubenswrapper[4846]: I1005 07:11:26.520742 4846 reconciler_common.go:293] "Volume detached for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/751734cb-38c8-46a0-9bac-05bcf24cc589-kolla-config\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:26 crc kubenswrapper[4846]: I1005 07:11:26.520755 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b8cz2\" (UniqueName: \"kubernetes.io/projected/751734cb-38c8-46a0-9bac-05bcf24cc589-kube-api-access-b8cz2\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:26 crc kubenswrapper[4846]: I1005 07:11:26.520765 4846 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/751734cb-38c8-46a0-9bac-05bcf24cc589-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:26 crc kubenswrapper[4846]: I1005 07:11:26.520774 4846 reconciler_common.go:293] "Volume detached for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/751734cb-38c8-46a0-9bac-05bcf24cc589-memcached-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:26 crc kubenswrapper[4846]: I1005 07:11:26.525875 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="086cf913-a4a2-41ee-910b-a83be95c69d6" path="/var/lib/kubelet/pods/086cf913-a4a2-41ee-910b-a83be95c69d6/volumes" Oct 05 07:11:26 crc kubenswrapper[4846]: I1005 07:11:26.526623 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0c0f8308-455e-4544-9f53-0c06a54c6e22" path="/var/lib/kubelet/pods/0c0f8308-455e-4544-9f53-0c06a54c6e22/volumes" Oct 05 07:11:26 crc kubenswrapper[4846]: I1005 07:11:26.527667 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2adc8ba0-2810-4bf4-bd24-172b97fb9b22" path="/var/lib/kubelet/pods/2adc8ba0-2810-4bf4-bd24-172b97fb9b22/volumes" Oct 05 07:11:26 crc kubenswrapper[4846]: I1005 07:11:26.529037 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="36dca705-2f72-411e-8d80-ed31d49989cc" path="/var/lib/kubelet/pods/36dca705-2f72-411e-8d80-ed31d49989cc/volumes" Oct 05 07:11:26 crc kubenswrapper[4846]: I1005 07:11:26.529700 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6660b9ff-df9c-478e-b852-82cb5fb2c60c" path="/var/lib/kubelet/pods/6660b9ff-df9c-478e-b852-82cb5fb2c60c/volumes" Oct 05 07:11:26 crc kubenswrapper[4846]: I1005 07:11:26.530340 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6e9e53c6-4cf2-4988-89b4-c69471f616aa" path="/var/lib/kubelet/pods/6e9e53c6-4cf2-4988-89b4-c69471f616aa/volumes" Oct 05 07:11:26 crc kubenswrapper[4846]: I1005 07:11:26.531793 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7fd963e7-aa89-4a46-8b1b-ac1bd6bc503d" path="/var/lib/kubelet/pods/7fd963e7-aa89-4a46-8b1b-ac1bd6bc503d/volumes" Oct 05 07:11:26 crc kubenswrapper[4846]: I1005 07:11:26.534755 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="979a46b2-96f3-498a-8838-4b73b4c8dc91" path="/var/lib/kubelet/pods/979a46b2-96f3-498a-8838-4b73b4c8dc91/volumes" Oct 05 07:11:26 crc kubenswrapper[4846]: I1005 07:11:26.535386 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a32cdd03-1398-4055-b051-62adde35200f" path="/var/lib/kubelet/pods/a32cdd03-1398-4055-b051-62adde35200f/volumes" Oct 05 07:11:26 crc kubenswrapper[4846]: I1005 07:11:26.536038 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b4b973ee-209f-477b-92ee-3f1c16f17f43" path="/var/lib/kubelet/pods/b4b973ee-209f-477b-92ee-3f1c16f17f43/volumes" Oct 05 07:11:26 crc kubenswrapper[4846]: I1005 07:11:26.536886 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bfe96e0c-b9fc-438d-9b17-bec246d48446" path="/var/lib/kubelet/pods/bfe96e0c-b9fc-438d-9b17-bec246d48446/volumes" Oct 05 07:11:26 crc kubenswrapper[4846]: I1005 07:11:26.540707 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c31937eb-9836-461b-9230-5e06ce7a99e8" path="/var/lib/kubelet/pods/c31937eb-9836-461b-9230-5e06ce7a99e8/volumes" Oct 05 07:11:26 crc kubenswrapper[4846]: I1005 07:11:26.541373 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c35affe0-da4e-448f-aac0-7bba26f89040" path="/var/lib/kubelet/pods/c35affe0-da4e-448f-aac0-7bba26f89040/volumes" Oct 05 07:11:26 crc kubenswrapper[4846]: I1005 07:11:26.541943 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d11b6df6-6da2-4cb8-a4f9-d87cfd5d4d26" path="/var/lib/kubelet/pods/d11b6df6-6da2-4cb8-a4f9-d87cfd5d4d26/volumes" Oct 05 07:11:26 crc kubenswrapper[4846]: I1005 07:11:26.543085 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e2f9f9a8-a25f-4a99-b425-4799964d4e24" path="/var/lib/kubelet/pods/e2f9f9a8-a25f-4a99-b425-4799964d4e24/volumes" Oct 05 07:11:26 crc kubenswrapper[4846]: I1005 07:11:26.543767 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e6e86b1c-87ac-4594-aeba-943947065374" path="/var/lib/kubelet/pods/e6e86b1c-87ac-4594-aeba-943947065374/volumes" Oct 05 07:11:26 crc kubenswrapper[4846]: I1005 07:11:26.544462 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f8c7a428-fb8a-4992-bf65-d603b4cfc323" path="/var/lib/kubelet/pods/f8c7a428-fb8a-4992-bf65-d603b4cfc323/volumes" Oct 05 07:11:26 crc kubenswrapper[4846]: I1005 07:11:26.568963 4846 scope.go:117] "RemoveContainer" containerID="7b4d83183acff2c37efb3d06dc23fc65569d8eb1f78f7305d68857a97455040e" Oct 05 07:11:26 crc kubenswrapper[4846]: I1005 07:11:26.665572 4846 generic.go:334] "Generic (PLEG): container finished" podID="b1e394dd-f55b-4af1-af0b-52cc7d09afa1" containerID="6682a3a723ef5336f655521943bfbd504709a37185a094744e9c63a0c851b903" exitCode=0 Oct 05 07:11:26 crc kubenswrapper[4846]: I1005 07:11:26.665742 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b1e394dd-f55b-4af1-af0b-52cc7d09afa1","Type":"ContainerDied","Data":"6682a3a723ef5336f655521943bfbd504709a37185a094744e9c63a0c851b903"} Oct 05 07:11:26 crc kubenswrapper[4846]: I1005 07:11:26.666026 4846 scope.go:117] "RemoveContainer" containerID="a2a13c5c05a60506397f6b2d96fab2cded887ebfc95de2c51ee9dd249fae3050" Oct 05 07:11:26 crc kubenswrapper[4846]: I1005 07:11:26.668550 4846 generic.go:334] "Generic (PLEG): container finished" podID="751734cb-38c8-46a0-9bac-05bcf24cc589" containerID="ed2b341c0cbc39aa9be3b1731eb5e7d43c786f18e7d5a05613b33a6282bdffcd" exitCode=0 Oct 05 07:11:26 crc kubenswrapper[4846]: I1005 07:11:26.668583 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"751734cb-38c8-46a0-9bac-05bcf24cc589","Type":"ContainerDied","Data":"ed2b341c0cbc39aa9be3b1731eb5e7d43c786f18e7d5a05613b33a6282bdffcd"} Oct 05 07:11:26 crc kubenswrapper[4846]: I1005 07:11:26.668627 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"751734cb-38c8-46a0-9bac-05bcf24cc589","Type":"ContainerDied","Data":"cc6b8981a9438d5d7acae93755280feb389f6382aef3df813588262d04e57ddb"} Oct 05 07:11:26 crc kubenswrapper[4846]: I1005 07:11:26.670492 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Oct 05 07:11:26 crc kubenswrapper[4846]: I1005 07:11:26.715625 4846 scope.go:117] "RemoveContainer" containerID="ed2b341c0cbc39aa9be3b1731eb5e7d43c786f18e7d5a05613b33a6282bdffcd" Oct 05 07:11:26 crc kubenswrapper[4846]: I1005 07:11:26.724764 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glancebc5f-account-delete-hkzcl" Oct 05 07:11:26 crc kubenswrapper[4846]: I1005 07:11:26.724768 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glancebc5f-account-delete-hkzcl" event={"ID":"4d3616b4-9f2c-4c62-9760-0300b02d0dbb","Type":"ContainerDied","Data":"1da9c21171f99446de2b2aa2c3001e53829eca259615385092f0f013986911bf"} Oct 05 07:11:26 crc kubenswrapper[4846]: I1005 07:11:26.724981 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1da9c21171f99446de2b2aa2c3001e53829eca259615385092f0f013986911bf" Oct 05 07:11:26 crc kubenswrapper[4846]: I1005 07:11:26.728788 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/memcached-0"] Oct 05 07:11:26 crc kubenswrapper[4846]: I1005 07:11:26.738626 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/memcached-0"] Oct 05 07:11:26 crc kubenswrapper[4846]: I1005 07:11:26.769292 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone67b1-account-delete-7lt6j" Oct 05 07:11:26 crc kubenswrapper[4846]: I1005 07:11:26.770411 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 05 07:11:26 crc kubenswrapper[4846]: I1005 07:11:26.770606 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"57b09608-08e5-4998-a3b3-e88b13587f58","Type":"ContainerDied","Data":"56ffad5ab9e3eaf08b4a3a4ff530c7b1dd85702163911520bca3b753e9901952"} Oct 05 07:11:26 crc kubenswrapper[4846]: I1005 07:11:26.792031 4846 scope.go:117] "RemoveContainer" containerID="ed2b341c0cbc39aa9be3b1731eb5e7d43c786f18e7d5a05613b33a6282bdffcd" Oct 05 07:11:26 crc kubenswrapper[4846]: I1005 07:11:26.792608 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone67b1-account-delete-7lt6j" Oct 05 07:11:26 crc kubenswrapper[4846]: E1005 07:11:26.792716 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ed2b341c0cbc39aa9be3b1731eb5e7d43c786f18e7d5a05613b33a6282bdffcd\": container with ID starting with ed2b341c0cbc39aa9be3b1731eb5e7d43c786f18e7d5a05613b33a6282bdffcd not found: ID does not exist" containerID="ed2b341c0cbc39aa9be3b1731eb5e7d43c786f18e7d5a05613b33a6282bdffcd" Oct 05 07:11:26 crc kubenswrapper[4846]: I1005 07:11:26.792745 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ed2b341c0cbc39aa9be3b1731eb5e7d43c786f18e7d5a05613b33a6282bdffcd"} err="failed to get container status \"ed2b341c0cbc39aa9be3b1731eb5e7d43c786f18e7d5a05613b33a6282bdffcd\": rpc error: code = NotFound desc = could not find container \"ed2b341c0cbc39aa9be3b1731eb5e7d43c786f18e7d5a05613b33a6282bdffcd\": container with ID starting with ed2b341c0cbc39aa9be3b1731eb5e7d43c786f18e7d5a05613b33a6282bdffcd not found: ID does not exist" Oct 05 07:11:26 crc kubenswrapper[4846]: I1005 07:11:26.792765 4846 scope.go:117] "RemoveContainer" containerID="0f36596d6da69b2a28102d8a73a67b3c631900ebfdb71663f52974e9dd99b1d1" Oct 05 07:11:26 crc kubenswrapper[4846]: I1005 07:11:26.795153 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glancebc5f-account-delete-hkzcl"] Oct 05 07:11:26 crc kubenswrapper[4846]: I1005 07:11:26.829917 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glancebc5f-account-delete-hkzcl"] Oct 05 07:11:26 crc kubenswrapper[4846]: I1005 07:11:26.862461 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 05 07:11:26 crc kubenswrapper[4846]: I1005 07:11:26.892088 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 05 07:11:26 crc kubenswrapper[4846]: I1005 07:11:26.935995 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9dmcg\" (UniqueName: \"kubernetes.io/projected/c3e9c65e-42bf-4cd6-a47d-b6ef0db5160e-kube-api-access-9dmcg\") pod \"keystone67b1-account-delete-7lt6j\" (UID: \"c3e9c65e-42bf-4cd6-a47d-b6ef0db5160e\") " pod="openstack/keystone67b1-account-delete-7lt6j" Oct 05 07:11:26 crc kubenswrapper[4846]: E1005 07:11:26.940945 4846 projected.go:194] Error preparing data for projected volume kube-api-access-9dmcg for pod openstack/keystone67b1-account-delete-7lt6j: failed to fetch token: serviceaccounts "galera-openstack" not found Oct 05 07:11:26 crc kubenswrapper[4846]: E1005 07:11:26.941004 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/c3e9c65e-42bf-4cd6-a47d-b6ef0db5160e-kube-api-access-9dmcg podName:c3e9c65e-42bf-4cd6-a47d-b6ef0db5160e nodeName:}" failed. No retries permitted until 2025-10-05 07:11:28.940987969 +0000 UTC m=+1411.181840744 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-9dmcg" (UniqueName: "kubernetes.io/projected/c3e9c65e-42bf-4cd6-a47d-b6ef0db5160e-kube-api-access-9dmcg") pod "keystone67b1-account-delete-7lt6j" (UID: "c3e9c65e-42bf-4cd6-a47d-b6ef0db5160e") : failed to fetch token: serviceaccounts "galera-openstack" not found Oct 05 07:11:26 crc kubenswrapper[4846]: E1005 07:11:26.956909 4846 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="055332eb57f5e6e889d548a1b4717caea296e84dc70616d5bdb53dd34aaa0697" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Oct 05 07:11:26 crc kubenswrapper[4846]: E1005 07:11:26.968162 4846 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="055332eb57f5e6e889d548a1b4717caea296e84dc70616d5bdb53dd34aaa0697" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Oct 05 07:11:26 crc kubenswrapper[4846]: E1005 07:11:26.987413 4846 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="56b8f764086c61683708e28e3f2929710527aada412eda06ddb4b14e487478c3" cmd=["/bin/bash","/var/lib/operator-scripts/mysql_probe.sh","readiness"] Oct 05 07:11:26 crc kubenswrapper[4846]: E1005 07:11:26.995386 4846 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="055332eb57f5e6e889d548a1b4717caea296e84dc70616d5bdb53dd34aaa0697" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Oct 05 07:11:26 crc kubenswrapper[4846]: E1005 07:11:26.995461 4846 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-cell0-conductor-0" podUID="98b50269-c6d4-42f2-8cef-afc127bfccb1" containerName="nova-cell0-conductor-conductor" Oct 05 07:11:26 crc kubenswrapper[4846]: E1005 07:11:26.999503 4846 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="56b8f764086c61683708e28e3f2929710527aada412eda06ddb4b14e487478c3" cmd=["/bin/bash","/var/lib/operator-scripts/mysql_probe.sh","readiness"] Oct 05 07:11:27 crc kubenswrapper[4846]: E1005 07:11:27.011377 4846 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="56b8f764086c61683708e28e3f2929710527aada412eda06ddb4b14e487478c3" cmd=["/bin/bash","/var/lib/operator-scripts/mysql_probe.sh","readiness"] Oct 05 07:11:27 crc kubenswrapper[4846]: E1005 07:11:27.011469 4846 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/openstack-galera-0" podUID="d88c48f9-948e-4888-986e-e00163ec93a6" containerName="galera" Oct 05 07:11:27 crc kubenswrapper[4846]: E1005 07:11:27.037652 4846 configmap.go:193] Couldn't get configMap openstack/rabbitmq-cell1-config-data: configmap "rabbitmq-cell1-config-data" not found Oct 05 07:11:27 crc kubenswrapper[4846]: E1005 07:11:27.038072 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/626e6bce-7422-4fbd-b978-dfa5cb93de34-config-data podName:626e6bce-7422-4fbd-b978-dfa5cb93de34 nodeName:}" failed. No retries permitted until 2025-10-05 07:11:35.038051729 +0000 UTC m=+1417.278904504 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/626e6bce-7422-4fbd-b978-dfa5cb93de34-config-data") pod "rabbitmq-cell1-server-0" (UID: "626e6bce-7422-4fbd-b978-dfa5cb93de34") : configmap "rabbitmq-cell1-config-data" not found Oct 05 07:11:27 crc kubenswrapper[4846]: E1005 07:11:27.372573 4846 handlers.go:78] "Exec lifecycle hook for Container in Pod failed" err=< Oct 05 07:11:27 crc kubenswrapper[4846]: command '/usr/share/ovn/scripts/ovn-ctl stop_controller' exited with 137: 2025-10-05T07:11:19Z|00001|fatal_signal|WARN|terminating with signal 14 (Alarm clock) Oct 05 07:11:27 crc kubenswrapper[4846]: /etc/init.d/functions: line 589: 442 Alarm clock "$@" Oct 05 07:11:27 crc kubenswrapper[4846]: > execCommand=["/usr/share/ovn/scripts/ovn-ctl","stop_controller"] containerName="ovn-controller" pod="openstack/ovn-controller-4kqnw" message=< Oct 05 07:11:27 crc kubenswrapper[4846]: Exiting ovn-controller (1) [FAILED] Oct 05 07:11:27 crc kubenswrapper[4846]: Killing ovn-controller (1) [ OK ] Oct 05 07:11:27 crc kubenswrapper[4846]: Killing ovn-controller (1) with SIGKILL [ OK ] Oct 05 07:11:27 crc kubenswrapper[4846]: 2025-10-05T07:11:19Z|00001|fatal_signal|WARN|terminating with signal 14 (Alarm clock) Oct 05 07:11:27 crc kubenswrapper[4846]: /etc/init.d/functions: line 589: 442 Alarm clock "$@" Oct 05 07:11:27 crc kubenswrapper[4846]: > Oct 05 07:11:27 crc kubenswrapper[4846]: E1005 07:11:27.372631 4846 kuberuntime_container.go:691] "PreStop hook failed" err=< Oct 05 07:11:27 crc kubenswrapper[4846]: command '/usr/share/ovn/scripts/ovn-ctl stop_controller' exited with 137: 2025-10-05T07:11:19Z|00001|fatal_signal|WARN|terminating with signal 14 (Alarm clock) Oct 05 07:11:27 crc kubenswrapper[4846]: /etc/init.d/functions: line 589: 442 Alarm clock "$@" Oct 05 07:11:27 crc kubenswrapper[4846]: > pod="openstack/ovn-controller-4kqnw" podUID="169918d8-597e-4ef4-b00c-fb14c4d0347f" containerName="ovn-controller" containerID="cri-o://d171365315f951cc3fe3c42e399c18ed39c3027267575259a8e68fb2d38283e1" Oct 05 07:11:27 crc kubenswrapper[4846]: I1005 07:11:27.372699 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-controller-4kqnw" podUID="169918d8-597e-4ef4-b00c-fb14c4d0347f" containerName="ovn-controller" containerID="cri-o://d171365315f951cc3fe3c42e399c18ed39c3027267575259a8e68fb2d38283e1" gracePeriod=22 Oct 05 07:11:27 crc kubenswrapper[4846]: I1005 07:11:27.422717 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 05 07:11:27 crc kubenswrapper[4846]: I1005 07:11:27.548287 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/35261dbc-cabe-4bf0-88f6-b8613ddec0f1-erlang-cookie-secret\") pod \"35261dbc-cabe-4bf0-88f6-b8613ddec0f1\" (UID: \"35261dbc-cabe-4bf0-88f6-b8613ddec0f1\") " Oct 05 07:11:27 crc kubenswrapper[4846]: I1005 07:11:27.548348 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/35261dbc-cabe-4bf0-88f6-b8613ddec0f1-rabbitmq-erlang-cookie\") pod \"35261dbc-cabe-4bf0-88f6-b8613ddec0f1\" (UID: \"35261dbc-cabe-4bf0-88f6-b8613ddec0f1\") " Oct 05 07:11:27 crc kubenswrapper[4846]: I1005 07:11:27.548759 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/35261dbc-cabe-4bf0-88f6-b8613ddec0f1-rabbitmq-confd\") pod \"35261dbc-cabe-4bf0-88f6-b8613ddec0f1\" (UID: \"35261dbc-cabe-4bf0-88f6-b8613ddec0f1\") " Oct 05 07:11:27 crc kubenswrapper[4846]: I1005 07:11:27.548801 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/35261dbc-cabe-4bf0-88f6-b8613ddec0f1-config-data\") pod \"35261dbc-cabe-4bf0-88f6-b8613ddec0f1\" (UID: \"35261dbc-cabe-4bf0-88f6-b8613ddec0f1\") " Oct 05 07:11:27 crc kubenswrapper[4846]: I1005 07:11:27.548847 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/35261dbc-cabe-4bf0-88f6-b8613ddec0f1-server-conf\") pod \"35261dbc-cabe-4bf0-88f6-b8613ddec0f1\" (UID: \"35261dbc-cabe-4bf0-88f6-b8613ddec0f1\") " Oct 05 07:11:27 crc kubenswrapper[4846]: I1005 07:11:27.548878 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/35261dbc-cabe-4bf0-88f6-b8613ddec0f1-rabbitmq-tls\") pod \"35261dbc-cabe-4bf0-88f6-b8613ddec0f1\" (UID: \"35261dbc-cabe-4bf0-88f6-b8613ddec0f1\") " Oct 05 07:11:27 crc kubenswrapper[4846]: I1005 07:11:27.548928 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j5pb6\" (UniqueName: \"kubernetes.io/projected/35261dbc-cabe-4bf0-88f6-b8613ddec0f1-kube-api-access-j5pb6\") pod \"35261dbc-cabe-4bf0-88f6-b8613ddec0f1\" (UID: \"35261dbc-cabe-4bf0-88f6-b8613ddec0f1\") " Oct 05 07:11:27 crc kubenswrapper[4846]: I1005 07:11:27.549012 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"35261dbc-cabe-4bf0-88f6-b8613ddec0f1\" (UID: \"35261dbc-cabe-4bf0-88f6-b8613ddec0f1\") " Oct 05 07:11:27 crc kubenswrapper[4846]: I1005 07:11:27.549056 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/35261dbc-cabe-4bf0-88f6-b8613ddec0f1-rabbitmq-plugins\") pod \"35261dbc-cabe-4bf0-88f6-b8613ddec0f1\" (UID: \"35261dbc-cabe-4bf0-88f6-b8613ddec0f1\") " Oct 05 07:11:27 crc kubenswrapper[4846]: I1005 07:11:27.549546 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/35261dbc-cabe-4bf0-88f6-b8613ddec0f1-pod-info\") pod \"35261dbc-cabe-4bf0-88f6-b8613ddec0f1\" (UID: \"35261dbc-cabe-4bf0-88f6-b8613ddec0f1\") " Oct 05 07:11:27 crc kubenswrapper[4846]: I1005 07:11:27.549593 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/35261dbc-cabe-4bf0-88f6-b8613ddec0f1-plugins-conf\") pod \"35261dbc-cabe-4bf0-88f6-b8613ddec0f1\" (UID: \"35261dbc-cabe-4bf0-88f6-b8613ddec0f1\") " Oct 05 07:11:27 crc kubenswrapper[4846]: I1005 07:11:27.551803 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/35261dbc-cabe-4bf0-88f6-b8613ddec0f1-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "35261dbc-cabe-4bf0-88f6-b8613ddec0f1" (UID: "35261dbc-cabe-4bf0-88f6-b8613ddec0f1"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 07:11:27 crc kubenswrapper[4846]: I1005 07:11:27.552768 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/35261dbc-cabe-4bf0-88f6-b8613ddec0f1-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "35261dbc-cabe-4bf0-88f6-b8613ddec0f1" (UID: "35261dbc-cabe-4bf0-88f6-b8613ddec0f1"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 07:11:27 crc kubenswrapper[4846]: I1005 07:11:27.556140 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/35261dbc-cabe-4bf0-88f6-b8613ddec0f1-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "35261dbc-cabe-4bf0-88f6-b8613ddec0f1" (UID: "35261dbc-cabe-4bf0-88f6-b8613ddec0f1"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 07:11:27 crc kubenswrapper[4846]: I1005 07:11:27.558340 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/35261dbc-cabe-4bf0-88f6-b8613ddec0f1-pod-info" (OuterVolumeSpecName: "pod-info") pod "35261dbc-cabe-4bf0-88f6-b8613ddec0f1" (UID: "35261dbc-cabe-4bf0-88f6-b8613ddec0f1"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Oct 05 07:11:27 crc kubenswrapper[4846]: I1005 07:11:27.560328 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/35261dbc-cabe-4bf0-88f6-b8613ddec0f1-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "35261dbc-cabe-4bf0-88f6-b8613ddec0f1" (UID: "35261dbc-cabe-4bf0-88f6-b8613ddec0f1"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:11:27 crc kubenswrapper[4846]: I1005 07:11:27.563883 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/35261dbc-cabe-4bf0-88f6-b8613ddec0f1-kube-api-access-j5pb6" (OuterVolumeSpecName: "kube-api-access-j5pb6") pod "35261dbc-cabe-4bf0-88f6-b8613ddec0f1" (UID: "35261dbc-cabe-4bf0-88f6-b8613ddec0f1"). InnerVolumeSpecName "kube-api-access-j5pb6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:11:27 crc kubenswrapper[4846]: I1005 07:11:27.565269 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/35261dbc-cabe-4bf0-88f6-b8613ddec0f1-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "35261dbc-cabe-4bf0-88f6-b8613ddec0f1" (UID: "35261dbc-cabe-4bf0-88f6-b8613ddec0f1"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:11:27 crc kubenswrapper[4846]: I1005 07:11:27.566035 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage10-crc" (OuterVolumeSpecName: "persistence") pod "35261dbc-cabe-4bf0-88f6-b8613ddec0f1" (UID: "35261dbc-cabe-4bf0-88f6-b8613ddec0f1"). InnerVolumeSpecName "local-storage10-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 05 07:11:27 crc kubenswrapper[4846]: I1005 07:11:27.566164 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 05 07:11:27 crc kubenswrapper[4846]: I1005 07:11:27.581755 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/35261dbc-cabe-4bf0-88f6-b8613ddec0f1-config-data" (OuterVolumeSpecName: "config-data") pod "35261dbc-cabe-4bf0-88f6-b8613ddec0f1" (UID: "35261dbc-cabe-4bf0-88f6-b8613ddec0f1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 07:11:27 crc kubenswrapper[4846]: I1005 07:11:27.624983 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/35261dbc-cabe-4bf0-88f6-b8613ddec0f1-server-conf" (OuterVolumeSpecName: "server-conf") pod "35261dbc-cabe-4bf0-88f6-b8613ddec0f1" (UID: "35261dbc-cabe-4bf0-88f6-b8613ddec0f1"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 07:11:27 crc kubenswrapper[4846]: I1005 07:11:27.653285 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/626e6bce-7422-4fbd-b978-dfa5cb93de34-rabbitmq-tls\") pod \"626e6bce-7422-4fbd-b978-dfa5cb93de34\" (UID: \"626e6bce-7422-4fbd-b978-dfa5cb93de34\") " Oct 05 07:11:27 crc kubenswrapper[4846]: I1005 07:11:27.653349 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/626e6bce-7422-4fbd-b978-dfa5cb93de34-server-conf\") pod \"626e6bce-7422-4fbd-b978-dfa5cb93de34\" (UID: \"626e6bce-7422-4fbd-b978-dfa5cb93de34\") " Oct 05 07:11:27 crc kubenswrapper[4846]: I1005 07:11:27.653420 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gtvkd\" (UniqueName: \"kubernetes.io/projected/626e6bce-7422-4fbd-b978-dfa5cb93de34-kube-api-access-gtvkd\") pod \"626e6bce-7422-4fbd-b978-dfa5cb93de34\" (UID: \"626e6bce-7422-4fbd-b978-dfa5cb93de34\") " Oct 05 07:11:27 crc kubenswrapper[4846]: I1005 07:11:27.653458 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/626e6bce-7422-4fbd-b978-dfa5cb93de34-pod-info\") pod \"626e6bce-7422-4fbd-b978-dfa5cb93de34\" (UID: \"626e6bce-7422-4fbd-b978-dfa5cb93de34\") " Oct 05 07:11:27 crc kubenswrapper[4846]: I1005 07:11:27.653521 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/626e6bce-7422-4fbd-b978-dfa5cb93de34-plugins-conf\") pod \"626e6bce-7422-4fbd-b978-dfa5cb93de34\" (UID: \"626e6bce-7422-4fbd-b978-dfa5cb93de34\") " Oct 05 07:11:27 crc kubenswrapper[4846]: I1005 07:11:27.653552 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/626e6bce-7422-4fbd-b978-dfa5cb93de34-rabbitmq-plugins\") pod \"626e6bce-7422-4fbd-b978-dfa5cb93de34\" (UID: \"626e6bce-7422-4fbd-b978-dfa5cb93de34\") " Oct 05 07:11:27 crc kubenswrapper[4846]: I1005 07:11:27.653607 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/626e6bce-7422-4fbd-b978-dfa5cb93de34-config-data\") pod \"626e6bce-7422-4fbd-b978-dfa5cb93de34\" (UID: \"626e6bce-7422-4fbd-b978-dfa5cb93de34\") " Oct 05 07:11:27 crc kubenswrapper[4846]: I1005 07:11:27.653630 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"626e6bce-7422-4fbd-b978-dfa5cb93de34\" (UID: \"626e6bce-7422-4fbd-b978-dfa5cb93de34\") " Oct 05 07:11:27 crc kubenswrapper[4846]: I1005 07:11:27.653660 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/626e6bce-7422-4fbd-b978-dfa5cb93de34-rabbitmq-confd\") pod \"626e6bce-7422-4fbd-b978-dfa5cb93de34\" (UID: \"626e6bce-7422-4fbd-b978-dfa5cb93de34\") " Oct 05 07:11:27 crc kubenswrapper[4846]: I1005 07:11:27.653697 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/626e6bce-7422-4fbd-b978-dfa5cb93de34-erlang-cookie-secret\") pod \"626e6bce-7422-4fbd-b978-dfa5cb93de34\" (UID: \"626e6bce-7422-4fbd-b978-dfa5cb93de34\") " Oct 05 07:11:27 crc kubenswrapper[4846]: I1005 07:11:27.653730 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/626e6bce-7422-4fbd-b978-dfa5cb93de34-rabbitmq-erlang-cookie\") pod \"626e6bce-7422-4fbd-b978-dfa5cb93de34\" (UID: \"626e6bce-7422-4fbd-b978-dfa5cb93de34\") " Oct 05 07:11:27 crc kubenswrapper[4846]: I1005 07:11:27.654150 4846 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" " Oct 05 07:11:27 crc kubenswrapper[4846]: I1005 07:11:27.654180 4846 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/35261dbc-cabe-4bf0-88f6-b8613ddec0f1-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:27 crc kubenswrapper[4846]: I1005 07:11:27.659914 4846 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/35261dbc-cabe-4bf0-88f6-b8613ddec0f1-pod-info\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:27 crc kubenswrapper[4846]: I1005 07:11:27.659932 4846 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/35261dbc-cabe-4bf0-88f6-b8613ddec0f1-plugins-conf\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:27 crc kubenswrapper[4846]: I1005 07:11:27.659944 4846 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/35261dbc-cabe-4bf0-88f6-b8613ddec0f1-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:27 crc kubenswrapper[4846]: I1005 07:11:27.659957 4846 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/35261dbc-cabe-4bf0-88f6-b8613ddec0f1-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:27 crc kubenswrapper[4846]: I1005 07:11:27.659976 4846 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/35261dbc-cabe-4bf0-88f6-b8613ddec0f1-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:27 crc kubenswrapper[4846]: I1005 07:11:27.659990 4846 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/35261dbc-cabe-4bf0-88f6-b8613ddec0f1-server-conf\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:27 crc kubenswrapper[4846]: I1005 07:11:27.660004 4846 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/35261dbc-cabe-4bf0-88f6-b8613ddec0f1-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:27 crc kubenswrapper[4846]: I1005 07:11:27.660015 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j5pb6\" (UniqueName: \"kubernetes.io/projected/35261dbc-cabe-4bf0-88f6-b8613ddec0f1-kube-api-access-j5pb6\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:27 crc kubenswrapper[4846]: I1005 07:11:27.660428 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/626e6bce-7422-4fbd-b978-dfa5cb93de34-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "626e6bce-7422-4fbd-b978-dfa5cb93de34" (UID: "626e6bce-7422-4fbd-b978-dfa5cb93de34"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 07:11:27 crc kubenswrapper[4846]: I1005 07:11:27.661944 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/626e6bce-7422-4fbd-b978-dfa5cb93de34-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "626e6bce-7422-4fbd-b978-dfa5cb93de34" (UID: "626e6bce-7422-4fbd-b978-dfa5cb93de34"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 07:11:27 crc kubenswrapper[4846]: I1005 07:11:27.662056 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/626e6bce-7422-4fbd-b978-dfa5cb93de34-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "626e6bce-7422-4fbd-b978-dfa5cb93de34" (UID: "626e6bce-7422-4fbd-b978-dfa5cb93de34"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 07:11:27 crc kubenswrapper[4846]: I1005 07:11:27.670293 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage02-crc" (OuterVolumeSpecName: "persistence") pod "626e6bce-7422-4fbd-b978-dfa5cb93de34" (UID: "626e6bce-7422-4fbd-b978-dfa5cb93de34"). InnerVolumeSpecName "local-storage02-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 05 07:11:27 crc kubenswrapper[4846]: I1005 07:11:27.670363 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/626e6bce-7422-4fbd-b978-dfa5cb93de34-pod-info" (OuterVolumeSpecName: "pod-info") pod "626e6bce-7422-4fbd-b978-dfa5cb93de34" (UID: "626e6bce-7422-4fbd-b978-dfa5cb93de34"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Oct 05 07:11:27 crc kubenswrapper[4846]: I1005 07:11:27.670415 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/626e6bce-7422-4fbd-b978-dfa5cb93de34-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "626e6bce-7422-4fbd-b978-dfa5cb93de34" (UID: "626e6bce-7422-4fbd-b978-dfa5cb93de34"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:11:27 crc kubenswrapper[4846]: I1005 07:11:27.670509 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/626e6bce-7422-4fbd-b978-dfa5cb93de34-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "626e6bce-7422-4fbd-b978-dfa5cb93de34" (UID: "626e6bce-7422-4fbd-b978-dfa5cb93de34"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:11:27 crc kubenswrapper[4846]: I1005 07:11:27.695223 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/626e6bce-7422-4fbd-b978-dfa5cb93de34-kube-api-access-gtvkd" (OuterVolumeSpecName: "kube-api-access-gtvkd") pod "626e6bce-7422-4fbd-b978-dfa5cb93de34" (UID: "626e6bce-7422-4fbd-b978-dfa5cb93de34"). InnerVolumeSpecName "kube-api-access-gtvkd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:11:27 crc kubenswrapper[4846]: I1005 07:11:27.696573 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/626e6bce-7422-4fbd-b978-dfa5cb93de34-config-data" (OuterVolumeSpecName: "config-data") pod "626e6bce-7422-4fbd-b978-dfa5cb93de34" (UID: "626e6bce-7422-4fbd-b978-dfa5cb93de34"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 07:11:27 crc kubenswrapper[4846]: I1005 07:11:27.708234 4846 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage10-crc" (UniqueName: "kubernetes.io/local-volume/local-storage10-crc") on node "crc" Oct 05 07:11:27 crc kubenswrapper[4846]: I1005 07:11:27.726089 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/35261dbc-cabe-4bf0-88f6-b8613ddec0f1-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "35261dbc-cabe-4bf0-88f6-b8613ddec0f1" (UID: "35261dbc-cabe-4bf0-88f6-b8613ddec0f1"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:11:27 crc kubenswrapper[4846]: I1005 07:11:27.745869 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/626e6bce-7422-4fbd-b978-dfa5cb93de34-server-conf" (OuterVolumeSpecName: "server-conf") pod "626e6bce-7422-4fbd-b978-dfa5cb93de34" (UID: "626e6bce-7422-4fbd-b978-dfa5cb93de34"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 07:11:27 crc kubenswrapper[4846]: I1005 07:11:27.761790 4846 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/35261dbc-cabe-4bf0-88f6-b8613ddec0f1-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:27 crc kubenswrapper[4846]: I1005 07:11:27.761815 4846 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/626e6bce-7422-4fbd-b978-dfa5cb93de34-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:27 crc kubenswrapper[4846]: I1005 07:11:27.761824 4846 reconciler_common.go:293] "Volume detached for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:27 crc kubenswrapper[4846]: I1005 07:11:27.761833 4846 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/626e6bce-7422-4fbd-b978-dfa5cb93de34-server-conf\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:27 crc kubenswrapper[4846]: I1005 07:11:27.761842 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gtvkd\" (UniqueName: \"kubernetes.io/projected/626e6bce-7422-4fbd-b978-dfa5cb93de34-kube-api-access-gtvkd\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:27 crc kubenswrapper[4846]: I1005 07:11:27.761851 4846 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/626e6bce-7422-4fbd-b978-dfa5cb93de34-pod-info\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:27 crc kubenswrapper[4846]: I1005 07:11:27.761859 4846 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/626e6bce-7422-4fbd-b978-dfa5cb93de34-plugins-conf\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:27 crc kubenswrapper[4846]: I1005 07:11:27.761867 4846 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/626e6bce-7422-4fbd-b978-dfa5cb93de34-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:27 crc kubenswrapper[4846]: I1005 07:11:27.761876 4846 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/626e6bce-7422-4fbd-b978-dfa5cb93de34-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:27 crc kubenswrapper[4846]: I1005 07:11:27.761899 4846 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" " Oct 05 07:11:27 crc kubenswrapper[4846]: I1005 07:11:27.761907 4846 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/626e6bce-7422-4fbd-b978-dfa5cb93de34-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:27 crc kubenswrapper[4846]: I1005 07:11:27.761917 4846 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/626e6bce-7422-4fbd-b978-dfa5cb93de34-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:27 crc kubenswrapper[4846]: I1005 07:11:27.782251 4846 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage02-crc" (UniqueName: "kubernetes.io/local-volume/local-storage02-crc") on node "crc" Oct 05 07:11:27 crc kubenswrapper[4846]: I1005 07:11:27.786277 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/626e6bce-7422-4fbd-b978-dfa5cb93de34-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "626e6bce-7422-4fbd-b978-dfa5cb93de34" (UID: "626e6bce-7422-4fbd-b978-dfa5cb93de34"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:11:27 crc kubenswrapper[4846]: I1005 07:11:27.809980 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_18a7be37-c8dc-4d8a-808f-2a63eb4d9692/ovn-northd/0.log" Oct 05 07:11:27 crc kubenswrapper[4846]: I1005 07:11:27.810041 4846 generic.go:334] "Generic (PLEG): container finished" podID="18a7be37-c8dc-4d8a-808f-2a63eb4d9692" containerID="4351908cf00eea1071aca5080f9d536de25e56a78e5d0713bec8f51caab84ecb" exitCode=139 Oct 05 07:11:27 crc kubenswrapper[4846]: I1005 07:11:27.810196 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"18a7be37-c8dc-4d8a-808f-2a63eb4d9692","Type":"ContainerDied","Data":"4351908cf00eea1071aca5080f9d536de25e56a78e5d0713bec8f51caab84ecb"} Oct 05 07:11:27 crc kubenswrapper[4846]: I1005 07:11:27.826623 4846 generic.go:334] "Generic (PLEG): container finished" podID="626e6bce-7422-4fbd-b978-dfa5cb93de34" containerID="c63a03c8ae2116953b605cde504ae09f678d59deda18e41527a386da67ee2ce4" exitCode=0 Oct 05 07:11:27 crc kubenswrapper[4846]: I1005 07:11:27.826811 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 05 07:11:27 crc kubenswrapper[4846]: I1005 07:11:27.827059 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"626e6bce-7422-4fbd-b978-dfa5cb93de34","Type":"ContainerDied","Data":"c63a03c8ae2116953b605cde504ae09f678d59deda18e41527a386da67ee2ce4"} Oct 05 07:11:27 crc kubenswrapper[4846]: I1005 07:11:27.827135 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"626e6bce-7422-4fbd-b978-dfa5cb93de34","Type":"ContainerDied","Data":"9a98fff94c37b3930255c3af93b8ce72516be04af351656fe24b90b08afac214"} Oct 05 07:11:27 crc kubenswrapper[4846]: I1005 07:11:27.827160 4846 scope.go:117] "RemoveContainer" containerID="c63a03c8ae2116953b605cde504ae09f678d59deda18e41527a386da67ee2ce4" Oct 05 07:11:27 crc kubenswrapper[4846]: I1005 07:11:27.840926 4846 generic.go:334] "Generic (PLEG): container finished" podID="35261dbc-cabe-4bf0-88f6-b8613ddec0f1" containerID="0b472a4ecaa620dd0581763afea4046cc8d15b55265470a90367ed2c9f54e39c" exitCode=0 Oct 05 07:11:27 crc kubenswrapper[4846]: I1005 07:11:27.841013 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"35261dbc-cabe-4bf0-88f6-b8613ddec0f1","Type":"ContainerDied","Data":"0b472a4ecaa620dd0581763afea4046cc8d15b55265470a90367ed2c9f54e39c"} Oct 05 07:11:27 crc kubenswrapper[4846]: I1005 07:11:27.841053 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"35261dbc-cabe-4bf0-88f6-b8613ddec0f1","Type":"ContainerDied","Data":"3c185c416d138656ea723c6753d54a6c7cca27516a80c1aaf2874043c2241621"} Oct 05 07:11:27 crc kubenswrapper[4846]: I1005 07:11:27.841133 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 05 07:11:27 crc kubenswrapper[4846]: I1005 07:11:27.864606 4846 reconciler_common.go:293] "Volume detached for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:27 crc kubenswrapper[4846]: I1005 07:11:27.864650 4846 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/626e6bce-7422-4fbd-b978-dfa5cb93de34-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:27 crc kubenswrapper[4846]: I1005 07:11:27.881316 4846 scope.go:117] "RemoveContainer" containerID="516d1ce74e479dfd0f7702334c1c73ac78a55e7ab16e082869f8bbbabc5a6c16" Oct 05 07:11:27 crc kubenswrapper[4846]: E1005 07:11:27.881507 4846 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of d171365315f951cc3fe3c42e399c18ed39c3027267575259a8e68fb2d38283e1 is running failed: container process not found" containerID="d171365315f951cc3fe3c42e399c18ed39c3027267575259a8e68fb2d38283e1" cmd=["/usr/local/bin/container-scripts/ovn_controller_readiness.sh"] Oct 05 07:11:27 crc kubenswrapper[4846]: I1005 07:11:27.881666 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-4kqnw_169918d8-597e-4ef4-b00c-fb14c4d0347f/ovn-controller/0.log" Oct 05 07:11:27 crc kubenswrapper[4846]: I1005 07:11:27.881707 4846 generic.go:334] "Generic (PLEG): container finished" podID="169918d8-597e-4ef4-b00c-fb14c4d0347f" containerID="d171365315f951cc3fe3c42e399c18ed39c3027267575259a8e68fb2d38283e1" exitCode=137 Oct 05 07:11:27 crc kubenswrapper[4846]: I1005 07:11:27.881759 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone67b1-account-delete-7lt6j" Oct 05 07:11:27 crc kubenswrapper[4846]: I1005 07:11:27.881858 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-4kqnw" event={"ID":"169918d8-597e-4ef4-b00c-fb14c4d0347f","Type":"ContainerDied","Data":"d171365315f951cc3fe3c42e399c18ed39c3027267575259a8e68fb2d38283e1"} Oct 05 07:11:27 crc kubenswrapper[4846]: E1005 07:11:27.888349 4846 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of d171365315f951cc3fe3c42e399c18ed39c3027267575259a8e68fb2d38283e1 is running failed: container process not found" containerID="d171365315f951cc3fe3c42e399c18ed39c3027267575259a8e68fb2d38283e1" cmd=["/usr/local/bin/container-scripts/ovn_controller_readiness.sh"] Oct 05 07:11:27 crc kubenswrapper[4846]: I1005 07:11:27.889543 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 05 07:11:27 crc kubenswrapper[4846]: E1005 07:11:27.890611 4846 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of d171365315f951cc3fe3c42e399c18ed39c3027267575259a8e68fb2d38283e1 is running failed: container process not found" containerID="d171365315f951cc3fe3c42e399c18ed39c3027267575259a8e68fb2d38283e1" cmd=["/usr/local/bin/container-scripts/ovn_controller_readiness.sh"] Oct 05 07:11:27 crc kubenswrapper[4846]: E1005 07:11:27.890647 4846 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of d171365315f951cc3fe3c42e399c18ed39c3027267575259a8e68fb2d38283e1 is running failed: container process not found" probeType="Readiness" pod="openstack/ovn-controller-4kqnw" podUID="169918d8-597e-4ef4-b00c-fb14c4d0347f" containerName="ovn-controller" Oct 05 07:11:27 crc kubenswrapper[4846]: I1005 07:11:27.909227 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 05 07:11:27 crc kubenswrapper[4846]: E1005 07:11:27.948405 4846 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of f4fa8732ebd0803a3d57af2df182505f645e900471768d31fb1d43cb0a83e0cf is running failed: container process not found" containerID="f4fa8732ebd0803a3d57af2df182505f645e900471768d31fb1d43cb0a83e0cf" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 05 07:11:27 crc kubenswrapper[4846]: E1005 07:11:27.948897 4846 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of f4fa8732ebd0803a3d57af2df182505f645e900471768d31fb1d43cb0a83e0cf is running failed: container process not found" containerID="f4fa8732ebd0803a3d57af2df182505f645e900471768d31fb1d43cb0a83e0cf" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 05 07:11:27 crc kubenswrapper[4846]: E1005 07:11:27.949226 4846 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of f4fa8732ebd0803a3d57af2df182505f645e900471768d31fb1d43cb0a83e0cf is running failed: container process not found" containerID="f4fa8732ebd0803a3d57af2df182505f645e900471768d31fb1d43cb0a83e0cf" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 05 07:11:27 crc kubenswrapper[4846]: E1005 07:11:27.949258 4846 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of f4fa8732ebd0803a3d57af2df182505f645e900471768d31fb1d43cb0a83e0cf is running failed: container process not found" probeType="Readiness" pod="openstack/ovn-controller-ovs-sg4vg" podUID="29c0f404-fc7d-4909-b6bf-e1826a092a72" containerName="ovsdb-server" Oct 05 07:11:27 crc kubenswrapper[4846]: I1005 07:11:27.956417 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 05 07:11:27 crc kubenswrapper[4846]: E1005 07:11:27.970750 4846 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="22929dba446bc2516169306e09540fa52839ffc6b8d024f494720cc39ad0ab7d" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 05 07:11:27 crc kubenswrapper[4846]: I1005 07:11:27.973131 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 05 07:11:27 crc kubenswrapper[4846]: E1005 07:11:27.974273 4846 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="22929dba446bc2516169306e09540fa52839ffc6b8d024f494720cc39ad0ab7d" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 05 07:11:27 crc kubenswrapper[4846]: E1005 07:11:27.984024 4846 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="22929dba446bc2516169306e09540fa52839ffc6b8d024f494720cc39ad0ab7d" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 05 07:11:27 crc kubenswrapper[4846]: E1005 07:11:27.984129 4846 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-controller-ovs-sg4vg" podUID="29c0f404-fc7d-4909-b6bf-e1826a092a72" containerName="ovs-vswitchd" Oct 05 07:11:28 crc kubenswrapper[4846]: I1005 07:11:28.014359 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone67b1-account-delete-7lt6j"] Oct 05 07:11:28 crc kubenswrapper[4846]: I1005 07:11:28.019736 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone67b1-account-delete-7lt6j"] Oct 05 07:11:28 crc kubenswrapper[4846]: I1005 07:11:28.021680 4846 scope.go:117] "RemoveContainer" containerID="c63a03c8ae2116953b605cde504ae09f678d59deda18e41527a386da67ee2ce4" Oct 05 07:11:28 crc kubenswrapper[4846]: E1005 07:11:28.022256 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c63a03c8ae2116953b605cde504ae09f678d59deda18e41527a386da67ee2ce4\": container with ID starting with c63a03c8ae2116953b605cde504ae09f678d59deda18e41527a386da67ee2ce4 not found: ID does not exist" containerID="c63a03c8ae2116953b605cde504ae09f678d59deda18e41527a386da67ee2ce4" Oct 05 07:11:28 crc kubenswrapper[4846]: I1005 07:11:28.022306 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c63a03c8ae2116953b605cde504ae09f678d59deda18e41527a386da67ee2ce4"} err="failed to get container status \"c63a03c8ae2116953b605cde504ae09f678d59deda18e41527a386da67ee2ce4\": rpc error: code = NotFound desc = could not find container \"c63a03c8ae2116953b605cde504ae09f678d59deda18e41527a386da67ee2ce4\": container with ID starting with c63a03c8ae2116953b605cde504ae09f678d59deda18e41527a386da67ee2ce4 not found: ID does not exist" Oct 05 07:11:28 crc kubenswrapper[4846]: I1005 07:11:28.022340 4846 scope.go:117] "RemoveContainer" containerID="516d1ce74e479dfd0f7702334c1c73ac78a55e7ab16e082869f8bbbabc5a6c16" Oct 05 07:11:28 crc kubenswrapper[4846]: E1005 07:11:28.022697 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"516d1ce74e479dfd0f7702334c1c73ac78a55e7ab16e082869f8bbbabc5a6c16\": container with ID starting with 516d1ce74e479dfd0f7702334c1c73ac78a55e7ab16e082869f8bbbabc5a6c16 not found: ID does not exist" containerID="516d1ce74e479dfd0f7702334c1c73ac78a55e7ab16e082869f8bbbabc5a6c16" Oct 05 07:11:28 crc kubenswrapper[4846]: I1005 07:11:28.022739 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"516d1ce74e479dfd0f7702334c1c73ac78a55e7ab16e082869f8bbbabc5a6c16"} err="failed to get container status \"516d1ce74e479dfd0f7702334c1c73ac78a55e7ab16e082869f8bbbabc5a6c16\": rpc error: code = NotFound desc = could not find container \"516d1ce74e479dfd0f7702334c1c73ac78a55e7ab16e082869f8bbbabc5a6c16\": container with ID starting with 516d1ce74e479dfd0f7702334c1c73ac78a55e7ab16e082869f8bbbabc5a6c16 not found: ID does not exist" Oct 05 07:11:28 crc kubenswrapper[4846]: I1005 07:11:28.022772 4846 scope.go:117] "RemoveContainer" containerID="0b472a4ecaa620dd0581763afea4046cc8d15b55265470a90367ed2c9f54e39c" Oct 05 07:11:28 crc kubenswrapper[4846]: I1005 07:11:28.067571 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9dmcg\" (UniqueName: \"kubernetes.io/projected/c3e9c65e-42bf-4cd6-a47d-b6ef0db5160e-kube-api-access-9dmcg\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:28 crc kubenswrapper[4846]: I1005 07:11:28.304953 4846 scope.go:117] "RemoveContainer" containerID="9919ee8b928899c8295fa6d8951cffdd5cffd78c7de5c1e7b1a230ab2550fbf6" Oct 05 07:11:28 crc kubenswrapper[4846]: I1005 07:11:28.325013 4846 scope.go:117] "RemoveContainer" containerID="0b472a4ecaa620dd0581763afea4046cc8d15b55265470a90367ed2c9f54e39c" Oct 05 07:11:28 crc kubenswrapper[4846]: E1005 07:11:28.325732 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0b472a4ecaa620dd0581763afea4046cc8d15b55265470a90367ed2c9f54e39c\": container with ID starting with 0b472a4ecaa620dd0581763afea4046cc8d15b55265470a90367ed2c9f54e39c not found: ID does not exist" containerID="0b472a4ecaa620dd0581763afea4046cc8d15b55265470a90367ed2c9f54e39c" Oct 05 07:11:28 crc kubenswrapper[4846]: I1005 07:11:28.325763 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0b472a4ecaa620dd0581763afea4046cc8d15b55265470a90367ed2c9f54e39c"} err="failed to get container status \"0b472a4ecaa620dd0581763afea4046cc8d15b55265470a90367ed2c9f54e39c\": rpc error: code = NotFound desc = could not find container \"0b472a4ecaa620dd0581763afea4046cc8d15b55265470a90367ed2c9f54e39c\": container with ID starting with 0b472a4ecaa620dd0581763afea4046cc8d15b55265470a90367ed2c9f54e39c not found: ID does not exist" Oct 05 07:11:28 crc kubenswrapper[4846]: I1005 07:11:28.325784 4846 scope.go:117] "RemoveContainer" containerID="9919ee8b928899c8295fa6d8951cffdd5cffd78c7de5c1e7b1a230ab2550fbf6" Oct 05 07:11:28 crc kubenswrapper[4846]: E1005 07:11:28.326687 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9919ee8b928899c8295fa6d8951cffdd5cffd78c7de5c1e7b1a230ab2550fbf6\": container with ID starting with 9919ee8b928899c8295fa6d8951cffdd5cffd78c7de5c1e7b1a230ab2550fbf6 not found: ID does not exist" containerID="9919ee8b928899c8295fa6d8951cffdd5cffd78c7de5c1e7b1a230ab2550fbf6" Oct 05 07:11:28 crc kubenswrapper[4846]: I1005 07:11:28.326733 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9919ee8b928899c8295fa6d8951cffdd5cffd78c7de5c1e7b1a230ab2550fbf6"} err="failed to get container status \"9919ee8b928899c8295fa6d8951cffdd5cffd78c7de5c1e7b1a230ab2550fbf6\": rpc error: code = NotFound desc = could not find container \"9919ee8b928899c8295fa6d8951cffdd5cffd78c7de5c1e7b1a230ab2550fbf6\": container with ID starting with 9919ee8b928899c8295fa6d8951cffdd5cffd78c7de5c1e7b1a230ab2550fbf6 not found: ID does not exist" Oct 05 07:11:28 crc kubenswrapper[4846]: I1005 07:11:28.331474 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-4kqnw_169918d8-597e-4ef4-b00c-fb14c4d0347f/ovn-controller/0.log" Oct 05 07:11:28 crc kubenswrapper[4846]: I1005 07:11:28.331556 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-4kqnw" Oct 05 07:11:28 crc kubenswrapper[4846]: I1005 07:11:28.372374 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/169918d8-597e-4ef4-b00c-fb14c4d0347f-var-run-ovn\") pod \"169918d8-597e-4ef4-b00c-fb14c4d0347f\" (UID: \"169918d8-597e-4ef4-b00c-fb14c4d0347f\") " Oct 05 07:11:28 crc kubenswrapper[4846]: I1005 07:11:28.372425 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/169918d8-597e-4ef4-b00c-fb14c4d0347f-var-run\") pod \"169918d8-597e-4ef4-b00c-fb14c4d0347f\" (UID: \"169918d8-597e-4ef4-b00c-fb14c4d0347f\") " Oct 05 07:11:28 crc kubenswrapper[4846]: I1005 07:11:28.372469 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/169918d8-597e-4ef4-b00c-fb14c4d0347f-ovn-controller-tls-certs\") pod \"169918d8-597e-4ef4-b00c-fb14c4d0347f\" (UID: \"169918d8-597e-4ef4-b00c-fb14c4d0347f\") " Oct 05 07:11:28 crc kubenswrapper[4846]: I1005 07:11:28.372533 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/169918d8-597e-4ef4-b00c-fb14c4d0347f-var-log-ovn\") pod \"169918d8-597e-4ef4-b00c-fb14c4d0347f\" (UID: \"169918d8-597e-4ef4-b00c-fb14c4d0347f\") " Oct 05 07:11:28 crc kubenswrapper[4846]: I1005 07:11:28.372562 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/169918d8-597e-4ef4-b00c-fb14c4d0347f-combined-ca-bundle\") pod \"169918d8-597e-4ef4-b00c-fb14c4d0347f\" (UID: \"169918d8-597e-4ef4-b00c-fb14c4d0347f\") " Oct 05 07:11:28 crc kubenswrapper[4846]: I1005 07:11:28.372631 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/169918d8-597e-4ef4-b00c-fb14c4d0347f-scripts\") pod \"169918d8-597e-4ef4-b00c-fb14c4d0347f\" (UID: \"169918d8-597e-4ef4-b00c-fb14c4d0347f\") " Oct 05 07:11:28 crc kubenswrapper[4846]: I1005 07:11:28.372669 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hmtzx\" (UniqueName: \"kubernetes.io/projected/169918d8-597e-4ef4-b00c-fb14c4d0347f-kube-api-access-hmtzx\") pod \"169918d8-597e-4ef4-b00c-fb14c4d0347f\" (UID: \"169918d8-597e-4ef4-b00c-fb14c4d0347f\") " Oct 05 07:11:28 crc kubenswrapper[4846]: I1005 07:11:28.374747 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/169918d8-597e-4ef4-b00c-fb14c4d0347f-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "169918d8-597e-4ef4-b00c-fb14c4d0347f" (UID: "169918d8-597e-4ef4-b00c-fb14c4d0347f"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 05 07:11:28 crc kubenswrapper[4846]: I1005 07:11:28.374786 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/169918d8-597e-4ef4-b00c-fb14c4d0347f-var-run" (OuterVolumeSpecName: "var-run") pod "169918d8-597e-4ef4-b00c-fb14c4d0347f" (UID: "169918d8-597e-4ef4-b00c-fb14c4d0347f"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 05 07:11:28 crc kubenswrapper[4846]: I1005 07:11:28.374849 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/169918d8-597e-4ef4-b00c-fb14c4d0347f-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "169918d8-597e-4ef4-b00c-fb14c4d0347f" (UID: "169918d8-597e-4ef4-b00c-fb14c4d0347f"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 05 07:11:28 crc kubenswrapper[4846]: I1005 07:11:28.376003 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/169918d8-597e-4ef4-b00c-fb14c4d0347f-scripts" (OuterVolumeSpecName: "scripts") pod "169918d8-597e-4ef4-b00c-fb14c4d0347f" (UID: "169918d8-597e-4ef4-b00c-fb14c4d0347f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 07:11:28 crc kubenswrapper[4846]: I1005 07:11:28.383448 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/169918d8-597e-4ef4-b00c-fb14c4d0347f-kube-api-access-hmtzx" (OuterVolumeSpecName: "kube-api-access-hmtzx") pod "169918d8-597e-4ef4-b00c-fb14c4d0347f" (UID: "169918d8-597e-4ef4-b00c-fb14c4d0347f"). InnerVolumeSpecName "kube-api-access-hmtzx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:11:28 crc kubenswrapper[4846]: I1005 07:11:28.414253 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/169918d8-597e-4ef4-b00c-fb14c4d0347f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "169918d8-597e-4ef4-b00c-fb14c4d0347f" (UID: "169918d8-597e-4ef4-b00c-fb14c4d0347f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:11:28 crc kubenswrapper[4846]: I1005 07:11:28.418581 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_18a7be37-c8dc-4d8a-808f-2a63eb4d9692/ovn-northd/0.log" Oct 05 07:11:28 crc kubenswrapper[4846]: I1005 07:11:28.418798 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Oct 05 07:11:28 crc kubenswrapper[4846]: I1005 07:11:28.475787 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/18a7be37-c8dc-4d8a-808f-2a63eb4d9692-ovn-rundir\") pod \"18a7be37-c8dc-4d8a-808f-2a63eb4d9692\" (UID: \"18a7be37-c8dc-4d8a-808f-2a63eb4d9692\") " Oct 05 07:11:28 crc kubenswrapper[4846]: I1005 07:11:28.476007 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/18a7be37-c8dc-4d8a-808f-2a63eb4d9692-config\") pod \"18a7be37-c8dc-4d8a-808f-2a63eb4d9692\" (UID: \"18a7be37-c8dc-4d8a-808f-2a63eb4d9692\") " Oct 05 07:11:28 crc kubenswrapper[4846]: I1005 07:11:28.476099 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/18a7be37-c8dc-4d8a-808f-2a63eb4d9692-combined-ca-bundle\") pod \"18a7be37-c8dc-4d8a-808f-2a63eb4d9692\" (UID: \"18a7be37-c8dc-4d8a-808f-2a63eb4d9692\") " Oct 05 07:11:28 crc kubenswrapper[4846]: I1005 07:11:28.476136 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p8kvb\" (UniqueName: \"kubernetes.io/projected/18a7be37-c8dc-4d8a-808f-2a63eb4d9692-kube-api-access-p8kvb\") pod \"18a7be37-c8dc-4d8a-808f-2a63eb4d9692\" (UID: \"18a7be37-c8dc-4d8a-808f-2a63eb4d9692\") " Oct 05 07:11:28 crc kubenswrapper[4846]: I1005 07:11:28.476210 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/18a7be37-c8dc-4d8a-808f-2a63eb4d9692-metrics-certs-tls-certs\") pod \"18a7be37-c8dc-4d8a-808f-2a63eb4d9692\" (UID: \"18a7be37-c8dc-4d8a-808f-2a63eb4d9692\") " Oct 05 07:11:28 crc kubenswrapper[4846]: I1005 07:11:28.476283 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/18a7be37-c8dc-4d8a-808f-2a63eb4d9692-scripts\") pod \"18a7be37-c8dc-4d8a-808f-2a63eb4d9692\" (UID: \"18a7be37-c8dc-4d8a-808f-2a63eb4d9692\") " Oct 05 07:11:28 crc kubenswrapper[4846]: I1005 07:11:28.476443 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/18a7be37-c8dc-4d8a-808f-2a63eb4d9692-ovn-northd-tls-certs\") pod \"18a7be37-c8dc-4d8a-808f-2a63eb4d9692\" (UID: \"18a7be37-c8dc-4d8a-808f-2a63eb4d9692\") " Oct 05 07:11:28 crc kubenswrapper[4846]: I1005 07:11:28.477164 4846 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/169918d8-597e-4ef4-b00c-fb14c4d0347f-var-run-ovn\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:28 crc kubenswrapper[4846]: I1005 07:11:28.477214 4846 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/169918d8-597e-4ef4-b00c-fb14c4d0347f-var-run\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:28 crc kubenswrapper[4846]: I1005 07:11:28.477225 4846 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/169918d8-597e-4ef4-b00c-fb14c4d0347f-var-log-ovn\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:28 crc kubenswrapper[4846]: I1005 07:11:28.477235 4846 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/169918d8-597e-4ef4-b00c-fb14c4d0347f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:28 crc kubenswrapper[4846]: I1005 07:11:28.477245 4846 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/169918d8-597e-4ef4-b00c-fb14c4d0347f-scripts\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:28 crc kubenswrapper[4846]: I1005 07:11:28.477254 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hmtzx\" (UniqueName: \"kubernetes.io/projected/169918d8-597e-4ef4-b00c-fb14c4d0347f-kube-api-access-hmtzx\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:28 crc kubenswrapper[4846]: I1005 07:11:28.478172 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/18a7be37-c8dc-4d8a-808f-2a63eb4d9692-ovn-rundir" (OuterVolumeSpecName: "ovn-rundir") pod "18a7be37-c8dc-4d8a-808f-2a63eb4d9692" (UID: "18a7be37-c8dc-4d8a-808f-2a63eb4d9692"). InnerVolumeSpecName "ovn-rundir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 07:11:28 crc kubenswrapper[4846]: I1005 07:11:28.478519 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/18a7be37-c8dc-4d8a-808f-2a63eb4d9692-scripts" (OuterVolumeSpecName: "scripts") pod "18a7be37-c8dc-4d8a-808f-2a63eb4d9692" (UID: "18a7be37-c8dc-4d8a-808f-2a63eb4d9692"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 07:11:28 crc kubenswrapper[4846]: I1005 07:11:28.478674 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/18a7be37-c8dc-4d8a-808f-2a63eb4d9692-config" (OuterVolumeSpecName: "config") pod "18a7be37-c8dc-4d8a-808f-2a63eb4d9692" (UID: "18a7be37-c8dc-4d8a-808f-2a63eb4d9692"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 07:11:28 crc kubenswrapper[4846]: I1005 07:11:28.503278 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/18a7be37-c8dc-4d8a-808f-2a63eb4d9692-kube-api-access-p8kvb" (OuterVolumeSpecName: "kube-api-access-p8kvb") pod "18a7be37-c8dc-4d8a-808f-2a63eb4d9692" (UID: "18a7be37-c8dc-4d8a-808f-2a63eb4d9692"). InnerVolumeSpecName "kube-api-access-p8kvb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:11:28 crc kubenswrapper[4846]: I1005 07:11:28.506704 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/18a7be37-c8dc-4d8a-808f-2a63eb4d9692-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "18a7be37-c8dc-4d8a-808f-2a63eb4d9692" (UID: "18a7be37-c8dc-4d8a-808f-2a63eb4d9692"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:11:28 crc kubenswrapper[4846]: I1005 07:11:28.525198 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="35261dbc-cabe-4bf0-88f6-b8613ddec0f1" path="/var/lib/kubelet/pods/35261dbc-cabe-4bf0-88f6-b8613ddec0f1/volumes" Oct 05 07:11:28 crc kubenswrapper[4846]: I1005 07:11:28.526332 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4d3616b4-9f2c-4c62-9760-0300b02d0dbb" path="/var/lib/kubelet/pods/4d3616b4-9f2c-4c62-9760-0300b02d0dbb/volumes" Oct 05 07:11:28 crc kubenswrapper[4846]: I1005 07:11:28.526912 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57b09608-08e5-4998-a3b3-e88b13587f58" path="/var/lib/kubelet/pods/57b09608-08e5-4998-a3b3-e88b13587f58/volumes" Oct 05 07:11:28 crc kubenswrapper[4846]: I1005 07:11:28.528300 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="626e6bce-7422-4fbd-b978-dfa5cb93de34" path="/var/lib/kubelet/pods/626e6bce-7422-4fbd-b978-dfa5cb93de34/volumes" Oct 05 07:11:28 crc kubenswrapper[4846]: I1005 07:11:28.529022 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="751734cb-38c8-46a0-9bac-05bcf24cc589" path="/var/lib/kubelet/pods/751734cb-38c8-46a0-9bac-05bcf24cc589/volumes" Oct 05 07:11:28 crc kubenswrapper[4846]: I1005 07:11:28.529493 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c3e9c65e-42bf-4cd6-a47d-b6ef0db5160e" path="/var/lib/kubelet/pods/c3e9c65e-42bf-4cd6-a47d-b6ef0db5160e/volumes" Oct 05 07:11:28 crc kubenswrapper[4846]: I1005 07:11:28.541393 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/169918d8-597e-4ef4-b00c-fb14c4d0347f-ovn-controller-tls-certs" (OuterVolumeSpecName: "ovn-controller-tls-certs") pod "169918d8-597e-4ef4-b00c-fb14c4d0347f" (UID: "169918d8-597e-4ef4-b00c-fb14c4d0347f"). InnerVolumeSpecName "ovn-controller-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:11:28 crc kubenswrapper[4846]: I1005 07:11:28.560142 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/18a7be37-c8dc-4d8a-808f-2a63eb4d9692-metrics-certs-tls-certs" (OuterVolumeSpecName: "metrics-certs-tls-certs") pod "18a7be37-c8dc-4d8a-808f-2a63eb4d9692" (UID: "18a7be37-c8dc-4d8a-808f-2a63eb4d9692"). InnerVolumeSpecName "metrics-certs-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:11:28 crc kubenswrapper[4846]: I1005 07:11:28.580448 4846 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/18a7be37-c8dc-4d8a-808f-2a63eb4d9692-config\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:28 crc kubenswrapper[4846]: I1005 07:11:28.580695 4846 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/18a7be37-c8dc-4d8a-808f-2a63eb4d9692-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:28 crc kubenswrapper[4846]: I1005 07:11:28.580789 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p8kvb\" (UniqueName: \"kubernetes.io/projected/18a7be37-c8dc-4d8a-808f-2a63eb4d9692-kube-api-access-p8kvb\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:28 crc kubenswrapper[4846]: I1005 07:11:28.580847 4846 reconciler_common.go:293] "Volume detached for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/169918d8-597e-4ef4-b00c-fb14c4d0347f-ovn-controller-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:28 crc kubenswrapper[4846]: I1005 07:11:28.580900 4846 reconciler_common.go:293] "Volume detached for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/18a7be37-c8dc-4d8a-808f-2a63eb4d9692-metrics-certs-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:28 crc kubenswrapper[4846]: I1005 07:11:28.580962 4846 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/18a7be37-c8dc-4d8a-808f-2a63eb4d9692-scripts\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:28 crc kubenswrapper[4846]: I1005 07:11:28.581016 4846 reconciler_common.go:293] "Volume detached for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/18a7be37-c8dc-4d8a-808f-2a63eb4d9692-ovn-rundir\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:28 crc kubenswrapper[4846]: I1005 07:11:28.594772 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/18a7be37-c8dc-4d8a-808f-2a63eb4d9692-ovn-northd-tls-certs" (OuterVolumeSpecName: "ovn-northd-tls-certs") pod "18a7be37-c8dc-4d8a-808f-2a63eb4d9692" (UID: "18a7be37-c8dc-4d8a-808f-2a63eb4d9692"). InnerVolumeSpecName "ovn-northd-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:11:28 crc kubenswrapper[4846]: I1005 07:11:28.684643 4846 reconciler_common.go:293] "Volume detached for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/18a7be37-c8dc-4d8a-808f-2a63eb4d9692-ovn-northd-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:28 crc kubenswrapper[4846]: I1005 07:11:28.823795 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-78995b9f79-nwl7v" Oct 05 07:11:28 crc kubenswrapper[4846]: I1005 07:11:28.840664 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Oct 05 07:11:28 crc kubenswrapper[4846]: I1005 07:11:28.894134 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3995994c-2a3f-400f-9a64-b564759b20b2-config-data-custom\") pod \"3995994c-2a3f-400f-9a64-b564759b20b2\" (UID: \"3995994c-2a3f-400f-9a64-b564759b20b2\") " Oct 05 07:11:28 crc kubenswrapper[4846]: I1005 07:11:28.894258 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/d88c48f9-948e-4888-986e-e00163ec93a6-config-data-generated\") pod \"d88c48f9-948e-4888-986e-e00163ec93a6\" (UID: \"d88c48f9-948e-4888-986e-e00163ec93a6\") " Oct 05 07:11:28 crc kubenswrapper[4846]: I1005 07:11:28.895898 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d88c48f9-948e-4888-986e-e00163ec93a6-config-data-generated" (OuterVolumeSpecName: "config-data-generated") pod "d88c48f9-948e-4888-986e-e00163ec93a6" (UID: "d88c48f9-948e-4888-986e-e00163ec93a6"). InnerVolumeSpecName "config-data-generated". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 07:11:28 crc kubenswrapper[4846]: I1005 07:11:28.898307 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-77664b8dfd-5hkqj" Oct 05 07:11:28 crc kubenswrapper[4846]: I1005 07:11:28.909733 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-85bbbcfb7f-kwfjt" Oct 05 07:11:28 crc kubenswrapper[4846]: I1005 07:11:28.914399 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/d88c48f9-948e-4888-986e-e00163ec93a6-secrets\") pod \"d88c48f9-948e-4888-986e-e00163ec93a6\" (UID: \"d88c48f9-948e-4888-986e-e00163ec93a6\") " Oct 05 07:11:28 crc kubenswrapper[4846]: I1005 07:11:28.914446 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f55kq\" (UniqueName: \"kubernetes.io/projected/3995994c-2a3f-400f-9a64-b564759b20b2-kube-api-access-f55kq\") pod \"3995994c-2a3f-400f-9a64-b564759b20b2\" (UID: \"3995994c-2a3f-400f-9a64-b564759b20b2\") " Oct 05 07:11:28 crc kubenswrapper[4846]: I1005 07:11:28.914510 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3995994c-2a3f-400f-9a64-b564759b20b2-combined-ca-bundle\") pod \"3995994c-2a3f-400f-9a64-b564759b20b2\" (UID: \"3995994c-2a3f-400f-9a64-b564759b20b2\") " Oct 05 07:11:28 crc kubenswrapper[4846]: I1005 07:11:28.914559 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pvxnh\" (UniqueName: \"kubernetes.io/projected/d88c48f9-948e-4888-986e-e00163ec93a6-kube-api-access-pvxnh\") pod \"d88c48f9-948e-4888-986e-e00163ec93a6\" (UID: \"d88c48f9-948e-4888-986e-e00163ec93a6\") " Oct 05 07:11:28 crc kubenswrapper[4846]: I1005 07:11:28.914592 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3995994c-2a3f-400f-9a64-b564759b20b2-config-data\") pod \"3995994c-2a3f-400f-9a64-b564759b20b2\" (UID: \"3995994c-2a3f-400f-9a64-b564759b20b2\") " Oct 05 07:11:28 crc kubenswrapper[4846]: I1005 07:11:28.914666 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d88c48f9-948e-4888-986e-e00163ec93a6-combined-ca-bundle\") pod \"d88c48f9-948e-4888-986e-e00163ec93a6\" (UID: \"d88c48f9-948e-4888-986e-e00163ec93a6\") " Oct 05 07:11:28 crc kubenswrapper[4846]: I1005 07:11:28.914692 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d88c48f9-948e-4888-986e-e00163ec93a6-operator-scripts\") pod \"d88c48f9-948e-4888-986e-e00163ec93a6\" (UID: \"d88c48f9-948e-4888-986e-e00163ec93a6\") " Oct 05 07:11:28 crc kubenswrapper[4846]: I1005 07:11:28.914739 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/d88c48f9-948e-4888-986e-e00163ec93a6-config-data-default\") pod \"d88c48f9-948e-4888-986e-e00163ec93a6\" (UID: \"d88c48f9-948e-4888-986e-e00163ec93a6\") " Oct 05 07:11:28 crc kubenswrapper[4846]: I1005 07:11:28.914762 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/d88c48f9-948e-4888-986e-e00163ec93a6-galera-tls-certs\") pod \"d88c48f9-948e-4888-986e-e00163ec93a6\" (UID: \"d88c48f9-948e-4888-986e-e00163ec93a6\") " Oct 05 07:11:28 crc kubenswrapper[4846]: I1005 07:11:28.914787 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mysql-db\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"d88c48f9-948e-4888-986e-e00163ec93a6\" (UID: \"d88c48f9-948e-4888-986e-e00163ec93a6\") " Oct 05 07:11:28 crc kubenswrapper[4846]: I1005 07:11:28.914818 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/d88c48f9-948e-4888-986e-e00163ec93a6-kolla-config\") pod \"d88c48f9-948e-4888-986e-e00163ec93a6\" (UID: \"d88c48f9-948e-4888-986e-e00163ec93a6\") " Oct 05 07:11:28 crc kubenswrapper[4846]: I1005 07:11:28.914870 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3995994c-2a3f-400f-9a64-b564759b20b2-logs\") pod \"3995994c-2a3f-400f-9a64-b564759b20b2\" (UID: \"3995994c-2a3f-400f-9a64-b564759b20b2\") " Oct 05 07:11:28 crc kubenswrapper[4846]: I1005 07:11:28.920454 4846 reconciler_common.go:293] "Volume detached for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/d88c48f9-948e-4888-986e-e00163ec93a6-config-data-generated\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:28 crc kubenswrapper[4846]: I1005 07:11:28.920875 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3995994c-2a3f-400f-9a64-b564759b20b2-logs" (OuterVolumeSpecName: "logs") pod "3995994c-2a3f-400f-9a64-b564759b20b2" (UID: "3995994c-2a3f-400f-9a64-b564759b20b2"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 07:11:28 crc kubenswrapper[4846]: I1005 07:11:28.923307 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d88c48f9-948e-4888-986e-e00163ec93a6-kolla-config" (OuterVolumeSpecName: "kolla-config") pod "d88c48f9-948e-4888-986e-e00163ec93a6" (UID: "d88c48f9-948e-4888-986e-e00163ec93a6"). InnerVolumeSpecName "kolla-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 07:11:28 crc kubenswrapper[4846]: I1005 07:11:28.924244 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d88c48f9-948e-4888-986e-e00163ec93a6-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "d88c48f9-948e-4888-986e-e00163ec93a6" (UID: "d88c48f9-948e-4888-986e-e00163ec93a6"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 07:11:28 crc kubenswrapper[4846]: I1005 07:11:28.927268 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d88c48f9-948e-4888-986e-e00163ec93a6-config-data-default" (OuterVolumeSpecName: "config-data-default") pod "d88c48f9-948e-4888-986e-e00163ec93a6" (UID: "d88c48f9-948e-4888-986e-e00163ec93a6"). InnerVolumeSpecName "config-data-default". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 07:11:28 crc kubenswrapper[4846]: I1005 07:11:28.927478 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3995994c-2a3f-400f-9a64-b564759b20b2-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "3995994c-2a3f-400f-9a64-b564759b20b2" (UID: "3995994c-2a3f-400f-9a64-b564759b20b2"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:11:28 crc kubenswrapper[4846]: I1005 07:11:28.935758 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d88c48f9-948e-4888-986e-e00163ec93a6-secrets" (OuterVolumeSpecName: "secrets") pod "d88c48f9-948e-4888-986e-e00163ec93a6" (UID: "d88c48f9-948e-4888-986e-e00163ec93a6"). InnerVolumeSpecName "secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:11:28 crc kubenswrapper[4846]: I1005 07:11:28.935973 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d88c48f9-948e-4888-986e-e00163ec93a6-kube-api-access-pvxnh" (OuterVolumeSpecName: "kube-api-access-pvxnh") pod "d88c48f9-948e-4888-986e-e00163ec93a6" (UID: "d88c48f9-948e-4888-986e-e00163ec93a6"). InnerVolumeSpecName "kube-api-access-pvxnh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:11:28 crc kubenswrapper[4846]: I1005 07:11:28.939555 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3995994c-2a3f-400f-9a64-b564759b20b2-kube-api-access-f55kq" (OuterVolumeSpecName: "kube-api-access-f55kq") pod "3995994c-2a3f-400f-9a64-b564759b20b2" (UID: "3995994c-2a3f-400f-9a64-b564759b20b2"). InnerVolumeSpecName "kube-api-access-f55kq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:11:28 crc kubenswrapper[4846]: I1005 07:11:28.973499 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage01-crc" (OuterVolumeSpecName: "mysql-db") pod "d88c48f9-948e-4888-986e-e00163ec93a6" (UID: "d88c48f9-948e-4888-986e-e00163ec93a6"). InnerVolumeSpecName "local-storage01-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 05 07:11:28 crc kubenswrapper[4846]: I1005 07:11:28.975293 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"d88c48f9-948e-4888-986e-e00163ec93a6","Type":"ContainerDied","Data":"56b8f764086c61683708e28e3f2929710527aada412eda06ddb4b14e487478c3"} Oct 05 07:11:28 crc kubenswrapper[4846]: I1005 07:11:28.975357 4846 scope.go:117] "RemoveContainer" containerID="56b8f764086c61683708e28e3f2929710527aada412eda06ddb4b14e487478c3" Oct 05 07:11:28 crc kubenswrapper[4846]: I1005 07:11:28.975561 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Oct 05 07:11:28 crc kubenswrapper[4846]: I1005 07:11:28.975269 4846 generic.go:334] "Generic (PLEG): container finished" podID="d88c48f9-948e-4888-986e-e00163ec93a6" containerID="56b8f764086c61683708e28e3f2929710527aada412eda06ddb4b14e487478c3" exitCode=0 Oct 05 07:11:28 crc kubenswrapper[4846]: I1005 07:11:28.975709 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"d88c48f9-948e-4888-986e-e00163ec93a6","Type":"ContainerDied","Data":"6bc8133968b75fac499facde615345124fd87e14184b7ee6c4a0b3c0f8eb0806"} Oct 05 07:11:28 crc kubenswrapper[4846]: I1005 07:11:28.992056 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3995994c-2a3f-400f-9a64-b564759b20b2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3995994c-2a3f-400f-9a64-b564759b20b2" (UID: "3995994c-2a3f-400f-9a64-b564759b20b2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:11:28 crc kubenswrapper[4846]: I1005 07:11:28.996215 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-4kqnw_169918d8-597e-4ef4-b00c-fb14c4d0347f/ovn-controller/0.log" Oct 05 07:11:28 crc kubenswrapper[4846]: I1005 07:11:28.996335 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-4kqnw" event={"ID":"169918d8-597e-4ef4-b00c-fb14c4d0347f","Type":"ContainerDied","Data":"404742809b09f98d85eb545d8420b8a5096b5cbf77f103b9390023f16993c807"} Oct 05 07:11:28 crc kubenswrapper[4846]: I1005 07:11:28.996455 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-4kqnw" Oct 05 07:11:28 crc kubenswrapper[4846]: I1005 07:11:28.997384 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d88c48f9-948e-4888-986e-e00163ec93a6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d88c48f9-948e-4888-986e-e00163ec93a6" (UID: "d88c48f9-948e-4888-986e-e00163ec93a6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:11:29 crc kubenswrapper[4846]: I1005 07:11:29.010069 4846 generic.go:334] "Generic (PLEG): container finished" podID="a72c0cf5-8cd9-4f5f-8a2b-25bf41ad0638" containerID="57243f61674b9f96faaeba668f666e7518c0f44adbe9b6c573a200820bbb58d3" exitCode=0 Oct 05 07:11:29 crc kubenswrapper[4846]: I1005 07:11:29.010308 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-85bbbcfb7f-kwfjt" Oct 05 07:11:29 crc kubenswrapper[4846]: I1005 07:11:29.011064 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-85bbbcfb7f-kwfjt" event={"ID":"a72c0cf5-8cd9-4f5f-8a2b-25bf41ad0638","Type":"ContainerDied","Data":"57243f61674b9f96faaeba668f666e7518c0f44adbe9b6c573a200820bbb58d3"} Oct 05 07:11:29 crc kubenswrapper[4846]: I1005 07:11:29.011201 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-85bbbcfb7f-kwfjt" event={"ID":"a72c0cf5-8cd9-4f5f-8a2b-25bf41ad0638","Type":"ContainerDied","Data":"f1cd9e6d2dcee73cc1f74d34d80432272e9c18c2ccc5fa1800a62740e3a25bb5"} Oct 05 07:11:29 crc kubenswrapper[4846]: I1005 07:11:29.016298 4846 generic.go:334] "Generic (PLEG): container finished" podID="8d55f55a-2d79-4c01-8b0f-d67cefb842ff" containerID="6a990887fb132123ab2aa7e7c062e5fbecb25e29a746e6d59d303ed6848ccde0" exitCode=0 Oct 05 07:11:29 crc kubenswrapper[4846]: I1005 07:11:29.016456 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-77664b8dfd-5hkqj" Oct 05 07:11:29 crc kubenswrapper[4846]: I1005 07:11:29.016515 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-77664b8dfd-5hkqj" event={"ID":"8d55f55a-2d79-4c01-8b0f-d67cefb842ff","Type":"ContainerDied","Data":"6a990887fb132123ab2aa7e7c062e5fbecb25e29a746e6d59d303ed6848ccde0"} Oct 05 07:11:29 crc kubenswrapper[4846]: I1005 07:11:29.016631 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Oct 05 07:11:29 crc kubenswrapper[4846]: I1005 07:11:29.016659 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-77664b8dfd-5hkqj" event={"ID":"8d55f55a-2d79-4c01-8b0f-d67cefb842ff","Type":"ContainerDied","Data":"3ddbf2a790be1dd9add94290c14ebc38db8686c6dd45f7665a169e191b7fa9a3"} Oct 05 07:11:29 crc kubenswrapper[4846]: I1005 07:11:29.019278 4846 generic.go:334] "Generic (PLEG): container finished" podID="98b50269-c6d4-42f2-8cef-afc127bfccb1" containerID="055332eb57f5e6e889d548a1b4717caea296e84dc70616d5bdb53dd34aaa0697" exitCode=0 Oct 05 07:11:29 crc kubenswrapper[4846]: I1005 07:11:29.019358 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"98b50269-c6d4-42f2-8cef-afc127bfccb1","Type":"ContainerDied","Data":"055332eb57f5e6e889d548a1b4717caea296e84dc70616d5bdb53dd34aaa0697"} Oct 05 07:11:29 crc kubenswrapper[4846]: I1005 07:11:29.021153 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a72c0cf5-8cd9-4f5f-8a2b-25bf41ad0638-public-tls-certs\") pod \"a72c0cf5-8cd9-4f5f-8a2b-25bf41ad0638\" (UID: \"a72c0cf5-8cd9-4f5f-8a2b-25bf41ad0638\") " Oct 05 07:11:29 crc kubenswrapper[4846]: I1005 07:11:29.021254 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8vlxz\" (UniqueName: \"kubernetes.io/projected/a72c0cf5-8cd9-4f5f-8a2b-25bf41ad0638-kube-api-access-8vlxz\") pod \"a72c0cf5-8cd9-4f5f-8a2b-25bf41ad0638\" (UID: \"a72c0cf5-8cd9-4f5f-8a2b-25bf41ad0638\") " Oct 05 07:11:29 crc kubenswrapper[4846]: I1005 07:11:29.021303 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/a72c0cf5-8cd9-4f5f-8a2b-25bf41ad0638-httpd-config\") pod \"a72c0cf5-8cd9-4f5f-8a2b-25bf41ad0638\" (UID: \"a72c0cf5-8cd9-4f5f-8a2b-25bf41ad0638\") " Oct 05 07:11:29 crc kubenswrapper[4846]: I1005 07:11:29.021347 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8d55f55a-2d79-4c01-8b0f-d67cefb842ff-logs\") pod \"8d55f55a-2d79-4c01-8b0f-d67cefb842ff\" (UID: \"8d55f55a-2d79-4c01-8b0f-d67cefb842ff\") " Oct 05 07:11:29 crc kubenswrapper[4846]: I1005 07:11:29.021381 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8d55f55a-2d79-4c01-8b0f-d67cefb842ff-combined-ca-bundle\") pod \"8d55f55a-2d79-4c01-8b0f-d67cefb842ff\" (UID: \"8d55f55a-2d79-4c01-8b0f-d67cefb842ff\") " Oct 05 07:11:29 crc kubenswrapper[4846]: I1005 07:11:29.021414 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a72c0cf5-8cd9-4f5f-8a2b-25bf41ad0638-combined-ca-bundle\") pod \"a72c0cf5-8cd9-4f5f-8a2b-25bf41ad0638\" (UID: \"a72c0cf5-8cd9-4f5f-8a2b-25bf41ad0638\") " Oct 05 07:11:29 crc kubenswrapper[4846]: I1005 07:11:29.021458 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/a72c0cf5-8cd9-4f5f-8a2b-25bf41ad0638-ovndb-tls-certs\") pod \"a72c0cf5-8cd9-4f5f-8a2b-25bf41ad0638\" (UID: \"a72c0cf5-8cd9-4f5f-8a2b-25bf41ad0638\") " Oct 05 07:11:29 crc kubenswrapper[4846]: I1005 07:11:29.021496 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8d55f55a-2d79-4c01-8b0f-d67cefb842ff-config-data\") pod \"8d55f55a-2d79-4c01-8b0f-d67cefb842ff\" (UID: \"8d55f55a-2d79-4c01-8b0f-d67cefb842ff\") " Oct 05 07:11:29 crc kubenswrapper[4846]: I1005 07:11:29.021523 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a72c0cf5-8cd9-4f5f-8a2b-25bf41ad0638-internal-tls-certs\") pod \"a72c0cf5-8cd9-4f5f-8a2b-25bf41ad0638\" (UID: \"a72c0cf5-8cd9-4f5f-8a2b-25bf41ad0638\") " Oct 05 07:11:29 crc kubenswrapper[4846]: I1005 07:11:29.021543 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/a72c0cf5-8cd9-4f5f-8a2b-25bf41ad0638-config\") pod \"a72c0cf5-8cd9-4f5f-8a2b-25bf41ad0638\" (UID: \"a72c0cf5-8cd9-4f5f-8a2b-25bf41ad0638\") " Oct 05 07:11:29 crc kubenswrapper[4846]: I1005 07:11:29.021558 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8d55f55a-2d79-4c01-8b0f-d67cefb842ff-config-data-custom\") pod \"8d55f55a-2d79-4c01-8b0f-d67cefb842ff\" (UID: \"8d55f55a-2d79-4c01-8b0f-d67cefb842ff\") " Oct 05 07:11:29 crc kubenswrapper[4846]: I1005 07:11:29.021615 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8n2gd\" (UniqueName: \"kubernetes.io/projected/8d55f55a-2d79-4c01-8b0f-d67cefb842ff-kube-api-access-8n2gd\") pod \"8d55f55a-2d79-4c01-8b0f-d67cefb842ff\" (UID: \"8d55f55a-2d79-4c01-8b0f-d67cefb842ff\") " Oct 05 07:11:29 crc kubenswrapper[4846]: I1005 07:11:29.021884 4846 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d88c48f9-948e-4888-986e-e00163ec93a6-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:29 crc kubenswrapper[4846]: I1005 07:11:29.022351 4846 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d88c48f9-948e-4888-986e-e00163ec93a6-operator-scripts\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:29 crc kubenswrapper[4846]: I1005 07:11:29.022366 4846 reconciler_common.go:293] "Volume detached for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/d88c48f9-948e-4888-986e-e00163ec93a6-config-data-default\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:29 crc kubenswrapper[4846]: I1005 07:11:29.022397 4846 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" " Oct 05 07:11:29 crc kubenswrapper[4846]: I1005 07:11:29.022408 4846 reconciler_common.go:293] "Volume detached for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/d88c48f9-948e-4888-986e-e00163ec93a6-kolla-config\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:29 crc kubenswrapper[4846]: I1005 07:11:29.022420 4846 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3995994c-2a3f-400f-9a64-b564759b20b2-logs\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:29 crc kubenswrapper[4846]: I1005 07:11:29.022607 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8d55f55a-2d79-4c01-8b0f-d67cefb842ff-logs" (OuterVolumeSpecName: "logs") pod "8d55f55a-2d79-4c01-8b0f-d67cefb842ff" (UID: "8d55f55a-2d79-4c01-8b0f-d67cefb842ff"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 07:11:29 crc kubenswrapper[4846]: I1005 07:11:29.025549 4846 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3995994c-2a3f-400f-9a64-b564759b20b2-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:29 crc kubenswrapper[4846]: I1005 07:11:29.025578 4846 reconciler_common.go:293] "Volume detached for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/d88c48f9-948e-4888-986e-e00163ec93a6-secrets\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:29 crc kubenswrapper[4846]: I1005 07:11:29.025589 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f55kq\" (UniqueName: \"kubernetes.io/projected/3995994c-2a3f-400f-9a64-b564759b20b2-kube-api-access-f55kq\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:29 crc kubenswrapper[4846]: I1005 07:11:29.025605 4846 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3995994c-2a3f-400f-9a64-b564759b20b2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:29 crc kubenswrapper[4846]: I1005 07:11:29.025614 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pvxnh\" (UniqueName: \"kubernetes.io/projected/d88c48f9-948e-4888-986e-e00163ec93a6-kube-api-access-pvxnh\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:29 crc kubenswrapper[4846]: I1005 07:11:29.025928 4846 scope.go:117] "RemoveContainer" containerID="bb847394eafe6318ca0e44440040ff96d7a8d7be64f605561ce5e3a0d06b6195" Oct 05 07:11:29 crc kubenswrapper[4846]: I1005 07:11:29.026467 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a72c0cf5-8cd9-4f5f-8a2b-25bf41ad0638-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "a72c0cf5-8cd9-4f5f-8a2b-25bf41ad0638" (UID: "a72c0cf5-8cd9-4f5f-8a2b-25bf41ad0638"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:11:29 crc kubenswrapper[4846]: I1005 07:11:29.027450 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a72c0cf5-8cd9-4f5f-8a2b-25bf41ad0638-kube-api-access-8vlxz" (OuterVolumeSpecName: "kube-api-access-8vlxz") pod "a72c0cf5-8cd9-4f5f-8a2b-25bf41ad0638" (UID: "a72c0cf5-8cd9-4f5f-8a2b-25bf41ad0638"). InnerVolumeSpecName "kube-api-access-8vlxz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:11:29 crc kubenswrapper[4846]: I1005 07:11:29.028154 4846 generic.go:334] "Generic (PLEG): container finished" podID="3995994c-2a3f-400f-9a64-b564759b20b2" containerID="5cae68aeaeb33c53f345bf863894fe1192fa80a8ec2b486a6c2d29e83dec72af" exitCode=0 Oct 05 07:11:29 crc kubenswrapper[4846]: I1005 07:11:29.028287 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-78995b9f79-nwl7v" event={"ID":"3995994c-2a3f-400f-9a64-b564759b20b2","Type":"ContainerDied","Data":"5cae68aeaeb33c53f345bf863894fe1192fa80a8ec2b486a6c2d29e83dec72af"} Oct 05 07:11:29 crc kubenswrapper[4846]: I1005 07:11:29.028327 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-78995b9f79-nwl7v" event={"ID":"3995994c-2a3f-400f-9a64-b564759b20b2","Type":"ContainerDied","Data":"5f1d95b021777d90a46ec2e78aa381055bb3200c115fbcc6b0575e04c4a2dae3"} Oct 05 07:11:29 crc kubenswrapper[4846]: I1005 07:11:29.028581 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-78995b9f79-nwl7v" Oct 05 07:11:29 crc kubenswrapper[4846]: I1005 07:11:29.049311 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8d55f55a-2d79-4c01-8b0f-d67cefb842ff-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "8d55f55a-2d79-4c01-8b0f-d67cefb842ff" (UID: "8d55f55a-2d79-4c01-8b0f-d67cefb842ff"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:11:29 crc kubenswrapper[4846]: I1005 07:11:29.052237 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8d55f55a-2d79-4c01-8b0f-d67cefb842ff-kube-api-access-8n2gd" (OuterVolumeSpecName: "kube-api-access-8n2gd") pod "8d55f55a-2d79-4c01-8b0f-d67cefb842ff" (UID: "8d55f55a-2d79-4c01-8b0f-d67cefb842ff"). InnerVolumeSpecName "kube-api-access-8n2gd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:11:29 crc kubenswrapper[4846]: I1005 07:11:29.053328 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3995994c-2a3f-400f-9a64-b564759b20b2-config-data" (OuterVolumeSpecName: "config-data") pod "3995994c-2a3f-400f-9a64-b564759b20b2" (UID: "3995994c-2a3f-400f-9a64-b564759b20b2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:11:29 crc kubenswrapper[4846]: I1005 07:11:29.053583 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_18a7be37-c8dc-4d8a-808f-2a63eb4d9692/ovn-northd/0.log" Oct 05 07:11:29 crc kubenswrapper[4846]: I1005 07:11:29.053669 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"18a7be37-c8dc-4d8a-808f-2a63eb4d9692","Type":"ContainerDied","Data":"4c157e28a99b6e1bcdc1f682bb2c7be41ccc019fcb6d08ba1de9c45e47d39e16"} Oct 05 07:11:29 crc kubenswrapper[4846]: I1005 07:11:29.053773 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Oct 05 07:11:29 crc kubenswrapper[4846]: I1005 07:11:29.083396 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-4kqnw"] Oct 05 07:11:29 crc kubenswrapper[4846]: I1005 07:11:29.084963 4846 generic.go:334] "Generic (PLEG): container finished" podID="1fa90810-c620-48c2-a49c-762341889743" containerID="26da8c60efda49d0ad9e38f18f39b3b3c7a75ffb60bb839d759345619a5d337c" exitCode=0 Oct 05 07:11:29 crc kubenswrapper[4846]: I1005 07:11:29.085021 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-79f67599fc-2sk76" event={"ID":"1fa90810-c620-48c2-a49c-762341889743","Type":"ContainerDied","Data":"26da8c60efda49d0ad9e38f18f39b3b3c7a75ffb60bb839d759345619a5d337c"} Oct 05 07:11:29 crc kubenswrapper[4846]: I1005 07:11:29.091687 4846 scope.go:117] "RemoveContainer" containerID="56b8f764086c61683708e28e3f2929710527aada412eda06ddb4b14e487478c3" Oct 05 07:11:29 crc kubenswrapper[4846]: I1005 07:11:29.097375 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d88c48f9-948e-4888-986e-e00163ec93a6-galera-tls-certs" (OuterVolumeSpecName: "galera-tls-certs") pod "d88c48f9-948e-4888-986e-e00163ec93a6" (UID: "d88c48f9-948e-4888-986e-e00163ec93a6"). InnerVolumeSpecName "galera-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:11:29 crc kubenswrapper[4846]: I1005 07:11:29.102133 4846 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage01-crc" (UniqueName: "kubernetes.io/local-volume/local-storage01-crc") on node "crc" Oct 05 07:11:29 crc kubenswrapper[4846]: E1005 07:11:29.102453 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"56b8f764086c61683708e28e3f2929710527aada412eda06ddb4b14e487478c3\": container with ID starting with 56b8f764086c61683708e28e3f2929710527aada412eda06ddb4b14e487478c3 not found: ID does not exist" containerID="56b8f764086c61683708e28e3f2929710527aada412eda06ddb4b14e487478c3" Oct 05 07:11:29 crc kubenswrapper[4846]: I1005 07:11:29.102571 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"56b8f764086c61683708e28e3f2929710527aada412eda06ddb4b14e487478c3"} err="failed to get container status \"56b8f764086c61683708e28e3f2929710527aada412eda06ddb4b14e487478c3\": rpc error: code = NotFound desc = could not find container \"56b8f764086c61683708e28e3f2929710527aada412eda06ddb4b14e487478c3\": container with ID starting with 56b8f764086c61683708e28e3f2929710527aada412eda06ddb4b14e487478c3 not found: ID does not exist" Oct 05 07:11:29 crc kubenswrapper[4846]: I1005 07:11:29.102746 4846 scope.go:117] "RemoveContainer" containerID="bb847394eafe6318ca0e44440040ff96d7a8d7be64f605561ce5e3a0d06b6195" Oct 05 07:11:29 crc kubenswrapper[4846]: E1005 07:11:29.103468 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bb847394eafe6318ca0e44440040ff96d7a8d7be64f605561ce5e3a0d06b6195\": container with ID starting with bb847394eafe6318ca0e44440040ff96d7a8d7be64f605561ce5e3a0d06b6195 not found: ID does not exist" containerID="bb847394eafe6318ca0e44440040ff96d7a8d7be64f605561ce5e3a0d06b6195" Oct 05 07:11:29 crc kubenswrapper[4846]: I1005 07:11:29.103496 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bb847394eafe6318ca0e44440040ff96d7a8d7be64f605561ce5e3a0d06b6195"} err="failed to get container status \"bb847394eafe6318ca0e44440040ff96d7a8d7be64f605561ce5e3a0d06b6195\": rpc error: code = NotFound desc = could not find container \"bb847394eafe6318ca0e44440040ff96d7a8d7be64f605561ce5e3a0d06b6195\": container with ID starting with bb847394eafe6318ca0e44440040ff96d7a8d7be64f605561ce5e3a0d06b6195 not found: ID does not exist" Oct 05 07:11:29 crc kubenswrapper[4846]: I1005 07:11:29.103510 4846 scope.go:117] "RemoveContainer" containerID="d171365315f951cc3fe3c42e399c18ed39c3027267575259a8e68fb2d38283e1" Oct 05 07:11:29 crc kubenswrapper[4846]: I1005 07:11:29.107917 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-4kqnw"] Oct 05 07:11:29 crc kubenswrapper[4846]: I1005 07:11:29.110352 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-79f67599fc-2sk76" Oct 05 07:11:29 crc kubenswrapper[4846]: I1005 07:11:29.111709 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8d55f55a-2d79-4c01-8b0f-d67cefb842ff-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8d55f55a-2d79-4c01-8b0f-d67cefb842ff" (UID: "8d55f55a-2d79-4c01-8b0f-d67cefb842ff"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:11:29 crc kubenswrapper[4846]: I1005 07:11:29.126001 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a72c0cf5-8cd9-4f5f-8a2b-25bf41ad0638-config" (OuterVolumeSpecName: "config") pod "a72c0cf5-8cd9-4f5f-8a2b-25bf41ad0638" (UID: "a72c0cf5-8cd9-4f5f-8a2b-25bf41ad0638"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:11:29 crc kubenswrapper[4846]: I1005 07:11:29.126401 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hfbjg\" (UniqueName: \"kubernetes.io/projected/98b50269-c6d4-42f2-8cef-afc127bfccb1-kube-api-access-hfbjg\") pod \"98b50269-c6d4-42f2-8cef-afc127bfccb1\" (UID: \"98b50269-c6d4-42f2-8cef-afc127bfccb1\") " Oct 05 07:11:29 crc kubenswrapper[4846]: I1005 07:11:29.126527 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/98b50269-c6d4-42f2-8cef-afc127bfccb1-combined-ca-bundle\") pod \"98b50269-c6d4-42f2-8cef-afc127bfccb1\" (UID: \"98b50269-c6d4-42f2-8cef-afc127bfccb1\") " Oct 05 07:11:29 crc kubenswrapper[4846]: I1005 07:11:29.126557 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/98b50269-c6d4-42f2-8cef-afc127bfccb1-config-data\") pod \"98b50269-c6d4-42f2-8cef-afc127bfccb1\" (UID: \"98b50269-c6d4-42f2-8cef-afc127bfccb1\") " Oct 05 07:11:29 crc kubenswrapper[4846]: I1005 07:11:29.126877 4846 reconciler_common.go:293] "Volume detached for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/d88c48f9-948e-4888-986e-e00163ec93a6-galera-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:29 crc kubenswrapper[4846]: I1005 07:11:29.126895 4846 reconciler_common.go:293] "Volume detached for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:29 crc kubenswrapper[4846]: I1005 07:11:29.126905 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8vlxz\" (UniqueName: \"kubernetes.io/projected/a72c0cf5-8cd9-4f5f-8a2b-25bf41ad0638-kube-api-access-8vlxz\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:29 crc kubenswrapper[4846]: I1005 07:11:29.126916 4846 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/a72c0cf5-8cd9-4f5f-8a2b-25bf41ad0638-httpd-config\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:29 crc kubenswrapper[4846]: I1005 07:11:29.126926 4846 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8d55f55a-2d79-4c01-8b0f-d67cefb842ff-logs\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:29 crc kubenswrapper[4846]: I1005 07:11:29.126934 4846 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8d55f55a-2d79-4c01-8b0f-d67cefb842ff-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:29 crc kubenswrapper[4846]: I1005 07:11:29.126942 4846 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/a72c0cf5-8cd9-4f5f-8a2b-25bf41ad0638-config\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:29 crc kubenswrapper[4846]: I1005 07:11:29.126950 4846 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8d55f55a-2d79-4c01-8b0f-d67cefb842ff-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:29 crc kubenswrapper[4846]: I1005 07:11:29.126959 4846 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3995994c-2a3f-400f-9a64-b564759b20b2-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:29 crc kubenswrapper[4846]: I1005 07:11:29.126969 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8n2gd\" (UniqueName: \"kubernetes.io/projected/8d55f55a-2d79-4c01-8b0f-d67cefb842ff-kube-api-access-8n2gd\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:29 crc kubenswrapper[4846]: I1005 07:11:29.131229 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-northd-0"] Oct 05 07:11:29 crc kubenswrapper[4846]: I1005 07:11:29.134251 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a72c0cf5-8cd9-4f5f-8a2b-25bf41ad0638-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "a72c0cf5-8cd9-4f5f-8a2b-25bf41ad0638" (UID: "a72c0cf5-8cd9-4f5f-8a2b-25bf41ad0638"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:11:29 crc kubenswrapper[4846]: I1005 07:11:29.136247 4846 scope.go:117] "RemoveContainer" containerID="1ac7cfeb9e27da195d8d27a5e4c8617a950c62fafc93500f5db134c1b1af30c2" Oct 05 07:11:29 crc kubenswrapper[4846]: I1005 07:11:29.157237 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/98b50269-c6d4-42f2-8cef-afc127bfccb1-kube-api-access-hfbjg" (OuterVolumeSpecName: "kube-api-access-hfbjg") pod "98b50269-c6d4-42f2-8cef-afc127bfccb1" (UID: "98b50269-c6d4-42f2-8cef-afc127bfccb1"). InnerVolumeSpecName "kube-api-access-hfbjg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:11:29 crc kubenswrapper[4846]: I1005 07:11:29.162517 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-northd-0"] Oct 05 07:11:29 crc kubenswrapper[4846]: I1005 07:11:29.162531 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a72c0cf5-8cd9-4f5f-8a2b-25bf41ad0638-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a72c0cf5-8cd9-4f5f-8a2b-25bf41ad0638" (UID: "a72c0cf5-8cd9-4f5f-8a2b-25bf41ad0638"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:11:29 crc kubenswrapper[4846]: I1005 07:11:29.166813 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8d55f55a-2d79-4c01-8b0f-d67cefb842ff-config-data" (OuterVolumeSpecName: "config-data") pod "8d55f55a-2d79-4c01-8b0f-d67cefb842ff" (UID: "8d55f55a-2d79-4c01-8b0f-d67cefb842ff"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:11:29 crc kubenswrapper[4846]: I1005 07:11:29.168714 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a72c0cf5-8cd9-4f5f-8a2b-25bf41ad0638-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "a72c0cf5-8cd9-4f5f-8a2b-25bf41ad0638" (UID: "a72c0cf5-8cd9-4f5f-8a2b-25bf41ad0638"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:11:29 crc kubenswrapper[4846]: I1005 07:11:29.168768 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/98b50269-c6d4-42f2-8cef-afc127bfccb1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "98b50269-c6d4-42f2-8cef-afc127bfccb1" (UID: "98b50269-c6d4-42f2-8cef-afc127bfccb1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:11:29 crc kubenswrapper[4846]: I1005 07:11:29.177729 4846 scope.go:117] "RemoveContainer" containerID="57243f61674b9f96faaeba668f666e7518c0f44adbe9b6c573a200820bbb58d3" Oct 05 07:11:29 crc kubenswrapper[4846]: I1005 07:11:29.178663 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/98b50269-c6d4-42f2-8cef-afc127bfccb1-config-data" (OuterVolumeSpecName: "config-data") pod "98b50269-c6d4-42f2-8cef-afc127bfccb1" (UID: "98b50269-c6d4-42f2-8cef-afc127bfccb1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:11:29 crc kubenswrapper[4846]: I1005 07:11:29.186238 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a72c0cf5-8cd9-4f5f-8a2b-25bf41ad0638-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "a72c0cf5-8cd9-4f5f-8a2b-25bf41ad0638" (UID: "a72c0cf5-8cd9-4f5f-8a2b-25bf41ad0638"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:11:29 crc kubenswrapper[4846]: I1005 07:11:29.201493 4846 scope.go:117] "RemoveContainer" containerID="1ac7cfeb9e27da195d8d27a5e4c8617a950c62fafc93500f5db134c1b1af30c2" Oct 05 07:11:29 crc kubenswrapper[4846]: E1005 07:11:29.202263 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1ac7cfeb9e27da195d8d27a5e4c8617a950c62fafc93500f5db134c1b1af30c2\": container with ID starting with 1ac7cfeb9e27da195d8d27a5e4c8617a950c62fafc93500f5db134c1b1af30c2 not found: ID does not exist" containerID="1ac7cfeb9e27da195d8d27a5e4c8617a950c62fafc93500f5db134c1b1af30c2" Oct 05 07:11:29 crc kubenswrapper[4846]: I1005 07:11:29.202407 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1ac7cfeb9e27da195d8d27a5e4c8617a950c62fafc93500f5db134c1b1af30c2"} err="failed to get container status \"1ac7cfeb9e27da195d8d27a5e4c8617a950c62fafc93500f5db134c1b1af30c2\": rpc error: code = NotFound desc = could not find container \"1ac7cfeb9e27da195d8d27a5e4c8617a950c62fafc93500f5db134c1b1af30c2\": container with ID starting with 1ac7cfeb9e27da195d8d27a5e4c8617a950c62fafc93500f5db134c1b1af30c2 not found: ID does not exist" Oct 05 07:11:29 crc kubenswrapper[4846]: I1005 07:11:29.202436 4846 scope.go:117] "RemoveContainer" containerID="57243f61674b9f96faaeba668f666e7518c0f44adbe9b6c573a200820bbb58d3" Oct 05 07:11:29 crc kubenswrapper[4846]: E1005 07:11:29.203083 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"57243f61674b9f96faaeba668f666e7518c0f44adbe9b6c573a200820bbb58d3\": container with ID starting with 57243f61674b9f96faaeba668f666e7518c0f44adbe9b6c573a200820bbb58d3 not found: ID does not exist" containerID="57243f61674b9f96faaeba668f666e7518c0f44adbe9b6c573a200820bbb58d3" Oct 05 07:11:29 crc kubenswrapper[4846]: I1005 07:11:29.203103 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"57243f61674b9f96faaeba668f666e7518c0f44adbe9b6c573a200820bbb58d3"} err="failed to get container status \"57243f61674b9f96faaeba668f666e7518c0f44adbe9b6c573a200820bbb58d3\": rpc error: code = NotFound desc = could not find container \"57243f61674b9f96faaeba668f666e7518c0f44adbe9b6c573a200820bbb58d3\": container with ID starting with 57243f61674b9f96faaeba668f666e7518c0f44adbe9b6c573a200820bbb58d3 not found: ID does not exist" Oct 05 07:11:29 crc kubenswrapper[4846]: I1005 07:11:29.203116 4846 scope.go:117] "RemoveContainer" containerID="6a990887fb132123ab2aa7e7c062e5fbecb25e29a746e6d59d303ed6848ccde0" Oct 05 07:11:29 crc kubenswrapper[4846]: I1005 07:11:29.228122 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1fa90810-c620-48c2-a49c-762341889743-internal-tls-certs\") pod \"1fa90810-c620-48c2-a49c-762341889743\" (UID: \"1fa90810-c620-48c2-a49c-762341889743\") " Oct 05 07:11:29 crc kubenswrapper[4846]: I1005 07:11:29.228263 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/1fa90810-c620-48c2-a49c-762341889743-fernet-keys\") pod \"1fa90810-c620-48c2-a49c-762341889743\" (UID: \"1fa90810-c620-48c2-a49c-762341889743\") " Oct 05 07:11:29 crc kubenswrapper[4846]: I1005 07:11:29.228295 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/1fa90810-c620-48c2-a49c-762341889743-credential-keys\") pod \"1fa90810-c620-48c2-a49c-762341889743\" (UID: \"1fa90810-c620-48c2-a49c-762341889743\") " Oct 05 07:11:29 crc kubenswrapper[4846]: I1005 07:11:29.228351 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1fa90810-c620-48c2-a49c-762341889743-scripts\") pod \"1fa90810-c620-48c2-a49c-762341889743\" (UID: \"1fa90810-c620-48c2-a49c-762341889743\") " Oct 05 07:11:29 crc kubenswrapper[4846]: I1005 07:11:29.228392 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m8n46\" (UniqueName: \"kubernetes.io/projected/1fa90810-c620-48c2-a49c-762341889743-kube-api-access-m8n46\") pod \"1fa90810-c620-48c2-a49c-762341889743\" (UID: \"1fa90810-c620-48c2-a49c-762341889743\") " Oct 05 07:11:29 crc kubenswrapper[4846]: I1005 07:11:29.228421 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1fa90810-c620-48c2-a49c-762341889743-combined-ca-bundle\") pod \"1fa90810-c620-48c2-a49c-762341889743\" (UID: \"1fa90810-c620-48c2-a49c-762341889743\") " Oct 05 07:11:29 crc kubenswrapper[4846]: I1005 07:11:29.228557 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1fa90810-c620-48c2-a49c-762341889743-config-data\") pod \"1fa90810-c620-48c2-a49c-762341889743\" (UID: \"1fa90810-c620-48c2-a49c-762341889743\") " Oct 05 07:11:29 crc kubenswrapper[4846]: I1005 07:11:29.228611 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1fa90810-c620-48c2-a49c-762341889743-public-tls-certs\") pod \"1fa90810-c620-48c2-a49c-762341889743\" (UID: \"1fa90810-c620-48c2-a49c-762341889743\") " Oct 05 07:11:29 crc kubenswrapper[4846]: I1005 07:11:29.228978 4846 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/a72c0cf5-8cd9-4f5f-8a2b-25bf41ad0638-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:29 crc kubenswrapper[4846]: I1005 07:11:29.229001 4846 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8d55f55a-2d79-4c01-8b0f-d67cefb842ff-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:29 crc kubenswrapper[4846]: I1005 07:11:29.229012 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hfbjg\" (UniqueName: \"kubernetes.io/projected/98b50269-c6d4-42f2-8cef-afc127bfccb1-kube-api-access-hfbjg\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:29 crc kubenswrapper[4846]: I1005 07:11:29.229026 4846 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a72c0cf5-8cd9-4f5f-8a2b-25bf41ad0638-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:29 crc kubenswrapper[4846]: I1005 07:11:29.229039 4846 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a72c0cf5-8cd9-4f5f-8a2b-25bf41ad0638-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:29 crc kubenswrapper[4846]: I1005 07:11:29.229049 4846 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/98b50269-c6d4-42f2-8cef-afc127bfccb1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:29 crc kubenswrapper[4846]: I1005 07:11:29.229059 4846 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/98b50269-c6d4-42f2-8cef-afc127bfccb1-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:29 crc kubenswrapper[4846]: I1005 07:11:29.229068 4846 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a72c0cf5-8cd9-4f5f-8a2b-25bf41ad0638-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:29 crc kubenswrapper[4846]: I1005 07:11:29.230716 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1fa90810-c620-48c2-a49c-762341889743-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "1fa90810-c620-48c2-a49c-762341889743" (UID: "1fa90810-c620-48c2-a49c-762341889743"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:11:29 crc kubenswrapper[4846]: I1005 07:11:29.230933 4846 scope.go:117] "RemoveContainer" containerID="bca1da26068e641006909835041107dfa8983891c244081dc35a6bf70c412d1d" Oct 05 07:11:29 crc kubenswrapper[4846]: I1005 07:11:29.233058 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1fa90810-c620-48c2-a49c-762341889743-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "1fa90810-c620-48c2-a49c-762341889743" (UID: "1fa90810-c620-48c2-a49c-762341889743"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:11:29 crc kubenswrapper[4846]: I1005 07:11:29.233728 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1fa90810-c620-48c2-a49c-762341889743-kube-api-access-m8n46" (OuterVolumeSpecName: "kube-api-access-m8n46") pod "1fa90810-c620-48c2-a49c-762341889743" (UID: "1fa90810-c620-48c2-a49c-762341889743"). InnerVolumeSpecName "kube-api-access-m8n46". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:11:29 crc kubenswrapper[4846]: I1005 07:11:29.233974 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1fa90810-c620-48c2-a49c-762341889743-scripts" (OuterVolumeSpecName: "scripts") pod "1fa90810-c620-48c2-a49c-762341889743" (UID: "1fa90810-c620-48c2-a49c-762341889743"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:11:29 crc kubenswrapper[4846]: I1005 07:11:29.257791 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1fa90810-c620-48c2-a49c-762341889743-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1fa90810-c620-48c2-a49c-762341889743" (UID: "1fa90810-c620-48c2-a49c-762341889743"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:11:29 crc kubenswrapper[4846]: I1005 07:11:29.258755 4846 scope.go:117] "RemoveContainer" containerID="6a990887fb132123ab2aa7e7c062e5fbecb25e29a746e6d59d303ed6848ccde0" Oct 05 07:11:29 crc kubenswrapper[4846]: E1005 07:11:29.259468 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6a990887fb132123ab2aa7e7c062e5fbecb25e29a746e6d59d303ed6848ccde0\": container with ID starting with 6a990887fb132123ab2aa7e7c062e5fbecb25e29a746e6d59d303ed6848ccde0 not found: ID does not exist" containerID="6a990887fb132123ab2aa7e7c062e5fbecb25e29a746e6d59d303ed6848ccde0" Oct 05 07:11:29 crc kubenswrapper[4846]: I1005 07:11:29.259542 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6a990887fb132123ab2aa7e7c062e5fbecb25e29a746e6d59d303ed6848ccde0"} err="failed to get container status \"6a990887fb132123ab2aa7e7c062e5fbecb25e29a746e6d59d303ed6848ccde0\": rpc error: code = NotFound desc = could not find container \"6a990887fb132123ab2aa7e7c062e5fbecb25e29a746e6d59d303ed6848ccde0\": container with ID starting with 6a990887fb132123ab2aa7e7c062e5fbecb25e29a746e6d59d303ed6848ccde0 not found: ID does not exist" Oct 05 07:11:29 crc kubenswrapper[4846]: I1005 07:11:29.259599 4846 scope.go:117] "RemoveContainer" containerID="bca1da26068e641006909835041107dfa8983891c244081dc35a6bf70c412d1d" Oct 05 07:11:29 crc kubenswrapper[4846]: E1005 07:11:29.260395 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bca1da26068e641006909835041107dfa8983891c244081dc35a6bf70c412d1d\": container with ID starting with bca1da26068e641006909835041107dfa8983891c244081dc35a6bf70c412d1d not found: ID does not exist" containerID="bca1da26068e641006909835041107dfa8983891c244081dc35a6bf70c412d1d" Oct 05 07:11:29 crc kubenswrapper[4846]: I1005 07:11:29.260440 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bca1da26068e641006909835041107dfa8983891c244081dc35a6bf70c412d1d"} err="failed to get container status \"bca1da26068e641006909835041107dfa8983891c244081dc35a6bf70c412d1d\": rpc error: code = NotFound desc = could not find container \"bca1da26068e641006909835041107dfa8983891c244081dc35a6bf70c412d1d\": container with ID starting with bca1da26068e641006909835041107dfa8983891c244081dc35a6bf70c412d1d not found: ID does not exist" Oct 05 07:11:29 crc kubenswrapper[4846]: I1005 07:11:29.260469 4846 scope.go:117] "RemoveContainer" containerID="055332eb57f5e6e889d548a1b4717caea296e84dc70616d5bdb53dd34aaa0697" Oct 05 07:11:29 crc kubenswrapper[4846]: I1005 07:11:29.261669 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1fa90810-c620-48c2-a49c-762341889743-config-data" (OuterVolumeSpecName: "config-data") pod "1fa90810-c620-48c2-a49c-762341889743" (UID: "1fa90810-c620-48c2-a49c-762341889743"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:11:29 crc kubenswrapper[4846]: I1005 07:11:29.269911 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1fa90810-c620-48c2-a49c-762341889743-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "1fa90810-c620-48c2-a49c-762341889743" (UID: "1fa90810-c620-48c2-a49c-762341889743"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:11:29 crc kubenswrapper[4846]: I1005 07:11:29.279705 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1fa90810-c620-48c2-a49c-762341889743-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "1fa90810-c620-48c2-a49c-762341889743" (UID: "1fa90810-c620-48c2-a49c-762341889743"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:11:29 crc kubenswrapper[4846]: I1005 07:11:29.286911 4846 scope.go:117] "RemoveContainer" containerID="5cae68aeaeb33c53f345bf863894fe1192fa80a8ec2b486a6c2d29e83dec72af" Oct 05 07:11:29 crc kubenswrapper[4846]: I1005 07:11:29.314348 4846 scope.go:117] "RemoveContainer" containerID="afd676f2477334d96744cd37d02c69587dea11053d28e2ae166c10deda1b1560" Oct 05 07:11:29 crc kubenswrapper[4846]: I1005 07:11:29.316228 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstack-galera-0"] Oct 05 07:11:29 crc kubenswrapper[4846]: I1005 07:11:29.322335 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstack-galera-0"] Oct 05 07:11:29 crc kubenswrapper[4846]: I1005 07:11:29.330119 4846 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1fa90810-c620-48c2-a49c-762341889743-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:29 crc kubenswrapper[4846]: I1005 07:11:29.330155 4846 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1fa90810-c620-48c2-a49c-762341889743-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:29 crc kubenswrapper[4846]: I1005 07:11:29.330170 4846 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/1fa90810-c620-48c2-a49c-762341889743-fernet-keys\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:29 crc kubenswrapper[4846]: I1005 07:11:29.330204 4846 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/1fa90810-c620-48c2-a49c-762341889743-credential-keys\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:29 crc kubenswrapper[4846]: I1005 07:11:29.330221 4846 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1fa90810-c620-48c2-a49c-762341889743-scripts\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:29 crc kubenswrapper[4846]: I1005 07:11:29.330231 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m8n46\" (UniqueName: \"kubernetes.io/projected/1fa90810-c620-48c2-a49c-762341889743-kube-api-access-m8n46\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:29 crc kubenswrapper[4846]: I1005 07:11:29.330241 4846 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1fa90810-c620-48c2-a49c-762341889743-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:29 crc kubenswrapper[4846]: I1005 07:11:29.330250 4846 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1fa90810-c620-48c2-a49c-762341889743-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:29 crc kubenswrapper[4846]: I1005 07:11:29.331799 4846 scope.go:117] "RemoveContainer" containerID="5cae68aeaeb33c53f345bf863894fe1192fa80a8ec2b486a6c2d29e83dec72af" Oct 05 07:11:29 crc kubenswrapper[4846]: E1005 07:11:29.332183 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5cae68aeaeb33c53f345bf863894fe1192fa80a8ec2b486a6c2d29e83dec72af\": container with ID starting with 5cae68aeaeb33c53f345bf863894fe1192fa80a8ec2b486a6c2d29e83dec72af not found: ID does not exist" containerID="5cae68aeaeb33c53f345bf863894fe1192fa80a8ec2b486a6c2d29e83dec72af" Oct 05 07:11:29 crc kubenswrapper[4846]: I1005 07:11:29.332247 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5cae68aeaeb33c53f345bf863894fe1192fa80a8ec2b486a6c2d29e83dec72af"} err="failed to get container status \"5cae68aeaeb33c53f345bf863894fe1192fa80a8ec2b486a6c2d29e83dec72af\": rpc error: code = NotFound desc = could not find container \"5cae68aeaeb33c53f345bf863894fe1192fa80a8ec2b486a6c2d29e83dec72af\": container with ID starting with 5cae68aeaeb33c53f345bf863894fe1192fa80a8ec2b486a6c2d29e83dec72af not found: ID does not exist" Oct 05 07:11:29 crc kubenswrapper[4846]: I1005 07:11:29.332270 4846 scope.go:117] "RemoveContainer" containerID="afd676f2477334d96744cd37d02c69587dea11053d28e2ae166c10deda1b1560" Oct 05 07:11:29 crc kubenswrapper[4846]: E1005 07:11:29.332617 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"afd676f2477334d96744cd37d02c69587dea11053d28e2ae166c10deda1b1560\": container with ID starting with afd676f2477334d96744cd37d02c69587dea11053d28e2ae166c10deda1b1560 not found: ID does not exist" containerID="afd676f2477334d96744cd37d02c69587dea11053d28e2ae166c10deda1b1560" Oct 05 07:11:29 crc kubenswrapper[4846]: I1005 07:11:29.332637 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"afd676f2477334d96744cd37d02c69587dea11053d28e2ae166c10deda1b1560"} err="failed to get container status \"afd676f2477334d96744cd37d02c69587dea11053d28e2ae166c10deda1b1560\": rpc error: code = NotFound desc = could not find container \"afd676f2477334d96744cd37d02c69587dea11053d28e2ae166c10deda1b1560\": container with ID starting with afd676f2477334d96744cd37d02c69587dea11053d28e2ae166c10deda1b1560 not found: ID does not exist" Oct 05 07:11:29 crc kubenswrapper[4846]: I1005 07:11:29.332652 4846 scope.go:117] "RemoveContainer" containerID="05e6ade463feffe3d2a4a45255b934b5f81dd79a5b4cd4809663d8a570eb260c" Oct 05 07:11:29 crc kubenswrapper[4846]: I1005 07:11:29.377339 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-85bbbcfb7f-kwfjt"] Oct 05 07:11:29 crc kubenswrapper[4846]: I1005 07:11:29.384464 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-85bbbcfb7f-kwfjt"] Oct 05 07:11:29 crc kubenswrapper[4846]: I1005 07:11:29.387024 4846 scope.go:117] "RemoveContainer" containerID="4351908cf00eea1071aca5080f9d536de25e56a78e5d0713bec8f51caab84ecb" Oct 05 07:11:29 crc kubenswrapper[4846]: I1005 07:11:29.396630 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-keystone-listener-77664b8dfd-5hkqj"] Oct 05 07:11:29 crc kubenswrapper[4846]: I1005 07:11:29.411592 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-keystone-listener-77664b8dfd-5hkqj"] Oct 05 07:11:29 crc kubenswrapper[4846]: I1005 07:11:29.418265 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-worker-78995b9f79-nwl7v"] Oct 05 07:11:29 crc kubenswrapper[4846]: I1005 07:11:29.424656 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-worker-78995b9f79-nwl7v"] Oct 05 07:11:30 crc kubenswrapper[4846]: I1005 07:11:30.096296 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"98b50269-c6d4-42f2-8cef-afc127bfccb1","Type":"ContainerDied","Data":"9f3658c37c9a27d088645083e4f9c2797423c225b6855b8581b1eff8ce69200e"} Oct 05 07:11:30 crc kubenswrapper[4846]: I1005 07:11:30.097385 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Oct 05 07:11:30 crc kubenswrapper[4846]: I1005 07:11:30.120329 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-79f67599fc-2sk76" event={"ID":"1fa90810-c620-48c2-a49c-762341889743","Type":"ContainerDied","Data":"0c1fa59eb2207671e67d590e69cac606e13f97317e06feb44ee17e3ee38b3559"} Oct 05 07:11:30 crc kubenswrapper[4846]: I1005 07:11:30.120350 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-79f67599fc-2sk76" Oct 05 07:11:30 crc kubenswrapper[4846]: I1005 07:11:30.120393 4846 scope.go:117] "RemoveContainer" containerID="26da8c60efda49d0ad9e38f18f39b3b3c7a75ffb60bb839d759345619a5d337c" Oct 05 07:11:30 crc kubenswrapper[4846]: I1005 07:11:30.155831 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-79f67599fc-2sk76"] Oct 05 07:11:30 crc kubenswrapper[4846]: I1005 07:11:30.162503 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-79f67599fc-2sk76"] Oct 05 07:11:30 crc kubenswrapper[4846]: I1005 07:11:30.176929 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 05 07:11:30 crc kubenswrapper[4846]: I1005 07:11:30.183512 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 05 07:11:30 crc kubenswrapper[4846]: I1005 07:11:30.510471 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="169918d8-597e-4ef4-b00c-fb14c4d0347f" path="/var/lib/kubelet/pods/169918d8-597e-4ef4-b00c-fb14c4d0347f/volumes" Oct 05 07:11:30 crc kubenswrapper[4846]: I1005 07:11:30.512293 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="18a7be37-c8dc-4d8a-808f-2a63eb4d9692" path="/var/lib/kubelet/pods/18a7be37-c8dc-4d8a-808f-2a63eb4d9692/volumes" Oct 05 07:11:30 crc kubenswrapper[4846]: I1005 07:11:30.513470 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1fa90810-c620-48c2-a49c-762341889743" path="/var/lib/kubelet/pods/1fa90810-c620-48c2-a49c-762341889743/volumes" Oct 05 07:11:30 crc kubenswrapper[4846]: I1005 07:11:30.515591 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3995994c-2a3f-400f-9a64-b564759b20b2" path="/var/lib/kubelet/pods/3995994c-2a3f-400f-9a64-b564759b20b2/volumes" Oct 05 07:11:30 crc kubenswrapper[4846]: I1005 07:11:30.516825 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8d55f55a-2d79-4c01-8b0f-d67cefb842ff" path="/var/lib/kubelet/pods/8d55f55a-2d79-4c01-8b0f-d67cefb842ff/volumes" Oct 05 07:11:30 crc kubenswrapper[4846]: I1005 07:11:30.517934 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="98b50269-c6d4-42f2-8cef-afc127bfccb1" path="/var/lib/kubelet/pods/98b50269-c6d4-42f2-8cef-afc127bfccb1/volumes" Oct 05 07:11:30 crc kubenswrapper[4846]: I1005 07:11:30.519997 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a72c0cf5-8cd9-4f5f-8a2b-25bf41ad0638" path="/var/lib/kubelet/pods/a72c0cf5-8cd9-4f5f-8a2b-25bf41ad0638/volumes" Oct 05 07:11:30 crc kubenswrapper[4846]: I1005 07:11:30.521415 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d88c48f9-948e-4888-986e-e00163ec93a6" path="/var/lib/kubelet/pods/d88c48f9-948e-4888-986e-e00163ec93a6/volumes" Oct 05 07:11:32 crc kubenswrapper[4846]: I1005 07:11:32.146746 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 05 07:11:32 crc kubenswrapper[4846]: I1005 07:11:32.150852 4846 generic.go:334] "Generic (PLEG): container finished" podID="b1e394dd-f55b-4af1-af0b-52cc7d09afa1" containerID="d0d0db8b415aac7ced1865679a2764d980d7574715bf4b8a924f0edb74f9f2fc" exitCode=0 Oct 05 07:11:32 crc kubenswrapper[4846]: I1005 07:11:32.150913 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b1e394dd-f55b-4af1-af0b-52cc7d09afa1","Type":"ContainerDied","Data":"d0d0db8b415aac7ced1865679a2764d980d7574715bf4b8a924f0edb74f9f2fc"} Oct 05 07:11:32 crc kubenswrapper[4846]: I1005 07:11:32.150950 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b1e394dd-f55b-4af1-af0b-52cc7d09afa1","Type":"ContainerDied","Data":"d8f6a2cda488a503a4cdbec1246f11f58542b60bc44c2b7a0b7ba804f3e98a70"} Oct 05 07:11:32 crc kubenswrapper[4846]: I1005 07:11:32.150972 4846 scope.go:117] "RemoveContainer" containerID="e92d479cbafa7786e2514b2d1e99c120ef33197a849bb0c13e2362354bcebd01" Oct 05 07:11:32 crc kubenswrapper[4846]: I1005 07:11:32.201881 4846 scope.go:117] "RemoveContainer" containerID="6db54d2ca22110333154adb6a23a0ec27e54e4f7118d44ea9636189fffd9e0ae" Oct 05 07:11:32 crc kubenswrapper[4846]: I1005 07:11:32.220130 4846 scope.go:117] "RemoveContainer" containerID="d0d0db8b415aac7ced1865679a2764d980d7574715bf4b8a924f0edb74f9f2fc" Oct 05 07:11:32 crc kubenswrapper[4846]: I1005 07:11:32.239663 4846 scope.go:117] "RemoveContainer" containerID="6682a3a723ef5336f655521943bfbd504709a37185a094744e9c63a0c851b903" Oct 05 07:11:32 crc kubenswrapper[4846]: I1005 07:11:32.263643 4846 scope.go:117] "RemoveContainer" containerID="e92d479cbafa7786e2514b2d1e99c120ef33197a849bb0c13e2362354bcebd01" Oct 05 07:11:32 crc kubenswrapper[4846]: E1005 07:11:32.264144 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e92d479cbafa7786e2514b2d1e99c120ef33197a849bb0c13e2362354bcebd01\": container with ID starting with e92d479cbafa7786e2514b2d1e99c120ef33197a849bb0c13e2362354bcebd01 not found: ID does not exist" containerID="e92d479cbafa7786e2514b2d1e99c120ef33197a849bb0c13e2362354bcebd01" Oct 05 07:11:32 crc kubenswrapper[4846]: I1005 07:11:32.264218 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e92d479cbafa7786e2514b2d1e99c120ef33197a849bb0c13e2362354bcebd01"} err="failed to get container status \"e92d479cbafa7786e2514b2d1e99c120ef33197a849bb0c13e2362354bcebd01\": rpc error: code = NotFound desc = could not find container \"e92d479cbafa7786e2514b2d1e99c120ef33197a849bb0c13e2362354bcebd01\": container with ID starting with e92d479cbafa7786e2514b2d1e99c120ef33197a849bb0c13e2362354bcebd01 not found: ID does not exist" Oct 05 07:11:32 crc kubenswrapper[4846]: I1005 07:11:32.264254 4846 scope.go:117] "RemoveContainer" containerID="6db54d2ca22110333154adb6a23a0ec27e54e4f7118d44ea9636189fffd9e0ae" Oct 05 07:11:32 crc kubenswrapper[4846]: E1005 07:11:32.264735 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6db54d2ca22110333154adb6a23a0ec27e54e4f7118d44ea9636189fffd9e0ae\": container with ID starting with 6db54d2ca22110333154adb6a23a0ec27e54e4f7118d44ea9636189fffd9e0ae not found: ID does not exist" containerID="6db54d2ca22110333154adb6a23a0ec27e54e4f7118d44ea9636189fffd9e0ae" Oct 05 07:11:32 crc kubenswrapper[4846]: I1005 07:11:32.264758 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6db54d2ca22110333154adb6a23a0ec27e54e4f7118d44ea9636189fffd9e0ae"} err="failed to get container status \"6db54d2ca22110333154adb6a23a0ec27e54e4f7118d44ea9636189fffd9e0ae\": rpc error: code = NotFound desc = could not find container \"6db54d2ca22110333154adb6a23a0ec27e54e4f7118d44ea9636189fffd9e0ae\": container with ID starting with 6db54d2ca22110333154adb6a23a0ec27e54e4f7118d44ea9636189fffd9e0ae not found: ID does not exist" Oct 05 07:11:32 crc kubenswrapper[4846]: I1005 07:11:32.264774 4846 scope.go:117] "RemoveContainer" containerID="d0d0db8b415aac7ced1865679a2764d980d7574715bf4b8a924f0edb74f9f2fc" Oct 05 07:11:32 crc kubenswrapper[4846]: E1005 07:11:32.265014 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d0d0db8b415aac7ced1865679a2764d980d7574715bf4b8a924f0edb74f9f2fc\": container with ID starting with d0d0db8b415aac7ced1865679a2764d980d7574715bf4b8a924f0edb74f9f2fc not found: ID does not exist" containerID="d0d0db8b415aac7ced1865679a2764d980d7574715bf4b8a924f0edb74f9f2fc" Oct 05 07:11:32 crc kubenswrapper[4846]: I1005 07:11:32.265073 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d0d0db8b415aac7ced1865679a2764d980d7574715bf4b8a924f0edb74f9f2fc"} err="failed to get container status \"d0d0db8b415aac7ced1865679a2764d980d7574715bf4b8a924f0edb74f9f2fc\": rpc error: code = NotFound desc = could not find container \"d0d0db8b415aac7ced1865679a2764d980d7574715bf4b8a924f0edb74f9f2fc\": container with ID starting with d0d0db8b415aac7ced1865679a2764d980d7574715bf4b8a924f0edb74f9f2fc not found: ID does not exist" Oct 05 07:11:32 crc kubenswrapper[4846]: I1005 07:11:32.265087 4846 scope.go:117] "RemoveContainer" containerID="6682a3a723ef5336f655521943bfbd504709a37185a094744e9c63a0c851b903" Oct 05 07:11:32 crc kubenswrapper[4846]: E1005 07:11:32.265292 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6682a3a723ef5336f655521943bfbd504709a37185a094744e9c63a0c851b903\": container with ID starting with 6682a3a723ef5336f655521943bfbd504709a37185a094744e9c63a0c851b903 not found: ID does not exist" containerID="6682a3a723ef5336f655521943bfbd504709a37185a094744e9c63a0c851b903" Oct 05 07:11:32 crc kubenswrapper[4846]: I1005 07:11:32.265315 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6682a3a723ef5336f655521943bfbd504709a37185a094744e9c63a0c851b903"} err="failed to get container status \"6682a3a723ef5336f655521943bfbd504709a37185a094744e9c63a0c851b903\": rpc error: code = NotFound desc = could not find container \"6682a3a723ef5336f655521943bfbd504709a37185a094744e9c63a0c851b903\": container with ID starting with 6682a3a723ef5336f655521943bfbd504709a37185a094744e9c63a0c851b903 not found: ID does not exist" Oct 05 07:11:32 crc kubenswrapper[4846]: I1005 07:11:32.302540 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b1e394dd-f55b-4af1-af0b-52cc7d09afa1-sg-core-conf-yaml\") pod \"b1e394dd-f55b-4af1-af0b-52cc7d09afa1\" (UID: \"b1e394dd-f55b-4af1-af0b-52cc7d09afa1\") " Oct 05 07:11:32 crc kubenswrapper[4846]: I1005 07:11:32.302636 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lggkw\" (UniqueName: \"kubernetes.io/projected/b1e394dd-f55b-4af1-af0b-52cc7d09afa1-kube-api-access-lggkw\") pod \"b1e394dd-f55b-4af1-af0b-52cc7d09afa1\" (UID: \"b1e394dd-f55b-4af1-af0b-52cc7d09afa1\") " Oct 05 07:11:32 crc kubenswrapper[4846]: I1005 07:11:32.302672 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b1e394dd-f55b-4af1-af0b-52cc7d09afa1-config-data\") pod \"b1e394dd-f55b-4af1-af0b-52cc7d09afa1\" (UID: \"b1e394dd-f55b-4af1-af0b-52cc7d09afa1\") " Oct 05 07:11:32 crc kubenswrapper[4846]: I1005 07:11:32.302693 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b1e394dd-f55b-4af1-af0b-52cc7d09afa1-run-httpd\") pod \"b1e394dd-f55b-4af1-af0b-52cc7d09afa1\" (UID: \"b1e394dd-f55b-4af1-af0b-52cc7d09afa1\") " Oct 05 07:11:32 crc kubenswrapper[4846]: I1005 07:11:32.302726 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b1e394dd-f55b-4af1-af0b-52cc7d09afa1-scripts\") pod \"b1e394dd-f55b-4af1-af0b-52cc7d09afa1\" (UID: \"b1e394dd-f55b-4af1-af0b-52cc7d09afa1\") " Oct 05 07:11:32 crc kubenswrapper[4846]: I1005 07:11:32.302751 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b1e394dd-f55b-4af1-af0b-52cc7d09afa1-combined-ca-bundle\") pod \"b1e394dd-f55b-4af1-af0b-52cc7d09afa1\" (UID: \"b1e394dd-f55b-4af1-af0b-52cc7d09afa1\") " Oct 05 07:11:32 crc kubenswrapper[4846]: I1005 07:11:32.302783 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b1e394dd-f55b-4af1-af0b-52cc7d09afa1-log-httpd\") pod \"b1e394dd-f55b-4af1-af0b-52cc7d09afa1\" (UID: \"b1e394dd-f55b-4af1-af0b-52cc7d09afa1\") " Oct 05 07:11:32 crc kubenswrapper[4846]: I1005 07:11:32.302832 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/b1e394dd-f55b-4af1-af0b-52cc7d09afa1-ceilometer-tls-certs\") pod \"b1e394dd-f55b-4af1-af0b-52cc7d09afa1\" (UID: \"b1e394dd-f55b-4af1-af0b-52cc7d09afa1\") " Oct 05 07:11:32 crc kubenswrapper[4846]: I1005 07:11:32.304554 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b1e394dd-f55b-4af1-af0b-52cc7d09afa1-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "b1e394dd-f55b-4af1-af0b-52cc7d09afa1" (UID: "b1e394dd-f55b-4af1-af0b-52cc7d09afa1"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 07:11:32 crc kubenswrapper[4846]: I1005 07:11:32.304569 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b1e394dd-f55b-4af1-af0b-52cc7d09afa1-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "b1e394dd-f55b-4af1-af0b-52cc7d09afa1" (UID: "b1e394dd-f55b-4af1-af0b-52cc7d09afa1"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 07:11:32 crc kubenswrapper[4846]: I1005 07:11:32.311082 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b1e394dd-f55b-4af1-af0b-52cc7d09afa1-kube-api-access-lggkw" (OuterVolumeSpecName: "kube-api-access-lggkw") pod "b1e394dd-f55b-4af1-af0b-52cc7d09afa1" (UID: "b1e394dd-f55b-4af1-af0b-52cc7d09afa1"). InnerVolumeSpecName "kube-api-access-lggkw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:11:32 crc kubenswrapper[4846]: I1005 07:11:32.312017 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b1e394dd-f55b-4af1-af0b-52cc7d09afa1-scripts" (OuterVolumeSpecName: "scripts") pod "b1e394dd-f55b-4af1-af0b-52cc7d09afa1" (UID: "b1e394dd-f55b-4af1-af0b-52cc7d09afa1"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:11:32 crc kubenswrapper[4846]: I1005 07:11:32.356441 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b1e394dd-f55b-4af1-af0b-52cc7d09afa1-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "b1e394dd-f55b-4af1-af0b-52cc7d09afa1" (UID: "b1e394dd-f55b-4af1-af0b-52cc7d09afa1"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:11:32 crc kubenswrapper[4846]: I1005 07:11:32.359802 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b1e394dd-f55b-4af1-af0b-52cc7d09afa1-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "b1e394dd-f55b-4af1-af0b-52cc7d09afa1" (UID: "b1e394dd-f55b-4af1-af0b-52cc7d09afa1"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:11:32 crc kubenswrapper[4846]: I1005 07:11:32.404945 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lggkw\" (UniqueName: \"kubernetes.io/projected/b1e394dd-f55b-4af1-af0b-52cc7d09afa1-kube-api-access-lggkw\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:32 crc kubenswrapper[4846]: I1005 07:11:32.404999 4846 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b1e394dd-f55b-4af1-af0b-52cc7d09afa1-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:32 crc kubenswrapper[4846]: I1005 07:11:32.405010 4846 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b1e394dd-f55b-4af1-af0b-52cc7d09afa1-scripts\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:32 crc kubenswrapper[4846]: I1005 07:11:32.405019 4846 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b1e394dd-f55b-4af1-af0b-52cc7d09afa1-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:32 crc kubenswrapper[4846]: I1005 07:11:32.405028 4846 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/b1e394dd-f55b-4af1-af0b-52cc7d09afa1-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:32 crc kubenswrapper[4846]: I1005 07:11:32.405246 4846 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b1e394dd-f55b-4af1-af0b-52cc7d09afa1-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:32 crc kubenswrapper[4846]: I1005 07:11:32.412439 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b1e394dd-f55b-4af1-af0b-52cc7d09afa1-config-data" (OuterVolumeSpecName: "config-data") pod "b1e394dd-f55b-4af1-af0b-52cc7d09afa1" (UID: "b1e394dd-f55b-4af1-af0b-52cc7d09afa1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:11:32 crc kubenswrapper[4846]: I1005 07:11:32.429410 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b1e394dd-f55b-4af1-af0b-52cc7d09afa1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b1e394dd-f55b-4af1-af0b-52cc7d09afa1" (UID: "b1e394dd-f55b-4af1-af0b-52cc7d09afa1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:11:32 crc kubenswrapper[4846]: I1005 07:11:32.507225 4846 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b1e394dd-f55b-4af1-af0b-52cc7d09afa1-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:32 crc kubenswrapper[4846]: I1005 07:11:32.507279 4846 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b1e394dd-f55b-4af1-af0b-52cc7d09afa1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:32 crc kubenswrapper[4846]: E1005 07:11:32.942918 4846 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of f4fa8732ebd0803a3d57af2df182505f645e900471768d31fb1d43cb0a83e0cf is running failed: container process not found" containerID="f4fa8732ebd0803a3d57af2df182505f645e900471768d31fb1d43cb0a83e0cf" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 05 07:11:32 crc kubenswrapper[4846]: E1005 07:11:32.943772 4846 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of f4fa8732ebd0803a3d57af2df182505f645e900471768d31fb1d43cb0a83e0cf is running failed: container process not found" containerID="f4fa8732ebd0803a3d57af2df182505f645e900471768d31fb1d43cb0a83e0cf" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 05 07:11:32 crc kubenswrapper[4846]: E1005 07:11:32.944479 4846 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of f4fa8732ebd0803a3d57af2df182505f645e900471768d31fb1d43cb0a83e0cf is running failed: container process not found" containerID="f4fa8732ebd0803a3d57af2df182505f645e900471768d31fb1d43cb0a83e0cf" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 05 07:11:32 crc kubenswrapper[4846]: E1005 07:11:32.944570 4846 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of f4fa8732ebd0803a3d57af2df182505f645e900471768d31fb1d43cb0a83e0cf is running failed: container process not found" probeType="Readiness" pod="openstack/ovn-controller-ovs-sg4vg" podUID="29c0f404-fc7d-4909-b6bf-e1826a092a72" containerName="ovsdb-server" Oct 05 07:11:32 crc kubenswrapper[4846]: E1005 07:11:32.944654 4846 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="22929dba446bc2516169306e09540fa52839ffc6b8d024f494720cc39ad0ab7d" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 05 07:11:32 crc kubenswrapper[4846]: E1005 07:11:32.946220 4846 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="22929dba446bc2516169306e09540fa52839ffc6b8d024f494720cc39ad0ab7d" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 05 07:11:32 crc kubenswrapper[4846]: E1005 07:11:32.951066 4846 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="22929dba446bc2516169306e09540fa52839ffc6b8d024f494720cc39ad0ab7d" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 05 07:11:32 crc kubenswrapper[4846]: E1005 07:11:32.951338 4846 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-controller-ovs-sg4vg" podUID="29c0f404-fc7d-4909-b6bf-e1826a092a72" containerName="ovs-vswitchd" Oct 05 07:11:33 crc kubenswrapper[4846]: I1005 07:11:33.179784 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 05 07:11:33 crc kubenswrapper[4846]: I1005 07:11:33.209351 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 05 07:11:33 crc kubenswrapper[4846]: I1005 07:11:33.215885 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 05 07:11:34 crc kubenswrapper[4846]: I1005 07:11:34.517018 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b1e394dd-f55b-4af1-af0b-52cc7d09afa1" path="/var/lib/kubelet/pods/b1e394dd-f55b-4af1-af0b-52cc7d09afa1/volumes" Oct 05 07:11:37 crc kubenswrapper[4846]: E1005 07:11:37.942612 4846 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of f4fa8732ebd0803a3d57af2df182505f645e900471768d31fb1d43cb0a83e0cf is running failed: container process not found" containerID="f4fa8732ebd0803a3d57af2df182505f645e900471768d31fb1d43cb0a83e0cf" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 05 07:11:37 crc kubenswrapper[4846]: E1005 07:11:37.944096 4846 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="22929dba446bc2516169306e09540fa52839ffc6b8d024f494720cc39ad0ab7d" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 05 07:11:37 crc kubenswrapper[4846]: E1005 07:11:37.944303 4846 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of f4fa8732ebd0803a3d57af2df182505f645e900471768d31fb1d43cb0a83e0cf is running failed: container process not found" containerID="f4fa8732ebd0803a3d57af2df182505f645e900471768d31fb1d43cb0a83e0cf" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 05 07:11:37 crc kubenswrapper[4846]: E1005 07:11:37.944935 4846 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of f4fa8732ebd0803a3d57af2df182505f645e900471768d31fb1d43cb0a83e0cf is running failed: container process not found" containerID="f4fa8732ebd0803a3d57af2df182505f645e900471768d31fb1d43cb0a83e0cf" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 05 07:11:37 crc kubenswrapper[4846]: E1005 07:11:37.945022 4846 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of f4fa8732ebd0803a3d57af2df182505f645e900471768d31fb1d43cb0a83e0cf is running failed: container process not found" probeType="Readiness" pod="openstack/ovn-controller-ovs-sg4vg" podUID="29c0f404-fc7d-4909-b6bf-e1826a092a72" containerName="ovsdb-server" Oct 05 07:11:37 crc kubenswrapper[4846]: E1005 07:11:37.946945 4846 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="22929dba446bc2516169306e09540fa52839ffc6b8d024f494720cc39ad0ab7d" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 05 07:11:37 crc kubenswrapper[4846]: E1005 07:11:37.949900 4846 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="22929dba446bc2516169306e09540fa52839ffc6b8d024f494720cc39ad0ab7d" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 05 07:11:37 crc kubenswrapper[4846]: E1005 07:11:37.950020 4846 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-controller-ovs-sg4vg" podUID="29c0f404-fc7d-4909-b6bf-e1826a092a72" containerName="ovs-vswitchd" Oct 05 07:11:42 crc kubenswrapper[4846]: E1005 07:11:42.943370 4846 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of f4fa8732ebd0803a3d57af2df182505f645e900471768d31fb1d43cb0a83e0cf is running failed: container process not found" containerID="f4fa8732ebd0803a3d57af2df182505f645e900471768d31fb1d43cb0a83e0cf" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 05 07:11:42 crc kubenswrapper[4846]: E1005 07:11:42.945031 4846 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of f4fa8732ebd0803a3d57af2df182505f645e900471768d31fb1d43cb0a83e0cf is running failed: container process not found" containerID="f4fa8732ebd0803a3d57af2df182505f645e900471768d31fb1d43cb0a83e0cf" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 05 07:11:42 crc kubenswrapper[4846]: E1005 07:11:42.945903 4846 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of f4fa8732ebd0803a3d57af2df182505f645e900471768d31fb1d43cb0a83e0cf is running failed: container process not found" containerID="f4fa8732ebd0803a3d57af2df182505f645e900471768d31fb1d43cb0a83e0cf" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 05 07:11:42 crc kubenswrapper[4846]: E1005 07:11:42.946001 4846 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of f4fa8732ebd0803a3d57af2df182505f645e900471768d31fb1d43cb0a83e0cf is running failed: container process not found" probeType="Readiness" pod="openstack/ovn-controller-ovs-sg4vg" podUID="29c0f404-fc7d-4909-b6bf-e1826a092a72" containerName="ovsdb-server" Oct 05 07:11:42 crc kubenswrapper[4846]: E1005 07:11:42.946892 4846 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="22929dba446bc2516169306e09540fa52839ffc6b8d024f494720cc39ad0ab7d" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 05 07:11:42 crc kubenswrapper[4846]: E1005 07:11:42.949167 4846 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="22929dba446bc2516169306e09540fa52839ffc6b8d024f494720cc39ad0ab7d" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 05 07:11:42 crc kubenswrapper[4846]: E1005 07:11:42.952054 4846 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="22929dba446bc2516169306e09540fa52839ffc6b8d024f494720cc39ad0ab7d" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 05 07:11:42 crc kubenswrapper[4846]: E1005 07:11:42.952128 4846 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-controller-ovs-sg4vg" podUID="29c0f404-fc7d-4909-b6bf-e1826a092a72" containerName="ovs-vswitchd" Oct 05 07:11:47 crc kubenswrapper[4846]: E1005 07:11:47.943868 4846 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of f4fa8732ebd0803a3d57af2df182505f645e900471768d31fb1d43cb0a83e0cf is running failed: container process not found" containerID="f4fa8732ebd0803a3d57af2df182505f645e900471768d31fb1d43cb0a83e0cf" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 05 07:11:47 crc kubenswrapper[4846]: E1005 07:11:47.945020 4846 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of f4fa8732ebd0803a3d57af2df182505f645e900471768d31fb1d43cb0a83e0cf is running failed: container process not found" containerID="f4fa8732ebd0803a3d57af2df182505f645e900471768d31fb1d43cb0a83e0cf" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 05 07:11:47 crc kubenswrapper[4846]: E1005 07:11:47.945809 4846 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="22929dba446bc2516169306e09540fa52839ffc6b8d024f494720cc39ad0ab7d" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 05 07:11:47 crc kubenswrapper[4846]: E1005 07:11:47.945828 4846 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of f4fa8732ebd0803a3d57af2df182505f645e900471768d31fb1d43cb0a83e0cf is running failed: container process not found" containerID="f4fa8732ebd0803a3d57af2df182505f645e900471768d31fb1d43cb0a83e0cf" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 05 07:11:47 crc kubenswrapper[4846]: E1005 07:11:47.946446 4846 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of f4fa8732ebd0803a3d57af2df182505f645e900471768d31fb1d43cb0a83e0cf is running failed: container process not found" probeType="Readiness" pod="openstack/ovn-controller-ovs-sg4vg" podUID="29c0f404-fc7d-4909-b6bf-e1826a092a72" containerName="ovsdb-server" Oct 05 07:11:47 crc kubenswrapper[4846]: E1005 07:11:47.949807 4846 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="22929dba446bc2516169306e09540fa52839ffc6b8d024f494720cc39ad0ab7d" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 05 07:11:47 crc kubenswrapper[4846]: E1005 07:11:47.951818 4846 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="22929dba446bc2516169306e09540fa52839ffc6b8d024f494720cc39ad0ab7d" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 05 07:11:47 crc kubenswrapper[4846]: E1005 07:11:47.951886 4846 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-controller-ovs-sg4vg" podUID="29c0f404-fc7d-4909-b6bf-e1826a092a72" containerName="ovs-vswitchd" Oct 05 07:11:49 crc kubenswrapper[4846]: I1005 07:11:49.373062 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-sg4vg_29c0f404-fc7d-4909-b6bf-e1826a092a72/ovs-vswitchd/0.log" Oct 05 07:11:49 crc kubenswrapper[4846]: I1005 07:11:49.374332 4846 generic.go:334] "Generic (PLEG): container finished" podID="29c0f404-fc7d-4909-b6bf-e1826a092a72" containerID="22929dba446bc2516169306e09540fa52839ffc6b8d024f494720cc39ad0ab7d" exitCode=137 Oct 05 07:11:49 crc kubenswrapper[4846]: I1005 07:11:49.374392 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-sg4vg" event={"ID":"29c0f404-fc7d-4909-b6bf-e1826a092a72","Type":"ContainerDied","Data":"22929dba446bc2516169306e09540fa52839ffc6b8d024f494720cc39ad0ab7d"} Oct 05 07:11:49 crc kubenswrapper[4846]: I1005 07:11:49.976015 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-sg4vg_29c0f404-fc7d-4909-b6bf-e1826a092a72/ovs-vswitchd/0.log" Oct 05 07:11:49 crc kubenswrapper[4846]: I1005 07:11:49.977751 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-sg4vg" Oct 05 07:11:49 crc kubenswrapper[4846]: I1005 07:11:49.992787 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Oct 05 07:11:50 crc kubenswrapper[4846]: I1005 07:11:50.065238 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/29c0f404-fc7d-4909-b6bf-e1826a092a72-scripts\") pod \"29c0f404-fc7d-4909-b6bf-e1826a092a72\" (UID: \"29c0f404-fc7d-4909-b6bf-e1826a092a72\") " Oct 05 07:11:50 crc kubenswrapper[4846]: I1005 07:11:50.065530 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/29c0f404-fc7d-4909-b6bf-e1826a092a72-var-log\") pod \"29c0f404-fc7d-4909-b6bf-e1826a092a72\" (UID: \"29c0f404-fc7d-4909-b6bf-e1826a092a72\") " Oct 05 07:11:50 crc kubenswrapper[4846]: I1005 07:11:50.065650 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/29c0f404-fc7d-4909-b6bf-e1826a092a72-etc-ovs\") pod \"29c0f404-fc7d-4909-b6bf-e1826a092a72\" (UID: \"29c0f404-fc7d-4909-b6bf-e1826a092a72\") " Oct 05 07:11:50 crc kubenswrapper[4846]: I1005 07:11:50.065733 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dv6nn\" (UniqueName: \"kubernetes.io/projected/62d2ef82-9372-4b55-990e-30eb8e16e439-kube-api-access-dv6nn\") pod \"62d2ef82-9372-4b55-990e-30eb8e16e439\" (UID: \"62d2ef82-9372-4b55-990e-30eb8e16e439\") " Oct 05 07:11:50 crc kubenswrapper[4846]: I1005 07:11:50.065822 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/62d2ef82-9372-4b55-990e-30eb8e16e439-lock\") pod \"62d2ef82-9372-4b55-990e-30eb8e16e439\" (UID: \"62d2ef82-9372-4b55-990e-30eb8e16e439\") " Oct 05 07:11:50 crc kubenswrapper[4846]: I1005 07:11:50.065910 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swift\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"62d2ef82-9372-4b55-990e-30eb8e16e439\" (UID: \"62d2ef82-9372-4b55-990e-30eb8e16e439\") " Oct 05 07:11:50 crc kubenswrapper[4846]: I1005 07:11:50.066006 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/62d2ef82-9372-4b55-990e-30eb8e16e439-etc-swift\") pod \"62d2ef82-9372-4b55-990e-30eb8e16e439\" (UID: \"62d2ef82-9372-4b55-990e-30eb8e16e439\") " Oct 05 07:11:50 crc kubenswrapper[4846]: I1005 07:11:50.066105 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x9dgh\" (UniqueName: \"kubernetes.io/projected/29c0f404-fc7d-4909-b6bf-e1826a092a72-kube-api-access-x9dgh\") pod \"29c0f404-fc7d-4909-b6bf-e1826a092a72\" (UID: \"29c0f404-fc7d-4909-b6bf-e1826a092a72\") " Oct 05 07:11:50 crc kubenswrapper[4846]: I1005 07:11:50.066266 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/62d2ef82-9372-4b55-990e-30eb8e16e439-cache\") pod \"62d2ef82-9372-4b55-990e-30eb8e16e439\" (UID: \"62d2ef82-9372-4b55-990e-30eb8e16e439\") " Oct 05 07:11:50 crc kubenswrapper[4846]: I1005 07:11:50.066345 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/29c0f404-fc7d-4909-b6bf-e1826a092a72-var-run\") pod \"29c0f404-fc7d-4909-b6bf-e1826a092a72\" (UID: \"29c0f404-fc7d-4909-b6bf-e1826a092a72\") " Oct 05 07:11:50 crc kubenswrapper[4846]: I1005 07:11:50.066437 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/29c0f404-fc7d-4909-b6bf-e1826a092a72-var-lib\") pod \"29c0f404-fc7d-4909-b6bf-e1826a092a72\" (UID: \"29c0f404-fc7d-4909-b6bf-e1826a092a72\") " Oct 05 07:11:50 crc kubenswrapper[4846]: I1005 07:11:50.066868 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/29c0f404-fc7d-4909-b6bf-e1826a092a72-var-lib" (OuterVolumeSpecName: "var-lib") pod "29c0f404-fc7d-4909-b6bf-e1826a092a72" (UID: "29c0f404-fc7d-4909-b6bf-e1826a092a72"). InnerVolumeSpecName "var-lib". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 05 07:11:50 crc kubenswrapper[4846]: I1005 07:11:50.067456 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/62d2ef82-9372-4b55-990e-30eb8e16e439-lock" (OuterVolumeSpecName: "lock") pod "62d2ef82-9372-4b55-990e-30eb8e16e439" (UID: "62d2ef82-9372-4b55-990e-30eb8e16e439"). InnerVolumeSpecName "lock". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 07:11:50 crc kubenswrapper[4846]: I1005 07:11:50.069380 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/29c0f404-fc7d-4909-b6bf-e1826a092a72-var-log" (OuterVolumeSpecName: "var-log") pod "29c0f404-fc7d-4909-b6bf-e1826a092a72" (UID: "29c0f404-fc7d-4909-b6bf-e1826a092a72"). InnerVolumeSpecName "var-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 05 07:11:50 crc kubenswrapper[4846]: I1005 07:11:50.070153 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/62d2ef82-9372-4b55-990e-30eb8e16e439-cache" (OuterVolumeSpecName: "cache") pod "62d2ef82-9372-4b55-990e-30eb8e16e439" (UID: "62d2ef82-9372-4b55-990e-30eb8e16e439"). InnerVolumeSpecName "cache". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 07:11:50 crc kubenswrapper[4846]: I1005 07:11:50.070414 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/29c0f404-fc7d-4909-b6bf-e1826a092a72-var-run" (OuterVolumeSpecName: "var-run") pod "29c0f404-fc7d-4909-b6bf-e1826a092a72" (UID: "29c0f404-fc7d-4909-b6bf-e1826a092a72"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 05 07:11:50 crc kubenswrapper[4846]: I1005 07:11:50.071240 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/29c0f404-fc7d-4909-b6bf-e1826a092a72-scripts" (OuterVolumeSpecName: "scripts") pod "29c0f404-fc7d-4909-b6bf-e1826a092a72" (UID: "29c0f404-fc7d-4909-b6bf-e1826a092a72"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 07:11:50 crc kubenswrapper[4846]: I1005 07:11:50.071357 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/29c0f404-fc7d-4909-b6bf-e1826a092a72-etc-ovs" (OuterVolumeSpecName: "etc-ovs") pod "29c0f404-fc7d-4909-b6bf-e1826a092a72" (UID: "29c0f404-fc7d-4909-b6bf-e1826a092a72"). InnerVolumeSpecName "etc-ovs". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 05 07:11:50 crc kubenswrapper[4846]: I1005 07:11:50.074454 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/62d2ef82-9372-4b55-990e-30eb8e16e439-kube-api-access-dv6nn" (OuterVolumeSpecName: "kube-api-access-dv6nn") pod "62d2ef82-9372-4b55-990e-30eb8e16e439" (UID: "62d2ef82-9372-4b55-990e-30eb8e16e439"). InnerVolumeSpecName "kube-api-access-dv6nn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:11:50 crc kubenswrapper[4846]: I1005 07:11:50.074480 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage06-crc" (OuterVolumeSpecName: "swift") pod "62d2ef82-9372-4b55-990e-30eb8e16e439" (UID: "62d2ef82-9372-4b55-990e-30eb8e16e439"). InnerVolumeSpecName "local-storage06-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 05 07:11:50 crc kubenswrapper[4846]: I1005 07:11:50.074569 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/29c0f404-fc7d-4909-b6bf-e1826a092a72-kube-api-access-x9dgh" (OuterVolumeSpecName: "kube-api-access-x9dgh") pod "29c0f404-fc7d-4909-b6bf-e1826a092a72" (UID: "29c0f404-fc7d-4909-b6bf-e1826a092a72"). InnerVolumeSpecName "kube-api-access-x9dgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:11:50 crc kubenswrapper[4846]: I1005 07:11:50.074690 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/62d2ef82-9372-4b55-990e-30eb8e16e439-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "62d2ef82-9372-4b55-990e-30eb8e16e439" (UID: "62d2ef82-9372-4b55-990e-30eb8e16e439"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:11:50 crc kubenswrapper[4846]: I1005 07:11:50.168216 4846 reconciler_common.go:293] "Volume detached for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/62d2ef82-9372-4b55-990e-30eb8e16e439-lock\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:50 crc kubenswrapper[4846]: I1005 07:11:50.168618 4846 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" " Oct 05 07:11:50 crc kubenswrapper[4846]: I1005 07:11:50.168682 4846 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/62d2ef82-9372-4b55-990e-30eb8e16e439-etc-swift\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:50 crc kubenswrapper[4846]: I1005 07:11:50.168752 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x9dgh\" (UniqueName: \"kubernetes.io/projected/29c0f404-fc7d-4909-b6bf-e1826a092a72-kube-api-access-x9dgh\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:50 crc kubenswrapper[4846]: I1005 07:11:50.168817 4846 reconciler_common.go:293] "Volume detached for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/62d2ef82-9372-4b55-990e-30eb8e16e439-cache\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:50 crc kubenswrapper[4846]: I1005 07:11:50.168901 4846 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/29c0f404-fc7d-4909-b6bf-e1826a092a72-var-run\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:50 crc kubenswrapper[4846]: I1005 07:11:50.168969 4846 reconciler_common.go:293] "Volume detached for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/29c0f404-fc7d-4909-b6bf-e1826a092a72-var-lib\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:50 crc kubenswrapper[4846]: I1005 07:11:50.169021 4846 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/29c0f404-fc7d-4909-b6bf-e1826a092a72-scripts\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:50 crc kubenswrapper[4846]: I1005 07:11:50.169074 4846 reconciler_common.go:293] "Volume detached for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/29c0f404-fc7d-4909-b6bf-e1826a092a72-var-log\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:50 crc kubenswrapper[4846]: I1005 07:11:50.169128 4846 reconciler_common.go:293] "Volume detached for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/29c0f404-fc7d-4909-b6bf-e1826a092a72-etc-ovs\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:50 crc kubenswrapper[4846]: I1005 07:11:50.169196 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dv6nn\" (UniqueName: \"kubernetes.io/projected/62d2ef82-9372-4b55-990e-30eb8e16e439-kube-api-access-dv6nn\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:50 crc kubenswrapper[4846]: I1005 07:11:50.195493 4846 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage06-crc" (UniqueName: "kubernetes.io/local-volume/local-storage06-crc") on node "crc" Oct 05 07:11:50 crc kubenswrapper[4846]: I1005 07:11:50.271453 4846 reconciler_common.go:293] "Volume detached for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" DevicePath \"\"" Oct 05 07:11:50 crc kubenswrapper[4846]: I1005 07:11:50.398416 4846 generic.go:334] "Generic (PLEG): container finished" podID="62d2ef82-9372-4b55-990e-30eb8e16e439" containerID="ccba487ad88a231e17d0bff3850ebdc3dc9887084fa35706507c9470dedbf5df" exitCode=137 Oct 05 07:11:50 crc kubenswrapper[4846]: I1005 07:11:50.398473 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"62d2ef82-9372-4b55-990e-30eb8e16e439","Type":"ContainerDied","Data":"ccba487ad88a231e17d0bff3850ebdc3dc9887084fa35706507c9470dedbf5df"} Oct 05 07:11:50 crc kubenswrapper[4846]: I1005 07:11:50.398650 4846 scope.go:117] "RemoveContainer" containerID="ccba487ad88a231e17d0bff3850ebdc3dc9887084fa35706507c9470dedbf5df" Oct 05 07:11:50 crc kubenswrapper[4846]: I1005 07:11:50.398665 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Oct 05 07:11:50 crc kubenswrapper[4846]: I1005 07:11:50.399215 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"62d2ef82-9372-4b55-990e-30eb8e16e439","Type":"ContainerDied","Data":"ea57f2844ffd33c32881942ebca5ce2841d9fc7ee91c27d3b5e9b78505a3f19b"} Oct 05 07:11:50 crc kubenswrapper[4846]: I1005 07:11:50.402243 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-sg4vg_29c0f404-fc7d-4909-b6bf-e1826a092a72/ovs-vswitchd/0.log" Oct 05 07:11:50 crc kubenswrapper[4846]: I1005 07:11:50.403526 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-sg4vg" event={"ID":"29c0f404-fc7d-4909-b6bf-e1826a092a72","Type":"ContainerDied","Data":"3a7e569f8b29ff1301494b1573cbe42db801dc5571d083cb02917ec6fba8c253"} Oct 05 07:11:50 crc kubenswrapper[4846]: I1005 07:11:50.403742 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-sg4vg" Oct 05 07:11:50 crc kubenswrapper[4846]: I1005 07:11:50.441576 4846 scope.go:117] "RemoveContainer" containerID="8b1645c0bb7485916cb16906718326320b1c0c52c16426fad29ece1124acb1d4" Oct 05 07:11:50 crc kubenswrapper[4846]: I1005 07:11:50.490860 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-storage-0"] Oct 05 07:11:50 crc kubenswrapper[4846]: I1005 07:11:50.492371 4846 scope.go:117] "RemoveContainer" containerID="9af99544264e591f2b45ec0655947d9163b35da963449be03de565a72c5e8e50" Oct 05 07:11:50 crc kubenswrapper[4846]: I1005 07:11:50.515898 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/swift-storage-0"] Oct 05 07:11:50 crc kubenswrapper[4846]: I1005 07:11:50.515982 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-ovs-sg4vg"] Oct 05 07:11:50 crc kubenswrapper[4846]: I1005 07:11:50.521894 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-ovs-sg4vg"] Oct 05 07:11:50 crc kubenswrapper[4846]: I1005 07:11:50.543356 4846 scope.go:117] "RemoveContainer" containerID="983443d1fc33fbdef8e7ca49faf419ed150fada692ed4b5e60d29ffcb9cbaf6c" Oct 05 07:11:50 crc kubenswrapper[4846]: I1005 07:11:50.575884 4846 scope.go:117] "RemoveContainer" containerID="0560e729496f16c1c92b49ae2209e5dcfd4cd0655c6017f90693d429daef0694" Oct 05 07:11:50 crc kubenswrapper[4846]: I1005 07:11:50.598350 4846 scope.go:117] "RemoveContainer" containerID="3e6f56f54a2bfbb877522ef200058490423742afbcf0b93313a239fae7415942" Oct 05 07:11:50 crc kubenswrapper[4846]: I1005 07:11:50.623978 4846 scope.go:117] "RemoveContainer" containerID="b448ad891d0bdb677c147854fe487e87246e96353b5d9e033e8ba9ab2dfa1486" Oct 05 07:11:50 crc kubenswrapper[4846]: I1005 07:11:50.647275 4846 scope.go:117] "RemoveContainer" containerID="3335b438bf1fc84c97c1fad661d249e0c3d1828dffb17a746efb34add2788d6d" Oct 05 07:11:50 crc kubenswrapper[4846]: I1005 07:11:50.676048 4846 scope.go:117] "RemoveContainer" containerID="1af615c099267db200d6c220550e8454d08ffe8449631c9c707f1cdad15b401e" Oct 05 07:11:50 crc kubenswrapper[4846]: I1005 07:11:50.703958 4846 scope.go:117] "RemoveContainer" containerID="19432cd2eb09eb116d05c3d630f9414b8755868167c8ad9d50242b823e184fe6" Oct 05 07:11:50 crc kubenswrapper[4846]: I1005 07:11:50.726156 4846 scope.go:117] "RemoveContainer" containerID="1892fe91962ddb7379e117ea24de43a317054f7a7cd0983a6e06d7d4a4eafef9" Oct 05 07:11:50 crc kubenswrapper[4846]: I1005 07:11:50.747487 4846 scope.go:117] "RemoveContainer" containerID="d7516230c3ea4b249ca682d8efe6e221f59dc03b58d2094da6e825d330b539b4" Oct 05 07:11:50 crc kubenswrapper[4846]: I1005 07:11:50.776394 4846 scope.go:117] "RemoveContainer" containerID="cb25bcb083c8fd00a96201f36477176be16589836d2ad4521c750f525555a9af" Oct 05 07:11:50 crc kubenswrapper[4846]: I1005 07:11:50.807486 4846 scope.go:117] "RemoveContainer" containerID="888a361d7601c9ec3ffa1d314d1a426b937569ba6430ad0b3420461b118db555" Oct 05 07:11:50 crc kubenswrapper[4846]: I1005 07:11:50.835880 4846 scope.go:117] "RemoveContainer" containerID="a4edeadbef779fa82267d6b817cf8a19cf59e7146c6c1a8f5c786ad4fea85d4d" Oct 05 07:11:50 crc kubenswrapper[4846]: I1005 07:11:50.869087 4846 scope.go:117] "RemoveContainer" containerID="ccba487ad88a231e17d0bff3850ebdc3dc9887084fa35706507c9470dedbf5df" Oct 05 07:11:50 crc kubenswrapper[4846]: E1005 07:11:50.869890 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ccba487ad88a231e17d0bff3850ebdc3dc9887084fa35706507c9470dedbf5df\": container with ID starting with ccba487ad88a231e17d0bff3850ebdc3dc9887084fa35706507c9470dedbf5df not found: ID does not exist" containerID="ccba487ad88a231e17d0bff3850ebdc3dc9887084fa35706507c9470dedbf5df" Oct 05 07:11:50 crc kubenswrapper[4846]: I1005 07:11:50.869964 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ccba487ad88a231e17d0bff3850ebdc3dc9887084fa35706507c9470dedbf5df"} err="failed to get container status \"ccba487ad88a231e17d0bff3850ebdc3dc9887084fa35706507c9470dedbf5df\": rpc error: code = NotFound desc = could not find container \"ccba487ad88a231e17d0bff3850ebdc3dc9887084fa35706507c9470dedbf5df\": container with ID starting with ccba487ad88a231e17d0bff3850ebdc3dc9887084fa35706507c9470dedbf5df not found: ID does not exist" Oct 05 07:11:50 crc kubenswrapper[4846]: I1005 07:11:50.870029 4846 scope.go:117] "RemoveContainer" containerID="8b1645c0bb7485916cb16906718326320b1c0c52c16426fad29ece1124acb1d4" Oct 05 07:11:50 crc kubenswrapper[4846]: E1005 07:11:50.870571 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8b1645c0bb7485916cb16906718326320b1c0c52c16426fad29ece1124acb1d4\": container with ID starting with 8b1645c0bb7485916cb16906718326320b1c0c52c16426fad29ece1124acb1d4 not found: ID does not exist" containerID="8b1645c0bb7485916cb16906718326320b1c0c52c16426fad29ece1124acb1d4" Oct 05 07:11:50 crc kubenswrapper[4846]: I1005 07:11:50.870615 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8b1645c0bb7485916cb16906718326320b1c0c52c16426fad29ece1124acb1d4"} err="failed to get container status \"8b1645c0bb7485916cb16906718326320b1c0c52c16426fad29ece1124acb1d4\": rpc error: code = NotFound desc = could not find container \"8b1645c0bb7485916cb16906718326320b1c0c52c16426fad29ece1124acb1d4\": container with ID starting with 8b1645c0bb7485916cb16906718326320b1c0c52c16426fad29ece1124acb1d4 not found: ID does not exist" Oct 05 07:11:50 crc kubenswrapper[4846]: I1005 07:11:50.870658 4846 scope.go:117] "RemoveContainer" containerID="9af99544264e591f2b45ec0655947d9163b35da963449be03de565a72c5e8e50" Oct 05 07:11:50 crc kubenswrapper[4846]: E1005 07:11:50.871772 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9af99544264e591f2b45ec0655947d9163b35da963449be03de565a72c5e8e50\": container with ID starting with 9af99544264e591f2b45ec0655947d9163b35da963449be03de565a72c5e8e50 not found: ID does not exist" containerID="9af99544264e591f2b45ec0655947d9163b35da963449be03de565a72c5e8e50" Oct 05 07:11:50 crc kubenswrapper[4846]: I1005 07:11:50.871815 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9af99544264e591f2b45ec0655947d9163b35da963449be03de565a72c5e8e50"} err="failed to get container status \"9af99544264e591f2b45ec0655947d9163b35da963449be03de565a72c5e8e50\": rpc error: code = NotFound desc = could not find container \"9af99544264e591f2b45ec0655947d9163b35da963449be03de565a72c5e8e50\": container with ID starting with 9af99544264e591f2b45ec0655947d9163b35da963449be03de565a72c5e8e50 not found: ID does not exist" Oct 05 07:11:50 crc kubenswrapper[4846]: I1005 07:11:50.871837 4846 scope.go:117] "RemoveContainer" containerID="983443d1fc33fbdef8e7ca49faf419ed150fada692ed4b5e60d29ffcb9cbaf6c" Oct 05 07:11:50 crc kubenswrapper[4846]: E1005 07:11:50.872203 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"983443d1fc33fbdef8e7ca49faf419ed150fada692ed4b5e60d29ffcb9cbaf6c\": container with ID starting with 983443d1fc33fbdef8e7ca49faf419ed150fada692ed4b5e60d29ffcb9cbaf6c not found: ID does not exist" containerID="983443d1fc33fbdef8e7ca49faf419ed150fada692ed4b5e60d29ffcb9cbaf6c" Oct 05 07:11:50 crc kubenswrapper[4846]: I1005 07:11:50.872240 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"983443d1fc33fbdef8e7ca49faf419ed150fada692ed4b5e60d29ffcb9cbaf6c"} err="failed to get container status \"983443d1fc33fbdef8e7ca49faf419ed150fada692ed4b5e60d29ffcb9cbaf6c\": rpc error: code = NotFound desc = could not find container \"983443d1fc33fbdef8e7ca49faf419ed150fada692ed4b5e60d29ffcb9cbaf6c\": container with ID starting with 983443d1fc33fbdef8e7ca49faf419ed150fada692ed4b5e60d29ffcb9cbaf6c not found: ID does not exist" Oct 05 07:11:50 crc kubenswrapper[4846]: I1005 07:11:50.872259 4846 scope.go:117] "RemoveContainer" containerID="0560e729496f16c1c92b49ae2209e5dcfd4cd0655c6017f90693d429daef0694" Oct 05 07:11:50 crc kubenswrapper[4846]: E1005 07:11:50.872594 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0560e729496f16c1c92b49ae2209e5dcfd4cd0655c6017f90693d429daef0694\": container with ID starting with 0560e729496f16c1c92b49ae2209e5dcfd4cd0655c6017f90693d429daef0694 not found: ID does not exist" containerID="0560e729496f16c1c92b49ae2209e5dcfd4cd0655c6017f90693d429daef0694" Oct 05 07:11:50 crc kubenswrapper[4846]: I1005 07:11:50.872671 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0560e729496f16c1c92b49ae2209e5dcfd4cd0655c6017f90693d429daef0694"} err="failed to get container status \"0560e729496f16c1c92b49ae2209e5dcfd4cd0655c6017f90693d429daef0694\": rpc error: code = NotFound desc = could not find container \"0560e729496f16c1c92b49ae2209e5dcfd4cd0655c6017f90693d429daef0694\": container with ID starting with 0560e729496f16c1c92b49ae2209e5dcfd4cd0655c6017f90693d429daef0694 not found: ID does not exist" Oct 05 07:11:50 crc kubenswrapper[4846]: I1005 07:11:50.872734 4846 scope.go:117] "RemoveContainer" containerID="3e6f56f54a2bfbb877522ef200058490423742afbcf0b93313a239fae7415942" Oct 05 07:11:50 crc kubenswrapper[4846]: E1005 07:11:50.873188 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3e6f56f54a2bfbb877522ef200058490423742afbcf0b93313a239fae7415942\": container with ID starting with 3e6f56f54a2bfbb877522ef200058490423742afbcf0b93313a239fae7415942 not found: ID does not exist" containerID="3e6f56f54a2bfbb877522ef200058490423742afbcf0b93313a239fae7415942" Oct 05 07:11:50 crc kubenswrapper[4846]: I1005 07:11:50.873221 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3e6f56f54a2bfbb877522ef200058490423742afbcf0b93313a239fae7415942"} err="failed to get container status \"3e6f56f54a2bfbb877522ef200058490423742afbcf0b93313a239fae7415942\": rpc error: code = NotFound desc = could not find container \"3e6f56f54a2bfbb877522ef200058490423742afbcf0b93313a239fae7415942\": container with ID starting with 3e6f56f54a2bfbb877522ef200058490423742afbcf0b93313a239fae7415942 not found: ID does not exist" Oct 05 07:11:50 crc kubenswrapper[4846]: I1005 07:11:50.873242 4846 scope.go:117] "RemoveContainer" containerID="b448ad891d0bdb677c147854fe487e87246e96353b5d9e033e8ba9ab2dfa1486" Oct 05 07:11:50 crc kubenswrapper[4846]: E1005 07:11:50.873628 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b448ad891d0bdb677c147854fe487e87246e96353b5d9e033e8ba9ab2dfa1486\": container with ID starting with b448ad891d0bdb677c147854fe487e87246e96353b5d9e033e8ba9ab2dfa1486 not found: ID does not exist" containerID="b448ad891d0bdb677c147854fe487e87246e96353b5d9e033e8ba9ab2dfa1486" Oct 05 07:11:50 crc kubenswrapper[4846]: I1005 07:11:50.873680 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b448ad891d0bdb677c147854fe487e87246e96353b5d9e033e8ba9ab2dfa1486"} err="failed to get container status \"b448ad891d0bdb677c147854fe487e87246e96353b5d9e033e8ba9ab2dfa1486\": rpc error: code = NotFound desc = could not find container \"b448ad891d0bdb677c147854fe487e87246e96353b5d9e033e8ba9ab2dfa1486\": container with ID starting with b448ad891d0bdb677c147854fe487e87246e96353b5d9e033e8ba9ab2dfa1486 not found: ID does not exist" Oct 05 07:11:50 crc kubenswrapper[4846]: I1005 07:11:50.873714 4846 scope.go:117] "RemoveContainer" containerID="3335b438bf1fc84c97c1fad661d249e0c3d1828dffb17a746efb34add2788d6d" Oct 05 07:11:50 crc kubenswrapper[4846]: E1005 07:11:50.874210 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3335b438bf1fc84c97c1fad661d249e0c3d1828dffb17a746efb34add2788d6d\": container with ID starting with 3335b438bf1fc84c97c1fad661d249e0c3d1828dffb17a746efb34add2788d6d not found: ID does not exist" containerID="3335b438bf1fc84c97c1fad661d249e0c3d1828dffb17a746efb34add2788d6d" Oct 05 07:11:50 crc kubenswrapper[4846]: I1005 07:11:50.874255 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3335b438bf1fc84c97c1fad661d249e0c3d1828dffb17a746efb34add2788d6d"} err="failed to get container status \"3335b438bf1fc84c97c1fad661d249e0c3d1828dffb17a746efb34add2788d6d\": rpc error: code = NotFound desc = could not find container \"3335b438bf1fc84c97c1fad661d249e0c3d1828dffb17a746efb34add2788d6d\": container with ID starting with 3335b438bf1fc84c97c1fad661d249e0c3d1828dffb17a746efb34add2788d6d not found: ID does not exist" Oct 05 07:11:50 crc kubenswrapper[4846]: I1005 07:11:50.874278 4846 scope.go:117] "RemoveContainer" containerID="1af615c099267db200d6c220550e8454d08ffe8449631c9c707f1cdad15b401e" Oct 05 07:11:50 crc kubenswrapper[4846]: E1005 07:11:50.874680 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1af615c099267db200d6c220550e8454d08ffe8449631c9c707f1cdad15b401e\": container with ID starting with 1af615c099267db200d6c220550e8454d08ffe8449631c9c707f1cdad15b401e not found: ID does not exist" containerID="1af615c099267db200d6c220550e8454d08ffe8449631c9c707f1cdad15b401e" Oct 05 07:11:50 crc kubenswrapper[4846]: I1005 07:11:50.874713 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1af615c099267db200d6c220550e8454d08ffe8449631c9c707f1cdad15b401e"} err="failed to get container status \"1af615c099267db200d6c220550e8454d08ffe8449631c9c707f1cdad15b401e\": rpc error: code = NotFound desc = could not find container \"1af615c099267db200d6c220550e8454d08ffe8449631c9c707f1cdad15b401e\": container with ID starting with 1af615c099267db200d6c220550e8454d08ffe8449631c9c707f1cdad15b401e not found: ID does not exist" Oct 05 07:11:50 crc kubenswrapper[4846]: I1005 07:11:50.874732 4846 scope.go:117] "RemoveContainer" containerID="19432cd2eb09eb116d05c3d630f9414b8755868167c8ad9d50242b823e184fe6" Oct 05 07:11:50 crc kubenswrapper[4846]: E1005 07:11:50.875318 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"19432cd2eb09eb116d05c3d630f9414b8755868167c8ad9d50242b823e184fe6\": container with ID starting with 19432cd2eb09eb116d05c3d630f9414b8755868167c8ad9d50242b823e184fe6 not found: ID does not exist" containerID="19432cd2eb09eb116d05c3d630f9414b8755868167c8ad9d50242b823e184fe6" Oct 05 07:11:50 crc kubenswrapper[4846]: I1005 07:11:50.875349 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"19432cd2eb09eb116d05c3d630f9414b8755868167c8ad9d50242b823e184fe6"} err="failed to get container status \"19432cd2eb09eb116d05c3d630f9414b8755868167c8ad9d50242b823e184fe6\": rpc error: code = NotFound desc = could not find container \"19432cd2eb09eb116d05c3d630f9414b8755868167c8ad9d50242b823e184fe6\": container with ID starting with 19432cd2eb09eb116d05c3d630f9414b8755868167c8ad9d50242b823e184fe6 not found: ID does not exist" Oct 05 07:11:50 crc kubenswrapper[4846]: I1005 07:11:50.875366 4846 scope.go:117] "RemoveContainer" containerID="1892fe91962ddb7379e117ea24de43a317054f7a7cd0983a6e06d7d4a4eafef9" Oct 05 07:11:50 crc kubenswrapper[4846]: E1005 07:11:50.875775 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1892fe91962ddb7379e117ea24de43a317054f7a7cd0983a6e06d7d4a4eafef9\": container with ID starting with 1892fe91962ddb7379e117ea24de43a317054f7a7cd0983a6e06d7d4a4eafef9 not found: ID does not exist" containerID="1892fe91962ddb7379e117ea24de43a317054f7a7cd0983a6e06d7d4a4eafef9" Oct 05 07:11:50 crc kubenswrapper[4846]: I1005 07:11:50.875851 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1892fe91962ddb7379e117ea24de43a317054f7a7cd0983a6e06d7d4a4eafef9"} err="failed to get container status \"1892fe91962ddb7379e117ea24de43a317054f7a7cd0983a6e06d7d4a4eafef9\": rpc error: code = NotFound desc = could not find container \"1892fe91962ddb7379e117ea24de43a317054f7a7cd0983a6e06d7d4a4eafef9\": container with ID starting with 1892fe91962ddb7379e117ea24de43a317054f7a7cd0983a6e06d7d4a4eafef9 not found: ID does not exist" Oct 05 07:11:50 crc kubenswrapper[4846]: I1005 07:11:50.875911 4846 scope.go:117] "RemoveContainer" containerID="d7516230c3ea4b249ca682d8efe6e221f59dc03b58d2094da6e825d330b539b4" Oct 05 07:11:50 crc kubenswrapper[4846]: E1005 07:11:50.876398 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d7516230c3ea4b249ca682d8efe6e221f59dc03b58d2094da6e825d330b539b4\": container with ID starting with d7516230c3ea4b249ca682d8efe6e221f59dc03b58d2094da6e825d330b539b4 not found: ID does not exist" containerID="d7516230c3ea4b249ca682d8efe6e221f59dc03b58d2094da6e825d330b539b4" Oct 05 07:11:50 crc kubenswrapper[4846]: I1005 07:11:50.876452 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d7516230c3ea4b249ca682d8efe6e221f59dc03b58d2094da6e825d330b539b4"} err="failed to get container status \"d7516230c3ea4b249ca682d8efe6e221f59dc03b58d2094da6e825d330b539b4\": rpc error: code = NotFound desc = could not find container \"d7516230c3ea4b249ca682d8efe6e221f59dc03b58d2094da6e825d330b539b4\": container with ID starting with d7516230c3ea4b249ca682d8efe6e221f59dc03b58d2094da6e825d330b539b4 not found: ID does not exist" Oct 05 07:11:50 crc kubenswrapper[4846]: I1005 07:11:50.876487 4846 scope.go:117] "RemoveContainer" containerID="cb25bcb083c8fd00a96201f36477176be16589836d2ad4521c750f525555a9af" Oct 05 07:11:50 crc kubenswrapper[4846]: E1005 07:11:50.876802 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cb25bcb083c8fd00a96201f36477176be16589836d2ad4521c750f525555a9af\": container with ID starting with cb25bcb083c8fd00a96201f36477176be16589836d2ad4521c750f525555a9af not found: ID does not exist" containerID="cb25bcb083c8fd00a96201f36477176be16589836d2ad4521c750f525555a9af" Oct 05 07:11:50 crc kubenswrapper[4846]: I1005 07:11:50.876836 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cb25bcb083c8fd00a96201f36477176be16589836d2ad4521c750f525555a9af"} err="failed to get container status \"cb25bcb083c8fd00a96201f36477176be16589836d2ad4521c750f525555a9af\": rpc error: code = NotFound desc = could not find container \"cb25bcb083c8fd00a96201f36477176be16589836d2ad4521c750f525555a9af\": container with ID starting with cb25bcb083c8fd00a96201f36477176be16589836d2ad4521c750f525555a9af not found: ID does not exist" Oct 05 07:11:50 crc kubenswrapper[4846]: I1005 07:11:50.876854 4846 scope.go:117] "RemoveContainer" containerID="888a361d7601c9ec3ffa1d314d1a426b937569ba6430ad0b3420461b118db555" Oct 05 07:11:50 crc kubenswrapper[4846]: E1005 07:11:50.877128 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"888a361d7601c9ec3ffa1d314d1a426b937569ba6430ad0b3420461b118db555\": container with ID starting with 888a361d7601c9ec3ffa1d314d1a426b937569ba6430ad0b3420461b118db555 not found: ID does not exist" containerID="888a361d7601c9ec3ffa1d314d1a426b937569ba6430ad0b3420461b118db555" Oct 05 07:11:50 crc kubenswrapper[4846]: I1005 07:11:50.877164 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"888a361d7601c9ec3ffa1d314d1a426b937569ba6430ad0b3420461b118db555"} err="failed to get container status \"888a361d7601c9ec3ffa1d314d1a426b937569ba6430ad0b3420461b118db555\": rpc error: code = NotFound desc = could not find container \"888a361d7601c9ec3ffa1d314d1a426b937569ba6430ad0b3420461b118db555\": container with ID starting with 888a361d7601c9ec3ffa1d314d1a426b937569ba6430ad0b3420461b118db555 not found: ID does not exist" Oct 05 07:11:50 crc kubenswrapper[4846]: I1005 07:11:50.877219 4846 scope.go:117] "RemoveContainer" containerID="a4edeadbef779fa82267d6b817cf8a19cf59e7146c6c1a8f5c786ad4fea85d4d" Oct 05 07:11:50 crc kubenswrapper[4846]: E1005 07:11:50.877555 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a4edeadbef779fa82267d6b817cf8a19cf59e7146c6c1a8f5c786ad4fea85d4d\": container with ID starting with a4edeadbef779fa82267d6b817cf8a19cf59e7146c6c1a8f5c786ad4fea85d4d not found: ID does not exist" containerID="a4edeadbef779fa82267d6b817cf8a19cf59e7146c6c1a8f5c786ad4fea85d4d" Oct 05 07:11:50 crc kubenswrapper[4846]: I1005 07:11:50.877580 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a4edeadbef779fa82267d6b817cf8a19cf59e7146c6c1a8f5c786ad4fea85d4d"} err="failed to get container status \"a4edeadbef779fa82267d6b817cf8a19cf59e7146c6c1a8f5c786ad4fea85d4d\": rpc error: code = NotFound desc = could not find container \"a4edeadbef779fa82267d6b817cf8a19cf59e7146c6c1a8f5c786ad4fea85d4d\": container with ID starting with a4edeadbef779fa82267d6b817cf8a19cf59e7146c6c1a8f5c786ad4fea85d4d not found: ID does not exist" Oct 05 07:11:50 crc kubenswrapper[4846]: I1005 07:11:50.877601 4846 scope.go:117] "RemoveContainer" containerID="22929dba446bc2516169306e09540fa52839ffc6b8d024f494720cc39ad0ab7d" Oct 05 07:11:50 crc kubenswrapper[4846]: I1005 07:11:50.900974 4846 scope.go:117] "RemoveContainer" containerID="f4fa8732ebd0803a3d57af2df182505f645e900471768d31fb1d43cb0a83e0cf" Oct 05 07:11:50 crc kubenswrapper[4846]: I1005 07:11:50.929526 4846 scope.go:117] "RemoveContainer" containerID="0bb4176d1021478fccab08b6b67c5b5b6ecac6ad30df0ab98cc66c771c4ba7a4" Oct 05 07:11:52 crc kubenswrapper[4846]: I1005 07:11:52.517523 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="29c0f404-fc7d-4909-b6bf-e1826a092a72" path="/var/lib/kubelet/pods/29c0f404-fc7d-4909-b6bf-e1826a092a72/volumes" Oct 05 07:11:52 crc kubenswrapper[4846]: I1005 07:11:52.518871 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="62d2ef82-9372-4b55-990e-30eb8e16e439" path="/var/lib/kubelet/pods/62d2ef82-9372-4b55-990e-30eb8e16e439/volumes" Oct 05 07:12:03 crc kubenswrapper[4846]: I1005 07:12:03.538476 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-r7lpx"] Oct 05 07:12:03 crc kubenswrapper[4846]: E1005 07:12:03.539486 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="62d2ef82-9372-4b55-990e-30eb8e16e439" containerName="account-reaper" Oct 05 07:12:03 crc kubenswrapper[4846]: I1005 07:12:03.539506 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="62d2ef82-9372-4b55-990e-30eb8e16e439" containerName="account-reaper" Oct 05 07:12:03 crc kubenswrapper[4846]: E1005 07:12:03.539520 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="18a7be37-c8dc-4d8a-808f-2a63eb4d9692" containerName="openstack-network-exporter" Oct 05 07:12:03 crc kubenswrapper[4846]: I1005 07:12:03.539528 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="18a7be37-c8dc-4d8a-808f-2a63eb4d9692" containerName="openstack-network-exporter" Oct 05 07:12:03 crc kubenswrapper[4846]: E1005 07:12:03.539538 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a72c0cf5-8cd9-4f5f-8a2b-25bf41ad0638" containerName="neutron-api" Oct 05 07:12:03 crc kubenswrapper[4846]: I1005 07:12:03.539547 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="a72c0cf5-8cd9-4f5f-8a2b-25bf41ad0638" containerName="neutron-api" Oct 05 07:12:03 crc kubenswrapper[4846]: E1005 07:12:03.539563 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8d55f55a-2d79-4c01-8b0f-d67cefb842ff" containerName="barbican-keystone-listener" Oct 05 07:12:03 crc kubenswrapper[4846]: I1005 07:12:03.539570 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="8d55f55a-2d79-4c01-8b0f-d67cefb842ff" containerName="barbican-keystone-listener" Oct 05 07:12:03 crc kubenswrapper[4846]: E1005 07:12:03.539583 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b1e394dd-f55b-4af1-af0b-52cc7d09afa1" containerName="sg-core" Oct 05 07:12:03 crc kubenswrapper[4846]: I1005 07:12:03.539592 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="b1e394dd-f55b-4af1-af0b-52cc7d09afa1" containerName="sg-core" Oct 05 07:12:03 crc kubenswrapper[4846]: E1005 07:12:03.539618 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="62d2ef82-9372-4b55-990e-30eb8e16e439" containerName="container-server" Oct 05 07:12:03 crc kubenswrapper[4846]: I1005 07:12:03.539630 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="62d2ef82-9372-4b55-990e-30eb8e16e439" containerName="container-server" Oct 05 07:12:03 crc kubenswrapper[4846]: E1005 07:12:03.539649 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4d3616b4-9f2c-4c62-9760-0300b02d0dbb" containerName="mariadb-account-delete" Oct 05 07:12:03 crc kubenswrapper[4846]: I1005 07:12:03.539659 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="4d3616b4-9f2c-4c62-9760-0300b02d0dbb" containerName="mariadb-account-delete" Oct 05 07:12:03 crc kubenswrapper[4846]: E1005 07:12:03.539672 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="62d2ef82-9372-4b55-990e-30eb8e16e439" containerName="object-replicator" Oct 05 07:12:03 crc kubenswrapper[4846]: I1005 07:12:03.539682 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="62d2ef82-9372-4b55-990e-30eb8e16e439" containerName="object-replicator" Oct 05 07:12:03 crc kubenswrapper[4846]: E1005 07:12:03.539705 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="57b09608-08e5-4998-a3b3-e88b13587f58" containerName="kube-state-metrics" Oct 05 07:12:03 crc kubenswrapper[4846]: I1005 07:12:03.539716 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="57b09608-08e5-4998-a3b3-e88b13587f58" containerName="kube-state-metrics" Oct 05 07:12:03 crc kubenswrapper[4846]: E1005 07:12:03.539727 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="62d2ef82-9372-4b55-990e-30eb8e16e439" containerName="swift-recon-cron" Oct 05 07:12:03 crc kubenswrapper[4846]: I1005 07:12:03.539739 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="62d2ef82-9372-4b55-990e-30eb8e16e439" containerName="swift-recon-cron" Oct 05 07:12:03 crc kubenswrapper[4846]: E1005 07:12:03.539760 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b1e394dd-f55b-4af1-af0b-52cc7d09afa1" containerName="proxy-httpd" Oct 05 07:12:03 crc kubenswrapper[4846]: I1005 07:12:03.539772 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="b1e394dd-f55b-4af1-af0b-52cc7d09afa1" containerName="proxy-httpd" Oct 05 07:12:03 crc kubenswrapper[4846]: E1005 07:12:03.539786 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="35261dbc-cabe-4bf0-88f6-b8613ddec0f1" containerName="rabbitmq" Oct 05 07:12:03 crc kubenswrapper[4846]: I1005 07:12:03.539797 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="35261dbc-cabe-4bf0-88f6-b8613ddec0f1" containerName="rabbitmq" Oct 05 07:12:03 crc kubenswrapper[4846]: E1005 07:12:03.539809 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="62d2ef82-9372-4b55-990e-30eb8e16e439" containerName="object-expirer" Oct 05 07:12:03 crc kubenswrapper[4846]: I1005 07:12:03.539819 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="62d2ef82-9372-4b55-990e-30eb8e16e439" containerName="object-expirer" Oct 05 07:12:03 crc kubenswrapper[4846]: E1005 07:12:03.539835 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="751734cb-38c8-46a0-9bac-05bcf24cc589" containerName="memcached" Oct 05 07:12:03 crc kubenswrapper[4846]: I1005 07:12:03.539845 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="751734cb-38c8-46a0-9bac-05bcf24cc589" containerName="memcached" Oct 05 07:12:03 crc kubenswrapper[4846]: E1005 07:12:03.539861 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a72c0cf5-8cd9-4f5f-8a2b-25bf41ad0638" containerName="neutron-httpd" Oct 05 07:12:03 crc kubenswrapper[4846]: I1005 07:12:03.539871 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="a72c0cf5-8cd9-4f5f-8a2b-25bf41ad0638" containerName="neutron-httpd" Oct 05 07:12:03 crc kubenswrapper[4846]: E1005 07:12:03.539886 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="35261dbc-cabe-4bf0-88f6-b8613ddec0f1" containerName="setup-container" Oct 05 07:12:03 crc kubenswrapper[4846]: I1005 07:12:03.539897 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="35261dbc-cabe-4bf0-88f6-b8613ddec0f1" containerName="setup-container" Oct 05 07:12:03 crc kubenswrapper[4846]: E1005 07:12:03.539908 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b1e394dd-f55b-4af1-af0b-52cc7d09afa1" containerName="ceilometer-notification-agent" Oct 05 07:12:03 crc kubenswrapper[4846]: I1005 07:12:03.539921 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="b1e394dd-f55b-4af1-af0b-52cc7d09afa1" containerName="ceilometer-notification-agent" Oct 05 07:12:03 crc kubenswrapper[4846]: E1005 07:12:03.539938 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d88c48f9-948e-4888-986e-e00163ec93a6" containerName="mysql-bootstrap" Oct 05 07:12:03 crc kubenswrapper[4846]: I1005 07:12:03.539948 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="d88c48f9-948e-4888-986e-e00163ec93a6" containerName="mysql-bootstrap" Oct 05 07:12:03 crc kubenswrapper[4846]: E1005 07:12:03.539964 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d88c48f9-948e-4888-986e-e00163ec93a6" containerName="galera" Oct 05 07:12:03 crc kubenswrapper[4846]: I1005 07:12:03.539980 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="d88c48f9-948e-4888-986e-e00163ec93a6" containerName="galera" Oct 05 07:12:03 crc kubenswrapper[4846]: E1005 07:12:03.539996 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="29c0f404-fc7d-4909-b6bf-e1826a092a72" containerName="ovsdb-server" Oct 05 07:12:03 crc kubenswrapper[4846]: I1005 07:12:03.540005 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="29c0f404-fc7d-4909-b6bf-e1826a092a72" containerName="ovsdb-server" Oct 05 07:12:03 crc kubenswrapper[4846]: E1005 07:12:03.540023 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="62d2ef82-9372-4b55-990e-30eb8e16e439" containerName="container-replicator" Oct 05 07:12:03 crc kubenswrapper[4846]: I1005 07:12:03.540032 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="62d2ef82-9372-4b55-990e-30eb8e16e439" containerName="container-replicator" Oct 05 07:12:03 crc kubenswrapper[4846]: E1005 07:12:03.540052 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="169918d8-597e-4ef4-b00c-fb14c4d0347f" containerName="ovn-controller" Oct 05 07:12:03 crc kubenswrapper[4846]: I1005 07:12:03.540061 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="169918d8-597e-4ef4-b00c-fb14c4d0347f" containerName="ovn-controller" Oct 05 07:12:03 crc kubenswrapper[4846]: E1005 07:12:03.540074 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8d55f55a-2d79-4c01-8b0f-d67cefb842ff" containerName="barbican-keystone-listener-log" Oct 05 07:12:03 crc kubenswrapper[4846]: I1005 07:12:03.540085 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="8d55f55a-2d79-4c01-8b0f-d67cefb842ff" containerName="barbican-keystone-listener-log" Oct 05 07:12:03 crc kubenswrapper[4846]: E1005 07:12:03.540102 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="98b50269-c6d4-42f2-8cef-afc127bfccb1" containerName="nova-cell0-conductor-conductor" Oct 05 07:12:03 crc kubenswrapper[4846]: I1005 07:12:03.540112 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="98b50269-c6d4-42f2-8cef-afc127bfccb1" containerName="nova-cell0-conductor-conductor" Oct 05 07:12:03 crc kubenswrapper[4846]: E1005 07:12:03.540128 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1fa90810-c620-48c2-a49c-762341889743" containerName="keystone-api" Oct 05 07:12:03 crc kubenswrapper[4846]: I1005 07:12:03.540138 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="1fa90810-c620-48c2-a49c-762341889743" containerName="keystone-api" Oct 05 07:12:03 crc kubenswrapper[4846]: E1005 07:12:03.540150 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="29c0f404-fc7d-4909-b6bf-e1826a092a72" containerName="ovsdb-server-init" Oct 05 07:12:03 crc kubenswrapper[4846]: I1005 07:12:03.540159 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="29c0f404-fc7d-4909-b6bf-e1826a092a72" containerName="ovsdb-server-init" Oct 05 07:12:03 crc kubenswrapper[4846]: E1005 07:12:03.540172 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="62d2ef82-9372-4b55-990e-30eb8e16e439" containerName="account-replicator" Oct 05 07:12:03 crc kubenswrapper[4846]: I1005 07:12:03.540211 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="62d2ef82-9372-4b55-990e-30eb8e16e439" containerName="account-replicator" Oct 05 07:12:03 crc kubenswrapper[4846]: E1005 07:12:03.540225 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="18a7be37-c8dc-4d8a-808f-2a63eb4d9692" containerName="ovn-northd" Oct 05 07:12:03 crc kubenswrapper[4846]: I1005 07:12:03.540236 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="18a7be37-c8dc-4d8a-808f-2a63eb4d9692" containerName="ovn-northd" Oct 05 07:12:03 crc kubenswrapper[4846]: E1005 07:12:03.540252 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="626e6bce-7422-4fbd-b978-dfa5cb93de34" containerName="rabbitmq" Oct 05 07:12:03 crc kubenswrapper[4846]: I1005 07:12:03.540262 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="626e6bce-7422-4fbd-b978-dfa5cb93de34" containerName="rabbitmq" Oct 05 07:12:03 crc kubenswrapper[4846]: E1005 07:12:03.540277 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="62d2ef82-9372-4b55-990e-30eb8e16e439" containerName="rsync" Oct 05 07:12:03 crc kubenswrapper[4846]: I1005 07:12:03.540287 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="62d2ef82-9372-4b55-990e-30eb8e16e439" containerName="rsync" Oct 05 07:12:03 crc kubenswrapper[4846]: E1005 07:12:03.540302 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3995994c-2a3f-400f-9a64-b564759b20b2" containerName="barbican-worker-log" Oct 05 07:12:03 crc kubenswrapper[4846]: I1005 07:12:03.540311 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="3995994c-2a3f-400f-9a64-b564759b20b2" containerName="barbican-worker-log" Oct 05 07:12:03 crc kubenswrapper[4846]: E1005 07:12:03.540324 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="29c0f404-fc7d-4909-b6bf-e1826a092a72" containerName="ovs-vswitchd" Oct 05 07:12:03 crc kubenswrapper[4846]: I1005 07:12:03.540336 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="29c0f404-fc7d-4909-b6bf-e1826a092a72" containerName="ovs-vswitchd" Oct 05 07:12:03 crc kubenswrapper[4846]: E1005 07:12:03.540349 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="62d2ef82-9372-4b55-990e-30eb8e16e439" containerName="account-auditor" Oct 05 07:12:03 crc kubenswrapper[4846]: I1005 07:12:03.540359 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="62d2ef82-9372-4b55-990e-30eb8e16e439" containerName="account-auditor" Oct 05 07:12:03 crc kubenswrapper[4846]: E1005 07:12:03.540375 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="62d2ef82-9372-4b55-990e-30eb8e16e439" containerName="object-updater" Oct 05 07:12:03 crc kubenswrapper[4846]: I1005 07:12:03.540385 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="62d2ef82-9372-4b55-990e-30eb8e16e439" containerName="object-updater" Oct 05 07:12:03 crc kubenswrapper[4846]: E1005 07:12:03.540400 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="62d2ef82-9372-4b55-990e-30eb8e16e439" containerName="account-server" Oct 05 07:12:03 crc kubenswrapper[4846]: I1005 07:12:03.540410 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="62d2ef82-9372-4b55-990e-30eb8e16e439" containerName="account-server" Oct 05 07:12:03 crc kubenswrapper[4846]: E1005 07:12:03.540427 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3995994c-2a3f-400f-9a64-b564759b20b2" containerName="barbican-worker" Oct 05 07:12:03 crc kubenswrapper[4846]: I1005 07:12:03.540437 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="3995994c-2a3f-400f-9a64-b564759b20b2" containerName="barbican-worker" Oct 05 07:12:03 crc kubenswrapper[4846]: E1005 07:12:03.540451 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="62d2ef82-9372-4b55-990e-30eb8e16e439" containerName="object-server" Oct 05 07:12:03 crc kubenswrapper[4846]: I1005 07:12:03.540462 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="62d2ef82-9372-4b55-990e-30eb8e16e439" containerName="object-server" Oct 05 07:12:03 crc kubenswrapper[4846]: E1005 07:12:03.540481 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="62d2ef82-9372-4b55-990e-30eb8e16e439" containerName="container-updater" Oct 05 07:12:03 crc kubenswrapper[4846]: I1005 07:12:03.540490 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="62d2ef82-9372-4b55-990e-30eb8e16e439" containerName="container-updater" Oct 05 07:12:03 crc kubenswrapper[4846]: E1005 07:12:03.540500 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="62d2ef82-9372-4b55-990e-30eb8e16e439" containerName="object-auditor" Oct 05 07:12:03 crc kubenswrapper[4846]: I1005 07:12:03.540510 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="62d2ef82-9372-4b55-990e-30eb8e16e439" containerName="object-auditor" Oct 05 07:12:03 crc kubenswrapper[4846]: E1005 07:12:03.540529 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b1e394dd-f55b-4af1-af0b-52cc7d09afa1" containerName="ceilometer-central-agent" Oct 05 07:12:03 crc kubenswrapper[4846]: I1005 07:12:03.540539 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="b1e394dd-f55b-4af1-af0b-52cc7d09afa1" containerName="ceilometer-central-agent" Oct 05 07:12:03 crc kubenswrapper[4846]: E1005 07:12:03.540559 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="626e6bce-7422-4fbd-b978-dfa5cb93de34" containerName="setup-container" Oct 05 07:12:03 crc kubenswrapper[4846]: I1005 07:12:03.540569 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="626e6bce-7422-4fbd-b978-dfa5cb93de34" containerName="setup-container" Oct 05 07:12:03 crc kubenswrapper[4846]: E1005 07:12:03.540587 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="62d2ef82-9372-4b55-990e-30eb8e16e439" containerName="container-auditor" Oct 05 07:12:03 crc kubenswrapper[4846]: I1005 07:12:03.540595 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="62d2ef82-9372-4b55-990e-30eb8e16e439" containerName="container-auditor" Oct 05 07:12:03 crc kubenswrapper[4846]: I1005 07:12:03.540809 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="751734cb-38c8-46a0-9bac-05bcf24cc589" containerName="memcached" Oct 05 07:12:03 crc kubenswrapper[4846]: I1005 07:12:03.540836 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="18a7be37-c8dc-4d8a-808f-2a63eb4d9692" containerName="ovn-northd" Oct 05 07:12:03 crc kubenswrapper[4846]: I1005 07:12:03.540847 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="62d2ef82-9372-4b55-990e-30eb8e16e439" containerName="account-reaper" Oct 05 07:12:03 crc kubenswrapper[4846]: I1005 07:12:03.540861 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="4d3616b4-9f2c-4c62-9760-0300b02d0dbb" containerName="mariadb-account-delete" Oct 05 07:12:03 crc kubenswrapper[4846]: I1005 07:12:03.540873 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="62d2ef82-9372-4b55-990e-30eb8e16e439" containerName="container-auditor" Oct 05 07:12:03 crc kubenswrapper[4846]: I1005 07:12:03.540886 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="62d2ef82-9372-4b55-990e-30eb8e16e439" containerName="account-auditor" Oct 05 07:12:03 crc kubenswrapper[4846]: I1005 07:12:03.540898 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="62d2ef82-9372-4b55-990e-30eb8e16e439" containerName="container-server" Oct 05 07:12:03 crc kubenswrapper[4846]: I1005 07:12:03.540912 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="62d2ef82-9372-4b55-990e-30eb8e16e439" containerName="container-replicator" Oct 05 07:12:03 crc kubenswrapper[4846]: I1005 07:12:03.540928 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="626e6bce-7422-4fbd-b978-dfa5cb93de34" containerName="rabbitmq" Oct 05 07:12:03 crc kubenswrapper[4846]: I1005 07:12:03.540941 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="29c0f404-fc7d-4909-b6bf-e1826a092a72" containerName="ovs-vswitchd" Oct 05 07:12:03 crc kubenswrapper[4846]: I1005 07:12:03.540955 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="b1e394dd-f55b-4af1-af0b-52cc7d09afa1" containerName="sg-core" Oct 05 07:12:03 crc kubenswrapper[4846]: I1005 07:12:03.540968 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="35261dbc-cabe-4bf0-88f6-b8613ddec0f1" containerName="rabbitmq" Oct 05 07:12:03 crc kubenswrapper[4846]: I1005 07:12:03.540979 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="1fa90810-c620-48c2-a49c-762341889743" containerName="keystone-api" Oct 05 07:12:03 crc kubenswrapper[4846]: I1005 07:12:03.540999 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="62d2ef82-9372-4b55-990e-30eb8e16e439" containerName="account-replicator" Oct 05 07:12:03 crc kubenswrapper[4846]: I1005 07:12:03.541011 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="b1e394dd-f55b-4af1-af0b-52cc7d09afa1" containerName="ceilometer-central-agent" Oct 05 07:12:03 crc kubenswrapper[4846]: I1005 07:12:03.541026 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="29c0f404-fc7d-4909-b6bf-e1826a092a72" containerName="ovsdb-server" Oct 05 07:12:03 crc kubenswrapper[4846]: I1005 07:12:03.541043 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="62d2ef82-9372-4b55-990e-30eb8e16e439" containerName="account-server" Oct 05 07:12:03 crc kubenswrapper[4846]: I1005 07:12:03.541067 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="62d2ef82-9372-4b55-990e-30eb8e16e439" containerName="rsync" Oct 05 07:12:03 crc kubenswrapper[4846]: I1005 07:12:03.541083 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="18a7be37-c8dc-4d8a-808f-2a63eb4d9692" containerName="openstack-network-exporter" Oct 05 07:12:03 crc kubenswrapper[4846]: I1005 07:12:03.541098 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="a72c0cf5-8cd9-4f5f-8a2b-25bf41ad0638" containerName="neutron-httpd" Oct 05 07:12:03 crc kubenswrapper[4846]: I1005 07:12:03.541114 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="62d2ef82-9372-4b55-990e-30eb8e16e439" containerName="swift-recon-cron" Oct 05 07:12:03 crc kubenswrapper[4846]: I1005 07:12:03.541128 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="b1e394dd-f55b-4af1-af0b-52cc7d09afa1" containerName="proxy-httpd" Oct 05 07:12:03 crc kubenswrapper[4846]: I1005 07:12:03.541140 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="8d55f55a-2d79-4c01-8b0f-d67cefb842ff" containerName="barbican-keystone-listener" Oct 05 07:12:03 crc kubenswrapper[4846]: I1005 07:12:03.541152 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="62d2ef82-9372-4b55-990e-30eb8e16e439" containerName="object-auditor" Oct 05 07:12:03 crc kubenswrapper[4846]: I1005 07:12:03.541168 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="62d2ef82-9372-4b55-990e-30eb8e16e439" containerName="object-expirer" Oct 05 07:12:03 crc kubenswrapper[4846]: I1005 07:12:03.541522 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="8d55f55a-2d79-4c01-8b0f-d67cefb842ff" containerName="barbican-keystone-listener-log" Oct 05 07:12:03 crc kubenswrapper[4846]: I1005 07:12:03.541547 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="62d2ef82-9372-4b55-990e-30eb8e16e439" containerName="object-server" Oct 05 07:12:03 crc kubenswrapper[4846]: I1005 07:12:03.541562 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="d88c48f9-948e-4888-986e-e00163ec93a6" containerName="galera" Oct 05 07:12:03 crc kubenswrapper[4846]: I1005 07:12:03.541578 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="a72c0cf5-8cd9-4f5f-8a2b-25bf41ad0638" containerName="neutron-api" Oct 05 07:12:03 crc kubenswrapper[4846]: I1005 07:12:03.541591 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="169918d8-597e-4ef4-b00c-fb14c4d0347f" containerName="ovn-controller" Oct 05 07:12:03 crc kubenswrapper[4846]: I1005 07:12:03.541608 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="98b50269-c6d4-42f2-8cef-afc127bfccb1" containerName="nova-cell0-conductor-conductor" Oct 05 07:12:03 crc kubenswrapper[4846]: I1005 07:12:03.541622 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="57b09608-08e5-4998-a3b3-e88b13587f58" containerName="kube-state-metrics" Oct 05 07:12:03 crc kubenswrapper[4846]: I1005 07:12:03.541636 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="62d2ef82-9372-4b55-990e-30eb8e16e439" containerName="object-replicator" Oct 05 07:12:03 crc kubenswrapper[4846]: I1005 07:12:03.541653 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="3995994c-2a3f-400f-9a64-b564759b20b2" containerName="barbican-worker" Oct 05 07:12:03 crc kubenswrapper[4846]: I1005 07:12:03.541666 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="3995994c-2a3f-400f-9a64-b564759b20b2" containerName="barbican-worker-log" Oct 05 07:12:03 crc kubenswrapper[4846]: I1005 07:12:03.541678 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="62d2ef82-9372-4b55-990e-30eb8e16e439" containerName="container-updater" Oct 05 07:12:03 crc kubenswrapper[4846]: I1005 07:12:03.541706 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="b1e394dd-f55b-4af1-af0b-52cc7d09afa1" containerName="ceilometer-notification-agent" Oct 05 07:12:03 crc kubenswrapper[4846]: I1005 07:12:03.541723 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="62d2ef82-9372-4b55-990e-30eb8e16e439" containerName="object-updater" Oct 05 07:12:03 crc kubenswrapper[4846]: I1005 07:12:03.543222 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-r7lpx" Oct 05 07:12:03 crc kubenswrapper[4846]: I1005 07:12:03.570377 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-r7lpx"] Oct 05 07:12:03 crc kubenswrapper[4846]: I1005 07:12:03.593258 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7q9mv\" (UniqueName: \"kubernetes.io/projected/882ccdb3-10bb-4e83-88e0-665534256b6d-kube-api-access-7q9mv\") pod \"certified-operators-r7lpx\" (UID: \"882ccdb3-10bb-4e83-88e0-665534256b6d\") " pod="openshift-marketplace/certified-operators-r7lpx" Oct 05 07:12:03 crc kubenswrapper[4846]: I1005 07:12:03.593307 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/882ccdb3-10bb-4e83-88e0-665534256b6d-utilities\") pod \"certified-operators-r7lpx\" (UID: \"882ccdb3-10bb-4e83-88e0-665534256b6d\") " pod="openshift-marketplace/certified-operators-r7lpx" Oct 05 07:12:03 crc kubenswrapper[4846]: I1005 07:12:03.593344 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/882ccdb3-10bb-4e83-88e0-665534256b6d-catalog-content\") pod \"certified-operators-r7lpx\" (UID: \"882ccdb3-10bb-4e83-88e0-665534256b6d\") " pod="openshift-marketplace/certified-operators-r7lpx" Oct 05 07:12:03 crc kubenswrapper[4846]: I1005 07:12:03.694087 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7q9mv\" (UniqueName: \"kubernetes.io/projected/882ccdb3-10bb-4e83-88e0-665534256b6d-kube-api-access-7q9mv\") pod \"certified-operators-r7lpx\" (UID: \"882ccdb3-10bb-4e83-88e0-665534256b6d\") " pod="openshift-marketplace/certified-operators-r7lpx" Oct 05 07:12:03 crc kubenswrapper[4846]: I1005 07:12:03.694165 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/882ccdb3-10bb-4e83-88e0-665534256b6d-utilities\") pod \"certified-operators-r7lpx\" (UID: \"882ccdb3-10bb-4e83-88e0-665534256b6d\") " pod="openshift-marketplace/certified-operators-r7lpx" Oct 05 07:12:03 crc kubenswrapper[4846]: I1005 07:12:03.694243 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/882ccdb3-10bb-4e83-88e0-665534256b6d-catalog-content\") pod \"certified-operators-r7lpx\" (UID: \"882ccdb3-10bb-4e83-88e0-665534256b6d\") " pod="openshift-marketplace/certified-operators-r7lpx" Oct 05 07:12:03 crc kubenswrapper[4846]: I1005 07:12:03.694811 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/882ccdb3-10bb-4e83-88e0-665534256b6d-utilities\") pod \"certified-operators-r7lpx\" (UID: \"882ccdb3-10bb-4e83-88e0-665534256b6d\") " pod="openshift-marketplace/certified-operators-r7lpx" Oct 05 07:12:03 crc kubenswrapper[4846]: I1005 07:12:03.694840 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/882ccdb3-10bb-4e83-88e0-665534256b6d-catalog-content\") pod \"certified-operators-r7lpx\" (UID: \"882ccdb3-10bb-4e83-88e0-665534256b6d\") " pod="openshift-marketplace/certified-operators-r7lpx" Oct 05 07:12:03 crc kubenswrapper[4846]: I1005 07:12:03.713579 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7q9mv\" (UniqueName: \"kubernetes.io/projected/882ccdb3-10bb-4e83-88e0-665534256b6d-kube-api-access-7q9mv\") pod \"certified-operators-r7lpx\" (UID: \"882ccdb3-10bb-4e83-88e0-665534256b6d\") " pod="openshift-marketplace/certified-operators-r7lpx" Oct 05 07:12:03 crc kubenswrapper[4846]: I1005 07:12:03.869109 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-r7lpx" Oct 05 07:12:04 crc kubenswrapper[4846]: I1005 07:12:04.134342 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-r7lpx"] Oct 05 07:12:04 crc kubenswrapper[4846]: I1005 07:12:04.576261 4846 generic.go:334] "Generic (PLEG): container finished" podID="882ccdb3-10bb-4e83-88e0-665534256b6d" containerID="457ac650d9c63c04be8b24daaaae5c00a30ddd923f4ad9229cc56064cd5870f6" exitCode=0 Oct 05 07:12:04 crc kubenswrapper[4846]: I1005 07:12:04.576318 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-r7lpx" event={"ID":"882ccdb3-10bb-4e83-88e0-665534256b6d","Type":"ContainerDied","Data":"457ac650d9c63c04be8b24daaaae5c00a30ddd923f4ad9229cc56064cd5870f6"} Oct 05 07:12:04 crc kubenswrapper[4846]: I1005 07:12:04.576353 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-r7lpx" event={"ID":"882ccdb3-10bb-4e83-88e0-665534256b6d","Type":"ContainerStarted","Data":"85dbd762b5e0a59a2a828d002e7ef43f8d2f38aac80fe9ed3be882b70dc06c07"} Oct 05 07:12:06 crc kubenswrapper[4846]: I1005 07:12:06.601626 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-r7lpx" event={"ID":"882ccdb3-10bb-4e83-88e0-665534256b6d","Type":"ContainerStarted","Data":"225856147ef27bd8a261e5a72eb970e1f4a121ca2941ec16993251e7c1cf7c3a"} Oct 05 07:12:06 crc kubenswrapper[4846]: E1005 07:12:06.827105 4846 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod882ccdb3_10bb_4e83_88e0_665534256b6d.slice/crio-225856147ef27bd8a261e5a72eb970e1f4a121ca2941ec16993251e7c1cf7c3a.scope\": RecentStats: unable to find data in memory cache]" Oct 05 07:12:07 crc kubenswrapper[4846]: I1005 07:12:07.101135 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-kgt85"] Oct 05 07:12:07 crc kubenswrapper[4846]: I1005 07:12:07.102968 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-kgt85" Oct 05 07:12:07 crc kubenswrapper[4846]: I1005 07:12:07.129755 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-kgt85"] Oct 05 07:12:07 crc kubenswrapper[4846]: I1005 07:12:07.248385 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zgkh9\" (UniqueName: \"kubernetes.io/projected/53845eba-2712-4fb5-927e-ca27a403a71b-kube-api-access-zgkh9\") pod \"community-operators-kgt85\" (UID: \"53845eba-2712-4fb5-927e-ca27a403a71b\") " pod="openshift-marketplace/community-operators-kgt85" Oct 05 07:12:07 crc kubenswrapper[4846]: I1005 07:12:07.248468 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/53845eba-2712-4fb5-927e-ca27a403a71b-utilities\") pod \"community-operators-kgt85\" (UID: \"53845eba-2712-4fb5-927e-ca27a403a71b\") " pod="openshift-marketplace/community-operators-kgt85" Oct 05 07:12:07 crc kubenswrapper[4846]: I1005 07:12:07.248515 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/53845eba-2712-4fb5-927e-ca27a403a71b-catalog-content\") pod \"community-operators-kgt85\" (UID: \"53845eba-2712-4fb5-927e-ca27a403a71b\") " pod="openshift-marketplace/community-operators-kgt85" Oct 05 07:12:07 crc kubenswrapper[4846]: I1005 07:12:07.349399 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/53845eba-2712-4fb5-927e-ca27a403a71b-utilities\") pod \"community-operators-kgt85\" (UID: \"53845eba-2712-4fb5-927e-ca27a403a71b\") " pod="openshift-marketplace/community-operators-kgt85" Oct 05 07:12:07 crc kubenswrapper[4846]: I1005 07:12:07.349456 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/53845eba-2712-4fb5-927e-ca27a403a71b-catalog-content\") pod \"community-operators-kgt85\" (UID: \"53845eba-2712-4fb5-927e-ca27a403a71b\") " pod="openshift-marketplace/community-operators-kgt85" Oct 05 07:12:07 crc kubenswrapper[4846]: I1005 07:12:07.349505 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zgkh9\" (UniqueName: \"kubernetes.io/projected/53845eba-2712-4fb5-927e-ca27a403a71b-kube-api-access-zgkh9\") pod \"community-operators-kgt85\" (UID: \"53845eba-2712-4fb5-927e-ca27a403a71b\") " pod="openshift-marketplace/community-operators-kgt85" Oct 05 07:12:07 crc kubenswrapper[4846]: I1005 07:12:07.350280 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/53845eba-2712-4fb5-927e-ca27a403a71b-utilities\") pod \"community-operators-kgt85\" (UID: \"53845eba-2712-4fb5-927e-ca27a403a71b\") " pod="openshift-marketplace/community-operators-kgt85" Oct 05 07:12:07 crc kubenswrapper[4846]: I1005 07:12:07.350664 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/53845eba-2712-4fb5-927e-ca27a403a71b-catalog-content\") pod \"community-operators-kgt85\" (UID: \"53845eba-2712-4fb5-927e-ca27a403a71b\") " pod="openshift-marketplace/community-operators-kgt85" Oct 05 07:12:07 crc kubenswrapper[4846]: I1005 07:12:07.375845 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zgkh9\" (UniqueName: \"kubernetes.io/projected/53845eba-2712-4fb5-927e-ca27a403a71b-kube-api-access-zgkh9\") pod \"community-operators-kgt85\" (UID: \"53845eba-2712-4fb5-927e-ca27a403a71b\") " pod="openshift-marketplace/community-operators-kgt85" Oct 05 07:12:07 crc kubenswrapper[4846]: I1005 07:12:07.436515 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-kgt85" Oct 05 07:12:07 crc kubenswrapper[4846]: I1005 07:12:07.614460 4846 generic.go:334] "Generic (PLEG): container finished" podID="882ccdb3-10bb-4e83-88e0-665534256b6d" containerID="225856147ef27bd8a261e5a72eb970e1f4a121ca2941ec16993251e7c1cf7c3a" exitCode=0 Oct 05 07:12:07 crc kubenswrapper[4846]: I1005 07:12:07.614502 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-r7lpx" event={"ID":"882ccdb3-10bb-4e83-88e0-665534256b6d","Type":"ContainerDied","Data":"225856147ef27bd8a261e5a72eb970e1f4a121ca2941ec16993251e7c1cf7c3a"} Oct 05 07:12:07 crc kubenswrapper[4846]: I1005 07:12:07.983590 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-kgt85"] Oct 05 07:12:07 crc kubenswrapper[4846]: W1005 07:12:07.993795 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod53845eba_2712_4fb5_927e_ca27a403a71b.slice/crio-9048329fa98308ebdcd23f74f6d59072a7de6690f5fc988482529f49dc103c18 WatchSource:0}: Error finding container 9048329fa98308ebdcd23f74f6d59072a7de6690f5fc988482529f49dc103c18: Status 404 returned error can't find the container with id 9048329fa98308ebdcd23f74f6d59072a7de6690f5fc988482529f49dc103c18 Oct 05 07:12:08 crc kubenswrapper[4846]: I1005 07:12:08.631214 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-r7lpx" event={"ID":"882ccdb3-10bb-4e83-88e0-665534256b6d","Type":"ContainerStarted","Data":"b6f598f6b679ee405f8aa7c3b8e55babcae4eb82954fe95388ee69f7a74421c7"} Oct 05 07:12:08 crc kubenswrapper[4846]: I1005 07:12:08.634068 4846 generic.go:334] "Generic (PLEG): container finished" podID="53845eba-2712-4fb5-927e-ca27a403a71b" containerID="3e41584d101b05c4a8f8374933025b096e4e14f4535c02c9c38629938845bc8a" exitCode=0 Oct 05 07:12:08 crc kubenswrapper[4846]: I1005 07:12:08.634189 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kgt85" event={"ID":"53845eba-2712-4fb5-927e-ca27a403a71b","Type":"ContainerDied","Data":"3e41584d101b05c4a8f8374933025b096e4e14f4535c02c9c38629938845bc8a"} Oct 05 07:12:08 crc kubenswrapper[4846]: I1005 07:12:08.634254 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kgt85" event={"ID":"53845eba-2712-4fb5-927e-ca27a403a71b","Type":"ContainerStarted","Data":"9048329fa98308ebdcd23f74f6d59072a7de6690f5fc988482529f49dc103c18"} Oct 05 07:12:08 crc kubenswrapper[4846]: I1005 07:12:08.664402 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-r7lpx" podStartSLOduration=2.205582644 podStartE2EDuration="5.664376208s" podCreationTimestamp="2025-10-05 07:12:03 +0000 UTC" firstStartedPulling="2025-10-05 07:12:04.578578719 +0000 UTC m=+1446.819431504" lastFinishedPulling="2025-10-05 07:12:08.037372273 +0000 UTC m=+1450.278225068" observedRunningTime="2025-10-05 07:12:08.659669383 +0000 UTC m=+1450.900522188" watchObservedRunningTime="2025-10-05 07:12:08.664376208 +0000 UTC m=+1450.905229023" Oct 05 07:12:09 crc kubenswrapper[4846]: I1005 07:12:09.648395 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kgt85" event={"ID":"53845eba-2712-4fb5-927e-ca27a403a71b","Type":"ContainerStarted","Data":"0a3fb016e06c932ae65cd403da844c686346cdb52927cc7ab025a51ce2555635"} Oct 05 07:12:10 crc kubenswrapper[4846]: I1005 07:12:10.662625 4846 generic.go:334] "Generic (PLEG): container finished" podID="53845eba-2712-4fb5-927e-ca27a403a71b" containerID="0a3fb016e06c932ae65cd403da844c686346cdb52927cc7ab025a51ce2555635" exitCode=0 Oct 05 07:12:10 crc kubenswrapper[4846]: I1005 07:12:10.662806 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kgt85" event={"ID":"53845eba-2712-4fb5-927e-ca27a403a71b","Type":"ContainerDied","Data":"0a3fb016e06c932ae65cd403da844c686346cdb52927cc7ab025a51ce2555635"} Oct 05 07:12:11 crc kubenswrapper[4846]: I1005 07:12:11.673011 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kgt85" event={"ID":"53845eba-2712-4fb5-927e-ca27a403a71b","Type":"ContainerStarted","Data":"aa0f9480ed8fbf448b866ab8d16ce40caad2627449bfb7225433a0c19798d53e"} Oct 05 07:12:11 crc kubenswrapper[4846]: I1005 07:12:11.691219 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-kgt85" podStartSLOduration=2.270633911 podStartE2EDuration="4.691193623s" podCreationTimestamp="2025-10-05 07:12:07 +0000 UTC" firstStartedPulling="2025-10-05 07:12:08.637956453 +0000 UTC m=+1450.878809228" lastFinishedPulling="2025-10-05 07:12:11.058516125 +0000 UTC m=+1453.299368940" observedRunningTime="2025-10-05 07:12:11.688602224 +0000 UTC m=+1453.929454999" watchObservedRunningTime="2025-10-05 07:12:11.691193623 +0000 UTC m=+1453.932046398" Oct 05 07:12:13 crc kubenswrapper[4846]: I1005 07:12:13.869603 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-r7lpx" Oct 05 07:12:13 crc kubenswrapper[4846]: I1005 07:12:13.869686 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-r7lpx" Oct 05 07:12:13 crc kubenswrapper[4846]: I1005 07:12:13.916039 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-r7lpx" Oct 05 07:12:14 crc kubenswrapper[4846]: I1005 07:12:14.742927 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-r7lpx" Oct 05 07:12:15 crc kubenswrapper[4846]: I1005 07:12:15.089778 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-r7lpx"] Oct 05 07:12:16 crc kubenswrapper[4846]: I1005 07:12:16.726849 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-r7lpx" podUID="882ccdb3-10bb-4e83-88e0-665534256b6d" containerName="registry-server" containerID="cri-o://b6f598f6b679ee405f8aa7c3b8e55babcae4eb82954fe95388ee69f7a74421c7" gracePeriod=2 Oct 05 07:12:17 crc kubenswrapper[4846]: I1005 07:12:17.437144 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-kgt85" Oct 05 07:12:17 crc kubenswrapper[4846]: I1005 07:12:17.437663 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-kgt85" Oct 05 07:12:17 crc kubenswrapper[4846]: I1005 07:12:17.522973 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-kgt85" Oct 05 07:12:17 crc kubenswrapper[4846]: I1005 07:12:17.742722 4846 generic.go:334] "Generic (PLEG): container finished" podID="882ccdb3-10bb-4e83-88e0-665534256b6d" containerID="b6f598f6b679ee405f8aa7c3b8e55babcae4eb82954fe95388ee69f7a74421c7" exitCode=0 Oct 05 07:12:17 crc kubenswrapper[4846]: I1005 07:12:17.743679 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-r7lpx" event={"ID":"882ccdb3-10bb-4e83-88e0-665534256b6d","Type":"ContainerDied","Data":"b6f598f6b679ee405f8aa7c3b8e55babcae4eb82954fe95388ee69f7a74421c7"} Oct 05 07:12:17 crc kubenswrapper[4846]: I1005 07:12:17.786392 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-kgt85" Oct 05 07:12:17 crc kubenswrapper[4846]: I1005 07:12:17.891215 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-r7lpx" Oct 05 07:12:17 crc kubenswrapper[4846]: I1005 07:12:17.934017 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/882ccdb3-10bb-4e83-88e0-665534256b6d-catalog-content\") pod \"882ccdb3-10bb-4e83-88e0-665534256b6d\" (UID: \"882ccdb3-10bb-4e83-88e0-665534256b6d\") " Oct 05 07:12:17 crc kubenswrapper[4846]: I1005 07:12:17.934081 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7q9mv\" (UniqueName: \"kubernetes.io/projected/882ccdb3-10bb-4e83-88e0-665534256b6d-kube-api-access-7q9mv\") pod \"882ccdb3-10bb-4e83-88e0-665534256b6d\" (UID: \"882ccdb3-10bb-4e83-88e0-665534256b6d\") " Oct 05 07:12:17 crc kubenswrapper[4846]: I1005 07:12:17.934232 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/882ccdb3-10bb-4e83-88e0-665534256b6d-utilities\") pod \"882ccdb3-10bb-4e83-88e0-665534256b6d\" (UID: \"882ccdb3-10bb-4e83-88e0-665534256b6d\") " Oct 05 07:12:17 crc kubenswrapper[4846]: I1005 07:12:17.935423 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/882ccdb3-10bb-4e83-88e0-665534256b6d-utilities" (OuterVolumeSpecName: "utilities") pod "882ccdb3-10bb-4e83-88e0-665534256b6d" (UID: "882ccdb3-10bb-4e83-88e0-665534256b6d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 07:12:17 crc kubenswrapper[4846]: I1005 07:12:17.940316 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/882ccdb3-10bb-4e83-88e0-665534256b6d-kube-api-access-7q9mv" (OuterVolumeSpecName: "kube-api-access-7q9mv") pod "882ccdb3-10bb-4e83-88e0-665534256b6d" (UID: "882ccdb3-10bb-4e83-88e0-665534256b6d"). InnerVolumeSpecName "kube-api-access-7q9mv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:12:17 crc kubenswrapper[4846]: I1005 07:12:17.989831 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/882ccdb3-10bb-4e83-88e0-665534256b6d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "882ccdb3-10bb-4e83-88e0-665534256b6d" (UID: "882ccdb3-10bb-4e83-88e0-665534256b6d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 07:12:18 crc kubenswrapper[4846]: I1005 07:12:18.035714 4846 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/882ccdb3-10bb-4e83-88e0-665534256b6d-utilities\") on node \"crc\" DevicePath \"\"" Oct 05 07:12:18 crc kubenswrapper[4846]: I1005 07:12:18.035755 4846 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/882ccdb3-10bb-4e83-88e0-665534256b6d-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 05 07:12:18 crc kubenswrapper[4846]: I1005 07:12:18.035777 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7q9mv\" (UniqueName: \"kubernetes.io/projected/882ccdb3-10bb-4e83-88e0-665534256b6d-kube-api-access-7q9mv\") on node \"crc\" DevicePath \"\"" Oct 05 07:12:18 crc kubenswrapper[4846]: I1005 07:12:18.764898 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-r7lpx" event={"ID":"882ccdb3-10bb-4e83-88e0-665534256b6d","Type":"ContainerDied","Data":"85dbd762b5e0a59a2a828d002e7ef43f8d2f38aac80fe9ed3be882b70dc06c07"} Oct 05 07:12:18 crc kubenswrapper[4846]: I1005 07:12:18.764977 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-r7lpx" Oct 05 07:12:18 crc kubenswrapper[4846]: I1005 07:12:18.765515 4846 scope.go:117] "RemoveContainer" containerID="b6f598f6b679ee405f8aa7c3b8e55babcae4eb82954fe95388ee69f7a74421c7" Oct 05 07:12:18 crc kubenswrapper[4846]: I1005 07:12:18.808613 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-r7lpx"] Oct 05 07:12:18 crc kubenswrapper[4846]: I1005 07:12:18.824638 4846 scope.go:117] "RemoveContainer" containerID="225856147ef27bd8a261e5a72eb970e1f4a121ca2941ec16993251e7c1cf7c3a" Oct 05 07:12:18 crc kubenswrapper[4846]: I1005 07:12:18.830233 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-r7lpx"] Oct 05 07:12:18 crc kubenswrapper[4846]: I1005 07:12:18.862033 4846 scope.go:117] "RemoveContainer" containerID="457ac650d9c63c04be8b24daaaae5c00a30ddd923f4ad9229cc56064cd5870f6" Oct 05 07:12:19 crc kubenswrapper[4846]: I1005 07:12:19.285950 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-kgt85"] Oct 05 07:12:20 crc kubenswrapper[4846]: I1005 07:12:20.514219 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="882ccdb3-10bb-4e83-88e0-665534256b6d" path="/var/lib/kubelet/pods/882ccdb3-10bb-4e83-88e0-665534256b6d/volumes" Oct 05 07:12:20 crc kubenswrapper[4846]: I1005 07:12:20.797489 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-kgt85" podUID="53845eba-2712-4fb5-927e-ca27a403a71b" containerName="registry-server" containerID="cri-o://aa0f9480ed8fbf448b866ab8d16ce40caad2627449bfb7225433a0c19798d53e" gracePeriod=2 Oct 05 07:12:21 crc kubenswrapper[4846]: I1005 07:12:21.281057 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-kgt85" Oct 05 07:12:21 crc kubenswrapper[4846]: I1005 07:12:21.440703 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/53845eba-2712-4fb5-927e-ca27a403a71b-utilities\") pod \"53845eba-2712-4fb5-927e-ca27a403a71b\" (UID: \"53845eba-2712-4fb5-927e-ca27a403a71b\") " Oct 05 07:12:21 crc kubenswrapper[4846]: I1005 07:12:21.440791 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgkh9\" (UniqueName: \"kubernetes.io/projected/53845eba-2712-4fb5-927e-ca27a403a71b-kube-api-access-zgkh9\") pod \"53845eba-2712-4fb5-927e-ca27a403a71b\" (UID: \"53845eba-2712-4fb5-927e-ca27a403a71b\") " Oct 05 07:12:21 crc kubenswrapper[4846]: I1005 07:12:21.440855 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/53845eba-2712-4fb5-927e-ca27a403a71b-catalog-content\") pod \"53845eba-2712-4fb5-927e-ca27a403a71b\" (UID: \"53845eba-2712-4fb5-927e-ca27a403a71b\") " Oct 05 07:12:21 crc kubenswrapper[4846]: I1005 07:12:21.442138 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/53845eba-2712-4fb5-927e-ca27a403a71b-utilities" (OuterVolumeSpecName: "utilities") pod "53845eba-2712-4fb5-927e-ca27a403a71b" (UID: "53845eba-2712-4fb5-927e-ca27a403a71b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 07:12:21 crc kubenswrapper[4846]: I1005 07:12:21.448434 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/53845eba-2712-4fb5-927e-ca27a403a71b-kube-api-access-zgkh9" (OuterVolumeSpecName: "kube-api-access-zgkh9") pod "53845eba-2712-4fb5-927e-ca27a403a71b" (UID: "53845eba-2712-4fb5-927e-ca27a403a71b"). InnerVolumeSpecName "kube-api-access-zgkh9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:12:21 crc kubenswrapper[4846]: I1005 07:12:21.519714 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/53845eba-2712-4fb5-927e-ca27a403a71b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "53845eba-2712-4fb5-927e-ca27a403a71b" (UID: "53845eba-2712-4fb5-927e-ca27a403a71b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 07:12:21 crc kubenswrapper[4846]: I1005 07:12:21.542010 4846 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/53845eba-2712-4fb5-927e-ca27a403a71b-utilities\") on node \"crc\" DevicePath \"\"" Oct 05 07:12:21 crc kubenswrapper[4846]: I1005 07:12:21.542048 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgkh9\" (UniqueName: \"kubernetes.io/projected/53845eba-2712-4fb5-927e-ca27a403a71b-kube-api-access-zgkh9\") on node \"crc\" DevicePath \"\"" Oct 05 07:12:21 crc kubenswrapper[4846]: I1005 07:12:21.542062 4846 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/53845eba-2712-4fb5-927e-ca27a403a71b-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 05 07:12:21 crc kubenswrapper[4846]: I1005 07:12:21.811150 4846 generic.go:334] "Generic (PLEG): container finished" podID="53845eba-2712-4fb5-927e-ca27a403a71b" containerID="aa0f9480ed8fbf448b866ab8d16ce40caad2627449bfb7225433a0c19798d53e" exitCode=0 Oct 05 07:12:21 crc kubenswrapper[4846]: I1005 07:12:21.811218 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kgt85" event={"ID":"53845eba-2712-4fb5-927e-ca27a403a71b","Type":"ContainerDied","Data":"aa0f9480ed8fbf448b866ab8d16ce40caad2627449bfb7225433a0c19798d53e"} Oct 05 07:12:21 crc kubenswrapper[4846]: I1005 07:12:21.811271 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kgt85" event={"ID":"53845eba-2712-4fb5-927e-ca27a403a71b","Type":"ContainerDied","Data":"9048329fa98308ebdcd23f74f6d59072a7de6690f5fc988482529f49dc103c18"} Oct 05 07:12:21 crc kubenswrapper[4846]: I1005 07:12:21.811298 4846 scope.go:117] "RemoveContainer" containerID="aa0f9480ed8fbf448b866ab8d16ce40caad2627449bfb7225433a0c19798d53e" Oct 05 07:12:21 crc kubenswrapper[4846]: I1005 07:12:21.811300 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-kgt85" Oct 05 07:12:21 crc kubenswrapper[4846]: I1005 07:12:21.847161 4846 scope.go:117] "RemoveContainer" containerID="0a3fb016e06c932ae65cd403da844c686346cdb52927cc7ab025a51ce2555635" Oct 05 07:12:21 crc kubenswrapper[4846]: I1005 07:12:21.876395 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-kgt85"] Oct 05 07:12:21 crc kubenswrapper[4846]: I1005 07:12:21.885123 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-kgt85"] Oct 05 07:12:21 crc kubenswrapper[4846]: I1005 07:12:21.890460 4846 scope.go:117] "RemoveContainer" containerID="3e41584d101b05c4a8f8374933025b096e4e14f4535c02c9c38629938845bc8a" Oct 05 07:12:21 crc kubenswrapper[4846]: I1005 07:12:21.911973 4846 scope.go:117] "RemoveContainer" containerID="aa0f9480ed8fbf448b866ab8d16ce40caad2627449bfb7225433a0c19798d53e" Oct 05 07:12:21 crc kubenswrapper[4846]: E1005 07:12:21.912581 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"aa0f9480ed8fbf448b866ab8d16ce40caad2627449bfb7225433a0c19798d53e\": container with ID starting with aa0f9480ed8fbf448b866ab8d16ce40caad2627449bfb7225433a0c19798d53e not found: ID does not exist" containerID="aa0f9480ed8fbf448b866ab8d16ce40caad2627449bfb7225433a0c19798d53e" Oct 05 07:12:21 crc kubenswrapper[4846]: I1005 07:12:21.912622 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aa0f9480ed8fbf448b866ab8d16ce40caad2627449bfb7225433a0c19798d53e"} err="failed to get container status \"aa0f9480ed8fbf448b866ab8d16ce40caad2627449bfb7225433a0c19798d53e\": rpc error: code = NotFound desc = could not find container \"aa0f9480ed8fbf448b866ab8d16ce40caad2627449bfb7225433a0c19798d53e\": container with ID starting with aa0f9480ed8fbf448b866ab8d16ce40caad2627449bfb7225433a0c19798d53e not found: ID does not exist" Oct 05 07:12:21 crc kubenswrapper[4846]: I1005 07:12:21.912658 4846 scope.go:117] "RemoveContainer" containerID="0a3fb016e06c932ae65cd403da844c686346cdb52927cc7ab025a51ce2555635" Oct 05 07:12:21 crc kubenswrapper[4846]: E1005 07:12:21.913100 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0a3fb016e06c932ae65cd403da844c686346cdb52927cc7ab025a51ce2555635\": container with ID starting with 0a3fb016e06c932ae65cd403da844c686346cdb52927cc7ab025a51ce2555635 not found: ID does not exist" containerID="0a3fb016e06c932ae65cd403da844c686346cdb52927cc7ab025a51ce2555635" Oct 05 07:12:21 crc kubenswrapper[4846]: I1005 07:12:21.913144 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0a3fb016e06c932ae65cd403da844c686346cdb52927cc7ab025a51ce2555635"} err="failed to get container status \"0a3fb016e06c932ae65cd403da844c686346cdb52927cc7ab025a51ce2555635\": rpc error: code = NotFound desc = could not find container \"0a3fb016e06c932ae65cd403da844c686346cdb52927cc7ab025a51ce2555635\": container with ID starting with 0a3fb016e06c932ae65cd403da844c686346cdb52927cc7ab025a51ce2555635 not found: ID does not exist" Oct 05 07:12:21 crc kubenswrapper[4846]: I1005 07:12:21.913162 4846 scope.go:117] "RemoveContainer" containerID="3e41584d101b05c4a8f8374933025b096e4e14f4535c02c9c38629938845bc8a" Oct 05 07:12:21 crc kubenswrapper[4846]: E1005 07:12:21.913585 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3e41584d101b05c4a8f8374933025b096e4e14f4535c02c9c38629938845bc8a\": container with ID starting with 3e41584d101b05c4a8f8374933025b096e4e14f4535c02c9c38629938845bc8a not found: ID does not exist" containerID="3e41584d101b05c4a8f8374933025b096e4e14f4535c02c9c38629938845bc8a" Oct 05 07:12:21 crc kubenswrapper[4846]: I1005 07:12:21.913633 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3e41584d101b05c4a8f8374933025b096e4e14f4535c02c9c38629938845bc8a"} err="failed to get container status \"3e41584d101b05c4a8f8374933025b096e4e14f4535c02c9c38629938845bc8a\": rpc error: code = NotFound desc = could not find container \"3e41584d101b05c4a8f8374933025b096e4e14f4535c02c9c38629938845bc8a\": container with ID starting with 3e41584d101b05c4a8f8374933025b096e4e14f4535c02c9c38629938845bc8a not found: ID does not exist" Oct 05 07:12:22 crc kubenswrapper[4846]: I1005 07:12:22.516908 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="53845eba-2712-4fb5-927e-ca27a403a71b" path="/var/lib/kubelet/pods/53845eba-2712-4fb5-927e-ca27a403a71b/volumes" Oct 05 07:12:23 crc kubenswrapper[4846]: I1005 07:12:23.325199 4846 patch_prober.go:28] interesting pod/machine-config-daemon-fscvf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 07:12:23 crc kubenswrapper[4846]: I1005 07:12:23.325288 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 07:12:26 crc kubenswrapper[4846]: I1005 07:12:26.050398 4846 scope.go:117] "RemoveContainer" containerID="1026ae5247bafab34579d4dd5c0cd5d8bd570d34ab240d0ce80bdef49ddaa60a" Oct 05 07:12:26 crc kubenswrapper[4846]: I1005 07:12:26.092735 4846 scope.go:117] "RemoveContainer" containerID="841de69a750baa7db33c72de16defdadd02908d58b435db7b55539b3d4710699" Oct 05 07:12:26 crc kubenswrapper[4846]: I1005 07:12:26.138682 4846 scope.go:117] "RemoveContainer" containerID="9da68a94c483e18bf05dd62a5687e31c8c93c5ae056bb98a27b899f3ba190d7e" Oct 05 07:12:26 crc kubenswrapper[4846]: I1005 07:12:26.175567 4846 scope.go:117] "RemoveContainer" containerID="1997b10872b4f4d32c524583e381e6591ecdae66bc71a4d1d18521fe506a1eb7" Oct 05 07:12:32 crc kubenswrapper[4846]: I1005 07:12:32.652892 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-cfv9j"] Oct 05 07:12:32 crc kubenswrapper[4846]: E1005 07:12:32.654086 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="53845eba-2712-4fb5-927e-ca27a403a71b" containerName="extract-utilities" Oct 05 07:12:32 crc kubenswrapper[4846]: I1005 07:12:32.654110 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="53845eba-2712-4fb5-927e-ca27a403a71b" containerName="extract-utilities" Oct 05 07:12:32 crc kubenswrapper[4846]: E1005 07:12:32.654131 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="53845eba-2712-4fb5-927e-ca27a403a71b" containerName="registry-server" Oct 05 07:12:32 crc kubenswrapper[4846]: I1005 07:12:32.654143 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="53845eba-2712-4fb5-927e-ca27a403a71b" containerName="registry-server" Oct 05 07:12:32 crc kubenswrapper[4846]: E1005 07:12:32.654164 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="882ccdb3-10bb-4e83-88e0-665534256b6d" containerName="extract-utilities" Oct 05 07:12:32 crc kubenswrapper[4846]: I1005 07:12:32.654182 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="882ccdb3-10bb-4e83-88e0-665534256b6d" containerName="extract-utilities" Oct 05 07:12:32 crc kubenswrapper[4846]: E1005 07:12:32.654226 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="53845eba-2712-4fb5-927e-ca27a403a71b" containerName="extract-content" Oct 05 07:12:32 crc kubenswrapper[4846]: I1005 07:12:32.654238 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="53845eba-2712-4fb5-927e-ca27a403a71b" containerName="extract-content" Oct 05 07:12:32 crc kubenswrapper[4846]: E1005 07:12:32.654258 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="882ccdb3-10bb-4e83-88e0-665534256b6d" containerName="registry-server" Oct 05 07:12:32 crc kubenswrapper[4846]: I1005 07:12:32.654268 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="882ccdb3-10bb-4e83-88e0-665534256b6d" containerName="registry-server" Oct 05 07:12:32 crc kubenswrapper[4846]: E1005 07:12:32.654299 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="882ccdb3-10bb-4e83-88e0-665534256b6d" containerName="extract-content" Oct 05 07:12:32 crc kubenswrapper[4846]: I1005 07:12:32.654309 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="882ccdb3-10bb-4e83-88e0-665534256b6d" containerName="extract-content" Oct 05 07:12:32 crc kubenswrapper[4846]: I1005 07:12:32.654524 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="53845eba-2712-4fb5-927e-ca27a403a71b" containerName="registry-server" Oct 05 07:12:32 crc kubenswrapper[4846]: I1005 07:12:32.654557 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="882ccdb3-10bb-4e83-88e0-665534256b6d" containerName="registry-server" Oct 05 07:12:32 crc kubenswrapper[4846]: I1005 07:12:32.656309 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-cfv9j" Oct 05 07:12:32 crc kubenswrapper[4846]: I1005 07:12:32.665364 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-cfv9j"] Oct 05 07:12:32 crc kubenswrapper[4846]: I1005 07:12:32.719869 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a0eb9b2f-f93b-41f5-a790-558a6bcdf0be-utilities\") pod \"redhat-marketplace-cfv9j\" (UID: \"a0eb9b2f-f93b-41f5-a790-558a6bcdf0be\") " pod="openshift-marketplace/redhat-marketplace-cfv9j" Oct 05 07:12:32 crc kubenswrapper[4846]: I1005 07:12:32.719960 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a0eb9b2f-f93b-41f5-a790-558a6bcdf0be-catalog-content\") pod \"redhat-marketplace-cfv9j\" (UID: \"a0eb9b2f-f93b-41f5-a790-558a6bcdf0be\") " pod="openshift-marketplace/redhat-marketplace-cfv9j" Oct 05 07:12:32 crc kubenswrapper[4846]: I1005 07:12:32.720039 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-csx9n\" (UniqueName: \"kubernetes.io/projected/a0eb9b2f-f93b-41f5-a790-558a6bcdf0be-kube-api-access-csx9n\") pod \"redhat-marketplace-cfv9j\" (UID: \"a0eb9b2f-f93b-41f5-a790-558a6bcdf0be\") " pod="openshift-marketplace/redhat-marketplace-cfv9j" Oct 05 07:12:32 crc kubenswrapper[4846]: I1005 07:12:32.820882 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a0eb9b2f-f93b-41f5-a790-558a6bcdf0be-catalog-content\") pod \"redhat-marketplace-cfv9j\" (UID: \"a0eb9b2f-f93b-41f5-a790-558a6bcdf0be\") " pod="openshift-marketplace/redhat-marketplace-cfv9j" Oct 05 07:12:32 crc kubenswrapper[4846]: I1005 07:12:32.820973 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-csx9n\" (UniqueName: \"kubernetes.io/projected/a0eb9b2f-f93b-41f5-a790-558a6bcdf0be-kube-api-access-csx9n\") pod \"redhat-marketplace-cfv9j\" (UID: \"a0eb9b2f-f93b-41f5-a790-558a6bcdf0be\") " pod="openshift-marketplace/redhat-marketplace-cfv9j" Oct 05 07:12:32 crc kubenswrapper[4846]: I1005 07:12:32.821011 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a0eb9b2f-f93b-41f5-a790-558a6bcdf0be-utilities\") pod \"redhat-marketplace-cfv9j\" (UID: \"a0eb9b2f-f93b-41f5-a790-558a6bcdf0be\") " pod="openshift-marketplace/redhat-marketplace-cfv9j" Oct 05 07:12:32 crc kubenswrapper[4846]: I1005 07:12:32.821799 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a0eb9b2f-f93b-41f5-a790-558a6bcdf0be-utilities\") pod \"redhat-marketplace-cfv9j\" (UID: \"a0eb9b2f-f93b-41f5-a790-558a6bcdf0be\") " pod="openshift-marketplace/redhat-marketplace-cfv9j" Oct 05 07:12:32 crc kubenswrapper[4846]: I1005 07:12:32.821831 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a0eb9b2f-f93b-41f5-a790-558a6bcdf0be-catalog-content\") pod \"redhat-marketplace-cfv9j\" (UID: \"a0eb9b2f-f93b-41f5-a790-558a6bcdf0be\") " pod="openshift-marketplace/redhat-marketplace-cfv9j" Oct 05 07:12:32 crc kubenswrapper[4846]: I1005 07:12:32.847533 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-csx9n\" (UniqueName: \"kubernetes.io/projected/a0eb9b2f-f93b-41f5-a790-558a6bcdf0be-kube-api-access-csx9n\") pod \"redhat-marketplace-cfv9j\" (UID: \"a0eb9b2f-f93b-41f5-a790-558a6bcdf0be\") " pod="openshift-marketplace/redhat-marketplace-cfv9j" Oct 05 07:12:32 crc kubenswrapper[4846]: I1005 07:12:32.998559 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-cfv9j" Oct 05 07:12:33 crc kubenswrapper[4846]: I1005 07:12:33.424990 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-cfv9j"] Oct 05 07:12:33 crc kubenswrapper[4846]: I1005 07:12:33.945721 4846 generic.go:334] "Generic (PLEG): container finished" podID="a0eb9b2f-f93b-41f5-a790-558a6bcdf0be" containerID="69253aadbb0eb74c8ce5ee557dbed2535764f5d7a9a4bedd49b0b0dd420ad20a" exitCode=0 Oct 05 07:12:33 crc kubenswrapper[4846]: I1005 07:12:33.945795 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-cfv9j" event={"ID":"a0eb9b2f-f93b-41f5-a790-558a6bcdf0be","Type":"ContainerDied","Data":"69253aadbb0eb74c8ce5ee557dbed2535764f5d7a9a4bedd49b0b0dd420ad20a"} Oct 05 07:12:33 crc kubenswrapper[4846]: I1005 07:12:33.945836 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-cfv9j" event={"ID":"a0eb9b2f-f93b-41f5-a790-558a6bcdf0be","Type":"ContainerStarted","Data":"dfe2d7a1f68bc8d94e2fbe84e1bdfe4bb3db9b85ae96b4b43aa43a7f689ba3b3"} Oct 05 07:12:34 crc kubenswrapper[4846]: I1005 07:12:34.955831 4846 generic.go:334] "Generic (PLEG): container finished" podID="a0eb9b2f-f93b-41f5-a790-558a6bcdf0be" containerID="62b5bb44740009feb2f5c21f8547aa59f5f5439ad90cfa4571c6b9344b97a1ec" exitCode=0 Oct 05 07:12:34 crc kubenswrapper[4846]: I1005 07:12:34.955950 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-cfv9j" event={"ID":"a0eb9b2f-f93b-41f5-a790-558a6bcdf0be","Type":"ContainerDied","Data":"62b5bb44740009feb2f5c21f8547aa59f5f5439ad90cfa4571c6b9344b97a1ec"} Oct 05 07:12:35 crc kubenswrapper[4846]: I1005 07:12:35.970628 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-cfv9j" event={"ID":"a0eb9b2f-f93b-41f5-a790-558a6bcdf0be","Type":"ContainerStarted","Data":"9e277a08d354bb9c946687ece40ec6ee03169204de5e7e9a333df3f360771fb8"} Oct 05 07:12:35 crc kubenswrapper[4846]: I1005 07:12:35.993872 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-cfv9j" podStartSLOduration=2.2405808 podStartE2EDuration="3.9938487s" podCreationTimestamp="2025-10-05 07:12:32 +0000 UTC" firstStartedPulling="2025-10-05 07:12:33.950166429 +0000 UTC m=+1476.191019204" lastFinishedPulling="2025-10-05 07:12:35.703434328 +0000 UTC m=+1477.944287104" observedRunningTime="2025-10-05 07:12:35.99347563 +0000 UTC m=+1478.234328405" watchObservedRunningTime="2025-10-05 07:12:35.9938487 +0000 UTC m=+1478.234701485" Oct 05 07:12:43 crc kubenswrapper[4846]: I1005 07:12:42.999397 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-cfv9j" Oct 05 07:12:43 crc kubenswrapper[4846]: I1005 07:12:43.000156 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-cfv9j" Oct 05 07:12:43 crc kubenswrapper[4846]: I1005 07:12:43.077738 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-cfv9j" Oct 05 07:12:43 crc kubenswrapper[4846]: I1005 07:12:43.146928 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-cfv9j" Oct 05 07:12:43 crc kubenswrapper[4846]: I1005 07:12:43.340522 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-cfv9j"] Oct 05 07:12:45 crc kubenswrapper[4846]: I1005 07:12:45.061572 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-cfv9j" podUID="a0eb9b2f-f93b-41f5-a790-558a6bcdf0be" containerName="registry-server" containerID="cri-o://9e277a08d354bb9c946687ece40ec6ee03169204de5e7e9a333df3f360771fb8" gracePeriod=2 Oct 05 07:12:45 crc kubenswrapper[4846]: I1005 07:12:45.516773 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-cfv9j" Oct 05 07:12:45 crc kubenswrapper[4846]: I1005 07:12:45.625587 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a0eb9b2f-f93b-41f5-a790-558a6bcdf0be-utilities\") pod \"a0eb9b2f-f93b-41f5-a790-558a6bcdf0be\" (UID: \"a0eb9b2f-f93b-41f5-a790-558a6bcdf0be\") " Oct 05 07:12:45 crc kubenswrapper[4846]: I1005 07:12:45.625642 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-csx9n\" (UniqueName: \"kubernetes.io/projected/a0eb9b2f-f93b-41f5-a790-558a6bcdf0be-kube-api-access-csx9n\") pod \"a0eb9b2f-f93b-41f5-a790-558a6bcdf0be\" (UID: \"a0eb9b2f-f93b-41f5-a790-558a6bcdf0be\") " Oct 05 07:12:45 crc kubenswrapper[4846]: I1005 07:12:45.625729 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a0eb9b2f-f93b-41f5-a790-558a6bcdf0be-catalog-content\") pod \"a0eb9b2f-f93b-41f5-a790-558a6bcdf0be\" (UID: \"a0eb9b2f-f93b-41f5-a790-558a6bcdf0be\") " Oct 05 07:12:45 crc kubenswrapper[4846]: I1005 07:12:45.629507 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a0eb9b2f-f93b-41f5-a790-558a6bcdf0be-utilities" (OuterVolumeSpecName: "utilities") pod "a0eb9b2f-f93b-41f5-a790-558a6bcdf0be" (UID: "a0eb9b2f-f93b-41f5-a790-558a6bcdf0be"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 07:12:45 crc kubenswrapper[4846]: I1005 07:12:45.632211 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0eb9b2f-f93b-41f5-a790-558a6bcdf0be-kube-api-access-csx9n" (OuterVolumeSpecName: "kube-api-access-csx9n") pod "a0eb9b2f-f93b-41f5-a790-558a6bcdf0be" (UID: "a0eb9b2f-f93b-41f5-a790-558a6bcdf0be"). InnerVolumeSpecName "kube-api-access-csx9n". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:12:45 crc kubenswrapper[4846]: I1005 07:12:45.640836 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a0eb9b2f-f93b-41f5-a790-558a6bcdf0be-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a0eb9b2f-f93b-41f5-a790-558a6bcdf0be" (UID: "a0eb9b2f-f93b-41f5-a790-558a6bcdf0be"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 07:12:45 crc kubenswrapper[4846]: I1005 07:12:45.727331 4846 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a0eb9b2f-f93b-41f5-a790-558a6bcdf0be-utilities\") on node \"crc\" DevicePath \"\"" Oct 05 07:12:45 crc kubenswrapper[4846]: I1005 07:12:45.727376 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-csx9n\" (UniqueName: \"kubernetes.io/projected/a0eb9b2f-f93b-41f5-a790-558a6bcdf0be-kube-api-access-csx9n\") on node \"crc\" DevicePath \"\"" Oct 05 07:12:45 crc kubenswrapper[4846]: I1005 07:12:45.727393 4846 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a0eb9b2f-f93b-41f5-a790-558a6bcdf0be-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 05 07:12:46 crc kubenswrapper[4846]: I1005 07:12:46.077260 4846 generic.go:334] "Generic (PLEG): container finished" podID="a0eb9b2f-f93b-41f5-a790-558a6bcdf0be" containerID="9e277a08d354bb9c946687ece40ec6ee03169204de5e7e9a333df3f360771fb8" exitCode=0 Oct 05 07:12:46 crc kubenswrapper[4846]: I1005 07:12:46.077328 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-cfv9j" Oct 05 07:12:46 crc kubenswrapper[4846]: I1005 07:12:46.077323 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-cfv9j" event={"ID":"a0eb9b2f-f93b-41f5-a790-558a6bcdf0be","Type":"ContainerDied","Data":"9e277a08d354bb9c946687ece40ec6ee03169204de5e7e9a333df3f360771fb8"} Oct 05 07:12:46 crc kubenswrapper[4846]: I1005 07:12:46.077398 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-cfv9j" event={"ID":"a0eb9b2f-f93b-41f5-a790-558a6bcdf0be","Type":"ContainerDied","Data":"dfe2d7a1f68bc8d94e2fbe84e1bdfe4bb3db9b85ae96b4b43aa43a7f689ba3b3"} Oct 05 07:12:46 crc kubenswrapper[4846]: I1005 07:12:46.077423 4846 scope.go:117] "RemoveContainer" containerID="9e277a08d354bb9c946687ece40ec6ee03169204de5e7e9a333df3f360771fb8" Oct 05 07:12:46 crc kubenswrapper[4846]: I1005 07:12:46.112266 4846 scope.go:117] "RemoveContainer" containerID="62b5bb44740009feb2f5c21f8547aa59f5f5439ad90cfa4571c6b9344b97a1ec" Oct 05 07:12:46 crc kubenswrapper[4846]: I1005 07:12:46.138003 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-cfv9j"] Oct 05 07:12:46 crc kubenswrapper[4846]: I1005 07:12:46.148845 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-cfv9j"] Oct 05 07:12:46 crc kubenswrapper[4846]: I1005 07:12:46.156116 4846 scope.go:117] "RemoveContainer" containerID="69253aadbb0eb74c8ce5ee557dbed2535764f5d7a9a4bedd49b0b0dd420ad20a" Oct 05 07:12:46 crc kubenswrapper[4846]: I1005 07:12:46.180570 4846 scope.go:117] "RemoveContainer" containerID="9e277a08d354bb9c946687ece40ec6ee03169204de5e7e9a333df3f360771fb8" Oct 05 07:12:46 crc kubenswrapper[4846]: E1005 07:12:46.181100 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9e277a08d354bb9c946687ece40ec6ee03169204de5e7e9a333df3f360771fb8\": container with ID starting with 9e277a08d354bb9c946687ece40ec6ee03169204de5e7e9a333df3f360771fb8 not found: ID does not exist" containerID="9e277a08d354bb9c946687ece40ec6ee03169204de5e7e9a333df3f360771fb8" Oct 05 07:12:46 crc kubenswrapper[4846]: I1005 07:12:46.181134 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9e277a08d354bb9c946687ece40ec6ee03169204de5e7e9a333df3f360771fb8"} err="failed to get container status \"9e277a08d354bb9c946687ece40ec6ee03169204de5e7e9a333df3f360771fb8\": rpc error: code = NotFound desc = could not find container \"9e277a08d354bb9c946687ece40ec6ee03169204de5e7e9a333df3f360771fb8\": container with ID starting with 9e277a08d354bb9c946687ece40ec6ee03169204de5e7e9a333df3f360771fb8 not found: ID does not exist" Oct 05 07:12:46 crc kubenswrapper[4846]: I1005 07:12:46.181159 4846 scope.go:117] "RemoveContainer" containerID="62b5bb44740009feb2f5c21f8547aa59f5f5439ad90cfa4571c6b9344b97a1ec" Oct 05 07:12:46 crc kubenswrapper[4846]: E1005 07:12:46.181531 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"62b5bb44740009feb2f5c21f8547aa59f5f5439ad90cfa4571c6b9344b97a1ec\": container with ID starting with 62b5bb44740009feb2f5c21f8547aa59f5f5439ad90cfa4571c6b9344b97a1ec not found: ID does not exist" containerID="62b5bb44740009feb2f5c21f8547aa59f5f5439ad90cfa4571c6b9344b97a1ec" Oct 05 07:12:46 crc kubenswrapper[4846]: I1005 07:12:46.181564 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"62b5bb44740009feb2f5c21f8547aa59f5f5439ad90cfa4571c6b9344b97a1ec"} err="failed to get container status \"62b5bb44740009feb2f5c21f8547aa59f5f5439ad90cfa4571c6b9344b97a1ec\": rpc error: code = NotFound desc = could not find container \"62b5bb44740009feb2f5c21f8547aa59f5f5439ad90cfa4571c6b9344b97a1ec\": container with ID starting with 62b5bb44740009feb2f5c21f8547aa59f5f5439ad90cfa4571c6b9344b97a1ec not found: ID does not exist" Oct 05 07:12:46 crc kubenswrapper[4846]: I1005 07:12:46.181585 4846 scope.go:117] "RemoveContainer" containerID="69253aadbb0eb74c8ce5ee557dbed2535764f5d7a9a4bedd49b0b0dd420ad20a" Oct 05 07:12:46 crc kubenswrapper[4846]: E1005 07:12:46.181968 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"69253aadbb0eb74c8ce5ee557dbed2535764f5d7a9a4bedd49b0b0dd420ad20a\": container with ID starting with 69253aadbb0eb74c8ce5ee557dbed2535764f5d7a9a4bedd49b0b0dd420ad20a not found: ID does not exist" containerID="69253aadbb0eb74c8ce5ee557dbed2535764f5d7a9a4bedd49b0b0dd420ad20a" Oct 05 07:12:46 crc kubenswrapper[4846]: I1005 07:12:46.182003 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"69253aadbb0eb74c8ce5ee557dbed2535764f5d7a9a4bedd49b0b0dd420ad20a"} err="failed to get container status \"69253aadbb0eb74c8ce5ee557dbed2535764f5d7a9a4bedd49b0b0dd420ad20a\": rpc error: code = NotFound desc = could not find container \"69253aadbb0eb74c8ce5ee557dbed2535764f5d7a9a4bedd49b0b0dd420ad20a\": container with ID starting with 69253aadbb0eb74c8ce5ee557dbed2535764f5d7a9a4bedd49b0b0dd420ad20a not found: ID does not exist" Oct 05 07:12:46 crc kubenswrapper[4846]: I1005 07:12:46.510407 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0eb9b2f-f93b-41f5-a790-558a6bcdf0be" path="/var/lib/kubelet/pods/a0eb9b2f-f93b-41f5-a790-558a6bcdf0be/volumes" Oct 05 07:12:53 crc kubenswrapper[4846]: I1005 07:12:53.324950 4846 patch_prober.go:28] interesting pod/machine-config-daemon-fscvf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 07:12:53 crc kubenswrapper[4846]: I1005 07:12:53.325657 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 07:12:54 crc kubenswrapper[4846]: I1005 07:12:54.485634 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-59cxg"] Oct 05 07:12:54 crc kubenswrapper[4846]: E1005 07:12:54.486277 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a0eb9b2f-f93b-41f5-a790-558a6bcdf0be" containerName="extract-utilities" Oct 05 07:12:54 crc kubenswrapper[4846]: I1005 07:12:54.486295 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="a0eb9b2f-f93b-41f5-a790-558a6bcdf0be" containerName="extract-utilities" Oct 05 07:12:54 crc kubenswrapper[4846]: E1005 07:12:54.486318 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a0eb9b2f-f93b-41f5-a790-558a6bcdf0be" containerName="registry-server" Oct 05 07:12:54 crc kubenswrapper[4846]: I1005 07:12:54.486326 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="a0eb9b2f-f93b-41f5-a790-558a6bcdf0be" containerName="registry-server" Oct 05 07:12:54 crc kubenswrapper[4846]: E1005 07:12:54.486355 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a0eb9b2f-f93b-41f5-a790-558a6bcdf0be" containerName="extract-content" Oct 05 07:12:54 crc kubenswrapper[4846]: I1005 07:12:54.486363 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="a0eb9b2f-f93b-41f5-a790-558a6bcdf0be" containerName="extract-content" Oct 05 07:12:54 crc kubenswrapper[4846]: I1005 07:12:54.486555 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="a0eb9b2f-f93b-41f5-a790-558a6bcdf0be" containerName="registry-server" Oct 05 07:12:54 crc kubenswrapper[4846]: I1005 07:12:54.487775 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-59cxg" Oct 05 07:12:54 crc kubenswrapper[4846]: I1005 07:12:54.519660 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-59cxg"] Oct 05 07:12:54 crc kubenswrapper[4846]: I1005 07:12:54.567204 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/23b5747e-805f-4ea1-9868-927bac9150b2-utilities\") pod \"redhat-operators-59cxg\" (UID: \"23b5747e-805f-4ea1-9868-927bac9150b2\") " pod="openshift-marketplace/redhat-operators-59cxg" Oct 05 07:12:54 crc kubenswrapper[4846]: I1005 07:12:54.567425 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b5ghd\" (UniqueName: \"kubernetes.io/projected/23b5747e-805f-4ea1-9868-927bac9150b2-kube-api-access-b5ghd\") pod \"redhat-operators-59cxg\" (UID: \"23b5747e-805f-4ea1-9868-927bac9150b2\") " pod="openshift-marketplace/redhat-operators-59cxg" Oct 05 07:12:54 crc kubenswrapper[4846]: I1005 07:12:54.567537 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/23b5747e-805f-4ea1-9868-927bac9150b2-catalog-content\") pod \"redhat-operators-59cxg\" (UID: \"23b5747e-805f-4ea1-9868-927bac9150b2\") " pod="openshift-marketplace/redhat-operators-59cxg" Oct 05 07:12:54 crc kubenswrapper[4846]: I1005 07:12:54.669094 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/23b5747e-805f-4ea1-9868-927bac9150b2-utilities\") pod \"redhat-operators-59cxg\" (UID: \"23b5747e-805f-4ea1-9868-927bac9150b2\") " pod="openshift-marketplace/redhat-operators-59cxg" Oct 05 07:12:54 crc kubenswrapper[4846]: I1005 07:12:54.669224 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b5ghd\" (UniqueName: \"kubernetes.io/projected/23b5747e-805f-4ea1-9868-927bac9150b2-kube-api-access-b5ghd\") pod \"redhat-operators-59cxg\" (UID: \"23b5747e-805f-4ea1-9868-927bac9150b2\") " pod="openshift-marketplace/redhat-operators-59cxg" Oct 05 07:12:54 crc kubenswrapper[4846]: I1005 07:12:54.669252 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/23b5747e-805f-4ea1-9868-927bac9150b2-catalog-content\") pod \"redhat-operators-59cxg\" (UID: \"23b5747e-805f-4ea1-9868-927bac9150b2\") " pod="openshift-marketplace/redhat-operators-59cxg" Oct 05 07:12:54 crc kubenswrapper[4846]: I1005 07:12:54.669711 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/23b5747e-805f-4ea1-9868-927bac9150b2-utilities\") pod \"redhat-operators-59cxg\" (UID: \"23b5747e-805f-4ea1-9868-927bac9150b2\") " pod="openshift-marketplace/redhat-operators-59cxg" Oct 05 07:12:54 crc kubenswrapper[4846]: I1005 07:12:54.669748 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/23b5747e-805f-4ea1-9868-927bac9150b2-catalog-content\") pod \"redhat-operators-59cxg\" (UID: \"23b5747e-805f-4ea1-9868-927bac9150b2\") " pod="openshift-marketplace/redhat-operators-59cxg" Oct 05 07:12:54 crc kubenswrapper[4846]: I1005 07:12:54.688931 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b5ghd\" (UniqueName: \"kubernetes.io/projected/23b5747e-805f-4ea1-9868-927bac9150b2-kube-api-access-b5ghd\") pod \"redhat-operators-59cxg\" (UID: \"23b5747e-805f-4ea1-9868-927bac9150b2\") " pod="openshift-marketplace/redhat-operators-59cxg" Oct 05 07:12:54 crc kubenswrapper[4846]: I1005 07:12:54.812567 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-59cxg" Oct 05 07:12:55 crc kubenswrapper[4846]: I1005 07:12:55.250227 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-59cxg"] Oct 05 07:12:56 crc kubenswrapper[4846]: I1005 07:12:56.210980 4846 generic.go:334] "Generic (PLEG): container finished" podID="23b5747e-805f-4ea1-9868-927bac9150b2" containerID="ce1cc56f7162919482039ede6c396514f481e447a7deb05813c144692c4546e4" exitCode=0 Oct 05 07:12:56 crc kubenswrapper[4846]: I1005 07:12:56.211151 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-59cxg" event={"ID":"23b5747e-805f-4ea1-9868-927bac9150b2","Type":"ContainerDied","Data":"ce1cc56f7162919482039ede6c396514f481e447a7deb05813c144692c4546e4"} Oct 05 07:12:56 crc kubenswrapper[4846]: I1005 07:12:56.212858 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-59cxg" event={"ID":"23b5747e-805f-4ea1-9868-927bac9150b2","Type":"ContainerStarted","Data":"9fb0e8cc7f42ae86c9de5bf1ab988eab07c4574fcb2965afa7372615b352720f"} Oct 05 07:12:57 crc kubenswrapper[4846]: E1005 07:12:57.944419 4846 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod23b5747e_805f_4ea1_9868_927bac9150b2.slice/crio-conmon-412b911ab3d5418fb513afb7f10fb29e93cea8f20bac140789fb233eab19cb39.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod23b5747e_805f_4ea1_9868_927bac9150b2.slice/crio-412b911ab3d5418fb513afb7f10fb29e93cea8f20bac140789fb233eab19cb39.scope\": RecentStats: unable to find data in memory cache]" Oct 05 07:12:58 crc kubenswrapper[4846]: I1005 07:12:58.234598 4846 generic.go:334] "Generic (PLEG): container finished" podID="23b5747e-805f-4ea1-9868-927bac9150b2" containerID="412b911ab3d5418fb513afb7f10fb29e93cea8f20bac140789fb233eab19cb39" exitCode=0 Oct 05 07:12:58 crc kubenswrapper[4846]: I1005 07:12:58.234654 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-59cxg" event={"ID":"23b5747e-805f-4ea1-9868-927bac9150b2","Type":"ContainerDied","Data":"412b911ab3d5418fb513afb7f10fb29e93cea8f20bac140789fb233eab19cb39"} Oct 05 07:12:59 crc kubenswrapper[4846]: I1005 07:12:59.247287 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-59cxg" event={"ID":"23b5747e-805f-4ea1-9868-927bac9150b2","Type":"ContainerStarted","Data":"92a597dea70f98b8cf17bdf8707a2811e6d9b40f226760437ba00736c0a25230"} Oct 05 07:12:59 crc kubenswrapper[4846]: I1005 07:12:59.271675 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-59cxg" podStartSLOduration=2.721164824 podStartE2EDuration="5.271651183s" podCreationTimestamp="2025-10-05 07:12:54 +0000 UTC" firstStartedPulling="2025-10-05 07:12:56.214048608 +0000 UTC m=+1498.454901413" lastFinishedPulling="2025-10-05 07:12:58.764534987 +0000 UTC m=+1501.005387772" observedRunningTime="2025-10-05 07:12:59.26703378 +0000 UTC m=+1501.507886605" watchObservedRunningTime="2025-10-05 07:12:59.271651183 +0000 UTC m=+1501.512503968" Oct 05 07:13:04 crc kubenswrapper[4846]: I1005 07:13:04.813253 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-59cxg" Oct 05 07:13:04 crc kubenswrapper[4846]: I1005 07:13:04.814787 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-59cxg" Oct 05 07:13:04 crc kubenswrapper[4846]: I1005 07:13:04.903925 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-59cxg" Oct 05 07:13:05 crc kubenswrapper[4846]: I1005 07:13:05.352568 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-59cxg" Oct 05 07:13:05 crc kubenswrapper[4846]: I1005 07:13:05.395829 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-59cxg"] Oct 05 07:13:07 crc kubenswrapper[4846]: I1005 07:13:07.324830 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-59cxg" podUID="23b5747e-805f-4ea1-9868-927bac9150b2" containerName="registry-server" containerID="cri-o://92a597dea70f98b8cf17bdf8707a2811e6d9b40f226760437ba00736c0a25230" gracePeriod=2 Oct 05 07:13:07 crc kubenswrapper[4846]: I1005 07:13:07.755949 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-59cxg" Oct 05 07:13:07 crc kubenswrapper[4846]: I1005 07:13:07.883458 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/23b5747e-805f-4ea1-9868-927bac9150b2-catalog-content\") pod \"23b5747e-805f-4ea1-9868-927bac9150b2\" (UID: \"23b5747e-805f-4ea1-9868-927bac9150b2\") " Oct 05 07:13:07 crc kubenswrapper[4846]: I1005 07:13:07.883595 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/23b5747e-805f-4ea1-9868-927bac9150b2-utilities\") pod \"23b5747e-805f-4ea1-9868-927bac9150b2\" (UID: \"23b5747e-805f-4ea1-9868-927bac9150b2\") " Oct 05 07:13:07 crc kubenswrapper[4846]: I1005 07:13:07.883653 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b5ghd\" (UniqueName: \"kubernetes.io/projected/23b5747e-805f-4ea1-9868-927bac9150b2-kube-api-access-b5ghd\") pod \"23b5747e-805f-4ea1-9868-927bac9150b2\" (UID: \"23b5747e-805f-4ea1-9868-927bac9150b2\") " Oct 05 07:13:07 crc kubenswrapper[4846]: I1005 07:13:07.885148 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/23b5747e-805f-4ea1-9868-927bac9150b2-utilities" (OuterVolumeSpecName: "utilities") pod "23b5747e-805f-4ea1-9868-927bac9150b2" (UID: "23b5747e-805f-4ea1-9868-927bac9150b2"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 07:13:07 crc kubenswrapper[4846]: I1005 07:13:07.891340 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/23b5747e-805f-4ea1-9868-927bac9150b2-kube-api-access-b5ghd" (OuterVolumeSpecName: "kube-api-access-b5ghd") pod "23b5747e-805f-4ea1-9868-927bac9150b2" (UID: "23b5747e-805f-4ea1-9868-927bac9150b2"). InnerVolumeSpecName "kube-api-access-b5ghd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:13:07 crc kubenswrapper[4846]: I1005 07:13:07.986089 4846 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/23b5747e-805f-4ea1-9868-927bac9150b2-utilities\") on node \"crc\" DevicePath \"\"" Oct 05 07:13:07 crc kubenswrapper[4846]: I1005 07:13:07.986158 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b5ghd\" (UniqueName: \"kubernetes.io/projected/23b5747e-805f-4ea1-9868-927bac9150b2-kube-api-access-b5ghd\") on node \"crc\" DevicePath \"\"" Oct 05 07:13:08 crc kubenswrapper[4846]: I1005 07:13:08.339988 4846 generic.go:334] "Generic (PLEG): container finished" podID="23b5747e-805f-4ea1-9868-927bac9150b2" containerID="92a597dea70f98b8cf17bdf8707a2811e6d9b40f226760437ba00736c0a25230" exitCode=0 Oct 05 07:13:08 crc kubenswrapper[4846]: I1005 07:13:08.340037 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-59cxg" event={"ID":"23b5747e-805f-4ea1-9868-927bac9150b2","Type":"ContainerDied","Data":"92a597dea70f98b8cf17bdf8707a2811e6d9b40f226760437ba00736c0a25230"} Oct 05 07:13:08 crc kubenswrapper[4846]: I1005 07:13:08.340069 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-59cxg" event={"ID":"23b5747e-805f-4ea1-9868-927bac9150b2","Type":"ContainerDied","Data":"9fb0e8cc7f42ae86c9de5bf1ab988eab07c4574fcb2965afa7372615b352720f"} Oct 05 07:13:08 crc kubenswrapper[4846]: I1005 07:13:08.340071 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-59cxg" Oct 05 07:13:08 crc kubenswrapper[4846]: I1005 07:13:08.340089 4846 scope.go:117] "RemoveContainer" containerID="92a597dea70f98b8cf17bdf8707a2811e6d9b40f226760437ba00736c0a25230" Oct 05 07:13:08 crc kubenswrapper[4846]: I1005 07:13:08.380257 4846 scope.go:117] "RemoveContainer" containerID="412b911ab3d5418fb513afb7f10fb29e93cea8f20bac140789fb233eab19cb39" Oct 05 07:13:08 crc kubenswrapper[4846]: I1005 07:13:08.417747 4846 scope.go:117] "RemoveContainer" containerID="ce1cc56f7162919482039ede6c396514f481e447a7deb05813c144692c4546e4" Oct 05 07:13:08 crc kubenswrapper[4846]: I1005 07:13:08.453627 4846 scope.go:117] "RemoveContainer" containerID="92a597dea70f98b8cf17bdf8707a2811e6d9b40f226760437ba00736c0a25230" Oct 05 07:13:08 crc kubenswrapper[4846]: E1005 07:13:08.454321 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"92a597dea70f98b8cf17bdf8707a2811e6d9b40f226760437ba00736c0a25230\": container with ID starting with 92a597dea70f98b8cf17bdf8707a2811e6d9b40f226760437ba00736c0a25230 not found: ID does not exist" containerID="92a597dea70f98b8cf17bdf8707a2811e6d9b40f226760437ba00736c0a25230" Oct 05 07:13:08 crc kubenswrapper[4846]: I1005 07:13:08.454384 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"92a597dea70f98b8cf17bdf8707a2811e6d9b40f226760437ba00736c0a25230"} err="failed to get container status \"92a597dea70f98b8cf17bdf8707a2811e6d9b40f226760437ba00736c0a25230\": rpc error: code = NotFound desc = could not find container \"92a597dea70f98b8cf17bdf8707a2811e6d9b40f226760437ba00736c0a25230\": container with ID starting with 92a597dea70f98b8cf17bdf8707a2811e6d9b40f226760437ba00736c0a25230 not found: ID does not exist" Oct 05 07:13:08 crc kubenswrapper[4846]: I1005 07:13:08.454418 4846 scope.go:117] "RemoveContainer" containerID="412b911ab3d5418fb513afb7f10fb29e93cea8f20bac140789fb233eab19cb39" Oct 05 07:13:08 crc kubenswrapper[4846]: E1005 07:13:08.455041 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"412b911ab3d5418fb513afb7f10fb29e93cea8f20bac140789fb233eab19cb39\": container with ID starting with 412b911ab3d5418fb513afb7f10fb29e93cea8f20bac140789fb233eab19cb39 not found: ID does not exist" containerID="412b911ab3d5418fb513afb7f10fb29e93cea8f20bac140789fb233eab19cb39" Oct 05 07:13:08 crc kubenswrapper[4846]: I1005 07:13:08.455098 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"412b911ab3d5418fb513afb7f10fb29e93cea8f20bac140789fb233eab19cb39"} err="failed to get container status \"412b911ab3d5418fb513afb7f10fb29e93cea8f20bac140789fb233eab19cb39\": rpc error: code = NotFound desc = could not find container \"412b911ab3d5418fb513afb7f10fb29e93cea8f20bac140789fb233eab19cb39\": container with ID starting with 412b911ab3d5418fb513afb7f10fb29e93cea8f20bac140789fb233eab19cb39 not found: ID does not exist" Oct 05 07:13:08 crc kubenswrapper[4846]: I1005 07:13:08.455133 4846 scope.go:117] "RemoveContainer" containerID="ce1cc56f7162919482039ede6c396514f481e447a7deb05813c144692c4546e4" Oct 05 07:13:08 crc kubenswrapper[4846]: E1005 07:13:08.455521 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ce1cc56f7162919482039ede6c396514f481e447a7deb05813c144692c4546e4\": container with ID starting with ce1cc56f7162919482039ede6c396514f481e447a7deb05813c144692c4546e4 not found: ID does not exist" containerID="ce1cc56f7162919482039ede6c396514f481e447a7deb05813c144692c4546e4" Oct 05 07:13:08 crc kubenswrapper[4846]: I1005 07:13:08.455552 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ce1cc56f7162919482039ede6c396514f481e447a7deb05813c144692c4546e4"} err="failed to get container status \"ce1cc56f7162919482039ede6c396514f481e447a7deb05813c144692c4546e4\": rpc error: code = NotFound desc = could not find container \"ce1cc56f7162919482039ede6c396514f481e447a7deb05813c144692c4546e4\": container with ID starting with ce1cc56f7162919482039ede6c396514f481e447a7deb05813c144692c4546e4 not found: ID does not exist" Oct 05 07:13:09 crc kubenswrapper[4846]: I1005 07:13:09.097054 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/23b5747e-805f-4ea1-9868-927bac9150b2-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "23b5747e-805f-4ea1-9868-927bac9150b2" (UID: "23b5747e-805f-4ea1-9868-927bac9150b2"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 07:13:09 crc kubenswrapper[4846]: I1005 07:13:09.106809 4846 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/23b5747e-805f-4ea1-9868-927bac9150b2-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 05 07:13:09 crc kubenswrapper[4846]: I1005 07:13:09.293592 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-59cxg"] Oct 05 07:13:09 crc kubenswrapper[4846]: I1005 07:13:09.316118 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-59cxg"] Oct 05 07:13:10 crc kubenswrapper[4846]: I1005 07:13:10.516017 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="23b5747e-805f-4ea1-9868-927bac9150b2" path="/var/lib/kubelet/pods/23b5747e-805f-4ea1-9868-927bac9150b2/volumes" Oct 05 07:13:23 crc kubenswrapper[4846]: I1005 07:13:23.324523 4846 patch_prober.go:28] interesting pod/machine-config-daemon-fscvf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 07:13:23 crc kubenswrapper[4846]: I1005 07:13:23.325396 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 07:13:23 crc kubenswrapper[4846]: I1005 07:13:23.325484 4846 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" Oct 05 07:13:23 crc kubenswrapper[4846]: I1005 07:13:23.326584 4846 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"7d363dbd53ee92c0974009df77db8574365e53ef73df754e1c6fabb86787d68a"} pod="openshift-machine-config-operator/machine-config-daemon-fscvf" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 05 07:13:23 crc kubenswrapper[4846]: I1005 07:13:23.326709 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" containerName="machine-config-daemon" containerID="cri-o://7d363dbd53ee92c0974009df77db8574365e53ef73df754e1c6fabb86787d68a" gracePeriod=600 Oct 05 07:13:23 crc kubenswrapper[4846]: E1005 07:13:23.458482 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 07:13:23 crc kubenswrapper[4846]: I1005 07:13:23.510944 4846 generic.go:334] "Generic (PLEG): container finished" podID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" containerID="7d363dbd53ee92c0974009df77db8574365e53ef73df754e1c6fabb86787d68a" exitCode=0 Oct 05 07:13:23 crc kubenswrapper[4846]: I1005 07:13:23.510992 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" event={"ID":"ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7","Type":"ContainerDied","Data":"7d363dbd53ee92c0974009df77db8574365e53ef73df754e1c6fabb86787d68a"} Oct 05 07:13:23 crc kubenswrapper[4846]: I1005 07:13:23.511035 4846 scope.go:117] "RemoveContainer" containerID="11b9703888d7e63159496cf8c65783a81599439b16d1a1e5677afe95613a82de" Oct 05 07:13:23 crc kubenswrapper[4846]: I1005 07:13:23.511746 4846 scope.go:117] "RemoveContainer" containerID="7d363dbd53ee92c0974009df77db8574365e53ef73df754e1c6fabb86787d68a" Oct 05 07:13:23 crc kubenswrapper[4846]: E1005 07:13:23.512069 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 07:13:26 crc kubenswrapper[4846]: I1005 07:13:26.761729 4846 scope.go:117] "RemoveContainer" containerID="bb66ec24b24cb5604d450f4beb64f774dd42cb6a774b6536eca16fcc238ae31a" Oct 05 07:13:26 crc kubenswrapper[4846]: I1005 07:13:26.820211 4846 scope.go:117] "RemoveContainer" containerID="65cb7c9f9d81cbaf7f1a0cefd3a7c9535986612c6a0a836f3800e04819ebc0dc" Oct 05 07:13:26 crc kubenswrapper[4846]: I1005 07:13:26.857006 4846 scope.go:117] "RemoveContainer" containerID="b9645c4a90f8320793bcfa361c3d8c571e3730e1f40c3ec1b6cd107151e4d1c2" Oct 05 07:13:26 crc kubenswrapper[4846]: I1005 07:13:26.907855 4846 scope.go:117] "RemoveContainer" containerID="b1b00d22e5b0700b970a8485f74339e8139dbc0cb0edc8276f797b2d2f338545" Oct 05 07:13:26 crc kubenswrapper[4846]: I1005 07:13:26.936436 4846 scope.go:117] "RemoveContainer" containerID="a387124df9f8bc90bc9f6818e068798ea86dcd47589df9f1a1df098f8dfdab60" Oct 05 07:13:26 crc kubenswrapper[4846]: I1005 07:13:26.971207 4846 scope.go:117] "RemoveContainer" containerID="b902b0193cb363be23ae30173cd0b9b076f21d8ff2cb0927214f0825a53830ed" Oct 05 07:13:27 crc kubenswrapper[4846]: I1005 07:13:27.008055 4846 scope.go:117] "RemoveContainer" containerID="7cd6caeb5a6926e34fdec088400b24c5c5bc48b887c5e275a91ac1896351fdc3" Oct 05 07:13:27 crc kubenswrapper[4846]: I1005 07:13:27.056440 4846 scope.go:117] "RemoveContainer" containerID="75309f967d618b39cc1b78b8d76b5a811e908820abd07bd78a313ec106dcce6c" Oct 05 07:13:27 crc kubenswrapper[4846]: I1005 07:13:27.113773 4846 scope.go:117] "RemoveContainer" containerID="715bb8fb0bc7969f7b9a465afe9f43ee39ae63611715e80bbcb41d0b5b4445de" Oct 05 07:13:27 crc kubenswrapper[4846]: I1005 07:13:27.137506 4846 scope.go:117] "RemoveContainer" containerID="cf40f2c3e601ab1a12429ff595152653f7656b17ce27280b2f147555aeee04d2" Oct 05 07:13:27 crc kubenswrapper[4846]: I1005 07:13:27.159554 4846 scope.go:117] "RemoveContainer" containerID="1b20e0f9240e1bcc74644cccf0d33c9f6e702a6015a35a33c23dbd86eb5b3d91" Oct 05 07:13:27 crc kubenswrapper[4846]: I1005 07:13:27.178422 4846 scope.go:117] "RemoveContainer" containerID="2fd09fbd97edd50e2b9f40d0f48d643a047010b082f6e418ea82fb818d979cce" Oct 05 07:13:27 crc kubenswrapper[4846]: I1005 07:13:27.200799 4846 scope.go:117] "RemoveContainer" containerID="fd96bbff85b7b2a5303006a90a5cd90ed2a0e07acc25eb2d05bccf60b24cfee7" Oct 05 07:13:27 crc kubenswrapper[4846]: I1005 07:13:27.228316 4846 scope.go:117] "RemoveContainer" containerID="6f45c438b8b209cc382e7a14455205149b1abf123df3b2467fe12ebe7d71967f" Oct 05 07:13:27 crc kubenswrapper[4846]: I1005 07:13:27.271828 4846 scope.go:117] "RemoveContainer" containerID="bc74462851218e64b2af029b68e74068a836521458f244ad01e383cfa57ca002" Oct 05 07:13:27 crc kubenswrapper[4846]: I1005 07:13:27.303291 4846 scope.go:117] "RemoveContainer" containerID="f80ec72ff1ba79fd1d723988d040ed000dc6a49b01ce6daeba2f5db0549784a5" Oct 05 07:13:27 crc kubenswrapper[4846]: I1005 07:13:27.331076 4846 scope.go:117] "RemoveContainer" containerID="2cd30fb4a2bf37a8f593325296e9f9f11eaa096979e85337e1f058f23b5bd224" Oct 05 07:13:37 crc kubenswrapper[4846]: I1005 07:13:37.498641 4846 scope.go:117] "RemoveContainer" containerID="7d363dbd53ee92c0974009df77db8574365e53ef73df754e1c6fabb86787d68a" Oct 05 07:13:37 crc kubenswrapper[4846]: E1005 07:13:37.499752 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 07:13:50 crc kubenswrapper[4846]: I1005 07:13:50.498150 4846 scope.go:117] "RemoveContainer" containerID="7d363dbd53ee92c0974009df77db8574365e53ef73df754e1c6fabb86787d68a" Oct 05 07:13:50 crc kubenswrapper[4846]: E1005 07:13:50.499259 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 07:14:04 crc kubenswrapper[4846]: I1005 07:14:04.497825 4846 scope.go:117] "RemoveContainer" containerID="7d363dbd53ee92c0974009df77db8574365e53ef73df754e1c6fabb86787d68a" Oct 05 07:14:04 crc kubenswrapper[4846]: E1005 07:14:04.498996 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 07:14:18 crc kubenswrapper[4846]: I1005 07:14:18.503497 4846 scope.go:117] "RemoveContainer" containerID="7d363dbd53ee92c0974009df77db8574365e53ef73df754e1c6fabb86787d68a" Oct 05 07:14:18 crc kubenswrapper[4846]: E1005 07:14:18.504434 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 07:14:27 crc kubenswrapper[4846]: I1005 07:14:27.673762 4846 scope.go:117] "RemoveContainer" containerID="69871ad24cf3a1c323c9675966861d808a81181ea8ee899b1ff38c1e60c644f9" Oct 05 07:14:27 crc kubenswrapper[4846]: I1005 07:14:27.720872 4846 scope.go:117] "RemoveContainer" containerID="155de7225ff92721c9958b6c64a09e10453594fc7642a9165ae9799de8832fde" Oct 05 07:14:27 crc kubenswrapper[4846]: I1005 07:14:27.756847 4846 scope.go:117] "RemoveContainer" containerID="8adc95b909659a53237badc065088ebc75df6ae5e870528ac833dde781fc0d99" Oct 05 07:14:27 crc kubenswrapper[4846]: I1005 07:14:27.815650 4846 scope.go:117] "RemoveContainer" containerID="27a50aa96a7a8bb54961eddcd8037b7777a781e62487ae4213c4c228650c5b06" Oct 05 07:14:27 crc kubenswrapper[4846]: I1005 07:14:27.866857 4846 scope.go:117] "RemoveContainer" containerID="adb9e57c3af0eccf4eb1a4d4fa7671627385b5ebffd25943a05f6247adfca3ce" Oct 05 07:14:33 crc kubenswrapper[4846]: I1005 07:14:33.497815 4846 scope.go:117] "RemoveContainer" containerID="7d363dbd53ee92c0974009df77db8574365e53ef73df754e1c6fabb86787d68a" Oct 05 07:14:33 crc kubenswrapper[4846]: E1005 07:14:33.501015 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 07:14:46 crc kubenswrapper[4846]: I1005 07:14:46.497781 4846 scope.go:117] "RemoveContainer" containerID="7d363dbd53ee92c0974009df77db8574365e53ef73df754e1c6fabb86787d68a" Oct 05 07:14:46 crc kubenswrapper[4846]: E1005 07:14:46.498533 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 07:15:00 crc kubenswrapper[4846]: I1005 07:15:00.176408 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29327475-frl5l"] Oct 05 07:15:00 crc kubenswrapper[4846]: E1005 07:15:00.178603 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="23b5747e-805f-4ea1-9868-927bac9150b2" containerName="extract-utilities" Oct 05 07:15:00 crc kubenswrapper[4846]: I1005 07:15:00.178634 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="23b5747e-805f-4ea1-9868-927bac9150b2" containerName="extract-utilities" Oct 05 07:15:00 crc kubenswrapper[4846]: E1005 07:15:00.178671 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="23b5747e-805f-4ea1-9868-927bac9150b2" containerName="extract-content" Oct 05 07:15:00 crc kubenswrapper[4846]: I1005 07:15:00.178684 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="23b5747e-805f-4ea1-9868-927bac9150b2" containerName="extract-content" Oct 05 07:15:00 crc kubenswrapper[4846]: E1005 07:15:00.178730 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="23b5747e-805f-4ea1-9868-927bac9150b2" containerName="registry-server" Oct 05 07:15:00 crc kubenswrapper[4846]: I1005 07:15:00.178751 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="23b5747e-805f-4ea1-9868-927bac9150b2" containerName="registry-server" Oct 05 07:15:00 crc kubenswrapper[4846]: I1005 07:15:00.179002 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="23b5747e-805f-4ea1-9868-927bac9150b2" containerName="registry-server" Oct 05 07:15:00 crc kubenswrapper[4846]: I1005 07:15:00.179921 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29327475-frl5l" Oct 05 07:15:00 crc kubenswrapper[4846]: I1005 07:15:00.182572 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 05 07:15:00 crc kubenswrapper[4846]: I1005 07:15:00.182896 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 05 07:15:00 crc kubenswrapper[4846]: I1005 07:15:00.188945 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29327475-frl5l"] Oct 05 07:15:00 crc kubenswrapper[4846]: I1005 07:15:00.319680 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mt5br\" (UniqueName: \"kubernetes.io/projected/b6269daa-7302-4bf7-ae86-9d57935687ea-kube-api-access-mt5br\") pod \"collect-profiles-29327475-frl5l\" (UID: \"b6269daa-7302-4bf7-ae86-9d57935687ea\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327475-frl5l" Oct 05 07:15:00 crc kubenswrapper[4846]: I1005 07:15:00.319811 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b6269daa-7302-4bf7-ae86-9d57935687ea-secret-volume\") pod \"collect-profiles-29327475-frl5l\" (UID: \"b6269daa-7302-4bf7-ae86-9d57935687ea\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327475-frl5l" Oct 05 07:15:00 crc kubenswrapper[4846]: I1005 07:15:00.319869 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b6269daa-7302-4bf7-ae86-9d57935687ea-config-volume\") pod \"collect-profiles-29327475-frl5l\" (UID: \"b6269daa-7302-4bf7-ae86-9d57935687ea\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327475-frl5l" Oct 05 07:15:00 crc kubenswrapper[4846]: I1005 07:15:00.421740 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mt5br\" (UniqueName: \"kubernetes.io/projected/b6269daa-7302-4bf7-ae86-9d57935687ea-kube-api-access-mt5br\") pod \"collect-profiles-29327475-frl5l\" (UID: \"b6269daa-7302-4bf7-ae86-9d57935687ea\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327475-frl5l" Oct 05 07:15:00 crc kubenswrapper[4846]: I1005 07:15:00.421884 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b6269daa-7302-4bf7-ae86-9d57935687ea-secret-volume\") pod \"collect-profiles-29327475-frl5l\" (UID: \"b6269daa-7302-4bf7-ae86-9d57935687ea\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327475-frl5l" Oct 05 07:15:00 crc kubenswrapper[4846]: I1005 07:15:00.421981 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b6269daa-7302-4bf7-ae86-9d57935687ea-config-volume\") pod \"collect-profiles-29327475-frl5l\" (UID: \"b6269daa-7302-4bf7-ae86-9d57935687ea\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327475-frl5l" Oct 05 07:15:00 crc kubenswrapper[4846]: I1005 07:15:00.423820 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b6269daa-7302-4bf7-ae86-9d57935687ea-config-volume\") pod \"collect-profiles-29327475-frl5l\" (UID: \"b6269daa-7302-4bf7-ae86-9d57935687ea\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327475-frl5l" Oct 05 07:15:00 crc kubenswrapper[4846]: I1005 07:15:00.431843 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b6269daa-7302-4bf7-ae86-9d57935687ea-secret-volume\") pod \"collect-profiles-29327475-frl5l\" (UID: \"b6269daa-7302-4bf7-ae86-9d57935687ea\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327475-frl5l" Oct 05 07:15:00 crc kubenswrapper[4846]: I1005 07:15:00.454888 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mt5br\" (UniqueName: \"kubernetes.io/projected/b6269daa-7302-4bf7-ae86-9d57935687ea-kube-api-access-mt5br\") pod \"collect-profiles-29327475-frl5l\" (UID: \"b6269daa-7302-4bf7-ae86-9d57935687ea\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327475-frl5l" Oct 05 07:15:00 crc kubenswrapper[4846]: I1005 07:15:00.498854 4846 scope.go:117] "RemoveContainer" containerID="7d363dbd53ee92c0974009df77db8574365e53ef73df754e1c6fabb86787d68a" Oct 05 07:15:00 crc kubenswrapper[4846]: E1005 07:15:00.499270 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 07:15:00 crc kubenswrapper[4846]: I1005 07:15:00.508507 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29327475-frl5l" Oct 05 07:15:00 crc kubenswrapper[4846]: I1005 07:15:00.747674 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29327475-frl5l"] Oct 05 07:15:01 crc kubenswrapper[4846]: I1005 07:15:01.539009 4846 generic.go:334] "Generic (PLEG): container finished" podID="b6269daa-7302-4bf7-ae86-9d57935687ea" containerID="2581fecf028f7d3f4fef7ce6871f47512432aa79d692b1f91f468755e769e6e8" exitCode=0 Oct 05 07:15:01 crc kubenswrapper[4846]: I1005 07:15:01.539086 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29327475-frl5l" event={"ID":"b6269daa-7302-4bf7-ae86-9d57935687ea","Type":"ContainerDied","Data":"2581fecf028f7d3f4fef7ce6871f47512432aa79d692b1f91f468755e769e6e8"} Oct 05 07:15:01 crc kubenswrapper[4846]: I1005 07:15:01.539361 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29327475-frl5l" event={"ID":"b6269daa-7302-4bf7-ae86-9d57935687ea","Type":"ContainerStarted","Data":"dd4fcfad7637216d60c3451ab2ff8994e9172632e7c0a0b591d1a748aeb1ae0d"} Oct 05 07:15:02 crc kubenswrapper[4846]: I1005 07:15:02.881514 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29327475-frl5l" Oct 05 07:15:03 crc kubenswrapper[4846]: I1005 07:15:03.060905 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mt5br\" (UniqueName: \"kubernetes.io/projected/b6269daa-7302-4bf7-ae86-9d57935687ea-kube-api-access-mt5br\") pod \"b6269daa-7302-4bf7-ae86-9d57935687ea\" (UID: \"b6269daa-7302-4bf7-ae86-9d57935687ea\") " Oct 05 07:15:03 crc kubenswrapper[4846]: I1005 07:15:03.060981 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b6269daa-7302-4bf7-ae86-9d57935687ea-config-volume\") pod \"b6269daa-7302-4bf7-ae86-9d57935687ea\" (UID: \"b6269daa-7302-4bf7-ae86-9d57935687ea\") " Oct 05 07:15:03 crc kubenswrapper[4846]: I1005 07:15:03.061028 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b6269daa-7302-4bf7-ae86-9d57935687ea-secret-volume\") pod \"b6269daa-7302-4bf7-ae86-9d57935687ea\" (UID: \"b6269daa-7302-4bf7-ae86-9d57935687ea\") " Oct 05 07:15:03 crc kubenswrapper[4846]: I1005 07:15:03.062119 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6269daa-7302-4bf7-ae86-9d57935687ea-config-volume" (OuterVolumeSpecName: "config-volume") pod "b6269daa-7302-4bf7-ae86-9d57935687ea" (UID: "b6269daa-7302-4bf7-ae86-9d57935687ea"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 07:15:03 crc kubenswrapper[4846]: I1005 07:15:03.066561 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6269daa-7302-4bf7-ae86-9d57935687ea-kube-api-access-mt5br" (OuterVolumeSpecName: "kube-api-access-mt5br") pod "b6269daa-7302-4bf7-ae86-9d57935687ea" (UID: "b6269daa-7302-4bf7-ae86-9d57935687ea"). InnerVolumeSpecName "kube-api-access-mt5br". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:15:03 crc kubenswrapper[4846]: I1005 07:15:03.067150 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6269daa-7302-4bf7-ae86-9d57935687ea-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "b6269daa-7302-4bf7-ae86-9d57935687ea" (UID: "b6269daa-7302-4bf7-ae86-9d57935687ea"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:15:03 crc kubenswrapper[4846]: I1005 07:15:03.162875 4846 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b6269daa-7302-4bf7-ae86-9d57935687ea-config-volume\") on node \"crc\" DevicePath \"\"" Oct 05 07:15:03 crc kubenswrapper[4846]: I1005 07:15:03.162920 4846 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b6269daa-7302-4bf7-ae86-9d57935687ea-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 05 07:15:03 crc kubenswrapper[4846]: I1005 07:15:03.162931 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mt5br\" (UniqueName: \"kubernetes.io/projected/b6269daa-7302-4bf7-ae86-9d57935687ea-kube-api-access-mt5br\") on node \"crc\" DevicePath \"\"" Oct 05 07:15:03 crc kubenswrapper[4846]: I1005 07:15:03.558113 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29327475-frl5l" event={"ID":"b6269daa-7302-4bf7-ae86-9d57935687ea","Type":"ContainerDied","Data":"dd4fcfad7637216d60c3451ab2ff8994e9172632e7c0a0b591d1a748aeb1ae0d"} Oct 05 07:15:03 crc kubenswrapper[4846]: I1005 07:15:03.558156 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="dd4fcfad7637216d60c3451ab2ff8994e9172632e7c0a0b591d1a748aeb1ae0d" Oct 05 07:15:03 crc kubenswrapper[4846]: I1005 07:15:03.558238 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29327475-frl5l" Oct 05 07:15:13 crc kubenswrapper[4846]: I1005 07:15:13.498482 4846 scope.go:117] "RemoveContainer" containerID="7d363dbd53ee92c0974009df77db8574365e53ef73df754e1c6fabb86787d68a" Oct 05 07:15:13 crc kubenswrapper[4846]: E1005 07:15:13.499488 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 07:15:24 crc kubenswrapper[4846]: I1005 07:15:24.498260 4846 scope.go:117] "RemoveContainer" containerID="7d363dbd53ee92c0974009df77db8574365e53ef73df754e1c6fabb86787d68a" Oct 05 07:15:24 crc kubenswrapper[4846]: E1005 07:15:24.499633 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 07:15:27 crc kubenswrapper[4846]: I1005 07:15:27.972001 4846 scope.go:117] "RemoveContainer" containerID="787d7008e975b8e9978d9240b7d5756af79da41e854c43ea23614d9e93e503b5" Oct 05 07:15:28 crc kubenswrapper[4846]: I1005 07:15:28.038059 4846 scope.go:117] "RemoveContainer" containerID="49603cc789b1bc6ca79b06b68c268619d72a726745c51b51cb8b353e41fcf6b7" Oct 05 07:15:28 crc kubenswrapper[4846]: I1005 07:15:28.064343 4846 scope.go:117] "RemoveContainer" containerID="39e9c49a310f7904a188de2c7d263bafd273efb07adc6a0435e84e760f4fde4c" Oct 05 07:15:28 crc kubenswrapper[4846]: I1005 07:15:28.107203 4846 scope.go:117] "RemoveContainer" containerID="ec596b676d1e2029a427c619a2c09007c32cd26a7cd5b08b7f797a076c0a061e" Oct 05 07:15:28 crc kubenswrapper[4846]: I1005 07:15:28.135732 4846 scope.go:117] "RemoveContainer" containerID="14b045f6d72991175d4f2fa50e9298251749662cf7376add05f11dc4152f31ce" Oct 05 07:15:28 crc kubenswrapper[4846]: I1005 07:15:28.162425 4846 scope.go:117] "RemoveContainer" containerID="ffa470fa54fc1c858cb186b1545ff7805051d894cde1fb62d09980a0f72da492" Oct 05 07:15:28 crc kubenswrapper[4846]: I1005 07:15:28.195748 4846 scope.go:117] "RemoveContainer" containerID="3cf10e313f2fcf0243e3ffd0ebcf570c10c8a178135588f4b07bf18a38f012e1" Oct 05 07:15:36 crc kubenswrapper[4846]: I1005 07:15:36.498028 4846 scope.go:117] "RemoveContainer" containerID="7d363dbd53ee92c0974009df77db8574365e53ef73df754e1c6fabb86787d68a" Oct 05 07:15:36 crc kubenswrapper[4846]: E1005 07:15:36.499043 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 07:15:47 crc kubenswrapper[4846]: I1005 07:15:47.498551 4846 scope.go:117] "RemoveContainer" containerID="7d363dbd53ee92c0974009df77db8574365e53ef73df754e1c6fabb86787d68a" Oct 05 07:15:47 crc kubenswrapper[4846]: E1005 07:15:47.499585 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 07:15:58 crc kubenswrapper[4846]: I1005 07:15:58.507165 4846 scope.go:117] "RemoveContainer" containerID="7d363dbd53ee92c0974009df77db8574365e53ef73df754e1c6fabb86787d68a" Oct 05 07:15:58 crc kubenswrapper[4846]: E1005 07:15:58.508332 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 07:16:13 crc kubenswrapper[4846]: I1005 07:16:13.498001 4846 scope.go:117] "RemoveContainer" containerID="7d363dbd53ee92c0974009df77db8574365e53ef73df754e1c6fabb86787d68a" Oct 05 07:16:13 crc kubenswrapper[4846]: E1005 07:16:13.499586 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 07:16:24 crc kubenswrapper[4846]: I1005 07:16:24.498511 4846 scope.go:117] "RemoveContainer" containerID="7d363dbd53ee92c0974009df77db8574365e53ef73df754e1c6fabb86787d68a" Oct 05 07:16:24 crc kubenswrapper[4846]: E1005 07:16:24.499745 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 07:16:28 crc kubenswrapper[4846]: I1005 07:16:28.329462 4846 scope.go:117] "RemoveContainer" containerID="df3b6f1e9e992a4de6fcc136d2a372cb06b4dc91089da63e8a9ba4a7378ed72d" Oct 05 07:16:28 crc kubenswrapper[4846]: I1005 07:16:28.395966 4846 scope.go:117] "RemoveContainer" containerID="b3e205f106c607cc68f15ea01b24c2b915400791a93310bfef25ced3582cdd85" Oct 05 07:16:28 crc kubenswrapper[4846]: I1005 07:16:28.457766 4846 scope.go:117] "RemoveContainer" containerID="f0993f040dcd6d3145256a44cdc38fc8db9c44f334ee6dc7c33895edf97422b9" Oct 05 07:16:28 crc kubenswrapper[4846]: I1005 07:16:28.518156 4846 scope.go:117] "RemoveContainer" containerID="2a96112701d2d7042bce02efb3a776df18e837a8d8981829c635370de3016348" Oct 05 07:16:38 crc kubenswrapper[4846]: I1005 07:16:38.507071 4846 scope.go:117] "RemoveContainer" containerID="7d363dbd53ee92c0974009df77db8574365e53ef73df754e1c6fabb86787d68a" Oct 05 07:16:38 crc kubenswrapper[4846]: E1005 07:16:38.508282 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 07:16:50 crc kubenswrapper[4846]: I1005 07:16:50.498121 4846 scope.go:117] "RemoveContainer" containerID="7d363dbd53ee92c0974009df77db8574365e53ef73df754e1c6fabb86787d68a" Oct 05 07:16:50 crc kubenswrapper[4846]: E1005 07:16:50.499008 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 07:17:01 crc kubenswrapper[4846]: I1005 07:17:01.497928 4846 scope.go:117] "RemoveContainer" containerID="7d363dbd53ee92c0974009df77db8574365e53ef73df754e1c6fabb86787d68a" Oct 05 07:17:01 crc kubenswrapper[4846]: E1005 07:17:01.499132 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 07:17:16 crc kubenswrapper[4846]: I1005 07:17:16.498305 4846 scope.go:117] "RemoveContainer" containerID="7d363dbd53ee92c0974009df77db8574365e53ef73df754e1c6fabb86787d68a" Oct 05 07:17:16 crc kubenswrapper[4846]: E1005 07:17:16.499539 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 07:17:28 crc kubenswrapper[4846]: I1005 07:17:28.610880 4846 scope.go:117] "RemoveContainer" containerID="b30b71de14fefc7cbe669f3c30ee9c9b2aa5f68d63305a95ba9232ac4a893ad2" Oct 05 07:17:31 crc kubenswrapper[4846]: I1005 07:17:31.498479 4846 scope.go:117] "RemoveContainer" containerID="7d363dbd53ee92c0974009df77db8574365e53ef73df754e1c6fabb86787d68a" Oct 05 07:17:31 crc kubenswrapper[4846]: E1005 07:17:31.499539 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 07:17:43 crc kubenswrapper[4846]: I1005 07:17:43.498035 4846 scope.go:117] "RemoveContainer" containerID="7d363dbd53ee92c0974009df77db8574365e53ef73df754e1c6fabb86787d68a" Oct 05 07:17:43 crc kubenswrapper[4846]: E1005 07:17:43.499528 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 07:17:54 crc kubenswrapper[4846]: I1005 07:17:54.497701 4846 scope.go:117] "RemoveContainer" containerID="7d363dbd53ee92c0974009df77db8574365e53ef73df754e1c6fabb86787d68a" Oct 05 07:17:54 crc kubenswrapper[4846]: E1005 07:17:54.498964 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 07:18:09 crc kubenswrapper[4846]: I1005 07:18:09.497165 4846 scope.go:117] "RemoveContainer" containerID="7d363dbd53ee92c0974009df77db8574365e53ef73df754e1c6fabb86787d68a" Oct 05 07:18:09 crc kubenswrapper[4846]: E1005 07:18:09.498121 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 07:18:24 crc kubenswrapper[4846]: I1005 07:18:24.497296 4846 scope.go:117] "RemoveContainer" containerID="7d363dbd53ee92c0974009df77db8574365e53ef73df754e1c6fabb86787d68a" Oct 05 07:18:25 crc kubenswrapper[4846]: I1005 07:18:25.645427 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" event={"ID":"ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7","Type":"ContainerStarted","Data":"5dcee45f19c6e0d9810c4dda319bcb382224184464f9a3dbabad80d9e701b330"} Oct 05 07:20:53 crc kubenswrapper[4846]: I1005 07:20:53.325381 4846 patch_prober.go:28] interesting pod/machine-config-daemon-fscvf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 07:20:53 crc kubenswrapper[4846]: I1005 07:20:53.326336 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 07:21:23 crc kubenswrapper[4846]: I1005 07:21:23.325424 4846 patch_prober.go:28] interesting pod/machine-config-daemon-fscvf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 07:21:23 crc kubenswrapper[4846]: I1005 07:21:23.326477 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 07:21:53 crc kubenswrapper[4846]: I1005 07:21:53.324966 4846 patch_prober.go:28] interesting pod/machine-config-daemon-fscvf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 07:21:53 crc kubenswrapper[4846]: I1005 07:21:53.325631 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 07:21:53 crc kubenswrapper[4846]: I1005 07:21:53.325701 4846 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" Oct 05 07:21:53 crc kubenswrapper[4846]: I1005 07:21:53.326470 4846 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"5dcee45f19c6e0d9810c4dda319bcb382224184464f9a3dbabad80d9e701b330"} pod="openshift-machine-config-operator/machine-config-daemon-fscvf" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 05 07:21:53 crc kubenswrapper[4846]: I1005 07:21:53.326564 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" containerName="machine-config-daemon" containerID="cri-o://5dcee45f19c6e0d9810c4dda319bcb382224184464f9a3dbabad80d9e701b330" gracePeriod=600 Oct 05 07:21:53 crc kubenswrapper[4846]: I1005 07:21:53.662868 4846 generic.go:334] "Generic (PLEG): container finished" podID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" containerID="5dcee45f19c6e0d9810c4dda319bcb382224184464f9a3dbabad80d9e701b330" exitCode=0 Oct 05 07:21:53 crc kubenswrapper[4846]: I1005 07:21:53.662961 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" event={"ID":"ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7","Type":"ContainerDied","Data":"5dcee45f19c6e0d9810c4dda319bcb382224184464f9a3dbabad80d9e701b330"} Oct 05 07:21:53 crc kubenswrapper[4846]: I1005 07:21:53.663392 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" event={"ID":"ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7","Type":"ContainerStarted","Data":"4817903cb1e5987efe5a26ad35de6b3af2cbb8dfef6b67e286cafc5827c7ded4"} Oct 05 07:21:53 crc kubenswrapper[4846]: I1005 07:21:53.663431 4846 scope.go:117] "RemoveContainer" containerID="7d363dbd53ee92c0974009df77db8574365e53ef73df754e1c6fabb86787d68a" Oct 05 07:22:43 crc kubenswrapper[4846]: I1005 07:22:43.447837 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-8s2mz"] Oct 05 07:22:43 crc kubenswrapper[4846]: E1005 07:22:43.448875 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b6269daa-7302-4bf7-ae86-9d57935687ea" containerName="collect-profiles" Oct 05 07:22:43 crc kubenswrapper[4846]: I1005 07:22:43.448896 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="b6269daa-7302-4bf7-ae86-9d57935687ea" containerName="collect-profiles" Oct 05 07:22:43 crc kubenswrapper[4846]: I1005 07:22:43.449131 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="b6269daa-7302-4bf7-ae86-9d57935687ea" containerName="collect-profiles" Oct 05 07:22:43 crc kubenswrapper[4846]: I1005 07:22:43.450826 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8s2mz" Oct 05 07:22:43 crc kubenswrapper[4846]: I1005 07:22:43.471572 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-8s2mz"] Oct 05 07:22:43 crc kubenswrapper[4846]: I1005 07:22:43.560475 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5ghq5\" (UniqueName: \"kubernetes.io/projected/b005073c-5379-49a7-ade6-7b9611f3812e-kube-api-access-5ghq5\") pod \"certified-operators-8s2mz\" (UID: \"b005073c-5379-49a7-ade6-7b9611f3812e\") " pod="openshift-marketplace/certified-operators-8s2mz" Oct 05 07:22:43 crc kubenswrapper[4846]: I1005 07:22:43.560528 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b005073c-5379-49a7-ade6-7b9611f3812e-utilities\") pod \"certified-operators-8s2mz\" (UID: \"b005073c-5379-49a7-ade6-7b9611f3812e\") " pod="openshift-marketplace/certified-operators-8s2mz" Oct 05 07:22:43 crc kubenswrapper[4846]: I1005 07:22:43.560560 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b005073c-5379-49a7-ade6-7b9611f3812e-catalog-content\") pod \"certified-operators-8s2mz\" (UID: \"b005073c-5379-49a7-ade6-7b9611f3812e\") " pod="openshift-marketplace/certified-operators-8s2mz" Oct 05 07:22:43 crc kubenswrapper[4846]: I1005 07:22:43.661560 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5ghq5\" (UniqueName: \"kubernetes.io/projected/b005073c-5379-49a7-ade6-7b9611f3812e-kube-api-access-5ghq5\") pod \"certified-operators-8s2mz\" (UID: \"b005073c-5379-49a7-ade6-7b9611f3812e\") " pod="openshift-marketplace/certified-operators-8s2mz" Oct 05 07:22:43 crc kubenswrapper[4846]: I1005 07:22:43.661620 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b005073c-5379-49a7-ade6-7b9611f3812e-utilities\") pod \"certified-operators-8s2mz\" (UID: \"b005073c-5379-49a7-ade6-7b9611f3812e\") " pod="openshift-marketplace/certified-operators-8s2mz" Oct 05 07:22:43 crc kubenswrapper[4846]: I1005 07:22:43.661658 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b005073c-5379-49a7-ade6-7b9611f3812e-catalog-content\") pod \"certified-operators-8s2mz\" (UID: \"b005073c-5379-49a7-ade6-7b9611f3812e\") " pod="openshift-marketplace/certified-operators-8s2mz" Oct 05 07:22:43 crc kubenswrapper[4846]: I1005 07:22:43.662220 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b005073c-5379-49a7-ade6-7b9611f3812e-catalog-content\") pod \"certified-operators-8s2mz\" (UID: \"b005073c-5379-49a7-ade6-7b9611f3812e\") " pod="openshift-marketplace/certified-operators-8s2mz" Oct 05 07:22:43 crc kubenswrapper[4846]: I1005 07:22:43.662278 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b005073c-5379-49a7-ade6-7b9611f3812e-utilities\") pod \"certified-operators-8s2mz\" (UID: \"b005073c-5379-49a7-ade6-7b9611f3812e\") " pod="openshift-marketplace/certified-operators-8s2mz" Oct 05 07:22:43 crc kubenswrapper[4846]: I1005 07:22:43.691609 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5ghq5\" (UniqueName: \"kubernetes.io/projected/b005073c-5379-49a7-ade6-7b9611f3812e-kube-api-access-5ghq5\") pod \"certified-operators-8s2mz\" (UID: \"b005073c-5379-49a7-ade6-7b9611f3812e\") " pod="openshift-marketplace/certified-operators-8s2mz" Oct 05 07:22:43 crc kubenswrapper[4846]: I1005 07:22:43.772824 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8s2mz" Oct 05 07:22:44 crc kubenswrapper[4846]: I1005 07:22:44.295305 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-8s2mz"] Oct 05 07:22:45 crc kubenswrapper[4846]: I1005 07:22:45.172743 4846 generic.go:334] "Generic (PLEG): container finished" podID="b005073c-5379-49a7-ade6-7b9611f3812e" containerID="5d01dd021030098ec712093c48ed1b9348df36254e032e9ea757f9c025108df5" exitCode=0 Oct 05 07:22:45 crc kubenswrapper[4846]: I1005 07:22:45.172891 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8s2mz" event={"ID":"b005073c-5379-49a7-ade6-7b9611f3812e","Type":"ContainerDied","Data":"5d01dd021030098ec712093c48ed1b9348df36254e032e9ea757f9c025108df5"} Oct 05 07:22:45 crc kubenswrapper[4846]: I1005 07:22:45.173082 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8s2mz" event={"ID":"b005073c-5379-49a7-ade6-7b9611f3812e","Type":"ContainerStarted","Data":"f4bedce33747e0ae1dc1ec584855ceb7b24fd2354f27dc077c6a76ecf42500a7"} Oct 05 07:22:45 crc kubenswrapper[4846]: I1005 07:22:45.190557 4846 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 05 07:22:46 crc kubenswrapper[4846]: I1005 07:22:46.184114 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8s2mz" event={"ID":"b005073c-5379-49a7-ade6-7b9611f3812e","Type":"ContainerStarted","Data":"970c880171cc94085f786459b583edf4e9a111095a6f28247fb70d9d8486f2f1"} Oct 05 07:22:47 crc kubenswrapper[4846]: I1005 07:22:47.195028 4846 generic.go:334] "Generic (PLEG): container finished" podID="b005073c-5379-49a7-ade6-7b9611f3812e" containerID="970c880171cc94085f786459b583edf4e9a111095a6f28247fb70d9d8486f2f1" exitCode=0 Oct 05 07:22:47 crc kubenswrapper[4846]: I1005 07:22:47.195111 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8s2mz" event={"ID":"b005073c-5379-49a7-ade6-7b9611f3812e","Type":"ContainerDied","Data":"970c880171cc94085f786459b583edf4e9a111095a6f28247fb70d9d8486f2f1"} Oct 05 07:22:48 crc kubenswrapper[4846]: I1005 07:22:48.205129 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8s2mz" event={"ID":"b005073c-5379-49a7-ade6-7b9611f3812e","Type":"ContainerStarted","Data":"2a36f5d40a7acd7725c2706b5192f8b28cedd98869c6dbd4627bc2991d291715"} Oct 05 07:22:48 crc kubenswrapper[4846]: I1005 07:22:48.234506 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-8s2mz" podStartSLOduration=2.823829992 podStartE2EDuration="5.23448777s" podCreationTimestamp="2025-10-05 07:22:43 +0000 UTC" firstStartedPulling="2025-10-05 07:22:45.190146381 +0000 UTC m=+2087.430999166" lastFinishedPulling="2025-10-05 07:22:47.600804159 +0000 UTC m=+2089.841656944" observedRunningTime="2025-10-05 07:22:48.22664909 +0000 UTC m=+2090.467501895" watchObservedRunningTime="2025-10-05 07:22:48.23448777 +0000 UTC m=+2090.475340555" Oct 05 07:22:53 crc kubenswrapper[4846]: I1005 07:22:53.774398 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-8s2mz" Oct 05 07:22:53 crc kubenswrapper[4846]: I1005 07:22:53.774882 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-8s2mz" Oct 05 07:22:53 crc kubenswrapper[4846]: I1005 07:22:53.853630 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-8s2mz" Oct 05 07:22:54 crc kubenswrapper[4846]: I1005 07:22:54.319409 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-8s2mz" Oct 05 07:22:54 crc kubenswrapper[4846]: I1005 07:22:54.364803 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-8s2mz"] Oct 05 07:22:56 crc kubenswrapper[4846]: I1005 07:22:56.267870 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-8s2mz" podUID="b005073c-5379-49a7-ade6-7b9611f3812e" containerName="registry-server" containerID="cri-o://2a36f5d40a7acd7725c2706b5192f8b28cedd98869c6dbd4627bc2991d291715" gracePeriod=2 Oct 05 07:22:56 crc kubenswrapper[4846]: I1005 07:22:56.738907 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8s2mz" Oct 05 07:22:56 crc kubenswrapper[4846]: I1005 07:22:56.868677 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b005073c-5379-49a7-ade6-7b9611f3812e-catalog-content\") pod \"b005073c-5379-49a7-ade6-7b9611f3812e\" (UID: \"b005073c-5379-49a7-ade6-7b9611f3812e\") " Oct 05 07:22:56 crc kubenswrapper[4846]: I1005 07:22:56.868828 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b005073c-5379-49a7-ade6-7b9611f3812e-utilities\") pod \"b005073c-5379-49a7-ade6-7b9611f3812e\" (UID: \"b005073c-5379-49a7-ade6-7b9611f3812e\") " Oct 05 07:22:56 crc kubenswrapper[4846]: I1005 07:22:56.868905 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5ghq5\" (UniqueName: \"kubernetes.io/projected/b005073c-5379-49a7-ade6-7b9611f3812e-kube-api-access-5ghq5\") pod \"b005073c-5379-49a7-ade6-7b9611f3812e\" (UID: \"b005073c-5379-49a7-ade6-7b9611f3812e\") " Oct 05 07:22:56 crc kubenswrapper[4846]: I1005 07:22:56.869656 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b005073c-5379-49a7-ade6-7b9611f3812e-utilities" (OuterVolumeSpecName: "utilities") pod "b005073c-5379-49a7-ade6-7b9611f3812e" (UID: "b005073c-5379-49a7-ade6-7b9611f3812e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 07:22:56 crc kubenswrapper[4846]: I1005 07:22:56.870716 4846 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b005073c-5379-49a7-ade6-7b9611f3812e-utilities\") on node \"crc\" DevicePath \"\"" Oct 05 07:22:56 crc kubenswrapper[4846]: I1005 07:22:56.874658 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b005073c-5379-49a7-ade6-7b9611f3812e-kube-api-access-5ghq5" (OuterVolumeSpecName: "kube-api-access-5ghq5") pod "b005073c-5379-49a7-ade6-7b9611f3812e" (UID: "b005073c-5379-49a7-ade6-7b9611f3812e"). InnerVolumeSpecName "kube-api-access-5ghq5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:22:56 crc kubenswrapper[4846]: I1005 07:22:56.932984 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b005073c-5379-49a7-ade6-7b9611f3812e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b005073c-5379-49a7-ade6-7b9611f3812e" (UID: "b005073c-5379-49a7-ade6-7b9611f3812e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 07:22:56 crc kubenswrapper[4846]: I1005 07:22:56.973152 4846 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b005073c-5379-49a7-ade6-7b9611f3812e-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 05 07:22:56 crc kubenswrapper[4846]: I1005 07:22:56.973250 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5ghq5\" (UniqueName: \"kubernetes.io/projected/b005073c-5379-49a7-ade6-7b9611f3812e-kube-api-access-5ghq5\") on node \"crc\" DevicePath \"\"" Oct 05 07:22:57 crc kubenswrapper[4846]: I1005 07:22:57.285256 4846 generic.go:334] "Generic (PLEG): container finished" podID="b005073c-5379-49a7-ade6-7b9611f3812e" containerID="2a36f5d40a7acd7725c2706b5192f8b28cedd98869c6dbd4627bc2991d291715" exitCode=0 Oct 05 07:22:57 crc kubenswrapper[4846]: I1005 07:22:57.285344 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8s2mz" event={"ID":"b005073c-5379-49a7-ade6-7b9611f3812e","Type":"ContainerDied","Data":"2a36f5d40a7acd7725c2706b5192f8b28cedd98869c6dbd4627bc2991d291715"} Oct 05 07:22:57 crc kubenswrapper[4846]: I1005 07:22:57.285400 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8s2mz" event={"ID":"b005073c-5379-49a7-ade6-7b9611f3812e","Type":"ContainerDied","Data":"f4bedce33747e0ae1dc1ec584855ceb7b24fd2354f27dc077c6a76ecf42500a7"} Oct 05 07:22:57 crc kubenswrapper[4846]: I1005 07:22:57.285427 4846 scope.go:117] "RemoveContainer" containerID="2a36f5d40a7acd7725c2706b5192f8b28cedd98869c6dbd4627bc2991d291715" Oct 05 07:22:57 crc kubenswrapper[4846]: I1005 07:22:57.286368 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8s2mz" Oct 05 07:22:57 crc kubenswrapper[4846]: I1005 07:22:57.310662 4846 scope.go:117] "RemoveContainer" containerID="970c880171cc94085f786459b583edf4e9a111095a6f28247fb70d9d8486f2f1" Oct 05 07:22:57 crc kubenswrapper[4846]: I1005 07:22:57.354516 4846 scope.go:117] "RemoveContainer" containerID="5d01dd021030098ec712093c48ed1b9348df36254e032e9ea757f9c025108df5" Oct 05 07:22:57 crc kubenswrapper[4846]: I1005 07:22:57.355599 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-8s2mz"] Oct 05 07:22:57 crc kubenswrapper[4846]: I1005 07:22:57.375102 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-8s2mz"] Oct 05 07:22:57 crc kubenswrapper[4846]: I1005 07:22:57.382625 4846 scope.go:117] "RemoveContainer" containerID="2a36f5d40a7acd7725c2706b5192f8b28cedd98869c6dbd4627bc2991d291715" Oct 05 07:22:57 crc kubenswrapper[4846]: E1005 07:22:57.383065 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2a36f5d40a7acd7725c2706b5192f8b28cedd98869c6dbd4627bc2991d291715\": container with ID starting with 2a36f5d40a7acd7725c2706b5192f8b28cedd98869c6dbd4627bc2991d291715 not found: ID does not exist" containerID="2a36f5d40a7acd7725c2706b5192f8b28cedd98869c6dbd4627bc2991d291715" Oct 05 07:22:57 crc kubenswrapper[4846]: I1005 07:22:57.383115 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2a36f5d40a7acd7725c2706b5192f8b28cedd98869c6dbd4627bc2991d291715"} err="failed to get container status \"2a36f5d40a7acd7725c2706b5192f8b28cedd98869c6dbd4627bc2991d291715\": rpc error: code = NotFound desc = could not find container \"2a36f5d40a7acd7725c2706b5192f8b28cedd98869c6dbd4627bc2991d291715\": container with ID starting with 2a36f5d40a7acd7725c2706b5192f8b28cedd98869c6dbd4627bc2991d291715 not found: ID does not exist" Oct 05 07:22:57 crc kubenswrapper[4846]: I1005 07:22:57.383149 4846 scope.go:117] "RemoveContainer" containerID="970c880171cc94085f786459b583edf4e9a111095a6f28247fb70d9d8486f2f1" Oct 05 07:22:57 crc kubenswrapper[4846]: E1005 07:22:57.383575 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"970c880171cc94085f786459b583edf4e9a111095a6f28247fb70d9d8486f2f1\": container with ID starting with 970c880171cc94085f786459b583edf4e9a111095a6f28247fb70d9d8486f2f1 not found: ID does not exist" containerID="970c880171cc94085f786459b583edf4e9a111095a6f28247fb70d9d8486f2f1" Oct 05 07:22:57 crc kubenswrapper[4846]: I1005 07:22:57.383605 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"970c880171cc94085f786459b583edf4e9a111095a6f28247fb70d9d8486f2f1"} err="failed to get container status \"970c880171cc94085f786459b583edf4e9a111095a6f28247fb70d9d8486f2f1\": rpc error: code = NotFound desc = could not find container \"970c880171cc94085f786459b583edf4e9a111095a6f28247fb70d9d8486f2f1\": container with ID starting with 970c880171cc94085f786459b583edf4e9a111095a6f28247fb70d9d8486f2f1 not found: ID does not exist" Oct 05 07:22:57 crc kubenswrapper[4846]: I1005 07:22:57.383623 4846 scope.go:117] "RemoveContainer" containerID="5d01dd021030098ec712093c48ed1b9348df36254e032e9ea757f9c025108df5" Oct 05 07:22:57 crc kubenswrapper[4846]: E1005 07:22:57.383979 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5d01dd021030098ec712093c48ed1b9348df36254e032e9ea757f9c025108df5\": container with ID starting with 5d01dd021030098ec712093c48ed1b9348df36254e032e9ea757f9c025108df5 not found: ID does not exist" containerID="5d01dd021030098ec712093c48ed1b9348df36254e032e9ea757f9c025108df5" Oct 05 07:22:57 crc kubenswrapper[4846]: I1005 07:22:57.384017 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5d01dd021030098ec712093c48ed1b9348df36254e032e9ea757f9c025108df5"} err="failed to get container status \"5d01dd021030098ec712093c48ed1b9348df36254e032e9ea757f9c025108df5\": rpc error: code = NotFound desc = could not find container \"5d01dd021030098ec712093c48ed1b9348df36254e032e9ea757f9c025108df5\": container with ID starting with 5d01dd021030098ec712093c48ed1b9348df36254e032e9ea757f9c025108df5 not found: ID does not exist" Oct 05 07:22:58 crc kubenswrapper[4846]: I1005 07:22:58.511429 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b005073c-5379-49a7-ade6-7b9611f3812e" path="/var/lib/kubelet/pods/b005073c-5379-49a7-ade6-7b9611f3812e/volumes" Oct 05 07:23:25 crc kubenswrapper[4846]: I1005 07:23:25.629532 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-kb5vt"] Oct 05 07:23:25 crc kubenswrapper[4846]: E1005 07:23:25.632610 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b005073c-5379-49a7-ade6-7b9611f3812e" containerName="registry-server" Oct 05 07:23:25 crc kubenswrapper[4846]: I1005 07:23:25.632645 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="b005073c-5379-49a7-ade6-7b9611f3812e" containerName="registry-server" Oct 05 07:23:25 crc kubenswrapper[4846]: E1005 07:23:25.632692 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b005073c-5379-49a7-ade6-7b9611f3812e" containerName="extract-content" Oct 05 07:23:25 crc kubenswrapper[4846]: I1005 07:23:25.632704 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="b005073c-5379-49a7-ade6-7b9611f3812e" containerName="extract-content" Oct 05 07:23:25 crc kubenswrapper[4846]: E1005 07:23:25.632750 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b005073c-5379-49a7-ade6-7b9611f3812e" containerName="extract-utilities" Oct 05 07:23:25 crc kubenswrapper[4846]: I1005 07:23:25.632763 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="b005073c-5379-49a7-ade6-7b9611f3812e" containerName="extract-utilities" Oct 05 07:23:25 crc kubenswrapper[4846]: I1005 07:23:25.633065 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="b005073c-5379-49a7-ade6-7b9611f3812e" containerName="registry-server" Oct 05 07:23:25 crc kubenswrapper[4846]: I1005 07:23:25.634829 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-kb5vt" Oct 05 07:23:25 crc kubenswrapper[4846]: I1005 07:23:25.645246 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-kb5vt"] Oct 05 07:23:25 crc kubenswrapper[4846]: I1005 07:23:25.789878 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tk77b\" (UniqueName: \"kubernetes.io/projected/ae8772b1-dda6-4d4b-83de-941c0e90c6dd-kube-api-access-tk77b\") pod \"redhat-marketplace-kb5vt\" (UID: \"ae8772b1-dda6-4d4b-83de-941c0e90c6dd\") " pod="openshift-marketplace/redhat-marketplace-kb5vt" Oct 05 07:23:25 crc kubenswrapper[4846]: I1005 07:23:25.790271 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ae8772b1-dda6-4d4b-83de-941c0e90c6dd-utilities\") pod \"redhat-marketplace-kb5vt\" (UID: \"ae8772b1-dda6-4d4b-83de-941c0e90c6dd\") " pod="openshift-marketplace/redhat-marketplace-kb5vt" Oct 05 07:23:25 crc kubenswrapper[4846]: I1005 07:23:25.790457 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ae8772b1-dda6-4d4b-83de-941c0e90c6dd-catalog-content\") pod \"redhat-marketplace-kb5vt\" (UID: \"ae8772b1-dda6-4d4b-83de-941c0e90c6dd\") " pod="openshift-marketplace/redhat-marketplace-kb5vt" Oct 05 07:23:25 crc kubenswrapper[4846]: I1005 07:23:25.891278 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ae8772b1-dda6-4d4b-83de-941c0e90c6dd-catalog-content\") pod \"redhat-marketplace-kb5vt\" (UID: \"ae8772b1-dda6-4d4b-83de-941c0e90c6dd\") " pod="openshift-marketplace/redhat-marketplace-kb5vt" Oct 05 07:23:25 crc kubenswrapper[4846]: I1005 07:23:25.891755 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tk77b\" (UniqueName: \"kubernetes.io/projected/ae8772b1-dda6-4d4b-83de-941c0e90c6dd-kube-api-access-tk77b\") pod \"redhat-marketplace-kb5vt\" (UID: \"ae8772b1-dda6-4d4b-83de-941c0e90c6dd\") " pod="openshift-marketplace/redhat-marketplace-kb5vt" Oct 05 07:23:25 crc kubenswrapper[4846]: I1005 07:23:25.891884 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ae8772b1-dda6-4d4b-83de-941c0e90c6dd-catalog-content\") pod \"redhat-marketplace-kb5vt\" (UID: \"ae8772b1-dda6-4d4b-83de-941c0e90c6dd\") " pod="openshift-marketplace/redhat-marketplace-kb5vt" Oct 05 07:23:25 crc kubenswrapper[4846]: I1005 07:23:25.892045 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ae8772b1-dda6-4d4b-83de-941c0e90c6dd-utilities\") pod \"redhat-marketplace-kb5vt\" (UID: \"ae8772b1-dda6-4d4b-83de-941c0e90c6dd\") " pod="openshift-marketplace/redhat-marketplace-kb5vt" Oct 05 07:23:25 crc kubenswrapper[4846]: I1005 07:23:25.892423 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ae8772b1-dda6-4d4b-83de-941c0e90c6dd-utilities\") pod \"redhat-marketplace-kb5vt\" (UID: \"ae8772b1-dda6-4d4b-83de-941c0e90c6dd\") " pod="openshift-marketplace/redhat-marketplace-kb5vt" Oct 05 07:23:25 crc kubenswrapper[4846]: I1005 07:23:25.920212 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tk77b\" (UniqueName: \"kubernetes.io/projected/ae8772b1-dda6-4d4b-83de-941c0e90c6dd-kube-api-access-tk77b\") pod \"redhat-marketplace-kb5vt\" (UID: \"ae8772b1-dda6-4d4b-83de-941c0e90c6dd\") " pod="openshift-marketplace/redhat-marketplace-kb5vt" Oct 05 07:23:25 crc kubenswrapper[4846]: I1005 07:23:25.968171 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-kb5vt" Oct 05 07:23:26 crc kubenswrapper[4846]: I1005 07:23:26.224834 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-kb5vt"] Oct 05 07:23:26 crc kubenswrapper[4846]: I1005 07:23:26.567025 4846 generic.go:334] "Generic (PLEG): container finished" podID="ae8772b1-dda6-4d4b-83de-941c0e90c6dd" containerID="3d368cebdbc233fb02878efada0bf1ac0bd6543dad2b58f0c4c493583532f144" exitCode=0 Oct 05 07:23:26 crc kubenswrapper[4846]: I1005 07:23:26.567515 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kb5vt" event={"ID":"ae8772b1-dda6-4d4b-83de-941c0e90c6dd","Type":"ContainerDied","Data":"3d368cebdbc233fb02878efada0bf1ac0bd6543dad2b58f0c4c493583532f144"} Oct 05 07:23:26 crc kubenswrapper[4846]: I1005 07:23:26.567570 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kb5vt" event={"ID":"ae8772b1-dda6-4d4b-83de-941c0e90c6dd","Type":"ContainerStarted","Data":"e9edcbeaaa48d34351cb6ad7e40110eed48f8c64f2c1669d44928fc80bd49145"} Oct 05 07:23:27 crc kubenswrapper[4846]: I1005 07:23:27.576988 4846 generic.go:334] "Generic (PLEG): container finished" podID="ae8772b1-dda6-4d4b-83de-941c0e90c6dd" containerID="8f5b1c2023c8038706688e4c94f97ff471dc3d069e98732684b7bdf539577429" exitCode=0 Oct 05 07:23:27 crc kubenswrapper[4846]: I1005 07:23:27.577025 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kb5vt" event={"ID":"ae8772b1-dda6-4d4b-83de-941c0e90c6dd","Type":"ContainerDied","Data":"8f5b1c2023c8038706688e4c94f97ff471dc3d069e98732684b7bdf539577429"} Oct 05 07:23:28 crc kubenswrapper[4846]: I1005 07:23:28.588743 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kb5vt" event={"ID":"ae8772b1-dda6-4d4b-83de-941c0e90c6dd","Type":"ContainerStarted","Data":"1182d18c77b82a88b6aacf5abd2009029be99c061553d376ec93c634206ebc6c"} Oct 05 07:23:28 crc kubenswrapper[4846]: I1005 07:23:28.609521 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-kb5vt" podStartSLOduration=2.1296665949999998 podStartE2EDuration="3.609505686s" podCreationTimestamp="2025-10-05 07:23:25 +0000 UTC" firstStartedPulling="2025-10-05 07:23:26.570248128 +0000 UTC m=+2128.811100933" lastFinishedPulling="2025-10-05 07:23:28.050087219 +0000 UTC m=+2130.290940024" observedRunningTime="2025-10-05 07:23:28.605060277 +0000 UTC m=+2130.845913052" watchObservedRunningTime="2025-10-05 07:23:28.609505686 +0000 UTC m=+2130.850358461" Oct 05 07:23:35 crc kubenswrapper[4846]: I1005 07:23:35.969105 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-kb5vt" Oct 05 07:23:35 crc kubenswrapper[4846]: I1005 07:23:35.970997 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-kb5vt" Oct 05 07:23:36 crc kubenswrapper[4846]: I1005 07:23:36.048384 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-kb5vt" Oct 05 07:23:36 crc kubenswrapper[4846]: I1005 07:23:36.758595 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-kb5vt" Oct 05 07:23:36 crc kubenswrapper[4846]: I1005 07:23:36.815815 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-kb5vt"] Oct 05 07:23:38 crc kubenswrapper[4846]: I1005 07:23:38.709763 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-kb5vt" podUID="ae8772b1-dda6-4d4b-83de-941c0e90c6dd" containerName="registry-server" containerID="cri-o://1182d18c77b82a88b6aacf5abd2009029be99c061553d376ec93c634206ebc6c" gracePeriod=2 Oct 05 07:23:39 crc kubenswrapper[4846]: I1005 07:23:39.205750 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-kb5vt" Oct 05 07:23:39 crc kubenswrapper[4846]: I1005 07:23:39.307260 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ae8772b1-dda6-4d4b-83de-941c0e90c6dd-catalog-content\") pod \"ae8772b1-dda6-4d4b-83de-941c0e90c6dd\" (UID: \"ae8772b1-dda6-4d4b-83de-941c0e90c6dd\") " Oct 05 07:23:39 crc kubenswrapper[4846]: I1005 07:23:39.307567 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk77b\" (UniqueName: \"kubernetes.io/projected/ae8772b1-dda6-4d4b-83de-941c0e90c6dd-kube-api-access-tk77b\") pod \"ae8772b1-dda6-4d4b-83de-941c0e90c6dd\" (UID: \"ae8772b1-dda6-4d4b-83de-941c0e90c6dd\") " Oct 05 07:23:39 crc kubenswrapper[4846]: I1005 07:23:39.307651 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ae8772b1-dda6-4d4b-83de-941c0e90c6dd-utilities\") pod \"ae8772b1-dda6-4d4b-83de-941c0e90c6dd\" (UID: \"ae8772b1-dda6-4d4b-83de-941c0e90c6dd\") " Oct 05 07:23:39 crc kubenswrapper[4846]: I1005 07:23:39.308753 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ae8772b1-dda6-4d4b-83de-941c0e90c6dd-utilities" (OuterVolumeSpecName: "utilities") pod "ae8772b1-dda6-4d4b-83de-941c0e90c6dd" (UID: "ae8772b1-dda6-4d4b-83de-941c0e90c6dd"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 07:23:39 crc kubenswrapper[4846]: I1005 07:23:39.314349 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ae8772b1-dda6-4d4b-83de-941c0e90c6dd-kube-api-access-tk77b" (OuterVolumeSpecName: "kube-api-access-tk77b") pod "ae8772b1-dda6-4d4b-83de-941c0e90c6dd" (UID: "ae8772b1-dda6-4d4b-83de-941c0e90c6dd"). InnerVolumeSpecName "kube-api-access-tk77b". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:23:39 crc kubenswrapper[4846]: I1005 07:23:39.319454 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ae8772b1-dda6-4d4b-83de-941c0e90c6dd-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ae8772b1-dda6-4d4b-83de-941c0e90c6dd" (UID: "ae8772b1-dda6-4d4b-83de-941c0e90c6dd"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 07:23:39 crc kubenswrapper[4846]: I1005 07:23:39.410114 4846 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ae8772b1-dda6-4d4b-83de-941c0e90c6dd-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 05 07:23:39 crc kubenswrapper[4846]: I1005 07:23:39.410173 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk77b\" (UniqueName: \"kubernetes.io/projected/ae8772b1-dda6-4d4b-83de-941c0e90c6dd-kube-api-access-tk77b\") on node \"crc\" DevicePath \"\"" Oct 05 07:23:39 crc kubenswrapper[4846]: I1005 07:23:39.410302 4846 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ae8772b1-dda6-4d4b-83de-941c0e90c6dd-utilities\") on node \"crc\" DevicePath \"\"" Oct 05 07:23:39 crc kubenswrapper[4846]: I1005 07:23:39.738485 4846 generic.go:334] "Generic (PLEG): container finished" podID="ae8772b1-dda6-4d4b-83de-941c0e90c6dd" containerID="1182d18c77b82a88b6aacf5abd2009029be99c061553d376ec93c634206ebc6c" exitCode=0 Oct 05 07:23:39 crc kubenswrapper[4846]: I1005 07:23:39.738546 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kb5vt" event={"ID":"ae8772b1-dda6-4d4b-83de-941c0e90c6dd","Type":"ContainerDied","Data":"1182d18c77b82a88b6aacf5abd2009029be99c061553d376ec93c634206ebc6c"} Oct 05 07:23:39 crc kubenswrapper[4846]: I1005 07:23:39.738586 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kb5vt" event={"ID":"ae8772b1-dda6-4d4b-83de-941c0e90c6dd","Type":"ContainerDied","Data":"e9edcbeaaa48d34351cb6ad7e40110eed48f8c64f2c1669d44928fc80bd49145"} Oct 05 07:23:39 crc kubenswrapper[4846]: I1005 07:23:39.738615 4846 scope.go:117] "RemoveContainer" containerID="1182d18c77b82a88b6aacf5abd2009029be99c061553d376ec93c634206ebc6c" Oct 05 07:23:39 crc kubenswrapper[4846]: I1005 07:23:39.738785 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-kb5vt" Oct 05 07:23:39 crc kubenswrapper[4846]: I1005 07:23:39.779711 4846 scope.go:117] "RemoveContainer" containerID="8f5b1c2023c8038706688e4c94f97ff471dc3d069e98732684b7bdf539577429" Oct 05 07:23:39 crc kubenswrapper[4846]: I1005 07:23:39.797353 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-kb5vt"] Oct 05 07:23:39 crc kubenswrapper[4846]: I1005 07:23:39.807409 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-kb5vt"] Oct 05 07:23:39 crc kubenswrapper[4846]: I1005 07:23:39.810951 4846 scope.go:117] "RemoveContainer" containerID="3d368cebdbc233fb02878efada0bf1ac0bd6543dad2b58f0c4c493583532f144" Oct 05 07:23:39 crc kubenswrapper[4846]: I1005 07:23:39.840536 4846 scope.go:117] "RemoveContainer" containerID="1182d18c77b82a88b6aacf5abd2009029be99c061553d376ec93c634206ebc6c" Oct 05 07:23:39 crc kubenswrapper[4846]: E1005 07:23:39.841234 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1182d18c77b82a88b6aacf5abd2009029be99c061553d376ec93c634206ebc6c\": container with ID starting with 1182d18c77b82a88b6aacf5abd2009029be99c061553d376ec93c634206ebc6c not found: ID does not exist" containerID="1182d18c77b82a88b6aacf5abd2009029be99c061553d376ec93c634206ebc6c" Oct 05 07:23:39 crc kubenswrapper[4846]: I1005 07:23:39.841303 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1182d18c77b82a88b6aacf5abd2009029be99c061553d376ec93c634206ebc6c"} err="failed to get container status \"1182d18c77b82a88b6aacf5abd2009029be99c061553d376ec93c634206ebc6c\": rpc error: code = NotFound desc = could not find container \"1182d18c77b82a88b6aacf5abd2009029be99c061553d376ec93c634206ebc6c\": container with ID starting with 1182d18c77b82a88b6aacf5abd2009029be99c061553d376ec93c634206ebc6c not found: ID does not exist" Oct 05 07:23:39 crc kubenswrapper[4846]: I1005 07:23:39.841354 4846 scope.go:117] "RemoveContainer" containerID="8f5b1c2023c8038706688e4c94f97ff471dc3d069e98732684b7bdf539577429" Oct 05 07:23:39 crc kubenswrapper[4846]: E1005 07:23:39.842103 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8f5b1c2023c8038706688e4c94f97ff471dc3d069e98732684b7bdf539577429\": container with ID starting with 8f5b1c2023c8038706688e4c94f97ff471dc3d069e98732684b7bdf539577429 not found: ID does not exist" containerID="8f5b1c2023c8038706688e4c94f97ff471dc3d069e98732684b7bdf539577429" Oct 05 07:23:39 crc kubenswrapper[4846]: I1005 07:23:39.842166 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8f5b1c2023c8038706688e4c94f97ff471dc3d069e98732684b7bdf539577429"} err="failed to get container status \"8f5b1c2023c8038706688e4c94f97ff471dc3d069e98732684b7bdf539577429\": rpc error: code = NotFound desc = could not find container \"8f5b1c2023c8038706688e4c94f97ff471dc3d069e98732684b7bdf539577429\": container with ID starting with 8f5b1c2023c8038706688e4c94f97ff471dc3d069e98732684b7bdf539577429 not found: ID does not exist" Oct 05 07:23:39 crc kubenswrapper[4846]: I1005 07:23:39.842233 4846 scope.go:117] "RemoveContainer" containerID="3d368cebdbc233fb02878efada0bf1ac0bd6543dad2b58f0c4c493583532f144" Oct 05 07:23:39 crc kubenswrapper[4846]: E1005 07:23:39.842664 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3d368cebdbc233fb02878efada0bf1ac0bd6543dad2b58f0c4c493583532f144\": container with ID starting with 3d368cebdbc233fb02878efada0bf1ac0bd6543dad2b58f0c4c493583532f144 not found: ID does not exist" containerID="3d368cebdbc233fb02878efada0bf1ac0bd6543dad2b58f0c4c493583532f144" Oct 05 07:23:39 crc kubenswrapper[4846]: I1005 07:23:39.842708 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3d368cebdbc233fb02878efada0bf1ac0bd6543dad2b58f0c4c493583532f144"} err="failed to get container status \"3d368cebdbc233fb02878efada0bf1ac0bd6543dad2b58f0c4c493583532f144\": rpc error: code = NotFound desc = could not find container \"3d368cebdbc233fb02878efada0bf1ac0bd6543dad2b58f0c4c493583532f144\": container with ID starting with 3d368cebdbc233fb02878efada0bf1ac0bd6543dad2b58f0c4c493583532f144 not found: ID does not exist" Oct 05 07:23:40 crc kubenswrapper[4846]: I1005 07:23:40.517067 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ae8772b1-dda6-4d4b-83de-941c0e90c6dd" path="/var/lib/kubelet/pods/ae8772b1-dda6-4d4b-83de-941c0e90c6dd/volumes" Oct 05 07:23:53 crc kubenswrapper[4846]: I1005 07:23:53.325004 4846 patch_prober.go:28] interesting pod/machine-config-daemon-fscvf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 07:23:53 crc kubenswrapper[4846]: I1005 07:23:53.325643 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 07:24:23 crc kubenswrapper[4846]: I1005 07:24:23.325079 4846 patch_prober.go:28] interesting pod/machine-config-daemon-fscvf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 07:24:23 crc kubenswrapper[4846]: I1005 07:24:23.325829 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 07:24:24 crc kubenswrapper[4846]: I1005 07:24:24.736417 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-clnvm"] Oct 05 07:24:24 crc kubenswrapper[4846]: E1005 07:24:24.737413 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ae8772b1-dda6-4d4b-83de-941c0e90c6dd" containerName="registry-server" Oct 05 07:24:24 crc kubenswrapper[4846]: I1005 07:24:24.737443 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="ae8772b1-dda6-4d4b-83de-941c0e90c6dd" containerName="registry-server" Oct 05 07:24:24 crc kubenswrapper[4846]: E1005 07:24:24.737506 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ae8772b1-dda6-4d4b-83de-941c0e90c6dd" containerName="extract-content" Oct 05 07:24:24 crc kubenswrapper[4846]: I1005 07:24:24.737523 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="ae8772b1-dda6-4d4b-83de-941c0e90c6dd" containerName="extract-content" Oct 05 07:24:24 crc kubenswrapper[4846]: E1005 07:24:24.737545 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ae8772b1-dda6-4d4b-83de-941c0e90c6dd" containerName="extract-utilities" Oct 05 07:24:24 crc kubenswrapper[4846]: I1005 07:24:24.737562 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="ae8772b1-dda6-4d4b-83de-941c0e90c6dd" containerName="extract-utilities" Oct 05 07:24:24 crc kubenswrapper[4846]: I1005 07:24:24.737890 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="ae8772b1-dda6-4d4b-83de-941c0e90c6dd" containerName="registry-server" Oct 05 07:24:24 crc kubenswrapper[4846]: I1005 07:24:24.740366 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-clnvm" Oct 05 07:24:24 crc kubenswrapper[4846]: I1005 07:24:24.748451 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-clnvm"] Oct 05 07:24:24 crc kubenswrapper[4846]: I1005 07:24:24.852078 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4dh6f\" (UniqueName: \"kubernetes.io/projected/ed7834ca-b09f-4117-9d4a-75a401a43428-kube-api-access-4dh6f\") pod \"community-operators-clnvm\" (UID: \"ed7834ca-b09f-4117-9d4a-75a401a43428\") " pod="openshift-marketplace/community-operators-clnvm" Oct 05 07:24:24 crc kubenswrapper[4846]: I1005 07:24:24.852129 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ed7834ca-b09f-4117-9d4a-75a401a43428-catalog-content\") pod \"community-operators-clnvm\" (UID: \"ed7834ca-b09f-4117-9d4a-75a401a43428\") " pod="openshift-marketplace/community-operators-clnvm" Oct 05 07:24:24 crc kubenswrapper[4846]: I1005 07:24:24.852199 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ed7834ca-b09f-4117-9d4a-75a401a43428-utilities\") pod \"community-operators-clnvm\" (UID: \"ed7834ca-b09f-4117-9d4a-75a401a43428\") " pod="openshift-marketplace/community-operators-clnvm" Oct 05 07:24:24 crc kubenswrapper[4846]: I1005 07:24:24.953247 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4dh6f\" (UniqueName: \"kubernetes.io/projected/ed7834ca-b09f-4117-9d4a-75a401a43428-kube-api-access-4dh6f\") pod \"community-operators-clnvm\" (UID: \"ed7834ca-b09f-4117-9d4a-75a401a43428\") " pod="openshift-marketplace/community-operators-clnvm" Oct 05 07:24:24 crc kubenswrapper[4846]: I1005 07:24:24.953323 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ed7834ca-b09f-4117-9d4a-75a401a43428-catalog-content\") pod \"community-operators-clnvm\" (UID: \"ed7834ca-b09f-4117-9d4a-75a401a43428\") " pod="openshift-marketplace/community-operators-clnvm" Oct 05 07:24:24 crc kubenswrapper[4846]: I1005 07:24:24.953401 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ed7834ca-b09f-4117-9d4a-75a401a43428-utilities\") pod \"community-operators-clnvm\" (UID: \"ed7834ca-b09f-4117-9d4a-75a401a43428\") " pod="openshift-marketplace/community-operators-clnvm" Oct 05 07:24:24 crc kubenswrapper[4846]: I1005 07:24:24.953997 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ed7834ca-b09f-4117-9d4a-75a401a43428-catalog-content\") pod \"community-operators-clnvm\" (UID: \"ed7834ca-b09f-4117-9d4a-75a401a43428\") " pod="openshift-marketplace/community-operators-clnvm" Oct 05 07:24:24 crc kubenswrapper[4846]: I1005 07:24:24.954078 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ed7834ca-b09f-4117-9d4a-75a401a43428-utilities\") pod \"community-operators-clnvm\" (UID: \"ed7834ca-b09f-4117-9d4a-75a401a43428\") " pod="openshift-marketplace/community-operators-clnvm" Oct 05 07:24:24 crc kubenswrapper[4846]: I1005 07:24:24.987286 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4dh6f\" (UniqueName: \"kubernetes.io/projected/ed7834ca-b09f-4117-9d4a-75a401a43428-kube-api-access-4dh6f\") pod \"community-operators-clnvm\" (UID: \"ed7834ca-b09f-4117-9d4a-75a401a43428\") " pod="openshift-marketplace/community-operators-clnvm" Oct 05 07:24:25 crc kubenswrapper[4846]: I1005 07:24:25.080048 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-clnvm" Oct 05 07:24:25 crc kubenswrapper[4846]: I1005 07:24:25.333549 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-clnvm"] Oct 05 07:24:26 crc kubenswrapper[4846]: I1005 07:24:26.201973 4846 generic.go:334] "Generic (PLEG): container finished" podID="ed7834ca-b09f-4117-9d4a-75a401a43428" containerID="649a41953281b41a777ff15169a825997e791b828bd2cc6700cf6e96e9423dcf" exitCode=0 Oct 05 07:24:26 crc kubenswrapper[4846]: I1005 07:24:26.202132 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-clnvm" event={"ID":"ed7834ca-b09f-4117-9d4a-75a401a43428","Type":"ContainerDied","Data":"649a41953281b41a777ff15169a825997e791b828bd2cc6700cf6e96e9423dcf"} Oct 05 07:24:26 crc kubenswrapper[4846]: I1005 07:24:26.202449 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-clnvm" event={"ID":"ed7834ca-b09f-4117-9d4a-75a401a43428","Type":"ContainerStarted","Data":"82c05b47e24a20e45e284ebe695e6591382d280910173fbcb37fbcf93a48f2b6"} Oct 05 07:24:27 crc kubenswrapper[4846]: I1005 07:24:27.125308 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-rk8kt"] Oct 05 07:24:27 crc kubenswrapper[4846]: I1005 07:24:27.128655 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rk8kt" Oct 05 07:24:27 crc kubenswrapper[4846]: I1005 07:24:27.138526 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-rk8kt"] Oct 05 07:24:27 crc kubenswrapper[4846]: I1005 07:24:27.188274 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/88c88750-5101-49e7-85d7-47ad81bf9fe8-utilities\") pod \"redhat-operators-rk8kt\" (UID: \"88c88750-5101-49e7-85d7-47ad81bf9fe8\") " pod="openshift-marketplace/redhat-operators-rk8kt" Oct 05 07:24:27 crc kubenswrapper[4846]: I1005 07:24:27.188380 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/88c88750-5101-49e7-85d7-47ad81bf9fe8-catalog-content\") pod \"redhat-operators-rk8kt\" (UID: \"88c88750-5101-49e7-85d7-47ad81bf9fe8\") " pod="openshift-marketplace/redhat-operators-rk8kt" Oct 05 07:24:27 crc kubenswrapper[4846]: I1005 07:24:27.188452 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5lx8t\" (UniqueName: \"kubernetes.io/projected/88c88750-5101-49e7-85d7-47ad81bf9fe8-kube-api-access-5lx8t\") pod \"redhat-operators-rk8kt\" (UID: \"88c88750-5101-49e7-85d7-47ad81bf9fe8\") " pod="openshift-marketplace/redhat-operators-rk8kt" Oct 05 07:24:27 crc kubenswrapper[4846]: I1005 07:24:27.213048 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-clnvm" event={"ID":"ed7834ca-b09f-4117-9d4a-75a401a43428","Type":"ContainerStarted","Data":"b4ce0f1458455605c81ebc4adaea69e77e47d67bb88d7c6edd0a0fa2877c8265"} Oct 05 07:24:27 crc kubenswrapper[4846]: I1005 07:24:27.289942 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/88c88750-5101-49e7-85d7-47ad81bf9fe8-utilities\") pod \"redhat-operators-rk8kt\" (UID: \"88c88750-5101-49e7-85d7-47ad81bf9fe8\") " pod="openshift-marketplace/redhat-operators-rk8kt" Oct 05 07:24:27 crc kubenswrapper[4846]: I1005 07:24:27.290030 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/88c88750-5101-49e7-85d7-47ad81bf9fe8-catalog-content\") pod \"redhat-operators-rk8kt\" (UID: \"88c88750-5101-49e7-85d7-47ad81bf9fe8\") " pod="openshift-marketplace/redhat-operators-rk8kt" Oct 05 07:24:27 crc kubenswrapper[4846]: I1005 07:24:27.290073 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5lx8t\" (UniqueName: \"kubernetes.io/projected/88c88750-5101-49e7-85d7-47ad81bf9fe8-kube-api-access-5lx8t\") pod \"redhat-operators-rk8kt\" (UID: \"88c88750-5101-49e7-85d7-47ad81bf9fe8\") " pod="openshift-marketplace/redhat-operators-rk8kt" Oct 05 07:24:27 crc kubenswrapper[4846]: I1005 07:24:27.290550 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/88c88750-5101-49e7-85d7-47ad81bf9fe8-utilities\") pod \"redhat-operators-rk8kt\" (UID: \"88c88750-5101-49e7-85d7-47ad81bf9fe8\") " pod="openshift-marketplace/redhat-operators-rk8kt" Oct 05 07:24:27 crc kubenswrapper[4846]: I1005 07:24:27.290625 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/88c88750-5101-49e7-85d7-47ad81bf9fe8-catalog-content\") pod \"redhat-operators-rk8kt\" (UID: \"88c88750-5101-49e7-85d7-47ad81bf9fe8\") " pod="openshift-marketplace/redhat-operators-rk8kt" Oct 05 07:24:27 crc kubenswrapper[4846]: I1005 07:24:27.312623 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5lx8t\" (UniqueName: \"kubernetes.io/projected/88c88750-5101-49e7-85d7-47ad81bf9fe8-kube-api-access-5lx8t\") pod \"redhat-operators-rk8kt\" (UID: \"88c88750-5101-49e7-85d7-47ad81bf9fe8\") " pod="openshift-marketplace/redhat-operators-rk8kt" Oct 05 07:24:27 crc kubenswrapper[4846]: I1005 07:24:27.454873 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rk8kt" Oct 05 07:24:27 crc kubenswrapper[4846]: I1005 07:24:27.685343 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-rk8kt"] Oct 05 07:24:27 crc kubenswrapper[4846]: W1005 07:24:27.692083 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod88c88750_5101_49e7_85d7_47ad81bf9fe8.slice/crio-5b09246fe409e13dd1da4c1ae05459b1ed30c59291d0a5a8ef5a67108a4671d3 WatchSource:0}: Error finding container 5b09246fe409e13dd1da4c1ae05459b1ed30c59291d0a5a8ef5a67108a4671d3: Status 404 returned error can't find the container with id 5b09246fe409e13dd1da4c1ae05459b1ed30c59291d0a5a8ef5a67108a4671d3 Oct 05 07:24:28 crc kubenswrapper[4846]: I1005 07:24:28.220321 4846 generic.go:334] "Generic (PLEG): container finished" podID="ed7834ca-b09f-4117-9d4a-75a401a43428" containerID="b4ce0f1458455605c81ebc4adaea69e77e47d67bb88d7c6edd0a0fa2877c8265" exitCode=0 Oct 05 07:24:28 crc kubenswrapper[4846]: I1005 07:24:28.220429 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-clnvm" event={"ID":"ed7834ca-b09f-4117-9d4a-75a401a43428","Type":"ContainerDied","Data":"b4ce0f1458455605c81ebc4adaea69e77e47d67bb88d7c6edd0a0fa2877c8265"} Oct 05 07:24:28 crc kubenswrapper[4846]: I1005 07:24:28.223746 4846 generic.go:334] "Generic (PLEG): container finished" podID="88c88750-5101-49e7-85d7-47ad81bf9fe8" containerID="2b843a250df67ce4ac4896ab5d0326cce00f4ce5701a5686f0b2bb109acd8996" exitCode=0 Oct 05 07:24:28 crc kubenswrapper[4846]: I1005 07:24:28.223776 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rk8kt" event={"ID":"88c88750-5101-49e7-85d7-47ad81bf9fe8","Type":"ContainerDied","Data":"2b843a250df67ce4ac4896ab5d0326cce00f4ce5701a5686f0b2bb109acd8996"} Oct 05 07:24:28 crc kubenswrapper[4846]: I1005 07:24:28.223793 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rk8kt" event={"ID":"88c88750-5101-49e7-85d7-47ad81bf9fe8","Type":"ContainerStarted","Data":"5b09246fe409e13dd1da4c1ae05459b1ed30c59291d0a5a8ef5a67108a4671d3"} Oct 05 07:24:29 crc kubenswrapper[4846]: I1005 07:24:29.244194 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-clnvm" event={"ID":"ed7834ca-b09f-4117-9d4a-75a401a43428","Type":"ContainerStarted","Data":"81d1dc3b8df3a57352261ff505bc40a9742fadb86eb663f6bff1752eafc332b1"} Oct 05 07:24:29 crc kubenswrapper[4846]: I1005 07:24:29.247240 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rk8kt" event={"ID":"88c88750-5101-49e7-85d7-47ad81bf9fe8","Type":"ContainerStarted","Data":"8cfe4374cd02643f1ff0eb80affdcfb7b398ccb356c8451565930a5c2f37104f"} Oct 05 07:24:29 crc kubenswrapper[4846]: I1005 07:24:29.273146 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-clnvm" podStartSLOduration=2.723971553 podStartE2EDuration="5.273127068s" podCreationTimestamp="2025-10-05 07:24:24 +0000 UTC" firstStartedPulling="2025-10-05 07:24:26.204078969 +0000 UTC m=+2188.444931784" lastFinishedPulling="2025-10-05 07:24:28.753234484 +0000 UTC m=+2190.994087299" observedRunningTime="2025-10-05 07:24:29.266733847 +0000 UTC m=+2191.507586642" watchObservedRunningTime="2025-10-05 07:24:29.273127068 +0000 UTC m=+2191.513979853" Oct 05 07:24:30 crc kubenswrapper[4846]: I1005 07:24:30.261970 4846 generic.go:334] "Generic (PLEG): container finished" podID="88c88750-5101-49e7-85d7-47ad81bf9fe8" containerID="8cfe4374cd02643f1ff0eb80affdcfb7b398ccb356c8451565930a5c2f37104f" exitCode=0 Oct 05 07:24:30 crc kubenswrapper[4846]: I1005 07:24:30.262105 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rk8kt" event={"ID":"88c88750-5101-49e7-85d7-47ad81bf9fe8","Type":"ContainerDied","Data":"8cfe4374cd02643f1ff0eb80affdcfb7b398ccb356c8451565930a5c2f37104f"} Oct 05 07:24:31 crc kubenswrapper[4846]: I1005 07:24:31.273301 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rk8kt" event={"ID":"88c88750-5101-49e7-85d7-47ad81bf9fe8","Type":"ContainerStarted","Data":"0b708ade2544d24e946164684bebdeda2fd7f9e4c911deb00876f476ea9b0ce6"} Oct 05 07:24:31 crc kubenswrapper[4846]: I1005 07:24:31.292721 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-rk8kt" podStartSLOduration=1.721546939 podStartE2EDuration="4.292695791s" podCreationTimestamp="2025-10-05 07:24:27 +0000 UTC" firstStartedPulling="2025-10-05 07:24:28.225151172 +0000 UTC m=+2190.466003947" lastFinishedPulling="2025-10-05 07:24:30.796300014 +0000 UTC m=+2193.037152799" observedRunningTime="2025-10-05 07:24:31.29078758 +0000 UTC m=+2193.531640365" watchObservedRunningTime="2025-10-05 07:24:31.292695791 +0000 UTC m=+2193.533548586" Oct 05 07:24:35 crc kubenswrapper[4846]: I1005 07:24:35.080640 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-clnvm" Oct 05 07:24:35 crc kubenswrapper[4846]: I1005 07:24:35.081035 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-clnvm" Oct 05 07:24:35 crc kubenswrapper[4846]: I1005 07:24:35.165948 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-clnvm" Oct 05 07:24:35 crc kubenswrapper[4846]: I1005 07:24:35.389558 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-clnvm" Oct 05 07:24:35 crc kubenswrapper[4846]: I1005 07:24:35.512070 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-clnvm"] Oct 05 07:24:37 crc kubenswrapper[4846]: I1005 07:24:37.327090 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-clnvm" podUID="ed7834ca-b09f-4117-9d4a-75a401a43428" containerName="registry-server" containerID="cri-o://81d1dc3b8df3a57352261ff505bc40a9742fadb86eb663f6bff1752eafc332b1" gracePeriod=2 Oct 05 07:24:37 crc kubenswrapper[4846]: I1005 07:24:37.455517 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-rk8kt" Oct 05 07:24:37 crc kubenswrapper[4846]: I1005 07:24:37.456390 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-rk8kt" Oct 05 07:24:37 crc kubenswrapper[4846]: I1005 07:24:37.525405 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-rk8kt" Oct 05 07:24:37 crc kubenswrapper[4846]: I1005 07:24:37.747102 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-clnvm" Oct 05 07:24:37 crc kubenswrapper[4846]: I1005 07:24:37.844945 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4dh6f\" (UniqueName: \"kubernetes.io/projected/ed7834ca-b09f-4117-9d4a-75a401a43428-kube-api-access-4dh6f\") pod \"ed7834ca-b09f-4117-9d4a-75a401a43428\" (UID: \"ed7834ca-b09f-4117-9d4a-75a401a43428\") " Oct 05 07:24:37 crc kubenswrapper[4846]: I1005 07:24:37.845036 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ed7834ca-b09f-4117-9d4a-75a401a43428-catalog-content\") pod \"ed7834ca-b09f-4117-9d4a-75a401a43428\" (UID: \"ed7834ca-b09f-4117-9d4a-75a401a43428\") " Oct 05 07:24:37 crc kubenswrapper[4846]: I1005 07:24:37.845115 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ed7834ca-b09f-4117-9d4a-75a401a43428-utilities\") pod \"ed7834ca-b09f-4117-9d4a-75a401a43428\" (UID: \"ed7834ca-b09f-4117-9d4a-75a401a43428\") " Oct 05 07:24:37 crc kubenswrapper[4846]: I1005 07:24:37.846266 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ed7834ca-b09f-4117-9d4a-75a401a43428-utilities" (OuterVolumeSpecName: "utilities") pod "ed7834ca-b09f-4117-9d4a-75a401a43428" (UID: "ed7834ca-b09f-4117-9d4a-75a401a43428"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 07:24:37 crc kubenswrapper[4846]: I1005 07:24:37.850482 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ed7834ca-b09f-4117-9d4a-75a401a43428-kube-api-access-4dh6f" (OuterVolumeSpecName: "kube-api-access-4dh6f") pod "ed7834ca-b09f-4117-9d4a-75a401a43428" (UID: "ed7834ca-b09f-4117-9d4a-75a401a43428"). InnerVolumeSpecName "kube-api-access-4dh6f". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:24:37 crc kubenswrapper[4846]: I1005 07:24:37.910557 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ed7834ca-b09f-4117-9d4a-75a401a43428-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ed7834ca-b09f-4117-9d4a-75a401a43428" (UID: "ed7834ca-b09f-4117-9d4a-75a401a43428"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 07:24:37 crc kubenswrapper[4846]: I1005 07:24:37.947547 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4dh6f\" (UniqueName: \"kubernetes.io/projected/ed7834ca-b09f-4117-9d4a-75a401a43428-kube-api-access-4dh6f\") on node \"crc\" DevicePath \"\"" Oct 05 07:24:37 crc kubenswrapper[4846]: I1005 07:24:37.947578 4846 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ed7834ca-b09f-4117-9d4a-75a401a43428-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 05 07:24:37 crc kubenswrapper[4846]: I1005 07:24:37.947589 4846 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ed7834ca-b09f-4117-9d4a-75a401a43428-utilities\") on node \"crc\" DevicePath \"\"" Oct 05 07:24:38 crc kubenswrapper[4846]: I1005 07:24:38.341511 4846 generic.go:334] "Generic (PLEG): container finished" podID="ed7834ca-b09f-4117-9d4a-75a401a43428" containerID="81d1dc3b8df3a57352261ff505bc40a9742fadb86eb663f6bff1752eafc332b1" exitCode=0 Oct 05 07:24:38 crc kubenswrapper[4846]: I1005 07:24:38.341623 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-clnvm" event={"ID":"ed7834ca-b09f-4117-9d4a-75a401a43428","Type":"ContainerDied","Data":"81d1dc3b8df3a57352261ff505bc40a9742fadb86eb663f6bff1752eafc332b1"} Oct 05 07:24:38 crc kubenswrapper[4846]: I1005 07:24:38.341704 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-clnvm" event={"ID":"ed7834ca-b09f-4117-9d4a-75a401a43428","Type":"ContainerDied","Data":"82c05b47e24a20e45e284ebe695e6591382d280910173fbcb37fbcf93a48f2b6"} Oct 05 07:24:38 crc kubenswrapper[4846]: I1005 07:24:38.341744 4846 scope.go:117] "RemoveContainer" containerID="81d1dc3b8df3a57352261ff505bc40a9742fadb86eb663f6bff1752eafc332b1" Oct 05 07:24:38 crc kubenswrapper[4846]: I1005 07:24:38.343045 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-clnvm" Oct 05 07:24:38 crc kubenswrapper[4846]: I1005 07:24:38.374452 4846 scope.go:117] "RemoveContainer" containerID="b4ce0f1458455605c81ebc4adaea69e77e47d67bb88d7c6edd0a0fa2877c8265" Oct 05 07:24:38 crc kubenswrapper[4846]: I1005 07:24:38.414392 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-rk8kt" Oct 05 07:24:38 crc kubenswrapper[4846]: I1005 07:24:38.417544 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-clnvm"] Oct 05 07:24:38 crc kubenswrapper[4846]: I1005 07:24:38.418759 4846 scope.go:117] "RemoveContainer" containerID="649a41953281b41a777ff15169a825997e791b828bd2cc6700cf6e96e9423dcf" Oct 05 07:24:38 crc kubenswrapper[4846]: I1005 07:24:38.428244 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-clnvm"] Oct 05 07:24:38 crc kubenswrapper[4846]: I1005 07:24:38.448124 4846 scope.go:117] "RemoveContainer" containerID="81d1dc3b8df3a57352261ff505bc40a9742fadb86eb663f6bff1752eafc332b1" Oct 05 07:24:38 crc kubenswrapper[4846]: E1005 07:24:38.448610 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"81d1dc3b8df3a57352261ff505bc40a9742fadb86eb663f6bff1752eafc332b1\": container with ID starting with 81d1dc3b8df3a57352261ff505bc40a9742fadb86eb663f6bff1752eafc332b1 not found: ID does not exist" containerID="81d1dc3b8df3a57352261ff505bc40a9742fadb86eb663f6bff1752eafc332b1" Oct 05 07:24:38 crc kubenswrapper[4846]: I1005 07:24:38.448654 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"81d1dc3b8df3a57352261ff505bc40a9742fadb86eb663f6bff1752eafc332b1"} err="failed to get container status \"81d1dc3b8df3a57352261ff505bc40a9742fadb86eb663f6bff1752eafc332b1\": rpc error: code = NotFound desc = could not find container \"81d1dc3b8df3a57352261ff505bc40a9742fadb86eb663f6bff1752eafc332b1\": container with ID starting with 81d1dc3b8df3a57352261ff505bc40a9742fadb86eb663f6bff1752eafc332b1 not found: ID does not exist" Oct 05 07:24:38 crc kubenswrapper[4846]: I1005 07:24:38.448685 4846 scope.go:117] "RemoveContainer" containerID="b4ce0f1458455605c81ebc4adaea69e77e47d67bb88d7c6edd0a0fa2877c8265" Oct 05 07:24:38 crc kubenswrapper[4846]: E1005 07:24:38.449394 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b4ce0f1458455605c81ebc4adaea69e77e47d67bb88d7c6edd0a0fa2877c8265\": container with ID starting with b4ce0f1458455605c81ebc4adaea69e77e47d67bb88d7c6edd0a0fa2877c8265 not found: ID does not exist" containerID="b4ce0f1458455605c81ebc4adaea69e77e47d67bb88d7c6edd0a0fa2877c8265" Oct 05 07:24:38 crc kubenswrapper[4846]: I1005 07:24:38.449456 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b4ce0f1458455605c81ebc4adaea69e77e47d67bb88d7c6edd0a0fa2877c8265"} err="failed to get container status \"b4ce0f1458455605c81ebc4adaea69e77e47d67bb88d7c6edd0a0fa2877c8265\": rpc error: code = NotFound desc = could not find container \"b4ce0f1458455605c81ebc4adaea69e77e47d67bb88d7c6edd0a0fa2877c8265\": container with ID starting with b4ce0f1458455605c81ebc4adaea69e77e47d67bb88d7c6edd0a0fa2877c8265 not found: ID does not exist" Oct 05 07:24:38 crc kubenswrapper[4846]: I1005 07:24:38.449496 4846 scope.go:117] "RemoveContainer" containerID="649a41953281b41a777ff15169a825997e791b828bd2cc6700cf6e96e9423dcf" Oct 05 07:24:38 crc kubenswrapper[4846]: E1005 07:24:38.449896 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"649a41953281b41a777ff15169a825997e791b828bd2cc6700cf6e96e9423dcf\": container with ID starting with 649a41953281b41a777ff15169a825997e791b828bd2cc6700cf6e96e9423dcf not found: ID does not exist" containerID="649a41953281b41a777ff15169a825997e791b828bd2cc6700cf6e96e9423dcf" Oct 05 07:24:38 crc kubenswrapper[4846]: I1005 07:24:38.449932 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"649a41953281b41a777ff15169a825997e791b828bd2cc6700cf6e96e9423dcf"} err="failed to get container status \"649a41953281b41a777ff15169a825997e791b828bd2cc6700cf6e96e9423dcf\": rpc error: code = NotFound desc = could not find container \"649a41953281b41a777ff15169a825997e791b828bd2cc6700cf6e96e9423dcf\": container with ID starting with 649a41953281b41a777ff15169a825997e791b828bd2cc6700cf6e96e9423dcf not found: ID does not exist" Oct 05 07:24:38 crc kubenswrapper[4846]: I1005 07:24:38.510483 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ed7834ca-b09f-4117-9d4a-75a401a43428" path="/var/lib/kubelet/pods/ed7834ca-b09f-4117-9d4a-75a401a43428/volumes" Oct 05 07:24:40 crc kubenswrapper[4846]: I1005 07:24:40.710492 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-rk8kt"] Oct 05 07:24:41 crc kubenswrapper[4846]: I1005 07:24:41.371417 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-rk8kt" podUID="88c88750-5101-49e7-85d7-47ad81bf9fe8" containerName="registry-server" containerID="cri-o://0b708ade2544d24e946164684bebdeda2fd7f9e4c911deb00876f476ea9b0ce6" gracePeriod=2 Oct 05 07:24:41 crc kubenswrapper[4846]: I1005 07:24:41.816735 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rk8kt" Oct 05 07:24:41 crc kubenswrapper[4846]: I1005 07:24:41.933104 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/88c88750-5101-49e7-85d7-47ad81bf9fe8-utilities\") pod \"88c88750-5101-49e7-85d7-47ad81bf9fe8\" (UID: \"88c88750-5101-49e7-85d7-47ad81bf9fe8\") " Oct 05 07:24:41 crc kubenswrapper[4846]: I1005 07:24:41.933485 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/88c88750-5101-49e7-85d7-47ad81bf9fe8-catalog-content\") pod \"88c88750-5101-49e7-85d7-47ad81bf9fe8\" (UID: \"88c88750-5101-49e7-85d7-47ad81bf9fe8\") " Oct 05 07:24:41 crc kubenswrapper[4846]: I1005 07:24:41.933570 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5lx8t\" (UniqueName: \"kubernetes.io/projected/88c88750-5101-49e7-85d7-47ad81bf9fe8-kube-api-access-5lx8t\") pod \"88c88750-5101-49e7-85d7-47ad81bf9fe8\" (UID: \"88c88750-5101-49e7-85d7-47ad81bf9fe8\") " Oct 05 07:24:41 crc kubenswrapper[4846]: I1005 07:24:41.934784 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/88c88750-5101-49e7-85d7-47ad81bf9fe8-utilities" (OuterVolumeSpecName: "utilities") pod "88c88750-5101-49e7-85d7-47ad81bf9fe8" (UID: "88c88750-5101-49e7-85d7-47ad81bf9fe8"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 07:24:41 crc kubenswrapper[4846]: I1005 07:24:41.939440 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/88c88750-5101-49e7-85d7-47ad81bf9fe8-kube-api-access-5lx8t" (OuterVolumeSpecName: "kube-api-access-5lx8t") pod "88c88750-5101-49e7-85d7-47ad81bf9fe8" (UID: "88c88750-5101-49e7-85d7-47ad81bf9fe8"). InnerVolumeSpecName "kube-api-access-5lx8t". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:24:42 crc kubenswrapper[4846]: I1005 07:24:42.016282 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/88c88750-5101-49e7-85d7-47ad81bf9fe8-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "88c88750-5101-49e7-85d7-47ad81bf9fe8" (UID: "88c88750-5101-49e7-85d7-47ad81bf9fe8"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 07:24:42 crc kubenswrapper[4846]: I1005 07:24:42.035453 4846 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/88c88750-5101-49e7-85d7-47ad81bf9fe8-utilities\") on node \"crc\" DevicePath \"\"" Oct 05 07:24:42 crc kubenswrapper[4846]: I1005 07:24:42.035480 4846 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/88c88750-5101-49e7-85d7-47ad81bf9fe8-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 05 07:24:42 crc kubenswrapper[4846]: I1005 07:24:42.035493 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5lx8t\" (UniqueName: \"kubernetes.io/projected/88c88750-5101-49e7-85d7-47ad81bf9fe8-kube-api-access-5lx8t\") on node \"crc\" DevicePath \"\"" Oct 05 07:24:42 crc kubenswrapper[4846]: I1005 07:24:42.389532 4846 generic.go:334] "Generic (PLEG): container finished" podID="88c88750-5101-49e7-85d7-47ad81bf9fe8" containerID="0b708ade2544d24e946164684bebdeda2fd7f9e4c911deb00876f476ea9b0ce6" exitCode=0 Oct 05 07:24:42 crc kubenswrapper[4846]: I1005 07:24:42.389614 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rk8kt" event={"ID":"88c88750-5101-49e7-85d7-47ad81bf9fe8","Type":"ContainerDied","Data":"0b708ade2544d24e946164684bebdeda2fd7f9e4c911deb00876f476ea9b0ce6"} Oct 05 07:24:42 crc kubenswrapper[4846]: I1005 07:24:42.389620 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rk8kt" Oct 05 07:24:42 crc kubenswrapper[4846]: I1005 07:24:42.389670 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rk8kt" event={"ID":"88c88750-5101-49e7-85d7-47ad81bf9fe8","Type":"ContainerDied","Data":"5b09246fe409e13dd1da4c1ae05459b1ed30c59291d0a5a8ef5a67108a4671d3"} Oct 05 07:24:42 crc kubenswrapper[4846]: I1005 07:24:42.389710 4846 scope.go:117] "RemoveContainer" containerID="0b708ade2544d24e946164684bebdeda2fd7f9e4c911deb00876f476ea9b0ce6" Oct 05 07:24:42 crc kubenswrapper[4846]: I1005 07:24:42.425211 4846 scope.go:117] "RemoveContainer" containerID="8cfe4374cd02643f1ff0eb80affdcfb7b398ccb356c8451565930a5c2f37104f" Oct 05 07:24:42 crc kubenswrapper[4846]: I1005 07:24:42.441952 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-rk8kt"] Oct 05 07:24:42 crc kubenswrapper[4846]: I1005 07:24:42.448296 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-rk8kt"] Oct 05 07:24:42 crc kubenswrapper[4846]: I1005 07:24:42.468806 4846 scope.go:117] "RemoveContainer" containerID="2b843a250df67ce4ac4896ab5d0326cce00f4ce5701a5686f0b2bb109acd8996" Oct 05 07:24:42 crc kubenswrapper[4846]: I1005 07:24:42.489363 4846 scope.go:117] "RemoveContainer" containerID="0b708ade2544d24e946164684bebdeda2fd7f9e4c911deb00876f476ea9b0ce6" Oct 05 07:24:42 crc kubenswrapper[4846]: E1005 07:24:42.490012 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0b708ade2544d24e946164684bebdeda2fd7f9e4c911deb00876f476ea9b0ce6\": container with ID starting with 0b708ade2544d24e946164684bebdeda2fd7f9e4c911deb00876f476ea9b0ce6 not found: ID does not exist" containerID="0b708ade2544d24e946164684bebdeda2fd7f9e4c911deb00876f476ea9b0ce6" Oct 05 07:24:42 crc kubenswrapper[4846]: I1005 07:24:42.490110 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0b708ade2544d24e946164684bebdeda2fd7f9e4c911deb00876f476ea9b0ce6"} err="failed to get container status \"0b708ade2544d24e946164684bebdeda2fd7f9e4c911deb00876f476ea9b0ce6\": rpc error: code = NotFound desc = could not find container \"0b708ade2544d24e946164684bebdeda2fd7f9e4c911deb00876f476ea9b0ce6\": container with ID starting with 0b708ade2544d24e946164684bebdeda2fd7f9e4c911deb00876f476ea9b0ce6 not found: ID does not exist" Oct 05 07:24:42 crc kubenswrapper[4846]: I1005 07:24:42.490195 4846 scope.go:117] "RemoveContainer" containerID="8cfe4374cd02643f1ff0eb80affdcfb7b398ccb356c8451565930a5c2f37104f" Oct 05 07:24:42 crc kubenswrapper[4846]: E1005 07:24:42.491505 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8cfe4374cd02643f1ff0eb80affdcfb7b398ccb356c8451565930a5c2f37104f\": container with ID starting with 8cfe4374cd02643f1ff0eb80affdcfb7b398ccb356c8451565930a5c2f37104f not found: ID does not exist" containerID="8cfe4374cd02643f1ff0eb80affdcfb7b398ccb356c8451565930a5c2f37104f" Oct 05 07:24:42 crc kubenswrapper[4846]: I1005 07:24:42.491575 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8cfe4374cd02643f1ff0eb80affdcfb7b398ccb356c8451565930a5c2f37104f"} err="failed to get container status \"8cfe4374cd02643f1ff0eb80affdcfb7b398ccb356c8451565930a5c2f37104f\": rpc error: code = NotFound desc = could not find container \"8cfe4374cd02643f1ff0eb80affdcfb7b398ccb356c8451565930a5c2f37104f\": container with ID starting with 8cfe4374cd02643f1ff0eb80affdcfb7b398ccb356c8451565930a5c2f37104f not found: ID does not exist" Oct 05 07:24:42 crc kubenswrapper[4846]: I1005 07:24:42.491611 4846 scope.go:117] "RemoveContainer" containerID="2b843a250df67ce4ac4896ab5d0326cce00f4ce5701a5686f0b2bb109acd8996" Oct 05 07:24:42 crc kubenswrapper[4846]: E1005 07:24:42.492249 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2b843a250df67ce4ac4896ab5d0326cce00f4ce5701a5686f0b2bb109acd8996\": container with ID starting with 2b843a250df67ce4ac4896ab5d0326cce00f4ce5701a5686f0b2bb109acd8996 not found: ID does not exist" containerID="2b843a250df67ce4ac4896ab5d0326cce00f4ce5701a5686f0b2bb109acd8996" Oct 05 07:24:42 crc kubenswrapper[4846]: I1005 07:24:42.492302 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2b843a250df67ce4ac4896ab5d0326cce00f4ce5701a5686f0b2bb109acd8996"} err="failed to get container status \"2b843a250df67ce4ac4896ab5d0326cce00f4ce5701a5686f0b2bb109acd8996\": rpc error: code = NotFound desc = could not find container \"2b843a250df67ce4ac4896ab5d0326cce00f4ce5701a5686f0b2bb109acd8996\": container with ID starting with 2b843a250df67ce4ac4896ab5d0326cce00f4ce5701a5686f0b2bb109acd8996 not found: ID does not exist" Oct 05 07:24:42 crc kubenswrapper[4846]: I1005 07:24:42.512081 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="88c88750-5101-49e7-85d7-47ad81bf9fe8" path="/var/lib/kubelet/pods/88c88750-5101-49e7-85d7-47ad81bf9fe8/volumes" Oct 05 07:24:53 crc kubenswrapper[4846]: I1005 07:24:53.325114 4846 patch_prober.go:28] interesting pod/machine-config-daemon-fscvf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 07:24:53 crc kubenswrapper[4846]: I1005 07:24:53.326095 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 07:24:53 crc kubenswrapper[4846]: I1005 07:24:53.326214 4846 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" Oct 05 07:24:53 crc kubenswrapper[4846]: I1005 07:24:53.326944 4846 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"4817903cb1e5987efe5a26ad35de6b3af2cbb8dfef6b67e286cafc5827c7ded4"} pod="openshift-machine-config-operator/machine-config-daemon-fscvf" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 05 07:24:53 crc kubenswrapper[4846]: I1005 07:24:53.327060 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" containerName="machine-config-daemon" containerID="cri-o://4817903cb1e5987efe5a26ad35de6b3af2cbb8dfef6b67e286cafc5827c7ded4" gracePeriod=600 Oct 05 07:24:53 crc kubenswrapper[4846]: E1005 07:24:53.455727 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 07:24:53 crc kubenswrapper[4846]: I1005 07:24:53.507319 4846 generic.go:334] "Generic (PLEG): container finished" podID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" containerID="4817903cb1e5987efe5a26ad35de6b3af2cbb8dfef6b67e286cafc5827c7ded4" exitCode=0 Oct 05 07:24:53 crc kubenswrapper[4846]: I1005 07:24:53.507375 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" event={"ID":"ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7","Type":"ContainerDied","Data":"4817903cb1e5987efe5a26ad35de6b3af2cbb8dfef6b67e286cafc5827c7ded4"} Oct 05 07:24:53 crc kubenswrapper[4846]: I1005 07:24:53.507822 4846 scope.go:117] "RemoveContainer" containerID="5dcee45f19c6e0d9810c4dda319bcb382224184464f9a3dbabad80d9e701b330" Oct 05 07:24:53 crc kubenswrapper[4846]: I1005 07:24:53.508458 4846 scope.go:117] "RemoveContainer" containerID="4817903cb1e5987efe5a26ad35de6b3af2cbb8dfef6b67e286cafc5827c7ded4" Oct 05 07:24:53 crc kubenswrapper[4846]: E1005 07:24:53.508849 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 07:25:08 crc kubenswrapper[4846]: I1005 07:25:08.508933 4846 scope.go:117] "RemoveContainer" containerID="4817903cb1e5987efe5a26ad35de6b3af2cbb8dfef6b67e286cafc5827c7ded4" Oct 05 07:25:08 crc kubenswrapper[4846]: E1005 07:25:08.509581 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 07:25:19 crc kubenswrapper[4846]: I1005 07:25:19.497351 4846 scope.go:117] "RemoveContainer" containerID="4817903cb1e5987efe5a26ad35de6b3af2cbb8dfef6b67e286cafc5827c7ded4" Oct 05 07:25:19 crc kubenswrapper[4846]: E1005 07:25:19.498223 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 07:25:33 crc kubenswrapper[4846]: I1005 07:25:33.497608 4846 scope.go:117] "RemoveContainer" containerID="4817903cb1e5987efe5a26ad35de6b3af2cbb8dfef6b67e286cafc5827c7ded4" Oct 05 07:25:33 crc kubenswrapper[4846]: E1005 07:25:33.498833 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 07:25:46 crc kubenswrapper[4846]: I1005 07:25:46.497787 4846 scope.go:117] "RemoveContainer" containerID="4817903cb1e5987efe5a26ad35de6b3af2cbb8dfef6b67e286cafc5827c7ded4" Oct 05 07:25:46 crc kubenswrapper[4846]: E1005 07:25:46.498771 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 07:25:57 crc kubenswrapper[4846]: I1005 07:25:57.497723 4846 scope.go:117] "RemoveContainer" containerID="4817903cb1e5987efe5a26ad35de6b3af2cbb8dfef6b67e286cafc5827c7ded4" Oct 05 07:25:57 crc kubenswrapper[4846]: E1005 07:25:57.498938 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 07:26:11 crc kubenswrapper[4846]: I1005 07:26:11.497732 4846 scope.go:117] "RemoveContainer" containerID="4817903cb1e5987efe5a26ad35de6b3af2cbb8dfef6b67e286cafc5827c7ded4" Oct 05 07:26:11 crc kubenswrapper[4846]: E1005 07:26:11.499321 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 07:26:26 crc kubenswrapper[4846]: I1005 07:26:26.498522 4846 scope.go:117] "RemoveContainer" containerID="4817903cb1e5987efe5a26ad35de6b3af2cbb8dfef6b67e286cafc5827c7ded4" Oct 05 07:26:26 crc kubenswrapper[4846]: E1005 07:26:26.499694 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 07:26:40 crc kubenswrapper[4846]: I1005 07:26:40.498270 4846 scope.go:117] "RemoveContainer" containerID="4817903cb1e5987efe5a26ad35de6b3af2cbb8dfef6b67e286cafc5827c7ded4" Oct 05 07:26:40 crc kubenswrapper[4846]: E1005 07:26:40.501051 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 07:26:52 crc kubenswrapper[4846]: I1005 07:26:52.498488 4846 scope.go:117] "RemoveContainer" containerID="4817903cb1e5987efe5a26ad35de6b3af2cbb8dfef6b67e286cafc5827c7ded4" Oct 05 07:26:52 crc kubenswrapper[4846]: E1005 07:26:52.499292 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 07:27:03 crc kubenswrapper[4846]: I1005 07:27:03.497615 4846 scope.go:117] "RemoveContainer" containerID="4817903cb1e5987efe5a26ad35de6b3af2cbb8dfef6b67e286cafc5827c7ded4" Oct 05 07:27:03 crc kubenswrapper[4846]: E1005 07:27:03.498344 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 07:27:16 crc kubenswrapper[4846]: I1005 07:27:16.527209 4846 scope.go:117] "RemoveContainer" containerID="4817903cb1e5987efe5a26ad35de6b3af2cbb8dfef6b67e286cafc5827c7ded4" Oct 05 07:27:16 crc kubenswrapper[4846]: E1005 07:27:16.528001 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 07:27:27 crc kubenswrapper[4846]: I1005 07:27:27.498588 4846 scope.go:117] "RemoveContainer" containerID="4817903cb1e5987efe5a26ad35de6b3af2cbb8dfef6b67e286cafc5827c7ded4" Oct 05 07:27:27 crc kubenswrapper[4846]: E1005 07:27:27.499706 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 07:27:41 crc kubenswrapper[4846]: I1005 07:27:41.497864 4846 scope.go:117] "RemoveContainer" containerID="4817903cb1e5987efe5a26ad35de6b3af2cbb8dfef6b67e286cafc5827c7ded4" Oct 05 07:27:41 crc kubenswrapper[4846]: E1005 07:27:41.498755 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 07:27:56 crc kubenswrapper[4846]: I1005 07:27:56.497909 4846 scope.go:117] "RemoveContainer" containerID="4817903cb1e5987efe5a26ad35de6b3af2cbb8dfef6b67e286cafc5827c7ded4" Oct 05 07:27:56 crc kubenswrapper[4846]: E1005 07:27:56.499103 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 07:28:09 crc kubenswrapper[4846]: I1005 07:28:09.498348 4846 scope.go:117] "RemoveContainer" containerID="4817903cb1e5987efe5a26ad35de6b3af2cbb8dfef6b67e286cafc5827c7ded4" Oct 05 07:28:09 crc kubenswrapper[4846]: E1005 07:28:09.499124 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 07:28:21 crc kubenswrapper[4846]: I1005 07:28:21.497916 4846 scope.go:117] "RemoveContainer" containerID="4817903cb1e5987efe5a26ad35de6b3af2cbb8dfef6b67e286cafc5827c7ded4" Oct 05 07:28:21 crc kubenswrapper[4846]: E1005 07:28:21.500995 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 07:28:34 crc kubenswrapper[4846]: I1005 07:28:34.498226 4846 scope.go:117] "RemoveContainer" containerID="4817903cb1e5987efe5a26ad35de6b3af2cbb8dfef6b67e286cafc5827c7ded4" Oct 05 07:28:34 crc kubenswrapper[4846]: E1005 07:28:34.499140 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 07:28:45 crc kubenswrapper[4846]: I1005 07:28:45.497599 4846 scope.go:117] "RemoveContainer" containerID="4817903cb1e5987efe5a26ad35de6b3af2cbb8dfef6b67e286cafc5827c7ded4" Oct 05 07:28:45 crc kubenswrapper[4846]: E1005 07:28:45.498516 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 07:28:57 crc kubenswrapper[4846]: I1005 07:28:57.498990 4846 scope.go:117] "RemoveContainer" containerID="4817903cb1e5987efe5a26ad35de6b3af2cbb8dfef6b67e286cafc5827c7ded4" Oct 05 07:28:57 crc kubenswrapper[4846]: E1005 07:28:57.500898 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 07:29:11 crc kubenswrapper[4846]: I1005 07:29:11.498456 4846 scope.go:117] "RemoveContainer" containerID="4817903cb1e5987efe5a26ad35de6b3af2cbb8dfef6b67e286cafc5827c7ded4" Oct 05 07:29:11 crc kubenswrapper[4846]: E1005 07:29:11.499494 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 07:29:25 crc kubenswrapper[4846]: I1005 07:29:25.498240 4846 scope.go:117] "RemoveContainer" containerID="4817903cb1e5987efe5a26ad35de6b3af2cbb8dfef6b67e286cafc5827c7ded4" Oct 05 07:29:25 crc kubenswrapper[4846]: E1005 07:29:25.499405 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 07:29:38 crc kubenswrapper[4846]: I1005 07:29:38.505285 4846 scope.go:117] "RemoveContainer" containerID="4817903cb1e5987efe5a26ad35de6b3af2cbb8dfef6b67e286cafc5827c7ded4" Oct 05 07:29:38 crc kubenswrapper[4846]: E1005 07:29:38.506508 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 07:29:51 crc kubenswrapper[4846]: I1005 07:29:51.497810 4846 scope.go:117] "RemoveContainer" containerID="4817903cb1e5987efe5a26ad35de6b3af2cbb8dfef6b67e286cafc5827c7ded4" Oct 05 07:29:51 crc kubenswrapper[4846]: E1005 07:29:51.498997 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 07:30:00 crc kubenswrapper[4846]: I1005 07:30:00.167616 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29327490-6kpqj"] Oct 05 07:30:00 crc kubenswrapper[4846]: E1005 07:30:00.169108 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ed7834ca-b09f-4117-9d4a-75a401a43428" containerName="registry-server" Oct 05 07:30:00 crc kubenswrapper[4846]: I1005 07:30:00.169133 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="ed7834ca-b09f-4117-9d4a-75a401a43428" containerName="registry-server" Oct 05 07:30:00 crc kubenswrapper[4846]: E1005 07:30:00.169151 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="88c88750-5101-49e7-85d7-47ad81bf9fe8" containerName="extract-utilities" Oct 05 07:30:00 crc kubenswrapper[4846]: I1005 07:30:00.169164 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="88c88750-5101-49e7-85d7-47ad81bf9fe8" containerName="extract-utilities" Oct 05 07:30:00 crc kubenswrapper[4846]: E1005 07:30:00.169223 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="88c88750-5101-49e7-85d7-47ad81bf9fe8" containerName="extract-content" Oct 05 07:30:00 crc kubenswrapper[4846]: I1005 07:30:00.169240 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="88c88750-5101-49e7-85d7-47ad81bf9fe8" containerName="extract-content" Oct 05 07:30:00 crc kubenswrapper[4846]: E1005 07:30:00.169263 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ed7834ca-b09f-4117-9d4a-75a401a43428" containerName="extract-content" Oct 05 07:30:00 crc kubenswrapper[4846]: I1005 07:30:00.169274 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="ed7834ca-b09f-4117-9d4a-75a401a43428" containerName="extract-content" Oct 05 07:30:00 crc kubenswrapper[4846]: E1005 07:30:00.169313 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ed7834ca-b09f-4117-9d4a-75a401a43428" containerName="extract-utilities" Oct 05 07:30:00 crc kubenswrapper[4846]: I1005 07:30:00.169325 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="ed7834ca-b09f-4117-9d4a-75a401a43428" containerName="extract-utilities" Oct 05 07:30:00 crc kubenswrapper[4846]: E1005 07:30:00.169362 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="88c88750-5101-49e7-85d7-47ad81bf9fe8" containerName="registry-server" Oct 05 07:30:00 crc kubenswrapper[4846]: I1005 07:30:00.169374 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="88c88750-5101-49e7-85d7-47ad81bf9fe8" containerName="registry-server" Oct 05 07:30:00 crc kubenswrapper[4846]: I1005 07:30:00.169680 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="ed7834ca-b09f-4117-9d4a-75a401a43428" containerName="registry-server" Oct 05 07:30:00 crc kubenswrapper[4846]: I1005 07:30:00.169714 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="88c88750-5101-49e7-85d7-47ad81bf9fe8" containerName="registry-server" Oct 05 07:30:00 crc kubenswrapper[4846]: I1005 07:30:00.170717 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29327490-6kpqj" Oct 05 07:30:00 crc kubenswrapper[4846]: I1005 07:30:00.174162 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 05 07:30:00 crc kubenswrapper[4846]: I1005 07:30:00.175004 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 05 07:30:00 crc kubenswrapper[4846]: I1005 07:30:00.192861 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29327490-6kpqj"] Oct 05 07:30:00 crc kubenswrapper[4846]: I1005 07:30:00.292441 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5e0701f1-92c9-45fe-81d9-910fcf30ea6d-config-volume\") pod \"collect-profiles-29327490-6kpqj\" (UID: \"5e0701f1-92c9-45fe-81d9-910fcf30ea6d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327490-6kpqj" Oct 05 07:30:00 crc kubenswrapper[4846]: I1005 07:30:00.292784 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zwtc9\" (UniqueName: \"kubernetes.io/projected/5e0701f1-92c9-45fe-81d9-910fcf30ea6d-kube-api-access-zwtc9\") pod \"collect-profiles-29327490-6kpqj\" (UID: \"5e0701f1-92c9-45fe-81d9-910fcf30ea6d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327490-6kpqj" Oct 05 07:30:00 crc kubenswrapper[4846]: I1005 07:30:00.292882 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5e0701f1-92c9-45fe-81d9-910fcf30ea6d-secret-volume\") pod \"collect-profiles-29327490-6kpqj\" (UID: \"5e0701f1-92c9-45fe-81d9-910fcf30ea6d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327490-6kpqj" Oct 05 07:30:00 crc kubenswrapper[4846]: I1005 07:30:00.394933 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zwtc9\" (UniqueName: \"kubernetes.io/projected/5e0701f1-92c9-45fe-81d9-910fcf30ea6d-kube-api-access-zwtc9\") pod \"collect-profiles-29327490-6kpqj\" (UID: \"5e0701f1-92c9-45fe-81d9-910fcf30ea6d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327490-6kpqj" Oct 05 07:30:00 crc kubenswrapper[4846]: I1005 07:30:00.395019 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5e0701f1-92c9-45fe-81d9-910fcf30ea6d-secret-volume\") pod \"collect-profiles-29327490-6kpqj\" (UID: \"5e0701f1-92c9-45fe-81d9-910fcf30ea6d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327490-6kpqj" Oct 05 07:30:00 crc kubenswrapper[4846]: I1005 07:30:00.395276 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5e0701f1-92c9-45fe-81d9-910fcf30ea6d-config-volume\") pod \"collect-profiles-29327490-6kpqj\" (UID: \"5e0701f1-92c9-45fe-81d9-910fcf30ea6d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327490-6kpqj" Oct 05 07:30:00 crc kubenswrapper[4846]: I1005 07:30:00.397264 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5e0701f1-92c9-45fe-81d9-910fcf30ea6d-config-volume\") pod \"collect-profiles-29327490-6kpqj\" (UID: \"5e0701f1-92c9-45fe-81d9-910fcf30ea6d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327490-6kpqj" Oct 05 07:30:00 crc kubenswrapper[4846]: I1005 07:30:00.409324 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5e0701f1-92c9-45fe-81d9-910fcf30ea6d-secret-volume\") pod \"collect-profiles-29327490-6kpqj\" (UID: \"5e0701f1-92c9-45fe-81d9-910fcf30ea6d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327490-6kpqj" Oct 05 07:30:00 crc kubenswrapper[4846]: I1005 07:30:00.426131 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zwtc9\" (UniqueName: \"kubernetes.io/projected/5e0701f1-92c9-45fe-81d9-910fcf30ea6d-kube-api-access-zwtc9\") pod \"collect-profiles-29327490-6kpqj\" (UID: \"5e0701f1-92c9-45fe-81d9-910fcf30ea6d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327490-6kpqj" Oct 05 07:30:00 crc kubenswrapper[4846]: I1005 07:30:00.505636 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29327490-6kpqj" Oct 05 07:30:00 crc kubenswrapper[4846]: I1005 07:30:00.752445 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29327490-6kpqj"] Oct 05 07:30:01 crc kubenswrapper[4846]: W1005 07:30:01.119980 4846 helpers.go:245] readString: Failed to read "/sys/fs/cgroup/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5e0701f1_92c9_45fe_81d9_910fcf30ea6d.slice/crio-684ae39745ff6d2990ca068ccd0c568189894228b378e8b8f52ea762a2fc2e61.scope/pids.max": read /sys/fs/cgroup/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5e0701f1_92c9_45fe_81d9_910fcf30ea6d.slice/crio-684ae39745ff6d2990ca068ccd0c568189894228b378e8b8f52ea762a2fc2e61.scope/pids.max: no such device Oct 05 07:30:01 crc kubenswrapper[4846]: I1005 07:30:01.554229 4846 generic.go:334] "Generic (PLEG): container finished" podID="5e0701f1-92c9-45fe-81d9-910fcf30ea6d" containerID="684ae39745ff6d2990ca068ccd0c568189894228b378e8b8f52ea762a2fc2e61" exitCode=0 Oct 05 07:30:01 crc kubenswrapper[4846]: I1005 07:30:01.554331 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29327490-6kpqj" event={"ID":"5e0701f1-92c9-45fe-81d9-910fcf30ea6d","Type":"ContainerDied","Data":"684ae39745ff6d2990ca068ccd0c568189894228b378e8b8f52ea762a2fc2e61"} Oct 05 07:30:01 crc kubenswrapper[4846]: I1005 07:30:01.554554 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29327490-6kpqj" event={"ID":"5e0701f1-92c9-45fe-81d9-910fcf30ea6d","Type":"ContainerStarted","Data":"f7929322a139774458e0659da6e183a2afe039663d6aeb793916dc06e93990fa"} Oct 05 07:30:02 crc kubenswrapper[4846]: I1005 07:30:02.927334 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29327490-6kpqj" Oct 05 07:30:03 crc kubenswrapper[4846]: I1005 07:30:03.043369 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5e0701f1-92c9-45fe-81d9-910fcf30ea6d-config-volume\") pod \"5e0701f1-92c9-45fe-81d9-910fcf30ea6d\" (UID: \"5e0701f1-92c9-45fe-81d9-910fcf30ea6d\") " Oct 05 07:30:03 crc kubenswrapper[4846]: I1005 07:30:03.043565 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zwtc9\" (UniqueName: \"kubernetes.io/projected/5e0701f1-92c9-45fe-81d9-910fcf30ea6d-kube-api-access-zwtc9\") pod \"5e0701f1-92c9-45fe-81d9-910fcf30ea6d\" (UID: \"5e0701f1-92c9-45fe-81d9-910fcf30ea6d\") " Oct 05 07:30:03 crc kubenswrapper[4846]: I1005 07:30:03.043600 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5e0701f1-92c9-45fe-81d9-910fcf30ea6d-secret-volume\") pod \"5e0701f1-92c9-45fe-81d9-910fcf30ea6d\" (UID: \"5e0701f1-92c9-45fe-81d9-910fcf30ea6d\") " Oct 05 07:30:03 crc kubenswrapper[4846]: I1005 07:30:03.045735 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5e0701f1-92c9-45fe-81d9-910fcf30ea6d-config-volume" (OuterVolumeSpecName: "config-volume") pod "5e0701f1-92c9-45fe-81d9-910fcf30ea6d" (UID: "5e0701f1-92c9-45fe-81d9-910fcf30ea6d"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 07:30:03 crc kubenswrapper[4846]: I1005 07:30:03.048991 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5e0701f1-92c9-45fe-81d9-910fcf30ea6d-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "5e0701f1-92c9-45fe-81d9-910fcf30ea6d" (UID: "5e0701f1-92c9-45fe-81d9-910fcf30ea6d"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:30:03 crc kubenswrapper[4846]: I1005 07:30:03.049242 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5e0701f1-92c9-45fe-81d9-910fcf30ea6d-kube-api-access-zwtc9" (OuterVolumeSpecName: "kube-api-access-zwtc9") pod "5e0701f1-92c9-45fe-81d9-910fcf30ea6d" (UID: "5e0701f1-92c9-45fe-81d9-910fcf30ea6d"). InnerVolumeSpecName "kube-api-access-zwtc9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:30:03 crc kubenswrapper[4846]: I1005 07:30:03.145292 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zwtc9\" (UniqueName: \"kubernetes.io/projected/5e0701f1-92c9-45fe-81d9-910fcf30ea6d-kube-api-access-zwtc9\") on node \"crc\" DevicePath \"\"" Oct 05 07:30:03 crc kubenswrapper[4846]: I1005 07:30:03.145330 4846 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5e0701f1-92c9-45fe-81d9-910fcf30ea6d-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 05 07:30:03 crc kubenswrapper[4846]: I1005 07:30:03.145343 4846 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5e0701f1-92c9-45fe-81d9-910fcf30ea6d-config-volume\") on node \"crc\" DevicePath \"\"" Oct 05 07:30:03 crc kubenswrapper[4846]: I1005 07:30:03.497827 4846 scope.go:117] "RemoveContainer" containerID="4817903cb1e5987efe5a26ad35de6b3af2cbb8dfef6b67e286cafc5827c7ded4" Oct 05 07:30:03 crc kubenswrapper[4846]: I1005 07:30:03.575810 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29327490-6kpqj" event={"ID":"5e0701f1-92c9-45fe-81d9-910fcf30ea6d","Type":"ContainerDied","Data":"f7929322a139774458e0659da6e183a2afe039663d6aeb793916dc06e93990fa"} Oct 05 07:30:03 crc kubenswrapper[4846]: I1005 07:30:03.575886 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f7929322a139774458e0659da6e183a2afe039663d6aeb793916dc06e93990fa" Oct 05 07:30:03 crc kubenswrapper[4846]: I1005 07:30:03.575933 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29327490-6kpqj" Oct 05 07:30:04 crc kubenswrapper[4846]: I1005 07:30:04.048078 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29327445-dn5mf"] Oct 05 07:30:04 crc kubenswrapper[4846]: I1005 07:30:04.057499 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29327445-dn5mf"] Oct 05 07:30:04 crc kubenswrapper[4846]: I1005 07:30:04.508828 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="41aa056a-34d5-463b-8352-c71e0f0db1a4" path="/var/lib/kubelet/pods/41aa056a-34d5-463b-8352-c71e0f0db1a4/volumes" Oct 05 07:30:04 crc kubenswrapper[4846]: I1005 07:30:04.587627 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" event={"ID":"ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7","Type":"ContainerStarted","Data":"b6694c2f42f5d02bf9e82272eb86940e1c6a3c4bb6e25979d8d7da5c72b02506"} Oct 05 07:30:28 crc kubenswrapper[4846]: I1005 07:30:28.972947 4846 scope.go:117] "RemoveContainer" containerID="e46a0ae8f0a310b6735443855f8119ee77787d11058d4f9af366550a7374d17d" Oct 05 07:32:23 crc kubenswrapper[4846]: I1005 07:32:23.326065 4846 patch_prober.go:28] interesting pod/machine-config-daemon-fscvf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 07:32:23 crc kubenswrapper[4846]: I1005 07:32:23.326553 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 07:32:53 crc kubenswrapper[4846]: I1005 07:32:53.324753 4846 patch_prober.go:28] interesting pod/machine-config-daemon-fscvf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 07:32:53 crc kubenswrapper[4846]: I1005 07:32:53.325317 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 07:33:23 crc kubenswrapper[4846]: I1005 07:33:23.324946 4846 patch_prober.go:28] interesting pod/machine-config-daemon-fscvf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 07:33:23 crc kubenswrapper[4846]: I1005 07:33:23.326399 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 07:33:23 crc kubenswrapper[4846]: I1005 07:33:23.326501 4846 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" Oct 05 07:33:23 crc kubenswrapper[4846]: I1005 07:33:23.327503 4846 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"b6694c2f42f5d02bf9e82272eb86940e1c6a3c4bb6e25979d8d7da5c72b02506"} pod="openshift-machine-config-operator/machine-config-daemon-fscvf" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 05 07:33:23 crc kubenswrapper[4846]: I1005 07:33:23.327605 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" containerName="machine-config-daemon" containerID="cri-o://b6694c2f42f5d02bf9e82272eb86940e1c6a3c4bb6e25979d8d7da5c72b02506" gracePeriod=600 Oct 05 07:33:23 crc kubenswrapper[4846]: I1005 07:33:23.495330 4846 generic.go:334] "Generic (PLEG): container finished" podID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" containerID="b6694c2f42f5d02bf9e82272eb86940e1c6a3c4bb6e25979d8d7da5c72b02506" exitCode=0 Oct 05 07:33:23 crc kubenswrapper[4846]: I1005 07:33:23.495377 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" event={"ID":"ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7","Type":"ContainerDied","Data":"b6694c2f42f5d02bf9e82272eb86940e1c6a3c4bb6e25979d8d7da5c72b02506"} Oct 05 07:33:23 crc kubenswrapper[4846]: I1005 07:33:23.495413 4846 scope.go:117] "RemoveContainer" containerID="4817903cb1e5987efe5a26ad35de6b3af2cbb8dfef6b67e286cafc5827c7ded4" Oct 05 07:33:24 crc kubenswrapper[4846]: I1005 07:33:24.514467 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" event={"ID":"ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7","Type":"ContainerStarted","Data":"702f9b90152900919451502be251dfa1fc91d3fc2af6439c709b401f8ac5d520"} Oct 05 07:33:42 crc kubenswrapper[4846]: I1005 07:33:42.379813 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-jjtr9"] Oct 05 07:33:42 crc kubenswrapper[4846]: E1005 07:33:42.382806 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5e0701f1-92c9-45fe-81d9-910fcf30ea6d" containerName="collect-profiles" Oct 05 07:33:42 crc kubenswrapper[4846]: I1005 07:33:42.382970 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="5e0701f1-92c9-45fe-81d9-910fcf30ea6d" containerName="collect-profiles" Oct 05 07:33:42 crc kubenswrapper[4846]: I1005 07:33:42.383434 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="5e0701f1-92c9-45fe-81d9-910fcf30ea6d" containerName="collect-profiles" Oct 05 07:33:42 crc kubenswrapper[4846]: I1005 07:33:42.386229 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jjtr9" Oct 05 07:33:42 crc kubenswrapper[4846]: I1005 07:33:42.393689 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-jjtr9"] Oct 05 07:33:42 crc kubenswrapper[4846]: I1005 07:33:42.576973 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f724a69c-9f76-495d-8b4c-222177c6f591-catalog-content\") pod \"certified-operators-jjtr9\" (UID: \"f724a69c-9f76-495d-8b4c-222177c6f591\") " pod="openshift-marketplace/certified-operators-jjtr9" Oct 05 07:33:42 crc kubenswrapper[4846]: I1005 07:33:42.577366 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pcblk\" (UniqueName: \"kubernetes.io/projected/f724a69c-9f76-495d-8b4c-222177c6f591-kube-api-access-pcblk\") pod \"certified-operators-jjtr9\" (UID: \"f724a69c-9f76-495d-8b4c-222177c6f591\") " pod="openshift-marketplace/certified-operators-jjtr9" Oct 05 07:33:42 crc kubenswrapper[4846]: I1005 07:33:42.577471 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f724a69c-9f76-495d-8b4c-222177c6f591-utilities\") pod \"certified-operators-jjtr9\" (UID: \"f724a69c-9f76-495d-8b4c-222177c6f591\") " pod="openshift-marketplace/certified-operators-jjtr9" Oct 05 07:33:42 crc kubenswrapper[4846]: I1005 07:33:42.678908 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f724a69c-9f76-495d-8b4c-222177c6f591-catalog-content\") pod \"certified-operators-jjtr9\" (UID: \"f724a69c-9f76-495d-8b4c-222177c6f591\") " pod="openshift-marketplace/certified-operators-jjtr9" Oct 05 07:33:42 crc kubenswrapper[4846]: I1005 07:33:42.679052 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pcblk\" (UniqueName: \"kubernetes.io/projected/f724a69c-9f76-495d-8b4c-222177c6f591-kube-api-access-pcblk\") pod \"certified-operators-jjtr9\" (UID: \"f724a69c-9f76-495d-8b4c-222177c6f591\") " pod="openshift-marketplace/certified-operators-jjtr9" Oct 05 07:33:42 crc kubenswrapper[4846]: I1005 07:33:42.679148 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f724a69c-9f76-495d-8b4c-222177c6f591-utilities\") pod \"certified-operators-jjtr9\" (UID: \"f724a69c-9f76-495d-8b4c-222177c6f591\") " pod="openshift-marketplace/certified-operators-jjtr9" Oct 05 07:33:42 crc kubenswrapper[4846]: I1005 07:33:42.679547 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f724a69c-9f76-495d-8b4c-222177c6f591-catalog-content\") pod \"certified-operators-jjtr9\" (UID: \"f724a69c-9f76-495d-8b4c-222177c6f591\") " pod="openshift-marketplace/certified-operators-jjtr9" Oct 05 07:33:42 crc kubenswrapper[4846]: I1005 07:33:42.679867 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f724a69c-9f76-495d-8b4c-222177c6f591-utilities\") pod \"certified-operators-jjtr9\" (UID: \"f724a69c-9f76-495d-8b4c-222177c6f591\") " pod="openshift-marketplace/certified-operators-jjtr9" Oct 05 07:33:42 crc kubenswrapper[4846]: I1005 07:33:42.716563 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pcblk\" (UniqueName: \"kubernetes.io/projected/f724a69c-9f76-495d-8b4c-222177c6f591-kube-api-access-pcblk\") pod \"certified-operators-jjtr9\" (UID: \"f724a69c-9f76-495d-8b4c-222177c6f591\") " pod="openshift-marketplace/certified-operators-jjtr9" Oct 05 07:33:42 crc kubenswrapper[4846]: I1005 07:33:42.727171 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jjtr9" Oct 05 07:33:43 crc kubenswrapper[4846]: I1005 07:33:43.044281 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-jjtr9"] Oct 05 07:33:43 crc kubenswrapper[4846]: W1005 07:33:43.053298 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf724a69c_9f76_495d_8b4c_222177c6f591.slice/crio-c26642951bdaff30e54d596f242fca0386e0af2bfd060016d548d153edd26d89 WatchSource:0}: Error finding container c26642951bdaff30e54d596f242fca0386e0af2bfd060016d548d153edd26d89: Status 404 returned error can't find the container with id c26642951bdaff30e54d596f242fca0386e0af2bfd060016d548d153edd26d89 Oct 05 07:33:43 crc kubenswrapper[4846]: I1005 07:33:43.689561 4846 generic.go:334] "Generic (PLEG): container finished" podID="f724a69c-9f76-495d-8b4c-222177c6f591" containerID="37ac93a7793e4706594579a6d78ff19eca3fa87a0a40182f274a8dd110f9f121" exitCode=0 Oct 05 07:33:43 crc kubenswrapper[4846]: I1005 07:33:43.689628 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jjtr9" event={"ID":"f724a69c-9f76-495d-8b4c-222177c6f591","Type":"ContainerDied","Data":"37ac93a7793e4706594579a6d78ff19eca3fa87a0a40182f274a8dd110f9f121"} Oct 05 07:33:43 crc kubenswrapper[4846]: I1005 07:33:43.689670 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jjtr9" event={"ID":"f724a69c-9f76-495d-8b4c-222177c6f591","Type":"ContainerStarted","Data":"c26642951bdaff30e54d596f242fca0386e0af2bfd060016d548d153edd26d89"} Oct 05 07:33:43 crc kubenswrapper[4846]: I1005 07:33:43.692709 4846 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 05 07:33:45 crc kubenswrapper[4846]: I1005 07:33:45.718472 4846 generic.go:334] "Generic (PLEG): container finished" podID="f724a69c-9f76-495d-8b4c-222177c6f591" containerID="44abd4b1a6f214543a32650b81193323b80c1ed3236e3bee2d0d20aa6a5b1f14" exitCode=0 Oct 05 07:33:45 crc kubenswrapper[4846]: I1005 07:33:45.719316 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jjtr9" event={"ID":"f724a69c-9f76-495d-8b4c-222177c6f591","Type":"ContainerDied","Data":"44abd4b1a6f214543a32650b81193323b80c1ed3236e3bee2d0d20aa6a5b1f14"} Oct 05 07:33:46 crc kubenswrapper[4846]: I1005 07:33:46.739408 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jjtr9" event={"ID":"f724a69c-9f76-495d-8b4c-222177c6f591","Type":"ContainerStarted","Data":"d1232715265b8327f5210ff673ef3362140f3bec0d185e1b62be7a192ff15740"} Oct 05 07:33:46 crc kubenswrapper[4846]: I1005 07:33:46.773136 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-jjtr9" podStartSLOduration=2.290910844 podStartE2EDuration="4.773088942s" podCreationTimestamp="2025-10-05 07:33:42 +0000 UTC" firstStartedPulling="2025-10-05 07:33:43.69231604 +0000 UTC m=+2745.933168855" lastFinishedPulling="2025-10-05 07:33:46.174494168 +0000 UTC m=+2748.415346953" observedRunningTime="2025-10-05 07:33:46.766471745 +0000 UTC m=+2749.007324590" watchObservedRunningTime="2025-10-05 07:33:46.773088942 +0000 UTC m=+2749.013941727" Oct 05 07:33:52 crc kubenswrapper[4846]: I1005 07:33:52.728422 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-jjtr9" Oct 05 07:33:52 crc kubenswrapper[4846]: I1005 07:33:52.728817 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-jjtr9" Oct 05 07:33:52 crc kubenswrapper[4846]: I1005 07:33:52.804010 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-jjtr9" Oct 05 07:33:52 crc kubenswrapper[4846]: I1005 07:33:52.889105 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-jjtr9" Oct 05 07:33:53 crc kubenswrapper[4846]: I1005 07:33:53.064494 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-jjtr9"] Oct 05 07:33:54 crc kubenswrapper[4846]: I1005 07:33:54.813123 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-jjtr9" podUID="f724a69c-9f76-495d-8b4c-222177c6f591" containerName="registry-server" containerID="cri-o://d1232715265b8327f5210ff673ef3362140f3bec0d185e1b62be7a192ff15740" gracePeriod=2 Oct 05 07:33:55 crc kubenswrapper[4846]: I1005 07:33:55.341380 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jjtr9" Oct 05 07:33:55 crc kubenswrapper[4846]: I1005 07:33:55.408058 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcblk\" (UniqueName: \"kubernetes.io/projected/f724a69c-9f76-495d-8b4c-222177c6f591-kube-api-access-pcblk\") pod \"f724a69c-9f76-495d-8b4c-222177c6f591\" (UID: \"f724a69c-9f76-495d-8b4c-222177c6f591\") " Oct 05 07:33:55 crc kubenswrapper[4846]: I1005 07:33:55.408112 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f724a69c-9f76-495d-8b4c-222177c6f591-catalog-content\") pod \"f724a69c-9f76-495d-8b4c-222177c6f591\" (UID: \"f724a69c-9f76-495d-8b4c-222177c6f591\") " Oct 05 07:33:55 crc kubenswrapper[4846]: I1005 07:33:55.408191 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f724a69c-9f76-495d-8b4c-222177c6f591-utilities\") pod \"f724a69c-9f76-495d-8b4c-222177c6f591\" (UID: \"f724a69c-9f76-495d-8b4c-222177c6f591\") " Oct 05 07:33:55 crc kubenswrapper[4846]: I1005 07:33:55.413853 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f724a69c-9f76-495d-8b4c-222177c6f591-utilities" (OuterVolumeSpecName: "utilities") pod "f724a69c-9f76-495d-8b4c-222177c6f591" (UID: "f724a69c-9f76-495d-8b4c-222177c6f591"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 07:33:55 crc kubenswrapper[4846]: I1005 07:33:55.423352 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f724a69c-9f76-495d-8b4c-222177c6f591-kube-api-access-pcblk" (OuterVolumeSpecName: "kube-api-access-pcblk") pod "f724a69c-9f76-495d-8b4c-222177c6f591" (UID: "f724a69c-9f76-495d-8b4c-222177c6f591"). InnerVolumeSpecName "kube-api-access-pcblk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:33:55 crc kubenswrapper[4846]: I1005 07:33:55.473234 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-w67tx"] Oct 05 07:33:55 crc kubenswrapper[4846]: E1005 07:33:55.473599 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f724a69c-9f76-495d-8b4c-222177c6f591" containerName="extract-content" Oct 05 07:33:55 crc kubenswrapper[4846]: I1005 07:33:55.473619 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="f724a69c-9f76-495d-8b4c-222177c6f591" containerName="extract-content" Oct 05 07:33:55 crc kubenswrapper[4846]: E1005 07:33:55.473665 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f724a69c-9f76-495d-8b4c-222177c6f591" containerName="extract-utilities" Oct 05 07:33:55 crc kubenswrapper[4846]: I1005 07:33:55.473678 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="f724a69c-9f76-495d-8b4c-222177c6f591" containerName="extract-utilities" Oct 05 07:33:55 crc kubenswrapper[4846]: E1005 07:33:55.473697 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f724a69c-9f76-495d-8b4c-222177c6f591" containerName="registry-server" Oct 05 07:33:55 crc kubenswrapper[4846]: I1005 07:33:55.473706 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="f724a69c-9f76-495d-8b4c-222177c6f591" containerName="registry-server" Oct 05 07:33:55 crc kubenswrapper[4846]: I1005 07:33:55.473895 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="f724a69c-9f76-495d-8b4c-222177c6f591" containerName="registry-server" Oct 05 07:33:55 crc kubenswrapper[4846]: I1005 07:33:55.475286 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-w67tx" Oct 05 07:33:55 crc kubenswrapper[4846]: I1005 07:33:55.489421 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-w67tx"] Oct 05 07:33:55 crc kubenswrapper[4846]: I1005 07:33:55.491251 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f724a69c-9f76-495d-8b4c-222177c6f591-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f724a69c-9f76-495d-8b4c-222177c6f591" (UID: "f724a69c-9f76-495d-8b4c-222177c6f591"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 07:33:55 crc kubenswrapper[4846]: I1005 07:33:55.510012 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/afd3a686-66a9-413d-8072-af2a11c2d93a-catalog-content\") pod \"redhat-marketplace-w67tx\" (UID: \"afd3a686-66a9-413d-8072-af2a11c2d93a\") " pod="openshift-marketplace/redhat-marketplace-w67tx" Oct 05 07:33:55 crc kubenswrapper[4846]: I1005 07:33:55.510243 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/afd3a686-66a9-413d-8072-af2a11c2d93a-utilities\") pod \"redhat-marketplace-w67tx\" (UID: \"afd3a686-66a9-413d-8072-af2a11c2d93a\") " pod="openshift-marketplace/redhat-marketplace-w67tx" Oct 05 07:33:55 crc kubenswrapper[4846]: I1005 07:33:55.510548 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zpjmt\" (UniqueName: \"kubernetes.io/projected/afd3a686-66a9-413d-8072-af2a11c2d93a-kube-api-access-zpjmt\") pod \"redhat-marketplace-w67tx\" (UID: \"afd3a686-66a9-413d-8072-af2a11c2d93a\") " pod="openshift-marketplace/redhat-marketplace-w67tx" Oct 05 07:33:55 crc kubenswrapper[4846]: I1005 07:33:55.510719 4846 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f724a69c-9f76-495d-8b4c-222177c6f591-utilities\") on node \"crc\" DevicePath \"\"" Oct 05 07:33:55 crc kubenswrapper[4846]: I1005 07:33:55.510753 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcblk\" (UniqueName: \"kubernetes.io/projected/f724a69c-9f76-495d-8b4c-222177c6f591-kube-api-access-pcblk\") on node \"crc\" DevicePath \"\"" Oct 05 07:33:55 crc kubenswrapper[4846]: I1005 07:33:55.510774 4846 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f724a69c-9f76-495d-8b4c-222177c6f591-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 05 07:33:55 crc kubenswrapper[4846]: I1005 07:33:55.611698 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zpjmt\" (UniqueName: \"kubernetes.io/projected/afd3a686-66a9-413d-8072-af2a11c2d93a-kube-api-access-zpjmt\") pod \"redhat-marketplace-w67tx\" (UID: \"afd3a686-66a9-413d-8072-af2a11c2d93a\") " pod="openshift-marketplace/redhat-marketplace-w67tx" Oct 05 07:33:55 crc kubenswrapper[4846]: I1005 07:33:55.611761 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/afd3a686-66a9-413d-8072-af2a11c2d93a-catalog-content\") pod \"redhat-marketplace-w67tx\" (UID: \"afd3a686-66a9-413d-8072-af2a11c2d93a\") " pod="openshift-marketplace/redhat-marketplace-w67tx" Oct 05 07:33:55 crc kubenswrapper[4846]: I1005 07:33:55.611783 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/afd3a686-66a9-413d-8072-af2a11c2d93a-utilities\") pod \"redhat-marketplace-w67tx\" (UID: \"afd3a686-66a9-413d-8072-af2a11c2d93a\") " pod="openshift-marketplace/redhat-marketplace-w67tx" Oct 05 07:33:55 crc kubenswrapper[4846]: I1005 07:33:55.612238 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/afd3a686-66a9-413d-8072-af2a11c2d93a-utilities\") pod \"redhat-marketplace-w67tx\" (UID: \"afd3a686-66a9-413d-8072-af2a11c2d93a\") " pod="openshift-marketplace/redhat-marketplace-w67tx" Oct 05 07:33:55 crc kubenswrapper[4846]: I1005 07:33:55.612352 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/afd3a686-66a9-413d-8072-af2a11c2d93a-catalog-content\") pod \"redhat-marketplace-w67tx\" (UID: \"afd3a686-66a9-413d-8072-af2a11c2d93a\") " pod="openshift-marketplace/redhat-marketplace-w67tx" Oct 05 07:33:55 crc kubenswrapper[4846]: I1005 07:33:55.645544 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zpjmt\" (UniqueName: \"kubernetes.io/projected/afd3a686-66a9-413d-8072-af2a11c2d93a-kube-api-access-zpjmt\") pod \"redhat-marketplace-w67tx\" (UID: \"afd3a686-66a9-413d-8072-af2a11c2d93a\") " pod="openshift-marketplace/redhat-marketplace-w67tx" Oct 05 07:33:55 crc kubenswrapper[4846]: I1005 07:33:55.804437 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-w67tx" Oct 05 07:33:55 crc kubenswrapper[4846]: I1005 07:33:55.821374 4846 generic.go:334] "Generic (PLEG): container finished" podID="f724a69c-9f76-495d-8b4c-222177c6f591" containerID="d1232715265b8327f5210ff673ef3362140f3bec0d185e1b62be7a192ff15740" exitCode=0 Oct 05 07:33:55 crc kubenswrapper[4846]: I1005 07:33:55.821412 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jjtr9" event={"ID":"f724a69c-9f76-495d-8b4c-222177c6f591","Type":"ContainerDied","Data":"d1232715265b8327f5210ff673ef3362140f3bec0d185e1b62be7a192ff15740"} Oct 05 07:33:55 crc kubenswrapper[4846]: I1005 07:33:55.821440 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jjtr9" event={"ID":"f724a69c-9f76-495d-8b4c-222177c6f591","Type":"ContainerDied","Data":"c26642951bdaff30e54d596f242fca0386e0af2bfd060016d548d153edd26d89"} Oct 05 07:33:55 crc kubenswrapper[4846]: I1005 07:33:55.821457 4846 scope.go:117] "RemoveContainer" containerID="d1232715265b8327f5210ff673ef3362140f3bec0d185e1b62be7a192ff15740" Oct 05 07:33:55 crc kubenswrapper[4846]: I1005 07:33:55.821483 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jjtr9" Oct 05 07:33:55 crc kubenswrapper[4846]: I1005 07:33:55.851505 4846 scope.go:117] "RemoveContainer" containerID="44abd4b1a6f214543a32650b81193323b80c1ed3236e3bee2d0d20aa6a5b1f14" Oct 05 07:33:55 crc kubenswrapper[4846]: I1005 07:33:55.874851 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-jjtr9"] Oct 05 07:33:55 crc kubenswrapper[4846]: I1005 07:33:55.884637 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-jjtr9"] Oct 05 07:33:55 crc kubenswrapper[4846]: I1005 07:33:55.899342 4846 scope.go:117] "RemoveContainer" containerID="37ac93a7793e4706594579a6d78ff19eca3fa87a0a40182f274a8dd110f9f121" Oct 05 07:33:55 crc kubenswrapper[4846]: I1005 07:33:55.924899 4846 scope.go:117] "RemoveContainer" containerID="d1232715265b8327f5210ff673ef3362140f3bec0d185e1b62be7a192ff15740" Oct 05 07:33:55 crc kubenswrapper[4846]: E1005 07:33:55.952786 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d1232715265b8327f5210ff673ef3362140f3bec0d185e1b62be7a192ff15740\": container with ID starting with d1232715265b8327f5210ff673ef3362140f3bec0d185e1b62be7a192ff15740 not found: ID does not exist" containerID="d1232715265b8327f5210ff673ef3362140f3bec0d185e1b62be7a192ff15740" Oct 05 07:33:55 crc kubenswrapper[4846]: I1005 07:33:55.952823 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d1232715265b8327f5210ff673ef3362140f3bec0d185e1b62be7a192ff15740"} err="failed to get container status \"d1232715265b8327f5210ff673ef3362140f3bec0d185e1b62be7a192ff15740\": rpc error: code = NotFound desc = could not find container \"d1232715265b8327f5210ff673ef3362140f3bec0d185e1b62be7a192ff15740\": container with ID starting with d1232715265b8327f5210ff673ef3362140f3bec0d185e1b62be7a192ff15740 not found: ID does not exist" Oct 05 07:33:55 crc kubenswrapper[4846]: I1005 07:33:55.952849 4846 scope.go:117] "RemoveContainer" containerID="44abd4b1a6f214543a32650b81193323b80c1ed3236e3bee2d0d20aa6a5b1f14" Oct 05 07:33:55 crc kubenswrapper[4846]: E1005 07:33:55.953131 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"44abd4b1a6f214543a32650b81193323b80c1ed3236e3bee2d0d20aa6a5b1f14\": container with ID starting with 44abd4b1a6f214543a32650b81193323b80c1ed3236e3bee2d0d20aa6a5b1f14 not found: ID does not exist" containerID="44abd4b1a6f214543a32650b81193323b80c1ed3236e3bee2d0d20aa6a5b1f14" Oct 05 07:33:55 crc kubenswrapper[4846]: I1005 07:33:55.953157 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"44abd4b1a6f214543a32650b81193323b80c1ed3236e3bee2d0d20aa6a5b1f14"} err="failed to get container status \"44abd4b1a6f214543a32650b81193323b80c1ed3236e3bee2d0d20aa6a5b1f14\": rpc error: code = NotFound desc = could not find container \"44abd4b1a6f214543a32650b81193323b80c1ed3236e3bee2d0d20aa6a5b1f14\": container with ID starting with 44abd4b1a6f214543a32650b81193323b80c1ed3236e3bee2d0d20aa6a5b1f14 not found: ID does not exist" Oct 05 07:33:55 crc kubenswrapper[4846]: I1005 07:33:55.953196 4846 scope.go:117] "RemoveContainer" containerID="37ac93a7793e4706594579a6d78ff19eca3fa87a0a40182f274a8dd110f9f121" Oct 05 07:33:55 crc kubenswrapper[4846]: E1005 07:33:55.953400 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"37ac93a7793e4706594579a6d78ff19eca3fa87a0a40182f274a8dd110f9f121\": container with ID starting with 37ac93a7793e4706594579a6d78ff19eca3fa87a0a40182f274a8dd110f9f121 not found: ID does not exist" containerID="37ac93a7793e4706594579a6d78ff19eca3fa87a0a40182f274a8dd110f9f121" Oct 05 07:33:55 crc kubenswrapper[4846]: I1005 07:33:55.953421 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"37ac93a7793e4706594579a6d78ff19eca3fa87a0a40182f274a8dd110f9f121"} err="failed to get container status \"37ac93a7793e4706594579a6d78ff19eca3fa87a0a40182f274a8dd110f9f121\": rpc error: code = NotFound desc = could not find container \"37ac93a7793e4706594579a6d78ff19eca3fa87a0a40182f274a8dd110f9f121\": container with ID starting with 37ac93a7793e4706594579a6d78ff19eca3fa87a0a40182f274a8dd110f9f121 not found: ID does not exist" Oct 05 07:33:56 crc kubenswrapper[4846]: I1005 07:33:56.083940 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-w67tx"] Oct 05 07:33:56 crc kubenswrapper[4846]: W1005 07:33:56.096486 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podafd3a686_66a9_413d_8072_af2a11c2d93a.slice/crio-d194f480719c44b8eb182141b53ba83f9d59fcb9918e2b5f7a933ee6fb90063b WatchSource:0}: Error finding container d194f480719c44b8eb182141b53ba83f9d59fcb9918e2b5f7a933ee6fb90063b: Status 404 returned error can't find the container with id d194f480719c44b8eb182141b53ba83f9d59fcb9918e2b5f7a933ee6fb90063b Oct 05 07:33:56 crc kubenswrapper[4846]: I1005 07:33:56.515258 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f724a69c-9f76-495d-8b4c-222177c6f591" path="/var/lib/kubelet/pods/f724a69c-9f76-495d-8b4c-222177c6f591/volumes" Oct 05 07:33:56 crc kubenswrapper[4846]: I1005 07:33:56.836927 4846 generic.go:334] "Generic (PLEG): container finished" podID="afd3a686-66a9-413d-8072-af2a11c2d93a" containerID="cdc7cdee32340b83ee0cee1b4b1c1dc4b421109307f7016cfddf0ce7a94365b2" exitCode=0 Oct 05 07:33:56 crc kubenswrapper[4846]: I1005 07:33:56.837004 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-w67tx" event={"ID":"afd3a686-66a9-413d-8072-af2a11c2d93a","Type":"ContainerDied","Data":"cdc7cdee32340b83ee0cee1b4b1c1dc4b421109307f7016cfddf0ce7a94365b2"} Oct 05 07:33:56 crc kubenswrapper[4846]: I1005 07:33:56.837035 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-w67tx" event={"ID":"afd3a686-66a9-413d-8072-af2a11c2d93a","Type":"ContainerStarted","Data":"d194f480719c44b8eb182141b53ba83f9d59fcb9918e2b5f7a933ee6fb90063b"} Oct 05 07:33:57 crc kubenswrapper[4846]: I1005 07:33:57.848025 4846 generic.go:334] "Generic (PLEG): container finished" podID="afd3a686-66a9-413d-8072-af2a11c2d93a" containerID="2a15f92b9b7f5de01e78bebcf46de2f3f6346a52dd56a44a7ef173326fb221fc" exitCode=0 Oct 05 07:33:57 crc kubenswrapper[4846]: I1005 07:33:57.848241 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-w67tx" event={"ID":"afd3a686-66a9-413d-8072-af2a11c2d93a","Type":"ContainerDied","Data":"2a15f92b9b7f5de01e78bebcf46de2f3f6346a52dd56a44a7ef173326fb221fc"} Oct 05 07:33:58 crc kubenswrapper[4846]: I1005 07:33:58.864049 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-w67tx" event={"ID":"afd3a686-66a9-413d-8072-af2a11c2d93a","Type":"ContainerStarted","Data":"6ef063a22320ce788750909c7c5c6a7a5d6a5d69a632d19e2df771fe4d700e9f"} Oct 05 07:33:58 crc kubenswrapper[4846]: I1005 07:33:58.899818 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-w67tx" podStartSLOduration=2.368069061 podStartE2EDuration="3.899784086s" podCreationTimestamp="2025-10-05 07:33:55 +0000 UTC" firstStartedPulling="2025-10-05 07:33:56.840251557 +0000 UTC m=+2759.081104342" lastFinishedPulling="2025-10-05 07:33:58.371966572 +0000 UTC m=+2760.612819367" observedRunningTime="2025-10-05 07:33:58.897958618 +0000 UTC m=+2761.138811433" watchObservedRunningTime="2025-10-05 07:33:58.899784086 +0000 UTC m=+2761.140636901" Oct 05 07:34:05 crc kubenswrapper[4846]: I1005 07:34:05.805537 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-w67tx" Oct 05 07:34:05 crc kubenswrapper[4846]: I1005 07:34:05.808044 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-w67tx" Oct 05 07:34:05 crc kubenswrapper[4846]: I1005 07:34:05.890990 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-w67tx" Oct 05 07:34:06 crc kubenswrapper[4846]: I1005 07:34:06.006678 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-w67tx" Oct 05 07:34:06 crc kubenswrapper[4846]: I1005 07:34:06.279225 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-w67tx"] Oct 05 07:34:07 crc kubenswrapper[4846]: I1005 07:34:07.946700 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-w67tx" podUID="afd3a686-66a9-413d-8072-af2a11c2d93a" containerName="registry-server" containerID="cri-o://6ef063a22320ce788750909c7c5c6a7a5d6a5d69a632d19e2df771fe4d700e9f" gracePeriod=2 Oct 05 07:34:08 crc kubenswrapper[4846]: I1005 07:34:08.434823 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-w67tx" Oct 05 07:34:08 crc kubenswrapper[4846]: I1005 07:34:08.612363 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/afd3a686-66a9-413d-8072-af2a11c2d93a-utilities\") pod \"afd3a686-66a9-413d-8072-af2a11c2d93a\" (UID: \"afd3a686-66a9-413d-8072-af2a11c2d93a\") " Oct 05 07:34:08 crc kubenswrapper[4846]: I1005 07:34:08.612483 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zpjmt\" (UniqueName: \"kubernetes.io/projected/afd3a686-66a9-413d-8072-af2a11c2d93a-kube-api-access-zpjmt\") pod \"afd3a686-66a9-413d-8072-af2a11c2d93a\" (UID: \"afd3a686-66a9-413d-8072-af2a11c2d93a\") " Oct 05 07:34:08 crc kubenswrapper[4846]: I1005 07:34:08.612749 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/afd3a686-66a9-413d-8072-af2a11c2d93a-catalog-content\") pod \"afd3a686-66a9-413d-8072-af2a11c2d93a\" (UID: \"afd3a686-66a9-413d-8072-af2a11c2d93a\") " Oct 05 07:34:08 crc kubenswrapper[4846]: I1005 07:34:08.615016 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/afd3a686-66a9-413d-8072-af2a11c2d93a-utilities" (OuterVolumeSpecName: "utilities") pod "afd3a686-66a9-413d-8072-af2a11c2d93a" (UID: "afd3a686-66a9-413d-8072-af2a11c2d93a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 07:34:08 crc kubenswrapper[4846]: I1005 07:34:08.622009 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/afd3a686-66a9-413d-8072-af2a11c2d93a-kube-api-access-zpjmt" (OuterVolumeSpecName: "kube-api-access-zpjmt") pod "afd3a686-66a9-413d-8072-af2a11c2d93a" (UID: "afd3a686-66a9-413d-8072-af2a11c2d93a"). InnerVolumeSpecName "kube-api-access-zpjmt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:34:08 crc kubenswrapper[4846]: I1005 07:34:08.637465 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/afd3a686-66a9-413d-8072-af2a11c2d93a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "afd3a686-66a9-413d-8072-af2a11c2d93a" (UID: "afd3a686-66a9-413d-8072-af2a11c2d93a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 07:34:08 crc kubenswrapper[4846]: I1005 07:34:08.714658 4846 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/afd3a686-66a9-413d-8072-af2a11c2d93a-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 05 07:34:08 crc kubenswrapper[4846]: I1005 07:34:08.714745 4846 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/afd3a686-66a9-413d-8072-af2a11c2d93a-utilities\") on node \"crc\" DevicePath \"\"" Oct 05 07:34:08 crc kubenswrapper[4846]: I1005 07:34:08.714776 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zpjmt\" (UniqueName: \"kubernetes.io/projected/afd3a686-66a9-413d-8072-af2a11c2d93a-kube-api-access-zpjmt\") on node \"crc\" DevicePath \"\"" Oct 05 07:34:08 crc kubenswrapper[4846]: I1005 07:34:08.956727 4846 generic.go:334] "Generic (PLEG): container finished" podID="afd3a686-66a9-413d-8072-af2a11c2d93a" containerID="6ef063a22320ce788750909c7c5c6a7a5d6a5d69a632d19e2df771fe4d700e9f" exitCode=0 Oct 05 07:34:08 crc kubenswrapper[4846]: I1005 07:34:08.956793 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-w67tx" event={"ID":"afd3a686-66a9-413d-8072-af2a11c2d93a","Type":"ContainerDied","Data":"6ef063a22320ce788750909c7c5c6a7a5d6a5d69a632d19e2df771fe4d700e9f"} Oct 05 07:34:08 crc kubenswrapper[4846]: I1005 07:34:08.956835 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-w67tx" event={"ID":"afd3a686-66a9-413d-8072-af2a11c2d93a","Type":"ContainerDied","Data":"d194f480719c44b8eb182141b53ba83f9d59fcb9918e2b5f7a933ee6fb90063b"} Oct 05 07:34:08 crc kubenswrapper[4846]: I1005 07:34:08.956864 4846 scope.go:117] "RemoveContainer" containerID="6ef063a22320ce788750909c7c5c6a7a5d6a5d69a632d19e2df771fe4d700e9f" Oct 05 07:34:08 crc kubenswrapper[4846]: I1005 07:34:08.956878 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-w67tx" Oct 05 07:34:08 crc kubenswrapper[4846]: I1005 07:34:08.976745 4846 scope.go:117] "RemoveContainer" containerID="2a15f92b9b7f5de01e78bebcf46de2f3f6346a52dd56a44a7ef173326fb221fc" Oct 05 07:34:09 crc kubenswrapper[4846]: I1005 07:34:09.016836 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-w67tx"] Oct 05 07:34:09 crc kubenswrapper[4846]: I1005 07:34:09.022385 4846 scope.go:117] "RemoveContainer" containerID="cdc7cdee32340b83ee0cee1b4b1c1dc4b421109307f7016cfddf0ce7a94365b2" Oct 05 07:34:09 crc kubenswrapper[4846]: I1005 07:34:09.025163 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-w67tx"] Oct 05 07:34:09 crc kubenswrapper[4846]: I1005 07:34:09.050946 4846 scope.go:117] "RemoveContainer" containerID="6ef063a22320ce788750909c7c5c6a7a5d6a5d69a632d19e2df771fe4d700e9f" Oct 05 07:34:09 crc kubenswrapper[4846]: E1005 07:34:09.051611 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6ef063a22320ce788750909c7c5c6a7a5d6a5d69a632d19e2df771fe4d700e9f\": container with ID starting with 6ef063a22320ce788750909c7c5c6a7a5d6a5d69a632d19e2df771fe4d700e9f not found: ID does not exist" containerID="6ef063a22320ce788750909c7c5c6a7a5d6a5d69a632d19e2df771fe4d700e9f" Oct 05 07:34:09 crc kubenswrapper[4846]: I1005 07:34:09.051669 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6ef063a22320ce788750909c7c5c6a7a5d6a5d69a632d19e2df771fe4d700e9f"} err="failed to get container status \"6ef063a22320ce788750909c7c5c6a7a5d6a5d69a632d19e2df771fe4d700e9f\": rpc error: code = NotFound desc = could not find container \"6ef063a22320ce788750909c7c5c6a7a5d6a5d69a632d19e2df771fe4d700e9f\": container with ID starting with 6ef063a22320ce788750909c7c5c6a7a5d6a5d69a632d19e2df771fe4d700e9f not found: ID does not exist" Oct 05 07:34:09 crc kubenswrapper[4846]: I1005 07:34:09.051708 4846 scope.go:117] "RemoveContainer" containerID="2a15f92b9b7f5de01e78bebcf46de2f3f6346a52dd56a44a7ef173326fb221fc" Oct 05 07:34:09 crc kubenswrapper[4846]: E1005 07:34:09.052049 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2a15f92b9b7f5de01e78bebcf46de2f3f6346a52dd56a44a7ef173326fb221fc\": container with ID starting with 2a15f92b9b7f5de01e78bebcf46de2f3f6346a52dd56a44a7ef173326fb221fc not found: ID does not exist" containerID="2a15f92b9b7f5de01e78bebcf46de2f3f6346a52dd56a44a7ef173326fb221fc" Oct 05 07:34:09 crc kubenswrapper[4846]: I1005 07:34:09.052200 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2a15f92b9b7f5de01e78bebcf46de2f3f6346a52dd56a44a7ef173326fb221fc"} err="failed to get container status \"2a15f92b9b7f5de01e78bebcf46de2f3f6346a52dd56a44a7ef173326fb221fc\": rpc error: code = NotFound desc = could not find container \"2a15f92b9b7f5de01e78bebcf46de2f3f6346a52dd56a44a7ef173326fb221fc\": container with ID starting with 2a15f92b9b7f5de01e78bebcf46de2f3f6346a52dd56a44a7ef173326fb221fc not found: ID does not exist" Oct 05 07:34:09 crc kubenswrapper[4846]: I1005 07:34:09.052323 4846 scope.go:117] "RemoveContainer" containerID="cdc7cdee32340b83ee0cee1b4b1c1dc4b421109307f7016cfddf0ce7a94365b2" Oct 05 07:34:09 crc kubenswrapper[4846]: E1005 07:34:09.052790 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cdc7cdee32340b83ee0cee1b4b1c1dc4b421109307f7016cfddf0ce7a94365b2\": container with ID starting with cdc7cdee32340b83ee0cee1b4b1c1dc4b421109307f7016cfddf0ce7a94365b2 not found: ID does not exist" containerID="cdc7cdee32340b83ee0cee1b4b1c1dc4b421109307f7016cfddf0ce7a94365b2" Oct 05 07:34:09 crc kubenswrapper[4846]: I1005 07:34:09.052866 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cdc7cdee32340b83ee0cee1b4b1c1dc4b421109307f7016cfddf0ce7a94365b2"} err="failed to get container status \"cdc7cdee32340b83ee0cee1b4b1c1dc4b421109307f7016cfddf0ce7a94365b2\": rpc error: code = NotFound desc = could not find container \"cdc7cdee32340b83ee0cee1b4b1c1dc4b421109307f7016cfddf0ce7a94365b2\": container with ID starting with cdc7cdee32340b83ee0cee1b4b1c1dc4b421109307f7016cfddf0ce7a94365b2 not found: ID does not exist" Oct 05 07:34:10 crc kubenswrapper[4846]: I1005 07:34:10.514075 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="afd3a686-66a9-413d-8072-af2a11c2d93a" path="/var/lib/kubelet/pods/afd3a686-66a9-413d-8072-af2a11c2d93a/volumes" Oct 05 07:34:51 crc kubenswrapper[4846]: I1005 07:34:51.098745 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-qqh5k"] Oct 05 07:34:51 crc kubenswrapper[4846]: E1005 07:34:51.099807 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="afd3a686-66a9-413d-8072-af2a11c2d93a" containerName="registry-server" Oct 05 07:34:51 crc kubenswrapper[4846]: I1005 07:34:51.099823 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="afd3a686-66a9-413d-8072-af2a11c2d93a" containerName="registry-server" Oct 05 07:34:51 crc kubenswrapper[4846]: E1005 07:34:51.099844 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="afd3a686-66a9-413d-8072-af2a11c2d93a" containerName="extract-utilities" Oct 05 07:34:51 crc kubenswrapper[4846]: I1005 07:34:51.099852 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="afd3a686-66a9-413d-8072-af2a11c2d93a" containerName="extract-utilities" Oct 05 07:34:51 crc kubenswrapper[4846]: E1005 07:34:51.099870 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="afd3a686-66a9-413d-8072-af2a11c2d93a" containerName="extract-content" Oct 05 07:34:51 crc kubenswrapper[4846]: I1005 07:34:51.099878 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="afd3a686-66a9-413d-8072-af2a11c2d93a" containerName="extract-content" Oct 05 07:34:51 crc kubenswrapper[4846]: I1005 07:34:51.100078 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="afd3a686-66a9-413d-8072-af2a11c2d93a" containerName="registry-server" Oct 05 07:34:51 crc kubenswrapper[4846]: I1005 07:34:51.106242 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qqh5k" Oct 05 07:34:51 crc kubenswrapper[4846]: I1005 07:34:51.139546 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/130b5769-e7e3-422d-8540-26e044e3500e-catalog-content\") pod \"redhat-operators-qqh5k\" (UID: \"130b5769-e7e3-422d-8540-26e044e3500e\") " pod="openshift-marketplace/redhat-operators-qqh5k" Oct 05 07:34:51 crc kubenswrapper[4846]: I1005 07:34:51.139625 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g5mxg\" (UniqueName: \"kubernetes.io/projected/130b5769-e7e3-422d-8540-26e044e3500e-kube-api-access-g5mxg\") pod \"redhat-operators-qqh5k\" (UID: \"130b5769-e7e3-422d-8540-26e044e3500e\") " pod="openshift-marketplace/redhat-operators-qqh5k" Oct 05 07:34:51 crc kubenswrapper[4846]: I1005 07:34:51.139853 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/130b5769-e7e3-422d-8540-26e044e3500e-utilities\") pod \"redhat-operators-qqh5k\" (UID: \"130b5769-e7e3-422d-8540-26e044e3500e\") " pod="openshift-marketplace/redhat-operators-qqh5k" Oct 05 07:34:51 crc kubenswrapper[4846]: I1005 07:34:51.151633 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-qqh5k"] Oct 05 07:34:51 crc kubenswrapper[4846]: I1005 07:34:51.240721 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g5mxg\" (UniqueName: \"kubernetes.io/projected/130b5769-e7e3-422d-8540-26e044e3500e-kube-api-access-g5mxg\") pod \"redhat-operators-qqh5k\" (UID: \"130b5769-e7e3-422d-8540-26e044e3500e\") " pod="openshift-marketplace/redhat-operators-qqh5k" Oct 05 07:34:51 crc kubenswrapper[4846]: I1005 07:34:51.240781 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/130b5769-e7e3-422d-8540-26e044e3500e-utilities\") pod \"redhat-operators-qqh5k\" (UID: \"130b5769-e7e3-422d-8540-26e044e3500e\") " pod="openshift-marketplace/redhat-operators-qqh5k" Oct 05 07:34:51 crc kubenswrapper[4846]: I1005 07:34:51.240856 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/130b5769-e7e3-422d-8540-26e044e3500e-catalog-content\") pod \"redhat-operators-qqh5k\" (UID: \"130b5769-e7e3-422d-8540-26e044e3500e\") " pod="openshift-marketplace/redhat-operators-qqh5k" Oct 05 07:34:51 crc kubenswrapper[4846]: I1005 07:34:51.241436 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/130b5769-e7e3-422d-8540-26e044e3500e-catalog-content\") pod \"redhat-operators-qqh5k\" (UID: \"130b5769-e7e3-422d-8540-26e044e3500e\") " pod="openshift-marketplace/redhat-operators-qqh5k" Oct 05 07:34:51 crc kubenswrapper[4846]: I1005 07:34:51.241672 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/130b5769-e7e3-422d-8540-26e044e3500e-utilities\") pod \"redhat-operators-qqh5k\" (UID: \"130b5769-e7e3-422d-8540-26e044e3500e\") " pod="openshift-marketplace/redhat-operators-qqh5k" Oct 05 07:34:51 crc kubenswrapper[4846]: I1005 07:34:51.267338 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g5mxg\" (UniqueName: \"kubernetes.io/projected/130b5769-e7e3-422d-8540-26e044e3500e-kube-api-access-g5mxg\") pod \"redhat-operators-qqh5k\" (UID: \"130b5769-e7e3-422d-8540-26e044e3500e\") " pod="openshift-marketplace/redhat-operators-qqh5k" Oct 05 07:34:51 crc kubenswrapper[4846]: I1005 07:34:51.451254 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qqh5k" Oct 05 07:34:51 crc kubenswrapper[4846]: I1005 07:34:51.919568 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-qqh5k"] Oct 05 07:34:51 crc kubenswrapper[4846]: W1005 07:34:51.926968 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod130b5769_e7e3_422d_8540_26e044e3500e.slice/crio-43e60c6c07a1af478f5df406105b0a47117efc204c0ab40033d86f43b706742b WatchSource:0}: Error finding container 43e60c6c07a1af478f5df406105b0a47117efc204c0ab40033d86f43b706742b: Status 404 returned error can't find the container with id 43e60c6c07a1af478f5df406105b0a47117efc204c0ab40033d86f43b706742b Oct 05 07:34:52 crc kubenswrapper[4846]: I1005 07:34:52.420106 4846 generic.go:334] "Generic (PLEG): container finished" podID="130b5769-e7e3-422d-8540-26e044e3500e" containerID="faf7839283bf07e8e315d1dfd254923e7b84924d76ed6f995b626748327e6a7e" exitCode=0 Oct 05 07:34:52 crc kubenswrapper[4846]: I1005 07:34:52.420525 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qqh5k" event={"ID":"130b5769-e7e3-422d-8540-26e044e3500e","Type":"ContainerDied","Data":"faf7839283bf07e8e315d1dfd254923e7b84924d76ed6f995b626748327e6a7e"} Oct 05 07:34:52 crc kubenswrapper[4846]: I1005 07:34:52.420969 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qqh5k" event={"ID":"130b5769-e7e3-422d-8540-26e044e3500e","Type":"ContainerStarted","Data":"43e60c6c07a1af478f5df406105b0a47117efc204c0ab40033d86f43b706742b"} Oct 05 07:34:53 crc kubenswrapper[4846]: I1005 07:34:53.433311 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qqh5k" event={"ID":"130b5769-e7e3-422d-8540-26e044e3500e","Type":"ContainerStarted","Data":"d62e8f6a2f5d1ccc127a8f5d19eb8289aed6f09bdfaf5f281442d502f1759177"} Oct 05 07:34:54 crc kubenswrapper[4846]: I1005 07:34:54.447640 4846 generic.go:334] "Generic (PLEG): container finished" podID="130b5769-e7e3-422d-8540-26e044e3500e" containerID="d62e8f6a2f5d1ccc127a8f5d19eb8289aed6f09bdfaf5f281442d502f1759177" exitCode=0 Oct 05 07:34:54 crc kubenswrapper[4846]: I1005 07:34:54.447754 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qqh5k" event={"ID":"130b5769-e7e3-422d-8540-26e044e3500e","Type":"ContainerDied","Data":"d62e8f6a2f5d1ccc127a8f5d19eb8289aed6f09bdfaf5f281442d502f1759177"} Oct 05 07:34:55 crc kubenswrapper[4846]: I1005 07:34:55.458436 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qqh5k" event={"ID":"130b5769-e7e3-422d-8540-26e044e3500e","Type":"ContainerStarted","Data":"fd678423600418231bcaa9f5e7953c99b5947350f919b2102f8321d9e58559cd"} Oct 05 07:35:01 crc kubenswrapper[4846]: I1005 07:35:01.451628 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-qqh5k" Oct 05 07:35:01 crc kubenswrapper[4846]: I1005 07:35:01.452493 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-qqh5k" Oct 05 07:35:01 crc kubenswrapper[4846]: I1005 07:35:01.527442 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-qqh5k" Oct 05 07:35:01 crc kubenswrapper[4846]: I1005 07:35:01.567298 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-qqh5k" podStartSLOduration=8.081407157 podStartE2EDuration="10.567263644s" podCreationTimestamp="2025-10-05 07:34:51 +0000 UTC" firstStartedPulling="2025-10-05 07:34:52.422385899 +0000 UTC m=+2814.663238704" lastFinishedPulling="2025-10-05 07:34:54.908242376 +0000 UTC m=+2817.149095191" observedRunningTime="2025-10-05 07:34:55.485642254 +0000 UTC m=+2817.726495039" watchObservedRunningTime="2025-10-05 07:35:01.567263644 +0000 UTC m=+2823.808116449" Oct 05 07:35:01 crc kubenswrapper[4846]: I1005 07:35:01.593094 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-qqh5k" Oct 05 07:35:01 crc kubenswrapper[4846]: I1005 07:35:01.774769 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-qqh5k"] Oct 05 07:35:03 crc kubenswrapper[4846]: I1005 07:35:03.532247 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-qqh5k" podUID="130b5769-e7e3-422d-8540-26e044e3500e" containerName="registry-server" containerID="cri-o://fd678423600418231bcaa9f5e7953c99b5947350f919b2102f8321d9e58559cd" gracePeriod=2 Oct 05 07:35:03 crc kubenswrapper[4846]: I1005 07:35:03.940123 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qqh5k" Oct 05 07:35:04 crc kubenswrapper[4846]: I1005 07:35:04.048594 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/130b5769-e7e3-422d-8540-26e044e3500e-utilities\") pod \"130b5769-e7e3-422d-8540-26e044e3500e\" (UID: \"130b5769-e7e3-422d-8540-26e044e3500e\") " Oct 05 07:35:04 crc kubenswrapper[4846]: I1005 07:35:04.048658 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/130b5769-e7e3-422d-8540-26e044e3500e-catalog-content\") pod \"130b5769-e7e3-422d-8540-26e044e3500e\" (UID: \"130b5769-e7e3-422d-8540-26e044e3500e\") " Oct 05 07:35:04 crc kubenswrapper[4846]: I1005 07:35:04.049619 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/130b5769-e7e3-422d-8540-26e044e3500e-utilities" (OuterVolumeSpecName: "utilities") pod "130b5769-e7e3-422d-8540-26e044e3500e" (UID: "130b5769-e7e3-422d-8540-26e044e3500e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 07:35:04 crc kubenswrapper[4846]: I1005 07:35:04.052610 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g5mxg\" (UniqueName: \"kubernetes.io/projected/130b5769-e7e3-422d-8540-26e044e3500e-kube-api-access-g5mxg\") pod \"130b5769-e7e3-422d-8540-26e044e3500e\" (UID: \"130b5769-e7e3-422d-8540-26e044e3500e\") " Oct 05 07:35:04 crc kubenswrapper[4846]: I1005 07:35:04.053214 4846 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/130b5769-e7e3-422d-8540-26e044e3500e-utilities\") on node \"crc\" DevicePath \"\"" Oct 05 07:35:04 crc kubenswrapper[4846]: I1005 07:35:04.059254 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/130b5769-e7e3-422d-8540-26e044e3500e-kube-api-access-g5mxg" (OuterVolumeSpecName: "kube-api-access-g5mxg") pod "130b5769-e7e3-422d-8540-26e044e3500e" (UID: "130b5769-e7e3-422d-8540-26e044e3500e"). InnerVolumeSpecName "kube-api-access-g5mxg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:35:04 crc kubenswrapper[4846]: I1005 07:35:04.154435 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g5mxg\" (UniqueName: \"kubernetes.io/projected/130b5769-e7e3-422d-8540-26e044e3500e-kube-api-access-g5mxg\") on node \"crc\" DevicePath \"\"" Oct 05 07:35:04 crc kubenswrapper[4846]: I1005 07:35:04.543016 4846 generic.go:334] "Generic (PLEG): container finished" podID="130b5769-e7e3-422d-8540-26e044e3500e" containerID="fd678423600418231bcaa9f5e7953c99b5947350f919b2102f8321d9e58559cd" exitCode=0 Oct 05 07:35:04 crc kubenswrapper[4846]: I1005 07:35:04.543067 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qqh5k" event={"ID":"130b5769-e7e3-422d-8540-26e044e3500e","Type":"ContainerDied","Data":"fd678423600418231bcaa9f5e7953c99b5947350f919b2102f8321d9e58559cd"} Oct 05 07:35:04 crc kubenswrapper[4846]: I1005 07:35:04.543098 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qqh5k" event={"ID":"130b5769-e7e3-422d-8540-26e044e3500e","Type":"ContainerDied","Data":"43e60c6c07a1af478f5df406105b0a47117efc204c0ab40033d86f43b706742b"} Oct 05 07:35:04 crc kubenswrapper[4846]: I1005 07:35:04.543125 4846 scope.go:117] "RemoveContainer" containerID="fd678423600418231bcaa9f5e7953c99b5947350f919b2102f8321d9e58559cd" Oct 05 07:35:04 crc kubenswrapper[4846]: I1005 07:35:04.543280 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qqh5k" Oct 05 07:35:04 crc kubenswrapper[4846]: I1005 07:35:04.579109 4846 scope.go:117] "RemoveContainer" containerID="d62e8f6a2f5d1ccc127a8f5d19eb8289aed6f09bdfaf5f281442d502f1759177" Oct 05 07:35:04 crc kubenswrapper[4846]: I1005 07:35:04.610634 4846 scope.go:117] "RemoveContainer" containerID="faf7839283bf07e8e315d1dfd254923e7b84924d76ed6f995b626748327e6a7e" Oct 05 07:35:04 crc kubenswrapper[4846]: I1005 07:35:04.639082 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/130b5769-e7e3-422d-8540-26e044e3500e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "130b5769-e7e3-422d-8540-26e044e3500e" (UID: "130b5769-e7e3-422d-8540-26e044e3500e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 07:35:04 crc kubenswrapper[4846]: I1005 07:35:04.641881 4846 scope.go:117] "RemoveContainer" containerID="fd678423600418231bcaa9f5e7953c99b5947350f919b2102f8321d9e58559cd" Oct 05 07:35:04 crc kubenswrapper[4846]: E1005 07:35:04.642544 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fd678423600418231bcaa9f5e7953c99b5947350f919b2102f8321d9e58559cd\": container with ID starting with fd678423600418231bcaa9f5e7953c99b5947350f919b2102f8321d9e58559cd not found: ID does not exist" containerID="fd678423600418231bcaa9f5e7953c99b5947350f919b2102f8321d9e58559cd" Oct 05 07:35:04 crc kubenswrapper[4846]: I1005 07:35:04.642620 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fd678423600418231bcaa9f5e7953c99b5947350f919b2102f8321d9e58559cd"} err="failed to get container status \"fd678423600418231bcaa9f5e7953c99b5947350f919b2102f8321d9e58559cd\": rpc error: code = NotFound desc = could not find container \"fd678423600418231bcaa9f5e7953c99b5947350f919b2102f8321d9e58559cd\": container with ID starting with fd678423600418231bcaa9f5e7953c99b5947350f919b2102f8321d9e58559cd not found: ID does not exist" Oct 05 07:35:04 crc kubenswrapper[4846]: I1005 07:35:04.642660 4846 scope.go:117] "RemoveContainer" containerID="d62e8f6a2f5d1ccc127a8f5d19eb8289aed6f09bdfaf5f281442d502f1759177" Oct 05 07:35:04 crc kubenswrapper[4846]: E1005 07:35:04.643033 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d62e8f6a2f5d1ccc127a8f5d19eb8289aed6f09bdfaf5f281442d502f1759177\": container with ID starting with d62e8f6a2f5d1ccc127a8f5d19eb8289aed6f09bdfaf5f281442d502f1759177 not found: ID does not exist" containerID="d62e8f6a2f5d1ccc127a8f5d19eb8289aed6f09bdfaf5f281442d502f1759177" Oct 05 07:35:04 crc kubenswrapper[4846]: I1005 07:35:04.643069 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d62e8f6a2f5d1ccc127a8f5d19eb8289aed6f09bdfaf5f281442d502f1759177"} err="failed to get container status \"d62e8f6a2f5d1ccc127a8f5d19eb8289aed6f09bdfaf5f281442d502f1759177\": rpc error: code = NotFound desc = could not find container \"d62e8f6a2f5d1ccc127a8f5d19eb8289aed6f09bdfaf5f281442d502f1759177\": container with ID starting with d62e8f6a2f5d1ccc127a8f5d19eb8289aed6f09bdfaf5f281442d502f1759177 not found: ID does not exist" Oct 05 07:35:04 crc kubenswrapper[4846]: I1005 07:35:04.643089 4846 scope.go:117] "RemoveContainer" containerID="faf7839283bf07e8e315d1dfd254923e7b84924d76ed6f995b626748327e6a7e" Oct 05 07:35:04 crc kubenswrapper[4846]: E1005 07:35:04.644434 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"faf7839283bf07e8e315d1dfd254923e7b84924d76ed6f995b626748327e6a7e\": container with ID starting with faf7839283bf07e8e315d1dfd254923e7b84924d76ed6f995b626748327e6a7e not found: ID does not exist" containerID="faf7839283bf07e8e315d1dfd254923e7b84924d76ed6f995b626748327e6a7e" Oct 05 07:35:04 crc kubenswrapper[4846]: I1005 07:35:04.644493 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"faf7839283bf07e8e315d1dfd254923e7b84924d76ed6f995b626748327e6a7e"} err="failed to get container status \"faf7839283bf07e8e315d1dfd254923e7b84924d76ed6f995b626748327e6a7e\": rpc error: code = NotFound desc = could not find container \"faf7839283bf07e8e315d1dfd254923e7b84924d76ed6f995b626748327e6a7e\": container with ID starting with faf7839283bf07e8e315d1dfd254923e7b84924d76ed6f995b626748327e6a7e not found: ID does not exist" Oct 05 07:35:04 crc kubenswrapper[4846]: I1005 07:35:04.662000 4846 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/130b5769-e7e3-422d-8540-26e044e3500e-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 05 07:35:04 crc kubenswrapper[4846]: I1005 07:35:04.892213 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-qqh5k"] Oct 05 07:35:04 crc kubenswrapper[4846]: I1005 07:35:04.902160 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-qqh5k"] Oct 05 07:35:06 crc kubenswrapper[4846]: I1005 07:35:06.507347 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="130b5769-e7e3-422d-8540-26e044e3500e" path="/var/lib/kubelet/pods/130b5769-e7e3-422d-8540-26e044e3500e/volumes" Oct 05 07:35:23 crc kubenswrapper[4846]: I1005 07:35:23.325417 4846 patch_prober.go:28] interesting pod/machine-config-daemon-fscvf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 07:35:23 crc kubenswrapper[4846]: I1005 07:35:23.326039 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 07:35:26 crc kubenswrapper[4846]: I1005 07:35:26.075046 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-7fc4h"] Oct 05 07:35:26 crc kubenswrapper[4846]: E1005 07:35:26.075681 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="130b5769-e7e3-422d-8540-26e044e3500e" containerName="extract-content" Oct 05 07:35:26 crc kubenswrapper[4846]: I1005 07:35:26.075696 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="130b5769-e7e3-422d-8540-26e044e3500e" containerName="extract-content" Oct 05 07:35:26 crc kubenswrapper[4846]: E1005 07:35:26.075707 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="130b5769-e7e3-422d-8540-26e044e3500e" containerName="registry-server" Oct 05 07:35:26 crc kubenswrapper[4846]: I1005 07:35:26.075714 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="130b5769-e7e3-422d-8540-26e044e3500e" containerName="registry-server" Oct 05 07:35:26 crc kubenswrapper[4846]: E1005 07:35:26.075747 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="130b5769-e7e3-422d-8540-26e044e3500e" containerName="extract-utilities" Oct 05 07:35:26 crc kubenswrapper[4846]: I1005 07:35:26.075757 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="130b5769-e7e3-422d-8540-26e044e3500e" containerName="extract-utilities" Oct 05 07:35:26 crc kubenswrapper[4846]: I1005 07:35:26.075918 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="130b5769-e7e3-422d-8540-26e044e3500e" containerName="registry-server" Oct 05 07:35:26 crc kubenswrapper[4846]: I1005 07:35:26.077174 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7fc4h" Oct 05 07:35:26 crc kubenswrapper[4846]: I1005 07:35:26.094111 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-7fc4h"] Oct 05 07:35:26 crc kubenswrapper[4846]: I1005 07:35:26.230954 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f11d6823-4622-49bf-aa2a-41fcd37b7f96-catalog-content\") pod \"community-operators-7fc4h\" (UID: \"f11d6823-4622-49bf-aa2a-41fcd37b7f96\") " pod="openshift-marketplace/community-operators-7fc4h" Oct 05 07:35:26 crc kubenswrapper[4846]: I1005 07:35:26.231142 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dws7g\" (UniqueName: \"kubernetes.io/projected/f11d6823-4622-49bf-aa2a-41fcd37b7f96-kube-api-access-dws7g\") pod \"community-operators-7fc4h\" (UID: \"f11d6823-4622-49bf-aa2a-41fcd37b7f96\") " pod="openshift-marketplace/community-operators-7fc4h" Oct 05 07:35:26 crc kubenswrapper[4846]: I1005 07:35:26.231487 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f11d6823-4622-49bf-aa2a-41fcd37b7f96-utilities\") pod \"community-operators-7fc4h\" (UID: \"f11d6823-4622-49bf-aa2a-41fcd37b7f96\") " pod="openshift-marketplace/community-operators-7fc4h" Oct 05 07:35:26 crc kubenswrapper[4846]: I1005 07:35:26.333335 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f11d6823-4622-49bf-aa2a-41fcd37b7f96-utilities\") pod \"community-operators-7fc4h\" (UID: \"f11d6823-4622-49bf-aa2a-41fcd37b7f96\") " pod="openshift-marketplace/community-operators-7fc4h" Oct 05 07:35:26 crc kubenswrapper[4846]: I1005 07:35:26.333660 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f11d6823-4622-49bf-aa2a-41fcd37b7f96-catalog-content\") pod \"community-operators-7fc4h\" (UID: \"f11d6823-4622-49bf-aa2a-41fcd37b7f96\") " pod="openshift-marketplace/community-operators-7fc4h" Oct 05 07:35:26 crc kubenswrapper[4846]: I1005 07:35:26.333806 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dws7g\" (UniqueName: \"kubernetes.io/projected/f11d6823-4622-49bf-aa2a-41fcd37b7f96-kube-api-access-dws7g\") pod \"community-operators-7fc4h\" (UID: \"f11d6823-4622-49bf-aa2a-41fcd37b7f96\") " pod="openshift-marketplace/community-operators-7fc4h" Oct 05 07:35:26 crc kubenswrapper[4846]: I1005 07:35:26.333880 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f11d6823-4622-49bf-aa2a-41fcd37b7f96-utilities\") pod \"community-operators-7fc4h\" (UID: \"f11d6823-4622-49bf-aa2a-41fcd37b7f96\") " pod="openshift-marketplace/community-operators-7fc4h" Oct 05 07:35:26 crc kubenswrapper[4846]: I1005 07:35:26.334163 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f11d6823-4622-49bf-aa2a-41fcd37b7f96-catalog-content\") pod \"community-operators-7fc4h\" (UID: \"f11d6823-4622-49bf-aa2a-41fcd37b7f96\") " pod="openshift-marketplace/community-operators-7fc4h" Oct 05 07:35:26 crc kubenswrapper[4846]: I1005 07:35:26.356003 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dws7g\" (UniqueName: \"kubernetes.io/projected/f11d6823-4622-49bf-aa2a-41fcd37b7f96-kube-api-access-dws7g\") pod \"community-operators-7fc4h\" (UID: \"f11d6823-4622-49bf-aa2a-41fcd37b7f96\") " pod="openshift-marketplace/community-operators-7fc4h" Oct 05 07:35:26 crc kubenswrapper[4846]: I1005 07:35:26.414013 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7fc4h" Oct 05 07:35:26 crc kubenswrapper[4846]: I1005 07:35:26.952130 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-7fc4h"] Oct 05 07:35:27 crc kubenswrapper[4846]: I1005 07:35:27.758910 4846 generic.go:334] "Generic (PLEG): container finished" podID="f11d6823-4622-49bf-aa2a-41fcd37b7f96" containerID="39004b7c3b9d5642f010771a831d6d5c9c3e714ac17a13360425cb6f3c49724e" exitCode=0 Oct 05 07:35:27 crc kubenswrapper[4846]: I1005 07:35:27.758966 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7fc4h" event={"ID":"f11d6823-4622-49bf-aa2a-41fcd37b7f96","Type":"ContainerDied","Data":"39004b7c3b9d5642f010771a831d6d5c9c3e714ac17a13360425cb6f3c49724e"} Oct 05 07:35:27 crc kubenswrapper[4846]: I1005 07:35:27.759372 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7fc4h" event={"ID":"f11d6823-4622-49bf-aa2a-41fcd37b7f96","Type":"ContainerStarted","Data":"cb10b2fdce482fb7ca70748b45c3f5a0743884dedc30543c2f558ae4a2dc47a4"} Oct 05 07:35:28 crc kubenswrapper[4846]: I1005 07:35:28.770860 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7fc4h" event={"ID":"f11d6823-4622-49bf-aa2a-41fcd37b7f96","Type":"ContainerStarted","Data":"f03751c84fd861fa08c0457d06e28ce7ceed0225beaa8bf228db887600ea62b9"} Oct 05 07:35:29 crc kubenswrapper[4846]: I1005 07:35:29.785351 4846 generic.go:334] "Generic (PLEG): container finished" podID="f11d6823-4622-49bf-aa2a-41fcd37b7f96" containerID="f03751c84fd861fa08c0457d06e28ce7ceed0225beaa8bf228db887600ea62b9" exitCode=0 Oct 05 07:35:29 crc kubenswrapper[4846]: I1005 07:35:29.785431 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7fc4h" event={"ID":"f11d6823-4622-49bf-aa2a-41fcd37b7f96","Type":"ContainerDied","Data":"f03751c84fd861fa08c0457d06e28ce7ceed0225beaa8bf228db887600ea62b9"} Oct 05 07:35:30 crc kubenswrapper[4846]: I1005 07:35:30.816914 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7fc4h" event={"ID":"f11d6823-4622-49bf-aa2a-41fcd37b7f96","Type":"ContainerStarted","Data":"8bfed29f6f45ddb93a9b3821d93fd307776ae1cdec28fabf9558acb362060a21"} Oct 05 07:35:30 crc kubenswrapper[4846]: I1005 07:35:30.841890 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-7fc4h" podStartSLOduration=2.318453877 podStartE2EDuration="4.84187503s" podCreationTimestamp="2025-10-05 07:35:26 +0000 UTC" firstStartedPulling="2025-10-05 07:35:27.760823738 +0000 UTC m=+2850.001676503" lastFinishedPulling="2025-10-05 07:35:30.284244881 +0000 UTC m=+2852.525097656" observedRunningTime="2025-10-05 07:35:30.841358096 +0000 UTC m=+2853.082210881" watchObservedRunningTime="2025-10-05 07:35:30.84187503 +0000 UTC m=+2853.082727805" Oct 05 07:35:36 crc kubenswrapper[4846]: I1005 07:35:36.414364 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-7fc4h" Oct 05 07:35:36 crc kubenswrapper[4846]: I1005 07:35:36.414900 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-7fc4h" Oct 05 07:35:36 crc kubenswrapper[4846]: I1005 07:35:36.515830 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-7fc4h" Oct 05 07:35:36 crc kubenswrapper[4846]: I1005 07:35:36.941886 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-7fc4h" Oct 05 07:35:37 crc kubenswrapper[4846]: I1005 07:35:37.002913 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-7fc4h"] Oct 05 07:35:38 crc kubenswrapper[4846]: I1005 07:35:38.893085 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-7fc4h" podUID="f11d6823-4622-49bf-aa2a-41fcd37b7f96" containerName="registry-server" containerID="cri-o://8bfed29f6f45ddb93a9b3821d93fd307776ae1cdec28fabf9558acb362060a21" gracePeriod=2 Oct 05 07:35:39 crc kubenswrapper[4846]: I1005 07:35:39.478255 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7fc4h" Oct 05 07:35:39 crc kubenswrapper[4846]: I1005 07:35:39.548608 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dws7g\" (UniqueName: \"kubernetes.io/projected/f11d6823-4622-49bf-aa2a-41fcd37b7f96-kube-api-access-dws7g\") pod \"f11d6823-4622-49bf-aa2a-41fcd37b7f96\" (UID: \"f11d6823-4622-49bf-aa2a-41fcd37b7f96\") " Oct 05 07:35:39 crc kubenswrapper[4846]: I1005 07:35:39.548863 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f11d6823-4622-49bf-aa2a-41fcd37b7f96-utilities\") pod \"f11d6823-4622-49bf-aa2a-41fcd37b7f96\" (UID: \"f11d6823-4622-49bf-aa2a-41fcd37b7f96\") " Oct 05 07:35:39 crc kubenswrapper[4846]: I1005 07:35:39.548917 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f11d6823-4622-49bf-aa2a-41fcd37b7f96-catalog-content\") pod \"f11d6823-4622-49bf-aa2a-41fcd37b7f96\" (UID: \"f11d6823-4622-49bf-aa2a-41fcd37b7f96\") " Oct 05 07:35:39 crc kubenswrapper[4846]: I1005 07:35:39.549967 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f11d6823-4622-49bf-aa2a-41fcd37b7f96-utilities" (OuterVolumeSpecName: "utilities") pod "f11d6823-4622-49bf-aa2a-41fcd37b7f96" (UID: "f11d6823-4622-49bf-aa2a-41fcd37b7f96"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 07:35:39 crc kubenswrapper[4846]: I1005 07:35:39.555737 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f11d6823-4622-49bf-aa2a-41fcd37b7f96-kube-api-access-dws7g" (OuterVolumeSpecName: "kube-api-access-dws7g") pod "f11d6823-4622-49bf-aa2a-41fcd37b7f96" (UID: "f11d6823-4622-49bf-aa2a-41fcd37b7f96"). InnerVolumeSpecName "kube-api-access-dws7g". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:35:39 crc kubenswrapper[4846]: I1005 07:35:39.621575 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f11d6823-4622-49bf-aa2a-41fcd37b7f96-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f11d6823-4622-49bf-aa2a-41fcd37b7f96" (UID: "f11d6823-4622-49bf-aa2a-41fcd37b7f96"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 07:35:39 crc kubenswrapper[4846]: I1005 07:35:39.650936 4846 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f11d6823-4622-49bf-aa2a-41fcd37b7f96-utilities\") on node \"crc\" DevicePath \"\"" Oct 05 07:35:39 crc kubenswrapper[4846]: I1005 07:35:39.650978 4846 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f11d6823-4622-49bf-aa2a-41fcd37b7f96-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 05 07:35:39 crc kubenswrapper[4846]: I1005 07:35:39.650998 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dws7g\" (UniqueName: \"kubernetes.io/projected/f11d6823-4622-49bf-aa2a-41fcd37b7f96-kube-api-access-dws7g\") on node \"crc\" DevicePath \"\"" Oct 05 07:35:39 crc kubenswrapper[4846]: I1005 07:35:39.905312 4846 generic.go:334] "Generic (PLEG): container finished" podID="f11d6823-4622-49bf-aa2a-41fcd37b7f96" containerID="8bfed29f6f45ddb93a9b3821d93fd307776ae1cdec28fabf9558acb362060a21" exitCode=0 Oct 05 07:35:39 crc kubenswrapper[4846]: I1005 07:35:39.905368 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7fc4h" event={"ID":"f11d6823-4622-49bf-aa2a-41fcd37b7f96","Type":"ContainerDied","Data":"8bfed29f6f45ddb93a9b3821d93fd307776ae1cdec28fabf9558acb362060a21"} Oct 05 07:35:39 crc kubenswrapper[4846]: I1005 07:35:39.905408 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7fc4h" event={"ID":"f11d6823-4622-49bf-aa2a-41fcd37b7f96","Type":"ContainerDied","Data":"cb10b2fdce482fb7ca70748b45c3f5a0743884dedc30543c2f558ae4a2dc47a4"} Oct 05 07:35:39 crc kubenswrapper[4846]: I1005 07:35:39.905435 4846 scope.go:117] "RemoveContainer" containerID="8bfed29f6f45ddb93a9b3821d93fd307776ae1cdec28fabf9558acb362060a21" Oct 05 07:35:39 crc kubenswrapper[4846]: I1005 07:35:39.905426 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7fc4h" Oct 05 07:35:39 crc kubenswrapper[4846]: I1005 07:35:39.927676 4846 scope.go:117] "RemoveContainer" containerID="f03751c84fd861fa08c0457d06e28ce7ceed0225beaa8bf228db887600ea62b9" Oct 05 07:35:39 crc kubenswrapper[4846]: I1005 07:35:39.952951 4846 scope.go:117] "RemoveContainer" containerID="39004b7c3b9d5642f010771a831d6d5c9c3e714ac17a13360425cb6f3c49724e" Oct 05 07:35:39 crc kubenswrapper[4846]: I1005 07:35:39.980334 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-7fc4h"] Oct 05 07:35:39 crc kubenswrapper[4846]: I1005 07:35:39.995111 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-7fc4h"] Oct 05 07:35:40 crc kubenswrapper[4846]: I1005 07:35:40.004051 4846 scope.go:117] "RemoveContainer" containerID="8bfed29f6f45ddb93a9b3821d93fd307776ae1cdec28fabf9558acb362060a21" Oct 05 07:35:40 crc kubenswrapper[4846]: E1005 07:35:40.004652 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8bfed29f6f45ddb93a9b3821d93fd307776ae1cdec28fabf9558acb362060a21\": container with ID starting with 8bfed29f6f45ddb93a9b3821d93fd307776ae1cdec28fabf9558acb362060a21 not found: ID does not exist" containerID="8bfed29f6f45ddb93a9b3821d93fd307776ae1cdec28fabf9558acb362060a21" Oct 05 07:35:40 crc kubenswrapper[4846]: I1005 07:35:40.004731 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8bfed29f6f45ddb93a9b3821d93fd307776ae1cdec28fabf9558acb362060a21"} err="failed to get container status \"8bfed29f6f45ddb93a9b3821d93fd307776ae1cdec28fabf9558acb362060a21\": rpc error: code = NotFound desc = could not find container \"8bfed29f6f45ddb93a9b3821d93fd307776ae1cdec28fabf9558acb362060a21\": container with ID starting with 8bfed29f6f45ddb93a9b3821d93fd307776ae1cdec28fabf9558acb362060a21 not found: ID does not exist" Oct 05 07:35:40 crc kubenswrapper[4846]: I1005 07:35:40.004813 4846 scope.go:117] "RemoveContainer" containerID="f03751c84fd861fa08c0457d06e28ce7ceed0225beaa8bf228db887600ea62b9" Oct 05 07:35:40 crc kubenswrapper[4846]: E1005 07:35:40.005149 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f03751c84fd861fa08c0457d06e28ce7ceed0225beaa8bf228db887600ea62b9\": container with ID starting with f03751c84fd861fa08c0457d06e28ce7ceed0225beaa8bf228db887600ea62b9 not found: ID does not exist" containerID="f03751c84fd861fa08c0457d06e28ce7ceed0225beaa8bf228db887600ea62b9" Oct 05 07:35:40 crc kubenswrapper[4846]: I1005 07:35:40.005212 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f03751c84fd861fa08c0457d06e28ce7ceed0225beaa8bf228db887600ea62b9"} err="failed to get container status \"f03751c84fd861fa08c0457d06e28ce7ceed0225beaa8bf228db887600ea62b9\": rpc error: code = NotFound desc = could not find container \"f03751c84fd861fa08c0457d06e28ce7ceed0225beaa8bf228db887600ea62b9\": container with ID starting with f03751c84fd861fa08c0457d06e28ce7ceed0225beaa8bf228db887600ea62b9 not found: ID does not exist" Oct 05 07:35:40 crc kubenswrapper[4846]: I1005 07:35:40.005241 4846 scope.go:117] "RemoveContainer" containerID="39004b7c3b9d5642f010771a831d6d5c9c3e714ac17a13360425cb6f3c49724e" Oct 05 07:35:40 crc kubenswrapper[4846]: E1005 07:35:40.005599 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"39004b7c3b9d5642f010771a831d6d5c9c3e714ac17a13360425cb6f3c49724e\": container with ID starting with 39004b7c3b9d5642f010771a831d6d5c9c3e714ac17a13360425cb6f3c49724e not found: ID does not exist" containerID="39004b7c3b9d5642f010771a831d6d5c9c3e714ac17a13360425cb6f3c49724e" Oct 05 07:35:40 crc kubenswrapper[4846]: I1005 07:35:40.005631 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"39004b7c3b9d5642f010771a831d6d5c9c3e714ac17a13360425cb6f3c49724e"} err="failed to get container status \"39004b7c3b9d5642f010771a831d6d5c9c3e714ac17a13360425cb6f3c49724e\": rpc error: code = NotFound desc = could not find container \"39004b7c3b9d5642f010771a831d6d5c9c3e714ac17a13360425cb6f3c49724e\": container with ID starting with 39004b7c3b9d5642f010771a831d6d5c9c3e714ac17a13360425cb6f3c49724e not found: ID does not exist" Oct 05 07:35:40 crc kubenswrapper[4846]: I1005 07:35:40.532601 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f11d6823-4622-49bf-aa2a-41fcd37b7f96" path="/var/lib/kubelet/pods/f11d6823-4622-49bf-aa2a-41fcd37b7f96/volumes" Oct 05 07:35:53 crc kubenswrapper[4846]: I1005 07:35:53.325148 4846 patch_prober.go:28] interesting pod/machine-config-daemon-fscvf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 07:35:53 crc kubenswrapper[4846]: I1005 07:35:53.325962 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 07:36:23 crc kubenswrapper[4846]: I1005 07:36:23.325059 4846 patch_prober.go:28] interesting pod/machine-config-daemon-fscvf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 07:36:23 crc kubenswrapper[4846]: I1005 07:36:23.325757 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 07:36:23 crc kubenswrapper[4846]: I1005 07:36:23.325892 4846 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" Oct 05 07:36:23 crc kubenswrapper[4846]: I1005 07:36:23.327458 4846 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"702f9b90152900919451502be251dfa1fc91d3fc2af6439c709b401f8ac5d520"} pod="openshift-machine-config-operator/machine-config-daemon-fscvf" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 05 07:36:23 crc kubenswrapper[4846]: I1005 07:36:23.327574 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" containerName="machine-config-daemon" containerID="cri-o://702f9b90152900919451502be251dfa1fc91d3fc2af6439c709b401f8ac5d520" gracePeriod=600 Oct 05 07:36:23 crc kubenswrapper[4846]: E1005 07:36:23.455273 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 07:36:24 crc kubenswrapper[4846]: I1005 07:36:24.343119 4846 generic.go:334] "Generic (PLEG): container finished" podID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" containerID="702f9b90152900919451502be251dfa1fc91d3fc2af6439c709b401f8ac5d520" exitCode=0 Oct 05 07:36:24 crc kubenswrapper[4846]: I1005 07:36:24.343228 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" event={"ID":"ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7","Type":"ContainerDied","Data":"702f9b90152900919451502be251dfa1fc91d3fc2af6439c709b401f8ac5d520"} Oct 05 07:36:24 crc kubenswrapper[4846]: I1005 07:36:24.343585 4846 scope.go:117] "RemoveContainer" containerID="b6694c2f42f5d02bf9e82272eb86940e1c6a3c4bb6e25979d8d7da5c72b02506" Oct 05 07:36:24 crc kubenswrapper[4846]: I1005 07:36:24.344438 4846 scope.go:117] "RemoveContainer" containerID="702f9b90152900919451502be251dfa1fc91d3fc2af6439c709b401f8ac5d520" Oct 05 07:36:24 crc kubenswrapper[4846]: E1005 07:36:24.344923 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 07:36:38 crc kubenswrapper[4846]: I1005 07:36:38.505585 4846 scope.go:117] "RemoveContainer" containerID="702f9b90152900919451502be251dfa1fc91d3fc2af6439c709b401f8ac5d520" Oct 05 07:36:38 crc kubenswrapper[4846]: E1005 07:36:38.506614 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 07:36:51 crc kubenswrapper[4846]: I1005 07:36:51.498752 4846 scope.go:117] "RemoveContainer" containerID="702f9b90152900919451502be251dfa1fc91d3fc2af6439c709b401f8ac5d520" Oct 05 07:36:51 crc kubenswrapper[4846]: E1005 07:36:51.500403 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 07:37:05 crc kubenswrapper[4846]: I1005 07:37:05.498566 4846 scope.go:117] "RemoveContainer" containerID="702f9b90152900919451502be251dfa1fc91d3fc2af6439c709b401f8ac5d520" Oct 05 07:37:05 crc kubenswrapper[4846]: E1005 07:37:05.499754 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 07:37:18 crc kubenswrapper[4846]: I1005 07:37:18.501431 4846 scope.go:117] "RemoveContainer" containerID="702f9b90152900919451502be251dfa1fc91d3fc2af6439c709b401f8ac5d520" Oct 05 07:37:18 crc kubenswrapper[4846]: E1005 07:37:18.502249 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 07:37:33 crc kubenswrapper[4846]: I1005 07:37:33.498019 4846 scope.go:117] "RemoveContainer" containerID="702f9b90152900919451502be251dfa1fc91d3fc2af6439c709b401f8ac5d520" Oct 05 07:37:33 crc kubenswrapper[4846]: E1005 07:37:33.499146 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 07:37:44 crc kubenswrapper[4846]: I1005 07:37:44.499100 4846 scope.go:117] "RemoveContainer" containerID="702f9b90152900919451502be251dfa1fc91d3fc2af6439c709b401f8ac5d520" Oct 05 07:37:44 crc kubenswrapper[4846]: E1005 07:37:44.500491 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 07:37:57 crc kubenswrapper[4846]: I1005 07:37:57.498062 4846 scope.go:117] "RemoveContainer" containerID="702f9b90152900919451502be251dfa1fc91d3fc2af6439c709b401f8ac5d520" Oct 05 07:37:57 crc kubenswrapper[4846]: E1005 07:37:57.499399 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 07:38:10 crc kubenswrapper[4846]: I1005 07:38:10.498277 4846 scope.go:117] "RemoveContainer" containerID="702f9b90152900919451502be251dfa1fc91d3fc2af6439c709b401f8ac5d520" Oct 05 07:38:10 crc kubenswrapper[4846]: E1005 07:38:10.499595 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 07:38:23 crc kubenswrapper[4846]: I1005 07:38:23.498465 4846 scope.go:117] "RemoveContainer" containerID="702f9b90152900919451502be251dfa1fc91d3fc2af6439c709b401f8ac5d520" Oct 05 07:38:23 crc kubenswrapper[4846]: E1005 07:38:23.499607 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 07:38:38 crc kubenswrapper[4846]: I1005 07:38:38.506155 4846 scope.go:117] "RemoveContainer" containerID="702f9b90152900919451502be251dfa1fc91d3fc2af6439c709b401f8ac5d520" Oct 05 07:38:38 crc kubenswrapper[4846]: E1005 07:38:38.507451 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 07:38:52 crc kubenswrapper[4846]: I1005 07:38:52.499010 4846 scope.go:117] "RemoveContainer" containerID="702f9b90152900919451502be251dfa1fc91d3fc2af6439c709b401f8ac5d520" Oct 05 07:38:52 crc kubenswrapper[4846]: E1005 07:38:52.500289 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 07:39:07 crc kubenswrapper[4846]: I1005 07:39:07.498142 4846 scope.go:117] "RemoveContainer" containerID="702f9b90152900919451502be251dfa1fc91d3fc2af6439c709b401f8ac5d520" Oct 05 07:39:07 crc kubenswrapper[4846]: E1005 07:39:07.499416 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 07:39:19 crc kubenswrapper[4846]: I1005 07:39:19.497869 4846 scope.go:117] "RemoveContainer" containerID="702f9b90152900919451502be251dfa1fc91d3fc2af6439c709b401f8ac5d520" Oct 05 07:39:19 crc kubenswrapper[4846]: E1005 07:39:19.498830 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 07:39:31 crc kubenswrapper[4846]: I1005 07:39:31.497467 4846 scope.go:117] "RemoveContainer" containerID="702f9b90152900919451502be251dfa1fc91d3fc2af6439c709b401f8ac5d520" Oct 05 07:39:31 crc kubenswrapper[4846]: E1005 07:39:31.498483 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 07:39:44 crc kubenswrapper[4846]: I1005 07:39:44.497845 4846 scope.go:117] "RemoveContainer" containerID="702f9b90152900919451502be251dfa1fc91d3fc2af6439c709b401f8ac5d520" Oct 05 07:39:44 crc kubenswrapper[4846]: E1005 07:39:44.499267 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 07:39:59 crc kubenswrapper[4846]: I1005 07:39:59.498011 4846 scope.go:117] "RemoveContainer" containerID="702f9b90152900919451502be251dfa1fc91d3fc2af6439c709b401f8ac5d520" Oct 05 07:39:59 crc kubenswrapper[4846]: E1005 07:39:59.499043 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 07:40:11 crc kubenswrapper[4846]: I1005 07:40:11.497445 4846 scope.go:117] "RemoveContainer" containerID="702f9b90152900919451502be251dfa1fc91d3fc2af6439c709b401f8ac5d520" Oct 05 07:40:11 crc kubenswrapper[4846]: E1005 07:40:11.498174 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 07:40:22 crc kubenswrapper[4846]: I1005 07:40:22.503941 4846 scope.go:117] "RemoveContainer" containerID="702f9b90152900919451502be251dfa1fc91d3fc2af6439c709b401f8ac5d520" Oct 05 07:40:22 crc kubenswrapper[4846]: E1005 07:40:22.506366 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 07:40:35 crc kubenswrapper[4846]: I1005 07:40:35.496977 4846 scope.go:117] "RemoveContainer" containerID="702f9b90152900919451502be251dfa1fc91d3fc2af6439c709b401f8ac5d520" Oct 05 07:40:35 crc kubenswrapper[4846]: E1005 07:40:35.497769 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 07:40:50 crc kubenswrapper[4846]: I1005 07:40:50.498133 4846 scope.go:117] "RemoveContainer" containerID="702f9b90152900919451502be251dfa1fc91d3fc2af6439c709b401f8ac5d520" Oct 05 07:40:50 crc kubenswrapper[4846]: E1005 07:40:50.499235 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 07:41:04 crc kubenswrapper[4846]: I1005 07:41:04.497854 4846 scope.go:117] "RemoveContainer" containerID="702f9b90152900919451502be251dfa1fc91d3fc2af6439c709b401f8ac5d520" Oct 05 07:41:04 crc kubenswrapper[4846]: E1005 07:41:04.498705 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 07:41:16 crc kubenswrapper[4846]: I1005 07:41:16.497599 4846 scope.go:117] "RemoveContainer" containerID="702f9b90152900919451502be251dfa1fc91d3fc2af6439c709b401f8ac5d520" Oct 05 07:41:16 crc kubenswrapper[4846]: E1005 07:41:16.498826 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 07:41:31 crc kubenswrapper[4846]: I1005 07:41:31.498011 4846 scope.go:117] "RemoveContainer" containerID="702f9b90152900919451502be251dfa1fc91d3fc2af6439c709b401f8ac5d520" Oct 05 07:41:32 crc kubenswrapper[4846]: I1005 07:41:32.228169 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" event={"ID":"ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7","Type":"ContainerStarted","Data":"725a93409ff4757fae7733195dbbc6de2ee38624fdbf934d0b72e1d2e3418171"} Oct 05 07:43:53 crc kubenswrapper[4846]: I1005 07:43:53.325649 4846 patch_prober.go:28] interesting pod/machine-config-daemon-fscvf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 07:43:53 crc kubenswrapper[4846]: I1005 07:43:53.326353 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 07:43:54 crc kubenswrapper[4846]: I1005 07:43:54.949817 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-hqxdx"] Oct 05 07:43:54 crc kubenswrapper[4846]: E1005 07:43:54.952629 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f11d6823-4622-49bf-aa2a-41fcd37b7f96" containerName="extract-utilities" Oct 05 07:43:54 crc kubenswrapper[4846]: I1005 07:43:54.952688 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="f11d6823-4622-49bf-aa2a-41fcd37b7f96" containerName="extract-utilities" Oct 05 07:43:54 crc kubenswrapper[4846]: E1005 07:43:54.952733 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f11d6823-4622-49bf-aa2a-41fcd37b7f96" containerName="extract-content" Oct 05 07:43:54 crc kubenswrapper[4846]: I1005 07:43:54.952759 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="f11d6823-4622-49bf-aa2a-41fcd37b7f96" containerName="extract-content" Oct 05 07:43:54 crc kubenswrapper[4846]: E1005 07:43:54.952783 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f11d6823-4622-49bf-aa2a-41fcd37b7f96" containerName="registry-server" Oct 05 07:43:54 crc kubenswrapper[4846]: I1005 07:43:54.952793 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="f11d6823-4622-49bf-aa2a-41fcd37b7f96" containerName="registry-server" Oct 05 07:43:54 crc kubenswrapper[4846]: I1005 07:43:54.953166 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="f11d6823-4622-49bf-aa2a-41fcd37b7f96" containerName="registry-server" Oct 05 07:43:54 crc kubenswrapper[4846]: I1005 07:43:54.954948 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hqxdx" Oct 05 07:43:54 crc kubenswrapper[4846]: I1005 07:43:54.970092 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tsnsx\" (UniqueName: \"kubernetes.io/projected/25529293-99d7-4c74-97ec-31a8e3fed742-kube-api-access-tsnsx\") pod \"certified-operators-hqxdx\" (UID: \"25529293-99d7-4c74-97ec-31a8e3fed742\") " pod="openshift-marketplace/certified-operators-hqxdx" Oct 05 07:43:54 crc kubenswrapper[4846]: I1005 07:43:54.970238 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/25529293-99d7-4c74-97ec-31a8e3fed742-catalog-content\") pod \"certified-operators-hqxdx\" (UID: \"25529293-99d7-4c74-97ec-31a8e3fed742\") " pod="openshift-marketplace/certified-operators-hqxdx" Oct 05 07:43:54 crc kubenswrapper[4846]: I1005 07:43:54.970288 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/25529293-99d7-4c74-97ec-31a8e3fed742-utilities\") pod \"certified-operators-hqxdx\" (UID: \"25529293-99d7-4c74-97ec-31a8e3fed742\") " pod="openshift-marketplace/certified-operators-hqxdx" Oct 05 07:43:54 crc kubenswrapper[4846]: I1005 07:43:54.979884 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-hqxdx"] Oct 05 07:43:55 crc kubenswrapper[4846]: I1005 07:43:55.071645 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/25529293-99d7-4c74-97ec-31a8e3fed742-catalog-content\") pod \"certified-operators-hqxdx\" (UID: \"25529293-99d7-4c74-97ec-31a8e3fed742\") " pod="openshift-marketplace/certified-operators-hqxdx" Oct 05 07:43:55 crc kubenswrapper[4846]: I1005 07:43:55.071711 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/25529293-99d7-4c74-97ec-31a8e3fed742-utilities\") pod \"certified-operators-hqxdx\" (UID: \"25529293-99d7-4c74-97ec-31a8e3fed742\") " pod="openshift-marketplace/certified-operators-hqxdx" Oct 05 07:43:55 crc kubenswrapper[4846]: I1005 07:43:55.071772 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tsnsx\" (UniqueName: \"kubernetes.io/projected/25529293-99d7-4c74-97ec-31a8e3fed742-kube-api-access-tsnsx\") pod \"certified-operators-hqxdx\" (UID: \"25529293-99d7-4c74-97ec-31a8e3fed742\") " pod="openshift-marketplace/certified-operators-hqxdx" Oct 05 07:43:55 crc kubenswrapper[4846]: I1005 07:43:55.072362 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/25529293-99d7-4c74-97ec-31a8e3fed742-catalog-content\") pod \"certified-operators-hqxdx\" (UID: \"25529293-99d7-4c74-97ec-31a8e3fed742\") " pod="openshift-marketplace/certified-operators-hqxdx" Oct 05 07:43:55 crc kubenswrapper[4846]: I1005 07:43:55.072576 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/25529293-99d7-4c74-97ec-31a8e3fed742-utilities\") pod \"certified-operators-hqxdx\" (UID: \"25529293-99d7-4c74-97ec-31a8e3fed742\") " pod="openshift-marketplace/certified-operators-hqxdx" Oct 05 07:43:55 crc kubenswrapper[4846]: I1005 07:43:55.102331 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tsnsx\" (UniqueName: \"kubernetes.io/projected/25529293-99d7-4c74-97ec-31a8e3fed742-kube-api-access-tsnsx\") pod \"certified-operators-hqxdx\" (UID: \"25529293-99d7-4c74-97ec-31a8e3fed742\") " pod="openshift-marketplace/certified-operators-hqxdx" Oct 05 07:43:55 crc kubenswrapper[4846]: I1005 07:43:55.293068 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hqxdx" Oct 05 07:43:55 crc kubenswrapper[4846]: I1005 07:43:55.603974 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-hqxdx"] Oct 05 07:43:55 crc kubenswrapper[4846]: I1005 07:43:55.662992 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hqxdx" event={"ID":"25529293-99d7-4c74-97ec-31a8e3fed742","Type":"ContainerStarted","Data":"ceb05f4050313bec07de3192366b81ece161af38de4eb79c63daeb7e6d0a8938"} Oct 05 07:43:56 crc kubenswrapper[4846]: I1005 07:43:56.678260 4846 generic.go:334] "Generic (PLEG): container finished" podID="25529293-99d7-4c74-97ec-31a8e3fed742" containerID="d651b03b52ca5b19510a41c74746e0aa330229c761a2259de176195549b84009" exitCode=0 Oct 05 07:43:56 crc kubenswrapper[4846]: I1005 07:43:56.678595 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hqxdx" event={"ID":"25529293-99d7-4c74-97ec-31a8e3fed742","Type":"ContainerDied","Data":"d651b03b52ca5b19510a41c74746e0aa330229c761a2259de176195549b84009"} Oct 05 07:43:56 crc kubenswrapper[4846]: I1005 07:43:56.681844 4846 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 05 07:43:57 crc kubenswrapper[4846]: I1005 07:43:57.692945 4846 generic.go:334] "Generic (PLEG): container finished" podID="25529293-99d7-4c74-97ec-31a8e3fed742" containerID="935b07387f0457a303c4479c9b0fd1f70a389fccfb24029662b7faff770e4f77" exitCode=0 Oct 05 07:43:57 crc kubenswrapper[4846]: I1005 07:43:57.693064 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hqxdx" event={"ID":"25529293-99d7-4c74-97ec-31a8e3fed742","Type":"ContainerDied","Data":"935b07387f0457a303c4479c9b0fd1f70a389fccfb24029662b7faff770e4f77"} Oct 05 07:43:58 crc kubenswrapper[4846]: I1005 07:43:58.703581 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hqxdx" event={"ID":"25529293-99d7-4c74-97ec-31a8e3fed742","Type":"ContainerStarted","Data":"8e20f3b1aca1c643d396a60ada4747a610a1b142978718e8fab06ffd931677d2"} Oct 05 07:43:58 crc kubenswrapper[4846]: I1005 07:43:58.727645 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-hqxdx" podStartSLOduration=3.322910743 podStartE2EDuration="4.727624908s" podCreationTimestamp="2025-10-05 07:43:54 +0000 UTC" firstStartedPulling="2025-10-05 07:43:56.681446457 +0000 UTC m=+3358.922299262" lastFinishedPulling="2025-10-05 07:43:58.086160642 +0000 UTC m=+3360.327013427" observedRunningTime="2025-10-05 07:43:58.721567267 +0000 UTC m=+3360.962420092" watchObservedRunningTime="2025-10-05 07:43:58.727624908 +0000 UTC m=+3360.968477693" Oct 05 07:44:05 crc kubenswrapper[4846]: I1005 07:44:05.293777 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-hqxdx" Oct 05 07:44:05 crc kubenswrapper[4846]: I1005 07:44:05.295482 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-hqxdx" Oct 05 07:44:05 crc kubenswrapper[4846]: I1005 07:44:05.378690 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-hqxdx" Oct 05 07:44:05 crc kubenswrapper[4846]: I1005 07:44:05.808668 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-hqxdx" Oct 05 07:44:05 crc kubenswrapper[4846]: I1005 07:44:05.862542 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-hqxdx"] Oct 05 07:44:07 crc kubenswrapper[4846]: I1005 07:44:07.781054 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-hqxdx" podUID="25529293-99d7-4c74-97ec-31a8e3fed742" containerName="registry-server" containerID="cri-o://8e20f3b1aca1c643d396a60ada4747a610a1b142978718e8fab06ffd931677d2" gracePeriod=2 Oct 05 07:44:08 crc kubenswrapper[4846]: I1005 07:44:08.364678 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hqxdx" Oct 05 07:44:08 crc kubenswrapper[4846]: I1005 07:44:08.515274 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tsnsx\" (UniqueName: \"kubernetes.io/projected/25529293-99d7-4c74-97ec-31a8e3fed742-kube-api-access-tsnsx\") pod \"25529293-99d7-4c74-97ec-31a8e3fed742\" (UID: \"25529293-99d7-4c74-97ec-31a8e3fed742\") " Oct 05 07:44:08 crc kubenswrapper[4846]: I1005 07:44:08.515391 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/25529293-99d7-4c74-97ec-31a8e3fed742-utilities\") pod \"25529293-99d7-4c74-97ec-31a8e3fed742\" (UID: \"25529293-99d7-4c74-97ec-31a8e3fed742\") " Oct 05 07:44:08 crc kubenswrapper[4846]: I1005 07:44:08.515537 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/25529293-99d7-4c74-97ec-31a8e3fed742-catalog-content\") pod \"25529293-99d7-4c74-97ec-31a8e3fed742\" (UID: \"25529293-99d7-4c74-97ec-31a8e3fed742\") " Oct 05 07:44:08 crc kubenswrapper[4846]: I1005 07:44:08.516458 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/25529293-99d7-4c74-97ec-31a8e3fed742-utilities" (OuterVolumeSpecName: "utilities") pod "25529293-99d7-4c74-97ec-31a8e3fed742" (UID: "25529293-99d7-4c74-97ec-31a8e3fed742"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 07:44:08 crc kubenswrapper[4846]: I1005 07:44:08.526358 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25529293-99d7-4c74-97ec-31a8e3fed742-kube-api-access-tsnsx" (OuterVolumeSpecName: "kube-api-access-tsnsx") pod "25529293-99d7-4c74-97ec-31a8e3fed742" (UID: "25529293-99d7-4c74-97ec-31a8e3fed742"). InnerVolumeSpecName "kube-api-access-tsnsx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:44:08 crc kubenswrapper[4846]: I1005 07:44:08.591551 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/25529293-99d7-4c74-97ec-31a8e3fed742-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "25529293-99d7-4c74-97ec-31a8e3fed742" (UID: "25529293-99d7-4c74-97ec-31a8e3fed742"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 07:44:08 crc kubenswrapper[4846]: I1005 07:44:08.617134 4846 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/25529293-99d7-4c74-97ec-31a8e3fed742-utilities\") on node \"crc\" DevicePath \"\"" Oct 05 07:44:08 crc kubenswrapper[4846]: I1005 07:44:08.617168 4846 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/25529293-99d7-4c74-97ec-31a8e3fed742-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 05 07:44:08 crc kubenswrapper[4846]: I1005 07:44:08.617198 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tsnsx\" (UniqueName: \"kubernetes.io/projected/25529293-99d7-4c74-97ec-31a8e3fed742-kube-api-access-tsnsx\") on node \"crc\" DevicePath \"\"" Oct 05 07:44:08 crc kubenswrapper[4846]: I1005 07:44:08.795566 4846 generic.go:334] "Generic (PLEG): container finished" podID="25529293-99d7-4c74-97ec-31a8e3fed742" containerID="8e20f3b1aca1c643d396a60ada4747a610a1b142978718e8fab06ffd931677d2" exitCode=0 Oct 05 07:44:08 crc kubenswrapper[4846]: I1005 07:44:08.795643 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hqxdx" event={"ID":"25529293-99d7-4c74-97ec-31a8e3fed742","Type":"ContainerDied","Data":"8e20f3b1aca1c643d396a60ada4747a610a1b142978718e8fab06ffd931677d2"} Oct 05 07:44:08 crc kubenswrapper[4846]: I1005 07:44:08.795704 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hqxdx" Oct 05 07:44:08 crc kubenswrapper[4846]: I1005 07:44:08.795729 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hqxdx" event={"ID":"25529293-99d7-4c74-97ec-31a8e3fed742","Type":"ContainerDied","Data":"ceb05f4050313bec07de3192366b81ece161af38de4eb79c63daeb7e6d0a8938"} Oct 05 07:44:08 crc kubenswrapper[4846]: I1005 07:44:08.795767 4846 scope.go:117] "RemoveContainer" containerID="8e20f3b1aca1c643d396a60ada4747a610a1b142978718e8fab06ffd931677d2" Oct 05 07:44:08 crc kubenswrapper[4846]: I1005 07:44:08.822879 4846 scope.go:117] "RemoveContainer" containerID="935b07387f0457a303c4479c9b0fd1f70a389fccfb24029662b7faff770e4f77" Oct 05 07:44:08 crc kubenswrapper[4846]: I1005 07:44:08.857490 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-hqxdx"] Oct 05 07:44:08 crc kubenswrapper[4846]: I1005 07:44:08.868477 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-hqxdx"] Oct 05 07:44:08 crc kubenswrapper[4846]: I1005 07:44:08.891021 4846 scope.go:117] "RemoveContainer" containerID="d651b03b52ca5b19510a41c74746e0aa330229c761a2259de176195549b84009" Oct 05 07:44:08 crc kubenswrapper[4846]: I1005 07:44:08.923956 4846 scope.go:117] "RemoveContainer" containerID="8e20f3b1aca1c643d396a60ada4747a610a1b142978718e8fab06ffd931677d2" Oct 05 07:44:08 crc kubenswrapper[4846]: E1005 07:44:08.924598 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8e20f3b1aca1c643d396a60ada4747a610a1b142978718e8fab06ffd931677d2\": container with ID starting with 8e20f3b1aca1c643d396a60ada4747a610a1b142978718e8fab06ffd931677d2 not found: ID does not exist" containerID="8e20f3b1aca1c643d396a60ada4747a610a1b142978718e8fab06ffd931677d2" Oct 05 07:44:08 crc kubenswrapper[4846]: I1005 07:44:08.924654 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8e20f3b1aca1c643d396a60ada4747a610a1b142978718e8fab06ffd931677d2"} err="failed to get container status \"8e20f3b1aca1c643d396a60ada4747a610a1b142978718e8fab06ffd931677d2\": rpc error: code = NotFound desc = could not find container \"8e20f3b1aca1c643d396a60ada4747a610a1b142978718e8fab06ffd931677d2\": container with ID starting with 8e20f3b1aca1c643d396a60ada4747a610a1b142978718e8fab06ffd931677d2 not found: ID does not exist" Oct 05 07:44:08 crc kubenswrapper[4846]: I1005 07:44:08.924691 4846 scope.go:117] "RemoveContainer" containerID="935b07387f0457a303c4479c9b0fd1f70a389fccfb24029662b7faff770e4f77" Oct 05 07:44:08 crc kubenswrapper[4846]: E1005 07:44:08.925140 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"935b07387f0457a303c4479c9b0fd1f70a389fccfb24029662b7faff770e4f77\": container with ID starting with 935b07387f0457a303c4479c9b0fd1f70a389fccfb24029662b7faff770e4f77 not found: ID does not exist" containerID="935b07387f0457a303c4479c9b0fd1f70a389fccfb24029662b7faff770e4f77" Oct 05 07:44:08 crc kubenswrapper[4846]: I1005 07:44:08.925200 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"935b07387f0457a303c4479c9b0fd1f70a389fccfb24029662b7faff770e4f77"} err="failed to get container status \"935b07387f0457a303c4479c9b0fd1f70a389fccfb24029662b7faff770e4f77\": rpc error: code = NotFound desc = could not find container \"935b07387f0457a303c4479c9b0fd1f70a389fccfb24029662b7faff770e4f77\": container with ID starting with 935b07387f0457a303c4479c9b0fd1f70a389fccfb24029662b7faff770e4f77 not found: ID does not exist" Oct 05 07:44:08 crc kubenswrapper[4846]: I1005 07:44:08.925228 4846 scope.go:117] "RemoveContainer" containerID="d651b03b52ca5b19510a41c74746e0aa330229c761a2259de176195549b84009" Oct 05 07:44:08 crc kubenswrapper[4846]: E1005 07:44:08.925628 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d651b03b52ca5b19510a41c74746e0aa330229c761a2259de176195549b84009\": container with ID starting with d651b03b52ca5b19510a41c74746e0aa330229c761a2259de176195549b84009 not found: ID does not exist" containerID="d651b03b52ca5b19510a41c74746e0aa330229c761a2259de176195549b84009" Oct 05 07:44:08 crc kubenswrapper[4846]: I1005 07:44:08.925677 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d651b03b52ca5b19510a41c74746e0aa330229c761a2259de176195549b84009"} err="failed to get container status \"d651b03b52ca5b19510a41c74746e0aa330229c761a2259de176195549b84009\": rpc error: code = NotFound desc = could not find container \"d651b03b52ca5b19510a41c74746e0aa330229c761a2259de176195549b84009\": container with ID starting with d651b03b52ca5b19510a41c74746e0aa330229c761a2259de176195549b84009 not found: ID does not exist" Oct 05 07:44:10 crc kubenswrapper[4846]: I1005 07:44:10.508899 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25529293-99d7-4c74-97ec-31a8e3fed742" path="/var/lib/kubelet/pods/25529293-99d7-4c74-97ec-31a8e3fed742/volumes" Oct 05 07:44:14 crc kubenswrapper[4846]: I1005 07:44:14.880637 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-hthsd"] Oct 05 07:44:14 crc kubenswrapper[4846]: E1005 07:44:14.881194 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="25529293-99d7-4c74-97ec-31a8e3fed742" containerName="registry-server" Oct 05 07:44:14 crc kubenswrapper[4846]: I1005 07:44:14.881208 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="25529293-99d7-4c74-97ec-31a8e3fed742" containerName="registry-server" Oct 05 07:44:14 crc kubenswrapper[4846]: E1005 07:44:14.881225 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="25529293-99d7-4c74-97ec-31a8e3fed742" containerName="extract-utilities" Oct 05 07:44:14 crc kubenswrapper[4846]: I1005 07:44:14.881231 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="25529293-99d7-4c74-97ec-31a8e3fed742" containerName="extract-utilities" Oct 05 07:44:14 crc kubenswrapper[4846]: E1005 07:44:14.881245 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="25529293-99d7-4c74-97ec-31a8e3fed742" containerName="extract-content" Oct 05 07:44:14 crc kubenswrapper[4846]: I1005 07:44:14.881250 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="25529293-99d7-4c74-97ec-31a8e3fed742" containerName="extract-content" Oct 05 07:44:14 crc kubenswrapper[4846]: I1005 07:44:14.881397 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="25529293-99d7-4c74-97ec-31a8e3fed742" containerName="registry-server" Oct 05 07:44:14 crc kubenswrapper[4846]: I1005 07:44:14.882476 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-hthsd" Oct 05 07:44:14 crc kubenswrapper[4846]: I1005 07:44:14.891530 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-hthsd"] Oct 05 07:44:15 crc kubenswrapper[4846]: I1005 07:44:15.021616 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wg2qc\" (UniqueName: \"kubernetes.io/projected/49b3f33d-e853-4d88-a56a-99e8e0e535df-kube-api-access-wg2qc\") pod \"redhat-marketplace-hthsd\" (UID: \"49b3f33d-e853-4d88-a56a-99e8e0e535df\") " pod="openshift-marketplace/redhat-marketplace-hthsd" Oct 05 07:44:15 crc kubenswrapper[4846]: I1005 07:44:15.021727 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/49b3f33d-e853-4d88-a56a-99e8e0e535df-utilities\") pod \"redhat-marketplace-hthsd\" (UID: \"49b3f33d-e853-4d88-a56a-99e8e0e535df\") " pod="openshift-marketplace/redhat-marketplace-hthsd" Oct 05 07:44:15 crc kubenswrapper[4846]: I1005 07:44:15.021755 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/49b3f33d-e853-4d88-a56a-99e8e0e535df-catalog-content\") pod \"redhat-marketplace-hthsd\" (UID: \"49b3f33d-e853-4d88-a56a-99e8e0e535df\") " pod="openshift-marketplace/redhat-marketplace-hthsd" Oct 05 07:44:15 crc kubenswrapper[4846]: I1005 07:44:15.122635 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/49b3f33d-e853-4d88-a56a-99e8e0e535df-utilities\") pod \"redhat-marketplace-hthsd\" (UID: \"49b3f33d-e853-4d88-a56a-99e8e0e535df\") " pod="openshift-marketplace/redhat-marketplace-hthsd" Oct 05 07:44:15 crc kubenswrapper[4846]: I1005 07:44:15.122693 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/49b3f33d-e853-4d88-a56a-99e8e0e535df-catalog-content\") pod \"redhat-marketplace-hthsd\" (UID: \"49b3f33d-e853-4d88-a56a-99e8e0e535df\") " pod="openshift-marketplace/redhat-marketplace-hthsd" Oct 05 07:44:15 crc kubenswrapper[4846]: I1005 07:44:15.122764 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wg2qc\" (UniqueName: \"kubernetes.io/projected/49b3f33d-e853-4d88-a56a-99e8e0e535df-kube-api-access-wg2qc\") pod \"redhat-marketplace-hthsd\" (UID: \"49b3f33d-e853-4d88-a56a-99e8e0e535df\") " pod="openshift-marketplace/redhat-marketplace-hthsd" Oct 05 07:44:15 crc kubenswrapper[4846]: I1005 07:44:15.123530 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/49b3f33d-e853-4d88-a56a-99e8e0e535df-utilities\") pod \"redhat-marketplace-hthsd\" (UID: \"49b3f33d-e853-4d88-a56a-99e8e0e535df\") " pod="openshift-marketplace/redhat-marketplace-hthsd" Oct 05 07:44:15 crc kubenswrapper[4846]: I1005 07:44:15.123606 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/49b3f33d-e853-4d88-a56a-99e8e0e535df-catalog-content\") pod \"redhat-marketplace-hthsd\" (UID: \"49b3f33d-e853-4d88-a56a-99e8e0e535df\") " pod="openshift-marketplace/redhat-marketplace-hthsd" Oct 05 07:44:15 crc kubenswrapper[4846]: I1005 07:44:15.151694 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wg2qc\" (UniqueName: \"kubernetes.io/projected/49b3f33d-e853-4d88-a56a-99e8e0e535df-kube-api-access-wg2qc\") pod \"redhat-marketplace-hthsd\" (UID: \"49b3f33d-e853-4d88-a56a-99e8e0e535df\") " pod="openshift-marketplace/redhat-marketplace-hthsd" Oct 05 07:44:15 crc kubenswrapper[4846]: I1005 07:44:15.202123 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-hthsd" Oct 05 07:44:15 crc kubenswrapper[4846]: I1005 07:44:15.698777 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-hthsd"] Oct 05 07:44:15 crc kubenswrapper[4846]: W1005 07:44:15.715533 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod49b3f33d_e853_4d88_a56a_99e8e0e535df.slice/crio-e5958a27091c10404c96ba342d90d9ff6c58795cab58c0de4825702be85a5e75 WatchSource:0}: Error finding container e5958a27091c10404c96ba342d90d9ff6c58795cab58c0de4825702be85a5e75: Status 404 returned error can't find the container with id e5958a27091c10404c96ba342d90d9ff6c58795cab58c0de4825702be85a5e75 Oct 05 07:44:15 crc kubenswrapper[4846]: I1005 07:44:15.859583 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hthsd" event={"ID":"49b3f33d-e853-4d88-a56a-99e8e0e535df","Type":"ContainerStarted","Data":"e5958a27091c10404c96ba342d90d9ff6c58795cab58c0de4825702be85a5e75"} Oct 05 07:44:16 crc kubenswrapper[4846]: I1005 07:44:16.884504 4846 generic.go:334] "Generic (PLEG): container finished" podID="49b3f33d-e853-4d88-a56a-99e8e0e535df" containerID="13c1b2b19000b6fbed76b1346f68f57edbf80588b4b478eb0a84bd46da70fe97" exitCode=0 Oct 05 07:44:16 crc kubenswrapper[4846]: I1005 07:44:16.884566 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hthsd" event={"ID":"49b3f33d-e853-4d88-a56a-99e8e0e535df","Type":"ContainerDied","Data":"13c1b2b19000b6fbed76b1346f68f57edbf80588b4b478eb0a84bd46da70fe97"} Oct 05 07:44:18 crc kubenswrapper[4846]: I1005 07:44:18.919995 4846 generic.go:334] "Generic (PLEG): container finished" podID="49b3f33d-e853-4d88-a56a-99e8e0e535df" containerID="354daf4166ddbb3bef7f60de6909fe06f6913bba273f5295a4db95c06fa5c466" exitCode=0 Oct 05 07:44:18 crc kubenswrapper[4846]: I1005 07:44:18.920427 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hthsd" event={"ID":"49b3f33d-e853-4d88-a56a-99e8e0e535df","Type":"ContainerDied","Data":"354daf4166ddbb3bef7f60de6909fe06f6913bba273f5295a4db95c06fa5c466"} Oct 05 07:44:19 crc kubenswrapper[4846]: I1005 07:44:19.934445 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hthsd" event={"ID":"49b3f33d-e853-4d88-a56a-99e8e0e535df","Type":"ContainerStarted","Data":"5001ce64b5807a77360f85619fc2f563d8ca1ee2214018cc4b027bae6bc9c106"} Oct 05 07:44:19 crc kubenswrapper[4846]: I1005 07:44:19.958584 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-hthsd" podStartSLOduration=3.235410271 podStartE2EDuration="5.95856847s" podCreationTimestamp="2025-10-05 07:44:14 +0000 UTC" firstStartedPulling="2025-10-05 07:44:16.886707726 +0000 UTC m=+3379.127560511" lastFinishedPulling="2025-10-05 07:44:19.609865935 +0000 UTC m=+3381.850718710" observedRunningTime="2025-10-05 07:44:19.953456643 +0000 UTC m=+3382.194309468" watchObservedRunningTime="2025-10-05 07:44:19.95856847 +0000 UTC m=+3382.199421245" Oct 05 07:44:23 crc kubenswrapper[4846]: I1005 07:44:23.325198 4846 patch_prober.go:28] interesting pod/machine-config-daemon-fscvf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 07:44:23 crc kubenswrapper[4846]: I1005 07:44:23.325526 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 07:44:25 crc kubenswrapper[4846]: I1005 07:44:25.202394 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-hthsd" Oct 05 07:44:25 crc kubenswrapper[4846]: I1005 07:44:25.203887 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-hthsd" Oct 05 07:44:25 crc kubenswrapper[4846]: I1005 07:44:25.282747 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-hthsd" Oct 05 07:44:26 crc kubenswrapper[4846]: I1005 07:44:26.066078 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-hthsd" Oct 05 07:44:26 crc kubenswrapper[4846]: I1005 07:44:26.154347 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-hthsd"] Oct 05 07:44:28 crc kubenswrapper[4846]: I1005 07:44:28.013945 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-hthsd" podUID="49b3f33d-e853-4d88-a56a-99e8e0e535df" containerName="registry-server" containerID="cri-o://5001ce64b5807a77360f85619fc2f563d8ca1ee2214018cc4b027bae6bc9c106" gracePeriod=2 Oct 05 07:44:28 crc kubenswrapper[4846]: I1005 07:44:28.545098 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-hthsd" Oct 05 07:44:28 crc kubenswrapper[4846]: I1005 07:44:28.645900 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/49b3f33d-e853-4d88-a56a-99e8e0e535df-catalog-content\") pod \"49b3f33d-e853-4d88-a56a-99e8e0e535df\" (UID: \"49b3f33d-e853-4d88-a56a-99e8e0e535df\") " Oct 05 07:44:28 crc kubenswrapper[4846]: I1005 07:44:28.646043 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wg2qc\" (UniqueName: \"kubernetes.io/projected/49b3f33d-e853-4d88-a56a-99e8e0e535df-kube-api-access-wg2qc\") pod \"49b3f33d-e853-4d88-a56a-99e8e0e535df\" (UID: \"49b3f33d-e853-4d88-a56a-99e8e0e535df\") " Oct 05 07:44:28 crc kubenswrapper[4846]: I1005 07:44:28.646074 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/49b3f33d-e853-4d88-a56a-99e8e0e535df-utilities\") pod \"49b3f33d-e853-4d88-a56a-99e8e0e535df\" (UID: \"49b3f33d-e853-4d88-a56a-99e8e0e535df\") " Oct 05 07:44:28 crc kubenswrapper[4846]: I1005 07:44:28.647757 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/49b3f33d-e853-4d88-a56a-99e8e0e535df-utilities" (OuterVolumeSpecName: "utilities") pod "49b3f33d-e853-4d88-a56a-99e8e0e535df" (UID: "49b3f33d-e853-4d88-a56a-99e8e0e535df"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 07:44:28 crc kubenswrapper[4846]: I1005 07:44:28.652778 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49b3f33d-e853-4d88-a56a-99e8e0e535df-kube-api-access-wg2qc" (OuterVolumeSpecName: "kube-api-access-wg2qc") pod "49b3f33d-e853-4d88-a56a-99e8e0e535df" (UID: "49b3f33d-e853-4d88-a56a-99e8e0e535df"). InnerVolumeSpecName "kube-api-access-wg2qc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:44:28 crc kubenswrapper[4846]: I1005 07:44:28.674881 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/49b3f33d-e853-4d88-a56a-99e8e0e535df-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "49b3f33d-e853-4d88-a56a-99e8e0e535df" (UID: "49b3f33d-e853-4d88-a56a-99e8e0e535df"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 07:44:28 crc kubenswrapper[4846]: I1005 07:44:28.747548 4846 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/49b3f33d-e853-4d88-a56a-99e8e0e535df-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 05 07:44:28 crc kubenswrapper[4846]: I1005 07:44:28.747584 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wg2qc\" (UniqueName: \"kubernetes.io/projected/49b3f33d-e853-4d88-a56a-99e8e0e535df-kube-api-access-wg2qc\") on node \"crc\" DevicePath \"\"" Oct 05 07:44:28 crc kubenswrapper[4846]: I1005 07:44:28.747598 4846 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/49b3f33d-e853-4d88-a56a-99e8e0e535df-utilities\") on node \"crc\" DevicePath \"\"" Oct 05 07:44:29 crc kubenswrapper[4846]: I1005 07:44:29.027118 4846 generic.go:334] "Generic (PLEG): container finished" podID="49b3f33d-e853-4d88-a56a-99e8e0e535df" containerID="5001ce64b5807a77360f85619fc2f563d8ca1ee2214018cc4b027bae6bc9c106" exitCode=0 Oct 05 07:44:29 crc kubenswrapper[4846]: I1005 07:44:29.027206 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hthsd" event={"ID":"49b3f33d-e853-4d88-a56a-99e8e0e535df","Type":"ContainerDied","Data":"5001ce64b5807a77360f85619fc2f563d8ca1ee2214018cc4b027bae6bc9c106"} Oct 05 07:44:29 crc kubenswrapper[4846]: I1005 07:44:29.027249 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hthsd" event={"ID":"49b3f33d-e853-4d88-a56a-99e8e0e535df","Type":"ContainerDied","Data":"e5958a27091c10404c96ba342d90d9ff6c58795cab58c0de4825702be85a5e75"} Oct 05 07:44:29 crc kubenswrapper[4846]: I1005 07:44:29.027249 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-hthsd" Oct 05 07:44:29 crc kubenswrapper[4846]: I1005 07:44:29.027275 4846 scope.go:117] "RemoveContainer" containerID="5001ce64b5807a77360f85619fc2f563d8ca1ee2214018cc4b027bae6bc9c106" Oct 05 07:44:29 crc kubenswrapper[4846]: I1005 07:44:29.055558 4846 scope.go:117] "RemoveContainer" containerID="354daf4166ddbb3bef7f60de6909fe06f6913bba273f5295a4db95c06fa5c466" Oct 05 07:44:29 crc kubenswrapper[4846]: I1005 07:44:29.094959 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-hthsd"] Oct 05 07:44:29 crc kubenswrapper[4846]: I1005 07:44:29.096585 4846 scope.go:117] "RemoveContainer" containerID="13c1b2b19000b6fbed76b1346f68f57edbf80588b4b478eb0a84bd46da70fe97" Oct 05 07:44:29 crc kubenswrapper[4846]: I1005 07:44:29.112808 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-hthsd"] Oct 05 07:44:29 crc kubenswrapper[4846]: I1005 07:44:29.120981 4846 scope.go:117] "RemoveContainer" containerID="5001ce64b5807a77360f85619fc2f563d8ca1ee2214018cc4b027bae6bc9c106" Oct 05 07:44:29 crc kubenswrapper[4846]: E1005 07:44:29.121602 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5001ce64b5807a77360f85619fc2f563d8ca1ee2214018cc4b027bae6bc9c106\": container with ID starting with 5001ce64b5807a77360f85619fc2f563d8ca1ee2214018cc4b027bae6bc9c106 not found: ID does not exist" containerID="5001ce64b5807a77360f85619fc2f563d8ca1ee2214018cc4b027bae6bc9c106" Oct 05 07:44:29 crc kubenswrapper[4846]: I1005 07:44:29.121630 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5001ce64b5807a77360f85619fc2f563d8ca1ee2214018cc4b027bae6bc9c106"} err="failed to get container status \"5001ce64b5807a77360f85619fc2f563d8ca1ee2214018cc4b027bae6bc9c106\": rpc error: code = NotFound desc = could not find container \"5001ce64b5807a77360f85619fc2f563d8ca1ee2214018cc4b027bae6bc9c106\": container with ID starting with 5001ce64b5807a77360f85619fc2f563d8ca1ee2214018cc4b027bae6bc9c106 not found: ID does not exist" Oct 05 07:44:29 crc kubenswrapper[4846]: I1005 07:44:29.121650 4846 scope.go:117] "RemoveContainer" containerID="354daf4166ddbb3bef7f60de6909fe06f6913bba273f5295a4db95c06fa5c466" Oct 05 07:44:29 crc kubenswrapper[4846]: E1005 07:44:29.121917 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"354daf4166ddbb3bef7f60de6909fe06f6913bba273f5295a4db95c06fa5c466\": container with ID starting with 354daf4166ddbb3bef7f60de6909fe06f6913bba273f5295a4db95c06fa5c466 not found: ID does not exist" containerID="354daf4166ddbb3bef7f60de6909fe06f6913bba273f5295a4db95c06fa5c466" Oct 05 07:44:29 crc kubenswrapper[4846]: I1005 07:44:29.121941 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"354daf4166ddbb3bef7f60de6909fe06f6913bba273f5295a4db95c06fa5c466"} err="failed to get container status \"354daf4166ddbb3bef7f60de6909fe06f6913bba273f5295a4db95c06fa5c466\": rpc error: code = NotFound desc = could not find container \"354daf4166ddbb3bef7f60de6909fe06f6913bba273f5295a4db95c06fa5c466\": container with ID starting with 354daf4166ddbb3bef7f60de6909fe06f6913bba273f5295a4db95c06fa5c466 not found: ID does not exist" Oct 05 07:44:29 crc kubenswrapper[4846]: I1005 07:44:29.121955 4846 scope.go:117] "RemoveContainer" containerID="13c1b2b19000b6fbed76b1346f68f57edbf80588b4b478eb0a84bd46da70fe97" Oct 05 07:44:29 crc kubenswrapper[4846]: E1005 07:44:29.122132 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"13c1b2b19000b6fbed76b1346f68f57edbf80588b4b478eb0a84bd46da70fe97\": container with ID starting with 13c1b2b19000b6fbed76b1346f68f57edbf80588b4b478eb0a84bd46da70fe97 not found: ID does not exist" containerID="13c1b2b19000b6fbed76b1346f68f57edbf80588b4b478eb0a84bd46da70fe97" Oct 05 07:44:29 crc kubenswrapper[4846]: I1005 07:44:29.122159 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"13c1b2b19000b6fbed76b1346f68f57edbf80588b4b478eb0a84bd46da70fe97"} err="failed to get container status \"13c1b2b19000b6fbed76b1346f68f57edbf80588b4b478eb0a84bd46da70fe97\": rpc error: code = NotFound desc = could not find container \"13c1b2b19000b6fbed76b1346f68f57edbf80588b4b478eb0a84bd46da70fe97\": container with ID starting with 13c1b2b19000b6fbed76b1346f68f57edbf80588b4b478eb0a84bd46da70fe97 not found: ID does not exist" Oct 05 07:44:30 crc kubenswrapper[4846]: I1005 07:44:30.516592 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49b3f33d-e853-4d88-a56a-99e8e0e535df" path="/var/lib/kubelet/pods/49b3f33d-e853-4d88-a56a-99e8e0e535df/volumes" Oct 05 07:44:53 crc kubenswrapper[4846]: I1005 07:44:53.324856 4846 patch_prober.go:28] interesting pod/machine-config-daemon-fscvf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 07:44:53 crc kubenswrapper[4846]: I1005 07:44:53.325958 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 07:44:53 crc kubenswrapper[4846]: I1005 07:44:53.326034 4846 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" Oct 05 07:44:53 crc kubenswrapper[4846]: I1005 07:44:53.327065 4846 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"725a93409ff4757fae7733195dbbc6de2ee38624fdbf934d0b72e1d2e3418171"} pod="openshift-machine-config-operator/machine-config-daemon-fscvf" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 05 07:44:53 crc kubenswrapper[4846]: I1005 07:44:53.327166 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" containerName="machine-config-daemon" containerID="cri-o://725a93409ff4757fae7733195dbbc6de2ee38624fdbf934d0b72e1d2e3418171" gracePeriod=600 Oct 05 07:44:54 crc kubenswrapper[4846]: I1005 07:44:54.285065 4846 generic.go:334] "Generic (PLEG): container finished" podID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" containerID="725a93409ff4757fae7733195dbbc6de2ee38624fdbf934d0b72e1d2e3418171" exitCode=0 Oct 05 07:44:54 crc kubenswrapper[4846]: I1005 07:44:54.285168 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" event={"ID":"ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7","Type":"ContainerDied","Data":"725a93409ff4757fae7733195dbbc6de2ee38624fdbf934d0b72e1d2e3418171"} Oct 05 07:44:54 crc kubenswrapper[4846]: I1005 07:44:54.285506 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" event={"ID":"ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7","Type":"ContainerStarted","Data":"a8dd7af9a49a3eba2b3ea4d7586f51b807c63a3b657f800b5182c8288f292750"} Oct 05 07:44:54 crc kubenswrapper[4846]: I1005 07:44:54.285538 4846 scope.go:117] "RemoveContainer" containerID="702f9b90152900919451502be251dfa1fc91d3fc2af6439c709b401f8ac5d520" Oct 05 07:45:00 crc kubenswrapper[4846]: I1005 07:45:00.198710 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29327505-gczm6"] Oct 05 07:45:00 crc kubenswrapper[4846]: E1005 07:45:00.199637 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="49b3f33d-e853-4d88-a56a-99e8e0e535df" containerName="registry-server" Oct 05 07:45:00 crc kubenswrapper[4846]: I1005 07:45:00.199653 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="49b3f33d-e853-4d88-a56a-99e8e0e535df" containerName="registry-server" Oct 05 07:45:00 crc kubenswrapper[4846]: E1005 07:45:00.199665 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="49b3f33d-e853-4d88-a56a-99e8e0e535df" containerName="extract-utilities" Oct 05 07:45:00 crc kubenswrapper[4846]: I1005 07:45:00.199673 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="49b3f33d-e853-4d88-a56a-99e8e0e535df" containerName="extract-utilities" Oct 05 07:45:00 crc kubenswrapper[4846]: E1005 07:45:00.199686 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="49b3f33d-e853-4d88-a56a-99e8e0e535df" containerName="extract-content" Oct 05 07:45:00 crc kubenswrapper[4846]: I1005 07:45:00.199697 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="49b3f33d-e853-4d88-a56a-99e8e0e535df" containerName="extract-content" Oct 05 07:45:00 crc kubenswrapper[4846]: I1005 07:45:00.199889 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="49b3f33d-e853-4d88-a56a-99e8e0e535df" containerName="registry-server" Oct 05 07:45:00 crc kubenswrapper[4846]: I1005 07:45:00.200528 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29327505-gczm6" Oct 05 07:45:00 crc kubenswrapper[4846]: I1005 07:45:00.202619 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 05 07:45:00 crc kubenswrapper[4846]: I1005 07:45:00.202744 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 05 07:45:00 crc kubenswrapper[4846]: I1005 07:45:00.212833 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29327505-gczm6"] Oct 05 07:45:00 crc kubenswrapper[4846]: I1005 07:45:00.281172 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9k8mt\" (UniqueName: \"kubernetes.io/projected/a4d2ee77-b77c-4ff4-a27a-164efbf89c7a-kube-api-access-9k8mt\") pod \"collect-profiles-29327505-gczm6\" (UID: \"a4d2ee77-b77c-4ff4-a27a-164efbf89c7a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327505-gczm6" Oct 05 07:45:00 crc kubenswrapper[4846]: I1005 07:45:00.281267 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a4d2ee77-b77c-4ff4-a27a-164efbf89c7a-secret-volume\") pod \"collect-profiles-29327505-gczm6\" (UID: \"a4d2ee77-b77c-4ff4-a27a-164efbf89c7a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327505-gczm6" Oct 05 07:45:00 crc kubenswrapper[4846]: I1005 07:45:00.281328 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a4d2ee77-b77c-4ff4-a27a-164efbf89c7a-config-volume\") pod \"collect-profiles-29327505-gczm6\" (UID: \"a4d2ee77-b77c-4ff4-a27a-164efbf89c7a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327505-gczm6" Oct 05 07:45:00 crc kubenswrapper[4846]: I1005 07:45:00.382216 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9k8mt\" (UniqueName: \"kubernetes.io/projected/a4d2ee77-b77c-4ff4-a27a-164efbf89c7a-kube-api-access-9k8mt\") pod \"collect-profiles-29327505-gczm6\" (UID: \"a4d2ee77-b77c-4ff4-a27a-164efbf89c7a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327505-gczm6" Oct 05 07:45:00 crc kubenswrapper[4846]: I1005 07:45:00.382352 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a4d2ee77-b77c-4ff4-a27a-164efbf89c7a-secret-volume\") pod \"collect-profiles-29327505-gczm6\" (UID: \"a4d2ee77-b77c-4ff4-a27a-164efbf89c7a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327505-gczm6" Oct 05 07:45:00 crc kubenswrapper[4846]: I1005 07:45:00.382436 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a4d2ee77-b77c-4ff4-a27a-164efbf89c7a-config-volume\") pod \"collect-profiles-29327505-gczm6\" (UID: \"a4d2ee77-b77c-4ff4-a27a-164efbf89c7a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327505-gczm6" Oct 05 07:45:00 crc kubenswrapper[4846]: I1005 07:45:00.384814 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a4d2ee77-b77c-4ff4-a27a-164efbf89c7a-config-volume\") pod \"collect-profiles-29327505-gczm6\" (UID: \"a4d2ee77-b77c-4ff4-a27a-164efbf89c7a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327505-gczm6" Oct 05 07:45:00 crc kubenswrapper[4846]: I1005 07:45:00.393475 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a4d2ee77-b77c-4ff4-a27a-164efbf89c7a-secret-volume\") pod \"collect-profiles-29327505-gczm6\" (UID: \"a4d2ee77-b77c-4ff4-a27a-164efbf89c7a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327505-gczm6" Oct 05 07:45:00 crc kubenswrapper[4846]: I1005 07:45:00.404323 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9k8mt\" (UniqueName: \"kubernetes.io/projected/a4d2ee77-b77c-4ff4-a27a-164efbf89c7a-kube-api-access-9k8mt\") pod \"collect-profiles-29327505-gczm6\" (UID: \"a4d2ee77-b77c-4ff4-a27a-164efbf89c7a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327505-gczm6" Oct 05 07:45:00 crc kubenswrapper[4846]: I1005 07:45:00.523029 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29327505-gczm6" Oct 05 07:45:01 crc kubenswrapper[4846]: W1005 07:45:01.037001 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda4d2ee77_b77c_4ff4_a27a_164efbf89c7a.slice/crio-49950b0b02418d5f92f7830cac9136bf48989de932da9125586221f45f55c8f7 WatchSource:0}: Error finding container 49950b0b02418d5f92f7830cac9136bf48989de932da9125586221f45f55c8f7: Status 404 returned error can't find the container with id 49950b0b02418d5f92f7830cac9136bf48989de932da9125586221f45f55c8f7 Oct 05 07:45:01 crc kubenswrapper[4846]: I1005 07:45:01.039247 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29327505-gczm6"] Oct 05 07:45:01 crc kubenswrapper[4846]: I1005 07:45:01.349847 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29327505-gczm6" event={"ID":"a4d2ee77-b77c-4ff4-a27a-164efbf89c7a","Type":"ContainerStarted","Data":"2aaed7a87773121e7ebff2f4276b7f6b86cd9bf885d338e114304462977b8737"} Oct 05 07:45:01 crc kubenswrapper[4846]: I1005 07:45:01.349883 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29327505-gczm6" event={"ID":"a4d2ee77-b77c-4ff4-a27a-164efbf89c7a","Type":"ContainerStarted","Data":"49950b0b02418d5f92f7830cac9136bf48989de932da9125586221f45f55c8f7"} Oct 05 07:45:01 crc kubenswrapper[4846]: I1005 07:45:01.373039 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29327505-gczm6" podStartSLOduration=1.373023405 podStartE2EDuration="1.373023405s" podCreationTimestamp="2025-10-05 07:45:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 07:45:01.368392081 +0000 UTC m=+3423.609244866" watchObservedRunningTime="2025-10-05 07:45:01.373023405 +0000 UTC m=+3423.613876180" Oct 05 07:45:02 crc kubenswrapper[4846]: I1005 07:45:02.361222 4846 generic.go:334] "Generic (PLEG): container finished" podID="a4d2ee77-b77c-4ff4-a27a-164efbf89c7a" containerID="2aaed7a87773121e7ebff2f4276b7f6b86cd9bf885d338e114304462977b8737" exitCode=0 Oct 05 07:45:02 crc kubenswrapper[4846]: I1005 07:45:02.361352 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29327505-gczm6" event={"ID":"a4d2ee77-b77c-4ff4-a27a-164efbf89c7a","Type":"ContainerDied","Data":"2aaed7a87773121e7ebff2f4276b7f6b86cd9bf885d338e114304462977b8737"} Oct 05 07:45:03 crc kubenswrapper[4846]: I1005 07:45:03.703019 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29327505-gczm6" Oct 05 07:45:03 crc kubenswrapper[4846]: I1005 07:45:03.733929 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a4d2ee77-b77c-4ff4-a27a-164efbf89c7a-secret-volume\") pod \"a4d2ee77-b77c-4ff4-a27a-164efbf89c7a\" (UID: \"a4d2ee77-b77c-4ff4-a27a-164efbf89c7a\") " Oct 05 07:45:03 crc kubenswrapper[4846]: I1005 07:45:03.734140 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9k8mt\" (UniqueName: \"kubernetes.io/projected/a4d2ee77-b77c-4ff4-a27a-164efbf89c7a-kube-api-access-9k8mt\") pod \"a4d2ee77-b77c-4ff4-a27a-164efbf89c7a\" (UID: \"a4d2ee77-b77c-4ff4-a27a-164efbf89c7a\") " Oct 05 07:45:03 crc kubenswrapper[4846]: I1005 07:45:03.734242 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a4d2ee77-b77c-4ff4-a27a-164efbf89c7a-config-volume\") pod \"a4d2ee77-b77c-4ff4-a27a-164efbf89c7a\" (UID: \"a4d2ee77-b77c-4ff4-a27a-164efbf89c7a\") " Oct 05 07:45:03 crc kubenswrapper[4846]: I1005 07:45:03.734931 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a4d2ee77-b77c-4ff4-a27a-164efbf89c7a-config-volume" (OuterVolumeSpecName: "config-volume") pod "a4d2ee77-b77c-4ff4-a27a-164efbf89c7a" (UID: "a4d2ee77-b77c-4ff4-a27a-164efbf89c7a"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 07:45:03 crc kubenswrapper[4846]: I1005 07:45:03.740628 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a4d2ee77-b77c-4ff4-a27a-164efbf89c7a-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "a4d2ee77-b77c-4ff4-a27a-164efbf89c7a" (UID: "a4d2ee77-b77c-4ff4-a27a-164efbf89c7a"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 07:45:03 crc kubenswrapper[4846]: I1005 07:45:03.740831 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a4d2ee77-b77c-4ff4-a27a-164efbf89c7a-kube-api-access-9k8mt" (OuterVolumeSpecName: "kube-api-access-9k8mt") pod "a4d2ee77-b77c-4ff4-a27a-164efbf89c7a" (UID: "a4d2ee77-b77c-4ff4-a27a-164efbf89c7a"). InnerVolumeSpecName "kube-api-access-9k8mt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:45:03 crc kubenswrapper[4846]: I1005 07:45:03.835950 4846 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a4d2ee77-b77c-4ff4-a27a-164efbf89c7a-config-volume\") on node \"crc\" DevicePath \"\"" Oct 05 07:45:03 crc kubenswrapper[4846]: I1005 07:45:03.835984 4846 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a4d2ee77-b77c-4ff4-a27a-164efbf89c7a-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 05 07:45:03 crc kubenswrapper[4846]: I1005 07:45:03.835999 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9k8mt\" (UniqueName: \"kubernetes.io/projected/a4d2ee77-b77c-4ff4-a27a-164efbf89c7a-kube-api-access-9k8mt\") on node \"crc\" DevicePath \"\"" Oct 05 07:45:04 crc kubenswrapper[4846]: I1005 07:45:04.380013 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29327505-gczm6" event={"ID":"a4d2ee77-b77c-4ff4-a27a-164efbf89c7a","Type":"ContainerDied","Data":"49950b0b02418d5f92f7830cac9136bf48989de932da9125586221f45f55c8f7"} Oct 05 07:45:04 crc kubenswrapper[4846]: I1005 07:45:04.380068 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="49950b0b02418d5f92f7830cac9136bf48989de932da9125586221f45f55c8f7" Oct 05 07:45:04 crc kubenswrapper[4846]: I1005 07:45:04.380159 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29327505-gczm6" Oct 05 07:45:04 crc kubenswrapper[4846]: I1005 07:45:04.454249 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29327460-x7z7p"] Oct 05 07:45:04 crc kubenswrapper[4846]: I1005 07:45:04.459274 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29327460-x7z7p"] Oct 05 07:45:04 crc kubenswrapper[4846]: I1005 07:45:04.505954 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6fc79ab2-6284-412f-b641-01e7d1f4b1c6" path="/var/lib/kubelet/pods/6fc79ab2-6284-412f-b641-01e7d1f4b1c6/volumes" Oct 05 07:45:23 crc kubenswrapper[4846]: I1005 07:45:23.916098 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-zbdlf"] Oct 05 07:45:23 crc kubenswrapper[4846]: E1005 07:45:23.917605 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a4d2ee77-b77c-4ff4-a27a-164efbf89c7a" containerName="collect-profiles" Oct 05 07:45:23 crc kubenswrapper[4846]: I1005 07:45:23.917643 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="a4d2ee77-b77c-4ff4-a27a-164efbf89c7a" containerName="collect-profiles" Oct 05 07:45:23 crc kubenswrapper[4846]: I1005 07:45:23.917997 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="a4d2ee77-b77c-4ff4-a27a-164efbf89c7a" containerName="collect-profiles" Oct 05 07:45:23 crc kubenswrapper[4846]: I1005 07:45:23.923854 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-zbdlf" Oct 05 07:45:23 crc kubenswrapper[4846]: I1005 07:45:23.925235 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-zbdlf"] Oct 05 07:45:24 crc kubenswrapper[4846]: I1005 07:45:24.042105 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fzfdz\" (UniqueName: \"kubernetes.io/projected/f29bf0d4-54c4-4d11-87e3-d153d5f3fbb8-kube-api-access-fzfdz\") pod \"redhat-operators-zbdlf\" (UID: \"f29bf0d4-54c4-4d11-87e3-d153d5f3fbb8\") " pod="openshift-marketplace/redhat-operators-zbdlf" Oct 05 07:45:24 crc kubenswrapper[4846]: I1005 07:45:24.042212 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f29bf0d4-54c4-4d11-87e3-d153d5f3fbb8-catalog-content\") pod \"redhat-operators-zbdlf\" (UID: \"f29bf0d4-54c4-4d11-87e3-d153d5f3fbb8\") " pod="openshift-marketplace/redhat-operators-zbdlf" Oct 05 07:45:24 crc kubenswrapper[4846]: I1005 07:45:24.042298 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f29bf0d4-54c4-4d11-87e3-d153d5f3fbb8-utilities\") pod \"redhat-operators-zbdlf\" (UID: \"f29bf0d4-54c4-4d11-87e3-d153d5f3fbb8\") " pod="openshift-marketplace/redhat-operators-zbdlf" Oct 05 07:45:24 crc kubenswrapper[4846]: I1005 07:45:24.143904 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f29bf0d4-54c4-4d11-87e3-d153d5f3fbb8-utilities\") pod \"redhat-operators-zbdlf\" (UID: \"f29bf0d4-54c4-4d11-87e3-d153d5f3fbb8\") " pod="openshift-marketplace/redhat-operators-zbdlf" Oct 05 07:45:24 crc kubenswrapper[4846]: I1005 07:45:24.143970 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fzfdz\" (UniqueName: \"kubernetes.io/projected/f29bf0d4-54c4-4d11-87e3-d153d5f3fbb8-kube-api-access-fzfdz\") pod \"redhat-operators-zbdlf\" (UID: \"f29bf0d4-54c4-4d11-87e3-d153d5f3fbb8\") " pod="openshift-marketplace/redhat-operators-zbdlf" Oct 05 07:45:24 crc kubenswrapper[4846]: I1005 07:45:24.144007 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f29bf0d4-54c4-4d11-87e3-d153d5f3fbb8-catalog-content\") pod \"redhat-operators-zbdlf\" (UID: \"f29bf0d4-54c4-4d11-87e3-d153d5f3fbb8\") " pod="openshift-marketplace/redhat-operators-zbdlf" Oct 05 07:45:24 crc kubenswrapper[4846]: I1005 07:45:24.144581 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f29bf0d4-54c4-4d11-87e3-d153d5f3fbb8-utilities\") pod \"redhat-operators-zbdlf\" (UID: \"f29bf0d4-54c4-4d11-87e3-d153d5f3fbb8\") " pod="openshift-marketplace/redhat-operators-zbdlf" Oct 05 07:45:24 crc kubenswrapper[4846]: I1005 07:45:24.144601 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f29bf0d4-54c4-4d11-87e3-d153d5f3fbb8-catalog-content\") pod \"redhat-operators-zbdlf\" (UID: \"f29bf0d4-54c4-4d11-87e3-d153d5f3fbb8\") " pod="openshift-marketplace/redhat-operators-zbdlf" Oct 05 07:45:24 crc kubenswrapper[4846]: I1005 07:45:24.175539 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fzfdz\" (UniqueName: \"kubernetes.io/projected/f29bf0d4-54c4-4d11-87e3-d153d5f3fbb8-kube-api-access-fzfdz\") pod \"redhat-operators-zbdlf\" (UID: \"f29bf0d4-54c4-4d11-87e3-d153d5f3fbb8\") " pod="openshift-marketplace/redhat-operators-zbdlf" Oct 05 07:45:24 crc kubenswrapper[4846]: I1005 07:45:24.267725 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-zbdlf" Oct 05 07:45:24 crc kubenswrapper[4846]: I1005 07:45:24.688094 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-zbdlf"] Oct 05 07:45:25 crc kubenswrapper[4846]: I1005 07:45:25.588468 4846 generic.go:334] "Generic (PLEG): container finished" podID="f29bf0d4-54c4-4d11-87e3-d153d5f3fbb8" containerID="c6e05032bd4a96b4de60fc1077adf019fe4093876cbf4ca68bb29782f395da36" exitCode=0 Oct 05 07:45:25 crc kubenswrapper[4846]: I1005 07:45:25.588571 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zbdlf" event={"ID":"f29bf0d4-54c4-4d11-87e3-d153d5f3fbb8","Type":"ContainerDied","Data":"c6e05032bd4a96b4de60fc1077adf019fe4093876cbf4ca68bb29782f395da36"} Oct 05 07:45:25 crc kubenswrapper[4846]: I1005 07:45:25.591065 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zbdlf" event={"ID":"f29bf0d4-54c4-4d11-87e3-d153d5f3fbb8","Type":"ContainerStarted","Data":"031302afa19c4f978c145937c7270a72aabbdb58e2ac2b25ecb295dfc29325e7"} Oct 05 07:45:27 crc kubenswrapper[4846]: I1005 07:45:27.609081 4846 generic.go:334] "Generic (PLEG): container finished" podID="f29bf0d4-54c4-4d11-87e3-d153d5f3fbb8" containerID="7d8f27b872a19c60802af1911701bd97e0d3d2bb93d5f28d7ee16ce6ce7fac52" exitCode=0 Oct 05 07:45:27 crc kubenswrapper[4846]: I1005 07:45:27.609249 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zbdlf" event={"ID":"f29bf0d4-54c4-4d11-87e3-d153d5f3fbb8","Type":"ContainerDied","Data":"7d8f27b872a19c60802af1911701bd97e0d3d2bb93d5f28d7ee16ce6ce7fac52"} Oct 05 07:45:28 crc kubenswrapper[4846]: I1005 07:45:28.620256 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zbdlf" event={"ID":"f29bf0d4-54c4-4d11-87e3-d153d5f3fbb8","Type":"ContainerStarted","Data":"e84efa9ac80ddac17f300cdb3eabe6354816ee7dc895886f49f311b75eec455a"} Oct 05 07:45:28 crc kubenswrapper[4846]: I1005 07:45:28.644755 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-zbdlf" podStartSLOduration=3.22141924 podStartE2EDuration="5.644731985s" podCreationTimestamp="2025-10-05 07:45:23 +0000 UTC" firstStartedPulling="2025-10-05 07:45:25.590546354 +0000 UTC m=+3447.831399149" lastFinishedPulling="2025-10-05 07:45:28.013859119 +0000 UTC m=+3450.254711894" observedRunningTime="2025-10-05 07:45:28.641735374 +0000 UTC m=+3450.882588159" watchObservedRunningTime="2025-10-05 07:45:28.644731985 +0000 UTC m=+3450.885584770" Oct 05 07:45:29 crc kubenswrapper[4846]: I1005 07:45:29.497056 4846 scope.go:117] "RemoveContainer" containerID="4d929a7b2e9007b4930575af61db05d35381bb42b7507051e5aadda242a701bb" Oct 05 07:45:34 crc kubenswrapper[4846]: I1005 07:45:34.268290 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-zbdlf" Oct 05 07:45:34 crc kubenswrapper[4846]: I1005 07:45:34.269689 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-zbdlf" Oct 05 07:45:34 crc kubenswrapper[4846]: I1005 07:45:34.346344 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-zbdlf" Oct 05 07:45:34 crc kubenswrapper[4846]: I1005 07:45:34.759261 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-zbdlf" Oct 05 07:45:34 crc kubenswrapper[4846]: I1005 07:45:34.823209 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-zbdlf"] Oct 05 07:45:36 crc kubenswrapper[4846]: I1005 07:45:36.704954 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-zbdlf" podUID="f29bf0d4-54c4-4d11-87e3-d153d5f3fbb8" containerName="registry-server" containerID="cri-o://e84efa9ac80ddac17f300cdb3eabe6354816ee7dc895886f49f311b75eec455a" gracePeriod=2 Oct 05 07:45:37 crc kubenswrapper[4846]: I1005 07:45:37.231387 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-zbdlf" Oct 05 07:45:37 crc kubenswrapper[4846]: I1005 07:45:37.273138 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f29bf0d4-54c4-4d11-87e3-d153d5f3fbb8-utilities\") pod \"f29bf0d4-54c4-4d11-87e3-d153d5f3fbb8\" (UID: \"f29bf0d4-54c4-4d11-87e3-d153d5f3fbb8\") " Oct 05 07:45:37 crc kubenswrapper[4846]: I1005 07:45:37.273317 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f29bf0d4-54c4-4d11-87e3-d153d5f3fbb8-catalog-content\") pod \"f29bf0d4-54c4-4d11-87e3-d153d5f3fbb8\" (UID: \"f29bf0d4-54c4-4d11-87e3-d153d5f3fbb8\") " Oct 05 07:45:37 crc kubenswrapper[4846]: I1005 07:45:37.274204 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f29bf0d4-54c4-4d11-87e3-d153d5f3fbb8-utilities" (OuterVolumeSpecName: "utilities") pod "f29bf0d4-54c4-4d11-87e3-d153d5f3fbb8" (UID: "f29bf0d4-54c4-4d11-87e3-d153d5f3fbb8"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 07:45:37 crc kubenswrapper[4846]: I1005 07:45:37.374968 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fzfdz\" (UniqueName: \"kubernetes.io/projected/f29bf0d4-54c4-4d11-87e3-d153d5f3fbb8-kube-api-access-fzfdz\") pod \"f29bf0d4-54c4-4d11-87e3-d153d5f3fbb8\" (UID: \"f29bf0d4-54c4-4d11-87e3-d153d5f3fbb8\") " Oct 05 07:45:37 crc kubenswrapper[4846]: I1005 07:45:37.376857 4846 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f29bf0d4-54c4-4d11-87e3-d153d5f3fbb8-utilities\") on node \"crc\" DevicePath \"\"" Oct 05 07:45:37 crc kubenswrapper[4846]: I1005 07:45:37.381798 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f29bf0d4-54c4-4d11-87e3-d153d5f3fbb8-kube-api-access-fzfdz" (OuterVolumeSpecName: "kube-api-access-fzfdz") pod "f29bf0d4-54c4-4d11-87e3-d153d5f3fbb8" (UID: "f29bf0d4-54c4-4d11-87e3-d153d5f3fbb8"). InnerVolumeSpecName "kube-api-access-fzfdz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:45:37 crc kubenswrapper[4846]: I1005 07:45:37.406443 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f29bf0d4-54c4-4d11-87e3-d153d5f3fbb8-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f29bf0d4-54c4-4d11-87e3-d153d5f3fbb8" (UID: "f29bf0d4-54c4-4d11-87e3-d153d5f3fbb8"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 07:45:37 crc kubenswrapper[4846]: I1005 07:45:37.479267 4846 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f29bf0d4-54c4-4d11-87e3-d153d5f3fbb8-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 05 07:45:37 crc kubenswrapper[4846]: I1005 07:45:37.479373 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fzfdz\" (UniqueName: \"kubernetes.io/projected/f29bf0d4-54c4-4d11-87e3-d153d5f3fbb8-kube-api-access-fzfdz\") on node \"crc\" DevicePath \"\"" Oct 05 07:45:37 crc kubenswrapper[4846]: I1005 07:45:37.719689 4846 generic.go:334] "Generic (PLEG): container finished" podID="f29bf0d4-54c4-4d11-87e3-d153d5f3fbb8" containerID="e84efa9ac80ddac17f300cdb3eabe6354816ee7dc895886f49f311b75eec455a" exitCode=0 Oct 05 07:45:37 crc kubenswrapper[4846]: I1005 07:45:37.719761 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zbdlf" event={"ID":"f29bf0d4-54c4-4d11-87e3-d153d5f3fbb8","Type":"ContainerDied","Data":"e84efa9ac80ddac17f300cdb3eabe6354816ee7dc895886f49f311b75eec455a"} Oct 05 07:45:37 crc kubenswrapper[4846]: I1005 07:45:37.719815 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-zbdlf" Oct 05 07:45:37 crc kubenswrapper[4846]: I1005 07:45:37.719811 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zbdlf" event={"ID":"f29bf0d4-54c4-4d11-87e3-d153d5f3fbb8","Type":"ContainerDied","Data":"031302afa19c4f978c145937c7270a72aabbdb58e2ac2b25ecb295dfc29325e7"} Oct 05 07:45:37 crc kubenswrapper[4846]: I1005 07:45:37.719946 4846 scope.go:117] "RemoveContainer" containerID="e84efa9ac80ddac17f300cdb3eabe6354816ee7dc895886f49f311b75eec455a" Oct 05 07:45:37 crc kubenswrapper[4846]: I1005 07:45:37.753706 4846 scope.go:117] "RemoveContainer" containerID="7d8f27b872a19c60802af1911701bd97e0d3d2bb93d5f28d7ee16ce6ce7fac52" Oct 05 07:45:37 crc kubenswrapper[4846]: I1005 07:45:37.787498 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-zbdlf"] Oct 05 07:45:37 crc kubenswrapper[4846]: I1005 07:45:37.795561 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-zbdlf"] Oct 05 07:45:37 crc kubenswrapper[4846]: I1005 07:45:37.796290 4846 scope.go:117] "RemoveContainer" containerID="c6e05032bd4a96b4de60fc1077adf019fe4093876cbf4ca68bb29782f395da36" Oct 05 07:45:37 crc kubenswrapper[4846]: I1005 07:45:37.842498 4846 scope.go:117] "RemoveContainer" containerID="e84efa9ac80ddac17f300cdb3eabe6354816ee7dc895886f49f311b75eec455a" Oct 05 07:45:37 crc kubenswrapper[4846]: E1005 07:45:37.843169 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e84efa9ac80ddac17f300cdb3eabe6354816ee7dc895886f49f311b75eec455a\": container with ID starting with e84efa9ac80ddac17f300cdb3eabe6354816ee7dc895886f49f311b75eec455a not found: ID does not exist" containerID="e84efa9ac80ddac17f300cdb3eabe6354816ee7dc895886f49f311b75eec455a" Oct 05 07:45:37 crc kubenswrapper[4846]: I1005 07:45:37.843255 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e84efa9ac80ddac17f300cdb3eabe6354816ee7dc895886f49f311b75eec455a"} err="failed to get container status \"e84efa9ac80ddac17f300cdb3eabe6354816ee7dc895886f49f311b75eec455a\": rpc error: code = NotFound desc = could not find container \"e84efa9ac80ddac17f300cdb3eabe6354816ee7dc895886f49f311b75eec455a\": container with ID starting with e84efa9ac80ddac17f300cdb3eabe6354816ee7dc895886f49f311b75eec455a not found: ID does not exist" Oct 05 07:45:37 crc kubenswrapper[4846]: I1005 07:45:37.843299 4846 scope.go:117] "RemoveContainer" containerID="7d8f27b872a19c60802af1911701bd97e0d3d2bb93d5f28d7ee16ce6ce7fac52" Oct 05 07:45:37 crc kubenswrapper[4846]: E1005 07:45:37.844235 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7d8f27b872a19c60802af1911701bd97e0d3d2bb93d5f28d7ee16ce6ce7fac52\": container with ID starting with 7d8f27b872a19c60802af1911701bd97e0d3d2bb93d5f28d7ee16ce6ce7fac52 not found: ID does not exist" containerID="7d8f27b872a19c60802af1911701bd97e0d3d2bb93d5f28d7ee16ce6ce7fac52" Oct 05 07:45:37 crc kubenswrapper[4846]: I1005 07:45:37.844321 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7d8f27b872a19c60802af1911701bd97e0d3d2bb93d5f28d7ee16ce6ce7fac52"} err="failed to get container status \"7d8f27b872a19c60802af1911701bd97e0d3d2bb93d5f28d7ee16ce6ce7fac52\": rpc error: code = NotFound desc = could not find container \"7d8f27b872a19c60802af1911701bd97e0d3d2bb93d5f28d7ee16ce6ce7fac52\": container with ID starting with 7d8f27b872a19c60802af1911701bd97e0d3d2bb93d5f28d7ee16ce6ce7fac52 not found: ID does not exist" Oct 05 07:45:37 crc kubenswrapper[4846]: I1005 07:45:37.844367 4846 scope.go:117] "RemoveContainer" containerID="c6e05032bd4a96b4de60fc1077adf019fe4093876cbf4ca68bb29782f395da36" Oct 05 07:45:37 crc kubenswrapper[4846]: E1005 07:45:37.844918 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c6e05032bd4a96b4de60fc1077adf019fe4093876cbf4ca68bb29782f395da36\": container with ID starting with c6e05032bd4a96b4de60fc1077adf019fe4093876cbf4ca68bb29782f395da36 not found: ID does not exist" containerID="c6e05032bd4a96b4de60fc1077adf019fe4093876cbf4ca68bb29782f395da36" Oct 05 07:45:37 crc kubenswrapper[4846]: I1005 07:45:37.844963 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c6e05032bd4a96b4de60fc1077adf019fe4093876cbf4ca68bb29782f395da36"} err="failed to get container status \"c6e05032bd4a96b4de60fc1077adf019fe4093876cbf4ca68bb29782f395da36\": rpc error: code = NotFound desc = could not find container \"c6e05032bd4a96b4de60fc1077adf019fe4093876cbf4ca68bb29782f395da36\": container with ID starting with c6e05032bd4a96b4de60fc1077adf019fe4093876cbf4ca68bb29782f395da36 not found: ID does not exist" Oct 05 07:45:38 crc kubenswrapper[4846]: I1005 07:45:38.514793 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f29bf0d4-54c4-4d11-87e3-d153d5f3fbb8" path="/var/lib/kubelet/pods/f29bf0d4-54c4-4d11-87e3-d153d5f3fbb8/volumes" Oct 05 07:46:53 crc kubenswrapper[4846]: I1005 07:46:53.325714 4846 patch_prober.go:28] interesting pod/machine-config-daemon-fscvf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 07:46:53 crc kubenswrapper[4846]: I1005 07:46:53.326348 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 07:47:23 crc kubenswrapper[4846]: I1005 07:47:23.325737 4846 patch_prober.go:28] interesting pod/machine-config-daemon-fscvf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 07:47:23 crc kubenswrapper[4846]: I1005 07:47:23.326418 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 07:47:53 crc kubenswrapper[4846]: I1005 07:47:53.325666 4846 patch_prober.go:28] interesting pod/machine-config-daemon-fscvf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 07:47:53 crc kubenswrapper[4846]: I1005 07:47:53.326348 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 07:47:53 crc kubenswrapper[4846]: I1005 07:47:53.326419 4846 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" Oct 05 07:47:53 crc kubenswrapper[4846]: I1005 07:47:53.327334 4846 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"a8dd7af9a49a3eba2b3ea4d7586f51b807c63a3b657f800b5182c8288f292750"} pod="openshift-machine-config-operator/machine-config-daemon-fscvf" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 05 07:47:53 crc kubenswrapper[4846]: I1005 07:47:53.327427 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" containerName="machine-config-daemon" containerID="cri-o://a8dd7af9a49a3eba2b3ea4d7586f51b807c63a3b657f800b5182c8288f292750" gracePeriod=600 Oct 05 07:47:53 crc kubenswrapper[4846]: E1005 07:47:53.461812 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 07:47:54 crc kubenswrapper[4846]: I1005 07:47:54.042486 4846 generic.go:334] "Generic (PLEG): container finished" podID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" containerID="a8dd7af9a49a3eba2b3ea4d7586f51b807c63a3b657f800b5182c8288f292750" exitCode=0 Oct 05 07:47:54 crc kubenswrapper[4846]: I1005 07:47:54.042576 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" event={"ID":"ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7","Type":"ContainerDied","Data":"a8dd7af9a49a3eba2b3ea4d7586f51b807c63a3b657f800b5182c8288f292750"} Oct 05 07:47:54 crc kubenswrapper[4846]: I1005 07:47:54.042638 4846 scope.go:117] "RemoveContainer" containerID="725a93409ff4757fae7733195dbbc6de2ee38624fdbf934d0b72e1d2e3418171" Oct 05 07:47:54 crc kubenswrapper[4846]: I1005 07:47:54.043611 4846 scope.go:117] "RemoveContainer" containerID="a8dd7af9a49a3eba2b3ea4d7586f51b807c63a3b657f800b5182c8288f292750" Oct 05 07:47:54 crc kubenswrapper[4846]: E1005 07:47:54.044099 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 07:48:06 crc kubenswrapper[4846]: I1005 07:48:06.497481 4846 scope.go:117] "RemoveContainer" containerID="a8dd7af9a49a3eba2b3ea4d7586f51b807c63a3b657f800b5182c8288f292750" Oct 05 07:48:06 crc kubenswrapper[4846]: E1005 07:48:06.498528 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 07:48:19 crc kubenswrapper[4846]: I1005 07:48:19.498590 4846 scope.go:117] "RemoveContainer" containerID="a8dd7af9a49a3eba2b3ea4d7586f51b807c63a3b657f800b5182c8288f292750" Oct 05 07:48:19 crc kubenswrapper[4846]: E1005 07:48:19.499728 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 07:48:31 crc kubenswrapper[4846]: I1005 07:48:31.498548 4846 scope.go:117] "RemoveContainer" containerID="a8dd7af9a49a3eba2b3ea4d7586f51b807c63a3b657f800b5182c8288f292750" Oct 05 07:48:31 crc kubenswrapper[4846]: E1005 07:48:31.499694 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 07:48:37 crc kubenswrapper[4846]: I1005 07:48:37.275746 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-smlkc"] Oct 05 07:48:37 crc kubenswrapper[4846]: E1005 07:48:37.276545 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f29bf0d4-54c4-4d11-87e3-d153d5f3fbb8" containerName="registry-server" Oct 05 07:48:37 crc kubenswrapper[4846]: I1005 07:48:37.276567 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="f29bf0d4-54c4-4d11-87e3-d153d5f3fbb8" containerName="registry-server" Oct 05 07:48:37 crc kubenswrapper[4846]: E1005 07:48:37.276602 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f29bf0d4-54c4-4d11-87e3-d153d5f3fbb8" containerName="extract-content" Oct 05 07:48:37 crc kubenswrapper[4846]: I1005 07:48:37.276617 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="f29bf0d4-54c4-4d11-87e3-d153d5f3fbb8" containerName="extract-content" Oct 05 07:48:37 crc kubenswrapper[4846]: E1005 07:48:37.276642 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f29bf0d4-54c4-4d11-87e3-d153d5f3fbb8" containerName="extract-utilities" Oct 05 07:48:37 crc kubenswrapper[4846]: I1005 07:48:37.276656 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="f29bf0d4-54c4-4d11-87e3-d153d5f3fbb8" containerName="extract-utilities" Oct 05 07:48:37 crc kubenswrapper[4846]: I1005 07:48:37.276935 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="f29bf0d4-54c4-4d11-87e3-d153d5f3fbb8" containerName="registry-server" Oct 05 07:48:37 crc kubenswrapper[4846]: I1005 07:48:37.279414 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-smlkc" Oct 05 07:48:37 crc kubenswrapper[4846]: I1005 07:48:37.297759 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-smlkc"] Oct 05 07:48:37 crc kubenswrapper[4846]: I1005 07:48:37.338636 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fdda562f-1d0b-4f9b-979c-a51547f809e3-utilities\") pod \"community-operators-smlkc\" (UID: \"fdda562f-1d0b-4f9b-979c-a51547f809e3\") " pod="openshift-marketplace/community-operators-smlkc" Oct 05 07:48:37 crc kubenswrapper[4846]: I1005 07:48:37.339108 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-62nsc\" (UniqueName: \"kubernetes.io/projected/fdda562f-1d0b-4f9b-979c-a51547f809e3-kube-api-access-62nsc\") pod \"community-operators-smlkc\" (UID: \"fdda562f-1d0b-4f9b-979c-a51547f809e3\") " pod="openshift-marketplace/community-operators-smlkc" Oct 05 07:48:37 crc kubenswrapper[4846]: I1005 07:48:37.341270 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fdda562f-1d0b-4f9b-979c-a51547f809e3-catalog-content\") pod \"community-operators-smlkc\" (UID: \"fdda562f-1d0b-4f9b-979c-a51547f809e3\") " pod="openshift-marketplace/community-operators-smlkc" Oct 05 07:48:37 crc kubenswrapper[4846]: I1005 07:48:37.443629 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fdda562f-1d0b-4f9b-979c-a51547f809e3-catalog-content\") pod \"community-operators-smlkc\" (UID: \"fdda562f-1d0b-4f9b-979c-a51547f809e3\") " pod="openshift-marketplace/community-operators-smlkc" Oct 05 07:48:37 crc kubenswrapper[4846]: I1005 07:48:37.443723 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fdda562f-1d0b-4f9b-979c-a51547f809e3-utilities\") pod \"community-operators-smlkc\" (UID: \"fdda562f-1d0b-4f9b-979c-a51547f809e3\") " pod="openshift-marketplace/community-operators-smlkc" Oct 05 07:48:37 crc kubenswrapper[4846]: I1005 07:48:37.443781 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-62nsc\" (UniqueName: \"kubernetes.io/projected/fdda562f-1d0b-4f9b-979c-a51547f809e3-kube-api-access-62nsc\") pod \"community-operators-smlkc\" (UID: \"fdda562f-1d0b-4f9b-979c-a51547f809e3\") " pod="openshift-marketplace/community-operators-smlkc" Oct 05 07:48:37 crc kubenswrapper[4846]: I1005 07:48:37.444550 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fdda562f-1d0b-4f9b-979c-a51547f809e3-utilities\") pod \"community-operators-smlkc\" (UID: \"fdda562f-1d0b-4f9b-979c-a51547f809e3\") " pod="openshift-marketplace/community-operators-smlkc" Oct 05 07:48:37 crc kubenswrapper[4846]: I1005 07:48:37.444594 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fdda562f-1d0b-4f9b-979c-a51547f809e3-catalog-content\") pod \"community-operators-smlkc\" (UID: \"fdda562f-1d0b-4f9b-979c-a51547f809e3\") " pod="openshift-marketplace/community-operators-smlkc" Oct 05 07:48:37 crc kubenswrapper[4846]: I1005 07:48:37.465437 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-62nsc\" (UniqueName: \"kubernetes.io/projected/fdda562f-1d0b-4f9b-979c-a51547f809e3-kube-api-access-62nsc\") pod \"community-operators-smlkc\" (UID: \"fdda562f-1d0b-4f9b-979c-a51547f809e3\") " pod="openshift-marketplace/community-operators-smlkc" Oct 05 07:48:37 crc kubenswrapper[4846]: I1005 07:48:37.654103 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-smlkc" Oct 05 07:48:38 crc kubenswrapper[4846]: I1005 07:48:38.173727 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-smlkc"] Oct 05 07:48:38 crc kubenswrapper[4846]: W1005 07:48:38.181069 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfdda562f_1d0b_4f9b_979c_a51547f809e3.slice/crio-a68d595a306f3cd71ea9fe0f1da4dbafc610147aea1dbfc6c7a1409bb80a42b0 WatchSource:0}: Error finding container a68d595a306f3cd71ea9fe0f1da4dbafc610147aea1dbfc6c7a1409bb80a42b0: Status 404 returned error can't find the container with id a68d595a306f3cd71ea9fe0f1da4dbafc610147aea1dbfc6c7a1409bb80a42b0 Oct 05 07:48:38 crc kubenswrapper[4846]: I1005 07:48:38.477854 4846 generic.go:334] "Generic (PLEG): container finished" podID="fdda562f-1d0b-4f9b-979c-a51547f809e3" containerID="2837752354a0ae0cdc74e7ad9fea3448b53a6109f5e2a2e08c2ecdd775cedf3b" exitCode=0 Oct 05 07:48:38 crc kubenswrapper[4846]: I1005 07:48:38.478455 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-smlkc" event={"ID":"fdda562f-1d0b-4f9b-979c-a51547f809e3","Type":"ContainerDied","Data":"2837752354a0ae0cdc74e7ad9fea3448b53a6109f5e2a2e08c2ecdd775cedf3b"} Oct 05 07:48:38 crc kubenswrapper[4846]: I1005 07:48:38.478580 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-smlkc" event={"ID":"fdda562f-1d0b-4f9b-979c-a51547f809e3","Type":"ContainerStarted","Data":"a68d595a306f3cd71ea9fe0f1da4dbafc610147aea1dbfc6c7a1409bb80a42b0"} Oct 05 07:48:39 crc kubenswrapper[4846]: I1005 07:48:39.488430 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-smlkc" event={"ID":"fdda562f-1d0b-4f9b-979c-a51547f809e3","Type":"ContainerStarted","Data":"9c9e0040a16751073d1a5f9eb1bb6d076ba153d2032c09651881009f23e29ca8"} Oct 05 07:48:40 crc kubenswrapper[4846]: I1005 07:48:40.501161 4846 generic.go:334] "Generic (PLEG): container finished" podID="fdda562f-1d0b-4f9b-979c-a51547f809e3" containerID="9c9e0040a16751073d1a5f9eb1bb6d076ba153d2032c09651881009f23e29ca8" exitCode=0 Oct 05 07:48:40 crc kubenswrapper[4846]: I1005 07:48:40.511443 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-smlkc" event={"ID":"fdda562f-1d0b-4f9b-979c-a51547f809e3","Type":"ContainerDied","Data":"9c9e0040a16751073d1a5f9eb1bb6d076ba153d2032c09651881009f23e29ca8"} Oct 05 07:48:41 crc kubenswrapper[4846]: I1005 07:48:41.514882 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-smlkc" event={"ID":"fdda562f-1d0b-4f9b-979c-a51547f809e3","Type":"ContainerStarted","Data":"c102aea078e185ea8da72a76f77d27a58dcc29a8c552f3d619f692c7006ea8b7"} Oct 05 07:48:41 crc kubenswrapper[4846]: I1005 07:48:41.537872 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-smlkc" podStartSLOduration=2.046721494 podStartE2EDuration="4.537855278s" podCreationTimestamp="2025-10-05 07:48:37 +0000 UTC" firstStartedPulling="2025-10-05 07:48:38.4841383 +0000 UTC m=+3640.724991105" lastFinishedPulling="2025-10-05 07:48:40.975272104 +0000 UTC m=+3643.216124889" observedRunningTime="2025-10-05 07:48:41.533229324 +0000 UTC m=+3643.774082159" watchObservedRunningTime="2025-10-05 07:48:41.537855278 +0000 UTC m=+3643.778708053" Oct 05 07:48:46 crc kubenswrapper[4846]: I1005 07:48:46.499940 4846 scope.go:117] "RemoveContainer" containerID="a8dd7af9a49a3eba2b3ea4d7586f51b807c63a3b657f800b5182c8288f292750" Oct 05 07:48:46 crc kubenswrapper[4846]: E1005 07:48:46.500982 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 07:48:47 crc kubenswrapper[4846]: I1005 07:48:47.654551 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-smlkc" Oct 05 07:48:47 crc kubenswrapper[4846]: I1005 07:48:47.654659 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-smlkc" Oct 05 07:48:47 crc kubenswrapper[4846]: I1005 07:48:47.730615 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-smlkc" Oct 05 07:48:48 crc kubenswrapper[4846]: I1005 07:48:48.656506 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-smlkc" Oct 05 07:48:48 crc kubenswrapper[4846]: I1005 07:48:48.718930 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-smlkc"] Oct 05 07:48:50 crc kubenswrapper[4846]: I1005 07:48:50.598935 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-smlkc" podUID="fdda562f-1d0b-4f9b-979c-a51547f809e3" containerName="registry-server" containerID="cri-o://c102aea078e185ea8da72a76f77d27a58dcc29a8c552f3d619f692c7006ea8b7" gracePeriod=2 Oct 05 07:48:51 crc kubenswrapper[4846]: I1005 07:48:51.609758 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-smlkc" Oct 05 07:48:51 crc kubenswrapper[4846]: I1005 07:48:51.610083 4846 generic.go:334] "Generic (PLEG): container finished" podID="fdda562f-1d0b-4f9b-979c-a51547f809e3" containerID="c102aea078e185ea8da72a76f77d27a58dcc29a8c552f3d619f692c7006ea8b7" exitCode=0 Oct 05 07:48:51 crc kubenswrapper[4846]: I1005 07:48:51.610105 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-smlkc" event={"ID":"fdda562f-1d0b-4f9b-979c-a51547f809e3","Type":"ContainerDied","Data":"c102aea078e185ea8da72a76f77d27a58dcc29a8c552f3d619f692c7006ea8b7"} Oct 05 07:48:51 crc kubenswrapper[4846]: I1005 07:48:51.611013 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-smlkc" event={"ID":"fdda562f-1d0b-4f9b-979c-a51547f809e3","Type":"ContainerDied","Data":"a68d595a306f3cd71ea9fe0f1da4dbafc610147aea1dbfc6c7a1409bb80a42b0"} Oct 05 07:48:51 crc kubenswrapper[4846]: I1005 07:48:51.611057 4846 scope.go:117] "RemoveContainer" containerID="c102aea078e185ea8da72a76f77d27a58dcc29a8c552f3d619f692c7006ea8b7" Oct 05 07:48:51 crc kubenswrapper[4846]: I1005 07:48:51.637962 4846 scope.go:117] "RemoveContainer" containerID="9c9e0040a16751073d1a5f9eb1bb6d076ba153d2032c09651881009f23e29ca8" Oct 05 07:48:51 crc kubenswrapper[4846]: I1005 07:48:51.669612 4846 scope.go:117] "RemoveContainer" containerID="2837752354a0ae0cdc74e7ad9fea3448b53a6109f5e2a2e08c2ecdd775cedf3b" Oct 05 07:48:51 crc kubenswrapper[4846]: I1005 07:48:51.679251 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fdda562f-1d0b-4f9b-979c-a51547f809e3-utilities\") pod \"fdda562f-1d0b-4f9b-979c-a51547f809e3\" (UID: \"fdda562f-1d0b-4f9b-979c-a51547f809e3\") " Oct 05 07:48:51 crc kubenswrapper[4846]: I1005 07:48:51.679434 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-62nsc\" (UniqueName: \"kubernetes.io/projected/fdda562f-1d0b-4f9b-979c-a51547f809e3-kube-api-access-62nsc\") pod \"fdda562f-1d0b-4f9b-979c-a51547f809e3\" (UID: \"fdda562f-1d0b-4f9b-979c-a51547f809e3\") " Oct 05 07:48:51 crc kubenswrapper[4846]: I1005 07:48:51.679515 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fdda562f-1d0b-4f9b-979c-a51547f809e3-catalog-content\") pod \"fdda562f-1d0b-4f9b-979c-a51547f809e3\" (UID: \"fdda562f-1d0b-4f9b-979c-a51547f809e3\") " Oct 05 07:48:51 crc kubenswrapper[4846]: I1005 07:48:51.680534 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fdda562f-1d0b-4f9b-979c-a51547f809e3-utilities" (OuterVolumeSpecName: "utilities") pod "fdda562f-1d0b-4f9b-979c-a51547f809e3" (UID: "fdda562f-1d0b-4f9b-979c-a51547f809e3"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 07:48:51 crc kubenswrapper[4846]: I1005 07:48:51.687068 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fdda562f-1d0b-4f9b-979c-a51547f809e3-kube-api-access-62nsc" (OuterVolumeSpecName: "kube-api-access-62nsc") pod "fdda562f-1d0b-4f9b-979c-a51547f809e3" (UID: "fdda562f-1d0b-4f9b-979c-a51547f809e3"). InnerVolumeSpecName "kube-api-access-62nsc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:48:51 crc kubenswrapper[4846]: I1005 07:48:51.695648 4846 scope.go:117] "RemoveContainer" containerID="c102aea078e185ea8da72a76f77d27a58dcc29a8c552f3d619f692c7006ea8b7" Oct 05 07:48:51 crc kubenswrapper[4846]: E1005 07:48:51.696166 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c102aea078e185ea8da72a76f77d27a58dcc29a8c552f3d619f692c7006ea8b7\": container with ID starting with c102aea078e185ea8da72a76f77d27a58dcc29a8c552f3d619f692c7006ea8b7 not found: ID does not exist" containerID="c102aea078e185ea8da72a76f77d27a58dcc29a8c552f3d619f692c7006ea8b7" Oct 05 07:48:51 crc kubenswrapper[4846]: I1005 07:48:51.696270 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c102aea078e185ea8da72a76f77d27a58dcc29a8c552f3d619f692c7006ea8b7"} err="failed to get container status \"c102aea078e185ea8da72a76f77d27a58dcc29a8c552f3d619f692c7006ea8b7\": rpc error: code = NotFound desc = could not find container \"c102aea078e185ea8da72a76f77d27a58dcc29a8c552f3d619f692c7006ea8b7\": container with ID starting with c102aea078e185ea8da72a76f77d27a58dcc29a8c552f3d619f692c7006ea8b7 not found: ID does not exist" Oct 05 07:48:51 crc kubenswrapper[4846]: I1005 07:48:51.696304 4846 scope.go:117] "RemoveContainer" containerID="9c9e0040a16751073d1a5f9eb1bb6d076ba153d2032c09651881009f23e29ca8" Oct 05 07:48:51 crc kubenswrapper[4846]: E1005 07:48:51.696759 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9c9e0040a16751073d1a5f9eb1bb6d076ba153d2032c09651881009f23e29ca8\": container with ID starting with 9c9e0040a16751073d1a5f9eb1bb6d076ba153d2032c09651881009f23e29ca8 not found: ID does not exist" containerID="9c9e0040a16751073d1a5f9eb1bb6d076ba153d2032c09651881009f23e29ca8" Oct 05 07:48:51 crc kubenswrapper[4846]: I1005 07:48:51.696810 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9c9e0040a16751073d1a5f9eb1bb6d076ba153d2032c09651881009f23e29ca8"} err="failed to get container status \"9c9e0040a16751073d1a5f9eb1bb6d076ba153d2032c09651881009f23e29ca8\": rpc error: code = NotFound desc = could not find container \"9c9e0040a16751073d1a5f9eb1bb6d076ba153d2032c09651881009f23e29ca8\": container with ID starting with 9c9e0040a16751073d1a5f9eb1bb6d076ba153d2032c09651881009f23e29ca8 not found: ID does not exist" Oct 05 07:48:51 crc kubenswrapper[4846]: I1005 07:48:51.696843 4846 scope.go:117] "RemoveContainer" containerID="2837752354a0ae0cdc74e7ad9fea3448b53a6109f5e2a2e08c2ecdd775cedf3b" Oct 05 07:48:51 crc kubenswrapper[4846]: E1005 07:48:51.697109 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2837752354a0ae0cdc74e7ad9fea3448b53a6109f5e2a2e08c2ecdd775cedf3b\": container with ID starting with 2837752354a0ae0cdc74e7ad9fea3448b53a6109f5e2a2e08c2ecdd775cedf3b not found: ID does not exist" containerID="2837752354a0ae0cdc74e7ad9fea3448b53a6109f5e2a2e08c2ecdd775cedf3b" Oct 05 07:48:51 crc kubenswrapper[4846]: I1005 07:48:51.697131 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2837752354a0ae0cdc74e7ad9fea3448b53a6109f5e2a2e08c2ecdd775cedf3b"} err="failed to get container status \"2837752354a0ae0cdc74e7ad9fea3448b53a6109f5e2a2e08c2ecdd775cedf3b\": rpc error: code = NotFound desc = could not find container \"2837752354a0ae0cdc74e7ad9fea3448b53a6109f5e2a2e08c2ecdd775cedf3b\": container with ID starting with 2837752354a0ae0cdc74e7ad9fea3448b53a6109f5e2a2e08c2ecdd775cedf3b not found: ID does not exist" Oct 05 07:48:51 crc kubenswrapper[4846]: I1005 07:48:51.740018 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fdda562f-1d0b-4f9b-979c-a51547f809e3-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "fdda562f-1d0b-4f9b-979c-a51547f809e3" (UID: "fdda562f-1d0b-4f9b-979c-a51547f809e3"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 07:48:51 crc kubenswrapper[4846]: I1005 07:48:51.781753 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-62nsc\" (UniqueName: \"kubernetes.io/projected/fdda562f-1d0b-4f9b-979c-a51547f809e3-kube-api-access-62nsc\") on node \"crc\" DevicePath \"\"" Oct 05 07:48:51 crc kubenswrapper[4846]: I1005 07:48:51.782044 4846 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fdda562f-1d0b-4f9b-979c-a51547f809e3-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 05 07:48:51 crc kubenswrapper[4846]: I1005 07:48:51.782135 4846 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fdda562f-1d0b-4f9b-979c-a51547f809e3-utilities\") on node \"crc\" DevicePath \"\"" Oct 05 07:48:52 crc kubenswrapper[4846]: I1005 07:48:52.624154 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-smlkc" Oct 05 07:48:52 crc kubenswrapper[4846]: I1005 07:48:52.660006 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-smlkc"] Oct 05 07:48:52 crc kubenswrapper[4846]: I1005 07:48:52.669712 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-smlkc"] Oct 05 07:48:54 crc kubenswrapper[4846]: I1005 07:48:54.510835 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fdda562f-1d0b-4f9b-979c-a51547f809e3" path="/var/lib/kubelet/pods/fdda562f-1d0b-4f9b-979c-a51547f809e3/volumes" Oct 05 07:48:59 crc kubenswrapper[4846]: I1005 07:48:59.498151 4846 scope.go:117] "RemoveContainer" containerID="a8dd7af9a49a3eba2b3ea4d7586f51b807c63a3b657f800b5182c8288f292750" Oct 05 07:48:59 crc kubenswrapper[4846]: E1005 07:48:59.498995 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 07:49:11 crc kubenswrapper[4846]: I1005 07:49:11.497792 4846 scope.go:117] "RemoveContainer" containerID="a8dd7af9a49a3eba2b3ea4d7586f51b807c63a3b657f800b5182c8288f292750" Oct 05 07:49:11 crc kubenswrapper[4846]: E1005 07:49:11.498856 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 07:49:22 crc kubenswrapper[4846]: I1005 07:49:22.497778 4846 scope.go:117] "RemoveContainer" containerID="a8dd7af9a49a3eba2b3ea4d7586f51b807c63a3b657f800b5182c8288f292750" Oct 05 07:49:22 crc kubenswrapper[4846]: E1005 07:49:22.498812 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 07:49:36 crc kubenswrapper[4846]: I1005 07:49:36.497921 4846 scope.go:117] "RemoveContainer" containerID="a8dd7af9a49a3eba2b3ea4d7586f51b807c63a3b657f800b5182c8288f292750" Oct 05 07:49:36 crc kubenswrapper[4846]: E1005 07:49:36.499110 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 07:49:47 crc kubenswrapper[4846]: I1005 07:49:47.497691 4846 scope.go:117] "RemoveContainer" containerID="a8dd7af9a49a3eba2b3ea4d7586f51b807c63a3b657f800b5182c8288f292750" Oct 05 07:49:47 crc kubenswrapper[4846]: E1005 07:49:47.498675 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 07:50:00 crc kubenswrapper[4846]: I1005 07:50:00.497816 4846 scope.go:117] "RemoveContainer" containerID="a8dd7af9a49a3eba2b3ea4d7586f51b807c63a3b657f800b5182c8288f292750" Oct 05 07:50:00 crc kubenswrapper[4846]: E1005 07:50:00.498973 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 07:50:12 crc kubenswrapper[4846]: I1005 07:50:12.497803 4846 scope.go:117] "RemoveContainer" containerID="a8dd7af9a49a3eba2b3ea4d7586f51b807c63a3b657f800b5182c8288f292750" Oct 05 07:50:12 crc kubenswrapper[4846]: E1005 07:50:12.498770 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 07:50:23 crc kubenswrapper[4846]: I1005 07:50:23.498409 4846 scope.go:117] "RemoveContainer" containerID="a8dd7af9a49a3eba2b3ea4d7586f51b807c63a3b657f800b5182c8288f292750" Oct 05 07:50:23 crc kubenswrapper[4846]: E1005 07:50:23.499527 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 07:50:36 crc kubenswrapper[4846]: I1005 07:50:36.497988 4846 scope.go:117] "RemoveContainer" containerID="a8dd7af9a49a3eba2b3ea4d7586f51b807c63a3b657f800b5182c8288f292750" Oct 05 07:50:36 crc kubenswrapper[4846]: E1005 07:50:36.498987 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 07:50:49 crc kubenswrapper[4846]: I1005 07:50:49.497749 4846 scope.go:117] "RemoveContainer" containerID="a8dd7af9a49a3eba2b3ea4d7586f51b807c63a3b657f800b5182c8288f292750" Oct 05 07:50:49 crc kubenswrapper[4846]: E1005 07:50:49.498717 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 07:51:02 crc kubenswrapper[4846]: I1005 07:51:02.497734 4846 scope.go:117] "RemoveContainer" containerID="a8dd7af9a49a3eba2b3ea4d7586f51b807c63a3b657f800b5182c8288f292750" Oct 05 07:51:02 crc kubenswrapper[4846]: E1005 07:51:02.500261 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 07:51:17 crc kubenswrapper[4846]: I1005 07:51:17.498494 4846 scope.go:117] "RemoveContainer" containerID="a8dd7af9a49a3eba2b3ea4d7586f51b807c63a3b657f800b5182c8288f292750" Oct 05 07:51:17 crc kubenswrapper[4846]: E1005 07:51:17.499680 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 07:51:29 crc kubenswrapper[4846]: I1005 07:51:29.497942 4846 scope.go:117] "RemoveContainer" containerID="a8dd7af9a49a3eba2b3ea4d7586f51b807c63a3b657f800b5182c8288f292750" Oct 05 07:51:29 crc kubenswrapper[4846]: E1005 07:51:29.498724 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 07:51:42 crc kubenswrapper[4846]: I1005 07:51:42.498002 4846 scope.go:117] "RemoveContainer" containerID="a8dd7af9a49a3eba2b3ea4d7586f51b807c63a3b657f800b5182c8288f292750" Oct 05 07:51:42 crc kubenswrapper[4846]: E1005 07:51:42.499276 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 07:51:56 crc kubenswrapper[4846]: I1005 07:51:56.498491 4846 scope.go:117] "RemoveContainer" containerID="a8dd7af9a49a3eba2b3ea4d7586f51b807c63a3b657f800b5182c8288f292750" Oct 05 07:51:56 crc kubenswrapper[4846]: E1005 07:51:56.499624 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 07:52:11 crc kubenswrapper[4846]: I1005 07:52:11.497894 4846 scope.go:117] "RemoveContainer" containerID="a8dd7af9a49a3eba2b3ea4d7586f51b807c63a3b657f800b5182c8288f292750" Oct 05 07:52:11 crc kubenswrapper[4846]: E1005 07:52:11.499114 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 07:52:25 crc kubenswrapper[4846]: I1005 07:52:25.497549 4846 scope.go:117] "RemoveContainer" containerID="a8dd7af9a49a3eba2b3ea4d7586f51b807c63a3b657f800b5182c8288f292750" Oct 05 07:52:25 crc kubenswrapper[4846]: E1005 07:52:25.498165 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 07:52:37 crc kubenswrapper[4846]: I1005 07:52:37.498326 4846 scope.go:117] "RemoveContainer" containerID="a8dd7af9a49a3eba2b3ea4d7586f51b807c63a3b657f800b5182c8288f292750" Oct 05 07:52:37 crc kubenswrapper[4846]: E1005 07:52:37.499258 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 07:52:51 crc kubenswrapper[4846]: I1005 07:52:51.497694 4846 scope.go:117] "RemoveContainer" containerID="a8dd7af9a49a3eba2b3ea4d7586f51b807c63a3b657f800b5182c8288f292750" Oct 05 07:52:51 crc kubenswrapper[4846]: E1005 07:52:51.498444 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 07:53:06 crc kubenswrapper[4846]: I1005 07:53:06.498433 4846 scope.go:117] "RemoveContainer" containerID="a8dd7af9a49a3eba2b3ea4d7586f51b807c63a3b657f800b5182c8288f292750" Oct 05 07:53:07 crc kubenswrapper[4846]: I1005 07:53:07.006748 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" event={"ID":"ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7","Type":"ContainerStarted","Data":"dccf837d14b883290775f3ba1aae37388eb10cbc64bfe9ddcfac744876bcccff"} Oct 05 07:54:31 crc kubenswrapper[4846]: I1005 07:54:31.139271 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-v9d5k"] Oct 05 07:54:31 crc kubenswrapper[4846]: E1005 07:54:31.140397 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fdda562f-1d0b-4f9b-979c-a51547f809e3" containerName="extract-utilities" Oct 05 07:54:31 crc kubenswrapper[4846]: I1005 07:54:31.140424 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="fdda562f-1d0b-4f9b-979c-a51547f809e3" containerName="extract-utilities" Oct 05 07:54:31 crc kubenswrapper[4846]: E1005 07:54:31.140449 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fdda562f-1d0b-4f9b-979c-a51547f809e3" containerName="extract-content" Oct 05 07:54:31 crc kubenswrapper[4846]: I1005 07:54:31.140461 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="fdda562f-1d0b-4f9b-979c-a51547f809e3" containerName="extract-content" Oct 05 07:54:31 crc kubenswrapper[4846]: E1005 07:54:31.140508 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fdda562f-1d0b-4f9b-979c-a51547f809e3" containerName="registry-server" Oct 05 07:54:31 crc kubenswrapper[4846]: I1005 07:54:31.140520 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="fdda562f-1d0b-4f9b-979c-a51547f809e3" containerName="registry-server" Oct 05 07:54:31 crc kubenswrapper[4846]: I1005 07:54:31.140781 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="fdda562f-1d0b-4f9b-979c-a51547f809e3" containerName="registry-server" Oct 05 07:54:31 crc kubenswrapper[4846]: I1005 07:54:31.142721 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-v9d5k" Oct 05 07:54:31 crc kubenswrapper[4846]: I1005 07:54:31.154621 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-v9d5k"] Oct 05 07:54:31 crc kubenswrapper[4846]: I1005 07:54:31.186565 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/34905cec-e457-43a4-ba2e-acbb11e9c637-catalog-content\") pod \"certified-operators-v9d5k\" (UID: \"34905cec-e457-43a4-ba2e-acbb11e9c637\") " pod="openshift-marketplace/certified-operators-v9d5k" Oct 05 07:54:31 crc kubenswrapper[4846]: I1005 07:54:31.186655 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6sh55\" (UniqueName: \"kubernetes.io/projected/34905cec-e457-43a4-ba2e-acbb11e9c637-kube-api-access-6sh55\") pod \"certified-operators-v9d5k\" (UID: \"34905cec-e457-43a4-ba2e-acbb11e9c637\") " pod="openshift-marketplace/certified-operators-v9d5k" Oct 05 07:54:31 crc kubenswrapper[4846]: I1005 07:54:31.186743 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/34905cec-e457-43a4-ba2e-acbb11e9c637-utilities\") pod \"certified-operators-v9d5k\" (UID: \"34905cec-e457-43a4-ba2e-acbb11e9c637\") " pod="openshift-marketplace/certified-operators-v9d5k" Oct 05 07:54:31 crc kubenswrapper[4846]: I1005 07:54:31.287810 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/34905cec-e457-43a4-ba2e-acbb11e9c637-catalog-content\") pod \"certified-operators-v9d5k\" (UID: \"34905cec-e457-43a4-ba2e-acbb11e9c637\") " pod="openshift-marketplace/certified-operators-v9d5k" Oct 05 07:54:31 crc kubenswrapper[4846]: I1005 07:54:31.287877 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6sh55\" (UniqueName: \"kubernetes.io/projected/34905cec-e457-43a4-ba2e-acbb11e9c637-kube-api-access-6sh55\") pod \"certified-operators-v9d5k\" (UID: \"34905cec-e457-43a4-ba2e-acbb11e9c637\") " pod="openshift-marketplace/certified-operators-v9d5k" Oct 05 07:54:31 crc kubenswrapper[4846]: I1005 07:54:31.288263 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/34905cec-e457-43a4-ba2e-acbb11e9c637-utilities\") pod \"certified-operators-v9d5k\" (UID: \"34905cec-e457-43a4-ba2e-acbb11e9c637\") " pod="openshift-marketplace/certified-operators-v9d5k" Oct 05 07:54:31 crc kubenswrapper[4846]: I1005 07:54:31.288913 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/34905cec-e457-43a4-ba2e-acbb11e9c637-catalog-content\") pod \"certified-operators-v9d5k\" (UID: \"34905cec-e457-43a4-ba2e-acbb11e9c637\") " pod="openshift-marketplace/certified-operators-v9d5k" Oct 05 07:54:31 crc kubenswrapper[4846]: I1005 07:54:31.288932 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/34905cec-e457-43a4-ba2e-acbb11e9c637-utilities\") pod \"certified-operators-v9d5k\" (UID: \"34905cec-e457-43a4-ba2e-acbb11e9c637\") " pod="openshift-marketplace/certified-operators-v9d5k" Oct 05 07:54:31 crc kubenswrapper[4846]: I1005 07:54:31.310322 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6sh55\" (UniqueName: \"kubernetes.io/projected/34905cec-e457-43a4-ba2e-acbb11e9c637-kube-api-access-6sh55\") pod \"certified-operators-v9d5k\" (UID: \"34905cec-e457-43a4-ba2e-acbb11e9c637\") " pod="openshift-marketplace/certified-operators-v9d5k" Oct 05 07:54:31 crc kubenswrapper[4846]: I1005 07:54:31.518816 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-v9d5k" Oct 05 07:54:31 crc kubenswrapper[4846]: I1005 07:54:31.961628 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-v9d5k"] Oct 05 07:54:32 crc kubenswrapper[4846]: I1005 07:54:32.809043 4846 generic.go:334] "Generic (PLEG): container finished" podID="34905cec-e457-43a4-ba2e-acbb11e9c637" containerID="23fd8b3e9310bf309d01fae62df1ac42dd336a9522a815293d6600447e145990" exitCode=0 Oct 05 07:54:32 crc kubenswrapper[4846]: I1005 07:54:32.809231 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-v9d5k" event={"ID":"34905cec-e457-43a4-ba2e-acbb11e9c637","Type":"ContainerDied","Data":"23fd8b3e9310bf309d01fae62df1ac42dd336a9522a815293d6600447e145990"} Oct 05 07:54:32 crc kubenswrapper[4846]: I1005 07:54:32.809507 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-v9d5k" event={"ID":"34905cec-e457-43a4-ba2e-acbb11e9c637","Type":"ContainerStarted","Data":"adbf0f980085fee15fd27380c0a275e94c635dc032c527ca58198cd54e59a31c"} Oct 05 07:54:32 crc kubenswrapper[4846]: I1005 07:54:32.812896 4846 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 05 07:54:33 crc kubenswrapper[4846]: I1005 07:54:33.822797 4846 generic.go:334] "Generic (PLEG): container finished" podID="34905cec-e457-43a4-ba2e-acbb11e9c637" containerID="c2a319fa9190a1b8c3f678cc0c79c8f911624f0436faad73fa07931d997c65ec" exitCode=0 Oct 05 07:54:33 crc kubenswrapper[4846]: I1005 07:54:33.822902 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-v9d5k" event={"ID":"34905cec-e457-43a4-ba2e-acbb11e9c637","Type":"ContainerDied","Data":"c2a319fa9190a1b8c3f678cc0c79c8f911624f0436faad73fa07931d997c65ec"} Oct 05 07:54:34 crc kubenswrapper[4846]: I1005 07:54:34.836608 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-v9d5k" event={"ID":"34905cec-e457-43a4-ba2e-acbb11e9c637","Type":"ContainerStarted","Data":"71f21dab5f133a6c8cfdc20d2e8f62f925ff42bf2f99b766251a617c141b7dd7"} Oct 05 07:54:34 crc kubenswrapper[4846]: I1005 07:54:34.877653 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-v9d5k" podStartSLOduration=2.471907935 podStartE2EDuration="3.877622985s" podCreationTimestamp="2025-10-05 07:54:31 +0000 UTC" firstStartedPulling="2025-10-05 07:54:32.811966986 +0000 UTC m=+3995.052819791" lastFinishedPulling="2025-10-05 07:54:34.217682026 +0000 UTC m=+3996.458534841" observedRunningTime="2025-10-05 07:54:34.869026951 +0000 UTC m=+3997.109879786" watchObservedRunningTime="2025-10-05 07:54:34.877622985 +0000 UTC m=+3997.118475800" Oct 05 07:54:41 crc kubenswrapper[4846]: I1005 07:54:41.519169 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-v9d5k" Oct 05 07:54:41 crc kubenswrapper[4846]: I1005 07:54:41.520567 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-v9d5k" Oct 05 07:54:41 crc kubenswrapper[4846]: I1005 07:54:41.605161 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-v9d5k" Oct 05 07:54:41 crc kubenswrapper[4846]: I1005 07:54:41.974067 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-v9d5k" Oct 05 07:54:42 crc kubenswrapper[4846]: I1005 07:54:42.034498 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-v9d5k"] Oct 05 07:54:43 crc kubenswrapper[4846]: I1005 07:54:43.925881 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-v9d5k" podUID="34905cec-e457-43a4-ba2e-acbb11e9c637" containerName="registry-server" containerID="cri-o://71f21dab5f133a6c8cfdc20d2e8f62f925ff42bf2f99b766251a617c141b7dd7" gracePeriod=2 Oct 05 07:54:44 crc kubenswrapper[4846]: I1005 07:54:44.869565 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-v9d5k" Oct 05 07:54:44 crc kubenswrapper[4846]: I1005 07:54:44.939242 4846 generic.go:334] "Generic (PLEG): container finished" podID="34905cec-e457-43a4-ba2e-acbb11e9c637" containerID="71f21dab5f133a6c8cfdc20d2e8f62f925ff42bf2f99b766251a617c141b7dd7" exitCode=0 Oct 05 07:54:44 crc kubenswrapper[4846]: I1005 07:54:44.939285 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-v9d5k" event={"ID":"34905cec-e457-43a4-ba2e-acbb11e9c637","Type":"ContainerDied","Data":"71f21dab5f133a6c8cfdc20d2e8f62f925ff42bf2f99b766251a617c141b7dd7"} Oct 05 07:54:44 crc kubenswrapper[4846]: I1005 07:54:44.939310 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-v9d5k" event={"ID":"34905cec-e457-43a4-ba2e-acbb11e9c637","Type":"ContainerDied","Data":"adbf0f980085fee15fd27380c0a275e94c635dc032c527ca58198cd54e59a31c"} Oct 05 07:54:44 crc kubenswrapper[4846]: I1005 07:54:44.939308 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-v9d5k" Oct 05 07:54:44 crc kubenswrapper[4846]: I1005 07:54:44.939323 4846 scope.go:117] "RemoveContainer" containerID="71f21dab5f133a6c8cfdc20d2e8f62f925ff42bf2f99b766251a617c141b7dd7" Oct 05 07:54:44 crc kubenswrapper[4846]: I1005 07:54:44.960646 4846 scope.go:117] "RemoveContainer" containerID="c2a319fa9190a1b8c3f678cc0c79c8f911624f0436faad73fa07931d997c65ec" Oct 05 07:54:44 crc kubenswrapper[4846]: I1005 07:54:44.976490 4846 scope.go:117] "RemoveContainer" containerID="23fd8b3e9310bf309d01fae62df1ac42dd336a9522a815293d6600447e145990" Oct 05 07:54:45 crc kubenswrapper[4846]: I1005 07:54:45.007429 4846 scope.go:117] "RemoveContainer" containerID="71f21dab5f133a6c8cfdc20d2e8f62f925ff42bf2f99b766251a617c141b7dd7" Oct 05 07:54:45 crc kubenswrapper[4846]: I1005 07:54:45.007720 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6sh55\" (UniqueName: \"kubernetes.io/projected/34905cec-e457-43a4-ba2e-acbb11e9c637-kube-api-access-6sh55\") pod \"34905cec-e457-43a4-ba2e-acbb11e9c637\" (UID: \"34905cec-e457-43a4-ba2e-acbb11e9c637\") " Oct 05 07:54:45 crc kubenswrapper[4846]: I1005 07:54:45.007761 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/34905cec-e457-43a4-ba2e-acbb11e9c637-catalog-content\") pod \"34905cec-e457-43a4-ba2e-acbb11e9c637\" (UID: \"34905cec-e457-43a4-ba2e-acbb11e9c637\") " Oct 05 07:54:45 crc kubenswrapper[4846]: I1005 07:54:45.007802 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/34905cec-e457-43a4-ba2e-acbb11e9c637-utilities\") pod \"34905cec-e457-43a4-ba2e-acbb11e9c637\" (UID: \"34905cec-e457-43a4-ba2e-acbb11e9c637\") " Oct 05 07:54:45 crc kubenswrapper[4846]: I1005 07:54:45.008543 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/34905cec-e457-43a4-ba2e-acbb11e9c637-utilities" (OuterVolumeSpecName: "utilities") pod "34905cec-e457-43a4-ba2e-acbb11e9c637" (UID: "34905cec-e457-43a4-ba2e-acbb11e9c637"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 07:54:45 crc kubenswrapper[4846]: E1005 07:54:45.007981 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"71f21dab5f133a6c8cfdc20d2e8f62f925ff42bf2f99b766251a617c141b7dd7\": container with ID starting with 71f21dab5f133a6c8cfdc20d2e8f62f925ff42bf2f99b766251a617c141b7dd7 not found: ID does not exist" containerID="71f21dab5f133a6c8cfdc20d2e8f62f925ff42bf2f99b766251a617c141b7dd7" Oct 05 07:54:45 crc kubenswrapper[4846]: I1005 07:54:45.008603 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"71f21dab5f133a6c8cfdc20d2e8f62f925ff42bf2f99b766251a617c141b7dd7"} err="failed to get container status \"71f21dab5f133a6c8cfdc20d2e8f62f925ff42bf2f99b766251a617c141b7dd7\": rpc error: code = NotFound desc = could not find container \"71f21dab5f133a6c8cfdc20d2e8f62f925ff42bf2f99b766251a617c141b7dd7\": container with ID starting with 71f21dab5f133a6c8cfdc20d2e8f62f925ff42bf2f99b766251a617c141b7dd7 not found: ID does not exist" Oct 05 07:54:45 crc kubenswrapper[4846]: I1005 07:54:45.008623 4846 scope.go:117] "RemoveContainer" containerID="c2a319fa9190a1b8c3f678cc0c79c8f911624f0436faad73fa07931d997c65ec" Oct 05 07:54:45 crc kubenswrapper[4846]: E1005 07:54:45.015500 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c2a319fa9190a1b8c3f678cc0c79c8f911624f0436faad73fa07931d997c65ec\": container with ID starting with c2a319fa9190a1b8c3f678cc0c79c8f911624f0436faad73fa07931d997c65ec not found: ID does not exist" containerID="c2a319fa9190a1b8c3f678cc0c79c8f911624f0436faad73fa07931d997c65ec" Oct 05 07:54:45 crc kubenswrapper[4846]: I1005 07:54:45.015534 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c2a319fa9190a1b8c3f678cc0c79c8f911624f0436faad73fa07931d997c65ec"} err="failed to get container status \"c2a319fa9190a1b8c3f678cc0c79c8f911624f0436faad73fa07931d997c65ec\": rpc error: code = NotFound desc = could not find container \"c2a319fa9190a1b8c3f678cc0c79c8f911624f0436faad73fa07931d997c65ec\": container with ID starting with c2a319fa9190a1b8c3f678cc0c79c8f911624f0436faad73fa07931d997c65ec not found: ID does not exist" Oct 05 07:54:45 crc kubenswrapper[4846]: I1005 07:54:45.015556 4846 scope.go:117] "RemoveContainer" containerID="23fd8b3e9310bf309d01fae62df1ac42dd336a9522a815293d6600447e145990" Oct 05 07:54:45 crc kubenswrapper[4846]: I1005 07:54:45.015563 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/34905cec-e457-43a4-ba2e-acbb11e9c637-kube-api-access-6sh55" (OuterVolumeSpecName: "kube-api-access-6sh55") pod "34905cec-e457-43a4-ba2e-acbb11e9c637" (UID: "34905cec-e457-43a4-ba2e-acbb11e9c637"). InnerVolumeSpecName "kube-api-access-6sh55". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:54:45 crc kubenswrapper[4846]: E1005 07:54:45.015937 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"23fd8b3e9310bf309d01fae62df1ac42dd336a9522a815293d6600447e145990\": container with ID starting with 23fd8b3e9310bf309d01fae62df1ac42dd336a9522a815293d6600447e145990 not found: ID does not exist" containerID="23fd8b3e9310bf309d01fae62df1ac42dd336a9522a815293d6600447e145990" Oct 05 07:54:45 crc kubenswrapper[4846]: I1005 07:54:45.015966 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"23fd8b3e9310bf309d01fae62df1ac42dd336a9522a815293d6600447e145990"} err="failed to get container status \"23fd8b3e9310bf309d01fae62df1ac42dd336a9522a815293d6600447e145990\": rpc error: code = NotFound desc = could not find container \"23fd8b3e9310bf309d01fae62df1ac42dd336a9522a815293d6600447e145990\": container with ID starting with 23fd8b3e9310bf309d01fae62df1ac42dd336a9522a815293d6600447e145990 not found: ID does not exist" Oct 05 07:54:45 crc kubenswrapper[4846]: I1005 07:54:45.050159 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/34905cec-e457-43a4-ba2e-acbb11e9c637-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "34905cec-e457-43a4-ba2e-acbb11e9c637" (UID: "34905cec-e457-43a4-ba2e-acbb11e9c637"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 07:54:45 crc kubenswrapper[4846]: I1005 07:54:45.108973 4846 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/34905cec-e457-43a4-ba2e-acbb11e9c637-utilities\") on node \"crc\" DevicePath \"\"" Oct 05 07:54:45 crc kubenswrapper[4846]: I1005 07:54:45.109009 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6sh55\" (UniqueName: \"kubernetes.io/projected/34905cec-e457-43a4-ba2e-acbb11e9c637-kube-api-access-6sh55\") on node \"crc\" DevicePath \"\"" Oct 05 07:54:45 crc kubenswrapper[4846]: I1005 07:54:45.109020 4846 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/34905cec-e457-43a4-ba2e-acbb11e9c637-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 05 07:54:45 crc kubenswrapper[4846]: I1005 07:54:45.287141 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-v9d5k"] Oct 05 07:54:45 crc kubenswrapper[4846]: I1005 07:54:45.295240 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-v9d5k"] Oct 05 07:54:46 crc kubenswrapper[4846]: I1005 07:54:46.513077 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="34905cec-e457-43a4-ba2e-acbb11e9c637" path="/var/lib/kubelet/pods/34905cec-e457-43a4-ba2e-acbb11e9c637/volumes" Oct 05 07:55:16 crc kubenswrapper[4846]: I1005 07:55:16.905060 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-f8r9s"] Oct 05 07:55:16 crc kubenswrapper[4846]: E1005 07:55:16.906573 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="34905cec-e457-43a4-ba2e-acbb11e9c637" containerName="registry-server" Oct 05 07:55:16 crc kubenswrapper[4846]: I1005 07:55:16.906599 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="34905cec-e457-43a4-ba2e-acbb11e9c637" containerName="registry-server" Oct 05 07:55:16 crc kubenswrapper[4846]: E1005 07:55:16.906625 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="34905cec-e457-43a4-ba2e-acbb11e9c637" containerName="extract-content" Oct 05 07:55:16 crc kubenswrapper[4846]: I1005 07:55:16.906640 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="34905cec-e457-43a4-ba2e-acbb11e9c637" containerName="extract-content" Oct 05 07:55:16 crc kubenswrapper[4846]: E1005 07:55:16.906704 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="34905cec-e457-43a4-ba2e-acbb11e9c637" containerName="extract-utilities" Oct 05 07:55:16 crc kubenswrapper[4846]: I1005 07:55:16.906716 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="34905cec-e457-43a4-ba2e-acbb11e9c637" containerName="extract-utilities" Oct 05 07:55:16 crc kubenswrapper[4846]: I1005 07:55:16.907152 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="34905cec-e457-43a4-ba2e-acbb11e9c637" containerName="registry-server" Oct 05 07:55:16 crc kubenswrapper[4846]: I1005 07:55:16.908953 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-f8r9s" Oct 05 07:55:16 crc kubenswrapper[4846]: I1005 07:55:16.936459 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-f8r9s"] Oct 05 07:55:17 crc kubenswrapper[4846]: I1005 07:55:17.094881 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r7lw5\" (UniqueName: \"kubernetes.io/projected/b0f2b619-e294-4a3a-935d-28b00ca65c33-kube-api-access-r7lw5\") pod \"redhat-marketplace-f8r9s\" (UID: \"b0f2b619-e294-4a3a-935d-28b00ca65c33\") " pod="openshift-marketplace/redhat-marketplace-f8r9s" Oct 05 07:55:17 crc kubenswrapper[4846]: I1005 07:55:17.094942 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b0f2b619-e294-4a3a-935d-28b00ca65c33-utilities\") pod \"redhat-marketplace-f8r9s\" (UID: \"b0f2b619-e294-4a3a-935d-28b00ca65c33\") " pod="openshift-marketplace/redhat-marketplace-f8r9s" Oct 05 07:55:17 crc kubenswrapper[4846]: I1005 07:55:17.095006 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b0f2b619-e294-4a3a-935d-28b00ca65c33-catalog-content\") pod \"redhat-marketplace-f8r9s\" (UID: \"b0f2b619-e294-4a3a-935d-28b00ca65c33\") " pod="openshift-marketplace/redhat-marketplace-f8r9s" Oct 05 07:55:17 crc kubenswrapper[4846]: I1005 07:55:17.197846 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r7lw5\" (UniqueName: \"kubernetes.io/projected/b0f2b619-e294-4a3a-935d-28b00ca65c33-kube-api-access-r7lw5\") pod \"redhat-marketplace-f8r9s\" (UID: \"b0f2b619-e294-4a3a-935d-28b00ca65c33\") " pod="openshift-marketplace/redhat-marketplace-f8r9s" Oct 05 07:55:17 crc kubenswrapper[4846]: I1005 07:55:17.197927 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b0f2b619-e294-4a3a-935d-28b00ca65c33-utilities\") pod \"redhat-marketplace-f8r9s\" (UID: \"b0f2b619-e294-4a3a-935d-28b00ca65c33\") " pod="openshift-marketplace/redhat-marketplace-f8r9s" Oct 05 07:55:17 crc kubenswrapper[4846]: I1005 07:55:17.197986 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b0f2b619-e294-4a3a-935d-28b00ca65c33-catalog-content\") pod \"redhat-marketplace-f8r9s\" (UID: \"b0f2b619-e294-4a3a-935d-28b00ca65c33\") " pod="openshift-marketplace/redhat-marketplace-f8r9s" Oct 05 07:55:17 crc kubenswrapper[4846]: I1005 07:55:17.198704 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b0f2b619-e294-4a3a-935d-28b00ca65c33-utilities\") pod \"redhat-marketplace-f8r9s\" (UID: \"b0f2b619-e294-4a3a-935d-28b00ca65c33\") " pod="openshift-marketplace/redhat-marketplace-f8r9s" Oct 05 07:55:17 crc kubenswrapper[4846]: I1005 07:55:17.198741 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b0f2b619-e294-4a3a-935d-28b00ca65c33-catalog-content\") pod \"redhat-marketplace-f8r9s\" (UID: \"b0f2b619-e294-4a3a-935d-28b00ca65c33\") " pod="openshift-marketplace/redhat-marketplace-f8r9s" Oct 05 07:55:17 crc kubenswrapper[4846]: I1005 07:55:17.236540 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r7lw5\" (UniqueName: \"kubernetes.io/projected/b0f2b619-e294-4a3a-935d-28b00ca65c33-kube-api-access-r7lw5\") pod \"redhat-marketplace-f8r9s\" (UID: \"b0f2b619-e294-4a3a-935d-28b00ca65c33\") " pod="openshift-marketplace/redhat-marketplace-f8r9s" Oct 05 07:55:17 crc kubenswrapper[4846]: I1005 07:55:17.238402 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-f8r9s" Oct 05 07:55:17 crc kubenswrapper[4846]: I1005 07:55:17.695670 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-f8r9s"] Oct 05 07:55:18 crc kubenswrapper[4846]: I1005 07:55:18.228045 4846 generic.go:334] "Generic (PLEG): container finished" podID="b0f2b619-e294-4a3a-935d-28b00ca65c33" containerID="6cd99914646adf17724d8bad0f5bd4b01d9dab3d128548fe6c07156301263d27" exitCode=0 Oct 05 07:55:18 crc kubenswrapper[4846]: I1005 07:55:18.228109 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-f8r9s" event={"ID":"b0f2b619-e294-4a3a-935d-28b00ca65c33","Type":"ContainerDied","Data":"6cd99914646adf17724d8bad0f5bd4b01d9dab3d128548fe6c07156301263d27"} Oct 05 07:55:18 crc kubenswrapper[4846]: I1005 07:55:18.228446 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-f8r9s" event={"ID":"b0f2b619-e294-4a3a-935d-28b00ca65c33","Type":"ContainerStarted","Data":"957ca81f3af86f53db52994dab5e2397f145ac3e75c12459ff42bd309bc962a6"} Oct 05 07:55:19 crc kubenswrapper[4846]: I1005 07:55:19.240138 4846 generic.go:334] "Generic (PLEG): container finished" podID="b0f2b619-e294-4a3a-935d-28b00ca65c33" containerID="21fca854ca56b6b44b30431d01528f7ec410e69554736dce1d01d8ee65196668" exitCode=0 Oct 05 07:55:19 crc kubenswrapper[4846]: I1005 07:55:19.240231 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-f8r9s" event={"ID":"b0f2b619-e294-4a3a-935d-28b00ca65c33","Type":"ContainerDied","Data":"21fca854ca56b6b44b30431d01528f7ec410e69554736dce1d01d8ee65196668"} Oct 05 07:55:20 crc kubenswrapper[4846]: I1005 07:55:20.252620 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-f8r9s" event={"ID":"b0f2b619-e294-4a3a-935d-28b00ca65c33","Type":"ContainerStarted","Data":"1397f1539826ddbda2b8ba2162fd067e00723d01d98fd4f9a7652da404575b8c"} Oct 05 07:55:20 crc kubenswrapper[4846]: I1005 07:55:20.273158 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-f8r9s" podStartSLOduration=2.573234593 podStartE2EDuration="4.273137571s" podCreationTimestamp="2025-10-05 07:55:16 +0000 UTC" firstStartedPulling="2025-10-05 07:55:18.229873083 +0000 UTC m=+4040.470725908" lastFinishedPulling="2025-10-05 07:55:19.929776071 +0000 UTC m=+4042.170628886" observedRunningTime="2025-10-05 07:55:20.271036284 +0000 UTC m=+4042.511889079" watchObservedRunningTime="2025-10-05 07:55:20.273137571 +0000 UTC m=+4042.513990366" Oct 05 07:55:23 crc kubenswrapper[4846]: I1005 07:55:23.325469 4846 patch_prober.go:28] interesting pod/machine-config-daemon-fscvf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 07:55:23 crc kubenswrapper[4846]: I1005 07:55:23.326283 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 07:55:27 crc kubenswrapper[4846]: I1005 07:55:27.239402 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-f8r9s" Oct 05 07:55:27 crc kubenswrapper[4846]: I1005 07:55:27.239836 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-f8r9s" Oct 05 07:55:27 crc kubenswrapper[4846]: I1005 07:55:27.314594 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-f8r9s" Oct 05 07:55:27 crc kubenswrapper[4846]: I1005 07:55:27.401129 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-f8r9s" Oct 05 07:55:27 crc kubenswrapper[4846]: I1005 07:55:27.561500 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-f8r9s"] Oct 05 07:55:29 crc kubenswrapper[4846]: I1005 07:55:29.339520 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-f8r9s" podUID="b0f2b619-e294-4a3a-935d-28b00ca65c33" containerName="registry-server" containerID="cri-o://1397f1539826ddbda2b8ba2162fd067e00723d01d98fd4f9a7652da404575b8c" gracePeriod=2 Oct 05 07:55:29 crc kubenswrapper[4846]: I1005 07:55:29.777006 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-f8r9s" Oct 05 07:55:29 crc kubenswrapper[4846]: I1005 07:55:29.912768 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r7lw5\" (UniqueName: \"kubernetes.io/projected/b0f2b619-e294-4a3a-935d-28b00ca65c33-kube-api-access-r7lw5\") pod \"b0f2b619-e294-4a3a-935d-28b00ca65c33\" (UID: \"b0f2b619-e294-4a3a-935d-28b00ca65c33\") " Oct 05 07:55:29 crc kubenswrapper[4846]: I1005 07:55:29.912827 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b0f2b619-e294-4a3a-935d-28b00ca65c33-catalog-content\") pod \"b0f2b619-e294-4a3a-935d-28b00ca65c33\" (UID: \"b0f2b619-e294-4a3a-935d-28b00ca65c33\") " Oct 05 07:55:29 crc kubenswrapper[4846]: I1005 07:55:29.912939 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b0f2b619-e294-4a3a-935d-28b00ca65c33-utilities\") pod \"b0f2b619-e294-4a3a-935d-28b00ca65c33\" (UID: \"b0f2b619-e294-4a3a-935d-28b00ca65c33\") " Oct 05 07:55:29 crc kubenswrapper[4846]: I1005 07:55:29.914374 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b0f2b619-e294-4a3a-935d-28b00ca65c33-utilities" (OuterVolumeSpecName: "utilities") pod "b0f2b619-e294-4a3a-935d-28b00ca65c33" (UID: "b0f2b619-e294-4a3a-935d-28b00ca65c33"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 07:55:29 crc kubenswrapper[4846]: I1005 07:55:29.928425 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b0f2b619-e294-4a3a-935d-28b00ca65c33-kube-api-access-r7lw5" (OuterVolumeSpecName: "kube-api-access-r7lw5") pod "b0f2b619-e294-4a3a-935d-28b00ca65c33" (UID: "b0f2b619-e294-4a3a-935d-28b00ca65c33"). InnerVolumeSpecName "kube-api-access-r7lw5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:55:29 crc kubenswrapper[4846]: I1005 07:55:29.932409 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b0f2b619-e294-4a3a-935d-28b00ca65c33-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b0f2b619-e294-4a3a-935d-28b00ca65c33" (UID: "b0f2b619-e294-4a3a-935d-28b00ca65c33"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 07:55:30 crc kubenswrapper[4846]: I1005 07:55:30.014620 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r7lw5\" (UniqueName: \"kubernetes.io/projected/b0f2b619-e294-4a3a-935d-28b00ca65c33-kube-api-access-r7lw5\") on node \"crc\" DevicePath \"\"" Oct 05 07:55:30 crc kubenswrapper[4846]: I1005 07:55:30.014654 4846 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b0f2b619-e294-4a3a-935d-28b00ca65c33-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 05 07:55:30 crc kubenswrapper[4846]: I1005 07:55:30.014680 4846 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b0f2b619-e294-4a3a-935d-28b00ca65c33-utilities\") on node \"crc\" DevicePath \"\"" Oct 05 07:55:30 crc kubenswrapper[4846]: I1005 07:55:30.356577 4846 generic.go:334] "Generic (PLEG): container finished" podID="b0f2b619-e294-4a3a-935d-28b00ca65c33" containerID="1397f1539826ddbda2b8ba2162fd067e00723d01d98fd4f9a7652da404575b8c" exitCode=0 Oct 05 07:55:30 crc kubenswrapper[4846]: I1005 07:55:30.356644 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-f8r9s" Oct 05 07:55:30 crc kubenswrapper[4846]: I1005 07:55:30.356660 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-f8r9s" event={"ID":"b0f2b619-e294-4a3a-935d-28b00ca65c33","Type":"ContainerDied","Data":"1397f1539826ddbda2b8ba2162fd067e00723d01d98fd4f9a7652da404575b8c"} Oct 05 07:55:30 crc kubenswrapper[4846]: I1005 07:55:30.356825 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-f8r9s" event={"ID":"b0f2b619-e294-4a3a-935d-28b00ca65c33","Type":"ContainerDied","Data":"957ca81f3af86f53db52994dab5e2397f145ac3e75c12459ff42bd309bc962a6"} Oct 05 07:55:30 crc kubenswrapper[4846]: I1005 07:55:30.356859 4846 scope.go:117] "RemoveContainer" containerID="1397f1539826ddbda2b8ba2162fd067e00723d01d98fd4f9a7652da404575b8c" Oct 05 07:55:30 crc kubenswrapper[4846]: I1005 07:55:30.408248 4846 scope.go:117] "RemoveContainer" containerID="21fca854ca56b6b44b30431d01528f7ec410e69554736dce1d01d8ee65196668" Oct 05 07:55:30 crc kubenswrapper[4846]: I1005 07:55:30.452232 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-f8r9s"] Oct 05 07:55:30 crc kubenswrapper[4846]: I1005 07:55:30.459373 4846 scope.go:117] "RemoveContainer" containerID="6cd99914646adf17724d8bad0f5bd4b01d9dab3d128548fe6c07156301263d27" Oct 05 07:55:30 crc kubenswrapper[4846]: I1005 07:55:30.466631 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-f8r9s"] Oct 05 07:55:30 crc kubenswrapper[4846]: I1005 07:55:30.516466 4846 scope.go:117] "RemoveContainer" containerID="1397f1539826ddbda2b8ba2162fd067e00723d01d98fd4f9a7652da404575b8c" Oct 05 07:55:30 crc kubenswrapper[4846]: E1005 07:55:30.522961 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1397f1539826ddbda2b8ba2162fd067e00723d01d98fd4f9a7652da404575b8c\": container with ID starting with 1397f1539826ddbda2b8ba2162fd067e00723d01d98fd4f9a7652da404575b8c not found: ID does not exist" containerID="1397f1539826ddbda2b8ba2162fd067e00723d01d98fd4f9a7652da404575b8c" Oct 05 07:55:30 crc kubenswrapper[4846]: I1005 07:55:30.523003 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1397f1539826ddbda2b8ba2162fd067e00723d01d98fd4f9a7652da404575b8c"} err="failed to get container status \"1397f1539826ddbda2b8ba2162fd067e00723d01d98fd4f9a7652da404575b8c\": rpc error: code = NotFound desc = could not find container \"1397f1539826ddbda2b8ba2162fd067e00723d01d98fd4f9a7652da404575b8c\": container with ID starting with 1397f1539826ddbda2b8ba2162fd067e00723d01d98fd4f9a7652da404575b8c not found: ID does not exist" Oct 05 07:55:30 crc kubenswrapper[4846]: I1005 07:55:30.523024 4846 scope.go:117] "RemoveContainer" containerID="21fca854ca56b6b44b30431d01528f7ec410e69554736dce1d01d8ee65196668" Oct 05 07:55:30 crc kubenswrapper[4846]: I1005 07:55:30.524462 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b0f2b619-e294-4a3a-935d-28b00ca65c33" path="/var/lib/kubelet/pods/b0f2b619-e294-4a3a-935d-28b00ca65c33/volumes" Oct 05 07:55:30 crc kubenswrapper[4846]: E1005 07:55:30.527383 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"21fca854ca56b6b44b30431d01528f7ec410e69554736dce1d01d8ee65196668\": container with ID starting with 21fca854ca56b6b44b30431d01528f7ec410e69554736dce1d01d8ee65196668 not found: ID does not exist" containerID="21fca854ca56b6b44b30431d01528f7ec410e69554736dce1d01d8ee65196668" Oct 05 07:55:30 crc kubenswrapper[4846]: I1005 07:55:30.527444 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"21fca854ca56b6b44b30431d01528f7ec410e69554736dce1d01d8ee65196668"} err="failed to get container status \"21fca854ca56b6b44b30431d01528f7ec410e69554736dce1d01d8ee65196668\": rpc error: code = NotFound desc = could not find container \"21fca854ca56b6b44b30431d01528f7ec410e69554736dce1d01d8ee65196668\": container with ID starting with 21fca854ca56b6b44b30431d01528f7ec410e69554736dce1d01d8ee65196668 not found: ID does not exist" Oct 05 07:55:30 crc kubenswrapper[4846]: I1005 07:55:30.527478 4846 scope.go:117] "RemoveContainer" containerID="6cd99914646adf17724d8bad0f5bd4b01d9dab3d128548fe6c07156301263d27" Oct 05 07:55:30 crc kubenswrapper[4846]: E1005 07:55:30.529476 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6cd99914646adf17724d8bad0f5bd4b01d9dab3d128548fe6c07156301263d27\": container with ID starting with 6cd99914646adf17724d8bad0f5bd4b01d9dab3d128548fe6c07156301263d27 not found: ID does not exist" containerID="6cd99914646adf17724d8bad0f5bd4b01d9dab3d128548fe6c07156301263d27" Oct 05 07:55:30 crc kubenswrapper[4846]: I1005 07:55:30.529512 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6cd99914646adf17724d8bad0f5bd4b01d9dab3d128548fe6c07156301263d27"} err="failed to get container status \"6cd99914646adf17724d8bad0f5bd4b01d9dab3d128548fe6c07156301263d27\": rpc error: code = NotFound desc = could not find container \"6cd99914646adf17724d8bad0f5bd4b01d9dab3d128548fe6c07156301263d27\": container with ID starting with 6cd99914646adf17724d8bad0f5bd4b01d9dab3d128548fe6c07156301263d27 not found: ID does not exist" Oct 05 07:55:53 crc kubenswrapper[4846]: I1005 07:55:53.325645 4846 patch_prober.go:28] interesting pod/machine-config-daemon-fscvf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 07:55:53 crc kubenswrapper[4846]: I1005 07:55:53.326282 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 07:56:23 crc kubenswrapper[4846]: I1005 07:56:23.325741 4846 patch_prober.go:28] interesting pod/machine-config-daemon-fscvf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 07:56:23 crc kubenswrapper[4846]: I1005 07:56:23.326618 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 07:56:23 crc kubenswrapper[4846]: I1005 07:56:23.326723 4846 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" Oct 05 07:56:23 crc kubenswrapper[4846]: I1005 07:56:23.327655 4846 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"dccf837d14b883290775f3ba1aae37388eb10cbc64bfe9ddcfac744876bcccff"} pod="openshift-machine-config-operator/machine-config-daemon-fscvf" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 05 07:56:23 crc kubenswrapper[4846]: I1005 07:56:23.327771 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" containerName="machine-config-daemon" containerID="cri-o://dccf837d14b883290775f3ba1aae37388eb10cbc64bfe9ddcfac744876bcccff" gracePeriod=600 Oct 05 07:56:23 crc kubenswrapper[4846]: I1005 07:56:23.883607 4846 generic.go:334] "Generic (PLEG): container finished" podID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" containerID="dccf837d14b883290775f3ba1aae37388eb10cbc64bfe9ddcfac744876bcccff" exitCode=0 Oct 05 07:56:23 crc kubenswrapper[4846]: I1005 07:56:23.883718 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" event={"ID":"ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7","Type":"ContainerDied","Data":"dccf837d14b883290775f3ba1aae37388eb10cbc64bfe9ddcfac744876bcccff"} Oct 05 07:56:23 crc kubenswrapper[4846]: I1005 07:56:23.884458 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" event={"ID":"ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7","Type":"ContainerStarted","Data":"2d2e59cb51a3cd322bc7aca627ddfa4d48c6f5f95b26a7264c472515babd8d10"} Oct 05 07:56:23 crc kubenswrapper[4846]: I1005 07:56:23.884506 4846 scope.go:117] "RemoveContainer" containerID="a8dd7af9a49a3eba2b3ea4d7586f51b807c63a3b657f800b5182c8288f292750" Oct 05 07:58:23 crc kubenswrapper[4846]: I1005 07:58:23.324714 4846 patch_prober.go:28] interesting pod/machine-config-daemon-fscvf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 07:58:23 crc kubenswrapper[4846]: I1005 07:58:23.325363 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 07:58:23 crc kubenswrapper[4846]: I1005 07:58:23.706622 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-7jnwm"] Oct 05 07:58:23 crc kubenswrapper[4846]: E1005 07:58:23.707704 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b0f2b619-e294-4a3a-935d-28b00ca65c33" containerName="extract-utilities" Oct 05 07:58:23 crc kubenswrapper[4846]: I1005 07:58:23.707900 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="b0f2b619-e294-4a3a-935d-28b00ca65c33" containerName="extract-utilities" Oct 05 07:58:23 crc kubenswrapper[4846]: E1005 07:58:23.708044 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b0f2b619-e294-4a3a-935d-28b00ca65c33" containerName="registry-server" Oct 05 07:58:23 crc kubenswrapper[4846]: I1005 07:58:23.708166 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="b0f2b619-e294-4a3a-935d-28b00ca65c33" containerName="registry-server" Oct 05 07:58:23 crc kubenswrapper[4846]: E1005 07:58:23.708340 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b0f2b619-e294-4a3a-935d-28b00ca65c33" containerName="extract-content" Oct 05 07:58:23 crc kubenswrapper[4846]: I1005 07:58:23.708467 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="b0f2b619-e294-4a3a-935d-28b00ca65c33" containerName="extract-content" Oct 05 07:58:23 crc kubenswrapper[4846]: I1005 07:58:23.708811 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="b0f2b619-e294-4a3a-935d-28b00ca65c33" containerName="registry-server" Oct 05 07:58:23 crc kubenswrapper[4846]: I1005 07:58:23.711511 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7jnwm" Oct 05 07:58:23 crc kubenswrapper[4846]: I1005 07:58:23.731725 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-7jnwm"] Oct 05 07:58:23 crc kubenswrapper[4846]: I1005 07:58:23.823345 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-btmtm\" (UniqueName: \"kubernetes.io/projected/666047b1-bc37-4f34-b26b-80934e688269-kube-api-access-btmtm\") pod \"redhat-operators-7jnwm\" (UID: \"666047b1-bc37-4f34-b26b-80934e688269\") " pod="openshift-marketplace/redhat-operators-7jnwm" Oct 05 07:58:23 crc kubenswrapper[4846]: I1005 07:58:23.823414 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/666047b1-bc37-4f34-b26b-80934e688269-catalog-content\") pod \"redhat-operators-7jnwm\" (UID: \"666047b1-bc37-4f34-b26b-80934e688269\") " pod="openshift-marketplace/redhat-operators-7jnwm" Oct 05 07:58:23 crc kubenswrapper[4846]: I1005 07:58:23.823863 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/666047b1-bc37-4f34-b26b-80934e688269-utilities\") pod \"redhat-operators-7jnwm\" (UID: \"666047b1-bc37-4f34-b26b-80934e688269\") " pod="openshift-marketplace/redhat-operators-7jnwm" Oct 05 07:58:23 crc kubenswrapper[4846]: I1005 07:58:23.925745 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/666047b1-bc37-4f34-b26b-80934e688269-utilities\") pod \"redhat-operators-7jnwm\" (UID: \"666047b1-bc37-4f34-b26b-80934e688269\") " pod="openshift-marketplace/redhat-operators-7jnwm" Oct 05 07:58:23 crc kubenswrapper[4846]: I1005 07:58:23.925864 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-btmtm\" (UniqueName: \"kubernetes.io/projected/666047b1-bc37-4f34-b26b-80934e688269-kube-api-access-btmtm\") pod \"redhat-operators-7jnwm\" (UID: \"666047b1-bc37-4f34-b26b-80934e688269\") " pod="openshift-marketplace/redhat-operators-7jnwm" Oct 05 07:58:23 crc kubenswrapper[4846]: I1005 07:58:23.925922 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/666047b1-bc37-4f34-b26b-80934e688269-catalog-content\") pod \"redhat-operators-7jnwm\" (UID: \"666047b1-bc37-4f34-b26b-80934e688269\") " pod="openshift-marketplace/redhat-operators-7jnwm" Oct 05 07:58:23 crc kubenswrapper[4846]: I1005 07:58:23.926752 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/666047b1-bc37-4f34-b26b-80934e688269-catalog-content\") pod \"redhat-operators-7jnwm\" (UID: \"666047b1-bc37-4f34-b26b-80934e688269\") " pod="openshift-marketplace/redhat-operators-7jnwm" Oct 05 07:58:23 crc kubenswrapper[4846]: I1005 07:58:23.926782 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/666047b1-bc37-4f34-b26b-80934e688269-utilities\") pod \"redhat-operators-7jnwm\" (UID: \"666047b1-bc37-4f34-b26b-80934e688269\") " pod="openshift-marketplace/redhat-operators-7jnwm" Oct 05 07:58:23 crc kubenswrapper[4846]: I1005 07:58:23.958317 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-btmtm\" (UniqueName: \"kubernetes.io/projected/666047b1-bc37-4f34-b26b-80934e688269-kube-api-access-btmtm\") pod \"redhat-operators-7jnwm\" (UID: \"666047b1-bc37-4f34-b26b-80934e688269\") " pod="openshift-marketplace/redhat-operators-7jnwm" Oct 05 07:58:24 crc kubenswrapper[4846]: I1005 07:58:24.038543 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7jnwm" Oct 05 07:58:24 crc kubenswrapper[4846]: I1005 07:58:24.261098 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-7jnwm"] Oct 05 07:58:25 crc kubenswrapper[4846]: I1005 07:58:25.062386 4846 generic.go:334] "Generic (PLEG): container finished" podID="666047b1-bc37-4f34-b26b-80934e688269" containerID="55d82e3806aa891bfc9caf95082338217cb5d8dc7f8a55c29c8897d3f27b5fc6" exitCode=0 Oct 05 07:58:25 crc kubenswrapper[4846]: I1005 07:58:25.062503 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7jnwm" event={"ID":"666047b1-bc37-4f34-b26b-80934e688269","Type":"ContainerDied","Data":"55d82e3806aa891bfc9caf95082338217cb5d8dc7f8a55c29c8897d3f27b5fc6"} Oct 05 07:58:25 crc kubenswrapper[4846]: I1005 07:58:25.062757 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7jnwm" event={"ID":"666047b1-bc37-4f34-b26b-80934e688269","Type":"ContainerStarted","Data":"2c4a0d299a76a0f8f83516f1ae70201a03e64f6287105e39c815fecdf113adb0"} Oct 05 07:58:27 crc kubenswrapper[4846]: I1005 07:58:27.088542 4846 generic.go:334] "Generic (PLEG): container finished" podID="666047b1-bc37-4f34-b26b-80934e688269" containerID="905551d23a7a7c6984c7d2fc111dbc7e307fff6e99cd2b62d660b5fbe9e41d5f" exitCode=0 Oct 05 07:58:27 crc kubenswrapper[4846]: I1005 07:58:27.088623 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7jnwm" event={"ID":"666047b1-bc37-4f34-b26b-80934e688269","Type":"ContainerDied","Data":"905551d23a7a7c6984c7d2fc111dbc7e307fff6e99cd2b62d660b5fbe9e41d5f"} Oct 05 07:58:28 crc kubenswrapper[4846]: I1005 07:58:28.100885 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7jnwm" event={"ID":"666047b1-bc37-4f34-b26b-80934e688269","Type":"ContainerStarted","Data":"7730108d8dbaffa871e221cdb8173e4e15bca26d5df0fe485afc5d6966a62682"} Oct 05 07:58:28 crc kubenswrapper[4846]: I1005 07:58:28.122017 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-7jnwm" podStartSLOduration=2.637659092 podStartE2EDuration="5.121993611s" podCreationTimestamp="2025-10-05 07:58:23 +0000 UTC" firstStartedPulling="2025-10-05 07:58:25.065452983 +0000 UTC m=+4227.306305798" lastFinishedPulling="2025-10-05 07:58:27.549787502 +0000 UTC m=+4229.790640317" observedRunningTime="2025-10-05 07:58:28.116953326 +0000 UTC m=+4230.357806151" watchObservedRunningTime="2025-10-05 07:58:28.121993611 +0000 UTC m=+4230.362846426" Oct 05 07:58:34 crc kubenswrapper[4846]: I1005 07:58:34.038818 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-7jnwm" Oct 05 07:58:34 crc kubenswrapper[4846]: I1005 07:58:34.039660 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-7jnwm" Oct 05 07:58:34 crc kubenswrapper[4846]: I1005 07:58:34.106873 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-7jnwm" Oct 05 07:58:34 crc kubenswrapper[4846]: I1005 07:58:34.238085 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-7jnwm" Oct 05 07:58:34 crc kubenswrapper[4846]: I1005 07:58:34.358581 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-7jnwm"] Oct 05 07:58:36 crc kubenswrapper[4846]: I1005 07:58:36.177450 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-7jnwm" podUID="666047b1-bc37-4f34-b26b-80934e688269" containerName="registry-server" containerID="cri-o://7730108d8dbaffa871e221cdb8173e4e15bca26d5df0fe485afc5d6966a62682" gracePeriod=2 Oct 05 07:58:36 crc kubenswrapper[4846]: I1005 07:58:36.691691 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7jnwm" Oct 05 07:58:36 crc kubenswrapper[4846]: I1005 07:58:36.822050 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/666047b1-bc37-4f34-b26b-80934e688269-catalog-content\") pod \"666047b1-bc37-4f34-b26b-80934e688269\" (UID: \"666047b1-bc37-4f34-b26b-80934e688269\") " Oct 05 07:58:36 crc kubenswrapper[4846]: I1005 07:58:36.822240 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-btmtm\" (UniqueName: \"kubernetes.io/projected/666047b1-bc37-4f34-b26b-80934e688269-kube-api-access-btmtm\") pod \"666047b1-bc37-4f34-b26b-80934e688269\" (UID: \"666047b1-bc37-4f34-b26b-80934e688269\") " Oct 05 07:58:36 crc kubenswrapper[4846]: I1005 07:58:36.822441 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/666047b1-bc37-4f34-b26b-80934e688269-utilities\") pod \"666047b1-bc37-4f34-b26b-80934e688269\" (UID: \"666047b1-bc37-4f34-b26b-80934e688269\") " Oct 05 07:58:36 crc kubenswrapper[4846]: I1005 07:58:36.823372 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/666047b1-bc37-4f34-b26b-80934e688269-utilities" (OuterVolumeSpecName: "utilities") pod "666047b1-bc37-4f34-b26b-80934e688269" (UID: "666047b1-bc37-4f34-b26b-80934e688269"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 07:58:36 crc kubenswrapper[4846]: I1005 07:58:36.828121 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/666047b1-bc37-4f34-b26b-80934e688269-kube-api-access-btmtm" (OuterVolumeSpecName: "kube-api-access-btmtm") pod "666047b1-bc37-4f34-b26b-80934e688269" (UID: "666047b1-bc37-4f34-b26b-80934e688269"). InnerVolumeSpecName "kube-api-access-btmtm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:58:36 crc kubenswrapper[4846]: I1005 07:58:36.924931 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-btmtm\" (UniqueName: \"kubernetes.io/projected/666047b1-bc37-4f34-b26b-80934e688269-kube-api-access-btmtm\") on node \"crc\" DevicePath \"\"" Oct 05 07:58:36 crc kubenswrapper[4846]: I1005 07:58:36.925339 4846 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/666047b1-bc37-4f34-b26b-80934e688269-utilities\") on node \"crc\" DevicePath \"\"" Oct 05 07:58:37 crc kubenswrapper[4846]: I1005 07:58:37.191808 4846 generic.go:334] "Generic (PLEG): container finished" podID="666047b1-bc37-4f34-b26b-80934e688269" containerID="7730108d8dbaffa871e221cdb8173e4e15bca26d5df0fe485afc5d6966a62682" exitCode=0 Oct 05 07:58:37 crc kubenswrapper[4846]: I1005 07:58:37.191896 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7jnwm" event={"ID":"666047b1-bc37-4f34-b26b-80934e688269","Type":"ContainerDied","Data":"7730108d8dbaffa871e221cdb8173e4e15bca26d5df0fe485afc5d6966a62682"} Oct 05 07:58:37 crc kubenswrapper[4846]: I1005 07:58:37.191959 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7jnwm" event={"ID":"666047b1-bc37-4f34-b26b-80934e688269","Type":"ContainerDied","Data":"2c4a0d299a76a0f8f83516f1ae70201a03e64f6287105e39c815fecdf113adb0"} Oct 05 07:58:37 crc kubenswrapper[4846]: I1005 07:58:37.191995 4846 scope.go:117] "RemoveContainer" containerID="7730108d8dbaffa871e221cdb8173e4e15bca26d5df0fe485afc5d6966a62682" Oct 05 07:58:37 crc kubenswrapper[4846]: I1005 07:58:37.192000 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7jnwm" Oct 05 07:58:37 crc kubenswrapper[4846]: I1005 07:58:37.220016 4846 scope.go:117] "RemoveContainer" containerID="905551d23a7a7c6984c7d2fc111dbc7e307fff6e99cd2b62d660b5fbe9e41d5f" Oct 05 07:58:37 crc kubenswrapper[4846]: I1005 07:58:37.243077 4846 scope.go:117] "RemoveContainer" containerID="55d82e3806aa891bfc9caf95082338217cb5d8dc7f8a55c29c8897d3f27b5fc6" Oct 05 07:58:37 crc kubenswrapper[4846]: I1005 07:58:37.292585 4846 scope.go:117] "RemoveContainer" containerID="7730108d8dbaffa871e221cdb8173e4e15bca26d5df0fe485afc5d6966a62682" Oct 05 07:58:37 crc kubenswrapper[4846]: E1005 07:58:37.293240 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7730108d8dbaffa871e221cdb8173e4e15bca26d5df0fe485afc5d6966a62682\": container with ID starting with 7730108d8dbaffa871e221cdb8173e4e15bca26d5df0fe485afc5d6966a62682 not found: ID does not exist" containerID="7730108d8dbaffa871e221cdb8173e4e15bca26d5df0fe485afc5d6966a62682" Oct 05 07:58:37 crc kubenswrapper[4846]: I1005 07:58:37.293270 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7730108d8dbaffa871e221cdb8173e4e15bca26d5df0fe485afc5d6966a62682"} err="failed to get container status \"7730108d8dbaffa871e221cdb8173e4e15bca26d5df0fe485afc5d6966a62682\": rpc error: code = NotFound desc = could not find container \"7730108d8dbaffa871e221cdb8173e4e15bca26d5df0fe485afc5d6966a62682\": container with ID starting with 7730108d8dbaffa871e221cdb8173e4e15bca26d5df0fe485afc5d6966a62682 not found: ID does not exist" Oct 05 07:58:37 crc kubenswrapper[4846]: I1005 07:58:37.293292 4846 scope.go:117] "RemoveContainer" containerID="905551d23a7a7c6984c7d2fc111dbc7e307fff6e99cd2b62d660b5fbe9e41d5f" Oct 05 07:58:37 crc kubenswrapper[4846]: E1005 07:58:37.293825 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"905551d23a7a7c6984c7d2fc111dbc7e307fff6e99cd2b62d660b5fbe9e41d5f\": container with ID starting with 905551d23a7a7c6984c7d2fc111dbc7e307fff6e99cd2b62d660b5fbe9e41d5f not found: ID does not exist" containerID="905551d23a7a7c6984c7d2fc111dbc7e307fff6e99cd2b62d660b5fbe9e41d5f" Oct 05 07:58:37 crc kubenswrapper[4846]: I1005 07:58:37.293846 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"905551d23a7a7c6984c7d2fc111dbc7e307fff6e99cd2b62d660b5fbe9e41d5f"} err="failed to get container status \"905551d23a7a7c6984c7d2fc111dbc7e307fff6e99cd2b62d660b5fbe9e41d5f\": rpc error: code = NotFound desc = could not find container \"905551d23a7a7c6984c7d2fc111dbc7e307fff6e99cd2b62d660b5fbe9e41d5f\": container with ID starting with 905551d23a7a7c6984c7d2fc111dbc7e307fff6e99cd2b62d660b5fbe9e41d5f not found: ID does not exist" Oct 05 07:58:37 crc kubenswrapper[4846]: I1005 07:58:37.293862 4846 scope.go:117] "RemoveContainer" containerID="55d82e3806aa891bfc9caf95082338217cb5d8dc7f8a55c29c8897d3f27b5fc6" Oct 05 07:58:37 crc kubenswrapper[4846]: E1005 07:58:37.294445 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"55d82e3806aa891bfc9caf95082338217cb5d8dc7f8a55c29c8897d3f27b5fc6\": container with ID starting with 55d82e3806aa891bfc9caf95082338217cb5d8dc7f8a55c29c8897d3f27b5fc6 not found: ID does not exist" containerID="55d82e3806aa891bfc9caf95082338217cb5d8dc7f8a55c29c8897d3f27b5fc6" Oct 05 07:58:37 crc kubenswrapper[4846]: I1005 07:58:37.294543 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"55d82e3806aa891bfc9caf95082338217cb5d8dc7f8a55c29c8897d3f27b5fc6"} err="failed to get container status \"55d82e3806aa891bfc9caf95082338217cb5d8dc7f8a55c29c8897d3f27b5fc6\": rpc error: code = NotFound desc = could not find container \"55d82e3806aa891bfc9caf95082338217cb5d8dc7f8a55c29c8897d3f27b5fc6\": container with ID starting with 55d82e3806aa891bfc9caf95082338217cb5d8dc7f8a55c29c8897d3f27b5fc6 not found: ID does not exist" Oct 05 07:58:37 crc kubenswrapper[4846]: I1005 07:58:37.876470 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/666047b1-bc37-4f34-b26b-80934e688269-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "666047b1-bc37-4f34-b26b-80934e688269" (UID: "666047b1-bc37-4f34-b26b-80934e688269"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 07:58:37 crc kubenswrapper[4846]: I1005 07:58:37.942828 4846 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/666047b1-bc37-4f34-b26b-80934e688269-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 05 07:58:38 crc kubenswrapper[4846]: I1005 07:58:38.152905 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-7jnwm"] Oct 05 07:58:38 crc kubenswrapper[4846]: I1005 07:58:38.163405 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-7jnwm"] Oct 05 07:58:38 crc kubenswrapper[4846]: I1005 07:58:38.515668 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="666047b1-bc37-4f34-b26b-80934e688269" path="/var/lib/kubelet/pods/666047b1-bc37-4f34-b26b-80934e688269/volumes" Oct 05 07:58:53 crc kubenswrapper[4846]: I1005 07:58:53.325319 4846 patch_prober.go:28] interesting pod/machine-config-daemon-fscvf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 07:58:53 crc kubenswrapper[4846]: I1005 07:58:53.325965 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 07:59:02 crc kubenswrapper[4846]: I1005 07:59:02.192071 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-jhglm"] Oct 05 07:59:02 crc kubenswrapper[4846]: E1005 07:59:02.192881 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="666047b1-bc37-4f34-b26b-80934e688269" containerName="extract-utilities" Oct 05 07:59:02 crc kubenswrapper[4846]: I1005 07:59:02.192902 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="666047b1-bc37-4f34-b26b-80934e688269" containerName="extract-utilities" Oct 05 07:59:02 crc kubenswrapper[4846]: E1005 07:59:02.192924 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="666047b1-bc37-4f34-b26b-80934e688269" containerName="extract-content" Oct 05 07:59:02 crc kubenswrapper[4846]: I1005 07:59:02.192935 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="666047b1-bc37-4f34-b26b-80934e688269" containerName="extract-content" Oct 05 07:59:02 crc kubenswrapper[4846]: E1005 07:59:02.192974 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="666047b1-bc37-4f34-b26b-80934e688269" containerName="registry-server" Oct 05 07:59:02 crc kubenswrapper[4846]: I1005 07:59:02.192985 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="666047b1-bc37-4f34-b26b-80934e688269" containerName="registry-server" Oct 05 07:59:02 crc kubenswrapper[4846]: I1005 07:59:02.193265 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="666047b1-bc37-4f34-b26b-80934e688269" containerName="registry-server" Oct 05 07:59:02 crc kubenswrapper[4846]: I1005 07:59:02.194863 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jhglm" Oct 05 07:59:02 crc kubenswrapper[4846]: I1005 07:59:02.209665 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-jhglm"] Oct 05 07:59:02 crc kubenswrapper[4846]: I1005 07:59:02.342812 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5bc4e54a-e28d-4cdd-8acf-c9a2238b3aeb-catalog-content\") pod \"community-operators-jhglm\" (UID: \"5bc4e54a-e28d-4cdd-8acf-c9a2238b3aeb\") " pod="openshift-marketplace/community-operators-jhglm" Oct 05 07:59:02 crc kubenswrapper[4846]: I1005 07:59:02.342895 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9g8lv\" (UniqueName: \"kubernetes.io/projected/5bc4e54a-e28d-4cdd-8acf-c9a2238b3aeb-kube-api-access-9g8lv\") pod \"community-operators-jhglm\" (UID: \"5bc4e54a-e28d-4cdd-8acf-c9a2238b3aeb\") " pod="openshift-marketplace/community-operators-jhglm" Oct 05 07:59:02 crc kubenswrapper[4846]: I1005 07:59:02.342945 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5bc4e54a-e28d-4cdd-8acf-c9a2238b3aeb-utilities\") pod \"community-operators-jhglm\" (UID: \"5bc4e54a-e28d-4cdd-8acf-c9a2238b3aeb\") " pod="openshift-marketplace/community-operators-jhglm" Oct 05 07:59:02 crc kubenswrapper[4846]: I1005 07:59:02.444551 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5bc4e54a-e28d-4cdd-8acf-c9a2238b3aeb-catalog-content\") pod \"community-operators-jhglm\" (UID: \"5bc4e54a-e28d-4cdd-8acf-c9a2238b3aeb\") " pod="openshift-marketplace/community-operators-jhglm" Oct 05 07:59:02 crc kubenswrapper[4846]: I1005 07:59:02.444631 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9g8lv\" (UniqueName: \"kubernetes.io/projected/5bc4e54a-e28d-4cdd-8acf-c9a2238b3aeb-kube-api-access-9g8lv\") pod \"community-operators-jhglm\" (UID: \"5bc4e54a-e28d-4cdd-8acf-c9a2238b3aeb\") " pod="openshift-marketplace/community-operators-jhglm" Oct 05 07:59:02 crc kubenswrapper[4846]: I1005 07:59:02.444675 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5bc4e54a-e28d-4cdd-8acf-c9a2238b3aeb-utilities\") pod \"community-operators-jhglm\" (UID: \"5bc4e54a-e28d-4cdd-8acf-c9a2238b3aeb\") " pod="openshift-marketplace/community-operators-jhglm" Oct 05 07:59:02 crc kubenswrapper[4846]: I1005 07:59:02.445553 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5bc4e54a-e28d-4cdd-8acf-c9a2238b3aeb-utilities\") pod \"community-operators-jhglm\" (UID: \"5bc4e54a-e28d-4cdd-8acf-c9a2238b3aeb\") " pod="openshift-marketplace/community-operators-jhglm" Oct 05 07:59:02 crc kubenswrapper[4846]: I1005 07:59:02.445589 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5bc4e54a-e28d-4cdd-8acf-c9a2238b3aeb-catalog-content\") pod \"community-operators-jhglm\" (UID: \"5bc4e54a-e28d-4cdd-8acf-c9a2238b3aeb\") " pod="openshift-marketplace/community-operators-jhglm" Oct 05 07:59:02 crc kubenswrapper[4846]: I1005 07:59:02.477483 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9g8lv\" (UniqueName: \"kubernetes.io/projected/5bc4e54a-e28d-4cdd-8acf-c9a2238b3aeb-kube-api-access-9g8lv\") pod \"community-operators-jhglm\" (UID: \"5bc4e54a-e28d-4cdd-8acf-c9a2238b3aeb\") " pod="openshift-marketplace/community-operators-jhglm" Oct 05 07:59:02 crc kubenswrapper[4846]: I1005 07:59:02.551858 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jhglm" Oct 05 07:59:02 crc kubenswrapper[4846]: I1005 07:59:02.808504 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-jhglm"] Oct 05 07:59:03 crc kubenswrapper[4846]: I1005 07:59:03.451787 4846 generic.go:334] "Generic (PLEG): container finished" podID="5bc4e54a-e28d-4cdd-8acf-c9a2238b3aeb" containerID="1b1edc9615e724a347abfae2508a81f5c057c39449aad40e4069893472c3b25f" exitCode=0 Oct 05 07:59:03 crc kubenswrapper[4846]: I1005 07:59:03.451889 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jhglm" event={"ID":"5bc4e54a-e28d-4cdd-8acf-c9a2238b3aeb","Type":"ContainerDied","Data":"1b1edc9615e724a347abfae2508a81f5c057c39449aad40e4069893472c3b25f"} Oct 05 07:59:03 crc kubenswrapper[4846]: I1005 07:59:03.452207 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jhglm" event={"ID":"5bc4e54a-e28d-4cdd-8acf-c9a2238b3aeb","Type":"ContainerStarted","Data":"23b247f41cf37c9eb571a033ffba0c5f446fc40673fc8cfb21465d32fc4813f3"} Oct 05 07:59:04 crc kubenswrapper[4846]: I1005 07:59:04.465040 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jhglm" event={"ID":"5bc4e54a-e28d-4cdd-8acf-c9a2238b3aeb","Type":"ContainerStarted","Data":"e7b3ae473b39d5eaa13dde68a9eefac0508171a21d552bec05a8df6785de5acf"} Oct 05 07:59:05 crc kubenswrapper[4846]: I1005 07:59:05.477119 4846 generic.go:334] "Generic (PLEG): container finished" podID="5bc4e54a-e28d-4cdd-8acf-c9a2238b3aeb" containerID="e7b3ae473b39d5eaa13dde68a9eefac0508171a21d552bec05a8df6785de5acf" exitCode=0 Oct 05 07:59:05 crc kubenswrapper[4846]: I1005 07:59:05.477245 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jhglm" event={"ID":"5bc4e54a-e28d-4cdd-8acf-c9a2238b3aeb","Type":"ContainerDied","Data":"e7b3ae473b39d5eaa13dde68a9eefac0508171a21d552bec05a8df6785de5acf"} Oct 05 07:59:06 crc kubenswrapper[4846]: I1005 07:59:06.490630 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jhglm" event={"ID":"5bc4e54a-e28d-4cdd-8acf-c9a2238b3aeb","Type":"ContainerStarted","Data":"ae8acfeb158b5e8f59bd7457a0729bc1e2465ecd90e0a4fe17493cb64c4552e1"} Oct 05 07:59:06 crc kubenswrapper[4846]: I1005 07:59:06.519705 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-jhglm" podStartSLOduration=2.106810369 podStartE2EDuration="4.519686305s" podCreationTimestamp="2025-10-05 07:59:02 +0000 UTC" firstStartedPulling="2025-10-05 07:59:03.455263546 +0000 UTC m=+4265.696116351" lastFinishedPulling="2025-10-05 07:59:05.868139492 +0000 UTC m=+4268.108992287" observedRunningTime="2025-10-05 07:59:06.514670041 +0000 UTC m=+4268.755522826" watchObservedRunningTime="2025-10-05 07:59:06.519686305 +0000 UTC m=+4268.760539090" Oct 05 07:59:12 crc kubenswrapper[4846]: I1005 07:59:12.552279 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-jhglm" Oct 05 07:59:12 crc kubenswrapper[4846]: I1005 07:59:12.553220 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-jhglm" Oct 05 07:59:12 crc kubenswrapper[4846]: I1005 07:59:12.627747 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-jhglm" Oct 05 07:59:13 crc kubenswrapper[4846]: I1005 07:59:13.624270 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-jhglm" Oct 05 07:59:13 crc kubenswrapper[4846]: I1005 07:59:13.696662 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-jhglm"] Oct 05 07:59:15 crc kubenswrapper[4846]: I1005 07:59:15.581427 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-jhglm" podUID="5bc4e54a-e28d-4cdd-8acf-c9a2238b3aeb" containerName="registry-server" containerID="cri-o://ae8acfeb158b5e8f59bd7457a0729bc1e2465ecd90e0a4fe17493cb64c4552e1" gracePeriod=2 Oct 05 07:59:16 crc kubenswrapper[4846]: I1005 07:59:16.057870 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jhglm" Oct 05 07:59:16 crc kubenswrapper[4846]: I1005 07:59:16.167934 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5bc4e54a-e28d-4cdd-8acf-c9a2238b3aeb-catalog-content\") pod \"5bc4e54a-e28d-4cdd-8acf-c9a2238b3aeb\" (UID: \"5bc4e54a-e28d-4cdd-8acf-c9a2238b3aeb\") " Oct 05 07:59:16 crc kubenswrapper[4846]: I1005 07:59:16.168237 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5bc4e54a-e28d-4cdd-8acf-c9a2238b3aeb-utilities\") pod \"5bc4e54a-e28d-4cdd-8acf-c9a2238b3aeb\" (UID: \"5bc4e54a-e28d-4cdd-8acf-c9a2238b3aeb\") " Oct 05 07:59:16 crc kubenswrapper[4846]: I1005 07:59:16.168347 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9g8lv\" (UniqueName: \"kubernetes.io/projected/5bc4e54a-e28d-4cdd-8acf-c9a2238b3aeb-kube-api-access-9g8lv\") pod \"5bc4e54a-e28d-4cdd-8acf-c9a2238b3aeb\" (UID: \"5bc4e54a-e28d-4cdd-8acf-c9a2238b3aeb\") " Oct 05 07:59:16 crc kubenswrapper[4846]: I1005 07:59:16.169995 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5bc4e54a-e28d-4cdd-8acf-c9a2238b3aeb-utilities" (OuterVolumeSpecName: "utilities") pod "5bc4e54a-e28d-4cdd-8acf-c9a2238b3aeb" (UID: "5bc4e54a-e28d-4cdd-8acf-c9a2238b3aeb"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 07:59:16 crc kubenswrapper[4846]: I1005 07:59:16.173677 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5bc4e54a-e28d-4cdd-8acf-c9a2238b3aeb-kube-api-access-9g8lv" (OuterVolumeSpecName: "kube-api-access-9g8lv") pod "5bc4e54a-e28d-4cdd-8acf-c9a2238b3aeb" (UID: "5bc4e54a-e28d-4cdd-8acf-c9a2238b3aeb"). InnerVolumeSpecName "kube-api-access-9g8lv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 07:59:16 crc kubenswrapper[4846]: I1005 07:59:16.238856 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5bc4e54a-e28d-4cdd-8acf-c9a2238b3aeb-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5bc4e54a-e28d-4cdd-8acf-c9a2238b3aeb" (UID: "5bc4e54a-e28d-4cdd-8acf-c9a2238b3aeb"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 07:59:16 crc kubenswrapper[4846]: I1005 07:59:16.270389 4846 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5bc4e54a-e28d-4cdd-8acf-c9a2238b3aeb-utilities\") on node \"crc\" DevicePath \"\"" Oct 05 07:59:16 crc kubenswrapper[4846]: I1005 07:59:16.270444 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9g8lv\" (UniqueName: \"kubernetes.io/projected/5bc4e54a-e28d-4cdd-8acf-c9a2238b3aeb-kube-api-access-9g8lv\") on node \"crc\" DevicePath \"\"" Oct 05 07:59:16 crc kubenswrapper[4846]: I1005 07:59:16.270458 4846 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5bc4e54a-e28d-4cdd-8acf-c9a2238b3aeb-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 05 07:59:16 crc kubenswrapper[4846]: I1005 07:59:16.593837 4846 generic.go:334] "Generic (PLEG): container finished" podID="5bc4e54a-e28d-4cdd-8acf-c9a2238b3aeb" containerID="ae8acfeb158b5e8f59bd7457a0729bc1e2465ecd90e0a4fe17493cb64c4552e1" exitCode=0 Oct 05 07:59:16 crc kubenswrapper[4846]: I1005 07:59:16.593898 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jhglm" event={"ID":"5bc4e54a-e28d-4cdd-8acf-c9a2238b3aeb","Type":"ContainerDied","Data":"ae8acfeb158b5e8f59bd7457a0729bc1e2465ecd90e0a4fe17493cb64c4552e1"} Oct 05 07:59:16 crc kubenswrapper[4846]: I1005 07:59:16.593938 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jhglm" event={"ID":"5bc4e54a-e28d-4cdd-8acf-c9a2238b3aeb","Type":"ContainerDied","Data":"23b247f41cf37c9eb571a033ffba0c5f446fc40673fc8cfb21465d32fc4813f3"} Oct 05 07:59:16 crc kubenswrapper[4846]: I1005 07:59:16.593971 4846 scope.go:117] "RemoveContainer" containerID="ae8acfeb158b5e8f59bd7457a0729bc1e2465ecd90e0a4fe17493cb64c4552e1" Oct 05 07:59:16 crc kubenswrapper[4846]: I1005 07:59:16.594226 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jhglm" Oct 05 07:59:16 crc kubenswrapper[4846]: I1005 07:59:16.636837 4846 scope.go:117] "RemoveContainer" containerID="e7b3ae473b39d5eaa13dde68a9eefac0508171a21d552bec05a8df6785de5acf" Oct 05 07:59:16 crc kubenswrapper[4846]: I1005 07:59:16.644726 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-jhglm"] Oct 05 07:59:16 crc kubenswrapper[4846]: I1005 07:59:16.656927 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-jhglm"] Oct 05 07:59:16 crc kubenswrapper[4846]: I1005 07:59:16.675605 4846 scope.go:117] "RemoveContainer" containerID="1b1edc9615e724a347abfae2508a81f5c057c39449aad40e4069893472c3b25f" Oct 05 07:59:16 crc kubenswrapper[4846]: I1005 07:59:16.705629 4846 scope.go:117] "RemoveContainer" containerID="ae8acfeb158b5e8f59bd7457a0729bc1e2465ecd90e0a4fe17493cb64c4552e1" Oct 05 07:59:16 crc kubenswrapper[4846]: E1005 07:59:16.707990 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ae8acfeb158b5e8f59bd7457a0729bc1e2465ecd90e0a4fe17493cb64c4552e1\": container with ID starting with ae8acfeb158b5e8f59bd7457a0729bc1e2465ecd90e0a4fe17493cb64c4552e1 not found: ID does not exist" containerID="ae8acfeb158b5e8f59bd7457a0729bc1e2465ecd90e0a4fe17493cb64c4552e1" Oct 05 07:59:16 crc kubenswrapper[4846]: I1005 07:59:16.708030 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ae8acfeb158b5e8f59bd7457a0729bc1e2465ecd90e0a4fe17493cb64c4552e1"} err="failed to get container status \"ae8acfeb158b5e8f59bd7457a0729bc1e2465ecd90e0a4fe17493cb64c4552e1\": rpc error: code = NotFound desc = could not find container \"ae8acfeb158b5e8f59bd7457a0729bc1e2465ecd90e0a4fe17493cb64c4552e1\": container with ID starting with ae8acfeb158b5e8f59bd7457a0729bc1e2465ecd90e0a4fe17493cb64c4552e1 not found: ID does not exist" Oct 05 07:59:16 crc kubenswrapper[4846]: I1005 07:59:16.708057 4846 scope.go:117] "RemoveContainer" containerID="e7b3ae473b39d5eaa13dde68a9eefac0508171a21d552bec05a8df6785de5acf" Oct 05 07:59:16 crc kubenswrapper[4846]: E1005 07:59:16.708938 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e7b3ae473b39d5eaa13dde68a9eefac0508171a21d552bec05a8df6785de5acf\": container with ID starting with e7b3ae473b39d5eaa13dde68a9eefac0508171a21d552bec05a8df6785de5acf not found: ID does not exist" containerID="e7b3ae473b39d5eaa13dde68a9eefac0508171a21d552bec05a8df6785de5acf" Oct 05 07:59:16 crc kubenswrapper[4846]: I1005 07:59:16.709004 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e7b3ae473b39d5eaa13dde68a9eefac0508171a21d552bec05a8df6785de5acf"} err="failed to get container status \"e7b3ae473b39d5eaa13dde68a9eefac0508171a21d552bec05a8df6785de5acf\": rpc error: code = NotFound desc = could not find container \"e7b3ae473b39d5eaa13dde68a9eefac0508171a21d552bec05a8df6785de5acf\": container with ID starting with e7b3ae473b39d5eaa13dde68a9eefac0508171a21d552bec05a8df6785de5acf not found: ID does not exist" Oct 05 07:59:16 crc kubenswrapper[4846]: I1005 07:59:16.709049 4846 scope.go:117] "RemoveContainer" containerID="1b1edc9615e724a347abfae2508a81f5c057c39449aad40e4069893472c3b25f" Oct 05 07:59:16 crc kubenswrapper[4846]: E1005 07:59:16.709575 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1b1edc9615e724a347abfae2508a81f5c057c39449aad40e4069893472c3b25f\": container with ID starting with 1b1edc9615e724a347abfae2508a81f5c057c39449aad40e4069893472c3b25f not found: ID does not exist" containerID="1b1edc9615e724a347abfae2508a81f5c057c39449aad40e4069893472c3b25f" Oct 05 07:59:16 crc kubenswrapper[4846]: I1005 07:59:16.709615 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1b1edc9615e724a347abfae2508a81f5c057c39449aad40e4069893472c3b25f"} err="failed to get container status \"1b1edc9615e724a347abfae2508a81f5c057c39449aad40e4069893472c3b25f\": rpc error: code = NotFound desc = could not find container \"1b1edc9615e724a347abfae2508a81f5c057c39449aad40e4069893472c3b25f\": container with ID starting with 1b1edc9615e724a347abfae2508a81f5c057c39449aad40e4069893472c3b25f not found: ID does not exist" Oct 05 07:59:18 crc kubenswrapper[4846]: I1005 07:59:18.512470 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5bc4e54a-e28d-4cdd-8acf-c9a2238b3aeb" path="/var/lib/kubelet/pods/5bc4e54a-e28d-4cdd-8acf-c9a2238b3aeb/volumes" Oct 05 07:59:23 crc kubenswrapper[4846]: I1005 07:59:23.325874 4846 patch_prober.go:28] interesting pod/machine-config-daemon-fscvf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 07:59:23 crc kubenswrapper[4846]: I1005 07:59:23.326386 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 07:59:23 crc kubenswrapper[4846]: I1005 07:59:23.326479 4846 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" Oct 05 07:59:23 crc kubenswrapper[4846]: I1005 07:59:23.327692 4846 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"2d2e59cb51a3cd322bc7aca627ddfa4d48c6f5f95b26a7264c472515babd8d10"} pod="openshift-machine-config-operator/machine-config-daemon-fscvf" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 05 07:59:23 crc kubenswrapper[4846]: I1005 07:59:23.327806 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" containerName="machine-config-daemon" containerID="cri-o://2d2e59cb51a3cd322bc7aca627ddfa4d48c6f5f95b26a7264c472515babd8d10" gracePeriod=600 Oct 05 07:59:23 crc kubenswrapper[4846]: E1005 07:59:23.459623 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 07:59:23 crc kubenswrapper[4846]: I1005 07:59:23.660163 4846 generic.go:334] "Generic (PLEG): container finished" podID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" containerID="2d2e59cb51a3cd322bc7aca627ddfa4d48c6f5f95b26a7264c472515babd8d10" exitCode=0 Oct 05 07:59:23 crc kubenswrapper[4846]: I1005 07:59:23.660206 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" event={"ID":"ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7","Type":"ContainerDied","Data":"2d2e59cb51a3cd322bc7aca627ddfa4d48c6f5f95b26a7264c472515babd8d10"} Oct 05 07:59:23 crc kubenswrapper[4846]: I1005 07:59:23.660261 4846 scope.go:117] "RemoveContainer" containerID="dccf837d14b883290775f3ba1aae37388eb10cbc64bfe9ddcfac744876bcccff" Oct 05 07:59:23 crc kubenswrapper[4846]: I1005 07:59:23.660930 4846 scope.go:117] "RemoveContainer" containerID="2d2e59cb51a3cd322bc7aca627ddfa4d48c6f5f95b26a7264c472515babd8d10" Oct 05 07:59:23 crc kubenswrapper[4846]: E1005 07:59:23.661483 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 07:59:37 crc kubenswrapper[4846]: I1005 07:59:37.498222 4846 scope.go:117] "RemoveContainer" containerID="2d2e59cb51a3cd322bc7aca627ddfa4d48c6f5f95b26a7264c472515babd8d10" Oct 05 07:59:37 crc kubenswrapper[4846]: E1005 07:59:37.500594 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 07:59:50 crc kubenswrapper[4846]: I1005 07:59:50.498560 4846 scope.go:117] "RemoveContainer" containerID="2d2e59cb51a3cd322bc7aca627ddfa4d48c6f5f95b26a7264c472515babd8d10" Oct 05 07:59:50 crc kubenswrapper[4846]: E1005 07:59:50.499804 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 08:00:00 crc kubenswrapper[4846]: I1005 08:00:00.163946 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29327520-dcpnc"] Oct 05 08:00:00 crc kubenswrapper[4846]: E1005 08:00:00.165354 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5bc4e54a-e28d-4cdd-8acf-c9a2238b3aeb" containerName="extract-content" Oct 05 08:00:00 crc kubenswrapper[4846]: I1005 08:00:00.165388 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="5bc4e54a-e28d-4cdd-8acf-c9a2238b3aeb" containerName="extract-content" Oct 05 08:00:00 crc kubenswrapper[4846]: E1005 08:00:00.165418 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5bc4e54a-e28d-4cdd-8acf-c9a2238b3aeb" containerName="extract-utilities" Oct 05 08:00:00 crc kubenswrapper[4846]: I1005 08:00:00.165438 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="5bc4e54a-e28d-4cdd-8acf-c9a2238b3aeb" containerName="extract-utilities" Oct 05 08:00:00 crc kubenswrapper[4846]: E1005 08:00:00.165469 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5bc4e54a-e28d-4cdd-8acf-c9a2238b3aeb" containerName="registry-server" Oct 05 08:00:00 crc kubenswrapper[4846]: I1005 08:00:00.165488 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="5bc4e54a-e28d-4cdd-8acf-c9a2238b3aeb" containerName="registry-server" Oct 05 08:00:00 crc kubenswrapper[4846]: I1005 08:00:00.165902 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="5bc4e54a-e28d-4cdd-8acf-c9a2238b3aeb" containerName="registry-server" Oct 05 08:00:00 crc kubenswrapper[4846]: I1005 08:00:00.167027 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29327520-dcpnc" Oct 05 08:00:00 crc kubenswrapper[4846]: I1005 08:00:00.184902 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 05 08:00:00 crc kubenswrapper[4846]: I1005 08:00:00.188025 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29327520-dcpnc"] Oct 05 08:00:00 crc kubenswrapper[4846]: I1005 08:00:00.193493 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 05 08:00:00 crc kubenswrapper[4846]: I1005 08:00:00.286138 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4tpx2\" (UniqueName: \"kubernetes.io/projected/4dc64c01-83c3-4468-9df8-4d0b6562d26a-kube-api-access-4tpx2\") pod \"collect-profiles-29327520-dcpnc\" (UID: \"4dc64c01-83c3-4468-9df8-4d0b6562d26a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327520-dcpnc" Oct 05 08:00:00 crc kubenswrapper[4846]: I1005 08:00:00.286319 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4dc64c01-83c3-4468-9df8-4d0b6562d26a-secret-volume\") pod \"collect-profiles-29327520-dcpnc\" (UID: \"4dc64c01-83c3-4468-9df8-4d0b6562d26a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327520-dcpnc" Oct 05 08:00:00 crc kubenswrapper[4846]: I1005 08:00:00.286440 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4dc64c01-83c3-4468-9df8-4d0b6562d26a-config-volume\") pod \"collect-profiles-29327520-dcpnc\" (UID: \"4dc64c01-83c3-4468-9df8-4d0b6562d26a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327520-dcpnc" Oct 05 08:00:00 crc kubenswrapper[4846]: I1005 08:00:00.388435 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4tpx2\" (UniqueName: \"kubernetes.io/projected/4dc64c01-83c3-4468-9df8-4d0b6562d26a-kube-api-access-4tpx2\") pod \"collect-profiles-29327520-dcpnc\" (UID: \"4dc64c01-83c3-4468-9df8-4d0b6562d26a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327520-dcpnc" Oct 05 08:00:00 crc kubenswrapper[4846]: I1005 08:00:00.388540 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4dc64c01-83c3-4468-9df8-4d0b6562d26a-secret-volume\") pod \"collect-profiles-29327520-dcpnc\" (UID: \"4dc64c01-83c3-4468-9df8-4d0b6562d26a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327520-dcpnc" Oct 05 08:00:00 crc kubenswrapper[4846]: I1005 08:00:00.388648 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4dc64c01-83c3-4468-9df8-4d0b6562d26a-config-volume\") pod \"collect-profiles-29327520-dcpnc\" (UID: \"4dc64c01-83c3-4468-9df8-4d0b6562d26a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327520-dcpnc" Oct 05 08:00:00 crc kubenswrapper[4846]: I1005 08:00:00.390412 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4dc64c01-83c3-4468-9df8-4d0b6562d26a-config-volume\") pod \"collect-profiles-29327520-dcpnc\" (UID: \"4dc64c01-83c3-4468-9df8-4d0b6562d26a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327520-dcpnc" Oct 05 08:00:00 crc kubenswrapper[4846]: I1005 08:00:00.395827 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4dc64c01-83c3-4468-9df8-4d0b6562d26a-secret-volume\") pod \"collect-profiles-29327520-dcpnc\" (UID: \"4dc64c01-83c3-4468-9df8-4d0b6562d26a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327520-dcpnc" Oct 05 08:00:00 crc kubenswrapper[4846]: I1005 08:00:00.413847 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4tpx2\" (UniqueName: \"kubernetes.io/projected/4dc64c01-83c3-4468-9df8-4d0b6562d26a-kube-api-access-4tpx2\") pod \"collect-profiles-29327520-dcpnc\" (UID: \"4dc64c01-83c3-4468-9df8-4d0b6562d26a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327520-dcpnc" Oct 05 08:00:00 crc kubenswrapper[4846]: I1005 08:00:00.510621 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29327520-dcpnc" Oct 05 08:00:00 crc kubenswrapper[4846]: I1005 08:00:00.957486 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29327520-dcpnc"] Oct 05 08:00:01 crc kubenswrapper[4846]: I1005 08:00:01.000161 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29327520-dcpnc" event={"ID":"4dc64c01-83c3-4468-9df8-4d0b6562d26a","Type":"ContainerStarted","Data":"d06c178922f9e7b087663095a8f9fb70e5e26f01edd9072f2316d55d91ced9b3"} Oct 05 08:00:02 crc kubenswrapper[4846]: I1005 08:00:02.009946 4846 generic.go:334] "Generic (PLEG): container finished" podID="4dc64c01-83c3-4468-9df8-4d0b6562d26a" containerID="ffde80bfd45c9c3a6ca4b4c833fddc3220ff829cea27ad3ca24a63a713789e68" exitCode=0 Oct 05 08:00:02 crc kubenswrapper[4846]: I1005 08:00:02.010018 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29327520-dcpnc" event={"ID":"4dc64c01-83c3-4468-9df8-4d0b6562d26a","Type":"ContainerDied","Data":"ffde80bfd45c9c3a6ca4b4c833fddc3220ff829cea27ad3ca24a63a713789e68"} Oct 05 08:00:03 crc kubenswrapper[4846]: I1005 08:00:03.324282 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29327520-dcpnc" Oct 05 08:00:03 crc kubenswrapper[4846]: I1005 08:00:03.429450 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4tpx2\" (UniqueName: \"kubernetes.io/projected/4dc64c01-83c3-4468-9df8-4d0b6562d26a-kube-api-access-4tpx2\") pod \"4dc64c01-83c3-4468-9df8-4d0b6562d26a\" (UID: \"4dc64c01-83c3-4468-9df8-4d0b6562d26a\") " Oct 05 08:00:03 crc kubenswrapper[4846]: I1005 08:00:03.429524 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4dc64c01-83c3-4468-9df8-4d0b6562d26a-config-volume\") pod \"4dc64c01-83c3-4468-9df8-4d0b6562d26a\" (UID: \"4dc64c01-83c3-4468-9df8-4d0b6562d26a\") " Oct 05 08:00:03 crc kubenswrapper[4846]: I1005 08:00:03.429638 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4dc64c01-83c3-4468-9df8-4d0b6562d26a-secret-volume\") pod \"4dc64c01-83c3-4468-9df8-4d0b6562d26a\" (UID: \"4dc64c01-83c3-4468-9df8-4d0b6562d26a\") " Oct 05 08:00:03 crc kubenswrapper[4846]: I1005 08:00:03.430369 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4dc64c01-83c3-4468-9df8-4d0b6562d26a-config-volume" (OuterVolumeSpecName: "config-volume") pod "4dc64c01-83c3-4468-9df8-4d0b6562d26a" (UID: "4dc64c01-83c3-4468-9df8-4d0b6562d26a"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 08:00:03 crc kubenswrapper[4846]: I1005 08:00:03.434532 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4dc64c01-83c3-4468-9df8-4d0b6562d26a-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "4dc64c01-83c3-4468-9df8-4d0b6562d26a" (UID: "4dc64c01-83c3-4468-9df8-4d0b6562d26a"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:00:03 crc kubenswrapper[4846]: I1005 08:00:03.434610 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4dc64c01-83c3-4468-9df8-4d0b6562d26a-kube-api-access-4tpx2" (OuterVolumeSpecName: "kube-api-access-4tpx2") pod "4dc64c01-83c3-4468-9df8-4d0b6562d26a" (UID: "4dc64c01-83c3-4468-9df8-4d0b6562d26a"). InnerVolumeSpecName "kube-api-access-4tpx2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:00:03 crc kubenswrapper[4846]: I1005 08:00:03.498384 4846 scope.go:117] "RemoveContainer" containerID="2d2e59cb51a3cd322bc7aca627ddfa4d48c6f5f95b26a7264c472515babd8d10" Oct 05 08:00:03 crc kubenswrapper[4846]: E1005 08:00:03.498634 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 08:00:03 crc kubenswrapper[4846]: I1005 08:00:03.531493 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4tpx2\" (UniqueName: \"kubernetes.io/projected/4dc64c01-83c3-4468-9df8-4d0b6562d26a-kube-api-access-4tpx2\") on node \"crc\" DevicePath \"\"" Oct 05 08:00:03 crc kubenswrapper[4846]: I1005 08:00:03.531541 4846 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4dc64c01-83c3-4468-9df8-4d0b6562d26a-config-volume\") on node \"crc\" DevicePath \"\"" Oct 05 08:00:03 crc kubenswrapper[4846]: I1005 08:00:03.531560 4846 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4dc64c01-83c3-4468-9df8-4d0b6562d26a-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 05 08:00:04 crc kubenswrapper[4846]: I1005 08:00:04.028219 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29327520-dcpnc" event={"ID":"4dc64c01-83c3-4468-9df8-4d0b6562d26a","Type":"ContainerDied","Data":"d06c178922f9e7b087663095a8f9fb70e5e26f01edd9072f2316d55d91ced9b3"} Oct 05 08:00:04 crc kubenswrapper[4846]: I1005 08:00:04.028560 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d06c178922f9e7b087663095a8f9fb70e5e26f01edd9072f2316d55d91ced9b3" Oct 05 08:00:04 crc kubenswrapper[4846]: I1005 08:00:04.028334 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29327520-dcpnc" Oct 05 08:00:04 crc kubenswrapper[4846]: I1005 08:00:04.430607 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29327475-frl5l"] Oct 05 08:00:04 crc kubenswrapper[4846]: I1005 08:00:04.442941 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29327475-frl5l"] Oct 05 08:00:04 crc kubenswrapper[4846]: I1005 08:00:04.531078 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6269daa-7302-4bf7-ae86-9d57935687ea" path="/var/lib/kubelet/pods/b6269daa-7302-4bf7-ae86-9d57935687ea/volumes" Oct 05 08:00:18 crc kubenswrapper[4846]: I1005 08:00:18.505591 4846 scope.go:117] "RemoveContainer" containerID="2d2e59cb51a3cd322bc7aca627ddfa4d48c6f5f95b26a7264c472515babd8d10" Oct 05 08:00:18 crc kubenswrapper[4846]: E1005 08:00:18.506609 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 08:00:29 crc kubenswrapper[4846]: I1005 08:00:29.498957 4846 scope.go:117] "RemoveContainer" containerID="2d2e59cb51a3cd322bc7aca627ddfa4d48c6f5f95b26a7264c472515babd8d10" Oct 05 08:00:29 crc kubenswrapper[4846]: E1005 08:00:29.500474 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 08:00:29 crc kubenswrapper[4846]: I1005 08:00:29.933532 4846 scope.go:117] "RemoveContainer" containerID="2581fecf028f7d3f4fef7ce6871f47512432aa79d692b1f91f468755e769e6e8" Oct 05 08:00:42 crc kubenswrapper[4846]: I1005 08:00:42.497973 4846 scope.go:117] "RemoveContainer" containerID="2d2e59cb51a3cd322bc7aca627ddfa4d48c6f5f95b26a7264c472515babd8d10" Oct 05 08:00:42 crc kubenswrapper[4846]: E1005 08:00:42.499082 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 08:00:57 crc kubenswrapper[4846]: I1005 08:00:57.497971 4846 scope.go:117] "RemoveContainer" containerID="2d2e59cb51a3cd322bc7aca627ddfa4d48c6f5f95b26a7264c472515babd8d10" Oct 05 08:00:57 crc kubenswrapper[4846]: E1005 08:00:57.499149 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 08:01:08 crc kubenswrapper[4846]: I1005 08:01:08.509849 4846 scope.go:117] "RemoveContainer" containerID="2d2e59cb51a3cd322bc7aca627ddfa4d48c6f5f95b26a7264c472515babd8d10" Oct 05 08:01:08 crc kubenswrapper[4846]: E1005 08:01:08.511289 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 08:01:21 crc kubenswrapper[4846]: I1005 08:01:21.498699 4846 scope.go:117] "RemoveContainer" containerID="2d2e59cb51a3cd322bc7aca627ddfa4d48c6f5f95b26a7264c472515babd8d10" Oct 05 08:01:21 crc kubenswrapper[4846]: E1005 08:01:21.499453 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 08:01:35 crc kubenswrapper[4846]: I1005 08:01:35.498358 4846 scope.go:117] "RemoveContainer" containerID="2d2e59cb51a3cd322bc7aca627ddfa4d48c6f5f95b26a7264c472515babd8d10" Oct 05 08:01:35 crc kubenswrapper[4846]: E1005 08:01:35.499759 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 08:01:46 crc kubenswrapper[4846]: I1005 08:01:46.498079 4846 scope.go:117] "RemoveContainer" containerID="2d2e59cb51a3cd322bc7aca627ddfa4d48c6f5f95b26a7264c472515babd8d10" Oct 05 08:01:46 crc kubenswrapper[4846]: E1005 08:01:46.498902 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 08:01:57 crc kubenswrapper[4846]: I1005 08:01:57.498263 4846 scope.go:117] "RemoveContainer" containerID="2d2e59cb51a3cd322bc7aca627ddfa4d48c6f5f95b26a7264c472515babd8d10" Oct 05 08:01:57 crc kubenswrapper[4846]: E1005 08:01:57.499162 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 08:02:09 crc kubenswrapper[4846]: I1005 08:02:09.499027 4846 scope.go:117] "RemoveContainer" containerID="2d2e59cb51a3cd322bc7aca627ddfa4d48c6f5f95b26a7264c472515babd8d10" Oct 05 08:02:09 crc kubenswrapper[4846]: E1005 08:02:09.500033 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 08:02:23 crc kubenswrapper[4846]: I1005 08:02:23.497876 4846 scope.go:117] "RemoveContainer" containerID="2d2e59cb51a3cd322bc7aca627ddfa4d48c6f5f95b26a7264c472515babd8d10" Oct 05 08:02:23 crc kubenswrapper[4846]: E1005 08:02:23.499852 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 08:02:34 crc kubenswrapper[4846]: I1005 08:02:34.498429 4846 scope.go:117] "RemoveContainer" containerID="2d2e59cb51a3cd322bc7aca627ddfa4d48c6f5f95b26a7264c472515babd8d10" Oct 05 08:02:34 crc kubenswrapper[4846]: E1005 08:02:34.499311 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 08:02:49 crc kubenswrapper[4846]: I1005 08:02:49.497481 4846 scope.go:117] "RemoveContainer" containerID="2d2e59cb51a3cd322bc7aca627ddfa4d48c6f5f95b26a7264c472515babd8d10" Oct 05 08:02:49 crc kubenswrapper[4846]: E1005 08:02:49.498732 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 08:03:03 crc kubenswrapper[4846]: I1005 08:03:03.497625 4846 scope.go:117] "RemoveContainer" containerID="2d2e59cb51a3cd322bc7aca627ddfa4d48c6f5f95b26a7264c472515babd8d10" Oct 05 08:03:03 crc kubenswrapper[4846]: E1005 08:03:03.498306 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 08:03:14 crc kubenswrapper[4846]: I1005 08:03:14.497964 4846 scope.go:117] "RemoveContainer" containerID="2d2e59cb51a3cd322bc7aca627ddfa4d48c6f5f95b26a7264c472515babd8d10" Oct 05 08:03:14 crc kubenswrapper[4846]: E1005 08:03:14.498747 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 08:03:27 crc kubenswrapper[4846]: I1005 08:03:27.497593 4846 scope.go:117] "RemoveContainer" containerID="2d2e59cb51a3cd322bc7aca627ddfa4d48c6f5f95b26a7264c472515babd8d10" Oct 05 08:03:27 crc kubenswrapper[4846]: E1005 08:03:27.498450 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 08:03:39 crc kubenswrapper[4846]: I1005 08:03:39.188246 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["crc-storage/crc-storage-crc-dr8j2"] Oct 05 08:03:39 crc kubenswrapper[4846]: I1005 08:03:39.195107 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["crc-storage/crc-storage-crc-dr8j2"] Oct 05 08:03:39 crc kubenswrapper[4846]: I1005 08:03:39.308014 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["crc-storage/crc-storage-crc-r5trh"] Oct 05 08:03:39 crc kubenswrapper[4846]: E1005 08:03:39.308360 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4dc64c01-83c3-4468-9df8-4d0b6562d26a" containerName="collect-profiles" Oct 05 08:03:39 crc kubenswrapper[4846]: I1005 08:03:39.308372 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="4dc64c01-83c3-4468-9df8-4d0b6562d26a" containerName="collect-profiles" Oct 05 08:03:39 crc kubenswrapper[4846]: I1005 08:03:39.308623 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="4dc64c01-83c3-4468-9df8-4d0b6562d26a" containerName="collect-profiles" Oct 05 08:03:39 crc kubenswrapper[4846]: I1005 08:03:39.309162 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-r5trh" Oct 05 08:03:39 crc kubenswrapper[4846]: I1005 08:03:39.320778 4846 reflector.go:368] Caches populated for *v1.Secret from object-"crc-storage"/"crc-storage-dockercfg-w2v6h" Oct 05 08:03:39 crc kubenswrapper[4846]: I1005 08:03:39.320809 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"openshift-service-ca.crt" Oct 05 08:03:39 crc kubenswrapper[4846]: I1005 08:03:39.320826 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"kube-root-ca.crt" Oct 05 08:03:39 crc kubenswrapper[4846]: I1005 08:03:39.321060 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"crc-storage" Oct 05 08:03:39 crc kubenswrapper[4846]: I1005 08:03:39.324566 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["crc-storage/crc-storage-crc-r5trh"] Oct 05 08:03:39 crc kubenswrapper[4846]: I1005 08:03:39.409653 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/4df1be73-aa09-4c83-a580-6454e6df05dd-crc-storage\") pod \"crc-storage-crc-r5trh\" (UID: \"4df1be73-aa09-4c83-a580-6454e6df05dd\") " pod="crc-storage/crc-storage-crc-r5trh" Oct 05 08:03:39 crc kubenswrapper[4846]: I1005 08:03:39.409721 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/4df1be73-aa09-4c83-a580-6454e6df05dd-node-mnt\") pod \"crc-storage-crc-r5trh\" (UID: \"4df1be73-aa09-4c83-a580-6454e6df05dd\") " pod="crc-storage/crc-storage-crc-r5trh" Oct 05 08:03:39 crc kubenswrapper[4846]: I1005 08:03:39.410074 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6wpqt\" (UniqueName: \"kubernetes.io/projected/4df1be73-aa09-4c83-a580-6454e6df05dd-kube-api-access-6wpqt\") pod \"crc-storage-crc-r5trh\" (UID: \"4df1be73-aa09-4c83-a580-6454e6df05dd\") " pod="crc-storage/crc-storage-crc-r5trh" Oct 05 08:03:39 crc kubenswrapper[4846]: I1005 08:03:39.511780 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/4df1be73-aa09-4c83-a580-6454e6df05dd-crc-storage\") pod \"crc-storage-crc-r5trh\" (UID: \"4df1be73-aa09-4c83-a580-6454e6df05dd\") " pod="crc-storage/crc-storage-crc-r5trh" Oct 05 08:03:39 crc kubenswrapper[4846]: I1005 08:03:39.511837 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/4df1be73-aa09-4c83-a580-6454e6df05dd-node-mnt\") pod \"crc-storage-crc-r5trh\" (UID: \"4df1be73-aa09-4c83-a580-6454e6df05dd\") " pod="crc-storage/crc-storage-crc-r5trh" Oct 05 08:03:39 crc kubenswrapper[4846]: I1005 08:03:39.511939 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6wpqt\" (UniqueName: \"kubernetes.io/projected/4df1be73-aa09-4c83-a580-6454e6df05dd-kube-api-access-6wpqt\") pod \"crc-storage-crc-r5trh\" (UID: \"4df1be73-aa09-4c83-a580-6454e6df05dd\") " pod="crc-storage/crc-storage-crc-r5trh" Oct 05 08:03:39 crc kubenswrapper[4846]: I1005 08:03:39.512382 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/4df1be73-aa09-4c83-a580-6454e6df05dd-node-mnt\") pod \"crc-storage-crc-r5trh\" (UID: \"4df1be73-aa09-4c83-a580-6454e6df05dd\") " pod="crc-storage/crc-storage-crc-r5trh" Oct 05 08:03:39 crc kubenswrapper[4846]: I1005 08:03:39.513333 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/4df1be73-aa09-4c83-a580-6454e6df05dd-crc-storage\") pod \"crc-storage-crc-r5trh\" (UID: \"4df1be73-aa09-4c83-a580-6454e6df05dd\") " pod="crc-storage/crc-storage-crc-r5trh" Oct 05 08:03:39 crc kubenswrapper[4846]: I1005 08:03:39.545745 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6wpqt\" (UniqueName: \"kubernetes.io/projected/4df1be73-aa09-4c83-a580-6454e6df05dd-kube-api-access-6wpqt\") pod \"crc-storage-crc-r5trh\" (UID: \"4df1be73-aa09-4c83-a580-6454e6df05dd\") " pod="crc-storage/crc-storage-crc-r5trh" Oct 05 08:03:39 crc kubenswrapper[4846]: I1005 08:03:39.631902 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-r5trh" Oct 05 08:03:40 crc kubenswrapper[4846]: I1005 08:03:40.114026 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["crc-storage/crc-storage-crc-r5trh"] Oct 05 08:03:40 crc kubenswrapper[4846]: I1005 08:03:40.124242 4846 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 05 08:03:40 crc kubenswrapper[4846]: I1005 08:03:40.512795 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a3699d3c-ac3f-4e15-983c-1b6332c9accc" path="/var/lib/kubelet/pods/a3699d3c-ac3f-4e15-983c-1b6332c9accc/volumes" Oct 05 08:03:41 crc kubenswrapper[4846]: I1005 08:03:41.075266 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-r5trh" event={"ID":"4df1be73-aa09-4c83-a580-6454e6df05dd","Type":"ContainerStarted","Data":"98e5c44de4fcb0e5e35062d5d1b192dd535a1dd475399e57eb0037da03fc51cb"} Oct 05 08:03:41 crc kubenswrapper[4846]: I1005 08:03:41.075707 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-r5trh" event={"ID":"4df1be73-aa09-4c83-a580-6454e6df05dd","Type":"ContainerStarted","Data":"60ca2104cc2195af32cac5550c211385647a2f9ad4c64b25453540ce94aacd60"} Oct 05 08:03:41 crc kubenswrapper[4846]: I1005 08:03:41.097509 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="crc-storage/crc-storage-crc-r5trh" podStartSLOduration=1.514487191 podStartE2EDuration="2.097485958s" podCreationTimestamp="2025-10-05 08:03:39 +0000 UTC" firstStartedPulling="2025-10-05 08:03:40.123803056 +0000 UTC m=+4542.364655871" lastFinishedPulling="2025-10-05 08:03:40.706801853 +0000 UTC m=+4542.947654638" observedRunningTime="2025-10-05 08:03:41.096547613 +0000 UTC m=+4543.337400388" watchObservedRunningTime="2025-10-05 08:03:41.097485958 +0000 UTC m=+4543.338338733" Oct 05 08:03:42 crc kubenswrapper[4846]: I1005 08:03:42.082422 4846 generic.go:334] "Generic (PLEG): container finished" podID="4df1be73-aa09-4c83-a580-6454e6df05dd" containerID="98e5c44de4fcb0e5e35062d5d1b192dd535a1dd475399e57eb0037da03fc51cb" exitCode=0 Oct 05 08:03:42 crc kubenswrapper[4846]: I1005 08:03:42.082460 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-r5trh" event={"ID":"4df1be73-aa09-4c83-a580-6454e6df05dd","Type":"ContainerDied","Data":"98e5c44de4fcb0e5e35062d5d1b192dd535a1dd475399e57eb0037da03fc51cb"} Oct 05 08:03:42 crc kubenswrapper[4846]: I1005 08:03:42.498624 4846 scope.go:117] "RemoveContainer" containerID="2d2e59cb51a3cd322bc7aca627ddfa4d48c6f5f95b26a7264c472515babd8d10" Oct 05 08:03:42 crc kubenswrapper[4846]: E1005 08:03:42.498912 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 08:03:43 crc kubenswrapper[4846]: I1005 08:03:43.460786 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-r5trh" Oct 05 08:03:43 crc kubenswrapper[4846]: I1005 08:03:43.572499 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6wpqt\" (UniqueName: \"kubernetes.io/projected/4df1be73-aa09-4c83-a580-6454e6df05dd-kube-api-access-6wpqt\") pod \"4df1be73-aa09-4c83-a580-6454e6df05dd\" (UID: \"4df1be73-aa09-4c83-a580-6454e6df05dd\") " Oct 05 08:03:43 crc kubenswrapper[4846]: I1005 08:03:43.572661 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/4df1be73-aa09-4c83-a580-6454e6df05dd-crc-storage\") pod \"4df1be73-aa09-4c83-a580-6454e6df05dd\" (UID: \"4df1be73-aa09-4c83-a580-6454e6df05dd\") " Oct 05 08:03:43 crc kubenswrapper[4846]: I1005 08:03:43.572798 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/4df1be73-aa09-4c83-a580-6454e6df05dd-node-mnt\") pod \"4df1be73-aa09-4c83-a580-6454e6df05dd\" (UID: \"4df1be73-aa09-4c83-a580-6454e6df05dd\") " Oct 05 08:03:43 crc kubenswrapper[4846]: I1005 08:03:43.572873 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4df1be73-aa09-4c83-a580-6454e6df05dd-node-mnt" (OuterVolumeSpecName: "node-mnt") pod "4df1be73-aa09-4c83-a580-6454e6df05dd" (UID: "4df1be73-aa09-4c83-a580-6454e6df05dd"). InnerVolumeSpecName "node-mnt". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 05 08:03:43 crc kubenswrapper[4846]: I1005 08:03:43.573332 4846 reconciler_common.go:293] "Volume detached for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/4df1be73-aa09-4c83-a580-6454e6df05dd-node-mnt\") on node \"crc\" DevicePath \"\"" Oct 05 08:03:43 crc kubenswrapper[4846]: I1005 08:03:43.578485 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4df1be73-aa09-4c83-a580-6454e6df05dd-kube-api-access-6wpqt" (OuterVolumeSpecName: "kube-api-access-6wpqt") pod "4df1be73-aa09-4c83-a580-6454e6df05dd" (UID: "4df1be73-aa09-4c83-a580-6454e6df05dd"). InnerVolumeSpecName "kube-api-access-6wpqt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:03:43 crc kubenswrapper[4846]: I1005 08:03:43.597990 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4df1be73-aa09-4c83-a580-6454e6df05dd-crc-storage" (OuterVolumeSpecName: "crc-storage") pod "4df1be73-aa09-4c83-a580-6454e6df05dd" (UID: "4df1be73-aa09-4c83-a580-6454e6df05dd"). InnerVolumeSpecName "crc-storage". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 08:03:43 crc kubenswrapper[4846]: I1005 08:03:43.675078 4846 reconciler_common.go:293] "Volume detached for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/4df1be73-aa09-4c83-a580-6454e6df05dd-crc-storage\") on node \"crc\" DevicePath \"\"" Oct 05 08:03:43 crc kubenswrapper[4846]: I1005 08:03:43.675120 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6wpqt\" (UniqueName: \"kubernetes.io/projected/4df1be73-aa09-4c83-a580-6454e6df05dd-kube-api-access-6wpqt\") on node \"crc\" DevicePath \"\"" Oct 05 08:03:44 crc kubenswrapper[4846]: I1005 08:03:44.100870 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-r5trh" event={"ID":"4df1be73-aa09-4c83-a580-6454e6df05dd","Type":"ContainerDied","Data":"60ca2104cc2195af32cac5550c211385647a2f9ad4c64b25453540ce94aacd60"} Oct 05 08:03:44 crc kubenswrapper[4846]: I1005 08:03:44.100909 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="60ca2104cc2195af32cac5550c211385647a2f9ad4c64b25453540ce94aacd60" Oct 05 08:03:44 crc kubenswrapper[4846]: I1005 08:03:44.100978 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-r5trh" Oct 05 08:03:45 crc kubenswrapper[4846]: I1005 08:03:45.415755 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["crc-storage/crc-storage-crc-r5trh"] Oct 05 08:03:45 crc kubenswrapper[4846]: I1005 08:03:45.421081 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["crc-storage/crc-storage-crc-r5trh"] Oct 05 08:03:45 crc kubenswrapper[4846]: I1005 08:03:45.587788 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["crc-storage/crc-storage-crc-nmlhl"] Oct 05 08:03:45 crc kubenswrapper[4846]: E1005 08:03:45.588307 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4df1be73-aa09-4c83-a580-6454e6df05dd" containerName="storage" Oct 05 08:03:45 crc kubenswrapper[4846]: I1005 08:03:45.588337 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="4df1be73-aa09-4c83-a580-6454e6df05dd" containerName="storage" Oct 05 08:03:45 crc kubenswrapper[4846]: I1005 08:03:45.588610 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="4df1be73-aa09-4c83-a580-6454e6df05dd" containerName="storage" Oct 05 08:03:45 crc kubenswrapper[4846]: I1005 08:03:45.589387 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-nmlhl" Oct 05 08:03:45 crc kubenswrapper[4846]: I1005 08:03:45.593555 4846 reflector.go:368] Caches populated for *v1.Secret from object-"crc-storage"/"crc-storage-dockercfg-w2v6h" Oct 05 08:03:45 crc kubenswrapper[4846]: I1005 08:03:45.593961 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"openshift-service-ca.crt" Oct 05 08:03:45 crc kubenswrapper[4846]: I1005 08:03:45.594007 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"kube-root-ca.crt" Oct 05 08:03:45 crc kubenswrapper[4846]: I1005 08:03:45.594389 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"crc-storage" Oct 05 08:03:45 crc kubenswrapper[4846]: I1005 08:03:45.601267 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["crc-storage/crc-storage-crc-nmlhl"] Oct 05 08:03:45 crc kubenswrapper[4846]: I1005 08:03:45.705079 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/150a3c25-af3b-47ec-8c95-bf5c3934d8a2-node-mnt\") pod \"crc-storage-crc-nmlhl\" (UID: \"150a3c25-af3b-47ec-8c95-bf5c3934d8a2\") " pod="crc-storage/crc-storage-crc-nmlhl" Oct 05 08:03:45 crc kubenswrapper[4846]: I1005 08:03:45.705347 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-86xqg\" (UniqueName: \"kubernetes.io/projected/150a3c25-af3b-47ec-8c95-bf5c3934d8a2-kube-api-access-86xqg\") pod \"crc-storage-crc-nmlhl\" (UID: \"150a3c25-af3b-47ec-8c95-bf5c3934d8a2\") " pod="crc-storage/crc-storage-crc-nmlhl" Oct 05 08:03:45 crc kubenswrapper[4846]: I1005 08:03:45.705574 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/150a3c25-af3b-47ec-8c95-bf5c3934d8a2-crc-storage\") pod \"crc-storage-crc-nmlhl\" (UID: \"150a3c25-af3b-47ec-8c95-bf5c3934d8a2\") " pod="crc-storage/crc-storage-crc-nmlhl" Oct 05 08:03:45 crc kubenswrapper[4846]: I1005 08:03:45.806967 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-86xqg\" (UniqueName: \"kubernetes.io/projected/150a3c25-af3b-47ec-8c95-bf5c3934d8a2-kube-api-access-86xqg\") pod \"crc-storage-crc-nmlhl\" (UID: \"150a3c25-af3b-47ec-8c95-bf5c3934d8a2\") " pod="crc-storage/crc-storage-crc-nmlhl" Oct 05 08:03:45 crc kubenswrapper[4846]: I1005 08:03:45.807070 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/150a3c25-af3b-47ec-8c95-bf5c3934d8a2-crc-storage\") pod \"crc-storage-crc-nmlhl\" (UID: \"150a3c25-af3b-47ec-8c95-bf5c3934d8a2\") " pod="crc-storage/crc-storage-crc-nmlhl" Oct 05 08:03:45 crc kubenswrapper[4846]: I1005 08:03:45.807206 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/150a3c25-af3b-47ec-8c95-bf5c3934d8a2-node-mnt\") pod \"crc-storage-crc-nmlhl\" (UID: \"150a3c25-af3b-47ec-8c95-bf5c3934d8a2\") " pod="crc-storage/crc-storage-crc-nmlhl" Oct 05 08:03:45 crc kubenswrapper[4846]: I1005 08:03:45.807518 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/150a3c25-af3b-47ec-8c95-bf5c3934d8a2-node-mnt\") pod \"crc-storage-crc-nmlhl\" (UID: \"150a3c25-af3b-47ec-8c95-bf5c3934d8a2\") " pod="crc-storage/crc-storage-crc-nmlhl" Oct 05 08:03:45 crc kubenswrapper[4846]: I1005 08:03:45.808588 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/150a3c25-af3b-47ec-8c95-bf5c3934d8a2-crc-storage\") pod \"crc-storage-crc-nmlhl\" (UID: \"150a3c25-af3b-47ec-8c95-bf5c3934d8a2\") " pod="crc-storage/crc-storage-crc-nmlhl" Oct 05 08:03:45 crc kubenswrapper[4846]: I1005 08:03:45.827390 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-86xqg\" (UniqueName: \"kubernetes.io/projected/150a3c25-af3b-47ec-8c95-bf5c3934d8a2-kube-api-access-86xqg\") pod \"crc-storage-crc-nmlhl\" (UID: \"150a3c25-af3b-47ec-8c95-bf5c3934d8a2\") " pod="crc-storage/crc-storage-crc-nmlhl" Oct 05 08:03:45 crc kubenswrapper[4846]: I1005 08:03:45.945752 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-nmlhl" Oct 05 08:03:46 crc kubenswrapper[4846]: I1005 08:03:46.425816 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["crc-storage/crc-storage-crc-nmlhl"] Oct 05 08:03:46 crc kubenswrapper[4846]: I1005 08:03:46.514400 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4df1be73-aa09-4c83-a580-6454e6df05dd" path="/var/lib/kubelet/pods/4df1be73-aa09-4c83-a580-6454e6df05dd/volumes" Oct 05 08:03:47 crc kubenswrapper[4846]: I1005 08:03:47.130163 4846 generic.go:334] "Generic (PLEG): container finished" podID="150a3c25-af3b-47ec-8c95-bf5c3934d8a2" containerID="c3e86ca01231af2b6b4540d4c6b1078f396f226237fe8a747896bd8290638e27" exitCode=0 Oct 05 08:03:47 crc kubenswrapper[4846]: I1005 08:03:47.130384 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-nmlhl" event={"ID":"150a3c25-af3b-47ec-8c95-bf5c3934d8a2","Type":"ContainerDied","Data":"c3e86ca01231af2b6b4540d4c6b1078f396f226237fe8a747896bd8290638e27"} Oct 05 08:03:47 crc kubenswrapper[4846]: I1005 08:03:47.130797 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-nmlhl" event={"ID":"150a3c25-af3b-47ec-8c95-bf5c3934d8a2","Type":"ContainerStarted","Data":"8242ee3a08ae60bb1292b8516c8a590f1e6bc10ded27870af3773f369aadcd90"} Oct 05 08:03:48 crc kubenswrapper[4846]: I1005 08:03:48.458093 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-nmlhl" Oct 05 08:03:48 crc kubenswrapper[4846]: I1005 08:03:48.556930 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-86xqg\" (UniqueName: \"kubernetes.io/projected/150a3c25-af3b-47ec-8c95-bf5c3934d8a2-kube-api-access-86xqg\") pod \"150a3c25-af3b-47ec-8c95-bf5c3934d8a2\" (UID: \"150a3c25-af3b-47ec-8c95-bf5c3934d8a2\") " Oct 05 08:03:48 crc kubenswrapper[4846]: I1005 08:03:48.558192 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/150a3c25-af3b-47ec-8c95-bf5c3934d8a2-node-mnt\") pod \"150a3c25-af3b-47ec-8c95-bf5c3934d8a2\" (UID: \"150a3c25-af3b-47ec-8c95-bf5c3934d8a2\") " Oct 05 08:03:48 crc kubenswrapper[4846]: I1005 08:03:48.558511 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/150a3c25-af3b-47ec-8c95-bf5c3934d8a2-crc-storage\") pod \"150a3c25-af3b-47ec-8c95-bf5c3934d8a2\" (UID: \"150a3c25-af3b-47ec-8c95-bf5c3934d8a2\") " Oct 05 08:03:48 crc kubenswrapper[4846]: I1005 08:03:48.559534 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/150a3c25-af3b-47ec-8c95-bf5c3934d8a2-node-mnt" (OuterVolumeSpecName: "node-mnt") pod "150a3c25-af3b-47ec-8c95-bf5c3934d8a2" (UID: "150a3c25-af3b-47ec-8c95-bf5c3934d8a2"). InnerVolumeSpecName "node-mnt". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 05 08:03:48 crc kubenswrapper[4846]: I1005 08:03:48.573709 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/150a3c25-af3b-47ec-8c95-bf5c3934d8a2-kube-api-access-86xqg" (OuterVolumeSpecName: "kube-api-access-86xqg") pod "150a3c25-af3b-47ec-8c95-bf5c3934d8a2" (UID: "150a3c25-af3b-47ec-8c95-bf5c3934d8a2"). InnerVolumeSpecName "kube-api-access-86xqg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:03:48 crc kubenswrapper[4846]: I1005 08:03:48.586651 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/150a3c25-af3b-47ec-8c95-bf5c3934d8a2-crc-storage" (OuterVolumeSpecName: "crc-storage") pod "150a3c25-af3b-47ec-8c95-bf5c3934d8a2" (UID: "150a3c25-af3b-47ec-8c95-bf5c3934d8a2"). InnerVolumeSpecName "crc-storage". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 08:03:48 crc kubenswrapper[4846]: I1005 08:03:48.660144 4846 reconciler_common.go:293] "Volume detached for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/150a3c25-af3b-47ec-8c95-bf5c3934d8a2-node-mnt\") on node \"crc\" DevicePath \"\"" Oct 05 08:03:48 crc kubenswrapper[4846]: I1005 08:03:48.660710 4846 reconciler_common.go:293] "Volume detached for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/150a3c25-af3b-47ec-8c95-bf5c3934d8a2-crc-storage\") on node \"crc\" DevicePath \"\"" Oct 05 08:03:48 crc kubenswrapper[4846]: I1005 08:03:48.660792 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-86xqg\" (UniqueName: \"kubernetes.io/projected/150a3c25-af3b-47ec-8c95-bf5c3934d8a2-kube-api-access-86xqg\") on node \"crc\" DevicePath \"\"" Oct 05 08:03:49 crc kubenswrapper[4846]: I1005 08:03:49.176537 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-nmlhl" event={"ID":"150a3c25-af3b-47ec-8c95-bf5c3934d8a2","Type":"ContainerDied","Data":"8242ee3a08ae60bb1292b8516c8a590f1e6bc10ded27870af3773f369aadcd90"} Oct 05 08:03:49 crc kubenswrapper[4846]: I1005 08:03:49.176598 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8242ee3a08ae60bb1292b8516c8a590f1e6bc10ded27870af3773f369aadcd90" Oct 05 08:03:49 crc kubenswrapper[4846]: I1005 08:03:49.176610 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-nmlhl" Oct 05 08:03:54 crc kubenswrapper[4846]: I1005 08:03:54.498526 4846 scope.go:117] "RemoveContainer" containerID="2d2e59cb51a3cd322bc7aca627ddfa4d48c6f5f95b26a7264c472515babd8d10" Oct 05 08:03:54 crc kubenswrapper[4846]: E1005 08:03:54.499416 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 08:04:08 crc kubenswrapper[4846]: I1005 08:04:08.505653 4846 scope.go:117] "RemoveContainer" containerID="2d2e59cb51a3cd322bc7aca627ddfa4d48c6f5f95b26a7264c472515babd8d10" Oct 05 08:04:08 crc kubenswrapper[4846]: E1005 08:04:08.507083 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 08:04:23 crc kubenswrapper[4846]: I1005 08:04:23.497531 4846 scope.go:117] "RemoveContainer" containerID="2d2e59cb51a3cd322bc7aca627ddfa4d48c6f5f95b26a7264c472515babd8d10" Oct 05 08:04:24 crc kubenswrapper[4846]: I1005 08:04:24.522772 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" event={"ID":"ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7","Type":"ContainerStarted","Data":"bad87dae9648a1776fbb2ee38b4fedce1273888d4f401172760726d09bb965f3"} Oct 05 08:04:30 crc kubenswrapper[4846]: I1005 08:04:30.080125 4846 scope.go:117] "RemoveContainer" containerID="c428789eb7e249054d1de1e757e22c2d6318371fc59227c254b993a1c774bb6c" Oct 05 08:04:54 crc kubenswrapper[4846]: I1005 08:04:54.025104 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-sb52r"] Oct 05 08:04:54 crc kubenswrapper[4846]: E1005 08:04:54.028056 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="150a3c25-af3b-47ec-8c95-bf5c3934d8a2" containerName="storage" Oct 05 08:04:54 crc kubenswrapper[4846]: I1005 08:04:54.028134 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="150a3c25-af3b-47ec-8c95-bf5c3934d8a2" containerName="storage" Oct 05 08:04:54 crc kubenswrapper[4846]: I1005 08:04:54.029231 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="150a3c25-af3b-47ec-8c95-bf5c3934d8a2" containerName="storage" Oct 05 08:04:54 crc kubenswrapper[4846]: I1005 08:04:54.032131 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-sb52r" Oct 05 08:04:54 crc kubenswrapper[4846]: I1005 08:04:54.040267 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-sb52r"] Oct 05 08:04:54 crc kubenswrapper[4846]: I1005 08:04:54.162279 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b4wcv\" (UniqueName: \"kubernetes.io/projected/ff3207f0-4ff4-455f-b3e7-1a6afcce2576-kube-api-access-b4wcv\") pod \"certified-operators-sb52r\" (UID: \"ff3207f0-4ff4-455f-b3e7-1a6afcce2576\") " pod="openshift-marketplace/certified-operators-sb52r" Oct 05 08:04:54 crc kubenswrapper[4846]: I1005 08:04:54.162361 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ff3207f0-4ff4-455f-b3e7-1a6afcce2576-catalog-content\") pod \"certified-operators-sb52r\" (UID: \"ff3207f0-4ff4-455f-b3e7-1a6afcce2576\") " pod="openshift-marketplace/certified-operators-sb52r" Oct 05 08:04:54 crc kubenswrapper[4846]: I1005 08:04:54.162560 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ff3207f0-4ff4-455f-b3e7-1a6afcce2576-utilities\") pod \"certified-operators-sb52r\" (UID: \"ff3207f0-4ff4-455f-b3e7-1a6afcce2576\") " pod="openshift-marketplace/certified-operators-sb52r" Oct 05 08:04:54 crc kubenswrapper[4846]: I1005 08:04:54.263749 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b4wcv\" (UniqueName: \"kubernetes.io/projected/ff3207f0-4ff4-455f-b3e7-1a6afcce2576-kube-api-access-b4wcv\") pod \"certified-operators-sb52r\" (UID: \"ff3207f0-4ff4-455f-b3e7-1a6afcce2576\") " pod="openshift-marketplace/certified-operators-sb52r" Oct 05 08:04:54 crc kubenswrapper[4846]: I1005 08:04:54.263911 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ff3207f0-4ff4-455f-b3e7-1a6afcce2576-catalog-content\") pod \"certified-operators-sb52r\" (UID: \"ff3207f0-4ff4-455f-b3e7-1a6afcce2576\") " pod="openshift-marketplace/certified-operators-sb52r" Oct 05 08:04:54 crc kubenswrapper[4846]: I1005 08:04:54.264009 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ff3207f0-4ff4-455f-b3e7-1a6afcce2576-utilities\") pod \"certified-operators-sb52r\" (UID: \"ff3207f0-4ff4-455f-b3e7-1a6afcce2576\") " pod="openshift-marketplace/certified-operators-sb52r" Oct 05 08:04:54 crc kubenswrapper[4846]: I1005 08:04:54.264972 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ff3207f0-4ff4-455f-b3e7-1a6afcce2576-catalog-content\") pod \"certified-operators-sb52r\" (UID: \"ff3207f0-4ff4-455f-b3e7-1a6afcce2576\") " pod="openshift-marketplace/certified-operators-sb52r" Oct 05 08:04:54 crc kubenswrapper[4846]: I1005 08:04:54.264994 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ff3207f0-4ff4-455f-b3e7-1a6afcce2576-utilities\") pod \"certified-operators-sb52r\" (UID: \"ff3207f0-4ff4-455f-b3e7-1a6afcce2576\") " pod="openshift-marketplace/certified-operators-sb52r" Oct 05 08:04:54 crc kubenswrapper[4846]: I1005 08:04:54.297105 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b4wcv\" (UniqueName: \"kubernetes.io/projected/ff3207f0-4ff4-455f-b3e7-1a6afcce2576-kube-api-access-b4wcv\") pod \"certified-operators-sb52r\" (UID: \"ff3207f0-4ff4-455f-b3e7-1a6afcce2576\") " pod="openshift-marketplace/certified-operators-sb52r" Oct 05 08:04:54 crc kubenswrapper[4846]: I1005 08:04:54.359749 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-sb52r" Oct 05 08:04:54 crc kubenswrapper[4846]: I1005 08:04:54.668149 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-sb52r"] Oct 05 08:04:54 crc kubenswrapper[4846]: I1005 08:04:54.817588 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-sb52r" event={"ID":"ff3207f0-4ff4-455f-b3e7-1a6afcce2576","Type":"ContainerStarted","Data":"fee137fb889928d8d2a99dab78b2b33a135662461beae686e1dc5a423a309e7f"} Oct 05 08:04:55 crc kubenswrapper[4846]: I1005 08:04:55.828382 4846 generic.go:334] "Generic (PLEG): container finished" podID="ff3207f0-4ff4-455f-b3e7-1a6afcce2576" containerID="18f787787c03f613060c73e3b55df0efed664694b254aaac2c02309b8a714513" exitCode=0 Oct 05 08:04:55 crc kubenswrapper[4846]: I1005 08:04:55.828447 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-sb52r" event={"ID":"ff3207f0-4ff4-455f-b3e7-1a6afcce2576","Type":"ContainerDied","Data":"18f787787c03f613060c73e3b55df0efed664694b254aaac2c02309b8a714513"} Oct 05 08:04:56 crc kubenswrapper[4846]: I1005 08:04:56.838551 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-sb52r" event={"ID":"ff3207f0-4ff4-455f-b3e7-1a6afcce2576","Type":"ContainerStarted","Data":"528e94c042e54a195beda32cf735d030d618ee22831b33fef7195a827034254d"} Oct 05 08:04:57 crc kubenswrapper[4846]: I1005 08:04:57.846000 4846 generic.go:334] "Generic (PLEG): container finished" podID="ff3207f0-4ff4-455f-b3e7-1a6afcce2576" containerID="528e94c042e54a195beda32cf735d030d618ee22831b33fef7195a827034254d" exitCode=0 Oct 05 08:04:57 crc kubenswrapper[4846]: I1005 08:04:57.846042 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-sb52r" event={"ID":"ff3207f0-4ff4-455f-b3e7-1a6afcce2576","Type":"ContainerDied","Data":"528e94c042e54a195beda32cf735d030d618ee22831b33fef7195a827034254d"} Oct 05 08:04:58 crc kubenswrapper[4846]: I1005 08:04:58.857697 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-sb52r" event={"ID":"ff3207f0-4ff4-455f-b3e7-1a6afcce2576","Type":"ContainerStarted","Data":"6f57191ec60cfdc03764d92f8e6b80eee1953178e29ff5ff7b83ea814ee3ed47"} Oct 05 08:04:58 crc kubenswrapper[4846]: I1005 08:04:58.886236 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-sb52r" podStartSLOduration=3.469481492 podStartE2EDuration="5.886210271s" podCreationTimestamp="2025-10-05 08:04:53 +0000 UTC" firstStartedPulling="2025-10-05 08:04:55.831698187 +0000 UTC m=+4618.072551012" lastFinishedPulling="2025-10-05 08:04:58.248427016 +0000 UTC m=+4620.489279791" observedRunningTime="2025-10-05 08:04:58.880771116 +0000 UTC m=+4621.121623941" watchObservedRunningTime="2025-10-05 08:04:58.886210271 +0000 UTC m=+4621.127063086" Oct 05 08:05:04 crc kubenswrapper[4846]: I1005 08:05:04.360906 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-sb52r" Oct 05 08:05:04 crc kubenswrapper[4846]: I1005 08:05:04.361765 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-sb52r" Oct 05 08:05:04 crc kubenswrapper[4846]: I1005 08:05:04.448520 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-sb52r" Oct 05 08:05:04 crc kubenswrapper[4846]: I1005 08:05:04.977818 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-sb52r" Oct 05 08:05:07 crc kubenswrapper[4846]: I1005 08:05:07.600099 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-sb52r"] Oct 05 08:05:07 crc kubenswrapper[4846]: I1005 08:05:07.600323 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-sb52r" podUID="ff3207f0-4ff4-455f-b3e7-1a6afcce2576" containerName="registry-server" containerID="cri-o://6f57191ec60cfdc03764d92f8e6b80eee1953178e29ff5ff7b83ea814ee3ed47" gracePeriod=2 Oct 05 08:05:07 crc kubenswrapper[4846]: I1005 08:05:07.946345 4846 generic.go:334] "Generic (PLEG): container finished" podID="ff3207f0-4ff4-455f-b3e7-1a6afcce2576" containerID="6f57191ec60cfdc03764d92f8e6b80eee1953178e29ff5ff7b83ea814ee3ed47" exitCode=0 Oct 05 08:05:07 crc kubenswrapper[4846]: I1005 08:05:07.946468 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-sb52r" event={"ID":"ff3207f0-4ff4-455f-b3e7-1a6afcce2576","Type":"ContainerDied","Data":"6f57191ec60cfdc03764d92f8e6b80eee1953178e29ff5ff7b83ea814ee3ed47"} Oct 05 08:05:08 crc kubenswrapper[4846]: I1005 08:05:08.485473 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-sb52r" Oct 05 08:05:08 crc kubenswrapper[4846]: I1005 08:05:08.608372 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ff3207f0-4ff4-455f-b3e7-1a6afcce2576-utilities\") pod \"ff3207f0-4ff4-455f-b3e7-1a6afcce2576\" (UID: \"ff3207f0-4ff4-455f-b3e7-1a6afcce2576\") " Oct 05 08:05:08 crc kubenswrapper[4846]: I1005 08:05:08.608469 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b4wcv\" (UniqueName: \"kubernetes.io/projected/ff3207f0-4ff4-455f-b3e7-1a6afcce2576-kube-api-access-b4wcv\") pod \"ff3207f0-4ff4-455f-b3e7-1a6afcce2576\" (UID: \"ff3207f0-4ff4-455f-b3e7-1a6afcce2576\") " Oct 05 08:05:08 crc kubenswrapper[4846]: I1005 08:05:08.608544 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ff3207f0-4ff4-455f-b3e7-1a6afcce2576-catalog-content\") pod \"ff3207f0-4ff4-455f-b3e7-1a6afcce2576\" (UID: \"ff3207f0-4ff4-455f-b3e7-1a6afcce2576\") " Oct 05 08:05:08 crc kubenswrapper[4846]: I1005 08:05:08.610382 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ff3207f0-4ff4-455f-b3e7-1a6afcce2576-utilities" (OuterVolumeSpecName: "utilities") pod "ff3207f0-4ff4-455f-b3e7-1a6afcce2576" (UID: "ff3207f0-4ff4-455f-b3e7-1a6afcce2576"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 08:05:08 crc kubenswrapper[4846]: I1005 08:05:08.617644 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ff3207f0-4ff4-455f-b3e7-1a6afcce2576-kube-api-access-b4wcv" (OuterVolumeSpecName: "kube-api-access-b4wcv") pod "ff3207f0-4ff4-455f-b3e7-1a6afcce2576" (UID: "ff3207f0-4ff4-455f-b3e7-1a6afcce2576"). InnerVolumeSpecName "kube-api-access-b4wcv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:05:08 crc kubenswrapper[4846]: I1005 08:05:08.688717 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ff3207f0-4ff4-455f-b3e7-1a6afcce2576-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ff3207f0-4ff4-455f-b3e7-1a6afcce2576" (UID: "ff3207f0-4ff4-455f-b3e7-1a6afcce2576"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 08:05:08 crc kubenswrapper[4846]: I1005 08:05:08.709906 4846 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ff3207f0-4ff4-455f-b3e7-1a6afcce2576-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 05 08:05:08 crc kubenswrapper[4846]: I1005 08:05:08.709951 4846 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ff3207f0-4ff4-455f-b3e7-1a6afcce2576-utilities\") on node \"crc\" DevicePath \"\"" Oct 05 08:05:08 crc kubenswrapper[4846]: I1005 08:05:08.709965 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b4wcv\" (UniqueName: \"kubernetes.io/projected/ff3207f0-4ff4-455f-b3e7-1a6afcce2576-kube-api-access-b4wcv\") on node \"crc\" DevicePath \"\"" Oct 05 08:05:08 crc kubenswrapper[4846]: I1005 08:05:08.957934 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-sb52r" event={"ID":"ff3207f0-4ff4-455f-b3e7-1a6afcce2576","Type":"ContainerDied","Data":"fee137fb889928d8d2a99dab78b2b33a135662461beae686e1dc5a423a309e7f"} Oct 05 08:05:08 crc kubenswrapper[4846]: I1005 08:05:08.958037 4846 scope.go:117] "RemoveContainer" containerID="6f57191ec60cfdc03764d92f8e6b80eee1953178e29ff5ff7b83ea814ee3ed47" Oct 05 08:05:08 crc kubenswrapper[4846]: I1005 08:05:08.958079 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-sb52r" Oct 05 08:05:08 crc kubenswrapper[4846]: I1005 08:05:08.985791 4846 scope.go:117] "RemoveContainer" containerID="528e94c042e54a195beda32cf735d030d618ee22831b33fef7195a827034254d" Oct 05 08:05:09 crc kubenswrapper[4846]: I1005 08:05:09.014914 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-sb52r"] Oct 05 08:05:09 crc kubenswrapper[4846]: I1005 08:05:09.024047 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-sb52r"] Oct 05 08:05:09 crc kubenswrapper[4846]: I1005 08:05:09.026717 4846 scope.go:117] "RemoveContainer" containerID="18f787787c03f613060c73e3b55df0efed664694b254aaac2c02309b8a714513" Oct 05 08:05:10 crc kubenswrapper[4846]: I1005 08:05:10.516996 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ff3207f0-4ff4-455f-b3e7-1a6afcce2576" path="/var/lib/kubelet/pods/ff3207f0-4ff4-455f-b3e7-1a6afcce2576/volumes" Oct 05 08:05:57 crc kubenswrapper[4846]: I1005 08:05:57.856924 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6697c69467-27t47"] Oct 05 08:05:57 crc kubenswrapper[4846]: E1005 08:05:57.857788 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ff3207f0-4ff4-455f-b3e7-1a6afcce2576" containerName="registry-server" Oct 05 08:05:57 crc kubenswrapper[4846]: I1005 08:05:57.857802 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="ff3207f0-4ff4-455f-b3e7-1a6afcce2576" containerName="registry-server" Oct 05 08:05:57 crc kubenswrapper[4846]: E1005 08:05:57.857826 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ff3207f0-4ff4-455f-b3e7-1a6afcce2576" containerName="extract-utilities" Oct 05 08:05:57 crc kubenswrapper[4846]: I1005 08:05:57.857834 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="ff3207f0-4ff4-455f-b3e7-1a6afcce2576" containerName="extract-utilities" Oct 05 08:05:57 crc kubenswrapper[4846]: E1005 08:05:57.857870 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ff3207f0-4ff4-455f-b3e7-1a6afcce2576" containerName="extract-content" Oct 05 08:05:57 crc kubenswrapper[4846]: I1005 08:05:57.857878 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="ff3207f0-4ff4-455f-b3e7-1a6afcce2576" containerName="extract-content" Oct 05 08:05:57 crc kubenswrapper[4846]: I1005 08:05:57.858065 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="ff3207f0-4ff4-455f-b3e7-1a6afcce2576" containerName="registry-server" Oct 05 08:05:57 crc kubenswrapper[4846]: I1005 08:05:57.863933 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-55cbcdb7f9-cftk8"] Oct 05 08:05:57 crc kubenswrapper[4846]: I1005 08:05:57.864056 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6697c69467-27t47" Oct 05 08:05:57 crc kubenswrapper[4846]: I1005 08:05:57.865271 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55cbcdb7f9-cftk8" Oct 05 08:05:57 crc kubenswrapper[4846]: I1005 08:05:57.872155 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-svc" Oct 05 08:05:57 crc kubenswrapper[4846]: I1005 08:05:57.872578 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dnsmasq-dns-dockercfg-szbb8" Oct 05 08:05:57 crc kubenswrapper[4846]: I1005 08:05:57.872757 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns" Oct 05 08:05:57 crc kubenswrapper[4846]: I1005 08:05:57.872894 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openshift-service-ca.crt" Oct 05 08:05:57 crc kubenswrapper[4846]: I1005 08:05:57.876789 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"kube-root-ca.crt" Oct 05 08:05:57 crc kubenswrapper[4846]: I1005 08:05:57.881314 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-55cbcdb7f9-cftk8"] Oct 05 08:05:57 crc kubenswrapper[4846]: I1005 08:05:57.887586 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6697c69467-27t47"] Oct 05 08:05:57 crc kubenswrapper[4846]: I1005 08:05:57.894911 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d6d6783a-7816-402e-b38e-935ffcb7314e-config\") pod \"dnsmasq-dns-55cbcdb7f9-cftk8\" (UID: \"d6d6783a-7816-402e-b38e-935ffcb7314e\") " pod="openstack/dnsmasq-dns-55cbcdb7f9-cftk8" Oct 05 08:05:57 crc kubenswrapper[4846]: I1005 08:05:57.894954 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7sbjj\" (UniqueName: \"kubernetes.io/projected/d6d6783a-7816-402e-b38e-935ffcb7314e-kube-api-access-7sbjj\") pod \"dnsmasq-dns-55cbcdb7f9-cftk8\" (UID: \"d6d6783a-7816-402e-b38e-935ffcb7314e\") " pod="openstack/dnsmasq-dns-55cbcdb7f9-cftk8" Oct 05 08:05:57 crc kubenswrapper[4846]: I1005 08:05:57.894994 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/037a0419-1e47-429c-ac6f-e2fe6e7db5d9-dns-svc\") pod \"dnsmasq-dns-6697c69467-27t47\" (UID: \"037a0419-1e47-429c-ac6f-e2fe6e7db5d9\") " pod="openstack/dnsmasq-dns-6697c69467-27t47" Oct 05 08:05:57 crc kubenswrapper[4846]: I1005 08:05:57.895025 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v65gp\" (UniqueName: \"kubernetes.io/projected/037a0419-1e47-429c-ac6f-e2fe6e7db5d9-kube-api-access-v65gp\") pod \"dnsmasq-dns-6697c69467-27t47\" (UID: \"037a0419-1e47-429c-ac6f-e2fe6e7db5d9\") " pod="openstack/dnsmasq-dns-6697c69467-27t47" Oct 05 08:05:57 crc kubenswrapper[4846]: I1005 08:05:57.895078 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/037a0419-1e47-429c-ac6f-e2fe6e7db5d9-config\") pod \"dnsmasq-dns-6697c69467-27t47\" (UID: \"037a0419-1e47-429c-ac6f-e2fe6e7db5d9\") " pod="openstack/dnsmasq-dns-6697c69467-27t47" Oct 05 08:05:57 crc kubenswrapper[4846]: I1005 08:05:57.995769 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v65gp\" (UniqueName: \"kubernetes.io/projected/037a0419-1e47-429c-ac6f-e2fe6e7db5d9-kube-api-access-v65gp\") pod \"dnsmasq-dns-6697c69467-27t47\" (UID: \"037a0419-1e47-429c-ac6f-e2fe6e7db5d9\") " pod="openstack/dnsmasq-dns-6697c69467-27t47" Oct 05 08:05:57 crc kubenswrapper[4846]: I1005 08:05:57.995876 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/037a0419-1e47-429c-ac6f-e2fe6e7db5d9-config\") pod \"dnsmasq-dns-6697c69467-27t47\" (UID: \"037a0419-1e47-429c-ac6f-e2fe6e7db5d9\") " pod="openstack/dnsmasq-dns-6697c69467-27t47" Oct 05 08:05:57 crc kubenswrapper[4846]: I1005 08:05:57.995920 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d6d6783a-7816-402e-b38e-935ffcb7314e-config\") pod \"dnsmasq-dns-55cbcdb7f9-cftk8\" (UID: \"d6d6783a-7816-402e-b38e-935ffcb7314e\") " pod="openstack/dnsmasq-dns-55cbcdb7f9-cftk8" Oct 05 08:05:57 crc kubenswrapper[4846]: I1005 08:05:57.995946 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7sbjj\" (UniqueName: \"kubernetes.io/projected/d6d6783a-7816-402e-b38e-935ffcb7314e-kube-api-access-7sbjj\") pod \"dnsmasq-dns-55cbcdb7f9-cftk8\" (UID: \"d6d6783a-7816-402e-b38e-935ffcb7314e\") " pod="openstack/dnsmasq-dns-55cbcdb7f9-cftk8" Oct 05 08:05:57 crc kubenswrapper[4846]: I1005 08:05:57.995989 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/037a0419-1e47-429c-ac6f-e2fe6e7db5d9-dns-svc\") pod \"dnsmasq-dns-6697c69467-27t47\" (UID: \"037a0419-1e47-429c-ac6f-e2fe6e7db5d9\") " pod="openstack/dnsmasq-dns-6697c69467-27t47" Oct 05 08:05:57 crc kubenswrapper[4846]: I1005 08:05:57.997062 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/037a0419-1e47-429c-ac6f-e2fe6e7db5d9-config\") pod \"dnsmasq-dns-6697c69467-27t47\" (UID: \"037a0419-1e47-429c-ac6f-e2fe6e7db5d9\") " pod="openstack/dnsmasq-dns-6697c69467-27t47" Oct 05 08:05:57 crc kubenswrapper[4846]: I1005 08:05:57.997069 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/037a0419-1e47-429c-ac6f-e2fe6e7db5d9-dns-svc\") pod \"dnsmasq-dns-6697c69467-27t47\" (UID: \"037a0419-1e47-429c-ac6f-e2fe6e7db5d9\") " pod="openstack/dnsmasq-dns-6697c69467-27t47" Oct 05 08:05:57 crc kubenswrapper[4846]: I1005 08:05:57.997814 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d6d6783a-7816-402e-b38e-935ffcb7314e-config\") pod \"dnsmasq-dns-55cbcdb7f9-cftk8\" (UID: \"d6d6783a-7816-402e-b38e-935ffcb7314e\") " pod="openstack/dnsmasq-dns-55cbcdb7f9-cftk8" Oct 05 08:05:58 crc kubenswrapper[4846]: I1005 08:05:58.018945 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7sbjj\" (UniqueName: \"kubernetes.io/projected/d6d6783a-7816-402e-b38e-935ffcb7314e-kube-api-access-7sbjj\") pod \"dnsmasq-dns-55cbcdb7f9-cftk8\" (UID: \"d6d6783a-7816-402e-b38e-935ffcb7314e\") " pod="openstack/dnsmasq-dns-55cbcdb7f9-cftk8" Oct 05 08:05:58 crc kubenswrapper[4846]: I1005 08:05:58.030240 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v65gp\" (UniqueName: \"kubernetes.io/projected/037a0419-1e47-429c-ac6f-e2fe6e7db5d9-kube-api-access-v65gp\") pod \"dnsmasq-dns-6697c69467-27t47\" (UID: \"037a0419-1e47-429c-ac6f-e2fe6e7db5d9\") " pod="openstack/dnsmasq-dns-6697c69467-27t47" Oct 05 08:05:58 crc kubenswrapper[4846]: I1005 08:05:58.090608 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-55cbcdb7f9-cftk8"] Oct 05 08:05:58 crc kubenswrapper[4846]: I1005 08:05:58.091128 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55cbcdb7f9-cftk8" Oct 05 08:05:58 crc kubenswrapper[4846]: I1005 08:05:58.117023 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-78dbf54cbc-c2z7x"] Oct 05 08:05:58 crc kubenswrapper[4846]: I1005 08:05:58.118126 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dbf54cbc-c2z7x" Oct 05 08:05:58 crc kubenswrapper[4846]: I1005 08:05:58.141798 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dbf54cbc-c2z7x"] Oct 05 08:05:58 crc kubenswrapper[4846]: I1005 08:05:58.188709 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6697c69467-27t47" Oct 05 08:05:58 crc kubenswrapper[4846]: I1005 08:05:58.310228 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ebf7ce11-32a8-4fea-9488-8f48b08991a9-config\") pod \"dnsmasq-dns-78dbf54cbc-c2z7x\" (UID: \"ebf7ce11-32a8-4fea-9488-8f48b08991a9\") " pod="openstack/dnsmasq-dns-78dbf54cbc-c2z7x" Oct 05 08:05:58 crc kubenswrapper[4846]: I1005 08:05:58.310512 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zm54k\" (UniqueName: \"kubernetes.io/projected/ebf7ce11-32a8-4fea-9488-8f48b08991a9-kube-api-access-zm54k\") pod \"dnsmasq-dns-78dbf54cbc-c2z7x\" (UID: \"ebf7ce11-32a8-4fea-9488-8f48b08991a9\") " pod="openstack/dnsmasq-dns-78dbf54cbc-c2z7x" Oct 05 08:05:58 crc kubenswrapper[4846]: I1005 08:05:58.310536 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ebf7ce11-32a8-4fea-9488-8f48b08991a9-dns-svc\") pod \"dnsmasq-dns-78dbf54cbc-c2z7x\" (UID: \"ebf7ce11-32a8-4fea-9488-8f48b08991a9\") " pod="openstack/dnsmasq-dns-78dbf54cbc-c2z7x" Oct 05 08:05:58 crc kubenswrapper[4846]: I1005 08:05:58.390375 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6697c69467-27t47"] Oct 05 08:05:58 crc kubenswrapper[4846]: I1005 08:05:58.405653 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6cfd7b4c45-gjjg8"] Oct 05 08:05:58 crc kubenswrapper[4846]: I1005 08:05:58.410115 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6cfd7b4c45-gjjg8" Oct 05 08:05:58 crc kubenswrapper[4846]: I1005 08:05:58.417977 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ebf7ce11-32a8-4fea-9488-8f48b08991a9-config\") pod \"dnsmasq-dns-78dbf54cbc-c2z7x\" (UID: \"ebf7ce11-32a8-4fea-9488-8f48b08991a9\") " pod="openstack/dnsmasq-dns-78dbf54cbc-c2z7x" Oct 05 08:05:58 crc kubenswrapper[4846]: I1005 08:05:58.418030 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zm54k\" (UniqueName: \"kubernetes.io/projected/ebf7ce11-32a8-4fea-9488-8f48b08991a9-kube-api-access-zm54k\") pod \"dnsmasq-dns-78dbf54cbc-c2z7x\" (UID: \"ebf7ce11-32a8-4fea-9488-8f48b08991a9\") " pod="openstack/dnsmasq-dns-78dbf54cbc-c2z7x" Oct 05 08:05:58 crc kubenswrapper[4846]: I1005 08:05:58.418220 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ebf7ce11-32a8-4fea-9488-8f48b08991a9-dns-svc\") pod \"dnsmasq-dns-78dbf54cbc-c2z7x\" (UID: \"ebf7ce11-32a8-4fea-9488-8f48b08991a9\") " pod="openstack/dnsmasq-dns-78dbf54cbc-c2z7x" Oct 05 08:05:58 crc kubenswrapper[4846]: I1005 08:05:58.419331 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ebf7ce11-32a8-4fea-9488-8f48b08991a9-dns-svc\") pod \"dnsmasq-dns-78dbf54cbc-c2z7x\" (UID: \"ebf7ce11-32a8-4fea-9488-8f48b08991a9\") " pod="openstack/dnsmasq-dns-78dbf54cbc-c2z7x" Oct 05 08:05:58 crc kubenswrapper[4846]: I1005 08:05:58.421137 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ebf7ce11-32a8-4fea-9488-8f48b08991a9-config\") pod \"dnsmasq-dns-78dbf54cbc-c2z7x\" (UID: \"ebf7ce11-32a8-4fea-9488-8f48b08991a9\") " pod="openstack/dnsmasq-dns-78dbf54cbc-c2z7x" Oct 05 08:05:58 crc kubenswrapper[4846]: I1005 08:05:58.439083 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6cfd7b4c45-gjjg8"] Oct 05 08:05:58 crc kubenswrapper[4846]: I1005 08:05:58.480831 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zm54k\" (UniqueName: \"kubernetes.io/projected/ebf7ce11-32a8-4fea-9488-8f48b08991a9-kube-api-access-zm54k\") pod \"dnsmasq-dns-78dbf54cbc-c2z7x\" (UID: \"ebf7ce11-32a8-4fea-9488-8f48b08991a9\") " pod="openstack/dnsmasq-dns-78dbf54cbc-c2z7x" Oct 05 08:05:58 crc kubenswrapper[4846]: I1005 08:05:58.520154 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6hslb\" (UniqueName: \"kubernetes.io/projected/e4bbd7e3-1e67-4145-943d-fe911ac8f3ba-kube-api-access-6hslb\") pod \"dnsmasq-dns-6cfd7b4c45-gjjg8\" (UID: \"e4bbd7e3-1e67-4145-943d-fe911ac8f3ba\") " pod="openstack/dnsmasq-dns-6cfd7b4c45-gjjg8" Oct 05 08:05:58 crc kubenswrapper[4846]: I1005 08:05:58.520252 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e4bbd7e3-1e67-4145-943d-fe911ac8f3ba-dns-svc\") pod \"dnsmasq-dns-6cfd7b4c45-gjjg8\" (UID: \"e4bbd7e3-1e67-4145-943d-fe911ac8f3ba\") " pod="openstack/dnsmasq-dns-6cfd7b4c45-gjjg8" Oct 05 08:05:58 crc kubenswrapper[4846]: I1005 08:05:58.520315 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e4bbd7e3-1e67-4145-943d-fe911ac8f3ba-config\") pod \"dnsmasq-dns-6cfd7b4c45-gjjg8\" (UID: \"e4bbd7e3-1e67-4145-943d-fe911ac8f3ba\") " pod="openstack/dnsmasq-dns-6cfd7b4c45-gjjg8" Oct 05 08:05:58 crc kubenswrapper[4846]: I1005 08:05:58.622874 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e4bbd7e3-1e67-4145-943d-fe911ac8f3ba-dns-svc\") pod \"dnsmasq-dns-6cfd7b4c45-gjjg8\" (UID: \"e4bbd7e3-1e67-4145-943d-fe911ac8f3ba\") " pod="openstack/dnsmasq-dns-6cfd7b4c45-gjjg8" Oct 05 08:05:58 crc kubenswrapper[4846]: I1005 08:05:58.622989 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e4bbd7e3-1e67-4145-943d-fe911ac8f3ba-config\") pod \"dnsmasq-dns-6cfd7b4c45-gjjg8\" (UID: \"e4bbd7e3-1e67-4145-943d-fe911ac8f3ba\") " pod="openstack/dnsmasq-dns-6cfd7b4c45-gjjg8" Oct 05 08:05:58 crc kubenswrapper[4846]: I1005 08:05:58.623062 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6hslb\" (UniqueName: \"kubernetes.io/projected/e4bbd7e3-1e67-4145-943d-fe911ac8f3ba-kube-api-access-6hslb\") pod \"dnsmasq-dns-6cfd7b4c45-gjjg8\" (UID: \"e4bbd7e3-1e67-4145-943d-fe911ac8f3ba\") " pod="openstack/dnsmasq-dns-6cfd7b4c45-gjjg8" Oct 05 08:05:58 crc kubenswrapper[4846]: I1005 08:05:58.624555 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e4bbd7e3-1e67-4145-943d-fe911ac8f3ba-dns-svc\") pod \"dnsmasq-dns-6cfd7b4c45-gjjg8\" (UID: \"e4bbd7e3-1e67-4145-943d-fe911ac8f3ba\") " pod="openstack/dnsmasq-dns-6cfd7b4c45-gjjg8" Oct 05 08:05:58 crc kubenswrapper[4846]: I1005 08:05:58.625062 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e4bbd7e3-1e67-4145-943d-fe911ac8f3ba-config\") pod \"dnsmasq-dns-6cfd7b4c45-gjjg8\" (UID: \"e4bbd7e3-1e67-4145-943d-fe911ac8f3ba\") " pod="openstack/dnsmasq-dns-6cfd7b4c45-gjjg8" Oct 05 08:05:58 crc kubenswrapper[4846]: I1005 08:05:58.646019 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6hslb\" (UniqueName: \"kubernetes.io/projected/e4bbd7e3-1e67-4145-943d-fe911ac8f3ba-kube-api-access-6hslb\") pod \"dnsmasq-dns-6cfd7b4c45-gjjg8\" (UID: \"e4bbd7e3-1e67-4145-943d-fe911ac8f3ba\") " pod="openstack/dnsmasq-dns-6cfd7b4c45-gjjg8" Oct 05 08:05:58 crc kubenswrapper[4846]: I1005 08:05:58.663460 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-55cbcdb7f9-cftk8"] Oct 05 08:05:58 crc kubenswrapper[4846]: I1005 08:05:58.743812 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6cfd7b4c45-gjjg8" Oct 05 08:05:58 crc kubenswrapper[4846]: I1005 08:05:58.767319 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dbf54cbc-c2z7x" Oct 05 08:05:58 crc kubenswrapper[4846]: I1005 08:05:58.797800 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6697c69467-27t47"] Oct 05 08:05:59 crc kubenswrapper[4846]: I1005 08:05:59.125541 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dbf54cbc-c2z7x"] Oct 05 08:05:59 crc kubenswrapper[4846]: W1005 08:05:59.148240 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podebf7ce11_32a8_4fea_9488_8f48b08991a9.slice/crio-959478198ae09db61f2d861e262d4ec5924205b6ef01a0842eed82be2acc56f4 WatchSource:0}: Error finding container 959478198ae09db61f2d861e262d4ec5924205b6ef01a0842eed82be2acc56f4: Status 404 returned error can't find the container with id 959478198ae09db61f2d861e262d4ec5924205b6ef01a0842eed82be2acc56f4 Oct 05 08:05:59 crc kubenswrapper[4846]: I1005 08:05:59.215023 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6cfd7b4c45-gjjg8"] Oct 05 08:05:59 crc kubenswrapper[4846]: W1005 08:05:59.233936 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode4bbd7e3_1e67_4145_943d_fe911ac8f3ba.slice/crio-6e5d96123807b9dd396efecc22ea1f353c095e39422045835342c86cf7641290 WatchSource:0}: Error finding container 6e5d96123807b9dd396efecc22ea1f353c095e39422045835342c86cf7641290: Status 404 returned error can't find the container with id 6e5d96123807b9dd396efecc22ea1f353c095e39422045835342c86cf7641290 Oct 05 08:05:59 crc kubenswrapper[4846]: I1005 08:05:59.254770 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Oct 05 08:05:59 crc kubenswrapper[4846]: I1005 08:05:59.256112 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 05 08:05:59 crc kubenswrapper[4846]: I1005 08:05:59.260239 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Oct 05 08:05:59 crc kubenswrapper[4846]: I1005 08:05:59.260353 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Oct 05 08:05:59 crc kubenswrapper[4846]: I1005 08:05:59.260455 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Oct 05 08:05:59 crc kubenswrapper[4846]: I1005 08:05:59.261002 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Oct 05 08:05:59 crc kubenswrapper[4846]: I1005 08:05:59.264820 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Oct 05 08:05:59 crc kubenswrapper[4846]: I1005 08:05:59.265064 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Oct 05 08:05:59 crc kubenswrapper[4846]: I1005 08:05:59.269825 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 05 08:05:59 crc kubenswrapper[4846]: I1005 08:05:59.270273 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-2k7rg" Oct 05 08:05:59 crc kubenswrapper[4846]: I1005 08:05:59.433641 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dx9rf\" (UniqueName: \"kubernetes.io/projected/0f8171a9-82f5-443a-a8fc-d42b575d498a-kube-api-access-dx9rf\") pod \"rabbitmq-server-0\" (UID: \"0f8171a9-82f5-443a-a8fc-d42b575d498a\") " pod="openstack/rabbitmq-server-0" Oct 05 08:05:59 crc kubenswrapper[4846]: I1005 08:05:59.433694 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/0f8171a9-82f5-443a-a8fc-d42b575d498a-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"0f8171a9-82f5-443a-a8fc-d42b575d498a\") " pod="openstack/rabbitmq-server-0" Oct 05 08:05:59 crc kubenswrapper[4846]: I1005 08:05:59.433722 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0f8171a9-82f5-443a-a8fc-d42b575d498a-config-data\") pod \"rabbitmq-server-0\" (UID: \"0f8171a9-82f5-443a-a8fc-d42b575d498a\") " pod="openstack/rabbitmq-server-0" Oct 05 08:05:59 crc kubenswrapper[4846]: I1005 08:05:59.433740 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/0f8171a9-82f5-443a-a8fc-d42b575d498a-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"0f8171a9-82f5-443a-a8fc-d42b575d498a\") " pod="openstack/rabbitmq-server-0" Oct 05 08:05:59 crc kubenswrapper[4846]: I1005 08:05:59.433765 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/0f8171a9-82f5-443a-a8fc-d42b575d498a-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"0f8171a9-82f5-443a-a8fc-d42b575d498a\") " pod="openstack/rabbitmq-server-0" Oct 05 08:05:59 crc kubenswrapper[4846]: I1005 08:05:59.433845 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/0f8171a9-82f5-443a-a8fc-d42b575d498a-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"0f8171a9-82f5-443a-a8fc-d42b575d498a\") " pod="openstack/rabbitmq-server-0" Oct 05 08:05:59 crc kubenswrapper[4846]: I1005 08:05:59.433917 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-c0fb7d88-e3e3-43b2-a6ce-f0d1fa614bdb\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-c0fb7d88-e3e3-43b2-a6ce-f0d1fa614bdb\") pod \"rabbitmq-server-0\" (UID: \"0f8171a9-82f5-443a-a8fc-d42b575d498a\") " pod="openstack/rabbitmq-server-0" Oct 05 08:05:59 crc kubenswrapper[4846]: I1005 08:05:59.433960 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/0f8171a9-82f5-443a-a8fc-d42b575d498a-server-conf\") pod \"rabbitmq-server-0\" (UID: \"0f8171a9-82f5-443a-a8fc-d42b575d498a\") " pod="openstack/rabbitmq-server-0" Oct 05 08:05:59 crc kubenswrapper[4846]: I1005 08:05:59.433981 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/0f8171a9-82f5-443a-a8fc-d42b575d498a-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"0f8171a9-82f5-443a-a8fc-d42b575d498a\") " pod="openstack/rabbitmq-server-0" Oct 05 08:05:59 crc kubenswrapper[4846]: I1005 08:05:59.434009 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/0f8171a9-82f5-443a-a8fc-d42b575d498a-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"0f8171a9-82f5-443a-a8fc-d42b575d498a\") " pod="openstack/rabbitmq-server-0" Oct 05 08:05:59 crc kubenswrapper[4846]: I1005 08:05:59.434034 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/0f8171a9-82f5-443a-a8fc-d42b575d498a-pod-info\") pod \"rabbitmq-server-0\" (UID: \"0f8171a9-82f5-443a-a8fc-d42b575d498a\") " pod="openstack/rabbitmq-server-0" Oct 05 08:05:59 crc kubenswrapper[4846]: I1005 08:05:59.466208 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6697c69467-27t47" event={"ID":"037a0419-1e47-429c-ac6f-e2fe6e7db5d9","Type":"ContainerStarted","Data":"287993d9589773357eb676da45a84796bcc861e31aeaa91e62382c72f29f0a10"} Oct 05 08:05:59 crc kubenswrapper[4846]: I1005 08:05:59.468143 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55cbcdb7f9-cftk8" event={"ID":"d6d6783a-7816-402e-b38e-935ffcb7314e","Type":"ContainerStarted","Data":"8aea79dfaa6807aa570d88e3076b71c703c9d3cc6ebcd51f9235172a2c11faf6"} Oct 05 08:05:59 crc kubenswrapper[4846]: I1005 08:05:59.469740 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6cfd7b4c45-gjjg8" event={"ID":"e4bbd7e3-1e67-4145-943d-fe911ac8f3ba","Type":"ContainerStarted","Data":"6e5d96123807b9dd396efecc22ea1f353c095e39422045835342c86cf7641290"} Oct 05 08:05:59 crc kubenswrapper[4846]: I1005 08:05:59.472215 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dbf54cbc-c2z7x" event={"ID":"ebf7ce11-32a8-4fea-9488-8f48b08991a9","Type":"ContainerStarted","Data":"959478198ae09db61f2d861e262d4ec5924205b6ef01a0842eed82be2acc56f4"} Oct 05 08:05:59 crc kubenswrapper[4846]: I1005 08:05:59.534984 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dx9rf\" (UniqueName: \"kubernetes.io/projected/0f8171a9-82f5-443a-a8fc-d42b575d498a-kube-api-access-dx9rf\") pod \"rabbitmq-server-0\" (UID: \"0f8171a9-82f5-443a-a8fc-d42b575d498a\") " pod="openstack/rabbitmq-server-0" Oct 05 08:05:59 crc kubenswrapper[4846]: I1005 08:05:59.535057 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/0f8171a9-82f5-443a-a8fc-d42b575d498a-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"0f8171a9-82f5-443a-a8fc-d42b575d498a\") " pod="openstack/rabbitmq-server-0" Oct 05 08:05:59 crc kubenswrapper[4846]: I1005 08:05:59.535107 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0f8171a9-82f5-443a-a8fc-d42b575d498a-config-data\") pod \"rabbitmq-server-0\" (UID: \"0f8171a9-82f5-443a-a8fc-d42b575d498a\") " pod="openstack/rabbitmq-server-0" Oct 05 08:05:59 crc kubenswrapper[4846]: I1005 08:05:59.535130 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/0f8171a9-82f5-443a-a8fc-d42b575d498a-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"0f8171a9-82f5-443a-a8fc-d42b575d498a\") " pod="openstack/rabbitmq-server-0" Oct 05 08:05:59 crc kubenswrapper[4846]: I1005 08:05:59.535147 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/0f8171a9-82f5-443a-a8fc-d42b575d498a-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"0f8171a9-82f5-443a-a8fc-d42b575d498a\") " pod="openstack/rabbitmq-server-0" Oct 05 08:05:59 crc kubenswrapper[4846]: I1005 08:05:59.535204 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/0f8171a9-82f5-443a-a8fc-d42b575d498a-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"0f8171a9-82f5-443a-a8fc-d42b575d498a\") " pod="openstack/rabbitmq-server-0" Oct 05 08:05:59 crc kubenswrapper[4846]: I1005 08:05:59.535243 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-c0fb7d88-e3e3-43b2-a6ce-f0d1fa614bdb\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-c0fb7d88-e3e3-43b2-a6ce-f0d1fa614bdb\") pod \"rabbitmq-server-0\" (UID: \"0f8171a9-82f5-443a-a8fc-d42b575d498a\") " pod="openstack/rabbitmq-server-0" Oct 05 08:05:59 crc kubenswrapper[4846]: I1005 08:05:59.535311 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/0f8171a9-82f5-443a-a8fc-d42b575d498a-server-conf\") pod \"rabbitmq-server-0\" (UID: \"0f8171a9-82f5-443a-a8fc-d42b575d498a\") " pod="openstack/rabbitmq-server-0" Oct 05 08:05:59 crc kubenswrapper[4846]: I1005 08:05:59.535330 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/0f8171a9-82f5-443a-a8fc-d42b575d498a-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"0f8171a9-82f5-443a-a8fc-d42b575d498a\") " pod="openstack/rabbitmq-server-0" Oct 05 08:05:59 crc kubenswrapper[4846]: I1005 08:05:59.535408 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/0f8171a9-82f5-443a-a8fc-d42b575d498a-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"0f8171a9-82f5-443a-a8fc-d42b575d498a\") " pod="openstack/rabbitmq-server-0" Oct 05 08:05:59 crc kubenswrapper[4846]: I1005 08:05:59.535452 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/0f8171a9-82f5-443a-a8fc-d42b575d498a-pod-info\") pod \"rabbitmq-server-0\" (UID: \"0f8171a9-82f5-443a-a8fc-d42b575d498a\") " pod="openstack/rabbitmq-server-0" Oct 05 08:05:59 crc kubenswrapper[4846]: I1005 08:05:59.536910 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/0f8171a9-82f5-443a-a8fc-d42b575d498a-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"0f8171a9-82f5-443a-a8fc-d42b575d498a\") " pod="openstack/rabbitmq-server-0" Oct 05 08:05:59 crc kubenswrapper[4846]: I1005 08:05:59.537160 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/0f8171a9-82f5-443a-a8fc-d42b575d498a-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"0f8171a9-82f5-443a-a8fc-d42b575d498a\") " pod="openstack/rabbitmq-server-0" Oct 05 08:05:59 crc kubenswrapper[4846]: I1005 08:05:59.537648 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/0f8171a9-82f5-443a-a8fc-d42b575d498a-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"0f8171a9-82f5-443a-a8fc-d42b575d498a\") " pod="openstack/rabbitmq-server-0" Oct 05 08:05:59 crc kubenswrapper[4846]: I1005 08:05:59.538336 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0f8171a9-82f5-443a-a8fc-d42b575d498a-config-data\") pod \"rabbitmq-server-0\" (UID: \"0f8171a9-82f5-443a-a8fc-d42b575d498a\") " pod="openstack/rabbitmq-server-0" Oct 05 08:05:59 crc kubenswrapper[4846]: I1005 08:05:59.538791 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/0f8171a9-82f5-443a-a8fc-d42b575d498a-server-conf\") pod \"rabbitmq-server-0\" (UID: \"0f8171a9-82f5-443a-a8fc-d42b575d498a\") " pod="openstack/rabbitmq-server-0" Oct 05 08:05:59 crc kubenswrapper[4846]: I1005 08:05:59.540168 4846 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 05 08:05:59 crc kubenswrapper[4846]: I1005 08:05:59.540218 4846 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-c0fb7d88-e3e3-43b2-a6ce-f0d1fa614bdb\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-c0fb7d88-e3e3-43b2-a6ce-f0d1fa614bdb\") pod \"rabbitmq-server-0\" (UID: \"0f8171a9-82f5-443a-a8fc-d42b575d498a\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/d6d150f95ef01577cee9c24c1b65571c71a6ebd80cf8fb72945050d88d52ce8c/globalmount\"" pod="openstack/rabbitmq-server-0" Oct 05 08:05:59 crc kubenswrapper[4846]: I1005 08:05:59.548880 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/0f8171a9-82f5-443a-a8fc-d42b575d498a-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"0f8171a9-82f5-443a-a8fc-d42b575d498a\") " pod="openstack/rabbitmq-server-0" Oct 05 08:05:59 crc kubenswrapper[4846]: I1005 08:05:59.549295 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/0f8171a9-82f5-443a-a8fc-d42b575d498a-pod-info\") pod \"rabbitmq-server-0\" (UID: \"0f8171a9-82f5-443a-a8fc-d42b575d498a\") " pod="openstack/rabbitmq-server-0" Oct 05 08:05:59 crc kubenswrapper[4846]: I1005 08:05:59.550069 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/0f8171a9-82f5-443a-a8fc-d42b575d498a-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"0f8171a9-82f5-443a-a8fc-d42b575d498a\") " pod="openstack/rabbitmq-server-0" Oct 05 08:05:59 crc kubenswrapper[4846]: I1005 08:05:59.568449 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dx9rf\" (UniqueName: \"kubernetes.io/projected/0f8171a9-82f5-443a-a8fc-d42b575d498a-kube-api-access-dx9rf\") pod \"rabbitmq-server-0\" (UID: \"0f8171a9-82f5-443a-a8fc-d42b575d498a\") " pod="openstack/rabbitmq-server-0" Oct 05 08:05:59 crc kubenswrapper[4846]: I1005 08:05:59.575994 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 05 08:05:59 crc kubenswrapper[4846]: I1005 08:05:59.580036 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 05 08:05:59 crc kubenswrapper[4846]: I1005 08:05:59.580393 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/0f8171a9-82f5-443a-a8fc-d42b575d498a-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"0f8171a9-82f5-443a-a8fc-d42b575d498a\") " pod="openstack/rabbitmq-server-0" Oct 05 08:05:59 crc kubenswrapper[4846]: I1005 08:05:59.586449 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Oct 05 08:05:59 crc kubenswrapper[4846]: I1005 08:05:59.586606 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Oct 05 08:05:59 crc kubenswrapper[4846]: I1005 08:05:59.586770 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Oct 05 08:05:59 crc kubenswrapper[4846]: I1005 08:05:59.586893 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Oct 05 08:05:59 crc kubenswrapper[4846]: I1005 08:05:59.586996 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Oct 05 08:05:59 crc kubenswrapper[4846]: I1005 08:05:59.587110 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Oct 05 08:05:59 crc kubenswrapper[4846]: I1005 08:05:59.587268 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-r9ldz" Oct 05 08:05:59 crc kubenswrapper[4846]: I1005 08:05:59.594551 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 05 08:05:59 crc kubenswrapper[4846]: I1005 08:05:59.623096 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-c0fb7d88-e3e3-43b2-a6ce-f0d1fa614bdb\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-c0fb7d88-e3e3-43b2-a6ce-f0d1fa614bdb\") pod \"rabbitmq-server-0\" (UID: \"0f8171a9-82f5-443a-a8fc-d42b575d498a\") " pod="openstack/rabbitmq-server-0" Oct 05 08:05:59 crc kubenswrapper[4846]: I1005 08:05:59.738411 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-0b1f215a-31af-4eb1-a3f9-ccbfe14235d1\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-0b1f215a-31af-4eb1-a3f9-ccbfe14235d1\") pod \"rabbitmq-cell1-server-0\" (UID: \"a288716c-b683-47c0-bf44-c0bbc997da97\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 08:05:59 crc kubenswrapper[4846]: I1005 08:05:59.739239 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/a288716c-b683-47c0-bf44-c0bbc997da97-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"a288716c-b683-47c0-bf44-c0bbc997da97\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 08:05:59 crc kubenswrapper[4846]: I1005 08:05:59.739286 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a288716c-b683-47c0-bf44-c0bbc997da97-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"a288716c-b683-47c0-bf44-c0bbc997da97\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 08:05:59 crc kubenswrapper[4846]: I1005 08:05:59.739342 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/a288716c-b683-47c0-bf44-c0bbc997da97-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"a288716c-b683-47c0-bf44-c0bbc997da97\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 08:05:59 crc kubenswrapper[4846]: I1005 08:05:59.739383 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bdwsq\" (UniqueName: \"kubernetes.io/projected/a288716c-b683-47c0-bf44-c0bbc997da97-kube-api-access-bdwsq\") pod \"rabbitmq-cell1-server-0\" (UID: \"a288716c-b683-47c0-bf44-c0bbc997da97\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 08:05:59 crc kubenswrapper[4846]: I1005 08:05:59.739448 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/a288716c-b683-47c0-bf44-c0bbc997da97-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"a288716c-b683-47c0-bf44-c0bbc997da97\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 08:05:59 crc kubenswrapper[4846]: I1005 08:05:59.739818 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/a288716c-b683-47c0-bf44-c0bbc997da97-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"a288716c-b683-47c0-bf44-c0bbc997da97\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 08:05:59 crc kubenswrapper[4846]: I1005 08:05:59.739917 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/a288716c-b683-47c0-bf44-c0bbc997da97-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"a288716c-b683-47c0-bf44-c0bbc997da97\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 08:05:59 crc kubenswrapper[4846]: I1005 08:05:59.739976 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/a288716c-b683-47c0-bf44-c0bbc997da97-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"a288716c-b683-47c0-bf44-c0bbc997da97\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 08:05:59 crc kubenswrapper[4846]: I1005 08:05:59.740042 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/a288716c-b683-47c0-bf44-c0bbc997da97-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"a288716c-b683-47c0-bf44-c0bbc997da97\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 08:05:59 crc kubenswrapper[4846]: I1005 08:05:59.740241 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/a288716c-b683-47c0-bf44-c0bbc997da97-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"a288716c-b683-47c0-bf44-c0bbc997da97\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 08:05:59 crc kubenswrapper[4846]: I1005 08:05:59.845688 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/a288716c-b683-47c0-bf44-c0bbc997da97-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"a288716c-b683-47c0-bf44-c0bbc997da97\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 08:05:59 crc kubenswrapper[4846]: I1005 08:05:59.845790 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-0b1f215a-31af-4eb1-a3f9-ccbfe14235d1\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-0b1f215a-31af-4eb1-a3f9-ccbfe14235d1\") pod \"rabbitmq-cell1-server-0\" (UID: \"a288716c-b683-47c0-bf44-c0bbc997da97\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 08:05:59 crc kubenswrapper[4846]: I1005 08:05:59.845831 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/a288716c-b683-47c0-bf44-c0bbc997da97-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"a288716c-b683-47c0-bf44-c0bbc997da97\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 08:05:59 crc kubenswrapper[4846]: I1005 08:05:59.845855 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a288716c-b683-47c0-bf44-c0bbc997da97-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"a288716c-b683-47c0-bf44-c0bbc997da97\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 08:05:59 crc kubenswrapper[4846]: I1005 08:05:59.845902 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/a288716c-b683-47c0-bf44-c0bbc997da97-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"a288716c-b683-47c0-bf44-c0bbc997da97\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 08:05:59 crc kubenswrapper[4846]: I1005 08:05:59.845965 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bdwsq\" (UniqueName: \"kubernetes.io/projected/a288716c-b683-47c0-bf44-c0bbc997da97-kube-api-access-bdwsq\") pod \"rabbitmq-cell1-server-0\" (UID: \"a288716c-b683-47c0-bf44-c0bbc997da97\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 08:05:59 crc kubenswrapper[4846]: I1005 08:05:59.846005 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/a288716c-b683-47c0-bf44-c0bbc997da97-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"a288716c-b683-47c0-bf44-c0bbc997da97\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 08:05:59 crc kubenswrapper[4846]: I1005 08:05:59.846134 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/a288716c-b683-47c0-bf44-c0bbc997da97-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"a288716c-b683-47c0-bf44-c0bbc997da97\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 08:05:59 crc kubenswrapper[4846]: I1005 08:05:59.846162 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/a288716c-b683-47c0-bf44-c0bbc997da97-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"a288716c-b683-47c0-bf44-c0bbc997da97\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 08:05:59 crc kubenswrapper[4846]: I1005 08:05:59.846233 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/a288716c-b683-47c0-bf44-c0bbc997da97-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"a288716c-b683-47c0-bf44-c0bbc997da97\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 08:05:59 crc kubenswrapper[4846]: I1005 08:05:59.846279 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/a288716c-b683-47c0-bf44-c0bbc997da97-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"a288716c-b683-47c0-bf44-c0bbc997da97\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 08:05:59 crc kubenswrapper[4846]: I1005 08:05:59.847768 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/a288716c-b683-47c0-bf44-c0bbc997da97-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"a288716c-b683-47c0-bf44-c0bbc997da97\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 08:05:59 crc kubenswrapper[4846]: I1005 08:05:59.849781 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/a288716c-b683-47c0-bf44-c0bbc997da97-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"a288716c-b683-47c0-bf44-c0bbc997da97\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 08:05:59 crc kubenswrapper[4846]: I1005 08:05:59.854840 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/a288716c-b683-47c0-bf44-c0bbc997da97-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"a288716c-b683-47c0-bf44-c0bbc997da97\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 08:05:59 crc kubenswrapper[4846]: I1005 08:05:59.856165 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a288716c-b683-47c0-bf44-c0bbc997da97-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"a288716c-b683-47c0-bf44-c0bbc997da97\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 08:05:59 crc kubenswrapper[4846]: I1005 08:05:59.856579 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/a288716c-b683-47c0-bf44-c0bbc997da97-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"a288716c-b683-47c0-bf44-c0bbc997da97\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 08:05:59 crc kubenswrapper[4846]: I1005 08:05:59.856719 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/a288716c-b683-47c0-bf44-c0bbc997da97-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"a288716c-b683-47c0-bf44-c0bbc997da97\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 08:05:59 crc kubenswrapper[4846]: I1005 08:05:59.859353 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/a288716c-b683-47c0-bf44-c0bbc997da97-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"a288716c-b683-47c0-bf44-c0bbc997da97\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 08:05:59 crc kubenswrapper[4846]: I1005 08:05:59.860027 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/a288716c-b683-47c0-bf44-c0bbc997da97-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"a288716c-b683-47c0-bf44-c0bbc997da97\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 08:05:59 crc kubenswrapper[4846]: I1005 08:05:59.860537 4846 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 05 08:05:59 crc kubenswrapper[4846]: I1005 08:05:59.860612 4846 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-0b1f215a-31af-4eb1-a3f9-ccbfe14235d1\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-0b1f215a-31af-4eb1-a3f9-ccbfe14235d1\") pod \"rabbitmq-cell1-server-0\" (UID: \"a288716c-b683-47c0-bf44-c0bbc997da97\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/46422321c9fc4982dc959ec2b6aca5699513296fc205359a8d3e45d1d4b35fe5/globalmount\"" pod="openstack/rabbitmq-cell1-server-0" Oct 05 08:05:59 crc kubenswrapper[4846]: I1005 08:05:59.861315 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/a288716c-b683-47c0-bf44-c0bbc997da97-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"a288716c-b683-47c0-bf44-c0bbc997da97\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 08:05:59 crc kubenswrapper[4846]: I1005 08:05:59.874328 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bdwsq\" (UniqueName: \"kubernetes.io/projected/a288716c-b683-47c0-bf44-c0bbc997da97-kube-api-access-bdwsq\") pod \"rabbitmq-cell1-server-0\" (UID: \"a288716c-b683-47c0-bf44-c0bbc997da97\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 08:05:59 crc kubenswrapper[4846]: I1005 08:05:59.892109 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 05 08:05:59 crc kubenswrapper[4846]: I1005 08:05:59.896541 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-0b1f215a-31af-4eb1-a3f9-ccbfe14235d1\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-0b1f215a-31af-4eb1-a3f9-ccbfe14235d1\") pod \"rabbitmq-cell1-server-0\" (UID: \"a288716c-b683-47c0-bf44-c0bbc997da97\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 08:05:59 crc kubenswrapper[4846]: I1005 08:05:59.942763 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 05 08:06:00 crc kubenswrapper[4846]: I1005 08:06:00.381699 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-galera-0"] Oct 05 08:06:00 crc kubenswrapper[4846]: I1005 08:06:00.382967 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Oct 05 08:06:00 crc kubenswrapper[4846]: I1005 08:06:00.388666 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Oct 05 08:06:00 crc kubenswrapper[4846]: I1005 08:06:00.389106 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-dockercfg-xjxql" Oct 05 08:06:00 crc kubenswrapper[4846]: I1005 08:06:00.389388 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-scripts" Oct 05 08:06:00 crc kubenswrapper[4846]: I1005 08:06:00.389407 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-svc" Oct 05 08:06:00 crc kubenswrapper[4846]: I1005 08:06:00.389441 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config-data" Oct 05 08:06:00 crc kubenswrapper[4846]: I1005 08:06:00.398624 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"combined-ca-bundle" Oct 05 08:06:00 crc kubenswrapper[4846]: I1005 08:06:00.404676 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Oct 05 08:06:00 crc kubenswrapper[4846]: I1005 08:06:00.557630 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/771f44c5-2cc5-431e-9569-eb7da776756d-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"771f44c5-2cc5-431e-9569-eb7da776756d\") " pod="openstack/openstack-galera-0" Oct 05 08:06:00 crc kubenswrapper[4846]: I1005 08:06:00.557676 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/771f44c5-2cc5-431e-9569-eb7da776756d-config-data-generated\") pod \"openstack-galera-0\" (UID: \"771f44c5-2cc5-431e-9569-eb7da776756d\") " pod="openstack/openstack-galera-0" Oct 05 08:06:00 crc kubenswrapper[4846]: I1005 08:06:00.557708 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/771f44c5-2cc5-431e-9569-eb7da776756d-secrets\") pod \"openstack-galera-0\" (UID: \"771f44c5-2cc5-431e-9569-eb7da776756d\") " pod="openstack/openstack-galera-0" Oct 05 08:06:00 crc kubenswrapper[4846]: I1005 08:06:00.557730 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-a5a7cf00-4e61-4a6b-a6c8-1f3dd3b5a859\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-a5a7cf00-4e61-4a6b-a6c8-1f3dd3b5a859\") pod \"openstack-galera-0\" (UID: \"771f44c5-2cc5-431e-9569-eb7da776756d\") " pod="openstack/openstack-galera-0" Oct 05 08:06:00 crc kubenswrapper[4846]: I1005 08:06:00.557749 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/771f44c5-2cc5-431e-9569-eb7da776756d-config-data-default\") pod \"openstack-galera-0\" (UID: \"771f44c5-2cc5-431e-9569-eb7da776756d\") " pod="openstack/openstack-galera-0" Oct 05 08:06:00 crc kubenswrapper[4846]: I1005 08:06:00.557782 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/771f44c5-2cc5-431e-9569-eb7da776756d-kolla-config\") pod \"openstack-galera-0\" (UID: \"771f44c5-2cc5-431e-9569-eb7da776756d\") " pod="openstack/openstack-galera-0" Oct 05 08:06:00 crc kubenswrapper[4846]: I1005 08:06:00.557822 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w8p5s\" (UniqueName: \"kubernetes.io/projected/771f44c5-2cc5-431e-9569-eb7da776756d-kube-api-access-w8p5s\") pod \"openstack-galera-0\" (UID: \"771f44c5-2cc5-431e-9569-eb7da776756d\") " pod="openstack/openstack-galera-0" Oct 05 08:06:00 crc kubenswrapper[4846]: I1005 08:06:00.557843 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/771f44c5-2cc5-431e-9569-eb7da776756d-operator-scripts\") pod \"openstack-galera-0\" (UID: \"771f44c5-2cc5-431e-9569-eb7da776756d\") " pod="openstack/openstack-galera-0" Oct 05 08:06:00 crc kubenswrapper[4846]: I1005 08:06:00.557879 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/771f44c5-2cc5-431e-9569-eb7da776756d-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"771f44c5-2cc5-431e-9569-eb7da776756d\") " pod="openstack/openstack-galera-0" Oct 05 08:06:00 crc kubenswrapper[4846]: I1005 08:06:00.660534 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/771f44c5-2cc5-431e-9569-eb7da776756d-operator-scripts\") pod \"openstack-galera-0\" (UID: \"771f44c5-2cc5-431e-9569-eb7da776756d\") " pod="openstack/openstack-galera-0" Oct 05 08:06:00 crc kubenswrapper[4846]: I1005 08:06:00.660605 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/771f44c5-2cc5-431e-9569-eb7da776756d-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"771f44c5-2cc5-431e-9569-eb7da776756d\") " pod="openstack/openstack-galera-0" Oct 05 08:06:00 crc kubenswrapper[4846]: I1005 08:06:00.660662 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/771f44c5-2cc5-431e-9569-eb7da776756d-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"771f44c5-2cc5-431e-9569-eb7da776756d\") " pod="openstack/openstack-galera-0" Oct 05 08:06:00 crc kubenswrapper[4846]: I1005 08:06:00.660683 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/771f44c5-2cc5-431e-9569-eb7da776756d-config-data-generated\") pod \"openstack-galera-0\" (UID: \"771f44c5-2cc5-431e-9569-eb7da776756d\") " pod="openstack/openstack-galera-0" Oct 05 08:06:00 crc kubenswrapper[4846]: I1005 08:06:00.660718 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/771f44c5-2cc5-431e-9569-eb7da776756d-secrets\") pod \"openstack-galera-0\" (UID: \"771f44c5-2cc5-431e-9569-eb7da776756d\") " pod="openstack/openstack-galera-0" Oct 05 08:06:00 crc kubenswrapper[4846]: I1005 08:06:00.660747 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-a5a7cf00-4e61-4a6b-a6c8-1f3dd3b5a859\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-a5a7cf00-4e61-4a6b-a6c8-1f3dd3b5a859\") pod \"openstack-galera-0\" (UID: \"771f44c5-2cc5-431e-9569-eb7da776756d\") " pod="openstack/openstack-galera-0" Oct 05 08:06:00 crc kubenswrapper[4846]: I1005 08:06:00.661152 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/771f44c5-2cc5-431e-9569-eb7da776756d-config-data-default\") pod \"openstack-galera-0\" (UID: \"771f44c5-2cc5-431e-9569-eb7da776756d\") " pod="openstack/openstack-galera-0" Oct 05 08:06:00 crc kubenswrapper[4846]: I1005 08:06:00.661299 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/771f44c5-2cc5-431e-9569-eb7da776756d-kolla-config\") pod \"openstack-galera-0\" (UID: \"771f44c5-2cc5-431e-9569-eb7da776756d\") " pod="openstack/openstack-galera-0" Oct 05 08:06:00 crc kubenswrapper[4846]: I1005 08:06:00.661372 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w8p5s\" (UniqueName: \"kubernetes.io/projected/771f44c5-2cc5-431e-9569-eb7da776756d-kube-api-access-w8p5s\") pod \"openstack-galera-0\" (UID: \"771f44c5-2cc5-431e-9569-eb7da776756d\") " pod="openstack/openstack-galera-0" Oct 05 08:06:00 crc kubenswrapper[4846]: I1005 08:06:00.662905 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/771f44c5-2cc5-431e-9569-eb7da776756d-config-data-default\") pod \"openstack-galera-0\" (UID: \"771f44c5-2cc5-431e-9569-eb7da776756d\") " pod="openstack/openstack-galera-0" Oct 05 08:06:00 crc kubenswrapper[4846]: I1005 08:06:00.664196 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/771f44c5-2cc5-431e-9569-eb7da776756d-kolla-config\") pod \"openstack-galera-0\" (UID: \"771f44c5-2cc5-431e-9569-eb7da776756d\") " pod="openstack/openstack-galera-0" Oct 05 08:06:00 crc kubenswrapper[4846]: I1005 08:06:00.664763 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/771f44c5-2cc5-431e-9569-eb7da776756d-operator-scripts\") pod \"openstack-galera-0\" (UID: \"771f44c5-2cc5-431e-9569-eb7da776756d\") " pod="openstack/openstack-galera-0" Oct 05 08:06:00 crc kubenswrapper[4846]: I1005 08:06:00.665461 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/771f44c5-2cc5-431e-9569-eb7da776756d-config-data-generated\") pod \"openstack-galera-0\" (UID: \"771f44c5-2cc5-431e-9569-eb7da776756d\") " pod="openstack/openstack-galera-0" Oct 05 08:06:00 crc kubenswrapper[4846]: I1005 08:06:00.666939 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/771f44c5-2cc5-431e-9569-eb7da776756d-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"771f44c5-2cc5-431e-9569-eb7da776756d\") " pod="openstack/openstack-galera-0" Oct 05 08:06:00 crc kubenswrapper[4846]: I1005 08:06:00.667444 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/771f44c5-2cc5-431e-9569-eb7da776756d-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"771f44c5-2cc5-431e-9569-eb7da776756d\") " pod="openstack/openstack-galera-0" Oct 05 08:06:00 crc kubenswrapper[4846]: I1005 08:06:00.671654 4846 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 05 08:06:00 crc kubenswrapper[4846]: I1005 08:06:00.671680 4846 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-a5a7cf00-4e61-4a6b-a6c8-1f3dd3b5a859\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-a5a7cf00-4e61-4a6b-a6c8-1f3dd3b5a859\") pod \"openstack-galera-0\" (UID: \"771f44c5-2cc5-431e-9569-eb7da776756d\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/9c9504f336d65034eab2633d4c46d115cf261a82bd44994e9f3bd31aa128ef7d/globalmount\"" pod="openstack/openstack-galera-0" Oct 05 08:06:00 crc kubenswrapper[4846]: I1005 08:06:00.674053 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/771f44c5-2cc5-431e-9569-eb7da776756d-secrets\") pod \"openstack-galera-0\" (UID: \"771f44c5-2cc5-431e-9569-eb7da776756d\") " pod="openstack/openstack-galera-0" Oct 05 08:06:00 crc kubenswrapper[4846]: I1005 08:06:00.697880 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w8p5s\" (UniqueName: \"kubernetes.io/projected/771f44c5-2cc5-431e-9569-eb7da776756d-kube-api-access-w8p5s\") pod \"openstack-galera-0\" (UID: \"771f44c5-2cc5-431e-9569-eb7da776756d\") " pod="openstack/openstack-galera-0" Oct 05 08:06:00 crc kubenswrapper[4846]: I1005 08:06:00.740707 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-a5a7cf00-4e61-4a6b-a6c8-1f3dd3b5a859\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-a5a7cf00-4e61-4a6b-a6c8-1f3dd3b5a859\") pod \"openstack-galera-0\" (UID: \"771f44c5-2cc5-431e-9569-eb7da776756d\") " pod="openstack/openstack-galera-0" Oct 05 08:06:00 crc kubenswrapper[4846]: I1005 08:06:00.879878 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 05 08:06:01 crc kubenswrapper[4846]: I1005 08:06:01.034020 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Oct 05 08:06:01 crc kubenswrapper[4846]: I1005 08:06:01.208759 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 05 08:06:01 crc kubenswrapper[4846]: I1005 08:06:01.503035 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"0f8171a9-82f5-443a-a8fc-d42b575d498a","Type":"ContainerStarted","Data":"50e4eb4686ae493476e9c60dc3ad1b3556039dd5c0694bc9a4853855b11e27be"} Oct 05 08:06:01 crc kubenswrapper[4846]: I1005 08:06:01.504443 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"a288716c-b683-47c0-bf44-c0bbc997da97","Type":"ContainerStarted","Data":"3c0f92164a082c2da547bc954edbe40f8475270017609029921e3d666bbda069"} Oct 05 08:06:01 crc kubenswrapper[4846]: I1005 08:06:01.515094 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Oct 05 08:06:01 crc kubenswrapper[4846]: W1005 08:06:01.521507 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod771f44c5_2cc5_431e_9569_eb7da776756d.slice/crio-5842da1b8e2e56632cbf7404f5b2fe29f742b978b9762834166bd4db490aa643 WatchSource:0}: Error finding container 5842da1b8e2e56632cbf7404f5b2fe29f742b978b9762834166bd4db490aa643: Status 404 returned error can't find the container with id 5842da1b8e2e56632cbf7404f5b2fe29f742b978b9762834166bd4db490aa643 Oct 05 08:06:01 crc kubenswrapper[4846]: I1005 08:06:01.797581 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-cell1-galera-0"] Oct 05 08:06:01 crc kubenswrapper[4846]: I1005 08:06:01.798759 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Oct 05 08:06:01 crc kubenswrapper[4846]: I1005 08:06:01.801381 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-cell1-dockercfg-tf8qs" Oct 05 08:06:01 crc kubenswrapper[4846]: I1005 08:06:01.802555 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-scripts" Oct 05 08:06:01 crc kubenswrapper[4846]: I1005 08:06:01.802607 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-cell1-svc" Oct 05 08:06:01 crc kubenswrapper[4846]: I1005 08:06:01.802561 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-config-data" Oct 05 08:06:01 crc kubenswrapper[4846]: I1005 08:06:01.806309 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Oct 05 08:06:01 crc kubenswrapper[4846]: I1005 08:06:01.885542 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/643edc27-5b17-45e9-9286-f74591843690-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"643edc27-5b17-45e9-9286-f74591843690\") " pod="openstack/openstack-cell1-galera-0" Oct 05 08:06:01 crc kubenswrapper[4846]: I1005 08:06:01.885647 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/643edc27-5b17-45e9-9286-f74591843690-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"643edc27-5b17-45e9-9286-f74591843690\") " pod="openstack/openstack-cell1-galera-0" Oct 05 08:06:01 crc kubenswrapper[4846]: I1005 08:06:01.885720 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/643edc27-5b17-45e9-9286-f74591843690-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"643edc27-5b17-45e9-9286-f74591843690\") " pod="openstack/openstack-cell1-galera-0" Oct 05 08:06:01 crc kubenswrapper[4846]: I1005 08:06:01.885780 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/643edc27-5b17-45e9-9286-f74591843690-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"643edc27-5b17-45e9-9286-f74591843690\") " pod="openstack/openstack-cell1-galera-0" Oct 05 08:06:01 crc kubenswrapper[4846]: I1005 08:06:01.885824 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/643edc27-5b17-45e9-9286-f74591843690-secrets\") pod \"openstack-cell1-galera-0\" (UID: \"643edc27-5b17-45e9-9286-f74591843690\") " pod="openstack/openstack-cell1-galera-0" Oct 05 08:06:01 crc kubenswrapper[4846]: I1005 08:06:01.885860 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-ae29e329-f4e2-4411-816f-7b8aeb9a6f29\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ae29e329-f4e2-4411-816f-7b8aeb9a6f29\") pod \"openstack-cell1-galera-0\" (UID: \"643edc27-5b17-45e9-9286-f74591843690\") " pod="openstack/openstack-cell1-galera-0" Oct 05 08:06:01 crc kubenswrapper[4846]: I1005 08:06:01.885911 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8vl8q\" (UniqueName: \"kubernetes.io/projected/643edc27-5b17-45e9-9286-f74591843690-kube-api-access-8vl8q\") pod \"openstack-cell1-galera-0\" (UID: \"643edc27-5b17-45e9-9286-f74591843690\") " pod="openstack/openstack-cell1-galera-0" Oct 05 08:06:01 crc kubenswrapper[4846]: I1005 08:06:01.885947 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/643edc27-5b17-45e9-9286-f74591843690-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"643edc27-5b17-45e9-9286-f74591843690\") " pod="openstack/openstack-cell1-galera-0" Oct 05 08:06:01 crc kubenswrapper[4846]: I1005 08:06:01.885998 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/643edc27-5b17-45e9-9286-f74591843690-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"643edc27-5b17-45e9-9286-f74591843690\") " pod="openstack/openstack-cell1-galera-0" Oct 05 08:06:01 crc kubenswrapper[4846]: I1005 08:06:01.988138 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/643edc27-5b17-45e9-9286-f74591843690-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"643edc27-5b17-45e9-9286-f74591843690\") " pod="openstack/openstack-cell1-galera-0" Oct 05 08:06:01 crc kubenswrapper[4846]: I1005 08:06:01.988261 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/643edc27-5b17-45e9-9286-f74591843690-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"643edc27-5b17-45e9-9286-f74591843690\") " pod="openstack/openstack-cell1-galera-0" Oct 05 08:06:01 crc kubenswrapper[4846]: I1005 08:06:01.988313 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/643edc27-5b17-45e9-9286-f74591843690-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"643edc27-5b17-45e9-9286-f74591843690\") " pod="openstack/openstack-cell1-galera-0" Oct 05 08:06:01 crc kubenswrapper[4846]: I1005 08:06:01.988330 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/643edc27-5b17-45e9-9286-f74591843690-secrets\") pod \"openstack-cell1-galera-0\" (UID: \"643edc27-5b17-45e9-9286-f74591843690\") " pod="openstack/openstack-cell1-galera-0" Oct 05 08:06:01 crc kubenswrapper[4846]: I1005 08:06:01.988400 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-ae29e329-f4e2-4411-816f-7b8aeb9a6f29\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ae29e329-f4e2-4411-816f-7b8aeb9a6f29\") pod \"openstack-cell1-galera-0\" (UID: \"643edc27-5b17-45e9-9286-f74591843690\") " pod="openstack/openstack-cell1-galera-0" Oct 05 08:06:01 crc kubenswrapper[4846]: I1005 08:06:01.988464 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8vl8q\" (UniqueName: \"kubernetes.io/projected/643edc27-5b17-45e9-9286-f74591843690-kube-api-access-8vl8q\") pod \"openstack-cell1-galera-0\" (UID: \"643edc27-5b17-45e9-9286-f74591843690\") " pod="openstack/openstack-cell1-galera-0" Oct 05 08:06:01 crc kubenswrapper[4846]: I1005 08:06:01.988518 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/643edc27-5b17-45e9-9286-f74591843690-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"643edc27-5b17-45e9-9286-f74591843690\") " pod="openstack/openstack-cell1-galera-0" Oct 05 08:06:01 crc kubenswrapper[4846]: I1005 08:06:01.988550 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/643edc27-5b17-45e9-9286-f74591843690-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"643edc27-5b17-45e9-9286-f74591843690\") " pod="openstack/openstack-cell1-galera-0" Oct 05 08:06:01 crc kubenswrapper[4846]: I1005 08:06:01.988607 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/643edc27-5b17-45e9-9286-f74591843690-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"643edc27-5b17-45e9-9286-f74591843690\") " pod="openstack/openstack-cell1-galera-0" Oct 05 08:06:01 crc kubenswrapper[4846]: I1005 08:06:01.988631 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/643edc27-5b17-45e9-9286-f74591843690-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"643edc27-5b17-45e9-9286-f74591843690\") " pod="openstack/openstack-cell1-galera-0" Oct 05 08:06:01 crc kubenswrapper[4846]: I1005 08:06:01.989428 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/643edc27-5b17-45e9-9286-f74591843690-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"643edc27-5b17-45e9-9286-f74591843690\") " pod="openstack/openstack-cell1-galera-0" Oct 05 08:06:01 crc kubenswrapper[4846]: I1005 08:06:01.990507 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/643edc27-5b17-45e9-9286-f74591843690-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"643edc27-5b17-45e9-9286-f74591843690\") " pod="openstack/openstack-cell1-galera-0" Oct 05 08:06:01 crc kubenswrapper[4846]: I1005 08:06:01.991313 4846 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 05 08:06:01 crc kubenswrapper[4846]: I1005 08:06:01.991346 4846 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-ae29e329-f4e2-4411-816f-7b8aeb9a6f29\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ae29e329-f4e2-4411-816f-7b8aeb9a6f29\") pod \"openstack-cell1-galera-0\" (UID: \"643edc27-5b17-45e9-9286-f74591843690\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/8f5ddecc1807d191297313dc6af265552bc4f7373ce1f0cb53415cbf4466274f/globalmount\"" pod="openstack/openstack-cell1-galera-0" Oct 05 08:06:01 crc kubenswrapper[4846]: I1005 08:06:01.991329 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/643edc27-5b17-45e9-9286-f74591843690-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"643edc27-5b17-45e9-9286-f74591843690\") " pod="openstack/openstack-cell1-galera-0" Oct 05 08:06:01 crc kubenswrapper[4846]: I1005 08:06:01.994935 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/643edc27-5b17-45e9-9286-f74591843690-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"643edc27-5b17-45e9-9286-f74591843690\") " pod="openstack/openstack-cell1-galera-0" Oct 05 08:06:01 crc kubenswrapper[4846]: I1005 08:06:01.995041 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/643edc27-5b17-45e9-9286-f74591843690-secrets\") pod \"openstack-cell1-galera-0\" (UID: \"643edc27-5b17-45e9-9286-f74591843690\") " pod="openstack/openstack-cell1-galera-0" Oct 05 08:06:01 crc kubenswrapper[4846]: I1005 08:06:01.998866 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/643edc27-5b17-45e9-9286-f74591843690-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"643edc27-5b17-45e9-9286-f74591843690\") " pod="openstack/openstack-cell1-galera-0" Oct 05 08:06:02 crc kubenswrapper[4846]: I1005 08:06:02.004709 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8vl8q\" (UniqueName: \"kubernetes.io/projected/643edc27-5b17-45e9-9286-f74591843690-kube-api-access-8vl8q\") pod \"openstack-cell1-galera-0\" (UID: \"643edc27-5b17-45e9-9286-f74591843690\") " pod="openstack/openstack-cell1-galera-0" Oct 05 08:06:02 crc kubenswrapper[4846]: I1005 08:06:02.022063 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-ae29e329-f4e2-4411-816f-7b8aeb9a6f29\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ae29e329-f4e2-4411-816f-7b8aeb9a6f29\") pod \"openstack-cell1-galera-0\" (UID: \"643edc27-5b17-45e9-9286-f74591843690\") " pod="openstack/openstack-cell1-galera-0" Oct 05 08:06:02 crc kubenswrapper[4846]: I1005 08:06:02.129147 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Oct 05 08:06:02 crc kubenswrapper[4846]: I1005 08:06:02.258961 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/memcached-0"] Oct 05 08:06:02 crc kubenswrapper[4846]: I1005 08:06:02.259907 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Oct 05 08:06:02 crc kubenswrapper[4846]: I1005 08:06:02.265575 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"memcached-memcached-dockercfg-q6pxx" Oct 05 08:06:02 crc kubenswrapper[4846]: I1005 08:06:02.265586 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"memcached-config-data" Oct 05 08:06:02 crc kubenswrapper[4846]: I1005 08:06:02.266113 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-memcached-svc" Oct 05 08:06:02 crc kubenswrapper[4846]: I1005 08:06:02.281421 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Oct 05 08:06:02 crc kubenswrapper[4846]: I1005 08:06:02.396249 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8b5a8f28-85f4-41e5-8da2-f7fe7a0d1220-combined-ca-bundle\") pod \"memcached-0\" (UID: \"8b5a8f28-85f4-41e5-8da2-f7fe7a0d1220\") " pod="openstack/memcached-0" Oct 05 08:06:02 crc kubenswrapper[4846]: I1005 08:06:02.396312 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/8b5a8f28-85f4-41e5-8da2-f7fe7a0d1220-memcached-tls-certs\") pod \"memcached-0\" (UID: \"8b5a8f28-85f4-41e5-8da2-f7fe7a0d1220\") " pod="openstack/memcached-0" Oct 05 08:06:02 crc kubenswrapper[4846]: I1005 08:06:02.396374 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/8b5a8f28-85f4-41e5-8da2-f7fe7a0d1220-kolla-config\") pod \"memcached-0\" (UID: \"8b5a8f28-85f4-41e5-8da2-f7fe7a0d1220\") " pod="openstack/memcached-0" Oct 05 08:06:02 crc kubenswrapper[4846]: I1005 08:06:02.396451 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-krn2k\" (UniqueName: \"kubernetes.io/projected/8b5a8f28-85f4-41e5-8da2-f7fe7a0d1220-kube-api-access-krn2k\") pod \"memcached-0\" (UID: \"8b5a8f28-85f4-41e5-8da2-f7fe7a0d1220\") " pod="openstack/memcached-0" Oct 05 08:06:02 crc kubenswrapper[4846]: I1005 08:06:02.396720 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8b5a8f28-85f4-41e5-8da2-f7fe7a0d1220-config-data\") pod \"memcached-0\" (UID: \"8b5a8f28-85f4-41e5-8da2-f7fe7a0d1220\") " pod="openstack/memcached-0" Oct 05 08:06:02 crc kubenswrapper[4846]: I1005 08:06:02.498036 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8b5a8f28-85f4-41e5-8da2-f7fe7a0d1220-combined-ca-bundle\") pod \"memcached-0\" (UID: \"8b5a8f28-85f4-41e5-8da2-f7fe7a0d1220\") " pod="openstack/memcached-0" Oct 05 08:06:02 crc kubenswrapper[4846]: I1005 08:06:02.498077 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/8b5a8f28-85f4-41e5-8da2-f7fe7a0d1220-memcached-tls-certs\") pod \"memcached-0\" (UID: \"8b5a8f28-85f4-41e5-8da2-f7fe7a0d1220\") " pod="openstack/memcached-0" Oct 05 08:06:02 crc kubenswrapper[4846]: I1005 08:06:02.498097 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/8b5a8f28-85f4-41e5-8da2-f7fe7a0d1220-kolla-config\") pod \"memcached-0\" (UID: \"8b5a8f28-85f4-41e5-8da2-f7fe7a0d1220\") " pod="openstack/memcached-0" Oct 05 08:06:02 crc kubenswrapper[4846]: I1005 08:06:02.498119 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-krn2k\" (UniqueName: \"kubernetes.io/projected/8b5a8f28-85f4-41e5-8da2-f7fe7a0d1220-kube-api-access-krn2k\") pod \"memcached-0\" (UID: \"8b5a8f28-85f4-41e5-8da2-f7fe7a0d1220\") " pod="openstack/memcached-0" Oct 05 08:06:02 crc kubenswrapper[4846]: I1005 08:06:02.498142 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8b5a8f28-85f4-41e5-8da2-f7fe7a0d1220-config-data\") pod \"memcached-0\" (UID: \"8b5a8f28-85f4-41e5-8da2-f7fe7a0d1220\") " pod="openstack/memcached-0" Oct 05 08:06:02 crc kubenswrapper[4846]: I1005 08:06:02.498973 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8b5a8f28-85f4-41e5-8da2-f7fe7a0d1220-config-data\") pod \"memcached-0\" (UID: \"8b5a8f28-85f4-41e5-8da2-f7fe7a0d1220\") " pod="openstack/memcached-0" Oct 05 08:06:02 crc kubenswrapper[4846]: I1005 08:06:02.499679 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/8b5a8f28-85f4-41e5-8da2-f7fe7a0d1220-kolla-config\") pod \"memcached-0\" (UID: \"8b5a8f28-85f4-41e5-8da2-f7fe7a0d1220\") " pod="openstack/memcached-0" Oct 05 08:06:02 crc kubenswrapper[4846]: I1005 08:06:02.511856 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/8b5a8f28-85f4-41e5-8da2-f7fe7a0d1220-memcached-tls-certs\") pod \"memcached-0\" (UID: \"8b5a8f28-85f4-41e5-8da2-f7fe7a0d1220\") " pod="openstack/memcached-0" Oct 05 08:06:02 crc kubenswrapper[4846]: I1005 08:06:02.512590 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8b5a8f28-85f4-41e5-8da2-f7fe7a0d1220-combined-ca-bundle\") pod \"memcached-0\" (UID: \"8b5a8f28-85f4-41e5-8da2-f7fe7a0d1220\") " pod="openstack/memcached-0" Oct 05 08:06:02 crc kubenswrapper[4846]: I1005 08:06:02.532369 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"771f44c5-2cc5-431e-9569-eb7da776756d","Type":"ContainerStarted","Data":"5842da1b8e2e56632cbf7404f5b2fe29f742b978b9762834166bd4db490aa643"} Oct 05 08:06:02 crc kubenswrapper[4846]: I1005 08:06:02.536070 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-krn2k\" (UniqueName: \"kubernetes.io/projected/8b5a8f28-85f4-41e5-8da2-f7fe7a0d1220-kube-api-access-krn2k\") pod \"memcached-0\" (UID: \"8b5a8f28-85f4-41e5-8da2-f7fe7a0d1220\") " pod="openstack/memcached-0" Oct 05 08:06:02 crc kubenswrapper[4846]: I1005 08:06:02.604316 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Oct 05 08:06:02 crc kubenswrapper[4846]: I1005 08:06:02.625429 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Oct 05 08:06:03 crc kubenswrapper[4846]: I1005 08:06:03.064479 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Oct 05 08:06:03 crc kubenswrapper[4846]: I1005 08:06:03.539428 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"643edc27-5b17-45e9-9286-f74591843690","Type":"ContainerStarted","Data":"eae6b0d64da1a38e7da1364998e09211a0b76fcf93888868cdef59375269099a"} Oct 05 08:06:04 crc kubenswrapper[4846]: I1005 08:06:04.047747 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-h76f7"] Oct 05 08:06:04 crc kubenswrapper[4846]: I1005 08:06:04.051348 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-h76f7" Oct 05 08:06:04 crc kubenswrapper[4846]: I1005 08:06:04.055409 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-h76f7"] Oct 05 08:06:04 crc kubenswrapper[4846]: I1005 08:06:04.129702 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a242d9d6-e865-48c3-bcc0-ca4084acad69-catalog-content\") pod \"redhat-marketplace-h76f7\" (UID: \"a242d9d6-e865-48c3-bcc0-ca4084acad69\") " pod="openshift-marketplace/redhat-marketplace-h76f7" Oct 05 08:06:04 crc kubenswrapper[4846]: I1005 08:06:04.129748 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hthdl\" (UniqueName: \"kubernetes.io/projected/a242d9d6-e865-48c3-bcc0-ca4084acad69-kube-api-access-hthdl\") pod \"redhat-marketplace-h76f7\" (UID: \"a242d9d6-e865-48c3-bcc0-ca4084acad69\") " pod="openshift-marketplace/redhat-marketplace-h76f7" Oct 05 08:06:04 crc kubenswrapper[4846]: I1005 08:06:04.129778 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a242d9d6-e865-48c3-bcc0-ca4084acad69-utilities\") pod \"redhat-marketplace-h76f7\" (UID: \"a242d9d6-e865-48c3-bcc0-ca4084acad69\") " pod="openshift-marketplace/redhat-marketplace-h76f7" Oct 05 08:06:04 crc kubenswrapper[4846]: I1005 08:06:04.231411 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a242d9d6-e865-48c3-bcc0-ca4084acad69-catalog-content\") pod \"redhat-marketplace-h76f7\" (UID: \"a242d9d6-e865-48c3-bcc0-ca4084acad69\") " pod="openshift-marketplace/redhat-marketplace-h76f7" Oct 05 08:06:04 crc kubenswrapper[4846]: I1005 08:06:04.231460 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hthdl\" (UniqueName: \"kubernetes.io/projected/a242d9d6-e865-48c3-bcc0-ca4084acad69-kube-api-access-hthdl\") pod \"redhat-marketplace-h76f7\" (UID: \"a242d9d6-e865-48c3-bcc0-ca4084acad69\") " pod="openshift-marketplace/redhat-marketplace-h76f7" Oct 05 08:06:04 crc kubenswrapper[4846]: I1005 08:06:04.231482 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a242d9d6-e865-48c3-bcc0-ca4084acad69-utilities\") pod \"redhat-marketplace-h76f7\" (UID: \"a242d9d6-e865-48c3-bcc0-ca4084acad69\") " pod="openshift-marketplace/redhat-marketplace-h76f7" Oct 05 08:06:04 crc kubenswrapper[4846]: I1005 08:06:04.231912 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a242d9d6-e865-48c3-bcc0-ca4084acad69-utilities\") pod \"redhat-marketplace-h76f7\" (UID: \"a242d9d6-e865-48c3-bcc0-ca4084acad69\") " pod="openshift-marketplace/redhat-marketplace-h76f7" Oct 05 08:06:04 crc kubenswrapper[4846]: I1005 08:06:04.231928 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a242d9d6-e865-48c3-bcc0-ca4084acad69-catalog-content\") pod \"redhat-marketplace-h76f7\" (UID: \"a242d9d6-e865-48c3-bcc0-ca4084acad69\") " pod="openshift-marketplace/redhat-marketplace-h76f7" Oct 05 08:06:04 crc kubenswrapper[4846]: I1005 08:06:04.249204 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hthdl\" (UniqueName: \"kubernetes.io/projected/a242d9d6-e865-48c3-bcc0-ca4084acad69-kube-api-access-hthdl\") pod \"redhat-marketplace-h76f7\" (UID: \"a242d9d6-e865-48c3-bcc0-ca4084acad69\") " pod="openshift-marketplace/redhat-marketplace-h76f7" Oct 05 08:06:04 crc kubenswrapper[4846]: I1005 08:06:04.382542 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-h76f7" Oct 05 08:06:05 crc kubenswrapper[4846]: I1005 08:06:05.555601 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"8b5a8f28-85f4-41e5-8da2-f7fe7a0d1220","Type":"ContainerStarted","Data":"b1ba7073f2682f6f7fda7feb24bdb7e390141ef5945fa55a5e3791d1ce824e57"} Oct 05 08:06:13 crc kubenswrapper[4846]: I1005 08:06:13.636713 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-h76f7"] Oct 05 08:06:15 crc kubenswrapper[4846]: W1005 08:06:15.921488 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda242d9d6_e865_48c3_bcc0_ca4084acad69.slice/crio-3d8113bb8247e92b2508fcce6466a578fa535cfc736ef5be7d4814e7f13e3785 WatchSource:0}: Error finding container 3d8113bb8247e92b2508fcce6466a578fa535cfc736ef5be7d4814e7f13e3785: Status 404 returned error can't find the container with id 3d8113bb8247e92b2508fcce6466a578fa535cfc736ef5be7d4814e7f13e3785 Oct 05 08:06:16 crc kubenswrapper[4846]: I1005 08:06:16.671424 4846 generic.go:334] "Generic (PLEG): container finished" podID="a242d9d6-e865-48c3-bcc0-ca4084acad69" containerID="87bf61f7d51f07f4353dea295ef025b9369270cb2a4a5a701063f339b58f4c9c" exitCode=0 Oct 05 08:06:16 crc kubenswrapper[4846]: I1005 08:06:16.671596 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-h76f7" event={"ID":"a242d9d6-e865-48c3-bcc0-ca4084acad69","Type":"ContainerDied","Data":"87bf61f7d51f07f4353dea295ef025b9369270cb2a4a5a701063f339b58f4c9c"} Oct 05 08:06:16 crc kubenswrapper[4846]: I1005 08:06:16.671953 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-h76f7" event={"ID":"a242d9d6-e865-48c3-bcc0-ca4084acad69","Type":"ContainerStarted","Data":"3d8113bb8247e92b2508fcce6466a578fa535cfc736ef5be7d4814e7f13e3785"} Oct 05 08:06:16 crc kubenswrapper[4846]: I1005 08:06:16.677416 4846 generic.go:334] "Generic (PLEG): container finished" podID="d6d6783a-7816-402e-b38e-935ffcb7314e" containerID="8876aec95a893518e103665f2a4586f7d40b487a9668dbdb64a9fe128886b884" exitCode=0 Oct 05 08:06:16 crc kubenswrapper[4846]: I1005 08:06:16.677472 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55cbcdb7f9-cftk8" event={"ID":"d6d6783a-7816-402e-b38e-935ffcb7314e","Type":"ContainerDied","Data":"8876aec95a893518e103665f2a4586f7d40b487a9668dbdb64a9fe128886b884"} Oct 05 08:06:17 crc kubenswrapper[4846]: I1005 08:06:17.687262 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"771f44c5-2cc5-431e-9569-eb7da776756d","Type":"ContainerStarted","Data":"20c39c043ee809b5b4035e000fa9aba6f9f793016c3bbc02d280fc6ff4f6777d"} Oct 05 08:06:17 crc kubenswrapper[4846]: I1005 08:06:17.688985 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"a288716c-b683-47c0-bf44-c0bbc997da97","Type":"ContainerStarted","Data":"acf58c0d7da784189c7fe44a098cc3fcf43783cf328b1e35c6673ffa86b97407"} Oct 05 08:06:17 crc kubenswrapper[4846]: I1005 08:06:17.691300 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"643edc27-5b17-45e9-9286-f74591843690","Type":"ContainerStarted","Data":"b017a81a4af597abea1ce5f4dd6450196e7caef4004eb5c24a7847adc8e827fd"} Oct 05 08:06:17 crc kubenswrapper[4846]: I1005 08:06:17.694136 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55cbcdb7f9-cftk8" event={"ID":"d6d6783a-7816-402e-b38e-935ffcb7314e","Type":"ContainerDied","Data":"8aea79dfaa6807aa570d88e3076b71c703c9d3cc6ebcd51f9235172a2c11faf6"} Oct 05 08:06:17 crc kubenswrapper[4846]: I1005 08:06:17.694166 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8aea79dfaa6807aa570d88e3076b71c703c9d3cc6ebcd51f9235172a2c11faf6" Oct 05 08:06:17 crc kubenswrapper[4846]: I1005 08:06:17.699087 4846 generic.go:334] "Generic (PLEG): container finished" podID="e4bbd7e3-1e67-4145-943d-fe911ac8f3ba" containerID="b442aa9794bf57d50231ee0e39bf7081a6948efd11b46994a86d66e4e417a840" exitCode=0 Oct 05 08:06:17 crc kubenswrapper[4846]: I1005 08:06:17.699191 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6cfd7b4c45-gjjg8" event={"ID":"e4bbd7e3-1e67-4145-943d-fe911ac8f3ba","Type":"ContainerDied","Data":"b442aa9794bf57d50231ee0e39bf7081a6948efd11b46994a86d66e4e417a840"} Oct 05 08:06:17 crc kubenswrapper[4846]: I1005 08:06:17.700602 4846 generic.go:334] "Generic (PLEG): container finished" podID="ebf7ce11-32a8-4fea-9488-8f48b08991a9" containerID="5bf8c50fb8d81adf3a96df86ca7ec827673ce8a866e72c4f06768eca5873ac4c" exitCode=0 Oct 05 08:06:17 crc kubenswrapper[4846]: I1005 08:06:17.700651 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dbf54cbc-c2z7x" event={"ID":"ebf7ce11-32a8-4fea-9488-8f48b08991a9","Type":"ContainerDied","Data":"5bf8c50fb8d81adf3a96df86ca7ec827673ce8a866e72c4f06768eca5873ac4c"} Oct 05 08:06:17 crc kubenswrapper[4846]: I1005 08:06:17.702323 4846 generic.go:334] "Generic (PLEG): container finished" podID="037a0419-1e47-429c-ac6f-e2fe6e7db5d9" containerID="8b8a2037cbaae640928b9e66be056953b808d40ddca2e72272402cc2f9e01f70" exitCode=0 Oct 05 08:06:17 crc kubenswrapper[4846]: I1005 08:06:17.702369 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6697c69467-27t47" event={"ID":"037a0419-1e47-429c-ac6f-e2fe6e7db5d9","Type":"ContainerDied","Data":"8b8a2037cbaae640928b9e66be056953b808d40ddca2e72272402cc2f9e01f70"} Oct 05 08:06:17 crc kubenswrapper[4846]: I1005 08:06:17.705747 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"8b5a8f28-85f4-41e5-8da2-f7fe7a0d1220","Type":"ContainerStarted","Data":"ed1a47c654737a603e0056562098a36c51622a8af3320b3e817aac3c164a9c95"} Oct 05 08:06:17 crc kubenswrapper[4846]: I1005 08:06:17.705934 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/memcached-0" Oct 05 08:06:17 crc kubenswrapper[4846]: I1005 08:06:17.707438 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"0f8171a9-82f5-443a-a8fc-d42b575d498a","Type":"ContainerStarted","Data":"2db7c55ce2a5c2dd4b47e3a9f2c447ed7148777d04a0deeecedd8f7015c83e5e"} Oct 05 08:06:17 crc kubenswrapper[4846]: I1005 08:06:17.868574 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55cbcdb7f9-cftk8" Oct 05 08:06:17 crc kubenswrapper[4846]: I1005 08:06:17.887660 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/memcached-0" podStartSLOduration=5.253286082 podStartE2EDuration="15.887642528s" podCreationTimestamp="2025-10-05 08:06:02 +0000 UTC" firstStartedPulling="2025-10-05 08:06:05.349812552 +0000 UTC m=+4687.590665327" lastFinishedPulling="2025-10-05 08:06:15.984168998 +0000 UTC m=+4698.225021773" observedRunningTime="2025-10-05 08:06:17.878196515 +0000 UTC m=+4700.119049300" watchObservedRunningTime="2025-10-05 08:06:17.887642528 +0000 UTC m=+4700.128495303" Oct 05 08:06:18 crc kubenswrapper[4846]: I1005 08:06:18.024096 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d6d6783a-7816-402e-b38e-935ffcb7314e-config\") pod \"d6d6783a-7816-402e-b38e-935ffcb7314e\" (UID: \"d6d6783a-7816-402e-b38e-935ffcb7314e\") " Oct 05 08:06:18 crc kubenswrapper[4846]: I1005 08:06:18.024144 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7sbjj\" (UniqueName: \"kubernetes.io/projected/d6d6783a-7816-402e-b38e-935ffcb7314e-kube-api-access-7sbjj\") pod \"d6d6783a-7816-402e-b38e-935ffcb7314e\" (UID: \"d6d6783a-7816-402e-b38e-935ffcb7314e\") " Oct 05 08:06:18 crc kubenswrapper[4846]: I1005 08:06:18.040369 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d6d6783a-7816-402e-b38e-935ffcb7314e-kube-api-access-7sbjj" (OuterVolumeSpecName: "kube-api-access-7sbjj") pod "d6d6783a-7816-402e-b38e-935ffcb7314e" (UID: "d6d6783a-7816-402e-b38e-935ffcb7314e"). InnerVolumeSpecName "kube-api-access-7sbjj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:06:18 crc kubenswrapper[4846]: I1005 08:06:18.050102 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d6d6783a-7816-402e-b38e-935ffcb7314e-config" (OuterVolumeSpecName: "config") pod "d6d6783a-7816-402e-b38e-935ffcb7314e" (UID: "d6d6783a-7816-402e-b38e-935ffcb7314e"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 08:06:18 crc kubenswrapper[4846]: I1005 08:06:18.125818 4846 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d6d6783a-7816-402e-b38e-935ffcb7314e-config\") on node \"crc\" DevicePath \"\"" Oct 05 08:06:18 crc kubenswrapper[4846]: I1005 08:06:18.125858 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7sbjj\" (UniqueName: \"kubernetes.io/projected/d6d6783a-7816-402e-b38e-935ffcb7314e-kube-api-access-7sbjj\") on node \"crc\" DevicePath \"\"" Oct 05 08:06:18 crc kubenswrapper[4846]: I1005 08:06:18.162531 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6697c69467-27t47" Oct 05 08:06:18 crc kubenswrapper[4846]: I1005 08:06:18.328843 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/037a0419-1e47-429c-ac6f-e2fe6e7db5d9-config\") pod \"037a0419-1e47-429c-ac6f-e2fe6e7db5d9\" (UID: \"037a0419-1e47-429c-ac6f-e2fe6e7db5d9\") " Oct 05 08:06:18 crc kubenswrapper[4846]: I1005 08:06:18.329037 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/037a0419-1e47-429c-ac6f-e2fe6e7db5d9-dns-svc\") pod \"037a0419-1e47-429c-ac6f-e2fe6e7db5d9\" (UID: \"037a0419-1e47-429c-ac6f-e2fe6e7db5d9\") " Oct 05 08:06:18 crc kubenswrapper[4846]: I1005 08:06:18.329099 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v65gp\" (UniqueName: \"kubernetes.io/projected/037a0419-1e47-429c-ac6f-e2fe6e7db5d9-kube-api-access-v65gp\") pod \"037a0419-1e47-429c-ac6f-e2fe6e7db5d9\" (UID: \"037a0419-1e47-429c-ac6f-e2fe6e7db5d9\") " Oct 05 08:06:18 crc kubenswrapper[4846]: I1005 08:06:18.331707 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/037a0419-1e47-429c-ac6f-e2fe6e7db5d9-kube-api-access-v65gp" (OuterVolumeSpecName: "kube-api-access-v65gp") pod "037a0419-1e47-429c-ac6f-e2fe6e7db5d9" (UID: "037a0419-1e47-429c-ac6f-e2fe6e7db5d9"). InnerVolumeSpecName "kube-api-access-v65gp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:06:18 crc kubenswrapper[4846]: I1005 08:06:18.346606 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/037a0419-1e47-429c-ac6f-e2fe6e7db5d9-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "037a0419-1e47-429c-ac6f-e2fe6e7db5d9" (UID: "037a0419-1e47-429c-ac6f-e2fe6e7db5d9"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 08:06:18 crc kubenswrapper[4846]: I1005 08:06:18.359025 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/037a0419-1e47-429c-ac6f-e2fe6e7db5d9-config" (OuterVolumeSpecName: "config") pod "037a0419-1e47-429c-ac6f-e2fe6e7db5d9" (UID: "037a0419-1e47-429c-ac6f-e2fe6e7db5d9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 08:06:18 crc kubenswrapper[4846]: I1005 08:06:18.431101 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v65gp\" (UniqueName: \"kubernetes.io/projected/037a0419-1e47-429c-ac6f-e2fe6e7db5d9-kube-api-access-v65gp\") on node \"crc\" DevicePath \"\"" Oct 05 08:06:18 crc kubenswrapper[4846]: I1005 08:06:18.431200 4846 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/037a0419-1e47-429c-ac6f-e2fe6e7db5d9-config\") on node \"crc\" DevicePath \"\"" Oct 05 08:06:18 crc kubenswrapper[4846]: I1005 08:06:18.431229 4846 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/037a0419-1e47-429c-ac6f-e2fe6e7db5d9-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 05 08:06:18 crc kubenswrapper[4846]: I1005 08:06:18.718899 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6697c69467-27t47" event={"ID":"037a0419-1e47-429c-ac6f-e2fe6e7db5d9","Type":"ContainerDied","Data":"287993d9589773357eb676da45a84796bcc861e31aeaa91e62382c72f29f0a10"} Oct 05 08:06:18 crc kubenswrapper[4846]: I1005 08:06:18.718912 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6697c69467-27t47" Oct 05 08:06:18 crc kubenswrapper[4846]: I1005 08:06:18.718953 4846 scope.go:117] "RemoveContainer" containerID="8b8a2037cbaae640928b9e66be056953b808d40ddca2e72272402cc2f9e01f70" Oct 05 08:06:18 crc kubenswrapper[4846]: I1005 08:06:18.721563 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6cfd7b4c45-gjjg8" event={"ID":"e4bbd7e3-1e67-4145-943d-fe911ac8f3ba","Type":"ContainerStarted","Data":"91dc787a32b9b22d8b7fcd9ee93ac0d3a459f590ee48fbb2411b26a679ca11b2"} Oct 05 08:06:18 crc kubenswrapper[4846]: I1005 08:06:18.721935 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6cfd7b4c45-gjjg8" Oct 05 08:06:18 crc kubenswrapper[4846]: I1005 08:06:18.724053 4846 generic.go:334] "Generic (PLEG): container finished" podID="a242d9d6-e865-48c3-bcc0-ca4084acad69" containerID="6ce7d39854e8e55e7757003e846b0cdee8a724513eb7f483ee82022f558a6045" exitCode=0 Oct 05 08:06:18 crc kubenswrapper[4846]: I1005 08:06:18.724244 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-h76f7" event={"ID":"a242d9d6-e865-48c3-bcc0-ca4084acad69","Type":"ContainerDied","Data":"6ce7d39854e8e55e7757003e846b0cdee8a724513eb7f483ee82022f558a6045"} Oct 05 08:06:18 crc kubenswrapper[4846]: I1005 08:06:18.726799 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dbf54cbc-c2z7x" event={"ID":"ebf7ce11-32a8-4fea-9488-8f48b08991a9","Type":"ContainerStarted","Data":"e0df5e2f1beecb5f16efbdcef40251285f0b42e4ed72d03baf2e892412b980d6"} Oct 05 08:06:18 crc kubenswrapper[4846]: I1005 08:06:18.726987 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55cbcdb7f9-cftk8" Oct 05 08:06:18 crc kubenswrapper[4846]: I1005 08:06:18.761137 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6cfd7b4c45-gjjg8" podStartSLOduration=4.045072781 podStartE2EDuration="20.761105685s" podCreationTimestamp="2025-10-05 08:05:58 +0000 UTC" firstStartedPulling="2025-10-05 08:05:59.240313029 +0000 UTC m=+4681.481165804" lastFinishedPulling="2025-10-05 08:06:15.956345913 +0000 UTC m=+4698.197198708" observedRunningTime="2025-10-05 08:06:18.749031412 +0000 UTC m=+4700.989884217" watchObservedRunningTime="2025-10-05 08:06:18.761105685 +0000 UTC m=+4701.001958500" Oct 05 08:06:18 crc kubenswrapper[4846]: I1005 08:06:18.771412 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-78dbf54cbc-c2z7x" Oct 05 08:06:18 crc kubenswrapper[4846]: I1005 08:06:18.814685 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6697c69467-27t47"] Oct 05 08:06:18 crc kubenswrapper[4846]: I1005 08:06:18.821872 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6697c69467-27t47"] Oct 05 08:06:18 crc kubenswrapper[4846]: I1005 08:06:18.830156 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-78dbf54cbc-c2z7x" podStartSLOduration=3.899829372 podStartE2EDuration="20.830095943s" podCreationTimestamp="2025-10-05 08:05:58 +0000 UTC" firstStartedPulling="2025-10-05 08:05:59.155365224 +0000 UTC m=+4681.396217999" lastFinishedPulling="2025-10-05 08:06:16.085631775 +0000 UTC m=+4698.326484570" observedRunningTime="2025-10-05 08:06:18.809898632 +0000 UTC m=+4701.050751407" watchObservedRunningTime="2025-10-05 08:06:18.830095943 +0000 UTC m=+4701.070948768" Oct 05 08:06:18 crc kubenswrapper[4846]: I1005 08:06:18.881416 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-55cbcdb7f9-cftk8"] Oct 05 08:06:18 crc kubenswrapper[4846]: I1005 08:06:18.886769 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-55cbcdb7f9-cftk8"] Oct 05 08:06:19 crc kubenswrapper[4846]: I1005 08:06:19.738311 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-h76f7" event={"ID":"a242d9d6-e865-48c3-bcc0-ca4084acad69","Type":"ContainerStarted","Data":"b3bca8f0f7e189f11e49d72dcd8125ff55f26898af90be1e73e8266ec3f0fa2e"} Oct 05 08:06:19 crc kubenswrapper[4846]: I1005 08:06:19.776599 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-h76f7" podStartSLOduration=13.307060461 podStartE2EDuration="15.776580387s" podCreationTimestamp="2025-10-05 08:06:04 +0000 UTC" firstStartedPulling="2025-10-05 08:06:16.674704088 +0000 UTC m=+4698.915556893" lastFinishedPulling="2025-10-05 08:06:19.144224014 +0000 UTC m=+4701.385076819" observedRunningTime="2025-10-05 08:06:19.771665025 +0000 UTC m=+4702.012517820" watchObservedRunningTime="2025-10-05 08:06:19.776580387 +0000 UTC m=+4702.017433172" Oct 05 08:06:20 crc kubenswrapper[4846]: I1005 08:06:20.516842 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="037a0419-1e47-429c-ac6f-e2fe6e7db5d9" path="/var/lib/kubelet/pods/037a0419-1e47-429c-ac6f-e2fe6e7db5d9/volumes" Oct 05 08:06:20 crc kubenswrapper[4846]: I1005 08:06:20.518529 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d6d6783a-7816-402e-b38e-935ffcb7314e" path="/var/lib/kubelet/pods/d6d6783a-7816-402e-b38e-935ffcb7314e/volumes" Oct 05 08:06:20 crc kubenswrapper[4846]: I1005 08:06:20.754139 4846 generic.go:334] "Generic (PLEG): container finished" podID="771f44c5-2cc5-431e-9569-eb7da776756d" containerID="20c39c043ee809b5b4035e000fa9aba6f9f793016c3bbc02d280fc6ff4f6777d" exitCode=0 Oct 05 08:06:20 crc kubenswrapper[4846]: I1005 08:06:20.754272 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"771f44c5-2cc5-431e-9569-eb7da776756d","Type":"ContainerDied","Data":"20c39c043ee809b5b4035e000fa9aba6f9f793016c3bbc02d280fc6ff4f6777d"} Oct 05 08:06:20 crc kubenswrapper[4846]: I1005 08:06:20.757804 4846 generic.go:334] "Generic (PLEG): container finished" podID="643edc27-5b17-45e9-9286-f74591843690" containerID="b017a81a4af597abea1ce5f4dd6450196e7caef4004eb5c24a7847adc8e827fd" exitCode=0 Oct 05 08:06:20 crc kubenswrapper[4846]: I1005 08:06:20.758467 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"643edc27-5b17-45e9-9286-f74591843690","Type":"ContainerDied","Data":"b017a81a4af597abea1ce5f4dd6450196e7caef4004eb5c24a7847adc8e827fd"} Oct 05 08:06:21 crc kubenswrapper[4846]: I1005 08:06:21.772884 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"771f44c5-2cc5-431e-9569-eb7da776756d","Type":"ContainerStarted","Data":"4eb30eb6298f47c8ba67d6982c45bfbf6da9cc77b3c767a0731eab6b0318b123"} Oct 05 08:06:21 crc kubenswrapper[4846]: I1005 08:06:21.775971 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"643edc27-5b17-45e9-9286-f74591843690","Type":"ContainerStarted","Data":"bdcc9b7ecbdd3a0ce8ea4b2417738c9b8cea61b848cd1e5e4a3eaeace6f52d10"} Oct 05 08:06:21 crc kubenswrapper[4846]: I1005 08:06:21.807576 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-galera-0" podStartSLOduration=8.40222648 podStartE2EDuration="22.80754394s" podCreationTimestamp="2025-10-05 08:05:59 +0000 UTC" firstStartedPulling="2025-10-05 08:06:01.52403713 +0000 UTC m=+4683.764889905" lastFinishedPulling="2025-10-05 08:06:15.92935459 +0000 UTC m=+4698.170207365" observedRunningTime="2025-10-05 08:06:21.805682201 +0000 UTC m=+4704.046535026" watchObservedRunningTime="2025-10-05 08:06:21.80754394 +0000 UTC m=+4704.048396765" Oct 05 08:06:21 crc kubenswrapper[4846]: I1005 08:06:21.843015 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-cell1-galera-0" podStartSLOduration=8.49458032 podStartE2EDuration="21.84298906s" podCreationTimestamp="2025-10-05 08:06:00 +0000 UTC" firstStartedPulling="2025-10-05 08:06:02.635121051 +0000 UTC m=+4684.875973826" lastFinishedPulling="2025-10-05 08:06:15.983529771 +0000 UTC m=+4698.224382566" observedRunningTime="2025-10-05 08:06:21.83592401 +0000 UTC m=+4704.076776825" watchObservedRunningTime="2025-10-05 08:06:21.84298906 +0000 UTC m=+4704.083841865" Oct 05 08:06:22 crc kubenswrapper[4846]: I1005 08:06:22.129508 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-cell1-galera-0" Oct 05 08:06:22 crc kubenswrapper[4846]: I1005 08:06:22.129590 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-cell1-galera-0" Oct 05 08:06:22 crc kubenswrapper[4846]: I1005 08:06:22.607771 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/memcached-0" Oct 05 08:06:23 crc kubenswrapper[4846]: I1005 08:06:23.325679 4846 patch_prober.go:28] interesting pod/machine-config-daemon-fscvf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 08:06:23 crc kubenswrapper[4846]: I1005 08:06:23.326134 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 08:06:23 crc kubenswrapper[4846]: I1005 08:06:23.745430 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6cfd7b4c45-gjjg8" Oct 05 08:06:23 crc kubenswrapper[4846]: I1005 08:06:23.770145 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-78dbf54cbc-c2z7x" Oct 05 08:06:23 crc kubenswrapper[4846]: I1005 08:06:23.855648 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dbf54cbc-c2z7x"] Oct 05 08:06:23 crc kubenswrapper[4846]: I1005 08:06:23.856122 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-78dbf54cbc-c2z7x" podUID="ebf7ce11-32a8-4fea-9488-8f48b08991a9" containerName="dnsmasq-dns" containerID="cri-o://e0df5e2f1beecb5f16efbdcef40251285f0b42e4ed72d03baf2e892412b980d6" gracePeriod=10 Oct 05 08:06:24 crc kubenswrapper[4846]: I1005 08:06:24.348114 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dbf54cbc-c2z7x" Oct 05 08:06:24 crc kubenswrapper[4846]: I1005 08:06:24.383281 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-h76f7" Oct 05 08:06:24 crc kubenswrapper[4846]: I1005 08:06:24.383324 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-h76f7" Oct 05 08:06:24 crc kubenswrapper[4846]: I1005 08:06:24.425193 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-h76f7" Oct 05 08:06:24 crc kubenswrapper[4846]: I1005 08:06:24.454399 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zm54k\" (UniqueName: \"kubernetes.io/projected/ebf7ce11-32a8-4fea-9488-8f48b08991a9-kube-api-access-zm54k\") pod \"ebf7ce11-32a8-4fea-9488-8f48b08991a9\" (UID: \"ebf7ce11-32a8-4fea-9488-8f48b08991a9\") " Oct 05 08:06:24 crc kubenswrapper[4846]: I1005 08:06:24.454479 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ebf7ce11-32a8-4fea-9488-8f48b08991a9-dns-svc\") pod \"ebf7ce11-32a8-4fea-9488-8f48b08991a9\" (UID: \"ebf7ce11-32a8-4fea-9488-8f48b08991a9\") " Oct 05 08:06:24 crc kubenswrapper[4846]: I1005 08:06:24.454568 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ebf7ce11-32a8-4fea-9488-8f48b08991a9-config\") pod \"ebf7ce11-32a8-4fea-9488-8f48b08991a9\" (UID: \"ebf7ce11-32a8-4fea-9488-8f48b08991a9\") " Oct 05 08:06:24 crc kubenswrapper[4846]: I1005 08:06:24.461380 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ebf7ce11-32a8-4fea-9488-8f48b08991a9-kube-api-access-zm54k" (OuterVolumeSpecName: "kube-api-access-zm54k") pod "ebf7ce11-32a8-4fea-9488-8f48b08991a9" (UID: "ebf7ce11-32a8-4fea-9488-8f48b08991a9"). InnerVolumeSpecName "kube-api-access-zm54k". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:06:24 crc kubenswrapper[4846]: E1005 08:06:24.492872 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/ebf7ce11-32a8-4fea-9488-8f48b08991a9-dns-svc podName:ebf7ce11-32a8-4fea-9488-8f48b08991a9 nodeName:}" failed. No retries permitted until 2025-10-05 08:06:24.992841794 +0000 UTC m=+4707.233694579 (durationBeforeRetry 500ms). Error: error cleaning subPath mounts for volume "dns-svc" (UniqueName: "kubernetes.io/configmap/ebf7ce11-32a8-4fea-9488-8f48b08991a9-dns-svc") pod "ebf7ce11-32a8-4fea-9488-8f48b08991a9" (UID: "ebf7ce11-32a8-4fea-9488-8f48b08991a9") : error deleting /var/lib/kubelet/pods/ebf7ce11-32a8-4fea-9488-8f48b08991a9/volume-subpaths: remove /var/lib/kubelet/pods/ebf7ce11-32a8-4fea-9488-8f48b08991a9/volume-subpaths: no such file or directory Oct 05 08:06:24 crc kubenswrapper[4846]: I1005 08:06:24.493337 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ebf7ce11-32a8-4fea-9488-8f48b08991a9-config" (OuterVolumeSpecName: "config") pod "ebf7ce11-32a8-4fea-9488-8f48b08991a9" (UID: "ebf7ce11-32a8-4fea-9488-8f48b08991a9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 08:06:24 crc kubenswrapper[4846]: I1005 08:06:24.556394 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zm54k\" (UniqueName: \"kubernetes.io/projected/ebf7ce11-32a8-4fea-9488-8f48b08991a9-kube-api-access-zm54k\") on node \"crc\" DevicePath \"\"" Oct 05 08:06:24 crc kubenswrapper[4846]: I1005 08:06:24.557039 4846 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ebf7ce11-32a8-4fea-9488-8f48b08991a9-config\") on node \"crc\" DevicePath \"\"" Oct 05 08:06:24 crc kubenswrapper[4846]: I1005 08:06:24.808064 4846 generic.go:334] "Generic (PLEG): container finished" podID="ebf7ce11-32a8-4fea-9488-8f48b08991a9" containerID="e0df5e2f1beecb5f16efbdcef40251285f0b42e4ed72d03baf2e892412b980d6" exitCode=0 Oct 05 08:06:24 crc kubenswrapper[4846]: I1005 08:06:24.808149 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dbf54cbc-c2z7x" Oct 05 08:06:24 crc kubenswrapper[4846]: I1005 08:06:24.808131 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dbf54cbc-c2z7x" event={"ID":"ebf7ce11-32a8-4fea-9488-8f48b08991a9","Type":"ContainerDied","Data":"e0df5e2f1beecb5f16efbdcef40251285f0b42e4ed72d03baf2e892412b980d6"} Oct 05 08:06:24 crc kubenswrapper[4846]: I1005 08:06:24.809236 4846 scope.go:117] "RemoveContainer" containerID="e0df5e2f1beecb5f16efbdcef40251285f0b42e4ed72d03baf2e892412b980d6" Oct 05 08:06:24 crc kubenswrapper[4846]: I1005 08:06:24.809011 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dbf54cbc-c2z7x" event={"ID":"ebf7ce11-32a8-4fea-9488-8f48b08991a9","Type":"ContainerDied","Data":"959478198ae09db61f2d861e262d4ec5924205b6ef01a0842eed82be2acc56f4"} Oct 05 08:06:24 crc kubenswrapper[4846]: I1005 08:06:24.849412 4846 scope.go:117] "RemoveContainer" containerID="5bf8c50fb8d81adf3a96df86ca7ec827673ce8a866e72c4f06768eca5873ac4c" Oct 05 08:06:24 crc kubenswrapper[4846]: I1005 08:06:24.898302 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-h76f7" Oct 05 08:06:24 crc kubenswrapper[4846]: I1005 08:06:24.901731 4846 scope.go:117] "RemoveContainer" containerID="e0df5e2f1beecb5f16efbdcef40251285f0b42e4ed72d03baf2e892412b980d6" Oct 05 08:06:24 crc kubenswrapper[4846]: E1005 08:06:24.903090 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e0df5e2f1beecb5f16efbdcef40251285f0b42e4ed72d03baf2e892412b980d6\": container with ID starting with e0df5e2f1beecb5f16efbdcef40251285f0b42e4ed72d03baf2e892412b980d6 not found: ID does not exist" containerID="e0df5e2f1beecb5f16efbdcef40251285f0b42e4ed72d03baf2e892412b980d6" Oct 05 08:06:24 crc kubenswrapper[4846]: I1005 08:06:24.903158 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e0df5e2f1beecb5f16efbdcef40251285f0b42e4ed72d03baf2e892412b980d6"} err="failed to get container status \"e0df5e2f1beecb5f16efbdcef40251285f0b42e4ed72d03baf2e892412b980d6\": rpc error: code = NotFound desc = could not find container \"e0df5e2f1beecb5f16efbdcef40251285f0b42e4ed72d03baf2e892412b980d6\": container with ID starting with e0df5e2f1beecb5f16efbdcef40251285f0b42e4ed72d03baf2e892412b980d6 not found: ID does not exist" Oct 05 08:06:24 crc kubenswrapper[4846]: I1005 08:06:24.903216 4846 scope.go:117] "RemoveContainer" containerID="5bf8c50fb8d81adf3a96df86ca7ec827673ce8a866e72c4f06768eca5873ac4c" Oct 05 08:06:24 crc kubenswrapper[4846]: E1005 08:06:24.903802 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5bf8c50fb8d81adf3a96df86ca7ec827673ce8a866e72c4f06768eca5873ac4c\": container with ID starting with 5bf8c50fb8d81adf3a96df86ca7ec827673ce8a866e72c4f06768eca5873ac4c not found: ID does not exist" containerID="5bf8c50fb8d81adf3a96df86ca7ec827673ce8a866e72c4f06768eca5873ac4c" Oct 05 08:06:24 crc kubenswrapper[4846]: I1005 08:06:24.903866 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5bf8c50fb8d81adf3a96df86ca7ec827673ce8a866e72c4f06768eca5873ac4c"} err="failed to get container status \"5bf8c50fb8d81adf3a96df86ca7ec827673ce8a866e72c4f06768eca5873ac4c\": rpc error: code = NotFound desc = could not find container \"5bf8c50fb8d81adf3a96df86ca7ec827673ce8a866e72c4f06768eca5873ac4c\": container with ID starting with 5bf8c50fb8d81adf3a96df86ca7ec827673ce8a866e72c4f06768eca5873ac4c not found: ID does not exist" Oct 05 08:06:24 crc kubenswrapper[4846]: I1005 08:06:24.958129 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-h76f7"] Oct 05 08:06:25 crc kubenswrapper[4846]: I1005 08:06:25.065361 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ebf7ce11-32a8-4fea-9488-8f48b08991a9-dns-svc\") pod \"ebf7ce11-32a8-4fea-9488-8f48b08991a9\" (UID: \"ebf7ce11-32a8-4fea-9488-8f48b08991a9\") " Oct 05 08:06:25 crc kubenswrapper[4846]: I1005 08:06:25.065739 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ebf7ce11-32a8-4fea-9488-8f48b08991a9-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "ebf7ce11-32a8-4fea-9488-8f48b08991a9" (UID: "ebf7ce11-32a8-4fea-9488-8f48b08991a9"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 08:06:25 crc kubenswrapper[4846]: I1005 08:06:25.066210 4846 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ebf7ce11-32a8-4fea-9488-8f48b08991a9-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 05 08:06:25 crc kubenswrapper[4846]: I1005 08:06:25.152925 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dbf54cbc-c2z7x"] Oct 05 08:06:25 crc kubenswrapper[4846]: I1005 08:06:25.159800 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-78dbf54cbc-c2z7x"] Oct 05 08:06:26 crc kubenswrapper[4846]: I1005 08:06:26.248899 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-cell1-galera-0" Oct 05 08:06:26 crc kubenswrapper[4846]: I1005 08:06:26.308277 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-cell1-galera-0" Oct 05 08:06:26 crc kubenswrapper[4846]: I1005 08:06:26.520664 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ebf7ce11-32a8-4fea-9488-8f48b08991a9" path="/var/lib/kubelet/pods/ebf7ce11-32a8-4fea-9488-8f48b08991a9/volumes" Oct 05 08:06:26 crc kubenswrapper[4846]: I1005 08:06:26.827450 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-h76f7" podUID="a242d9d6-e865-48c3-bcc0-ca4084acad69" containerName="registry-server" containerID="cri-o://b3bca8f0f7e189f11e49d72dcd8125ff55f26898af90be1e73e8266ec3f0fa2e" gracePeriod=2 Oct 05 08:06:27 crc kubenswrapper[4846]: I1005 08:06:27.308202 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-h76f7" Oct 05 08:06:27 crc kubenswrapper[4846]: I1005 08:06:27.506842 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hthdl\" (UniqueName: \"kubernetes.io/projected/a242d9d6-e865-48c3-bcc0-ca4084acad69-kube-api-access-hthdl\") pod \"a242d9d6-e865-48c3-bcc0-ca4084acad69\" (UID: \"a242d9d6-e865-48c3-bcc0-ca4084acad69\") " Oct 05 08:06:27 crc kubenswrapper[4846]: I1005 08:06:27.506924 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a242d9d6-e865-48c3-bcc0-ca4084acad69-utilities\") pod \"a242d9d6-e865-48c3-bcc0-ca4084acad69\" (UID: \"a242d9d6-e865-48c3-bcc0-ca4084acad69\") " Oct 05 08:06:27 crc kubenswrapper[4846]: I1005 08:06:27.506964 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a242d9d6-e865-48c3-bcc0-ca4084acad69-catalog-content\") pod \"a242d9d6-e865-48c3-bcc0-ca4084acad69\" (UID: \"a242d9d6-e865-48c3-bcc0-ca4084acad69\") " Oct 05 08:06:27 crc kubenswrapper[4846]: I1005 08:06:27.508557 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a242d9d6-e865-48c3-bcc0-ca4084acad69-utilities" (OuterVolumeSpecName: "utilities") pod "a242d9d6-e865-48c3-bcc0-ca4084acad69" (UID: "a242d9d6-e865-48c3-bcc0-ca4084acad69"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 08:06:27 crc kubenswrapper[4846]: I1005 08:06:27.524835 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a242d9d6-e865-48c3-bcc0-ca4084acad69-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a242d9d6-e865-48c3-bcc0-ca4084acad69" (UID: "a242d9d6-e865-48c3-bcc0-ca4084acad69"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 08:06:27 crc kubenswrapper[4846]: I1005 08:06:27.527413 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a242d9d6-e865-48c3-bcc0-ca4084acad69-kube-api-access-hthdl" (OuterVolumeSpecName: "kube-api-access-hthdl") pod "a242d9d6-e865-48c3-bcc0-ca4084acad69" (UID: "a242d9d6-e865-48c3-bcc0-ca4084acad69"). InnerVolumeSpecName "kube-api-access-hthdl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:06:27 crc kubenswrapper[4846]: I1005 08:06:27.609203 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hthdl\" (UniqueName: \"kubernetes.io/projected/a242d9d6-e865-48c3-bcc0-ca4084acad69-kube-api-access-hthdl\") on node \"crc\" DevicePath \"\"" Oct 05 08:06:27 crc kubenswrapper[4846]: I1005 08:06:27.609246 4846 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a242d9d6-e865-48c3-bcc0-ca4084acad69-utilities\") on node \"crc\" DevicePath \"\"" Oct 05 08:06:27 crc kubenswrapper[4846]: I1005 08:06:27.609261 4846 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a242d9d6-e865-48c3-bcc0-ca4084acad69-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 05 08:06:27 crc kubenswrapper[4846]: I1005 08:06:27.839724 4846 generic.go:334] "Generic (PLEG): container finished" podID="a242d9d6-e865-48c3-bcc0-ca4084acad69" containerID="b3bca8f0f7e189f11e49d72dcd8125ff55f26898af90be1e73e8266ec3f0fa2e" exitCode=0 Oct 05 08:06:27 crc kubenswrapper[4846]: I1005 08:06:27.839856 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-h76f7" Oct 05 08:06:27 crc kubenswrapper[4846]: I1005 08:06:27.839853 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-h76f7" event={"ID":"a242d9d6-e865-48c3-bcc0-ca4084acad69","Type":"ContainerDied","Data":"b3bca8f0f7e189f11e49d72dcd8125ff55f26898af90be1e73e8266ec3f0fa2e"} Oct 05 08:06:27 crc kubenswrapper[4846]: I1005 08:06:27.840438 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-h76f7" event={"ID":"a242d9d6-e865-48c3-bcc0-ca4084acad69","Type":"ContainerDied","Data":"3d8113bb8247e92b2508fcce6466a578fa535cfc736ef5be7d4814e7f13e3785"} Oct 05 08:06:27 crc kubenswrapper[4846]: I1005 08:06:27.840493 4846 scope.go:117] "RemoveContainer" containerID="b3bca8f0f7e189f11e49d72dcd8125ff55f26898af90be1e73e8266ec3f0fa2e" Oct 05 08:06:27 crc kubenswrapper[4846]: I1005 08:06:27.892318 4846 scope.go:117] "RemoveContainer" containerID="6ce7d39854e8e55e7757003e846b0cdee8a724513eb7f483ee82022f558a6045" Oct 05 08:06:27 crc kubenswrapper[4846]: I1005 08:06:27.897441 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-h76f7"] Oct 05 08:06:27 crc kubenswrapper[4846]: I1005 08:06:27.911884 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-h76f7"] Oct 05 08:06:27 crc kubenswrapper[4846]: I1005 08:06:27.919781 4846 scope.go:117] "RemoveContainer" containerID="87bf61f7d51f07f4353dea295ef025b9369270cb2a4a5a701063f339b58f4c9c" Oct 05 08:06:27 crc kubenswrapper[4846]: I1005 08:06:27.960902 4846 scope.go:117] "RemoveContainer" containerID="b3bca8f0f7e189f11e49d72dcd8125ff55f26898af90be1e73e8266ec3f0fa2e" Oct 05 08:06:27 crc kubenswrapper[4846]: E1005 08:06:27.961467 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b3bca8f0f7e189f11e49d72dcd8125ff55f26898af90be1e73e8266ec3f0fa2e\": container with ID starting with b3bca8f0f7e189f11e49d72dcd8125ff55f26898af90be1e73e8266ec3f0fa2e not found: ID does not exist" containerID="b3bca8f0f7e189f11e49d72dcd8125ff55f26898af90be1e73e8266ec3f0fa2e" Oct 05 08:06:27 crc kubenswrapper[4846]: I1005 08:06:27.961503 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b3bca8f0f7e189f11e49d72dcd8125ff55f26898af90be1e73e8266ec3f0fa2e"} err="failed to get container status \"b3bca8f0f7e189f11e49d72dcd8125ff55f26898af90be1e73e8266ec3f0fa2e\": rpc error: code = NotFound desc = could not find container \"b3bca8f0f7e189f11e49d72dcd8125ff55f26898af90be1e73e8266ec3f0fa2e\": container with ID starting with b3bca8f0f7e189f11e49d72dcd8125ff55f26898af90be1e73e8266ec3f0fa2e not found: ID does not exist" Oct 05 08:06:27 crc kubenswrapper[4846]: I1005 08:06:27.961528 4846 scope.go:117] "RemoveContainer" containerID="6ce7d39854e8e55e7757003e846b0cdee8a724513eb7f483ee82022f558a6045" Oct 05 08:06:27 crc kubenswrapper[4846]: E1005 08:06:27.962054 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6ce7d39854e8e55e7757003e846b0cdee8a724513eb7f483ee82022f558a6045\": container with ID starting with 6ce7d39854e8e55e7757003e846b0cdee8a724513eb7f483ee82022f558a6045 not found: ID does not exist" containerID="6ce7d39854e8e55e7757003e846b0cdee8a724513eb7f483ee82022f558a6045" Oct 05 08:06:27 crc kubenswrapper[4846]: I1005 08:06:27.962081 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6ce7d39854e8e55e7757003e846b0cdee8a724513eb7f483ee82022f558a6045"} err="failed to get container status \"6ce7d39854e8e55e7757003e846b0cdee8a724513eb7f483ee82022f558a6045\": rpc error: code = NotFound desc = could not find container \"6ce7d39854e8e55e7757003e846b0cdee8a724513eb7f483ee82022f558a6045\": container with ID starting with 6ce7d39854e8e55e7757003e846b0cdee8a724513eb7f483ee82022f558a6045 not found: ID does not exist" Oct 05 08:06:27 crc kubenswrapper[4846]: I1005 08:06:27.962099 4846 scope.go:117] "RemoveContainer" containerID="87bf61f7d51f07f4353dea295ef025b9369270cb2a4a5a701063f339b58f4c9c" Oct 05 08:06:27 crc kubenswrapper[4846]: E1005 08:06:27.962556 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"87bf61f7d51f07f4353dea295ef025b9369270cb2a4a5a701063f339b58f4c9c\": container with ID starting with 87bf61f7d51f07f4353dea295ef025b9369270cb2a4a5a701063f339b58f4c9c not found: ID does not exist" containerID="87bf61f7d51f07f4353dea295ef025b9369270cb2a4a5a701063f339b58f4c9c" Oct 05 08:06:27 crc kubenswrapper[4846]: I1005 08:06:27.962581 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"87bf61f7d51f07f4353dea295ef025b9369270cb2a4a5a701063f339b58f4c9c"} err="failed to get container status \"87bf61f7d51f07f4353dea295ef025b9369270cb2a4a5a701063f339b58f4c9c\": rpc error: code = NotFound desc = could not find container \"87bf61f7d51f07f4353dea295ef025b9369270cb2a4a5a701063f339b58f4c9c\": container with ID starting with 87bf61f7d51f07f4353dea295ef025b9369270cb2a4a5a701063f339b58f4c9c not found: ID does not exist" Oct 05 08:06:28 crc kubenswrapper[4846]: I1005 08:06:28.513348 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a242d9d6-e865-48c3-bcc0-ca4084acad69" path="/var/lib/kubelet/pods/a242d9d6-e865-48c3-bcc0-ca4084acad69/volumes" Oct 05 08:06:31 crc kubenswrapper[4846]: I1005 08:06:31.034418 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-galera-0" Oct 05 08:06:31 crc kubenswrapper[4846]: I1005 08:06:31.034692 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-galera-0" Oct 05 08:06:31 crc kubenswrapper[4846]: I1005 08:06:31.102648 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-galera-0" Oct 05 08:06:31 crc kubenswrapper[4846]: I1005 08:06:31.957028 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-galera-0" Oct 05 08:06:51 crc kubenswrapper[4846]: I1005 08:06:51.064908 4846 generic.go:334] "Generic (PLEG): container finished" podID="a288716c-b683-47c0-bf44-c0bbc997da97" containerID="acf58c0d7da784189c7fe44a098cc3fcf43783cf328b1e35c6673ffa86b97407" exitCode=0 Oct 05 08:06:51 crc kubenswrapper[4846]: I1005 08:06:51.065043 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"a288716c-b683-47c0-bf44-c0bbc997da97","Type":"ContainerDied","Data":"acf58c0d7da784189c7fe44a098cc3fcf43783cf328b1e35c6673ffa86b97407"} Oct 05 08:06:51 crc kubenswrapper[4846]: I1005 08:06:51.075687 4846 generic.go:334] "Generic (PLEG): container finished" podID="0f8171a9-82f5-443a-a8fc-d42b575d498a" containerID="2db7c55ce2a5c2dd4b47e3a9f2c447ed7148777d04a0deeecedd8f7015c83e5e" exitCode=0 Oct 05 08:06:51 crc kubenswrapper[4846]: I1005 08:06:51.075769 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"0f8171a9-82f5-443a-a8fc-d42b575d498a","Type":"ContainerDied","Data":"2db7c55ce2a5c2dd4b47e3a9f2c447ed7148777d04a0deeecedd8f7015c83e5e"} Oct 05 08:06:52 crc kubenswrapper[4846]: I1005 08:06:52.109880 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"0f8171a9-82f5-443a-a8fc-d42b575d498a","Type":"ContainerStarted","Data":"8a2ed397be99ca33dbab31eee6c0d309eb6aa160a01f425fb7b56c7b0fafea38"} Oct 05 08:06:52 crc kubenswrapper[4846]: I1005 08:06:52.110551 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Oct 05 08:06:52 crc kubenswrapper[4846]: I1005 08:06:52.114384 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"a288716c-b683-47c0-bf44-c0bbc997da97","Type":"ContainerStarted","Data":"108ce2b65180236e38b6ea4b73d17aee3f7109fbcf46a4478eb708a7280488d3"} Oct 05 08:06:52 crc kubenswrapper[4846]: I1005 08:06:52.114676 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Oct 05 08:06:52 crc kubenswrapper[4846]: I1005 08:06:52.179401 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=39.519157107 podStartE2EDuration="54.179372793s" podCreationTimestamp="2025-10-05 08:05:58 +0000 UTC" firstStartedPulling="2025-10-05 08:06:01.269270898 +0000 UTC m=+4683.510123663" lastFinishedPulling="2025-10-05 08:06:15.929486584 +0000 UTC m=+4698.170339349" observedRunningTime="2025-10-05 08:06:52.173838145 +0000 UTC m=+4734.414690920" watchObservedRunningTime="2025-10-05 08:06:52.179372793 +0000 UTC m=+4734.420225578" Oct 05 08:06:52 crc kubenswrapper[4846]: I1005 08:06:52.184786 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=41.902110296000004 podStartE2EDuration="54.184768308s" podCreationTimestamp="2025-10-05 08:05:58 +0000 UTC" firstStartedPulling="2025-10-05 08:06:00.903141504 +0000 UTC m=+4683.143994279" lastFinishedPulling="2025-10-05 08:06:13.185799516 +0000 UTC m=+4695.426652291" observedRunningTime="2025-10-05 08:06:52.14453631 +0000 UTC m=+4734.385389095" watchObservedRunningTime="2025-10-05 08:06:52.184768308 +0000 UTC m=+4734.425621103" Oct 05 08:06:53 crc kubenswrapper[4846]: I1005 08:06:53.325293 4846 patch_prober.go:28] interesting pod/machine-config-daemon-fscvf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 08:06:53 crc kubenswrapper[4846]: I1005 08:06:53.325790 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 08:07:09 crc kubenswrapper[4846]: I1005 08:07:09.895786 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Oct 05 08:07:09 crc kubenswrapper[4846]: I1005 08:07:09.951435 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Oct 05 08:07:16 crc kubenswrapper[4846]: I1005 08:07:16.575141 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-96d5866c7-cpp4z"] Oct 05 08:07:16 crc kubenswrapper[4846]: E1005 08:07:16.576179 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a242d9d6-e865-48c3-bcc0-ca4084acad69" containerName="extract-content" Oct 05 08:07:16 crc kubenswrapper[4846]: I1005 08:07:16.576219 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="a242d9d6-e865-48c3-bcc0-ca4084acad69" containerName="extract-content" Oct 05 08:07:16 crc kubenswrapper[4846]: E1005 08:07:16.576232 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="037a0419-1e47-429c-ac6f-e2fe6e7db5d9" containerName="init" Oct 05 08:07:16 crc kubenswrapper[4846]: I1005 08:07:16.576242 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="037a0419-1e47-429c-ac6f-e2fe6e7db5d9" containerName="init" Oct 05 08:07:16 crc kubenswrapper[4846]: E1005 08:07:16.576265 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a242d9d6-e865-48c3-bcc0-ca4084acad69" containerName="extract-utilities" Oct 05 08:07:16 crc kubenswrapper[4846]: I1005 08:07:16.576274 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="a242d9d6-e865-48c3-bcc0-ca4084acad69" containerName="extract-utilities" Oct 05 08:07:16 crc kubenswrapper[4846]: E1005 08:07:16.576293 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d6d6783a-7816-402e-b38e-935ffcb7314e" containerName="init" Oct 05 08:07:16 crc kubenswrapper[4846]: I1005 08:07:16.576301 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="d6d6783a-7816-402e-b38e-935ffcb7314e" containerName="init" Oct 05 08:07:16 crc kubenswrapper[4846]: E1005 08:07:16.576315 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ebf7ce11-32a8-4fea-9488-8f48b08991a9" containerName="init" Oct 05 08:07:16 crc kubenswrapper[4846]: I1005 08:07:16.576325 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="ebf7ce11-32a8-4fea-9488-8f48b08991a9" containerName="init" Oct 05 08:07:16 crc kubenswrapper[4846]: E1005 08:07:16.576343 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a242d9d6-e865-48c3-bcc0-ca4084acad69" containerName="registry-server" Oct 05 08:07:16 crc kubenswrapper[4846]: I1005 08:07:16.576354 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="a242d9d6-e865-48c3-bcc0-ca4084acad69" containerName="registry-server" Oct 05 08:07:16 crc kubenswrapper[4846]: E1005 08:07:16.576379 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ebf7ce11-32a8-4fea-9488-8f48b08991a9" containerName="dnsmasq-dns" Oct 05 08:07:16 crc kubenswrapper[4846]: I1005 08:07:16.576390 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="ebf7ce11-32a8-4fea-9488-8f48b08991a9" containerName="dnsmasq-dns" Oct 05 08:07:16 crc kubenswrapper[4846]: I1005 08:07:16.576569 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="037a0419-1e47-429c-ac6f-e2fe6e7db5d9" containerName="init" Oct 05 08:07:16 crc kubenswrapper[4846]: I1005 08:07:16.576581 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="d6d6783a-7816-402e-b38e-935ffcb7314e" containerName="init" Oct 05 08:07:16 crc kubenswrapper[4846]: I1005 08:07:16.576599 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="a242d9d6-e865-48c3-bcc0-ca4084acad69" containerName="registry-server" Oct 05 08:07:16 crc kubenswrapper[4846]: I1005 08:07:16.576626 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="ebf7ce11-32a8-4fea-9488-8f48b08991a9" containerName="dnsmasq-dns" Oct 05 08:07:16 crc kubenswrapper[4846]: I1005 08:07:16.577826 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-96d5866c7-cpp4z" Oct 05 08:07:16 crc kubenswrapper[4846]: I1005 08:07:16.596768 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-96d5866c7-cpp4z"] Oct 05 08:07:16 crc kubenswrapper[4846]: I1005 08:07:16.692336 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2001fca1-97a6-4420-ad29-3b83ab632284-dns-svc\") pod \"dnsmasq-dns-96d5866c7-cpp4z\" (UID: \"2001fca1-97a6-4420-ad29-3b83ab632284\") " pod="openstack/dnsmasq-dns-96d5866c7-cpp4z" Oct 05 08:07:16 crc kubenswrapper[4846]: I1005 08:07:16.692821 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-psx4k\" (UniqueName: \"kubernetes.io/projected/2001fca1-97a6-4420-ad29-3b83ab632284-kube-api-access-psx4k\") pod \"dnsmasq-dns-96d5866c7-cpp4z\" (UID: \"2001fca1-97a6-4420-ad29-3b83ab632284\") " pod="openstack/dnsmasq-dns-96d5866c7-cpp4z" Oct 05 08:07:16 crc kubenswrapper[4846]: I1005 08:07:16.692884 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2001fca1-97a6-4420-ad29-3b83ab632284-config\") pod \"dnsmasq-dns-96d5866c7-cpp4z\" (UID: \"2001fca1-97a6-4420-ad29-3b83ab632284\") " pod="openstack/dnsmasq-dns-96d5866c7-cpp4z" Oct 05 08:07:16 crc kubenswrapper[4846]: I1005 08:07:16.794496 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-psx4k\" (UniqueName: \"kubernetes.io/projected/2001fca1-97a6-4420-ad29-3b83ab632284-kube-api-access-psx4k\") pod \"dnsmasq-dns-96d5866c7-cpp4z\" (UID: \"2001fca1-97a6-4420-ad29-3b83ab632284\") " pod="openstack/dnsmasq-dns-96d5866c7-cpp4z" Oct 05 08:07:16 crc kubenswrapper[4846]: I1005 08:07:16.794625 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2001fca1-97a6-4420-ad29-3b83ab632284-config\") pod \"dnsmasq-dns-96d5866c7-cpp4z\" (UID: \"2001fca1-97a6-4420-ad29-3b83ab632284\") " pod="openstack/dnsmasq-dns-96d5866c7-cpp4z" Oct 05 08:07:16 crc kubenswrapper[4846]: I1005 08:07:16.794709 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2001fca1-97a6-4420-ad29-3b83ab632284-dns-svc\") pod \"dnsmasq-dns-96d5866c7-cpp4z\" (UID: \"2001fca1-97a6-4420-ad29-3b83ab632284\") " pod="openstack/dnsmasq-dns-96d5866c7-cpp4z" Oct 05 08:07:16 crc kubenswrapper[4846]: I1005 08:07:16.795742 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2001fca1-97a6-4420-ad29-3b83ab632284-dns-svc\") pod \"dnsmasq-dns-96d5866c7-cpp4z\" (UID: \"2001fca1-97a6-4420-ad29-3b83ab632284\") " pod="openstack/dnsmasq-dns-96d5866c7-cpp4z" Oct 05 08:07:16 crc kubenswrapper[4846]: I1005 08:07:16.795781 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2001fca1-97a6-4420-ad29-3b83ab632284-config\") pod \"dnsmasq-dns-96d5866c7-cpp4z\" (UID: \"2001fca1-97a6-4420-ad29-3b83ab632284\") " pod="openstack/dnsmasq-dns-96d5866c7-cpp4z" Oct 05 08:07:16 crc kubenswrapper[4846]: I1005 08:07:16.826185 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-psx4k\" (UniqueName: \"kubernetes.io/projected/2001fca1-97a6-4420-ad29-3b83ab632284-kube-api-access-psx4k\") pod \"dnsmasq-dns-96d5866c7-cpp4z\" (UID: \"2001fca1-97a6-4420-ad29-3b83ab632284\") " pod="openstack/dnsmasq-dns-96d5866c7-cpp4z" Oct 05 08:07:16 crc kubenswrapper[4846]: I1005 08:07:16.915750 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-96d5866c7-cpp4z" Oct 05 08:07:17 crc kubenswrapper[4846]: I1005 08:07:17.445286 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-96d5866c7-cpp4z"] Oct 05 08:07:17 crc kubenswrapper[4846]: I1005 08:07:17.532424 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 05 08:07:18 crc kubenswrapper[4846]: I1005 08:07:18.048864 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 05 08:07:18 crc kubenswrapper[4846]: I1005 08:07:18.359780 4846 generic.go:334] "Generic (PLEG): container finished" podID="2001fca1-97a6-4420-ad29-3b83ab632284" containerID="af710e4cd5ecb39d20b237388f1eccc6ee4c8d63599d39ae3fcc508fa36c1fe6" exitCode=0 Oct 05 08:07:18 crc kubenswrapper[4846]: I1005 08:07:18.359860 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-96d5866c7-cpp4z" event={"ID":"2001fca1-97a6-4420-ad29-3b83ab632284","Type":"ContainerDied","Data":"af710e4cd5ecb39d20b237388f1eccc6ee4c8d63599d39ae3fcc508fa36c1fe6"} Oct 05 08:07:18 crc kubenswrapper[4846]: I1005 08:07:18.359906 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-96d5866c7-cpp4z" event={"ID":"2001fca1-97a6-4420-ad29-3b83ab632284","Type":"ContainerStarted","Data":"e2497b036f252756187de8b44fb734a3c7b5ca587b07a2b4ffc45b9dce57edc7"} Oct 05 08:07:19 crc kubenswrapper[4846]: I1005 08:07:19.368800 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-96d5866c7-cpp4z" event={"ID":"2001fca1-97a6-4420-ad29-3b83ab632284","Type":"ContainerStarted","Data":"9d28829bb3200c54bbf39175a22dc488c7bca2383ac6674b8c071b500c3299c4"} Oct 05 08:07:19 crc kubenswrapper[4846]: I1005 08:07:19.370215 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-96d5866c7-cpp4z" Oct 05 08:07:19 crc kubenswrapper[4846]: I1005 08:07:19.391154 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-96d5866c7-cpp4z" podStartSLOduration=3.391125677 podStartE2EDuration="3.391125677s" podCreationTimestamp="2025-10-05 08:07:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 08:07:19.384593812 +0000 UTC m=+4761.625446597" watchObservedRunningTime="2025-10-05 08:07:19.391125677 +0000 UTC m=+4761.631978492" Oct 05 08:07:22 crc kubenswrapper[4846]: I1005 08:07:22.264457 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-cell1-server-0" podUID="a288716c-b683-47c0-bf44-c0bbc997da97" containerName="rabbitmq" containerID="cri-o://108ce2b65180236e38b6ea4b73d17aee3f7109fbcf46a4478eb708a7280488d3" gracePeriod=604796 Oct 05 08:07:22 crc kubenswrapper[4846]: I1005 08:07:22.323973 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-server-0" podUID="0f8171a9-82f5-443a-a8fc-d42b575d498a" containerName="rabbitmq" containerID="cri-o://8a2ed397be99ca33dbab31eee6c0d309eb6aa160a01f425fb7b56c7b0fafea38" gracePeriod=604796 Oct 05 08:07:23 crc kubenswrapper[4846]: I1005 08:07:23.325996 4846 patch_prober.go:28] interesting pod/machine-config-daemon-fscvf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 08:07:23 crc kubenswrapper[4846]: I1005 08:07:23.326121 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 08:07:23 crc kubenswrapper[4846]: I1005 08:07:23.326254 4846 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" Oct 05 08:07:23 crc kubenswrapper[4846]: I1005 08:07:23.327533 4846 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"bad87dae9648a1776fbb2ee38b4fedce1273888d4f401172760726d09bb965f3"} pod="openshift-machine-config-operator/machine-config-daemon-fscvf" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 05 08:07:23 crc kubenswrapper[4846]: I1005 08:07:23.327673 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" containerName="machine-config-daemon" containerID="cri-o://bad87dae9648a1776fbb2ee38b4fedce1273888d4f401172760726d09bb965f3" gracePeriod=600 Oct 05 08:07:24 crc kubenswrapper[4846]: I1005 08:07:24.420224 4846 generic.go:334] "Generic (PLEG): container finished" podID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" containerID="bad87dae9648a1776fbb2ee38b4fedce1273888d4f401172760726d09bb965f3" exitCode=0 Oct 05 08:07:24 crc kubenswrapper[4846]: I1005 08:07:24.420283 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" event={"ID":"ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7","Type":"ContainerDied","Data":"bad87dae9648a1776fbb2ee38b4fedce1273888d4f401172760726d09bb965f3"} Oct 05 08:07:24 crc kubenswrapper[4846]: I1005 08:07:24.420681 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" event={"ID":"ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7","Type":"ContainerStarted","Data":"8fa72be735dff8e623a4a4958f452896ef79e72dd44c2d714a3019187462de47"} Oct 05 08:07:24 crc kubenswrapper[4846]: I1005 08:07:24.420726 4846 scope.go:117] "RemoveContainer" containerID="2d2e59cb51a3cd322bc7aca627ddfa4d48c6f5f95b26a7264c472515babd8d10" Oct 05 08:07:26 crc kubenswrapper[4846]: I1005 08:07:26.918516 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-96d5866c7-cpp4z" Oct 05 08:07:27 crc kubenswrapper[4846]: I1005 08:07:27.013610 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6cfd7b4c45-gjjg8"] Oct 05 08:07:27 crc kubenswrapper[4846]: I1005 08:07:27.013975 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6cfd7b4c45-gjjg8" podUID="e4bbd7e3-1e67-4145-943d-fe911ac8f3ba" containerName="dnsmasq-dns" containerID="cri-o://91dc787a32b9b22d8b7fcd9ee93ac0d3a459f590ee48fbb2411b26a679ca11b2" gracePeriod=10 Oct 05 08:07:27 crc kubenswrapper[4846]: I1005 08:07:27.458048 4846 generic.go:334] "Generic (PLEG): container finished" podID="e4bbd7e3-1e67-4145-943d-fe911ac8f3ba" containerID="91dc787a32b9b22d8b7fcd9ee93ac0d3a459f590ee48fbb2411b26a679ca11b2" exitCode=0 Oct 05 08:07:27 crc kubenswrapper[4846]: I1005 08:07:27.458138 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6cfd7b4c45-gjjg8" event={"ID":"e4bbd7e3-1e67-4145-943d-fe911ac8f3ba","Type":"ContainerDied","Data":"91dc787a32b9b22d8b7fcd9ee93ac0d3a459f590ee48fbb2411b26a679ca11b2"} Oct 05 08:07:27 crc kubenswrapper[4846]: I1005 08:07:27.458422 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6cfd7b4c45-gjjg8" event={"ID":"e4bbd7e3-1e67-4145-943d-fe911ac8f3ba","Type":"ContainerDied","Data":"6e5d96123807b9dd396efecc22ea1f353c095e39422045835342c86cf7641290"} Oct 05 08:07:27 crc kubenswrapper[4846]: I1005 08:07:27.458441 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6e5d96123807b9dd396efecc22ea1f353c095e39422045835342c86cf7641290" Oct 05 08:07:27 crc kubenswrapper[4846]: I1005 08:07:27.492704 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6cfd7b4c45-gjjg8" Oct 05 08:07:27 crc kubenswrapper[4846]: I1005 08:07:27.575979 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6hslb\" (UniqueName: \"kubernetes.io/projected/e4bbd7e3-1e67-4145-943d-fe911ac8f3ba-kube-api-access-6hslb\") pod \"e4bbd7e3-1e67-4145-943d-fe911ac8f3ba\" (UID: \"e4bbd7e3-1e67-4145-943d-fe911ac8f3ba\") " Oct 05 08:07:27 crc kubenswrapper[4846]: I1005 08:07:27.576066 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e4bbd7e3-1e67-4145-943d-fe911ac8f3ba-dns-svc\") pod \"e4bbd7e3-1e67-4145-943d-fe911ac8f3ba\" (UID: \"e4bbd7e3-1e67-4145-943d-fe911ac8f3ba\") " Oct 05 08:07:27 crc kubenswrapper[4846]: I1005 08:07:27.576135 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e4bbd7e3-1e67-4145-943d-fe911ac8f3ba-config\") pod \"e4bbd7e3-1e67-4145-943d-fe911ac8f3ba\" (UID: \"e4bbd7e3-1e67-4145-943d-fe911ac8f3ba\") " Oct 05 08:07:27 crc kubenswrapper[4846]: I1005 08:07:27.581125 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e4bbd7e3-1e67-4145-943d-fe911ac8f3ba-kube-api-access-6hslb" (OuterVolumeSpecName: "kube-api-access-6hslb") pod "e4bbd7e3-1e67-4145-943d-fe911ac8f3ba" (UID: "e4bbd7e3-1e67-4145-943d-fe911ac8f3ba"). InnerVolumeSpecName "kube-api-access-6hslb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:07:27 crc kubenswrapper[4846]: I1005 08:07:27.617321 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e4bbd7e3-1e67-4145-943d-fe911ac8f3ba-config" (OuterVolumeSpecName: "config") pod "e4bbd7e3-1e67-4145-943d-fe911ac8f3ba" (UID: "e4bbd7e3-1e67-4145-943d-fe911ac8f3ba"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 08:07:27 crc kubenswrapper[4846]: I1005 08:07:27.617537 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e4bbd7e3-1e67-4145-943d-fe911ac8f3ba-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "e4bbd7e3-1e67-4145-943d-fe911ac8f3ba" (UID: "e4bbd7e3-1e67-4145-943d-fe911ac8f3ba"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 08:07:27 crc kubenswrapper[4846]: I1005 08:07:27.677978 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6hslb\" (UniqueName: \"kubernetes.io/projected/e4bbd7e3-1e67-4145-943d-fe911ac8f3ba-kube-api-access-6hslb\") on node \"crc\" DevicePath \"\"" Oct 05 08:07:27 crc kubenswrapper[4846]: I1005 08:07:27.678026 4846 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e4bbd7e3-1e67-4145-943d-fe911ac8f3ba-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 05 08:07:27 crc kubenswrapper[4846]: I1005 08:07:27.678035 4846 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e4bbd7e3-1e67-4145-943d-fe911ac8f3ba-config\") on node \"crc\" DevicePath \"\"" Oct 05 08:07:28 crc kubenswrapper[4846]: I1005 08:07:28.481423 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6cfd7b4c45-gjjg8" Oct 05 08:07:28 crc kubenswrapper[4846]: I1005 08:07:28.659580 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6cfd7b4c45-gjjg8"] Oct 05 08:07:28 crc kubenswrapper[4846]: I1005 08:07:28.665424 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6cfd7b4c45-gjjg8"] Oct 05 08:07:29 crc kubenswrapper[4846]: I1005 08:07:29.324245 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 05 08:07:29 crc kubenswrapper[4846]: I1005 08:07:29.332703 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 05 08:07:29 crc kubenswrapper[4846]: I1005 08:07:29.415591 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/0f8171a9-82f5-443a-a8fc-d42b575d498a-rabbitmq-erlang-cookie\") pod \"0f8171a9-82f5-443a-a8fc-d42b575d498a\" (UID: \"0f8171a9-82f5-443a-a8fc-d42b575d498a\") " Oct 05 08:07:29 crc kubenswrapper[4846]: I1005 08:07:29.416500 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0f8171a9-82f5-443a-a8fc-d42b575d498a-config-data\") pod \"0f8171a9-82f5-443a-a8fc-d42b575d498a\" (UID: \"0f8171a9-82f5-443a-a8fc-d42b575d498a\") " Oct 05 08:07:29 crc kubenswrapper[4846]: I1005 08:07:29.416546 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/0f8171a9-82f5-443a-a8fc-d42b575d498a-erlang-cookie-secret\") pod \"0f8171a9-82f5-443a-a8fc-d42b575d498a\" (UID: \"0f8171a9-82f5-443a-a8fc-d42b575d498a\") " Oct 05 08:07:29 crc kubenswrapper[4846]: I1005 08:07:29.416568 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/a288716c-b683-47c0-bf44-c0bbc997da97-rabbitmq-confd\") pod \"a288716c-b683-47c0-bf44-c0bbc997da97\" (UID: \"a288716c-b683-47c0-bf44-c0bbc997da97\") " Oct 05 08:07:29 crc kubenswrapper[4846]: I1005 08:07:29.416595 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/a288716c-b683-47c0-bf44-c0bbc997da97-plugins-conf\") pod \"a288716c-b683-47c0-bf44-c0bbc997da97\" (UID: \"a288716c-b683-47c0-bf44-c0bbc997da97\") " Oct 05 08:07:29 crc kubenswrapper[4846]: I1005 08:07:29.416619 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/a288716c-b683-47c0-bf44-c0bbc997da97-rabbitmq-erlang-cookie\") pod \"a288716c-b683-47c0-bf44-c0bbc997da97\" (UID: \"a288716c-b683-47c0-bf44-c0bbc997da97\") " Oct 05 08:07:29 crc kubenswrapper[4846]: I1005 08:07:29.416643 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/0f8171a9-82f5-443a-a8fc-d42b575d498a-plugins-conf\") pod \"0f8171a9-82f5-443a-a8fc-d42b575d498a\" (UID: \"0f8171a9-82f5-443a-a8fc-d42b575d498a\") " Oct 05 08:07:29 crc kubenswrapper[4846]: I1005 08:07:29.416666 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/0f8171a9-82f5-443a-a8fc-d42b575d498a-rabbitmq-tls\") pod \"0f8171a9-82f5-443a-a8fc-d42b575d498a\" (UID: \"0f8171a9-82f5-443a-a8fc-d42b575d498a\") " Oct 05 08:07:29 crc kubenswrapper[4846]: I1005 08:07:29.416702 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bdwsq\" (UniqueName: \"kubernetes.io/projected/a288716c-b683-47c0-bf44-c0bbc997da97-kube-api-access-bdwsq\") pod \"a288716c-b683-47c0-bf44-c0bbc997da97\" (UID: \"a288716c-b683-47c0-bf44-c0bbc997da97\") " Oct 05 08:07:29 crc kubenswrapper[4846]: I1005 08:07:29.416734 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/0f8171a9-82f5-443a-a8fc-d42b575d498a-rabbitmq-plugins\") pod \"0f8171a9-82f5-443a-a8fc-d42b575d498a\" (UID: \"0f8171a9-82f5-443a-a8fc-d42b575d498a\") " Oct 05 08:07:29 crc kubenswrapper[4846]: I1005 08:07:29.416762 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/0f8171a9-82f5-443a-a8fc-d42b575d498a-server-conf\") pod \"0f8171a9-82f5-443a-a8fc-d42b575d498a\" (UID: \"0f8171a9-82f5-443a-a8fc-d42b575d498a\") " Oct 05 08:07:29 crc kubenswrapper[4846]: I1005 08:07:29.416788 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/a288716c-b683-47c0-bf44-c0bbc997da97-erlang-cookie-secret\") pod \"a288716c-b683-47c0-bf44-c0bbc997da97\" (UID: \"a288716c-b683-47c0-bf44-c0bbc997da97\") " Oct 05 08:07:29 crc kubenswrapper[4846]: I1005 08:07:29.416817 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dx9rf\" (UniqueName: \"kubernetes.io/projected/0f8171a9-82f5-443a-a8fc-d42b575d498a-kube-api-access-dx9rf\") pod \"0f8171a9-82f5-443a-a8fc-d42b575d498a\" (UID: \"0f8171a9-82f5-443a-a8fc-d42b575d498a\") " Oct 05 08:07:29 crc kubenswrapper[4846]: I1005 08:07:29.416846 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a288716c-b683-47c0-bf44-c0bbc997da97-config-data\") pod \"a288716c-b683-47c0-bf44-c0bbc997da97\" (UID: \"a288716c-b683-47c0-bf44-c0bbc997da97\") " Oct 05 08:07:29 crc kubenswrapper[4846]: I1005 08:07:29.416868 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/0f8171a9-82f5-443a-a8fc-d42b575d498a-pod-info\") pod \"0f8171a9-82f5-443a-a8fc-d42b575d498a\" (UID: \"0f8171a9-82f5-443a-a8fc-d42b575d498a\") " Oct 05 08:07:29 crc kubenswrapper[4846]: I1005 08:07:29.416898 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/a288716c-b683-47c0-bf44-c0bbc997da97-pod-info\") pod \"a288716c-b683-47c0-bf44-c0bbc997da97\" (UID: \"a288716c-b683-47c0-bf44-c0bbc997da97\") " Oct 05 08:07:29 crc kubenswrapper[4846]: I1005 08:07:29.416934 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/a288716c-b683-47c0-bf44-c0bbc997da97-rabbitmq-plugins\") pod \"a288716c-b683-47c0-bf44-c0bbc997da97\" (UID: \"a288716c-b683-47c0-bf44-c0bbc997da97\") " Oct 05 08:07:29 crc kubenswrapper[4846]: I1005 08:07:29.417026 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-c0fb7d88-e3e3-43b2-a6ce-f0d1fa614bdb\") pod \"0f8171a9-82f5-443a-a8fc-d42b575d498a\" (UID: \"0f8171a9-82f5-443a-a8fc-d42b575d498a\") " Oct 05 08:07:29 crc kubenswrapper[4846]: I1005 08:07:29.417050 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/0f8171a9-82f5-443a-a8fc-d42b575d498a-rabbitmq-confd\") pod \"0f8171a9-82f5-443a-a8fc-d42b575d498a\" (UID: \"0f8171a9-82f5-443a-a8fc-d42b575d498a\") " Oct 05 08:07:29 crc kubenswrapper[4846]: I1005 08:07:29.417098 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/a288716c-b683-47c0-bf44-c0bbc997da97-server-conf\") pod \"a288716c-b683-47c0-bf44-c0bbc997da97\" (UID: \"a288716c-b683-47c0-bf44-c0bbc997da97\") " Oct 05 08:07:29 crc kubenswrapper[4846]: I1005 08:07:29.417139 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/a288716c-b683-47c0-bf44-c0bbc997da97-rabbitmq-tls\") pod \"a288716c-b683-47c0-bf44-c0bbc997da97\" (UID: \"a288716c-b683-47c0-bf44-c0bbc997da97\") " Oct 05 08:07:29 crc kubenswrapper[4846]: I1005 08:07:29.417275 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-0b1f215a-31af-4eb1-a3f9-ccbfe14235d1\") pod \"a288716c-b683-47c0-bf44-c0bbc997da97\" (UID: \"a288716c-b683-47c0-bf44-c0bbc997da97\") " Oct 05 08:07:29 crc kubenswrapper[4846]: I1005 08:07:29.416453 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0f8171a9-82f5-443a-a8fc-d42b575d498a-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "0f8171a9-82f5-443a-a8fc-d42b575d498a" (UID: "0f8171a9-82f5-443a-a8fc-d42b575d498a"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 08:07:29 crc kubenswrapper[4846]: I1005 08:07:29.419272 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a288716c-b683-47c0-bf44-c0bbc997da97-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "a288716c-b683-47c0-bf44-c0bbc997da97" (UID: "a288716c-b683-47c0-bf44-c0bbc997da97"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 08:07:29 crc kubenswrapper[4846]: I1005 08:07:29.419662 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a288716c-b683-47c0-bf44-c0bbc997da97-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "a288716c-b683-47c0-bf44-c0bbc997da97" (UID: "a288716c-b683-47c0-bf44-c0bbc997da97"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 08:07:29 crc kubenswrapper[4846]: I1005 08:07:29.430172 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/a288716c-b683-47c0-bf44-c0bbc997da97-pod-info" (OuterVolumeSpecName: "pod-info") pod "a288716c-b683-47c0-bf44-c0bbc997da97" (UID: "a288716c-b683-47c0-bf44-c0bbc997da97"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Oct 05 08:07:29 crc kubenswrapper[4846]: I1005 08:07:29.431146 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/0f8171a9-82f5-443a-a8fc-d42b575d498a-pod-info" (OuterVolumeSpecName: "pod-info") pod "0f8171a9-82f5-443a-a8fc-d42b575d498a" (UID: "0f8171a9-82f5-443a-a8fc-d42b575d498a"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Oct 05 08:07:29 crc kubenswrapper[4846]: I1005 08:07:29.432027 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0f8171a9-82f5-443a-a8fc-d42b575d498a-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "0f8171a9-82f5-443a-a8fc-d42b575d498a" (UID: "0f8171a9-82f5-443a-a8fc-d42b575d498a"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 08:07:29 crc kubenswrapper[4846]: I1005 08:07:29.432564 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a288716c-b683-47c0-bf44-c0bbc997da97-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "a288716c-b683-47c0-bf44-c0bbc997da97" (UID: "a288716c-b683-47c0-bf44-c0bbc997da97"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:07:29 crc kubenswrapper[4846]: I1005 08:07:29.433125 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a288716c-b683-47c0-bf44-c0bbc997da97-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "a288716c-b683-47c0-bf44-c0bbc997da97" (UID: "a288716c-b683-47c0-bf44-c0bbc997da97"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 08:07:29 crc kubenswrapper[4846]: I1005 08:07:29.433390 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0f8171a9-82f5-443a-a8fc-d42b575d498a-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "0f8171a9-82f5-443a-a8fc-d42b575d498a" (UID: "0f8171a9-82f5-443a-a8fc-d42b575d498a"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:07:29 crc kubenswrapper[4846]: I1005 08:07:29.433848 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0f8171a9-82f5-443a-a8fc-d42b575d498a-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "0f8171a9-82f5-443a-a8fc-d42b575d498a" (UID: "0f8171a9-82f5-443a-a8fc-d42b575d498a"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 08:07:29 crc kubenswrapper[4846]: I1005 08:07:29.434328 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0f8171a9-82f5-443a-a8fc-d42b575d498a-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "0f8171a9-82f5-443a-a8fc-d42b575d498a" (UID: "0f8171a9-82f5-443a-a8fc-d42b575d498a"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:07:29 crc kubenswrapper[4846]: I1005 08:07:29.446220 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0f8171a9-82f5-443a-a8fc-d42b575d498a-kube-api-access-dx9rf" (OuterVolumeSpecName: "kube-api-access-dx9rf") pod "0f8171a9-82f5-443a-a8fc-d42b575d498a" (UID: "0f8171a9-82f5-443a-a8fc-d42b575d498a"). InnerVolumeSpecName "kube-api-access-dx9rf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:07:29 crc kubenswrapper[4846]: I1005 08:07:29.451371 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a288716c-b683-47c0-bf44-c0bbc997da97-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "a288716c-b683-47c0-bf44-c0bbc997da97" (UID: "a288716c-b683-47c0-bf44-c0bbc997da97"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:07:29 crc kubenswrapper[4846]: I1005 08:07:29.456566 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a288716c-b683-47c0-bf44-c0bbc997da97-kube-api-access-bdwsq" (OuterVolumeSpecName: "kube-api-access-bdwsq") pod "a288716c-b683-47c0-bf44-c0bbc997da97" (UID: "a288716c-b683-47c0-bf44-c0bbc997da97"). InnerVolumeSpecName "kube-api-access-bdwsq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:07:29 crc kubenswrapper[4846]: I1005 08:07:29.465919 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0f8171a9-82f5-443a-a8fc-d42b575d498a-config-data" (OuterVolumeSpecName: "config-data") pod "0f8171a9-82f5-443a-a8fc-d42b575d498a" (UID: "0f8171a9-82f5-443a-a8fc-d42b575d498a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 08:07:29 crc kubenswrapper[4846]: I1005 08:07:29.478890 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-c0fb7d88-e3e3-43b2-a6ce-f0d1fa614bdb" (OuterVolumeSpecName: "persistence") pod "0f8171a9-82f5-443a-a8fc-d42b575d498a" (UID: "0f8171a9-82f5-443a-a8fc-d42b575d498a"). InnerVolumeSpecName "pvc-c0fb7d88-e3e3-43b2-a6ce-f0d1fa614bdb". PluginName "kubernetes.io/csi", VolumeGidValue "" Oct 05 08:07:29 crc kubenswrapper[4846]: I1005 08:07:29.479259 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-0b1f215a-31af-4eb1-a3f9-ccbfe14235d1" (OuterVolumeSpecName: "persistence") pod "a288716c-b683-47c0-bf44-c0bbc997da97" (UID: "a288716c-b683-47c0-bf44-c0bbc997da97"). InnerVolumeSpecName "pvc-0b1f215a-31af-4eb1-a3f9-ccbfe14235d1". PluginName "kubernetes.io/csi", VolumeGidValue "" Oct 05 08:07:29 crc kubenswrapper[4846]: I1005 08:07:29.494723 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a288716c-b683-47c0-bf44-c0bbc997da97-config-data" (OuterVolumeSpecName: "config-data") pod "a288716c-b683-47c0-bf44-c0bbc997da97" (UID: "a288716c-b683-47c0-bf44-c0bbc997da97"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 08:07:29 crc kubenswrapper[4846]: I1005 08:07:29.496745 4846 generic.go:334] "Generic (PLEG): container finished" podID="0f8171a9-82f5-443a-a8fc-d42b575d498a" containerID="8a2ed397be99ca33dbab31eee6c0d309eb6aa160a01f425fb7b56c7b0fafea38" exitCode=0 Oct 05 08:07:29 crc kubenswrapper[4846]: I1005 08:07:29.496862 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"0f8171a9-82f5-443a-a8fc-d42b575d498a","Type":"ContainerDied","Data":"8a2ed397be99ca33dbab31eee6c0d309eb6aa160a01f425fb7b56c7b0fafea38"} Oct 05 08:07:29 crc kubenswrapper[4846]: I1005 08:07:29.496897 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"0f8171a9-82f5-443a-a8fc-d42b575d498a","Type":"ContainerDied","Data":"50e4eb4686ae493476e9c60dc3ad1b3556039dd5c0694bc9a4853855b11e27be"} Oct 05 08:07:29 crc kubenswrapper[4846]: I1005 08:07:29.496917 4846 scope.go:117] "RemoveContainer" containerID="8a2ed397be99ca33dbab31eee6c0d309eb6aa160a01f425fb7b56c7b0fafea38" Oct 05 08:07:29 crc kubenswrapper[4846]: I1005 08:07:29.497990 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 05 08:07:29 crc kubenswrapper[4846]: I1005 08:07:29.503322 4846 generic.go:334] "Generic (PLEG): container finished" podID="a288716c-b683-47c0-bf44-c0bbc997da97" containerID="108ce2b65180236e38b6ea4b73d17aee3f7109fbcf46a4478eb708a7280488d3" exitCode=0 Oct 05 08:07:29 crc kubenswrapper[4846]: I1005 08:07:29.503354 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"a288716c-b683-47c0-bf44-c0bbc997da97","Type":"ContainerDied","Data":"108ce2b65180236e38b6ea4b73d17aee3f7109fbcf46a4478eb708a7280488d3"} Oct 05 08:07:29 crc kubenswrapper[4846]: I1005 08:07:29.503389 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"a288716c-b683-47c0-bf44-c0bbc997da97","Type":"ContainerDied","Data":"3c0f92164a082c2da547bc954edbe40f8475270017609029921e3d666bbda069"} Oct 05 08:07:29 crc kubenswrapper[4846]: I1005 08:07:29.503447 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 05 08:07:29 crc kubenswrapper[4846]: I1005 08:07:29.505615 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a288716c-b683-47c0-bf44-c0bbc997da97-server-conf" (OuterVolumeSpecName: "server-conf") pod "a288716c-b683-47c0-bf44-c0bbc997da97" (UID: "a288716c-b683-47c0-bf44-c0bbc997da97"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 08:07:29 crc kubenswrapper[4846]: I1005 08:07:29.519124 4846 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/a288716c-b683-47c0-bf44-c0bbc997da97-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Oct 05 08:07:29 crc kubenswrapper[4846]: I1005 08:07:29.519153 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dx9rf\" (UniqueName: \"kubernetes.io/projected/0f8171a9-82f5-443a-a8fc-d42b575d498a-kube-api-access-dx9rf\") on node \"crc\" DevicePath \"\"" Oct 05 08:07:29 crc kubenswrapper[4846]: I1005 08:07:29.519166 4846 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a288716c-b683-47c0-bf44-c0bbc997da97-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 08:07:29 crc kubenswrapper[4846]: I1005 08:07:29.519187 4846 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/0f8171a9-82f5-443a-a8fc-d42b575d498a-pod-info\") on node \"crc\" DevicePath \"\"" Oct 05 08:07:29 crc kubenswrapper[4846]: I1005 08:07:29.519195 4846 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/a288716c-b683-47c0-bf44-c0bbc997da97-pod-info\") on node \"crc\" DevicePath \"\"" Oct 05 08:07:29 crc kubenswrapper[4846]: I1005 08:07:29.519203 4846 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/a288716c-b683-47c0-bf44-c0bbc997da97-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Oct 05 08:07:29 crc kubenswrapper[4846]: I1005 08:07:29.519229 4846 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-c0fb7d88-e3e3-43b2-a6ce-f0d1fa614bdb\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-c0fb7d88-e3e3-43b2-a6ce-f0d1fa614bdb\") on node \"crc\" " Oct 05 08:07:29 crc kubenswrapper[4846]: I1005 08:07:29.519240 4846 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/a288716c-b683-47c0-bf44-c0bbc997da97-server-conf\") on node \"crc\" DevicePath \"\"" Oct 05 08:07:29 crc kubenswrapper[4846]: I1005 08:07:29.519249 4846 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/a288716c-b683-47c0-bf44-c0bbc997da97-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Oct 05 08:07:29 crc kubenswrapper[4846]: I1005 08:07:29.519264 4846 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-0b1f215a-31af-4eb1-a3f9-ccbfe14235d1\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-0b1f215a-31af-4eb1-a3f9-ccbfe14235d1\") on node \"crc\" " Oct 05 08:07:29 crc kubenswrapper[4846]: I1005 08:07:29.519273 4846 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/0f8171a9-82f5-443a-a8fc-d42b575d498a-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Oct 05 08:07:29 crc kubenswrapper[4846]: I1005 08:07:29.519282 4846 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0f8171a9-82f5-443a-a8fc-d42b575d498a-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 08:07:29 crc kubenswrapper[4846]: I1005 08:07:29.519290 4846 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/0f8171a9-82f5-443a-a8fc-d42b575d498a-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Oct 05 08:07:29 crc kubenswrapper[4846]: I1005 08:07:29.519298 4846 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/a288716c-b683-47c0-bf44-c0bbc997da97-plugins-conf\") on node \"crc\" DevicePath \"\"" Oct 05 08:07:29 crc kubenswrapper[4846]: I1005 08:07:29.519306 4846 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/a288716c-b683-47c0-bf44-c0bbc997da97-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Oct 05 08:07:29 crc kubenswrapper[4846]: I1005 08:07:29.519314 4846 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/0f8171a9-82f5-443a-a8fc-d42b575d498a-plugins-conf\") on node \"crc\" DevicePath \"\"" Oct 05 08:07:29 crc kubenswrapper[4846]: I1005 08:07:29.519323 4846 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/0f8171a9-82f5-443a-a8fc-d42b575d498a-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Oct 05 08:07:29 crc kubenswrapper[4846]: I1005 08:07:29.519331 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bdwsq\" (UniqueName: \"kubernetes.io/projected/a288716c-b683-47c0-bf44-c0bbc997da97-kube-api-access-bdwsq\") on node \"crc\" DevicePath \"\"" Oct 05 08:07:29 crc kubenswrapper[4846]: I1005 08:07:29.519339 4846 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/0f8171a9-82f5-443a-a8fc-d42b575d498a-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Oct 05 08:07:29 crc kubenswrapper[4846]: I1005 08:07:29.523653 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0f8171a9-82f5-443a-a8fc-d42b575d498a-server-conf" (OuterVolumeSpecName: "server-conf") pod "0f8171a9-82f5-443a-a8fc-d42b575d498a" (UID: "0f8171a9-82f5-443a-a8fc-d42b575d498a"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 08:07:29 crc kubenswrapper[4846]: I1005 08:07:29.536084 4846 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Oct 05 08:07:29 crc kubenswrapper[4846]: I1005 08:07:29.536343 4846 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-c0fb7d88-e3e3-43b2-a6ce-f0d1fa614bdb" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-c0fb7d88-e3e3-43b2-a6ce-f0d1fa614bdb") on node "crc" Oct 05 08:07:29 crc kubenswrapper[4846]: I1005 08:07:29.548068 4846 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Oct 05 08:07:29 crc kubenswrapper[4846]: I1005 08:07:29.548266 4846 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-0b1f215a-31af-4eb1-a3f9-ccbfe14235d1" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-0b1f215a-31af-4eb1-a3f9-ccbfe14235d1") on node "crc" Oct 05 08:07:29 crc kubenswrapper[4846]: I1005 08:07:29.560536 4846 scope.go:117] "RemoveContainer" containerID="2db7c55ce2a5c2dd4b47e3a9f2c447ed7148777d04a0deeecedd8f7015c83e5e" Oct 05 08:07:29 crc kubenswrapper[4846]: I1005 08:07:29.579470 4846 scope.go:117] "RemoveContainer" containerID="8a2ed397be99ca33dbab31eee6c0d309eb6aa160a01f425fb7b56c7b0fafea38" Oct 05 08:07:29 crc kubenswrapper[4846]: E1005 08:07:29.579976 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8a2ed397be99ca33dbab31eee6c0d309eb6aa160a01f425fb7b56c7b0fafea38\": container with ID starting with 8a2ed397be99ca33dbab31eee6c0d309eb6aa160a01f425fb7b56c7b0fafea38 not found: ID does not exist" containerID="8a2ed397be99ca33dbab31eee6c0d309eb6aa160a01f425fb7b56c7b0fafea38" Oct 05 08:07:29 crc kubenswrapper[4846]: I1005 08:07:29.580016 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8a2ed397be99ca33dbab31eee6c0d309eb6aa160a01f425fb7b56c7b0fafea38"} err="failed to get container status \"8a2ed397be99ca33dbab31eee6c0d309eb6aa160a01f425fb7b56c7b0fafea38\": rpc error: code = NotFound desc = could not find container \"8a2ed397be99ca33dbab31eee6c0d309eb6aa160a01f425fb7b56c7b0fafea38\": container with ID starting with 8a2ed397be99ca33dbab31eee6c0d309eb6aa160a01f425fb7b56c7b0fafea38 not found: ID does not exist" Oct 05 08:07:29 crc kubenswrapper[4846]: I1005 08:07:29.580046 4846 scope.go:117] "RemoveContainer" containerID="2db7c55ce2a5c2dd4b47e3a9f2c447ed7148777d04a0deeecedd8f7015c83e5e" Oct 05 08:07:29 crc kubenswrapper[4846]: E1005 08:07:29.580701 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2db7c55ce2a5c2dd4b47e3a9f2c447ed7148777d04a0deeecedd8f7015c83e5e\": container with ID starting with 2db7c55ce2a5c2dd4b47e3a9f2c447ed7148777d04a0deeecedd8f7015c83e5e not found: ID does not exist" containerID="2db7c55ce2a5c2dd4b47e3a9f2c447ed7148777d04a0deeecedd8f7015c83e5e" Oct 05 08:07:29 crc kubenswrapper[4846]: I1005 08:07:29.580724 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2db7c55ce2a5c2dd4b47e3a9f2c447ed7148777d04a0deeecedd8f7015c83e5e"} err="failed to get container status \"2db7c55ce2a5c2dd4b47e3a9f2c447ed7148777d04a0deeecedd8f7015c83e5e\": rpc error: code = NotFound desc = could not find container \"2db7c55ce2a5c2dd4b47e3a9f2c447ed7148777d04a0deeecedd8f7015c83e5e\": container with ID starting with 2db7c55ce2a5c2dd4b47e3a9f2c447ed7148777d04a0deeecedd8f7015c83e5e not found: ID does not exist" Oct 05 08:07:29 crc kubenswrapper[4846]: I1005 08:07:29.580739 4846 scope.go:117] "RemoveContainer" containerID="108ce2b65180236e38b6ea4b73d17aee3f7109fbcf46a4478eb708a7280488d3" Oct 05 08:07:29 crc kubenswrapper[4846]: I1005 08:07:29.582424 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a288716c-b683-47c0-bf44-c0bbc997da97-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "a288716c-b683-47c0-bf44-c0bbc997da97" (UID: "a288716c-b683-47c0-bf44-c0bbc997da97"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:07:29 crc kubenswrapper[4846]: I1005 08:07:29.588818 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0f8171a9-82f5-443a-a8fc-d42b575d498a-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "0f8171a9-82f5-443a-a8fc-d42b575d498a" (UID: "0f8171a9-82f5-443a-a8fc-d42b575d498a"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:07:29 crc kubenswrapper[4846]: I1005 08:07:29.594487 4846 scope.go:117] "RemoveContainer" containerID="acf58c0d7da784189c7fe44a098cc3fcf43783cf328b1e35c6673ffa86b97407" Oct 05 08:07:29 crc kubenswrapper[4846]: I1005 08:07:29.620628 4846 reconciler_common.go:293] "Volume detached for volume \"pvc-0b1f215a-31af-4eb1-a3f9-ccbfe14235d1\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-0b1f215a-31af-4eb1-a3f9-ccbfe14235d1\") on node \"crc\" DevicePath \"\"" Oct 05 08:07:29 crc kubenswrapper[4846]: I1005 08:07:29.620679 4846 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/a288716c-b683-47c0-bf44-c0bbc997da97-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Oct 05 08:07:29 crc kubenswrapper[4846]: I1005 08:07:29.620694 4846 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/0f8171a9-82f5-443a-a8fc-d42b575d498a-server-conf\") on node \"crc\" DevicePath \"\"" Oct 05 08:07:29 crc kubenswrapper[4846]: I1005 08:07:29.620706 4846 reconciler_common.go:293] "Volume detached for volume \"pvc-c0fb7d88-e3e3-43b2-a6ce-f0d1fa614bdb\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-c0fb7d88-e3e3-43b2-a6ce-f0d1fa614bdb\") on node \"crc\" DevicePath \"\"" Oct 05 08:07:29 crc kubenswrapper[4846]: I1005 08:07:29.620719 4846 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/0f8171a9-82f5-443a-a8fc-d42b575d498a-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Oct 05 08:07:29 crc kubenswrapper[4846]: I1005 08:07:29.627405 4846 scope.go:117] "RemoveContainer" containerID="108ce2b65180236e38b6ea4b73d17aee3f7109fbcf46a4478eb708a7280488d3" Oct 05 08:07:29 crc kubenswrapper[4846]: E1005 08:07:29.628407 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"108ce2b65180236e38b6ea4b73d17aee3f7109fbcf46a4478eb708a7280488d3\": container with ID starting with 108ce2b65180236e38b6ea4b73d17aee3f7109fbcf46a4478eb708a7280488d3 not found: ID does not exist" containerID="108ce2b65180236e38b6ea4b73d17aee3f7109fbcf46a4478eb708a7280488d3" Oct 05 08:07:29 crc kubenswrapper[4846]: I1005 08:07:29.628448 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"108ce2b65180236e38b6ea4b73d17aee3f7109fbcf46a4478eb708a7280488d3"} err="failed to get container status \"108ce2b65180236e38b6ea4b73d17aee3f7109fbcf46a4478eb708a7280488d3\": rpc error: code = NotFound desc = could not find container \"108ce2b65180236e38b6ea4b73d17aee3f7109fbcf46a4478eb708a7280488d3\": container with ID starting with 108ce2b65180236e38b6ea4b73d17aee3f7109fbcf46a4478eb708a7280488d3 not found: ID does not exist" Oct 05 08:07:29 crc kubenswrapper[4846]: I1005 08:07:29.628478 4846 scope.go:117] "RemoveContainer" containerID="acf58c0d7da784189c7fe44a098cc3fcf43783cf328b1e35c6673ffa86b97407" Oct 05 08:07:29 crc kubenswrapper[4846]: E1005 08:07:29.628798 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"acf58c0d7da784189c7fe44a098cc3fcf43783cf328b1e35c6673ffa86b97407\": container with ID starting with acf58c0d7da784189c7fe44a098cc3fcf43783cf328b1e35c6673ffa86b97407 not found: ID does not exist" containerID="acf58c0d7da784189c7fe44a098cc3fcf43783cf328b1e35c6673ffa86b97407" Oct 05 08:07:29 crc kubenswrapper[4846]: I1005 08:07:29.628827 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"acf58c0d7da784189c7fe44a098cc3fcf43783cf328b1e35c6673ffa86b97407"} err="failed to get container status \"acf58c0d7da784189c7fe44a098cc3fcf43783cf328b1e35c6673ffa86b97407\": rpc error: code = NotFound desc = could not find container \"acf58c0d7da784189c7fe44a098cc3fcf43783cf328b1e35c6673ffa86b97407\": container with ID starting with acf58c0d7da784189c7fe44a098cc3fcf43783cf328b1e35c6673ffa86b97407 not found: ID does not exist" Oct 05 08:07:29 crc kubenswrapper[4846]: I1005 08:07:29.826909 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 05 08:07:29 crc kubenswrapper[4846]: I1005 08:07:29.832334 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 05 08:07:29 crc kubenswrapper[4846]: I1005 08:07:29.850338 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Oct 05 08:07:29 crc kubenswrapper[4846]: E1005 08:07:29.850622 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0f8171a9-82f5-443a-a8fc-d42b575d498a" containerName="setup-container" Oct 05 08:07:29 crc kubenswrapper[4846]: I1005 08:07:29.850639 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="0f8171a9-82f5-443a-a8fc-d42b575d498a" containerName="setup-container" Oct 05 08:07:29 crc kubenswrapper[4846]: E1005 08:07:29.850648 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0f8171a9-82f5-443a-a8fc-d42b575d498a" containerName="rabbitmq" Oct 05 08:07:29 crc kubenswrapper[4846]: I1005 08:07:29.850656 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="0f8171a9-82f5-443a-a8fc-d42b575d498a" containerName="rabbitmq" Oct 05 08:07:29 crc kubenswrapper[4846]: E1005 08:07:29.850679 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e4bbd7e3-1e67-4145-943d-fe911ac8f3ba" containerName="dnsmasq-dns" Oct 05 08:07:29 crc kubenswrapper[4846]: I1005 08:07:29.850685 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="e4bbd7e3-1e67-4145-943d-fe911ac8f3ba" containerName="dnsmasq-dns" Oct 05 08:07:29 crc kubenswrapper[4846]: E1005 08:07:29.850699 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e4bbd7e3-1e67-4145-943d-fe911ac8f3ba" containerName="init" Oct 05 08:07:29 crc kubenswrapper[4846]: I1005 08:07:29.850705 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="e4bbd7e3-1e67-4145-943d-fe911ac8f3ba" containerName="init" Oct 05 08:07:29 crc kubenswrapper[4846]: E1005 08:07:29.850721 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a288716c-b683-47c0-bf44-c0bbc997da97" containerName="rabbitmq" Oct 05 08:07:29 crc kubenswrapper[4846]: I1005 08:07:29.850726 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="a288716c-b683-47c0-bf44-c0bbc997da97" containerName="rabbitmq" Oct 05 08:07:29 crc kubenswrapper[4846]: E1005 08:07:29.850737 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a288716c-b683-47c0-bf44-c0bbc997da97" containerName="setup-container" Oct 05 08:07:29 crc kubenswrapper[4846]: I1005 08:07:29.850743 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="a288716c-b683-47c0-bf44-c0bbc997da97" containerName="setup-container" Oct 05 08:07:29 crc kubenswrapper[4846]: I1005 08:07:29.850871 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="e4bbd7e3-1e67-4145-943d-fe911ac8f3ba" containerName="dnsmasq-dns" Oct 05 08:07:29 crc kubenswrapper[4846]: I1005 08:07:29.850886 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="a288716c-b683-47c0-bf44-c0bbc997da97" containerName="rabbitmq" Oct 05 08:07:29 crc kubenswrapper[4846]: I1005 08:07:29.850900 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="0f8171a9-82f5-443a-a8fc-d42b575d498a" containerName="rabbitmq" Oct 05 08:07:29 crc kubenswrapper[4846]: I1005 08:07:29.851634 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 05 08:07:29 crc kubenswrapper[4846]: W1005 08:07:29.855529 4846 reflector.go:561] object-"openstack"/"rabbitmq-default-user": failed to list *v1.Secret: secrets "rabbitmq-default-user" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openstack": no relationship found between node 'crc' and this object Oct 05 08:07:29 crc kubenswrapper[4846]: W1005 08:07:29.855568 4846 reflector.go:561] object-"openstack"/"rabbitmq-server-dockercfg-2k7rg": failed to list *v1.Secret: secrets "rabbitmq-server-dockercfg-2k7rg" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openstack": no relationship found between node 'crc' and this object Oct 05 08:07:29 crc kubenswrapper[4846]: W1005 08:07:29.855568 4846 reflector.go:561] object-"openstack"/"rabbitmq-erlang-cookie": failed to list *v1.Secret: secrets "rabbitmq-erlang-cookie" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openstack": no relationship found between node 'crc' and this object Oct 05 08:07:29 crc kubenswrapper[4846]: W1005 08:07:29.855598 4846 reflector.go:561] object-"openstack"/"cert-rabbitmq-svc": failed to list *v1.Secret: secrets "cert-rabbitmq-svc" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openstack": no relationship found between node 'crc' and this object Oct 05 08:07:29 crc kubenswrapper[4846]: E1005 08:07:29.855617 4846 reflector.go:158] "Unhandled Error" err="object-\"openstack\"/\"rabbitmq-server-dockercfg-2k7rg\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"rabbitmq-server-dockercfg-2k7rg\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openstack\": no relationship found between node 'crc' and this object" logger="UnhandledError" Oct 05 08:07:29 crc kubenswrapper[4846]: E1005 08:07:29.855619 4846 reflector.go:158] "Unhandled Error" err="object-\"openstack\"/\"rabbitmq-erlang-cookie\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"rabbitmq-erlang-cookie\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openstack\": no relationship found between node 'crc' and this object" logger="UnhandledError" Oct 05 08:07:29 crc kubenswrapper[4846]: E1005 08:07:29.855639 4846 reflector.go:158] "Unhandled Error" err="object-\"openstack\"/\"cert-rabbitmq-svc\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"cert-rabbitmq-svc\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openstack\": no relationship found between node 'crc' and this object" logger="UnhandledError" Oct 05 08:07:29 crc kubenswrapper[4846]: E1005 08:07:29.855598 4846 reflector.go:158] "Unhandled Error" err="object-\"openstack\"/\"rabbitmq-default-user\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"rabbitmq-default-user\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openstack\": no relationship found between node 'crc' and this object" logger="UnhandledError" Oct 05 08:07:29 crc kubenswrapper[4846]: I1005 08:07:29.857765 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Oct 05 08:07:29 crc kubenswrapper[4846]: I1005 08:07:29.858080 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Oct 05 08:07:29 crc kubenswrapper[4846]: I1005 08:07:29.858953 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Oct 05 08:07:29 crc kubenswrapper[4846]: I1005 08:07:29.861417 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 05 08:07:29 crc kubenswrapper[4846]: I1005 08:07:29.867591 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 05 08:07:29 crc kubenswrapper[4846]: I1005 08:07:29.878281 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 05 08:07:29 crc kubenswrapper[4846]: I1005 08:07:29.883826 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 05 08:07:29 crc kubenswrapper[4846]: I1005 08:07:29.885068 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 05 08:07:29 crc kubenswrapper[4846]: I1005 08:07:29.888018 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Oct 05 08:07:29 crc kubenswrapper[4846]: I1005 08:07:29.888274 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Oct 05 08:07:29 crc kubenswrapper[4846]: I1005 08:07:29.888370 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Oct 05 08:07:29 crc kubenswrapper[4846]: I1005 08:07:29.888501 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-r9ldz" Oct 05 08:07:29 crc kubenswrapper[4846]: I1005 08:07:29.888712 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Oct 05 08:07:29 crc kubenswrapper[4846]: I1005 08:07:29.888875 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Oct 05 08:07:29 crc kubenswrapper[4846]: I1005 08:07:29.891155 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Oct 05 08:07:29 crc kubenswrapper[4846]: I1005 08:07:29.900213 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 05 08:07:29 crc kubenswrapper[4846]: I1005 08:07:29.925719 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/a8b2a717-7ead-4166-b460-3d9de6f3acec-pod-info\") pod \"rabbitmq-server-0\" (UID: \"a8b2a717-7ead-4166-b460-3d9de6f3acec\") " pod="openstack/rabbitmq-server-0" Oct 05 08:07:29 crc kubenswrapper[4846]: I1005 08:07:29.925828 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a8b2a717-7ead-4166-b460-3d9de6f3acec-config-data\") pod \"rabbitmq-server-0\" (UID: \"a8b2a717-7ead-4166-b460-3d9de6f3acec\") " pod="openstack/rabbitmq-server-0" Oct 05 08:07:29 crc kubenswrapper[4846]: I1005 08:07:29.925877 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/a8b2a717-7ead-4166-b460-3d9de6f3acec-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"a8b2a717-7ead-4166-b460-3d9de6f3acec\") " pod="openstack/rabbitmq-server-0" Oct 05 08:07:29 crc kubenswrapper[4846]: I1005 08:07:29.925946 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/a8b2a717-7ead-4166-b460-3d9de6f3acec-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"a8b2a717-7ead-4166-b460-3d9de6f3acec\") " pod="openstack/rabbitmq-server-0" Oct 05 08:07:29 crc kubenswrapper[4846]: I1005 08:07:29.926167 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/a8b2a717-7ead-4166-b460-3d9de6f3acec-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"a8b2a717-7ead-4166-b460-3d9de6f3acec\") " pod="openstack/rabbitmq-server-0" Oct 05 08:07:29 crc kubenswrapper[4846]: I1005 08:07:29.926255 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/a8b2a717-7ead-4166-b460-3d9de6f3acec-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"a8b2a717-7ead-4166-b460-3d9de6f3acec\") " pod="openstack/rabbitmq-server-0" Oct 05 08:07:29 crc kubenswrapper[4846]: I1005 08:07:29.926305 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/a8b2a717-7ead-4166-b460-3d9de6f3acec-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"a8b2a717-7ead-4166-b460-3d9de6f3acec\") " pod="openstack/rabbitmq-server-0" Oct 05 08:07:29 crc kubenswrapper[4846]: I1005 08:07:29.926341 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-c0fb7d88-e3e3-43b2-a6ce-f0d1fa614bdb\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-c0fb7d88-e3e3-43b2-a6ce-f0d1fa614bdb\") pod \"rabbitmq-server-0\" (UID: \"a8b2a717-7ead-4166-b460-3d9de6f3acec\") " pod="openstack/rabbitmq-server-0" Oct 05 08:07:29 crc kubenswrapper[4846]: I1005 08:07:29.926373 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tc258\" (UniqueName: \"kubernetes.io/projected/a8b2a717-7ead-4166-b460-3d9de6f3acec-kube-api-access-tc258\") pod \"rabbitmq-server-0\" (UID: \"a8b2a717-7ead-4166-b460-3d9de6f3acec\") " pod="openstack/rabbitmq-server-0" Oct 05 08:07:29 crc kubenswrapper[4846]: I1005 08:07:29.926452 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/a8b2a717-7ead-4166-b460-3d9de6f3acec-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"a8b2a717-7ead-4166-b460-3d9de6f3acec\") " pod="openstack/rabbitmq-server-0" Oct 05 08:07:29 crc kubenswrapper[4846]: I1005 08:07:29.926479 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/a8b2a717-7ead-4166-b460-3d9de6f3acec-server-conf\") pod \"rabbitmq-server-0\" (UID: \"a8b2a717-7ead-4166-b460-3d9de6f3acec\") " pod="openstack/rabbitmq-server-0" Oct 05 08:07:30 crc kubenswrapper[4846]: I1005 08:07:30.028131 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/df56f237-d65a-4dbf-a3cf-ee1c496d7ce4-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"df56f237-d65a-4dbf-a3cf-ee1c496d7ce4\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 08:07:30 crc kubenswrapper[4846]: I1005 08:07:30.028226 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/df56f237-d65a-4dbf-a3cf-ee1c496d7ce4-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"df56f237-d65a-4dbf-a3cf-ee1c496d7ce4\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 08:07:30 crc kubenswrapper[4846]: I1005 08:07:30.028359 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/a8b2a717-7ead-4166-b460-3d9de6f3acec-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"a8b2a717-7ead-4166-b460-3d9de6f3acec\") " pod="openstack/rabbitmq-server-0" Oct 05 08:07:30 crc kubenswrapper[4846]: I1005 08:07:30.028455 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/df56f237-d65a-4dbf-a3cf-ee1c496d7ce4-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"df56f237-d65a-4dbf-a3cf-ee1c496d7ce4\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 08:07:30 crc kubenswrapper[4846]: I1005 08:07:30.028503 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/a8b2a717-7ead-4166-b460-3d9de6f3acec-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"a8b2a717-7ead-4166-b460-3d9de6f3acec\") " pod="openstack/rabbitmq-server-0" Oct 05 08:07:30 crc kubenswrapper[4846]: I1005 08:07:30.028544 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/a8b2a717-7ead-4166-b460-3d9de6f3acec-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"a8b2a717-7ead-4166-b460-3d9de6f3acec\") " pod="openstack/rabbitmq-server-0" Oct 05 08:07:30 crc kubenswrapper[4846]: I1005 08:07:30.028573 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/df56f237-d65a-4dbf-a3cf-ee1c496d7ce4-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"df56f237-d65a-4dbf-a3cf-ee1c496d7ce4\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 08:07:30 crc kubenswrapper[4846]: I1005 08:07:30.028627 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-c0fb7d88-e3e3-43b2-a6ce-f0d1fa614bdb\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-c0fb7d88-e3e3-43b2-a6ce-f0d1fa614bdb\") pod \"rabbitmq-server-0\" (UID: \"a8b2a717-7ead-4166-b460-3d9de6f3acec\") " pod="openstack/rabbitmq-server-0" Oct 05 08:07:30 crc kubenswrapper[4846]: I1005 08:07:30.028681 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tc258\" (UniqueName: \"kubernetes.io/projected/a8b2a717-7ead-4166-b460-3d9de6f3acec-kube-api-access-tc258\") pod \"rabbitmq-server-0\" (UID: \"a8b2a717-7ead-4166-b460-3d9de6f3acec\") " pod="openstack/rabbitmq-server-0" Oct 05 08:07:30 crc kubenswrapper[4846]: I1005 08:07:30.028723 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/df56f237-d65a-4dbf-a3cf-ee1c496d7ce4-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"df56f237-d65a-4dbf-a3cf-ee1c496d7ce4\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 08:07:30 crc kubenswrapper[4846]: I1005 08:07:30.028777 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-0b1f215a-31af-4eb1-a3f9-ccbfe14235d1\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-0b1f215a-31af-4eb1-a3f9-ccbfe14235d1\") pod \"rabbitmq-cell1-server-0\" (UID: \"df56f237-d65a-4dbf-a3cf-ee1c496d7ce4\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 08:07:30 crc kubenswrapper[4846]: I1005 08:07:30.028835 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/df56f237-d65a-4dbf-a3cf-ee1c496d7ce4-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"df56f237-d65a-4dbf-a3cf-ee1c496d7ce4\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 08:07:30 crc kubenswrapper[4846]: I1005 08:07:30.028872 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/a8b2a717-7ead-4166-b460-3d9de6f3acec-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"a8b2a717-7ead-4166-b460-3d9de6f3acec\") " pod="openstack/rabbitmq-server-0" Oct 05 08:07:30 crc kubenswrapper[4846]: I1005 08:07:30.028960 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/a8b2a717-7ead-4166-b460-3d9de6f3acec-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"a8b2a717-7ead-4166-b460-3d9de6f3acec\") " pod="openstack/rabbitmq-server-0" Oct 05 08:07:30 crc kubenswrapper[4846]: I1005 08:07:30.028963 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/a8b2a717-7ead-4166-b460-3d9de6f3acec-server-conf\") pod \"rabbitmq-server-0\" (UID: \"a8b2a717-7ead-4166-b460-3d9de6f3acec\") " pod="openstack/rabbitmq-server-0" Oct 05 08:07:30 crc kubenswrapper[4846]: I1005 08:07:30.029019 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/a8b2a717-7ead-4166-b460-3d9de6f3acec-pod-info\") pod \"rabbitmq-server-0\" (UID: \"a8b2a717-7ead-4166-b460-3d9de6f3acec\") " pod="openstack/rabbitmq-server-0" Oct 05 08:07:30 crc kubenswrapper[4846]: I1005 08:07:30.029070 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/df56f237-d65a-4dbf-a3cf-ee1c496d7ce4-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"df56f237-d65a-4dbf-a3cf-ee1c496d7ce4\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 08:07:30 crc kubenswrapper[4846]: I1005 08:07:30.029097 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h6k5s\" (UniqueName: \"kubernetes.io/projected/df56f237-d65a-4dbf-a3cf-ee1c496d7ce4-kube-api-access-h6k5s\") pod \"rabbitmq-cell1-server-0\" (UID: \"df56f237-d65a-4dbf-a3cf-ee1c496d7ce4\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 08:07:30 crc kubenswrapper[4846]: I1005 08:07:30.029205 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a8b2a717-7ead-4166-b460-3d9de6f3acec-config-data\") pod \"rabbitmq-server-0\" (UID: \"a8b2a717-7ead-4166-b460-3d9de6f3acec\") " pod="openstack/rabbitmq-server-0" Oct 05 08:07:30 crc kubenswrapper[4846]: I1005 08:07:30.029244 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/a8b2a717-7ead-4166-b460-3d9de6f3acec-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"a8b2a717-7ead-4166-b460-3d9de6f3acec\") " pod="openstack/rabbitmq-server-0" Oct 05 08:07:30 crc kubenswrapper[4846]: I1005 08:07:30.029295 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/a8b2a717-7ead-4166-b460-3d9de6f3acec-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"a8b2a717-7ead-4166-b460-3d9de6f3acec\") " pod="openstack/rabbitmq-server-0" Oct 05 08:07:30 crc kubenswrapper[4846]: I1005 08:07:30.029418 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/a8b2a717-7ead-4166-b460-3d9de6f3acec-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"a8b2a717-7ead-4166-b460-3d9de6f3acec\") " pod="openstack/rabbitmq-server-0" Oct 05 08:07:30 crc kubenswrapper[4846]: I1005 08:07:30.029564 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/df56f237-d65a-4dbf-a3cf-ee1c496d7ce4-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"df56f237-d65a-4dbf-a3cf-ee1c496d7ce4\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 08:07:30 crc kubenswrapper[4846]: I1005 08:07:30.029622 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/df56f237-d65a-4dbf-a3cf-ee1c496d7ce4-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"df56f237-d65a-4dbf-a3cf-ee1c496d7ce4\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 08:07:30 crc kubenswrapper[4846]: I1005 08:07:30.029820 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a8b2a717-7ead-4166-b460-3d9de6f3acec-config-data\") pod \"rabbitmq-server-0\" (UID: \"a8b2a717-7ead-4166-b460-3d9de6f3acec\") " pod="openstack/rabbitmq-server-0" Oct 05 08:07:30 crc kubenswrapper[4846]: I1005 08:07:30.030164 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/a8b2a717-7ead-4166-b460-3d9de6f3acec-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"a8b2a717-7ead-4166-b460-3d9de6f3acec\") " pod="openstack/rabbitmq-server-0" Oct 05 08:07:30 crc kubenswrapper[4846]: I1005 08:07:30.030353 4846 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 05 08:07:30 crc kubenswrapper[4846]: I1005 08:07:30.030378 4846 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-c0fb7d88-e3e3-43b2-a6ce-f0d1fa614bdb\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-c0fb7d88-e3e3-43b2-a6ce-f0d1fa614bdb\") pod \"rabbitmq-server-0\" (UID: \"a8b2a717-7ead-4166-b460-3d9de6f3acec\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/d6d150f95ef01577cee9c24c1b65571c71a6ebd80cf8fb72945050d88d52ce8c/globalmount\"" pod="openstack/rabbitmq-server-0" Oct 05 08:07:30 crc kubenswrapper[4846]: I1005 08:07:30.030389 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/a8b2a717-7ead-4166-b460-3d9de6f3acec-server-conf\") pod \"rabbitmq-server-0\" (UID: \"a8b2a717-7ead-4166-b460-3d9de6f3acec\") " pod="openstack/rabbitmq-server-0" Oct 05 08:07:30 crc kubenswrapper[4846]: I1005 08:07:30.033512 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/a8b2a717-7ead-4166-b460-3d9de6f3acec-pod-info\") pod \"rabbitmq-server-0\" (UID: \"a8b2a717-7ead-4166-b460-3d9de6f3acec\") " pod="openstack/rabbitmq-server-0" Oct 05 08:07:30 crc kubenswrapper[4846]: I1005 08:07:30.044619 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tc258\" (UniqueName: \"kubernetes.io/projected/a8b2a717-7ead-4166-b460-3d9de6f3acec-kube-api-access-tc258\") pod \"rabbitmq-server-0\" (UID: \"a8b2a717-7ead-4166-b460-3d9de6f3acec\") " pod="openstack/rabbitmq-server-0" Oct 05 08:07:30 crc kubenswrapper[4846]: I1005 08:07:30.056037 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-c0fb7d88-e3e3-43b2-a6ce-f0d1fa614bdb\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-c0fb7d88-e3e3-43b2-a6ce-f0d1fa614bdb\") pod \"rabbitmq-server-0\" (UID: \"a8b2a717-7ead-4166-b460-3d9de6f3acec\") " pod="openstack/rabbitmq-server-0" Oct 05 08:07:30 crc kubenswrapper[4846]: I1005 08:07:30.131588 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/df56f237-d65a-4dbf-a3cf-ee1c496d7ce4-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"df56f237-d65a-4dbf-a3cf-ee1c496d7ce4\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 08:07:30 crc kubenswrapper[4846]: I1005 08:07:30.131663 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-0b1f215a-31af-4eb1-a3f9-ccbfe14235d1\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-0b1f215a-31af-4eb1-a3f9-ccbfe14235d1\") pod \"rabbitmq-cell1-server-0\" (UID: \"df56f237-d65a-4dbf-a3cf-ee1c496d7ce4\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 08:07:30 crc kubenswrapper[4846]: I1005 08:07:30.131708 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/df56f237-d65a-4dbf-a3cf-ee1c496d7ce4-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"df56f237-d65a-4dbf-a3cf-ee1c496d7ce4\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 08:07:30 crc kubenswrapper[4846]: I1005 08:07:30.131765 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/df56f237-d65a-4dbf-a3cf-ee1c496d7ce4-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"df56f237-d65a-4dbf-a3cf-ee1c496d7ce4\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 08:07:30 crc kubenswrapper[4846]: I1005 08:07:30.131801 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h6k5s\" (UniqueName: \"kubernetes.io/projected/df56f237-d65a-4dbf-a3cf-ee1c496d7ce4-kube-api-access-h6k5s\") pod \"rabbitmq-cell1-server-0\" (UID: \"df56f237-d65a-4dbf-a3cf-ee1c496d7ce4\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 08:07:30 crc kubenswrapper[4846]: I1005 08:07:30.131893 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/df56f237-d65a-4dbf-a3cf-ee1c496d7ce4-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"df56f237-d65a-4dbf-a3cf-ee1c496d7ce4\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 08:07:30 crc kubenswrapper[4846]: I1005 08:07:30.131927 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/df56f237-d65a-4dbf-a3cf-ee1c496d7ce4-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"df56f237-d65a-4dbf-a3cf-ee1c496d7ce4\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 08:07:30 crc kubenswrapper[4846]: I1005 08:07:30.131966 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/df56f237-d65a-4dbf-a3cf-ee1c496d7ce4-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"df56f237-d65a-4dbf-a3cf-ee1c496d7ce4\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 08:07:30 crc kubenswrapper[4846]: I1005 08:07:30.132018 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/df56f237-d65a-4dbf-a3cf-ee1c496d7ce4-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"df56f237-d65a-4dbf-a3cf-ee1c496d7ce4\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 08:07:30 crc kubenswrapper[4846]: I1005 08:07:30.132091 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/df56f237-d65a-4dbf-a3cf-ee1c496d7ce4-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"df56f237-d65a-4dbf-a3cf-ee1c496d7ce4\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 08:07:30 crc kubenswrapper[4846]: I1005 08:07:30.132140 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/df56f237-d65a-4dbf-a3cf-ee1c496d7ce4-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"df56f237-d65a-4dbf-a3cf-ee1c496d7ce4\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 08:07:30 crc kubenswrapper[4846]: I1005 08:07:30.132961 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/df56f237-d65a-4dbf-a3cf-ee1c496d7ce4-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"df56f237-d65a-4dbf-a3cf-ee1c496d7ce4\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 08:07:30 crc kubenswrapper[4846]: I1005 08:07:30.132979 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/df56f237-d65a-4dbf-a3cf-ee1c496d7ce4-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"df56f237-d65a-4dbf-a3cf-ee1c496d7ce4\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 08:07:30 crc kubenswrapper[4846]: I1005 08:07:30.133477 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/df56f237-d65a-4dbf-a3cf-ee1c496d7ce4-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"df56f237-d65a-4dbf-a3cf-ee1c496d7ce4\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 08:07:30 crc kubenswrapper[4846]: I1005 08:07:30.134199 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/df56f237-d65a-4dbf-a3cf-ee1c496d7ce4-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"df56f237-d65a-4dbf-a3cf-ee1c496d7ce4\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 08:07:30 crc kubenswrapper[4846]: I1005 08:07:30.134930 4846 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 05 08:07:30 crc kubenswrapper[4846]: I1005 08:07:30.135010 4846 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-0b1f215a-31af-4eb1-a3f9-ccbfe14235d1\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-0b1f215a-31af-4eb1-a3f9-ccbfe14235d1\") pod \"rabbitmq-cell1-server-0\" (UID: \"df56f237-d65a-4dbf-a3cf-ee1c496d7ce4\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/46422321c9fc4982dc959ec2b6aca5699513296fc205359a8d3e45d1d4b35fe5/globalmount\"" pod="openstack/rabbitmq-cell1-server-0" Oct 05 08:07:30 crc kubenswrapper[4846]: I1005 08:07:30.134933 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/df56f237-d65a-4dbf-a3cf-ee1c496d7ce4-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"df56f237-d65a-4dbf-a3cf-ee1c496d7ce4\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 08:07:30 crc kubenswrapper[4846]: I1005 08:07:30.135994 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/df56f237-d65a-4dbf-a3cf-ee1c496d7ce4-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"df56f237-d65a-4dbf-a3cf-ee1c496d7ce4\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 08:07:30 crc kubenswrapper[4846]: I1005 08:07:30.137350 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/df56f237-d65a-4dbf-a3cf-ee1c496d7ce4-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"df56f237-d65a-4dbf-a3cf-ee1c496d7ce4\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 08:07:30 crc kubenswrapper[4846]: I1005 08:07:30.138059 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/df56f237-d65a-4dbf-a3cf-ee1c496d7ce4-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"df56f237-d65a-4dbf-a3cf-ee1c496d7ce4\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 08:07:30 crc kubenswrapper[4846]: I1005 08:07:30.141053 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/df56f237-d65a-4dbf-a3cf-ee1c496d7ce4-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"df56f237-d65a-4dbf-a3cf-ee1c496d7ce4\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 08:07:30 crc kubenswrapper[4846]: I1005 08:07:30.150309 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h6k5s\" (UniqueName: \"kubernetes.io/projected/df56f237-d65a-4dbf-a3cf-ee1c496d7ce4-kube-api-access-h6k5s\") pod \"rabbitmq-cell1-server-0\" (UID: \"df56f237-d65a-4dbf-a3cf-ee1c496d7ce4\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 08:07:30 crc kubenswrapper[4846]: I1005 08:07:30.181665 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-0b1f215a-31af-4eb1-a3f9-ccbfe14235d1\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-0b1f215a-31af-4eb1-a3f9-ccbfe14235d1\") pod \"rabbitmq-cell1-server-0\" (UID: \"df56f237-d65a-4dbf-a3cf-ee1c496d7ce4\") " pod="openstack/rabbitmq-cell1-server-0" Oct 05 08:07:30 crc kubenswrapper[4846]: I1005 08:07:30.203508 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 05 08:07:30 crc kubenswrapper[4846]: I1005 08:07:30.496402 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 05 08:07:30 crc kubenswrapper[4846]: I1005 08:07:30.507756 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0f8171a9-82f5-443a-a8fc-d42b575d498a" path="/var/lib/kubelet/pods/0f8171a9-82f5-443a-a8fc-d42b575d498a/volumes" Oct 05 08:07:30 crc kubenswrapper[4846]: I1005 08:07:30.508499 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a288716c-b683-47c0-bf44-c0bbc997da97" path="/var/lib/kubelet/pods/a288716c-b683-47c0-bf44-c0bbc997da97/volumes" Oct 05 08:07:30 crc kubenswrapper[4846]: I1005 08:07:30.509513 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e4bbd7e3-1e67-4145-943d-fe911ac8f3ba" path="/var/lib/kubelet/pods/e4bbd7e3-1e67-4145-943d-fe911ac8f3ba/volumes" Oct 05 08:07:30 crc kubenswrapper[4846]: I1005 08:07:30.727302 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Oct 05 08:07:30 crc kubenswrapper[4846]: I1005 08:07:30.796223 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/a8b2a717-7ead-4166-b460-3d9de6f3acec-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"a8b2a717-7ead-4166-b460-3d9de6f3acec\") " pod="openstack/rabbitmq-server-0" Oct 05 08:07:30 crc kubenswrapper[4846]: W1005 08:07:30.801564 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddf56f237_d65a_4dbf_a3cf_ee1c496d7ce4.slice/crio-b1308fc8e3007d912ae6d1ca0d9aa2699e9e59fb0c0a93e13b5cea59ff2b92bd WatchSource:0}: Error finding container b1308fc8e3007d912ae6d1ca0d9aa2699e9e59fb0c0a93e13b5cea59ff2b92bd: Status 404 returned error can't find the container with id b1308fc8e3007d912ae6d1ca0d9aa2699e9e59fb0c0a93e13b5cea59ff2b92bd Oct 05 08:07:30 crc kubenswrapper[4846]: I1005 08:07:30.835247 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Oct 05 08:07:30 crc kubenswrapper[4846]: I1005 08:07:30.843706 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/a8b2a717-7ead-4166-b460-3d9de6f3acec-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"a8b2a717-7ead-4166-b460-3d9de6f3acec\") " pod="openstack/rabbitmq-server-0" Oct 05 08:07:30 crc kubenswrapper[4846]: I1005 08:07:30.994822 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-2k7rg" Oct 05 08:07:31 crc kubenswrapper[4846]: E1005 08:07:31.029535 4846 projected.go:263] Couldn't get secret openstack/cert-rabbitmq-svc: failed to sync secret cache: timed out waiting for the condition Oct 05 08:07:31 crc kubenswrapper[4846]: I1005 08:07:31.382367 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Oct 05 08:07:31 crc kubenswrapper[4846]: E1005 08:07:31.390301 4846 projected.go:194] Error preparing data for projected volume rabbitmq-tls for pod openstack/rabbitmq-server-0: failed to sync secret cache: timed out waiting for the condition Oct 05 08:07:31 crc kubenswrapper[4846]: E1005 08:07:31.390454 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/a8b2a717-7ead-4166-b460-3d9de6f3acec-rabbitmq-tls podName:a8b2a717-7ead-4166-b460-3d9de6f3acec nodeName:}" failed. No retries permitted until 2025-10-05 08:07:31.890412751 +0000 UTC m=+4774.131265566 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "rabbitmq-tls" (UniqueName: "kubernetes.io/projected/a8b2a717-7ead-4166-b460-3d9de6f3acec-rabbitmq-tls") pod "rabbitmq-server-0" (UID: "a8b2a717-7ead-4166-b460-3d9de6f3acec") : failed to sync secret cache: timed out waiting for the condition Oct 05 08:07:31 crc kubenswrapper[4846]: I1005 08:07:31.539868 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"df56f237-d65a-4dbf-a3cf-ee1c496d7ce4","Type":"ContainerStarted","Data":"b1308fc8e3007d912ae6d1ca0d9aa2699e9e59fb0c0a93e13b5cea59ff2b92bd"} Oct 05 08:07:31 crc kubenswrapper[4846]: I1005 08:07:31.960931 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/a8b2a717-7ead-4166-b460-3d9de6f3acec-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"a8b2a717-7ead-4166-b460-3d9de6f3acec\") " pod="openstack/rabbitmq-server-0" Oct 05 08:07:31 crc kubenswrapper[4846]: I1005 08:07:31.986346 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/a8b2a717-7ead-4166-b460-3d9de6f3acec-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"a8b2a717-7ead-4166-b460-3d9de6f3acec\") " pod="openstack/rabbitmq-server-0" Oct 05 08:07:32 crc kubenswrapper[4846]: I1005 08:07:32.268813 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 05 08:07:32 crc kubenswrapper[4846]: I1005 08:07:32.549742 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 05 08:07:32 crc kubenswrapper[4846]: I1005 08:07:32.554091 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"df56f237-d65a-4dbf-a3cf-ee1c496d7ce4","Type":"ContainerStarted","Data":"b5c4f60c62475cc4bd9ed8e6bed14001c6ddd38545c44fca82e0b1a3c8954dff"} Oct 05 08:07:32 crc kubenswrapper[4846]: W1005 08:07:32.558820 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda8b2a717_7ead_4166_b460_3d9de6f3acec.slice/crio-304988e2d70598c5a8d240001fa3ea9a7a8f2e404d510590d999b8505cf9a208 WatchSource:0}: Error finding container 304988e2d70598c5a8d240001fa3ea9a7a8f2e404d510590d999b8505cf9a208: Status 404 returned error can't find the container with id 304988e2d70598c5a8d240001fa3ea9a7a8f2e404d510590d999b8505cf9a208 Oct 05 08:07:33 crc kubenswrapper[4846]: I1005 08:07:33.561884 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"a8b2a717-7ead-4166-b460-3d9de6f3acec","Type":"ContainerStarted","Data":"304988e2d70598c5a8d240001fa3ea9a7a8f2e404d510590d999b8505cf9a208"} Oct 05 08:07:34 crc kubenswrapper[4846]: I1005 08:07:34.580523 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"a8b2a717-7ead-4166-b460-3d9de6f3acec","Type":"ContainerStarted","Data":"f3552013ade091bafbb82316effe9d2dee35fc3b0d705dc488ffe382e58d851f"} Oct 05 08:08:05 crc kubenswrapper[4846]: I1005 08:08:05.871502 4846 generic.go:334] "Generic (PLEG): container finished" podID="df56f237-d65a-4dbf-a3cf-ee1c496d7ce4" containerID="b5c4f60c62475cc4bd9ed8e6bed14001c6ddd38545c44fca82e0b1a3c8954dff" exitCode=0 Oct 05 08:08:05 crc kubenswrapper[4846]: I1005 08:08:05.871615 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"df56f237-d65a-4dbf-a3cf-ee1c496d7ce4","Type":"ContainerDied","Data":"b5c4f60c62475cc4bd9ed8e6bed14001c6ddd38545c44fca82e0b1a3c8954dff"} Oct 05 08:08:06 crc kubenswrapper[4846]: I1005 08:08:06.884608 4846 generic.go:334] "Generic (PLEG): container finished" podID="a8b2a717-7ead-4166-b460-3d9de6f3acec" containerID="f3552013ade091bafbb82316effe9d2dee35fc3b0d705dc488ffe382e58d851f" exitCode=0 Oct 05 08:08:06 crc kubenswrapper[4846]: I1005 08:08:06.884655 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"a8b2a717-7ead-4166-b460-3d9de6f3acec","Type":"ContainerDied","Data":"f3552013ade091bafbb82316effe9d2dee35fc3b0d705dc488ffe382e58d851f"} Oct 05 08:08:06 crc kubenswrapper[4846]: I1005 08:08:06.888530 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"df56f237-d65a-4dbf-a3cf-ee1c496d7ce4","Type":"ContainerStarted","Data":"f0b8a1f09c17377c691218afe2eb8b6360679c1085e78ebeeb347d7f8cc60ee4"} Oct 05 08:08:06 crc kubenswrapper[4846]: I1005 08:08:06.888765 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Oct 05 08:08:07 crc kubenswrapper[4846]: I1005 08:08:07.901104 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"a8b2a717-7ead-4166-b460-3d9de6f3acec","Type":"ContainerStarted","Data":"e1f62621dd742f33fc19654936c94d2f9da775441bdc905862b0a14a437c9511"} Oct 05 08:08:07 crc kubenswrapper[4846]: I1005 08:08:07.901889 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Oct 05 08:08:07 crc kubenswrapper[4846]: I1005 08:08:07.937242 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=38.937214158 podStartE2EDuration="38.937214158s" podCreationTimestamp="2025-10-05 08:07:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 08:08:07.933671473 +0000 UTC m=+4810.174524308" watchObservedRunningTime="2025-10-05 08:08:07.937214158 +0000 UTC m=+4810.178066973" Oct 05 08:08:07 crc kubenswrapper[4846]: I1005 08:08:07.945368 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=38.945343325 podStartE2EDuration="38.945343325s" podCreationTimestamp="2025-10-05 08:07:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 08:08:06.957079182 +0000 UTC m=+4809.197931987" watchObservedRunningTime="2025-10-05 08:08:07.945343325 +0000 UTC m=+4810.186196140" Oct 05 08:08:20 crc kubenswrapper[4846]: I1005 08:08:20.208540 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Oct 05 08:08:22 crc kubenswrapper[4846]: I1005 08:08:22.272376 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Oct 05 08:08:25 crc kubenswrapper[4846]: I1005 08:08:25.476406 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client-1-default"] Oct 05 08:08:25 crc kubenswrapper[4846]: I1005 08:08:25.478504 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-1-default" Oct 05 08:08:25 crc kubenswrapper[4846]: I1005 08:08:25.482823 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-4c82f" Oct 05 08:08:25 crc kubenswrapper[4846]: I1005 08:08:25.490401 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-1-default"] Oct 05 08:08:25 crc kubenswrapper[4846]: I1005 08:08:25.563049 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rs4hc\" (UniqueName: \"kubernetes.io/projected/4bd1482a-e598-4428-ab5e-7997547e4647-kube-api-access-rs4hc\") pod \"mariadb-client-1-default\" (UID: \"4bd1482a-e598-4428-ab5e-7997547e4647\") " pod="openstack/mariadb-client-1-default" Oct 05 08:08:25 crc kubenswrapper[4846]: I1005 08:08:25.665297 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rs4hc\" (UniqueName: \"kubernetes.io/projected/4bd1482a-e598-4428-ab5e-7997547e4647-kube-api-access-rs4hc\") pod \"mariadb-client-1-default\" (UID: \"4bd1482a-e598-4428-ab5e-7997547e4647\") " pod="openstack/mariadb-client-1-default" Oct 05 08:08:25 crc kubenswrapper[4846]: I1005 08:08:25.695520 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rs4hc\" (UniqueName: \"kubernetes.io/projected/4bd1482a-e598-4428-ab5e-7997547e4647-kube-api-access-rs4hc\") pod \"mariadb-client-1-default\" (UID: \"4bd1482a-e598-4428-ab5e-7997547e4647\") " pod="openstack/mariadb-client-1-default" Oct 05 08:08:25 crc kubenswrapper[4846]: I1005 08:08:25.795879 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-1-default" Oct 05 08:08:26 crc kubenswrapper[4846]: I1005 08:08:26.422388 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-1-default"] Oct 05 08:08:27 crc kubenswrapper[4846]: I1005 08:08:27.083498 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-1-default" event={"ID":"4bd1482a-e598-4428-ab5e-7997547e4647","Type":"ContainerStarted","Data":"c39d9fc5f97d3ee86e47ee9c538411f38d147f6b00bb63260eddb07b3ecd7129"} Oct 05 08:08:28 crc kubenswrapper[4846]: I1005 08:08:28.090844 4846 generic.go:334] "Generic (PLEG): container finished" podID="4bd1482a-e598-4428-ab5e-7997547e4647" containerID="4b99194058de1e2b0eeda1df7895b490c5e4cc1d359f20acfac1596bfa476a84" exitCode=0 Oct 05 08:08:28 crc kubenswrapper[4846]: I1005 08:08:28.090889 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-1-default" event={"ID":"4bd1482a-e598-4428-ab5e-7997547e4647","Type":"ContainerDied","Data":"4b99194058de1e2b0eeda1df7895b490c5e4cc1d359f20acfac1596bfa476a84"} Oct 05 08:08:29 crc kubenswrapper[4846]: I1005 08:08:29.495510 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-1-default" Oct 05 08:08:29 crc kubenswrapper[4846]: I1005 08:08:29.536013 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mariadb-client-1-default_4bd1482a-e598-4428-ab5e-7997547e4647/mariadb-client-1-default/0.log" Oct 05 08:08:29 crc kubenswrapper[4846]: I1005 08:08:29.577987 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client-1-default"] Oct 05 08:08:29 crc kubenswrapper[4846]: I1005 08:08:29.581685 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client-1-default"] Oct 05 08:08:29 crc kubenswrapper[4846]: I1005 08:08:29.633847 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rs4hc\" (UniqueName: \"kubernetes.io/projected/4bd1482a-e598-4428-ab5e-7997547e4647-kube-api-access-rs4hc\") pod \"4bd1482a-e598-4428-ab5e-7997547e4647\" (UID: \"4bd1482a-e598-4428-ab5e-7997547e4647\") " Oct 05 08:08:29 crc kubenswrapper[4846]: I1005 08:08:29.640606 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bd1482a-e598-4428-ab5e-7997547e4647-kube-api-access-rs4hc" (OuterVolumeSpecName: "kube-api-access-rs4hc") pod "4bd1482a-e598-4428-ab5e-7997547e4647" (UID: "4bd1482a-e598-4428-ab5e-7997547e4647"). InnerVolumeSpecName "kube-api-access-rs4hc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:08:29 crc kubenswrapper[4846]: I1005 08:08:29.737476 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rs4hc\" (UniqueName: \"kubernetes.io/projected/4bd1482a-e598-4428-ab5e-7997547e4647-kube-api-access-rs4hc\") on node \"crc\" DevicePath \"\"" Oct 05 08:08:30 crc kubenswrapper[4846]: I1005 08:08:30.062603 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client-2-default"] Oct 05 08:08:30 crc kubenswrapper[4846]: E1005 08:08:30.062918 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4bd1482a-e598-4428-ab5e-7997547e4647" containerName="mariadb-client-1-default" Oct 05 08:08:30 crc kubenswrapper[4846]: I1005 08:08:30.062935 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="4bd1482a-e598-4428-ab5e-7997547e4647" containerName="mariadb-client-1-default" Oct 05 08:08:30 crc kubenswrapper[4846]: I1005 08:08:30.063081 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="4bd1482a-e598-4428-ab5e-7997547e4647" containerName="mariadb-client-1-default" Oct 05 08:08:30 crc kubenswrapper[4846]: I1005 08:08:30.063584 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-2-default" Oct 05 08:08:30 crc kubenswrapper[4846]: I1005 08:08:30.078458 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-2-default"] Oct 05 08:08:30 crc kubenswrapper[4846]: I1005 08:08:30.107786 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c39d9fc5f97d3ee86e47ee9c538411f38d147f6b00bb63260eddb07b3ecd7129" Oct 05 08:08:30 crc kubenswrapper[4846]: I1005 08:08:30.107841 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-1-default" Oct 05 08:08:30 crc kubenswrapper[4846]: I1005 08:08:30.145027 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wbt78\" (UniqueName: \"kubernetes.io/projected/9e7a80d7-f91f-4d09-a3fd-fd4c1d609daf-kube-api-access-wbt78\") pod \"mariadb-client-2-default\" (UID: \"9e7a80d7-f91f-4d09-a3fd-fd4c1d609daf\") " pod="openstack/mariadb-client-2-default" Oct 05 08:08:30 crc kubenswrapper[4846]: I1005 08:08:30.246661 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wbt78\" (UniqueName: \"kubernetes.io/projected/9e7a80d7-f91f-4d09-a3fd-fd4c1d609daf-kube-api-access-wbt78\") pod \"mariadb-client-2-default\" (UID: \"9e7a80d7-f91f-4d09-a3fd-fd4c1d609daf\") " pod="openstack/mariadb-client-2-default" Oct 05 08:08:30 crc kubenswrapper[4846]: I1005 08:08:30.266574 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wbt78\" (UniqueName: \"kubernetes.io/projected/9e7a80d7-f91f-4d09-a3fd-fd4c1d609daf-kube-api-access-wbt78\") pod \"mariadb-client-2-default\" (UID: \"9e7a80d7-f91f-4d09-a3fd-fd4c1d609daf\") " pod="openstack/mariadb-client-2-default" Oct 05 08:08:30 crc kubenswrapper[4846]: I1005 08:08:30.379717 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-2-default" Oct 05 08:08:30 crc kubenswrapper[4846]: I1005 08:08:30.531738 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bd1482a-e598-4428-ab5e-7997547e4647" path="/var/lib/kubelet/pods/4bd1482a-e598-4428-ab5e-7997547e4647/volumes" Oct 05 08:08:30 crc kubenswrapper[4846]: I1005 08:08:30.941157 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-2-default"] Oct 05 08:08:30 crc kubenswrapper[4846]: W1005 08:08:30.954561 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9e7a80d7_f91f_4d09_a3fd_fd4c1d609daf.slice/crio-0f6815f15a6ee84596409a6d1335e0b1f553ddbb63d04e1ef6081fb0ae4e3356 WatchSource:0}: Error finding container 0f6815f15a6ee84596409a6d1335e0b1f553ddbb63d04e1ef6081fb0ae4e3356: Status 404 returned error can't find the container with id 0f6815f15a6ee84596409a6d1335e0b1f553ddbb63d04e1ef6081fb0ae4e3356 Oct 05 08:08:31 crc kubenswrapper[4846]: I1005 08:08:31.118342 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-2-default" event={"ID":"9e7a80d7-f91f-4d09-a3fd-fd4c1d609daf","Type":"ContainerStarted","Data":"0f6815f15a6ee84596409a6d1335e0b1f553ddbb63d04e1ef6081fb0ae4e3356"} Oct 05 08:08:32 crc kubenswrapper[4846]: I1005 08:08:32.133747 4846 generic.go:334] "Generic (PLEG): container finished" podID="9e7a80d7-f91f-4d09-a3fd-fd4c1d609daf" containerID="8d7578bd15f9380fd60c5291373e36dbe5257462d8b2c81e3243098241213ddf" exitCode=0 Oct 05 08:08:32 crc kubenswrapper[4846]: I1005 08:08:32.133834 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-2-default" event={"ID":"9e7a80d7-f91f-4d09-a3fd-fd4c1d609daf","Type":"ContainerDied","Data":"8d7578bd15f9380fd60c5291373e36dbe5257462d8b2c81e3243098241213ddf"} Oct 05 08:08:33 crc kubenswrapper[4846]: I1005 08:08:33.630372 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-2-default" Oct 05 08:08:33 crc kubenswrapper[4846]: I1005 08:08:33.689000 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mariadb-client-2-default_9e7a80d7-f91f-4d09-a3fd-fd4c1d609daf/mariadb-client-2-default/0.log" Oct 05 08:08:33 crc kubenswrapper[4846]: I1005 08:08:33.699885 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wbt78\" (UniqueName: \"kubernetes.io/projected/9e7a80d7-f91f-4d09-a3fd-fd4c1d609daf-kube-api-access-wbt78\") pod \"9e7a80d7-f91f-4d09-a3fd-fd4c1d609daf\" (UID: \"9e7a80d7-f91f-4d09-a3fd-fd4c1d609daf\") " Oct 05 08:08:33 crc kubenswrapper[4846]: I1005 08:08:33.708363 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9e7a80d7-f91f-4d09-a3fd-fd4c1d609daf-kube-api-access-wbt78" (OuterVolumeSpecName: "kube-api-access-wbt78") pod "9e7a80d7-f91f-4d09-a3fd-fd4c1d609daf" (UID: "9e7a80d7-f91f-4d09-a3fd-fd4c1d609daf"). InnerVolumeSpecName "kube-api-access-wbt78". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:08:33 crc kubenswrapper[4846]: I1005 08:08:33.720161 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client-2-default"] Oct 05 08:08:33 crc kubenswrapper[4846]: I1005 08:08:33.735566 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client-2-default"] Oct 05 08:08:33 crc kubenswrapper[4846]: I1005 08:08:33.802395 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wbt78\" (UniqueName: \"kubernetes.io/projected/9e7a80d7-f91f-4d09-a3fd-fd4c1d609daf-kube-api-access-wbt78\") on node \"crc\" DevicePath \"\"" Oct 05 08:08:34 crc kubenswrapper[4846]: I1005 08:08:34.160971 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0f6815f15a6ee84596409a6d1335e0b1f553ddbb63d04e1ef6081fb0ae4e3356" Oct 05 08:08:34 crc kubenswrapper[4846]: I1005 08:08:34.161057 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-2-default" Oct 05 08:08:34 crc kubenswrapper[4846]: I1005 08:08:34.273600 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client-1"] Oct 05 08:08:34 crc kubenswrapper[4846]: E1005 08:08:34.274037 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9e7a80d7-f91f-4d09-a3fd-fd4c1d609daf" containerName="mariadb-client-2-default" Oct 05 08:08:34 crc kubenswrapper[4846]: I1005 08:08:34.274070 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="9e7a80d7-f91f-4d09-a3fd-fd4c1d609daf" containerName="mariadb-client-2-default" Oct 05 08:08:34 crc kubenswrapper[4846]: I1005 08:08:34.274383 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="9e7a80d7-f91f-4d09-a3fd-fd4c1d609daf" containerName="mariadb-client-2-default" Oct 05 08:08:34 crc kubenswrapper[4846]: I1005 08:08:34.275171 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-1" Oct 05 08:08:34 crc kubenswrapper[4846]: I1005 08:08:34.279142 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-4c82f" Oct 05 08:08:34 crc kubenswrapper[4846]: I1005 08:08:34.285633 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-1"] Oct 05 08:08:34 crc kubenswrapper[4846]: I1005 08:08:34.411753 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pmqmg\" (UniqueName: \"kubernetes.io/projected/96324806-7f5b-437b-a68a-9141e0c58e2a-kube-api-access-pmqmg\") pod \"mariadb-client-1\" (UID: \"96324806-7f5b-437b-a68a-9141e0c58e2a\") " pod="openstack/mariadb-client-1" Oct 05 08:08:34 crc kubenswrapper[4846]: I1005 08:08:34.518428 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pmqmg\" (UniqueName: \"kubernetes.io/projected/96324806-7f5b-437b-a68a-9141e0c58e2a-kube-api-access-pmqmg\") pod \"mariadb-client-1\" (UID: \"96324806-7f5b-437b-a68a-9141e0c58e2a\") " pod="openstack/mariadb-client-1" Oct 05 08:08:34 crc kubenswrapper[4846]: I1005 08:08:34.532825 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9e7a80d7-f91f-4d09-a3fd-fd4c1d609daf" path="/var/lib/kubelet/pods/9e7a80d7-f91f-4d09-a3fd-fd4c1d609daf/volumes" Oct 05 08:08:34 crc kubenswrapper[4846]: I1005 08:08:34.540626 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pmqmg\" (UniqueName: \"kubernetes.io/projected/96324806-7f5b-437b-a68a-9141e0c58e2a-kube-api-access-pmqmg\") pod \"mariadb-client-1\" (UID: \"96324806-7f5b-437b-a68a-9141e0c58e2a\") " pod="openstack/mariadb-client-1" Oct 05 08:08:34 crc kubenswrapper[4846]: I1005 08:08:34.610450 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-1" Oct 05 08:08:34 crc kubenswrapper[4846]: I1005 08:08:34.945713 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-1"] Oct 05 08:08:34 crc kubenswrapper[4846]: W1005 08:08:34.950706 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod96324806_7f5b_437b_a68a_9141e0c58e2a.slice/crio-c03c8c06ae70d8e2ca22208240838dd53f0fb8b2e6347f9384628810d47d4d66 WatchSource:0}: Error finding container c03c8c06ae70d8e2ca22208240838dd53f0fb8b2e6347f9384628810d47d4d66: Status 404 returned error can't find the container with id c03c8c06ae70d8e2ca22208240838dd53f0fb8b2e6347f9384628810d47d4d66 Oct 05 08:08:35 crc kubenswrapper[4846]: I1005 08:08:35.172419 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-1" event={"ID":"96324806-7f5b-437b-a68a-9141e0c58e2a","Type":"ContainerStarted","Data":"c03c8c06ae70d8e2ca22208240838dd53f0fb8b2e6347f9384628810d47d4d66"} Oct 05 08:08:36 crc kubenswrapper[4846]: I1005 08:08:36.180709 4846 generic.go:334] "Generic (PLEG): container finished" podID="96324806-7f5b-437b-a68a-9141e0c58e2a" containerID="44f8a74c28755cc48927ad4cf2d25620cd6873c2155dfaabe6e8cf0e0b2bca25" exitCode=0 Oct 05 08:08:36 crc kubenswrapper[4846]: I1005 08:08:36.180763 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-1" event={"ID":"96324806-7f5b-437b-a68a-9141e0c58e2a","Type":"ContainerDied","Data":"44f8a74c28755cc48927ad4cf2d25620cd6873c2155dfaabe6e8cf0e0b2bca25"} Oct 05 08:08:37 crc kubenswrapper[4846]: I1005 08:08:37.864903 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-1" Oct 05 08:08:37 crc kubenswrapper[4846]: I1005 08:08:37.882428 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mariadb-client-1_96324806-7f5b-437b-a68a-9141e0c58e2a/mariadb-client-1/0.log" Oct 05 08:08:37 crc kubenswrapper[4846]: I1005 08:08:37.914107 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client-1"] Oct 05 08:08:37 crc kubenswrapper[4846]: I1005 08:08:37.924395 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client-1"] Oct 05 08:08:37 crc kubenswrapper[4846]: I1005 08:08:37.975876 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pmqmg\" (UniqueName: \"kubernetes.io/projected/96324806-7f5b-437b-a68a-9141e0c58e2a-kube-api-access-pmqmg\") pod \"96324806-7f5b-437b-a68a-9141e0c58e2a\" (UID: \"96324806-7f5b-437b-a68a-9141e0c58e2a\") " Oct 05 08:08:37 crc kubenswrapper[4846]: I1005 08:08:37.984376 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96324806-7f5b-437b-a68a-9141e0c58e2a-kube-api-access-pmqmg" (OuterVolumeSpecName: "kube-api-access-pmqmg") pod "96324806-7f5b-437b-a68a-9141e0c58e2a" (UID: "96324806-7f5b-437b-a68a-9141e0c58e2a"). InnerVolumeSpecName "kube-api-access-pmqmg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:08:38 crc kubenswrapper[4846]: I1005 08:08:38.082629 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pmqmg\" (UniqueName: \"kubernetes.io/projected/96324806-7f5b-437b-a68a-9141e0c58e2a-kube-api-access-pmqmg\") on node \"crc\" DevicePath \"\"" Oct 05 08:08:38 crc kubenswrapper[4846]: I1005 08:08:38.200604 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c03c8c06ae70d8e2ca22208240838dd53f0fb8b2e6347f9384628810d47d4d66" Oct 05 08:08:38 crc kubenswrapper[4846]: I1005 08:08:38.200677 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-1" Oct 05 08:08:38 crc kubenswrapper[4846]: I1005 08:08:38.440861 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client-4-default"] Oct 05 08:08:38 crc kubenswrapper[4846]: E1005 08:08:38.441978 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="96324806-7f5b-437b-a68a-9141e0c58e2a" containerName="mariadb-client-1" Oct 05 08:08:38 crc kubenswrapper[4846]: I1005 08:08:38.442000 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="96324806-7f5b-437b-a68a-9141e0c58e2a" containerName="mariadb-client-1" Oct 05 08:08:38 crc kubenswrapper[4846]: I1005 08:08:38.442254 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="96324806-7f5b-437b-a68a-9141e0c58e2a" containerName="mariadb-client-1" Oct 05 08:08:38 crc kubenswrapper[4846]: I1005 08:08:38.443059 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-4-default" Oct 05 08:08:38 crc kubenswrapper[4846]: I1005 08:08:38.446042 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-4c82f" Oct 05 08:08:38 crc kubenswrapper[4846]: I1005 08:08:38.453459 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-4-default"] Oct 05 08:08:38 crc kubenswrapper[4846]: I1005 08:08:38.515785 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96324806-7f5b-437b-a68a-9141e0c58e2a" path="/var/lib/kubelet/pods/96324806-7f5b-437b-a68a-9141e0c58e2a/volumes" Oct 05 08:08:38 crc kubenswrapper[4846]: I1005 08:08:38.592372 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ht57x\" (UniqueName: \"kubernetes.io/projected/4a7b706d-845e-405f-92b3-f30ed5989614-kube-api-access-ht57x\") pod \"mariadb-client-4-default\" (UID: \"4a7b706d-845e-405f-92b3-f30ed5989614\") " pod="openstack/mariadb-client-4-default" Oct 05 08:08:38 crc kubenswrapper[4846]: I1005 08:08:38.694844 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ht57x\" (UniqueName: \"kubernetes.io/projected/4a7b706d-845e-405f-92b3-f30ed5989614-kube-api-access-ht57x\") pod \"mariadb-client-4-default\" (UID: \"4a7b706d-845e-405f-92b3-f30ed5989614\") " pod="openstack/mariadb-client-4-default" Oct 05 08:08:38 crc kubenswrapper[4846]: I1005 08:08:38.714950 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ht57x\" (UniqueName: \"kubernetes.io/projected/4a7b706d-845e-405f-92b3-f30ed5989614-kube-api-access-ht57x\") pod \"mariadb-client-4-default\" (UID: \"4a7b706d-845e-405f-92b3-f30ed5989614\") " pod="openstack/mariadb-client-4-default" Oct 05 08:08:38 crc kubenswrapper[4846]: I1005 08:08:38.778463 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-4-default" Oct 05 08:08:39 crc kubenswrapper[4846]: I1005 08:08:39.270137 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-4-default"] Oct 05 08:08:39 crc kubenswrapper[4846]: W1005 08:08:39.273430 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4a7b706d_845e_405f_92b3_f30ed5989614.slice/crio-860f349807ed236e433137cc257c5263d444219d0f4867b30a8dba537a30c9f1 WatchSource:0}: Error finding container 860f349807ed236e433137cc257c5263d444219d0f4867b30a8dba537a30c9f1: Status 404 returned error can't find the container with id 860f349807ed236e433137cc257c5263d444219d0f4867b30a8dba537a30c9f1 Oct 05 08:08:40 crc kubenswrapper[4846]: I1005 08:08:40.228109 4846 generic.go:334] "Generic (PLEG): container finished" podID="4a7b706d-845e-405f-92b3-f30ed5989614" containerID="9779d77d4e31773e26dd50d0a2be0bdf70553cc6b89afccf608ff6a584d82ca9" exitCode=0 Oct 05 08:08:40 crc kubenswrapper[4846]: I1005 08:08:40.228507 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-4-default" event={"ID":"4a7b706d-845e-405f-92b3-f30ed5989614","Type":"ContainerDied","Data":"9779d77d4e31773e26dd50d0a2be0bdf70553cc6b89afccf608ff6a584d82ca9"} Oct 05 08:08:40 crc kubenswrapper[4846]: I1005 08:08:40.229712 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-4-default" event={"ID":"4a7b706d-845e-405f-92b3-f30ed5989614","Type":"ContainerStarted","Data":"860f349807ed236e433137cc257c5263d444219d0f4867b30a8dba537a30c9f1"} Oct 05 08:08:41 crc kubenswrapper[4846]: I1005 08:08:41.712781 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-4-default" Oct 05 08:08:41 crc kubenswrapper[4846]: I1005 08:08:41.731755 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mariadb-client-4-default_4a7b706d-845e-405f-92b3-f30ed5989614/mariadb-client-4-default/0.log" Oct 05 08:08:41 crc kubenswrapper[4846]: I1005 08:08:41.769798 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client-4-default"] Oct 05 08:08:41 crc kubenswrapper[4846]: I1005 08:08:41.776371 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client-4-default"] Oct 05 08:08:41 crc kubenswrapper[4846]: I1005 08:08:41.843780 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ht57x\" (UniqueName: \"kubernetes.io/projected/4a7b706d-845e-405f-92b3-f30ed5989614-kube-api-access-ht57x\") pod \"4a7b706d-845e-405f-92b3-f30ed5989614\" (UID: \"4a7b706d-845e-405f-92b3-f30ed5989614\") " Oct 05 08:08:41 crc kubenswrapper[4846]: I1005 08:08:41.854554 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4a7b706d-845e-405f-92b3-f30ed5989614-kube-api-access-ht57x" (OuterVolumeSpecName: "kube-api-access-ht57x") pod "4a7b706d-845e-405f-92b3-f30ed5989614" (UID: "4a7b706d-845e-405f-92b3-f30ed5989614"). InnerVolumeSpecName "kube-api-access-ht57x". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:08:41 crc kubenswrapper[4846]: I1005 08:08:41.945862 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ht57x\" (UniqueName: \"kubernetes.io/projected/4a7b706d-845e-405f-92b3-f30ed5989614-kube-api-access-ht57x\") on node \"crc\" DevicePath \"\"" Oct 05 08:08:42 crc kubenswrapper[4846]: I1005 08:08:42.246685 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="860f349807ed236e433137cc257c5263d444219d0f4867b30a8dba537a30c9f1" Oct 05 08:08:42 crc kubenswrapper[4846]: I1005 08:08:42.246765 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-4-default" Oct 05 08:08:42 crc kubenswrapper[4846]: I1005 08:08:42.511298 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4a7b706d-845e-405f-92b3-f30ed5989614" path="/var/lib/kubelet/pods/4a7b706d-845e-405f-92b3-f30ed5989614/volumes" Oct 05 08:08:45 crc kubenswrapper[4846]: I1005 08:08:45.218349 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client-5-default"] Oct 05 08:08:45 crc kubenswrapper[4846]: E1005 08:08:45.219457 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4a7b706d-845e-405f-92b3-f30ed5989614" containerName="mariadb-client-4-default" Oct 05 08:08:45 crc kubenswrapper[4846]: I1005 08:08:45.219492 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="4a7b706d-845e-405f-92b3-f30ed5989614" containerName="mariadb-client-4-default" Oct 05 08:08:45 crc kubenswrapper[4846]: I1005 08:08:45.219855 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="4a7b706d-845e-405f-92b3-f30ed5989614" containerName="mariadb-client-4-default" Oct 05 08:08:45 crc kubenswrapper[4846]: I1005 08:08:45.221053 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-5-default" Oct 05 08:08:45 crc kubenswrapper[4846]: I1005 08:08:45.223854 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-4c82f" Oct 05 08:08:45 crc kubenswrapper[4846]: I1005 08:08:45.230712 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-5-default"] Oct 05 08:08:45 crc kubenswrapper[4846]: I1005 08:08:45.299652 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lmbwg\" (UniqueName: \"kubernetes.io/projected/bc9150db-68ec-44c7-9830-dac5b207e8ca-kube-api-access-lmbwg\") pod \"mariadb-client-5-default\" (UID: \"bc9150db-68ec-44c7-9830-dac5b207e8ca\") " pod="openstack/mariadb-client-5-default" Oct 05 08:08:45 crc kubenswrapper[4846]: I1005 08:08:45.401088 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lmbwg\" (UniqueName: \"kubernetes.io/projected/bc9150db-68ec-44c7-9830-dac5b207e8ca-kube-api-access-lmbwg\") pod \"mariadb-client-5-default\" (UID: \"bc9150db-68ec-44c7-9830-dac5b207e8ca\") " pod="openstack/mariadb-client-5-default" Oct 05 08:08:45 crc kubenswrapper[4846]: I1005 08:08:45.428793 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lmbwg\" (UniqueName: \"kubernetes.io/projected/bc9150db-68ec-44c7-9830-dac5b207e8ca-kube-api-access-lmbwg\") pod \"mariadb-client-5-default\" (UID: \"bc9150db-68ec-44c7-9830-dac5b207e8ca\") " pod="openstack/mariadb-client-5-default" Oct 05 08:08:45 crc kubenswrapper[4846]: I1005 08:08:45.553778 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-5-default" Oct 05 08:08:45 crc kubenswrapper[4846]: I1005 08:08:45.906357 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-5-default"] Oct 05 08:08:46 crc kubenswrapper[4846]: I1005 08:08:46.283849 4846 generic.go:334] "Generic (PLEG): container finished" podID="bc9150db-68ec-44c7-9830-dac5b207e8ca" containerID="de7e0ed4f28e168538cad8a0da58698b223ebf7d02900eef61442a3cde6cff4e" exitCode=0 Oct 05 08:08:46 crc kubenswrapper[4846]: I1005 08:08:46.283889 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-5-default" event={"ID":"bc9150db-68ec-44c7-9830-dac5b207e8ca","Type":"ContainerDied","Data":"de7e0ed4f28e168538cad8a0da58698b223ebf7d02900eef61442a3cde6cff4e"} Oct 05 08:08:46 crc kubenswrapper[4846]: I1005 08:08:46.283920 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-5-default" event={"ID":"bc9150db-68ec-44c7-9830-dac5b207e8ca","Type":"ContainerStarted","Data":"3c348aeb370529644a3e790d92a142fc46dbc197d66446d9db6c41c771e83a56"} Oct 05 08:08:47 crc kubenswrapper[4846]: I1005 08:08:47.750570 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-5-default" Oct 05 08:08:47 crc kubenswrapper[4846]: I1005 08:08:47.779393 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mariadb-client-5-default_bc9150db-68ec-44c7-9830-dac5b207e8ca/mariadb-client-5-default/0.log" Oct 05 08:08:47 crc kubenswrapper[4846]: I1005 08:08:47.813288 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client-5-default"] Oct 05 08:08:47 crc kubenswrapper[4846]: I1005 08:08:47.817975 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client-5-default"] Oct 05 08:08:47 crc kubenswrapper[4846]: I1005 08:08:47.843112 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lmbwg\" (UniqueName: \"kubernetes.io/projected/bc9150db-68ec-44c7-9830-dac5b207e8ca-kube-api-access-lmbwg\") pod \"bc9150db-68ec-44c7-9830-dac5b207e8ca\" (UID: \"bc9150db-68ec-44c7-9830-dac5b207e8ca\") " Oct 05 08:08:47 crc kubenswrapper[4846]: I1005 08:08:47.848392 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc9150db-68ec-44c7-9830-dac5b207e8ca-kube-api-access-lmbwg" (OuterVolumeSpecName: "kube-api-access-lmbwg") pod "bc9150db-68ec-44c7-9830-dac5b207e8ca" (UID: "bc9150db-68ec-44c7-9830-dac5b207e8ca"). InnerVolumeSpecName "kube-api-access-lmbwg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:08:47 crc kubenswrapper[4846]: I1005 08:08:47.945031 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lmbwg\" (UniqueName: \"kubernetes.io/projected/bc9150db-68ec-44c7-9830-dac5b207e8ca-kube-api-access-lmbwg\") on node \"crc\" DevicePath \"\"" Oct 05 08:08:48 crc kubenswrapper[4846]: I1005 08:08:48.013586 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client-6-default"] Oct 05 08:08:48 crc kubenswrapper[4846]: E1005 08:08:48.014107 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bc9150db-68ec-44c7-9830-dac5b207e8ca" containerName="mariadb-client-5-default" Oct 05 08:08:48 crc kubenswrapper[4846]: I1005 08:08:48.014142 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="bc9150db-68ec-44c7-9830-dac5b207e8ca" containerName="mariadb-client-5-default" Oct 05 08:08:48 crc kubenswrapper[4846]: I1005 08:08:48.014458 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="bc9150db-68ec-44c7-9830-dac5b207e8ca" containerName="mariadb-client-5-default" Oct 05 08:08:48 crc kubenswrapper[4846]: I1005 08:08:48.015229 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-6-default" Oct 05 08:08:48 crc kubenswrapper[4846]: I1005 08:08:48.025923 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-6-default"] Oct 05 08:08:48 crc kubenswrapper[4846]: I1005 08:08:48.169451 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m2qff\" (UniqueName: \"kubernetes.io/projected/d2fd7ae8-dad6-4975-96aa-11471a17a15d-kube-api-access-m2qff\") pod \"mariadb-client-6-default\" (UID: \"d2fd7ae8-dad6-4975-96aa-11471a17a15d\") " pod="openstack/mariadb-client-6-default" Oct 05 08:08:48 crc kubenswrapper[4846]: I1005 08:08:48.271245 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m2qff\" (UniqueName: \"kubernetes.io/projected/d2fd7ae8-dad6-4975-96aa-11471a17a15d-kube-api-access-m2qff\") pod \"mariadb-client-6-default\" (UID: \"d2fd7ae8-dad6-4975-96aa-11471a17a15d\") " pod="openstack/mariadb-client-6-default" Oct 05 08:08:48 crc kubenswrapper[4846]: I1005 08:08:48.287412 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m2qff\" (UniqueName: \"kubernetes.io/projected/d2fd7ae8-dad6-4975-96aa-11471a17a15d-kube-api-access-m2qff\") pod \"mariadb-client-6-default\" (UID: \"d2fd7ae8-dad6-4975-96aa-11471a17a15d\") " pod="openstack/mariadb-client-6-default" Oct 05 08:08:48 crc kubenswrapper[4846]: I1005 08:08:48.308376 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3c348aeb370529644a3e790d92a142fc46dbc197d66446d9db6c41c771e83a56" Oct 05 08:08:48 crc kubenswrapper[4846]: I1005 08:08:48.308646 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-5-default" Oct 05 08:08:48 crc kubenswrapper[4846]: I1005 08:08:48.384939 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-6-default" Oct 05 08:08:48 crc kubenswrapper[4846]: I1005 08:08:48.522657 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc9150db-68ec-44c7-9830-dac5b207e8ca" path="/var/lib/kubelet/pods/bc9150db-68ec-44c7-9830-dac5b207e8ca/volumes" Oct 05 08:08:48 crc kubenswrapper[4846]: I1005 08:08:48.729023 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-6-default"] Oct 05 08:08:49 crc kubenswrapper[4846]: W1005 08:08:49.105506 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd2fd7ae8_dad6_4975_96aa_11471a17a15d.slice/crio-e4508e9ab8938888e379ee73e07cb58229b213b5b0f9d08aad8d149790a2bd34 WatchSource:0}: Error finding container e4508e9ab8938888e379ee73e07cb58229b213b5b0f9d08aad8d149790a2bd34: Status 404 returned error can't find the container with id e4508e9ab8938888e379ee73e07cb58229b213b5b0f9d08aad8d149790a2bd34 Oct 05 08:08:49 crc kubenswrapper[4846]: I1005 08:08:49.316693 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-6-default" event={"ID":"d2fd7ae8-dad6-4975-96aa-11471a17a15d","Type":"ContainerStarted","Data":"e4508e9ab8938888e379ee73e07cb58229b213b5b0f9d08aad8d149790a2bd34"} Oct 05 08:08:50 crc kubenswrapper[4846]: I1005 08:08:50.329368 4846 generic.go:334] "Generic (PLEG): container finished" podID="d2fd7ae8-dad6-4975-96aa-11471a17a15d" containerID="2bb61e3ac89e52c35f391432b40c44b7bb3c7a7c8e6307dbd82465c4f38b2c90" exitCode=0 Oct 05 08:08:50 crc kubenswrapper[4846]: I1005 08:08:50.329428 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-6-default" event={"ID":"d2fd7ae8-dad6-4975-96aa-11471a17a15d","Type":"ContainerDied","Data":"2bb61e3ac89e52c35f391432b40c44b7bb3c7a7c8e6307dbd82465c4f38b2c90"} Oct 05 08:08:51 crc kubenswrapper[4846]: I1005 08:08:51.865157 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-6-default" Oct 05 08:08:51 crc kubenswrapper[4846]: I1005 08:08:51.921813 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mariadb-client-6-default_d2fd7ae8-dad6-4975-96aa-11471a17a15d/mariadb-client-6-default/0.log" Oct 05 08:08:51 crc kubenswrapper[4846]: I1005 08:08:51.951371 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client-6-default"] Oct 05 08:08:51 crc kubenswrapper[4846]: I1005 08:08:51.956692 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client-6-default"] Oct 05 08:08:51 crc kubenswrapper[4846]: I1005 08:08:51.961729 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m2qff\" (UniqueName: \"kubernetes.io/projected/d2fd7ae8-dad6-4975-96aa-11471a17a15d-kube-api-access-m2qff\") pod \"d2fd7ae8-dad6-4975-96aa-11471a17a15d\" (UID: \"d2fd7ae8-dad6-4975-96aa-11471a17a15d\") " Oct 05 08:08:51 crc kubenswrapper[4846]: I1005 08:08:51.971621 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d2fd7ae8-dad6-4975-96aa-11471a17a15d-kube-api-access-m2qff" (OuterVolumeSpecName: "kube-api-access-m2qff") pod "d2fd7ae8-dad6-4975-96aa-11471a17a15d" (UID: "d2fd7ae8-dad6-4975-96aa-11471a17a15d"). InnerVolumeSpecName "kube-api-access-m2qff". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:08:52 crc kubenswrapper[4846]: I1005 08:08:52.063941 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m2qff\" (UniqueName: \"kubernetes.io/projected/d2fd7ae8-dad6-4975-96aa-11471a17a15d-kube-api-access-m2qff\") on node \"crc\" DevicePath \"\"" Oct 05 08:08:52 crc kubenswrapper[4846]: I1005 08:08:52.130836 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client-7-default"] Oct 05 08:08:52 crc kubenswrapper[4846]: E1005 08:08:52.131513 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d2fd7ae8-dad6-4975-96aa-11471a17a15d" containerName="mariadb-client-6-default" Oct 05 08:08:52 crc kubenswrapper[4846]: I1005 08:08:52.131536 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="d2fd7ae8-dad6-4975-96aa-11471a17a15d" containerName="mariadb-client-6-default" Oct 05 08:08:52 crc kubenswrapper[4846]: I1005 08:08:52.131759 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="d2fd7ae8-dad6-4975-96aa-11471a17a15d" containerName="mariadb-client-6-default" Oct 05 08:08:52 crc kubenswrapper[4846]: I1005 08:08:52.134571 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-7-default" Oct 05 08:08:52 crc kubenswrapper[4846]: I1005 08:08:52.143637 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-7-default"] Oct 05 08:08:52 crc kubenswrapper[4846]: I1005 08:08:52.266348 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wdktc\" (UniqueName: \"kubernetes.io/projected/39d4f97a-1675-4326-972b-a571325f4ec9-kube-api-access-wdktc\") pod \"mariadb-client-7-default\" (UID: \"39d4f97a-1675-4326-972b-a571325f4ec9\") " pod="openstack/mariadb-client-7-default" Oct 05 08:08:52 crc kubenswrapper[4846]: I1005 08:08:52.352803 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e4508e9ab8938888e379ee73e07cb58229b213b5b0f9d08aad8d149790a2bd34" Oct 05 08:08:52 crc kubenswrapper[4846]: I1005 08:08:52.352880 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-6-default" Oct 05 08:08:52 crc kubenswrapper[4846]: I1005 08:08:52.369076 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wdktc\" (UniqueName: \"kubernetes.io/projected/39d4f97a-1675-4326-972b-a571325f4ec9-kube-api-access-wdktc\") pod \"mariadb-client-7-default\" (UID: \"39d4f97a-1675-4326-972b-a571325f4ec9\") " pod="openstack/mariadb-client-7-default" Oct 05 08:08:52 crc kubenswrapper[4846]: I1005 08:08:52.401402 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wdktc\" (UniqueName: \"kubernetes.io/projected/39d4f97a-1675-4326-972b-a571325f4ec9-kube-api-access-wdktc\") pod \"mariadb-client-7-default\" (UID: \"39d4f97a-1675-4326-972b-a571325f4ec9\") " pod="openstack/mariadb-client-7-default" Oct 05 08:08:52 crc kubenswrapper[4846]: I1005 08:08:52.458869 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-7-default" Oct 05 08:08:52 crc kubenswrapper[4846]: I1005 08:08:52.516133 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d2fd7ae8-dad6-4975-96aa-11471a17a15d" path="/var/lib/kubelet/pods/d2fd7ae8-dad6-4975-96aa-11471a17a15d/volumes" Oct 05 08:08:53 crc kubenswrapper[4846]: I1005 08:08:53.109009 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-7-default"] Oct 05 08:08:53 crc kubenswrapper[4846]: I1005 08:08:53.366950 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-7-default" event={"ID":"39d4f97a-1675-4326-972b-a571325f4ec9","Type":"ContainerStarted","Data":"e89802c66747160446da70dbc0328ef5e9d5603b4c8558d9b3a8cced6ea69467"} Oct 05 08:08:53 crc kubenswrapper[4846]: I1005 08:08:53.367490 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-7-default" event={"ID":"39d4f97a-1675-4326-972b-a571325f4ec9","Type":"ContainerStarted","Data":"988af27a2639ff89601d22472e644a56a843714af1ef3b9fc81d151da89b7282"} Oct 05 08:08:53 crc kubenswrapper[4846]: E1005 08:08:53.489618 4846 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod39d4f97a_1675_4326_972b_a571325f4ec9.slice/crio-conmon-e89802c66747160446da70dbc0328ef5e9d5603b4c8558d9b3a8cced6ea69467.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod39d4f97a_1675_4326_972b_a571325f4ec9.slice/crio-e89802c66747160446da70dbc0328ef5e9d5603b4c8558d9b3a8cced6ea69467.scope\": RecentStats: unable to find data in memory cache]" Oct 05 08:08:54 crc kubenswrapper[4846]: I1005 08:08:54.379965 4846 generic.go:334] "Generic (PLEG): container finished" podID="39d4f97a-1675-4326-972b-a571325f4ec9" containerID="e89802c66747160446da70dbc0328ef5e9d5603b4c8558d9b3a8cced6ea69467" exitCode=0 Oct 05 08:08:54 crc kubenswrapper[4846]: I1005 08:08:54.380036 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-7-default" event={"ID":"39d4f97a-1675-4326-972b-a571325f4ec9","Type":"ContainerDied","Data":"e89802c66747160446da70dbc0328ef5e9d5603b4c8558d9b3a8cced6ea69467"} Oct 05 08:08:54 crc kubenswrapper[4846]: I1005 08:08:54.884810 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-7-default" Oct 05 08:08:54 crc kubenswrapper[4846]: I1005 08:08:54.906687 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mariadb-client-7-default_39d4f97a-1675-4326-972b-a571325f4ec9/mariadb-client-7-default/0.log" Oct 05 08:08:54 crc kubenswrapper[4846]: I1005 08:08:54.940559 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client-7-default"] Oct 05 08:08:54 crc kubenswrapper[4846]: I1005 08:08:54.947006 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client-7-default"] Oct 05 08:08:55 crc kubenswrapper[4846]: I1005 08:08:55.015658 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wdktc\" (UniqueName: \"kubernetes.io/projected/39d4f97a-1675-4326-972b-a571325f4ec9-kube-api-access-wdktc\") pod \"39d4f97a-1675-4326-972b-a571325f4ec9\" (UID: \"39d4f97a-1675-4326-972b-a571325f4ec9\") " Oct 05 08:08:55 crc kubenswrapper[4846]: I1005 08:08:55.025353 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/39d4f97a-1675-4326-972b-a571325f4ec9-kube-api-access-wdktc" (OuterVolumeSpecName: "kube-api-access-wdktc") pod "39d4f97a-1675-4326-972b-a571325f4ec9" (UID: "39d4f97a-1675-4326-972b-a571325f4ec9"). InnerVolumeSpecName "kube-api-access-wdktc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:08:55 crc kubenswrapper[4846]: I1005 08:08:55.118225 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wdktc\" (UniqueName: \"kubernetes.io/projected/39d4f97a-1675-4326-972b-a571325f4ec9-kube-api-access-wdktc\") on node \"crc\" DevicePath \"\"" Oct 05 08:08:55 crc kubenswrapper[4846]: I1005 08:08:55.136853 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client-2"] Oct 05 08:08:55 crc kubenswrapper[4846]: E1005 08:08:55.137404 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="39d4f97a-1675-4326-972b-a571325f4ec9" containerName="mariadb-client-7-default" Oct 05 08:08:55 crc kubenswrapper[4846]: I1005 08:08:55.137431 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="39d4f97a-1675-4326-972b-a571325f4ec9" containerName="mariadb-client-7-default" Oct 05 08:08:55 crc kubenswrapper[4846]: I1005 08:08:55.137700 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="39d4f97a-1675-4326-972b-a571325f4ec9" containerName="mariadb-client-7-default" Oct 05 08:08:55 crc kubenswrapper[4846]: I1005 08:08:55.138451 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-2" Oct 05 08:08:55 crc kubenswrapper[4846]: I1005 08:08:55.148899 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-2"] Oct 05 08:08:55 crc kubenswrapper[4846]: I1005 08:08:55.219960 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h8zsn\" (UniqueName: \"kubernetes.io/projected/bc60f646-a481-40fc-91ee-c7d000c9646e-kube-api-access-h8zsn\") pod \"mariadb-client-2\" (UID: \"bc60f646-a481-40fc-91ee-c7d000c9646e\") " pod="openstack/mariadb-client-2" Oct 05 08:08:55 crc kubenswrapper[4846]: I1005 08:08:55.322090 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h8zsn\" (UniqueName: \"kubernetes.io/projected/bc60f646-a481-40fc-91ee-c7d000c9646e-kube-api-access-h8zsn\") pod \"mariadb-client-2\" (UID: \"bc60f646-a481-40fc-91ee-c7d000c9646e\") " pod="openstack/mariadb-client-2" Oct 05 08:08:55 crc kubenswrapper[4846]: I1005 08:08:55.354982 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h8zsn\" (UniqueName: \"kubernetes.io/projected/bc60f646-a481-40fc-91ee-c7d000c9646e-kube-api-access-h8zsn\") pod \"mariadb-client-2\" (UID: \"bc60f646-a481-40fc-91ee-c7d000c9646e\") " pod="openstack/mariadb-client-2" Oct 05 08:08:55 crc kubenswrapper[4846]: I1005 08:08:55.399496 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="988af27a2639ff89601d22472e644a56a843714af1ef3b9fc81d151da89b7282" Oct 05 08:08:55 crc kubenswrapper[4846]: I1005 08:08:55.399571 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-7-default" Oct 05 08:08:55 crc kubenswrapper[4846]: I1005 08:08:55.474248 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-2" Oct 05 08:08:56 crc kubenswrapper[4846]: I1005 08:08:56.051511 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-2"] Oct 05 08:08:56 crc kubenswrapper[4846]: I1005 08:08:56.413761 4846 generic.go:334] "Generic (PLEG): container finished" podID="bc60f646-a481-40fc-91ee-c7d000c9646e" containerID="ba8753d9a7d9578062a3d1bc4fea0af38c3bae6bbe52523e9a0a59a10da8c886" exitCode=0 Oct 05 08:08:56 crc kubenswrapper[4846]: I1005 08:08:56.413878 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-2" event={"ID":"bc60f646-a481-40fc-91ee-c7d000c9646e","Type":"ContainerDied","Data":"ba8753d9a7d9578062a3d1bc4fea0af38c3bae6bbe52523e9a0a59a10da8c886"} Oct 05 08:08:56 crc kubenswrapper[4846]: I1005 08:08:56.414212 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-2" event={"ID":"bc60f646-a481-40fc-91ee-c7d000c9646e","Type":"ContainerStarted","Data":"ab7ce226453594782f9bda68af0683cd0738b63588134eea1d22f5e79317576e"} Oct 05 08:08:56 crc kubenswrapper[4846]: I1005 08:08:56.512368 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="39d4f97a-1675-4326-972b-a571325f4ec9" path="/var/lib/kubelet/pods/39d4f97a-1675-4326-972b-a571325f4ec9/volumes" Oct 05 08:08:57 crc kubenswrapper[4846]: I1005 08:08:57.875488 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-2" Oct 05 08:08:57 crc kubenswrapper[4846]: I1005 08:08:57.899386 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mariadb-client-2_bc60f646-a481-40fc-91ee-c7d000c9646e/mariadb-client-2/0.log" Oct 05 08:08:57 crc kubenswrapper[4846]: I1005 08:08:57.960139 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client-2"] Oct 05 08:08:57 crc kubenswrapper[4846]: I1005 08:08:57.964878 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h8zsn\" (UniqueName: \"kubernetes.io/projected/bc60f646-a481-40fc-91ee-c7d000c9646e-kube-api-access-h8zsn\") pod \"bc60f646-a481-40fc-91ee-c7d000c9646e\" (UID: \"bc60f646-a481-40fc-91ee-c7d000c9646e\") " Oct 05 08:08:57 crc kubenswrapper[4846]: I1005 08:08:57.972235 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client-2"] Oct 05 08:08:57 crc kubenswrapper[4846]: I1005 08:08:57.973970 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc60f646-a481-40fc-91ee-c7d000c9646e-kube-api-access-h8zsn" (OuterVolumeSpecName: "kube-api-access-h8zsn") pod "bc60f646-a481-40fc-91ee-c7d000c9646e" (UID: "bc60f646-a481-40fc-91ee-c7d000c9646e"). InnerVolumeSpecName "kube-api-access-h8zsn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:08:58 crc kubenswrapper[4846]: I1005 08:08:58.068312 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h8zsn\" (UniqueName: \"kubernetes.io/projected/bc60f646-a481-40fc-91ee-c7d000c9646e-kube-api-access-h8zsn\") on node \"crc\" DevicePath \"\"" Oct 05 08:08:58 crc kubenswrapper[4846]: I1005 08:08:58.436409 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ab7ce226453594782f9bda68af0683cd0738b63588134eea1d22f5e79317576e" Oct 05 08:08:58 crc kubenswrapper[4846]: I1005 08:08:58.436539 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-2" Oct 05 08:08:58 crc kubenswrapper[4846]: I1005 08:08:58.516156 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc60f646-a481-40fc-91ee-c7d000c9646e" path="/var/lib/kubelet/pods/bc60f646-a481-40fc-91ee-c7d000c9646e/volumes" Oct 05 08:09:23 crc kubenswrapper[4846]: I1005 08:09:23.324760 4846 patch_prober.go:28] interesting pod/machine-config-daemon-fscvf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 08:09:23 crc kubenswrapper[4846]: I1005 08:09:23.325426 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 08:09:53 crc kubenswrapper[4846]: I1005 08:09:53.325010 4846 patch_prober.go:28] interesting pod/machine-config-daemon-fscvf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 08:09:53 crc kubenswrapper[4846]: I1005 08:09:53.325809 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 08:10:23 crc kubenswrapper[4846]: I1005 08:10:23.325173 4846 patch_prober.go:28] interesting pod/machine-config-daemon-fscvf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 08:10:23 crc kubenswrapper[4846]: I1005 08:10:23.326097 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 08:10:23 crc kubenswrapper[4846]: I1005 08:10:23.326170 4846 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" Oct 05 08:10:23 crc kubenswrapper[4846]: I1005 08:10:23.327475 4846 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"8fa72be735dff8e623a4a4958f452896ef79e72dd44c2d714a3019187462de47"} pod="openshift-machine-config-operator/machine-config-daemon-fscvf" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 05 08:10:23 crc kubenswrapper[4846]: I1005 08:10:23.327586 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" containerName="machine-config-daemon" containerID="cri-o://8fa72be735dff8e623a4a4958f452896ef79e72dd44c2d714a3019187462de47" gracePeriod=600 Oct 05 08:10:23 crc kubenswrapper[4846]: E1005 08:10:23.478149 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 08:10:24 crc kubenswrapper[4846]: I1005 08:10:24.307398 4846 generic.go:334] "Generic (PLEG): container finished" podID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" containerID="8fa72be735dff8e623a4a4958f452896ef79e72dd44c2d714a3019187462de47" exitCode=0 Oct 05 08:10:24 crc kubenswrapper[4846]: I1005 08:10:24.307501 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" event={"ID":"ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7","Type":"ContainerDied","Data":"8fa72be735dff8e623a4a4958f452896ef79e72dd44c2d714a3019187462de47"} Oct 05 08:10:24 crc kubenswrapper[4846]: I1005 08:10:24.307631 4846 scope.go:117] "RemoveContainer" containerID="bad87dae9648a1776fbb2ee38b4fedce1273888d4f401172760726d09bb965f3" Oct 05 08:10:24 crc kubenswrapper[4846]: I1005 08:10:24.308607 4846 scope.go:117] "RemoveContainer" containerID="8fa72be735dff8e623a4a4958f452896ef79e72dd44c2d714a3019187462de47" Oct 05 08:10:24 crc kubenswrapper[4846]: E1005 08:10:24.309142 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 08:10:30 crc kubenswrapper[4846]: I1005 08:10:30.415270 4846 scope.go:117] "RemoveContainer" containerID="98e5c44de4fcb0e5e35062d5d1b192dd535a1dd475399e57eb0037da03fc51cb" Oct 05 08:10:35 crc kubenswrapper[4846]: I1005 08:10:35.498566 4846 scope.go:117] "RemoveContainer" containerID="8fa72be735dff8e623a4a4958f452896ef79e72dd44c2d714a3019187462de47" Oct 05 08:10:35 crc kubenswrapper[4846]: E1005 08:10:35.499418 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 08:10:48 crc kubenswrapper[4846]: I1005 08:10:48.508289 4846 scope.go:117] "RemoveContainer" containerID="8fa72be735dff8e623a4a4958f452896ef79e72dd44c2d714a3019187462de47" Oct 05 08:10:48 crc kubenswrapper[4846]: E1005 08:10:48.509685 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 08:11:03 crc kubenswrapper[4846]: I1005 08:11:03.497888 4846 scope.go:117] "RemoveContainer" containerID="8fa72be735dff8e623a4a4958f452896ef79e72dd44c2d714a3019187462de47" Oct 05 08:11:03 crc kubenswrapper[4846]: E1005 08:11:03.498748 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 08:11:16 crc kubenswrapper[4846]: I1005 08:11:16.497361 4846 scope.go:117] "RemoveContainer" containerID="8fa72be735dff8e623a4a4958f452896ef79e72dd44c2d714a3019187462de47" Oct 05 08:11:16 crc kubenswrapper[4846]: E1005 08:11:16.498428 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 08:11:27 crc kubenswrapper[4846]: I1005 08:11:27.498040 4846 scope.go:117] "RemoveContainer" containerID="8fa72be735dff8e623a4a4958f452896ef79e72dd44c2d714a3019187462de47" Oct 05 08:11:27 crc kubenswrapper[4846]: E1005 08:11:27.499358 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 08:11:42 crc kubenswrapper[4846]: I1005 08:11:42.498994 4846 scope.go:117] "RemoveContainer" containerID="8fa72be735dff8e623a4a4958f452896ef79e72dd44c2d714a3019187462de47" Oct 05 08:11:42 crc kubenswrapper[4846]: E1005 08:11:42.499936 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 08:11:44 crc kubenswrapper[4846]: I1005 08:11:44.136493 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-copy-data"] Oct 05 08:11:44 crc kubenswrapper[4846]: E1005 08:11:44.137361 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bc60f646-a481-40fc-91ee-c7d000c9646e" containerName="mariadb-client-2" Oct 05 08:11:44 crc kubenswrapper[4846]: I1005 08:11:44.137390 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="bc60f646-a481-40fc-91ee-c7d000c9646e" containerName="mariadb-client-2" Oct 05 08:11:44 crc kubenswrapper[4846]: I1005 08:11:44.137717 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="bc60f646-a481-40fc-91ee-c7d000c9646e" containerName="mariadb-client-2" Oct 05 08:11:44 crc kubenswrapper[4846]: I1005 08:11:44.138857 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-copy-data" Oct 05 08:11:44 crc kubenswrapper[4846]: I1005 08:11:44.142808 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-4c82f" Oct 05 08:11:44 crc kubenswrapper[4846]: I1005 08:11:44.158567 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-copy-data"] Oct 05 08:11:44 crc kubenswrapper[4846]: I1005 08:11:44.291028 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-8c01e9e7-1c79-4d8c-a6a6-9aa1977a5bc1\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-8c01e9e7-1c79-4d8c-a6a6-9aa1977a5bc1\") pod \"mariadb-copy-data\" (UID: \"1edceaaf-ed3d-4595-985f-91d045d101c7\") " pod="openstack/mariadb-copy-data" Oct 05 08:11:44 crc kubenswrapper[4846]: I1005 08:11:44.291218 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t7gqx\" (UniqueName: \"kubernetes.io/projected/1edceaaf-ed3d-4595-985f-91d045d101c7-kube-api-access-t7gqx\") pod \"mariadb-copy-data\" (UID: \"1edceaaf-ed3d-4595-985f-91d045d101c7\") " pod="openstack/mariadb-copy-data" Oct 05 08:11:44 crc kubenswrapper[4846]: I1005 08:11:44.392622 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-8c01e9e7-1c79-4d8c-a6a6-9aa1977a5bc1\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-8c01e9e7-1c79-4d8c-a6a6-9aa1977a5bc1\") pod \"mariadb-copy-data\" (UID: \"1edceaaf-ed3d-4595-985f-91d045d101c7\") " pod="openstack/mariadb-copy-data" Oct 05 08:11:44 crc kubenswrapper[4846]: I1005 08:11:44.392758 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t7gqx\" (UniqueName: \"kubernetes.io/projected/1edceaaf-ed3d-4595-985f-91d045d101c7-kube-api-access-t7gqx\") pod \"mariadb-copy-data\" (UID: \"1edceaaf-ed3d-4595-985f-91d045d101c7\") " pod="openstack/mariadb-copy-data" Oct 05 08:11:44 crc kubenswrapper[4846]: I1005 08:11:44.396157 4846 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 05 08:11:44 crc kubenswrapper[4846]: I1005 08:11:44.396217 4846 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-8c01e9e7-1c79-4d8c-a6a6-9aa1977a5bc1\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-8c01e9e7-1c79-4d8c-a6a6-9aa1977a5bc1\") pod \"mariadb-copy-data\" (UID: \"1edceaaf-ed3d-4595-985f-91d045d101c7\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1a56a2fcb104a5b762040a1e96254786cd1b34e7024ba051dc382707671d7f3e/globalmount\"" pod="openstack/mariadb-copy-data" Oct 05 08:11:44 crc kubenswrapper[4846]: I1005 08:11:44.416894 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t7gqx\" (UniqueName: \"kubernetes.io/projected/1edceaaf-ed3d-4595-985f-91d045d101c7-kube-api-access-t7gqx\") pod \"mariadb-copy-data\" (UID: \"1edceaaf-ed3d-4595-985f-91d045d101c7\") " pod="openstack/mariadb-copy-data" Oct 05 08:11:44 crc kubenswrapper[4846]: I1005 08:11:44.424262 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-8c01e9e7-1c79-4d8c-a6a6-9aa1977a5bc1\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-8c01e9e7-1c79-4d8c-a6a6-9aa1977a5bc1\") pod \"mariadb-copy-data\" (UID: \"1edceaaf-ed3d-4595-985f-91d045d101c7\") " pod="openstack/mariadb-copy-data" Oct 05 08:11:44 crc kubenswrapper[4846]: I1005 08:11:44.485629 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-copy-data" Oct 05 08:11:45 crc kubenswrapper[4846]: I1005 08:11:45.125382 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-copy-data"] Oct 05 08:11:45 crc kubenswrapper[4846]: W1005 08:11:45.129216 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1edceaaf_ed3d_4595_985f_91d045d101c7.slice/crio-3aa0c20bc7a0191ec1ca1e2908aac1495dcb12691c1b3c0032ff93e51adb1046 WatchSource:0}: Error finding container 3aa0c20bc7a0191ec1ca1e2908aac1495dcb12691c1b3c0032ff93e51adb1046: Status 404 returned error can't find the container with id 3aa0c20bc7a0191ec1ca1e2908aac1495dcb12691c1b3c0032ff93e51adb1046 Oct 05 08:11:46 crc kubenswrapper[4846]: I1005 08:11:46.126392 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-copy-data" event={"ID":"1edceaaf-ed3d-4595-985f-91d045d101c7","Type":"ContainerStarted","Data":"1647da17d92398b042c0915aded45cad06c01a4fd9210b9cf619d434c72384e0"} Oct 05 08:11:46 crc kubenswrapper[4846]: I1005 08:11:46.126792 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-copy-data" event={"ID":"1edceaaf-ed3d-4595-985f-91d045d101c7","Type":"ContainerStarted","Data":"3aa0c20bc7a0191ec1ca1e2908aac1495dcb12691c1b3c0032ff93e51adb1046"} Oct 05 08:11:46 crc kubenswrapper[4846]: I1005 08:11:46.147065 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/mariadb-copy-data" podStartSLOduration=3.147049806 podStartE2EDuration="3.147049806s" podCreationTimestamp="2025-10-05 08:11:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 08:11:46.141453196 +0000 UTC m=+5028.382305981" watchObservedRunningTime="2025-10-05 08:11:46.147049806 +0000 UTC m=+5028.387902581" Oct 05 08:11:48 crc kubenswrapper[4846]: I1005 08:11:48.438462 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client"] Oct 05 08:11:48 crc kubenswrapper[4846]: I1005 08:11:48.440481 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Oct 05 08:11:48 crc kubenswrapper[4846]: I1005 08:11:48.445717 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client"] Oct 05 08:11:48 crc kubenswrapper[4846]: I1005 08:11:48.566630 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tfpzh\" (UniqueName: \"kubernetes.io/projected/c2c743c4-101c-4662-bf00-f6a192561c7a-kube-api-access-tfpzh\") pod \"mariadb-client\" (UID: \"c2c743c4-101c-4662-bf00-f6a192561c7a\") " pod="openstack/mariadb-client" Oct 05 08:11:48 crc kubenswrapper[4846]: I1005 08:11:48.668650 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tfpzh\" (UniqueName: \"kubernetes.io/projected/c2c743c4-101c-4662-bf00-f6a192561c7a-kube-api-access-tfpzh\") pod \"mariadb-client\" (UID: \"c2c743c4-101c-4662-bf00-f6a192561c7a\") " pod="openstack/mariadb-client" Oct 05 08:11:48 crc kubenswrapper[4846]: I1005 08:11:48.691608 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tfpzh\" (UniqueName: \"kubernetes.io/projected/c2c743c4-101c-4662-bf00-f6a192561c7a-kube-api-access-tfpzh\") pod \"mariadb-client\" (UID: \"c2c743c4-101c-4662-bf00-f6a192561c7a\") " pod="openstack/mariadb-client" Oct 05 08:11:48 crc kubenswrapper[4846]: I1005 08:11:48.773762 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Oct 05 08:11:49 crc kubenswrapper[4846]: I1005 08:11:49.318478 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client"] Oct 05 08:11:50 crc kubenswrapper[4846]: I1005 08:11:50.165011 4846 generic.go:334] "Generic (PLEG): container finished" podID="c2c743c4-101c-4662-bf00-f6a192561c7a" containerID="83d22853a3b1f66f4f6d0f405a062e2172344cab88ae034b26b2f8a384bc74b0" exitCode=0 Oct 05 08:11:50 crc kubenswrapper[4846]: I1005 08:11:50.165060 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client" event={"ID":"c2c743c4-101c-4662-bf00-f6a192561c7a","Type":"ContainerDied","Data":"83d22853a3b1f66f4f6d0f405a062e2172344cab88ae034b26b2f8a384bc74b0"} Oct 05 08:11:50 crc kubenswrapper[4846]: I1005 08:11:50.165091 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client" event={"ID":"c2c743c4-101c-4662-bf00-f6a192561c7a","Type":"ContainerStarted","Data":"9683a791a3700cd1faeceb87f3d25885f2fdaa0e01dc680b8bfd2d3d4269db07"} Oct 05 08:11:51 crc kubenswrapper[4846]: I1005 08:11:51.495421 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Oct 05 08:11:51 crc kubenswrapper[4846]: I1005 08:11:51.524439 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mariadb-client_c2c743c4-101c-4662-bf00-f6a192561c7a/mariadb-client/0.log" Oct 05 08:11:51 crc kubenswrapper[4846]: I1005 08:11:51.550900 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client"] Oct 05 08:11:51 crc kubenswrapper[4846]: I1005 08:11:51.557507 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client"] Oct 05 08:11:51 crc kubenswrapper[4846]: I1005 08:11:51.615258 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tfpzh\" (UniqueName: \"kubernetes.io/projected/c2c743c4-101c-4662-bf00-f6a192561c7a-kube-api-access-tfpzh\") pod \"c2c743c4-101c-4662-bf00-f6a192561c7a\" (UID: \"c2c743c4-101c-4662-bf00-f6a192561c7a\") " Oct 05 08:11:51 crc kubenswrapper[4846]: I1005 08:11:51.625498 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c2c743c4-101c-4662-bf00-f6a192561c7a-kube-api-access-tfpzh" (OuterVolumeSpecName: "kube-api-access-tfpzh") pod "c2c743c4-101c-4662-bf00-f6a192561c7a" (UID: "c2c743c4-101c-4662-bf00-f6a192561c7a"). InnerVolumeSpecName "kube-api-access-tfpzh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:11:51 crc kubenswrapper[4846]: I1005 08:11:51.716956 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tfpzh\" (UniqueName: \"kubernetes.io/projected/c2c743c4-101c-4662-bf00-f6a192561c7a-kube-api-access-tfpzh\") on node \"crc\" DevicePath \"\"" Oct 05 08:11:51 crc kubenswrapper[4846]: I1005 08:11:51.750959 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client"] Oct 05 08:11:51 crc kubenswrapper[4846]: E1005 08:11:51.751501 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c2c743c4-101c-4662-bf00-f6a192561c7a" containerName="mariadb-client" Oct 05 08:11:51 crc kubenswrapper[4846]: I1005 08:11:51.751528 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="c2c743c4-101c-4662-bf00-f6a192561c7a" containerName="mariadb-client" Oct 05 08:11:51 crc kubenswrapper[4846]: I1005 08:11:51.751684 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="c2c743c4-101c-4662-bf00-f6a192561c7a" containerName="mariadb-client" Oct 05 08:11:51 crc kubenswrapper[4846]: I1005 08:11:51.752328 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Oct 05 08:11:51 crc kubenswrapper[4846]: I1005 08:11:51.759779 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client"] Oct 05 08:11:51 crc kubenswrapper[4846]: I1005 08:11:51.818424 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j6znd\" (UniqueName: \"kubernetes.io/projected/1a681e6c-1ba5-4b9a-aeb6-0dd0b07ffc18-kube-api-access-j6znd\") pod \"mariadb-client\" (UID: \"1a681e6c-1ba5-4b9a-aeb6-0dd0b07ffc18\") " pod="openstack/mariadb-client" Oct 05 08:11:51 crc kubenswrapper[4846]: I1005 08:11:51.919838 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j6znd\" (UniqueName: \"kubernetes.io/projected/1a681e6c-1ba5-4b9a-aeb6-0dd0b07ffc18-kube-api-access-j6znd\") pod \"mariadb-client\" (UID: \"1a681e6c-1ba5-4b9a-aeb6-0dd0b07ffc18\") " pod="openstack/mariadb-client" Oct 05 08:11:51 crc kubenswrapper[4846]: I1005 08:11:51.942565 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j6znd\" (UniqueName: \"kubernetes.io/projected/1a681e6c-1ba5-4b9a-aeb6-0dd0b07ffc18-kube-api-access-j6znd\") pod \"mariadb-client\" (UID: \"1a681e6c-1ba5-4b9a-aeb6-0dd0b07ffc18\") " pod="openstack/mariadb-client" Oct 05 08:11:52 crc kubenswrapper[4846]: I1005 08:11:52.080788 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Oct 05 08:11:52 crc kubenswrapper[4846]: I1005 08:11:52.191443 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9683a791a3700cd1faeceb87f3d25885f2fdaa0e01dc680b8bfd2d3d4269db07" Oct 05 08:11:52 crc kubenswrapper[4846]: I1005 08:11:52.191699 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Oct 05 08:11:52 crc kubenswrapper[4846]: I1005 08:11:52.218194 4846 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/mariadb-client" oldPodUID="c2c743c4-101c-4662-bf00-f6a192561c7a" podUID="1a681e6c-1ba5-4b9a-aeb6-0dd0b07ffc18" Oct 05 08:11:52 crc kubenswrapper[4846]: I1005 08:11:52.340766 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client"] Oct 05 08:11:52 crc kubenswrapper[4846]: W1005 08:11:52.343294 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1a681e6c_1ba5_4b9a_aeb6_0dd0b07ffc18.slice/crio-b42039e351c47b2cf64896492c7cb0282f3f3c303eb8654a7d56a6d582dab21f WatchSource:0}: Error finding container b42039e351c47b2cf64896492c7cb0282f3f3c303eb8654a7d56a6d582dab21f: Status 404 returned error can't find the container with id b42039e351c47b2cf64896492c7cb0282f3f3c303eb8654a7d56a6d582dab21f Oct 05 08:11:52 crc kubenswrapper[4846]: I1005 08:11:52.519323 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c2c743c4-101c-4662-bf00-f6a192561c7a" path="/var/lib/kubelet/pods/c2c743c4-101c-4662-bf00-f6a192561c7a/volumes" Oct 05 08:11:53 crc kubenswrapper[4846]: I1005 08:11:53.211253 4846 generic.go:334] "Generic (PLEG): container finished" podID="1a681e6c-1ba5-4b9a-aeb6-0dd0b07ffc18" containerID="908006bdcb1c13dd53d795ae9d9c6e3a93be46060c089674ede060bc0113e262" exitCode=0 Oct 05 08:11:53 crc kubenswrapper[4846]: I1005 08:11:53.211311 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client" event={"ID":"1a681e6c-1ba5-4b9a-aeb6-0dd0b07ffc18","Type":"ContainerDied","Data":"908006bdcb1c13dd53d795ae9d9c6e3a93be46060c089674ede060bc0113e262"} Oct 05 08:11:53 crc kubenswrapper[4846]: I1005 08:11:53.211348 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client" event={"ID":"1a681e6c-1ba5-4b9a-aeb6-0dd0b07ffc18","Type":"ContainerStarted","Data":"b42039e351c47b2cf64896492c7cb0282f3f3c303eb8654a7d56a6d582dab21f"} Oct 05 08:11:54 crc kubenswrapper[4846]: I1005 08:11:54.596465 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Oct 05 08:11:54 crc kubenswrapper[4846]: I1005 08:11:54.616506 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mariadb-client_1a681e6c-1ba5-4b9a-aeb6-0dd0b07ffc18/mariadb-client/0.log" Oct 05 08:11:54 crc kubenswrapper[4846]: I1005 08:11:54.644602 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client"] Oct 05 08:11:54 crc kubenswrapper[4846]: I1005 08:11:54.650198 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client"] Oct 05 08:11:54 crc kubenswrapper[4846]: I1005 08:11:54.681146 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j6znd\" (UniqueName: \"kubernetes.io/projected/1a681e6c-1ba5-4b9a-aeb6-0dd0b07ffc18-kube-api-access-j6znd\") pod \"1a681e6c-1ba5-4b9a-aeb6-0dd0b07ffc18\" (UID: \"1a681e6c-1ba5-4b9a-aeb6-0dd0b07ffc18\") " Oct 05 08:11:54 crc kubenswrapper[4846]: I1005 08:11:54.685709 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1a681e6c-1ba5-4b9a-aeb6-0dd0b07ffc18-kube-api-access-j6znd" (OuterVolumeSpecName: "kube-api-access-j6znd") pod "1a681e6c-1ba5-4b9a-aeb6-0dd0b07ffc18" (UID: "1a681e6c-1ba5-4b9a-aeb6-0dd0b07ffc18"). InnerVolumeSpecName "kube-api-access-j6znd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:11:54 crc kubenswrapper[4846]: I1005 08:11:54.783977 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j6znd\" (UniqueName: \"kubernetes.io/projected/1a681e6c-1ba5-4b9a-aeb6-0dd0b07ffc18-kube-api-access-j6znd\") on node \"crc\" DevicePath \"\"" Oct 05 08:11:55 crc kubenswrapper[4846]: I1005 08:11:55.234254 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b42039e351c47b2cf64896492c7cb0282f3f3c303eb8654a7d56a6d582dab21f" Oct 05 08:11:55 crc kubenswrapper[4846]: I1005 08:11:55.234443 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Oct 05 08:11:55 crc kubenswrapper[4846]: I1005 08:11:55.497863 4846 scope.go:117] "RemoveContainer" containerID="8fa72be735dff8e623a4a4958f452896ef79e72dd44c2d714a3019187462de47" Oct 05 08:11:55 crc kubenswrapper[4846]: E1005 08:11:55.498455 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 08:11:56 crc kubenswrapper[4846]: I1005 08:11:56.515517 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1a681e6c-1ba5-4b9a-aeb6-0dd0b07ffc18" path="/var/lib/kubelet/pods/1a681e6c-1ba5-4b9a-aeb6-0dd0b07ffc18/volumes" Oct 05 08:12:07 crc kubenswrapper[4846]: I1005 08:12:07.497911 4846 scope.go:117] "RemoveContainer" containerID="8fa72be735dff8e623a4a4958f452896ef79e72dd44c2d714a3019187462de47" Oct 05 08:12:07 crc kubenswrapper[4846]: E1005 08:12:07.499222 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 08:12:18 crc kubenswrapper[4846]: I1005 08:12:18.517986 4846 scope.go:117] "RemoveContainer" containerID="8fa72be735dff8e623a4a4958f452896ef79e72dd44c2d714a3019187462de47" Oct 05 08:12:18 crc kubenswrapper[4846]: E1005 08:12:18.519031 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 08:12:28 crc kubenswrapper[4846]: I1005 08:12:28.483359 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-0"] Oct 05 08:12:28 crc kubenswrapper[4846]: E1005 08:12:28.488463 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1a681e6c-1ba5-4b9a-aeb6-0dd0b07ffc18" containerName="mariadb-client" Oct 05 08:12:28 crc kubenswrapper[4846]: I1005 08:12:28.488839 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="1a681e6c-1ba5-4b9a-aeb6-0dd0b07ffc18" containerName="mariadb-client" Oct 05 08:12:28 crc kubenswrapper[4846]: I1005 08:12:28.490427 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="1a681e6c-1ba5-4b9a-aeb6-0dd0b07ffc18" containerName="mariadb-client" Oct 05 08:12:28 crc kubenswrapper[4846]: I1005 08:12:28.491783 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Oct 05 08:12:28 crc kubenswrapper[4846]: I1005 08:12:28.497448 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-scripts" Oct 05 08:12:28 crc kubenswrapper[4846]: I1005 08:12:28.497542 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovn-metrics" Oct 05 08:12:28 crc kubenswrapper[4846]: I1005 08:12:28.497717 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-nb-ovndbs" Oct 05 08:12:28 crc kubenswrapper[4846]: I1005 08:12:28.497872 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-config" Oct 05 08:12:28 crc kubenswrapper[4846]: I1005 08:12:28.498039 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-nb-dockercfg-knxtk" Oct 05 08:12:28 crc kubenswrapper[4846]: I1005 08:12:28.569472 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Oct 05 08:12:28 crc kubenswrapper[4846]: I1005 08:12:28.570860 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-1"] Oct 05 08:12:28 crc kubenswrapper[4846]: I1005 08:12:28.573637 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-1" Oct 05 08:12:28 crc kubenswrapper[4846]: I1005 08:12:28.579722 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-2"] Oct 05 08:12:28 crc kubenswrapper[4846]: I1005 08:12:28.581005 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-2" Oct 05 08:12:28 crc kubenswrapper[4846]: I1005 08:12:28.585151 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-1"] Oct 05 08:12:28 crc kubenswrapper[4846]: I1005 08:12:28.591308 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-2"] Oct 05 08:12:28 crc kubenswrapper[4846]: I1005 08:12:28.621920 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/52e7bf3e-f881-4ff2-aae6-40aed0788d54-config\") pod \"ovsdbserver-nb-0\" (UID: \"52e7bf3e-f881-4ff2-aae6-40aed0788d54\") " pod="openstack/ovsdbserver-nb-0" Oct 05 08:12:28 crc kubenswrapper[4846]: I1005 08:12:28.622042 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/52e7bf3e-f881-4ff2-aae6-40aed0788d54-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"52e7bf3e-f881-4ff2-aae6-40aed0788d54\") " pod="openstack/ovsdbserver-nb-0" Oct 05 08:12:28 crc kubenswrapper[4846]: I1005 08:12:28.622083 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g4mm2\" (UniqueName: \"kubernetes.io/projected/52e7bf3e-f881-4ff2-aae6-40aed0788d54-kube-api-access-g4mm2\") pod \"ovsdbserver-nb-0\" (UID: \"52e7bf3e-f881-4ff2-aae6-40aed0788d54\") " pod="openstack/ovsdbserver-nb-0" Oct 05 08:12:28 crc kubenswrapper[4846]: I1005 08:12:28.622106 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/52e7bf3e-f881-4ff2-aae6-40aed0788d54-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"52e7bf3e-f881-4ff2-aae6-40aed0788d54\") " pod="openstack/ovsdbserver-nb-0" Oct 05 08:12:28 crc kubenswrapper[4846]: I1005 08:12:28.622164 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-ac493672-d90d-4cff-aa5a-f3fd0a2eab10\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ac493672-d90d-4cff-aa5a-f3fd0a2eab10\") pod \"ovsdbserver-nb-0\" (UID: \"52e7bf3e-f881-4ff2-aae6-40aed0788d54\") " pod="openstack/ovsdbserver-nb-0" Oct 05 08:12:28 crc kubenswrapper[4846]: I1005 08:12:28.622228 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/52e7bf3e-f881-4ff2-aae6-40aed0788d54-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"52e7bf3e-f881-4ff2-aae6-40aed0788d54\") " pod="openstack/ovsdbserver-nb-0" Oct 05 08:12:28 crc kubenswrapper[4846]: I1005 08:12:28.622245 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/52e7bf3e-f881-4ff2-aae6-40aed0788d54-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"52e7bf3e-f881-4ff2-aae6-40aed0788d54\") " pod="openstack/ovsdbserver-nb-0" Oct 05 08:12:28 crc kubenswrapper[4846]: I1005 08:12:28.622338 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/52e7bf3e-f881-4ff2-aae6-40aed0788d54-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"52e7bf3e-f881-4ff2-aae6-40aed0788d54\") " pod="openstack/ovsdbserver-nb-0" Oct 05 08:12:28 crc kubenswrapper[4846]: I1005 08:12:28.723883 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8sphp\" (UniqueName: \"kubernetes.io/projected/fcda94d9-8d6f-4aa3-a60b-2ed20e1b3552-kube-api-access-8sphp\") pod \"ovsdbserver-nb-2\" (UID: \"fcda94d9-8d6f-4aa3-a60b-2ed20e1b3552\") " pod="openstack/ovsdbserver-nb-2" Oct 05 08:12:28 crc kubenswrapper[4846]: I1005 08:12:28.723997 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/52e7bf3e-f881-4ff2-aae6-40aed0788d54-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"52e7bf3e-f881-4ff2-aae6-40aed0788d54\") " pod="openstack/ovsdbserver-nb-0" Oct 05 08:12:28 crc kubenswrapper[4846]: I1005 08:12:28.724042 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g4mm2\" (UniqueName: \"kubernetes.io/projected/52e7bf3e-f881-4ff2-aae6-40aed0788d54-kube-api-access-g4mm2\") pod \"ovsdbserver-nb-0\" (UID: \"52e7bf3e-f881-4ff2-aae6-40aed0788d54\") " pod="openstack/ovsdbserver-nb-0" Oct 05 08:12:28 crc kubenswrapper[4846]: I1005 08:12:28.724070 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fcda94d9-8d6f-4aa3-a60b-2ed20e1b3552-config\") pod \"ovsdbserver-nb-2\" (UID: \"fcda94d9-8d6f-4aa3-a60b-2ed20e1b3552\") " pod="openstack/ovsdbserver-nb-2" Oct 05 08:12:28 crc kubenswrapper[4846]: I1005 08:12:28.724090 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/f46c4128-97fe-45d4-ac45-dc6e018294ba-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-1\" (UID: \"f46c4128-97fe-45d4-ac45-dc6e018294ba\") " pod="openstack/ovsdbserver-nb-1" Oct 05 08:12:28 crc kubenswrapper[4846]: I1005 08:12:28.724119 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/52e7bf3e-f881-4ff2-aae6-40aed0788d54-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"52e7bf3e-f881-4ff2-aae6-40aed0788d54\") " pod="openstack/ovsdbserver-nb-0" Oct 05 08:12:28 crc kubenswrapper[4846]: I1005 08:12:28.724152 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-ac493672-d90d-4cff-aa5a-f3fd0a2eab10\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ac493672-d90d-4cff-aa5a-f3fd0a2eab10\") pod \"ovsdbserver-nb-0\" (UID: \"52e7bf3e-f881-4ff2-aae6-40aed0788d54\") " pod="openstack/ovsdbserver-nb-0" Oct 05 08:12:28 crc kubenswrapper[4846]: I1005 08:12:28.724201 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/52e7bf3e-f881-4ff2-aae6-40aed0788d54-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"52e7bf3e-f881-4ff2-aae6-40aed0788d54\") " pod="openstack/ovsdbserver-nb-0" Oct 05 08:12:28 crc kubenswrapper[4846]: I1005 08:12:28.724233 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/52e7bf3e-f881-4ff2-aae6-40aed0788d54-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"52e7bf3e-f881-4ff2-aae6-40aed0788d54\") " pod="openstack/ovsdbserver-nb-0" Oct 05 08:12:28 crc kubenswrapper[4846]: I1005 08:12:28.724269 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-943eb05c-d1d6-469f-9003-179086b6659c\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-943eb05c-d1d6-469f-9003-179086b6659c\") pod \"ovsdbserver-nb-1\" (UID: \"f46c4128-97fe-45d4-ac45-dc6e018294ba\") " pod="openstack/ovsdbserver-nb-1" Oct 05 08:12:28 crc kubenswrapper[4846]: I1005 08:12:28.724301 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/fcda94d9-8d6f-4aa3-a60b-2ed20e1b3552-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-2\" (UID: \"fcda94d9-8d6f-4aa3-a60b-2ed20e1b3552\") " pod="openstack/ovsdbserver-nb-2" Oct 05 08:12:28 crc kubenswrapper[4846]: I1005 08:12:28.724333 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/f46c4128-97fe-45d4-ac45-dc6e018294ba-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-1\" (UID: \"f46c4128-97fe-45d4-ac45-dc6e018294ba\") " pod="openstack/ovsdbserver-nb-1" Oct 05 08:12:28 crc kubenswrapper[4846]: I1005 08:12:28.724360 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f46c4128-97fe-45d4-ac45-dc6e018294ba-combined-ca-bundle\") pod \"ovsdbserver-nb-1\" (UID: \"f46c4128-97fe-45d4-ac45-dc6e018294ba\") " pod="openstack/ovsdbserver-nb-1" Oct 05 08:12:28 crc kubenswrapper[4846]: I1005 08:12:28.724388 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/fcda94d9-8d6f-4aa3-a60b-2ed20e1b3552-ovsdb-rundir\") pod \"ovsdbserver-nb-2\" (UID: \"fcda94d9-8d6f-4aa3-a60b-2ed20e1b3552\") " pod="openstack/ovsdbserver-nb-2" Oct 05 08:12:28 crc kubenswrapper[4846]: I1005 08:12:28.724428 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-a2fcc757-7d1e-455c-8c8a-1e2238a9f14b\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-a2fcc757-7d1e-455c-8c8a-1e2238a9f14b\") pod \"ovsdbserver-nb-2\" (UID: \"fcda94d9-8d6f-4aa3-a60b-2ed20e1b3552\") " pod="openstack/ovsdbserver-nb-2" Oct 05 08:12:28 crc kubenswrapper[4846]: I1005 08:12:28.724493 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/52e7bf3e-f881-4ff2-aae6-40aed0788d54-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"52e7bf3e-f881-4ff2-aae6-40aed0788d54\") " pod="openstack/ovsdbserver-nb-0" Oct 05 08:12:28 crc kubenswrapper[4846]: I1005 08:12:28.724536 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l4ddt\" (UniqueName: \"kubernetes.io/projected/f46c4128-97fe-45d4-ac45-dc6e018294ba-kube-api-access-l4ddt\") pod \"ovsdbserver-nb-1\" (UID: \"f46c4128-97fe-45d4-ac45-dc6e018294ba\") " pod="openstack/ovsdbserver-nb-1" Oct 05 08:12:28 crc kubenswrapper[4846]: I1005 08:12:28.724568 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/fcda94d9-8d6f-4aa3-a60b-2ed20e1b3552-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-2\" (UID: \"fcda94d9-8d6f-4aa3-a60b-2ed20e1b3552\") " pod="openstack/ovsdbserver-nb-2" Oct 05 08:12:28 crc kubenswrapper[4846]: I1005 08:12:28.724602 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f46c4128-97fe-45d4-ac45-dc6e018294ba-scripts\") pod \"ovsdbserver-nb-1\" (UID: \"f46c4128-97fe-45d4-ac45-dc6e018294ba\") " pod="openstack/ovsdbserver-nb-1" Oct 05 08:12:28 crc kubenswrapper[4846]: I1005 08:12:28.724645 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/f46c4128-97fe-45d4-ac45-dc6e018294ba-ovsdb-rundir\") pod \"ovsdbserver-nb-1\" (UID: \"f46c4128-97fe-45d4-ac45-dc6e018294ba\") " pod="openstack/ovsdbserver-nb-1" Oct 05 08:12:28 crc kubenswrapper[4846]: I1005 08:12:28.724673 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f46c4128-97fe-45d4-ac45-dc6e018294ba-config\") pod \"ovsdbserver-nb-1\" (UID: \"f46c4128-97fe-45d4-ac45-dc6e018294ba\") " pod="openstack/ovsdbserver-nb-1" Oct 05 08:12:28 crc kubenswrapper[4846]: I1005 08:12:28.724709 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/52e7bf3e-f881-4ff2-aae6-40aed0788d54-config\") pod \"ovsdbserver-nb-0\" (UID: \"52e7bf3e-f881-4ff2-aae6-40aed0788d54\") " pod="openstack/ovsdbserver-nb-0" Oct 05 08:12:28 crc kubenswrapper[4846]: I1005 08:12:28.724747 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fcda94d9-8d6f-4aa3-a60b-2ed20e1b3552-combined-ca-bundle\") pod \"ovsdbserver-nb-2\" (UID: \"fcda94d9-8d6f-4aa3-a60b-2ed20e1b3552\") " pod="openstack/ovsdbserver-nb-2" Oct 05 08:12:28 crc kubenswrapper[4846]: I1005 08:12:28.724781 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/fcda94d9-8d6f-4aa3-a60b-2ed20e1b3552-scripts\") pod \"ovsdbserver-nb-2\" (UID: \"fcda94d9-8d6f-4aa3-a60b-2ed20e1b3552\") " pod="openstack/ovsdbserver-nb-2" Oct 05 08:12:28 crc kubenswrapper[4846]: I1005 08:12:28.724845 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/52e7bf3e-f881-4ff2-aae6-40aed0788d54-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"52e7bf3e-f881-4ff2-aae6-40aed0788d54\") " pod="openstack/ovsdbserver-nb-0" Oct 05 08:12:28 crc kubenswrapper[4846]: I1005 08:12:28.725943 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/52e7bf3e-f881-4ff2-aae6-40aed0788d54-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"52e7bf3e-f881-4ff2-aae6-40aed0788d54\") " pod="openstack/ovsdbserver-nb-0" Oct 05 08:12:28 crc kubenswrapper[4846]: I1005 08:12:28.726088 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/52e7bf3e-f881-4ff2-aae6-40aed0788d54-config\") pod \"ovsdbserver-nb-0\" (UID: \"52e7bf3e-f881-4ff2-aae6-40aed0788d54\") " pod="openstack/ovsdbserver-nb-0" Oct 05 08:12:28 crc kubenswrapper[4846]: I1005 08:12:28.727805 4846 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 05 08:12:28 crc kubenswrapper[4846]: I1005 08:12:28.727835 4846 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-ac493672-d90d-4cff-aa5a-f3fd0a2eab10\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ac493672-d90d-4cff-aa5a-f3fd0a2eab10\") pod \"ovsdbserver-nb-0\" (UID: \"52e7bf3e-f881-4ff2-aae6-40aed0788d54\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/5310f695e42fe8fa426e0f999e6ba3bf938619c010c66c09fe9415e04cef3868/globalmount\"" pod="openstack/ovsdbserver-nb-0" Oct 05 08:12:28 crc kubenswrapper[4846]: I1005 08:12:28.826267 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fcda94d9-8d6f-4aa3-a60b-2ed20e1b3552-config\") pod \"ovsdbserver-nb-2\" (UID: \"fcda94d9-8d6f-4aa3-a60b-2ed20e1b3552\") " pod="openstack/ovsdbserver-nb-2" Oct 05 08:12:28 crc kubenswrapper[4846]: I1005 08:12:28.826328 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/f46c4128-97fe-45d4-ac45-dc6e018294ba-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-1\" (UID: \"f46c4128-97fe-45d4-ac45-dc6e018294ba\") " pod="openstack/ovsdbserver-nb-1" Oct 05 08:12:28 crc kubenswrapper[4846]: I1005 08:12:28.826371 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-943eb05c-d1d6-469f-9003-179086b6659c\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-943eb05c-d1d6-469f-9003-179086b6659c\") pod \"ovsdbserver-nb-1\" (UID: \"f46c4128-97fe-45d4-ac45-dc6e018294ba\") " pod="openstack/ovsdbserver-nb-1" Oct 05 08:12:28 crc kubenswrapper[4846]: I1005 08:12:28.826390 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/fcda94d9-8d6f-4aa3-a60b-2ed20e1b3552-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-2\" (UID: \"fcda94d9-8d6f-4aa3-a60b-2ed20e1b3552\") " pod="openstack/ovsdbserver-nb-2" Oct 05 08:12:28 crc kubenswrapper[4846]: I1005 08:12:28.826406 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/f46c4128-97fe-45d4-ac45-dc6e018294ba-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-1\" (UID: \"f46c4128-97fe-45d4-ac45-dc6e018294ba\") " pod="openstack/ovsdbserver-nb-1" Oct 05 08:12:28 crc kubenswrapper[4846]: I1005 08:12:28.826421 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f46c4128-97fe-45d4-ac45-dc6e018294ba-combined-ca-bundle\") pod \"ovsdbserver-nb-1\" (UID: \"f46c4128-97fe-45d4-ac45-dc6e018294ba\") " pod="openstack/ovsdbserver-nb-1" Oct 05 08:12:28 crc kubenswrapper[4846]: I1005 08:12:28.826439 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/fcda94d9-8d6f-4aa3-a60b-2ed20e1b3552-ovsdb-rundir\") pod \"ovsdbserver-nb-2\" (UID: \"fcda94d9-8d6f-4aa3-a60b-2ed20e1b3552\") " pod="openstack/ovsdbserver-nb-2" Oct 05 08:12:28 crc kubenswrapper[4846]: I1005 08:12:28.826460 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-a2fcc757-7d1e-455c-8c8a-1e2238a9f14b\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-a2fcc757-7d1e-455c-8c8a-1e2238a9f14b\") pod \"ovsdbserver-nb-2\" (UID: \"fcda94d9-8d6f-4aa3-a60b-2ed20e1b3552\") " pod="openstack/ovsdbserver-nb-2" Oct 05 08:12:28 crc kubenswrapper[4846]: I1005 08:12:28.827794 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l4ddt\" (UniqueName: \"kubernetes.io/projected/f46c4128-97fe-45d4-ac45-dc6e018294ba-kube-api-access-l4ddt\") pod \"ovsdbserver-nb-1\" (UID: \"f46c4128-97fe-45d4-ac45-dc6e018294ba\") " pod="openstack/ovsdbserver-nb-1" Oct 05 08:12:28 crc kubenswrapper[4846]: I1005 08:12:28.827830 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/fcda94d9-8d6f-4aa3-a60b-2ed20e1b3552-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-2\" (UID: \"fcda94d9-8d6f-4aa3-a60b-2ed20e1b3552\") " pod="openstack/ovsdbserver-nb-2" Oct 05 08:12:28 crc kubenswrapper[4846]: I1005 08:12:28.827856 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f46c4128-97fe-45d4-ac45-dc6e018294ba-scripts\") pod \"ovsdbserver-nb-1\" (UID: \"f46c4128-97fe-45d4-ac45-dc6e018294ba\") " pod="openstack/ovsdbserver-nb-1" Oct 05 08:12:28 crc kubenswrapper[4846]: I1005 08:12:28.827899 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/f46c4128-97fe-45d4-ac45-dc6e018294ba-ovsdb-rundir\") pod \"ovsdbserver-nb-1\" (UID: \"f46c4128-97fe-45d4-ac45-dc6e018294ba\") " pod="openstack/ovsdbserver-nb-1" Oct 05 08:12:28 crc kubenswrapper[4846]: I1005 08:12:28.827925 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f46c4128-97fe-45d4-ac45-dc6e018294ba-config\") pod \"ovsdbserver-nb-1\" (UID: \"f46c4128-97fe-45d4-ac45-dc6e018294ba\") " pod="openstack/ovsdbserver-nb-1" Oct 05 08:12:28 crc kubenswrapper[4846]: I1005 08:12:28.827985 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/fcda94d9-8d6f-4aa3-a60b-2ed20e1b3552-ovsdb-rundir\") pod \"ovsdbserver-nb-2\" (UID: \"fcda94d9-8d6f-4aa3-a60b-2ed20e1b3552\") " pod="openstack/ovsdbserver-nb-2" Oct 05 08:12:28 crc kubenswrapper[4846]: I1005 08:12:28.828114 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fcda94d9-8d6f-4aa3-a60b-2ed20e1b3552-config\") pod \"ovsdbserver-nb-2\" (UID: \"fcda94d9-8d6f-4aa3-a60b-2ed20e1b3552\") " pod="openstack/ovsdbserver-nb-2" Oct 05 08:12:28 crc kubenswrapper[4846]: I1005 08:12:28.828369 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/f46c4128-97fe-45d4-ac45-dc6e018294ba-ovsdb-rundir\") pod \"ovsdbserver-nb-1\" (UID: \"f46c4128-97fe-45d4-ac45-dc6e018294ba\") " pod="openstack/ovsdbserver-nb-1" Oct 05 08:12:28 crc kubenswrapper[4846]: I1005 08:12:28.828418 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fcda94d9-8d6f-4aa3-a60b-2ed20e1b3552-combined-ca-bundle\") pod \"ovsdbserver-nb-2\" (UID: \"fcda94d9-8d6f-4aa3-a60b-2ed20e1b3552\") " pod="openstack/ovsdbserver-nb-2" Oct 05 08:12:28 crc kubenswrapper[4846]: I1005 08:12:28.828485 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/fcda94d9-8d6f-4aa3-a60b-2ed20e1b3552-scripts\") pod \"ovsdbserver-nb-2\" (UID: \"fcda94d9-8d6f-4aa3-a60b-2ed20e1b3552\") " pod="openstack/ovsdbserver-nb-2" Oct 05 08:12:28 crc kubenswrapper[4846]: I1005 08:12:28.828536 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8sphp\" (UniqueName: \"kubernetes.io/projected/fcda94d9-8d6f-4aa3-a60b-2ed20e1b3552-kube-api-access-8sphp\") pod \"ovsdbserver-nb-2\" (UID: \"fcda94d9-8d6f-4aa3-a60b-2ed20e1b3552\") " pod="openstack/ovsdbserver-nb-2" Oct 05 08:12:28 crc kubenswrapper[4846]: I1005 08:12:28.828947 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f46c4128-97fe-45d4-ac45-dc6e018294ba-config\") pod \"ovsdbserver-nb-1\" (UID: \"f46c4128-97fe-45d4-ac45-dc6e018294ba\") " pod="openstack/ovsdbserver-nb-1" Oct 05 08:12:28 crc kubenswrapper[4846]: I1005 08:12:28.829392 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/fcda94d9-8d6f-4aa3-a60b-2ed20e1b3552-scripts\") pod \"ovsdbserver-nb-2\" (UID: \"fcda94d9-8d6f-4aa3-a60b-2ed20e1b3552\") " pod="openstack/ovsdbserver-nb-2" Oct 05 08:12:28 crc kubenswrapper[4846]: I1005 08:12:28.829744 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f46c4128-97fe-45d4-ac45-dc6e018294ba-scripts\") pod \"ovsdbserver-nb-1\" (UID: \"f46c4128-97fe-45d4-ac45-dc6e018294ba\") " pod="openstack/ovsdbserver-nb-1" Oct 05 08:12:28 crc kubenswrapper[4846]: I1005 08:12:28.833097 4846 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 05 08:12:28 crc kubenswrapper[4846]: I1005 08:12:28.833127 4846 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-943eb05c-d1d6-469f-9003-179086b6659c\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-943eb05c-d1d6-469f-9003-179086b6659c\") pod \"ovsdbserver-nb-1\" (UID: \"f46c4128-97fe-45d4-ac45-dc6e018294ba\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/4c775798502a4c9dc3acf312a5384bf1ea99c8c9ab5d5f01aa189c90c23f1b14/globalmount\"" pod="openstack/ovsdbserver-nb-1" Oct 05 08:12:28 crc kubenswrapper[4846]: I1005 08:12:28.834347 4846 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 05 08:12:28 crc kubenswrapper[4846]: I1005 08:12:28.834607 4846 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-a2fcc757-7d1e-455c-8c8a-1e2238a9f14b\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-a2fcc757-7d1e-455c-8c8a-1e2238a9f14b\") pod \"ovsdbserver-nb-2\" (UID: \"fcda94d9-8d6f-4aa3-a60b-2ed20e1b3552\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/fe78b29c3294957cbb18e79dc278f50e386fe474e8c7220456f6f43066896d6d/globalmount\"" pod="openstack/ovsdbserver-nb-2" Oct 05 08:12:29 crc kubenswrapper[4846]: I1005 08:12:29.098063 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/52e7bf3e-f881-4ff2-aae6-40aed0788d54-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"52e7bf3e-f881-4ff2-aae6-40aed0788d54\") " pod="openstack/ovsdbserver-nb-0" Oct 05 08:12:29 crc kubenswrapper[4846]: I1005 08:12:29.098759 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/52e7bf3e-f881-4ff2-aae6-40aed0788d54-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"52e7bf3e-f881-4ff2-aae6-40aed0788d54\") " pod="openstack/ovsdbserver-nb-0" Oct 05 08:12:29 crc kubenswrapper[4846]: I1005 08:12:29.098786 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/52e7bf3e-f881-4ff2-aae6-40aed0788d54-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"52e7bf3e-f881-4ff2-aae6-40aed0788d54\") " pod="openstack/ovsdbserver-nb-0" Oct 05 08:12:29 crc kubenswrapper[4846]: I1005 08:12:29.103959 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fcda94d9-8d6f-4aa3-a60b-2ed20e1b3552-combined-ca-bundle\") pod \"ovsdbserver-nb-2\" (UID: \"fcda94d9-8d6f-4aa3-a60b-2ed20e1b3552\") " pod="openstack/ovsdbserver-nb-2" Oct 05 08:12:29 crc kubenswrapper[4846]: I1005 08:12:29.103980 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/fcda94d9-8d6f-4aa3-a60b-2ed20e1b3552-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-2\" (UID: \"fcda94d9-8d6f-4aa3-a60b-2ed20e1b3552\") " pod="openstack/ovsdbserver-nb-2" Oct 05 08:12:29 crc kubenswrapper[4846]: I1005 08:12:29.104101 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/f46c4128-97fe-45d4-ac45-dc6e018294ba-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-1\" (UID: \"f46c4128-97fe-45d4-ac45-dc6e018294ba\") " pod="openstack/ovsdbserver-nb-1" Oct 05 08:12:29 crc kubenswrapper[4846]: I1005 08:12:29.104220 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/f46c4128-97fe-45d4-ac45-dc6e018294ba-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-1\" (UID: \"f46c4128-97fe-45d4-ac45-dc6e018294ba\") " pod="openstack/ovsdbserver-nb-1" Oct 05 08:12:29 crc kubenswrapper[4846]: I1005 08:12:29.104238 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/fcda94d9-8d6f-4aa3-a60b-2ed20e1b3552-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-2\" (UID: \"fcda94d9-8d6f-4aa3-a60b-2ed20e1b3552\") " pod="openstack/ovsdbserver-nb-2" Oct 05 08:12:29 crc kubenswrapper[4846]: I1005 08:12:29.105632 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f46c4128-97fe-45d4-ac45-dc6e018294ba-combined-ca-bundle\") pod \"ovsdbserver-nb-1\" (UID: \"f46c4128-97fe-45d4-ac45-dc6e018294ba\") " pod="openstack/ovsdbserver-nb-1" Oct 05 08:12:29 crc kubenswrapper[4846]: I1005 08:12:29.106411 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8sphp\" (UniqueName: \"kubernetes.io/projected/fcda94d9-8d6f-4aa3-a60b-2ed20e1b3552-kube-api-access-8sphp\") pod \"ovsdbserver-nb-2\" (UID: \"fcda94d9-8d6f-4aa3-a60b-2ed20e1b3552\") " pod="openstack/ovsdbserver-nb-2" Oct 05 08:12:29 crc kubenswrapper[4846]: I1005 08:12:29.107101 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g4mm2\" (UniqueName: \"kubernetes.io/projected/52e7bf3e-f881-4ff2-aae6-40aed0788d54-kube-api-access-g4mm2\") pod \"ovsdbserver-nb-0\" (UID: \"52e7bf3e-f881-4ff2-aae6-40aed0788d54\") " pod="openstack/ovsdbserver-nb-0" Oct 05 08:12:29 crc kubenswrapper[4846]: I1005 08:12:29.107407 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l4ddt\" (UniqueName: \"kubernetes.io/projected/f46c4128-97fe-45d4-ac45-dc6e018294ba-kube-api-access-l4ddt\") pod \"ovsdbserver-nb-1\" (UID: \"f46c4128-97fe-45d4-ac45-dc6e018294ba\") " pod="openstack/ovsdbserver-nb-1" Oct 05 08:12:29 crc kubenswrapper[4846]: I1005 08:12:29.128982 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-943eb05c-d1d6-469f-9003-179086b6659c\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-943eb05c-d1d6-469f-9003-179086b6659c\") pod \"ovsdbserver-nb-1\" (UID: \"f46c4128-97fe-45d4-ac45-dc6e018294ba\") " pod="openstack/ovsdbserver-nb-1" Oct 05 08:12:29 crc kubenswrapper[4846]: I1005 08:12:29.134360 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-ac493672-d90d-4cff-aa5a-f3fd0a2eab10\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ac493672-d90d-4cff-aa5a-f3fd0a2eab10\") pod \"ovsdbserver-nb-0\" (UID: \"52e7bf3e-f881-4ff2-aae6-40aed0788d54\") " pod="openstack/ovsdbserver-nb-0" Oct 05 08:12:29 crc kubenswrapper[4846]: I1005 08:12:29.134810 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-a2fcc757-7d1e-455c-8c8a-1e2238a9f14b\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-a2fcc757-7d1e-455c-8c8a-1e2238a9f14b\") pod \"ovsdbserver-nb-2\" (UID: \"fcda94d9-8d6f-4aa3-a60b-2ed20e1b3552\") " pod="openstack/ovsdbserver-nb-2" Oct 05 08:12:29 crc kubenswrapper[4846]: I1005 08:12:29.150843 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Oct 05 08:12:29 crc kubenswrapper[4846]: I1005 08:12:29.190410 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-1" Oct 05 08:12:29 crc kubenswrapper[4846]: I1005 08:12:29.198897 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-2" Oct 05 08:12:29 crc kubenswrapper[4846]: I1005 08:12:29.782963 4846 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 05 08:12:29 crc kubenswrapper[4846]: I1005 08:12:29.790120 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Oct 05 08:12:29 crc kubenswrapper[4846]: I1005 08:12:29.867361 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-2"] Oct 05 08:12:29 crc kubenswrapper[4846]: W1005 08:12:29.875838 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfcda94d9_8d6f_4aa3_a60b_2ed20e1b3552.slice/crio-407dc7ba16e000a922d24e6f2954e2bee6b77490631306362f439edf908d80d7 WatchSource:0}: Error finding container 407dc7ba16e000a922d24e6f2954e2bee6b77490631306362f439edf908d80d7: Status 404 returned error can't find the container with id 407dc7ba16e000a922d24e6f2954e2bee6b77490631306362f439edf908d80d7 Oct 05 08:12:29 crc kubenswrapper[4846]: I1005 08:12:29.964581 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-0"] Oct 05 08:12:29 crc kubenswrapper[4846]: I1005 08:12:29.966276 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Oct 05 08:12:29 crc kubenswrapper[4846]: I1005 08:12:29.968273 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-config" Oct 05 08:12:29 crc kubenswrapper[4846]: I1005 08:12:29.968547 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-scripts" Oct 05 08:12:29 crc kubenswrapper[4846]: I1005 08:12:29.979666 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-sb-ovndbs" Oct 05 08:12:29 crc kubenswrapper[4846]: I1005 08:12:29.979771 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-sb-dockercfg-2dx7x" Oct 05 08:12:29 crc kubenswrapper[4846]: I1005 08:12:29.990073 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Oct 05 08:12:30 crc kubenswrapper[4846]: I1005 08:12:30.006112 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-2"] Oct 05 08:12:30 crc kubenswrapper[4846]: I1005 08:12:30.008343 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-2" Oct 05 08:12:30 crc kubenswrapper[4846]: I1005 08:12:30.013512 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-1"] Oct 05 08:12:30 crc kubenswrapper[4846]: I1005 08:12:30.014910 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-1" Oct 05 08:12:30 crc kubenswrapper[4846]: I1005 08:12:30.027825 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-2"] Oct 05 08:12:30 crc kubenswrapper[4846]: I1005 08:12:30.051885 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cfc772ff-5d34-4838-8077-aa4da607881d-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"cfc772ff-5d34-4838-8077-aa4da607881d\") " pod="openstack/ovsdbserver-sb-0" Oct 05 08:12:30 crc kubenswrapper[4846]: I1005 08:12:30.051968 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fwfzm\" (UniqueName: \"kubernetes.io/projected/cfc772ff-5d34-4838-8077-aa4da607881d-kube-api-access-fwfzm\") pod \"ovsdbserver-sb-0\" (UID: \"cfc772ff-5d34-4838-8077-aa4da607881d\") " pod="openstack/ovsdbserver-sb-0" Oct 05 08:12:30 crc kubenswrapper[4846]: I1005 08:12:30.051992 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/cfc772ff-5d34-4838-8077-aa4da607881d-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"cfc772ff-5d34-4838-8077-aa4da607881d\") " pod="openstack/ovsdbserver-sb-0" Oct 05 08:12:30 crc kubenswrapper[4846]: I1005 08:12:30.052009 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cfc772ff-5d34-4838-8077-aa4da607881d-config\") pod \"ovsdbserver-sb-0\" (UID: \"cfc772ff-5d34-4838-8077-aa4da607881d\") " pod="openstack/ovsdbserver-sb-0" Oct 05 08:12:30 crc kubenswrapper[4846]: I1005 08:12:30.052031 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/cfc772ff-5d34-4838-8077-aa4da607881d-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"cfc772ff-5d34-4838-8077-aa4da607881d\") " pod="openstack/ovsdbserver-sb-0" Oct 05 08:12:30 crc kubenswrapper[4846]: I1005 08:12:30.052051 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/cfc772ff-5d34-4838-8077-aa4da607881d-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"cfc772ff-5d34-4838-8077-aa4da607881d\") " pod="openstack/ovsdbserver-sb-0" Oct 05 08:12:30 crc kubenswrapper[4846]: I1005 08:12:30.052079 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/cfc772ff-5d34-4838-8077-aa4da607881d-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"cfc772ff-5d34-4838-8077-aa4da607881d\") " pod="openstack/ovsdbserver-sb-0" Oct 05 08:12:30 crc kubenswrapper[4846]: I1005 08:12:30.052097 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-a9e5e0c5-537e-4cf8-a15d-7f2f03321aba\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-a9e5e0c5-537e-4cf8-a15d-7f2f03321aba\") pod \"ovsdbserver-sb-0\" (UID: \"cfc772ff-5d34-4838-8077-aa4da607881d\") " pod="openstack/ovsdbserver-sb-0" Oct 05 08:12:30 crc kubenswrapper[4846]: I1005 08:12:30.057987 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-1"] Oct 05 08:12:30 crc kubenswrapper[4846]: I1005 08:12:30.154248 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6d99cbc7-c521-437e-9225-eaf5d0abdc51-combined-ca-bundle\") pod \"ovsdbserver-sb-2\" (UID: \"6d99cbc7-c521-437e-9225-eaf5d0abdc51\") " pod="openstack/ovsdbserver-sb-2" Oct 05 08:12:30 crc kubenswrapper[4846]: I1005 08:12:30.154329 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cfc772ff-5d34-4838-8077-aa4da607881d-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"cfc772ff-5d34-4838-8077-aa4da607881d\") " pod="openstack/ovsdbserver-sb-0" Oct 05 08:12:30 crc kubenswrapper[4846]: I1005 08:12:30.154390 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/cb861427-61ff-4885-bfc3-af718ddd3b7e-ovsdb-rundir\") pod \"ovsdbserver-sb-1\" (UID: \"cb861427-61ff-4885-bfc3-af718ddd3b7e\") " pod="openstack/ovsdbserver-sb-1" Oct 05 08:12:30 crc kubenswrapper[4846]: I1005 08:12:30.154414 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cb861427-61ff-4885-bfc3-af718ddd3b7e-config\") pod \"ovsdbserver-sb-1\" (UID: \"cb861427-61ff-4885-bfc3-af718ddd3b7e\") " pod="openstack/ovsdbserver-sb-1" Oct 05 08:12:30 crc kubenswrapper[4846]: I1005 08:12:30.154486 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-a72121f7-cff8-4ec2-8284-85f8523e2b7d\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-a72121f7-cff8-4ec2-8284-85f8523e2b7d\") pod \"ovsdbserver-sb-1\" (UID: \"cb861427-61ff-4885-bfc3-af718ddd3b7e\") " pod="openstack/ovsdbserver-sb-1" Oct 05 08:12:30 crc kubenswrapper[4846]: I1005 08:12:30.154525 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/cb861427-61ff-4885-bfc3-af718ddd3b7e-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-1\" (UID: \"cb861427-61ff-4885-bfc3-af718ddd3b7e\") " pod="openstack/ovsdbserver-sb-1" Oct 05 08:12:30 crc kubenswrapper[4846]: I1005 08:12:30.154556 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6d99cbc7-c521-437e-9225-eaf5d0abdc51-config\") pod \"ovsdbserver-sb-2\" (UID: \"6d99cbc7-c521-437e-9225-eaf5d0abdc51\") " pod="openstack/ovsdbserver-sb-2" Oct 05 08:12:30 crc kubenswrapper[4846]: I1005 08:12:30.154808 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fwfzm\" (UniqueName: \"kubernetes.io/projected/cfc772ff-5d34-4838-8077-aa4da607881d-kube-api-access-fwfzm\") pod \"ovsdbserver-sb-0\" (UID: \"cfc772ff-5d34-4838-8077-aa4da607881d\") " pod="openstack/ovsdbserver-sb-0" Oct 05 08:12:30 crc kubenswrapper[4846]: I1005 08:12:30.154882 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cfc772ff-5d34-4838-8077-aa4da607881d-config\") pod \"ovsdbserver-sb-0\" (UID: \"cfc772ff-5d34-4838-8077-aa4da607881d\") " pod="openstack/ovsdbserver-sb-0" Oct 05 08:12:30 crc kubenswrapper[4846]: I1005 08:12:30.154912 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/cfc772ff-5d34-4838-8077-aa4da607881d-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"cfc772ff-5d34-4838-8077-aa4da607881d\") " pod="openstack/ovsdbserver-sb-0" Oct 05 08:12:30 crc kubenswrapper[4846]: I1005 08:12:30.154959 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-9789c7fe-8396-4aa2-9d0e-dbf9f6ee6404\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-9789c7fe-8396-4aa2-9d0e-dbf9f6ee6404\") pod \"ovsdbserver-sb-2\" (UID: \"6d99cbc7-c521-437e-9225-eaf5d0abdc51\") " pod="openstack/ovsdbserver-sb-2" Oct 05 08:12:30 crc kubenswrapper[4846]: I1005 08:12:30.155007 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/cfc772ff-5d34-4838-8077-aa4da607881d-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"cfc772ff-5d34-4838-8077-aa4da607881d\") " pod="openstack/ovsdbserver-sb-0" Oct 05 08:12:30 crc kubenswrapper[4846]: I1005 08:12:30.155037 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/6d99cbc7-c521-437e-9225-eaf5d0abdc51-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-2\" (UID: \"6d99cbc7-c521-437e-9225-eaf5d0abdc51\") " pod="openstack/ovsdbserver-sb-2" Oct 05 08:12:30 crc kubenswrapper[4846]: I1005 08:12:30.155068 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5hk2r\" (UniqueName: \"kubernetes.io/projected/6d99cbc7-c521-437e-9225-eaf5d0abdc51-kube-api-access-5hk2r\") pod \"ovsdbserver-sb-2\" (UID: \"6d99cbc7-c521-437e-9225-eaf5d0abdc51\") " pod="openstack/ovsdbserver-sb-2" Oct 05 08:12:30 crc kubenswrapper[4846]: I1005 08:12:30.155100 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/6d99cbc7-c521-437e-9225-eaf5d0abdc51-ovsdb-rundir\") pod \"ovsdbserver-sb-2\" (UID: \"6d99cbc7-c521-437e-9225-eaf5d0abdc51\") " pod="openstack/ovsdbserver-sb-2" Oct 05 08:12:30 crc kubenswrapper[4846]: I1005 08:12:30.155127 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/cfc772ff-5d34-4838-8077-aa4da607881d-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"cfc772ff-5d34-4838-8077-aa4da607881d\") " pod="openstack/ovsdbserver-sb-0" Oct 05 08:12:30 crc kubenswrapper[4846]: I1005 08:12:30.155164 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cb861427-61ff-4885-bfc3-af718ddd3b7e-combined-ca-bundle\") pod \"ovsdbserver-sb-1\" (UID: \"cb861427-61ff-4885-bfc3-af718ddd3b7e\") " pod="openstack/ovsdbserver-sb-1" Oct 05 08:12:30 crc kubenswrapper[4846]: I1005 08:12:30.155634 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/cb861427-61ff-4885-bfc3-af718ddd3b7e-scripts\") pod \"ovsdbserver-sb-1\" (UID: \"cb861427-61ff-4885-bfc3-af718ddd3b7e\") " pod="openstack/ovsdbserver-sb-1" Oct 05 08:12:30 crc kubenswrapper[4846]: I1005 08:12:30.155685 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/cfc772ff-5d34-4838-8077-aa4da607881d-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"cfc772ff-5d34-4838-8077-aa4da607881d\") " pod="openstack/ovsdbserver-sb-0" Oct 05 08:12:30 crc kubenswrapper[4846]: I1005 08:12:30.155706 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/cb861427-61ff-4885-bfc3-af718ddd3b7e-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-1\" (UID: \"cb861427-61ff-4885-bfc3-af718ddd3b7e\") " pod="openstack/ovsdbserver-sb-1" Oct 05 08:12:30 crc kubenswrapper[4846]: I1005 08:12:30.155781 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-a9e5e0c5-537e-4cf8-a15d-7f2f03321aba\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-a9e5e0c5-537e-4cf8-a15d-7f2f03321aba\") pod \"ovsdbserver-sb-0\" (UID: \"cfc772ff-5d34-4838-8077-aa4da607881d\") " pod="openstack/ovsdbserver-sb-0" Oct 05 08:12:30 crc kubenswrapper[4846]: I1005 08:12:30.155830 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f2n28\" (UniqueName: \"kubernetes.io/projected/cb861427-61ff-4885-bfc3-af718ddd3b7e-kube-api-access-f2n28\") pod \"ovsdbserver-sb-1\" (UID: \"cb861427-61ff-4885-bfc3-af718ddd3b7e\") " pod="openstack/ovsdbserver-sb-1" Oct 05 08:12:30 crc kubenswrapper[4846]: I1005 08:12:30.155856 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6d99cbc7-c521-437e-9225-eaf5d0abdc51-scripts\") pod \"ovsdbserver-sb-2\" (UID: \"6d99cbc7-c521-437e-9225-eaf5d0abdc51\") " pod="openstack/ovsdbserver-sb-2" Oct 05 08:12:30 crc kubenswrapper[4846]: I1005 08:12:30.155897 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/6d99cbc7-c521-437e-9225-eaf5d0abdc51-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-2\" (UID: \"6d99cbc7-c521-437e-9225-eaf5d0abdc51\") " pod="openstack/ovsdbserver-sb-2" Oct 05 08:12:30 crc kubenswrapper[4846]: I1005 08:12:30.156847 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/cfc772ff-5d34-4838-8077-aa4da607881d-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"cfc772ff-5d34-4838-8077-aa4da607881d\") " pod="openstack/ovsdbserver-sb-0" Oct 05 08:12:30 crc kubenswrapper[4846]: I1005 08:12:30.156941 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/cfc772ff-5d34-4838-8077-aa4da607881d-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"cfc772ff-5d34-4838-8077-aa4da607881d\") " pod="openstack/ovsdbserver-sb-0" Oct 05 08:12:30 crc kubenswrapper[4846]: I1005 08:12:30.156993 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cfc772ff-5d34-4838-8077-aa4da607881d-config\") pod \"ovsdbserver-sb-0\" (UID: \"cfc772ff-5d34-4838-8077-aa4da607881d\") " pod="openstack/ovsdbserver-sb-0" Oct 05 08:12:30 crc kubenswrapper[4846]: I1005 08:12:30.160150 4846 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 05 08:12:30 crc kubenswrapper[4846]: I1005 08:12:30.160222 4846 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-a9e5e0c5-537e-4cf8-a15d-7f2f03321aba\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-a9e5e0c5-537e-4cf8-a15d-7f2f03321aba\") pod \"ovsdbserver-sb-0\" (UID: \"cfc772ff-5d34-4838-8077-aa4da607881d\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/9dd1261feaf0409470b914c6d2de90d7d91c8f33a42a2b928bb058c50e007e81/globalmount\"" pod="openstack/ovsdbserver-sb-0" Oct 05 08:12:30 crc kubenswrapper[4846]: I1005 08:12:30.161155 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/cfc772ff-5d34-4838-8077-aa4da607881d-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"cfc772ff-5d34-4838-8077-aa4da607881d\") " pod="openstack/ovsdbserver-sb-0" Oct 05 08:12:30 crc kubenswrapper[4846]: I1005 08:12:30.163612 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/cfc772ff-5d34-4838-8077-aa4da607881d-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"cfc772ff-5d34-4838-8077-aa4da607881d\") " pod="openstack/ovsdbserver-sb-0" Oct 05 08:12:30 crc kubenswrapper[4846]: I1005 08:12:30.165159 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cfc772ff-5d34-4838-8077-aa4da607881d-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"cfc772ff-5d34-4838-8077-aa4da607881d\") " pod="openstack/ovsdbserver-sb-0" Oct 05 08:12:30 crc kubenswrapper[4846]: I1005 08:12:30.172495 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fwfzm\" (UniqueName: \"kubernetes.io/projected/cfc772ff-5d34-4838-8077-aa4da607881d-kube-api-access-fwfzm\") pod \"ovsdbserver-sb-0\" (UID: \"cfc772ff-5d34-4838-8077-aa4da607881d\") " pod="openstack/ovsdbserver-sb-0" Oct 05 08:12:30 crc kubenswrapper[4846]: I1005 08:12:30.187659 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-a9e5e0c5-537e-4cf8-a15d-7f2f03321aba\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-a9e5e0c5-537e-4cf8-a15d-7f2f03321aba\") pod \"ovsdbserver-sb-0\" (UID: \"cfc772ff-5d34-4838-8077-aa4da607881d\") " pod="openstack/ovsdbserver-sb-0" Oct 05 08:12:30 crc kubenswrapper[4846]: I1005 08:12:30.257153 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-9789c7fe-8396-4aa2-9d0e-dbf9f6ee6404\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-9789c7fe-8396-4aa2-9d0e-dbf9f6ee6404\") pod \"ovsdbserver-sb-2\" (UID: \"6d99cbc7-c521-437e-9225-eaf5d0abdc51\") " pod="openstack/ovsdbserver-sb-2" Oct 05 08:12:30 crc kubenswrapper[4846]: I1005 08:12:30.257245 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/6d99cbc7-c521-437e-9225-eaf5d0abdc51-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-2\" (UID: \"6d99cbc7-c521-437e-9225-eaf5d0abdc51\") " pod="openstack/ovsdbserver-sb-2" Oct 05 08:12:30 crc kubenswrapper[4846]: I1005 08:12:30.257364 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5hk2r\" (UniqueName: \"kubernetes.io/projected/6d99cbc7-c521-437e-9225-eaf5d0abdc51-kube-api-access-5hk2r\") pod \"ovsdbserver-sb-2\" (UID: \"6d99cbc7-c521-437e-9225-eaf5d0abdc51\") " pod="openstack/ovsdbserver-sb-2" Oct 05 08:12:30 crc kubenswrapper[4846]: I1005 08:12:30.257393 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/6d99cbc7-c521-437e-9225-eaf5d0abdc51-ovsdb-rundir\") pod \"ovsdbserver-sb-2\" (UID: \"6d99cbc7-c521-437e-9225-eaf5d0abdc51\") " pod="openstack/ovsdbserver-sb-2" Oct 05 08:12:30 crc kubenswrapper[4846]: I1005 08:12:30.257940 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cb861427-61ff-4885-bfc3-af718ddd3b7e-combined-ca-bundle\") pod \"ovsdbserver-sb-1\" (UID: \"cb861427-61ff-4885-bfc3-af718ddd3b7e\") " pod="openstack/ovsdbserver-sb-1" Oct 05 08:12:30 crc kubenswrapper[4846]: I1005 08:12:30.258166 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/6d99cbc7-c521-437e-9225-eaf5d0abdc51-ovsdb-rundir\") pod \"ovsdbserver-sb-2\" (UID: \"6d99cbc7-c521-437e-9225-eaf5d0abdc51\") " pod="openstack/ovsdbserver-sb-2" Oct 05 08:12:30 crc kubenswrapper[4846]: I1005 08:12:30.259370 4846 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 05 08:12:30 crc kubenswrapper[4846]: I1005 08:12:30.259426 4846 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-9789c7fe-8396-4aa2-9d0e-dbf9f6ee6404\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-9789c7fe-8396-4aa2-9d0e-dbf9f6ee6404\") pod \"ovsdbserver-sb-2\" (UID: \"6d99cbc7-c521-437e-9225-eaf5d0abdc51\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/5abf56f7303727abbc30f5ad7c610a7c423f6005738ec064d001419536c3a5ba/globalmount\"" pod="openstack/ovsdbserver-sb-2" Oct 05 08:12:30 crc kubenswrapper[4846]: I1005 08:12:30.260020 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/cb861427-61ff-4885-bfc3-af718ddd3b7e-scripts\") pod \"ovsdbserver-sb-1\" (UID: \"cb861427-61ff-4885-bfc3-af718ddd3b7e\") " pod="openstack/ovsdbserver-sb-1" Oct 05 08:12:30 crc kubenswrapper[4846]: I1005 08:12:30.260952 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/cb861427-61ff-4885-bfc3-af718ddd3b7e-scripts\") pod \"ovsdbserver-sb-1\" (UID: \"cb861427-61ff-4885-bfc3-af718ddd3b7e\") " pod="openstack/ovsdbserver-sb-1" Oct 05 08:12:30 crc kubenswrapper[4846]: I1005 08:12:30.261382 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/cb861427-61ff-4885-bfc3-af718ddd3b7e-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-1\" (UID: \"cb861427-61ff-4885-bfc3-af718ddd3b7e\") " pod="openstack/ovsdbserver-sb-1" Oct 05 08:12:30 crc kubenswrapper[4846]: I1005 08:12:30.261454 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f2n28\" (UniqueName: \"kubernetes.io/projected/cb861427-61ff-4885-bfc3-af718ddd3b7e-kube-api-access-f2n28\") pod \"ovsdbserver-sb-1\" (UID: \"cb861427-61ff-4885-bfc3-af718ddd3b7e\") " pod="openstack/ovsdbserver-sb-1" Oct 05 08:12:30 crc kubenswrapper[4846]: I1005 08:12:30.261486 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6d99cbc7-c521-437e-9225-eaf5d0abdc51-scripts\") pod \"ovsdbserver-sb-2\" (UID: \"6d99cbc7-c521-437e-9225-eaf5d0abdc51\") " pod="openstack/ovsdbserver-sb-2" Oct 05 08:12:30 crc kubenswrapper[4846]: I1005 08:12:30.261548 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cb861427-61ff-4885-bfc3-af718ddd3b7e-combined-ca-bundle\") pod \"ovsdbserver-sb-1\" (UID: \"cb861427-61ff-4885-bfc3-af718ddd3b7e\") " pod="openstack/ovsdbserver-sb-1" Oct 05 08:12:30 crc kubenswrapper[4846]: I1005 08:12:30.261573 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/6d99cbc7-c521-437e-9225-eaf5d0abdc51-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-2\" (UID: \"6d99cbc7-c521-437e-9225-eaf5d0abdc51\") " pod="openstack/ovsdbserver-sb-2" Oct 05 08:12:30 crc kubenswrapper[4846]: I1005 08:12:30.261709 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6d99cbc7-c521-437e-9225-eaf5d0abdc51-combined-ca-bundle\") pod \"ovsdbserver-sb-2\" (UID: \"6d99cbc7-c521-437e-9225-eaf5d0abdc51\") " pod="openstack/ovsdbserver-sb-2" Oct 05 08:12:30 crc kubenswrapper[4846]: I1005 08:12:30.262921 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/cb861427-61ff-4885-bfc3-af718ddd3b7e-ovsdb-rundir\") pod \"ovsdbserver-sb-1\" (UID: \"cb861427-61ff-4885-bfc3-af718ddd3b7e\") " pod="openstack/ovsdbserver-sb-1" Oct 05 08:12:30 crc kubenswrapper[4846]: I1005 08:12:30.263299 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/6d99cbc7-c521-437e-9225-eaf5d0abdc51-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-2\" (UID: \"6d99cbc7-c521-437e-9225-eaf5d0abdc51\") " pod="openstack/ovsdbserver-sb-2" Oct 05 08:12:30 crc kubenswrapper[4846]: I1005 08:12:30.263570 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6d99cbc7-c521-437e-9225-eaf5d0abdc51-scripts\") pod \"ovsdbserver-sb-2\" (UID: \"6d99cbc7-c521-437e-9225-eaf5d0abdc51\") " pod="openstack/ovsdbserver-sb-2" Oct 05 08:12:30 crc kubenswrapper[4846]: I1005 08:12:30.265892 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/6d99cbc7-c521-437e-9225-eaf5d0abdc51-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-2\" (UID: \"6d99cbc7-c521-437e-9225-eaf5d0abdc51\") " pod="openstack/ovsdbserver-sb-2" Oct 05 08:12:30 crc kubenswrapper[4846]: I1005 08:12:30.266128 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6d99cbc7-c521-437e-9225-eaf5d0abdc51-combined-ca-bundle\") pod \"ovsdbserver-sb-2\" (UID: \"6d99cbc7-c521-437e-9225-eaf5d0abdc51\") " pod="openstack/ovsdbserver-sb-2" Oct 05 08:12:30 crc kubenswrapper[4846]: I1005 08:12:30.271311 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/cb861427-61ff-4885-bfc3-af718ddd3b7e-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-1\" (UID: \"cb861427-61ff-4885-bfc3-af718ddd3b7e\") " pod="openstack/ovsdbserver-sb-1" Oct 05 08:12:30 crc kubenswrapper[4846]: I1005 08:12:30.271385 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/cb861427-61ff-4885-bfc3-af718ddd3b7e-ovsdb-rundir\") pod \"ovsdbserver-sb-1\" (UID: \"cb861427-61ff-4885-bfc3-af718ddd3b7e\") " pod="openstack/ovsdbserver-sb-1" Oct 05 08:12:30 crc kubenswrapper[4846]: I1005 08:12:30.271455 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cb861427-61ff-4885-bfc3-af718ddd3b7e-config\") pod \"ovsdbserver-sb-1\" (UID: \"cb861427-61ff-4885-bfc3-af718ddd3b7e\") " pod="openstack/ovsdbserver-sb-1" Oct 05 08:12:30 crc kubenswrapper[4846]: I1005 08:12:30.272042 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cb861427-61ff-4885-bfc3-af718ddd3b7e-config\") pod \"ovsdbserver-sb-1\" (UID: \"cb861427-61ff-4885-bfc3-af718ddd3b7e\") " pod="openstack/ovsdbserver-sb-1" Oct 05 08:12:30 crc kubenswrapper[4846]: I1005 08:12:30.272108 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-a72121f7-cff8-4ec2-8284-85f8523e2b7d\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-a72121f7-cff8-4ec2-8284-85f8523e2b7d\") pod \"ovsdbserver-sb-1\" (UID: \"cb861427-61ff-4885-bfc3-af718ddd3b7e\") " pod="openstack/ovsdbserver-sb-1" Oct 05 08:12:30 crc kubenswrapper[4846]: I1005 08:12:30.272134 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/cb861427-61ff-4885-bfc3-af718ddd3b7e-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-1\" (UID: \"cb861427-61ff-4885-bfc3-af718ddd3b7e\") " pod="openstack/ovsdbserver-sb-1" Oct 05 08:12:30 crc kubenswrapper[4846]: I1005 08:12:30.273273 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6d99cbc7-c521-437e-9225-eaf5d0abdc51-config\") pod \"ovsdbserver-sb-2\" (UID: \"6d99cbc7-c521-437e-9225-eaf5d0abdc51\") " pod="openstack/ovsdbserver-sb-2" Oct 05 08:12:30 crc kubenswrapper[4846]: I1005 08:12:30.273928 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6d99cbc7-c521-437e-9225-eaf5d0abdc51-config\") pod \"ovsdbserver-sb-2\" (UID: \"6d99cbc7-c521-437e-9225-eaf5d0abdc51\") " pod="openstack/ovsdbserver-sb-2" Oct 05 08:12:30 crc kubenswrapper[4846]: I1005 08:12:30.274058 4846 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 05 08:12:30 crc kubenswrapper[4846]: I1005 08:12:30.274082 4846 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-a72121f7-cff8-4ec2-8284-85f8523e2b7d\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-a72121f7-cff8-4ec2-8284-85f8523e2b7d\") pod \"ovsdbserver-sb-1\" (UID: \"cb861427-61ff-4885-bfc3-af718ddd3b7e\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/4b3053078226b47ccc99e7f567bdcd1092d5e52dce22f0facb57d71906d023ab/globalmount\"" pod="openstack/ovsdbserver-sb-1" Oct 05 08:12:30 crc kubenswrapper[4846]: I1005 08:12:30.275983 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/cb861427-61ff-4885-bfc3-af718ddd3b7e-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-1\" (UID: \"cb861427-61ff-4885-bfc3-af718ddd3b7e\") " pod="openstack/ovsdbserver-sb-1" Oct 05 08:12:30 crc kubenswrapper[4846]: I1005 08:12:30.277011 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f2n28\" (UniqueName: \"kubernetes.io/projected/cb861427-61ff-4885-bfc3-af718ddd3b7e-kube-api-access-f2n28\") pod \"ovsdbserver-sb-1\" (UID: \"cb861427-61ff-4885-bfc3-af718ddd3b7e\") " pod="openstack/ovsdbserver-sb-1" Oct 05 08:12:30 crc kubenswrapper[4846]: I1005 08:12:30.278846 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5hk2r\" (UniqueName: \"kubernetes.io/projected/6d99cbc7-c521-437e-9225-eaf5d0abdc51-kube-api-access-5hk2r\") pod \"ovsdbserver-sb-2\" (UID: \"6d99cbc7-c521-437e-9225-eaf5d0abdc51\") " pod="openstack/ovsdbserver-sb-2" Oct 05 08:12:30 crc kubenswrapper[4846]: I1005 08:12:30.294241 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-9789c7fe-8396-4aa2-9d0e-dbf9f6ee6404\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-9789c7fe-8396-4aa2-9d0e-dbf9f6ee6404\") pod \"ovsdbserver-sb-2\" (UID: \"6d99cbc7-c521-437e-9225-eaf5d0abdc51\") " pod="openstack/ovsdbserver-sb-2" Oct 05 08:12:30 crc kubenswrapper[4846]: I1005 08:12:30.305686 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Oct 05 08:12:30 crc kubenswrapper[4846]: I1005 08:12:30.312469 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-a72121f7-cff8-4ec2-8284-85f8523e2b7d\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-a72121f7-cff8-4ec2-8284-85f8523e2b7d\") pod \"ovsdbserver-sb-1\" (UID: \"cb861427-61ff-4885-bfc3-af718ddd3b7e\") " pod="openstack/ovsdbserver-sb-1" Oct 05 08:12:30 crc kubenswrapper[4846]: I1005 08:12:30.352090 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-2" Oct 05 08:12:30 crc kubenswrapper[4846]: I1005 08:12:30.361230 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-1" Oct 05 08:12:30 crc kubenswrapper[4846]: I1005 08:12:30.505411 4846 scope.go:117] "RemoveContainer" containerID="b442aa9794bf57d50231ee0e39bf7081a6948efd11b46994a86d66e4e417a840" Oct 05 08:12:30 crc kubenswrapper[4846]: I1005 08:12:30.572662 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"52e7bf3e-f881-4ff2-aae6-40aed0788d54","Type":"ContainerStarted","Data":"94859a964355aebd0a6333e8990d4366bbe582b3c0c7dc535dc7b215908128c7"} Oct 05 08:12:30 crc kubenswrapper[4846]: I1005 08:12:30.574459 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-2" event={"ID":"fcda94d9-8d6f-4aa3-a60b-2ed20e1b3552","Type":"ContainerStarted","Data":"407dc7ba16e000a922d24e6f2954e2bee6b77490631306362f439edf908d80d7"} Oct 05 08:12:30 crc kubenswrapper[4846]: I1005 08:12:30.876171 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Oct 05 08:12:30 crc kubenswrapper[4846]: I1005 08:12:30.964788 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-1"] Oct 05 08:12:31 crc kubenswrapper[4846]: W1005 08:12:31.003884 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcfc772ff_5d34_4838_8077_aa4da607881d.slice/crio-19806e90464b89cfdd6c672c35bd1977f24b226741bf31d7361e088c13e39cc7 WatchSource:0}: Error finding container 19806e90464b89cfdd6c672c35bd1977f24b226741bf31d7361e088c13e39cc7: Status 404 returned error can't find the container with id 19806e90464b89cfdd6c672c35bd1977f24b226741bf31d7361e088c13e39cc7 Oct 05 08:12:31 crc kubenswrapper[4846]: W1005 08:12:31.006047 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcb861427_61ff_4885_bfc3_af718ddd3b7e.slice/crio-6f13b28be3707bdbb03719079a6bcc7c83c3ec03b7d40559755d81797eb3ed6d WatchSource:0}: Error finding container 6f13b28be3707bdbb03719079a6bcc7c83c3ec03b7d40559755d81797eb3ed6d: Status 404 returned error can't find the container with id 6f13b28be3707bdbb03719079a6bcc7c83c3ec03b7d40559755d81797eb3ed6d Oct 05 08:12:31 crc kubenswrapper[4846]: I1005 08:12:31.017286 4846 scope.go:117] "RemoveContainer" containerID="91dc787a32b9b22d8b7fcd9ee93ac0d3a459f590ee48fbb2411b26a679ca11b2" Oct 05 08:12:31 crc kubenswrapper[4846]: I1005 08:12:31.049036 4846 scope.go:117] "RemoveContainer" containerID="8876aec95a893518e103665f2a4586f7d40b487a9668dbdb64a9fe128886b884" Oct 05 08:12:31 crc kubenswrapper[4846]: I1005 08:12:31.497549 4846 scope.go:117] "RemoveContainer" containerID="8fa72be735dff8e623a4a4958f452896ef79e72dd44c2d714a3019187462de47" Oct 05 08:12:31 crc kubenswrapper[4846]: E1005 08:12:31.497888 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 08:12:31 crc kubenswrapper[4846]: I1005 08:12:31.549964 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-2"] Oct 05 08:12:31 crc kubenswrapper[4846]: W1005 08:12:31.559550 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6d99cbc7_c521_437e_9225_eaf5d0abdc51.slice/crio-50019d7bd90efbfdd36a975119f9075d0884d9397cb7026400d1276d05de91cc WatchSource:0}: Error finding container 50019d7bd90efbfdd36a975119f9075d0884d9397cb7026400d1276d05de91cc: Status 404 returned error can't find the container with id 50019d7bd90efbfdd36a975119f9075d0884d9397cb7026400d1276d05de91cc Oct 05 08:12:31 crc kubenswrapper[4846]: I1005 08:12:31.591426 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-2" event={"ID":"6d99cbc7-c521-437e-9225-eaf5d0abdc51","Type":"ContainerStarted","Data":"50019d7bd90efbfdd36a975119f9075d0884d9397cb7026400d1276d05de91cc"} Oct 05 08:12:31 crc kubenswrapper[4846]: I1005 08:12:31.593196 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"cfc772ff-5d34-4838-8077-aa4da607881d","Type":"ContainerStarted","Data":"19806e90464b89cfdd6c672c35bd1977f24b226741bf31d7361e088c13e39cc7"} Oct 05 08:12:31 crc kubenswrapper[4846]: I1005 08:12:31.596297 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-1" event={"ID":"cb861427-61ff-4885-bfc3-af718ddd3b7e","Type":"ContainerStarted","Data":"6f13b28be3707bdbb03719079a6bcc7c83c3ec03b7d40559755d81797eb3ed6d"} Oct 05 08:12:32 crc kubenswrapper[4846]: I1005 08:12:32.016375 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-1"] Oct 05 08:12:33 crc kubenswrapper[4846]: W1005 08:12:33.275521 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf46c4128_97fe_45d4_ac45_dc6e018294ba.slice/crio-b47109379c174078ee58b0dca6976f769d33605ec7ea90130246e630c7dc903d WatchSource:0}: Error finding container b47109379c174078ee58b0dca6976f769d33605ec7ea90130246e630c7dc903d: Status 404 returned error can't find the container with id b47109379c174078ee58b0dca6976f769d33605ec7ea90130246e630c7dc903d Oct 05 08:12:33 crc kubenswrapper[4846]: I1005 08:12:33.617193 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-1" event={"ID":"f46c4128-97fe-45d4-ac45-dc6e018294ba","Type":"ContainerStarted","Data":"b47109379c174078ee58b0dca6976f769d33605ec7ea90130246e630c7dc903d"} Oct 05 08:12:34 crc kubenswrapper[4846]: I1005 08:12:34.625283 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"cfc772ff-5d34-4838-8077-aa4da607881d","Type":"ContainerStarted","Data":"f77fe82893d3f41211d973808aa2326f89185b3193791114d7ebdd2e59521593"} Oct 05 08:12:34 crc kubenswrapper[4846]: I1005 08:12:34.625657 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"cfc772ff-5d34-4838-8077-aa4da607881d","Type":"ContainerStarted","Data":"8d91c05a4094ac6054eec3fddf7424df7fc5949b8439928566fd988e4d91a020"} Oct 05 08:12:34 crc kubenswrapper[4846]: I1005 08:12:34.628956 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-1" event={"ID":"cb861427-61ff-4885-bfc3-af718ddd3b7e","Type":"ContainerStarted","Data":"90e3b8d9c44c98810d2fb28a52de1fae935fb6899da75f60559dad5d28a94800"} Oct 05 08:12:34 crc kubenswrapper[4846]: I1005 08:12:34.628981 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-1" event={"ID":"cb861427-61ff-4885-bfc3-af718ddd3b7e","Type":"ContainerStarted","Data":"6afb2e08ebd83fb7469fc5a8a16fa8ed3f4b973689b8620315f48dd0cb6b67f5"} Oct 05 08:12:34 crc kubenswrapper[4846]: I1005 08:12:34.631353 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"52e7bf3e-f881-4ff2-aae6-40aed0788d54","Type":"ContainerStarted","Data":"52241702ef4153555c98962b5aa43df0aecfa4a084b21897b19b79c95830ec85"} Oct 05 08:12:34 crc kubenswrapper[4846]: I1005 08:12:34.631402 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"52e7bf3e-f881-4ff2-aae6-40aed0788d54","Type":"ContainerStarted","Data":"0c575614f912347ae113d38c99e9e4b8049bb43f21f2dafbde0f72ee56cc8f62"} Oct 05 08:12:34 crc kubenswrapper[4846]: I1005 08:12:34.633984 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-2" event={"ID":"6d99cbc7-c521-437e-9225-eaf5d0abdc51","Type":"ContainerStarted","Data":"27d2894de111d8a44e76614030e9614ab19af6deef685e81664c5dba092c319c"} Oct 05 08:12:34 crc kubenswrapper[4846]: I1005 08:12:34.634029 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-2" event={"ID":"6d99cbc7-c521-437e-9225-eaf5d0abdc51","Type":"ContainerStarted","Data":"5f1e72653bf2172ab6c7377d8597e4436d406a6daeaddf0c7fc163defe32de93"} Oct 05 08:12:34 crc kubenswrapper[4846]: I1005 08:12:34.636042 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-2" event={"ID":"fcda94d9-8d6f-4aa3-a60b-2ed20e1b3552","Type":"ContainerStarted","Data":"29d8aa1fa7d6b8325267f44cbf2a66042166b3ec70f67ff2eb17090b9c261b74"} Oct 05 08:12:34 crc kubenswrapper[4846]: I1005 08:12:34.636074 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-2" event={"ID":"fcda94d9-8d6f-4aa3-a60b-2ed20e1b3552","Type":"ContainerStarted","Data":"dff1c0668c959c77a997acdc0209077795d8539d4cfe8f572e8157ce31cab8d9"} Oct 05 08:12:34 crc kubenswrapper[4846]: I1005 08:12:34.638355 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-1" event={"ID":"f46c4128-97fe-45d4-ac45-dc6e018294ba","Type":"ContainerStarted","Data":"1850bfcfabca5235f1bfbfd71cf7dc3da2ec6d44d52062965fb4658d5f221dde"} Oct 05 08:12:34 crc kubenswrapper[4846]: I1005 08:12:34.638379 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-1" event={"ID":"f46c4128-97fe-45d4-ac45-dc6e018294ba","Type":"ContainerStarted","Data":"525235659e04c8177e51a8b243214e2dcc3aa762a526ab0801d1efeee6452ec1"} Oct 05 08:12:34 crc kubenswrapper[4846]: I1005 08:12:34.653628 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-0" podStartSLOduration=3.8485477599999998 podStartE2EDuration="6.653611652s" podCreationTimestamp="2025-10-05 08:12:28 +0000 UTC" firstStartedPulling="2025-10-05 08:12:31.017403645 +0000 UTC m=+5073.258256420" lastFinishedPulling="2025-10-05 08:12:33.822467497 +0000 UTC m=+5076.063320312" observedRunningTime="2025-10-05 08:12:34.652686177 +0000 UTC m=+5076.893539032" watchObservedRunningTime="2025-10-05 08:12:34.653611652 +0000 UTC m=+5076.894464427" Oct 05 08:12:34 crc kubenswrapper[4846]: I1005 08:12:34.677616 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-0" podStartSLOduration=3.5898478259999997 podStartE2EDuration="7.677601944s" podCreationTimestamp="2025-10-05 08:12:27 +0000 UTC" firstStartedPulling="2025-10-05 08:12:29.782647523 +0000 UTC m=+5072.023500318" lastFinishedPulling="2025-10-05 08:12:33.870401651 +0000 UTC m=+5076.111254436" observedRunningTime="2025-10-05 08:12:34.674941513 +0000 UTC m=+5076.915794288" watchObservedRunningTime="2025-10-05 08:12:34.677601944 +0000 UTC m=+5076.918454709" Oct 05 08:12:34 crc kubenswrapper[4846]: I1005 08:12:34.695111 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-1" podStartSLOduration=3.842615251 podStartE2EDuration="6.695090142s" podCreationTimestamp="2025-10-05 08:12:28 +0000 UTC" firstStartedPulling="2025-10-05 08:12:31.017828817 +0000 UTC m=+5073.258681592" lastFinishedPulling="2025-10-05 08:12:33.870303708 +0000 UTC m=+5076.111156483" observedRunningTime="2025-10-05 08:12:34.691546768 +0000 UTC m=+5076.932399543" watchObservedRunningTime="2025-10-05 08:12:34.695090142 +0000 UTC m=+5076.935942917" Oct 05 08:12:34 crc kubenswrapper[4846]: I1005 08:12:34.714267 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-1" podStartSLOduration=7.087440511 podStartE2EDuration="7.714245925s" podCreationTimestamp="2025-10-05 08:12:27 +0000 UTC" firstStartedPulling="2025-10-05 08:12:33.277966787 +0000 UTC m=+5075.518819562" lastFinishedPulling="2025-10-05 08:12:33.904772201 +0000 UTC m=+5076.145624976" observedRunningTime="2025-10-05 08:12:34.709824207 +0000 UTC m=+5076.950677002" watchObservedRunningTime="2025-10-05 08:12:34.714245925 +0000 UTC m=+5076.955098690" Oct 05 08:12:34 crc kubenswrapper[4846]: I1005 08:12:34.733406 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-2" podStartSLOduration=4.288124841 podStartE2EDuration="6.733387488s" podCreationTimestamp="2025-10-05 08:12:28 +0000 UTC" firstStartedPulling="2025-10-05 08:12:31.561170976 +0000 UTC m=+5073.802023741" lastFinishedPulling="2025-10-05 08:12:34.006433613 +0000 UTC m=+5076.247286388" observedRunningTime="2025-10-05 08:12:34.727143181 +0000 UTC m=+5076.967995976" watchObservedRunningTime="2025-10-05 08:12:34.733387488 +0000 UTC m=+5076.974240263" Oct 05 08:12:34 crc kubenswrapper[4846]: I1005 08:12:34.760405 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-2" podStartSLOduration=3.815986032 podStartE2EDuration="7.760379341s" podCreationTimestamp="2025-10-05 08:12:27 +0000 UTC" firstStartedPulling="2025-10-05 08:12:29.878030467 +0000 UTC m=+5072.118883282" lastFinishedPulling="2025-10-05 08:12:33.822423816 +0000 UTC m=+5076.063276591" observedRunningTime="2025-10-05 08:12:34.754268027 +0000 UTC m=+5076.995120822" watchObservedRunningTime="2025-10-05 08:12:34.760379341 +0000 UTC m=+5077.001232106" Oct 05 08:12:35 crc kubenswrapper[4846]: I1005 08:12:35.151326 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-0" Oct 05 08:12:35 crc kubenswrapper[4846]: I1005 08:12:35.191068 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-1" Oct 05 08:12:35 crc kubenswrapper[4846]: I1005 08:12:35.199719 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-2" Oct 05 08:12:35 crc kubenswrapper[4846]: I1005 08:12:35.308909 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-0" Oct 05 08:12:35 crc kubenswrapper[4846]: I1005 08:12:35.352318 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-2" Oct 05 08:12:35 crc kubenswrapper[4846]: I1005 08:12:35.361932 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-1" Oct 05 08:12:36 crc kubenswrapper[4846]: I1005 08:12:36.307657 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-0" Oct 05 08:12:36 crc kubenswrapper[4846]: I1005 08:12:36.352768 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-2" Oct 05 08:12:36 crc kubenswrapper[4846]: I1005 08:12:36.362076 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-1" Oct 05 08:12:38 crc kubenswrapper[4846]: I1005 08:12:38.203336 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-0" Oct 05 08:12:38 crc kubenswrapper[4846]: I1005 08:12:38.204579 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-0" Oct 05 08:12:38 crc kubenswrapper[4846]: I1005 08:12:38.257073 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-2" Oct 05 08:12:38 crc kubenswrapper[4846]: I1005 08:12:38.257705 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-2" Oct 05 08:12:38 crc kubenswrapper[4846]: I1005 08:12:38.273465 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-1" Oct 05 08:12:38 crc kubenswrapper[4846]: I1005 08:12:38.274321 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-1" Oct 05 08:12:39 crc kubenswrapper[4846]: I1005 08:12:39.219993 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-0" Oct 05 08:12:39 crc kubenswrapper[4846]: I1005 08:12:39.274774 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-2" Oct 05 08:12:39 crc kubenswrapper[4846]: I1005 08:12:39.274931 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-1" Oct 05 08:12:39 crc kubenswrapper[4846]: I1005 08:12:39.359619 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-0" Oct 05 08:12:39 crc kubenswrapper[4846]: I1005 08:12:39.465922 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-2" Oct 05 08:12:39 crc kubenswrapper[4846]: I1005 08:12:39.553968 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-2" Oct 05 08:12:39 crc kubenswrapper[4846]: I1005 08:12:39.569300 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-1" Oct 05 08:12:39 crc kubenswrapper[4846]: I1005 08:12:39.576923 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-0" Oct 05 08:12:39 crc kubenswrapper[4846]: I1005 08:12:39.608948 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-9b45cbc8f-btndm"] Oct 05 08:12:39 crc kubenswrapper[4846]: I1005 08:12:39.611501 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-1" Oct 05 08:12:39 crc kubenswrapper[4846]: I1005 08:12:39.611527 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-9b45cbc8f-btndm" Oct 05 08:12:39 crc kubenswrapper[4846]: I1005 08:12:39.613396 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-nb" Oct 05 08:12:39 crc kubenswrapper[4846]: I1005 08:12:39.629269 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-9b45cbc8f-btndm"] Oct 05 08:12:39 crc kubenswrapper[4846]: I1005 08:12:39.647157 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/98829bda-56bb-4ef9-8c95-3c8fd7d57bc2-dns-svc\") pod \"dnsmasq-dns-9b45cbc8f-btndm\" (UID: \"98829bda-56bb-4ef9-8c95-3c8fd7d57bc2\") " pod="openstack/dnsmasq-dns-9b45cbc8f-btndm" Oct 05 08:12:39 crc kubenswrapper[4846]: I1005 08:12:39.647239 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/98829bda-56bb-4ef9-8c95-3c8fd7d57bc2-config\") pod \"dnsmasq-dns-9b45cbc8f-btndm\" (UID: \"98829bda-56bb-4ef9-8c95-3c8fd7d57bc2\") " pod="openstack/dnsmasq-dns-9b45cbc8f-btndm" Oct 05 08:12:39 crc kubenswrapper[4846]: I1005 08:12:39.647800 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/98829bda-56bb-4ef9-8c95-3c8fd7d57bc2-ovsdbserver-nb\") pod \"dnsmasq-dns-9b45cbc8f-btndm\" (UID: \"98829bda-56bb-4ef9-8c95-3c8fd7d57bc2\") " pod="openstack/dnsmasq-dns-9b45cbc8f-btndm" Oct 05 08:12:39 crc kubenswrapper[4846]: I1005 08:12:39.647927 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rzgbd\" (UniqueName: \"kubernetes.io/projected/98829bda-56bb-4ef9-8c95-3c8fd7d57bc2-kube-api-access-rzgbd\") pod \"dnsmasq-dns-9b45cbc8f-btndm\" (UID: \"98829bda-56bb-4ef9-8c95-3c8fd7d57bc2\") " pod="openstack/dnsmasq-dns-9b45cbc8f-btndm" Oct 05 08:12:39 crc kubenswrapper[4846]: I1005 08:12:39.749973 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/98829bda-56bb-4ef9-8c95-3c8fd7d57bc2-ovsdbserver-nb\") pod \"dnsmasq-dns-9b45cbc8f-btndm\" (UID: \"98829bda-56bb-4ef9-8c95-3c8fd7d57bc2\") " pod="openstack/dnsmasq-dns-9b45cbc8f-btndm" Oct 05 08:12:39 crc kubenswrapper[4846]: I1005 08:12:39.750233 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rzgbd\" (UniqueName: \"kubernetes.io/projected/98829bda-56bb-4ef9-8c95-3c8fd7d57bc2-kube-api-access-rzgbd\") pod \"dnsmasq-dns-9b45cbc8f-btndm\" (UID: \"98829bda-56bb-4ef9-8c95-3c8fd7d57bc2\") " pod="openstack/dnsmasq-dns-9b45cbc8f-btndm" Oct 05 08:12:39 crc kubenswrapper[4846]: I1005 08:12:39.750363 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/98829bda-56bb-4ef9-8c95-3c8fd7d57bc2-dns-svc\") pod \"dnsmasq-dns-9b45cbc8f-btndm\" (UID: \"98829bda-56bb-4ef9-8c95-3c8fd7d57bc2\") " pod="openstack/dnsmasq-dns-9b45cbc8f-btndm" Oct 05 08:12:39 crc kubenswrapper[4846]: I1005 08:12:39.750597 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/98829bda-56bb-4ef9-8c95-3c8fd7d57bc2-config\") pod \"dnsmasq-dns-9b45cbc8f-btndm\" (UID: \"98829bda-56bb-4ef9-8c95-3c8fd7d57bc2\") " pod="openstack/dnsmasq-dns-9b45cbc8f-btndm" Oct 05 08:12:39 crc kubenswrapper[4846]: I1005 08:12:39.750891 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/98829bda-56bb-4ef9-8c95-3c8fd7d57bc2-ovsdbserver-nb\") pod \"dnsmasq-dns-9b45cbc8f-btndm\" (UID: \"98829bda-56bb-4ef9-8c95-3c8fd7d57bc2\") " pod="openstack/dnsmasq-dns-9b45cbc8f-btndm" Oct 05 08:12:39 crc kubenswrapper[4846]: I1005 08:12:39.751257 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/98829bda-56bb-4ef9-8c95-3c8fd7d57bc2-config\") pod \"dnsmasq-dns-9b45cbc8f-btndm\" (UID: \"98829bda-56bb-4ef9-8c95-3c8fd7d57bc2\") " pod="openstack/dnsmasq-dns-9b45cbc8f-btndm" Oct 05 08:12:39 crc kubenswrapper[4846]: I1005 08:12:39.751593 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/98829bda-56bb-4ef9-8c95-3c8fd7d57bc2-dns-svc\") pod \"dnsmasq-dns-9b45cbc8f-btndm\" (UID: \"98829bda-56bb-4ef9-8c95-3c8fd7d57bc2\") " pod="openstack/dnsmasq-dns-9b45cbc8f-btndm" Oct 05 08:12:39 crc kubenswrapper[4846]: I1005 08:12:39.780491 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rzgbd\" (UniqueName: \"kubernetes.io/projected/98829bda-56bb-4ef9-8c95-3c8fd7d57bc2-kube-api-access-rzgbd\") pod \"dnsmasq-dns-9b45cbc8f-btndm\" (UID: \"98829bda-56bb-4ef9-8c95-3c8fd7d57bc2\") " pod="openstack/dnsmasq-dns-9b45cbc8f-btndm" Oct 05 08:12:39 crc kubenswrapper[4846]: I1005 08:12:39.945948 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-9b45cbc8f-btndm" Oct 05 08:12:40 crc kubenswrapper[4846]: I1005 08:12:40.018694 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-9b45cbc8f-btndm"] Oct 05 08:12:40 crc kubenswrapper[4846]: I1005 08:12:40.033892 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-697c78d8d5-2f9zd"] Oct 05 08:12:40 crc kubenswrapper[4846]: I1005 08:12:40.046653 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-697c78d8d5-2f9zd"] Oct 05 08:12:40 crc kubenswrapper[4846]: I1005 08:12:40.046759 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-697c78d8d5-2f9zd" Oct 05 08:12:40 crc kubenswrapper[4846]: I1005 08:12:40.050999 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-sb" Oct 05 08:12:40 crc kubenswrapper[4846]: I1005 08:12:40.156831 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0a9ba111-4d92-4903-8c97-f386e3c29d5c-config\") pod \"dnsmasq-dns-697c78d8d5-2f9zd\" (UID: \"0a9ba111-4d92-4903-8c97-f386e3c29d5c\") " pod="openstack/dnsmasq-dns-697c78d8d5-2f9zd" Oct 05 08:12:40 crc kubenswrapper[4846]: I1005 08:12:40.156875 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0a9ba111-4d92-4903-8c97-f386e3c29d5c-ovsdbserver-nb\") pod \"dnsmasq-dns-697c78d8d5-2f9zd\" (UID: \"0a9ba111-4d92-4903-8c97-f386e3c29d5c\") " pod="openstack/dnsmasq-dns-697c78d8d5-2f9zd" Oct 05 08:12:40 crc kubenswrapper[4846]: I1005 08:12:40.156907 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zfcc2\" (UniqueName: \"kubernetes.io/projected/0a9ba111-4d92-4903-8c97-f386e3c29d5c-kube-api-access-zfcc2\") pod \"dnsmasq-dns-697c78d8d5-2f9zd\" (UID: \"0a9ba111-4d92-4903-8c97-f386e3c29d5c\") " pod="openstack/dnsmasq-dns-697c78d8d5-2f9zd" Oct 05 08:12:40 crc kubenswrapper[4846]: I1005 08:12:40.156973 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0a9ba111-4d92-4903-8c97-f386e3c29d5c-dns-svc\") pod \"dnsmasq-dns-697c78d8d5-2f9zd\" (UID: \"0a9ba111-4d92-4903-8c97-f386e3c29d5c\") " pod="openstack/dnsmasq-dns-697c78d8d5-2f9zd" Oct 05 08:12:40 crc kubenswrapper[4846]: I1005 08:12:40.157303 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0a9ba111-4d92-4903-8c97-f386e3c29d5c-ovsdbserver-sb\") pod \"dnsmasq-dns-697c78d8d5-2f9zd\" (UID: \"0a9ba111-4d92-4903-8c97-f386e3c29d5c\") " pod="openstack/dnsmasq-dns-697c78d8d5-2f9zd" Oct 05 08:12:40 crc kubenswrapper[4846]: I1005 08:12:40.259081 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0a9ba111-4d92-4903-8c97-f386e3c29d5c-ovsdbserver-sb\") pod \"dnsmasq-dns-697c78d8d5-2f9zd\" (UID: \"0a9ba111-4d92-4903-8c97-f386e3c29d5c\") " pod="openstack/dnsmasq-dns-697c78d8d5-2f9zd" Oct 05 08:12:40 crc kubenswrapper[4846]: I1005 08:12:40.259136 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0a9ba111-4d92-4903-8c97-f386e3c29d5c-config\") pod \"dnsmasq-dns-697c78d8d5-2f9zd\" (UID: \"0a9ba111-4d92-4903-8c97-f386e3c29d5c\") " pod="openstack/dnsmasq-dns-697c78d8d5-2f9zd" Oct 05 08:12:40 crc kubenswrapper[4846]: I1005 08:12:40.259157 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0a9ba111-4d92-4903-8c97-f386e3c29d5c-ovsdbserver-nb\") pod \"dnsmasq-dns-697c78d8d5-2f9zd\" (UID: \"0a9ba111-4d92-4903-8c97-f386e3c29d5c\") " pod="openstack/dnsmasq-dns-697c78d8d5-2f9zd" Oct 05 08:12:40 crc kubenswrapper[4846]: I1005 08:12:40.259187 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zfcc2\" (UniqueName: \"kubernetes.io/projected/0a9ba111-4d92-4903-8c97-f386e3c29d5c-kube-api-access-zfcc2\") pod \"dnsmasq-dns-697c78d8d5-2f9zd\" (UID: \"0a9ba111-4d92-4903-8c97-f386e3c29d5c\") " pod="openstack/dnsmasq-dns-697c78d8d5-2f9zd" Oct 05 08:12:40 crc kubenswrapper[4846]: I1005 08:12:40.259231 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0a9ba111-4d92-4903-8c97-f386e3c29d5c-dns-svc\") pod \"dnsmasq-dns-697c78d8d5-2f9zd\" (UID: \"0a9ba111-4d92-4903-8c97-f386e3c29d5c\") " pod="openstack/dnsmasq-dns-697c78d8d5-2f9zd" Oct 05 08:12:40 crc kubenswrapper[4846]: I1005 08:12:40.260247 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0a9ba111-4d92-4903-8c97-f386e3c29d5c-dns-svc\") pod \"dnsmasq-dns-697c78d8d5-2f9zd\" (UID: \"0a9ba111-4d92-4903-8c97-f386e3c29d5c\") " pod="openstack/dnsmasq-dns-697c78d8d5-2f9zd" Oct 05 08:12:40 crc kubenswrapper[4846]: I1005 08:12:40.260282 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0a9ba111-4d92-4903-8c97-f386e3c29d5c-ovsdbserver-nb\") pod \"dnsmasq-dns-697c78d8d5-2f9zd\" (UID: \"0a9ba111-4d92-4903-8c97-f386e3c29d5c\") " pod="openstack/dnsmasq-dns-697c78d8d5-2f9zd" Oct 05 08:12:40 crc kubenswrapper[4846]: I1005 08:12:40.260485 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0a9ba111-4d92-4903-8c97-f386e3c29d5c-config\") pod \"dnsmasq-dns-697c78d8d5-2f9zd\" (UID: \"0a9ba111-4d92-4903-8c97-f386e3c29d5c\") " pod="openstack/dnsmasq-dns-697c78d8d5-2f9zd" Oct 05 08:12:40 crc kubenswrapper[4846]: I1005 08:12:40.260638 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0a9ba111-4d92-4903-8c97-f386e3c29d5c-ovsdbserver-sb\") pod \"dnsmasq-dns-697c78d8d5-2f9zd\" (UID: \"0a9ba111-4d92-4903-8c97-f386e3c29d5c\") " pod="openstack/dnsmasq-dns-697c78d8d5-2f9zd" Oct 05 08:12:40 crc kubenswrapper[4846]: I1005 08:12:40.422730 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-9b45cbc8f-btndm"] Oct 05 08:12:40 crc kubenswrapper[4846]: I1005 08:12:40.697555 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zfcc2\" (UniqueName: \"kubernetes.io/projected/0a9ba111-4d92-4903-8c97-f386e3c29d5c-kube-api-access-zfcc2\") pod \"dnsmasq-dns-697c78d8d5-2f9zd\" (UID: \"0a9ba111-4d92-4903-8c97-f386e3c29d5c\") " pod="openstack/dnsmasq-dns-697c78d8d5-2f9zd" Oct 05 08:12:40 crc kubenswrapper[4846]: I1005 08:12:40.969254 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-697c78d8d5-2f9zd" Oct 05 08:12:41 crc kubenswrapper[4846]: I1005 08:12:41.214767 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-697c78d8d5-2f9zd"] Oct 05 08:12:41 crc kubenswrapper[4846]: I1005 08:12:41.712071 4846 generic.go:334] "Generic (PLEG): container finished" podID="98829bda-56bb-4ef9-8c95-3c8fd7d57bc2" containerID="3561e6e55e5652d5c8f1f9f3b810e3b7723ed3da4f888c2264301b4b507862eb" exitCode=0 Oct 05 08:12:41 crc kubenswrapper[4846]: I1005 08:12:41.712264 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-9b45cbc8f-btndm" event={"ID":"98829bda-56bb-4ef9-8c95-3c8fd7d57bc2","Type":"ContainerDied","Data":"3561e6e55e5652d5c8f1f9f3b810e3b7723ed3da4f888c2264301b4b507862eb"} Oct 05 08:12:41 crc kubenswrapper[4846]: I1005 08:12:41.714540 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-9b45cbc8f-btndm" event={"ID":"98829bda-56bb-4ef9-8c95-3c8fd7d57bc2","Type":"ContainerStarted","Data":"b11c4b05f04c9e0e50ff215dd2e4d4e638ac98e35b77339d6943ce5805881b7d"} Oct 05 08:12:41 crc kubenswrapper[4846]: I1005 08:12:41.719732 4846 generic.go:334] "Generic (PLEG): container finished" podID="0a9ba111-4d92-4903-8c97-f386e3c29d5c" containerID="8b575118c9b34bd243118cf9a593a887bd7ab1577a66cd225b525d2ff127ac68" exitCode=0 Oct 05 08:12:41 crc kubenswrapper[4846]: I1005 08:12:41.719819 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-697c78d8d5-2f9zd" event={"ID":"0a9ba111-4d92-4903-8c97-f386e3c29d5c","Type":"ContainerDied","Data":"8b575118c9b34bd243118cf9a593a887bd7ab1577a66cd225b525d2ff127ac68"} Oct 05 08:12:41 crc kubenswrapper[4846]: I1005 08:12:41.719854 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-697c78d8d5-2f9zd" event={"ID":"0a9ba111-4d92-4903-8c97-f386e3c29d5c","Type":"ContainerStarted","Data":"b07264268023104a93afeaff29fdf3393f87dbf3c85b0e9124f603946f0eb3ac"} Oct 05 08:12:42 crc kubenswrapper[4846]: I1005 08:12:42.033779 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-9b45cbc8f-btndm" Oct 05 08:12:42 crc kubenswrapper[4846]: I1005 08:12:42.192307 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/98829bda-56bb-4ef9-8c95-3c8fd7d57bc2-config\") pod \"98829bda-56bb-4ef9-8c95-3c8fd7d57bc2\" (UID: \"98829bda-56bb-4ef9-8c95-3c8fd7d57bc2\") " Oct 05 08:12:42 crc kubenswrapper[4846]: I1005 08:12:42.192576 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rzgbd\" (UniqueName: \"kubernetes.io/projected/98829bda-56bb-4ef9-8c95-3c8fd7d57bc2-kube-api-access-rzgbd\") pod \"98829bda-56bb-4ef9-8c95-3c8fd7d57bc2\" (UID: \"98829bda-56bb-4ef9-8c95-3c8fd7d57bc2\") " Oct 05 08:12:42 crc kubenswrapper[4846]: I1005 08:12:42.192630 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/98829bda-56bb-4ef9-8c95-3c8fd7d57bc2-dns-svc\") pod \"98829bda-56bb-4ef9-8c95-3c8fd7d57bc2\" (UID: \"98829bda-56bb-4ef9-8c95-3c8fd7d57bc2\") " Oct 05 08:12:42 crc kubenswrapper[4846]: I1005 08:12:42.192776 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/98829bda-56bb-4ef9-8c95-3c8fd7d57bc2-ovsdbserver-nb\") pod \"98829bda-56bb-4ef9-8c95-3c8fd7d57bc2\" (UID: \"98829bda-56bb-4ef9-8c95-3c8fd7d57bc2\") " Oct 05 08:12:42 crc kubenswrapper[4846]: I1005 08:12:42.497021 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/98829bda-56bb-4ef9-8c95-3c8fd7d57bc2-kube-api-access-rzgbd" (OuterVolumeSpecName: "kube-api-access-rzgbd") pod "98829bda-56bb-4ef9-8c95-3c8fd7d57bc2" (UID: "98829bda-56bb-4ef9-8c95-3c8fd7d57bc2"). InnerVolumeSpecName "kube-api-access-rzgbd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:12:42 crc kubenswrapper[4846]: I1005 08:12:42.501105 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rzgbd\" (UniqueName: \"kubernetes.io/projected/98829bda-56bb-4ef9-8c95-3c8fd7d57bc2-kube-api-access-rzgbd\") on node \"crc\" DevicePath \"\"" Oct 05 08:12:42 crc kubenswrapper[4846]: I1005 08:12:42.512699 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/98829bda-56bb-4ef9-8c95-3c8fd7d57bc2-config" (OuterVolumeSpecName: "config") pod "98829bda-56bb-4ef9-8c95-3c8fd7d57bc2" (UID: "98829bda-56bb-4ef9-8c95-3c8fd7d57bc2"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 08:12:42 crc kubenswrapper[4846]: I1005 08:12:42.527039 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/98829bda-56bb-4ef9-8c95-3c8fd7d57bc2-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "98829bda-56bb-4ef9-8c95-3c8fd7d57bc2" (UID: "98829bda-56bb-4ef9-8c95-3c8fd7d57bc2"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 08:12:42 crc kubenswrapper[4846]: I1005 08:12:42.529842 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/98829bda-56bb-4ef9-8c95-3c8fd7d57bc2-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "98829bda-56bb-4ef9-8c95-3c8fd7d57bc2" (UID: "98829bda-56bb-4ef9-8c95-3c8fd7d57bc2"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 08:12:42 crc kubenswrapper[4846]: I1005 08:12:42.602242 4846 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/98829bda-56bb-4ef9-8c95-3c8fd7d57bc2-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 05 08:12:42 crc kubenswrapper[4846]: I1005 08:12:42.602281 4846 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/98829bda-56bb-4ef9-8c95-3c8fd7d57bc2-config\") on node \"crc\" DevicePath \"\"" Oct 05 08:12:42 crc kubenswrapper[4846]: I1005 08:12:42.602291 4846 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/98829bda-56bb-4ef9-8c95-3c8fd7d57bc2-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 05 08:12:42 crc kubenswrapper[4846]: I1005 08:12:42.615662 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-copy-data"] Oct 05 08:12:42 crc kubenswrapper[4846]: E1005 08:12:42.616286 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="98829bda-56bb-4ef9-8c95-3c8fd7d57bc2" containerName="init" Oct 05 08:12:42 crc kubenswrapper[4846]: I1005 08:12:42.616385 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="98829bda-56bb-4ef9-8c95-3c8fd7d57bc2" containerName="init" Oct 05 08:12:42 crc kubenswrapper[4846]: I1005 08:12:42.616761 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="98829bda-56bb-4ef9-8c95-3c8fd7d57bc2" containerName="init" Oct 05 08:12:42 crc kubenswrapper[4846]: I1005 08:12:42.617726 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-copy-data"] Oct 05 08:12:42 crc kubenswrapper[4846]: I1005 08:12:42.617818 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-copy-data" Oct 05 08:12:42 crc kubenswrapper[4846]: I1005 08:12:42.619432 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovn-data-cert" Oct 05 08:12:42 crc kubenswrapper[4846]: I1005 08:12:42.703570 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-data-cert\" (UniqueName: \"kubernetes.io/secret/f0768736-73ad-4ce5-be3c-1941d56cd760-ovn-data-cert\") pod \"ovn-copy-data\" (UID: \"f0768736-73ad-4ce5-be3c-1941d56cd760\") " pod="openstack/ovn-copy-data" Oct 05 08:12:42 crc kubenswrapper[4846]: I1005 08:12:42.703720 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fdfrw\" (UniqueName: \"kubernetes.io/projected/f0768736-73ad-4ce5-be3c-1941d56cd760-kube-api-access-fdfrw\") pod \"ovn-copy-data\" (UID: \"f0768736-73ad-4ce5-be3c-1941d56cd760\") " pod="openstack/ovn-copy-data" Oct 05 08:12:42 crc kubenswrapper[4846]: I1005 08:12:42.703873 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-7d55f83c-6f76-43e2-9773-bc401946e9b8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-7d55f83c-6f76-43e2-9773-bc401946e9b8\") pod \"ovn-copy-data\" (UID: \"f0768736-73ad-4ce5-be3c-1941d56cd760\") " pod="openstack/ovn-copy-data" Oct 05 08:12:42 crc kubenswrapper[4846]: I1005 08:12:42.731630 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-697c78d8d5-2f9zd" event={"ID":"0a9ba111-4d92-4903-8c97-f386e3c29d5c","Type":"ContainerStarted","Data":"6ab8d5d38e9026b25124afb01ed5109b8d460824f788775d79e9a584be3f2edd"} Oct 05 08:12:42 crc kubenswrapper[4846]: I1005 08:12:42.731777 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-697c78d8d5-2f9zd" Oct 05 08:12:42 crc kubenswrapper[4846]: I1005 08:12:42.734060 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-9b45cbc8f-btndm" event={"ID":"98829bda-56bb-4ef9-8c95-3c8fd7d57bc2","Type":"ContainerDied","Data":"b11c4b05f04c9e0e50ff215dd2e4d4e638ac98e35b77339d6943ce5805881b7d"} Oct 05 08:12:42 crc kubenswrapper[4846]: I1005 08:12:42.734110 4846 scope.go:117] "RemoveContainer" containerID="3561e6e55e5652d5c8f1f9f3b810e3b7723ed3da4f888c2264301b4b507862eb" Oct 05 08:12:42 crc kubenswrapper[4846]: I1005 08:12:42.734113 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-9b45cbc8f-btndm" Oct 05 08:12:42 crc kubenswrapper[4846]: I1005 08:12:42.764892 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-697c78d8d5-2f9zd" podStartSLOduration=2.764873647 podStartE2EDuration="2.764873647s" podCreationTimestamp="2025-10-05 08:12:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 08:12:42.759877964 +0000 UTC m=+5085.000730779" watchObservedRunningTime="2025-10-05 08:12:42.764873647 +0000 UTC m=+5085.005726432" Oct 05 08:12:42 crc kubenswrapper[4846]: I1005 08:12:42.805294 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fdfrw\" (UniqueName: \"kubernetes.io/projected/f0768736-73ad-4ce5-be3c-1941d56cd760-kube-api-access-fdfrw\") pod \"ovn-copy-data\" (UID: \"f0768736-73ad-4ce5-be3c-1941d56cd760\") " pod="openstack/ovn-copy-data" Oct 05 08:12:42 crc kubenswrapper[4846]: I1005 08:12:42.805369 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-7d55f83c-6f76-43e2-9773-bc401946e9b8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-7d55f83c-6f76-43e2-9773-bc401946e9b8\") pod \"ovn-copy-data\" (UID: \"f0768736-73ad-4ce5-be3c-1941d56cd760\") " pod="openstack/ovn-copy-data" Oct 05 08:12:42 crc kubenswrapper[4846]: I1005 08:12:42.805505 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-data-cert\" (UniqueName: \"kubernetes.io/secret/f0768736-73ad-4ce5-be3c-1941d56cd760-ovn-data-cert\") pod \"ovn-copy-data\" (UID: \"f0768736-73ad-4ce5-be3c-1941d56cd760\") " pod="openstack/ovn-copy-data" Oct 05 08:12:42 crc kubenswrapper[4846]: I1005 08:12:42.807166 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-9b45cbc8f-btndm"] Oct 05 08:12:42 crc kubenswrapper[4846]: I1005 08:12:42.809689 4846 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 05 08:12:42 crc kubenswrapper[4846]: I1005 08:12:42.809728 4846 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-7d55f83c-6f76-43e2-9773-bc401946e9b8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-7d55f83c-6f76-43e2-9773-bc401946e9b8\") pod \"ovn-copy-data\" (UID: \"f0768736-73ad-4ce5-be3c-1941d56cd760\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/790e7d26a9df00d9880d6421e697ded23b1c76a8bd6c8a45e769f68df9b60dc5/globalmount\"" pod="openstack/ovn-copy-data" Oct 05 08:12:42 crc kubenswrapper[4846]: I1005 08:12:42.816609 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-9b45cbc8f-btndm"] Oct 05 08:12:42 crc kubenswrapper[4846]: I1005 08:12:42.817674 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-data-cert\" (UniqueName: \"kubernetes.io/secret/f0768736-73ad-4ce5-be3c-1941d56cd760-ovn-data-cert\") pod \"ovn-copy-data\" (UID: \"f0768736-73ad-4ce5-be3c-1941d56cd760\") " pod="openstack/ovn-copy-data" Oct 05 08:12:42 crc kubenswrapper[4846]: I1005 08:12:42.835283 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fdfrw\" (UniqueName: \"kubernetes.io/projected/f0768736-73ad-4ce5-be3c-1941d56cd760-kube-api-access-fdfrw\") pod \"ovn-copy-data\" (UID: \"f0768736-73ad-4ce5-be3c-1941d56cd760\") " pod="openstack/ovn-copy-data" Oct 05 08:12:42 crc kubenswrapper[4846]: I1005 08:12:42.854452 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-7d55f83c-6f76-43e2-9773-bc401946e9b8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-7d55f83c-6f76-43e2-9773-bc401946e9b8\") pod \"ovn-copy-data\" (UID: \"f0768736-73ad-4ce5-be3c-1941d56cd760\") " pod="openstack/ovn-copy-data" Oct 05 08:12:42 crc kubenswrapper[4846]: I1005 08:12:42.938526 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-copy-data" Oct 05 08:12:43 crc kubenswrapper[4846]: I1005 08:12:43.474864 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-copy-data"] Oct 05 08:12:43 crc kubenswrapper[4846]: I1005 08:12:43.743738 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-copy-data" event={"ID":"f0768736-73ad-4ce5-be3c-1941d56cd760","Type":"ContainerStarted","Data":"199d01d1d7efc979b2c00c0d62bafdfbab10d8cf0ac1b73fd55f71c4407448ab"} Oct 05 08:12:44 crc kubenswrapper[4846]: I1005 08:12:44.498498 4846 scope.go:117] "RemoveContainer" containerID="8fa72be735dff8e623a4a4958f452896ef79e72dd44c2d714a3019187462de47" Oct 05 08:12:44 crc kubenswrapper[4846]: E1005 08:12:44.499349 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 08:12:44 crc kubenswrapper[4846]: I1005 08:12:44.520505 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="98829bda-56bb-4ef9-8c95-3c8fd7d57bc2" path="/var/lib/kubelet/pods/98829bda-56bb-4ef9-8c95-3c8fd7d57bc2/volumes" Oct 05 08:12:44 crc kubenswrapper[4846]: I1005 08:12:44.761064 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-copy-data" event={"ID":"f0768736-73ad-4ce5-be3c-1941d56cd760","Type":"ContainerStarted","Data":"b2793fb5da2d48dbd09af9440210742861a05910349aabc2730a5269a46eecc4"} Oct 05 08:12:44 crc kubenswrapper[4846]: I1005 08:12:44.801478 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-copy-data" podStartSLOduration=3.625575322 podStartE2EDuration="3.801441711s" podCreationTimestamp="2025-10-05 08:12:41 +0000 UTC" firstStartedPulling="2025-10-05 08:12:43.48423442 +0000 UTC m=+5085.725087225" lastFinishedPulling="2025-10-05 08:12:43.660100819 +0000 UTC m=+5085.900953614" observedRunningTime="2025-10-05 08:12:44.790550299 +0000 UTC m=+5087.031403134" watchObservedRunningTime="2025-10-05 08:12:44.801441711 +0000 UTC m=+5087.042294536" Oct 05 08:12:50 crc kubenswrapper[4846]: I1005 08:12:50.528670 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-northd-0"] Oct 05 08:12:50 crc kubenswrapper[4846]: I1005 08:12:50.530714 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Oct 05 08:12:50 crc kubenswrapper[4846]: I1005 08:12:50.533407 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-scripts" Oct 05 08:12:50 crc kubenswrapper[4846]: I1005 08:12:50.533680 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovnnorthd-ovnnorthd-dockercfg-ltpkd" Oct 05 08:12:50 crc kubenswrapper[4846]: I1005 08:12:50.533815 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-config" Oct 05 08:12:50 crc kubenswrapper[4846]: I1005 08:12:50.537637 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovnnorthd-ovndbs" Oct 05 08:12:50 crc kubenswrapper[4846]: I1005 08:12:50.541126 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Oct 05 08:12:50 crc kubenswrapper[4846]: I1005 08:12:50.659344 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4bbb24b3-544b-47db-9058-47aa640afa64-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"4bbb24b3-544b-47db-9058-47aa640afa64\") " pod="openstack/ovn-northd-0" Oct 05 08:12:50 crc kubenswrapper[4846]: I1005 08:12:50.659429 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4bl5t\" (UniqueName: \"kubernetes.io/projected/4bbb24b3-544b-47db-9058-47aa640afa64-kube-api-access-4bl5t\") pod \"ovn-northd-0\" (UID: \"4bbb24b3-544b-47db-9058-47aa640afa64\") " pod="openstack/ovn-northd-0" Oct 05 08:12:50 crc kubenswrapper[4846]: I1005 08:12:50.659471 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/4bbb24b3-544b-47db-9058-47aa640afa64-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"4bbb24b3-544b-47db-9058-47aa640afa64\") " pod="openstack/ovn-northd-0" Oct 05 08:12:50 crc kubenswrapper[4846]: I1005 08:12:50.659499 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/4bbb24b3-544b-47db-9058-47aa640afa64-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"4bbb24b3-544b-47db-9058-47aa640afa64\") " pod="openstack/ovn-northd-0" Oct 05 08:12:50 crc kubenswrapper[4846]: I1005 08:12:50.659574 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/4bbb24b3-544b-47db-9058-47aa640afa64-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"4bbb24b3-544b-47db-9058-47aa640afa64\") " pod="openstack/ovn-northd-0" Oct 05 08:12:50 crc kubenswrapper[4846]: I1005 08:12:50.659615 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4bbb24b3-544b-47db-9058-47aa640afa64-config\") pod \"ovn-northd-0\" (UID: \"4bbb24b3-544b-47db-9058-47aa640afa64\") " pod="openstack/ovn-northd-0" Oct 05 08:12:50 crc kubenswrapper[4846]: I1005 08:12:50.659656 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4bbb24b3-544b-47db-9058-47aa640afa64-scripts\") pod \"ovn-northd-0\" (UID: \"4bbb24b3-544b-47db-9058-47aa640afa64\") " pod="openstack/ovn-northd-0" Oct 05 08:12:50 crc kubenswrapper[4846]: I1005 08:12:50.761065 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4bl5t\" (UniqueName: \"kubernetes.io/projected/4bbb24b3-544b-47db-9058-47aa640afa64-kube-api-access-4bl5t\") pod \"ovn-northd-0\" (UID: \"4bbb24b3-544b-47db-9058-47aa640afa64\") " pod="openstack/ovn-northd-0" Oct 05 08:12:50 crc kubenswrapper[4846]: I1005 08:12:50.761114 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/4bbb24b3-544b-47db-9058-47aa640afa64-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"4bbb24b3-544b-47db-9058-47aa640afa64\") " pod="openstack/ovn-northd-0" Oct 05 08:12:50 crc kubenswrapper[4846]: I1005 08:12:50.761141 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/4bbb24b3-544b-47db-9058-47aa640afa64-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"4bbb24b3-544b-47db-9058-47aa640afa64\") " pod="openstack/ovn-northd-0" Oct 05 08:12:50 crc kubenswrapper[4846]: I1005 08:12:50.761216 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/4bbb24b3-544b-47db-9058-47aa640afa64-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"4bbb24b3-544b-47db-9058-47aa640afa64\") " pod="openstack/ovn-northd-0" Oct 05 08:12:50 crc kubenswrapper[4846]: I1005 08:12:50.761250 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4bbb24b3-544b-47db-9058-47aa640afa64-config\") pod \"ovn-northd-0\" (UID: \"4bbb24b3-544b-47db-9058-47aa640afa64\") " pod="openstack/ovn-northd-0" Oct 05 08:12:50 crc kubenswrapper[4846]: I1005 08:12:50.761280 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4bbb24b3-544b-47db-9058-47aa640afa64-scripts\") pod \"ovn-northd-0\" (UID: \"4bbb24b3-544b-47db-9058-47aa640afa64\") " pod="openstack/ovn-northd-0" Oct 05 08:12:50 crc kubenswrapper[4846]: I1005 08:12:50.761307 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4bbb24b3-544b-47db-9058-47aa640afa64-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"4bbb24b3-544b-47db-9058-47aa640afa64\") " pod="openstack/ovn-northd-0" Oct 05 08:12:50 crc kubenswrapper[4846]: I1005 08:12:50.762529 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/4bbb24b3-544b-47db-9058-47aa640afa64-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"4bbb24b3-544b-47db-9058-47aa640afa64\") " pod="openstack/ovn-northd-0" Oct 05 08:12:50 crc kubenswrapper[4846]: I1005 08:12:50.763051 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4bbb24b3-544b-47db-9058-47aa640afa64-scripts\") pod \"ovn-northd-0\" (UID: \"4bbb24b3-544b-47db-9058-47aa640afa64\") " pod="openstack/ovn-northd-0" Oct 05 08:12:50 crc kubenswrapper[4846]: I1005 08:12:50.764301 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4bbb24b3-544b-47db-9058-47aa640afa64-config\") pod \"ovn-northd-0\" (UID: \"4bbb24b3-544b-47db-9058-47aa640afa64\") " pod="openstack/ovn-northd-0" Oct 05 08:12:50 crc kubenswrapper[4846]: I1005 08:12:50.766827 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4bbb24b3-544b-47db-9058-47aa640afa64-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"4bbb24b3-544b-47db-9058-47aa640afa64\") " pod="openstack/ovn-northd-0" Oct 05 08:12:50 crc kubenswrapper[4846]: I1005 08:12:50.767690 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/4bbb24b3-544b-47db-9058-47aa640afa64-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"4bbb24b3-544b-47db-9058-47aa640afa64\") " pod="openstack/ovn-northd-0" Oct 05 08:12:50 crc kubenswrapper[4846]: I1005 08:12:50.769232 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/4bbb24b3-544b-47db-9058-47aa640afa64-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"4bbb24b3-544b-47db-9058-47aa640afa64\") " pod="openstack/ovn-northd-0" Oct 05 08:12:50 crc kubenswrapper[4846]: I1005 08:12:50.795244 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4bl5t\" (UniqueName: \"kubernetes.io/projected/4bbb24b3-544b-47db-9058-47aa640afa64-kube-api-access-4bl5t\") pod \"ovn-northd-0\" (UID: \"4bbb24b3-544b-47db-9058-47aa640afa64\") " pod="openstack/ovn-northd-0" Oct 05 08:12:50 crc kubenswrapper[4846]: I1005 08:12:50.853710 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Oct 05 08:12:50 crc kubenswrapper[4846]: I1005 08:12:50.972787 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-697c78d8d5-2f9zd" Oct 05 08:12:51 crc kubenswrapper[4846]: I1005 08:12:51.040548 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-96d5866c7-cpp4z"] Oct 05 08:12:51 crc kubenswrapper[4846]: I1005 08:12:51.040826 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-96d5866c7-cpp4z" podUID="2001fca1-97a6-4420-ad29-3b83ab632284" containerName="dnsmasq-dns" containerID="cri-o://9d28829bb3200c54bbf39175a22dc488c7bca2383ac6674b8c071b500c3299c4" gracePeriod=10 Oct 05 08:12:51 crc kubenswrapper[4846]: I1005 08:12:51.334316 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Oct 05 08:12:51 crc kubenswrapper[4846]: I1005 08:12:51.438513 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-96d5866c7-cpp4z" Oct 05 08:12:51 crc kubenswrapper[4846]: I1005 08:12:51.574491 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2001fca1-97a6-4420-ad29-3b83ab632284-dns-svc\") pod \"2001fca1-97a6-4420-ad29-3b83ab632284\" (UID: \"2001fca1-97a6-4420-ad29-3b83ab632284\") " Oct 05 08:12:51 crc kubenswrapper[4846]: I1005 08:12:51.574581 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2001fca1-97a6-4420-ad29-3b83ab632284-config\") pod \"2001fca1-97a6-4420-ad29-3b83ab632284\" (UID: \"2001fca1-97a6-4420-ad29-3b83ab632284\") " Oct 05 08:12:51 crc kubenswrapper[4846]: I1005 08:12:51.574674 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-psx4k\" (UniqueName: \"kubernetes.io/projected/2001fca1-97a6-4420-ad29-3b83ab632284-kube-api-access-psx4k\") pod \"2001fca1-97a6-4420-ad29-3b83ab632284\" (UID: \"2001fca1-97a6-4420-ad29-3b83ab632284\") " Oct 05 08:12:51 crc kubenswrapper[4846]: I1005 08:12:51.579713 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2001fca1-97a6-4420-ad29-3b83ab632284-kube-api-access-psx4k" (OuterVolumeSpecName: "kube-api-access-psx4k") pod "2001fca1-97a6-4420-ad29-3b83ab632284" (UID: "2001fca1-97a6-4420-ad29-3b83ab632284"). InnerVolumeSpecName "kube-api-access-psx4k". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:12:51 crc kubenswrapper[4846]: I1005 08:12:51.614719 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2001fca1-97a6-4420-ad29-3b83ab632284-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "2001fca1-97a6-4420-ad29-3b83ab632284" (UID: "2001fca1-97a6-4420-ad29-3b83ab632284"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 08:12:51 crc kubenswrapper[4846]: I1005 08:12:51.616035 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2001fca1-97a6-4420-ad29-3b83ab632284-config" (OuterVolumeSpecName: "config") pod "2001fca1-97a6-4420-ad29-3b83ab632284" (UID: "2001fca1-97a6-4420-ad29-3b83ab632284"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 08:12:51 crc kubenswrapper[4846]: I1005 08:12:51.676101 4846 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2001fca1-97a6-4420-ad29-3b83ab632284-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 05 08:12:51 crc kubenswrapper[4846]: I1005 08:12:51.676129 4846 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2001fca1-97a6-4420-ad29-3b83ab632284-config\") on node \"crc\" DevicePath \"\"" Oct 05 08:12:51 crc kubenswrapper[4846]: I1005 08:12:51.676138 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-psx4k\" (UniqueName: \"kubernetes.io/projected/2001fca1-97a6-4420-ad29-3b83ab632284-kube-api-access-psx4k\") on node \"crc\" DevicePath \"\"" Oct 05 08:12:51 crc kubenswrapper[4846]: E1005 08:12:51.689705 4846 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.111:57940->38.102.83.111:43661: write tcp 38.102.83.111:57940->38.102.83.111:43661: write: broken pipe Oct 05 08:12:51 crc kubenswrapper[4846]: I1005 08:12:51.825631 4846 generic.go:334] "Generic (PLEG): container finished" podID="2001fca1-97a6-4420-ad29-3b83ab632284" containerID="9d28829bb3200c54bbf39175a22dc488c7bca2383ac6674b8c071b500c3299c4" exitCode=0 Oct 05 08:12:51 crc kubenswrapper[4846]: I1005 08:12:51.825682 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-96d5866c7-cpp4z" event={"ID":"2001fca1-97a6-4420-ad29-3b83ab632284","Type":"ContainerDied","Data":"9d28829bb3200c54bbf39175a22dc488c7bca2383ac6674b8c071b500c3299c4"} Oct 05 08:12:51 crc kubenswrapper[4846]: I1005 08:12:51.825707 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-96d5866c7-cpp4z" event={"ID":"2001fca1-97a6-4420-ad29-3b83ab632284","Type":"ContainerDied","Data":"e2497b036f252756187de8b44fb734a3c7b5ca587b07a2b4ffc45b9dce57edc7"} Oct 05 08:12:51 crc kubenswrapper[4846]: I1005 08:12:51.825724 4846 scope.go:117] "RemoveContainer" containerID="9d28829bb3200c54bbf39175a22dc488c7bca2383ac6674b8c071b500c3299c4" Oct 05 08:12:51 crc kubenswrapper[4846]: I1005 08:12:51.825823 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-96d5866c7-cpp4z" Oct 05 08:12:51 crc kubenswrapper[4846]: I1005 08:12:51.832158 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"4bbb24b3-544b-47db-9058-47aa640afa64","Type":"ContainerStarted","Data":"53ccf9decf9df381c847608ad07dae117356d50df6d1631eac65e910946c6a7e"} Oct 05 08:12:51 crc kubenswrapper[4846]: I1005 08:12:51.860088 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-96d5866c7-cpp4z"] Oct 05 08:12:51 crc kubenswrapper[4846]: I1005 08:12:51.865220 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-96d5866c7-cpp4z"] Oct 05 08:12:51 crc kubenswrapper[4846]: I1005 08:12:51.996591 4846 scope.go:117] "RemoveContainer" containerID="af710e4cd5ecb39d20b237388f1eccc6ee4c8d63599d39ae3fcc508fa36c1fe6" Oct 05 08:12:52 crc kubenswrapper[4846]: I1005 08:12:52.050651 4846 scope.go:117] "RemoveContainer" containerID="9d28829bb3200c54bbf39175a22dc488c7bca2383ac6674b8c071b500c3299c4" Oct 05 08:12:52 crc kubenswrapper[4846]: E1005 08:12:52.051278 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9d28829bb3200c54bbf39175a22dc488c7bca2383ac6674b8c071b500c3299c4\": container with ID starting with 9d28829bb3200c54bbf39175a22dc488c7bca2383ac6674b8c071b500c3299c4 not found: ID does not exist" containerID="9d28829bb3200c54bbf39175a22dc488c7bca2383ac6674b8c071b500c3299c4" Oct 05 08:12:52 crc kubenswrapper[4846]: I1005 08:12:52.051344 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9d28829bb3200c54bbf39175a22dc488c7bca2383ac6674b8c071b500c3299c4"} err="failed to get container status \"9d28829bb3200c54bbf39175a22dc488c7bca2383ac6674b8c071b500c3299c4\": rpc error: code = NotFound desc = could not find container \"9d28829bb3200c54bbf39175a22dc488c7bca2383ac6674b8c071b500c3299c4\": container with ID starting with 9d28829bb3200c54bbf39175a22dc488c7bca2383ac6674b8c071b500c3299c4 not found: ID does not exist" Oct 05 08:12:52 crc kubenswrapper[4846]: I1005 08:12:52.051378 4846 scope.go:117] "RemoveContainer" containerID="af710e4cd5ecb39d20b237388f1eccc6ee4c8d63599d39ae3fcc508fa36c1fe6" Oct 05 08:12:52 crc kubenswrapper[4846]: E1005 08:12:52.051706 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"af710e4cd5ecb39d20b237388f1eccc6ee4c8d63599d39ae3fcc508fa36c1fe6\": container with ID starting with af710e4cd5ecb39d20b237388f1eccc6ee4c8d63599d39ae3fcc508fa36c1fe6 not found: ID does not exist" containerID="af710e4cd5ecb39d20b237388f1eccc6ee4c8d63599d39ae3fcc508fa36c1fe6" Oct 05 08:12:52 crc kubenswrapper[4846]: I1005 08:12:52.051729 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"af710e4cd5ecb39d20b237388f1eccc6ee4c8d63599d39ae3fcc508fa36c1fe6"} err="failed to get container status \"af710e4cd5ecb39d20b237388f1eccc6ee4c8d63599d39ae3fcc508fa36c1fe6\": rpc error: code = NotFound desc = could not find container \"af710e4cd5ecb39d20b237388f1eccc6ee4c8d63599d39ae3fcc508fa36c1fe6\": container with ID starting with af710e4cd5ecb39d20b237388f1eccc6ee4c8d63599d39ae3fcc508fa36c1fe6 not found: ID does not exist" Oct 05 08:12:52 crc kubenswrapper[4846]: I1005 08:12:52.511064 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2001fca1-97a6-4420-ad29-3b83ab632284" path="/var/lib/kubelet/pods/2001fca1-97a6-4420-ad29-3b83ab632284/volumes" Oct 05 08:12:52 crc kubenswrapper[4846]: I1005 08:12:52.846847 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"4bbb24b3-544b-47db-9058-47aa640afa64","Type":"ContainerStarted","Data":"fe7c36356b1637624424ce1aa11e63f0477ac8b4ce011665da6ca597abadf471"} Oct 05 08:12:52 crc kubenswrapper[4846]: I1005 08:12:52.846904 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"4bbb24b3-544b-47db-9058-47aa640afa64","Type":"ContainerStarted","Data":"6ba370d9448660b177419c92e451dbb7c95c9eddd7dc699ced570b17c1f19030"} Oct 05 08:12:52 crc kubenswrapper[4846]: I1005 08:12:52.847069 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-northd-0" Oct 05 08:12:52 crc kubenswrapper[4846]: I1005 08:12:52.884748 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-northd-0" podStartSLOduration=2.1853984300000002 podStartE2EDuration="2.884716795s" podCreationTimestamp="2025-10-05 08:12:50 +0000 UTC" firstStartedPulling="2025-10-05 08:12:51.352722004 +0000 UTC m=+5093.593574789" lastFinishedPulling="2025-10-05 08:12:52.052040339 +0000 UTC m=+5094.292893154" observedRunningTime="2025-10-05 08:12:52.877559063 +0000 UTC m=+5095.118411879" watchObservedRunningTime="2025-10-05 08:12:52.884716795 +0000 UTC m=+5095.125569610" Oct 05 08:12:55 crc kubenswrapper[4846]: I1005 08:12:55.498680 4846 scope.go:117] "RemoveContainer" containerID="8fa72be735dff8e623a4a4958f452896ef79e72dd44c2d714a3019187462de47" Oct 05 08:12:55 crc kubenswrapper[4846]: E1005 08:12:55.500086 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 08:12:56 crc kubenswrapper[4846]: I1005 08:12:56.007345 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-create-mblxr"] Oct 05 08:12:56 crc kubenswrapper[4846]: E1005 08:12:56.007688 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2001fca1-97a6-4420-ad29-3b83ab632284" containerName="init" Oct 05 08:12:56 crc kubenswrapper[4846]: I1005 08:12:56.007707 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="2001fca1-97a6-4420-ad29-3b83ab632284" containerName="init" Oct 05 08:12:56 crc kubenswrapper[4846]: E1005 08:12:56.007730 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2001fca1-97a6-4420-ad29-3b83ab632284" containerName="dnsmasq-dns" Oct 05 08:12:56 crc kubenswrapper[4846]: I1005 08:12:56.007738 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="2001fca1-97a6-4420-ad29-3b83ab632284" containerName="dnsmasq-dns" Oct 05 08:12:56 crc kubenswrapper[4846]: I1005 08:12:56.007976 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="2001fca1-97a6-4420-ad29-3b83ab632284" containerName="dnsmasq-dns" Oct 05 08:12:56 crc kubenswrapper[4846]: I1005 08:12:56.008509 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-mblxr" Oct 05 08:12:56 crc kubenswrapper[4846]: I1005 08:12:56.018324 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-mblxr"] Oct 05 08:12:56 crc kubenswrapper[4846]: I1005 08:12:56.073106 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4ffbp\" (UniqueName: \"kubernetes.io/projected/bd37756c-8442-4121-8c0d-fa534d95edab-kube-api-access-4ffbp\") pod \"keystone-db-create-mblxr\" (UID: \"bd37756c-8442-4121-8c0d-fa534d95edab\") " pod="openstack/keystone-db-create-mblxr" Oct 05 08:12:56 crc kubenswrapper[4846]: I1005 08:12:56.174778 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4ffbp\" (UniqueName: \"kubernetes.io/projected/bd37756c-8442-4121-8c0d-fa534d95edab-kube-api-access-4ffbp\") pod \"keystone-db-create-mblxr\" (UID: \"bd37756c-8442-4121-8c0d-fa534d95edab\") " pod="openstack/keystone-db-create-mblxr" Oct 05 08:12:56 crc kubenswrapper[4846]: I1005 08:12:56.196859 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4ffbp\" (UniqueName: \"kubernetes.io/projected/bd37756c-8442-4121-8c0d-fa534d95edab-kube-api-access-4ffbp\") pod \"keystone-db-create-mblxr\" (UID: \"bd37756c-8442-4121-8c0d-fa534d95edab\") " pod="openstack/keystone-db-create-mblxr" Oct 05 08:12:56 crc kubenswrapper[4846]: I1005 08:12:56.349927 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-mblxr" Oct 05 08:12:56 crc kubenswrapper[4846]: I1005 08:12:56.642155 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-mblxr"] Oct 05 08:12:56 crc kubenswrapper[4846]: I1005 08:12:56.891163 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-mblxr" event={"ID":"bd37756c-8442-4121-8c0d-fa534d95edab","Type":"ContainerStarted","Data":"c721c574769f3dc4ebcc372911f37c071866043013c026ac0fc7582fa24ddb00"} Oct 05 08:12:56 crc kubenswrapper[4846]: I1005 08:12:56.891436 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-mblxr" event={"ID":"bd37756c-8442-4121-8c0d-fa534d95edab","Type":"ContainerStarted","Data":"6d9f956277d9099d052421af486f679638008d139d62d1b11223d1fc520d60fb"} Oct 05 08:12:57 crc kubenswrapper[4846]: I1005 08:12:57.904692 4846 generic.go:334] "Generic (PLEG): container finished" podID="bd37756c-8442-4121-8c0d-fa534d95edab" containerID="c721c574769f3dc4ebcc372911f37c071866043013c026ac0fc7582fa24ddb00" exitCode=0 Oct 05 08:12:57 crc kubenswrapper[4846]: I1005 08:12:57.904769 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-mblxr" event={"ID":"bd37756c-8442-4121-8c0d-fa534d95edab","Type":"ContainerDied","Data":"c721c574769f3dc4ebcc372911f37c071866043013c026ac0fc7582fa24ddb00"} Oct 05 08:12:58 crc kubenswrapper[4846]: I1005 08:12:58.274590 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-mblxr" Oct 05 08:12:58 crc kubenswrapper[4846]: I1005 08:12:58.330146 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4ffbp\" (UniqueName: \"kubernetes.io/projected/bd37756c-8442-4121-8c0d-fa534d95edab-kube-api-access-4ffbp\") pod \"bd37756c-8442-4121-8c0d-fa534d95edab\" (UID: \"bd37756c-8442-4121-8c0d-fa534d95edab\") " Oct 05 08:12:58 crc kubenswrapper[4846]: I1005 08:12:58.345532 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd37756c-8442-4121-8c0d-fa534d95edab-kube-api-access-4ffbp" (OuterVolumeSpecName: "kube-api-access-4ffbp") pod "bd37756c-8442-4121-8c0d-fa534d95edab" (UID: "bd37756c-8442-4121-8c0d-fa534d95edab"). InnerVolumeSpecName "kube-api-access-4ffbp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:12:58 crc kubenswrapper[4846]: I1005 08:12:58.431948 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4ffbp\" (UniqueName: \"kubernetes.io/projected/bd37756c-8442-4121-8c0d-fa534d95edab-kube-api-access-4ffbp\") on node \"crc\" DevicePath \"\"" Oct 05 08:12:58 crc kubenswrapper[4846]: I1005 08:12:58.925674 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-mblxr" event={"ID":"bd37756c-8442-4121-8c0d-fa534d95edab","Type":"ContainerDied","Data":"6d9f956277d9099d052421af486f679638008d139d62d1b11223d1fc520d60fb"} Oct 05 08:12:58 crc kubenswrapper[4846]: I1005 08:12:58.925725 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6d9f956277d9099d052421af486f679638008d139d62d1b11223d1fc520d60fb" Oct 05 08:12:58 crc kubenswrapper[4846]: I1005 08:12:58.925880 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-mblxr" Oct 05 08:13:05 crc kubenswrapper[4846]: I1005 08:13:05.931250 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-northd-0" Oct 05 08:13:06 crc kubenswrapper[4846]: I1005 08:13:06.106030 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-d973-account-create-nt7ql"] Oct 05 08:13:06 crc kubenswrapper[4846]: E1005 08:13:06.106371 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bd37756c-8442-4121-8c0d-fa534d95edab" containerName="mariadb-database-create" Oct 05 08:13:06 crc kubenswrapper[4846]: I1005 08:13:06.106393 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="bd37756c-8442-4121-8c0d-fa534d95edab" containerName="mariadb-database-create" Oct 05 08:13:06 crc kubenswrapper[4846]: I1005 08:13:06.106559 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="bd37756c-8442-4121-8c0d-fa534d95edab" containerName="mariadb-database-create" Oct 05 08:13:06 crc kubenswrapper[4846]: I1005 08:13:06.107080 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-d973-account-create-nt7ql" Oct 05 08:13:06 crc kubenswrapper[4846]: I1005 08:13:06.109077 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-db-secret" Oct 05 08:13:06 crc kubenswrapper[4846]: I1005 08:13:06.117847 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-d973-account-create-nt7ql"] Oct 05 08:13:06 crc kubenswrapper[4846]: I1005 08:13:06.172206 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h9ffc\" (UniqueName: \"kubernetes.io/projected/6f59c11b-1a78-464d-b6c5-2367019b1140-kube-api-access-h9ffc\") pod \"keystone-d973-account-create-nt7ql\" (UID: \"6f59c11b-1a78-464d-b6c5-2367019b1140\") " pod="openstack/keystone-d973-account-create-nt7ql" Oct 05 08:13:06 crc kubenswrapper[4846]: I1005 08:13:06.275079 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h9ffc\" (UniqueName: \"kubernetes.io/projected/6f59c11b-1a78-464d-b6c5-2367019b1140-kube-api-access-h9ffc\") pod \"keystone-d973-account-create-nt7ql\" (UID: \"6f59c11b-1a78-464d-b6c5-2367019b1140\") " pod="openstack/keystone-d973-account-create-nt7ql" Oct 05 08:13:06 crc kubenswrapper[4846]: I1005 08:13:06.299016 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h9ffc\" (UniqueName: \"kubernetes.io/projected/6f59c11b-1a78-464d-b6c5-2367019b1140-kube-api-access-h9ffc\") pod \"keystone-d973-account-create-nt7ql\" (UID: \"6f59c11b-1a78-464d-b6c5-2367019b1140\") " pod="openstack/keystone-d973-account-create-nt7ql" Oct 05 08:13:06 crc kubenswrapper[4846]: I1005 08:13:06.426089 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-d973-account-create-nt7ql" Oct 05 08:13:06 crc kubenswrapper[4846]: I1005 08:13:06.497874 4846 scope.go:117] "RemoveContainer" containerID="8fa72be735dff8e623a4a4958f452896ef79e72dd44c2d714a3019187462de47" Oct 05 08:13:06 crc kubenswrapper[4846]: E1005 08:13:06.498355 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 08:13:06 crc kubenswrapper[4846]: I1005 08:13:06.719937 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-d973-account-create-nt7ql"] Oct 05 08:13:06 crc kubenswrapper[4846]: W1005 08:13:06.728327 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6f59c11b_1a78_464d_b6c5_2367019b1140.slice/crio-901f5d24dd163857f2924c1ed9fc5aeefcece222ae2e9e1582518325349bb992 WatchSource:0}: Error finding container 901f5d24dd163857f2924c1ed9fc5aeefcece222ae2e9e1582518325349bb992: Status 404 returned error can't find the container with id 901f5d24dd163857f2924c1ed9fc5aeefcece222ae2e9e1582518325349bb992 Oct 05 08:13:07 crc kubenswrapper[4846]: I1005 08:13:07.002727 4846 generic.go:334] "Generic (PLEG): container finished" podID="6f59c11b-1a78-464d-b6c5-2367019b1140" containerID="5d0729fccbe964bd9f23bb56ceb1eac397bcd850556d8691a91d1c009adb5c87" exitCode=0 Oct 05 08:13:07 crc kubenswrapper[4846]: I1005 08:13:07.002843 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-d973-account-create-nt7ql" event={"ID":"6f59c11b-1a78-464d-b6c5-2367019b1140","Type":"ContainerDied","Data":"5d0729fccbe964bd9f23bb56ceb1eac397bcd850556d8691a91d1c009adb5c87"} Oct 05 08:13:07 crc kubenswrapper[4846]: I1005 08:13:07.003561 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-d973-account-create-nt7ql" event={"ID":"6f59c11b-1a78-464d-b6c5-2367019b1140","Type":"ContainerStarted","Data":"901f5d24dd163857f2924c1ed9fc5aeefcece222ae2e9e1582518325349bb992"} Oct 05 08:13:08 crc kubenswrapper[4846]: I1005 08:13:08.439620 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-d973-account-create-nt7ql" Oct 05 08:13:08 crc kubenswrapper[4846]: I1005 08:13:08.517079 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h9ffc\" (UniqueName: \"kubernetes.io/projected/6f59c11b-1a78-464d-b6c5-2367019b1140-kube-api-access-h9ffc\") pod \"6f59c11b-1a78-464d-b6c5-2367019b1140\" (UID: \"6f59c11b-1a78-464d-b6c5-2367019b1140\") " Oct 05 08:13:08 crc kubenswrapper[4846]: I1005 08:13:08.524289 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6f59c11b-1a78-464d-b6c5-2367019b1140-kube-api-access-h9ffc" (OuterVolumeSpecName: "kube-api-access-h9ffc") pod "6f59c11b-1a78-464d-b6c5-2367019b1140" (UID: "6f59c11b-1a78-464d-b6c5-2367019b1140"). InnerVolumeSpecName "kube-api-access-h9ffc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:13:08 crc kubenswrapper[4846]: I1005 08:13:08.620872 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h9ffc\" (UniqueName: \"kubernetes.io/projected/6f59c11b-1a78-464d-b6c5-2367019b1140-kube-api-access-h9ffc\") on node \"crc\" DevicePath \"\"" Oct 05 08:13:09 crc kubenswrapper[4846]: I1005 08:13:09.024455 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-d973-account-create-nt7ql" event={"ID":"6f59c11b-1a78-464d-b6c5-2367019b1140","Type":"ContainerDied","Data":"901f5d24dd163857f2924c1ed9fc5aeefcece222ae2e9e1582518325349bb992"} Oct 05 08:13:09 crc kubenswrapper[4846]: I1005 08:13:09.024509 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="901f5d24dd163857f2924c1ed9fc5aeefcece222ae2e9e1582518325349bb992" Oct 05 08:13:09 crc kubenswrapper[4846]: I1005 08:13:09.024554 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-d973-account-create-nt7ql" Oct 05 08:13:11 crc kubenswrapper[4846]: I1005 08:13:11.505221 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-sync-g76dd"] Oct 05 08:13:11 crc kubenswrapper[4846]: E1005 08:13:11.506078 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6f59c11b-1a78-464d-b6c5-2367019b1140" containerName="mariadb-account-create" Oct 05 08:13:11 crc kubenswrapper[4846]: I1005 08:13:11.506090 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="6f59c11b-1a78-464d-b6c5-2367019b1140" containerName="mariadb-account-create" Oct 05 08:13:11 crc kubenswrapper[4846]: I1005 08:13:11.506281 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="6f59c11b-1a78-464d-b6c5-2367019b1140" containerName="mariadb-account-create" Oct 05 08:13:11 crc kubenswrapper[4846]: I1005 08:13:11.506830 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-g76dd" Oct 05 08:13:11 crc kubenswrapper[4846]: I1005 08:13:11.508981 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 05 08:13:11 crc kubenswrapper[4846]: I1005 08:13:11.510658 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-zhh8k" Oct 05 08:13:11 crc kubenswrapper[4846]: I1005 08:13:11.510684 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 05 08:13:11 crc kubenswrapper[4846]: I1005 08:13:11.511203 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 05 08:13:11 crc kubenswrapper[4846]: I1005 08:13:11.517400 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-g76dd"] Oct 05 08:13:11 crc kubenswrapper[4846]: I1005 08:13:11.572758 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8548783c-915e-4230-8b94-04cffa8ae2d9-combined-ca-bundle\") pod \"keystone-db-sync-g76dd\" (UID: \"8548783c-915e-4230-8b94-04cffa8ae2d9\") " pod="openstack/keystone-db-sync-g76dd" Oct 05 08:13:11 crc kubenswrapper[4846]: I1005 08:13:11.573050 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8548783c-915e-4230-8b94-04cffa8ae2d9-config-data\") pod \"keystone-db-sync-g76dd\" (UID: \"8548783c-915e-4230-8b94-04cffa8ae2d9\") " pod="openstack/keystone-db-sync-g76dd" Oct 05 08:13:11 crc kubenswrapper[4846]: I1005 08:13:11.573168 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h8jwh\" (UniqueName: \"kubernetes.io/projected/8548783c-915e-4230-8b94-04cffa8ae2d9-kube-api-access-h8jwh\") pod \"keystone-db-sync-g76dd\" (UID: \"8548783c-915e-4230-8b94-04cffa8ae2d9\") " pod="openstack/keystone-db-sync-g76dd" Oct 05 08:13:11 crc kubenswrapper[4846]: I1005 08:13:11.674167 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h8jwh\" (UniqueName: \"kubernetes.io/projected/8548783c-915e-4230-8b94-04cffa8ae2d9-kube-api-access-h8jwh\") pod \"keystone-db-sync-g76dd\" (UID: \"8548783c-915e-4230-8b94-04cffa8ae2d9\") " pod="openstack/keystone-db-sync-g76dd" Oct 05 08:13:11 crc kubenswrapper[4846]: I1005 08:13:11.674268 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8548783c-915e-4230-8b94-04cffa8ae2d9-combined-ca-bundle\") pod \"keystone-db-sync-g76dd\" (UID: \"8548783c-915e-4230-8b94-04cffa8ae2d9\") " pod="openstack/keystone-db-sync-g76dd" Oct 05 08:13:11 crc kubenswrapper[4846]: I1005 08:13:11.674373 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8548783c-915e-4230-8b94-04cffa8ae2d9-config-data\") pod \"keystone-db-sync-g76dd\" (UID: \"8548783c-915e-4230-8b94-04cffa8ae2d9\") " pod="openstack/keystone-db-sync-g76dd" Oct 05 08:13:11 crc kubenswrapper[4846]: I1005 08:13:11.681473 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8548783c-915e-4230-8b94-04cffa8ae2d9-config-data\") pod \"keystone-db-sync-g76dd\" (UID: \"8548783c-915e-4230-8b94-04cffa8ae2d9\") " pod="openstack/keystone-db-sync-g76dd" Oct 05 08:13:11 crc kubenswrapper[4846]: I1005 08:13:11.683100 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8548783c-915e-4230-8b94-04cffa8ae2d9-combined-ca-bundle\") pod \"keystone-db-sync-g76dd\" (UID: \"8548783c-915e-4230-8b94-04cffa8ae2d9\") " pod="openstack/keystone-db-sync-g76dd" Oct 05 08:13:11 crc kubenswrapper[4846]: I1005 08:13:11.700473 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h8jwh\" (UniqueName: \"kubernetes.io/projected/8548783c-915e-4230-8b94-04cffa8ae2d9-kube-api-access-h8jwh\") pod \"keystone-db-sync-g76dd\" (UID: \"8548783c-915e-4230-8b94-04cffa8ae2d9\") " pod="openstack/keystone-db-sync-g76dd" Oct 05 08:13:11 crc kubenswrapper[4846]: I1005 08:13:11.823492 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-g76dd" Oct 05 08:13:12 crc kubenswrapper[4846]: W1005 08:13:12.353025 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8548783c_915e_4230_8b94_04cffa8ae2d9.slice/crio-11e7db51b8bc696f778dd2f0a922c5bfab46a14ecd66c0360aee61f274794fc5 WatchSource:0}: Error finding container 11e7db51b8bc696f778dd2f0a922c5bfab46a14ecd66c0360aee61f274794fc5: Status 404 returned error can't find the container with id 11e7db51b8bc696f778dd2f0a922c5bfab46a14ecd66c0360aee61f274794fc5 Oct 05 08:13:12 crc kubenswrapper[4846]: I1005 08:13:12.359571 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-g76dd"] Oct 05 08:13:13 crc kubenswrapper[4846]: I1005 08:13:13.059721 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-g76dd" event={"ID":"8548783c-915e-4230-8b94-04cffa8ae2d9","Type":"ContainerStarted","Data":"11e7db51b8bc696f778dd2f0a922c5bfab46a14ecd66c0360aee61f274794fc5"} Oct 05 08:13:18 crc kubenswrapper[4846]: I1005 08:13:18.127767 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-g76dd" event={"ID":"8548783c-915e-4230-8b94-04cffa8ae2d9","Type":"ContainerStarted","Data":"b27004a16006eacec2ac7eb7962288a705d7cb1acb5d972dac5454fc0759e990"} Oct 05 08:13:18 crc kubenswrapper[4846]: I1005 08:13:18.149276 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-sync-g76dd" podStartSLOduration=2.192130345 podStartE2EDuration="7.1492611s" podCreationTimestamp="2025-10-05 08:13:11 +0000 UTC" firstStartedPulling="2025-10-05 08:13:12.355617356 +0000 UTC m=+5114.596470131" lastFinishedPulling="2025-10-05 08:13:17.312748081 +0000 UTC m=+5119.553600886" observedRunningTime="2025-10-05 08:13:18.143618619 +0000 UTC m=+5120.384471404" watchObservedRunningTime="2025-10-05 08:13:18.1492611 +0000 UTC m=+5120.390113875" Oct 05 08:13:19 crc kubenswrapper[4846]: I1005 08:13:19.147905 4846 generic.go:334] "Generic (PLEG): container finished" podID="8548783c-915e-4230-8b94-04cffa8ae2d9" containerID="b27004a16006eacec2ac7eb7962288a705d7cb1acb5d972dac5454fc0759e990" exitCode=0 Oct 05 08:13:19 crc kubenswrapper[4846]: I1005 08:13:19.148049 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-g76dd" event={"ID":"8548783c-915e-4230-8b94-04cffa8ae2d9","Type":"ContainerDied","Data":"b27004a16006eacec2ac7eb7962288a705d7cb1acb5d972dac5454fc0759e990"} Oct 05 08:13:19 crc kubenswrapper[4846]: I1005 08:13:19.497051 4846 scope.go:117] "RemoveContainer" containerID="8fa72be735dff8e623a4a4958f452896ef79e72dd44c2d714a3019187462de47" Oct 05 08:13:19 crc kubenswrapper[4846]: E1005 08:13:19.497382 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 08:13:20 crc kubenswrapper[4846]: I1005 08:13:20.549872 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-g76dd" Oct 05 08:13:20 crc kubenswrapper[4846]: I1005 08:13:20.692338 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8548783c-915e-4230-8b94-04cffa8ae2d9-config-data\") pod \"8548783c-915e-4230-8b94-04cffa8ae2d9\" (UID: \"8548783c-915e-4230-8b94-04cffa8ae2d9\") " Oct 05 08:13:20 crc kubenswrapper[4846]: I1005 08:13:20.692393 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8548783c-915e-4230-8b94-04cffa8ae2d9-combined-ca-bundle\") pod \"8548783c-915e-4230-8b94-04cffa8ae2d9\" (UID: \"8548783c-915e-4230-8b94-04cffa8ae2d9\") " Oct 05 08:13:20 crc kubenswrapper[4846]: I1005 08:13:20.692422 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h8jwh\" (UniqueName: \"kubernetes.io/projected/8548783c-915e-4230-8b94-04cffa8ae2d9-kube-api-access-h8jwh\") pod \"8548783c-915e-4230-8b94-04cffa8ae2d9\" (UID: \"8548783c-915e-4230-8b94-04cffa8ae2d9\") " Oct 05 08:13:20 crc kubenswrapper[4846]: I1005 08:13:20.698212 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8548783c-915e-4230-8b94-04cffa8ae2d9-kube-api-access-h8jwh" (OuterVolumeSpecName: "kube-api-access-h8jwh") pod "8548783c-915e-4230-8b94-04cffa8ae2d9" (UID: "8548783c-915e-4230-8b94-04cffa8ae2d9"). InnerVolumeSpecName "kube-api-access-h8jwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:13:20 crc kubenswrapper[4846]: I1005 08:13:20.716021 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8548783c-915e-4230-8b94-04cffa8ae2d9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8548783c-915e-4230-8b94-04cffa8ae2d9" (UID: "8548783c-915e-4230-8b94-04cffa8ae2d9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:13:20 crc kubenswrapper[4846]: I1005 08:13:20.753308 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8548783c-915e-4230-8b94-04cffa8ae2d9-config-data" (OuterVolumeSpecName: "config-data") pod "8548783c-915e-4230-8b94-04cffa8ae2d9" (UID: "8548783c-915e-4230-8b94-04cffa8ae2d9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:13:20 crc kubenswrapper[4846]: I1005 08:13:20.795368 4846 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8548783c-915e-4230-8b94-04cffa8ae2d9-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 08:13:20 crc kubenswrapper[4846]: I1005 08:13:20.795784 4846 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8548783c-915e-4230-8b94-04cffa8ae2d9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 08:13:20 crc kubenswrapper[4846]: I1005 08:13:20.795813 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h8jwh\" (UniqueName: \"kubernetes.io/projected/8548783c-915e-4230-8b94-04cffa8ae2d9-kube-api-access-h8jwh\") on node \"crc\" DevicePath \"\"" Oct 05 08:13:21 crc kubenswrapper[4846]: I1005 08:13:21.165845 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-g76dd" event={"ID":"8548783c-915e-4230-8b94-04cffa8ae2d9","Type":"ContainerDied","Data":"11e7db51b8bc696f778dd2f0a922c5bfab46a14ecd66c0360aee61f274794fc5"} Oct 05 08:13:21 crc kubenswrapper[4846]: I1005 08:13:21.165897 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="11e7db51b8bc696f778dd2f0a922c5bfab46a14ecd66c0360aee61f274794fc5" Oct 05 08:13:21 crc kubenswrapper[4846]: I1005 08:13:21.165958 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-g76dd" Oct 05 08:13:21 crc kubenswrapper[4846]: I1005 08:13:21.415867 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6d65c684ff-nnw4f"] Oct 05 08:13:21 crc kubenswrapper[4846]: E1005 08:13:21.416222 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8548783c-915e-4230-8b94-04cffa8ae2d9" containerName="keystone-db-sync" Oct 05 08:13:21 crc kubenswrapper[4846]: I1005 08:13:21.416236 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="8548783c-915e-4230-8b94-04cffa8ae2d9" containerName="keystone-db-sync" Oct 05 08:13:21 crc kubenswrapper[4846]: I1005 08:13:21.416617 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="8548783c-915e-4230-8b94-04cffa8ae2d9" containerName="keystone-db-sync" Oct 05 08:13:21 crc kubenswrapper[4846]: I1005 08:13:21.417534 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6d65c684ff-nnw4f" Oct 05 08:13:21 crc kubenswrapper[4846]: I1005 08:13:21.461039 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6d65c684ff-nnw4f"] Oct 05 08:13:21 crc kubenswrapper[4846]: I1005 08:13:21.510290 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-sf56p"] Oct 05 08:13:21 crc kubenswrapper[4846]: I1005 08:13:21.511229 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fd7jq\" (UniqueName: \"kubernetes.io/projected/d061e728-24eb-47ea-bb6e-6495722308ab-kube-api-access-fd7jq\") pod \"dnsmasq-dns-6d65c684ff-nnw4f\" (UID: \"d061e728-24eb-47ea-bb6e-6495722308ab\") " pod="openstack/dnsmasq-dns-6d65c684ff-nnw4f" Oct 05 08:13:21 crc kubenswrapper[4846]: I1005 08:13:21.511293 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d061e728-24eb-47ea-bb6e-6495722308ab-config\") pod \"dnsmasq-dns-6d65c684ff-nnw4f\" (UID: \"d061e728-24eb-47ea-bb6e-6495722308ab\") " pod="openstack/dnsmasq-dns-6d65c684ff-nnw4f" Oct 05 08:13:21 crc kubenswrapper[4846]: I1005 08:13:21.511325 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d061e728-24eb-47ea-bb6e-6495722308ab-ovsdbserver-nb\") pod \"dnsmasq-dns-6d65c684ff-nnw4f\" (UID: \"d061e728-24eb-47ea-bb6e-6495722308ab\") " pod="openstack/dnsmasq-dns-6d65c684ff-nnw4f" Oct 05 08:13:21 crc kubenswrapper[4846]: I1005 08:13:21.511442 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d061e728-24eb-47ea-bb6e-6495722308ab-dns-svc\") pod \"dnsmasq-dns-6d65c684ff-nnw4f\" (UID: \"d061e728-24eb-47ea-bb6e-6495722308ab\") " pod="openstack/dnsmasq-dns-6d65c684ff-nnw4f" Oct 05 08:13:21 crc kubenswrapper[4846]: I1005 08:13:21.511488 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d061e728-24eb-47ea-bb6e-6495722308ab-ovsdbserver-sb\") pod \"dnsmasq-dns-6d65c684ff-nnw4f\" (UID: \"d061e728-24eb-47ea-bb6e-6495722308ab\") " pod="openstack/dnsmasq-dns-6d65c684ff-nnw4f" Oct 05 08:13:21 crc kubenswrapper[4846]: I1005 08:13:21.511689 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-sf56p" Oct 05 08:13:21 crc kubenswrapper[4846]: I1005 08:13:21.520110 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 05 08:13:21 crc kubenswrapper[4846]: I1005 08:13:21.520359 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 05 08:13:21 crc kubenswrapper[4846]: I1005 08:13:21.520487 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-zhh8k" Oct 05 08:13:21 crc kubenswrapper[4846]: I1005 08:13:21.527562 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 05 08:13:21 crc kubenswrapper[4846]: I1005 08:13:21.538394 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-sf56p"] Oct 05 08:13:21 crc kubenswrapper[4846]: I1005 08:13:21.613487 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d061e728-24eb-47ea-bb6e-6495722308ab-ovsdbserver-sb\") pod \"dnsmasq-dns-6d65c684ff-nnw4f\" (UID: \"d061e728-24eb-47ea-bb6e-6495722308ab\") " pod="openstack/dnsmasq-dns-6d65c684ff-nnw4f" Oct 05 08:13:21 crc kubenswrapper[4846]: I1005 08:13:21.613573 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/b7472427-0989-4097-b523-5cc21d2f349b-fernet-keys\") pod \"keystone-bootstrap-sf56p\" (UID: \"b7472427-0989-4097-b523-5cc21d2f349b\") " pod="openstack/keystone-bootstrap-sf56p" Oct 05 08:13:21 crc kubenswrapper[4846]: I1005 08:13:21.613603 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b7472427-0989-4097-b523-5cc21d2f349b-combined-ca-bundle\") pod \"keystone-bootstrap-sf56p\" (UID: \"b7472427-0989-4097-b523-5cc21d2f349b\") " pod="openstack/keystone-bootstrap-sf56p" Oct 05 08:13:21 crc kubenswrapper[4846]: I1005 08:13:21.613654 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xgcv8\" (UniqueName: \"kubernetes.io/projected/b7472427-0989-4097-b523-5cc21d2f349b-kube-api-access-xgcv8\") pod \"keystone-bootstrap-sf56p\" (UID: \"b7472427-0989-4097-b523-5cc21d2f349b\") " pod="openstack/keystone-bootstrap-sf56p" Oct 05 08:13:21 crc kubenswrapper[4846]: I1005 08:13:21.613691 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fd7jq\" (UniqueName: \"kubernetes.io/projected/d061e728-24eb-47ea-bb6e-6495722308ab-kube-api-access-fd7jq\") pod \"dnsmasq-dns-6d65c684ff-nnw4f\" (UID: \"d061e728-24eb-47ea-bb6e-6495722308ab\") " pod="openstack/dnsmasq-dns-6d65c684ff-nnw4f" Oct 05 08:13:21 crc kubenswrapper[4846]: I1005 08:13:21.613718 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b7472427-0989-4097-b523-5cc21d2f349b-scripts\") pod \"keystone-bootstrap-sf56p\" (UID: \"b7472427-0989-4097-b523-5cc21d2f349b\") " pod="openstack/keystone-bootstrap-sf56p" Oct 05 08:13:21 crc kubenswrapper[4846]: I1005 08:13:21.613743 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b7472427-0989-4097-b523-5cc21d2f349b-config-data\") pod \"keystone-bootstrap-sf56p\" (UID: \"b7472427-0989-4097-b523-5cc21d2f349b\") " pod="openstack/keystone-bootstrap-sf56p" Oct 05 08:13:21 crc kubenswrapper[4846]: I1005 08:13:21.613771 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d061e728-24eb-47ea-bb6e-6495722308ab-config\") pod \"dnsmasq-dns-6d65c684ff-nnw4f\" (UID: \"d061e728-24eb-47ea-bb6e-6495722308ab\") " pod="openstack/dnsmasq-dns-6d65c684ff-nnw4f" Oct 05 08:13:21 crc kubenswrapper[4846]: I1005 08:13:21.613800 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d061e728-24eb-47ea-bb6e-6495722308ab-ovsdbserver-nb\") pod \"dnsmasq-dns-6d65c684ff-nnw4f\" (UID: \"d061e728-24eb-47ea-bb6e-6495722308ab\") " pod="openstack/dnsmasq-dns-6d65c684ff-nnw4f" Oct 05 08:13:21 crc kubenswrapper[4846]: I1005 08:13:21.613905 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/b7472427-0989-4097-b523-5cc21d2f349b-credential-keys\") pod \"keystone-bootstrap-sf56p\" (UID: \"b7472427-0989-4097-b523-5cc21d2f349b\") " pod="openstack/keystone-bootstrap-sf56p" Oct 05 08:13:21 crc kubenswrapper[4846]: I1005 08:13:21.613970 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d061e728-24eb-47ea-bb6e-6495722308ab-dns-svc\") pod \"dnsmasq-dns-6d65c684ff-nnw4f\" (UID: \"d061e728-24eb-47ea-bb6e-6495722308ab\") " pod="openstack/dnsmasq-dns-6d65c684ff-nnw4f" Oct 05 08:13:21 crc kubenswrapper[4846]: I1005 08:13:21.615207 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d061e728-24eb-47ea-bb6e-6495722308ab-ovsdbserver-sb\") pod \"dnsmasq-dns-6d65c684ff-nnw4f\" (UID: \"d061e728-24eb-47ea-bb6e-6495722308ab\") " pod="openstack/dnsmasq-dns-6d65c684ff-nnw4f" Oct 05 08:13:21 crc kubenswrapper[4846]: I1005 08:13:21.618486 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d061e728-24eb-47ea-bb6e-6495722308ab-dns-svc\") pod \"dnsmasq-dns-6d65c684ff-nnw4f\" (UID: \"d061e728-24eb-47ea-bb6e-6495722308ab\") " pod="openstack/dnsmasq-dns-6d65c684ff-nnw4f" Oct 05 08:13:21 crc kubenswrapper[4846]: I1005 08:13:21.618699 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d061e728-24eb-47ea-bb6e-6495722308ab-config\") pod \"dnsmasq-dns-6d65c684ff-nnw4f\" (UID: \"d061e728-24eb-47ea-bb6e-6495722308ab\") " pod="openstack/dnsmasq-dns-6d65c684ff-nnw4f" Oct 05 08:13:21 crc kubenswrapper[4846]: I1005 08:13:21.618759 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d061e728-24eb-47ea-bb6e-6495722308ab-ovsdbserver-nb\") pod \"dnsmasq-dns-6d65c684ff-nnw4f\" (UID: \"d061e728-24eb-47ea-bb6e-6495722308ab\") " pod="openstack/dnsmasq-dns-6d65c684ff-nnw4f" Oct 05 08:13:21 crc kubenswrapper[4846]: I1005 08:13:21.635466 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fd7jq\" (UniqueName: \"kubernetes.io/projected/d061e728-24eb-47ea-bb6e-6495722308ab-kube-api-access-fd7jq\") pod \"dnsmasq-dns-6d65c684ff-nnw4f\" (UID: \"d061e728-24eb-47ea-bb6e-6495722308ab\") " pod="openstack/dnsmasq-dns-6d65c684ff-nnw4f" Oct 05 08:13:21 crc kubenswrapper[4846]: I1005 08:13:21.716104 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/b7472427-0989-4097-b523-5cc21d2f349b-fernet-keys\") pod \"keystone-bootstrap-sf56p\" (UID: \"b7472427-0989-4097-b523-5cc21d2f349b\") " pod="openstack/keystone-bootstrap-sf56p" Oct 05 08:13:21 crc kubenswrapper[4846]: I1005 08:13:21.716165 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b7472427-0989-4097-b523-5cc21d2f349b-combined-ca-bundle\") pod \"keystone-bootstrap-sf56p\" (UID: \"b7472427-0989-4097-b523-5cc21d2f349b\") " pod="openstack/keystone-bootstrap-sf56p" Oct 05 08:13:21 crc kubenswrapper[4846]: I1005 08:13:21.716230 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xgcv8\" (UniqueName: \"kubernetes.io/projected/b7472427-0989-4097-b523-5cc21d2f349b-kube-api-access-xgcv8\") pod \"keystone-bootstrap-sf56p\" (UID: \"b7472427-0989-4097-b523-5cc21d2f349b\") " pod="openstack/keystone-bootstrap-sf56p" Oct 05 08:13:21 crc kubenswrapper[4846]: I1005 08:13:21.716270 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b7472427-0989-4097-b523-5cc21d2f349b-config-data\") pod \"keystone-bootstrap-sf56p\" (UID: \"b7472427-0989-4097-b523-5cc21d2f349b\") " pod="openstack/keystone-bootstrap-sf56p" Oct 05 08:13:21 crc kubenswrapper[4846]: I1005 08:13:21.716294 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b7472427-0989-4097-b523-5cc21d2f349b-scripts\") pod \"keystone-bootstrap-sf56p\" (UID: \"b7472427-0989-4097-b523-5cc21d2f349b\") " pod="openstack/keystone-bootstrap-sf56p" Oct 05 08:13:21 crc kubenswrapper[4846]: I1005 08:13:21.716368 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/b7472427-0989-4097-b523-5cc21d2f349b-credential-keys\") pod \"keystone-bootstrap-sf56p\" (UID: \"b7472427-0989-4097-b523-5cc21d2f349b\") " pod="openstack/keystone-bootstrap-sf56p" Oct 05 08:13:21 crc kubenswrapper[4846]: I1005 08:13:21.721276 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/b7472427-0989-4097-b523-5cc21d2f349b-credential-keys\") pod \"keystone-bootstrap-sf56p\" (UID: \"b7472427-0989-4097-b523-5cc21d2f349b\") " pod="openstack/keystone-bootstrap-sf56p" Oct 05 08:13:21 crc kubenswrapper[4846]: I1005 08:13:21.721613 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/b7472427-0989-4097-b523-5cc21d2f349b-fernet-keys\") pod \"keystone-bootstrap-sf56p\" (UID: \"b7472427-0989-4097-b523-5cc21d2f349b\") " pod="openstack/keystone-bootstrap-sf56p" Oct 05 08:13:21 crc kubenswrapper[4846]: I1005 08:13:21.727854 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b7472427-0989-4097-b523-5cc21d2f349b-combined-ca-bundle\") pod \"keystone-bootstrap-sf56p\" (UID: \"b7472427-0989-4097-b523-5cc21d2f349b\") " pod="openstack/keystone-bootstrap-sf56p" Oct 05 08:13:21 crc kubenswrapper[4846]: I1005 08:13:21.727960 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b7472427-0989-4097-b523-5cc21d2f349b-config-data\") pod \"keystone-bootstrap-sf56p\" (UID: \"b7472427-0989-4097-b523-5cc21d2f349b\") " pod="openstack/keystone-bootstrap-sf56p" Oct 05 08:13:21 crc kubenswrapper[4846]: I1005 08:13:21.728160 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b7472427-0989-4097-b523-5cc21d2f349b-scripts\") pod \"keystone-bootstrap-sf56p\" (UID: \"b7472427-0989-4097-b523-5cc21d2f349b\") " pod="openstack/keystone-bootstrap-sf56p" Oct 05 08:13:21 crc kubenswrapper[4846]: I1005 08:13:21.735418 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xgcv8\" (UniqueName: \"kubernetes.io/projected/b7472427-0989-4097-b523-5cc21d2f349b-kube-api-access-xgcv8\") pod \"keystone-bootstrap-sf56p\" (UID: \"b7472427-0989-4097-b523-5cc21d2f349b\") " pod="openstack/keystone-bootstrap-sf56p" Oct 05 08:13:21 crc kubenswrapper[4846]: I1005 08:13:21.736378 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6d65c684ff-nnw4f" Oct 05 08:13:21 crc kubenswrapper[4846]: I1005 08:13:21.835952 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-sf56p" Oct 05 08:13:22 crc kubenswrapper[4846]: I1005 08:13:22.167750 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6d65c684ff-nnw4f"] Oct 05 08:13:22 crc kubenswrapper[4846]: W1005 08:13:22.179614 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd061e728_24eb_47ea_bb6e_6495722308ab.slice/crio-f3cbc8645f852bba0537543aa8dbf4e6a6c407c928059bd0c9517e5e535a852c WatchSource:0}: Error finding container f3cbc8645f852bba0537543aa8dbf4e6a6c407c928059bd0c9517e5e535a852c: Status 404 returned error can't find the container with id f3cbc8645f852bba0537543aa8dbf4e6a6c407c928059bd0c9517e5e535a852c Oct 05 08:13:22 crc kubenswrapper[4846]: I1005 08:13:22.313683 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-sf56p"] Oct 05 08:13:22 crc kubenswrapper[4846]: W1005 08:13:22.323656 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb7472427_0989_4097_b523_5cc21d2f349b.slice/crio-b885daa245d6d4c59e4ca2dddd06b03fa7ba33f41f84eab47ef844f1ad9e8e8b WatchSource:0}: Error finding container b885daa245d6d4c59e4ca2dddd06b03fa7ba33f41f84eab47ef844f1ad9e8e8b: Status 404 returned error can't find the container with id b885daa245d6d4c59e4ca2dddd06b03fa7ba33f41f84eab47ef844f1ad9e8e8b Oct 05 08:13:23 crc kubenswrapper[4846]: I1005 08:13:23.199129 4846 generic.go:334] "Generic (PLEG): container finished" podID="d061e728-24eb-47ea-bb6e-6495722308ab" containerID="009ddf7e8407f1a461cf317efe56eadb1df61d474d03e628631d305c904dca0f" exitCode=0 Oct 05 08:13:23 crc kubenswrapper[4846]: I1005 08:13:23.199328 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d65c684ff-nnw4f" event={"ID":"d061e728-24eb-47ea-bb6e-6495722308ab","Type":"ContainerDied","Data":"009ddf7e8407f1a461cf317efe56eadb1df61d474d03e628631d305c904dca0f"} Oct 05 08:13:23 crc kubenswrapper[4846]: I1005 08:13:23.199636 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d65c684ff-nnw4f" event={"ID":"d061e728-24eb-47ea-bb6e-6495722308ab","Type":"ContainerStarted","Data":"f3cbc8645f852bba0537543aa8dbf4e6a6c407c928059bd0c9517e5e535a852c"} Oct 05 08:13:23 crc kubenswrapper[4846]: I1005 08:13:23.205780 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-sf56p" event={"ID":"b7472427-0989-4097-b523-5cc21d2f349b","Type":"ContainerStarted","Data":"9c3799f2b0fcce5160534a43d7e8520327c48bdb700c97480f49756f36d0b7fb"} Oct 05 08:13:23 crc kubenswrapper[4846]: I1005 08:13:23.205858 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-sf56p" event={"ID":"b7472427-0989-4097-b523-5cc21d2f349b","Type":"ContainerStarted","Data":"b885daa245d6d4c59e4ca2dddd06b03fa7ba33f41f84eab47ef844f1ad9e8e8b"} Oct 05 08:13:23 crc kubenswrapper[4846]: I1005 08:13:23.253389 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-sf56p" podStartSLOduration=2.253372583 podStartE2EDuration="2.253372583s" podCreationTimestamp="2025-10-05 08:13:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 08:13:23.251157513 +0000 UTC m=+5125.492010288" watchObservedRunningTime="2025-10-05 08:13:23.253372583 +0000 UTC m=+5125.494225358" Oct 05 08:13:24 crc kubenswrapper[4846]: I1005 08:13:24.254280 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d65c684ff-nnw4f" event={"ID":"d061e728-24eb-47ea-bb6e-6495722308ab","Type":"ContainerStarted","Data":"9dcbc8dca17e9726f55658375a4dbc3b9d27778f907b45a9136b9ab37b149e77"} Oct 05 08:13:24 crc kubenswrapper[4846]: I1005 08:13:24.254846 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6d65c684ff-nnw4f" Oct 05 08:13:24 crc kubenswrapper[4846]: I1005 08:13:24.288527 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6d65c684ff-nnw4f" podStartSLOduration=3.288504449 podStartE2EDuration="3.288504449s" podCreationTimestamp="2025-10-05 08:13:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 08:13:24.27614531 +0000 UTC m=+5126.516998095" watchObservedRunningTime="2025-10-05 08:13:24.288504449 +0000 UTC m=+5126.529357224" Oct 05 08:13:26 crc kubenswrapper[4846]: I1005 08:13:26.277237 4846 generic.go:334] "Generic (PLEG): container finished" podID="b7472427-0989-4097-b523-5cc21d2f349b" containerID="9c3799f2b0fcce5160534a43d7e8520327c48bdb700c97480f49756f36d0b7fb" exitCode=0 Oct 05 08:13:26 crc kubenswrapper[4846]: I1005 08:13:26.277381 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-sf56p" event={"ID":"b7472427-0989-4097-b523-5cc21d2f349b","Type":"ContainerDied","Data":"9c3799f2b0fcce5160534a43d7e8520327c48bdb700c97480f49756f36d0b7fb"} Oct 05 08:13:27 crc kubenswrapper[4846]: I1005 08:13:27.670923 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-sf56p" Oct 05 08:13:27 crc kubenswrapper[4846]: I1005 08:13:27.734731 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b7472427-0989-4097-b523-5cc21d2f349b-config-data\") pod \"b7472427-0989-4097-b523-5cc21d2f349b\" (UID: \"b7472427-0989-4097-b523-5cc21d2f349b\") " Oct 05 08:13:27 crc kubenswrapper[4846]: I1005 08:13:27.734787 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b7472427-0989-4097-b523-5cc21d2f349b-combined-ca-bundle\") pod \"b7472427-0989-4097-b523-5cc21d2f349b\" (UID: \"b7472427-0989-4097-b523-5cc21d2f349b\") " Oct 05 08:13:27 crc kubenswrapper[4846]: I1005 08:13:27.734840 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/b7472427-0989-4097-b523-5cc21d2f349b-credential-keys\") pod \"b7472427-0989-4097-b523-5cc21d2f349b\" (UID: \"b7472427-0989-4097-b523-5cc21d2f349b\") " Oct 05 08:13:27 crc kubenswrapper[4846]: I1005 08:13:27.734961 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/b7472427-0989-4097-b523-5cc21d2f349b-fernet-keys\") pod \"b7472427-0989-4097-b523-5cc21d2f349b\" (UID: \"b7472427-0989-4097-b523-5cc21d2f349b\") " Oct 05 08:13:27 crc kubenswrapper[4846]: I1005 08:13:27.735026 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b7472427-0989-4097-b523-5cc21d2f349b-scripts\") pod \"b7472427-0989-4097-b523-5cc21d2f349b\" (UID: \"b7472427-0989-4097-b523-5cc21d2f349b\") " Oct 05 08:13:27 crc kubenswrapper[4846]: I1005 08:13:27.735059 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xgcv8\" (UniqueName: \"kubernetes.io/projected/b7472427-0989-4097-b523-5cc21d2f349b-kube-api-access-xgcv8\") pod \"b7472427-0989-4097-b523-5cc21d2f349b\" (UID: \"b7472427-0989-4097-b523-5cc21d2f349b\") " Oct 05 08:13:27 crc kubenswrapper[4846]: I1005 08:13:27.740724 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b7472427-0989-4097-b523-5cc21d2f349b-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "b7472427-0989-4097-b523-5cc21d2f349b" (UID: "b7472427-0989-4097-b523-5cc21d2f349b"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:13:27 crc kubenswrapper[4846]: I1005 08:13:27.741197 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b7472427-0989-4097-b523-5cc21d2f349b-kube-api-access-xgcv8" (OuterVolumeSpecName: "kube-api-access-xgcv8") pod "b7472427-0989-4097-b523-5cc21d2f349b" (UID: "b7472427-0989-4097-b523-5cc21d2f349b"). InnerVolumeSpecName "kube-api-access-xgcv8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:13:27 crc kubenswrapper[4846]: I1005 08:13:27.742806 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b7472427-0989-4097-b523-5cc21d2f349b-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "b7472427-0989-4097-b523-5cc21d2f349b" (UID: "b7472427-0989-4097-b523-5cc21d2f349b"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:13:27 crc kubenswrapper[4846]: I1005 08:13:27.743471 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b7472427-0989-4097-b523-5cc21d2f349b-scripts" (OuterVolumeSpecName: "scripts") pod "b7472427-0989-4097-b523-5cc21d2f349b" (UID: "b7472427-0989-4097-b523-5cc21d2f349b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:13:27 crc kubenswrapper[4846]: I1005 08:13:27.762106 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b7472427-0989-4097-b523-5cc21d2f349b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b7472427-0989-4097-b523-5cc21d2f349b" (UID: "b7472427-0989-4097-b523-5cc21d2f349b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:13:27 crc kubenswrapper[4846]: I1005 08:13:27.777948 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b7472427-0989-4097-b523-5cc21d2f349b-config-data" (OuterVolumeSpecName: "config-data") pod "b7472427-0989-4097-b523-5cc21d2f349b" (UID: "b7472427-0989-4097-b523-5cc21d2f349b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:13:27 crc kubenswrapper[4846]: I1005 08:13:27.841404 4846 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b7472427-0989-4097-b523-5cc21d2f349b-scripts\") on node \"crc\" DevicePath \"\"" Oct 05 08:13:27 crc kubenswrapper[4846]: I1005 08:13:27.841466 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xgcv8\" (UniqueName: \"kubernetes.io/projected/b7472427-0989-4097-b523-5cc21d2f349b-kube-api-access-xgcv8\") on node \"crc\" DevicePath \"\"" Oct 05 08:13:27 crc kubenswrapper[4846]: I1005 08:13:27.841481 4846 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b7472427-0989-4097-b523-5cc21d2f349b-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 08:13:27 crc kubenswrapper[4846]: I1005 08:13:27.841493 4846 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b7472427-0989-4097-b523-5cc21d2f349b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 08:13:27 crc kubenswrapper[4846]: I1005 08:13:27.841503 4846 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/b7472427-0989-4097-b523-5cc21d2f349b-credential-keys\") on node \"crc\" DevicePath \"\"" Oct 05 08:13:27 crc kubenswrapper[4846]: I1005 08:13:27.841513 4846 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/b7472427-0989-4097-b523-5cc21d2f349b-fernet-keys\") on node \"crc\" DevicePath \"\"" Oct 05 08:13:28 crc kubenswrapper[4846]: I1005 08:13:28.297789 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-sf56p" event={"ID":"b7472427-0989-4097-b523-5cc21d2f349b","Type":"ContainerDied","Data":"b885daa245d6d4c59e4ca2dddd06b03fa7ba33f41f84eab47ef844f1ad9e8e8b"} Oct 05 08:13:28 crc kubenswrapper[4846]: I1005 08:13:28.297845 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b885daa245d6d4c59e4ca2dddd06b03fa7ba33f41f84eab47ef844f1ad9e8e8b" Oct 05 08:13:28 crc kubenswrapper[4846]: I1005 08:13:28.297894 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-sf56p" Oct 05 08:13:28 crc kubenswrapper[4846]: I1005 08:13:28.379964 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-sf56p"] Oct 05 08:13:28 crc kubenswrapper[4846]: I1005 08:13:28.385455 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-sf56p"] Oct 05 08:13:28 crc kubenswrapper[4846]: I1005 08:13:28.457100 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-fw275"] Oct 05 08:13:28 crc kubenswrapper[4846]: E1005 08:13:28.457448 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b7472427-0989-4097-b523-5cc21d2f349b" containerName="keystone-bootstrap" Oct 05 08:13:28 crc kubenswrapper[4846]: I1005 08:13:28.457465 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="b7472427-0989-4097-b523-5cc21d2f349b" containerName="keystone-bootstrap" Oct 05 08:13:28 crc kubenswrapper[4846]: I1005 08:13:28.457637 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="b7472427-0989-4097-b523-5cc21d2f349b" containerName="keystone-bootstrap" Oct 05 08:13:28 crc kubenswrapper[4846]: I1005 08:13:28.458165 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-fw275" Oct 05 08:13:28 crc kubenswrapper[4846]: I1005 08:13:28.460116 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 05 08:13:28 crc kubenswrapper[4846]: I1005 08:13:28.460211 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 05 08:13:28 crc kubenswrapper[4846]: I1005 08:13:28.460282 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 05 08:13:28 crc kubenswrapper[4846]: I1005 08:13:28.466630 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-zhh8k" Oct 05 08:13:28 crc kubenswrapper[4846]: I1005 08:13:28.485041 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-fw275"] Oct 05 08:13:28 crc kubenswrapper[4846]: I1005 08:13:28.518092 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b7472427-0989-4097-b523-5cc21d2f349b" path="/var/lib/kubelet/pods/b7472427-0989-4097-b523-5cc21d2f349b/volumes" Oct 05 08:13:28 crc kubenswrapper[4846]: I1005 08:13:28.553641 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/f49efe05-7f4e-4fb1-b523-ccdc75791838-fernet-keys\") pod \"keystone-bootstrap-fw275\" (UID: \"f49efe05-7f4e-4fb1-b523-ccdc75791838\") " pod="openstack/keystone-bootstrap-fw275" Oct 05 08:13:28 crc kubenswrapper[4846]: I1005 08:13:28.553725 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f49efe05-7f4e-4fb1-b523-ccdc75791838-combined-ca-bundle\") pod \"keystone-bootstrap-fw275\" (UID: \"f49efe05-7f4e-4fb1-b523-ccdc75791838\") " pod="openstack/keystone-bootstrap-fw275" Oct 05 08:13:28 crc kubenswrapper[4846]: I1005 08:13:28.553766 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h48r5\" (UniqueName: \"kubernetes.io/projected/f49efe05-7f4e-4fb1-b523-ccdc75791838-kube-api-access-h48r5\") pod \"keystone-bootstrap-fw275\" (UID: \"f49efe05-7f4e-4fb1-b523-ccdc75791838\") " pod="openstack/keystone-bootstrap-fw275" Oct 05 08:13:28 crc kubenswrapper[4846]: I1005 08:13:28.553837 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f49efe05-7f4e-4fb1-b523-ccdc75791838-config-data\") pod \"keystone-bootstrap-fw275\" (UID: \"f49efe05-7f4e-4fb1-b523-ccdc75791838\") " pod="openstack/keystone-bootstrap-fw275" Oct 05 08:13:28 crc kubenswrapper[4846]: I1005 08:13:28.553864 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f49efe05-7f4e-4fb1-b523-ccdc75791838-scripts\") pod \"keystone-bootstrap-fw275\" (UID: \"f49efe05-7f4e-4fb1-b523-ccdc75791838\") " pod="openstack/keystone-bootstrap-fw275" Oct 05 08:13:28 crc kubenswrapper[4846]: I1005 08:13:28.553921 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/f49efe05-7f4e-4fb1-b523-ccdc75791838-credential-keys\") pod \"keystone-bootstrap-fw275\" (UID: \"f49efe05-7f4e-4fb1-b523-ccdc75791838\") " pod="openstack/keystone-bootstrap-fw275" Oct 05 08:13:28 crc kubenswrapper[4846]: I1005 08:13:28.655314 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f49efe05-7f4e-4fb1-b523-ccdc75791838-config-data\") pod \"keystone-bootstrap-fw275\" (UID: \"f49efe05-7f4e-4fb1-b523-ccdc75791838\") " pod="openstack/keystone-bootstrap-fw275" Oct 05 08:13:28 crc kubenswrapper[4846]: I1005 08:13:28.655363 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f49efe05-7f4e-4fb1-b523-ccdc75791838-scripts\") pod \"keystone-bootstrap-fw275\" (UID: \"f49efe05-7f4e-4fb1-b523-ccdc75791838\") " pod="openstack/keystone-bootstrap-fw275" Oct 05 08:13:28 crc kubenswrapper[4846]: I1005 08:13:28.655418 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/f49efe05-7f4e-4fb1-b523-ccdc75791838-credential-keys\") pod \"keystone-bootstrap-fw275\" (UID: \"f49efe05-7f4e-4fb1-b523-ccdc75791838\") " pod="openstack/keystone-bootstrap-fw275" Oct 05 08:13:28 crc kubenswrapper[4846]: I1005 08:13:28.655467 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/f49efe05-7f4e-4fb1-b523-ccdc75791838-fernet-keys\") pod \"keystone-bootstrap-fw275\" (UID: \"f49efe05-7f4e-4fb1-b523-ccdc75791838\") " pod="openstack/keystone-bootstrap-fw275" Oct 05 08:13:28 crc kubenswrapper[4846]: I1005 08:13:28.655502 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f49efe05-7f4e-4fb1-b523-ccdc75791838-combined-ca-bundle\") pod \"keystone-bootstrap-fw275\" (UID: \"f49efe05-7f4e-4fb1-b523-ccdc75791838\") " pod="openstack/keystone-bootstrap-fw275" Oct 05 08:13:28 crc kubenswrapper[4846]: I1005 08:13:28.655523 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h48r5\" (UniqueName: \"kubernetes.io/projected/f49efe05-7f4e-4fb1-b523-ccdc75791838-kube-api-access-h48r5\") pod \"keystone-bootstrap-fw275\" (UID: \"f49efe05-7f4e-4fb1-b523-ccdc75791838\") " pod="openstack/keystone-bootstrap-fw275" Oct 05 08:13:28 crc kubenswrapper[4846]: I1005 08:13:28.661006 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/f49efe05-7f4e-4fb1-b523-ccdc75791838-fernet-keys\") pod \"keystone-bootstrap-fw275\" (UID: \"f49efe05-7f4e-4fb1-b523-ccdc75791838\") " pod="openstack/keystone-bootstrap-fw275" Oct 05 08:13:28 crc kubenswrapper[4846]: I1005 08:13:28.661120 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f49efe05-7f4e-4fb1-b523-ccdc75791838-config-data\") pod \"keystone-bootstrap-fw275\" (UID: \"f49efe05-7f4e-4fb1-b523-ccdc75791838\") " pod="openstack/keystone-bootstrap-fw275" Oct 05 08:13:28 crc kubenswrapper[4846]: I1005 08:13:28.662805 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f49efe05-7f4e-4fb1-b523-ccdc75791838-combined-ca-bundle\") pod \"keystone-bootstrap-fw275\" (UID: \"f49efe05-7f4e-4fb1-b523-ccdc75791838\") " pod="openstack/keystone-bootstrap-fw275" Oct 05 08:13:28 crc kubenswrapper[4846]: I1005 08:13:28.665776 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f49efe05-7f4e-4fb1-b523-ccdc75791838-scripts\") pod \"keystone-bootstrap-fw275\" (UID: \"f49efe05-7f4e-4fb1-b523-ccdc75791838\") " pod="openstack/keystone-bootstrap-fw275" Oct 05 08:13:28 crc kubenswrapper[4846]: I1005 08:13:28.666952 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/f49efe05-7f4e-4fb1-b523-ccdc75791838-credential-keys\") pod \"keystone-bootstrap-fw275\" (UID: \"f49efe05-7f4e-4fb1-b523-ccdc75791838\") " pod="openstack/keystone-bootstrap-fw275" Oct 05 08:13:28 crc kubenswrapper[4846]: I1005 08:13:28.687816 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h48r5\" (UniqueName: \"kubernetes.io/projected/f49efe05-7f4e-4fb1-b523-ccdc75791838-kube-api-access-h48r5\") pod \"keystone-bootstrap-fw275\" (UID: \"f49efe05-7f4e-4fb1-b523-ccdc75791838\") " pod="openstack/keystone-bootstrap-fw275" Oct 05 08:13:28 crc kubenswrapper[4846]: I1005 08:13:28.772953 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-fw275" Oct 05 08:13:29 crc kubenswrapper[4846]: I1005 08:13:29.299524 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-fw275"] Oct 05 08:13:29 crc kubenswrapper[4846]: W1005 08:13:29.309199 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf49efe05_7f4e_4fb1_b523_ccdc75791838.slice/crio-0e4c4357178382997ef3ad15209695ea4c074f51a537309debdd04670e0192cd WatchSource:0}: Error finding container 0e4c4357178382997ef3ad15209695ea4c074f51a537309debdd04670e0192cd: Status 404 returned error can't find the container with id 0e4c4357178382997ef3ad15209695ea4c074f51a537309debdd04670e0192cd Oct 05 08:13:30 crc kubenswrapper[4846]: I1005 08:13:30.319553 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-fw275" event={"ID":"f49efe05-7f4e-4fb1-b523-ccdc75791838","Type":"ContainerStarted","Data":"beaae9624261c763dbf004d0473c97d2be8428df0aeae1a54589c73d0a02db3f"} Oct 05 08:13:30 crc kubenswrapper[4846]: I1005 08:13:30.320034 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-fw275" event={"ID":"f49efe05-7f4e-4fb1-b523-ccdc75791838","Type":"ContainerStarted","Data":"0e4c4357178382997ef3ad15209695ea4c074f51a537309debdd04670e0192cd"} Oct 05 08:13:30 crc kubenswrapper[4846]: I1005 08:13:30.365701 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-fw275" podStartSLOduration=2.365660268 podStartE2EDuration="2.365660268s" podCreationTimestamp="2025-10-05 08:13:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 08:13:30.351371245 +0000 UTC m=+5132.592224050" watchObservedRunningTime="2025-10-05 08:13:30.365660268 +0000 UTC m=+5132.606513083" Oct 05 08:13:31 crc kubenswrapper[4846]: I1005 08:13:31.738065 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6d65c684ff-nnw4f" Oct 05 08:13:31 crc kubenswrapper[4846]: I1005 08:13:31.839257 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-697c78d8d5-2f9zd"] Oct 05 08:13:31 crc kubenswrapper[4846]: I1005 08:13:31.839679 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-697c78d8d5-2f9zd" podUID="0a9ba111-4d92-4903-8c97-f386e3c29d5c" containerName="dnsmasq-dns" containerID="cri-o://6ab8d5d38e9026b25124afb01ed5109b8d460824f788775d79e9a584be3f2edd" gracePeriod=10 Oct 05 08:13:32 crc kubenswrapper[4846]: I1005 08:13:32.292324 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-697c78d8d5-2f9zd" Oct 05 08:13:32 crc kubenswrapper[4846]: I1005 08:13:32.336992 4846 generic.go:334] "Generic (PLEG): container finished" podID="0a9ba111-4d92-4903-8c97-f386e3c29d5c" containerID="6ab8d5d38e9026b25124afb01ed5109b8d460824f788775d79e9a584be3f2edd" exitCode=0 Oct 05 08:13:32 crc kubenswrapper[4846]: I1005 08:13:32.337043 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-697c78d8d5-2f9zd" event={"ID":"0a9ba111-4d92-4903-8c97-f386e3c29d5c","Type":"ContainerDied","Data":"6ab8d5d38e9026b25124afb01ed5109b8d460824f788775d79e9a584be3f2edd"} Oct 05 08:13:32 crc kubenswrapper[4846]: I1005 08:13:32.337073 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-697c78d8d5-2f9zd" Oct 05 08:13:32 crc kubenswrapper[4846]: I1005 08:13:32.337100 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-697c78d8d5-2f9zd" event={"ID":"0a9ba111-4d92-4903-8c97-f386e3c29d5c","Type":"ContainerDied","Data":"b07264268023104a93afeaff29fdf3393f87dbf3c85b0e9124f603946f0eb3ac"} Oct 05 08:13:32 crc kubenswrapper[4846]: I1005 08:13:32.337125 4846 scope.go:117] "RemoveContainer" containerID="6ab8d5d38e9026b25124afb01ed5109b8d460824f788775d79e9a584be3f2edd" Oct 05 08:13:32 crc kubenswrapper[4846]: I1005 08:13:32.403163 4846 scope.go:117] "RemoveContainer" containerID="8b575118c9b34bd243118cf9a593a887bd7ab1577a66cd225b525d2ff127ac68" Oct 05 08:13:32 crc kubenswrapper[4846]: I1005 08:13:32.426593 4846 scope.go:117] "RemoveContainer" containerID="6ab8d5d38e9026b25124afb01ed5109b8d460824f788775d79e9a584be3f2edd" Oct 05 08:13:32 crc kubenswrapper[4846]: I1005 08:13:32.427036 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0a9ba111-4d92-4903-8c97-f386e3c29d5c-config\") pod \"0a9ba111-4d92-4903-8c97-f386e3c29d5c\" (UID: \"0a9ba111-4d92-4903-8c97-f386e3c29d5c\") " Oct 05 08:13:32 crc kubenswrapper[4846]: I1005 08:13:32.427115 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zfcc2\" (UniqueName: \"kubernetes.io/projected/0a9ba111-4d92-4903-8c97-f386e3c29d5c-kube-api-access-zfcc2\") pod \"0a9ba111-4d92-4903-8c97-f386e3c29d5c\" (UID: \"0a9ba111-4d92-4903-8c97-f386e3c29d5c\") " Oct 05 08:13:32 crc kubenswrapper[4846]: I1005 08:13:32.427171 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0a9ba111-4d92-4903-8c97-f386e3c29d5c-ovsdbserver-nb\") pod \"0a9ba111-4d92-4903-8c97-f386e3c29d5c\" (UID: \"0a9ba111-4d92-4903-8c97-f386e3c29d5c\") " Oct 05 08:13:32 crc kubenswrapper[4846]: I1005 08:13:32.427288 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0a9ba111-4d92-4903-8c97-f386e3c29d5c-dns-svc\") pod \"0a9ba111-4d92-4903-8c97-f386e3c29d5c\" (UID: \"0a9ba111-4d92-4903-8c97-f386e3c29d5c\") " Oct 05 08:13:32 crc kubenswrapper[4846]: I1005 08:13:32.427331 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0a9ba111-4d92-4903-8c97-f386e3c29d5c-ovsdbserver-sb\") pod \"0a9ba111-4d92-4903-8c97-f386e3c29d5c\" (UID: \"0a9ba111-4d92-4903-8c97-f386e3c29d5c\") " Oct 05 08:13:32 crc kubenswrapper[4846]: E1005 08:13:32.427108 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6ab8d5d38e9026b25124afb01ed5109b8d460824f788775d79e9a584be3f2edd\": container with ID starting with 6ab8d5d38e9026b25124afb01ed5109b8d460824f788775d79e9a584be3f2edd not found: ID does not exist" containerID="6ab8d5d38e9026b25124afb01ed5109b8d460824f788775d79e9a584be3f2edd" Oct 05 08:13:32 crc kubenswrapper[4846]: I1005 08:13:32.427860 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6ab8d5d38e9026b25124afb01ed5109b8d460824f788775d79e9a584be3f2edd"} err="failed to get container status \"6ab8d5d38e9026b25124afb01ed5109b8d460824f788775d79e9a584be3f2edd\": rpc error: code = NotFound desc = could not find container \"6ab8d5d38e9026b25124afb01ed5109b8d460824f788775d79e9a584be3f2edd\": container with ID starting with 6ab8d5d38e9026b25124afb01ed5109b8d460824f788775d79e9a584be3f2edd not found: ID does not exist" Oct 05 08:13:32 crc kubenswrapper[4846]: I1005 08:13:32.427887 4846 scope.go:117] "RemoveContainer" containerID="8b575118c9b34bd243118cf9a593a887bd7ab1577a66cd225b525d2ff127ac68" Oct 05 08:13:32 crc kubenswrapper[4846]: E1005 08:13:32.431987 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8b575118c9b34bd243118cf9a593a887bd7ab1577a66cd225b525d2ff127ac68\": container with ID starting with 8b575118c9b34bd243118cf9a593a887bd7ab1577a66cd225b525d2ff127ac68 not found: ID does not exist" containerID="8b575118c9b34bd243118cf9a593a887bd7ab1577a66cd225b525d2ff127ac68" Oct 05 08:13:32 crc kubenswrapper[4846]: I1005 08:13:32.432025 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8b575118c9b34bd243118cf9a593a887bd7ab1577a66cd225b525d2ff127ac68"} err="failed to get container status \"8b575118c9b34bd243118cf9a593a887bd7ab1577a66cd225b525d2ff127ac68\": rpc error: code = NotFound desc = could not find container \"8b575118c9b34bd243118cf9a593a887bd7ab1577a66cd225b525d2ff127ac68\": container with ID starting with 8b575118c9b34bd243118cf9a593a887bd7ab1577a66cd225b525d2ff127ac68 not found: ID does not exist" Oct 05 08:13:32 crc kubenswrapper[4846]: I1005 08:13:32.434829 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0a9ba111-4d92-4903-8c97-f386e3c29d5c-kube-api-access-zfcc2" (OuterVolumeSpecName: "kube-api-access-zfcc2") pod "0a9ba111-4d92-4903-8c97-f386e3c29d5c" (UID: "0a9ba111-4d92-4903-8c97-f386e3c29d5c"). InnerVolumeSpecName "kube-api-access-zfcc2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:13:32 crc kubenswrapper[4846]: I1005 08:13:32.484090 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0a9ba111-4d92-4903-8c97-f386e3c29d5c-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "0a9ba111-4d92-4903-8c97-f386e3c29d5c" (UID: "0a9ba111-4d92-4903-8c97-f386e3c29d5c"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 08:13:32 crc kubenswrapper[4846]: I1005 08:13:32.485380 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0a9ba111-4d92-4903-8c97-f386e3c29d5c-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "0a9ba111-4d92-4903-8c97-f386e3c29d5c" (UID: "0a9ba111-4d92-4903-8c97-f386e3c29d5c"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 08:13:32 crc kubenswrapper[4846]: I1005 08:13:32.487328 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0a9ba111-4d92-4903-8c97-f386e3c29d5c-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "0a9ba111-4d92-4903-8c97-f386e3c29d5c" (UID: "0a9ba111-4d92-4903-8c97-f386e3c29d5c"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 08:13:32 crc kubenswrapper[4846]: I1005 08:13:32.490293 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0a9ba111-4d92-4903-8c97-f386e3c29d5c-config" (OuterVolumeSpecName: "config") pod "0a9ba111-4d92-4903-8c97-f386e3c29d5c" (UID: "0a9ba111-4d92-4903-8c97-f386e3c29d5c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 08:13:32 crc kubenswrapper[4846]: I1005 08:13:32.534822 4846 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0a9ba111-4d92-4903-8c97-f386e3c29d5c-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 05 08:13:32 crc kubenswrapper[4846]: I1005 08:13:32.534871 4846 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0a9ba111-4d92-4903-8c97-f386e3c29d5c-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 05 08:13:32 crc kubenswrapper[4846]: I1005 08:13:32.534885 4846 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0a9ba111-4d92-4903-8c97-f386e3c29d5c-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 05 08:13:32 crc kubenswrapper[4846]: I1005 08:13:32.534905 4846 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0a9ba111-4d92-4903-8c97-f386e3c29d5c-config\") on node \"crc\" DevicePath \"\"" Oct 05 08:13:32 crc kubenswrapper[4846]: I1005 08:13:32.534918 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zfcc2\" (UniqueName: \"kubernetes.io/projected/0a9ba111-4d92-4903-8c97-f386e3c29d5c-kube-api-access-zfcc2\") on node \"crc\" DevicePath \"\"" Oct 05 08:13:32 crc kubenswrapper[4846]: I1005 08:13:32.671972 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-697c78d8d5-2f9zd"] Oct 05 08:13:32 crc kubenswrapper[4846]: I1005 08:13:32.681848 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-697c78d8d5-2f9zd"] Oct 05 08:13:33 crc kubenswrapper[4846]: I1005 08:13:33.354540 4846 generic.go:334] "Generic (PLEG): container finished" podID="f49efe05-7f4e-4fb1-b523-ccdc75791838" containerID="beaae9624261c763dbf004d0473c97d2be8428df0aeae1a54589c73d0a02db3f" exitCode=0 Oct 05 08:13:33 crc kubenswrapper[4846]: I1005 08:13:33.354614 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-fw275" event={"ID":"f49efe05-7f4e-4fb1-b523-ccdc75791838","Type":"ContainerDied","Data":"beaae9624261c763dbf004d0473c97d2be8428df0aeae1a54589c73d0a02db3f"} Oct 05 08:13:33 crc kubenswrapper[4846]: I1005 08:13:33.498348 4846 scope.go:117] "RemoveContainer" containerID="8fa72be735dff8e623a4a4958f452896ef79e72dd44c2d714a3019187462de47" Oct 05 08:13:33 crc kubenswrapper[4846]: E1005 08:13:33.498990 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 08:13:34 crc kubenswrapper[4846]: I1005 08:13:34.510519 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0a9ba111-4d92-4903-8c97-f386e3c29d5c" path="/var/lib/kubelet/pods/0a9ba111-4d92-4903-8c97-f386e3c29d5c/volumes" Oct 05 08:13:34 crc kubenswrapper[4846]: I1005 08:13:34.732139 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-fw275" Oct 05 08:13:34 crc kubenswrapper[4846]: I1005 08:13:34.882976 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f49efe05-7f4e-4fb1-b523-ccdc75791838-scripts\") pod \"f49efe05-7f4e-4fb1-b523-ccdc75791838\" (UID: \"f49efe05-7f4e-4fb1-b523-ccdc75791838\") " Oct 05 08:13:34 crc kubenswrapper[4846]: I1005 08:13:34.883368 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f49efe05-7f4e-4fb1-b523-ccdc75791838-config-data\") pod \"f49efe05-7f4e-4fb1-b523-ccdc75791838\" (UID: \"f49efe05-7f4e-4fb1-b523-ccdc75791838\") " Oct 05 08:13:34 crc kubenswrapper[4846]: I1005 08:13:34.883524 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h48r5\" (UniqueName: \"kubernetes.io/projected/f49efe05-7f4e-4fb1-b523-ccdc75791838-kube-api-access-h48r5\") pod \"f49efe05-7f4e-4fb1-b523-ccdc75791838\" (UID: \"f49efe05-7f4e-4fb1-b523-ccdc75791838\") " Oct 05 08:13:34 crc kubenswrapper[4846]: I1005 08:13:34.883578 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/f49efe05-7f4e-4fb1-b523-ccdc75791838-credential-keys\") pod \"f49efe05-7f4e-4fb1-b523-ccdc75791838\" (UID: \"f49efe05-7f4e-4fb1-b523-ccdc75791838\") " Oct 05 08:13:34 crc kubenswrapper[4846]: I1005 08:13:34.883602 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f49efe05-7f4e-4fb1-b523-ccdc75791838-combined-ca-bundle\") pod \"f49efe05-7f4e-4fb1-b523-ccdc75791838\" (UID: \"f49efe05-7f4e-4fb1-b523-ccdc75791838\") " Oct 05 08:13:34 crc kubenswrapper[4846]: I1005 08:13:34.883634 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/f49efe05-7f4e-4fb1-b523-ccdc75791838-fernet-keys\") pod \"f49efe05-7f4e-4fb1-b523-ccdc75791838\" (UID: \"f49efe05-7f4e-4fb1-b523-ccdc75791838\") " Oct 05 08:13:34 crc kubenswrapper[4846]: I1005 08:13:34.894363 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f49efe05-7f4e-4fb1-b523-ccdc75791838-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "f49efe05-7f4e-4fb1-b523-ccdc75791838" (UID: "f49efe05-7f4e-4fb1-b523-ccdc75791838"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:13:34 crc kubenswrapper[4846]: I1005 08:13:34.899472 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f49efe05-7f4e-4fb1-b523-ccdc75791838-scripts" (OuterVolumeSpecName: "scripts") pod "f49efe05-7f4e-4fb1-b523-ccdc75791838" (UID: "f49efe05-7f4e-4fb1-b523-ccdc75791838"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:13:34 crc kubenswrapper[4846]: I1005 08:13:34.906712 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f49efe05-7f4e-4fb1-b523-ccdc75791838-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "f49efe05-7f4e-4fb1-b523-ccdc75791838" (UID: "f49efe05-7f4e-4fb1-b523-ccdc75791838"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:13:34 crc kubenswrapper[4846]: I1005 08:13:34.916424 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f49efe05-7f4e-4fb1-b523-ccdc75791838-kube-api-access-h48r5" (OuterVolumeSpecName: "kube-api-access-h48r5") pod "f49efe05-7f4e-4fb1-b523-ccdc75791838" (UID: "f49efe05-7f4e-4fb1-b523-ccdc75791838"). InnerVolumeSpecName "kube-api-access-h48r5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:13:34 crc kubenswrapper[4846]: I1005 08:13:34.977775 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f49efe05-7f4e-4fb1-b523-ccdc75791838-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f49efe05-7f4e-4fb1-b523-ccdc75791838" (UID: "f49efe05-7f4e-4fb1-b523-ccdc75791838"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:13:34 crc kubenswrapper[4846]: I1005 08:13:34.986226 4846 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f49efe05-7f4e-4fb1-b523-ccdc75791838-scripts\") on node \"crc\" DevicePath \"\"" Oct 05 08:13:34 crc kubenswrapper[4846]: I1005 08:13:34.986257 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h48r5\" (UniqueName: \"kubernetes.io/projected/f49efe05-7f4e-4fb1-b523-ccdc75791838-kube-api-access-h48r5\") on node \"crc\" DevicePath \"\"" Oct 05 08:13:34 crc kubenswrapper[4846]: I1005 08:13:34.986267 4846 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/f49efe05-7f4e-4fb1-b523-ccdc75791838-credential-keys\") on node \"crc\" DevicePath \"\"" Oct 05 08:13:34 crc kubenswrapper[4846]: I1005 08:13:34.986275 4846 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f49efe05-7f4e-4fb1-b523-ccdc75791838-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 08:13:34 crc kubenswrapper[4846]: I1005 08:13:34.986286 4846 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/f49efe05-7f4e-4fb1-b523-ccdc75791838-fernet-keys\") on node \"crc\" DevicePath \"\"" Oct 05 08:13:35 crc kubenswrapper[4846]: I1005 08:13:35.009508 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f49efe05-7f4e-4fb1-b523-ccdc75791838-config-data" (OuterVolumeSpecName: "config-data") pod "f49efe05-7f4e-4fb1-b523-ccdc75791838" (UID: "f49efe05-7f4e-4fb1-b523-ccdc75791838"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:13:35 crc kubenswrapper[4846]: I1005 08:13:35.087437 4846 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f49efe05-7f4e-4fb1-b523-ccdc75791838-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 08:13:35 crc kubenswrapper[4846]: I1005 08:13:35.373824 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-fw275" event={"ID":"f49efe05-7f4e-4fb1-b523-ccdc75791838","Type":"ContainerDied","Data":"0e4c4357178382997ef3ad15209695ea4c074f51a537309debdd04670e0192cd"} Oct 05 08:13:35 crc kubenswrapper[4846]: I1005 08:13:35.373869 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0e4c4357178382997ef3ad15209695ea4c074f51a537309debdd04670e0192cd" Oct 05 08:13:35 crc kubenswrapper[4846]: I1005 08:13:35.373932 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-fw275" Oct 05 08:13:35 crc kubenswrapper[4846]: I1005 08:13:35.471782 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-575454b8f7-64pz9"] Oct 05 08:13:35 crc kubenswrapper[4846]: E1005 08:13:35.472261 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f49efe05-7f4e-4fb1-b523-ccdc75791838" containerName="keystone-bootstrap" Oct 05 08:13:35 crc kubenswrapper[4846]: I1005 08:13:35.472286 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="f49efe05-7f4e-4fb1-b523-ccdc75791838" containerName="keystone-bootstrap" Oct 05 08:13:35 crc kubenswrapper[4846]: E1005 08:13:35.472324 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0a9ba111-4d92-4903-8c97-f386e3c29d5c" containerName="init" Oct 05 08:13:35 crc kubenswrapper[4846]: I1005 08:13:35.472335 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="0a9ba111-4d92-4903-8c97-f386e3c29d5c" containerName="init" Oct 05 08:13:35 crc kubenswrapper[4846]: E1005 08:13:35.472350 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0a9ba111-4d92-4903-8c97-f386e3c29d5c" containerName="dnsmasq-dns" Oct 05 08:13:35 crc kubenswrapper[4846]: I1005 08:13:35.472359 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="0a9ba111-4d92-4903-8c97-f386e3c29d5c" containerName="dnsmasq-dns" Oct 05 08:13:35 crc kubenswrapper[4846]: I1005 08:13:35.472558 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="0a9ba111-4d92-4903-8c97-f386e3c29d5c" containerName="dnsmasq-dns" Oct 05 08:13:35 crc kubenswrapper[4846]: I1005 08:13:35.472592 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="f49efe05-7f4e-4fb1-b523-ccdc75791838" containerName="keystone-bootstrap" Oct 05 08:13:35 crc kubenswrapper[4846]: I1005 08:13:35.473229 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-575454b8f7-64pz9" Oct 05 08:13:35 crc kubenswrapper[4846]: I1005 08:13:35.476469 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 05 08:13:35 crc kubenswrapper[4846]: I1005 08:13:35.476642 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 05 08:13:35 crc kubenswrapper[4846]: I1005 08:13:35.476785 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-zhh8k" Oct 05 08:13:35 crc kubenswrapper[4846]: I1005 08:13:35.476956 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-internal-svc" Oct 05 08:13:35 crc kubenswrapper[4846]: I1005 08:13:35.487634 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-575454b8f7-64pz9"] Oct 05 08:13:35 crc kubenswrapper[4846]: I1005 08:13:35.487828 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 05 08:13:35 crc kubenswrapper[4846]: I1005 08:13:35.488092 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-public-svc" Oct 05 08:13:35 crc kubenswrapper[4846]: I1005 08:13:35.602686 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/617bdbc8-22e4-43b8-b344-7a76842c40e8-combined-ca-bundle\") pod \"keystone-575454b8f7-64pz9\" (UID: \"617bdbc8-22e4-43b8-b344-7a76842c40e8\") " pod="openstack/keystone-575454b8f7-64pz9" Oct 05 08:13:35 crc kubenswrapper[4846]: I1005 08:13:35.602783 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/617bdbc8-22e4-43b8-b344-7a76842c40e8-credential-keys\") pod \"keystone-575454b8f7-64pz9\" (UID: \"617bdbc8-22e4-43b8-b344-7a76842c40e8\") " pod="openstack/keystone-575454b8f7-64pz9" Oct 05 08:13:35 crc kubenswrapper[4846]: I1005 08:13:35.602837 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/617bdbc8-22e4-43b8-b344-7a76842c40e8-public-tls-certs\") pod \"keystone-575454b8f7-64pz9\" (UID: \"617bdbc8-22e4-43b8-b344-7a76842c40e8\") " pod="openstack/keystone-575454b8f7-64pz9" Oct 05 08:13:35 crc kubenswrapper[4846]: I1005 08:13:35.602938 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/617bdbc8-22e4-43b8-b344-7a76842c40e8-config-data\") pod \"keystone-575454b8f7-64pz9\" (UID: \"617bdbc8-22e4-43b8-b344-7a76842c40e8\") " pod="openstack/keystone-575454b8f7-64pz9" Oct 05 08:13:35 crc kubenswrapper[4846]: I1005 08:13:35.602968 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/617bdbc8-22e4-43b8-b344-7a76842c40e8-internal-tls-certs\") pod \"keystone-575454b8f7-64pz9\" (UID: \"617bdbc8-22e4-43b8-b344-7a76842c40e8\") " pod="openstack/keystone-575454b8f7-64pz9" Oct 05 08:13:35 crc kubenswrapper[4846]: I1005 08:13:35.603050 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tlrqz\" (UniqueName: \"kubernetes.io/projected/617bdbc8-22e4-43b8-b344-7a76842c40e8-kube-api-access-tlrqz\") pod \"keystone-575454b8f7-64pz9\" (UID: \"617bdbc8-22e4-43b8-b344-7a76842c40e8\") " pod="openstack/keystone-575454b8f7-64pz9" Oct 05 08:13:35 crc kubenswrapper[4846]: I1005 08:13:35.603105 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/617bdbc8-22e4-43b8-b344-7a76842c40e8-scripts\") pod \"keystone-575454b8f7-64pz9\" (UID: \"617bdbc8-22e4-43b8-b344-7a76842c40e8\") " pod="openstack/keystone-575454b8f7-64pz9" Oct 05 08:13:35 crc kubenswrapper[4846]: I1005 08:13:35.603133 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/617bdbc8-22e4-43b8-b344-7a76842c40e8-fernet-keys\") pod \"keystone-575454b8f7-64pz9\" (UID: \"617bdbc8-22e4-43b8-b344-7a76842c40e8\") " pod="openstack/keystone-575454b8f7-64pz9" Oct 05 08:13:35 crc kubenswrapper[4846]: I1005 08:13:35.704255 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/617bdbc8-22e4-43b8-b344-7a76842c40e8-scripts\") pod \"keystone-575454b8f7-64pz9\" (UID: \"617bdbc8-22e4-43b8-b344-7a76842c40e8\") " pod="openstack/keystone-575454b8f7-64pz9" Oct 05 08:13:35 crc kubenswrapper[4846]: I1005 08:13:35.704301 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/617bdbc8-22e4-43b8-b344-7a76842c40e8-fernet-keys\") pod \"keystone-575454b8f7-64pz9\" (UID: \"617bdbc8-22e4-43b8-b344-7a76842c40e8\") " pod="openstack/keystone-575454b8f7-64pz9" Oct 05 08:13:35 crc kubenswrapper[4846]: I1005 08:13:35.704362 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/617bdbc8-22e4-43b8-b344-7a76842c40e8-combined-ca-bundle\") pod \"keystone-575454b8f7-64pz9\" (UID: \"617bdbc8-22e4-43b8-b344-7a76842c40e8\") " pod="openstack/keystone-575454b8f7-64pz9" Oct 05 08:13:35 crc kubenswrapper[4846]: I1005 08:13:35.704386 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/617bdbc8-22e4-43b8-b344-7a76842c40e8-credential-keys\") pod \"keystone-575454b8f7-64pz9\" (UID: \"617bdbc8-22e4-43b8-b344-7a76842c40e8\") " pod="openstack/keystone-575454b8f7-64pz9" Oct 05 08:13:35 crc kubenswrapper[4846]: I1005 08:13:35.704413 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/617bdbc8-22e4-43b8-b344-7a76842c40e8-public-tls-certs\") pod \"keystone-575454b8f7-64pz9\" (UID: \"617bdbc8-22e4-43b8-b344-7a76842c40e8\") " pod="openstack/keystone-575454b8f7-64pz9" Oct 05 08:13:35 crc kubenswrapper[4846]: I1005 08:13:35.704459 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/617bdbc8-22e4-43b8-b344-7a76842c40e8-config-data\") pod \"keystone-575454b8f7-64pz9\" (UID: \"617bdbc8-22e4-43b8-b344-7a76842c40e8\") " pod="openstack/keystone-575454b8f7-64pz9" Oct 05 08:13:35 crc kubenswrapper[4846]: I1005 08:13:35.704479 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/617bdbc8-22e4-43b8-b344-7a76842c40e8-internal-tls-certs\") pod \"keystone-575454b8f7-64pz9\" (UID: \"617bdbc8-22e4-43b8-b344-7a76842c40e8\") " pod="openstack/keystone-575454b8f7-64pz9" Oct 05 08:13:35 crc kubenswrapper[4846]: I1005 08:13:35.704509 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tlrqz\" (UniqueName: \"kubernetes.io/projected/617bdbc8-22e4-43b8-b344-7a76842c40e8-kube-api-access-tlrqz\") pod \"keystone-575454b8f7-64pz9\" (UID: \"617bdbc8-22e4-43b8-b344-7a76842c40e8\") " pod="openstack/keystone-575454b8f7-64pz9" Oct 05 08:13:35 crc kubenswrapper[4846]: I1005 08:13:35.710110 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/617bdbc8-22e4-43b8-b344-7a76842c40e8-combined-ca-bundle\") pod \"keystone-575454b8f7-64pz9\" (UID: \"617bdbc8-22e4-43b8-b344-7a76842c40e8\") " pod="openstack/keystone-575454b8f7-64pz9" Oct 05 08:13:35 crc kubenswrapper[4846]: I1005 08:13:35.710994 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/617bdbc8-22e4-43b8-b344-7a76842c40e8-config-data\") pod \"keystone-575454b8f7-64pz9\" (UID: \"617bdbc8-22e4-43b8-b344-7a76842c40e8\") " pod="openstack/keystone-575454b8f7-64pz9" Oct 05 08:13:35 crc kubenswrapper[4846]: I1005 08:13:35.712796 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/617bdbc8-22e4-43b8-b344-7a76842c40e8-fernet-keys\") pod \"keystone-575454b8f7-64pz9\" (UID: \"617bdbc8-22e4-43b8-b344-7a76842c40e8\") " pod="openstack/keystone-575454b8f7-64pz9" Oct 05 08:13:35 crc kubenswrapper[4846]: I1005 08:13:35.712806 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/617bdbc8-22e4-43b8-b344-7a76842c40e8-credential-keys\") pod \"keystone-575454b8f7-64pz9\" (UID: \"617bdbc8-22e4-43b8-b344-7a76842c40e8\") " pod="openstack/keystone-575454b8f7-64pz9" Oct 05 08:13:35 crc kubenswrapper[4846]: I1005 08:13:35.712926 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/617bdbc8-22e4-43b8-b344-7a76842c40e8-public-tls-certs\") pod \"keystone-575454b8f7-64pz9\" (UID: \"617bdbc8-22e4-43b8-b344-7a76842c40e8\") " pod="openstack/keystone-575454b8f7-64pz9" Oct 05 08:13:35 crc kubenswrapper[4846]: I1005 08:13:35.713067 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/617bdbc8-22e4-43b8-b344-7a76842c40e8-scripts\") pod \"keystone-575454b8f7-64pz9\" (UID: \"617bdbc8-22e4-43b8-b344-7a76842c40e8\") " pod="openstack/keystone-575454b8f7-64pz9" Oct 05 08:13:35 crc kubenswrapper[4846]: I1005 08:13:35.715097 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/617bdbc8-22e4-43b8-b344-7a76842c40e8-internal-tls-certs\") pod \"keystone-575454b8f7-64pz9\" (UID: \"617bdbc8-22e4-43b8-b344-7a76842c40e8\") " pod="openstack/keystone-575454b8f7-64pz9" Oct 05 08:13:35 crc kubenswrapper[4846]: I1005 08:13:35.723012 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tlrqz\" (UniqueName: \"kubernetes.io/projected/617bdbc8-22e4-43b8-b344-7a76842c40e8-kube-api-access-tlrqz\") pod \"keystone-575454b8f7-64pz9\" (UID: \"617bdbc8-22e4-43b8-b344-7a76842c40e8\") " pod="openstack/keystone-575454b8f7-64pz9" Oct 05 08:13:35 crc kubenswrapper[4846]: I1005 08:13:35.839872 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-575454b8f7-64pz9" Oct 05 08:13:36 crc kubenswrapper[4846]: I1005 08:13:36.329444 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-575454b8f7-64pz9"] Oct 05 08:13:36 crc kubenswrapper[4846]: W1005 08:13:36.352014 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod617bdbc8_22e4_43b8_b344_7a76842c40e8.slice/crio-7d0c33f28973d64a749233db8d26b4c81f4a5bbb2f643ee54f0659510cda51ff WatchSource:0}: Error finding container 7d0c33f28973d64a749233db8d26b4c81f4a5bbb2f643ee54f0659510cda51ff: Status 404 returned error can't find the container with id 7d0c33f28973d64a749233db8d26b4c81f4a5bbb2f643ee54f0659510cda51ff Oct 05 08:13:36 crc kubenswrapper[4846]: I1005 08:13:36.384712 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-575454b8f7-64pz9" event={"ID":"617bdbc8-22e4-43b8-b344-7a76842c40e8","Type":"ContainerStarted","Data":"7d0c33f28973d64a749233db8d26b4c81f4a5bbb2f643ee54f0659510cda51ff"} Oct 05 08:13:37 crc kubenswrapper[4846]: I1005 08:13:37.396892 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-575454b8f7-64pz9" event={"ID":"617bdbc8-22e4-43b8-b344-7a76842c40e8","Type":"ContainerStarted","Data":"043bf55932bd1578ceca7b3308add47d09f3da3e794d2034efc3722aa5e4679b"} Oct 05 08:13:37 crc kubenswrapper[4846]: I1005 08:13:37.398275 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/keystone-575454b8f7-64pz9" Oct 05 08:13:37 crc kubenswrapper[4846]: I1005 08:13:37.438862 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-575454b8f7-64pz9" podStartSLOduration=2.438830875 podStartE2EDuration="2.438830875s" podCreationTimestamp="2025-10-05 08:13:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 08:13:37.426112135 +0000 UTC m=+5139.666964960" watchObservedRunningTime="2025-10-05 08:13:37.438830875 +0000 UTC m=+5139.679683690" Oct 05 08:13:47 crc kubenswrapper[4846]: I1005 08:13:47.498002 4846 scope.go:117] "RemoveContainer" containerID="8fa72be735dff8e623a4a4958f452896ef79e72dd44c2d714a3019187462de47" Oct 05 08:13:47 crc kubenswrapper[4846]: E1005 08:13:47.499203 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 08:14:01 crc kubenswrapper[4846]: I1005 08:14:01.498684 4846 scope.go:117] "RemoveContainer" containerID="8fa72be735dff8e623a4a4958f452896ef79e72dd44c2d714a3019187462de47" Oct 05 08:14:01 crc kubenswrapper[4846]: E1005 08:14:01.499884 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 08:14:07 crc kubenswrapper[4846]: I1005 08:14:07.462332 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/keystone-575454b8f7-64pz9" Oct 05 08:14:11 crc kubenswrapper[4846]: I1005 08:14:11.198687 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Oct 05 08:14:11 crc kubenswrapper[4846]: I1005 08:14:11.202413 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 05 08:14:11 crc kubenswrapper[4846]: I1005 08:14:11.205112 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstackclient-openstackclient-dockercfg-8jz9w" Oct 05 08:14:11 crc kubenswrapper[4846]: I1005 08:14:11.207143 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-config-secret" Oct 05 08:14:11 crc kubenswrapper[4846]: I1005 08:14:11.207421 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config" Oct 05 08:14:11 crc kubenswrapper[4846]: I1005 08:14:11.221141 4846 status_manager.go:875] "Failed to update status for pod" pod="openstack/openstackclient" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32f73c66-aee5-4c33-9435-f6c1e00b7136\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T08:14:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T08:14:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T08:14:11Z\\\",\\\"message\\\":\\\"containers with unready status: [openstackclient]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-05T08:14:11Z\\\",\\\"message\\\":\\\"containers with unready status: [openstackclient]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.rdoproject.org/podified-antelope-centos9/openstack-openstackclient:88dc57612f447daadb492dcf3ad854ac\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"openstackclient\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/home/cloud-admin/.config/openstack/clouds.yaml\\\",\\\"name\\\":\\\"openstack-config\\\"},{\\\"mountPath\\\":\\\"/home/cloud-admin/.config/openstack/secure.yaml\\\",\\\"name\\\":\\\"openstack-config-secret\\\"},{\\\"mountPath\\\":\\\"/home/cloud-admin/cloudrc\\\",\\\"name\\\":\\\"openstack-config-secret\\\"},{\\\"mountPath\\\":\\\"/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem\\\",\\\"name\\\":\\\"combined-ca-bundle\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vfx6r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-05T08:14:11Z\\\"}}\" for pod \"openstack\"/\"openstackclient\": pods \"openstackclient\" not found" Oct 05 08:14:11 crc kubenswrapper[4846]: I1005 08:14:11.234531 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Oct 05 08:14:11 crc kubenswrapper[4846]: I1005 08:14:11.236987 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vfx6r\" (UniqueName: \"kubernetes.io/projected/32f73c66-aee5-4c33-9435-f6c1e00b7136-kube-api-access-vfx6r\") pod \"openstackclient\" (UID: \"32f73c66-aee5-4c33-9435-f6c1e00b7136\") " pod="openstack/openstackclient" Oct 05 08:14:11 crc kubenswrapper[4846]: I1005 08:14:11.237062 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/32f73c66-aee5-4c33-9435-f6c1e00b7136-openstack-config-secret\") pod \"openstackclient\" (UID: \"32f73c66-aee5-4c33-9435-f6c1e00b7136\") " pod="openstack/openstackclient" Oct 05 08:14:11 crc kubenswrapper[4846]: I1005 08:14:11.237166 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/32f73c66-aee5-4c33-9435-f6c1e00b7136-openstack-config\") pod \"openstackclient\" (UID: \"32f73c66-aee5-4c33-9435-f6c1e00b7136\") " pod="openstack/openstackclient" Oct 05 08:14:11 crc kubenswrapper[4846]: I1005 08:14:11.237268 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/32f73c66-aee5-4c33-9435-f6c1e00b7136-combined-ca-bundle\") pod \"openstackclient\" (UID: \"32f73c66-aee5-4c33-9435-f6c1e00b7136\") " pod="openstack/openstackclient" Oct 05 08:14:11 crc kubenswrapper[4846]: I1005 08:14:11.244517 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstackclient"] Oct 05 08:14:11 crc kubenswrapper[4846]: I1005 08:14:11.252876 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstackclient"] Oct 05 08:14:11 crc kubenswrapper[4846]: E1005 08:14:11.253692 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[combined-ca-bundle kube-api-access-vfx6r openstack-config openstack-config-secret], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openstack/openstackclient" podUID="32f73c66-aee5-4c33-9435-f6c1e00b7136" Oct 05 08:14:11 crc kubenswrapper[4846]: I1005 08:14:11.318836 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Oct 05 08:14:11 crc kubenswrapper[4846]: I1005 08:14:11.320136 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 05 08:14:11 crc kubenswrapper[4846]: I1005 08:14:11.331654 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Oct 05 08:14:11 crc kubenswrapper[4846]: I1005 08:14:11.338359 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vfx6r\" (UniqueName: \"kubernetes.io/projected/32f73c66-aee5-4c33-9435-f6c1e00b7136-kube-api-access-vfx6r\") pod \"openstackclient\" (UID: \"32f73c66-aee5-4c33-9435-f6c1e00b7136\") " pod="openstack/openstackclient" Oct 05 08:14:11 crc kubenswrapper[4846]: I1005 08:14:11.338437 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/32f73c66-aee5-4c33-9435-f6c1e00b7136-openstack-config-secret\") pod \"openstackclient\" (UID: \"32f73c66-aee5-4c33-9435-f6c1e00b7136\") " pod="openstack/openstackclient" Oct 05 08:14:11 crc kubenswrapper[4846]: I1005 08:14:11.338515 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/32f73c66-aee5-4c33-9435-f6c1e00b7136-openstack-config\") pod \"openstackclient\" (UID: \"32f73c66-aee5-4c33-9435-f6c1e00b7136\") " pod="openstack/openstackclient" Oct 05 08:14:11 crc kubenswrapper[4846]: I1005 08:14:11.338579 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/32f73c66-aee5-4c33-9435-f6c1e00b7136-combined-ca-bundle\") pod \"openstackclient\" (UID: \"32f73c66-aee5-4c33-9435-f6c1e00b7136\") " pod="openstack/openstackclient" Oct 05 08:14:11 crc kubenswrapper[4846]: E1005 08:14:11.340513 4846 projected.go:194] Error preparing data for projected volume kube-api-access-vfx6r for pod openstack/openstackclient: failed to fetch token: serviceaccounts "openstackclient-openstackclient" is forbidden: the UID in the bound object reference (32f73c66-aee5-4c33-9435-f6c1e00b7136) does not match the UID in record. The object might have been deleted and then recreated Oct 05 08:14:11 crc kubenswrapper[4846]: E1005 08:14:11.340608 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/32f73c66-aee5-4c33-9435-f6c1e00b7136-kube-api-access-vfx6r podName:32f73c66-aee5-4c33-9435-f6c1e00b7136 nodeName:}" failed. No retries permitted until 2025-10-05 08:14:11.840563395 +0000 UTC m=+5174.081416180 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-vfx6r" (UniqueName: "kubernetes.io/projected/32f73c66-aee5-4c33-9435-f6c1e00b7136-kube-api-access-vfx6r") pod "openstackclient" (UID: "32f73c66-aee5-4c33-9435-f6c1e00b7136") : failed to fetch token: serviceaccounts "openstackclient-openstackclient" is forbidden: the UID in the bound object reference (32f73c66-aee5-4c33-9435-f6c1e00b7136) does not match the UID in record. The object might have been deleted and then recreated Oct 05 08:14:11 crc kubenswrapper[4846]: I1005 08:14:11.341172 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/32f73c66-aee5-4c33-9435-f6c1e00b7136-openstack-config\") pod \"openstackclient\" (UID: \"32f73c66-aee5-4c33-9435-f6c1e00b7136\") " pod="openstack/openstackclient" Oct 05 08:14:11 crc kubenswrapper[4846]: I1005 08:14:11.347301 4846 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="32f73c66-aee5-4c33-9435-f6c1e00b7136" podUID="ef74de1f-e213-4ea6-8d5c-9ba3d8416f0b" Oct 05 08:14:11 crc kubenswrapper[4846]: I1005 08:14:11.348278 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/32f73c66-aee5-4c33-9435-f6c1e00b7136-openstack-config-secret\") pod \"openstackclient\" (UID: \"32f73c66-aee5-4c33-9435-f6c1e00b7136\") " pod="openstack/openstackclient" Oct 05 08:14:11 crc kubenswrapper[4846]: I1005 08:14:11.348840 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/32f73c66-aee5-4c33-9435-f6c1e00b7136-combined-ca-bundle\") pod \"openstackclient\" (UID: \"32f73c66-aee5-4c33-9435-f6c1e00b7136\") " pod="openstack/openstackclient" Oct 05 08:14:11 crc kubenswrapper[4846]: I1005 08:14:11.439752 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/ef74de1f-e213-4ea6-8d5c-9ba3d8416f0b-openstack-config\") pod \"openstackclient\" (UID: \"ef74de1f-e213-4ea6-8d5c-9ba3d8416f0b\") " pod="openstack/openstackclient" Oct 05 08:14:11 crc kubenswrapper[4846]: I1005 08:14:11.439797 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef74de1f-e213-4ea6-8d5c-9ba3d8416f0b-combined-ca-bundle\") pod \"openstackclient\" (UID: \"ef74de1f-e213-4ea6-8d5c-9ba3d8416f0b\") " pod="openstack/openstackclient" Oct 05 08:14:11 crc kubenswrapper[4846]: I1005 08:14:11.439838 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kjvfz\" (UniqueName: \"kubernetes.io/projected/ef74de1f-e213-4ea6-8d5c-9ba3d8416f0b-kube-api-access-kjvfz\") pod \"openstackclient\" (UID: \"ef74de1f-e213-4ea6-8d5c-9ba3d8416f0b\") " pod="openstack/openstackclient" Oct 05 08:14:11 crc kubenswrapper[4846]: I1005 08:14:11.439876 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/ef74de1f-e213-4ea6-8d5c-9ba3d8416f0b-openstack-config-secret\") pod \"openstackclient\" (UID: \"ef74de1f-e213-4ea6-8d5c-9ba3d8416f0b\") " pod="openstack/openstackclient" Oct 05 08:14:11 crc kubenswrapper[4846]: I1005 08:14:11.541609 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/ef74de1f-e213-4ea6-8d5c-9ba3d8416f0b-openstack-config\") pod \"openstackclient\" (UID: \"ef74de1f-e213-4ea6-8d5c-9ba3d8416f0b\") " pod="openstack/openstackclient" Oct 05 08:14:11 crc kubenswrapper[4846]: I1005 08:14:11.541667 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef74de1f-e213-4ea6-8d5c-9ba3d8416f0b-combined-ca-bundle\") pod \"openstackclient\" (UID: \"ef74de1f-e213-4ea6-8d5c-9ba3d8416f0b\") " pod="openstack/openstackclient" Oct 05 08:14:11 crc kubenswrapper[4846]: I1005 08:14:11.541726 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kjvfz\" (UniqueName: \"kubernetes.io/projected/ef74de1f-e213-4ea6-8d5c-9ba3d8416f0b-kube-api-access-kjvfz\") pod \"openstackclient\" (UID: \"ef74de1f-e213-4ea6-8d5c-9ba3d8416f0b\") " pod="openstack/openstackclient" Oct 05 08:14:11 crc kubenswrapper[4846]: I1005 08:14:11.541777 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/ef74de1f-e213-4ea6-8d5c-9ba3d8416f0b-openstack-config-secret\") pod \"openstackclient\" (UID: \"ef74de1f-e213-4ea6-8d5c-9ba3d8416f0b\") " pod="openstack/openstackclient" Oct 05 08:14:11 crc kubenswrapper[4846]: I1005 08:14:11.543603 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/ef74de1f-e213-4ea6-8d5c-9ba3d8416f0b-openstack-config\") pod \"openstackclient\" (UID: \"ef74de1f-e213-4ea6-8d5c-9ba3d8416f0b\") " pod="openstack/openstackclient" Oct 05 08:14:11 crc kubenswrapper[4846]: I1005 08:14:11.545257 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/ef74de1f-e213-4ea6-8d5c-9ba3d8416f0b-openstack-config-secret\") pod \"openstackclient\" (UID: \"ef74de1f-e213-4ea6-8d5c-9ba3d8416f0b\") " pod="openstack/openstackclient" Oct 05 08:14:11 crc kubenswrapper[4846]: I1005 08:14:11.547460 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef74de1f-e213-4ea6-8d5c-9ba3d8416f0b-combined-ca-bundle\") pod \"openstackclient\" (UID: \"ef74de1f-e213-4ea6-8d5c-9ba3d8416f0b\") " pod="openstack/openstackclient" Oct 05 08:14:11 crc kubenswrapper[4846]: I1005 08:14:11.559889 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kjvfz\" (UniqueName: \"kubernetes.io/projected/ef74de1f-e213-4ea6-8d5c-9ba3d8416f0b-kube-api-access-kjvfz\") pod \"openstackclient\" (UID: \"ef74de1f-e213-4ea6-8d5c-9ba3d8416f0b\") " pod="openstack/openstackclient" Oct 05 08:14:11 crc kubenswrapper[4846]: I1005 08:14:11.645313 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 05 08:14:11 crc kubenswrapper[4846]: I1005 08:14:11.764916 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 05 08:14:11 crc kubenswrapper[4846]: I1005 08:14:11.769249 4846 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="32f73c66-aee5-4c33-9435-f6c1e00b7136" podUID="ef74de1f-e213-4ea6-8d5c-9ba3d8416f0b" Oct 05 08:14:11 crc kubenswrapper[4846]: I1005 08:14:11.789302 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 05 08:14:11 crc kubenswrapper[4846]: I1005 08:14:11.793187 4846 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="32f73c66-aee5-4c33-9435-f6c1e00b7136" podUID="ef74de1f-e213-4ea6-8d5c-9ba3d8416f0b" Oct 05 08:14:11 crc kubenswrapper[4846]: I1005 08:14:11.847120 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vfx6r\" (UniqueName: \"kubernetes.io/projected/32f73c66-aee5-4c33-9435-f6c1e00b7136-kube-api-access-vfx6r\") pod \"openstackclient\" (UID: \"32f73c66-aee5-4c33-9435-f6c1e00b7136\") " pod="openstack/openstackclient" Oct 05 08:14:11 crc kubenswrapper[4846]: E1005 08:14:11.849015 4846 projected.go:194] Error preparing data for projected volume kube-api-access-vfx6r for pod openstack/openstackclient: failed to fetch token: serviceaccounts "openstackclient-openstackclient" is forbidden: the UID in the bound object reference (32f73c66-aee5-4c33-9435-f6c1e00b7136) does not match the UID in record. The object might have been deleted and then recreated Oct 05 08:14:11 crc kubenswrapper[4846]: E1005 08:14:11.849089 4846 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/32f73c66-aee5-4c33-9435-f6c1e00b7136-kube-api-access-vfx6r podName:32f73c66-aee5-4c33-9435-f6c1e00b7136 nodeName:}" failed. No retries permitted until 2025-10-05 08:14:12.849070871 +0000 UTC m=+5175.089923646 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-vfx6r" (UniqueName: "kubernetes.io/projected/32f73c66-aee5-4c33-9435-f6c1e00b7136-kube-api-access-vfx6r") pod "openstackclient" (UID: "32f73c66-aee5-4c33-9435-f6c1e00b7136") : failed to fetch token: serviceaccounts "openstackclient-openstackclient" is forbidden: the UID in the bound object reference (32f73c66-aee5-4c33-9435-f6c1e00b7136) does not match the UID in record. The object might have been deleted and then recreated Oct 05 08:14:11 crc kubenswrapper[4846]: I1005 08:14:11.948078 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/32f73c66-aee5-4c33-9435-f6c1e00b7136-combined-ca-bundle\") pod \"32f73c66-aee5-4c33-9435-f6c1e00b7136\" (UID: \"32f73c66-aee5-4c33-9435-f6c1e00b7136\") " Oct 05 08:14:11 crc kubenswrapper[4846]: I1005 08:14:11.948161 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/32f73c66-aee5-4c33-9435-f6c1e00b7136-openstack-config-secret\") pod \"32f73c66-aee5-4c33-9435-f6c1e00b7136\" (UID: \"32f73c66-aee5-4c33-9435-f6c1e00b7136\") " Oct 05 08:14:11 crc kubenswrapper[4846]: I1005 08:14:11.948244 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/32f73c66-aee5-4c33-9435-f6c1e00b7136-openstack-config\") pod \"32f73c66-aee5-4c33-9435-f6c1e00b7136\" (UID: \"32f73c66-aee5-4c33-9435-f6c1e00b7136\") " Oct 05 08:14:11 crc kubenswrapper[4846]: I1005 08:14:11.948631 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vfx6r\" (UniqueName: \"kubernetes.io/projected/32f73c66-aee5-4c33-9435-f6c1e00b7136-kube-api-access-vfx6r\") on node \"crc\" DevicePath \"\"" Oct 05 08:14:11 crc kubenswrapper[4846]: I1005 08:14:11.949618 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/32f73c66-aee5-4c33-9435-f6c1e00b7136-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "32f73c66-aee5-4c33-9435-f6c1e00b7136" (UID: "32f73c66-aee5-4c33-9435-f6c1e00b7136"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 08:14:11 crc kubenswrapper[4846]: I1005 08:14:11.955838 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/32f73c66-aee5-4c33-9435-f6c1e00b7136-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "32f73c66-aee5-4c33-9435-f6c1e00b7136" (UID: "32f73c66-aee5-4c33-9435-f6c1e00b7136"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:14:11 crc kubenswrapper[4846]: I1005 08:14:11.956178 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/32f73c66-aee5-4c33-9435-f6c1e00b7136-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "32f73c66-aee5-4c33-9435-f6c1e00b7136" (UID: "32f73c66-aee5-4c33-9435-f6c1e00b7136"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:14:12 crc kubenswrapper[4846]: I1005 08:14:12.050813 4846 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/32f73c66-aee5-4c33-9435-f6c1e00b7136-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Oct 05 08:14:12 crc kubenswrapper[4846]: I1005 08:14:12.050878 4846 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/32f73c66-aee5-4c33-9435-f6c1e00b7136-openstack-config\") on node \"crc\" DevicePath \"\"" Oct 05 08:14:12 crc kubenswrapper[4846]: I1005 08:14:12.050893 4846 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/32f73c66-aee5-4c33-9435-f6c1e00b7136-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 08:14:12 crc kubenswrapper[4846]: I1005 08:14:12.169547 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Oct 05 08:14:12 crc kubenswrapper[4846]: W1005 08:14:12.177182 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podef74de1f_e213_4ea6_8d5c_9ba3d8416f0b.slice/crio-41e3508498d6b4a540c01a6c5431f0929514b25a708240c3f2174378ef9fa7b9 WatchSource:0}: Error finding container 41e3508498d6b4a540c01a6c5431f0929514b25a708240c3f2174378ef9fa7b9: Status 404 returned error can't find the container with id 41e3508498d6b4a540c01a6c5431f0929514b25a708240c3f2174378ef9fa7b9 Oct 05 08:14:12 crc kubenswrapper[4846]: I1005 08:14:12.514285 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="32f73c66-aee5-4c33-9435-f6c1e00b7136" path="/var/lib/kubelet/pods/32f73c66-aee5-4c33-9435-f6c1e00b7136/volumes" Oct 05 08:14:12 crc kubenswrapper[4846]: I1005 08:14:12.774818 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 05 08:14:12 crc kubenswrapper[4846]: I1005 08:14:12.774822 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"ef74de1f-e213-4ea6-8d5c-9ba3d8416f0b","Type":"ContainerStarted","Data":"41e3508498d6b4a540c01a6c5431f0929514b25a708240c3f2174378ef9fa7b9"} Oct 05 08:14:12 crc kubenswrapper[4846]: I1005 08:14:12.778397 4846 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="32f73c66-aee5-4c33-9435-f6c1e00b7136" podUID="ef74de1f-e213-4ea6-8d5c-9ba3d8416f0b" Oct 05 08:14:12 crc kubenswrapper[4846]: I1005 08:14:12.782261 4846 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="32f73c66-aee5-4c33-9435-f6c1e00b7136" podUID="ef74de1f-e213-4ea6-8d5c-9ba3d8416f0b" Oct 05 08:14:14 crc kubenswrapper[4846]: I1005 08:14:14.497901 4846 scope.go:117] "RemoveContainer" containerID="8fa72be735dff8e623a4a4958f452896ef79e72dd44c2d714a3019187462de47" Oct 05 08:14:14 crc kubenswrapper[4846]: E1005 08:14:14.498822 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 08:14:22 crc kubenswrapper[4846]: I1005 08:14:22.868831 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"ef74de1f-e213-4ea6-8d5c-9ba3d8416f0b","Type":"ContainerStarted","Data":"d6bf0f646ef44b67def03d565e027cf316116ed0fa05b1e4122c3b048ee80e62"} Oct 05 08:14:22 crc kubenswrapper[4846]: I1005 08:14:22.896525 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=1.690305181 podStartE2EDuration="11.896495905s" podCreationTimestamp="2025-10-05 08:14:11 +0000 UTC" firstStartedPulling="2025-10-05 08:14:12.180193537 +0000 UTC m=+5174.421046312" lastFinishedPulling="2025-10-05 08:14:22.386384221 +0000 UTC m=+5184.627237036" observedRunningTime="2025-10-05 08:14:22.889309203 +0000 UTC m=+5185.130162018" watchObservedRunningTime="2025-10-05 08:14:22.896495905 +0000 UTC m=+5185.137348720" Oct 05 08:14:26 crc kubenswrapper[4846]: I1005 08:14:26.497570 4846 scope.go:117] "RemoveContainer" containerID="8fa72be735dff8e623a4a4958f452896ef79e72dd44c2d714a3019187462de47" Oct 05 08:14:26 crc kubenswrapper[4846]: E1005 08:14:26.498594 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 08:14:31 crc kubenswrapper[4846]: I1005 08:14:31.186505 4846 scope.go:117] "RemoveContainer" containerID="4b99194058de1e2b0eeda1df7895b490c5e4cc1d359f20acfac1596bfa476a84" Oct 05 08:14:31 crc kubenswrapper[4846]: I1005 08:14:31.223766 4846 scope.go:117] "RemoveContainer" containerID="8d7578bd15f9380fd60c5291373e36dbe5257462d8b2c81e3243098241213ddf" Oct 05 08:14:41 crc kubenswrapper[4846]: I1005 08:14:41.497868 4846 scope.go:117] "RemoveContainer" containerID="8fa72be735dff8e623a4a4958f452896ef79e72dd44c2d714a3019187462de47" Oct 05 08:14:41 crc kubenswrapper[4846]: E1005 08:14:41.498890 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 08:14:53 crc kubenswrapper[4846]: I1005 08:14:53.497590 4846 scope.go:117] "RemoveContainer" containerID="8fa72be735dff8e623a4a4958f452896ef79e72dd44c2d714a3019187462de47" Oct 05 08:14:53 crc kubenswrapper[4846]: E1005 08:14:53.498637 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 08:15:00 crc kubenswrapper[4846]: I1005 08:15:00.169738 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29327535-9qgh9"] Oct 05 08:15:00 crc kubenswrapper[4846]: I1005 08:15:00.171923 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29327535-9qgh9" Oct 05 08:15:00 crc kubenswrapper[4846]: I1005 08:15:00.174715 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 05 08:15:00 crc kubenswrapper[4846]: I1005 08:15:00.175272 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 05 08:15:00 crc kubenswrapper[4846]: I1005 08:15:00.193563 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29327535-9qgh9"] Oct 05 08:15:00 crc kubenswrapper[4846]: I1005 08:15:00.303782 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bhn4f\" (UniqueName: \"kubernetes.io/projected/5612d2ab-7b8a-452c-8b97-8a968620b14f-kube-api-access-bhn4f\") pod \"collect-profiles-29327535-9qgh9\" (UID: \"5612d2ab-7b8a-452c-8b97-8a968620b14f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327535-9qgh9" Oct 05 08:15:00 crc kubenswrapper[4846]: I1005 08:15:00.303979 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5612d2ab-7b8a-452c-8b97-8a968620b14f-secret-volume\") pod \"collect-profiles-29327535-9qgh9\" (UID: \"5612d2ab-7b8a-452c-8b97-8a968620b14f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327535-9qgh9" Oct 05 08:15:00 crc kubenswrapper[4846]: I1005 08:15:00.304038 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5612d2ab-7b8a-452c-8b97-8a968620b14f-config-volume\") pod \"collect-profiles-29327535-9qgh9\" (UID: \"5612d2ab-7b8a-452c-8b97-8a968620b14f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327535-9qgh9" Oct 05 08:15:00 crc kubenswrapper[4846]: I1005 08:15:00.405856 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bhn4f\" (UniqueName: \"kubernetes.io/projected/5612d2ab-7b8a-452c-8b97-8a968620b14f-kube-api-access-bhn4f\") pod \"collect-profiles-29327535-9qgh9\" (UID: \"5612d2ab-7b8a-452c-8b97-8a968620b14f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327535-9qgh9" Oct 05 08:15:00 crc kubenswrapper[4846]: I1005 08:15:00.405973 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5612d2ab-7b8a-452c-8b97-8a968620b14f-secret-volume\") pod \"collect-profiles-29327535-9qgh9\" (UID: \"5612d2ab-7b8a-452c-8b97-8a968620b14f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327535-9qgh9" Oct 05 08:15:00 crc kubenswrapper[4846]: I1005 08:15:00.406011 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5612d2ab-7b8a-452c-8b97-8a968620b14f-config-volume\") pod \"collect-profiles-29327535-9qgh9\" (UID: \"5612d2ab-7b8a-452c-8b97-8a968620b14f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327535-9qgh9" Oct 05 08:15:00 crc kubenswrapper[4846]: I1005 08:15:00.407700 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5612d2ab-7b8a-452c-8b97-8a968620b14f-config-volume\") pod \"collect-profiles-29327535-9qgh9\" (UID: \"5612d2ab-7b8a-452c-8b97-8a968620b14f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327535-9qgh9" Oct 05 08:15:00 crc kubenswrapper[4846]: I1005 08:15:00.413025 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5612d2ab-7b8a-452c-8b97-8a968620b14f-secret-volume\") pod \"collect-profiles-29327535-9qgh9\" (UID: \"5612d2ab-7b8a-452c-8b97-8a968620b14f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327535-9qgh9" Oct 05 08:15:00 crc kubenswrapper[4846]: I1005 08:15:00.431397 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bhn4f\" (UniqueName: \"kubernetes.io/projected/5612d2ab-7b8a-452c-8b97-8a968620b14f-kube-api-access-bhn4f\") pod \"collect-profiles-29327535-9qgh9\" (UID: \"5612d2ab-7b8a-452c-8b97-8a968620b14f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327535-9qgh9" Oct 05 08:15:00 crc kubenswrapper[4846]: I1005 08:15:00.504171 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29327535-9qgh9" Oct 05 08:15:00 crc kubenswrapper[4846]: W1005 08:15:00.971873 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5612d2ab_7b8a_452c_8b97_8a968620b14f.slice/crio-9c2b41aa59d46892d53614638188500450e50b37dbdc914917c54b8ede9a7e4b WatchSource:0}: Error finding container 9c2b41aa59d46892d53614638188500450e50b37dbdc914917c54b8ede9a7e4b: Status 404 returned error can't find the container with id 9c2b41aa59d46892d53614638188500450e50b37dbdc914917c54b8ede9a7e4b Oct 05 08:15:00 crc kubenswrapper[4846]: I1005 08:15:00.972141 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29327535-9qgh9"] Oct 05 08:15:01 crc kubenswrapper[4846]: I1005 08:15:01.306143 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29327535-9qgh9" event={"ID":"5612d2ab-7b8a-452c-8b97-8a968620b14f","Type":"ContainerStarted","Data":"38fc1323ac49b5bbf03cda77705f78df164686147e3bd467d7059633eddc03e6"} Oct 05 08:15:01 crc kubenswrapper[4846]: I1005 08:15:01.306464 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29327535-9qgh9" event={"ID":"5612d2ab-7b8a-452c-8b97-8a968620b14f","Type":"ContainerStarted","Data":"9c2b41aa59d46892d53614638188500450e50b37dbdc914917c54b8ede9a7e4b"} Oct 05 08:15:02 crc kubenswrapper[4846]: I1005 08:15:02.318139 4846 generic.go:334] "Generic (PLEG): container finished" podID="5612d2ab-7b8a-452c-8b97-8a968620b14f" containerID="38fc1323ac49b5bbf03cda77705f78df164686147e3bd467d7059633eddc03e6" exitCode=0 Oct 05 08:15:02 crc kubenswrapper[4846]: I1005 08:15:02.318233 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29327535-9qgh9" event={"ID":"5612d2ab-7b8a-452c-8b97-8a968620b14f","Type":"ContainerDied","Data":"38fc1323ac49b5bbf03cda77705f78df164686147e3bd467d7059633eddc03e6"} Oct 05 08:15:03 crc kubenswrapper[4846]: I1005 08:15:03.692031 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29327535-9qgh9" Oct 05 08:15:03 crc kubenswrapper[4846]: I1005 08:15:03.873877 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bhn4f\" (UniqueName: \"kubernetes.io/projected/5612d2ab-7b8a-452c-8b97-8a968620b14f-kube-api-access-bhn4f\") pod \"5612d2ab-7b8a-452c-8b97-8a968620b14f\" (UID: \"5612d2ab-7b8a-452c-8b97-8a968620b14f\") " Oct 05 08:15:03 crc kubenswrapper[4846]: I1005 08:15:03.874074 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5612d2ab-7b8a-452c-8b97-8a968620b14f-config-volume\") pod \"5612d2ab-7b8a-452c-8b97-8a968620b14f\" (UID: \"5612d2ab-7b8a-452c-8b97-8a968620b14f\") " Oct 05 08:15:03 crc kubenswrapper[4846]: I1005 08:15:03.874127 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5612d2ab-7b8a-452c-8b97-8a968620b14f-secret-volume\") pod \"5612d2ab-7b8a-452c-8b97-8a968620b14f\" (UID: \"5612d2ab-7b8a-452c-8b97-8a968620b14f\") " Oct 05 08:15:03 crc kubenswrapper[4846]: I1005 08:15:03.876503 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5612d2ab-7b8a-452c-8b97-8a968620b14f-config-volume" (OuterVolumeSpecName: "config-volume") pod "5612d2ab-7b8a-452c-8b97-8a968620b14f" (UID: "5612d2ab-7b8a-452c-8b97-8a968620b14f"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 08:15:03 crc kubenswrapper[4846]: I1005 08:15:03.881320 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5612d2ab-7b8a-452c-8b97-8a968620b14f-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "5612d2ab-7b8a-452c-8b97-8a968620b14f" (UID: "5612d2ab-7b8a-452c-8b97-8a968620b14f"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:15:03 crc kubenswrapper[4846]: I1005 08:15:03.882254 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5612d2ab-7b8a-452c-8b97-8a968620b14f-kube-api-access-bhn4f" (OuterVolumeSpecName: "kube-api-access-bhn4f") pod "5612d2ab-7b8a-452c-8b97-8a968620b14f" (UID: "5612d2ab-7b8a-452c-8b97-8a968620b14f"). InnerVolumeSpecName "kube-api-access-bhn4f". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:15:03 crc kubenswrapper[4846]: I1005 08:15:03.976359 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bhn4f\" (UniqueName: \"kubernetes.io/projected/5612d2ab-7b8a-452c-8b97-8a968620b14f-kube-api-access-bhn4f\") on node \"crc\" DevicePath \"\"" Oct 05 08:15:03 crc kubenswrapper[4846]: I1005 08:15:03.976414 4846 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5612d2ab-7b8a-452c-8b97-8a968620b14f-config-volume\") on node \"crc\" DevicePath \"\"" Oct 05 08:15:03 crc kubenswrapper[4846]: I1005 08:15:03.976434 4846 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5612d2ab-7b8a-452c-8b97-8a968620b14f-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 05 08:15:04 crc kubenswrapper[4846]: I1005 08:15:04.338928 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29327535-9qgh9" event={"ID":"5612d2ab-7b8a-452c-8b97-8a968620b14f","Type":"ContainerDied","Data":"9c2b41aa59d46892d53614638188500450e50b37dbdc914917c54b8ede9a7e4b"} Oct 05 08:15:04 crc kubenswrapper[4846]: I1005 08:15:04.338968 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9c2b41aa59d46892d53614638188500450e50b37dbdc914917c54b8ede9a7e4b" Oct 05 08:15:04 crc kubenswrapper[4846]: I1005 08:15:04.339351 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29327535-9qgh9" Oct 05 08:15:04 crc kubenswrapper[4846]: I1005 08:15:04.437619 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29327490-6kpqj"] Oct 05 08:15:04 crc kubenswrapper[4846]: I1005 08:15:04.447059 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29327490-6kpqj"] Oct 05 08:15:04 crc kubenswrapper[4846]: I1005 08:15:04.509264 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5e0701f1-92c9-45fe-81d9-910fcf30ea6d" path="/var/lib/kubelet/pods/5e0701f1-92c9-45fe-81d9-910fcf30ea6d/volumes" Oct 05 08:15:07 crc kubenswrapper[4846]: I1005 08:15:07.499141 4846 scope.go:117] "RemoveContainer" containerID="8fa72be735dff8e623a4a4958f452896ef79e72dd44c2d714a3019187462de47" Oct 05 08:15:07 crc kubenswrapper[4846]: E1005 08:15:07.499920 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 08:15:21 crc kubenswrapper[4846]: I1005 08:15:21.497606 4846 scope.go:117] "RemoveContainer" containerID="8fa72be735dff8e623a4a4958f452896ef79e72dd44c2d714a3019187462de47" Oct 05 08:15:21 crc kubenswrapper[4846]: E1005 08:15:21.498355 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 08:15:23 crc kubenswrapper[4846]: I1005 08:15:23.199875 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-mbz24"] Oct 05 08:15:23 crc kubenswrapper[4846]: E1005 08:15:23.206769 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5612d2ab-7b8a-452c-8b97-8a968620b14f" containerName="collect-profiles" Oct 05 08:15:23 crc kubenswrapper[4846]: I1005 08:15:23.207072 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="5612d2ab-7b8a-452c-8b97-8a968620b14f" containerName="collect-profiles" Oct 05 08:15:23 crc kubenswrapper[4846]: I1005 08:15:23.207791 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="5612d2ab-7b8a-452c-8b97-8a968620b14f" containerName="collect-profiles" Oct 05 08:15:23 crc kubenswrapper[4846]: I1005 08:15:23.210727 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mbz24" Oct 05 08:15:23 crc kubenswrapper[4846]: I1005 08:15:23.222908 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-mbz24"] Oct 05 08:15:23 crc kubenswrapper[4846]: I1005 08:15:23.270913 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dc9j4\" (UniqueName: \"kubernetes.io/projected/1e7701c8-038a-4a7e-9b25-96c059199bba-kube-api-access-dc9j4\") pod \"certified-operators-mbz24\" (UID: \"1e7701c8-038a-4a7e-9b25-96c059199bba\") " pod="openshift-marketplace/certified-operators-mbz24" Oct 05 08:15:23 crc kubenswrapper[4846]: I1005 08:15:23.270994 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1e7701c8-038a-4a7e-9b25-96c059199bba-catalog-content\") pod \"certified-operators-mbz24\" (UID: \"1e7701c8-038a-4a7e-9b25-96c059199bba\") " pod="openshift-marketplace/certified-operators-mbz24" Oct 05 08:15:23 crc kubenswrapper[4846]: I1005 08:15:23.271034 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1e7701c8-038a-4a7e-9b25-96c059199bba-utilities\") pod \"certified-operators-mbz24\" (UID: \"1e7701c8-038a-4a7e-9b25-96c059199bba\") " pod="openshift-marketplace/certified-operators-mbz24" Oct 05 08:15:23 crc kubenswrapper[4846]: I1005 08:15:23.372761 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dc9j4\" (UniqueName: \"kubernetes.io/projected/1e7701c8-038a-4a7e-9b25-96c059199bba-kube-api-access-dc9j4\") pod \"certified-operators-mbz24\" (UID: \"1e7701c8-038a-4a7e-9b25-96c059199bba\") " pod="openshift-marketplace/certified-operators-mbz24" Oct 05 08:15:23 crc kubenswrapper[4846]: I1005 08:15:23.372893 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1e7701c8-038a-4a7e-9b25-96c059199bba-catalog-content\") pod \"certified-operators-mbz24\" (UID: \"1e7701c8-038a-4a7e-9b25-96c059199bba\") " pod="openshift-marketplace/certified-operators-mbz24" Oct 05 08:15:23 crc kubenswrapper[4846]: I1005 08:15:23.372965 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1e7701c8-038a-4a7e-9b25-96c059199bba-utilities\") pod \"certified-operators-mbz24\" (UID: \"1e7701c8-038a-4a7e-9b25-96c059199bba\") " pod="openshift-marketplace/certified-operators-mbz24" Oct 05 08:15:23 crc kubenswrapper[4846]: I1005 08:15:23.373700 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1e7701c8-038a-4a7e-9b25-96c059199bba-utilities\") pod \"certified-operators-mbz24\" (UID: \"1e7701c8-038a-4a7e-9b25-96c059199bba\") " pod="openshift-marketplace/certified-operators-mbz24" Oct 05 08:15:23 crc kubenswrapper[4846]: I1005 08:15:23.374427 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1e7701c8-038a-4a7e-9b25-96c059199bba-catalog-content\") pod \"certified-operators-mbz24\" (UID: \"1e7701c8-038a-4a7e-9b25-96c059199bba\") " pod="openshift-marketplace/certified-operators-mbz24" Oct 05 08:15:23 crc kubenswrapper[4846]: I1005 08:15:23.390784 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-xgzh2"] Oct 05 08:15:23 crc kubenswrapper[4846]: I1005 08:15:23.395972 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-xgzh2" Oct 05 08:15:23 crc kubenswrapper[4846]: I1005 08:15:23.405419 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-xgzh2"] Oct 05 08:15:23 crc kubenswrapper[4846]: I1005 08:15:23.412316 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dc9j4\" (UniqueName: \"kubernetes.io/projected/1e7701c8-038a-4a7e-9b25-96c059199bba-kube-api-access-dc9j4\") pod \"certified-operators-mbz24\" (UID: \"1e7701c8-038a-4a7e-9b25-96c059199bba\") " pod="openshift-marketplace/certified-operators-mbz24" Oct 05 08:15:23 crc kubenswrapper[4846]: I1005 08:15:23.474927 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/be1426c3-8e7b-494e-9ef6-3ce557ef6999-utilities\") pod \"community-operators-xgzh2\" (UID: \"be1426c3-8e7b-494e-9ef6-3ce557ef6999\") " pod="openshift-marketplace/community-operators-xgzh2" Oct 05 08:15:23 crc kubenswrapper[4846]: I1005 08:15:23.474971 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wrl7p\" (UniqueName: \"kubernetes.io/projected/be1426c3-8e7b-494e-9ef6-3ce557ef6999-kube-api-access-wrl7p\") pod \"community-operators-xgzh2\" (UID: \"be1426c3-8e7b-494e-9ef6-3ce557ef6999\") " pod="openshift-marketplace/community-operators-xgzh2" Oct 05 08:15:23 crc kubenswrapper[4846]: I1005 08:15:23.475037 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/be1426c3-8e7b-494e-9ef6-3ce557ef6999-catalog-content\") pod \"community-operators-xgzh2\" (UID: \"be1426c3-8e7b-494e-9ef6-3ce557ef6999\") " pod="openshift-marketplace/community-operators-xgzh2" Oct 05 08:15:23 crc kubenswrapper[4846]: I1005 08:15:23.555670 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mbz24" Oct 05 08:15:23 crc kubenswrapper[4846]: I1005 08:15:23.577027 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/be1426c3-8e7b-494e-9ef6-3ce557ef6999-utilities\") pod \"community-operators-xgzh2\" (UID: \"be1426c3-8e7b-494e-9ef6-3ce557ef6999\") " pod="openshift-marketplace/community-operators-xgzh2" Oct 05 08:15:23 crc kubenswrapper[4846]: I1005 08:15:23.577082 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wrl7p\" (UniqueName: \"kubernetes.io/projected/be1426c3-8e7b-494e-9ef6-3ce557ef6999-kube-api-access-wrl7p\") pod \"community-operators-xgzh2\" (UID: \"be1426c3-8e7b-494e-9ef6-3ce557ef6999\") " pod="openshift-marketplace/community-operators-xgzh2" Oct 05 08:15:23 crc kubenswrapper[4846]: I1005 08:15:23.577155 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/be1426c3-8e7b-494e-9ef6-3ce557ef6999-catalog-content\") pod \"community-operators-xgzh2\" (UID: \"be1426c3-8e7b-494e-9ef6-3ce557ef6999\") " pod="openshift-marketplace/community-operators-xgzh2" Oct 05 08:15:23 crc kubenswrapper[4846]: I1005 08:15:23.577666 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/be1426c3-8e7b-494e-9ef6-3ce557ef6999-utilities\") pod \"community-operators-xgzh2\" (UID: \"be1426c3-8e7b-494e-9ef6-3ce557ef6999\") " pod="openshift-marketplace/community-operators-xgzh2" Oct 05 08:15:23 crc kubenswrapper[4846]: I1005 08:15:23.577709 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/be1426c3-8e7b-494e-9ef6-3ce557ef6999-catalog-content\") pod \"community-operators-xgzh2\" (UID: \"be1426c3-8e7b-494e-9ef6-3ce557ef6999\") " pod="openshift-marketplace/community-operators-xgzh2" Oct 05 08:15:23 crc kubenswrapper[4846]: I1005 08:15:23.598902 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wrl7p\" (UniqueName: \"kubernetes.io/projected/be1426c3-8e7b-494e-9ef6-3ce557ef6999-kube-api-access-wrl7p\") pod \"community-operators-xgzh2\" (UID: \"be1426c3-8e7b-494e-9ef6-3ce557ef6999\") " pod="openshift-marketplace/community-operators-xgzh2" Oct 05 08:15:23 crc kubenswrapper[4846]: I1005 08:15:23.756177 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-xgzh2" Oct 05 08:15:24 crc kubenswrapper[4846]: I1005 08:15:24.092800 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-mbz24"] Oct 05 08:15:24 crc kubenswrapper[4846]: I1005 08:15:24.330758 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-xgzh2"] Oct 05 08:15:24 crc kubenswrapper[4846]: W1005 08:15:24.332355 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbe1426c3_8e7b_494e_9ef6_3ce557ef6999.slice/crio-45e8488ecac6168e9395ea3a126dbf5dc49874cec910217227d06bf521cb30fd WatchSource:0}: Error finding container 45e8488ecac6168e9395ea3a126dbf5dc49874cec910217227d06bf521cb30fd: Status 404 returned error can't find the container with id 45e8488ecac6168e9395ea3a126dbf5dc49874cec910217227d06bf521cb30fd Oct 05 08:15:24 crc kubenswrapper[4846]: I1005 08:15:24.565982 4846 generic.go:334] "Generic (PLEG): container finished" podID="be1426c3-8e7b-494e-9ef6-3ce557ef6999" containerID="076676db43db829bbf778305a914211f9be97efc4aae42987575a9362d313700" exitCode=0 Oct 05 08:15:24 crc kubenswrapper[4846]: I1005 08:15:24.566074 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xgzh2" event={"ID":"be1426c3-8e7b-494e-9ef6-3ce557ef6999","Type":"ContainerDied","Data":"076676db43db829bbf778305a914211f9be97efc4aae42987575a9362d313700"} Oct 05 08:15:24 crc kubenswrapper[4846]: I1005 08:15:24.566118 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xgzh2" event={"ID":"be1426c3-8e7b-494e-9ef6-3ce557ef6999","Type":"ContainerStarted","Data":"45e8488ecac6168e9395ea3a126dbf5dc49874cec910217227d06bf521cb30fd"} Oct 05 08:15:24 crc kubenswrapper[4846]: I1005 08:15:24.567645 4846 generic.go:334] "Generic (PLEG): container finished" podID="1e7701c8-038a-4a7e-9b25-96c059199bba" containerID="2b4f4484952a6da87c239627a030a3cfcd91dc71eda2fd217654b8974c743dc5" exitCode=0 Oct 05 08:15:24 crc kubenswrapper[4846]: I1005 08:15:24.567671 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mbz24" event={"ID":"1e7701c8-038a-4a7e-9b25-96c059199bba","Type":"ContainerDied","Data":"2b4f4484952a6da87c239627a030a3cfcd91dc71eda2fd217654b8974c743dc5"} Oct 05 08:15:24 crc kubenswrapper[4846]: I1005 08:15:24.567688 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mbz24" event={"ID":"1e7701c8-038a-4a7e-9b25-96c059199bba","Type":"ContainerStarted","Data":"2c60023001771be953f9a994f812596022f5b38ad87b758f9c85695e98fac6f2"} Oct 05 08:15:25 crc kubenswrapper[4846]: I1005 08:15:25.589681 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mbz24" event={"ID":"1e7701c8-038a-4a7e-9b25-96c059199bba","Type":"ContainerStarted","Data":"6daf0df1174036065afa31dbfba00b9815003c98ae9df135d26c50cca12908d9"} Oct 05 08:15:25 crc kubenswrapper[4846]: I1005 08:15:25.793937 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-rcl8t"] Oct 05 08:15:25 crc kubenswrapper[4846]: I1005 08:15:25.796852 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rcl8t" Oct 05 08:15:25 crc kubenswrapper[4846]: I1005 08:15:25.801825 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-rcl8t"] Oct 05 08:15:25 crc kubenswrapper[4846]: I1005 08:15:25.934865 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b0ae695f-c146-46ac-a49b-440929db1ff5-utilities\") pod \"redhat-operators-rcl8t\" (UID: \"b0ae695f-c146-46ac-a49b-440929db1ff5\") " pod="openshift-marketplace/redhat-operators-rcl8t" Oct 05 08:15:25 crc kubenswrapper[4846]: I1005 08:15:25.935164 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bn7qs\" (UniqueName: \"kubernetes.io/projected/b0ae695f-c146-46ac-a49b-440929db1ff5-kube-api-access-bn7qs\") pod \"redhat-operators-rcl8t\" (UID: \"b0ae695f-c146-46ac-a49b-440929db1ff5\") " pod="openshift-marketplace/redhat-operators-rcl8t" Oct 05 08:15:25 crc kubenswrapper[4846]: I1005 08:15:25.935379 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b0ae695f-c146-46ac-a49b-440929db1ff5-catalog-content\") pod \"redhat-operators-rcl8t\" (UID: \"b0ae695f-c146-46ac-a49b-440929db1ff5\") " pod="openshift-marketplace/redhat-operators-rcl8t" Oct 05 08:15:26 crc kubenswrapper[4846]: I1005 08:15:26.037105 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bn7qs\" (UniqueName: \"kubernetes.io/projected/b0ae695f-c146-46ac-a49b-440929db1ff5-kube-api-access-bn7qs\") pod \"redhat-operators-rcl8t\" (UID: \"b0ae695f-c146-46ac-a49b-440929db1ff5\") " pod="openshift-marketplace/redhat-operators-rcl8t" Oct 05 08:15:26 crc kubenswrapper[4846]: I1005 08:15:26.037159 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b0ae695f-c146-46ac-a49b-440929db1ff5-catalog-content\") pod \"redhat-operators-rcl8t\" (UID: \"b0ae695f-c146-46ac-a49b-440929db1ff5\") " pod="openshift-marketplace/redhat-operators-rcl8t" Oct 05 08:15:26 crc kubenswrapper[4846]: I1005 08:15:26.037288 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b0ae695f-c146-46ac-a49b-440929db1ff5-utilities\") pod \"redhat-operators-rcl8t\" (UID: \"b0ae695f-c146-46ac-a49b-440929db1ff5\") " pod="openshift-marketplace/redhat-operators-rcl8t" Oct 05 08:15:26 crc kubenswrapper[4846]: I1005 08:15:26.037755 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b0ae695f-c146-46ac-a49b-440929db1ff5-catalog-content\") pod \"redhat-operators-rcl8t\" (UID: \"b0ae695f-c146-46ac-a49b-440929db1ff5\") " pod="openshift-marketplace/redhat-operators-rcl8t" Oct 05 08:15:26 crc kubenswrapper[4846]: I1005 08:15:26.037779 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b0ae695f-c146-46ac-a49b-440929db1ff5-utilities\") pod \"redhat-operators-rcl8t\" (UID: \"b0ae695f-c146-46ac-a49b-440929db1ff5\") " pod="openshift-marketplace/redhat-operators-rcl8t" Oct 05 08:15:26 crc kubenswrapper[4846]: I1005 08:15:26.063254 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bn7qs\" (UniqueName: \"kubernetes.io/projected/b0ae695f-c146-46ac-a49b-440929db1ff5-kube-api-access-bn7qs\") pod \"redhat-operators-rcl8t\" (UID: \"b0ae695f-c146-46ac-a49b-440929db1ff5\") " pod="openshift-marketplace/redhat-operators-rcl8t" Oct 05 08:15:26 crc kubenswrapper[4846]: I1005 08:15:26.156677 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rcl8t" Oct 05 08:15:26 crc kubenswrapper[4846]: I1005 08:15:26.415127 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-rcl8t"] Oct 05 08:15:26 crc kubenswrapper[4846]: I1005 08:15:26.601821 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rcl8t" event={"ID":"b0ae695f-c146-46ac-a49b-440929db1ff5","Type":"ContainerStarted","Data":"ff03a9f7c5c885e3b7e898b342563620320f4a78a87d3ffcb3fdef5b29e3d49c"} Oct 05 08:15:26 crc kubenswrapper[4846]: I1005 08:15:26.601866 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rcl8t" event={"ID":"b0ae695f-c146-46ac-a49b-440929db1ff5","Type":"ContainerStarted","Data":"c88790f7e9f361b041ae37b5f346379b12446f523c8459cacfe57c461f120d16"} Oct 05 08:15:26 crc kubenswrapper[4846]: I1005 08:15:26.604324 4846 generic.go:334] "Generic (PLEG): container finished" podID="be1426c3-8e7b-494e-9ef6-3ce557ef6999" containerID="06607cc4bf2d6b43c288d30709c51085325ff68c8b27da805e76d795177262eb" exitCode=0 Oct 05 08:15:26 crc kubenswrapper[4846]: I1005 08:15:26.604395 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xgzh2" event={"ID":"be1426c3-8e7b-494e-9ef6-3ce557ef6999","Type":"ContainerDied","Data":"06607cc4bf2d6b43c288d30709c51085325ff68c8b27da805e76d795177262eb"} Oct 05 08:15:26 crc kubenswrapper[4846]: I1005 08:15:26.609857 4846 generic.go:334] "Generic (PLEG): container finished" podID="1e7701c8-038a-4a7e-9b25-96c059199bba" containerID="6daf0df1174036065afa31dbfba00b9815003c98ae9df135d26c50cca12908d9" exitCode=0 Oct 05 08:15:26 crc kubenswrapper[4846]: I1005 08:15:26.609889 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mbz24" event={"ID":"1e7701c8-038a-4a7e-9b25-96c059199bba","Type":"ContainerDied","Data":"6daf0df1174036065afa31dbfba00b9815003c98ae9df135d26c50cca12908d9"} Oct 05 08:15:27 crc kubenswrapper[4846]: I1005 08:15:27.618307 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mbz24" event={"ID":"1e7701c8-038a-4a7e-9b25-96c059199bba","Type":"ContainerStarted","Data":"b8ae166929fded5140c26d7b2278fb033be3a4e71c6e06c59c2bb77f63a71dd6"} Oct 05 08:15:27 crc kubenswrapper[4846]: I1005 08:15:27.620216 4846 generic.go:334] "Generic (PLEG): container finished" podID="b0ae695f-c146-46ac-a49b-440929db1ff5" containerID="ff03a9f7c5c885e3b7e898b342563620320f4a78a87d3ffcb3fdef5b29e3d49c" exitCode=0 Oct 05 08:15:27 crc kubenswrapper[4846]: I1005 08:15:27.620347 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rcl8t" event={"ID":"b0ae695f-c146-46ac-a49b-440929db1ff5","Type":"ContainerDied","Data":"ff03a9f7c5c885e3b7e898b342563620320f4a78a87d3ffcb3fdef5b29e3d49c"} Oct 05 08:15:27 crc kubenswrapper[4846]: I1005 08:15:27.622358 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xgzh2" event={"ID":"be1426c3-8e7b-494e-9ef6-3ce557ef6999","Type":"ContainerStarted","Data":"07ca0824fea249d693227682bbbbae66e4d51e2bf5d77417f0432159e2aceb2b"} Oct 05 08:15:27 crc kubenswrapper[4846]: I1005 08:15:27.641982 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-mbz24" podStartSLOduration=2.041141195 podStartE2EDuration="4.641956719s" podCreationTimestamp="2025-10-05 08:15:23 +0000 UTC" firstStartedPulling="2025-10-05 08:15:24.569555183 +0000 UTC m=+5246.810407958" lastFinishedPulling="2025-10-05 08:15:27.170370697 +0000 UTC m=+5249.411223482" observedRunningTime="2025-10-05 08:15:27.637821858 +0000 UTC m=+5249.878674643" watchObservedRunningTime="2025-10-05 08:15:27.641956719 +0000 UTC m=+5249.882809504" Oct 05 08:15:27 crc kubenswrapper[4846]: I1005 08:15:27.665347 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-xgzh2" podStartSLOduration=2.272852282 podStartE2EDuration="4.665326285s" podCreationTimestamp="2025-10-05 08:15:23 +0000 UTC" firstStartedPulling="2025-10-05 08:15:24.56793851 +0000 UTC m=+5246.808791285" lastFinishedPulling="2025-10-05 08:15:26.960412513 +0000 UTC m=+5249.201265288" observedRunningTime="2025-10-05 08:15:27.659612812 +0000 UTC m=+5249.900465617" watchObservedRunningTime="2025-10-05 08:15:27.665326285 +0000 UTC m=+5249.906179050" Oct 05 08:15:28 crc kubenswrapper[4846]: I1005 08:15:28.640393 4846 generic.go:334] "Generic (PLEG): container finished" podID="b0ae695f-c146-46ac-a49b-440929db1ff5" containerID="1e372f1d3561a8aeca56ecf971a04d54385410625172688a1d3cfc1f51ebeb04" exitCode=0 Oct 05 08:15:28 crc kubenswrapper[4846]: I1005 08:15:28.642869 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rcl8t" event={"ID":"b0ae695f-c146-46ac-a49b-440929db1ff5","Type":"ContainerDied","Data":"1e372f1d3561a8aeca56ecf971a04d54385410625172688a1d3cfc1f51ebeb04"} Oct 05 08:15:29 crc kubenswrapper[4846]: I1005 08:15:29.651293 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rcl8t" event={"ID":"b0ae695f-c146-46ac-a49b-440929db1ff5","Type":"ContainerStarted","Data":"957fcb8cc94f3fe139574666876e3be7c94a26c2cc943e63d6f7e14d716e4328"} Oct 05 08:15:29 crc kubenswrapper[4846]: I1005 08:15:29.674583 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-rcl8t" podStartSLOduration=2.117215061 podStartE2EDuration="4.674554242s" podCreationTimestamp="2025-10-05 08:15:25 +0000 UTC" firstStartedPulling="2025-10-05 08:15:26.609973836 +0000 UTC m=+5248.850826611" lastFinishedPulling="2025-10-05 08:15:29.167313017 +0000 UTC m=+5251.408165792" observedRunningTime="2025-10-05 08:15:29.669430635 +0000 UTC m=+5251.910283420" watchObservedRunningTime="2025-10-05 08:15:29.674554242 +0000 UTC m=+5251.915407027" Oct 05 08:15:31 crc kubenswrapper[4846]: I1005 08:15:31.369302 4846 scope.go:117] "RemoveContainer" containerID="e89802c66747160446da70dbc0328ef5e9d5603b4c8558d9b3a8cced6ea69467" Oct 05 08:15:31 crc kubenswrapper[4846]: I1005 08:15:31.411357 4846 scope.go:117] "RemoveContainer" containerID="2bb61e3ac89e52c35f391432b40c44b7bb3c7a7c8e6307dbd82465c4f38b2c90" Oct 05 08:15:31 crc kubenswrapper[4846]: I1005 08:15:31.454117 4846 scope.go:117] "RemoveContainer" containerID="ba8753d9a7d9578062a3d1bc4fea0af38c3bae6bbe52523e9a0a59a10da8c886" Oct 05 08:15:31 crc kubenswrapper[4846]: I1005 08:15:31.511285 4846 scope.go:117] "RemoveContainer" containerID="44f8a74c28755cc48927ad4cf2d25620cd6873c2155dfaabe6e8cf0e0b2bca25" Oct 05 08:15:31 crc kubenswrapper[4846]: I1005 08:15:31.556088 4846 scope.go:117] "RemoveContainer" containerID="de7e0ed4f28e168538cad8a0da58698b223ebf7d02900eef61442a3cde6cff4e" Oct 05 08:15:31 crc kubenswrapper[4846]: I1005 08:15:31.592445 4846 scope.go:117] "RemoveContainer" containerID="684ae39745ff6d2990ca068ccd0c568189894228b378e8b8f52ea762a2fc2e61" Oct 05 08:15:31 crc kubenswrapper[4846]: I1005 08:15:31.625407 4846 scope.go:117] "RemoveContainer" containerID="9779d77d4e31773e26dd50d0a2be0bdf70553cc6b89afccf608ff6a584d82ca9" Oct 05 08:15:33 crc kubenswrapper[4846]: I1005 08:15:33.556495 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-mbz24" Oct 05 08:15:33 crc kubenswrapper[4846]: I1005 08:15:33.556999 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-mbz24" Oct 05 08:15:33 crc kubenswrapper[4846]: I1005 08:15:33.639547 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-mbz24" Oct 05 08:15:33 crc kubenswrapper[4846]: I1005 08:15:33.757079 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-mbz24" Oct 05 08:15:33 crc kubenswrapper[4846]: I1005 08:15:33.759382 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-xgzh2" Oct 05 08:15:33 crc kubenswrapper[4846]: I1005 08:15:33.759455 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-xgzh2" Oct 05 08:15:33 crc kubenswrapper[4846]: I1005 08:15:33.823520 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-xgzh2" Oct 05 08:15:34 crc kubenswrapper[4846]: I1005 08:15:34.379411 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-mbz24"] Oct 05 08:15:34 crc kubenswrapper[4846]: I1005 08:15:34.497803 4846 scope.go:117] "RemoveContainer" containerID="8fa72be735dff8e623a4a4958f452896ef79e72dd44c2d714a3019187462de47" Oct 05 08:15:34 crc kubenswrapper[4846]: I1005 08:15:34.791489 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-xgzh2" Oct 05 08:15:35 crc kubenswrapper[4846]: I1005 08:15:35.721846 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" event={"ID":"ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7","Type":"ContainerStarted","Data":"f8ac1a187331f3fff703acc696060a82d728f8e154fe3e11ba6c956d3d906397"} Oct 05 08:15:35 crc kubenswrapper[4846]: I1005 08:15:35.722215 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-mbz24" podUID="1e7701c8-038a-4a7e-9b25-96c059199bba" containerName="registry-server" containerID="cri-o://b8ae166929fded5140c26d7b2278fb033be3a4e71c6e06c59c2bb77f63a71dd6" gracePeriod=2 Oct 05 08:15:36 crc kubenswrapper[4846]: I1005 08:15:36.157117 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-rcl8t" Oct 05 08:15:36 crc kubenswrapper[4846]: I1005 08:15:36.157585 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-rcl8t" Oct 05 08:15:36 crc kubenswrapper[4846]: I1005 08:15:36.173912 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-xgzh2"] Oct 05 08:15:36 crc kubenswrapper[4846]: I1005 08:15:36.229582 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-rcl8t" Oct 05 08:15:36 crc kubenswrapper[4846]: I1005 08:15:36.751856 4846 generic.go:334] "Generic (PLEG): container finished" podID="1e7701c8-038a-4a7e-9b25-96c059199bba" containerID="b8ae166929fded5140c26d7b2278fb033be3a4e71c6e06c59c2bb77f63a71dd6" exitCode=0 Oct 05 08:15:36 crc kubenswrapper[4846]: I1005 08:15:36.752642 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-xgzh2" podUID="be1426c3-8e7b-494e-9ef6-3ce557ef6999" containerName="registry-server" containerID="cri-o://07ca0824fea249d693227682bbbbae66e4d51e2bf5d77417f0432159e2aceb2b" gracePeriod=2 Oct 05 08:15:36 crc kubenswrapper[4846]: I1005 08:15:36.752740 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mbz24" event={"ID":"1e7701c8-038a-4a7e-9b25-96c059199bba","Type":"ContainerDied","Data":"b8ae166929fded5140c26d7b2278fb033be3a4e71c6e06c59c2bb77f63a71dd6"} Oct 05 08:15:36 crc kubenswrapper[4846]: I1005 08:15:36.752778 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mbz24" event={"ID":"1e7701c8-038a-4a7e-9b25-96c059199bba","Type":"ContainerDied","Data":"2c60023001771be953f9a994f812596022f5b38ad87b758f9c85695e98fac6f2"} Oct 05 08:15:36 crc kubenswrapper[4846]: I1005 08:15:36.752792 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2c60023001771be953f9a994f812596022f5b38ad87b758f9c85695e98fac6f2" Oct 05 08:15:36 crc kubenswrapper[4846]: I1005 08:15:36.792345 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mbz24" Oct 05 08:15:36 crc kubenswrapper[4846]: I1005 08:15:36.801564 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-rcl8t" Oct 05 08:15:36 crc kubenswrapper[4846]: I1005 08:15:36.949000 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1e7701c8-038a-4a7e-9b25-96c059199bba-utilities\") pod \"1e7701c8-038a-4a7e-9b25-96c059199bba\" (UID: \"1e7701c8-038a-4a7e-9b25-96c059199bba\") " Oct 05 08:15:36 crc kubenswrapper[4846]: I1005 08:15:36.949144 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dc9j4\" (UniqueName: \"kubernetes.io/projected/1e7701c8-038a-4a7e-9b25-96c059199bba-kube-api-access-dc9j4\") pod \"1e7701c8-038a-4a7e-9b25-96c059199bba\" (UID: \"1e7701c8-038a-4a7e-9b25-96c059199bba\") " Oct 05 08:15:36 crc kubenswrapper[4846]: I1005 08:15:36.949202 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1e7701c8-038a-4a7e-9b25-96c059199bba-catalog-content\") pod \"1e7701c8-038a-4a7e-9b25-96c059199bba\" (UID: \"1e7701c8-038a-4a7e-9b25-96c059199bba\") " Oct 05 08:15:36 crc kubenswrapper[4846]: I1005 08:15:36.950212 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1e7701c8-038a-4a7e-9b25-96c059199bba-utilities" (OuterVolumeSpecName: "utilities") pod "1e7701c8-038a-4a7e-9b25-96c059199bba" (UID: "1e7701c8-038a-4a7e-9b25-96c059199bba"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 08:15:36 crc kubenswrapper[4846]: I1005 08:15:36.950764 4846 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1e7701c8-038a-4a7e-9b25-96c059199bba-utilities\") on node \"crc\" DevicePath \"\"" Oct 05 08:15:36 crc kubenswrapper[4846]: I1005 08:15:36.956091 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1e7701c8-038a-4a7e-9b25-96c059199bba-kube-api-access-dc9j4" (OuterVolumeSpecName: "kube-api-access-dc9j4") pod "1e7701c8-038a-4a7e-9b25-96c059199bba" (UID: "1e7701c8-038a-4a7e-9b25-96c059199bba"). InnerVolumeSpecName "kube-api-access-dc9j4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:15:36 crc kubenswrapper[4846]: I1005 08:15:36.999307 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1e7701c8-038a-4a7e-9b25-96c059199bba-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1e7701c8-038a-4a7e-9b25-96c059199bba" (UID: "1e7701c8-038a-4a7e-9b25-96c059199bba"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 08:15:37 crc kubenswrapper[4846]: I1005 08:15:37.053221 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dc9j4\" (UniqueName: \"kubernetes.io/projected/1e7701c8-038a-4a7e-9b25-96c059199bba-kube-api-access-dc9j4\") on node \"crc\" DevicePath \"\"" Oct 05 08:15:37 crc kubenswrapper[4846]: I1005 08:15:37.053676 4846 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1e7701c8-038a-4a7e-9b25-96c059199bba-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 05 08:15:37 crc kubenswrapper[4846]: I1005 08:15:37.247495 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-xgzh2" Oct 05 08:15:37 crc kubenswrapper[4846]: I1005 08:15:37.358436 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/be1426c3-8e7b-494e-9ef6-3ce557ef6999-catalog-content\") pod \"be1426c3-8e7b-494e-9ef6-3ce557ef6999\" (UID: \"be1426c3-8e7b-494e-9ef6-3ce557ef6999\") " Oct 05 08:15:37 crc kubenswrapper[4846]: I1005 08:15:37.358600 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/be1426c3-8e7b-494e-9ef6-3ce557ef6999-utilities\") pod \"be1426c3-8e7b-494e-9ef6-3ce557ef6999\" (UID: \"be1426c3-8e7b-494e-9ef6-3ce557ef6999\") " Oct 05 08:15:37 crc kubenswrapper[4846]: I1005 08:15:37.358629 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wrl7p\" (UniqueName: \"kubernetes.io/projected/be1426c3-8e7b-494e-9ef6-3ce557ef6999-kube-api-access-wrl7p\") pod \"be1426c3-8e7b-494e-9ef6-3ce557ef6999\" (UID: \"be1426c3-8e7b-494e-9ef6-3ce557ef6999\") " Oct 05 08:15:37 crc kubenswrapper[4846]: I1005 08:15:37.359555 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/be1426c3-8e7b-494e-9ef6-3ce557ef6999-utilities" (OuterVolumeSpecName: "utilities") pod "be1426c3-8e7b-494e-9ef6-3ce557ef6999" (UID: "be1426c3-8e7b-494e-9ef6-3ce557ef6999"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 08:15:37 crc kubenswrapper[4846]: I1005 08:15:37.363268 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/be1426c3-8e7b-494e-9ef6-3ce557ef6999-kube-api-access-wrl7p" (OuterVolumeSpecName: "kube-api-access-wrl7p") pod "be1426c3-8e7b-494e-9ef6-3ce557ef6999" (UID: "be1426c3-8e7b-494e-9ef6-3ce557ef6999"). InnerVolumeSpecName "kube-api-access-wrl7p". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:15:37 crc kubenswrapper[4846]: I1005 08:15:37.425730 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/be1426c3-8e7b-494e-9ef6-3ce557ef6999-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "be1426c3-8e7b-494e-9ef6-3ce557ef6999" (UID: "be1426c3-8e7b-494e-9ef6-3ce557ef6999"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 08:15:37 crc kubenswrapper[4846]: I1005 08:15:37.460892 4846 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/be1426c3-8e7b-494e-9ef6-3ce557ef6999-utilities\") on node \"crc\" DevicePath \"\"" Oct 05 08:15:37 crc kubenswrapper[4846]: I1005 08:15:37.460924 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wrl7p\" (UniqueName: \"kubernetes.io/projected/be1426c3-8e7b-494e-9ef6-3ce557ef6999-kube-api-access-wrl7p\") on node \"crc\" DevicePath \"\"" Oct 05 08:15:37 crc kubenswrapper[4846]: I1005 08:15:37.460936 4846 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/be1426c3-8e7b-494e-9ef6-3ce557ef6999-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 05 08:15:37 crc kubenswrapper[4846]: I1005 08:15:37.765522 4846 generic.go:334] "Generic (PLEG): container finished" podID="be1426c3-8e7b-494e-9ef6-3ce557ef6999" containerID="07ca0824fea249d693227682bbbbae66e4d51e2bf5d77417f0432159e2aceb2b" exitCode=0 Oct 05 08:15:37 crc kubenswrapper[4846]: I1005 08:15:37.765547 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xgzh2" event={"ID":"be1426c3-8e7b-494e-9ef6-3ce557ef6999","Type":"ContainerDied","Data":"07ca0824fea249d693227682bbbbae66e4d51e2bf5d77417f0432159e2aceb2b"} Oct 05 08:15:37 crc kubenswrapper[4846]: I1005 08:15:37.765600 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xgzh2" event={"ID":"be1426c3-8e7b-494e-9ef6-3ce557ef6999","Type":"ContainerDied","Data":"45e8488ecac6168e9395ea3a126dbf5dc49874cec910217227d06bf521cb30fd"} Oct 05 08:15:37 crc kubenswrapper[4846]: I1005 08:15:37.765612 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mbz24" Oct 05 08:15:37 crc kubenswrapper[4846]: I1005 08:15:37.765619 4846 scope.go:117] "RemoveContainer" containerID="07ca0824fea249d693227682bbbbae66e4d51e2bf5d77417f0432159e2aceb2b" Oct 05 08:15:37 crc kubenswrapper[4846]: I1005 08:15:37.765835 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-xgzh2" Oct 05 08:15:37 crc kubenswrapper[4846]: I1005 08:15:37.799074 4846 scope.go:117] "RemoveContainer" containerID="06607cc4bf2d6b43c288d30709c51085325ff68c8b27da805e76d795177262eb" Oct 05 08:15:37 crc kubenswrapper[4846]: I1005 08:15:37.810033 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-xgzh2"] Oct 05 08:15:37 crc kubenswrapper[4846]: I1005 08:15:37.832744 4846 scope.go:117] "RemoveContainer" containerID="076676db43db829bbf778305a914211f9be97efc4aae42987575a9362d313700" Oct 05 08:15:37 crc kubenswrapper[4846]: I1005 08:15:37.835294 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-xgzh2"] Oct 05 08:15:37 crc kubenswrapper[4846]: I1005 08:15:37.845096 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-mbz24"] Oct 05 08:15:37 crc kubenswrapper[4846]: I1005 08:15:37.852661 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-mbz24"] Oct 05 08:15:37 crc kubenswrapper[4846]: I1005 08:15:37.855042 4846 scope.go:117] "RemoveContainer" containerID="07ca0824fea249d693227682bbbbae66e4d51e2bf5d77417f0432159e2aceb2b" Oct 05 08:15:37 crc kubenswrapper[4846]: E1005 08:15:37.855477 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"07ca0824fea249d693227682bbbbae66e4d51e2bf5d77417f0432159e2aceb2b\": container with ID starting with 07ca0824fea249d693227682bbbbae66e4d51e2bf5d77417f0432159e2aceb2b not found: ID does not exist" containerID="07ca0824fea249d693227682bbbbae66e4d51e2bf5d77417f0432159e2aceb2b" Oct 05 08:15:37 crc kubenswrapper[4846]: I1005 08:15:37.855512 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"07ca0824fea249d693227682bbbbae66e4d51e2bf5d77417f0432159e2aceb2b"} err="failed to get container status \"07ca0824fea249d693227682bbbbae66e4d51e2bf5d77417f0432159e2aceb2b\": rpc error: code = NotFound desc = could not find container \"07ca0824fea249d693227682bbbbae66e4d51e2bf5d77417f0432159e2aceb2b\": container with ID starting with 07ca0824fea249d693227682bbbbae66e4d51e2bf5d77417f0432159e2aceb2b not found: ID does not exist" Oct 05 08:15:37 crc kubenswrapper[4846]: I1005 08:15:37.855536 4846 scope.go:117] "RemoveContainer" containerID="06607cc4bf2d6b43c288d30709c51085325ff68c8b27da805e76d795177262eb" Oct 05 08:15:37 crc kubenswrapper[4846]: E1005 08:15:37.856112 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"06607cc4bf2d6b43c288d30709c51085325ff68c8b27da805e76d795177262eb\": container with ID starting with 06607cc4bf2d6b43c288d30709c51085325ff68c8b27da805e76d795177262eb not found: ID does not exist" containerID="06607cc4bf2d6b43c288d30709c51085325ff68c8b27da805e76d795177262eb" Oct 05 08:15:37 crc kubenswrapper[4846]: I1005 08:15:37.856197 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"06607cc4bf2d6b43c288d30709c51085325ff68c8b27da805e76d795177262eb"} err="failed to get container status \"06607cc4bf2d6b43c288d30709c51085325ff68c8b27da805e76d795177262eb\": rpc error: code = NotFound desc = could not find container \"06607cc4bf2d6b43c288d30709c51085325ff68c8b27da805e76d795177262eb\": container with ID starting with 06607cc4bf2d6b43c288d30709c51085325ff68c8b27da805e76d795177262eb not found: ID does not exist" Oct 05 08:15:37 crc kubenswrapper[4846]: I1005 08:15:37.856239 4846 scope.go:117] "RemoveContainer" containerID="076676db43db829bbf778305a914211f9be97efc4aae42987575a9362d313700" Oct 05 08:15:37 crc kubenswrapper[4846]: E1005 08:15:37.856614 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"076676db43db829bbf778305a914211f9be97efc4aae42987575a9362d313700\": container with ID starting with 076676db43db829bbf778305a914211f9be97efc4aae42987575a9362d313700 not found: ID does not exist" containerID="076676db43db829bbf778305a914211f9be97efc4aae42987575a9362d313700" Oct 05 08:15:37 crc kubenswrapper[4846]: I1005 08:15:37.856638 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"076676db43db829bbf778305a914211f9be97efc4aae42987575a9362d313700"} err="failed to get container status \"076676db43db829bbf778305a914211f9be97efc4aae42987575a9362d313700\": rpc error: code = NotFound desc = could not find container \"076676db43db829bbf778305a914211f9be97efc4aae42987575a9362d313700\": container with ID starting with 076676db43db829bbf778305a914211f9be97efc4aae42987575a9362d313700 not found: ID does not exist" Oct 05 08:15:38 crc kubenswrapper[4846]: I1005 08:15:38.517622 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1e7701c8-038a-4a7e-9b25-96c059199bba" path="/var/lib/kubelet/pods/1e7701c8-038a-4a7e-9b25-96c059199bba/volumes" Oct 05 08:15:38 crc kubenswrapper[4846]: I1005 08:15:38.519730 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="be1426c3-8e7b-494e-9ef6-3ce557ef6999" path="/var/lib/kubelet/pods/be1426c3-8e7b-494e-9ef6-3ce557ef6999/volumes" Oct 05 08:15:39 crc kubenswrapper[4846]: I1005 08:15:39.174594 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-rcl8t"] Oct 05 08:15:39 crc kubenswrapper[4846]: I1005 08:15:39.174865 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-rcl8t" podUID="b0ae695f-c146-46ac-a49b-440929db1ff5" containerName="registry-server" containerID="cri-o://957fcb8cc94f3fe139574666876e3be7c94a26c2cc943e63d6f7e14d716e4328" gracePeriod=2 Oct 05 08:15:39 crc kubenswrapper[4846]: I1005 08:15:39.794484 4846 generic.go:334] "Generic (PLEG): container finished" podID="b0ae695f-c146-46ac-a49b-440929db1ff5" containerID="957fcb8cc94f3fe139574666876e3be7c94a26c2cc943e63d6f7e14d716e4328" exitCode=0 Oct 05 08:15:39 crc kubenswrapper[4846]: I1005 08:15:39.794668 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rcl8t" event={"ID":"b0ae695f-c146-46ac-a49b-440929db1ff5","Type":"ContainerDied","Data":"957fcb8cc94f3fe139574666876e3be7c94a26c2cc943e63d6f7e14d716e4328"} Oct 05 08:15:40 crc kubenswrapper[4846]: I1005 08:15:40.319915 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rcl8t" Oct 05 08:15:40 crc kubenswrapper[4846]: I1005 08:15:40.419102 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b0ae695f-c146-46ac-a49b-440929db1ff5-utilities\") pod \"b0ae695f-c146-46ac-a49b-440929db1ff5\" (UID: \"b0ae695f-c146-46ac-a49b-440929db1ff5\") " Oct 05 08:15:40 crc kubenswrapper[4846]: I1005 08:15:40.420769 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b0ae695f-c146-46ac-a49b-440929db1ff5-utilities" (OuterVolumeSpecName: "utilities") pod "b0ae695f-c146-46ac-a49b-440929db1ff5" (UID: "b0ae695f-c146-46ac-a49b-440929db1ff5"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 08:15:40 crc kubenswrapper[4846]: I1005 08:15:40.520085 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b0ae695f-c146-46ac-a49b-440929db1ff5-catalog-content\") pod \"b0ae695f-c146-46ac-a49b-440929db1ff5\" (UID: \"b0ae695f-c146-46ac-a49b-440929db1ff5\") " Oct 05 08:15:40 crc kubenswrapper[4846]: I1005 08:15:40.520224 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bn7qs\" (UniqueName: \"kubernetes.io/projected/b0ae695f-c146-46ac-a49b-440929db1ff5-kube-api-access-bn7qs\") pod \"b0ae695f-c146-46ac-a49b-440929db1ff5\" (UID: \"b0ae695f-c146-46ac-a49b-440929db1ff5\") " Oct 05 08:15:40 crc kubenswrapper[4846]: I1005 08:15:40.521421 4846 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b0ae695f-c146-46ac-a49b-440929db1ff5-utilities\") on node \"crc\" DevicePath \"\"" Oct 05 08:15:40 crc kubenswrapper[4846]: I1005 08:15:40.529359 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b0ae695f-c146-46ac-a49b-440929db1ff5-kube-api-access-bn7qs" (OuterVolumeSpecName: "kube-api-access-bn7qs") pod "b0ae695f-c146-46ac-a49b-440929db1ff5" (UID: "b0ae695f-c146-46ac-a49b-440929db1ff5"). InnerVolumeSpecName "kube-api-access-bn7qs". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:15:40 crc kubenswrapper[4846]: I1005 08:15:40.597147 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b0ae695f-c146-46ac-a49b-440929db1ff5-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b0ae695f-c146-46ac-a49b-440929db1ff5" (UID: "b0ae695f-c146-46ac-a49b-440929db1ff5"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 08:15:40 crc kubenswrapper[4846]: I1005 08:15:40.623461 4846 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b0ae695f-c146-46ac-a49b-440929db1ff5-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 05 08:15:40 crc kubenswrapper[4846]: I1005 08:15:40.623486 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bn7qs\" (UniqueName: \"kubernetes.io/projected/b0ae695f-c146-46ac-a49b-440929db1ff5-kube-api-access-bn7qs\") on node \"crc\" DevicePath \"\"" Oct 05 08:15:40 crc kubenswrapper[4846]: I1005 08:15:40.807649 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rcl8t" event={"ID":"b0ae695f-c146-46ac-a49b-440929db1ff5","Type":"ContainerDied","Data":"c88790f7e9f361b041ae37b5f346379b12446f523c8459cacfe57c461f120d16"} Oct 05 08:15:40 crc kubenswrapper[4846]: I1005 08:15:40.807700 4846 scope.go:117] "RemoveContainer" containerID="957fcb8cc94f3fe139574666876e3be7c94a26c2cc943e63d6f7e14d716e4328" Oct 05 08:15:40 crc kubenswrapper[4846]: I1005 08:15:40.807836 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rcl8t" Oct 05 08:15:40 crc kubenswrapper[4846]: I1005 08:15:40.848802 4846 scope.go:117] "RemoveContainer" containerID="1e372f1d3561a8aeca56ecf971a04d54385410625172688a1d3cfc1f51ebeb04" Oct 05 08:15:40 crc kubenswrapper[4846]: I1005 08:15:40.853845 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-rcl8t"] Oct 05 08:15:40 crc kubenswrapper[4846]: I1005 08:15:40.862701 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-rcl8t"] Oct 05 08:15:40 crc kubenswrapper[4846]: I1005 08:15:40.881765 4846 scope.go:117] "RemoveContainer" containerID="ff03a9f7c5c885e3b7e898b342563620320f4a78a87d3ffcb3fdef5b29e3d49c" Oct 05 08:15:42 crc kubenswrapper[4846]: I1005 08:15:42.507334 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b0ae695f-c146-46ac-a49b-440929db1ff5" path="/var/lib/kubelet/pods/b0ae695f-c146-46ac-a49b-440929db1ff5/volumes" Oct 05 08:15:49 crc kubenswrapper[4846]: I1005 08:15:49.489904 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-create-c2xrh"] Oct 05 08:15:49 crc kubenswrapper[4846]: E1005 08:15:49.491072 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b0ae695f-c146-46ac-a49b-440929db1ff5" containerName="extract-content" Oct 05 08:15:49 crc kubenswrapper[4846]: I1005 08:15:49.491093 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="b0ae695f-c146-46ac-a49b-440929db1ff5" containerName="extract-content" Oct 05 08:15:49 crc kubenswrapper[4846]: E1005 08:15:49.491128 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="be1426c3-8e7b-494e-9ef6-3ce557ef6999" containerName="extract-content" Oct 05 08:15:49 crc kubenswrapper[4846]: I1005 08:15:49.491140 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="be1426c3-8e7b-494e-9ef6-3ce557ef6999" containerName="extract-content" Oct 05 08:15:49 crc kubenswrapper[4846]: E1005 08:15:49.491154 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1e7701c8-038a-4a7e-9b25-96c059199bba" containerName="extract-content" Oct 05 08:15:49 crc kubenswrapper[4846]: I1005 08:15:49.491166 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="1e7701c8-038a-4a7e-9b25-96c059199bba" containerName="extract-content" Oct 05 08:15:49 crc kubenswrapper[4846]: E1005 08:15:49.491212 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b0ae695f-c146-46ac-a49b-440929db1ff5" containerName="registry-server" Oct 05 08:15:49 crc kubenswrapper[4846]: I1005 08:15:49.491224 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="b0ae695f-c146-46ac-a49b-440929db1ff5" containerName="registry-server" Oct 05 08:15:49 crc kubenswrapper[4846]: E1005 08:15:49.491260 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="be1426c3-8e7b-494e-9ef6-3ce557ef6999" containerName="extract-utilities" Oct 05 08:15:49 crc kubenswrapper[4846]: I1005 08:15:49.491271 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="be1426c3-8e7b-494e-9ef6-3ce557ef6999" containerName="extract-utilities" Oct 05 08:15:49 crc kubenswrapper[4846]: E1005 08:15:49.491284 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1e7701c8-038a-4a7e-9b25-96c059199bba" containerName="registry-server" Oct 05 08:15:49 crc kubenswrapper[4846]: I1005 08:15:49.491295 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="1e7701c8-038a-4a7e-9b25-96c059199bba" containerName="registry-server" Oct 05 08:15:49 crc kubenswrapper[4846]: E1005 08:15:49.491327 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b0ae695f-c146-46ac-a49b-440929db1ff5" containerName="extract-utilities" Oct 05 08:15:49 crc kubenswrapper[4846]: I1005 08:15:49.491341 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="b0ae695f-c146-46ac-a49b-440929db1ff5" containerName="extract-utilities" Oct 05 08:15:49 crc kubenswrapper[4846]: E1005 08:15:49.491365 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1e7701c8-038a-4a7e-9b25-96c059199bba" containerName="extract-utilities" Oct 05 08:15:49 crc kubenswrapper[4846]: I1005 08:15:49.491377 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="1e7701c8-038a-4a7e-9b25-96c059199bba" containerName="extract-utilities" Oct 05 08:15:49 crc kubenswrapper[4846]: E1005 08:15:49.491403 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="be1426c3-8e7b-494e-9ef6-3ce557ef6999" containerName="registry-server" Oct 05 08:15:49 crc kubenswrapper[4846]: I1005 08:15:49.491415 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="be1426c3-8e7b-494e-9ef6-3ce557ef6999" containerName="registry-server" Oct 05 08:15:49 crc kubenswrapper[4846]: I1005 08:15:49.491671 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="b0ae695f-c146-46ac-a49b-440929db1ff5" containerName="registry-server" Oct 05 08:15:49 crc kubenswrapper[4846]: I1005 08:15:49.491704 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="1e7701c8-038a-4a7e-9b25-96c059199bba" containerName="registry-server" Oct 05 08:15:49 crc kubenswrapper[4846]: I1005 08:15:49.491727 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="be1426c3-8e7b-494e-9ef6-3ce557ef6999" containerName="registry-server" Oct 05 08:15:49 crc kubenswrapper[4846]: I1005 08:15:49.492668 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-c2xrh" Oct 05 08:15:49 crc kubenswrapper[4846]: I1005 08:15:49.501043 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-c2xrh"] Oct 05 08:15:49 crc kubenswrapper[4846]: I1005 08:15:49.518658 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8q2bj\" (UniqueName: \"kubernetes.io/projected/e39a19ff-3716-4940-9945-3602c8cbcaf2-kube-api-access-8q2bj\") pod \"barbican-db-create-c2xrh\" (UID: \"e39a19ff-3716-4940-9945-3602c8cbcaf2\") " pod="openstack/barbican-db-create-c2xrh" Oct 05 08:15:49 crc kubenswrapper[4846]: I1005 08:15:49.620537 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8q2bj\" (UniqueName: \"kubernetes.io/projected/e39a19ff-3716-4940-9945-3602c8cbcaf2-kube-api-access-8q2bj\") pod \"barbican-db-create-c2xrh\" (UID: \"e39a19ff-3716-4940-9945-3602c8cbcaf2\") " pod="openstack/barbican-db-create-c2xrh" Oct 05 08:15:49 crc kubenswrapper[4846]: I1005 08:15:49.644243 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8q2bj\" (UniqueName: \"kubernetes.io/projected/e39a19ff-3716-4940-9945-3602c8cbcaf2-kube-api-access-8q2bj\") pod \"barbican-db-create-c2xrh\" (UID: \"e39a19ff-3716-4940-9945-3602c8cbcaf2\") " pod="openstack/barbican-db-create-c2xrh" Oct 05 08:15:49 crc kubenswrapper[4846]: I1005 08:15:49.816541 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-c2xrh" Oct 05 08:15:50 crc kubenswrapper[4846]: I1005 08:15:50.281361 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-c2xrh"] Oct 05 08:15:50 crc kubenswrapper[4846]: I1005 08:15:50.913252 4846 generic.go:334] "Generic (PLEG): container finished" podID="e39a19ff-3716-4940-9945-3602c8cbcaf2" containerID="0b0a481e09ec59c44725cebd2f7c6a88a11897b33401732318eb9b88a948f74f" exitCode=0 Oct 05 08:15:50 crc kubenswrapper[4846]: I1005 08:15:50.913349 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-c2xrh" event={"ID":"e39a19ff-3716-4940-9945-3602c8cbcaf2","Type":"ContainerDied","Data":"0b0a481e09ec59c44725cebd2f7c6a88a11897b33401732318eb9b88a948f74f"} Oct 05 08:15:50 crc kubenswrapper[4846]: I1005 08:15:50.914931 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-c2xrh" event={"ID":"e39a19ff-3716-4940-9945-3602c8cbcaf2","Type":"ContainerStarted","Data":"20d0d3290ced2fb95e0c7c06400ff113bc20464f170cf02101d800f0c7013496"} Oct 05 08:15:52 crc kubenswrapper[4846]: I1005 08:15:52.281098 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-c2xrh" Oct 05 08:15:52 crc kubenswrapper[4846]: I1005 08:15:52.372094 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8q2bj\" (UniqueName: \"kubernetes.io/projected/e39a19ff-3716-4940-9945-3602c8cbcaf2-kube-api-access-8q2bj\") pod \"e39a19ff-3716-4940-9945-3602c8cbcaf2\" (UID: \"e39a19ff-3716-4940-9945-3602c8cbcaf2\") " Oct 05 08:15:52 crc kubenswrapper[4846]: I1005 08:15:52.377145 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e39a19ff-3716-4940-9945-3602c8cbcaf2-kube-api-access-8q2bj" (OuterVolumeSpecName: "kube-api-access-8q2bj") pod "e39a19ff-3716-4940-9945-3602c8cbcaf2" (UID: "e39a19ff-3716-4940-9945-3602c8cbcaf2"). InnerVolumeSpecName "kube-api-access-8q2bj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:15:52 crc kubenswrapper[4846]: I1005 08:15:52.476037 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8q2bj\" (UniqueName: \"kubernetes.io/projected/e39a19ff-3716-4940-9945-3602c8cbcaf2-kube-api-access-8q2bj\") on node \"crc\" DevicePath \"\"" Oct 05 08:15:52 crc kubenswrapper[4846]: I1005 08:15:52.933347 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-c2xrh" event={"ID":"e39a19ff-3716-4940-9945-3602c8cbcaf2","Type":"ContainerDied","Data":"20d0d3290ced2fb95e0c7c06400ff113bc20464f170cf02101d800f0c7013496"} Oct 05 08:15:52 crc kubenswrapper[4846]: I1005 08:15:52.933395 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="20d0d3290ced2fb95e0c7c06400ff113bc20464f170cf02101d800f0c7013496" Oct 05 08:15:52 crc kubenswrapper[4846]: I1005 08:15:52.933409 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-c2xrh" Oct 05 08:15:59 crc kubenswrapper[4846]: I1005 08:15:59.617457 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-57a8-account-create-tjbkx"] Oct 05 08:15:59 crc kubenswrapper[4846]: E1005 08:15:59.620467 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e39a19ff-3716-4940-9945-3602c8cbcaf2" containerName="mariadb-database-create" Oct 05 08:15:59 crc kubenswrapper[4846]: I1005 08:15:59.620518 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="e39a19ff-3716-4940-9945-3602c8cbcaf2" containerName="mariadb-database-create" Oct 05 08:15:59 crc kubenswrapper[4846]: I1005 08:15:59.621000 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="e39a19ff-3716-4940-9945-3602c8cbcaf2" containerName="mariadb-database-create" Oct 05 08:15:59 crc kubenswrapper[4846]: I1005 08:15:59.622203 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-57a8-account-create-tjbkx" Oct 05 08:15:59 crc kubenswrapper[4846]: I1005 08:15:59.628875 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-db-secret" Oct 05 08:15:59 crc kubenswrapper[4846]: I1005 08:15:59.630578 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-57a8-account-create-tjbkx"] Oct 05 08:15:59 crc kubenswrapper[4846]: I1005 08:15:59.716888 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kx5s7\" (UniqueName: \"kubernetes.io/projected/674af39b-f67b-4e1c-872b-4115a622dc8c-kube-api-access-kx5s7\") pod \"barbican-57a8-account-create-tjbkx\" (UID: \"674af39b-f67b-4e1c-872b-4115a622dc8c\") " pod="openstack/barbican-57a8-account-create-tjbkx" Oct 05 08:15:59 crc kubenswrapper[4846]: I1005 08:15:59.818348 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kx5s7\" (UniqueName: \"kubernetes.io/projected/674af39b-f67b-4e1c-872b-4115a622dc8c-kube-api-access-kx5s7\") pod \"barbican-57a8-account-create-tjbkx\" (UID: \"674af39b-f67b-4e1c-872b-4115a622dc8c\") " pod="openstack/barbican-57a8-account-create-tjbkx" Oct 05 08:15:59 crc kubenswrapper[4846]: I1005 08:15:59.853492 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kx5s7\" (UniqueName: \"kubernetes.io/projected/674af39b-f67b-4e1c-872b-4115a622dc8c-kube-api-access-kx5s7\") pod \"barbican-57a8-account-create-tjbkx\" (UID: \"674af39b-f67b-4e1c-872b-4115a622dc8c\") " pod="openstack/barbican-57a8-account-create-tjbkx" Oct 05 08:15:59 crc kubenswrapper[4846]: I1005 08:15:59.960895 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-57a8-account-create-tjbkx" Oct 05 08:16:00 crc kubenswrapper[4846]: I1005 08:16:00.573723 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-57a8-account-create-tjbkx"] Oct 05 08:16:01 crc kubenswrapper[4846]: I1005 08:16:01.025087 4846 generic.go:334] "Generic (PLEG): container finished" podID="674af39b-f67b-4e1c-872b-4115a622dc8c" containerID="05037c12b243f1732aa2a993b48998c6aa0359cbab8104a684ee66ccdb12ea92" exitCode=0 Oct 05 08:16:01 crc kubenswrapper[4846]: I1005 08:16:01.025253 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-57a8-account-create-tjbkx" event={"ID":"674af39b-f67b-4e1c-872b-4115a622dc8c","Type":"ContainerDied","Data":"05037c12b243f1732aa2a993b48998c6aa0359cbab8104a684ee66ccdb12ea92"} Oct 05 08:16:01 crc kubenswrapper[4846]: I1005 08:16:01.026625 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-57a8-account-create-tjbkx" event={"ID":"674af39b-f67b-4e1c-872b-4115a622dc8c","Type":"ContainerStarted","Data":"13f0b88efe9b7e131b43bb59488a34f7c387c3b5dc058abd34711432ee106f92"} Oct 05 08:16:02 crc kubenswrapper[4846]: I1005 08:16:02.436524 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-57a8-account-create-tjbkx" Oct 05 08:16:02 crc kubenswrapper[4846]: I1005 08:16:02.568764 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kx5s7\" (UniqueName: \"kubernetes.io/projected/674af39b-f67b-4e1c-872b-4115a622dc8c-kube-api-access-kx5s7\") pod \"674af39b-f67b-4e1c-872b-4115a622dc8c\" (UID: \"674af39b-f67b-4e1c-872b-4115a622dc8c\") " Oct 05 08:16:02 crc kubenswrapper[4846]: I1005 08:16:02.578094 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/674af39b-f67b-4e1c-872b-4115a622dc8c-kube-api-access-kx5s7" (OuterVolumeSpecName: "kube-api-access-kx5s7") pod "674af39b-f67b-4e1c-872b-4115a622dc8c" (UID: "674af39b-f67b-4e1c-872b-4115a622dc8c"). InnerVolumeSpecName "kube-api-access-kx5s7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:16:02 crc kubenswrapper[4846]: I1005 08:16:02.671857 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kx5s7\" (UniqueName: \"kubernetes.io/projected/674af39b-f67b-4e1c-872b-4115a622dc8c-kube-api-access-kx5s7\") on node \"crc\" DevicePath \"\"" Oct 05 08:16:03 crc kubenswrapper[4846]: I1005 08:16:03.048163 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-57a8-account-create-tjbkx" event={"ID":"674af39b-f67b-4e1c-872b-4115a622dc8c","Type":"ContainerDied","Data":"13f0b88efe9b7e131b43bb59488a34f7c387c3b5dc058abd34711432ee106f92"} Oct 05 08:16:03 crc kubenswrapper[4846]: I1005 08:16:03.048765 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="13f0b88efe9b7e131b43bb59488a34f7c387c3b5dc058abd34711432ee106f92" Oct 05 08:16:03 crc kubenswrapper[4846]: I1005 08:16:03.048886 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-57a8-account-create-tjbkx" Oct 05 08:16:04 crc kubenswrapper[4846]: I1005 08:16:04.875295 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-sync-n84rx"] Oct 05 08:16:04 crc kubenswrapper[4846]: E1005 08:16:04.875984 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="674af39b-f67b-4e1c-872b-4115a622dc8c" containerName="mariadb-account-create" Oct 05 08:16:04 crc kubenswrapper[4846]: I1005 08:16:04.876000 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="674af39b-f67b-4e1c-872b-4115a622dc8c" containerName="mariadb-account-create" Oct 05 08:16:04 crc kubenswrapper[4846]: I1005 08:16:04.876199 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="674af39b-f67b-4e1c-872b-4115a622dc8c" containerName="mariadb-account-create" Oct 05 08:16:04 crc kubenswrapper[4846]: I1005 08:16:04.876799 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-n84rx" Oct 05 08:16:04 crc kubenswrapper[4846]: I1005 08:16:04.881022 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-ql2bq" Oct 05 08:16:04 crc kubenswrapper[4846]: I1005 08:16:04.881132 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Oct 05 08:16:04 crc kubenswrapper[4846]: I1005 08:16:04.885042 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-n84rx"] Oct 05 08:16:05 crc kubenswrapper[4846]: I1005 08:16:05.020195 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nn79q\" (UniqueName: \"kubernetes.io/projected/f92090a2-228b-41f7-b97b-11df5415877a-kube-api-access-nn79q\") pod \"barbican-db-sync-n84rx\" (UID: \"f92090a2-228b-41f7-b97b-11df5415877a\") " pod="openstack/barbican-db-sync-n84rx" Oct 05 08:16:05 crc kubenswrapper[4846]: I1005 08:16:05.020282 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/f92090a2-228b-41f7-b97b-11df5415877a-db-sync-config-data\") pod \"barbican-db-sync-n84rx\" (UID: \"f92090a2-228b-41f7-b97b-11df5415877a\") " pod="openstack/barbican-db-sync-n84rx" Oct 05 08:16:05 crc kubenswrapper[4846]: I1005 08:16:05.020413 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f92090a2-228b-41f7-b97b-11df5415877a-combined-ca-bundle\") pod \"barbican-db-sync-n84rx\" (UID: \"f92090a2-228b-41f7-b97b-11df5415877a\") " pod="openstack/barbican-db-sync-n84rx" Oct 05 08:16:05 crc kubenswrapper[4846]: I1005 08:16:05.122720 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f92090a2-228b-41f7-b97b-11df5415877a-combined-ca-bundle\") pod \"barbican-db-sync-n84rx\" (UID: \"f92090a2-228b-41f7-b97b-11df5415877a\") " pod="openstack/barbican-db-sync-n84rx" Oct 05 08:16:05 crc kubenswrapper[4846]: I1005 08:16:05.122910 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nn79q\" (UniqueName: \"kubernetes.io/projected/f92090a2-228b-41f7-b97b-11df5415877a-kube-api-access-nn79q\") pod \"barbican-db-sync-n84rx\" (UID: \"f92090a2-228b-41f7-b97b-11df5415877a\") " pod="openstack/barbican-db-sync-n84rx" Oct 05 08:16:05 crc kubenswrapper[4846]: I1005 08:16:05.123032 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/f92090a2-228b-41f7-b97b-11df5415877a-db-sync-config-data\") pod \"barbican-db-sync-n84rx\" (UID: \"f92090a2-228b-41f7-b97b-11df5415877a\") " pod="openstack/barbican-db-sync-n84rx" Oct 05 08:16:05 crc kubenswrapper[4846]: I1005 08:16:05.131930 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/f92090a2-228b-41f7-b97b-11df5415877a-db-sync-config-data\") pod \"barbican-db-sync-n84rx\" (UID: \"f92090a2-228b-41f7-b97b-11df5415877a\") " pod="openstack/barbican-db-sync-n84rx" Oct 05 08:16:05 crc kubenswrapper[4846]: I1005 08:16:05.141356 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f92090a2-228b-41f7-b97b-11df5415877a-combined-ca-bundle\") pod \"barbican-db-sync-n84rx\" (UID: \"f92090a2-228b-41f7-b97b-11df5415877a\") " pod="openstack/barbican-db-sync-n84rx" Oct 05 08:16:05 crc kubenswrapper[4846]: I1005 08:16:05.163622 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nn79q\" (UniqueName: \"kubernetes.io/projected/f92090a2-228b-41f7-b97b-11df5415877a-kube-api-access-nn79q\") pod \"barbican-db-sync-n84rx\" (UID: \"f92090a2-228b-41f7-b97b-11df5415877a\") " pod="openstack/barbican-db-sync-n84rx" Oct 05 08:16:05 crc kubenswrapper[4846]: I1005 08:16:05.206781 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-n84rx" Oct 05 08:16:05 crc kubenswrapper[4846]: I1005 08:16:05.662776 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-n84rx"] Oct 05 08:16:05 crc kubenswrapper[4846]: W1005 08:16:05.670595 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf92090a2_228b_41f7_b97b_11df5415877a.slice/crio-ed623649739cbac37d0d720644a1201ae30943cf7c80ec33ee0089d7ff20bff7 WatchSource:0}: Error finding container ed623649739cbac37d0d720644a1201ae30943cf7c80ec33ee0089d7ff20bff7: Status 404 returned error can't find the container with id ed623649739cbac37d0d720644a1201ae30943cf7c80ec33ee0089d7ff20bff7 Oct 05 08:16:06 crc kubenswrapper[4846]: I1005 08:16:06.083482 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-n84rx" event={"ID":"f92090a2-228b-41f7-b97b-11df5415877a","Type":"ContainerStarted","Data":"ed623649739cbac37d0d720644a1201ae30943cf7c80ec33ee0089d7ff20bff7"} Oct 05 08:16:11 crc kubenswrapper[4846]: I1005 08:16:11.142594 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-n84rx" event={"ID":"f92090a2-228b-41f7-b97b-11df5415877a","Type":"ContainerStarted","Data":"786aa4b04ab3c851cfcfe099c1b123e134328bd98f3a0fb361907803d7ff58ab"} Oct 05 08:16:11 crc kubenswrapper[4846]: I1005 08:16:11.170868 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-sync-n84rx" podStartSLOduration=2.528742798 podStartE2EDuration="7.170842489s" podCreationTimestamp="2025-10-05 08:16:04 +0000 UTC" firstStartedPulling="2025-10-05 08:16:05.673564942 +0000 UTC m=+5287.914417727" lastFinishedPulling="2025-10-05 08:16:10.315664643 +0000 UTC m=+5292.556517418" observedRunningTime="2025-10-05 08:16:11.162992739 +0000 UTC m=+5293.403845554" watchObservedRunningTime="2025-10-05 08:16:11.170842489 +0000 UTC m=+5293.411695304" Oct 05 08:16:12 crc kubenswrapper[4846]: I1005 08:16:12.157250 4846 generic.go:334] "Generic (PLEG): container finished" podID="f92090a2-228b-41f7-b97b-11df5415877a" containerID="786aa4b04ab3c851cfcfe099c1b123e134328bd98f3a0fb361907803d7ff58ab" exitCode=0 Oct 05 08:16:12 crc kubenswrapper[4846]: I1005 08:16:12.157306 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-n84rx" event={"ID":"f92090a2-228b-41f7-b97b-11df5415877a","Type":"ContainerDied","Data":"786aa4b04ab3c851cfcfe099c1b123e134328bd98f3a0fb361907803d7ff58ab"} Oct 05 08:16:13 crc kubenswrapper[4846]: I1005 08:16:13.509378 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-n84rx" Oct 05 08:16:13 crc kubenswrapper[4846]: I1005 08:16:13.683886 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/f92090a2-228b-41f7-b97b-11df5415877a-db-sync-config-data\") pod \"f92090a2-228b-41f7-b97b-11df5415877a\" (UID: \"f92090a2-228b-41f7-b97b-11df5415877a\") " Oct 05 08:16:13 crc kubenswrapper[4846]: I1005 08:16:13.684450 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nn79q\" (UniqueName: \"kubernetes.io/projected/f92090a2-228b-41f7-b97b-11df5415877a-kube-api-access-nn79q\") pod \"f92090a2-228b-41f7-b97b-11df5415877a\" (UID: \"f92090a2-228b-41f7-b97b-11df5415877a\") " Oct 05 08:16:13 crc kubenswrapper[4846]: I1005 08:16:13.684538 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f92090a2-228b-41f7-b97b-11df5415877a-combined-ca-bundle\") pod \"f92090a2-228b-41f7-b97b-11df5415877a\" (UID: \"f92090a2-228b-41f7-b97b-11df5415877a\") " Oct 05 08:16:13 crc kubenswrapper[4846]: I1005 08:16:13.689385 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f92090a2-228b-41f7-b97b-11df5415877a-kube-api-access-nn79q" (OuterVolumeSpecName: "kube-api-access-nn79q") pod "f92090a2-228b-41f7-b97b-11df5415877a" (UID: "f92090a2-228b-41f7-b97b-11df5415877a"). InnerVolumeSpecName "kube-api-access-nn79q". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:16:13 crc kubenswrapper[4846]: I1005 08:16:13.689575 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f92090a2-228b-41f7-b97b-11df5415877a-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "f92090a2-228b-41f7-b97b-11df5415877a" (UID: "f92090a2-228b-41f7-b97b-11df5415877a"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:16:13 crc kubenswrapper[4846]: I1005 08:16:13.710836 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f92090a2-228b-41f7-b97b-11df5415877a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f92090a2-228b-41f7-b97b-11df5415877a" (UID: "f92090a2-228b-41f7-b97b-11df5415877a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:16:13 crc kubenswrapper[4846]: I1005 08:16:13.787022 4846 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/f92090a2-228b-41f7-b97b-11df5415877a-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 08:16:13 crc kubenswrapper[4846]: I1005 08:16:13.787057 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nn79q\" (UniqueName: \"kubernetes.io/projected/f92090a2-228b-41f7-b97b-11df5415877a-kube-api-access-nn79q\") on node \"crc\" DevicePath \"\"" Oct 05 08:16:13 crc kubenswrapper[4846]: I1005 08:16:13.787068 4846 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f92090a2-228b-41f7-b97b-11df5415877a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 08:16:14 crc kubenswrapper[4846]: I1005 08:16:14.189945 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-n84rx" event={"ID":"f92090a2-228b-41f7-b97b-11df5415877a","Type":"ContainerDied","Data":"ed623649739cbac37d0d720644a1201ae30943cf7c80ec33ee0089d7ff20bff7"} Oct 05 08:16:14 crc kubenswrapper[4846]: I1005 08:16:14.189985 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ed623649739cbac37d0d720644a1201ae30943cf7c80ec33ee0089d7ff20bff7" Oct 05 08:16:14 crc kubenswrapper[4846]: I1005 08:16:14.189986 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-n84rx" Oct 05 08:16:14 crc kubenswrapper[4846]: I1005 08:16:14.458112 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-worker-67855dd78c-jwwd4"] Oct 05 08:16:14 crc kubenswrapper[4846]: E1005 08:16:14.458918 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f92090a2-228b-41f7-b97b-11df5415877a" containerName="barbican-db-sync" Oct 05 08:16:14 crc kubenswrapper[4846]: I1005 08:16:14.458935 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="f92090a2-228b-41f7-b97b-11df5415877a" containerName="barbican-db-sync" Oct 05 08:16:14 crc kubenswrapper[4846]: I1005 08:16:14.459162 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="f92090a2-228b-41f7-b97b-11df5415877a" containerName="barbican-db-sync" Oct 05 08:16:14 crc kubenswrapper[4846]: I1005 08:16:14.460450 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-67855dd78c-jwwd4" Oct 05 08:16:14 crc kubenswrapper[4846]: I1005 08:16:14.464437 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-worker-config-data" Oct 05 08:16:14 crc kubenswrapper[4846]: I1005 08:16:14.464537 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Oct 05 08:16:14 crc kubenswrapper[4846]: I1005 08:16:14.464669 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-ql2bq" Oct 05 08:16:14 crc kubenswrapper[4846]: I1005 08:16:14.483172 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-keystone-listener-595f76b44d-4pljq"] Oct 05 08:16:14 crc kubenswrapper[4846]: I1005 08:16:14.484923 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-595f76b44d-4pljq" Oct 05 08:16:14 crc kubenswrapper[4846]: I1005 08:16:14.490433 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-67855dd78c-jwwd4"] Oct 05 08:16:14 crc kubenswrapper[4846]: I1005 08:16:14.490499 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-keystone-listener-config-data" Oct 05 08:16:14 crc kubenswrapper[4846]: I1005 08:16:14.527750 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-595f76b44d-4pljq"] Oct 05 08:16:14 crc kubenswrapper[4846]: I1005 08:16:14.542420 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-c57df7b7c-w5pmp"] Oct 05 08:16:14 crc kubenswrapper[4846]: I1005 08:16:14.543628 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-c57df7b7c-w5pmp" Oct 05 08:16:14 crc kubenswrapper[4846]: I1005 08:16:14.558995 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-c57df7b7c-w5pmp"] Oct 05 08:16:14 crc kubenswrapper[4846]: I1005 08:16:14.599660 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2zqlz\" (UniqueName: \"kubernetes.io/projected/f93477d4-a2ba-4b9c-bfa9-2db3393596e5-kube-api-access-2zqlz\") pod \"barbican-keystone-listener-595f76b44d-4pljq\" (UID: \"f93477d4-a2ba-4b9c-bfa9-2db3393596e5\") " pod="openstack/barbican-keystone-listener-595f76b44d-4pljq" Oct 05 08:16:14 crc kubenswrapper[4846]: I1005 08:16:14.599720 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f93477d4-a2ba-4b9c-bfa9-2db3393596e5-config-data-custom\") pod \"barbican-keystone-listener-595f76b44d-4pljq\" (UID: \"f93477d4-a2ba-4b9c-bfa9-2db3393596e5\") " pod="openstack/barbican-keystone-listener-595f76b44d-4pljq" Oct 05 08:16:14 crc kubenswrapper[4846]: I1005 08:16:14.599764 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/43fb6020-9a27-4839-9a56-a23a03d21b84-combined-ca-bundle\") pod \"barbican-worker-67855dd78c-jwwd4\" (UID: \"43fb6020-9a27-4839-9a56-a23a03d21b84\") " pod="openstack/barbican-worker-67855dd78c-jwwd4" Oct 05 08:16:14 crc kubenswrapper[4846]: I1005 08:16:14.599796 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/43fb6020-9a27-4839-9a56-a23a03d21b84-config-data\") pod \"barbican-worker-67855dd78c-jwwd4\" (UID: \"43fb6020-9a27-4839-9a56-a23a03d21b84\") " pod="openstack/barbican-worker-67855dd78c-jwwd4" Oct 05 08:16:14 crc kubenswrapper[4846]: I1005 08:16:14.599832 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/43fb6020-9a27-4839-9a56-a23a03d21b84-logs\") pod \"barbican-worker-67855dd78c-jwwd4\" (UID: \"43fb6020-9a27-4839-9a56-a23a03d21b84\") " pod="openstack/barbican-worker-67855dd78c-jwwd4" Oct 05 08:16:14 crc kubenswrapper[4846]: I1005 08:16:14.599883 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jlw7s\" (UniqueName: \"kubernetes.io/projected/43fb6020-9a27-4839-9a56-a23a03d21b84-kube-api-access-jlw7s\") pod \"barbican-worker-67855dd78c-jwwd4\" (UID: \"43fb6020-9a27-4839-9a56-a23a03d21b84\") " pod="openstack/barbican-worker-67855dd78c-jwwd4" Oct 05 08:16:14 crc kubenswrapper[4846]: I1005 08:16:14.599916 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/43fb6020-9a27-4839-9a56-a23a03d21b84-config-data-custom\") pod \"barbican-worker-67855dd78c-jwwd4\" (UID: \"43fb6020-9a27-4839-9a56-a23a03d21b84\") " pod="openstack/barbican-worker-67855dd78c-jwwd4" Oct 05 08:16:14 crc kubenswrapper[4846]: I1005 08:16:14.600024 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f93477d4-a2ba-4b9c-bfa9-2db3393596e5-logs\") pod \"barbican-keystone-listener-595f76b44d-4pljq\" (UID: \"f93477d4-a2ba-4b9c-bfa9-2db3393596e5\") " pod="openstack/barbican-keystone-listener-595f76b44d-4pljq" Oct 05 08:16:14 crc kubenswrapper[4846]: I1005 08:16:14.600079 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f93477d4-a2ba-4b9c-bfa9-2db3393596e5-config-data\") pod \"barbican-keystone-listener-595f76b44d-4pljq\" (UID: \"f93477d4-a2ba-4b9c-bfa9-2db3393596e5\") " pod="openstack/barbican-keystone-listener-595f76b44d-4pljq" Oct 05 08:16:14 crc kubenswrapper[4846]: I1005 08:16:14.600153 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f93477d4-a2ba-4b9c-bfa9-2db3393596e5-combined-ca-bundle\") pod \"barbican-keystone-listener-595f76b44d-4pljq\" (UID: \"f93477d4-a2ba-4b9c-bfa9-2db3393596e5\") " pod="openstack/barbican-keystone-listener-595f76b44d-4pljq" Oct 05 08:16:14 crc kubenswrapper[4846]: I1005 08:16:14.617487 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-7546d54c9d-x8sn2"] Oct 05 08:16:14 crc kubenswrapper[4846]: I1005 08:16:14.621000 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-7546d54c9d-x8sn2" Oct 05 08:16:14 crc kubenswrapper[4846]: I1005 08:16:14.623234 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-api-config-data" Oct 05 08:16:14 crc kubenswrapper[4846]: I1005 08:16:14.625103 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-7546d54c9d-x8sn2"] Oct 05 08:16:14 crc kubenswrapper[4846]: I1005 08:16:14.701950 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9d8932cd-6871-4b34-90cd-3945e6ce6956-ovsdbserver-nb\") pod \"dnsmasq-dns-c57df7b7c-w5pmp\" (UID: \"9d8932cd-6871-4b34-90cd-3945e6ce6956\") " pod="openstack/dnsmasq-dns-c57df7b7c-w5pmp" Oct 05 08:16:14 crc kubenswrapper[4846]: I1005 08:16:14.702000 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/43fb6020-9a27-4839-9a56-a23a03d21b84-combined-ca-bundle\") pod \"barbican-worker-67855dd78c-jwwd4\" (UID: \"43fb6020-9a27-4839-9a56-a23a03d21b84\") " pod="openstack/barbican-worker-67855dd78c-jwwd4" Oct 05 08:16:14 crc kubenswrapper[4846]: I1005 08:16:14.702024 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/43fb6020-9a27-4839-9a56-a23a03d21b84-config-data\") pod \"barbican-worker-67855dd78c-jwwd4\" (UID: \"43fb6020-9a27-4839-9a56-a23a03d21b84\") " pod="openstack/barbican-worker-67855dd78c-jwwd4" Oct 05 08:16:14 crc kubenswrapper[4846]: I1005 08:16:14.702050 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2mb9d\" (UniqueName: \"kubernetes.io/projected/9d8932cd-6871-4b34-90cd-3945e6ce6956-kube-api-access-2mb9d\") pod \"dnsmasq-dns-c57df7b7c-w5pmp\" (UID: \"9d8932cd-6871-4b34-90cd-3945e6ce6956\") " pod="openstack/dnsmasq-dns-c57df7b7c-w5pmp" Oct 05 08:16:14 crc kubenswrapper[4846]: I1005 08:16:14.702118 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f5e8843a-2b8b-4e5e-8b32-3e1ff385a83e-logs\") pod \"barbican-api-7546d54c9d-x8sn2\" (UID: \"f5e8843a-2b8b-4e5e-8b32-3e1ff385a83e\") " pod="openstack/barbican-api-7546d54c9d-x8sn2" Oct 05 08:16:14 crc kubenswrapper[4846]: I1005 08:16:14.702213 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/43fb6020-9a27-4839-9a56-a23a03d21b84-logs\") pod \"barbican-worker-67855dd78c-jwwd4\" (UID: \"43fb6020-9a27-4839-9a56-a23a03d21b84\") " pod="openstack/barbican-worker-67855dd78c-jwwd4" Oct 05 08:16:14 crc kubenswrapper[4846]: I1005 08:16:14.702244 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jlw7s\" (UniqueName: \"kubernetes.io/projected/43fb6020-9a27-4839-9a56-a23a03d21b84-kube-api-access-jlw7s\") pod \"barbican-worker-67855dd78c-jwwd4\" (UID: \"43fb6020-9a27-4839-9a56-a23a03d21b84\") " pod="openstack/barbican-worker-67855dd78c-jwwd4" Oct 05 08:16:14 crc kubenswrapper[4846]: I1005 08:16:14.702267 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/43fb6020-9a27-4839-9a56-a23a03d21b84-config-data-custom\") pod \"barbican-worker-67855dd78c-jwwd4\" (UID: \"43fb6020-9a27-4839-9a56-a23a03d21b84\") " pod="openstack/barbican-worker-67855dd78c-jwwd4" Oct 05 08:16:14 crc kubenswrapper[4846]: I1005 08:16:14.702299 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9d8932cd-6871-4b34-90cd-3945e6ce6956-ovsdbserver-sb\") pod \"dnsmasq-dns-c57df7b7c-w5pmp\" (UID: \"9d8932cd-6871-4b34-90cd-3945e6ce6956\") " pod="openstack/dnsmasq-dns-c57df7b7c-w5pmp" Oct 05 08:16:14 crc kubenswrapper[4846]: I1005 08:16:14.702332 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d8932cd-6871-4b34-90cd-3945e6ce6956-config\") pod \"dnsmasq-dns-c57df7b7c-w5pmp\" (UID: \"9d8932cd-6871-4b34-90cd-3945e6ce6956\") " pod="openstack/dnsmasq-dns-c57df7b7c-w5pmp" Oct 05 08:16:14 crc kubenswrapper[4846]: I1005 08:16:14.702352 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f5e8843a-2b8b-4e5e-8b32-3e1ff385a83e-config-data-custom\") pod \"barbican-api-7546d54c9d-x8sn2\" (UID: \"f5e8843a-2b8b-4e5e-8b32-3e1ff385a83e\") " pod="openstack/barbican-api-7546d54c9d-x8sn2" Oct 05 08:16:14 crc kubenswrapper[4846]: I1005 08:16:14.702379 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f93477d4-a2ba-4b9c-bfa9-2db3393596e5-logs\") pod \"barbican-keystone-listener-595f76b44d-4pljq\" (UID: \"f93477d4-a2ba-4b9c-bfa9-2db3393596e5\") " pod="openstack/barbican-keystone-listener-595f76b44d-4pljq" Oct 05 08:16:14 crc kubenswrapper[4846]: I1005 08:16:14.702402 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f93477d4-a2ba-4b9c-bfa9-2db3393596e5-config-data\") pod \"barbican-keystone-listener-595f76b44d-4pljq\" (UID: \"f93477d4-a2ba-4b9c-bfa9-2db3393596e5\") " pod="openstack/barbican-keystone-listener-595f76b44d-4pljq" Oct 05 08:16:14 crc kubenswrapper[4846]: I1005 08:16:14.702421 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f5e8843a-2b8b-4e5e-8b32-3e1ff385a83e-combined-ca-bundle\") pod \"barbican-api-7546d54c9d-x8sn2\" (UID: \"f5e8843a-2b8b-4e5e-8b32-3e1ff385a83e\") " pod="openstack/barbican-api-7546d54c9d-x8sn2" Oct 05 08:16:14 crc kubenswrapper[4846]: I1005 08:16:14.702453 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9d8932cd-6871-4b34-90cd-3945e6ce6956-dns-svc\") pod \"dnsmasq-dns-c57df7b7c-w5pmp\" (UID: \"9d8932cd-6871-4b34-90cd-3945e6ce6956\") " pod="openstack/dnsmasq-dns-c57df7b7c-w5pmp" Oct 05 08:16:14 crc kubenswrapper[4846]: I1005 08:16:14.702474 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f93477d4-a2ba-4b9c-bfa9-2db3393596e5-combined-ca-bundle\") pod \"barbican-keystone-listener-595f76b44d-4pljq\" (UID: \"f93477d4-a2ba-4b9c-bfa9-2db3393596e5\") " pod="openstack/barbican-keystone-listener-595f76b44d-4pljq" Oct 05 08:16:14 crc kubenswrapper[4846]: I1005 08:16:14.702499 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wkxg7\" (UniqueName: \"kubernetes.io/projected/f5e8843a-2b8b-4e5e-8b32-3e1ff385a83e-kube-api-access-wkxg7\") pod \"barbican-api-7546d54c9d-x8sn2\" (UID: \"f5e8843a-2b8b-4e5e-8b32-3e1ff385a83e\") " pod="openstack/barbican-api-7546d54c9d-x8sn2" Oct 05 08:16:14 crc kubenswrapper[4846]: I1005 08:16:14.702525 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2zqlz\" (UniqueName: \"kubernetes.io/projected/f93477d4-a2ba-4b9c-bfa9-2db3393596e5-kube-api-access-2zqlz\") pod \"barbican-keystone-listener-595f76b44d-4pljq\" (UID: \"f93477d4-a2ba-4b9c-bfa9-2db3393596e5\") " pod="openstack/barbican-keystone-listener-595f76b44d-4pljq" Oct 05 08:16:14 crc kubenswrapper[4846]: I1005 08:16:14.702541 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f5e8843a-2b8b-4e5e-8b32-3e1ff385a83e-config-data\") pod \"barbican-api-7546d54c9d-x8sn2\" (UID: \"f5e8843a-2b8b-4e5e-8b32-3e1ff385a83e\") " pod="openstack/barbican-api-7546d54c9d-x8sn2" Oct 05 08:16:14 crc kubenswrapper[4846]: I1005 08:16:14.702564 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f93477d4-a2ba-4b9c-bfa9-2db3393596e5-config-data-custom\") pod \"barbican-keystone-listener-595f76b44d-4pljq\" (UID: \"f93477d4-a2ba-4b9c-bfa9-2db3393596e5\") " pod="openstack/barbican-keystone-listener-595f76b44d-4pljq" Oct 05 08:16:14 crc kubenswrapper[4846]: I1005 08:16:14.703162 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f93477d4-a2ba-4b9c-bfa9-2db3393596e5-logs\") pod \"barbican-keystone-listener-595f76b44d-4pljq\" (UID: \"f93477d4-a2ba-4b9c-bfa9-2db3393596e5\") " pod="openstack/barbican-keystone-listener-595f76b44d-4pljq" Oct 05 08:16:14 crc kubenswrapper[4846]: I1005 08:16:14.704949 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/43fb6020-9a27-4839-9a56-a23a03d21b84-logs\") pod \"barbican-worker-67855dd78c-jwwd4\" (UID: \"43fb6020-9a27-4839-9a56-a23a03d21b84\") " pod="openstack/barbican-worker-67855dd78c-jwwd4" Oct 05 08:16:14 crc kubenswrapper[4846]: I1005 08:16:14.706238 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/43fb6020-9a27-4839-9a56-a23a03d21b84-combined-ca-bundle\") pod \"barbican-worker-67855dd78c-jwwd4\" (UID: \"43fb6020-9a27-4839-9a56-a23a03d21b84\") " pod="openstack/barbican-worker-67855dd78c-jwwd4" Oct 05 08:16:14 crc kubenswrapper[4846]: I1005 08:16:14.707453 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f93477d4-a2ba-4b9c-bfa9-2db3393596e5-config-data\") pod \"barbican-keystone-listener-595f76b44d-4pljq\" (UID: \"f93477d4-a2ba-4b9c-bfa9-2db3393596e5\") " pod="openstack/barbican-keystone-listener-595f76b44d-4pljq" Oct 05 08:16:14 crc kubenswrapper[4846]: I1005 08:16:14.708412 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/43fb6020-9a27-4839-9a56-a23a03d21b84-config-data-custom\") pod \"barbican-worker-67855dd78c-jwwd4\" (UID: \"43fb6020-9a27-4839-9a56-a23a03d21b84\") " pod="openstack/barbican-worker-67855dd78c-jwwd4" Oct 05 08:16:14 crc kubenswrapper[4846]: I1005 08:16:14.712845 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f93477d4-a2ba-4b9c-bfa9-2db3393596e5-combined-ca-bundle\") pod \"barbican-keystone-listener-595f76b44d-4pljq\" (UID: \"f93477d4-a2ba-4b9c-bfa9-2db3393596e5\") " pod="openstack/barbican-keystone-listener-595f76b44d-4pljq" Oct 05 08:16:14 crc kubenswrapper[4846]: I1005 08:16:14.722090 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/43fb6020-9a27-4839-9a56-a23a03d21b84-config-data\") pod \"barbican-worker-67855dd78c-jwwd4\" (UID: \"43fb6020-9a27-4839-9a56-a23a03d21b84\") " pod="openstack/barbican-worker-67855dd78c-jwwd4" Oct 05 08:16:14 crc kubenswrapper[4846]: I1005 08:16:14.725311 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jlw7s\" (UniqueName: \"kubernetes.io/projected/43fb6020-9a27-4839-9a56-a23a03d21b84-kube-api-access-jlw7s\") pod \"barbican-worker-67855dd78c-jwwd4\" (UID: \"43fb6020-9a27-4839-9a56-a23a03d21b84\") " pod="openstack/barbican-worker-67855dd78c-jwwd4" Oct 05 08:16:14 crc kubenswrapper[4846]: I1005 08:16:14.732468 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2zqlz\" (UniqueName: \"kubernetes.io/projected/f93477d4-a2ba-4b9c-bfa9-2db3393596e5-kube-api-access-2zqlz\") pod \"barbican-keystone-listener-595f76b44d-4pljq\" (UID: \"f93477d4-a2ba-4b9c-bfa9-2db3393596e5\") " pod="openstack/barbican-keystone-listener-595f76b44d-4pljq" Oct 05 08:16:14 crc kubenswrapper[4846]: I1005 08:16:14.732512 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f93477d4-a2ba-4b9c-bfa9-2db3393596e5-config-data-custom\") pod \"barbican-keystone-listener-595f76b44d-4pljq\" (UID: \"f93477d4-a2ba-4b9c-bfa9-2db3393596e5\") " pod="openstack/barbican-keystone-listener-595f76b44d-4pljq" Oct 05 08:16:14 crc kubenswrapper[4846]: I1005 08:16:14.783819 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-67855dd78c-jwwd4" Oct 05 08:16:14 crc kubenswrapper[4846]: I1005 08:16:14.803811 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9d8932cd-6871-4b34-90cd-3945e6ce6956-ovsdbserver-sb\") pod \"dnsmasq-dns-c57df7b7c-w5pmp\" (UID: \"9d8932cd-6871-4b34-90cd-3945e6ce6956\") " pod="openstack/dnsmasq-dns-c57df7b7c-w5pmp" Oct 05 08:16:14 crc kubenswrapper[4846]: I1005 08:16:14.803857 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d8932cd-6871-4b34-90cd-3945e6ce6956-config\") pod \"dnsmasq-dns-c57df7b7c-w5pmp\" (UID: \"9d8932cd-6871-4b34-90cd-3945e6ce6956\") " pod="openstack/dnsmasq-dns-c57df7b7c-w5pmp" Oct 05 08:16:14 crc kubenswrapper[4846]: I1005 08:16:14.803880 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f5e8843a-2b8b-4e5e-8b32-3e1ff385a83e-config-data-custom\") pod \"barbican-api-7546d54c9d-x8sn2\" (UID: \"f5e8843a-2b8b-4e5e-8b32-3e1ff385a83e\") " pod="openstack/barbican-api-7546d54c9d-x8sn2" Oct 05 08:16:14 crc kubenswrapper[4846]: I1005 08:16:14.803919 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f5e8843a-2b8b-4e5e-8b32-3e1ff385a83e-combined-ca-bundle\") pod \"barbican-api-7546d54c9d-x8sn2\" (UID: \"f5e8843a-2b8b-4e5e-8b32-3e1ff385a83e\") " pod="openstack/barbican-api-7546d54c9d-x8sn2" Oct 05 08:16:14 crc kubenswrapper[4846]: I1005 08:16:14.803947 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9d8932cd-6871-4b34-90cd-3945e6ce6956-dns-svc\") pod \"dnsmasq-dns-c57df7b7c-w5pmp\" (UID: \"9d8932cd-6871-4b34-90cd-3945e6ce6956\") " pod="openstack/dnsmasq-dns-c57df7b7c-w5pmp" Oct 05 08:16:14 crc kubenswrapper[4846]: I1005 08:16:14.803976 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wkxg7\" (UniqueName: \"kubernetes.io/projected/f5e8843a-2b8b-4e5e-8b32-3e1ff385a83e-kube-api-access-wkxg7\") pod \"barbican-api-7546d54c9d-x8sn2\" (UID: \"f5e8843a-2b8b-4e5e-8b32-3e1ff385a83e\") " pod="openstack/barbican-api-7546d54c9d-x8sn2" Oct 05 08:16:14 crc kubenswrapper[4846]: I1005 08:16:14.804000 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f5e8843a-2b8b-4e5e-8b32-3e1ff385a83e-config-data\") pod \"barbican-api-7546d54c9d-x8sn2\" (UID: \"f5e8843a-2b8b-4e5e-8b32-3e1ff385a83e\") " pod="openstack/barbican-api-7546d54c9d-x8sn2" Oct 05 08:16:14 crc kubenswrapper[4846]: I1005 08:16:14.804028 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9d8932cd-6871-4b34-90cd-3945e6ce6956-ovsdbserver-nb\") pod \"dnsmasq-dns-c57df7b7c-w5pmp\" (UID: \"9d8932cd-6871-4b34-90cd-3945e6ce6956\") " pod="openstack/dnsmasq-dns-c57df7b7c-w5pmp" Oct 05 08:16:14 crc kubenswrapper[4846]: I1005 08:16:14.804055 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2mb9d\" (UniqueName: \"kubernetes.io/projected/9d8932cd-6871-4b34-90cd-3945e6ce6956-kube-api-access-2mb9d\") pod \"dnsmasq-dns-c57df7b7c-w5pmp\" (UID: \"9d8932cd-6871-4b34-90cd-3945e6ce6956\") " pod="openstack/dnsmasq-dns-c57df7b7c-w5pmp" Oct 05 08:16:14 crc kubenswrapper[4846]: I1005 08:16:14.804069 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f5e8843a-2b8b-4e5e-8b32-3e1ff385a83e-logs\") pod \"barbican-api-7546d54c9d-x8sn2\" (UID: \"f5e8843a-2b8b-4e5e-8b32-3e1ff385a83e\") " pod="openstack/barbican-api-7546d54c9d-x8sn2" Oct 05 08:16:14 crc kubenswrapper[4846]: I1005 08:16:14.804509 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f5e8843a-2b8b-4e5e-8b32-3e1ff385a83e-logs\") pod \"barbican-api-7546d54c9d-x8sn2\" (UID: \"f5e8843a-2b8b-4e5e-8b32-3e1ff385a83e\") " pod="openstack/barbican-api-7546d54c9d-x8sn2" Oct 05 08:16:14 crc kubenswrapper[4846]: I1005 08:16:14.805101 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9d8932cd-6871-4b34-90cd-3945e6ce6956-dns-svc\") pod \"dnsmasq-dns-c57df7b7c-w5pmp\" (UID: \"9d8932cd-6871-4b34-90cd-3945e6ce6956\") " pod="openstack/dnsmasq-dns-c57df7b7c-w5pmp" Oct 05 08:16:14 crc kubenswrapper[4846]: I1005 08:16:14.805638 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9d8932cd-6871-4b34-90cd-3945e6ce6956-ovsdbserver-sb\") pod \"dnsmasq-dns-c57df7b7c-w5pmp\" (UID: \"9d8932cd-6871-4b34-90cd-3945e6ce6956\") " pod="openstack/dnsmasq-dns-c57df7b7c-w5pmp" Oct 05 08:16:14 crc kubenswrapper[4846]: I1005 08:16:14.806198 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d8932cd-6871-4b34-90cd-3945e6ce6956-config\") pod \"dnsmasq-dns-c57df7b7c-w5pmp\" (UID: \"9d8932cd-6871-4b34-90cd-3945e6ce6956\") " pod="openstack/dnsmasq-dns-c57df7b7c-w5pmp" Oct 05 08:16:14 crc kubenswrapper[4846]: I1005 08:16:14.811158 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9d8932cd-6871-4b34-90cd-3945e6ce6956-ovsdbserver-nb\") pod \"dnsmasq-dns-c57df7b7c-w5pmp\" (UID: \"9d8932cd-6871-4b34-90cd-3945e6ce6956\") " pod="openstack/dnsmasq-dns-c57df7b7c-w5pmp" Oct 05 08:16:14 crc kubenswrapper[4846]: I1005 08:16:14.811778 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-595f76b44d-4pljq" Oct 05 08:16:14 crc kubenswrapper[4846]: I1005 08:16:14.811971 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f5e8843a-2b8b-4e5e-8b32-3e1ff385a83e-config-data\") pod \"barbican-api-7546d54c9d-x8sn2\" (UID: \"f5e8843a-2b8b-4e5e-8b32-3e1ff385a83e\") " pod="openstack/barbican-api-7546d54c9d-x8sn2" Oct 05 08:16:14 crc kubenswrapper[4846]: I1005 08:16:14.818152 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f5e8843a-2b8b-4e5e-8b32-3e1ff385a83e-combined-ca-bundle\") pod \"barbican-api-7546d54c9d-x8sn2\" (UID: \"f5e8843a-2b8b-4e5e-8b32-3e1ff385a83e\") " pod="openstack/barbican-api-7546d54c9d-x8sn2" Oct 05 08:16:14 crc kubenswrapper[4846]: I1005 08:16:14.818773 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f5e8843a-2b8b-4e5e-8b32-3e1ff385a83e-config-data-custom\") pod \"barbican-api-7546d54c9d-x8sn2\" (UID: \"f5e8843a-2b8b-4e5e-8b32-3e1ff385a83e\") " pod="openstack/barbican-api-7546d54c9d-x8sn2" Oct 05 08:16:14 crc kubenswrapper[4846]: I1005 08:16:14.820492 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wkxg7\" (UniqueName: \"kubernetes.io/projected/f5e8843a-2b8b-4e5e-8b32-3e1ff385a83e-kube-api-access-wkxg7\") pod \"barbican-api-7546d54c9d-x8sn2\" (UID: \"f5e8843a-2b8b-4e5e-8b32-3e1ff385a83e\") " pod="openstack/barbican-api-7546d54c9d-x8sn2" Oct 05 08:16:14 crc kubenswrapper[4846]: I1005 08:16:14.826644 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2mb9d\" (UniqueName: \"kubernetes.io/projected/9d8932cd-6871-4b34-90cd-3945e6ce6956-kube-api-access-2mb9d\") pod \"dnsmasq-dns-c57df7b7c-w5pmp\" (UID: \"9d8932cd-6871-4b34-90cd-3945e6ce6956\") " pod="openstack/dnsmasq-dns-c57df7b7c-w5pmp" Oct 05 08:16:14 crc kubenswrapper[4846]: I1005 08:16:14.862347 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-c57df7b7c-w5pmp" Oct 05 08:16:14 crc kubenswrapper[4846]: I1005 08:16:14.937902 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-7546d54c9d-x8sn2" Oct 05 08:16:15 crc kubenswrapper[4846]: I1005 08:16:15.432870 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-595f76b44d-4pljq"] Oct 05 08:16:15 crc kubenswrapper[4846]: W1005 08:16:15.436655 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf93477d4_a2ba_4b9c_bfa9_2db3393596e5.slice/crio-4df856686d082228feb7b17df14c448ceb0d8c9d1ac4d24826afa3c5aa55c924 WatchSource:0}: Error finding container 4df856686d082228feb7b17df14c448ceb0d8c9d1ac4d24826afa3c5aa55c924: Status 404 returned error can't find the container with id 4df856686d082228feb7b17df14c448ceb0d8c9d1ac4d24826afa3c5aa55c924 Oct 05 08:16:15 crc kubenswrapper[4846]: I1005 08:16:15.533091 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-7546d54c9d-x8sn2"] Oct 05 08:16:15 crc kubenswrapper[4846]: W1005 08:16:15.535271 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf5e8843a_2b8b_4e5e_8b32_3e1ff385a83e.slice/crio-dec9d3c43452efe5a7b6eaca3d23d537312e968f30dafff2be851b19caa6b43c WatchSource:0}: Error finding container dec9d3c43452efe5a7b6eaca3d23d537312e968f30dafff2be851b19caa6b43c: Status 404 returned error can't find the container with id dec9d3c43452efe5a7b6eaca3d23d537312e968f30dafff2be851b19caa6b43c Oct 05 08:16:15 crc kubenswrapper[4846]: I1005 08:16:15.555262 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-67855dd78c-jwwd4"] Oct 05 08:16:15 crc kubenswrapper[4846]: I1005 08:16:15.603671 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-c57df7b7c-w5pmp"] Oct 05 08:16:16 crc kubenswrapper[4846]: I1005 08:16:16.225019 4846 generic.go:334] "Generic (PLEG): container finished" podID="9d8932cd-6871-4b34-90cd-3945e6ce6956" containerID="b5a670ff8fd08b29599244922b266773ed660c53c819ce0149d0e95ceabb3973" exitCode=0 Oct 05 08:16:16 crc kubenswrapper[4846]: I1005 08:16:16.225131 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-c57df7b7c-w5pmp" event={"ID":"9d8932cd-6871-4b34-90cd-3945e6ce6956","Type":"ContainerDied","Data":"b5a670ff8fd08b29599244922b266773ed660c53c819ce0149d0e95ceabb3973"} Oct 05 08:16:16 crc kubenswrapper[4846]: I1005 08:16:16.225398 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-c57df7b7c-w5pmp" event={"ID":"9d8932cd-6871-4b34-90cd-3945e6ce6956","Type":"ContainerStarted","Data":"7a710bfec083ef531a8af0716eb93fe0c35022644927d90eac75a2b405650ecf"} Oct 05 08:16:16 crc kubenswrapper[4846]: I1005 08:16:16.226636 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-67855dd78c-jwwd4" event={"ID":"43fb6020-9a27-4839-9a56-a23a03d21b84","Type":"ContainerStarted","Data":"a630b106508d41d3a6c749f2ee8b8b02d1bbf4598cbc0dd63b804636a7ade163"} Oct 05 08:16:16 crc kubenswrapper[4846]: I1005 08:16:16.230808 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-595f76b44d-4pljq" event={"ID":"f93477d4-a2ba-4b9c-bfa9-2db3393596e5","Type":"ContainerStarted","Data":"4df856686d082228feb7b17df14c448ceb0d8c9d1ac4d24826afa3c5aa55c924"} Oct 05 08:16:16 crc kubenswrapper[4846]: I1005 08:16:16.235778 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7546d54c9d-x8sn2" event={"ID":"f5e8843a-2b8b-4e5e-8b32-3e1ff385a83e","Type":"ContainerStarted","Data":"8bf92f2421b601ab85762e0732103c649287c9804192bb0d0c05ba6abc8c0e60"} Oct 05 08:16:16 crc kubenswrapper[4846]: I1005 08:16:16.235821 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7546d54c9d-x8sn2" event={"ID":"f5e8843a-2b8b-4e5e-8b32-3e1ff385a83e","Type":"ContainerStarted","Data":"48ee0d9d054943125c654f38836d686d2255e61573b06f5d3288121665374229"} Oct 05 08:16:16 crc kubenswrapper[4846]: I1005 08:16:16.235832 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7546d54c9d-x8sn2" event={"ID":"f5e8843a-2b8b-4e5e-8b32-3e1ff385a83e","Type":"ContainerStarted","Data":"dec9d3c43452efe5a7b6eaca3d23d537312e968f30dafff2be851b19caa6b43c"} Oct 05 08:16:16 crc kubenswrapper[4846]: I1005 08:16:16.236379 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-7546d54c9d-x8sn2" Oct 05 08:16:16 crc kubenswrapper[4846]: I1005 08:16:16.236406 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-7546d54c9d-x8sn2" Oct 05 08:16:16 crc kubenswrapper[4846]: I1005 08:16:16.288047 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-7546d54c9d-x8sn2" podStartSLOduration=2.288022045 podStartE2EDuration="2.288022045s" podCreationTimestamp="2025-10-05 08:16:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 08:16:16.280296258 +0000 UTC m=+5298.521149033" watchObservedRunningTime="2025-10-05 08:16:16.288022045 +0000 UTC m=+5298.528874820" Oct 05 08:16:17 crc kubenswrapper[4846]: I1005 08:16:17.044591 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-5d595db478-zkhxm"] Oct 05 08:16:17 crc kubenswrapper[4846]: I1005 08:16:17.046165 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-5d595db478-zkhxm" Oct 05 08:16:17 crc kubenswrapper[4846]: I1005 08:16:17.047752 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-public-svc" Oct 05 08:16:17 crc kubenswrapper[4846]: I1005 08:16:17.048131 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-internal-svc" Oct 05 08:16:17 crc kubenswrapper[4846]: I1005 08:16:17.062988 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-5d595db478-zkhxm"] Oct 05 08:16:17 crc kubenswrapper[4846]: I1005 08:16:17.148936 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8dd27928-1e7a-42b1-ae7f-7957e42cf4c1-combined-ca-bundle\") pod \"barbican-api-5d595db478-zkhxm\" (UID: \"8dd27928-1e7a-42b1-ae7f-7957e42cf4c1\") " pod="openstack/barbican-api-5d595db478-zkhxm" Oct 05 08:16:17 crc kubenswrapper[4846]: I1005 08:16:17.149071 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8dd27928-1e7a-42b1-ae7f-7957e42cf4c1-config-data-custom\") pod \"barbican-api-5d595db478-zkhxm\" (UID: \"8dd27928-1e7a-42b1-ae7f-7957e42cf4c1\") " pod="openstack/barbican-api-5d595db478-zkhxm" Oct 05 08:16:17 crc kubenswrapper[4846]: I1005 08:16:17.149140 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8dd27928-1e7a-42b1-ae7f-7957e42cf4c1-internal-tls-certs\") pod \"barbican-api-5d595db478-zkhxm\" (UID: \"8dd27928-1e7a-42b1-ae7f-7957e42cf4c1\") " pod="openstack/barbican-api-5d595db478-zkhxm" Oct 05 08:16:17 crc kubenswrapper[4846]: I1005 08:16:17.149432 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8dd27928-1e7a-42b1-ae7f-7957e42cf4c1-config-data\") pod \"barbican-api-5d595db478-zkhxm\" (UID: \"8dd27928-1e7a-42b1-ae7f-7957e42cf4c1\") " pod="openstack/barbican-api-5d595db478-zkhxm" Oct 05 08:16:17 crc kubenswrapper[4846]: I1005 08:16:17.149495 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8dd27928-1e7a-42b1-ae7f-7957e42cf4c1-public-tls-certs\") pod \"barbican-api-5d595db478-zkhxm\" (UID: \"8dd27928-1e7a-42b1-ae7f-7957e42cf4c1\") " pod="openstack/barbican-api-5d595db478-zkhxm" Oct 05 08:16:17 crc kubenswrapper[4846]: I1005 08:16:17.149707 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6xm8x\" (UniqueName: \"kubernetes.io/projected/8dd27928-1e7a-42b1-ae7f-7957e42cf4c1-kube-api-access-6xm8x\") pod \"barbican-api-5d595db478-zkhxm\" (UID: \"8dd27928-1e7a-42b1-ae7f-7957e42cf4c1\") " pod="openstack/barbican-api-5d595db478-zkhxm" Oct 05 08:16:17 crc kubenswrapper[4846]: I1005 08:16:17.149759 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8dd27928-1e7a-42b1-ae7f-7957e42cf4c1-logs\") pod \"barbican-api-5d595db478-zkhxm\" (UID: \"8dd27928-1e7a-42b1-ae7f-7957e42cf4c1\") " pod="openstack/barbican-api-5d595db478-zkhxm" Oct 05 08:16:17 crc kubenswrapper[4846]: I1005 08:16:17.251292 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8dd27928-1e7a-42b1-ae7f-7957e42cf4c1-logs\") pod \"barbican-api-5d595db478-zkhxm\" (UID: \"8dd27928-1e7a-42b1-ae7f-7957e42cf4c1\") " pod="openstack/barbican-api-5d595db478-zkhxm" Oct 05 08:16:17 crc kubenswrapper[4846]: I1005 08:16:17.251363 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8dd27928-1e7a-42b1-ae7f-7957e42cf4c1-combined-ca-bundle\") pod \"barbican-api-5d595db478-zkhxm\" (UID: \"8dd27928-1e7a-42b1-ae7f-7957e42cf4c1\") " pod="openstack/barbican-api-5d595db478-zkhxm" Oct 05 08:16:17 crc kubenswrapper[4846]: I1005 08:16:17.251388 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8dd27928-1e7a-42b1-ae7f-7957e42cf4c1-config-data-custom\") pod \"barbican-api-5d595db478-zkhxm\" (UID: \"8dd27928-1e7a-42b1-ae7f-7957e42cf4c1\") " pod="openstack/barbican-api-5d595db478-zkhxm" Oct 05 08:16:17 crc kubenswrapper[4846]: I1005 08:16:17.251410 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8dd27928-1e7a-42b1-ae7f-7957e42cf4c1-internal-tls-certs\") pod \"barbican-api-5d595db478-zkhxm\" (UID: \"8dd27928-1e7a-42b1-ae7f-7957e42cf4c1\") " pod="openstack/barbican-api-5d595db478-zkhxm" Oct 05 08:16:17 crc kubenswrapper[4846]: I1005 08:16:17.251461 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8dd27928-1e7a-42b1-ae7f-7957e42cf4c1-config-data\") pod \"barbican-api-5d595db478-zkhxm\" (UID: \"8dd27928-1e7a-42b1-ae7f-7957e42cf4c1\") " pod="openstack/barbican-api-5d595db478-zkhxm" Oct 05 08:16:17 crc kubenswrapper[4846]: I1005 08:16:17.251479 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8dd27928-1e7a-42b1-ae7f-7957e42cf4c1-public-tls-certs\") pod \"barbican-api-5d595db478-zkhxm\" (UID: \"8dd27928-1e7a-42b1-ae7f-7957e42cf4c1\") " pod="openstack/barbican-api-5d595db478-zkhxm" Oct 05 08:16:17 crc kubenswrapper[4846]: I1005 08:16:17.251530 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6xm8x\" (UniqueName: \"kubernetes.io/projected/8dd27928-1e7a-42b1-ae7f-7957e42cf4c1-kube-api-access-6xm8x\") pod \"barbican-api-5d595db478-zkhxm\" (UID: \"8dd27928-1e7a-42b1-ae7f-7957e42cf4c1\") " pod="openstack/barbican-api-5d595db478-zkhxm" Oct 05 08:16:17 crc kubenswrapper[4846]: I1005 08:16:17.252106 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8dd27928-1e7a-42b1-ae7f-7957e42cf4c1-logs\") pod \"barbican-api-5d595db478-zkhxm\" (UID: \"8dd27928-1e7a-42b1-ae7f-7957e42cf4c1\") " pod="openstack/barbican-api-5d595db478-zkhxm" Oct 05 08:16:17 crc kubenswrapper[4846]: I1005 08:16:17.257532 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8dd27928-1e7a-42b1-ae7f-7957e42cf4c1-public-tls-certs\") pod \"barbican-api-5d595db478-zkhxm\" (UID: \"8dd27928-1e7a-42b1-ae7f-7957e42cf4c1\") " pod="openstack/barbican-api-5d595db478-zkhxm" Oct 05 08:16:17 crc kubenswrapper[4846]: I1005 08:16:17.257647 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8dd27928-1e7a-42b1-ae7f-7957e42cf4c1-config-data-custom\") pod \"barbican-api-5d595db478-zkhxm\" (UID: \"8dd27928-1e7a-42b1-ae7f-7957e42cf4c1\") " pod="openstack/barbican-api-5d595db478-zkhxm" Oct 05 08:16:17 crc kubenswrapper[4846]: I1005 08:16:17.257675 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8dd27928-1e7a-42b1-ae7f-7957e42cf4c1-combined-ca-bundle\") pod \"barbican-api-5d595db478-zkhxm\" (UID: \"8dd27928-1e7a-42b1-ae7f-7957e42cf4c1\") " pod="openstack/barbican-api-5d595db478-zkhxm" Oct 05 08:16:17 crc kubenswrapper[4846]: I1005 08:16:17.259003 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8dd27928-1e7a-42b1-ae7f-7957e42cf4c1-config-data\") pod \"barbican-api-5d595db478-zkhxm\" (UID: \"8dd27928-1e7a-42b1-ae7f-7957e42cf4c1\") " pod="openstack/barbican-api-5d595db478-zkhxm" Oct 05 08:16:17 crc kubenswrapper[4846]: I1005 08:16:17.265620 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8dd27928-1e7a-42b1-ae7f-7957e42cf4c1-internal-tls-certs\") pod \"barbican-api-5d595db478-zkhxm\" (UID: \"8dd27928-1e7a-42b1-ae7f-7957e42cf4c1\") " pod="openstack/barbican-api-5d595db478-zkhxm" Oct 05 08:16:17 crc kubenswrapper[4846]: I1005 08:16:17.277027 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6xm8x\" (UniqueName: \"kubernetes.io/projected/8dd27928-1e7a-42b1-ae7f-7957e42cf4c1-kube-api-access-6xm8x\") pod \"barbican-api-5d595db478-zkhxm\" (UID: \"8dd27928-1e7a-42b1-ae7f-7957e42cf4c1\") " pod="openstack/barbican-api-5d595db478-zkhxm" Oct 05 08:16:17 crc kubenswrapper[4846]: I1005 08:16:17.372758 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-5d595db478-zkhxm" Oct 05 08:16:17 crc kubenswrapper[4846]: I1005 08:16:17.800933 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-5d595db478-zkhxm"] Oct 05 08:16:17 crc kubenswrapper[4846]: W1005 08:16:17.807007 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8dd27928_1e7a_42b1_ae7f_7957e42cf4c1.slice/crio-e7b66d3a6254b3fb9f8979f0f84c69f767b0c73f4e9b94ef4fc5ddec0c983f9b WatchSource:0}: Error finding container e7b66d3a6254b3fb9f8979f0f84c69f767b0c73f4e9b94ef4fc5ddec0c983f9b: Status 404 returned error can't find the container with id e7b66d3a6254b3fb9f8979f0f84c69f767b0c73f4e9b94ef4fc5ddec0c983f9b Oct 05 08:16:18 crc kubenswrapper[4846]: I1005 08:16:18.251404 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-67855dd78c-jwwd4" event={"ID":"43fb6020-9a27-4839-9a56-a23a03d21b84","Type":"ContainerStarted","Data":"2a31ddec6ea3209cbe8d208b82b69f5a2688a28619bec49ca71181dc5d61fd82"} Oct 05 08:16:18 crc kubenswrapper[4846]: I1005 08:16:18.251443 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-67855dd78c-jwwd4" event={"ID":"43fb6020-9a27-4839-9a56-a23a03d21b84","Type":"ContainerStarted","Data":"d48a67ff789bb44a86990bcd1b0a047b41898f45e8959b1ec6a92d3ca440e579"} Oct 05 08:16:18 crc kubenswrapper[4846]: I1005 08:16:18.254770 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5d595db478-zkhxm" event={"ID":"8dd27928-1e7a-42b1-ae7f-7957e42cf4c1","Type":"ContainerStarted","Data":"825a4a692efc7814d13e594f001a912e87297daf052a346d0877f85b853fd6d8"} Oct 05 08:16:18 crc kubenswrapper[4846]: I1005 08:16:18.254824 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5d595db478-zkhxm" event={"ID":"8dd27928-1e7a-42b1-ae7f-7957e42cf4c1","Type":"ContainerStarted","Data":"bae50e15d3d5bbf91d08394121fafce5ab4f7e67f208bc940571ca7b98d6f62b"} Oct 05 08:16:18 crc kubenswrapper[4846]: I1005 08:16:18.254838 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5d595db478-zkhxm" event={"ID":"8dd27928-1e7a-42b1-ae7f-7957e42cf4c1","Type":"ContainerStarted","Data":"e7b66d3a6254b3fb9f8979f0f84c69f767b0c73f4e9b94ef4fc5ddec0c983f9b"} Oct 05 08:16:18 crc kubenswrapper[4846]: I1005 08:16:18.254882 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-5d595db478-zkhxm" Oct 05 08:16:18 crc kubenswrapper[4846]: I1005 08:16:18.254906 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-5d595db478-zkhxm" Oct 05 08:16:18 crc kubenswrapper[4846]: I1005 08:16:18.256692 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-595f76b44d-4pljq" event={"ID":"f93477d4-a2ba-4b9c-bfa9-2db3393596e5","Type":"ContainerStarted","Data":"5c55e0387d7ce6789137e70c6867574e582f34f4f5044582ab60b2796edcb2c2"} Oct 05 08:16:18 crc kubenswrapper[4846]: I1005 08:16:18.256723 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-595f76b44d-4pljq" event={"ID":"f93477d4-a2ba-4b9c-bfa9-2db3393596e5","Type":"ContainerStarted","Data":"695197681ffef8776dcea43229ea61f73b1209b98b0bd5443dade485441197a1"} Oct 05 08:16:18 crc kubenswrapper[4846]: I1005 08:16:18.259057 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-c57df7b7c-w5pmp" event={"ID":"9d8932cd-6871-4b34-90cd-3945e6ce6956","Type":"ContainerStarted","Data":"7832c0fde700ceb61df332f26acd0cb00cb3c3565d3185c78b3864e1dabf2b88"} Oct 05 08:16:18 crc kubenswrapper[4846]: I1005 08:16:18.259098 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-c57df7b7c-w5pmp" Oct 05 08:16:18 crc kubenswrapper[4846]: I1005 08:16:18.289074 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-worker-67855dd78c-jwwd4" podStartSLOduration=2.73990821 podStartE2EDuration="4.289059994s" podCreationTimestamp="2025-10-05 08:16:14 +0000 UTC" firstStartedPulling="2025-10-05 08:16:15.577800682 +0000 UTC m=+5297.818653457" lastFinishedPulling="2025-10-05 08:16:17.126952466 +0000 UTC m=+5299.367805241" observedRunningTime="2025-10-05 08:16:18.286879966 +0000 UTC m=+5300.527732741" watchObservedRunningTime="2025-10-05 08:16:18.289059994 +0000 UTC m=+5300.529912769" Oct 05 08:16:18 crc kubenswrapper[4846]: I1005 08:16:18.311873 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-c57df7b7c-w5pmp" podStartSLOduration=4.311847815 podStartE2EDuration="4.311847815s" podCreationTimestamp="2025-10-05 08:16:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 08:16:18.310151689 +0000 UTC m=+5300.551004464" watchObservedRunningTime="2025-10-05 08:16:18.311847815 +0000 UTC m=+5300.552700590" Oct 05 08:16:18 crc kubenswrapper[4846]: I1005 08:16:18.354046 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-5d595db478-zkhxm" podStartSLOduration=1.354028965 podStartE2EDuration="1.354028965s" podCreationTimestamp="2025-10-05 08:16:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 08:16:18.344342295 +0000 UTC m=+5300.585195070" watchObservedRunningTime="2025-10-05 08:16:18.354028965 +0000 UTC m=+5300.594881740" Oct 05 08:16:18 crc kubenswrapper[4846]: I1005 08:16:18.370406 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-keystone-listener-595f76b44d-4pljq" podStartSLOduration=2.68278951 podStartE2EDuration="4.370388133s" podCreationTimestamp="2025-10-05 08:16:14 +0000 UTC" firstStartedPulling="2025-10-05 08:16:15.438819999 +0000 UTC m=+5297.679672774" lastFinishedPulling="2025-10-05 08:16:17.126418622 +0000 UTC m=+5299.367271397" observedRunningTime="2025-10-05 08:16:18.365948424 +0000 UTC m=+5300.606801199" watchObservedRunningTime="2025-10-05 08:16:18.370388133 +0000 UTC m=+5300.611240908" Oct 05 08:16:23 crc kubenswrapper[4846]: I1005 08:16:23.840818 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-5d595db478-zkhxm" Oct 05 08:16:24 crc kubenswrapper[4846]: I1005 08:16:24.864419 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-c57df7b7c-w5pmp" Oct 05 08:16:24 crc kubenswrapper[4846]: I1005 08:16:24.961195 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6d65c684ff-nnw4f"] Oct 05 08:16:24 crc kubenswrapper[4846]: I1005 08:16:24.961509 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6d65c684ff-nnw4f" podUID="d061e728-24eb-47ea-bb6e-6495722308ab" containerName="dnsmasq-dns" containerID="cri-o://9dcbc8dca17e9726f55658375a4dbc3b9d27778f907b45a9136b9ab37b149e77" gracePeriod=10 Oct 05 08:16:25 crc kubenswrapper[4846]: I1005 08:16:25.256891 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-5d595db478-zkhxm" Oct 05 08:16:25 crc kubenswrapper[4846]: I1005 08:16:25.332601 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-7546d54c9d-x8sn2"] Oct 05 08:16:25 crc kubenswrapper[4846]: I1005 08:16:25.332849 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-7546d54c9d-x8sn2" podUID="f5e8843a-2b8b-4e5e-8b32-3e1ff385a83e" containerName="barbican-api-log" containerID="cri-o://48ee0d9d054943125c654f38836d686d2255e61573b06f5d3288121665374229" gracePeriod=30 Oct 05 08:16:25 crc kubenswrapper[4846]: I1005 08:16:25.333676 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-7546d54c9d-x8sn2" podUID="f5e8843a-2b8b-4e5e-8b32-3e1ff385a83e" containerName="barbican-api" containerID="cri-o://8bf92f2421b601ab85762e0732103c649287c9804192bb0d0c05ba6abc8c0e60" gracePeriod=30 Oct 05 08:16:25 crc kubenswrapper[4846]: I1005 08:16:25.339468 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/barbican-api-7546d54c9d-x8sn2" podUID="f5e8843a-2b8b-4e5e-8b32-3e1ff385a83e" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.1.38:9311/healthcheck\": EOF" Oct 05 08:16:25 crc kubenswrapper[4846]: I1005 08:16:25.349496 4846 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-7546d54c9d-x8sn2" podUID="f5e8843a-2b8b-4e5e-8b32-3e1ff385a83e" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.1.38:9311/healthcheck\": EOF" Oct 05 08:16:25 crc kubenswrapper[4846]: I1005 08:16:25.349594 4846 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-7546d54c9d-x8sn2" podUID="f5e8843a-2b8b-4e5e-8b32-3e1ff385a83e" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.1.38:9311/healthcheck\": EOF" Oct 05 08:16:25 crc kubenswrapper[4846]: I1005 08:16:25.349972 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/barbican-api-7546d54c9d-x8sn2" podUID="f5e8843a-2b8b-4e5e-8b32-3e1ff385a83e" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.1.38:9311/healthcheck\": EOF" Oct 05 08:16:25 crc kubenswrapper[4846]: I1005 08:16:25.384116 4846 generic.go:334] "Generic (PLEG): container finished" podID="d061e728-24eb-47ea-bb6e-6495722308ab" containerID="9dcbc8dca17e9726f55658375a4dbc3b9d27778f907b45a9136b9ab37b149e77" exitCode=0 Oct 05 08:16:25 crc kubenswrapper[4846]: I1005 08:16:25.384182 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d65c684ff-nnw4f" event={"ID":"d061e728-24eb-47ea-bb6e-6495722308ab","Type":"ContainerDied","Data":"9dcbc8dca17e9726f55658375a4dbc3b9d27778f907b45a9136b9ab37b149e77"} Oct 05 08:16:25 crc kubenswrapper[4846]: I1005 08:16:25.542526 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6d65c684ff-nnw4f" Oct 05 08:16:25 crc kubenswrapper[4846]: I1005 08:16:25.622705 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d061e728-24eb-47ea-bb6e-6495722308ab-ovsdbserver-nb\") pod \"d061e728-24eb-47ea-bb6e-6495722308ab\" (UID: \"d061e728-24eb-47ea-bb6e-6495722308ab\") " Oct 05 08:16:25 crc kubenswrapper[4846]: I1005 08:16:25.622933 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d061e728-24eb-47ea-bb6e-6495722308ab-dns-svc\") pod \"d061e728-24eb-47ea-bb6e-6495722308ab\" (UID: \"d061e728-24eb-47ea-bb6e-6495722308ab\") " Oct 05 08:16:25 crc kubenswrapper[4846]: I1005 08:16:25.622970 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d061e728-24eb-47ea-bb6e-6495722308ab-ovsdbserver-sb\") pod \"d061e728-24eb-47ea-bb6e-6495722308ab\" (UID: \"d061e728-24eb-47ea-bb6e-6495722308ab\") " Oct 05 08:16:25 crc kubenswrapper[4846]: I1005 08:16:25.623060 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d061e728-24eb-47ea-bb6e-6495722308ab-config\") pod \"d061e728-24eb-47ea-bb6e-6495722308ab\" (UID: \"d061e728-24eb-47ea-bb6e-6495722308ab\") " Oct 05 08:16:25 crc kubenswrapper[4846]: I1005 08:16:25.623092 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fd7jq\" (UniqueName: \"kubernetes.io/projected/d061e728-24eb-47ea-bb6e-6495722308ab-kube-api-access-fd7jq\") pod \"d061e728-24eb-47ea-bb6e-6495722308ab\" (UID: \"d061e728-24eb-47ea-bb6e-6495722308ab\") " Oct 05 08:16:25 crc kubenswrapper[4846]: I1005 08:16:25.639444 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d061e728-24eb-47ea-bb6e-6495722308ab-kube-api-access-fd7jq" (OuterVolumeSpecName: "kube-api-access-fd7jq") pod "d061e728-24eb-47ea-bb6e-6495722308ab" (UID: "d061e728-24eb-47ea-bb6e-6495722308ab"). InnerVolumeSpecName "kube-api-access-fd7jq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:16:25 crc kubenswrapper[4846]: I1005 08:16:25.663168 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d061e728-24eb-47ea-bb6e-6495722308ab-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "d061e728-24eb-47ea-bb6e-6495722308ab" (UID: "d061e728-24eb-47ea-bb6e-6495722308ab"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 08:16:25 crc kubenswrapper[4846]: I1005 08:16:25.663241 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d061e728-24eb-47ea-bb6e-6495722308ab-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "d061e728-24eb-47ea-bb6e-6495722308ab" (UID: "d061e728-24eb-47ea-bb6e-6495722308ab"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 08:16:25 crc kubenswrapper[4846]: I1005 08:16:25.663776 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d061e728-24eb-47ea-bb6e-6495722308ab-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "d061e728-24eb-47ea-bb6e-6495722308ab" (UID: "d061e728-24eb-47ea-bb6e-6495722308ab"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 08:16:25 crc kubenswrapper[4846]: I1005 08:16:25.684230 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d061e728-24eb-47ea-bb6e-6495722308ab-config" (OuterVolumeSpecName: "config") pod "d061e728-24eb-47ea-bb6e-6495722308ab" (UID: "d061e728-24eb-47ea-bb6e-6495722308ab"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 08:16:25 crc kubenswrapper[4846]: I1005 08:16:25.724540 4846 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d061e728-24eb-47ea-bb6e-6495722308ab-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 05 08:16:25 crc kubenswrapper[4846]: I1005 08:16:25.724736 4846 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d061e728-24eb-47ea-bb6e-6495722308ab-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 05 08:16:25 crc kubenswrapper[4846]: I1005 08:16:25.724824 4846 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d061e728-24eb-47ea-bb6e-6495722308ab-config\") on node \"crc\" DevicePath \"\"" Oct 05 08:16:25 crc kubenswrapper[4846]: I1005 08:16:25.724887 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fd7jq\" (UniqueName: \"kubernetes.io/projected/d061e728-24eb-47ea-bb6e-6495722308ab-kube-api-access-fd7jq\") on node \"crc\" DevicePath \"\"" Oct 05 08:16:25 crc kubenswrapper[4846]: I1005 08:16:25.724941 4846 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d061e728-24eb-47ea-bb6e-6495722308ab-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 05 08:16:26 crc kubenswrapper[4846]: I1005 08:16:26.395492 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d65c684ff-nnw4f" event={"ID":"d061e728-24eb-47ea-bb6e-6495722308ab","Type":"ContainerDied","Data":"f3cbc8645f852bba0537543aa8dbf4e6a6c407c928059bd0c9517e5e535a852c"} Oct 05 08:16:26 crc kubenswrapper[4846]: I1005 08:16:26.397067 4846 scope.go:117] "RemoveContainer" containerID="9dcbc8dca17e9726f55658375a4dbc3b9d27778f907b45a9136b9ab37b149e77" Oct 05 08:16:26 crc kubenswrapper[4846]: I1005 08:16:26.397470 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6d65c684ff-nnw4f" Oct 05 08:16:26 crc kubenswrapper[4846]: I1005 08:16:26.403552 4846 generic.go:334] "Generic (PLEG): container finished" podID="f5e8843a-2b8b-4e5e-8b32-3e1ff385a83e" containerID="48ee0d9d054943125c654f38836d686d2255e61573b06f5d3288121665374229" exitCode=143 Oct 05 08:16:26 crc kubenswrapper[4846]: I1005 08:16:26.403602 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7546d54c9d-x8sn2" event={"ID":"f5e8843a-2b8b-4e5e-8b32-3e1ff385a83e","Type":"ContainerDied","Data":"48ee0d9d054943125c654f38836d686d2255e61573b06f5d3288121665374229"} Oct 05 08:16:26 crc kubenswrapper[4846]: I1005 08:16:26.432781 4846 scope.go:117] "RemoveContainer" containerID="009ddf7e8407f1a461cf317efe56eadb1df61d474d03e628631d305c904dca0f" Oct 05 08:16:26 crc kubenswrapper[4846]: I1005 08:16:26.457015 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6d65c684ff-nnw4f"] Oct 05 08:16:26 crc kubenswrapper[4846]: I1005 08:16:26.465320 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6d65c684ff-nnw4f"] Oct 05 08:16:26 crc kubenswrapper[4846]: I1005 08:16:26.517732 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d061e728-24eb-47ea-bb6e-6495722308ab" path="/var/lib/kubelet/pods/d061e728-24eb-47ea-bb6e-6495722308ab/volumes" Oct 05 08:16:29 crc kubenswrapper[4846]: I1005 08:16:29.744038 4846 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-7546d54c9d-x8sn2" podUID="f5e8843a-2b8b-4e5e-8b32-3e1ff385a83e" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.1.38:9311/healthcheck\": read tcp 10.217.0.2:47116->10.217.1.38:9311: read: connection reset by peer" Oct 05 08:16:29 crc kubenswrapper[4846]: I1005 08:16:29.744101 4846 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-7546d54c9d-x8sn2" podUID="f5e8843a-2b8b-4e5e-8b32-3e1ff385a83e" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.1.38:9311/healthcheck\": read tcp 10.217.0.2:47108->10.217.1.38:9311: read: connection reset by peer" Oct 05 08:16:30 crc kubenswrapper[4846]: I1005 08:16:30.192827 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-7546d54c9d-x8sn2" Oct 05 08:16:30 crc kubenswrapper[4846]: I1005 08:16:30.311610 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f5e8843a-2b8b-4e5e-8b32-3e1ff385a83e-config-data\") pod \"f5e8843a-2b8b-4e5e-8b32-3e1ff385a83e\" (UID: \"f5e8843a-2b8b-4e5e-8b32-3e1ff385a83e\") " Oct 05 08:16:30 crc kubenswrapper[4846]: I1005 08:16:30.311675 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wkxg7\" (UniqueName: \"kubernetes.io/projected/f5e8843a-2b8b-4e5e-8b32-3e1ff385a83e-kube-api-access-wkxg7\") pod \"f5e8843a-2b8b-4e5e-8b32-3e1ff385a83e\" (UID: \"f5e8843a-2b8b-4e5e-8b32-3e1ff385a83e\") " Oct 05 08:16:30 crc kubenswrapper[4846]: I1005 08:16:30.311718 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f5e8843a-2b8b-4e5e-8b32-3e1ff385a83e-logs\") pod \"f5e8843a-2b8b-4e5e-8b32-3e1ff385a83e\" (UID: \"f5e8843a-2b8b-4e5e-8b32-3e1ff385a83e\") " Oct 05 08:16:30 crc kubenswrapper[4846]: I1005 08:16:30.311757 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f5e8843a-2b8b-4e5e-8b32-3e1ff385a83e-combined-ca-bundle\") pod \"f5e8843a-2b8b-4e5e-8b32-3e1ff385a83e\" (UID: \"f5e8843a-2b8b-4e5e-8b32-3e1ff385a83e\") " Oct 05 08:16:30 crc kubenswrapper[4846]: I1005 08:16:30.311776 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f5e8843a-2b8b-4e5e-8b32-3e1ff385a83e-config-data-custom\") pod \"f5e8843a-2b8b-4e5e-8b32-3e1ff385a83e\" (UID: \"f5e8843a-2b8b-4e5e-8b32-3e1ff385a83e\") " Oct 05 08:16:30 crc kubenswrapper[4846]: I1005 08:16:30.312730 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f5e8843a-2b8b-4e5e-8b32-3e1ff385a83e-logs" (OuterVolumeSpecName: "logs") pod "f5e8843a-2b8b-4e5e-8b32-3e1ff385a83e" (UID: "f5e8843a-2b8b-4e5e-8b32-3e1ff385a83e"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 08:16:30 crc kubenswrapper[4846]: I1005 08:16:30.326005 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f5e8843a-2b8b-4e5e-8b32-3e1ff385a83e-kube-api-access-wkxg7" (OuterVolumeSpecName: "kube-api-access-wkxg7") pod "f5e8843a-2b8b-4e5e-8b32-3e1ff385a83e" (UID: "f5e8843a-2b8b-4e5e-8b32-3e1ff385a83e"). InnerVolumeSpecName "kube-api-access-wkxg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:16:30 crc kubenswrapper[4846]: I1005 08:16:30.326021 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f5e8843a-2b8b-4e5e-8b32-3e1ff385a83e-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "f5e8843a-2b8b-4e5e-8b32-3e1ff385a83e" (UID: "f5e8843a-2b8b-4e5e-8b32-3e1ff385a83e"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:16:30 crc kubenswrapper[4846]: I1005 08:16:30.344349 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f5e8843a-2b8b-4e5e-8b32-3e1ff385a83e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f5e8843a-2b8b-4e5e-8b32-3e1ff385a83e" (UID: "f5e8843a-2b8b-4e5e-8b32-3e1ff385a83e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:16:30 crc kubenswrapper[4846]: I1005 08:16:30.380846 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f5e8843a-2b8b-4e5e-8b32-3e1ff385a83e-config-data" (OuterVolumeSpecName: "config-data") pod "f5e8843a-2b8b-4e5e-8b32-3e1ff385a83e" (UID: "f5e8843a-2b8b-4e5e-8b32-3e1ff385a83e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:16:30 crc kubenswrapper[4846]: I1005 08:16:30.413552 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wkxg7\" (UniqueName: \"kubernetes.io/projected/f5e8843a-2b8b-4e5e-8b32-3e1ff385a83e-kube-api-access-wkxg7\") on node \"crc\" DevicePath \"\"" Oct 05 08:16:30 crc kubenswrapper[4846]: I1005 08:16:30.413591 4846 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f5e8843a-2b8b-4e5e-8b32-3e1ff385a83e-logs\") on node \"crc\" DevicePath \"\"" Oct 05 08:16:30 crc kubenswrapper[4846]: I1005 08:16:30.413606 4846 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f5e8843a-2b8b-4e5e-8b32-3e1ff385a83e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 08:16:30 crc kubenswrapper[4846]: I1005 08:16:30.413620 4846 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f5e8843a-2b8b-4e5e-8b32-3e1ff385a83e-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 05 08:16:30 crc kubenswrapper[4846]: I1005 08:16:30.413631 4846 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f5e8843a-2b8b-4e5e-8b32-3e1ff385a83e-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 08:16:30 crc kubenswrapper[4846]: I1005 08:16:30.463654 4846 generic.go:334] "Generic (PLEG): container finished" podID="f5e8843a-2b8b-4e5e-8b32-3e1ff385a83e" containerID="8bf92f2421b601ab85762e0732103c649287c9804192bb0d0c05ba6abc8c0e60" exitCode=0 Oct 05 08:16:30 crc kubenswrapper[4846]: I1005 08:16:30.463732 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-7546d54c9d-x8sn2" Oct 05 08:16:30 crc kubenswrapper[4846]: I1005 08:16:30.463730 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7546d54c9d-x8sn2" event={"ID":"f5e8843a-2b8b-4e5e-8b32-3e1ff385a83e","Type":"ContainerDied","Data":"8bf92f2421b601ab85762e0732103c649287c9804192bb0d0c05ba6abc8c0e60"} Oct 05 08:16:30 crc kubenswrapper[4846]: I1005 08:16:30.463855 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7546d54c9d-x8sn2" event={"ID":"f5e8843a-2b8b-4e5e-8b32-3e1ff385a83e","Type":"ContainerDied","Data":"dec9d3c43452efe5a7b6eaca3d23d537312e968f30dafff2be851b19caa6b43c"} Oct 05 08:16:30 crc kubenswrapper[4846]: I1005 08:16:30.463902 4846 scope.go:117] "RemoveContainer" containerID="8bf92f2421b601ab85762e0732103c649287c9804192bb0d0c05ba6abc8c0e60" Oct 05 08:16:30 crc kubenswrapper[4846]: I1005 08:16:30.483282 4846 scope.go:117] "RemoveContainer" containerID="48ee0d9d054943125c654f38836d686d2255e61573b06f5d3288121665374229" Oct 05 08:16:30 crc kubenswrapper[4846]: I1005 08:16:30.502134 4846 scope.go:117] "RemoveContainer" containerID="8bf92f2421b601ab85762e0732103c649287c9804192bb0d0c05ba6abc8c0e60" Oct 05 08:16:30 crc kubenswrapper[4846]: E1005 08:16:30.502620 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8bf92f2421b601ab85762e0732103c649287c9804192bb0d0c05ba6abc8c0e60\": container with ID starting with 8bf92f2421b601ab85762e0732103c649287c9804192bb0d0c05ba6abc8c0e60 not found: ID does not exist" containerID="8bf92f2421b601ab85762e0732103c649287c9804192bb0d0c05ba6abc8c0e60" Oct 05 08:16:30 crc kubenswrapper[4846]: I1005 08:16:30.502673 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8bf92f2421b601ab85762e0732103c649287c9804192bb0d0c05ba6abc8c0e60"} err="failed to get container status \"8bf92f2421b601ab85762e0732103c649287c9804192bb0d0c05ba6abc8c0e60\": rpc error: code = NotFound desc = could not find container \"8bf92f2421b601ab85762e0732103c649287c9804192bb0d0c05ba6abc8c0e60\": container with ID starting with 8bf92f2421b601ab85762e0732103c649287c9804192bb0d0c05ba6abc8c0e60 not found: ID does not exist" Oct 05 08:16:30 crc kubenswrapper[4846]: I1005 08:16:30.502707 4846 scope.go:117] "RemoveContainer" containerID="48ee0d9d054943125c654f38836d686d2255e61573b06f5d3288121665374229" Oct 05 08:16:30 crc kubenswrapper[4846]: E1005 08:16:30.502982 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"48ee0d9d054943125c654f38836d686d2255e61573b06f5d3288121665374229\": container with ID starting with 48ee0d9d054943125c654f38836d686d2255e61573b06f5d3288121665374229 not found: ID does not exist" containerID="48ee0d9d054943125c654f38836d686d2255e61573b06f5d3288121665374229" Oct 05 08:16:30 crc kubenswrapper[4846]: I1005 08:16:30.503024 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"48ee0d9d054943125c654f38836d686d2255e61573b06f5d3288121665374229"} err="failed to get container status \"48ee0d9d054943125c654f38836d686d2255e61573b06f5d3288121665374229\": rpc error: code = NotFound desc = could not find container \"48ee0d9d054943125c654f38836d686d2255e61573b06f5d3288121665374229\": container with ID starting with 48ee0d9d054943125c654f38836d686d2255e61573b06f5d3288121665374229 not found: ID does not exist" Oct 05 08:16:30 crc kubenswrapper[4846]: I1005 08:16:30.508655 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-7546d54c9d-x8sn2"] Oct 05 08:16:30 crc kubenswrapper[4846]: I1005 08:16:30.513249 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-7546d54c9d-x8sn2"] Oct 05 08:16:32 crc kubenswrapper[4846]: I1005 08:16:32.506867 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f5e8843a-2b8b-4e5e-8b32-3e1ff385a83e" path="/var/lib/kubelet/pods/f5e8843a-2b8b-4e5e-8b32-3e1ff385a83e/volumes" Oct 05 08:16:34 crc kubenswrapper[4846]: I1005 08:16:34.938819 4846 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-7546d54c9d-x8sn2" podUID="f5e8843a-2b8b-4e5e-8b32-3e1ff385a83e" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.1.38:9311/healthcheck\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 05 08:16:34 crc kubenswrapper[4846]: I1005 08:16:34.938986 4846 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-7546d54c9d-x8sn2" podUID="f5e8843a-2b8b-4e5e-8b32-3e1ff385a83e" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.1.38:9311/healthcheck\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 05 08:17:02 crc kubenswrapper[4846]: I1005 08:17:02.302839 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-create-p9zcj"] Oct 05 08:17:02 crc kubenswrapper[4846]: E1005 08:17:02.303784 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f5e8843a-2b8b-4e5e-8b32-3e1ff385a83e" containerName="barbican-api" Oct 05 08:17:02 crc kubenswrapper[4846]: I1005 08:17:02.303800 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="f5e8843a-2b8b-4e5e-8b32-3e1ff385a83e" containerName="barbican-api" Oct 05 08:17:02 crc kubenswrapper[4846]: E1005 08:17:02.303820 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f5e8843a-2b8b-4e5e-8b32-3e1ff385a83e" containerName="barbican-api-log" Oct 05 08:17:02 crc kubenswrapper[4846]: I1005 08:17:02.303830 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="f5e8843a-2b8b-4e5e-8b32-3e1ff385a83e" containerName="barbican-api-log" Oct 05 08:17:02 crc kubenswrapper[4846]: E1005 08:17:02.303854 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d061e728-24eb-47ea-bb6e-6495722308ab" containerName="dnsmasq-dns" Oct 05 08:17:02 crc kubenswrapper[4846]: I1005 08:17:02.303863 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="d061e728-24eb-47ea-bb6e-6495722308ab" containerName="dnsmasq-dns" Oct 05 08:17:02 crc kubenswrapper[4846]: E1005 08:17:02.303882 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d061e728-24eb-47ea-bb6e-6495722308ab" containerName="init" Oct 05 08:17:02 crc kubenswrapper[4846]: I1005 08:17:02.303890 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="d061e728-24eb-47ea-bb6e-6495722308ab" containerName="init" Oct 05 08:17:02 crc kubenswrapper[4846]: I1005 08:17:02.304110 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="f5e8843a-2b8b-4e5e-8b32-3e1ff385a83e" containerName="barbican-api-log" Oct 05 08:17:02 crc kubenswrapper[4846]: I1005 08:17:02.304126 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="d061e728-24eb-47ea-bb6e-6495722308ab" containerName="dnsmasq-dns" Oct 05 08:17:02 crc kubenswrapper[4846]: I1005 08:17:02.304140 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="f5e8843a-2b8b-4e5e-8b32-3e1ff385a83e" containerName="barbican-api" Oct 05 08:17:02 crc kubenswrapper[4846]: I1005 08:17:02.304825 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-p9zcj" Oct 05 08:17:02 crc kubenswrapper[4846]: I1005 08:17:02.316334 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-p9zcj"] Oct 05 08:17:02 crc kubenswrapper[4846]: I1005 08:17:02.430383 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g8dvz\" (UniqueName: \"kubernetes.io/projected/0cc5829c-be80-4bf4-adec-b18b8dd4aca5-kube-api-access-g8dvz\") pod \"neutron-db-create-p9zcj\" (UID: \"0cc5829c-be80-4bf4-adec-b18b8dd4aca5\") " pod="openstack/neutron-db-create-p9zcj" Oct 05 08:17:02 crc kubenswrapper[4846]: I1005 08:17:02.531767 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g8dvz\" (UniqueName: \"kubernetes.io/projected/0cc5829c-be80-4bf4-adec-b18b8dd4aca5-kube-api-access-g8dvz\") pod \"neutron-db-create-p9zcj\" (UID: \"0cc5829c-be80-4bf4-adec-b18b8dd4aca5\") " pod="openstack/neutron-db-create-p9zcj" Oct 05 08:17:02 crc kubenswrapper[4846]: I1005 08:17:02.550335 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g8dvz\" (UniqueName: \"kubernetes.io/projected/0cc5829c-be80-4bf4-adec-b18b8dd4aca5-kube-api-access-g8dvz\") pod \"neutron-db-create-p9zcj\" (UID: \"0cc5829c-be80-4bf4-adec-b18b8dd4aca5\") " pod="openstack/neutron-db-create-p9zcj" Oct 05 08:17:02 crc kubenswrapper[4846]: I1005 08:17:02.646030 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-p9zcj" Oct 05 08:17:03 crc kubenswrapper[4846]: I1005 08:17:03.102999 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-p9zcj"] Oct 05 08:17:03 crc kubenswrapper[4846]: W1005 08:17:03.107483 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0cc5829c_be80_4bf4_adec_b18b8dd4aca5.slice/crio-0d585b4f268d9a29d58e9c89c99ec35fed90eb86f34ee8c8e0d5263be8065789 WatchSource:0}: Error finding container 0d585b4f268d9a29d58e9c89c99ec35fed90eb86f34ee8c8e0d5263be8065789: Status 404 returned error can't find the container with id 0d585b4f268d9a29d58e9c89c99ec35fed90eb86f34ee8c8e0d5263be8065789 Oct 05 08:17:03 crc kubenswrapper[4846]: I1005 08:17:03.812885 4846 generic.go:334] "Generic (PLEG): container finished" podID="0cc5829c-be80-4bf4-adec-b18b8dd4aca5" containerID="16eb4369807860d0dae16b3f48adac19c01ac1eea08e5434c521d62557752894" exitCode=0 Oct 05 08:17:03 crc kubenswrapper[4846]: I1005 08:17:03.813114 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-p9zcj" event={"ID":"0cc5829c-be80-4bf4-adec-b18b8dd4aca5","Type":"ContainerDied","Data":"16eb4369807860d0dae16b3f48adac19c01ac1eea08e5434c521d62557752894"} Oct 05 08:17:03 crc kubenswrapper[4846]: I1005 08:17:03.813250 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-p9zcj" event={"ID":"0cc5829c-be80-4bf4-adec-b18b8dd4aca5","Type":"ContainerStarted","Data":"0d585b4f268d9a29d58e9c89c99ec35fed90eb86f34ee8c8e0d5263be8065789"} Oct 05 08:17:05 crc kubenswrapper[4846]: I1005 08:17:05.330783 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-p9zcj" Oct 05 08:17:05 crc kubenswrapper[4846]: I1005 08:17:05.507976 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g8dvz\" (UniqueName: \"kubernetes.io/projected/0cc5829c-be80-4bf4-adec-b18b8dd4aca5-kube-api-access-g8dvz\") pod \"0cc5829c-be80-4bf4-adec-b18b8dd4aca5\" (UID: \"0cc5829c-be80-4bf4-adec-b18b8dd4aca5\") " Oct 05 08:17:05 crc kubenswrapper[4846]: I1005 08:17:05.514419 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0cc5829c-be80-4bf4-adec-b18b8dd4aca5-kube-api-access-g8dvz" (OuterVolumeSpecName: "kube-api-access-g8dvz") pod "0cc5829c-be80-4bf4-adec-b18b8dd4aca5" (UID: "0cc5829c-be80-4bf4-adec-b18b8dd4aca5"). InnerVolumeSpecName "kube-api-access-g8dvz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:17:05 crc kubenswrapper[4846]: I1005 08:17:05.610014 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g8dvz\" (UniqueName: \"kubernetes.io/projected/0cc5829c-be80-4bf4-adec-b18b8dd4aca5-kube-api-access-g8dvz\") on node \"crc\" DevicePath \"\"" Oct 05 08:17:05 crc kubenswrapper[4846]: I1005 08:17:05.839617 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-p9zcj" event={"ID":"0cc5829c-be80-4bf4-adec-b18b8dd4aca5","Type":"ContainerDied","Data":"0d585b4f268d9a29d58e9c89c99ec35fed90eb86f34ee8c8e0d5263be8065789"} Oct 05 08:17:05 crc kubenswrapper[4846]: I1005 08:17:05.839687 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0d585b4f268d9a29d58e9c89c99ec35fed90eb86f34ee8c8e0d5263be8065789" Oct 05 08:17:05 crc kubenswrapper[4846]: I1005 08:17:05.839685 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-p9zcj" Oct 05 08:17:12 crc kubenswrapper[4846]: I1005 08:17:12.371747 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-e233-account-create-h5tj4"] Oct 05 08:17:12 crc kubenswrapper[4846]: E1005 08:17:12.372805 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0cc5829c-be80-4bf4-adec-b18b8dd4aca5" containerName="mariadb-database-create" Oct 05 08:17:12 crc kubenswrapper[4846]: I1005 08:17:12.372825 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="0cc5829c-be80-4bf4-adec-b18b8dd4aca5" containerName="mariadb-database-create" Oct 05 08:17:12 crc kubenswrapper[4846]: I1005 08:17:12.373090 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="0cc5829c-be80-4bf4-adec-b18b8dd4aca5" containerName="mariadb-database-create" Oct 05 08:17:12 crc kubenswrapper[4846]: I1005 08:17:12.373981 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-e233-account-create-h5tj4" Oct 05 08:17:12 crc kubenswrapper[4846]: I1005 08:17:12.376566 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-db-secret" Oct 05 08:17:12 crc kubenswrapper[4846]: I1005 08:17:12.401419 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-e233-account-create-h5tj4"] Oct 05 08:17:12 crc kubenswrapper[4846]: I1005 08:17:12.550618 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s4fw6\" (UniqueName: \"kubernetes.io/projected/9af60fac-0c55-45b4-9752-955f8df372d5-kube-api-access-s4fw6\") pod \"neutron-e233-account-create-h5tj4\" (UID: \"9af60fac-0c55-45b4-9752-955f8df372d5\") " pod="openstack/neutron-e233-account-create-h5tj4" Oct 05 08:17:12 crc kubenswrapper[4846]: I1005 08:17:12.652020 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s4fw6\" (UniqueName: \"kubernetes.io/projected/9af60fac-0c55-45b4-9752-955f8df372d5-kube-api-access-s4fw6\") pod \"neutron-e233-account-create-h5tj4\" (UID: \"9af60fac-0c55-45b4-9752-955f8df372d5\") " pod="openstack/neutron-e233-account-create-h5tj4" Oct 05 08:17:12 crc kubenswrapper[4846]: I1005 08:17:12.675862 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s4fw6\" (UniqueName: \"kubernetes.io/projected/9af60fac-0c55-45b4-9752-955f8df372d5-kube-api-access-s4fw6\") pod \"neutron-e233-account-create-h5tj4\" (UID: \"9af60fac-0c55-45b4-9752-955f8df372d5\") " pod="openstack/neutron-e233-account-create-h5tj4" Oct 05 08:17:12 crc kubenswrapper[4846]: I1005 08:17:12.723883 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-e233-account-create-h5tj4" Oct 05 08:17:13 crc kubenswrapper[4846]: I1005 08:17:13.200059 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-e233-account-create-h5tj4"] Oct 05 08:17:13 crc kubenswrapper[4846]: I1005 08:17:13.937443 4846 generic.go:334] "Generic (PLEG): container finished" podID="9af60fac-0c55-45b4-9752-955f8df372d5" containerID="dfa0b5dfa965bb1c9af8af1b32b0e6a7337a6350d305c54ed95ea8a8ad9ee9c6" exitCode=0 Oct 05 08:17:13 crc kubenswrapper[4846]: I1005 08:17:13.937552 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-e233-account-create-h5tj4" event={"ID":"9af60fac-0c55-45b4-9752-955f8df372d5","Type":"ContainerDied","Data":"dfa0b5dfa965bb1c9af8af1b32b0e6a7337a6350d305c54ed95ea8a8ad9ee9c6"} Oct 05 08:17:13 crc kubenswrapper[4846]: I1005 08:17:13.937923 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-e233-account-create-h5tj4" event={"ID":"9af60fac-0c55-45b4-9752-955f8df372d5","Type":"ContainerStarted","Data":"78e6dff8367042b692718f64b5322dfacc66ab383d703a50c281babd68cfd421"} Oct 05 08:17:15 crc kubenswrapper[4846]: I1005 08:17:15.254668 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-e233-account-create-h5tj4" Oct 05 08:17:15 crc kubenswrapper[4846]: I1005 08:17:15.397782 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4fw6\" (UniqueName: \"kubernetes.io/projected/9af60fac-0c55-45b4-9752-955f8df372d5-kube-api-access-s4fw6\") pod \"9af60fac-0c55-45b4-9752-955f8df372d5\" (UID: \"9af60fac-0c55-45b4-9752-955f8df372d5\") " Oct 05 08:17:15 crc kubenswrapper[4846]: I1005 08:17:15.404937 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9af60fac-0c55-45b4-9752-955f8df372d5-kube-api-access-s4fw6" (OuterVolumeSpecName: "kube-api-access-s4fw6") pod "9af60fac-0c55-45b4-9752-955f8df372d5" (UID: "9af60fac-0c55-45b4-9752-955f8df372d5"). InnerVolumeSpecName "kube-api-access-s4fw6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:17:15 crc kubenswrapper[4846]: I1005 08:17:15.500849 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4fw6\" (UniqueName: \"kubernetes.io/projected/9af60fac-0c55-45b4-9752-955f8df372d5-kube-api-access-s4fw6\") on node \"crc\" DevicePath \"\"" Oct 05 08:17:15 crc kubenswrapper[4846]: I1005 08:17:15.966506 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-e233-account-create-h5tj4" event={"ID":"9af60fac-0c55-45b4-9752-955f8df372d5","Type":"ContainerDied","Data":"78e6dff8367042b692718f64b5322dfacc66ab383d703a50c281babd68cfd421"} Oct 05 08:17:15 crc kubenswrapper[4846]: I1005 08:17:15.966672 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="78e6dff8367042b692718f64b5322dfacc66ab383d703a50c281babd68cfd421" Oct 05 08:17:15 crc kubenswrapper[4846]: I1005 08:17:15.966857 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-e233-account-create-h5tj4" Oct 05 08:17:17 crc kubenswrapper[4846]: I1005 08:17:17.582890 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-sync-42dlq"] Oct 05 08:17:17 crc kubenswrapper[4846]: E1005 08:17:17.583544 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9af60fac-0c55-45b4-9752-955f8df372d5" containerName="mariadb-account-create" Oct 05 08:17:17 crc kubenswrapper[4846]: I1005 08:17:17.583562 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="9af60fac-0c55-45b4-9752-955f8df372d5" containerName="mariadb-account-create" Oct 05 08:17:17 crc kubenswrapper[4846]: I1005 08:17:17.583804 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="9af60fac-0c55-45b4-9752-955f8df372d5" containerName="mariadb-account-create" Oct 05 08:17:17 crc kubenswrapper[4846]: I1005 08:17:17.584699 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-42dlq" Oct 05 08:17:17 crc kubenswrapper[4846]: I1005 08:17:17.587382 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Oct 05 08:17:17 crc kubenswrapper[4846]: I1005 08:17:17.587472 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-nmbv2" Oct 05 08:17:17 crc kubenswrapper[4846]: I1005 08:17:17.587658 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Oct 05 08:17:17 crc kubenswrapper[4846]: I1005 08:17:17.603400 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-42dlq"] Oct 05 08:17:17 crc kubenswrapper[4846]: I1005 08:17:17.738141 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e2077da3-902b-48e2-93ef-be196e85b88d-combined-ca-bundle\") pod \"neutron-db-sync-42dlq\" (UID: \"e2077da3-902b-48e2-93ef-be196e85b88d\") " pod="openstack/neutron-db-sync-42dlq" Oct 05 08:17:17 crc kubenswrapper[4846]: I1005 08:17:17.738346 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/e2077da3-902b-48e2-93ef-be196e85b88d-config\") pod \"neutron-db-sync-42dlq\" (UID: \"e2077da3-902b-48e2-93ef-be196e85b88d\") " pod="openstack/neutron-db-sync-42dlq" Oct 05 08:17:17 crc kubenswrapper[4846]: I1005 08:17:17.738409 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ztbbm\" (UniqueName: \"kubernetes.io/projected/e2077da3-902b-48e2-93ef-be196e85b88d-kube-api-access-ztbbm\") pod \"neutron-db-sync-42dlq\" (UID: \"e2077da3-902b-48e2-93ef-be196e85b88d\") " pod="openstack/neutron-db-sync-42dlq" Oct 05 08:17:17 crc kubenswrapper[4846]: I1005 08:17:17.840304 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e2077da3-902b-48e2-93ef-be196e85b88d-combined-ca-bundle\") pod \"neutron-db-sync-42dlq\" (UID: \"e2077da3-902b-48e2-93ef-be196e85b88d\") " pod="openstack/neutron-db-sync-42dlq" Oct 05 08:17:17 crc kubenswrapper[4846]: I1005 08:17:17.840389 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/e2077da3-902b-48e2-93ef-be196e85b88d-config\") pod \"neutron-db-sync-42dlq\" (UID: \"e2077da3-902b-48e2-93ef-be196e85b88d\") " pod="openstack/neutron-db-sync-42dlq" Oct 05 08:17:17 crc kubenswrapper[4846]: I1005 08:17:17.840416 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ztbbm\" (UniqueName: \"kubernetes.io/projected/e2077da3-902b-48e2-93ef-be196e85b88d-kube-api-access-ztbbm\") pod \"neutron-db-sync-42dlq\" (UID: \"e2077da3-902b-48e2-93ef-be196e85b88d\") " pod="openstack/neutron-db-sync-42dlq" Oct 05 08:17:17 crc kubenswrapper[4846]: I1005 08:17:17.848044 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/e2077da3-902b-48e2-93ef-be196e85b88d-config\") pod \"neutron-db-sync-42dlq\" (UID: \"e2077da3-902b-48e2-93ef-be196e85b88d\") " pod="openstack/neutron-db-sync-42dlq" Oct 05 08:17:17 crc kubenswrapper[4846]: I1005 08:17:17.848468 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e2077da3-902b-48e2-93ef-be196e85b88d-combined-ca-bundle\") pod \"neutron-db-sync-42dlq\" (UID: \"e2077da3-902b-48e2-93ef-be196e85b88d\") " pod="openstack/neutron-db-sync-42dlq" Oct 05 08:17:17 crc kubenswrapper[4846]: I1005 08:17:17.869779 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ztbbm\" (UniqueName: \"kubernetes.io/projected/e2077da3-902b-48e2-93ef-be196e85b88d-kube-api-access-ztbbm\") pod \"neutron-db-sync-42dlq\" (UID: \"e2077da3-902b-48e2-93ef-be196e85b88d\") " pod="openstack/neutron-db-sync-42dlq" Oct 05 08:17:17 crc kubenswrapper[4846]: I1005 08:17:17.912459 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-42dlq" Oct 05 08:17:18 crc kubenswrapper[4846]: I1005 08:17:18.366447 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-mlfl2"] Oct 05 08:17:18 crc kubenswrapper[4846]: I1005 08:17:18.368826 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mlfl2" Oct 05 08:17:18 crc kubenswrapper[4846]: I1005 08:17:18.382327 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-42dlq"] Oct 05 08:17:18 crc kubenswrapper[4846]: I1005 08:17:18.393135 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-mlfl2"] Oct 05 08:17:18 crc kubenswrapper[4846]: I1005 08:17:18.552070 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mh5ck\" (UniqueName: \"kubernetes.io/projected/bcfb4bd3-9712-466d-bc48-317857cdd151-kube-api-access-mh5ck\") pod \"redhat-marketplace-mlfl2\" (UID: \"bcfb4bd3-9712-466d-bc48-317857cdd151\") " pod="openshift-marketplace/redhat-marketplace-mlfl2" Oct 05 08:17:18 crc kubenswrapper[4846]: I1005 08:17:18.552472 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bcfb4bd3-9712-466d-bc48-317857cdd151-catalog-content\") pod \"redhat-marketplace-mlfl2\" (UID: \"bcfb4bd3-9712-466d-bc48-317857cdd151\") " pod="openshift-marketplace/redhat-marketplace-mlfl2" Oct 05 08:17:18 crc kubenswrapper[4846]: I1005 08:17:18.552505 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bcfb4bd3-9712-466d-bc48-317857cdd151-utilities\") pod \"redhat-marketplace-mlfl2\" (UID: \"bcfb4bd3-9712-466d-bc48-317857cdd151\") " pod="openshift-marketplace/redhat-marketplace-mlfl2" Oct 05 08:17:18 crc kubenswrapper[4846]: I1005 08:17:18.654396 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mh5ck\" (UniqueName: \"kubernetes.io/projected/bcfb4bd3-9712-466d-bc48-317857cdd151-kube-api-access-mh5ck\") pod \"redhat-marketplace-mlfl2\" (UID: \"bcfb4bd3-9712-466d-bc48-317857cdd151\") " pod="openshift-marketplace/redhat-marketplace-mlfl2" Oct 05 08:17:18 crc kubenswrapper[4846]: I1005 08:17:18.654482 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bcfb4bd3-9712-466d-bc48-317857cdd151-catalog-content\") pod \"redhat-marketplace-mlfl2\" (UID: \"bcfb4bd3-9712-466d-bc48-317857cdd151\") " pod="openshift-marketplace/redhat-marketplace-mlfl2" Oct 05 08:17:18 crc kubenswrapper[4846]: I1005 08:17:18.655078 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bcfb4bd3-9712-466d-bc48-317857cdd151-catalog-content\") pod \"redhat-marketplace-mlfl2\" (UID: \"bcfb4bd3-9712-466d-bc48-317857cdd151\") " pod="openshift-marketplace/redhat-marketplace-mlfl2" Oct 05 08:17:18 crc kubenswrapper[4846]: I1005 08:17:18.655119 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bcfb4bd3-9712-466d-bc48-317857cdd151-utilities\") pod \"redhat-marketplace-mlfl2\" (UID: \"bcfb4bd3-9712-466d-bc48-317857cdd151\") " pod="openshift-marketplace/redhat-marketplace-mlfl2" Oct 05 08:17:18 crc kubenswrapper[4846]: I1005 08:17:18.655124 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bcfb4bd3-9712-466d-bc48-317857cdd151-utilities\") pod \"redhat-marketplace-mlfl2\" (UID: \"bcfb4bd3-9712-466d-bc48-317857cdd151\") " pod="openshift-marketplace/redhat-marketplace-mlfl2" Oct 05 08:17:18 crc kubenswrapper[4846]: I1005 08:17:18.675841 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mh5ck\" (UniqueName: \"kubernetes.io/projected/bcfb4bd3-9712-466d-bc48-317857cdd151-kube-api-access-mh5ck\") pod \"redhat-marketplace-mlfl2\" (UID: \"bcfb4bd3-9712-466d-bc48-317857cdd151\") " pod="openshift-marketplace/redhat-marketplace-mlfl2" Oct 05 08:17:18 crc kubenswrapper[4846]: I1005 08:17:18.741134 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mlfl2" Oct 05 08:17:18 crc kubenswrapper[4846]: I1005 08:17:18.994512 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-42dlq" event={"ID":"e2077da3-902b-48e2-93ef-be196e85b88d","Type":"ContainerStarted","Data":"21f9d172a502fb5a7fb911386591365ff76f834fae51cb1409391d9b9c1ced56"} Oct 05 08:17:18 crc kubenswrapper[4846]: I1005 08:17:18.994750 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-42dlq" event={"ID":"e2077da3-902b-48e2-93ef-be196e85b88d","Type":"ContainerStarted","Data":"559214a25868d5630117f8a9e0cc4856e13176719e458ad0fe07b0650bb3206c"} Oct 05 08:17:19 crc kubenswrapper[4846]: I1005 08:17:19.011400 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-sync-42dlq" podStartSLOduration=2.011384893 podStartE2EDuration="2.011384893s" podCreationTimestamp="2025-10-05 08:17:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 08:17:19.00904827 +0000 UTC m=+5361.249901065" watchObservedRunningTime="2025-10-05 08:17:19.011384893 +0000 UTC m=+5361.252237668" Oct 05 08:17:19 crc kubenswrapper[4846]: I1005 08:17:19.198334 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-mlfl2"] Oct 05 08:17:19 crc kubenswrapper[4846]: W1005 08:17:19.211896 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbcfb4bd3_9712_466d_bc48_317857cdd151.slice/crio-c46046c40cb1c5230a9c2fd1bd8c484bf1f620bc8d16601261722529f20296cd WatchSource:0}: Error finding container c46046c40cb1c5230a9c2fd1bd8c484bf1f620bc8d16601261722529f20296cd: Status 404 returned error can't find the container with id c46046c40cb1c5230a9c2fd1bd8c484bf1f620bc8d16601261722529f20296cd Oct 05 08:17:20 crc kubenswrapper[4846]: I1005 08:17:20.005105 4846 generic.go:334] "Generic (PLEG): container finished" podID="bcfb4bd3-9712-466d-bc48-317857cdd151" containerID="8250648ccf9655e57ac2a743dfbaf5aa20a58bb6e6eb57f404fc31adad106b2d" exitCode=0 Oct 05 08:17:20 crc kubenswrapper[4846]: I1005 08:17:20.005202 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mlfl2" event={"ID":"bcfb4bd3-9712-466d-bc48-317857cdd151","Type":"ContainerDied","Data":"8250648ccf9655e57ac2a743dfbaf5aa20a58bb6e6eb57f404fc31adad106b2d"} Oct 05 08:17:20 crc kubenswrapper[4846]: I1005 08:17:20.005576 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mlfl2" event={"ID":"bcfb4bd3-9712-466d-bc48-317857cdd151","Type":"ContainerStarted","Data":"c46046c40cb1c5230a9c2fd1bd8c484bf1f620bc8d16601261722529f20296cd"} Oct 05 08:17:21 crc kubenswrapper[4846]: I1005 08:17:21.016588 4846 generic.go:334] "Generic (PLEG): container finished" podID="bcfb4bd3-9712-466d-bc48-317857cdd151" containerID="b7760653459fee5ad898728947509c6903e698373c8cf35e05416615b2b08cdb" exitCode=0 Oct 05 08:17:21 crc kubenswrapper[4846]: I1005 08:17:21.016639 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mlfl2" event={"ID":"bcfb4bd3-9712-466d-bc48-317857cdd151","Type":"ContainerDied","Data":"b7760653459fee5ad898728947509c6903e698373c8cf35e05416615b2b08cdb"} Oct 05 08:17:22 crc kubenswrapper[4846]: I1005 08:17:22.028152 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mlfl2" event={"ID":"bcfb4bd3-9712-466d-bc48-317857cdd151","Type":"ContainerStarted","Data":"df1bee18b97987a12e04c5fb5710b5c962a880e700a8dd484232762e157c4d1c"} Oct 05 08:17:22 crc kubenswrapper[4846]: I1005 08:17:22.051358 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-mlfl2" podStartSLOduration=2.580098902 podStartE2EDuration="4.051340439s" podCreationTimestamp="2025-10-05 08:17:18 +0000 UTC" firstStartedPulling="2025-10-05 08:17:20.007741681 +0000 UTC m=+5362.248594456" lastFinishedPulling="2025-10-05 08:17:21.478983218 +0000 UTC m=+5363.719835993" observedRunningTime="2025-10-05 08:17:22.050907188 +0000 UTC m=+5364.291759993" watchObservedRunningTime="2025-10-05 08:17:22.051340439 +0000 UTC m=+5364.292193214" Oct 05 08:17:23 crc kubenswrapper[4846]: I1005 08:17:23.041695 4846 generic.go:334] "Generic (PLEG): container finished" podID="e2077da3-902b-48e2-93ef-be196e85b88d" containerID="21f9d172a502fb5a7fb911386591365ff76f834fae51cb1409391d9b9c1ced56" exitCode=0 Oct 05 08:17:23 crc kubenswrapper[4846]: I1005 08:17:23.041798 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-42dlq" event={"ID":"e2077da3-902b-48e2-93ef-be196e85b88d","Type":"ContainerDied","Data":"21f9d172a502fb5a7fb911386591365ff76f834fae51cb1409391d9b9c1ced56"} Oct 05 08:17:24 crc kubenswrapper[4846]: I1005 08:17:24.452425 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-42dlq" Oct 05 08:17:24 crc kubenswrapper[4846]: I1005 08:17:24.571778 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/e2077da3-902b-48e2-93ef-be196e85b88d-config\") pod \"e2077da3-902b-48e2-93ef-be196e85b88d\" (UID: \"e2077da3-902b-48e2-93ef-be196e85b88d\") " Oct 05 08:17:24 crc kubenswrapper[4846]: I1005 08:17:24.572128 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e2077da3-902b-48e2-93ef-be196e85b88d-combined-ca-bundle\") pod \"e2077da3-902b-48e2-93ef-be196e85b88d\" (UID: \"e2077da3-902b-48e2-93ef-be196e85b88d\") " Oct 05 08:17:24 crc kubenswrapper[4846]: I1005 08:17:24.572366 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ztbbm\" (UniqueName: \"kubernetes.io/projected/e2077da3-902b-48e2-93ef-be196e85b88d-kube-api-access-ztbbm\") pod \"e2077da3-902b-48e2-93ef-be196e85b88d\" (UID: \"e2077da3-902b-48e2-93ef-be196e85b88d\") " Oct 05 08:17:24 crc kubenswrapper[4846]: I1005 08:17:24.576962 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e2077da3-902b-48e2-93ef-be196e85b88d-kube-api-access-ztbbm" (OuterVolumeSpecName: "kube-api-access-ztbbm") pod "e2077da3-902b-48e2-93ef-be196e85b88d" (UID: "e2077da3-902b-48e2-93ef-be196e85b88d"). InnerVolumeSpecName "kube-api-access-ztbbm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:17:24 crc kubenswrapper[4846]: I1005 08:17:24.595455 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e2077da3-902b-48e2-93ef-be196e85b88d-config" (OuterVolumeSpecName: "config") pod "e2077da3-902b-48e2-93ef-be196e85b88d" (UID: "e2077da3-902b-48e2-93ef-be196e85b88d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:17:24 crc kubenswrapper[4846]: I1005 08:17:24.595861 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e2077da3-902b-48e2-93ef-be196e85b88d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e2077da3-902b-48e2-93ef-be196e85b88d" (UID: "e2077da3-902b-48e2-93ef-be196e85b88d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:17:24 crc kubenswrapper[4846]: I1005 08:17:24.675372 4846 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/e2077da3-902b-48e2-93ef-be196e85b88d-config\") on node \"crc\" DevicePath \"\"" Oct 05 08:17:24 crc kubenswrapper[4846]: I1005 08:17:24.675423 4846 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e2077da3-902b-48e2-93ef-be196e85b88d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 08:17:24 crc kubenswrapper[4846]: I1005 08:17:24.675444 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ztbbm\" (UniqueName: \"kubernetes.io/projected/e2077da3-902b-48e2-93ef-be196e85b88d-kube-api-access-ztbbm\") on node \"crc\" DevicePath \"\"" Oct 05 08:17:25 crc kubenswrapper[4846]: I1005 08:17:25.076861 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-42dlq" event={"ID":"e2077da3-902b-48e2-93ef-be196e85b88d","Type":"ContainerDied","Data":"559214a25868d5630117f8a9e0cc4856e13176719e458ad0fe07b0650bb3206c"} Oct 05 08:17:25 crc kubenswrapper[4846]: I1005 08:17:25.076925 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-42dlq" Oct 05 08:17:25 crc kubenswrapper[4846]: I1005 08:17:25.076955 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="559214a25868d5630117f8a9e0cc4856e13176719e458ad0fe07b0650bb3206c" Oct 05 08:17:25 crc kubenswrapper[4846]: I1005 08:17:25.344377 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-58bc4f8fb5-flnpk"] Oct 05 08:17:25 crc kubenswrapper[4846]: E1005 08:17:25.344830 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e2077da3-902b-48e2-93ef-be196e85b88d" containerName="neutron-db-sync" Oct 05 08:17:25 crc kubenswrapper[4846]: I1005 08:17:25.344846 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="e2077da3-902b-48e2-93ef-be196e85b88d" containerName="neutron-db-sync" Oct 05 08:17:25 crc kubenswrapper[4846]: I1005 08:17:25.345060 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="e2077da3-902b-48e2-93ef-be196e85b88d" containerName="neutron-db-sync" Oct 05 08:17:25 crc kubenswrapper[4846]: I1005 08:17:25.346160 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-58bc4f8fb5-flnpk" Oct 05 08:17:25 crc kubenswrapper[4846]: I1005 08:17:25.379995 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-58bc4f8fb5-flnpk"] Oct 05 08:17:25 crc kubenswrapper[4846]: I1005 08:17:25.436239 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-6bf9574c4-zz2wx"] Oct 05 08:17:25 crc kubenswrapper[4846]: I1005 08:17:25.437672 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-6bf9574c4-zz2wx" Oct 05 08:17:25 crc kubenswrapper[4846]: I1005 08:17:25.449110 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Oct 05 08:17:25 crc kubenswrapper[4846]: I1005 08:17:25.449145 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Oct 05 08:17:25 crc kubenswrapper[4846]: I1005 08:17:25.449317 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-ovndbs" Oct 05 08:17:25 crc kubenswrapper[4846]: I1005 08:17:25.449359 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-nmbv2" Oct 05 08:17:25 crc kubenswrapper[4846]: I1005 08:17:25.452507 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-6bf9574c4-zz2wx"] Oct 05 08:17:25 crc kubenswrapper[4846]: I1005 08:17:25.495126 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nh2g2\" (UniqueName: \"kubernetes.io/projected/e21c80e9-984a-4882-8e73-ae3f272f75dd-kube-api-access-nh2g2\") pod \"dnsmasq-dns-58bc4f8fb5-flnpk\" (UID: \"e21c80e9-984a-4882-8e73-ae3f272f75dd\") " pod="openstack/dnsmasq-dns-58bc4f8fb5-flnpk" Oct 05 08:17:25 crc kubenswrapper[4846]: I1005 08:17:25.495355 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e21c80e9-984a-4882-8e73-ae3f272f75dd-dns-svc\") pod \"dnsmasq-dns-58bc4f8fb5-flnpk\" (UID: \"e21c80e9-984a-4882-8e73-ae3f272f75dd\") " pod="openstack/dnsmasq-dns-58bc4f8fb5-flnpk" Oct 05 08:17:25 crc kubenswrapper[4846]: I1005 08:17:25.495519 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e21c80e9-984a-4882-8e73-ae3f272f75dd-config\") pod \"dnsmasq-dns-58bc4f8fb5-flnpk\" (UID: \"e21c80e9-984a-4882-8e73-ae3f272f75dd\") " pod="openstack/dnsmasq-dns-58bc4f8fb5-flnpk" Oct 05 08:17:25 crc kubenswrapper[4846]: I1005 08:17:25.495597 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e21c80e9-984a-4882-8e73-ae3f272f75dd-ovsdbserver-nb\") pod \"dnsmasq-dns-58bc4f8fb5-flnpk\" (UID: \"e21c80e9-984a-4882-8e73-ae3f272f75dd\") " pod="openstack/dnsmasq-dns-58bc4f8fb5-flnpk" Oct 05 08:17:25 crc kubenswrapper[4846]: I1005 08:17:25.495643 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e21c80e9-984a-4882-8e73-ae3f272f75dd-ovsdbserver-sb\") pod \"dnsmasq-dns-58bc4f8fb5-flnpk\" (UID: \"e21c80e9-984a-4882-8e73-ae3f272f75dd\") " pod="openstack/dnsmasq-dns-58bc4f8fb5-flnpk" Oct 05 08:17:25 crc kubenswrapper[4846]: I1005 08:17:25.597063 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e4a6f807-103b-4390-8d1f-403cbde1ce2a-combined-ca-bundle\") pod \"neutron-6bf9574c4-zz2wx\" (UID: \"e4a6f807-103b-4390-8d1f-403cbde1ce2a\") " pod="openstack/neutron-6bf9574c4-zz2wx" Oct 05 08:17:25 crc kubenswrapper[4846]: I1005 08:17:25.597120 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e21c80e9-984a-4882-8e73-ae3f272f75dd-ovsdbserver-nb\") pod \"dnsmasq-dns-58bc4f8fb5-flnpk\" (UID: \"e21c80e9-984a-4882-8e73-ae3f272f75dd\") " pod="openstack/dnsmasq-dns-58bc4f8fb5-flnpk" Oct 05 08:17:25 crc kubenswrapper[4846]: I1005 08:17:25.597158 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/e4a6f807-103b-4390-8d1f-403cbde1ce2a-config\") pod \"neutron-6bf9574c4-zz2wx\" (UID: \"e4a6f807-103b-4390-8d1f-403cbde1ce2a\") " pod="openstack/neutron-6bf9574c4-zz2wx" Oct 05 08:17:25 crc kubenswrapper[4846]: I1005 08:17:25.597199 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tthcw\" (UniqueName: \"kubernetes.io/projected/e4a6f807-103b-4390-8d1f-403cbde1ce2a-kube-api-access-tthcw\") pod \"neutron-6bf9574c4-zz2wx\" (UID: \"e4a6f807-103b-4390-8d1f-403cbde1ce2a\") " pod="openstack/neutron-6bf9574c4-zz2wx" Oct 05 08:17:25 crc kubenswrapper[4846]: I1005 08:17:25.597223 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e21c80e9-984a-4882-8e73-ae3f272f75dd-ovsdbserver-sb\") pod \"dnsmasq-dns-58bc4f8fb5-flnpk\" (UID: \"e21c80e9-984a-4882-8e73-ae3f272f75dd\") " pod="openstack/dnsmasq-dns-58bc4f8fb5-flnpk" Oct 05 08:17:25 crc kubenswrapper[4846]: I1005 08:17:25.597357 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/e4a6f807-103b-4390-8d1f-403cbde1ce2a-httpd-config\") pod \"neutron-6bf9574c4-zz2wx\" (UID: \"e4a6f807-103b-4390-8d1f-403cbde1ce2a\") " pod="openstack/neutron-6bf9574c4-zz2wx" Oct 05 08:17:25 crc kubenswrapper[4846]: I1005 08:17:25.597429 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nh2g2\" (UniqueName: \"kubernetes.io/projected/e21c80e9-984a-4882-8e73-ae3f272f75dd-kube-api-access-nh2g2\") pod \"dnsmasq-dns-58bc4f8fb5-flnpk\" (UID: \"e21c80e9-984a-4882-8e73-ae3f272f75dd\") " pod="openstack/dnsmasq-dns-58bc4f8fb5-flnpk" Oct 05 08:17:25 crc kubenswrapper[4846]: I1005 08:17:25.597974 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e21c80e9-984a-4882-8e73-ae3f272f75dd-ovsdbserver-nb\") pod \"dnsmasq-dns-58bc4f8fb5-flnpk\" (UID: \"e21c80e9-984a-4882-8e73-ae3f272f75dd\") " pod="openstack/dnsmasq-dns-58bc4f8fb5-flnpk" Oct 05 08:17:25 crc kubenswrapper[4846]: I1005 08:17:25.598012 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e21c80e9-984a-4882-8e73-ae3f272f75dd-ovsdbserver-sb\") pod \"dnsmasq-dns-58bc4f8fb5-flnpk\" (UID: \"e21c80e9-984a-4882-8e73-ae3f272f75dd\") " pod="openstack/dnsmasq-dns-58bc4f8fb5-flnpk" Oct 05 08:17:25 crc kubenswrapper[4846]: I1005 08:17:25.598472 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e21c80e9-984a-4882-8e73-ae3f272f75dd-dns-svc\") pod \"dnsmasq-dns-58bc4f8fb5-flnpk\" (UID: \"e21c80e9-984a-4882-8e73-ae3f272f75dd\") " pod="openstack/dnsmasq-dns-58bc4f8fb5-flnpk" Oct 05 08:17:25 crc kubenswrapper[4846]: I1005 08:17:25.598575 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/e4a6f807-103b-4390-8d1f-403cbde1ce2a-ovndb-tls-certs\") pod \"neutron-6bf9574c4-zz2wx\" (UID: \"e4a6f807-103b-4390-8d1f-403cbde1ce2a\") " pod="openstack/neutron-6bf9574c4-zz2wx" Oct 05 08:17:25 crc kubenswrapper[4846]: I1005 08:17:25.598671 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e21c80e9-984a-4882-8e73-ae3f272f75dd-config\") pod \"dnsmasq-dns-58bc4f8fb5-flnpk\" (UID: \"e21c80e9-984a-4882-8e73-ae3f272f75dd\") " pod="openstack/dnsmasq-dns-58bc4f8fb5-flnpk" Oct 05 08:17:25 crc kubenswrapper[4846]: I1005 08:17:25.599291 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e21c80e9-984a-4882-8e73-ae3f272f75dd-dns-svc\") pod \"dnsmasq-dns-58bc4f8fb5-flnpk\" (UID: \"e21c80e9-984a-4882-8e73-ae3f272f75dd\") " pod="openstack/dnsmasq-dns-58bc4f8fb5-flnpk" Oct 05 08:17:25 crc kubenswrapper[4846]: I1005 08:17:25.599336 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e21c80e9-984a-4882-8e73-ae3f272f75dd-config\") pod \"dnsmasq-dns-58bc4f8fb5-flnpk\" (UID: \"e21c80e9-984a-4882-8e73-ae3f272f75dd\") " pod="openstack/dnsmasq-dns-58bc4f8fb5-flnpk" Oct 05 08:17:25 crc kubenswrapper[4846]: I1005 08:17:25.617263 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nh2g2\" (UniqueName: \"kubernetes.io/projected/e21c80e9-984a-4882-8e73-ae3f272f75dd-kube-api-access-nh2g2\") pod \"dnsmasq-dns-58bc4f8fb5-flnpk\" (UID: \"e21c80e9-984a-4882-8e73-ae3f272f75dd\") " pod="openstack/dnsmasq-dns-58bc4f8fb5-flnpk" Oct 05 08:17:25 crc kubenswrapper[4846]: I1005 08:17:25.700265 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/e4a6f807-103b-4390-8d1f-403cbde1ce2a-ovndb-tls-certs\") pod \"neutron-6bf9574c4-zz2wx\" (UID: \"e4a6f807-103b-4390-8d1f-403cbde1ce2a\") " pod="openstack/neutron-6bf9574c4-zz2wx" Oct 05 08:17:25 crc kubenswrapper[4846]: I1005 08:17:25.700339 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e4a6f807-103b-4390-8d1f-403cbde1ce2a-combined-ca-bundle\") pod \"neutron-6bf9574c4-zz2wx\" (UID: \"e4a6f807-103b-4390-8d1f-403cbde1ce2a\") " pod="openstack/neutron-6bf9574c4-zz2wx" Oct 05 08:17:25 crc kubenswrapper[4846]: I1005 08:17:25.700364 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/e4a6f807-103b-4390-8d1f-403cbde1ce2a-config\") pod \"neutron-6bf9574c4-zz2wx\" (UID: \"e4a6f807-103b-4390-8d1f-403cbde1ce2a\") " pod="openstack/neutron-6bf9574c4-zz2wx" Oct 05 08:17:25 crc kubenswrapper[4846]: I1005 08:17:25.700394 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tthcw\" (UniqueName: \"kubernetes.io/projected/e4a6f807-103b-4390-8d1f-403cbde1ce2a-kube-api-access-tthcw\") pod \"neutron-6bf9574c4-zz2wx\" (UID: \"e4a6f807-103b-4390-8d1f-403cbde1ce2a\") " pod="openstack/neutron-6bf9574c4-zz2wx" Oct 05 08:17:25 crc kubenswrapper[4846]: I1005 08:17:25.700436 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/e4a6f807-103b-4390-8d1f-403cbde1ce2a-httpd-config\") pod \"neutron-6bf9574c4-zz2wx\" (UID: \"e4a6f807-103b-4390-8d1f-403cbde1ce2a\") " pod="openstack/neutron-6bf9574c4-zz2wx" Oct 05 08:17:25 crc kubenswrapper[4846]: I1005 08:17:25.701289 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-58bc4f8fb5-flnpk" Oct 05 08:17:25 crc kubenswrapper[4846]: I1005 08:17:25.704409 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/e4a6f807-103b-4390-8d1f-403cbde1ce2a-config\") pod \"neutron-6bf9574c4-zz2wx\" (UID: \"e4a6f807-103b-4390-8d1f-403cbde1ce2a\") " pod="openstack/neutron-6bf9574c4-zz2wx" Oct 05 08:17:25 crc kubenswrapper[4846]: I1005 08:17:25.704575 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e4a6f807-103b-4390-8d1f-403cbde1ce2a-combined-ca-bundle\") pod \"neutron-6bf9574c4-zz2wx\" (UID: \"e4a6f807-103b-4390-8d1f-403cbde1ce2a\") " pod="openstack/neutron-6bf9574c4-zz2wx" Oct 05 08:17:25 crc kubenswrapper[4846]: I1005 08:17:25.705870 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/e4a6f807-103b-4390-8d1f-403cbde1ce2a-httpd-config\") pod \"neutron-6bf9574c4-zz2wx\" (UID: \"e4a6f807-103b-4390-8d1f-403cbde1ce2a\") " pod="openstack/neutron-6bf9574c4-zz2wx" Oct 05 08:17:25 crc kubenswrapper[4846]: I1005 08:17:25.707723 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/e4a6f807-103b-4390-8d1f-403cbde1ce2a-ovndb-tls-certs\") pod \"neutron-6bf9574c4-zz2wx\" (UID: \"e4a6f807-103b-4390-8d1f-403cbde1ce2a\") " pod="openstack/neutron-6bf9574c4-zz2wx" Oct 05 08:17:25 crc kubenswrapper[4846]: I1005 08:17:25.725848 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tthcw\" (UniqueName: \"kubernetes.io/projected/e4a6f807-103b-4390-8d1f-403cbde1ce2a-kube-api-access-tthcw\") pod \"neutron-6bf9574c4-zz2wx\" (UID: \"e4a6f807-103b-4390-8d1f-403cbde1ce2a\") " pod="openstack/neutron-6bf9574c4-zz2wx" Oct 05 08:17:25 crc kubenswrapper[4846]: I1005 08:17:25.761046 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-6bf9574c4-zz2wx" Oct 05 08:17:26 crc kubenswrapper[4846]: I1005 08:17:26.206418 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-58bc4f8fb5-flnpk"] Oct 05 08:17:26 crc kubenswrapper[4846]: W1005 08:17:26.223962 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode21c80e9_984a_4882_8e73_ae3f272f75dd.slice/crio-1595a606a41a1cf9bca5a0ff79544a45146e0525bb5ff708691877db8ef40cec WatchSource:0}: Error finding container 1595a606a41a1cf9bca5a0ff79544a45146e0525bb5ff708691877db8ef40cec: Status 404 returned error can't find the container with id 1595a606a41a1cf9bca5a0ff79544a45146e0525bb5ff708691877db8ef40cec Oct 05 08:17:26 crc kubenswrapper[4846]: I1005 08:17:26.396895 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-6bf9574c4-zz2wx"] Oct 05 08:17:26 crc kubenswrapper[4846]: W1005 08:17:26.398119 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode4a6f807_103b_4390_8d1f_403cbde1ce2a.slice/crio-1309afb0b88458efcb5ce8b337e263ef0542c1a9c2d01521ef3bf1123705830a WatchSource:0}: Error finding container 1309afb0b88458efcb5ce8b337e263ef0542c1a9c2d01521ef3bf1123705830a: Status 404 returned error can't find the container with id 1309afb0b88458efcb5ce8b337e263ef0542c1a9c2d01521ef3bf1123705830a Oct 05 08:17:27 crc kubenswrapper[4846]: I1005 08:17:27.101343 4846 generic.go:334] "Generic (PLEG): container finished" podID="e21c80e9-984a-4882-8e73-ae3f272f75dd" containerID="7e28abfd4128d4c7447893a5f05c5c758471f760f57f749ffff8a4ce0dbf6cf8" exitCode=0 Oct 05 08:17:27 crc kubenswrapper[4846]: I1005 08:17:27.101482 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-58bc4f8fb5-flnpk" event={"ID":"e21c80e9-984a-4882-8e73-ae3f272f75dd","Type":"ContainerDied","Data":"7e28abfd4128d4c7447893a5f05c5c758471f760f57f749ffff8a4ce0dbf6cf8"} Oct 05 08:17:27 crc kubenswrapper[4846]: I1005 08:17:27.101713 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-58bc4f8fb5-flnpk" event={"ID":"e21c80e9-984a-4882-8e73-ae3f272f75dd","Type":"ContainerStarted","Data":"1595a606a41a1cf9bca5a0ff79544a45146e0525bb5ff708691877db8ef40cec"} Oct 05 08:17:27 crc kubenswrapper[4846]: I1005 08:17:27.103761 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6bf9574c4-zz2wx" event={"ID":"e4a6f807-103b-4390-8d1f-403cbde1ce2a","Type":"ContainerStarted","Data":"9e4655e9e5c45ff6cd334ce0fef1df22513c2a5ad10159833f9e711efd01236d"} Oct 05 08:17:27 crc kubenswrapper[4846]: I1005 08:17:27.103804 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6bf9574c4-zz2wx" event={"ID":"e4a6f807-103b-4390-8d1f-403cbde1ce2a","Type":"ContainerStarted","Data":"4f8c70367acf4b7eba0bb129d23bdadb9e35eef249c09fba85346ce1a99f7a26"} Oct 05 08:17:27 crc kubenswrapper[4846]: I1005 08:17:27.103814 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6bf9574c4-zz2wx" event={"ID":"e4a6f807-103b-4390-8d1f-403cbde1ce2a","Type":"ContainerStarted","Data":"1309afb0b88458efcb5ce8b337e263ef0542c1a9c2d01521ef3bf1123705830a"} Oct 05 08:17:27 crc kubenswrapper[4846]: I1005 08:17:27.104580 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-6bf9574c4-zz2wx" Oct 05 08:17:27 crc kubenswrapper[4846]: I1005 08:17:27.157604 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-6bf9574c4-zz2wx" podStartSLOduration=2.157581642 podStartE2EDuration="2.157581642s" podCreationTimestamp="2025-10-05 08:17:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 08:17:27.145805727 +0000 UTC m=+5369.386658532" watchObservedRunningTime="2025-10-05 08:17:27.157581642 +0000 UTC m=+5369.398434427" Oct 05 08:17:27 crc kubenswrapper[4846]: I1005 08:17:27.578372 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-fdc58b465-qpttk"] Oct 05 08:17:27 crc kubenswrapper[4846]: I1005 08:17:27.581279 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-fdc58b465-qpttk" Oct 05 08:17:27 crc kubenswrapper[4846]: I1005 08:17:27.584395 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-public-svc" Oct 05 08:17:27 crc kubenswrapper[4846]: I1005 08:17:27.584709 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-internal-svc" Oct 05 08:17:27 crc kubenswrapper[4846]: I1005 08:17:27.603394 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-fdc58b465-qpttk"] Oct 05 08:17:27 crc kubenswrapper[4846]: I1005 08:17:27.745510 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e8c40de6-3f63-4f56-a480-0fbf007507a5-combined-ca-bundle\") pod \"neutron-fdc58b465-qpttk\" (UID: \"e8c40de6-3f63-4f56-a480-0fbf007507a5\") " pod="openstack/neutron-fdc58b465-qpttk" Oct 05 08:17:27 crc kubenswrapper[4846]: I1005 08:17:27.745725 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e8c40de6-3f63-4f56-a480-0fbf007507a5-public-tls-certs\") pod \"neutron-fdc58b465-qpttk\" (UID: \"e8c40de6-3f63-4f56-a480-0fbf007507a5\") " pod="openstack/neutron-fdc58b465-qpttk" Oct 05 08:17:27 crc kubenswrapper[4846]: I1005 08:17:27.745843 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e8c40de6-3f63-4f56-a480-0fbf007507a5-internal-tls-certs\") pod \"neutron-fdc58b465-qpttk\" (UID: \"e8c40de6-3f63-4f56-a480-0fbf007507a5\") " pod="openstack/neutron-fdc58b465-qpttk" Oct 05 08:17:27 crc kubenswrapper[4846]: I1005 08:17:27.745924 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/e8c40de6-3f63-4f56-a480-0fbf007507a5-ovndb-tls-certs\") pod \"neutron-fdc58b465-qpttk\" (UID: \"e8c40de6-3f63-4f56-a480-0fbf007507a5\") " pod="openstack/neutron-fdc58b465-qpttk" Oct 05 08:17:27 crc kubenswrapper[4846]: I1005 08:17:27.746044 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/e8c40de6-3f63-4f56-a480-0fbf007507a5-config\") pod \"neutron-fdc58b465-qpttk\" (UID: \"e8c40de6-3f63-4f56-a480-0fbf007507a5\") " pod="openstack/neutron-fdc58b465-qpttk" Oct 05 08:17:27 crc kubenswrapper[4846]: I1005 08:17:27.746143 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/e8c40de6-3f63-4f56-a480-0fbf007507a5-httpd-config\") pod \"neutron-fdc58b465-qpttk\" (UID: \"e8c40de6-3f63-4f56-a480-0fbf007507a5\") " pod="openstack/neutron-fdc58b465-qpttk" Oct 05 08:17:27 crc kubenswrapper[4846]: I1005 08:17:27.746286 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2svq9\" (UniqueName: \"kubernetes.io/projected/e8c40de6-3f63-4f56-a480-0fbf007507a5-kube-api-access-2svq9\") pod \"neutron-fdc58b465-qpttk\" (UID: \"e8c40de6-3f63-4f56-a480-0fbf007507a5\") " pod="openstack/neutron-fdc58b465-qpttk" Oct 05 08:17:27 crc kubenswrapper[4846]: I1005 08:17:27.848062 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e8c40de6-3f63-4f56-a480-0fbf007507a5-combined-ca-bundle\") pod \"neutron-fdc58b465-qpttk\" (UID: \"e8c40de6-3f63-4f56-a480-0fbf007507a5\") " pod="openstack/neutron-fdc58b465-qpttk" Oct 05 08:17:27 crc kubenswrapper[4846]: I1005 08:17:27.848139 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e8c40de6-3f63-4f56-a480-0fbf007507a5-public-tls-certs\") pod \"neutron-fdc58b465-qpttk\" (UID: \"e8c40de6-3f63-4f56-a480-0fbf007507a5\") " pod="openstack/neutron-fdc58b465-qpttk" Oct 05 08:17:27 crc kubenswrapper[4846]: I1005 08:17:27.848175 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e8c40de6-3f63-4f56-a480-0fbf007507a5-internal-tls-certs\") pod \"neutron-fdc58b465-qpttk\" (UID: \"e8c40de6-3f63-4f56-a480-0fbf007507a5\") " pod="openstack/neutron-fdc58b465-qpttk" Oct 05 08:17:27 crc kubenswrapper[4846]: I1005 08:17:27.848204 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/e8c40de6-3f63-4f56-a480-0fbf007507a5-ovndb-tls-certs\") pod \"neutron-fdc58b465-qpttk\" (UID: \"e8c40de6-3f63-4f56-a480-0fbf007507a5\") " pod="openstack/neutron-fdc58b465-qpttk" Oct 05 08:17:27 crc kubenswrapper[4846]: I1005 08:17:27.848310 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/e8c40de6-3f63-4f56-a480-0fbf007507a5-config\") pod \"neutron-fdc58b465-qpttk\" (UID: \"e8c40de6-3f63-4f56-a480-0fbf007507a5\") " pod="openstack/neutron-fdc58b465-qpttk" Oct 05 08:17:27 crc kubenswrapper[4846]: I1005 08:17:27.848348 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/e8c40de6-3f63-4f56-a480-0fbf007507a5-httpd-config\") pod \"neutron-fdc58b465-qpttk\" (UID: \"e8c40de6-3f63-4f56-a480-0fbf007507a5\") " pod="openstack/neutron-fdc58b465-qpttk" Oct 05 08:17:27 crc kubenswrapper[4846]: I1005 08:17:27.848396 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2svq9\" (UniqueName: \"kubernetes.io/projected/e8c40de6-3f63-4f56-a480-0fbf007507a5-kube-api-access-2svq9\") pod \"neutron-fdc58b465-qpttk\" (UID: \"e8c40de6-3f63-4f56-a480-0fbf007507a5\") " pod="openstack/neutron-fdc58b465-qpttk" Oct 05 08:17:27 crc kubenswrapper[4846]: I1005 08:17:27.853478 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e8c40de6-3f63-4f56-a480-0fbf007507a5-public-tls-certs\") pod \"neutron-fdc58b465-qpttk\" (UID: \"e8c40de6-3f63-4f56-a480-0fbf007507a5\") " pod="openstack/neutron-fdc58b465-qpttk" Oct 05 08:17:27 crc kubenswrapper[4846]: I1005 08:17:27.853478 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/e8c40de6-3f63-4f56-a480-0fbf007507a5-ovndb-tls-certs\") pod \"neutron-fdc58b465-qpttk\" (UID: \"e8c40de6-3f63-4f56-a480-0fbf007507a5\") " pod="openstack/neutron-fdc58b465-qpttk" Oct 05 08:17:27 crc kubenswrapper[4846]: I1005 08:17:27.853743 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e8c40de6-3f63-4f56-a480-0fbf007507a5-combined-ca-bundle\") pod \"neutron-fdc58b465-qpttk\" (UID: \"e8c40de6-3f63-4f56-a480-0fbf007507a5\") " pod="openstack/neutron-fdc58b465-qpttk" Oct 05 08:17:27 crc kubenswrapper[4846]: I1005 08:17:27.855925 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/e8c40de6-3f63-4f56-a480-0fbf007507a5-config\") pod \"neutron-fdc58b465-qpttk\" (UID: \"e8c40de6-3f63-4f56-a480-0fbf007507a5\") " pod="openstack/neutron-fdc58b465-qpttk" Oct 05 08:17:27 crc kubenswrapper[4846]: I1005 08:17:27.858948 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/e8c40de6-3f63-4f56-a480-0fbf007507a5-httpd-config\") pod \"neutron-fdc58b465-qpttk\" (UID: \"e8c40de6-3f63-4f56-a480-0fbf007507a5\") " pod="openstack/neutron-fdc58b465-qpttk" Oct 05 08:17:27 crc kubenswrapper[4846]: I1005 08:17:27.873684 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2svq9\" (UniqueName: \"kubernetes.io/projected/e8c40de6-3f63-4f56-a480-0fbf007507a5-kube-api-access-2svq9\") pod \"neutron-fdc58b465-qpttk\" (UID: \"e8c40de6-3f63-4f56-a480-0fbf007507a5\") " pod="openstack/neutron-fdc58b465-qpttk" Oct 05 08:17:27 crc kubenswrapper[4846]: I1005 08:17:27.875964 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e8c40de6-3f63-4f56-a480-0fbf007507a5-internal-tls-certs\") pod \"neutron-fdc58b465-qpttk\" (UID: \"e8c40de6-3f63-4f56-a480-0fbf007507a5\") " pod="openstack/neutron-fdc58b465-qpttk" Oct 05 08:17:27 crc kubenswrapper[4846]: I1005 08:17:27.902595 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-fdc58b465-qpttk" Oct 05 08:17:28 crc kubenswrapper[4846]: I1005 08:17:28.112928 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-58bc4f8fb5-flnpk" event={"ID":"e21c80e9-984a-4882-8e73-ae3f272f75dd","Type":"ContainerStarted","Data":"d785b2914ea934843acf56f4a3fd3d68e13dac02e609ece8c2104a5ac4ae9870"} Oct 05 08:17:28 crc kubenswrapper[4846]: I1005 08:17:28.113410 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-58bc4f8fb5-flnpk" Oct 05 08:17:28 crc kubenswrapper[4846]: I1005 08:17:28.132230 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-58bc4f8fb5-flnpk" podStartSLOduration=3.132213598 podStartE2EDuration="3.132213598s" podCreationTimestamp="2025-10-05 08:17:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 08:17:28.128354254 +0000 UTC m=+5370.369207029" watchObservedRunningTime="2025-10-05 08:17:28.132213598 +0000 UTC m=+5370.373066373" Oct 05 08:17:28 crc kubenswrapper[4846]: I1005 08:17:28.437870 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-fdc58b465-qpttk"] Oct 05 08:17:28 crc kubenswrapper[4846]: W1005 08:17:28.452569 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode8c40de6_3f63_4f56_a480_0fbf007507a5.slice/crio-45de6a3cadaa9275c5b1720edb9a473bca345dca87f26cf14eb2d3a9f4ff8353 WatchSource:0}: Error finding container 45de6a3cadaa9275c5b1720edb9a473bca345dca87f26cf14eb2d3a9f4ff8353: Status 404 returned error can't find the container with id 45de6a3cadaa9275c5b1720edb9a473bca345dca87f26cf14eb2d3a9f4ff8353 Oct 05 08:17:28 crc kubenswrapper[4846]: I1005 08:17:28.741408 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-mlfl2" Oct 05 08:17:28 crc kubenswrapper[4846]: I1005 08:17:28.742170 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-mlfl2" Oct 05 08:17:28 crc kubenswrapper[4846]: I1005 08:17:28.802875 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-mlfl2" Oct 05 08:17:29 crc kubenswrapper[4846]: I1005 08:17:29.124424 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-fdc58b465-qpttk" event={"ID":"e8c40de6-3f63-4f56-a480-0fbf007507a5","Type":"ContainerStarted","Data":"5657c1f0f5a9306e974a3c3fda32aea8c56b4341ffb14091f248a1c55d60b694"} Oct 05 08:17:29 crc kubenswrapper[4846]: I1005 08:17:29.124482 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-fdc58b465-qpttk" event={"ID":"e8c40de6-3f63-4f56-a480-0fbf007507a5","Type":"ContainerStarted","Data":"8dcdaccb88fd4d03b104625355707d8930257885c3f19d90419b27071dd99a52"} Oct 05 08:17:29 crc kubenswrapper[4846]: I1005 08:17:29.124499 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-fdc58b465-qpttk" event={"ID":"e8c40de6-3f63-4f56-a480-0fbf007507a5","Type":"ContainerStarted","Data":"45de6a3cadaa9275c5b1720edb9a473bca345dca87f26cf14eb2d3a9f4ff8353"} Oct 05 08:17:29 crc kubenswrapper[4846]: I1005 08:17:29.144926 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-fdc58b465-qpttk" podStartSLOduration=2.144907973 podStartE2EDuration="2.144907973s" podCreationTimestamp="2025-10-05 08:17:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 08:17:29.143722562 +0000 UTC m=+5371.384575337" watchObservedRunningTime="2025-10-05 08:17:29.144907973 +0000 UTC m=+5371.385760748" Oct 05 08:17:29 crc kubenswrapper[4846]: I1005 08:17:29.186630 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-mlfl2" Oct 05 08:17:30 crc kubenswrapper[4846]: I1005 08:17:30.147509 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-fdc58b465-qpttk" Oct 05 08:17:31 crc kubenswrapper[4846]: I1005 08:17:31.755699 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-mlfl2"] Oct 05 08:17:32 crc kubenswrapper[4846]: I1005 08:17:32.165436 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-mlfl2" podUID="bcfb4bd3-9712-466d-bc48-317857cdd151" containerName="registry-server" containerID="cri-o://df1bee18b97987a12e04c5fb5710b5c962a880e700a8dd484232762e157c4d1c" gracePeriod=2 Oct 05 08:17:32 crc kubenswrapper[4846]: I1005 08:17:32.670917 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mlfl2" Oct 05 08:17:32 crc kubenswrapper[4846]: I1005 08:17:32.744281 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mh5ck\" (UniqueName: \"kubernetes.io/projected/bcfb4bd3-9712-466d-bc48-317857cdd151-kube-api-access-mh5ck\") pod \"bcfb4bd3-9712-466d-bc48-317857cdd151\" (UID: \"bcfb4bd3-9712-466d-bc48-317857cdd151\") " Oct 05 08:17:32 crc kubenswrapper[4846]: I1005 08:17:32.744358 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bcfb4bd3-9712-466d-bc48-317857cdd151-utilities\") pod \"bcfb4bd3-9712-466d-bc48-317857cdd151\" (UID: \"bcfb4bd3-9712-466d-bc48-317857cdd151\") " Oct 05 08:17:32 crc kubenswrapper[4846]: I1005 08:17:32.744425 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bcfb4bd3-9712-466d-bc48-317857cdd151-catalog-content\") pod \"bcfb4bd3-9712-466d-bc48-317857cdd151\" (UID: \"bcfb4bd3-9712-466d-bc48-317857cdd151\") " Oct 05 08:17:32 crc kubenswrapper[4846]: I1005 08:17:32.754211 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bcfb4bd3-9712-466d-bc48-317857cdd151-utilities" (OuterVolumeSpecName: "utilities") pod "bcfb4bd3-9712-466d-bc48-317857cdd151" (UID: "bcfb4bd3-9712-466d-bc48-317857cdd151"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 08:17:32 crc kubenswrapper[4846]: I1005 08:17:32.758952 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bcfb4bd3-9712-466d-bc48-317857cdd151-kube-api-access-mh5ck" (OuterVolumeSpecName: "kube-api-access-mh5ck") pod "bcfb4bd3-9712-466d-bc48-317857cdd151" (UID: "bcfb4bd3-9712-466d-bc48-317857cdd151"). InnerVolumeSpecName "kube-api-access-mh5ck". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:17:32 crc kubenswrapper[4846]: I1005 08:17:32.759133 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bcfb4bd3-9712-466d-bc48-317857cdd151-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "bcfb4bd3-9712-466d-bc48-317857cdd151" (UID: "bcfb4bd3-9712-466d-bc48-317857cdd151"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 08:17:32 crc kubenswrapper[4846]: I1005 08:17:32.846734 4846 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bcfb4bd3-9712-466d-bc48-317857cdd151-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 05 08:17:32 crc kubenswrapper[4846]: I1005 08:17:32.846958 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mh5ck\" (UniqueName: \"kubernetes.io/projected/bcfb4bd3-9712-466d-bc48-317857cdd151-kube-api-access-mh5ck\") on node \"crc\" DevicePath \"\"" Oct 05 08:17:32 crc kubenswrapper[4846]: I1005 08:17:32.846970 4846 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bcfb4bd3-9712-466d-bc48-317857cdd151-utilities\") on node \"crc\" DevicePath \"\"" Oct 05 08:17:33 crc kubenswrapper[4846]: I1005 08:17:33.179907 4846 generic.go:334] "Generic (PLEG): container finished" podID="bcfb4bd3-9712-466d-bc48-317857cdd151" containerID="df1bee18b97987a12e04c5fb5710b5c962a880e700a8dd484232762e157c4d1c" exitCode=0 Oct 05 08:17:33 crc kubenswrapper[4846]: I1005 08:17:33.179959 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mlfl2" event={"ID":"bcfb4bd3-9712-466d-bc48-317857cdd151","Type":"ContainerDied","Data":"df1bee18b97987a12e04c5fb5710b5c962a880e700a8dd484232762e157c4d1c"} Oct 05 08:17:33 crc kubenswrapper[4846]: I1005 08:17:33.180004 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mlfl2" event={"ID":"bcfb4bd3-9712-466d-bc48-317857cdd151","Type":"ContainerDied","Data":"c46046c40cb1c5230a9c2fd1bd8c484bf1f620bc8d16601261722529f20296cd"} Oct 05 08:17:33 crc kubenswrapper[4846]: I1005 08:17:33.180024 4846 scope.go:117] "RemoveContainer" containerID="df1bee18b97987a12e04c5fb5710b5c962a880e700a8dd484232762e157c4d1c" Oct 05 08:17:33 crc kubenswrapper[4846]: I1005 08:17:33.180021 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mlfl2" Oct 05 08:17:33 crc kubenswrapper[4846]: I1005 08:17:33.201638 4846 scope.go:117] "RemoveContainer" containerID="b7760653459fee5ad898728947509c6903e698373c8cf35e05416615b2b08cdb" Oct 05 08:17:33 crc kubenswrapper[4846]: I1005 08:17:33.219711 4846 scope.go:117] "RemoveContainer" containerID="8250648ccf9655e57ac2a743dfbaf5aa20a58bb6e6eb57f404fc31adad106b2d" Oct 05 08:17:33 crc kubenswrapper[4846]: I1005 08:17:33.286466 4846 scope.go:117] "RemoveContainer" containerID="df1bee18b97987a12e04c5fb5710b5c962a880e700a8dd484232762e157c4d1c" Oct 05 08:17:33 crc kubenswrapper[4846]: E1005 08:17:33.286863 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"df1bee18b97987a12e04c5fb5710b5c962a880e700a8dd484232762e157c4d1c\": container with ID starting with df1bee18b97987a12e04c5fb5710b5c962a880e700a8dd484232762e157c4d1c not found: ID does not exist" containerID="df1bee18b97987a12e04c5fb5710b5c962a880e700a8dd484232762e157c4d1c" Oct 05 08:17:33 crc kubenswrapper[4846]: I1005 08:17:33.286892 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"df1bee18b97987a12e04c5fb5710b5c962a880e700a8dd484232762e157c4d1c"} err="failed to get container status \"df1bee18b97987a12e04c5fb5710b5c962a880e700a8dd484232762e157c4d1c\": rpc error: code = NotFound desc = could not find container \"df1bee18b97987a12e04c5fb5710b5c962a880e700a8dd484232762e157c4d1c\": container with ID starting with df1bee18b97987a12e04c5fb5710b5c962a880e700a8dd484232762e157c4d1c not found: ID does not exist" Oct 05 08:17:33 crc kubenswrapper[4846]: I1005 08:17:33.286917 4846 scope.go:117] "RemoveContainer" containerID="b7760653459fee5ad898728947509c6903e698373c8cf35e05416615b2b08cdb" Oct 05 08:17:33 crc kubenswrapper[4846]: E1005 08:17:33.287157 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b7760653459fee5ad898728947509c6903e698373c8cf35e05416615b2b08cdb\": container with ID starting with b7760653459fee5ad898728947509c6903e698373c8cf35e05416615b2b08cdb not found: ID does not exist" containerID="b7760653459fee5ad898728947509c6903e698373c8cf35e05416615b2b08cdb" Oct 05 08:17:33 crc kubenswrapper[4846]: I1005 08:17:33.287178 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b7760653459fee5ad898728947509c6903e698373c8cf35e05416615b2b08cdb"} err="failed to get container status \"b7760653459fee5ad898728947509c6903e698373c8cf35e05416615b2b08cdb\": rpc error: code = NotFound desc = could not find container \"b7760653459fee5ad898728947509c6903e698373c8cf35e05416615b2b08cdb\": container with ID starting with b7760653459fee5ad898728947509c6903e698373c8cf35e05416615b2b08cdb not found: ID does not exist" Oct 05 08:17:33 crc kubenswrapper[4846]: I1005 08:17:33.287209 4846 scope.go:117] "RemoveContainer" containerID="8250648ccf9655e57ac2a743dfbaf5aa20a58bb6e6eb57f404fc31adad106b2d" Oct 05 08:17:33 crc kubenswrapper[4846]: E1005 08:17:33.287573 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8250648ccf9655e57ac2a743dfbaf5aa20a58bb6e6eb57f404fc31adad106b2d\": container with ID starting with 8250648ccf9655e57ac2a743dfbaf5aa20a58bb6e6eb57f404fc31adad106b2d not found: ID does not exist" containerID="8250648ccf9655e57ac2a743dfbaf5aa20a58bb6e6eb57f404fc31adad106b2d" Oct 05 08:17:33 crc kubenswrapper[4846]: I1005 08:17:33.287594 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8250648ccf9655e57ac2a743dfbaf5aa20a58bb6e6eb57f404fc31adad106b2d"} err="failed to get container status \"8250648ccf9655e57ac2a743dfbaf5aa20a58bb6e6eb57f404fc31adad106b2d\": rpc error: code = NotFound desc = could not find container \"8250648ccf9655e57ac2a743dfbaf5aa20a58bb6e6eb57f404fc31adad106b2d\": container with ID starting with 8250648ccf9655e57ac2a743dfbaf5aa20a58bb6e6eb57f404fc31adad106b2d not found: ID does not exist" Oct 05 08:17:33 crc kubenswrapper[4846]: I1005 08:17:33.291289 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-mlfl2"] Oct 05 08:17:33 crc kubenswrapper[4846]: I1005 08:17:33.297631 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-mlfl2"] Oct 05 08:17:34 crc kubenswrapper[4846]: I1005 08:17:34.525330 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bcfb4bd3-9712-466d-bc48-317857cdd151" path="/var/lib/kubelet/pods/bcfb4bd3-9712-466d-bc48-317857cdd151/volumes" Oct 05 08:17:35 crc kubenswrapper[4846]: I1005 08:17:35.703374 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-58bc4f8fb5-flnpk" Oct 05 08:17:35 crc kubenswrapper[4846]: I1005 08:17:35.784736 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-c57df7b7c-w5pmp"] Oct 05 08:17:35 crc kubenswrapper[4846]: I1005 08:17:35.785636 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-c57df7b7c-w5pmp" podUID="9d8932cd-6871-4b34-90cd-3945e6ce6956" containerName="dnsmasq-dns" containerID="cri-o://7832c0fde700ceb61df332f26acd0cb00cb3c3565d3185c78b3864e1dabf2b88" gracePeriod=10 Oct 05 08:17:36 crc kubenswrapper[4846]: I1005 08:17:36.210525 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-c57df7b7c-w5pmp" event={"ID":"9d8932cd-6871-4b34-90cd-3945e6ce6956","Type":"ContainerDied","Data":"7832c0fde700ceb61df332f26acd0cb00cb3c3565d3185c78b3864e1dabf2b88"} Oct 05 08:17:36 crc kubenswrapper[4846]: I1005 08:17:36.210336 4846 generic.go:334] "Generic (PLEG): container finished" podID="9d8932cd-6871-4b34-90cd-3945e6ce6956" containerID="7832c0fde700ceb61df332f26acd0cb00cb3c3565d3185c78b3864e1dabf2b88" exitCode=0 Oct 05 08:17:36 crc kubenswrapper[4846]: I1005 08:17:36.211455 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-c57df7b7c-w5pmp" event={"ID":"9d8932cd-6871-4b34-90cd-3945e6ce6956","Type":"ContainerDied","Data":"7a710bfec083ef531a8af0716eb93fe0c35022644927d90eac75a2b405650ecf"} Oct 05 08:17:36 crc kubenswrapper[4846]: I1005 08:17:36.211487 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7a710bfec083ef531a8af0716eb93fe0c35022644927d90eac75a2b405650ecf" Oct 05 08:17:36 crc kubenswrapper[4846]: I1005 08:17:36.276939 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-c57df7b7c-w5pmp" Oct 05 08:17:36 crc kubenswrapper[4846]: I1005 08:17:36.305361 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9d8932cd-6871-4b34-90cd-3945e6ce6956-dns-svc\") pod \"9d8932cd-6871-4b34-90cd-3945e6ce6956\" (UID: \"9d8932cd-6871-4b34-90cd-3945e6ce6956\") " Oct 05 08:17:36 crc kubenswrapper[4846]: I1005 08:17:36.305449 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9d8932cd-6871-4b34-90cd-3945e6ce6956-ovsdbserver-nb\") pod \"9d8932cd-6871-4b34-90cd-3945e6ce6956\" (UID: \"9d8932cd-6871-4b34-90cd-3945e6ce6956\") " Oct 05 08:17:36 crc kubenswrapper[4846]: I1005 08:17:36.305493 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2mb9d\" (UniqueName: \"kubernetes.io/projected/9d8932cd-6871-4b34-90cd-3945e6ce6956-kube-api-access-2mb9d\") pod \"9d8932cd-6871-4b34-90cd-3945e6ce6956\" (UID: \"9d8932cd-6871-4b34-90cd-3945e6ce6956\") " Oct 05 08:17:36 crc kubenswrapper[4846]: I1005 08:17:36.305518 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9d8932cd-6871-4b34-90cd-3945e6ce6956-ovsdbserver-sb\") pod \"9d8932cd-6871-4b34-90cd-3945e6ce6956\" (UID: \"9d8932cd-6871-4b34-90cd-3945e6ce6956\") " Oct 05 08:17:36 crc kubenswrapper[4846]: I1005 08:17:36.305576 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d8932cd-6871-4b34-90cd-3945e6ce6956-config\") pod \"9d8932cd-6871-4b34-90cd-3945e6ce6956\" (UID: \"9d8932cd-6871-4b34-90cd-3945e6ce6956\") " Oct 05 08:17:36 crc kubenswrapper[4846]: I1005 08:17:36.310748 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d8932cd-6871-4b34-90cd-3945e6ce6956-kube-api-access-2mb9d" (OuterVolumeSpecName: "kube-api-access-2mb9d") pod "9d8932cd-6871-4b34-90cd-3945e6ce6956" (UID: "9d8932cd-6871-4b34-90cd-3945e6ce6956"). InnerVolumeSpecName "kube-api-access-2mb9d". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:17:36 crc kubenswrapper[4846]: I1005 08:17:36.360743 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d8932cd-6871-4b34-90cd-3945e6ce6956-config" (OuterVolumeSpecName: "config") pod "9d8932cd-6871-4b34-90cd-3945e6ce6956" (UID: "9d8932cd-6871-4b34-90cd-3945e6ce6956"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 08:17:36 crc kubenswrapper[4846]: I1005 08:17:36.371543 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d8932cd-6871-4b34-90cd-3945e6ce6956-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "9d8932cd-6871-4b34-90cd-3945e6ce6956" (UID: "9d8932cd-6871-4b34-90cd-3945e6ce6956"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 08:17:36 crc kubenswrapper[4846]: I1005 08:17:36.384519 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d8932cd-6871-4b34-90cd-3945e6ce6956-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "9d8932cd-6871-4b34-90cd-3945e6ce6956" (UID: "9d8932cd-6871-4b34-90cd-3945e6ce6956"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 08:17:36 crc kubenswrapper[4846]: I1005 08:17:36.386745 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d8932cd-6871-4b34-90cd-3945e6ce6956-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "9d8932cd-6871-4b34-90cd-3945e6ce6956" (UID: "9d8932cd-6871-4b34-90cd-3945e6ce6956"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 08:17:36 crc kubenswrapper[4846]: I1005 08:17:36.407067 4846 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9d8932cd-6871-4b34-90cd-3945e6ce6956-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 05 08:17:36 crc kubenswrapper[4846]: I1005 08:17:36.407097 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2mb9d\" (UniqueName: \"kubernetes.io/projected/9d8932cd-6871-4b34-90cd-3945e6ce6956-kube-api-access-2mb9d\") on node \"crc\" DevicePath \"\"" Oct 05 08:17:36 crc kubenswrapper[4846]: I1005 08:17:36.407110 4846 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9d8932cd-6871-4b34-90cd-3945e6ce6956-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 05 08:17:36 crc kubenswrapper[4846]: I1005 08:17:36.407121 4846 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d8932cd-6871-4b34-90cd-3945e6ce6956-config\") on node \"crc\" DevicePath \"\"" Oct 05 08:17:36 crc kubenswrapper[4846]: I1005 08:17:36.407130 4846 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9d8932cd-6871-4b34-90cd-3945e6ce6956-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 05 08:17:37 crc kubenswrapper[4846]: I1005 08:17:37.222893 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-c57df7b7c-w5pmp" Oct 05 08:17:37 crc kubenswrapper[4846]: I1005 08:17:37.257841 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-c57df7b7c-w5pmp"] Oct 05 08:17:37 crc kubenswrapper[4846]: I1005 08:17:37.272768 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-c57df7b7c-w5pmp"] Oct 05 08:17:38 crc kubenswrapper[4846]: I1005 08:17:38.527278 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d8932cd-6871-4b34-90cd-3945e6ce6956" path="/var/lib/kubelet/pods/9d8932cd-6871-4b34-90cd-3945e6ce6956/volumes" Oct 05 08:17:53 crc kubenswrapper[4846]: I1005 08:17:53.324776 4846 patch_prober.go:28] interesting pod/machine-config-daemon-fscvf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 08:17:53 crc kubenswrapper[4846]: I1005 08:17:53.325492 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 08:17:55 crc kubenswrapper[4846]: I1005 08:17:55.775289 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-6bf9574c4-zz2wx" Oct 05 08:17:57 crc kubenswrapper[4846]: I1005 08:17:57.924863 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-fdc58b465-qpttk" Oct 05 08:17:58 crc kubenswrapper[4846]: I1005 08:17:58.006538 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-6bf9574c4-zz2wx"] Oct 05 08:17:58 crc kubenswrapper[4846]: I1005 08:17:58.006795 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-6bf9574c4-zz2wx" podUID="e4a6f807-103b-4390-8d1f-403cbde1ce2a" containerName="neutron-api" containerID="cri-o://4f8c70367acf4b7eba0bb129d23bdadb9e35eef249c09fba85346ce1a99f7a26" gracePeriod=30 Oct 05 08:17:58 crc kubenswrapper[4846]: I1005 08:17:58.007204 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-6bf9574c4-zz2wx" podUID="e4a6f807-103b-4390-8d1f-403cbde1ce2a" containerName="neutron-httpd" containerID="cri-o://9e4655e9e5c45ff6cd334ce0fef1df22513c2a5ad10159833f9e711efd01236d" gracePeriod=30 Oct 05 08:17:58 crc kubenswrapper[4846]: I1005 08:17:58.448939 4846 generic.go:334] "Generic (PLEG): container finished" podID="e4a6f807-103b-4390-8d1f-403cbde1ce2a" containerID="9e4655e9e5c45ff6cd334ce0fef1df22513c2a5ad10159833f9e711efd01236d" exitCode=0 Oct 05 08:17:58 crc kubenswrapper[4846]: I1005 08:17:58.449027 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6bf9574c4-zz2wx" event={"ID":"e4a6f807-103b-4390-8d1f-403cbde1ce2a","Type":"ContainerDied","Data":"9e4655e9e5c45ff6cd334ce0fef1df22513c2a5ad10159833f9e711efd01236d"} Oct 05 08:18:01 crc kubenswrapper[4846]: I1005 08:18:01.500684 4846 generic.go:334] "Generic (PLEG): container finished" podID="e4a6f807-103b-4390-8d1f-403cbde1ce2a" containerID="4f8c70367acf4b7eba0bb129d23bdadb9e35eef249c09fba85346ce1a99f7a26" exitCode=0 Oct 05 08:18:01 crc kubenswrapper[4846]: I1005 08:18:01.500765 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6bf9574c4-zz2wx" event={"ID":"e4a6f807-103b-4390-8d1f-403cbde1ce2a","Type":"ContainerDied","Data":"4f8c70367acf4b7eba0bb129d23bdadb9e35eef249c09fba85346ce1a99f7a26"} Oct 05 08:18:01 crc kubenswrapper[4846]: I1005 08:18:01.857553 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-6bf9574c4-zz2wx" Oct 05 08:18:01 crc kubenswrapper[4846]: I1005 08:18:01.934697 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tthcw\" (UniqueName: \"kubernetes.io/projected/e4a6f807-103b-4390-8d1f-403cbde1ce2a-kube-api-access-tthcw\") pod \"e4a6f807-103b-4390-8d1f-403cbde1ce2a\" (UID: \"e4a6f807-103b-4390-8d1f-403cbde1ce2a\") " Oct 05 08:18:01 crc kubenswrapper[4846]: I1005 08:18:01.935015 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/e4a6f807-103b-4390-8d1f-403cbde1ce2a-config\") pod \"e4a6f807-103b-4390-8d1f-403cbde1ce2a\" (UID: \"e4a6f807-103b-4390-8d1f-403cbde1ce2a\") " Oct 05 08:18:01 crc kubenswrapper[4846]: I1005 08:18:01.935284 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/e4a6f807-103b-4390-8d1f-403cbde1ce2a-ovndb-tls-certs\") pod \"e4a6f807-103b-4390-8d1f-403cbde1ce2a\" (UID: \"e4a6f807-103b-4390-8d1f-403cbde1ce2a\") " Oct 05 08:18:01 crc kubenswrapper[4846]: I1005 08:18:01.935407 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e4a6f807-103b-4390-8d1f-403cbde1ce2a-combined-ca-bundle\") pod \"e4a6f807-103b-4390-8d1f-403cbde1ce2a\" (UID: \"e4a6f807-103b-4390-8d1f-403cbde1ce2a\") " Oct 05 08:18:01 crc kubenswrapper[4846]: I1005 08:18:01.935517 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/e4a6f807-103b-4390-8d1f-403cbde1ce2a-httpd-config\") pod \"e4a6f807-103b-4390-8d1f-403cbde1ce2a\" (UID: \"e4a6f807-103b-4390-8d1f-403cbde1ce2a\") " Oct 05 08:18:01 crc kubenswrapper[4846]: I1005 08:18:01.946049 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e4a6f807-103b-4390-8d1f-403cbde1ce2a-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "e4a6f807-103b-4390-8d1f-403cbde1ce2a" (UID: "e4a6f807-103b-4390-8d1f-403cbde1ce2a"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:18:01 crc kubenswrapper[4846]: I1005 08:18:01.951041 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e4a6f807-103b-4390-8d1f-403cbde1ce2a-kube-api-access-tthcw" (OuterVolumeSpecName: "kube-api-access-tthcw") pod "e4a6f807-103b-4390-8d1f-403cbde1ce2a" (UID: "e4a6f807-103b-4390-8d1f-403cbde1ce2a"). InnerVolumeSpecName "kube-api-access-tthcw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:18:02 crc kubenswrapper[4846]: I1005 08:18:02.010633 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e4a6f807-103b-4390-8d1f-403cbde1ce2a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e4a6f807-103b-4390-8d1f-403cbde1ce2a" (UID: "e4a6f807-103b-4390-8d1f-403cbde1ce2a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:18:02 crc kubenswrapper[4846]: I1005 08:18:02.019332 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e4a6f807-103b-4390-8d1f-403cbde1ce2a-config" (OuterVolumeSpecName: "config") pod "e4a6f807-103b-4390-8d1f-403cbde1ce2a" (UID: "e4a6f807-103b-4390-8d1f-403cbde1ce2a"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:18:02 crc kubenswrapper[4846]: I1005 08:18:02.037440 4846 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e4a6f807-103b-4390-8d1f-403cbde1ce2a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 08:18:02 crc kubenswrapper[4846]: I1005 08:18:02.037654 4846 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/e4a6f807-103b-4390-8d1f-403cbde1ce2a-httpd-config\") on node \"crc\" DevicePath \"\"" Oct 05 08:18:02 crc kubenswrapper[4846]: I1005 08:18:02.037724 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tthcw\" (UniqueName: \"kubernetes.io/projected/e4a6f807-103b-4390-8d1f-403cbde1ce2a-kube-api-access-tthcw\") on node \"crc\" DevicePath \"\"" Oct 05 08:18:02 crc kubenswrapper[4846]: I1005 08:18:02.037787 4846 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/e4a6f807-103b-4390-8d1f-403cbde1ce2a-config\") on node \"crc\" DevicePath \"\"" Oct 05 08:18:02 crc kubenswrapper[4846]: I1005 08:18:02.053955 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e4a6f807-103b-4390-8d1f-403cbde1ce2a-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "e4a6f807-103b-4390-8d1f-403cbde1ce2a" (UID: "e4a6f807-103b-4390-8d1f-403cbde1ce2a"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:18:02 crc kubenswrapper[4846]: I1005 08:18:02.139598 4846 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/e4a6f807-103b-4390-8d1f-403cbde1ce2a-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 05 08:18:02 crc kubenswrapper[4846]: I1005 08:18:02.515723 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-6bf9574c4-zz2wx" Oct 05 08:18:02 crc kubenswrapper[4846]: I1005 08:18:02.518435 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6bf9574c4-zz2wx" event={"ID":"e4a6f807-103b-4390-8d1f-403cbde1ce2a","Type":"ContainerDied","Data":"1309afb0b88458efcb5ce8b337e263ef0542c1a9c2d01521ef3bf1123705830a"} Oct 05 08:18:02 crc kubenswrapper[4846]: I1005 08:18:02.518487 4846 scope.go:117] "RemoveContainer" containerID="9e4655e9e5c45ff6cd334ce0fef1df22513c2a5ad10159833f9e711efd01236d" Oct 05 08:18:02 crc kubenswrapper[4846]: I1005 08:18:02.579985 4846 scope.go:117] "RemoveContainer" containerID="4f8c70367acf4b7eba0bb129d23bdadb9e35eef249c09fba85346ce1a99f7a26" Oct 05 08:18:02 crc kubenswrapper[4846]: I1005 08:18:02.586827 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-6bf9574c4-zz2wx"] Oct 05 08:18:02 crc kubenswrapper[4846]: I1005 08:18:02.596448 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-6bf9574c4-zz2wx"] Oct 05 08:18:04 crc kubenswrapper[4846]: I1005 08:18:04.516588 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e4a6f807-103b-4390-8d1f-403cbde1ce2a" path="/var/lib/kubelet/pods/e4a6f807-103b-4390-8d1f-403cbde1ce2a/volumes" Oct 05 08:18:07 crc kubenswrapper[4846]: I1005 08:18:07.564723 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-vr9g5"] Oct 05 08:18:07 crc kubenswrapper[4846]: E1005 08:18:07.565453 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bcfb4bd3-9712-466d-bc48-317857cdd151" containerName="extract-content" Oct 05 08:18:07 crc kubenswrapper[4846]: I1005 08:18:07.565467 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="bcfb4bd3-9712-466d-bc48-317857cdd151" containerName="extract-content" Oct 05 08:18:07 crc kubenswrapper[4846]: E1005 08:18:07.565484 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9d8932cd-6871-4b34-90cd-3945e6ce6956" containerName="init" Oct 05 08:18:07 crc kubenswrapper[4846]: I1005 08:18:07.565490 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="9d8932cd-6871-4b34-90cd-3945e6ce6956" containerName="init" Oct 05 08:18:07 crc kubenswrapper[4846]: E1005 08:18:07.565502 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9d8932cd-6871-4b34-90cd-3945e6ce6956" containerName="dnsmasq-dns" Oct 05 08:18:07 crc kubenswrapper[4846]: I1005 08:18:07.565509 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="9d8932cd-6871-4b34-90cd-3945e6ce6956" containerName="dnsmasq-dns" Oct 05 08:18:07 crc kubenswrapper[4846]: E1005 08:18:07.565520 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bcfb4bd3-9712-466d-bc48-317857cdd151" containerName="extract-utilities" Oct 05 08:18:07 crc kubenswrapper[4846]: I1005 08:18:07.565527 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="bcfb4bd3-9712-466d-bc48-317857cdd151" containerName="extract-utilities" Oct 05 08:18:07 crc kubenswrapper[4846]: E1005 08:18:07.565544 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bcfb4bd3-9712-466d-bc48-317857cdd151" containerName="registry-server" Oct 05 08:18:07 crc kubenswrapper[4846]: I1005 08:18:07.565550 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="bcfb4bd3-9712-466d-bc48-317857cdd151" containerName="registry-server" Oct 05 08:18:07 crc kubenswrapper[4846]: E1005 08:18:07.565562 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e4a6f807-103b-4390-8d1f-403cbde1ce2a" containerName="neutron-api" Oct 05 08:18:07 crc kubenswrapper[4846]: I1005 08:18:07.565568 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="e4a6f807-103b-4390-8d1f-403cbde1ce2a" containerName="neutron-api" Oct 05 08:18:07 crc kubenswrapper[4846]: E1005 08:18:07.565584 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e4a6f807-103b-4390-8d1f-403cbde1ce2a" containerName="neutron-httpd" Oct 05 08:18:07 crc kubenswrapper[4846]: I1005 08:18:07.565589 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="e4a6f807-103b-4390-8d1f-403cbde1ce2a" containerName="neutron-httpd" Oct 05 08:18:07 crc kubenswrapper[4846]: I1005 08:18:07.565763 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="9d8932cd-6871-4b34-90cd-3945e6ce6956" containerName="dnsmasq-dns" Oct 05 08:18:07 crc kubenswrapper[4846]: I1005 08:18:07.565781 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="bcfb4bd3-9712-466d-bc48-317857cdd151" containerName="registry-server" Oct 05 08:18:07 crc kubenswrapper[4846]: I1005 08:18:07.565792 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="e4a6f807-103b-4390-8d1f-403cbde1ce2a" containerName="neutron-httpd" Oct 05 08:18:07 crc kubenswrapper[4846]: I1005 08:18:07.565802 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="e4a6f807-103b-4390-8d1f-403cbde1ce2a" containerName="neutron-api" Oct 05 08:18:07 crc kubenswrapper[4846]: I1005 08:18:07.566459 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-vr9g5" Oct 05 08:18:07 crc kubenswrapper[4846]: I1005 08:18:07.570809 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Oct 05 08:18:07 crc kubenswrapper[4846]: I1005 08:18:07.571002 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-conf" Oct 05 08:18:07 crc kubenswrapper[4846]: I1005 08:18:07.571117 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-scripts" Oct 05 08:18:07 crc kubenswrapper[4846]: I1005 08:18:07.571306 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-swift-dockercfg-sp9lb" Oct 05 08:18:07 crc kubenswrapper[4846]: I1005 08:18:07.571424 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-config-data" Oct 05 08:18:07 crc kubenswrapper[4846]: I1005 08:18:07.574902 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-vr9g5"] Oct 05 08:18:07 crc kubenswrapper[4846]: I1005 08:18:07.625868 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6688fc9d5f-7jvr8"] Oct 05 08:18:07 crc kubenswrapper[4846]: I1005 08:18:07.632258 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6688fc9d5f-7jvr8" Oct 05 08:18:07 crc kubenswrapper[4846]: I1005 08:18:07.661168 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6688fc9d5f-7jvr8"] Oct 05 08:18:07 crc kubenswrapper[4846]: I1005 08:18:07.755381 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-svt7n\" (UniqueName: \"kubernetes.io/projected/77e9ec78-ecc3-4953-9c33-f04b3f50b890-kube-api-access-svt7n\") pod \"swift-ring-rebalance-vr9g5\" (UID: \"77e9ec78-ecc3-4953-9c33-f04b3f50b890\") " pod="openstack/swift-ring-rebalance-vr9g5" Oct 05 08:18:07 crc kubenswrapper[4846]: I1005 08:18:07.755468 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/77e9ec78-ecc3-4953-9c33-f04b3f50b890-swiftconf\") pod \"swift-ring-rebalance-vr9g5\" (UID: \"77e9ec78-ecc3-4953-9c33-f04b3f50b890\") " pod="openstack/swift-ring-rebalance-vr9g5" Oct 05 08:18:07 crc kubenswrapper[4846]: I1005 08:18:07.755515 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/77e9ec78-ecc3-4953-9c33-f04b3f50b890-ring-data-devices\") pod \"swift-ring-rebalance-vr9g5\" (UID: \"77e9ec78-ecc3-4953-9c33-f04b3f50b890\") " pod="openstack/swift-ring-rebalance-vr9g5" Oct 05 08:18:07 crc kubenswrapper[4846]: I1005 08:18:07.755550 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d7bace3b-db65-4724-808a-38bf98c604d0-ovsdbserver-nb\") pod \"dnsmasq-dns-6688fc9d5f-7jvr8\" (UID: \"d7bace3b-db65-4724-808a-38bf98c604d0\") " pod="openstack/dnsmasq-dns-6688fc9d5f-7jvr8" Oct 05 08:18:07 crc kubenswrapper[4846]: I1005 08:18:07.755618 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/77e9ec78-ecc3-4953-9c33-f04b3f50b890-dispersionconf\") pod \"swift-ring-rebalance-vr9g5\" (UID: \"77e9ec78-ecc3-4953-9c33-f04b3f50b890\") " pod="openstack/swift-ring-rebalance-vr9g5" Oct 05 08:18:07 crc kubenswrapper[4846]: I1005 08:18:07.755643 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d7bace3b-db65-4724-808a-38bf98c604d0-dns-svc\") pod \"dnsmasq-dns-6688fc9d5f-7jvr8\" (UID: \"d7bace3b-db65-4724-808a-38bf98c604d0\") " pod="openstack/dnsmasq-dns-6688fc9d5f-7jvr8" Oct 05 08:18:07 crc kubenswrapper[4846]: I1005 08:18:07.755668 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/77e9ec78-ecc3-4953-9c33-f04b3f50b890-combined-ca-bundle\") pod \"swift-ring-rebalance-vr9g5\" (UID: \"77e9ec78-ecc3-4953-9c33-f04b3f50b890\") " pod="openstack/swift-ring-rebalance-vr9g5" Oct 05 08:18:07 crc kubenswrapper[4846]: I1005 08:18:07.755805 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5jkn4\" (UniqueName: \"kubernetes.io/projected/d7bace3b-db65-4724-808a-38bf98c604d0-kube-api-access-5jkn4\") pod \"dnsmasq-dns-6688fc9d5f-7jvr8\" (UID: \"d7bace3b-db65-4724-808a-38bf98c604d0\") " pod="openstack/dnsmasq-dns-6688fc9d5f-7jvr8" Oct 05 08:18:07 crc kubenswrapper[4846]: I1005 08:18:07.755825 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/77e9ec78-ecc3-4953-9c33-f04b3f50b890-scripts\") pod \"swift-ring-rebalance-vr9g5\" (UID: \"77e9ec78-ecc3-4953-9c33-f04b3f50b890\") " pod="openstack/swift-ring-rebalance-vr9g5" Oct 05 08:18:07 crc kubenswrapper[4846]: I1005 08:18:07.755858 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d7bace3b-db65-4724-808a-38bf98c604d0-config\") pod \"dnsmasq-dns-6688fc9d5f-7jvr8\" (UID: \"d7bace3b-db65-4724-808a-38bf98c604d0\") " pod="openstack/dnsmasq-dns-6688fc9d5f-7jvr8" Oct 05 08:18:07 crc kubenswrapper[4846]: I1005 08:18:07.755875 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/77e9ec78-ecc3-4953-9c33-f04b3f50b890-etc-swift\") pod \"swift-ring-rebalance-vr9g5\" (UID: \"77e9ec78-ecc3-4953-9c33-f04b3f50b890\") " pod="openstack/swift-ring-rebalance-vr9g5" Oct 05 08:18:07 crc kubenswrapper[4846]: I1005 08:18:07.755900 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d7bace3b-db65-4724-808a-38bf98c604d0-ovsdbserver-sb\") pod \"dnsmasq-dns-6688fc9d5f-7jvr8\" (UID: \"d7bace3b-db65-4724-808a-38bf98c604d0\") " pod="openstack/dnsmasq-dns-6688fc9d5f-7jvr8" Oct 05 08:18:07 crc kubenswrapper[4846]: I1005 08:18:07.857385 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5jkn4\" (UniqueName: \"kubernetes.io/projected/d7bace3b-db65-4724-808a-38bf98c604d0-kube-api-access-5jkn4\") pod \"dnsmasq-dns-6688fc9d5f-7jvr8\" (UID: \"d7bace3b-db65-4724-808a-38bf98c604d0\") " pod="openstack/dnsmasq-dns-6688fc9d5f-7jvr8" Oct 05 08:18:07 crc kubenswrapper[4846]: I1005 08:18:07.857454 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/77e9ec78-ecc3-4953-9c33-f04b3f50b890-scripts\") pod \"swift-ring-rebalance-vr9g5\" (UID: \"77e9ec78-ecc3-4953-9c33-f04b3f50b890\") " pod="openstack/swift-ring-rebalance-vr9g5" Oct 05 08:18:07 crc kubenswrapper[4846]: I1005 08:18:07.857501 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d7bace3b-db65-4724-808a-38bf98c604d0-config\") pod \"dnsmasq-dns-6688fc9d5f-7jvr8\" (UID: \"d7bace3b-db65-4724-808a-38bf98c604d0\") " pod="openstack/dnsmasq-dns-6688fc9d5f-7jvr8" Oct 05 08:18:07 crc kubenswrapper[4846]: I1005 08:18:07.857537 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/77e9ec78-ecc3-4953-9c33-f04b3f50b890-etc-swift\") pod \"swift-ring-rebalance-vr9g5\" (UID: \"77e9ec78-ecc3-4953-9c33-f04b3f50b890\") " pod="openstack/swift-ring-rebalance-vr9g5" Oct 05 08:18:07 crc kubenswrapper[4846]: I1005 08:18:07.857575 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d7bace3b-db65-4724-808a-38bf98c604d0-ovsdbserver-sb\") pod \"dnsmasq-dns-6688fc9d5f-7jvr8\" (UID: \"d7bace3b-db65-4724-808a-38bf98c604d0\") " pod="openstack/dnsmasq-dns-6688fc9d5f-7jvr8" Oct 05 08:18:07 crc kubenswrapper[4846]: I1005 08:18:07.857635 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-svt7n\" (UniqueName: \"kubernetes.io/projected/77e9ec78-ecc3-4953-9c33-f04b3f50b890-kube-api-access-svt7n\") pod \"swift-ring-rebalance-vr9g5\" (UID: \"77e9ec78-ecc3-4953-9c33-f04b3f50b890\") " pod="openstack/swift-ring-rebalance-vr9g5" Oct 05 08:18:07 crc kubenswrapper[4846]: I1005 08:18:07.857670 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/77e9ec78-ecc3-4953-9c33-f04b3f50b890-swiftconf\") pod \"swift-ring-rebalance-vr9g5\" (UID: \"77e9ec78-ecc3-4953-9c33-f04b3f50b890\") " pod="openstack/swift-ring-rebalance-vr9g5" Oct 05 08:18:07 crc kubenswrapper[4846]: I1005 08:18:07.857712 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/77e9ec78-ecc3-4953-9c33-f04b3f50b890-ring-data-devices\") pod \"swift-ring-rebalance-vr9g5\" (UID: \"77e9ec78-ecc3-4953-9c33-f04b3f50b890\") " pod="openstack/swift-ring-rebalance-vr9g5" Oct 05 08:18:07 crc kubenswrapper[4846]: I1005 08:18:07.857750 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d7bace3b-db65-4724-808a-38bf98c604d0-ovsdbserver-nb\") pod \"dnsmasq-dns-6688fc9d5f-7jvr8\" (UID: \"d7bace3b-db65-4724-808a-38bf98c604d0\") " pod="openstack/dnsmasq-dns-6688fc9d5f-7jvr8" Oct 05 08:18:07 crc kubenswrapper[4846]: I1005 08:18:07.857788 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/77e9ec78-ecc3-4953-9c33-f04b3f50b890-dispersionconf\") pod \"swift-ring-rebalance-vr9g5\" (UID: \"77e9ec78-ecc3-4953-9c33-f04b3f50b890\") " pod="openstack/swift-ring-rebalance-vr9g5" Oct 05 08:18:07 crc kubenswrapper[4846]: I1005 08:18:07.857824 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d7bace3b-db65-4724-808a-38bf98c604d0-dns-svc\") pod \"dnsmasq-dns-6688fc9d5f-7jvr8\" (UID: \"d7bace3b-db65-4724-808a-38bf98c604d0\") " pod="openstack/dnsmasq-dns-6688fc9d5f-7jvr8" Oct 05 08:18:07 crc kubenswrapper[4846]: I1005 08:18:07.857869 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/77e9ec78-ecc3-4953-9c33-f04b3f50b890-combined-ca-bundle\") pod \"swift-ring-rebalance-vr9g5\" (UID: \"77e9ec78-ecc3-4953-9c33-f04b3f50b890\") " pod="openstack/swift-ring-rebalance-vr9g5" Oct 05 08:18:07 crc kubenswrapper[4846]: I1005 08:18:07.858247 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/77e9ec78-ecc3-4953-9c33-f04b3f50b890-etc-swift\") pod \"swift-ring-rebalance-vr9g5\" (UID: \"77e9ec78-ecc3-4953-9c33-f04b3f50b890\") " pod="openstack/swift-ring-rebalance-vr9g5" Oct 05 08:18:07 crc kubenswrapper[4846]: I1005 08:18:07.858500 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/77e9ec78-ecc3-4953-9c33-f04b3f50b890-scripts\") pod \"swift-ring-rebalance-vr9g5\" (UID: \"77e9ec78-ecc3-4953-9c33-f04b3f50b890\") " pod="openstack/swift-ring-rebalance-vr9g5" Oct 05 08:18:07 crc kubenswrapper[4846]: I1005 08:18:07.858548 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d7bace3b-db65-4724-808a-38bf98c604d0-config\") pod \"dnsmasq-dns-6688fc9d5f-7jvr8\" (UID: \"d7bace3b-db65-4724-808a-38bf98c604d0\") " pod="openstack/dnsmasq-dns-6688fc9d5f-7jvr8" Oct 05 08:18:07 crc kubenswrapper[4846]: I1005 08:18:07.858909 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d7bace3b-db65-4724-808a-38bf98c604d0-ovsdbserver-sb\") pod \"dnsmasq-dns-6688fc9d5f-7jvr8\" (UID: \"d7bace3b-db65-4724-808a-38bf98c604d0\") " pod="openstack/dnsmasq-dns-6688fc9d5f-7jvr8" Oct 05 08:18:07 crc kubenswrapper[4846]: I1005 08:18:07.859116 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/77e9ec78-ecc3-4953-9c33-f04b3f50b890-ring-data-devices\") pod \"swift-ring-rebalance-vr9g5\" (UID: \"77e9ec78-ecc3-4953-9c33-f04b3f50b890\") " pod="openstack/swift-ring-rebalance-vr9g5" Oct 05 08:18:07 crc kubenswrapper[4846]: I1005 08:18:07.859132 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d7bace3b-db65-4724-808a-38bf98c604d0-ovsdbserver-nb\") pod \"dnsmasq-dns-6688fc9d5f-7jvr8\" (UID: \"d7bace3b-db65-4724-808a-38bf98c604d0\") " pod="openstack/dnsmasq-dns-6688fc9d5f-7jvr8" Oct 05 08:18:07 crc kubenswrapper[4846]: I1005 08:18:07.859479 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d7bace3b-db65-4724-808a-38bf98c604d0-dns-svc\") pod \"dnsmasq-dns-6688fc9d5f-7jvr8\" (UID: \"d7bace3b-db65-4724-808a-38bf98c604d0\") " pod="openstack/dnsmasq-dns-6688fc9d5f-7jvr8" Oct 05 08:18:07 crc kubenswrapper[4846]: I1005 08:18:07.864446 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/77e9ec78-ecc3-4953-9c33-f04b3f50b890-combined-ca-bundle\") pod \"swift-ring-rebalance-vr9g5\" (UID: \"77e9ec78-ecc3-4953-9c33-f04b3f50b890\") " pod="openstack/swift-ring-rebalance-vr9g5" Oct 05 08:18:07 crc kubenswrapper[4846]: I1005 08:18:07.865720 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/77e9ec78-ecc3-4953-9c33-f04b3f50b890-dispersionconf\") pod \"swift-ring-rebalance-vr9g5\" (UID: \"77e9ec78-ecc3-4953-9c33-f04b3f50b890\") " pod="openstack/swift-ring-rebalance-vr9g5" Oct 05 08:18:07 crc kubenswrapper[4846]: I1005 08:18:07.875019 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5jkn4\" (UniqueName: \"kubernetes.io/projected/d7bace3b-db65-4724-808a-38bf98c604d0-kube-api-access-5jkn4\") pod \"dnsmasq-dns-6688fc9d5f-7jvr8\" (UID: \"d7bace3b-db65-4724-808a-38bf98c604d0\") " pod="openstack/dnsmasq-dns-6688fc9d5f-7jvr8" Oct 05 08:18:07 crc kubenswrapper[4846]: I1005 08:18:07.876972 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/77e9ec78-ecc3-4953-9c33-f04b3f50b890-swiftconf\") pod \"swift-ring-rebalance-vr9g5\" (UID: \"77e9ec78-ecc3-4953-9c33-f04b3f50b890\") " pod="openstack/swift-ring-rebalance-vr9g5" Oct 05 08:18:07 crc kubenswrapper[4846]: I1005 08:18:07.879246 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-svt7n\" (UniqueName: \"kubernetes.io/projected/77e9ec78-ecc3-4953-9c33-f04b3f50b890-kube-api-access-svt7n\") pod \"swift-ring-rebalance-vr9g5\" (UID: \"77e9ec78-ecc3-4953-9c33-f04b3f50b890\") " pod="openstack/swift-ring-rebalance-vr9g5" Oct 05 08:18:07 crc kubenswrapper[4846]: I1005 08:18:07.898312 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-vr9g5" Oct 05 08:18:07 crc kubenswrapper[4846]: I1005 08:18:07.971061 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6688fc9d5f-7jvr8" Oct 05 08:18:08 crc kubenswrapper[4846]: I1005 08:18:08.359999 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-vr9g5"] Oct 05 08:18:08 crc kubenswrapper[4846]: I1005 08:18:08.372915 4846 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 05 08:18:08 crc kubenswrapper[4846]: W1005 08:18:08.495624 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd7bace3b_db65_4724_808a_38bf98c604d0.slice/crio-26ca66d9cc821152cb02ae73d9a9494f13333d5977ffe73147217911388abeb1 WatchSource:0}: Error finding container 26ca66d9cc821152cb02ae73d9a9494f13333d5977ffe73147217911388abeb1: Status 404 returned error can't find the container with id 26ca66d9cc821152cb02ae73d9a9494f13333d5977ffe73147217911388abeb1 Oct 05 08:18:08 crc kubenswrapper[4846]: I1005 08:18:08.516707 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6688fc9d5f-7jvr8"] Oct 05 08:18:08 crc kubenswrapper[4846]: I1005 08:18:08.576328 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-vr9g5" event={"ID":"77e9ec78-ecc3-4953-9c33-f04b3f50b890","Type":"ContainerStarted","Data":"cb98540a2bd7bac69405e36967d82af679509c808c7cb938860dc8dab8fe3169"} Oct 05 08:18:08 crc kubenswrapper[4846]: I1005 08:18:08.582358 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6688fc9d5f-7jvr8" event={"ID":"d7bace3b-db65-4724-808a-38bf98c604d0","Type":"ContainerStarted","Data":"26ca66d9cc821152cb02ae73d9a9494f13333d5977ffe73147217911388abeb1"} Oct 05 08:18:09 crc kubenswrapper[4846]: I1005 08:18:09.599656 4846 generic.go:334] "Generic (PLEG): container finished" podID="d7bace3b-db65-4724-808a-38bf98c604d0" containerID="0180312d58237bdf8500652e38bbc8bcd0953e65fd6875b29c29ddaea15f7e47" exitCode=0 Oct 05 08:18:09 crc kubenswrapper[4846]: I1005 08:18:09.599725 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6688fc9d5f-7jvr8" event={"ID":"d7bace3b-db65-4724-808a-38bf98c604d0","Type":"ContainerDied","Data":"0180312d58237bdf8500652e38bbc8bcd0953e65fd6875b29c29ddaea15f7e47"} Oct 05 08:18:09 crc kubenswrapper[4846]: I1005 08:18:09.601975 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-proxy-5bcb5dcc55-6mzb8"] Oct 05 08:18:09 crc kubenswrapper[4846]: I1005 08:18:09.608091 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-5bcb5dcc55-6mzb8" Oct 05 08:18:09 crc kubenswrapper[4846]: I1005 08:18:09.612568 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-files" Oct 05 08:18:09 crc kubenswrapper[4846]: I1005 08:18:09.680135 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-5bcb5dcc55-6mzb8"] Oct 05 08:18:09 crc kubenswrapper[4846]: I1005 08:18:09.713358 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/9491b5bb-7012-4a5b-bc96-d96af06e9736-etc-swift\") pod \"swift-proxy-5bcb5dcc55-6mzb8\" (UID: \"9491b5bb-7012-4a5b-bc96-d96af06e9736\") " pod="openstack/swift-proxy-5bcb5dcc55-6mzb8" Oct 05 08:18:09 crc kubenswrapper[4846]: I1005 08:18:09.713439 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9491b5bb-7012-4a5b-bc96-d96af06e9736-run-httpd\") pod \"swift-proxy-5bcb5dcc55-6mzb8\" (UID: \"9491b5bb-7012-4a5b-bc96-d96af06e9736\") " pod="openstack/swift-proxy-5bcb5dcc55-6mzb8" Oct 05 08:18:09 crc kubenswrapper[4846]: I1005 08:18:09.713493 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9491b5bb-7012-4a5b-bc96-d96af06e9736-log-httpd\") pod \"swift-proxy-5bcb5dcc55-6mzb8\" (UID: \"9491b5bb-7012-4a5b-bc96-d96af06e9736\") " pod="openstack/swift-proxy-5bcb5dcc55-6mzb8" Oct 05 08:18:09 crc kubenswrapper[4846]: I1005 08:18:09.713513 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9491b5bb-7012-4a5b-bc96-d96af06e9736-combined-ca-bundle\") pod \"swift-proxy-5bcb5dcc55-6mzb8\" (UID: \"9491b5bb-7012-4a5b-bc96-d96af06e9736\") " pod="openstack/swift-proxy-5bcb5dcc55-6mzb8" Oct 05 08:18:09 crc kubenswrapper[4846]: I1005 08:18:09.713723 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d558p\" (UniqueName: \"kubernetes.io/projected/9491b5bb-7012-4a5b-bc96-d96af06e9736-kube-api-access-d558p\") pod \"swift-proxy-5bcb5dcc55-6mzb8\" (UID: \"9491b5bb-7012-4a5b-bc96-d96af06e9736\") " pod="openstack/swift-proxy-5bcb5dcc55-6mzb8" Oct 05 08:18:09 crc kubenswrapper[4846]: I1005 08:18:09.713757 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9491b5bb-7012-4a5b-bc96-d96af06e9736-config-data\") pod \"swift-proxy-5bcb5dcc55-6mzb8\" (UID: \"9491b5bb-7012-4a5b-bc96-d96af06e9736\") " pod="openstack/swift-proxy-5bcb5dcc55-6mzb8" Oct 05 08:18:09 crc kubenswrapper[4846]: I1005 08:18:09.816379 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d558p\" (UniqueName: \"kubernetes.io/projected/9491b5bb-7012-4a5b-bc96-d96af06e9736-kube-api-access-d558p\") pod \"swift-proxy-5bcb5dcc55-6mzb8\" (UID: \"9491b5bb-7012-4a5b-bc96-d96af06e9736\") " pod="openstack/swift-proxy-5bcb5dcc55-6mzb8" Oct 05 08:18:09 crc kubenswrapper[4846]: I1005 08:18:09.816435 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9491b5bb-7012-4a5b-bc96-d96af06e9736-config-data\") pod \"swift-proxy-5bcb5dcc55-6mzb8\" (UID: \"9491b5bb-7012-4a5b-bc96-d96af06e9736\") " pod="openstack/swift-proxy-5bcb5dcc55-6mzb8" Oct 05 08:18:09 crc kubenswrapper[4846]: I1005 08:18:09.816464 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/9491b5bb-7012-4a5b-bc96-d96af06e9736-etc-swift\") pod \"swift-proxy-5bcb5dcc55-6mzb8\" (UID: \"9491b5bb-7012-4a5b-bc96-d96af06e9736\") " pod="openstack/swift-proxy-5bcb5dcc55-6mzb8" Oct 05 08:18:09 crc kubenswrapper[4846]: I1005 08:18:09.816484 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9491b5bb-7012-4a5b-bc96-d96af06e9736-run-httpd\") pod \"swift-proxy-5bcb5dcc55-6mzb8\" (UID: \"9491b5bb-7012-4a5b-bc96-d96af06e9736\") " pod="openstack/swift-proxy-5bcb5dcc55-6mzb8" Oct 05 08:18:09 crc kubenswrapper[4846]: I1005 08:18:09.816515 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9491b5bb-7012-4a5b-bc96-d96af06e9736-log-httpd\") pod \"swift-proxy-5bcb5dcc55-6mzb8\" (UID: \"9491b5bb-7012-4a5b-bc96-d96af06e9736\") " pod="openstack/swift-proxy-5bcb5dcc55-6mzb8" Oct 05 08:18:09 crc kubenswrapper[4846]: I1005 08:18:09.816537 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9491b5bb-7012-4a5b-bc96-d96af06e9736-combined-ca-bundle\") pod \"swift-proxy-5bcb5dcc55-6mzb8\" (UID: \"9491b5bb-7012-4a5b-bc96-d96af06e9736\") " pod="openstack/swift-proxy-5bcb5dcc55-6mzb8" Oct 05 08:18:09 crc kubenswrapper[4846]: I1005 08:18:09.817773 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9491b5bb-7012-4a5b-bc96-d96af06e9736-run-httpd\") pod \"swift-proxy-5bcb5dcc55-6mzb8\" (UID: \"9491b5bb-7012-4a5b-bc96-d96af06e9736\") " pod="openstack/swift-proxy-5bcb5dcc55-6mzb8" Oct 05 08:18:09 crc kubenswrapper[4846]: I1005 08:18:09.818097 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9491b5bb-7012-4a5b-bc96-d96af06e9736-log-httpd\") pod \"swift-proxy-5bcb5dcc55-6mzb8\" (UID: \"9491b5bb-7012-4a5b-bc96-d96af06e9736\") " pod="openstack/swift-proxy-5bcb5dcc55-6mzb8" Oct 05 08:18:09 crc kubenswrapper[4846]: I1005 08:18:09.824362 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/9491b5bb-7012-4a5b-bc96-d96af06e9736-etc-swift\") pod \"swift-proxy-5bcb5dcc55-6mzb8\" (UID: \"9491b5bb-7012-4a5b-bc96-d96af06e9736\") " pod="openstack/swift-proxy-5bcb5dcc55-6mzb8" Oct 05 08:18:09 crc kubenswrapper[4846]: I1005 08:18:09.824881 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9491b5bb-7012-4a5b-bc96-d96af06e9736-combined-ca-bundle\") pod \"swift-proxy-5bcb5dcc55-6mzb8\" (UID: \"9491b5bb-7012-4a5b-bc96-d96af06e9736\") " pod="openstack/swift-proxy-5bcb5dcc55-6mzb8" Oct 05 08:18:09 crc kubenswrapper[4846]: I1005 08:18:09.826941 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9491b5bb-7012-4a5b-bc96-d96af06e9736-config-data\") pod \"swift-proxy-5bcb5dcc55-6mzb8\" (UID: \"9491b5bb-7012-4a5b-bc96-d96af06e9736\") " pod="openstack/swift-proxy-5bcb5dcc55-6mzb8" Oct 05 08:18:09 crc kubenswrapper[4846]: I1005 08:18:09.835972 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d558p\" (UniqueName: \"kubernetes.io/projected/9491b5bb-7012-4a5b-bc96-d96af06e9736-kube-api-access-d558p\") pod \"swift-proxy-5bcb5dcc55-6mzb8\" (UID: \"9491b5bb-7012-4a5b-bc96-d96af06e9736\") " pod="openstack/swift-proxy-5bcb5dcc55-6mzb8" Oct 05 08:18:10 crc kubenswrapper[4846]: I1005 08:18:10.128426 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-5bcb5dcc55-6mzb8" Oct 05 08:18:10 crc kubenswrapper[4846]: I1005 08:18:10.614722 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6688fc9d5f-7jvr8" event={"ID":"d7bace3b-db65-4724-808a-38bf98c604d0","Type":"ContainerStarted","Data":"97b62413208bc6bd360f03b49cb792fe05f09a1c2a037f6e28171a331a73e1d5"} Oct 05 08:18:10 crc kubenswrapper[4846]: I1005 08:18:10.616096 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6688fc9d5f-7jvr8" Oct 05 08:18:10 crc kubenswrapper[4846]: I1005 08:18:10.639381 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6688fc9d5f-7jvr8" podStartSLOduration=3.639358357 podStartE2EDuration="3.639358357s" podCreationTimestamp="2025-10-05 08:18:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 08:18:10.633013607 +0000 UTC m=+5412.873866382" watchObservedRunningTime="2025-10-05 08:18:10.639358357 +0000 UTC m=+5412.880211132" Oct 05 08:18:11 crc kubenswrapper[4846]: I1005 08:18:11.979490 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-proxy-5b5cc989cf-btgr5"] Oct 05 08:18:11 crc kubenswrapper[4846]: I1005 08:18:11.981526 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-5b5cc989cf-btgr5" Oct 05 08:18:11 crc kubenswrapper[4846]: I1005 08:18:11.983671 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-internal-svc" Oct 05 08:18:11 crc kubenswrapper[4846]: I1005 08:18:11.984054 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-public-svc" Oct 05 08:18:12 crc kubenswrapper[4846]: I1005 08:18:11.989920 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-5b5cc989cf-btgr5"] Oct 05 08:18:12 crc kubenswrapper[4846]: I1005 08:18:12.056884 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/d3b7b761-5203-4ed5-8fc1-4f4a3d572592-etc-swift\") pod \"swift-proxy-5b5cc989cf-btgr5\" (UID: \"d3b7b761-5203-4ed5-8fc1-4f4a3d572592\") " pod="openstack/swift-proxy-5b5cc989cf-btgr5" Oct 05 08:18:12 crc kubenswrapper[4846]: I1005 08:18:12.057005 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d3b7b761-5203-4ed5-8fc1-4f4a3d572592-run-httpd\") pod \"swift-proxy-5b5cc989cf-btgr5\" (UID: \"d3b7b761-5203-4ed5-8fc1-4f4a3d572592\") " pod="openstack/swift-proxy-5b5cc989cf-btgr5" Oct 05 08:18:12 crc kubenswrapper[4846]: I1005 08:18:12.057373 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d3b7b761-5203-4ed5-8fc1-4f4a3d572592-combined-ca-bundle\") pod \"swift-proxy-5b5cc989cf-btgr5\" (UID: \"d3b7b761-5203-4ed5-8fc1-4f4a3d572592\") " pod="openstack/swift-proxy-5b5cc989cf-btgr5" Oct 05 08:18:12 crc kubenswrapper[4846]: I1005 08:18:12.057443 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gp59v\" (UniqueName: \"kubernetes.io/projected/d3b7b761-5203-4ed5-8fc1-4f4a3d572592-kube-api-access-gp59v\") pod \"swift-proxy-5b5cc989cf-btgr5\" (UID: \"d3b7b761-5203-4ed5-8fc1-4f4a3d572592\") " pod="openstack/swift-proxy-5b5cc989cf-btgr5" Oct 05 08:18:12 crc kubenswrapper[4846]: I1005 08:18:12.057479 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d3b7b761-5203-4ed5-8fc1-4f4a3d572592-config-data\") pod \"swift-proxy-5b5cc989cf-btgr5\" (UID: \"d3b7b761-5203-4ed5-8fc1-4f4a3d572592\") " pod="openstack/swift-proxy-5b5cc989cf-btgr5" Oct 05 08:18:12 crc kubenswrapper[4846]: I1005 08:18:12.057515 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d3b7b761-5203-4ed5-8fc1-4f4a3d572592-internal-tls-certs\") pod \"swift-proxy-5b5cc989cf-btgr5\" (UID: \"d3b7b761-5203-4ed5-8fc1-4f4a3d572592\") " pod="openstack/swift-proxy-5b5cc989cf-btgr5" Oct 05 08:18:12 crc kubenswrapper[4846]: I1005 08:18:12.057629 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d3b7b761-5203-4ed5-8fc1-4f4a3d572592-public-tls-certs\") pod \"swift-proxy-5b5cc989cf-btgr5\" (UID: \"d3b7b761-5203-4ed5-8fc1-4f4a3d572592\") " pod="openstack/swift-proxy-5b5cc989cf-btgr5" Oct 05 08:18:12 crc kubenswrapper[4846]: I1005 08:18:12.057664 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d3b7b761-5203-4ed5-8fc1-4f4a3d572592-log-httpd\") pod \"swift-proxy-5b5cc989cf-btgr5\" (UID: \"d3b7b761-5203-4ed5-8fc1-4f4a3d572592\") " pod="openstack/swift-proxy-5b5cc989cf-btgr5" Oct 05 08:18:12 crc kubenswrapper[4846]: I1005 08:18:12.159721 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gp59v\" (UniqueName: \"kubernetes.io/projected/d3b7b761-5203-4ed5-8fc1-4f4a3d572592-kube-api-access-gp59v\") pod \"swift-proxy-5b5cc989cf-btgr5\" (UID: \"d3b7b761-5203-4ed5-8fc1-4f4a3d572592\") " pod="openstack/swift-proxy-5b5cc989cf-btgr5" Oct 05 08:18:12 crc kubenswrapper[4846]: I1005 08:18:12.160748 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d3b7b761-5203-4ed5-8fc1-4f4a3d572592-config-data\") pod \"swift-proxy-5b5cc989cf-btgr5\" (UID: \"d3b7b761-5203-4ed5-8fc1-4f4a3d572592\") " pod="openstack/swift-proxy-5b5cc989cf-btgr5" Oct 05 08:18:12 crc kubenswrapper[4846]: I1005 08:18:12.160781 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d3b7b761-5203-4ed5-8fc1-4f4a3d572592-internal-tls-certs\") pod \"swift-proxy-5b5cc989cf-btgr5\" (UID: \"d3b7b761-5203-4ed5-8fc1-4f4a3d572592\") " pod="openstack/swift-proxy-5b5cc989cf-btgr5" Oct 05 08:18:12 crc kubenswrapper[4846]: I1005 08:18:12.160824 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d3b7b761-5203-4ed5-8fc1-4f4a3d572592-public-tls-certs\") pod \"swift-proxy-5b5cc989cf-btgr5\" (UID: \"d3b7b761-5203-4ed5-8fc1-4f4a3d572592\") " pod="openstack/swift-proxy-5b5cc989cf-btgr5" Oct 05 08:18:12 crc kubenswrapper[4846]: I1005 08:18:12.160857 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d3b7b761-5203-4ed5-8fc1-4f4a3d572592-log-httpd\") pod \"swift-proxy-5b5cc989cf-btgr5\" (UID: \"d3b7b761-5203-4ed5-8fc1-4f4a3d572592\") " pod="openstack/swift-proxy-5b5cc989cf-btgr5" Oct 05 08:18:12 crc kubenswrapper[4846]: I1005 08:18:12.160936 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/d3b7b761-5203-4ed5-8fc1-4f4a3d572592-etc-swift\") pod \"swift-proxy-5b5cc989cf-btgr5\" (UID: \"d3b7b761-5203-4ed5-8fc1-4f4a3d572592\") " pod="openstack/swift-proxy-5b5cc989cf-btgr5" Oct 05 08:18:12 crc kubenswrapper[4846]: I1005 08:18:12.160961 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d3b7b761-5203-4ed5-8fc1-4f4a3d572592-run-httpd\") pod \"swift-proxy-5b5cc989cf-btgr5\" (UID: \"d3b7b761-5203-4ed5-8fc1-4f4a3d572592\") " pod="openstack/swift-proxy-5b5cc989cf-btgr5" Oct 05 08:18:12 crc kubenswrapper[4846]: I1005 08:18:12.161124 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d3b7b761-5203-4ed5-8fc1-4f4a3d572592-combined-ca-bundle\") pod \"swift-proxy-5b5cc989cf-btgr5\" (UID: \"d3b7b761-5203-4ed5-8fc1-4f4a3d572592\") " pod="openstack/swift-proxy-5b5cc989cf-btgr5" Oct 05 08:18:12 crc kubenswrapper[4846]: I1005 08:18:12.161927 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d3b7b761-5203-4ed5-8fc1-4f4a3d572592-log-httpd\") pod \"swift-proxy-5b5cc989cf-btgr5\" (UID: \"d3b7b761-5203-4ed5-8fc1-4f4a3d572592\") " pod="openstack/swift-proxy-5b5cc989cf-btgr5" Oct 05 08:18:12 crc kubenswrapper[4846]: I1005 08:18:12.162049 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d3b7b761-5203-4ed5-8fc1-4f4a3d572592-run-httpd\") pod \"swift-proxy-5b5cc989cf-btgr5\" (UID: \"d3b7b761-5203-4ed5-8fc1-4f4a3d572592\") " pod="openstack/swift-proxy-5b5cc989cf-btgr5" Oct 05 08:18:12 crc kubenswrapper[4846]: I1005 08:18:12.167670 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d3b7b761-5203-4ed5-8fc1-4f4a3d572592-config-data\") pod \"swift-proxy-5b5cc989cf-btgr5\" (UID: \"d3b7b761-5203-4ed5-8fc1-4f4a3d572592\") " pod="openstack/swift-proxy-5b5cc989cf-btgr5" Oct 05 08:18:12 crc kubenswrapper[4846]: I1005 08:18:12.169886 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d3b7b761-5203-4ed5-8fc1-4f4a3d572592-combined-ca-bundle\") pod \"swift-proxy-5b5cc989cf-btgr5\" (UID: \"d3b7b761-5203-4ed5-8fc1-4f4a3d572592\") " pod="openstack/swift-proxy-5b5cc989cf-btgr5" Oct 05 08:18:12 crc kubenswrapper[4846]: I1005 08:18:12.169970 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d3b7b761-5203-4ed5-8fc1-4f4a3d572592-internal-tls-certs\") pod \"swift-proxy-5b5cc989cf-btgr5\" (UID: \"d3b7b761-5203-4ed5-8fc1-4f4a3d572592\") " pod="openstack/swift-proxy-5b5cc989cf-btgr5" Oct 05 08:18:12 crc kubenswrapper[4846]: I1005 08:18:12.176870 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/d3b7b761-5203-4ed5-8fc1-4f4a3d572592-etc-swift\") pod \"swift-proxy-5b5cc989cf-btgr5\" (UID: \"d3b7b761-5203-4ed5-8fc1-4f4a3d572592\") " pod="openstack/swift-proxy-5b5cc989cf-btgr5" Oct 05 08:18:12 crc kubenswrapper[4846]: I1005 08:18:12.180250 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gp59v\" (UniqueName: \"kubernetes.io/projected/d3b7b761-5203-4ed5-8fc1-4f4a3d572592-kube-api-access-gp59v\") pod \"swift-proxy-5b5cc989cf-btgr5\" (UID: \"d3b7b761-5203-4ed5-8fc1-4f4a3d572592\") " pod="openstack/swift-proxy-5b5cc989cf-btgr5" Oct 05 08:18:12 crc kubenswrapper[4846]: I1005 08:18:12.180479 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d3b7b761-5203-4ed5-8fc1-4f4a3d572592-public-tls-certs\") pod \"swift-proxy-5b5cc989cf-btgr5\" (UID: \"d3b7b761-5203-4ed5-8fc1-4f4a3d572592\") " pod="openstack/swift-proxy-5b5cc989cf-btgr5" Oct 05 08:18:12 crc kubenswrapper[4846]: I1005 08:18:12.352415 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-5b5cc989cf-btgr5" Oct 05 08:18:13 crc kubenswrapper[4846]: I1005 08:18:13.234454 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-5b5cc989cf-btgr5"] Oct 05 08:18:13 crc kubenswrapper[4846]: W1005 08:18:13.235228 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd3b7b761_5203_4ed5_8fc1_4f4a3d572592.slice/crio-ba64f1f4663700cbfe588e30a3d6bbc3c842b345dfa05cdeae2d7cbf24b7daf4 WatchSource:0}: Error finding container ba64f1f4663700cbfe588e30a3d6bbc3c842b345dfa05cdeae2d7cbf24b7daf4: Status 404 returned error can't find the container with id ba64f1f4663700cbfe588e30a3d6bbc3c842b345dfa05cdeae2d7cbf24b7daf4 Oct 05 08:18:13 crc kubenswrapper[4846]: I1005 08:18:13.434584 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-5bcb5dcc55-6mzb8"] Oct 05 08:18:13 crc kubenswrapper[4846]: W1005 08:18:13.445104 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9491b5bb_7012_4a5b_bc96_d96af06e9736.slice/crio-3f39f2f918e339b94d5b7884386b37466147e94ce36a3ef8771df8d42d434761 WatchSource:0}: Error finding container 3f39f2f918e339b94d5b7884386b37466147e94ce36a3ef8771df8d42d434761: Status 404 returned error can't find the container with id 3f39f2f918e339b94d5b7884386b37466147e94ce36a3ef8771df8d42d434761 Oct 05 08:18:13 crc kubenswrapper[4846]: I1005 08:18:13.642025 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-5b5cc989cf-btgr5" event={"ID":"d3b7b761-5203-4ed5-8fc1-4f4a3d572592","Type":"ContainerStarted","Data":"25f057ec53643b31370f0924e3915a99098aa8ef0fa63ddd6d16d9fa3f91627f"} Oct 05 08:18:13 crc kubenswrapper[4846]: I1005 08:18:13.642436 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-5b5cc989cf-btgr5" event={"ID":"d3b7b761-5203-4ed5-8fc1-4f4a3d572592","Type":"ContainerStarted","Data":"ba64f1f4663700cbfe588e30a3d6bbc3c842b345dfa05cdeae2d7cbf24b7daf4"} Oct 05 08:18:13 crc kubenswrapper[4846]: I1005 08:18:13.646589 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-vr9g5" event={"ID":"77e9ec78-ecc3-4953-9c33-f04b3f50b890","Type":"ContainerStarted","Data":"1c8ba7871632b08ebc5d6660c125985fea45dc4d01405f7dc5889ba80159c845"} Oct 05 08:18:13 crc kubenswrapper[4846]: I1005 08:18:13.650805 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-5bcb5dcc55-6mzb8" event={"ID":"9491b5bb-7012-4a5b-bc96-d96af06e9736","Type":"ContainerStarted","Data":"3f39f2f918e339b94d5b7884386b37466147e94ce36a3ef8771df8d42d434761"} Oct 05 08:18:13 crc kubenswrapper[4846]: I1005 08:18:13.664869 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-ring-rebalance-vr9g5" podStartSLOduration=2.352250622 podStartE2EDuration="6.664835287s" podCreationTimestamp="2025-10-05 08:18:07 +0000 UTC" firstStartedPulling="2025-10-05 08:18:08.372657013 +0000 UTC m=+5410.613509788" lastFinishedPulling="2025-10-05 08:18:12.685241678 +0000 UTC m=+5414.926094453" observedRunningTime="2025-10-05 08:18:13.661426606 +0000 UTC m=+5415.902279381" watchObservedRunningTime="2025-10-05 08:18:13.664835287 +0000 UTC m=+5415.905688062" Oct 05 08:18:14 crc kubenswrapper[4846]: I1005 08:18:14.662012 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-5b5cc989cf-btgr5" event={"ID":"d3b7b761-5203-4ed5-8fc1-4f4a3d572592","Type":"ContainerStarted","Data":"f99660f92ea9781976f819faa7bd60419cf7c43ade57de84c4a88638b6d4a69b"} Oct 05 08:18:14 crc kubenswrapper[4846]: I1005 08:18:14.662543 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-5b5cc989cf-btgr5" Oct 05 08:18:14 crc kubenswrapper[4846]: I1005 08:18:14.662575 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-5b5cc989cf-btgr5" Oct 05 08:18:14 crc kubenswrapper[4846]: I1005 08:18:14.663920 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-5bcb5dcc55-6mzb8" event={"ID":"9491b5bb-7012-4a5b-bc96-d96af06e9736","Type":"ContainerStarted","Data":"57eba68fcf6103a98a3b8d4e865d4201771105baa99ec9df5f04cfbc180444e7"} Oct 05 08:18:14 crc kubenswrapper[4846]: I1005 08:18:14.663965 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-5bcb5dcc55-6mzb8" event={"ID":"9491b5bb-7012-4a5b-bc96-d96af06e9736","Type":"ContainerStarted","Data":"da5acb620c15a5032b7e72e7eda04f4d941d2284dc3ab66ba148757051a39ccb"} Oct 05 08:18:14 crc kubenswrapper[4846]: I1005 08:18:14.693511 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-proxy-5b5cc989cf-btgr5" podStartSLOduration=3.693482559 podStartE2EDuration="3.693482559s" podCreationTimestamp="2025-10-05 08:18:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 08:18:14.687354725 +0000 UTC m=+5416.928207500" watchObservedRunningTime="2025-10-05 08:18:14.693482559 +0000 UTC m=+5416.934335334" Oct 05 08:18:14 crc kubenswrapper[4846]: I1005 08:18:14.717933 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-proxy-5bcb5dcc55-6mzb8" podStartSLOduration=5.717913283 podStartE2EDuration="5.717913283s" podCreationTimestamp="2025-10-05 08:18:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 08:18:14.705947153 +0000 UTC m=+5416.946799928" watchObservedRunningTime="2025-10-05 08:18:14.717913283 +0000 UTC m=+5416.958766058" Oct 05 08:18:15 crc kubenswrapper[4846]: I1005 08:18:15.129249 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-5bcb5dcc55-6mzb8" Oct 05 08:18:15 crc kubenswrapper[4846]: I1005 08:18:15.129309 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-5bcb5dcc55-6mzb8" Oct 05 08:18:17 crc kubenswrapper[4846]: I1005 08:18:17.708604 4846 generic.go:334] "Generic (PLEG): container finished" podID="77e9ec78-ecc3-4953-9c33-f04b3f50b890" containerID="1c8ba7871632b08ebc5d6660c125985fea45dc4d01405f7dc5889ba80159c845" exitCode=0 Oct 05 08:18:17 crc kubenswrapper[4846]: I1005 08:18:17.709102 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-vr9g5" event={"ID":"77e9ec78-ecc3-4953-9c33-f04b3f50b890","Type":"ContainerDied","Data":"1c8ba7871632b08ebc5d6660c125985fea45dc4d01405f7dc5889ba80159c845"} Oct 05 08:18:17 crc kubenswrapper[4846]: I1005 08:18:17.972421 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6688fc9d5f-7jvr8" Oct 05 08:18:18 crc kubenswrapper[4846]: I1005 08:18:18.064877 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-58bc4f8fb5-flnpk"] Oct 05 08:18:18 crc kubenswrapper[4846]: I1005 08:18:18.067710 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-58bc4f8fb5-flnpk" podUID="e21c80e9-984a-4882-8e73-ae3f272f75dd" containerName="dnsmasq-dns" containerID="cri-o://d785b2914ea934843acf56f4a3fd3d68e13dac02e609ece8c2104a5ac4ae9870" gracePeriod=10 Oct 05 08:18:18 crc kubenswrapper[4846]: I1005 08:18:18.589438 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-58bc4f8fb5-flnpk" Oct 05 08:18:18 crc kubenswrapper[4846]: I1005 08:18:18.721484 4846 generic.go:334] "Generic (PLEG): container finished" podID="e21c80e9-984a-4882-8e73-ae3f272f75dd" containerID="d785b2914ea934843acf56f4a3fd3d68e13dac02e609ece8c2104a5ac4ae9870" exitCode=0 Oct 05 08:18:18 crc kubenswrapper[4846]: I1005 08:18:18.721750 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-58bc4f8fb5-flnpk" Oct 05 08:18:18 crc kubenswrapper[4846]: I1005 08:18:18.721668 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-58bc4f8fb5-flnpk" event={"ID":"e21c80e9-984a-4882-8e73-ae3f272f75dd","Type":"ContainerDied","Data":"d785b2914ea934843acf56f4a3fd3d68e13dac02e609ece8c2104a5ac4ae9870"} Oct 05 08:18:18 crc kubenswrapper[4846]: I1005 08:18:18.721823 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-58bc4f8fb5-flnpk" event={"ID":"e21c80e9-984a-4882-8e73-ae3f272f75dd","Type":"ContainerDied","Data":"1595a606a41a1cf9bca5a0ff79544a45146e0525bb5ff708691877db8ef40cec"} Oct 05 08:18:18 crc kubenswrapper[4846]: I1005 08:18:18.721846 4846 scope.go:117] "RemoveContainer" containerID="d785b2914ea934843acf56f4a3fd3d68e13dac02e609ece8c2104a5ac4ae9870" Oct 05 08:18:18 crc kubenswrapper[4846]: I1005 08:18:18.749301 4846 scope.go:117] "RemoveContainer" containerID="7e28abfd4128d4c7447893a5f05c5c758471f760f57f749ffff8a4ce0dbf6cf8" Oct 05 08:18:18 crc kubenswrapper[4846]: I1005 08:18:18.753634 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e21c80e9-984a-4882-8e73-ae3f272f75dd-dns-svc\") pod \"e21c80e9-984a-4882-8e73-ae3f272f75dd\" (UID: \"e21c80e9-984a-4882-8e73-ae3f272f75dd\") " Oct 05 08:18:18 crc kubenswrapper[4846]: I1005 08:18:18.753678 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e21c80e9-984a-4882-8e73-ae3f272f75dd-ovsdbserver-nb\") pod \"e21c80e9-984a-4882-8e73-ae3f272f75dd\" (UID: \"e21c80e9-984a-4882-8e73-ae3f272f75dd\") " Oct 05 08:18:18 crc kubenswrapper[4846]: I1005 08:18:18.753816 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e21c80e9-984a-4882-8e73-ae3f272f75dd-ovsdbserver-sb\") pod \"e21c80e9-984a-4882-8e73-ae3f272f75dd\" (UID: \"e21c80e9-984a-4882-8e73-ae3f272f75dd\") " Oct 05 08:18:18 crc kubenswrapper[4846]: I1005 08:18:18.753922 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e21c80e9-984a-4882-8e73-ae3f272f75dd-config\") pod \"e21c80e9-984a-4882-8e73-ae3f272f75dd\" (UID: \"e21c80e9-984a-4882-8e73-ae3f272f75dd\") " Oct 05 08:18:18 crc kubenswrapper[4846]: I1005 08:18:18.754030 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nh2g2\" (UniqueName: \"kubernetes.io/projected/e21c80e9-984a-4882-8e73-ae3f272f75dd-kube-api-access-nh2g2\") pod \"e21c80e9-984a-4882-8e73-ae3f272f75dd\" (UID: \"e21c80e9-984a-4882-8e73-ae3f272f75dd\") " Oct 05 08:18:18 crc kubenswrapper[4846]: I1005 08:18:18.759436 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e21c80e9-984a-4882-8e73-ae3f272f75dd-kube-api-access-nh2g2" (OuterVolumeSpecName: "kube-api-access-nh2g2") pod "e21c80e9-984a-4882-8e73-ae3f272f75dd" (UID: "e21c80e9-984a-4882-8e73-ae3f272f75dd"). InnerVolumeSpecName "kube-api-access-nh2g2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:18:18 crc kubenswrapper[4846]: I1005 08:18:18.814635 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e21c80e9-984a-4882-8e73-ae3f272f75dd-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "e21c80e9-984a-4882-8e73-ae3f272f75dd" (UID: "e21c80e9-984a-4882-8e73-ae3f272f75dd"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 08:18:18 crc kubenswrapper[4846]: I1005 08:18:18.814801 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e21c80e9-984a-4882-8e73-ae3f272f75dd-config" (OuterVolumeSpecName: "config") pod "e21c80e9-984a-4882-8e73-ae3f272f75dd" (UID: "e21c80e9-984a-4882-8e73-ae3f272f75dd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 08:18:18 crc kubenswrapper[4846]: I1005 08:18:18.820731 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e21c80e9-984a-4882-8e73-ae3f272f75dd-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "e21c80e9-984a-4882-8e73-ae3f272f75dd" (UID: "e21c80e9-984a-4882-8e73-ae3f272f75dd"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 08:18:18 crc kubenswrapper[4846]: I1005 08:18:18.843315 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e21c80e9-984a-4882-8e73-ae3f272f75dd-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "e21c80e9-984a-4882-8e73-ae3f272f75dd" (UID: "e21c80e9-984a-4882-8e73-ae3f272f75dd"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 08:18:18 crc kubenswrapper[4846]: I1005 08:18:18.856030 4846 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e21c80e9-984a-4882-8e73-ae3f272f75dd-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 05 08:18:18 crc kubenswrapper[4846]: I1005 08:18:18.856063 4846 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e21c80e9-984a-4882-8e73-ae3f272f75dd-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 05 08:18:18 crc kubenswrapper[4846]: I1005 08:18:18.856075 4846 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e21c80e9-984a-4882-8e73-ae3f272f75dd-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 05 08:18:18 crc kubenswrapper[4846]: I1005 08:18:18.856084 4846 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e21c80e9-984a-4882-8e73-ae3f272f75dd-config\") on node \"crc\" DevicePath \"\"" Oct 05 08:18:18 crc kubenswrapper[4846]: I1005 08:18:18.856094 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nh2g2\" (UniqueName: \"kubernetes.io/projected/e21c80e9-984a-4882-8e73-ae3f272f75dd-kube-api-access-nh2g2\") on node \"crc\" DevicePath \"\"" Oct 05 08:18:18 crc kubenswrapper[4846]: I1005 08:18:18.866679 4846 scope.go:117] "RemoveContainer" containerID="d785b2914ea934843acf56f4a3fd3d68e13dac02e609ece8c2104a5ac4ae9870" Oct 05 08:18:18 crc kubenswrapper[4846]: E1005 08:18:18.866983 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d785b2914ea934843acf56f4a3fd3d68e13dac02e609ece8c2104a5ac4ae9870\": container with ID starting with d785b2914ea934843acf56f4a3fd3d68e13dac02e609ece8c2104a5ac4ae9870 not found: ID does not exist" containerID="d785b2914ea934843acf56f4a3fd3d68e13dac02e609ece8c2104a5ac4ae9870" Oct 05 08:18:18 crc kubenswrapper[4846]: I1005 08:18:18.867011 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d785b2914ea934843acf56f4a3fd3d68e13dac02e609ece8c2104a5ac4ae9870"} err="failed to get container status \"d785b2914ea934843acf56f4a3fd3d68e13dac02e609ece8c2104a5ac4ae9870\": rpc error: code = NotFound desc = could not find container \"d785b2914ea934843acf56f4a3fd3d68e13dac02e609ece8c2104a5ac4ae9870\": container with ID starting with d785b2914ea934843acf56f4a3fd3d68e13dac02e609ece8c2104a5ac4ae9870 not found: ID does not exist" Oct 05 08:18:18 crc kubenswrapper[4846]: I1005 08:18:18.867039 4846 scope.go:117] "RemoveContainer" containerID="7e28abfd4128d4c7447893a5f05c5c758471f760f57f749ffff8a4ce0dbf6cf8" Oct 05 08:18:18 crc kubenswrapper[4846]: E1005 08:18:18.867293 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7e28abfd4128d4c7447893a5f05c5c758471f760f57f749ffff8a4ce0dbf6cf8\": container with ID starting with 7e28abfd4128d4c7447893a5f05c5c758471f760f57f749ffff8a4ce0dbf6cf8 not found: ID does not exist" containerID="7e28abfd4128d4c7447893a5f05c5c758471f760f57f749ffff8a4ce0dbf6cf8" Oct 05 08:18:18 crc kubenswrapper[4846]: I1005 08:18:18.867313 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7e28abfd4128d4c7447893a5f05c5c758471f760f57f749ffff8a4ce0dbf6cf8"} err="failed to get container status \"7e28abfd4128d4c7447893a5f05c5c758471f760f57f749ffff8a4ce0dbf6cf8\": rpc error: code = NotFound desc = could not find container \"7e28abfd4128d4c7447893a5f05c5c758471f760f57f749ffff8a4ce0dbf6cf8\": container with ID starting with 7e28abfd4128d4c7447893a5f05c5c758471f760f57f749ffff8a4ce0dbf6cf8 not found: ID does not exist" Oct 05 08:18:19 crc kubenswrapper[4846]: I1005 08:18:19.026682 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-vr9g5" Oct 05 08:18:19 crc kubenswrapper[4846]: I1005 08:18:19.067547 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-58bc4f8fb5-flnpk"] Oct 05 08:18:19 crc kubenswrapper[4846]: I1005 08:18:19.073499 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-58bc4f8fb5-flnpk"] Oct 05 08:18:19 crc kubenswrapper[4846]: I1005 08:18:19.160500 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/77e9ec78-ecc3-4953-9c33-f04b3f50b890-etc-swift\") pod \"77e9ec78-ecc3-4953-9c33-f04b3f50b890\" (UID: \"77e9ec78-ecc3-4953-9c33-f04b3f50b890\") " Oct 05 08:18:19 crc kubenswrapper[4846]: I1005 08:18:19.160558 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-svt7n\" (UniqueName: \"kubernetes.io/projected/77e9ec78-ecc3-4953-9c33-f04b3f50b890-kube-api-access-svt7n\") pod \"77e9ec78-ecc3-4953-9c33-f04b3f50b890\" (UID: \"77e9ec78-ecc3-4953-9c33-f04b3f50b890\") " Oct 05 08:18:19 crc kubenswrapper[4846]: I1005 08:18:19.160581 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/77e9ec78-ecc3-4953-9c33-f04b3f50b890-dispersionconf\") pod \"77e9ec78-ecc3-4953-9c33-f04b3f50b890\" (UID: \"77e9ec78-ecc3-4953-9c33-f04b3f50b890\") " Oct 05 08:18:19 crc kubenswrapper[4846]: I1005 08:18:19.160630 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/77e9ec78-ecc3-4953-9c33-f04b3f50b890-ring-data-devices\") pod \"77e9ec78-ecc3-4953-9c33-f04b3f50b890\" (UID: \"77e9ec78-ecc3-4953-9c33-f04b3f50b890\") " Oct 05 08:18:19 crc kubenswrapper[4846]: I1005 08:18:19.160732 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/77e9ec78-ecc3-4953-9c33-f04b3f50b890-swiftconf\") pod \"77e9ec78-ecc3-4953-9c33-f04b3f50b890\" (UID: \"77e9ec78-ecc3-4953-9c33-f04b3f50b890\") " Oct 05 08:18:19 crc kubenswrapper[4846]: I1005 08:18:19.160761 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/77e9ec78-ecc3-4953-9c33-f04b3f50b890-scripts\") pod \"77e9ec78-ecc3-4953-9c33-f04b3f50b890\" (UID: \"77e9ec78-ecc3-4953-9c33-f04b3f50b890\") " Oct 05 08:18:19 crc kubenswrapper[4846]: I1005 08:18:19.160811 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/77e9ec78-ecc3-4953-9c33-f04b3f50b890-combined-ca-bundle\") pod \"77e9ec78-ecc3-4953-9c33-f04b3f50b890\" (UID: \"77e9ec78-ecc3-4953-9c33-f04b3f50b890\") " Oct 05 08:18:19 crc kubenswrapper[4846]: I1005 08:18:19.161220 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/77e9ec78-ecc3-4953-9c33-f04b3f50b890-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "77e9ec78-ecc3-4953-9c33-f04b3f50b890" (UID: "77e9ec78-ecc3-4953-9c33-f04b3f50b890"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 08:18:19 crc kubenswrapper[4846]: I1005 08:18:19.161598 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/77e9ec78-ecc3-4953-9c33-f04b3f50b890-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "77e9ec78-ecc3-4953-9c33-f04b3f50b890" (UID: "77e9ec78-ecc3-4953-9c33-f04b3f50b890"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 08:18:19 crc kubenswrapper[4846]: I1005 08:18:19.171688 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/77e9ec78-ecc3-4953-9c33-f04b3f50b890-kube-api-access-svt7n" (OuterVolumeSpecName: "kube-api-access-svt7n") pod "77e9ec78-ecc3-4953-9c33-f04b3f50b890" (UID: "77e9ec78-ecc3-4953-9c33-f04b3f50b890"). InnerVolumeSpecName "kube-api-access-svt7n". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:18:19 crc kubenswrapper[4846]: I1005 08:18:19.174719 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/77e9ec78-ecc3-4953-9c33-f04b3f50b890-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "77e9ec78-ecc3-4953-9c33-f04b3f50b890" (UID: "77e9ec78-ecc3-4953-9c33-f04b3f50b890"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:18:19 crc kubenswrapper[4846]: I1005 08:18:19.184265 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/77e9ec78-ecc3-4953-9c33-f04b3f50b890-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "77e9ec78-ecc3-4953-9c33-f04b3f50b890" (UID: "77e9ec78-ecc3-4953-9c33-f04b3f50b890"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:18:19 crc kubenswrapper[4846]: I1005 08:18:19.191831 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/77e9ec78-ecc3-4953-9c33-f04b3f50b890-scripts" (OuterVolumeSpecName: "scripts") pod "77e9ec78-ecc3-4953-9c33-f04b3f50b890" (UID: "77e9ec78-ecc3-4953-9c33-f04b3f50b890"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 08:18:19 crc kubenswrapper[4846]: I1005 08:18:19.195485 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/77e9ec78-ecc3-4953-9c33-f04b3f50b890-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "77e9ec78-ecc3-4953-9c33-f04b3f50b890" (UID: "77e9ec78-ecc3-4953-9c33-f04b3f50b890"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:18:19 crc kubenswrapper[4846]: I1005 08:18:19.262998 4846 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/77e9ec78-ecc3-4953-9c33-f04b3f50b890-swiftconf\") on node \"crc\" DevicePath \"\"" Oct 05 08:18:19 crc kubenswrapper[4846]: I1005 08:18:19.263283 4846 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/77e9ec78-ecc3-4953-9c33-f04b3f50b890-scripts\") on node \"crc\" DevicePath \"\"" Oct 05 08:18:19 crc kubenswrapper[4846]: I1005 08:18:19.263293 4846 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/77e9ec78-ecc3-4953-9c33-f04b3f50b890-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 08:18:19 crc kubenswrapper[4846]: I1005 08:18:19.263301 4846 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/77e9ec78-ecc3-4953-9c33-f04b3f50b890-etc-swift\") on node \"crc\" DevicePath \"\"" Oct 05 08:18:19 crc kubenswrapper[4846]: I1005 08:18:19.263311 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-svt7n\" (UniqueName: \"kubernetes.io/projected/77e9ec78-ecc3-4953-9c33-f04b3f50b890-kube-api-access-svt7n\") on node \"crc\" DevicePath \"\"" Oct 05 08:18:19 crc kubenswrapper[4846]: I1005 08:18:19.263319 4846 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/77e9ec78-ecc3-4953-9c33-f04b3f50b890-dispersionconf\") on node \"crc\" DevicePath \"\"" Oct 05 08:18:19 crc kubenswrapper[4846]: I1005 08:18:19.263327 4846 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/77e9ec78-ecc3-4953-9c33-f04b3f50b890-ring-data-devices\") on node \"crc\" DevicePath \"\"" Oct 05 08:18:19 crc kubenswrapper[4846]: I1005 08:18:19.732582 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-vr9g5" event={"ID":"77e9ec78-ecc3-4953-9c33-f04b3f50b890","Type":"ContainerDied","Data":"cb98540a2bd7bac69405e36967d82af679509c808c7cb938860dc8dab8fe3169"} Oct 05 08:18:19 crc kubenswrapper[4846]: I1005 08:18:19.732615 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-vr9g5" Oct 05 08:18:19 crc kubenswrapper[4846]: I1005 08:18:19.732626 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cb98540a2bd7bac69405e36967d82af679509c808c7cb938860dc8dab8fe3169" Oct 05 08:18:20 crc kubenswrapper[4846]: I1005 08:18:20.132485 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-5bcb5dcc55-6mzb8" Oct 05 08:18:20 crc kubenswrapper[4846]: I1005 08:18:20.133515 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-5bcb5dcc55-6mzb8" Oct 05 08:18:20 crc kubenswrapper[4846]: I1005 08:18:20.508905 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e21c80e9-984a-4882-8e73-ae3f272f75dd" path="/var/lib/kubelet/pods/e21c80e9-984a-4882-8e73-ae3f272f75dd/volumes" Oct 05 08:18:22 crc kubenswrapper[4846]: I1005 08:18:22.357833 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-5b5cc989cf-btgr5" Oct 05 08:18:22 crc kubenswrapper[4846]: I1005 08:18:22.358918 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-5b5cc989cf-btgr5" Oct 05 08:18:22 crc kubenswrapper[4846]: I1005 08:18:22.487866 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-proxy-5bcb5dcc55-6mzb8"] Oct 05 08:18:22 crc kubenswrapper[4846]: I1005 08:18:22.488149 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-proxy-5bcb5dcc55-6mzb8" podUID="9491b5bb-7012-4a5b-bc96-d96af06e9736" containerName="proxy-httpd" containerID="cri-o://57eba68fcf6103a98a3b8d4e865d4201771105baa99ec9df5f04cfbc180444e7" gracePeriod=30 Oct 05 08:18:22 crc kubenswrapper[4846]: I1005 08:18:22.488572 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-proxy-5bcb5dcc55-6mzb8" podUID="9491b5bb-7012-4a5b-bc96-d96af06e9736" containerName="proxy-server" containerID="cri-o://da5acb620c15a5032b7e72e7eda04f4d941d2284dc3ab66ba148757051a39ccb" gracePeriod=30 Oct 05 08:18:22 crc kubenswrapper[4846]: I1005 08:18:22.770378 4846 generic.go:334] "Generic (PLEG): container finished" podID="9491b5bb-7012-4a5b-bc96-d96af06e9736" containerID="57eba68fcf6103a98a3b8d4e865d4201771105baa99ec9df5f04cfbc180444e7" exitCode=0 Oct 05 08:18:22 crc kubenswrapper[4846]: I1005 08:18:22.770572 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-5bcb5dcc55-6mzb8" event={"ID":"9491b5bb-7012-4a5b-bc96-d96af06e9736","Type":"ContainerDied","Data":"57eba68fcf6103a98a3b8d4e865d4201771105baa99ec9df5f04cfbc180444e7"} Oct 05 08:18:23 crc kubenswrapper[4846]: I1005 08:18:23.325059 4846 patch_prober.go:28] interesting pod/machine-config-daemon-fscvf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 08:18:23 crc kubenswrapper[4846]: I1005 08:18:23.325432 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 08:18:23 crc kubenswrapper[4846]: I1005 08:18:23.546295 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-5bcb5dcc55-6mzb8" Oct 05 08:18:23 crc kubenswrapper[4846]: I1005 08:18:23.645618 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9491b5bb-7012-4a5b-bc96-d96af06e9736-run-httpd\") pod \"9491b5bb-7012-4a5b-bc96-d96af06e9736\" (UID: \"9491b5bb-7012-4a5b-bc96-d96af06e9736\") " Oct 05 08:18:23 crc kubenswrapper[4846]: I1005 08:18:23.645698 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/9491b5bb-7012-4a5b-bc96-d96af06e9736-etc-swift\") pod \"9491b5bb-7012-4a5b-bc96-d96af06e9736\" (UID: \"9491b5bb-7012-4a5b-bc96-d96af06e9736\") " Oct 05 08:18:23 crc kubenswrapper[4846]: I1005 08:18:23.645793 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9491b5bb-7012-4a5b-bc96-d96af06e9736-combined-ca-bundle\") pod \"9491b5bb-7012-4a5b-bc96-d96af06e9736\" (UID: \"9491b5bb-7012-4a5b-bc96-d96af06e9736\") " Oct 05 08:18:23 crc kubenswrapper[4846]: I1005 08:18:23.645821 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9491b5bb-7012-4a5b-bc96-d96af06e9736-log-httpd\") pod \"9491b5bb-7012-4a5b-bc96-d96af06e9736\" (UID: \"9491b5bb-7012-4a5b-bc96-d96af06e9736\") " Oct 05 08:18:23 crc kubenswrapper[4846]: I1005 08:18:23.645960 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9491b5bb-7012-4a5b-bc96-d96af06e9736-config-data\") pod \"9491b5bb-7012-4a5b-bc96-d96af06e9736\" (UID: \"9491b5bb-7012-4a5b-bc96-d96af06e9736\") " Oct 05 08:18:23 crc kubenswrapper[4846]: I1005 08:18:23.645987 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d558p\" (UniqueName: \"kubernetes.io/projected/9491b5bb-7012-4a5b-bc96-d96af06e9736-kube-api-access-d558p\") pod \"9491b5bb-7012-4a5b-bc96-d96af06e9736\" (UID: \"9491b5bb-7012-4a5b-bc96-d96af06e9736\") " Oct 05 08:18:23 crc kubenswrapper[4846]: I1005 08:18:23.646130 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9491b5bb-7012-4a5b-bc96-d96af06e9736-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "9491b5bb-7012-4a5b-bc96-d96af06e9736" (UID: "9491b5bb-7012-4a5b-bc96-d96af06e9736"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 08:18:23 crc kubenswrapper[4846]: I1005 08:18:23.646236 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9491b5bb-7012-4a5b-bc96-d96af06e9736-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "9491b5bb-7012-4a5b-bc96-d96af06e9736" (UID: "9491b5bb-7012-4a5b-bc96-d96af06e9736"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 08:18:23 crc kubenswrapper[4846]: I1005 08:18:23.647154 4846 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9491b5bb-7012-4a5b-bc96-d96af06e9736-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 05 08:18:23 crc kubenswrapper[4846]: I1005 08:18:23.647267 4846 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9491b5bb-7012-4a5b-bc96-d96af06e9736-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 05 08:18:23 crc kubenswrapper[4846]: I1005 08:18:23.660417 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9491b5bb-7012-4a5b-bc96-d96af06e9736-kube-api-access-d558p" (OuterVolumeSpecName: "kube-api-access-d558p") pod "9491b5bb-7012-4a5b-bc96-d96af06e9736" (UID: "9491b5bb-7012-4a5b-bc96-d96af06e9736"). InnerVolumeSpecName "kube-api-access-d558p". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:18:23 crc kubenswrapper[4846]: I1005 08:18:23.662331 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9491b5bb-7012-4a5b-bc96-d96af06e9736-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "9491b5bb-7012-4a5b-bc96-d96af06e9736" (UID: "9491b5bb-7012-4a5b-bc96-d96af06e9736"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:18:23 crc kubenswrapper[4846]: I1005 08:18:23.719202 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9491b5bb-7012-4a5b-bc96-d96af06e9736-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9491b5bb-7012-4a5b-bc96-d96af06e9736" (UID: "9491b5bb-7012-4a5b-bc96-d96af06e9736"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:18:23 crc kubenswrapper[4846]: I1005 08:18:23.721326 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9491b5bb-7012-4a5b-bc96-d96af06e9736-config-data" (OuterVolumeSpecName: "config-data") pod "9491b5bb-7012-4a5b-bc96-d96af06e9736" (UID: "9491b5bb-7012-4a5b-bc96-d96af06e9736"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:18:23 crc kubenswrapper[4846]: I1005 08:18:23.748832 4846 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/9491b5bb-7012-4a5b-bc96-d96af06e9736-etc-swift\") on node \"crc\" DevicePath \"\"" Oct 05 08:18:23 crc kubenswrapper[4846]: I1005 08:18:23.748856 4846 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9491b5bb-7012-4a5b-bc96-d96af06e9736-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 08:18:23 crc kubenswrapper[4846]: I1005 08:18:23.748865 4846 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9491b5bb-7012-4a5b-bc96-d96af06e9736-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 08:18:23 crc kubenswrapper[4846]: I1005 08:18:23.748873 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d558p\" (UniqueName: \"kubernetes.io/projected/9491b5bb-7012-4a5b-bc96-d96af06e9736-kube-api-access-d558p\") on node \"crc\" DevicePath \"\"" Oct 05 08:18:23 crc kubenswrapper[4846]: I1005 08:18:23.779622 4846 generic.go:334] "Generic (PLEG): container finished" podID="9491b5bb-7012-4a5b-bc96-d96af06e9736" containerID="da5acb620c15a5032b7e72e7eda04f4d941d2284dc3ab66ba148757051a39ccb" exitCode=0 Oct 05 08:18:23 crc kubenswrapper[4846]: I1005 08:18:23.779661 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-5bcb5dcc55-6mzb8" event={"ID":"9491b5bb-7012-4a5b-bc96-d96af06e9736","Type":"ContainerDied","Data":"da5acb620c15a5032b7e72e7eda04f4d941d2284dc3ab66ba148757051a39ccb"} Oct 05 08:18:23 crc kubenswrapper[4846]: I1005 08:18:23.779672 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-5bcb5dcc55-6mzb8" Oct 05 08:18:23 crc kubenswrapper[4846]: I1005 08:18:23.779687 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-5bcb5dcc55-6mzb8" event={"ID":"9491b5bb-7012-4a5b-bc96-d96af06e9736","Type":"ContainerDied","Data":"3f39f2f918e339b94d5b7884386b37466147e94ce36a3ef8771df8d42d434761"} Oct 05 08:18:23 crc kubenswrapper[4846]: I1005 08:18:23.779702 4846 scope.go:117] "RemoveContainer" containerID="da5acb620c15a5032b7e72e7eda04f4d941d2284dc3ab66ba148757051a39ccb" Oct 05 08:18:23 crc kubenswrapper[4846]: I1005 08:18:23.797043 4846 scope.go:117] "RemoveContainer" containerID="57eba68fcf6103a98a3b8d4e865d4201771105baa99ec9df5f04cfbc180444e7" Oct 05 08:18:23 crc kubenswrapper[4846]: I1005 08:18:23.820256 4846 scope.go:117] "RemoveContainer" containerID="da5acb620c15a5032b7e72e7eda04f4d941d2284dc3ab66ba148757051a39ccb" Oct 05 08:18:23 crc kubenswrapper[4846]: E1005 08:18:23.820765 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"da5acb620c15a5032b7e72e7eda04f4d941d2284dc3ab66ba148757051a39ccb\": container with ID starting with da5acb620c15a5032b7e72e7eda04f4d941d2284dc3ab66ba148757051a39ccb not found: ID does not exist" containerID="da5acb620c15a5032b7e72e7eda04f4d941d2284dc3ab66ba148757051a39ccb" Oct 05 08:18:23 crc kubenswrapper[4846]: I1005 08:18:23.820797 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"da5acb620c15a5032b7e72e7eda04f4d941d2284dc3ab66ba148757051a39ccb"} err="failed to get container status \"da5acb620c15a5032b7e72e7eda04f4d941d2284dc3ab66ba148757051a39ccb\": rpc error: code = NotFound desc = could not find container \"da5acb620c15a5032b7e72e7eda04f4d941d2284dc3ab66ba148757051a39ccb\": container with ID starting with da5acb620c15a5032b7e72e7eda04f4d941d2284dc3ab66ba148757051a39ccb not found: ID does not exist" Oct 05 08:18:23 crc kubenswrapper[4846]: I1005 08:18:23.820845 4846 scope.go:117] "RemoveContainer" containerID="57eba68fcf6103a98a3b8d4e865d4201771105baa99ec9df5f04cfbc180444e7" Oct 05 08:18:23 crc kubenswrapper[4846]: E1005 08:18:23.821266 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"57eba68fcf6103a98a3b8d4e865d4201771105baa99ec9df5f04cfbc180444e7\": container with ID starting with 57eba68fcf6103a98a3b8d4e865d4201771105baa99ec9df5f04cfbc180444e7 not found: ID does not exist" containerID="57eba68fcf6103a98a3b8d4e865d4201771105baa99ec9df5f04cfbc180444e7" Oct 05 08:18:23 crc kubenswrapper[4846]: I1005 08:18:23.821311 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"57eba68fcf6103a98a3b8d4e865d4201771105baa99ec9df5f04cfbc180444e7"} err="failed to get container status \"57eba68fcf6103a98a3b8d4e865d4201771105baa99ec9df5f04cfbc180444e7\": rpc error: code = NotFound desc = could not find container \"57eba68fcf6103a98a3b8d4e865d4201771105baa99ec9df5f04cfbc180444e7\": container with ID starting with 57eba68fcf6103a98a3b8d4e865d4201771105baa99ec9df5f04cfbc180444e7 not found: ID does not exist" Oct 05 08:18:23 crc kubenswrapper[4846]: I1005 08:18:23.821574 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-proxy-5bcb5dcc55-6mzb8"] Oct 05 08:18:23 crc kubenswrapper[4846]: I1005 08:18:23.827695 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/swift-proxy-5bcb5dcc55-6mzb8"] Oct 05 08:18:24 crc kubenswrapper[4846]: I1005 08:18:24.509899 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9491b5bb-7012-4a5b-bc96-d96af06e9736" path="/var/lib/kubelet/pods/9491b5bb-7012-4a5b-bc96-d96af06e9736/volumes" Oct 05 08:18:31 crc kubenswrapper[4846]: I1005 08:18:31.968751 4846 scope.go:117] "RemoveContainer" containerID="83d22853a3b1f66f4f6d0f405a062e2172344cab88ae034b26b2f8a384bc74b0" Oct 05 08:18:32 crc kubenswrapper[4846]: I1005 08:18:32.014623 4846 scope.go:117] "RemoveContainer" containerID="908006bdcb1c13dd53d795ae9d9c6e3a93be46060c089674ede060bc0113e262" Oct 05 08:18:53 crc kubenswrapper[4846]: I1005 08:18:53.325259 4846 patch_prober.go:28] interesting pod/machine-config-daemon-fscvf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 08:18:53 crc kubenswrapper[4846]: I1005 08:18:53.325923 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 08:18:53 crc kubenswrapper[4846]: I1005 08:18:53.325983 4846 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" Oct 05 08:18:53 crc kubenswrapper[4846]: I1005 08:18:53.327943 4846 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"f8ac1a187331f3fff703acc696060a82d728f8e154fe3e11ba6c956d3d906397"} pod="openshift-machine-config-operator/machine-config-daemon-fscvf" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 05 08:18:53 crc kubenswrapper[4846]: I1005 08:18:53.328031 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" containerName="machine-config-daemon" containerID="cri-o://f8ac1a187331f3fff703acc696060a82d728f8e154fe3e11ba6c956d3d906397" gracePeriod=600 Oct 05 08:18:54 crc kubenswrapper[4846]: I1005 08:18:54.141232 4846 generic.go:334] "Generic (PLEG): container finished" podID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" containerID="f8ac1a187331f3fff703acc696060a82d728f8e154fe3e11ba6c956d3d906397" exitCode=0 Oct 05 08:18:54 crc kubenswrapper[4846]: I1005 08:18:54.141306 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" event={"ID":"ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7","Type":"ContainerDied","Data":"f8ac1a187331f3fff703acc696060a82d728f8e154fe3e11ba6c956d3d906397"} Oct 05 08:18:54 crc kubenswrapper[4846]: I1005 08:18:54.141858 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" event={"ID":"ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7","Type":"ContainerStarted","Data":"fdafa27da24296f9087c33b6b12a6598d2a8d61fc985b06ad680401828c9feb6"} Oct 05 08:18:54 crc kubenswrapper[4846]: I1005 08:18:54.141889 4846 scope.go:117] "RemoveContainer" containerID="8fa72be735dff8e623a4a4958f452896ef79e72dd44c2d714a3019187462de47" Oct 05 08:18:55 crc kubenswrapper[4846]: I1005 08:18:55.666325 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-create-4ccwk"] Oct 05 08:18:55 crc kubenswrapper[4846]: E1005 08:18:55.667079 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e21c80e9-984a-4882-8e73-ae3f272f75dd" containerName="dnsmasq-dns" Oct 05 08:18:55 crc kubenswrapper[4846]: I1005 08:18:55.667098 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="e21c80e9-984a-4882-8e73-ae3f272f75dd" containerName="dnsmasq-dns" Oct 05 08:18:55 crc kubenswrapper[4846]: E1005 08:18:55.667115 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9491b5bb-7012-4a5b-bc96-d96af06e9736" containerName="proxy-httpd" Oct 05 08:18:55 crc kubenswrapper[4846]: I1005 08:18:55.667124 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="9491b5bb-7012-4a5b-bc96-d96af06e9736" containerName="proxy-httpd" Oct 05 08:18:55 crc kubenswrapper[4846]: E1005 08:18:55.667159 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9491b5bb-7012-4a5b-bc96-d96af06e9736" containerName="proxy-server" Oct 05 08:18:55 crc kubenswrapper[4846]: I1005 08:18:55.667167 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="9491b5bb-7012-4a5b-bc96-d96af06e9736" containerName="proxy-server" Oct 05 08:18:55 crc kubenswrapper[4846]: E1005 08:18:55.667209 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="77e9ec78-ecc3-4953-9c33-f04b3f50b890" containerName="swift-ring-rebalance" Oct 05 08:18:55 crc kubenswrapper[4846]: I1005 08:18:55.667217 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="77e9ec78-ecc3-4953-9c33-f04b3f50b890" containerName="swift-ring-rebalance" Oct 05 08:18:55 crc kubenswrapper[4846]: E1005 08:18:55.667234 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e21c80e9-984a-4882-8e73-ae3f272f75dd" containerName="init" Oct 05 08:18:55 crc kubenswrapper[4846]: I1005 08:18:55.667241 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="e21c80e9-984a-4882-8e73-ae3f272f75dd" containerName="init" Oct 05 08:18:55 crc kubenswrapper[4846]: I1005 08:18:55.667453 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="e21c80e9-984a-4882-8e73-ae3f272f75dd" containerName="dnsmasq-dns" Oct 05 08:18:55 crc kubenswrapper[4846]: I1005 08:18:55.667469 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="77e9ec78-ecc3-4953-9c33-f04b3f50b890" containerName="swift-ring-rebalance" Oct 05 08:18:55 crc kubenswrapper[4846]: I1005 08:18:55.667498 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="9491b5bb-7012-4a5b-bc96-d96af06e9736" containerName="proxy-server" Oct 05 08:18:55 crc kubenswrapper[4846]: I1005 08:18:55.667508 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="9491b5bb-7012-4a5b-bc96-d96af06e9736" containerName="proxy-httpd" Oct 05 08:18:55 crc kubenswrapper[4846]: I1005 08:18:55.668252 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-4ccwk" Oct 05 08:18:55 crc kubenswrapper[4846]: I1005 08:18:55.674807 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-4ccwk"] Oct 05 08:18:55 crc kubenswrapper[4846]: I1005 08:18:55.747470 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vf5w6\" (UniqueName: \"kubernetes.io/projected/1bf32b0d-81b1-4774-bf03-995ba2dcfe2d-kube-api-access-vf5w6\") pod \"cinder-db-create-4ccwk\" (UID: \"1bf32b0d-81b1-4774-bf03-995ba2dcfe2d\") " pod="openstack/cinder-db-create-4ccwk" Oct 05 08:18:55 crc kubenswrapper[4846]: I1005 08:18:55.849362 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vf5w6\" (UniqueName: \"kubernetes.io/projected/1bf32b0d-81b1-4774-bf03-995ba2dcfe2d-kube-api-access-vf5w6\") pod \"cinder-db-create-4ccwk\" (UID: \"1bf32b0d-81b1-4774-bf03-995ba2dcfe2d\") " pod="openstack/cinder-db-create-4ccwk" Oct 05 08:18:55 crc kubenswrapper[4846]: I1005 08:18:55.872827 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vf5w6\" (UniqueName: \"kubernetes.io/projected/1bf32b0d-81b1-4774-bf03-995ba2dcfe2d-kube-api-access-vf5w6\") pod \"cinder-db-create-4ccwk\" (UID: \"1bf32b0d-81b1-4774-bf03-995ba2dcfe2d\") " pod="openstack/cinder-db-create-4ccwk" Oct 05 08:18:55 crc kubenswrapper[4846]: I1005 08:18:55.992864 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-4ccwk" Oct 05 08:18:56 crc kubenswrapper[4846]: W1005 08:18:56.464360 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1bf32b0d_81b1_4774_bf03_995ba2dcfe2d.slice/crio-ea03e5e59803b75112ad03cdb0433fa6a18fdcb360654b27927acd7b6fd2f5aa WatchSource:0}: Error finding container ea03e5e59803b75112ad03cdb0433fa6a18fdcb360654b27927acd7b6fd2f5aa: Status 404 returned error can't find the container with id ea03e5e59803b75112ad03cdb0433fa6a18fdcb360654b27927acd7b6fd2f5aa Oct 05 08:18:56 crc kubenswrapper[4846]: I1005 08:18:56.464420 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-4ccwk"] Oct 05 08:18:57 crc kubenswrapper[4846]: I1005 08:18:57.177959 4846 generic.go:334] "Generic (PLEG): container finished" podID="1bf32b0d-81b1-4774-bf03-995ba2dcfe2d" containerID="0bfea0204492bee3feb5c3ffb076b60bd6f751b3c236782fcfad5d9ec9d3c52f" exitCode=0 Oct 05 08:18:57 crc kubenswrapper[4846]: I1005 08:18:57.178019 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-4ccwk" event={"ID":"1bf32b0d-81b1-4774-bf03-995ba2dcfe2d","Type":"ContainerDied","Data":"0bfea0204492bee3feb5c3ffb076b60bd6f751b3c236782fcfad5d9ec9d3c52f"} Oct 05 08:18:57 crc kubenswrapper[4846]: I1005 08:18:57.178059 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-4ccwk" event={"ID":"1bf32b0d-81b1-4774-bf03-995ba2dcfe2d","Type":"ContainerStarted","Data":"ea03e5e59803b75112ad03cdb0433fa6a18fdcb360654b27927acd7b6fd2f5aa"} Oct 05 08:18:58 crc kubenswrapper[4846]: I1005 08:18:58.560322 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-4ccwk" Oct 05 08:18:58 crc kubenswrapper[4846]: I1005 08:18:58.719820 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vf5w6\" (UniqueName: \"kubernetes.io/projected/1bf32b0d-81b1-4774-bf03-995ba2dcfe2d-kube-api-access-vf5w6\") pod \"1bf32b0d-81b1-4774-bf03-995ba2dcfe2d\" (UID: \"1bf32b0d-81b1-4774-bf03-995ba2dcfe2d\") " Oct 05 08:18:58 crc kubenswrapper[4846]: I1005 08:18:58.725139 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf32b0d-81b1-4774-bf03-995ba2dcfe2d-kube-api-access-vf5w6" (OuterVolumeSpecName: "kube-api-access-vf5w6") pod "1bf32b0d-81b1-4774-bf03-995ba2dcfe2d" (UID: "1bf32b0d-81b1-4774-bf03-995ba2dcfe2d"). InnerVolumeSpecName "kube-api-access-vf5w6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:18:58 crc kubenswrapper[4846]: I1005 08:18:58.822203 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vf5w6\" (UniqueName: \"kubernetes.io/projected/1bf32b0d-81b1-4774-bf03-995ba2dcfe2d-kube-api-access-vf5w6\") on node \"crc\" DevicePath \"\"" Oct 05 08:18:59 crc kubenswrapper[4846]: I1005 08:18:59.198260 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-4ccwk" event={"ID":"1bf32b0d-81b1-4774-bf03-995ba2dcfe2d","Type":"ContainerDied","Data":"ea03e5e59803b75112ad03cdb0433fa6a18fdcb360654b27927acd7b6fd2f5aa"} Oct 05 08:18:59 crc kubenswrapper[4846]: I1005 08:18:59.198588 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ea03e5e59803b75112ad03cdb0433fa6a18fdcb360654b27927acd7b6fd2f5aa" Oct 05 08:18:59 crc kubenswrapper[4846]: I1005 08:18:59.198669 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-4ccwk" Oct 05 08:19:05 crc kubenswrapper[4846]: I1005 08:19:05.770165 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-dac8-account-create-gt5sl"] Oct 05 08:19:05 crc kubenswrapper[4846]: E1005 08:19:05.771214 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1bf32b0d-81b1-4774-bf03-995ba2dcfe2d" containerName="mariadb-database-create" Oct 05 08:19:05 crc kubenswrapper[4846]: I1005 08:19:05.771238 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="1bf32b0d-81b1-4774-bf03-995ba2dcfe2d" containerName="mariadb-database-create" Oct 05 08:19:05 crc kubenswrapper[4846]: I1005 08:19:05.771479 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="1bf32b0d-81b1-4774-bf03-995ba2dcfe2d" containerName="mariadb-database-create" Oct 05 08:19:05 crc kubenswrapper[4846]: I1005 08:19:05.772287 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-dac8-account-create-gt5sl" Oct 05 08:19:05 crc kubenswrapper[4846]: I1005 08:19:05.783072 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-db-secret" Oct 05 08:19:05 crc kubenswrapper[4846]: I1005 08:19:05.783230 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-dac8-account-create-gt5sl"] Oct 05 08:19:05 crc kubenswrapper[4846]: I1005 08:19:05.866070 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-57shc\" (UniqueName: \"kubernetes.io/projected/778049cd-b935-404f-887f-63cb4e7d1054-kube-api-access-57shc\") pod \"cinder-dac8-account-create-gt5sl\" (UID: \"778049cd-b935-404f-887f-63cb4e7d1054\") " pod="openstack/cinder-dac8-account-create-gt5sl" Oct 05 08:19:05 crc kubenswrapper[4846]: I1005 08:19:05.968046 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-57shc\" (UniqueName: \"kubernetes.io/projected/778049cd-b935-404f-887f-63cb4e7d1054-kube-api-access-57shc\") pod \"cinder-dac8-account-create-gt5sl\" (UID: \"778049cd-b935-404f-887f-63cb4e7d1054\") " pod="openstack/cinder-dac8-account-create-gt5sl" Oct 05 08:19:05 crc kubenswrapper[4846]: I1005 08:19:05.992072 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-57shc\" (UniqueName: \"kubernetes.io/projected/778049cd-b935-404f-887f-63cb4e7d1054-kube-api-access-57shc\") pod \"cinder-dac8-account-create-gt5sl\" (UID: \"778049cd-b935-404f-887f-63cb4e7d1054\") " pod="openstack/cinder-dac8-account-create-gt5sl" Oct 05 08:19:06 crc kubenswrapper[4846]: I1005 08:19:06.110402 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-dac8-account-create-gt5sl" Oct 05 08:19:06 crc kubenswrapper[4846]: I1005 08:19:06.689448 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-dac8-account-create-gt5sl"] Oct 05 08:19:07 crc kubenswrapper[4846]: I1005 08:19:07.304623 4846 generic.go:334] "Generic (PLEG): container finished" podID="778049cd-b935-404f-887f-63cb4e7d1054" containerID="64a7de2a593c540a6acafc8f361280fb2d3e1a2d97fb1d267443feeb2ea268ae" exitCode=0 Oct 05 08:19:07 crc kubenswrapper[4846]: I1005 08:19:07.304679 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-dac8-account-create-gt5sl" event={"ID":"778049cd-b935-404f-887f-63cb4e7d1054","Type":"ContainerDied","Data":"64a7de2a593c540a6acafc8f361280fb2d3e1a2d97fb1d267443feeb2ea268ae"} Oct 05 08:19:07 crc kubenswrapper[4846]: I1005 08:19:07.305091 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-dac8-account-create-gt5sl" event={"ID":"778049cd-b935-404f-887f-63cb4e7d1054","Type":"ContainerStarted","Data":"ed2c7cd9f40264f5f852aa9729f9157cabf1d7869e4c02106685e17f57df3246"} Oct 05 08:19:08 crc kubenswrapper[4846]: I1005 08:19:08.747699 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-dac8-account-create-gt5sl" Oct 05 08:19:08 crc kubenswrapper[4846]: I1005 08:19:08.836631 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-57shc\" (UniqueName: \"kubernetes.io/projected/778049cd-b935-404f-887f-63cb4e7d1054-kube-api-access-57shc\") pod \"778049cd-b935-404f-887f-63cb4e7d1054\" (UID: \"778049cd-b935-404f-887f-63cb4e7d1054\") " Oct 05 08:19:08 crc kubenswrapper[4846]: I1005 08:19:08.849757 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/778049cd-b935-404f-887f-63cb4e7d1054-kube-api-access-57shc" (OuterVolumeSpecName: "kube-api-access-57shc") pod "778049cd-b935-404f-887f-63cb4e7d1054" (UID: "778049cd-b935-404f-887f-63cb4e7d1054"). InnerVolumeSpecName "kube-api-access-57shc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:19:08 crc kubenswrapper[4846]: I1005 08:19:08.938965 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-57shc\" (UniqueName: \"kubernetes.io/projected/778049cd-b935-404f-887f-63cb4e7d1054-kube-api-access-57shc\") on node \"crc\" DevicePath \"\"" Oct 05 08:19:09 crc kubenswrapper[4846]: I1005 08:19:09.328885 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-dac8-account-create-gt5sl" event={"ID":"778049cd-b935-404f-887f-63cb4e7d1054","Type":"ContainerDied","Data":"ed2c7cd9f40264f5f852aa9729f9157cabf1d7869e4c02106685e17f57df3246"} Oct 05 08:19:09 crc kubenswrapper[4846]: I1005 08:19:09.328936 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ed2c7cd9f40264f5f852aa9729f9157cabf1d7869e4c02106685e17f57df3246" Oct 05 08:19:09 crc kubenswrapper[4846]: I1005 08:19:09.328947 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-dac8-account-create-gt5sl" Oct 05 08:19:10 crc kubenswrapper[4846]: I1005 08:19:10.900702 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-sync-4dvh4"] Oct 05 08:19:10 crc kubenswrapper[4846]: E1005 08:19:10.901331 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="778049cd-b935-404f-887f-63cb4e7d1054" containerName="mariadb-account-create" Oct 05 08:19:10 crc kubenswrapper[4846]: I1005 08:19:10.901344 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="778049cd-b935-404f-887f-63cb4e7d1054" containerName="mariadb-account-create" Oct 05 08:19:10 crc kubenswrapper[4846]: I1005 08:19:10.901521 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="778049cd-b935-404f-887f-63cb4e7d1054" containerName="mariadb-account-create" Oct 05 08:19:10 crc kubenswrapper[4846]: I1005 08:19:10.902117 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-4dvh4" Oct 05 08:19:10 crc kubenswrapper[4846]: I1005 08:19:10.907028 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Oct 05 08:19:10 crc kubenswrapper[4846]: I1005 08:19:10.907314 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Oct 05 08:19:10 crc kubenswrapper[4846]: I1005 08:19:10.907470 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-4htq2" Oct 05 08:19:10 crc kubenswrapper[4846]: I1005 08:19:10.910982 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-4dvh4"] Oct 05 08:19:10 crc kubenswrapper[4846]: I1005 08:19:10.979304 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d91d2d9a-0d5c-4bab-bae1-5b428da00095-combined-ca-bundle\") pod \"cinder-db-sync-4dvh4\" (UID: \"d91d2d9a-0d5c-4bab-bae1-5b428da00095\") " pod="openstack/cinder-db-sync-4dvh4" Oct 05 08:19:10 crc kubenswrapper[4846]: I1005 08:19:10.979388 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/d91d2d9a-0d5c-4bab-bae1-5b428da00095-db-sync-config-data\") pod \"cinder-db-sync-4dvh4\" (UID: \"d91d2d9a-0d5c-4bab-bae1-5b428da00095\") " pod="openstack/cinder-db-sync-4dvh4" Oct 05 08:19:10 crc kubenswrapper[4846]: I1005 08:19:10.979417 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d91d2d9a-0d5c-4bab-bae1-5b428da00095-config-data\") pod \"cinder-db-sync-4dvh4\" (UID: \"d91d2d9a-0d5c-4bab-bae1-5b428da00095\") " pod="openstack/cinder-db-sync-4dvh4" Oct 05 08:19:10 crc kubenswrapper[4846]: I1005 08:19:10.979440 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d91d2d9a-0d5c-4bab-bae1-5b428da00095-scripts\") pod \"cinder-db-sync-4dvh4\" (UID: \"d91d2d9a-0d5c-4bab-bae1-5b428da00095\") " pod="openstack/cinder-db-sync-4dvh4" Oct 05 08:19:10 crc kubenswrapper[4846]: I1005 08:19:10.979470 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6d7rq\" (UniqueName: \"kubernetes.io/projected/d91d2d9a-0d5c-4bab-bae1-5b428da00095-kube-api-access-6d7rq\") pod \"cinder-db-sync-4dvh4\" (UID: \"d91d2d9a-0d5c-4bab-bae1-5b428da00095\") " pod="openstack/cinder-db-sync-4dvh4" Oct 05 08:19:10 crc kubenswrapper[4846]: I1005 08:19:10.979489 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d91d2d9a-0d5c-4bab-bae1-5b428da00095-etc-machine-id\") pod \"cinder-db-sync-4dvh4\" (UID: \"d91d2d9a-0d5c-4bab-bae1-5b428da00095\") " pod="openstack/cinder-db-sync-4dvh4" Oct 05 08:19:11 crc kubenswrapper[4846]: I1005 08:19:11.080643 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d91d2d9a-0d5c-4bab-bae1-5b428da00095-combined-ca-bundle\") pod \"cinder-db-sync-4dvh4\" (UID: \"d91d2d9a-0d5c-4bab-bae1-5b428da00095\") " pod="openstack/cinder-db-sync-4dvh4" Oct 05 08:19:11 crc kubenswrapper[4846]: I1005 08:19:11.080717 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/d91d2d9a-0d5c-4bab-bae1-5b428da00095-db-sync-config-data\") pod \"cinder-db-sync-4dvh4\" (UID: \"d91d2d9a-0d5c-4bab-bae1-5b428da00095\") " pod="openstack/cinder-db-sync-4dvh4" Oct 05 08:19:11 crc kubenswrapper[4846]: I1005 08:19:11.080739 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d91d2d9a-0d5c-4bab-bae1-5b428da00095-config-data\") pod \"cinder-db-sync-4dvh4\" (UID: \"d91d2d9a-0d5c-4bab-bae1-5b428da00095\") " pod="openstack/cinder-db-sync-4dvh4" Oct 05 08:19:11 crc kubenswrapper[4846]: I1005 08:19:11.080757 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d91d2d9a-0d5c-4bab-bae1-5b428da00095-scripts\") pod \"cinder-db-sync-4dvh4\" (UID: \"d91d2d9a-0d5c-4bab-bae1-5b428da00095\") " pod="openstack/cinder-db-sync-4dvh4" Oct 05 08:19:11 crc kubenswrapper[4846]: I1005 08:19:11.080787 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6d7rq\" (UniqueName: \"kubernetes.io/projected/d91d2d9a-0d5c-4bab-bae1-5b428da00095-kube-api-access-6d7rq\") pod \"cinder-db-sync-4dvh4\" (UID: \"d91d2d9a-0d5c-4bab-bae1-5b428da00095\") " pod="openstack/cinder-db-sync-4dvh4" Oct 05 08:19:11 crc kubenswrapper[4846]: I1005 08:19:11.080807 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d91d2d9a-0d5c-4bab-bae1-5b428da00095-etc-machine-id\") pod \"cinder-db-sync-4dvh4\" (UID: \"d91d2d9a-0d5c-4bab-bae1-5b428da00095\") " pod="openstack/cinder-db-sync-4dvh4" Oct 05 08:19:11 crc kubenswrapper[4846]: I1005 08:19:11.080877 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d91d2d9a-0d5c-4bab-bae1-5b428da00095-etc-machine-id\") pod \"cinder-db-sync-4dvh4\" (UID: \"d91d2d9a-0d5c-4bab-bae1-5b428da00095\") " pod="openstack/cinder-db-sync-4dvh4" Oct 05 08:19:11 crc kubenswrapper[4846]: I1005 08:19:11.087445 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d91d2d9a-0d5c-4bab-bae1-5b428da00095-scripts\") pod \"cinder-db-sync-4dvh4\" (UID: \"d91d2d9a-0d5c-4bab-bae1-5b428da00095\") " pod="openstack/cinder-db-sync-4dvh4" Oct 05 08:19:11 crc kubenswrapper[4846]: I1005 08:19:11.087541 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d91d2d9a-0d5c-4bab-bae1-5b428da00095-config-data\") pod \"cinder-db-sync-4dvh4\" (UID: \"d91d2d9a-0d5c-4bab-bae1-5b428da00095\") " pod="openstack/cinder-db-sync-4dvh4" Oct 05 08:19:11 crc kubenswrapper[4846]: I1005 08:19:11.089875 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/d91d2d9a-0d5c-4bab-bae1-5b428da00095-db-sync-config-data\") pod \"cinder-db-sync-4dvh4\" (UID: \"d91d2d9a-0d5c-4bab-bae1-5b428da00095\") " pod="openstack/cinder-db-sync-4dvh4" Oct 05 08:19:11 crc kubenswrapper[4846]: I1005 08:19:11.094399 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d91d2d9a-0d5c-4bab-bae1-5b428da00095-combined-ca-bundle\") pod \"cinder-db-sync-4dvh4\" (UID: \"d91d2d9a-0d5c-4bab-bae1-5b428da00095\") " pod="openstack/cinder-db-sync-4dvh4" Oct 05 08:19:11 crc kubenswrapper[4846]: I1005 08:19:11.109099 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6d7rq\" (UniqueName: \"kubernetes.io/projected/d91d2d9a-0d5c-4bab-bae1-5b428da00095-kube-api-access-6d7rq\") pod \"cinder-db-sync-4dvh4\" (UID: \"d91d2d9a-0d5c-4bab-bae1-5b428da00095\") " pod="openstack/cinder-db-sync-4dvh4" Oct 05 08:19:11 crc kubenswrapper[4846]: I1005 08:19:11.237113 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-4dvh4" Oct 05 08:19:11 crc kubenswrapper[4846]: I1005 08:19:11.707306 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-4dvh4"] Oct 05 08:19:11 crc kubenswrapper[4846]: W1005 08:19:11.711870 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd91d2d9a_0d5c_4bab_bae1_5b428da00095.slice/crio-3a172a7d61217e050c5954468bbfe5beaa7646808292a117922172f4f60b4d69 WatchSource:0}: Error finding container 3a172a7d61217e050c5954468bbfe5beaa7646808292a117922172f4f60b4d69: Status 404 returned error can't find the container with id 3a172a7d61217e050c5954468bbfe5beaa7646808292a117922172f4f60b4d69 Oct 05 08:19:12 crc kubenswrapper[4846]: I1005 08:19:12.366808 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-4dvh4" event={"ID":"d91d2d9a-0d5c-4bab-bae1-5b428da00095","Type":"ContainerStarted","Data":"3a172a7d61217e050c5954468bbfe5beaa7646808292a117922172f4f60b4d69"} Oct 05 08:19:31 crc kubenswrapper[4846]: I1005 08:19:31.542161 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-4dvh4" event={"ID":"d91d2d9a-0d5c-4bab-bae1-5b428da00095","Type":"ContainerStarted","Data":"488875f124d0a3c38ab627e4c84ddc6477d5a7ae2ce09b96bd4b553918976a69"} Oct 05 08:19:31 crc kubenswrapper[4846]: I1005 08:19:31.578934 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-sync-4dvh4" podStartSLOduration=2.568125493 podStartE2EDuration="21.578909674s" podCreationTimestamp="2025-10-05 08:19:10 +0000 UTC" firstStartedPulling="2025-10-05 08:19:11.71492804 +0000 UTC m=+5473.955780835" lastFinishedPulling="2025-10-05 08:19:30.725712241 +0000 UTC m=+5492.966565016" observedRunningTime="2025-10-05 08:19:31.568594598 +0000 UTC m=+5493.809447443" watchObservedRunningTime="2025-10-05 08:19:31.578909674 +0000 UTC m=+5493.819762479" Oct 05 08:19:32 crc kubenswrapper[4846]: I1005 08:19:32.195851 4846 scope.go:117] "RemoveContainer" containerID="9c3799f2b0fcce5160534a43d7e8520327c48bdb700c97480f49756f36d0b7fb" Oct 05 08:19:34 crc kubenswrapper[4846]: I1005 08:19:34.571373 4846 generic.go:334] "Generic (PLEG): container finished" podID="d91d2d9a-0d5c-4bab-bae1-5b428da00095" containerID="488875f124d0a3c38ab627e4c84ddc6477d5a7ae2ce09b96bd4b553918976a69" exitCode=0 Oct 05 08:19:34 crc kubenswrapper[4846]: I1005 08:19:34.571443 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-4dvh4" event={"ID":"d91d2d9a-0d5c-4bab-bae1-5b428da00095","Type":"ContainerDied","Data":"488875f124d0a3c38ab627e4c84ddc6477d5a7ae2ce09b96bd4b553918976a69"} Oct 05 08:19:35 crc kubenswrapper[4846]: I1005 08:19:35.953406 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-4dvh4" Oct 05 08:19:36 crc kubenswrapper[4846]: I1005 08:19:36.074061 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d91d2d9a-0d5c-4bab-bae1-5b428da00095-etc-machine-id\") pod \"d91d2d9a-0d5c-4bab-bae1-5b428da00095\" (UID: \"d91d2d9a-0d5c-4bab-bae1-5b428da00095\") " Oct 05 08:19:36 crc kubenswrapper[4846]: I1005 08:19:36.074244 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d91d2d9a-0d5c-4bab-bae1-5b428da00095-config-data\") pod \"d91d2d9a-0d5c-4bab-bae1-5b428da00095\" (UID: \"d91d2d9a-0d5c-4bab-bae1-5b428da00095\") " Oct 05 08:19:36 crc kubenswrapper[4846]: I1005 08:19:36.074281 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d91d2d9a-0d5c-4bab-bae1-5b428da00095-combined-ca-bundle\") pod \"d91d2d9a-0d5c-4bab-bae1-5b428da00095\" (UID: \"d91d2d9a-0d5c-4bab-bae1-5b428da00095\") " Oct 05 08:19:36 crc kubenswrapper[4846]: I1005 08:19:36.074322 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/d91d2d9a-0d5c-4bab-bae1-5b428da00095-db-sync-config-data\") pod \"d91d2d9a-0d5c-4bab-bae1-5b428da00095\" (UID: \"d91d2d9a-0d5c-4bab-bae1-5b428da00095\") " Oct 05 08:19:36 crc kubenswrapper[4846]: I1005 08:19:36.074361 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d91d2d9a-0d5c-4bab-bae1-5b428da00095-scripts\") pod \"d91d2d9a-0d5c-4bab-bae1-5b428da00095\" (UID: \"d91d2d9a-0d5c-4bab-bae1-5b428da00095\") " Oct 05 08:19:36 crc kubenswrapper[4846]: I1005 08:19:36.074412 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6d7rq\" (UniqueName: \"kubernetes.io/projected/d91d2d9a-0d5c-4bab-bae1-5b428da00095-kube-api-access-6d7rq\") pod \"d91d2d9a-0d5c-4bab-bae1-5b428da00095\" (UID: \"d91d2d9a-0d5c-4bab-bae1-5b428da00095\") " Oct 05 08:19:36 crc kubenswrapper[4846]: I1005 08:19:36.074574 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d91d2d9a-0d5c-4bab-bae1-5b428da00095-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "d91d2d9a-0d5c-4bab-bae1-5b428da00095" (UID: "d91d2d9a-0d5c-4bab-bae1-5b428da00095"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 05 08:19:36 crc kubenswrapper[4846]: I1005 08:19:36.075220 4846 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d91d2d9a-0d5c-4bab-bae1-5b428da00095-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 05 08:19:36 crc kubenswrapper[4846]: I1005 08:19:36.080819 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d91d2d9a-0d5c-4bab-bae1-5b428da00095-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "d91d2d9a-0d5c-4bab-bae1-5b428da00095" (UID: "d91d2d9a-0d5c-4bab-bae1-5b428da00095"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:19:36 crc kubenswrapper[4846]: I1005 08:19:36.081656 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d91d2d9a-0d5c-4bab-bae1-5b428da00095-scripts" (OuterVolumeSpecName: "scripts") pod "d91d2d9a-0d5c-4bab-bae1-5b428da00095" (UID: "d91d2d9a-0d5c-4bab-bae1-5b428da00095"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:19:36 crc kubenswrapper[4846]: I1005 08:19:36.081982 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d91d2d9a-0d5c-4bab-bae1-5b428da00095-kube-api-access-6d7rq" (OuterVolumeSpecName: "kube-api-access-6d7rq") pod "d91d2d9a-0d5c-4bab-bae1-5b428da00095" (UID: "d91d2d9a-0d5c-4bab-bae1-5b428da00095"). InnerVolumeSpecName "kube-api-access-6d7rq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:19:36 crc kubenswrapper[4846]: I1005 08:19:36.104350 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d91d2d9a-0d5c-4bab-bae1-5b428da00095-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d91d2d9a-0d5c-4bab-bae1-5b428da00095" (UID: "d91d2d9a-0d5c-4bab-bae1-5b428da00095"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:19:36 crc kubenswrapper[4846]: I1005 08:19:36.155695 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d91d2d9a-0d5c-4bab-bae1-5b428da00095-config-data" (OuterVolumeSpecName: "config-data") pod "d91d2d9a-0d5c-4bab-bae1-5b428da00095" (UID: "d91d2d9a-0d5c-4bab-bae1-5b428da00095"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:19:36 crc kubenswrapper[4846]: I1005 08:19:36.177857 4846 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d91d2d9a-0d5c-4bab-bae1-5b428da00095-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 08:19:36 crc kubenswrapper[4846]: I1005 08:19:36.177908 4846 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d91d2d9a-0d5c-4bab-bae1-5b428da00095-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 08:19:36 crc kubenswrapper[4846]: I1005 08:19:36.177929 4846 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/d91d2d9a-0d5c-4bab-bae1-5b428da00095-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 08:19:36 crc kubenswrapper[4846]: I1005 08:19:36.177948 4846 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d91d2d9a-0d5c-4bab-bae1-5b428da00095-scripts\") on node \"crc\" DevicePath \"\"" Oct 05 08:19:36 crc kubenswrapper[4846]: I1005 08:19:36.177966 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6d7rq\" (UniqueName: \"kubernetes.io/projected/d91d2d9a-0d5c-4bab-bae1-5b428da00095-kube-api-access-6d7rq\") on node \"crc\" DevicePath \"\"" Oct 05 08:19:36 crc kubenswrapper[4846]: I1005 08:19:36.598987 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-4dvh4" event={"ID":"d91d2d9a-0d5c-4bab-bae1-5b428da00095","Type":"ContainerDied","Data":"3a172a7d61217e050c5954468bbfe5beaa7646808292a117922172f4f60b4d69"} Oct 05 08:19:36 crc kubenswrapper[4846]: I1005 08:19:36.599633 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3a172a7d61217e050c5954468bbfe5beaa7646808292a117922172f4f60b4d69" Oct 05 08:19:36 crc kubenswrapper[4846]: I1005 08:19:36.599088 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-4dvh4" Oct 05 08:19:36 crc kubenswrapper[4846]: I1005 08:19:36.954506 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-84896b68fc-njlvz"] Oct 05 08:19:36 crc kubenswrapper[4846]: E1005 08:19:36.956826 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d91d2d9a-0d5c-4bab-bae1-5b428da00095" containerName="cinder-db-sync" Oct 05 08:19:36 crc kubenswrapper[4846]: I1005 08:19:36.956864 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="d91d2d9a-0d5c-4bab-bae1-5b428da00095" containerName="cinder-db-sync" Oct 05 08:19:36 crc kubenswrapper[4846]: I1005 08:19:36.957104 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="d91d2d9a-0d5c-4bab-bae1-5b428da00095" containerName="cinder-db-sync" Oct 05 08:19:36 crc kubenswrapper[4846]: I1005 08:19:36.958492 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-84896b68fc-njlvz" Oct 05 08:19:36 crc kubenswrapper[4846]: I1005 08:19:36.978446 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-84896b68fc-njlvz"] Oct 05 08:19:37 crc kubenswrapper[4846]: I1005 08:19:37.099177 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5eccdabd-f2e9-4ff9-8fee-37db337983b5-ovsdbserver-nb\") pod \"dnsmasq-dns-84896b68fc-njlvz\" (UID: \"5eccdabd-f2e9-4ff9-8fee-37db337983b5\") " pod="openstack/dnsmasq-dns-84896b68fc-njlvz" Oct 05 08:19:37 crc kubenswrapper[4846]: I1005 08:19:37.099336 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jmfrz\" (UniqueName: \"kubernetes.io/projected/5eccdabd-f2e9-4ff9-8fee-37db337983b5-kube-api-access-jmfrz\") pod \"dnsmasq-dns-84896b68fc-njlvz\" (UID: \"5eccdabd-f2e9-4ff9-8fee-37db337983b5\") " pod="openstack/dnsmasq-dns-84896b68fc-njlvz" Oct 05 08:19:37 crc kubenswrapper[4846]: I1005 08:19:37.099388 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5eccdabd-f2e9-4ff9-8fee-37db337983b5-config\") pod \"dnsmasq-dns-84896b68fc-njlvz\" (UID: \"5eccdabd-f2e9-4ff9-8fee-37db337983b5\") " pod="openstack/dnsmasq-dns-84896b68fc-njlvz" Oct 05 08:19:37 crc kubenswrapper[4846]: I1005 08:19:37.099420 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5eccdabd-f2e9-4ff9-8fee-37db337983b5-ovsdbserver-sb\") pod \"dnsmasq-dns-84896b68fc-njlvz\" (UID: \"5eccdabd-f2e9-4ff9-8fee-37db337983b5\") " pod="openstack/dnsmasq-dns-84896b68fc-njlvz" Oct 05 08:19:37 crc kubenswrapper[4846]: I1005 08:19:37.099604 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5eccdabd-f2e9-4ff9-8fee-37db337983b5-dns-svc\") pod \"dnsmasq-dns-84896b68fc-njlvz\" (UID: \"5eccdabd-f2e9-4ff9-8fee-37db337983b5\") " pod="openstack/dnsmasq-dns-84896b68fc-njlvz" Oct 05 08:19:37 crc kubenswrapper[4846]: I1005 08:19:37.126339 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Oct 05 08:19:37 crc kubenswrapper[4846]: I1005 08:19:37.128943 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 05 08:19:37 crc kubenswrapper[4846]: I1005 08:19:37.131386 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Oct 05 08:19:37 crc kubenswrapper[4846]: I1005 08:19:37.131396 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-4htq2" Oct 05 08:19:37 crc kubenswrapper[4846]: I1005 08:19:37.131790 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Oct 05 08:19:37 crc kubenswrapper[4846]: I1005 08:19:37.132514 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Oct 05 08:19:37 crc kubenswrapper[4846]: I1005 08:19:37.149384 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 05 08:19:37 crc kubenswrapper[4846]: I1005 08:19:37.201567 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ae28b85d-e09d-4a4c-803f-913a79107a70-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"ae28b85d-e09d-4a4c-803f-913a79107a70\") " pod="openstack/cinder-api-0" Oct 05 08:19:37 crc kubenswrapper[4846]: I1005 08:19:37.201647 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5eccdabd-f2e9-4ff9-8fee-37db337983b5-dns-svc\") pod \"dnsmasq-dns-84896b68fc-njlvz\" (UID: \"5eccdabd-f2e9-4ff9-8fee-37db337983b5\") " pod="openstack/dnsmasq-dns-84896b68fc-njlvz" Oct 05 08:19:37 crc kubenswrapper[4846]: I1005 08:19:37.201712 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5eccdabd-f2e9-4ff9-8fee-37db337983b5-ovsdbserver-nb\") pod \"dnsmasq-dns-84896b68fc-njlvz\" (UID: \"5eccdabd-f2e9-4ff9-8fee-37db337983b5\") " pod="openstack/dnsmasq-dns-84896b68fc-njlvz" Oct 05 08:19:37 crc kubenswrapper[4846]: I1005 08:19:37.201738 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ae28b85d-e09d-4a4c-803f-913a79107a70-etc-machine-id\") pod \"cinder-api-0\" (UID: \"ae28b85d-e09d-4a4c-803f-913a79107a70\") " pod="openstack/cinder-api-0" Oct 05 08:19:37 crc kubenswrapper[4846]: I1005 08:19:37.201773 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ae28b85d-e09d-4a4c-803f-913a79107a70-logs\") pod \"cinder-api-0\" (UID: \"ae28b85d-e09d-4a4c-803f-913a79107a70\") " pod="openstack/cinder-api-0" Oct 05 08:19:37 crc kubenswrapper[4846]: I1005 08:19:37.201806 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jmfrz\" (UniqueName: \"kubernetes.io/projected/5eccdabd-f2e9-4ff9-8fee-37db337983b5-kube-api-access-jmfrz\") pod \"dnsmasq-dns-84896b68fc-njlvz\" (UID: \"5eccdabd-f2e9-4ff9-8fee-37db337983b5\") " pod="openstack/dnsmasq-dns-84896b68fc-njlvz" Oct 05 08:19:37 crc kubenswrapper[4846]: I1005 08:19:37.201829 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5eccdabd-f2e9-4ff9-8fee-37db337983b5-config\") pod \"dnsmasq-dns-84896b68fc-njlvz\" (UID: \"5eccdabd-f2e9-4ff9-8fee-37db337983b5\") " pod="openstack/dnsmasq-dns-84896b68fc-njlvz" Oct 05 08:19:37 crc kubenswrapper[4846]: I1005 08:19:37.201845 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4nwnl\" (UniqueName: \"kubernetes.io/projected/ae28b85d-e09d-4a4c-803f-913a79107a70-kube-api-access-4nwnl\") pod \"cinder-api-0\" (UID: \"ae28b85d-e09d-4a4c-803f-913a79107a70\") " pod="openstack/cinder-api-0" Oct 05 08:19:37 crc kubenswrapper[4846]: I1005 08:19:37.201867 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5eccdabd-f2e9-4ff9-8fee-37db337983b5-ovsdbserver-sb\") pod \"dnsmasq-dns-84896b68fc-njlvz\" (UID: \"5eccdabd-f2e9-4ff9-8fee-37db337983b5\") " pod="openstack/dnsmasq-dns-84896b68fc-njlvz" Oct 05 08:19:37 crc kubenswrapper[4846]: I1005 08:19:37.201935 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ae28b85d-e09d-4a4c-803f-913a79107a70-config-data\") pod \"cinder-api-0\" (UID: \"ae28b85d-e09d-4a4c-803f-913a79107a70\") " pod="openstack/cinder-api-0" Oct 05 08:19:37 crc kubenswrapper[4846]: I1005 08:19:37.201977 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ae28b85d-e09d-4a4c-803f-913a79107a70-config-data-custom\") pod \"cinder-api-0\" (UID: \"ae28b85d-e09d-4a4c-803f-913a79107a70\") " pod="openstack/cinder-api-0" Oct 05 08:19:37 crc kubenswrapper[4846]: I1005 08:19:37.202272 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ae28b85d-e09d-4a4c-803f-913a79107a70-scripts\") pod \"cinder-api-0\" (UID: \"ae28b85d-e09d-4a4c-803f-913a79107a70\") " pod="openstack/cinder-api-0" Oct 05 08:19:37 crc kubenswrapper[4846]: I1005 08:19:37.202911 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5eccdabd-f2e9-4ff9-8fee-37db337983b5-ovsdbserver-nb\") pod \"dnsmasq-dns-84896b68fc-njlvz\" (UID: \"5eccdabd-f2e9-4ff9-8fee-37db337983b5\") " pod="openstack/dnsmasq-dns-84896b68fc-njlvz" Oct 05 08:19:37 crc kubenswrapper[4846]: I1005 08:19:37.203166 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5eccdabd-f2e9-4ff9-8fee-37db337983b5-dns-svc\") pod \"dnsmasq-dns-84896b68fc-njlvz\" (UID: \"5eccdabd-f2e9-4ff9-8fee-37db337983b5\") " pod="openstack/dnsmasq-dns-84896b68fc-njlvz" Oct 05 08:19:37 crc kubenswrapper[4846]: I1005 08:19:37.203205 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5eccdabd-f2e9-4ff9-8fee-37db337983b5-ovsdbserver-sb\") pod \"dnsmasq-dns-84896b68fc-njlvz\" (UID: \"5eccdabd-f2e9-4ff9-8fee-37db337983b5\") " pod="openstack/dnsmasq-dns-84896b68fc-njlvz" Oct 05 08:19:37 crc kubenswrapper[4846]: I1005 08:19:37.203184 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5eccdabd-f2e9-4ff9-8fee-37db337983b5-config\") pod \"dnsmasq-dns-84896b68fc-njlvz\" (UID: \"5eccdabd-f2e9-4ff9-8fee-37db337983b5\") " pod="openstack/dnsmasq-dns-84896b68fc-njlvz" Oct 05 08:19:37 crc kubenswrapper[4846]: I1005 08:19:37.232088 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jmfrz\" (UniqueName: \"kubernetes.io/projected/5eccdabd-f2e9-4ff9-8fee-37db337983b5-kube-api-access-jmfrz\") pod \"dnsmasq-dns-84896b68fc-njlvz\" (UID: \"5eccdabd-f2e9-4ff9-8fee-37db337983b5\") " pod="openstack/dnsmasq-dns-84896b68fc-njlvz" Oct 05 08:19:37 crc kubenswrapper[4846]: I1005 08:19:37.273990 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-84896b68fc-njlvz" Oct 05 08:19:37 crc kubenswrapper[4846]: I1005 08:19:37.303985 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ae28b85d-e09d-4a4c-803f-913a79107a70-config-data\") pod \"cinder-api-0\" (UID: \"ae28b85d-e09d-4a4c-803f-913a79107a70\") " pod="openstack/cinder-api-0" Oct 05 08:19:37 crc kubenswrapper[4846]: I1005 08:19:37.304029 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ae28b85d-e09d-4a4c-803f-913a79107a70-config-data-custom\") pod \"cinder-api-0\" (UID: \"ae28b85d-e09d-4a4c-803f-913a79107a70\") " pod="openstack/cinder-api-0" Oct 05 08:19:37 crc kubenswrapper[4846]: I1005 08:19:37.304058 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ae28b85d-e09d-4a4c-803f-913a79107a70-scripts\") pod \"cinder-api-0\" (UID: \"ae28b85d-e09d-4a4c-803f-913a79107a70\") " pod="openstack/cinder-api-0" Oct 05 08:19:37 crc kubenswrapper[4846]: I1005 08:19:37.304105 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ae28b85d-e09d-4a4c-803f-913a79107a70-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"ae28b85d-e09d-4a4c-803f-913a79107a70\") " pod="openstack/cinder-api-0" Oct 05 08:19:37 crc kubenswrapper[4846]: I1005 08:19:37.304172 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ae28b85d-e09d-4a4c-803f-913a79107a70-etc-machine-id\") pod \"cinder-api-0\" (UID: \"ae28b85d-e09d-4a4c-803f-913a79107a70\") " pod="openstack/cinder-api-0" Oct 05 08:19:37 crc kubenswrapper[4846]: I1005 08:19:37.304209 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ae28b85d-e09d-4a4c-803f-913a79107a70-logs\") pod \"cinder-api-0\" (UID: \"ae28b85d-e09d-4a4c-803f-913a79107a70\") " pod="openstack/cinder-api-0" Oct 05 08:19:37 crc kubenswrapper[4846]: I1005 08:19:37.304257 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4nwnl\" (UniqueName: \"kubernetes.io/projected/ae28b85d-e09d-4a4c-803f-913a79107a70-kube-api-access-4nwnl\") pod \"cinder-api-0\" (UID: \"ae28b85d-e09d-4a4c-803f-913a79107a70\") " pod="openstack/cinder-api-0" Oct 05 08:19:37 crc kubenswrapper[4846]: I1005 08:19:37.306306 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ae28b85d-e09d-4a4c-803f-913a79107a70-etc-machine-id\") pod \"cinder-api-0\" (UID: \"ae28b85d-e09d-4a4c-803f-913a79107a70\") " pod="openstack/cinder-api-0" Oct 05 08:19:37 crc kubenswrapper[4846]: I1005 08:19:37.307026 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ae28b85d-e09d-4a4c-803f-913a79107a70-logs\") pod \"cinder-api-0\" (UID: \"ae28b85d-e09d-4a4c-803f-913a79107a70\") " pod="openstack/cinder-api-0" Oct 05 08:19:37 crc kubenswrapper[4846]: I1005 08:19:37.309788 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ae28b85d-e09d-4a4c-803f-913a79107a70-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"ae28b85d-e09d-4a4c-803f-913a79107a70\") " pod="openstack/cinder-api-0" Oct 05 08:19:37 crc kubenswrapper[4846]: I1005 08:19:37.309800 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ae28b85d-e09d-4a4c-803f-913a79107a70-config-data\") pod \"cinder-api-0\" (UID: \"ae28b85d-e09d-4a4c-803f-913a79107a70\") " pod="openstack/cinder-api-0" Oct 05 08:19:37 crc kubenswrapper[4846]: I1005 08:19:37.322605 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ae28b85d-e09d-4a4c-803f-913a79107a70-scripts\") pod \"cinder-api-0\" (UID: \"ae28b85d-e09d-4a4c-803f-913a79107a70\") " pod="openstack/cinder-api-0" Oct 05 08:19:37 crc kubenswrapper[4846]: I1005 08:19:37.322755 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ae28b85d-e09d-4a4c-803f-913a79107a70-config-data-custom\") pod \"cinder-api-0\" (UID: \"ae28b85d-e09d-4a4c-803f-913a79107a70\") " pod="openstack/cinder-api-0" Oct 05 08:19:37 crc kubenswrapper[4846]: I1005 08:19:37.330599 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4nwnl\" (UniqueName: \"kubernetes.io/projected/ae28b85d-e09d-4a4c-803f-913a79107a70-kube-api-access-4nwnl\") pod \"cinder-api-0\" (UID: \"ae28b85d-e09d-4a4c-803f-913a79107a70\") " pod="openstack/cinder-api-0" Oct 05 08:19:37 crc kubenswrapper[4846]: I1005 08:19:37.443964 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 05 08:19:38 crc kubenswrapper[4846]: I1005 08:19:37.731628 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-84896b68fc-njlvz"] Oct 05 08:19:38 crc kubenswrapper[4846]: W1005 08:19:37.732547 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5eccdabd_f2e9_4ff9_8fee_37db337983b5.slice/crio-c57603f6c6b243b53a81d00e89ff2d595bbf3338ef1cb015b90121f5d1870893 WatchSource:0}: Error finding container c57603f6c6b243b53a81d00e89ff2d595bbf3338ef1cb015b90121f5d1870893: Status 404 returned error can't find the container with id c57603f6c6b243b53a81d00e89ff2d595bbf3338ef1cb015b90121f5d1870893 Oct 05 08:19:38 crc kubenswrapper[4846]: I1005 08:19:37.929760 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 05 08:19:38 crc kubenswrapper[4846]: I1005 08:19:38.622107 4846 generic.go:334] "Generic (PLEG): container finished" podID="5eccdabd-f2e9-4ff9-8fee-37db337983b5" containerID="e6f91d0ac3643b5c00c030c3cd42dbdcfb6aa3bce679bf16f630b96882ef39b2" exitCode=0 Oct 05 08:19:38 crc kubenswrapper[4846]: I1005 08:19:38.622261 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-84896b68fc-njlvz" event={"ID":"5eccdabd-f2e9-4ff9-8fee-37db337983b5","Type":"ContainerDied","Data":"e6f91d0ac3643b5c00c030c3cd42dbdcfb6aa3bce679bf16f630b96882ef39b2"} Oct 05 08:19:38 crc kubenswrapper[4846]: I1005 08:19:38.622508 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-84896b68fc-njlvz" event={"ID":"5eccdabd-f2e9-4ff9-8fee-37db337983b5","Type":"ContainerStarted","Data":"c57603f6c6b243b53a81d00e89ff2d595bbf3338ef1cb015b90121f5d1870893"} Oct 05 08:19:38 crc kubenswrapper[4846]: I1005 08:19:38.624391 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"ae28b85d-e09d-4a4c-803f-913a79107a70","Type":"ContainerStarted","Data":"0fe9d2e24070d8d49915ddeecbba6e0b302d41477291efc6d31b6eb65f614cec"} Oct 05 08:19:38 crc kubenswrapper[4846]: I1005 08:19:38.624411 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"ae28b85d-e09d-4a4c-803f-913a79107a70","Type":"ContainerStarted","Data":"13be8d281766353f853a56c065a6df7de8029c150526ee2f62652e316a5ada69"} Oct 05 08:19:39 crc kubenswrapper[4846]: I1005 08:19:39.257681 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Oct 05 08:19:39 crc kubenswrapper[4846]: I1005 08:19:39.634575 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-84896b68fc-njlvz" event={"ID":"5eccdabd-f2e9-4ff9-8fee-37db337983b5","Type":"ContainerStarted","Data":"6306c230a0140fbc7b9d73056688b0174f8dade5645ebb85c57ab843ae00807e"} Oct 05 08:19:39 crc kubenswrapper[4846]: I1005 08:19:39.634945 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-84896b68fc-njlvz" Oct 05 08:19:39 crc kubenswrapper[4846]: I1005 08:19:39.636604 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"ae28b85d-e09d-4a4c-803f-913a79107a70","Type":"ContainerStarted","Data":"b9b082bd8ce137870748c390e693716f0db7d3a3edb71d138a48e66431e42a08"} Oct 05 08:19:39 crc kubenswrapper[4846]: I1005 08:19:39.636748 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="ae28b85d-e09d-4a4c-803f-913a79107a70" containerName="cinder-api-log" containerID="cri-o://0fe9d2e24070d8d49915ddeecbba6e0b302d41477291efc6d31b6eb65f614cec" gracePeriod=30 Oct 05 08:19:39 crc kubenswrapper[4846]: I1005 08:19:39.636772 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="ae28b85d-e09d-4a4c-803f-913a79107a70" containerName="cinder-api" containerID="cri-o://b9b082bd8ce137870748c390e693716f0db7d3a3edb71d138a48e66431e42a08" gracePeriod=30 Oct 05 08:19:39 crc kubenswrapper[4846]: I1005 08:19:39.636757 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Oct 05 08:19:39 crc kubenswrapper[4846]: I1005 08:19:39.657663 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-84896b68fc-njlvz" podStartSLOduration=3.6576462960000002 podStartE2EDuration="3.657646296s" podCreationTimestamp="2025-10-05 08:19:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 08:19:39.656751142 +0000 UTC m=+5501.897603937" watchObservedRunningTime="2025-10-05 08:19:39.657646296 +0000 UTC m=+5501.898499071" Oct 05 08:19:39 crc kubenswrapper[4846]: I1005 08:19:39.682964 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=2.682939873 podStartE2EDuration="2.682939873s" podCreationTimestamp="2025-10-05 08:19:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 08:19:39.676883861 +0000 UTC m=+5501.917736636" watchObservedRunningTime="2025-10-05 08:19:39.682939873 +0000 UTC m=+5501.923792648" Oct 05 08:19:40 crc kubenswrapper[4846]: I1005 08:19:40.520303 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 05 08:19:40 crc kubenswrapper[4846]: I1005 08:19:40.645235 4846 generic.go:334] "Generic (PLEG): container finished" podID="ae28b85d-e09d-4a4c-803f-913a79107a70" containerID="b9b082bd8ce137870748c390e693716f0db7d3a3edb71d138a48e66431e42a08" exitCode=0 Oct 05 08:19:40 crc kubenswrapper[4846]: I1005 08:19:40.645277 4846 generic.go:334] "Generic (PLEG): container finished" podID="ae28b85d-e09d-4a4c-803f-913a79107a70" containerID="0fe9d2e24070d8d49915ddeecbba6e0b302d41477291efc6d31b6eb65f614cec" exitCode=143 Oct 05 08:19:40 crc kubenswrapper[4846]: I1005 08:19:40.645285 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 05 08:19:40 crc kubenswrapper[4846]: I1005 08:19:40.645324 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"ae28b85d-e09d-4a4c-803f-913a79107a70","Type":"ContainerDied","Data":"b9b082bd8ce137870748c390e693716f0db7d3a3edb71d138a48e66431e42a08"} Oct 05 08:19:40 crc kubenswrapper[4846]: I1005 08:19:40.645379 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"ae28b85d-e09d-4a4c-803f-913a79107a70","Type":"ContainerDied","Data":"0fe9d2e24070d8d49915ddeecbba6e0b302d41477291efc6d31b6eb65f614cec"} Oct 05 08:19:40 crc kubenswrapper[4846]: I1005 08:19:40.645392 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"ae28b85d-e09d-4a4c-803f-913a79107a70","Type":"ContainerDied","Data":"13be8d281766353f853a56c065a6df7de8029c150526ee2f62652e316a5ada69"} Oct 05 08:19:40 crc kubenswrapper[4846]: I1005 08:19:40.645412 4846 scope.go:117] "RemoveContainer" containerID="b9b082bd8ce137870748c390e693716f0db7d3a3edb71d138a48e66431e42a08" Oct 05 08:19:40 crc kubenswrapper[4846]: I1005 08:19:40.668777 4846 scope.go:117] "RemoveContainer" containerID="0fe9d2e24070d8d49915ddeecbba6e0b302d41477291efc6d31b6eb65f614cec" Oct 05 08:19:40 crc kubenswrapper[4846]: I1005 08:19:40.687517 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ae28b85d-e09d-4a4c-803f-913a79107a70-scripts\") pod \"ae28b85d-e09d-4a4c-803f-913a79107a70\" (UID: \"ae28b85d-e09d-4a4c-803f-913a79107a70\") " Oct 05 08:19:40 crc kubenswrapper[4846]: I1005 08:19:40.687610 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4nwnl\" (UniqueName: \"kubernetes.io/projected/ae28b85d-e09d-4a4c-803f-913a79107a70-kube-api-access-4nwnl\") pod \"ae28b85d-e09d-4a4c-803f-913a79107a70\" (UID: \"ae28b85d-e09d-4a4c-803f-913a79107a70\") " Oct 05 08:19:40 crc kubenswrapper[4846]: I1005 08:19:40.687705 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ae28b85d-e09d-4a4c-803f-913a79107a70-config-data\") pod \"ae28b85d-e09d-4a4c-803f-913a79107a70\" (UID: \"ae28b85d-e09d-4a4c-803f-913a79107a70\") " Oct 05 08:19:40 crc kubenswrapper[4846]: I1005 08:19:40.687743 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ae28b85d-e09d-4a4c-803f-913a79107a70-etc-machine-id\") pod \"ae28b85d-e09d-4a4c-803f-913a79107a70\" (UID: \"ae28b85d-e09d-4a4c-803f-913a79107a70\") " Oct 05 08:19:40 crc kubenswrapper[4846]: I1005 08:19:40.687853 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ae28b85d-e09d-4a4c-803f-913a79107a70-logs\") pod \"ae28b85d-e09d-4a4c-803f-913a79107a70\" (UID: \"ae28b85d-e09d-4a4c-803f-913a79107a70\") " Oct 05 08:19:40 crc kubenswrapper[4846]: I1005 08:19:40.687884 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ae28b85d-e09d-4a4c-803f-913a79107a70-config-data-custom\") pod \"ae28b85d-e09d-4a4c-803f-913a79107a70\" (UID: \"ae28b85d-e09d-4a4c-803f-913a79107a70\") " Oct 05 08:19:40 crc kubenswrapper[4846]: I1005 08:19:40.687877 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ae28b85d-e09d-4a4c-803f-913a79107a70-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "ae28b85d-e09d-4a4c-803f-913a79107a70" (UID: "ae28b85d-e09d-4a4c-803f-913a79107a70"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 05 08:19:40 crc kubenswrapper[4846]: I1005 08:19:40.687916 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ae28b85d-e09d-4a4c-803f-913a79107a70-combined-ca-bundle\") pod \"ae28b85d-e09d-4a4c-803f-913a79107a70\" (UID: \"ae28b85d-e09d-4a4c-803f-913a79107a70\") " Oct 05 08:19:40 crc kubenswrapper[4846]: I1005 08:19:40.688145 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ae28b85d-e09d-4a4c-803f-913a79107a70-logs" (OuterVolumeSpecName: "logs") pod "ae28b85d-e09d-4a4c-803f-913a79107a70" (UID: "ae28b85d-e09d-4a4c-803f-913a79107a70"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 08:19:40 crc kubenswrapper[4846]: I1005 08:19:40.688617 4846 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ae28b85d-e09d-4a4c-803f-913a79107a70-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 05 08:19:40 crc kubenswrapper[4846]: I1005 08:19:40.688631 4846 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ae28b85d-e09d-4a4c-803f-913a79107a70-logs\") on node \"crc\" DevicePath \"\"" Oct 05 08:19:40 crc kubenswrapper[4846]: I1005 08:19:40.692956 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ae28b85d-e09d-4a4c-803f-913a79107a70-scripts" (OuterVolumeSpecName: "scripts") pod "ae28b85d-e09d-4a4c-803f-913a79107a70" (UID: "ae28b85d-e09d-4a4c-803f-913a79107a70"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:19:40 crc kubenswrapper[4846]: I1005 08:19:40.693120 4846 scope.go:117] "RemoveContainer" containerID="b9b082bd8ce137870748c390e693716f0db7d3a3edb71d138a48e66431e42a08" Oct 05 08:19:40 crc kubenswrapper[4846]: E1005 08:19:40.694447 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b9b082bd8ce137870748c390e693716f0db7d3a3edb71d138a48e66431e42a08\": container with ID starting with b9b082bd8ce137870748c390e693716f0db7d3a3edb71d138a48e66431e42a08 not found: ID does not exist" containerID="b9b082bd8ce137870748c390e693716f0db7d3a3edb71d138a48e66431e42a08" Oct 05 08:19:40 crc kubenswrapper[4846]: I1005 08:19:40.694496 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b9b082bd8ce137870748c390e693716f0db7d3a3edb71d138a48e66431e42a08"} err="failed to get container status \"b9b082bd8ce137870748c390e693716f0db7d3a3edb71d138a48e66431e42a08\": rpc error: code = NotFound desc = could not find container \"b9b082bd8ce137870748c390e693716f0db7d3a3edb71d138a48e66431e42a08\": container with ID starting with b9b082bd8ce137870748c390e693716f0db7d3a3edb71d138a48e66431e42a08 not found: ID does not exist" Oct 05 08:19:40 crc kubenswrapper[4846]: I1005 08:19:40.694537 4846 scope.go:117] "RemoveContainer" containerID="0fe9d2e24070d8d49915ddeecbba6e0b302d41477291efc6d31b6eb65f614cec" Oct 05 08:19:40 crc kubenswrapper[4846]: E1005 08:19:40.695951 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0fe9d2e24070d8d49915ddeecbba6e0b302d41477291efc6d31b6eb65f614cec\": container with ID starting with 0fe9d2e24070d8d49915ddeecbba6e0b302d41477291efc6d31b6eb65f614cec not found: ID does not exist" containerID="0fe9d2e24070d8d49915ddeecbba6e0b302d41477291efc6d31b6eb65f614cec" Oct 05 08:19:40 crc kubenswrapper[4846]: I1005 08:19:40.695997 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0fe9d2e24070d8d49915ddeecbba6e0b302d41477291efc6d31b6eb65f614cec"} err="failed to get container status \"0fe9d2e24070d8d49915ddeecbba6e0b302d41477291efc6d31b6eb65f614cec\": rpc error: code = NotFound desc = could not find container \"0fe9d2e24070d8d49915ddeecbba6e0b302d41477291efc6d31b6eb65f614cec\": container with ID starting with 0fe9d2e24070d8d49915ddeecbba6e0b302d41477291efc6d31b6eb65f614cec not found: ID does not exist" Oct 05 08:19:40 crc kubenswrapper[4846]: I1005 08:19:40.696028 4846 scope.go:117] "RemoveContainer" containerID="b9b082bd8ce137870748c390e693716f0db7d3a3edb71d138a48e66431e42a08" Oct 05 08:19:40 crc kubenswrapper[4846]: I1005 08:19:40.698550 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ae28b85d-e09d-4a4c-803f-913a79107a70-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "ae28b85d-e09d-4a4c-803f-913a79107a70" (UID: "ae28b85d-e09d-4a4c-803f-913a79107a70"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:19:40 crc kubenswrapper[4846]: I1005 08:19:40.700571 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b9b082bd8ce137870748c390e693716f0db7d3a3edb71d138a48e66431e42a08"} err="failed to get container status \"b9b082bd8ce137870748c390e693716f0db7d3a3edb71d138a48e66431e42a08\": rpc error: code = NotFound desc = could not find container \"b9b082bd8ce137870748c390e693716f0db7d3a3edb71d138a48e66431e42a08\": container with ID starting with b9b082bd8ce137870748c390e693716f0db7d3a3edb71d138a48e66431e42a08 not found: ID does not exist" Oct 05 08:19:40 crc kubenswrapper[4846]: I1005 08:19:40.700621 4846 scope.go:117] "RemoveContainer" containerID="0fe9d2e24070d8d49915ddeecbba6e0b302d41477291efc6d31b6eb65f614cec" Oct 05 08:19:40 crc kubenswrapper[4846]: I1005 08:19:40.701106 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0fe9d2e24070d8d49915ddeecbba6e0b302d41477291efc6d31b6eb65f614cec"} err="failed to get container status \"0fe9d2e24070d8d49915ddeecbba6e0b302d41477291efc6d31b6eb65f614cec\": rpc error: code = NotFound desc = could not find container \"0fe9d2e24070d8d49915ddeecbba6e0b302d41477291efc6d31b6eb65f614cec\": container with ID starting with 0fe9d2e24070d8d49915ddeecbba6e0b302d41477291efc6d31b6eb65f614cec not found: ID does not exist" Oct 05 08:19:40 crc kubenswrapper[4846]: I1005 08:19:40.718605 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ae28b85d-e09d-4a4c-803f-913a79107a70-kube-api-access-4nwnl" (OuterVolumeSpecName: "kube-api-access-4nwnl") pod "ae28b85d-e09d-4a4c-803f-913a79107a70" (UID: "ae28b85d-e09d-4a4c-803f-913a79107a70"). InnerVolumeSpecName "kube-api-access-4nwnl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:19:40 crc kubenswrapper[4846]: I1005 08:19:40.721352 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ae28b85d-e09d-4a4c-803f-913a79107a70-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ae28b85d-e09d-4a4c-803f-913a79107a70" (UID: "ae28b85d-e09d-4a4c-803f-913a79107a70"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:19:40 crc kubenswrapper[4846]: I1005 08:19:40.759838 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ae28b85d-e09d-4a4c-803f-913a79107a70-config-data" (OuterVolumeSpecName: "config-data") pod "ae28b85d-e09d-4a4c-803f-913a79107a70" (UID: "ae28b85d-e09d-4a4c-803f-913a79107a70"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:19:40 crc kubenswrapper[4846]: I1005 08:19:40.790806 4846 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ae28b85d-e09d-4a4c-803f-913a79107a70-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 08:19:40 crc kubenswrapper[4846]: I1005 08:19:40.790991 4846 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ae28b85d-e09d-4a4c-803f-913a79107a70-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 05 08:19:40 crc kubenswrapper[4846]: I1005 08:19:40.791007 4846 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ae28b85d-e09d-4a4c-803f-913a79107a70-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 08:19:40 crc kubenswrapper[4846]: I1005 08:19:40.791016 4846 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ae28b85d-e09d-4a4c-803f-913a79107a70-scripts\") on node \"crc\" DevicePath \"\"" Oct 05 08:19:40 crc kubenswrapper[4846]: I1005 08:19:40.791029 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4nwnl\" (UniqueName: \"kubernetes.io/projected/ae28b85d-e09d-4a4c-803f-913a79107a70-kube-api-access-4nwnl\") on node \"crc\" DevicePath \"\"" Oct 05 08:19:40 crc kubenswrapper[4846]: I1005 08:19:40.974161 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Oct 05 08:19:40 crc kubenswrapper[4846]: I1005 08:19:40.980582 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Oct 05 08:19:41 crc kubenswrapper[4846]: I1005 08:19:41.006096 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Oct 05 08:19:41 crc kubenswrapper[4846]: E1005 08:19:41.006625 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ae28b85d-e09d-4a4c-803f-913a79107a70" containerName="cinder-api" Oct 05 08:19:41 crc kubenswrapper[4846]: I1005 08:19:41.006651 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="ae28b85d-e09d-4a4c-803f-913a79107a70" containerName="cinder-api" Oct 05 08:19:41 crc kubenswrapper[4846]: E1005 08:19:41.006689 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ae28b85d-e09d-4a4c-803f-913a79107a70" containerName="cinder-api-log" Oct 05 08:19:41 crc kubenswrapper[4846]: I1005 08:19:41.006698 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="ae28b85d-e09d-4a4c-803f-913a79107a70" containerName="cinder-api-log" Oct 05 08:19:41 crc kubenswrapper[4846]: I1005 08:19:41.006901 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="ae28b85d-e09d-4a4c-803f-913a79107a70" containerName="cinder-api-log" Oct 05 08:19:41 crc kubenswrapper[4846]: I1005 08:19:41.006935 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="ae28b85d-e09d-4a4c-803f-913a79107a70" containerName="cinder-api" Oct 05 08:19:41 crc kubenswrapper[4846]: I1005 08:19:41.008245 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 05 08:19:41 crc kubenswrapper[4846]: I1005 08:19:41.010588 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-internal-svc" Oct 05 08:19:41 crc kubenswrapper[4846]: I1005 08:19:41.016729 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Oct 05 08:19:41 crc kubenswrapper[4846]: I1005 08:19:41.016910 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-public-svc" Oct 05 08:19:41 crc kubenswrapper[4846]: I1005 08:19:41.017158 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Oct 05 08:19:41 crc kubenswrapper[4846]: I1005 08:19:41.017459 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Oct 05 08:19:41 crc kubenswrapper[4846]: I1005 08:19:41.017584 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-4htq2" Oct 05 08:19:41 crc kubenswrapper[4846]: I1005 08:19:41.018183 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 05 08:19:41 crc kubenswrapper[4846]: I1005 08:19:41.094842 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/410aa2e5-a827-496c-a66c-836ad471f4ef-etc-machine-id\") pod \"cinder-api-0\" (UID: \"410aa2e5-a827-496c-a66c-836ad471f4ef\") " pod="openstack/cinder-api-0" Oct 05 08:19:41 crc kubenswrapper[4846]: I1005 08:19:41.094878 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/410aa2e5-a827-496c-a66c-836ad471f4ef-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"410aa2e5-a827-496c-a66c-836ad471f4ef\") " pod="openstack/cinder-api-0" Oct 05 08:19:41 crc kubenswrapper[4846]: I1005 08:19:41.094906 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/410aa2e5-a827-496c-a66c-836ad471f4ef-logs\") pod \"cinder-api-0\" (UID: \"410aa2e5-a827-496c-a66c-836ad471f4ef\") " pod="openstack/cinder-api-0" Oct 05 08:19:41 crc kubenswrapper[4846]: I1005 08:19:41.094933 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/410aa2e5-a827-496c-a66c-836ad471f4ef-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"410aa2e5-a827-496c-a66c-836ad471f4ef\") " pod="openstack/cinder-api-0" Oct 05 08:19:41 crc kubenswrapper[4846]: I1005 08:19:41.095108 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/410aa2e5-a827-496c-a66c-836ad471f4ef-config-data\") pod \"cinder-api-0\" (UID: \"410aa2e5-a827-496c-a66c-836ad471f4ef\") " pod="openstack/cinder-api-0" Oct 05 08:19:41 crc kubenswrapper[4846]: I1005 08:19:41.095162 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/410aa2e5-a827-496c-a66c-836ad471f4ef-scripts\") pod \"cinder-api-0\" (UID: \"410aa2e5-a827-496c-a66c-836ad471f4ef\") " pod="openstack/cinder-api-0" Oct 05 08:19:41 crc kubenswrapper[4846]: I1005 08:19:41.095548 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/410aa2e5-a827-496c-a66c-836ad471f4ef-public-tls-certs\") pod \"cinder-api-0\" (UID: \"410aa2e5-a827-496c-a66c-836ad471f4ef\") " pod="openstack/cinder-api-0" Oct 05 08:19:41 crc kubenswrapper[4846]: I1005 08:19:41.095681 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mxsr2\" (UniqueName: \"kubernetes.io/projected/410aa2e5-a827-496c-a66c-836ad471f4ef-kube-api-access-mxsr2\") pod \"cinder-api-0\" (UID: \"410aa2e5-a827-496c-a66c-836ad471f4ef\") " pod="openstack/cinder-api-0" Oct 05 08:19:41 crc kubenswrapper[4846]: I1005 08:19:41.095917 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/410aa2e5-a827-496c-a66c-836ad471f4ef-config-data-custom\") pod \"cinder-api-0\" (UID: \"410aa2e5-a827-496c-a66c-836ad471f4ef\") " pod="openstack/cinder-api-0" Oct 05 08:19:41 crc kubenswrapper[4846]: I1005 08:19:41.198020 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/410aa2e5-a827-496c-a66c-836ad471f4ef-etc-machine-id\") pod \"cinder-api-0\" (UID: \"410aa2e5-a827-496c-a66c-836ad471f4ef\") " pod="openstack/cinder-api-0" Oct 05 08:19:41 crc kubenswrapper[4846]: I1005 08:19:41.198059 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/410aa2e5-a827-496c-a66c-836ad471f4ef-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"410aa2e5-a827-496c-a66c-836ad471f4ef\") " pod="openstack/cinder-api-0" Oct 05 08:19:41 crc kubenswrapper[4846]: I1005 08:19:41.198085 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/410aa2e5-a827-496c-a66c-836ad471f4ef-logs\") pod \"cinder-api-0\" (UID: \"410aa2e5-a827-496c-a66c-836ad471f4ef\") " pod="openstack/cinder-api-0" Oct 05 08:19:41 crc kubenswrapper[4846]: I1005 08:19:41.198110 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/410aa2e5-a827-496c-a66c-836ad471f4ef-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"410aa2e5-a827-496c-a66c-836ad471f4ef\") " pod="openstack/cinder-api-0" Oct 05 08:19:41 crc kubenswrapper[4846]: I1005 08:19:41.198130 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/410aa2e5-a827-496c-a66c-836ad471f4ef-scripts\") pod \"cinder-api-0\" (UID: \"410aa2e5-a827-496c-a66c-836ad471f4ef\") " pod="openstack/cinder-api-0" Oct 05 08:19:41 crc kubenswrapper[4846]: I1005 08:19:41.198144 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/410aa2e5-a827-496c-a66c-836ad471f4ef-config-data\") pod \"cinder-api-0\" (UID: \"410aa2e5-a827-496c-a66c-836ad471f4ef\") " pod="openstack/cinder-api-0" Oct 05 08:19:41 crc kubenswrapper[4846]: I1005 08:19:41.198210 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/410aa2e5-a827-496c-a66c-836ad471f4ef-public-tls-certs\") pod \"cinder-api-0\" (UID: \"410aa2e5-a827-496c-a66c-836ad471f4ef\") " pod="openstack/cinder-api-0" Oct 05 08:19:41 crc kubenswrapper[4846]: I1005 08:19:41.198232 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mxsr2\" (UniqueName: \"kubernetes.io/projected/410aa2e5-a827-496c-a66c-836ad471f4ef-kube-api-access-mxsr2\") pod \"cinder-api-0\" (UID: \"410aa2e5-a827-496c-a66c-836ad471f4ef\") " pod="openstack/cinder-api-0" Oct 05 08:19:41 crc kubenswrapper[4846]: I1005 08:19:41.198278 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/410aa2e5-a827-496c-a66c-836ad471f4ef-config-data-custom\") pod \"cinder-api-0\" (UID: \"410aa2e5-a827-496c-a66c-836ad471f4ef\") " pod="openstack/cinder-api-0" Oct 05 08:19:41 crc kubenswrapper[4846]: I1005 08:19:41.200948 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/410aa2e5-a827-496c-a66c-836ad471f4ef-etc-machine-id\") pod \"cinder-api-0\" (UID: \"410aa2e5-a827-496c-a66c-836ad471f4ef\") " pod="openstack/cinder-api-0" Oct 05 08:19:41 crc kubenswrapper[4846]: I1005 08:19:41.201900 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/410aa2e5-a827-496c-a66c-836ad471f4ef-logs\") pod \"cinder-api-0\" (UID: \"410aa2e5-a827-496c-a66c-836ad471f4ef\") " pod="openstack/cinder-api-0" Oct 05 08:19:41 crc kubenswrapper[4846]: I1005 08:19:41.204661 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/410aa2e5-a827-496c-a66c-836ad471f4ef-scripts\") pod \"cinder-api-0\" (UID: \"410aa2e5-a827-496c-a66c-836ad471f4ef\") " pod="openstack/cinder-api-0" Oct 05 08:19:41 crc kubenswrapper[4846]: I1005 08:19:41.204811 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/410aa2e5-a827-496c-a66c-836ad471f4ef-config-data-custom\") pod \"cinder-api-0\" (UID: \"410aa2e5-a827-496c-a66c-836ad471f4ef\") " pod="openstack/cinder-api-0" Oct 05 08:19:41 crc kubenswrapper[4846]: I1005 08:19:41.205657 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/410aa2e5-a827-496c-a66c-836ad471f4ef-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"410aa2e5-a827-496c-a66c-836ad471f4ef\") " pod="openstack/cinder-api-0" Oct 05 08:19:41 crc kubenswrapper[4846]: I1005 08:19:41.205682 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/410aa2e5-a827-496c-a66c-836ad471f4ef-public-tls-certs\") pod \"cinder-api-0\" (UID: \"410aa2e5-a827-496c-a66c-836ad471f4ef\") " pod="openstack/cinder-api-0" Oct 05 08:19:41 crc kubenswrapper[4846]: I1005 08:19:41.206893 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/410aa2e5-a827-496c-a66c-836ad471f4ef-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"410aa2e5-a827-496c-a66c-836ad471f4ef\") " pod="openstack/cinder-api-0" Oct 05 08:19:41 crc kubenswrapper[4846]: I1005 08:19:41.210152 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/410aa2e5-a827-496c-a66c-836ad471f4ef-config-data\") pod \"cinder-api-0\" (UID: \"410aa2e5-a827-496c-a66c-836ad471f4ef\") " pod="openstack/cinder-api-0" Oct 05 08:19:41 crc kubenswrapper[4846]: I1005 08:19:41.222728 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mxsr2\" (UniqueName: \"kubernetes.io/projected/410aa2e5-a827-496c-a66c-836ad471f4ef-kube-api-access-mxsr2\") pod \"cinder-api-0\" (UID: \"410aa2e5-a827-496c-a66c-836ad471f4ef\") " pod="openstack/cinder-api-0" Oct 05 08:19:41 crc kubenswrapper[4846]: I1005 08:19:41.341351 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 05 08:19:42 crc kubenswrapper[4846]: I1005 08:19:42.011670 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 05 08:19:42 crc kubenswrapper[4846]: W1005 08:19:42.016254 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod410aa2e5_a827_496c_a66c_836ad471f4ef.slice/crio-a19d73ce34324cea19c529cede026f1edcd2c82ad6818602cac8eebbef1103df WatchSource:0}: Error finding container a19d73ce34324cea19c529cede026f1edcd2c82ad6818602cac8eebbef1103df: Status 404 returned error can't find the container with id a19d73ce34324cea19c529cede026f1edcd2c82ad6818602cac8eebbef1103df Oct 05 08:19:42 crc kubenswrapper[4846]: I1005 08:19:42.507688 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ae28b85d-e09d-4a4c-803f-913a79107a70" path="/var/lib/kubelet/pods/ae28b85d-e09d-4a4c-803f-913a79107a70/volumes" Oct 05 08:19:42 crc kubenswrapper[4846]: I1005 08:19:42.672136 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"410aa2e5-a827-496c-a66c-836ad471f4ef","Type":"ContainerStarted","Data":"ac6f489661e41135cba0a45fd0a617748fcab22215d43d83de0b8a45edd66f18"} Oct 05 08:19:42 crc kubenswrapper[4846]: I1005 08:19:42.672209 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"410aa2e5-a827-496c-a66c-836ad471f4ef","Type":"ContainerStarted","Data":"a19d73ce34324cea19c529cede026f1edcd2c82ad6818602cac8eebbef1103df"} Oct 05 08:19:43 crc kubenswrapper[4846]: I1005 08:19:43.681739 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"410aa2e5-a827-496c-a66c-836ad471f4ef","Type":"ContainerStarted","Data":"a6560c80d6bd0c3d21d954a5be67f4ebcde5ecafa4cae153e76f1894941fcdca"} Oct 05 08:19:43 crc kubenswrapper[4846]: I1005 08:19:43.682012 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Oct 05 08:19:43 crc kubenswrapper[4846]: I1005 08:19:43.703411 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=3.703387592 podStartE2EDuration="3.703387592s" podCreationTimestamp="2025-10-05 08:19:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 08:19:43.699449417 +0000 UTC m=+5505.940302202" watchObservedRunningTime="2025-10-05 08:19:43.703387592 +0000 UTC m=+5505.944240367" Oct 05 08:19:47 crc kubenswrapper[4846]: I1005 08:19:47.276038 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-84896b68fc-njlvz" Oct 05 08:19:47 crc kubenswrapper[4846]: I1005 08:19:47.363085 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6688fc9d5f-7jvr8"] Oct 05 08:19:47 crc kubenswrapper[4846]: I1005 08:19:47.363488 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6688fc9d5f-7jvr8" podUID="d7bace3b-db65-4724-808a-38bf98c604d0" containerName="dnsmasq-dns" containerID="cri-o://97b62413208bc6bd360f03b49cb792fe05f09a1c2a037f6e28171a331a73e1d5" gracePeriod=10 Oct 05 08:19:47 crc kubenswrapper[4846]: I1005 08:19:47.737179 4846 generic.go:334] "Generic (PLEG): container finished" podID="d7bace3b-db65-4724-808a-38bf98c604d0" containerID="97b62413208bc6bd360f03b49cb792fe05f09a1c2a037f6e28171a331a73e1d5" exitCode=0 Oct 05 08:19:47 crc kubenswrapper[4846]: I1005 08:19:47.737533 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6688fc9d5f-7jvr8" event={"ID":"d7bace3b-db65-4724-808a-38bf98c604d0","Type":"ContainerDied","Data":"97b62413208bc6bd360f03b49cb792fe05f09a1c2a037f6e28171a331a73e1d5"} Oct 05 08:19:47 crc kubenswrapper[4846]: I1005 08:19:47.972401 4846 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-6688fc9d5f-7jvr8" podUID="d7bace3b-db65-4724-808a-38bf98c604d0" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.1.48:5353: connect: connection refused" Oct 05 08:19:51 crc kubenswrapper[4846]: I1005 08:19:51.446386 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6688fc9d5f-7jvr8" Oct 05 08:19:51 crc kubenswrapper[4846]: I1005 08:19:51.604199 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d7bace3b-db65-4724-808a-38bf98c604d0-config\") pod \"d7bace3b-db65-4724-808a-38bf98c604d0\" (UID: \"d7bace3b-db65-4724-808a-38bf98c604d0\") " Oct 05 08:19:51 crc kubenswrapper[4846]: I1005 08:19:51.604270 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d7bace3b-db65-4724-808a-38bf98c604d0-ovsdbserver-nb\") pod \"d7bace3b-db65-4724-808a-38bf98c604d0\" (UID: \"d7bace3b-db65-4724-808a-38bf98c604d0\") " Oct 05 08:19:51 crc kubenswrapper[4846]: I1005 08:19:51.604403 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d7bace3b-db65-4724-808a-38bf98c604d0-ovsdbserver-sb\") pod \"d7bace3b-db65-4724-808a-38bf98c604d0\" (UID: \"d7bace3b-db65-4724-808a-38bf98c604d0\") " Oct 05 08:19:51 crc kubenswrapper[4846]: I1005 08:19:51.604494 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5jkn4\" (UniqueName: \"kubernetes.io/projected/d7bace3b-db65-4724-808a-38bf98c604d0-kube-api-access-5jkn4\") pod \"d7bace3b-db65-4724-808a-38bf98c604d0\" (UID: \"d7bace3b-db65-4724-808a-38bf98c604d0\") " Oct 05 08:19:51 crc kubenswrapper[4846]: I1005 08:19:51.604585 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d7bace3b-db65-4724-808a-38bf98c604d0-dns-svc\") pod \"d7bace3b-db65-4724-808a-38bf98c604d0\" (UID: \"d7bace3b-db65-4724-808a-38bf98c604d0\") " Oct 05 08:19:51 crc kubenswrapper[4846]: I1005 08:19:51.610127 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d7bace3b-db65-4724-808a-38bf98c604d0-kube-api-access-5jkn4" (OuterVolumeSpecName: "kube-api-access-5jkn4") pod "d7bace3b-db65-4724-808a-38bf98c604d0" (UID: "d7bace3b-db65-4724-808a-38bf98c604d0"). InnerVolumeSpecName "kube-api-access-5jkn4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:19:51 crc kubenswrapper[4846]: I1005 08:19:51.654070 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d7bace3b-db65-4724-808a-38bf98c604d0-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "d7bace3b-db65-4724-808a-38bf98c604d0" (UID: "d7bace3b-db65-4724-808a-38bf98c604d0"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 08:19:51 crc kubenswrapper[4846]: I1005 08:19:51.655252 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d7bace3b-db65-4724-808a-38bf98c604d0-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "d7bace3b-db65-4724-808a-38bf98c604d0" (UID: "d7bace3b-db65-4724-808a-38bf98c604d0"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 08:19:51 crc kubenswrapper[4846]: I1005 08:19:51.664923 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d7bace3b-db65-4724-808a-38bf98c604d0-config" (OuterVolumeSpecName: "config") pod "d7bace3b-db65-4724-808a-38bf98c604d0" (UID: "d7bace3b-db65-4724-808a-38bf98c604d0"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 08:19:51 crc kubenswrapper[4846]: I1005 08:19:51.676864 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d7bace3b-db65-4724-808a-38bf98c604d0-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "d7bace3b-db65-4724-808a-38bf98c604d0" (UID: "d7bace3b-db65-4724-808a-38bf98c604d0"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 08:19:51 crc kubenswrapper[4846]: I1005 08:19:51.707448 4846 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d7bace3b-db65-4724-808a-38bf98c604d0-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 05 08:19:51 crc kubenswrapper[4846]: I1005 08:19:51.707501 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5jkn4\" (UniqueName: \"kubernetes.io/projected/d7bace3b-db65-4724-808a-38bf98c604d0-kube-api-access-5jkn4\") on node \"crc\" DevicePath \"\"" Oct 05 08:19:51 crc kubenswrapper[4846]: I1005 08:19:51.707524 4846 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d7bace3b-db65-4724-808a-38bf98c604d0-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 05 08:19:51 crc kubenswrapper[4846]: I1005 08:19:51.707545 4846 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d7bace3b-db65-4724-808a-38bf98c604d0-config\") on node \"crc\" DevicePath \"\"" Oct 05 08:19:51 crc kubenswrapper[4846]: I1005 08:19:51.707563 4846 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d7bace3b-db65-4724-808a-38bf98c604d0-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 05 08:19:51 crc kubenswrapper[4846]: I1005 08:19:51.782655 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6688fc9d5f-7jvr8" event={"ID":"d7bace3b-db65-4724-808a-38bf98c604d0","Type":"ContainerDied","Data":"26ca66d9cc821152cb02ae73d9a9494f13333d5977ffe73147217911388abeb1"} Oct 05 08:19:51 crc kubenswrapper[4846]: I1005 08:19:51.782708 4846 scope.go:117] "RemoveContainer" containerID="97b62413208bc6bd360f03b49cb792fe05f09a1c2a037f6e28171a331a73e1d5" Oct 05 08:19:51 crc kubenswrapper[4846]: I1005 08:19:51.782715 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6688fc9d5f-7jvr8" Oct 05 08:19:51 crc kubenswrapper[4846]: I1005 08:19:51.811303 4846 scope.go:117] "RemoveContainer" containerID="0180312d58237bdf8500652e38bbc8bcd0953e65fd6875b29c29ddaea15f7e47" Oct 05 08:19:51 crc kubenswrapper[4846]: I1005 08:19:51.835075 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6688fc9d5f-7jvr8"] Oct 05 08:19:51 crc kubenswrapper[4846]: I1005 08:19:51.842155 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6688fc9d5f-7jvr8"] Oct 05 08:19:52 crc kubenswrapper[4846]: I1005 08:19:52.517837 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d7bace3b-db65-4724-808a-38bf98c604d0" path="/var/lib/kubelet/pods/d7bace3b-db65-4724-808a-38bf98c604d0/volumes" Oct 05 08:19:53 crc kubenswrapper[4846]: I1005 08:19:53.150814 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Oct 05 08:20:10 crc kubenswrapper[4846]: I1005 08:20:10.289414 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Oct 05 08:20:10 crc kubenswrapper[4846]: E1005 08:20:10.290592 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d7bace3b-db65-4724-808a-38bf98c604d0" containerName="dnsmasq-dns" Oct 05 08:20:10 crc kubenswrapper[4846]: I1005 08:20:10.290617 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="d7bace3b-db65-4724-808a-38bf98c604d0" containerName="dnsmasq-dns" Oct 05 08:20:10 crc kubenswrapper[4846]: E1005 08:20:10.290664 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d7bace3b-db65-4724-808a-38bf98c604d0" containerName="init" Oct 05 08:20:10 crc kubenswrapper[4846]: I1005 08:20:10.290676 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="d7bace3b-db65-4724-808a-38bf98c604d0" containerName="init" Oct 05 08:20:10 crc kubenswrapper[4846]: I1005 08:20:10.290981 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="d7bace3b-db65-4724-808a-38bf98c604d0" containerName="dnsmasq-dns" Oct 05 08:20:10 crc kubenswrapper[4846]: I1005 08:20:10.292617 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 05 08:20:10 crc kubenswrapper[4846]: I1005 08:20:10.295615 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Oct 05 08:20:10 crc kubenswrapper[4846]: I1005 08:20:10.300909 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 05 08:20:10 crc kubenswrapper[4846]: I1005 08:20:10.391945 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ae8c941d-3995-4aaf-9432-ce7cd855aee0-config-data\") pod \"cinder-scheduler-0\" (UID: \"ae8c941d-3995-4aaf-9432-ce7cd855aee0\") " pod="openstack/cinder-scheduler-0" Oct 05 08:20:10 crc kubenswrapper[4846]: I1005 08:20:10.392017 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ae8c941d-3995-4aaf-9432-ce7cd855aee0-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"ae8c941d-3995-4aaf-9432-ce7cd855aee0\") " pod="openstack/cinder-scheduler-0" Oct 05 08:20:10 crc kubenswrapper[4846]: I1005 08:20:10.392071 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ae8c941d-3995-4aaf-9432-ce7cd855aee0-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"ae8c941d-3995-4aaf-9432-ce7cd855aee0\") " pod="openstack/cinder-scheduler-0" Oct 05 08:20:10 crc kubenswrapper[4846]: I1005 08:20:10.392112 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ae8c941d-3995-4aaf-9432-ce7cd855aee0-scripts\") pod \"cinder-scheduler-0\" (UID: \"ae8c941d-3995-4aaf-9432-ce7cd855aee0\") " pod="openstack/cinder-scheduler-0" Oct 05 08:20:10 crc kubenswrapper[4846]: I1005 08:20:10.392159 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ae8c941d-3995-4aaf-9432-ce7cd855aee0-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"ae8c941d-3995-4aaf-9432-ce7cd855aee0\") " pod="openstack/cinder-scheduler-0" Oct 05 08:20:10 crc kubenswrapper[4846]: I1005 08:20:10.392255 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s8mtv\" (UniqueName: \"kubernetes.io/projected/ae8c941d-3995-4aaf-9432-ce7cd855aee0-kube-api-access-s8mtv\") pod \"cinder-scheduler-0\" (UID: \"ae8c941d-3995-4aaf-9432-ce7cd855aee0\") " pod="openstack/cinder-scheduler-0" Oct 05 08:20:10 crc kubenswrapper[4846]: I1005 08:20:10.494225 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s8mtv\" (UniqueName: \"kubernetes.io/projected/ae8c941d-3995-4aaf-9432-ce7cd855aee0-kube-api-access-s8mtv\") pod \"cinder-scheduler-0\" (UID: \"ae8c941d-3995-4aaf-9432-ce7cd855aee0\") " pod="openstack/cinder-scheduler-0" Oct 05 08:20:10 crc kubenswrapper[4846]: I1005 08:20:10.494308 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ae8c941d-3995-4aaf-9432-ce7cd855aee0-config-data\") pod \"cinder-scheduler-0\" (UID: \"ae8c941d-3995-4aaf-9432-ce7cd855aee0\") " pod="openstack/cinder-scheduler-0" Oct 05 08:20:10 crc kubenswrapper[4846]: I1005 08:20:10.494345 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ae8c941d-3995-4aaf-9432-ce7cd855aee0-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"ae8c941d-3995-4aaf-9432-ce7cd855aee0\") " pod="openstack/cinder-scheduler-0" Oct 05 08:20:10 crc kubenswrapper[4846]: I1005 08:20:10.494385 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ae8c941d-3995-4aaf-9432-ce7cd855aee0-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"ae8c941d-3995-4aaf-9432-ce7cd855aee0\") " pod="openstack/cinder-scheduler-0" Oct 05 08:20:10 crc kubenswrapper[4846]: I1005 08:20:10.494410 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ae8c941d-3995-4aaf-9432-ce7cd855aee0-scripts\") pod \"cinder-scheduler-0\" (UID: \"ae8c941d-3995-4aaf-9432-ce7cd855aee0\") " pod="openstack/cinder-scheduler-0" Oct 05 08:20:10 crc kubenswrapper[4846]: I1005 08:20:10.494456 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ae8c941d-3995-4aaf-9432-ce7cd855aee0-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"ae8c941d-3995-4aaf-9432-ce7cd855aee0\") " pod="openstack/cinder-scheduler-0" Oct 05 08:20:10 crc kubenswrapper[4846]: I1005 08:20:10.494471 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ae8c941d-3995-4aaf-9432-ce7cd855aee0-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"ae8c941d-3995-4aaf-9432-ce7cd855aee0\") " pod="openstack/cinder-scheduler-0" Oct 05 08:20:10 crc kubenswrapper[4846]: I1005 08:20:10.501302 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ae8c941d-3995-4aaf-9432-ce7cd855aee0-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"ae8c941d-3995-4aaf-9432-ce7cd855aee0\") " pod="openstack/cinder-scheduler-0" Oct 05 08:20:10 crc kubenswrapper[4846]: I1005 08:20:10.503215 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ae8c941d-3995-4aaf-9432-ce7cd855aee0-config-data\") pod \"cinder-scheduler-0\" (UID: \"ae8c941d-3995-4aaf-9432-ce7cd855aee0\") " pod="openstack/cinder-scheduler-0" Oct 05 08:20:10 crc kubenswrapper[4846]: I1005 08:20:10.503487 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ae8c941d-3995-4aaf-9432-ce7cd855aee0-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"ae8c941d-3995-4aaf-9432-ce7cd855aee0\") " pod="openstack/cinder-scheduler-0" Oct 05 08:20:10 crc kubenswrapper[4846]: I1005 08:20:10.509070 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ae8c941d-3995-4aaf-9432-ce7cd855aee0-scripts\") pod \"cinder-scheduler-0\" (UID: \"ae8c941d-3995-4aaf-9432-ce7cd855aee0\") " pod="openstack/cinder-scheduler-0" Oct 05 08:20:10 crc kubenswrapper[4846]: I1005 08:20:10.521700 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s8mtv\" (UniqueName: \"kubernetes.io/projected/ae8c941d-3995-4aaf-9432-ce7cd855aee0-kube-api-access-s8mtv\") pod \"cinder-scheduler-0\" (UID: \"ae8c941d-3995-4aaf-9432-ce7cd855aee0\") " pod="openstack/cinder-scheduler-0" Oct 05 08:20:10 crc kubenswrapper[4846]: I1005 08:20:10.622216 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 05 08:20:11 crc kubenswrapper[4846]: I1005 08:20:11.154558 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 05 08:20:11 crc kubenswrapper[4846]: I1005 08:20:11.649747 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Oct 05 08:20:11 crc kubenswrapper[4846]: I1005 08:20:11.650656 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="410aa2e5-a827-496c-a66c-836ad471f4ef" containerName="cinder-api-log" containerID="cri-o://ac6f489661e41135cba0a45fd0a617748fcab22215d43d83de0b8a45edd66f18" gracePeriod=30 Oct 05 08:20:11 crc kubenswrapper[4846]: I1005 08:20:11.650766 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="410aa2e5-a827-496c-a66c-836ad471f4ef" containerName="cinder-api" containerID="cri-o://a6560c80d6bd0c3d21d954a5be67f4ebcde5ecafa4cae153e76f1894941fcdca" gracePeriod=30 Oct 05 08:20:12 crc kubenswrapper[4846]: I1005 08:20:12.028655 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"ae8c941d-3995-4aaf-9432-ce7cd855aee0","Type":"ContainerStarted","Data":"b54f3fed06b5591e3efb0cac561abc5ff051c6c99e83a2b219268934ba8d206a"} Oct 05 08:20:12 crc kubenswrapper[4846]: I1005 08:20:12.028886 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"ae8c941d-3995-4aaf-9432-ce7cd855aee0","Type":"ContainerStarted","Data":"c881f7015006f72755a8251f93e87e60973c0eecf408bd44f5bdf4012fbdb18f"} Oct 05 08:20:12 crc kubenswrapper[4846]: I1005 08:20:12.034698 4846 generic.go:334] "Generic (PLEG): container finished" podID="410aa2e5-a827-496c-a66c-836ad471f4ef" containerID="ac6f489661e41135cba0a45fd0a617748fcab22215d43d83de0b8a45edd66f18" exitCode=143 Oct 05 08:20:12 crc kubenswrapper[4846]: I1005 08:20:12.034754 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"410aa2e5-a827-496c-a66c-836ad471f4ef","Type":"ContainerDied","Data":"ac6f489661e41135cba0a45fd0a617748fcab22215d43d83de0b8a45edd66f18"} Oct 05 08:20:13 crc kubenswrapper[4846]: I1005 08:20:13.080225 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"ae8c941d-3995-4aaf-9432-ce7cd855aee0","Type":"ContainerStarted","Data":"866f37b3bdf278a6f121b83f945737bf3eba536da595e56befea6f55cb60c478"} Oct 05 08:20:13 crc kubenswrapper[4846]: I1005 08:20:13.117155 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=2.8622453439999997 podStartE2EDuration="3.117130902s" podCreationTimestamp="2025-10-05 08:20:10 +0000 UTC" firstStartedPulling="2025-10-05 08:20:11.155720105 +0000 UTC m=+5533.396572880" lastFinishedPulling="2025-10-05 08:20:11.410605653 +0000 UTC m=+5533.651458438" observedRunningTime="2025-10-05 08:20:13.107828483 +0000 UTC m=+5535.348681268" watchObservedRunningTime="2025-10-05 08:20:13.117130902 +0000 UTC m=+5535.357983677" Oct 05 08:20:15 crc kubenswrapper[4846]: I1005 08:20:15.106412 4846 generic.go:334] "Generic (PLEG): container finished" podID="410aa2e5-a827-496c-a66c-836ad471f4ef" containerID="a6560c80d6bd0c3d21d954a5be67f4ebcde5ecafa4cae153e76f1894941fcdca" exitCode=0 Oct 05 08:20:15 crc kubenswrapper[4846]: I1005 08:20:15.106478 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"410aa2e5-a827-496c-a66c-836ad471f4ef","Type":"ContainerDied","Data":"a6560c80d6bd0c3d21d954a5be67f4ebcde5ecafa4cae153e76f1894941fcdca"} Oct 05 08:20:15 crc kubenswrapper[4846]: I1005 08:20:15.414882 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 05 08:20:15 crc kubenswrapper[4846]: I1005 08:20:15.511265 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/410aa2e5-a827-496c-a66c-836ad471f4ef-combined-ca-bundle\") pod \"410aa2e5-a827-496c-a66c-836ad471f4ef\" (UID: \"410aa2e5-a827-496c-a66c-836ad471f4ef\") " Oct 05 08:20:15 crc kubenswrapper[4846]: I1005 08:20:15.511328 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/410aa2e5-a827-496c-a66c-836ad471f4ef-public-tls-certs\") pod \"410aa2e5-a827-496c-a66c-836ad471f4ef\" (UID: \"410aa2e5-a827-496c-a66c-836ad471f4ef\") " Oct 05 08:20:15 crc kubenswrapper[4846]: I1005 08:20:15.511386 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/410aa2e5-a827-496c-a66c-836ad471f4ef-config-data-custom\") pod \"410aa2e5-a827-496c-a66c-836ad471f4ef\" (UID: \"410aa2e5-a827-496c-a66c-836ad471f4ef\") " Oct 05 08:20:15 crc kubenswrapper[4846]: I1005 08:20:15.511432 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/410aa2e5-a827-496c-a66c-836ad471f4ef-logs\") pod \"410aa2e5-a827-496c-a66c-836ad471f4ef\" (UID: \"410aa2e5-a827-496c-a66c-836ad471f4ef\") " Oct 05 08:20:15 crc kubenswrapper[4846]: I1005 08:20:15.511515 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/410aa2e5-a827-496c-a66c-836ad471f4ef-scripts\") pod \"410aa2e5-a827-496c-a66c-836ad471f4ef\" (UID: \"410aa2e5-a827-496c-a66c-836ad471f4ef\") " Oct 05 08:20:15 crc kubenswrapper[4846]: I1005 08:20:15.511623 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mxsr2\" (UniqueName: \"kubernetes.io/projected/410aa2e5-a827-496c-a66c-836ad471f4ef-kube-api-access-mxsr2\") pod \"410aa2e5-a827-496c-a66c-836ad471f4ef\" (UID: \"410aa2e5-a827-496c-a66c-836ad471f4ef\") " Oct 05 08:20:15 crc kubenswrapper[4846]: I1005 08:20:15.511691 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/410aa2e5-a827-496c-a66c-836ad471f4ef-config-data\") pod \"410aa2e5-a827-496c-a66c-836ad471f4ef\" (UID: \"410aa2e5-a827-496c-a66c-836ad471f4ef\") " Oct 05 08:20:15 crc kubenswrapper[4846]: I1005 08:20:15.511728 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/410aa2e5-a827-496c-a66c-836ad471f4ef-internal-tls-certs\") pod \"410aa2e5-a827-496c-a66c-836ad471f4ef\" (UID: \"410aa2e5-a827-496c-a66c-836ad471f4ef\") " Oct 05 08:20:15 crc kubenswrapper[4846]: I1005 08:20:15.511776 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/410aa2e5-a827-496c-a66c-836ad471f4ef-etc-machine-id\") pod \"410aa2e5-a827-496c-a66c-836ad471f4ef\" (UID: \"410aa2e5-a827-496c-a66c-836ad471f4ef\") " Oct 05 08:20:15 crc kubenswrapper[4846]: I1005 08:20:15.512047 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/410aa2e5-a827-496c-a66c-836ad471f4ef-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "410aa2e5-a827-496c-a66c-836ad471f4ef" (UID: "410aa2e5-a827-496c-a66c-836ad471f4ef"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 05 08:20:15 crc kubenswrapper[4846]: I1005 08:20:15.512404 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/410aa2e5-a827-496c-a66c-836ad471f4ef-logs" (OuterVolumeSpecName: "logs") pod "410aa2e5-a827-496c-a66c-836ad471f4ef" (UID: "410aa2e5-a827-496c-a66c-836ad471f4ef"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 08:20:15 crc kubenswrapper[4846]: I1005 08:20:15.512810 4846 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/410aa2e5-a827-496c-a66c-836ad471f4ef-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 05 08:20:15 crc kubenswrapper[4846]: I1005 08:20:15.512840 4846 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/410aa2e5-a827-496c-a66c-836ad471f4ef-logs\") on node \"crc\" DevicePath \"\"" Oct 05 08:20:15 crc kubenswrapper[4846]: I1005 08:20:15.516680 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/410aa2e5-a827-496c-a66c-836ad471f4ef-kube-api-access-mxsr2" (OuterVolumeSpecName: "kube-api-access-mxsr2") pod "410aa2e5-a827-496c-a66c-836ad471f4ef" (UID: "410aa2e5-a827-496c-a66c-836ad471f4ef"). InnerVolumeSpecName "kube-api-access-mxsr2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:20:15 crc kubenswrapper[4846]: I1005 08:20:15.530712 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/410aa2e5-a827-496c-a66c-836ad471f4ef-scripts" (OuterVolumeSpecName: "scripts") pod "410aa2e5-a827-496c-a66c-836ad471f4ef" (UID: "410aa2e5-a827-496c-a66c-836ad471f4ef"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:20:15 crc kubenswrapper[4846]: I1005 08:20:15.536558 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/410aa2e5-a827-496c-a66c-836ad471f4ef-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "410aa2e5-a827-496c-a66c-836ad471f4ef" (UID: "410aa2e5-a827-496c-a66c-836ad471f4ef"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:20:15 crc kubenswrapper[4846]: I1005 08:20:15.540351 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/410aa2e5-a827-496c-a66c-836ad471f4ef-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "410aa2e5-a827-496c-a66c-836ad471f4ef" (UID: "410aa2e5-a827-496c-a66c-836ad471f4ef"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:20:15 crc kubenswrapper[4846]: I1005 08:20:15.560934 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/410aa2e5-a827-496c-a66c-836ad471f4ef-config-data" (OuterVolumeSpecName: "config-data") pod "410aa2e5-a827-496c-a66c-836ad471f4ef" (UID: "410aa2e5-a827-496c-a66c-836ad471f4ef"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:20:15 crc kubenswrapper[4846]: I1005 08:20:15.579795 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/410aa2e5-a827-496c-a66c-836ad471f4ef-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "410aa2e5-a827-496c-a66c-836ad471f4ef" (UID: "410aa2e5-a827-496c-a66c-836ad471f4ef"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:20:15 crc kubenswrapper[4846]: I1005 08:20:15.585911 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/410aa2e5-a827-496c-a66c-836ad471f4ef-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "410aa2e5-a827-496c-a66c-836ad471f4ef" (UID: "410aa2e5-a827-496c-a66c-836ad471f4ef"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:20:15 crc kubenswrapper[4846]: I1005 08:20:15.616388 4846 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/410aa2e5-a827-496c-a66c-836ad471f4ef-scripts\") on node \"crc\" DevicePath \"\"" Oct 05 08:20:15 crc kubenswrapper[4846]: I1005 08:20:15.616442 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mxsr2\" (UniqueName: \"kubernetes.io/projected/410aa2e5-a827-496c-a66c-836ad471f4ef-kube-api-access-mxsr2\") on node \"crc\" DevicePath \"\"" Oct 05 08:20:15 crc kubenswrapper[4846]: I1005 08:20:15.616463 4846 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/410aa2e5-a827-496c-a66c-836ad471f4ef-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 08:20:15 crc kubenswrapper[4846]: I1005 08:20:15.616482 4846 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/410aa2e5-a827-496c-a66c-836ad471f4ef-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 05 08:20:15 crc kubenswrapper[4846]: I1005 08:20:15.616501 4846 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/410aa2e5-a827-496c-a66c-836ad471f4ef-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 08:20:15 crc kubenswrapper[4846]: I1005 08:20:15.616516 4846 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/410aa2e5-a827-496c-a66c-836ad471f4ef-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 05 08:20:15 crc kubenswrapper[4846]: I1005 08:20:15.616532 4846 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/410aa2e5-a827-496c-a66c-836ad471f4ef-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 05 08:20:15 crc kubenswrapper[4846]: I1005 08:20:15.622383 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Oct 05 08:20:16 crc kubenswrapper[4846]: I1005 08:20:16.124890 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"410aa2e5-a827-496c-a66c-836ad471f4ef","Type":"ContainerDied","Data":"a19d73ce34324cea19c529cede026f1edcd2c82ad6818602cac8eebbef1103df"} Oct 05 08:20:16 crc kubenswrapper[4846]: I1005 08:20:16.125009 4846 scope.go:117] "RemoveContainer" containerID="a6560c80d6bd0c3d21d954a5be67f4ebcde5ecafa4cae153e76f1894941fcdca" Oct 05 08:20:16 crc kubenswrapper[4846]: I1005 08:20:16.125048 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 05 08:20:16 crc kubenswrapper[4846]: I1005 08:20:16.160846 4846 scope.go:117] "RemoveContainer" containerID="ac6f489661e41135cba0a45fd0a617748fcab22215d43d83de0b8a45edd66f18" Oct 05 08:20:16 crc kubenswrapper[4846]: I1005 08:20:16.189246 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Oct 05 08:20:16 crc kubenswrapper[4846]: I1005 08:20:16.226899 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Oct 05 08:20:16 crc kubenswrapper[4846]: I1005 08:20:16.234290 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Oct 05 08:20:16 crc kubenswrapper[4846]: E1005 08:20:16.234826 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="410aa2e5-a827-496c-a66c-836ad471f4ef" containerName="cinder-api" Oct 05 08:20:16 crc kubenswrapper[4846]: I1005 08:20:16.234856 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="410aa2e5-a827-496c-a66c-836ad471f4ef" containerName="cinder-api" Oct 05 08:20:16 crc kubenswrapper[4846]: E1005 08:20:16.234899 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="410aa2e5-a827-496c-a66c-836ad471f4ef" containerName="cinder-api-log" Oct 05 08:20:16 crc kubenswrapper[4846]: I1005 08:20:16.234912 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="410aa2e5-a827-496c-a66c-836ad471f4ef" containerName="cinder-api-log" Oct 05 08:20:16 crc kubenswrapper[4846]: I1005 08:20:16.235254 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="410aa2e5-a827-496c-a66c-836ad471f4ef" containerName="cinder-api" Oct 05 08:20:16 crc kubenswrapper[4846]: I1005 08:20:16.235280 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="410aa2e5-a827-496c-a66c-836ad471f4ef" containerName="cinder-api-log" Oct 05 08:20:16 crc kubenswrapper[4846]: I1005 08:20:16.236421 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 05 08:20:16 crc kubenswrapper[4846]: I1005 08:20:16.242218 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Oct 05 08:20:16 crc kubenswrapper[4846]: I1005 08:20:16.242515 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-internal-svc" Oct 05 08:20:16 crc kubenswrapper[4846]: I1005 08:20:16.242726 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-public-svc" Oct 05 08:20:16 crc kubenswrapper[4846]: I1005 08:20:16.270729 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 05 08:20:16 crc kubenswrapper[4846]: I1005 08:20:16.432356 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8d38c463-603e-4e77-9af8-d9329875a9a3-logs\") pod \"cinder-api-0\" (UID: \"8d38c463-603e-4e77-9af8-d9329875a9a3\") " pod="openstack/cinder-api-0" Oct 05 08:20:16 crc kubenswrapper[4846]: I1005 08:20:16.432489 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/8d38c463-603e-4e77-9af8-d9329875a9a3-etc-machine-id\") pod \"cinder-api-0\" (UID: \"8d38c463-603e-4e77-9af8-d9329875a9a3\") " pod="openstack/cinder-api-0" Oct 05 08:20:16 crc kubenswrapper[4846]: I1005 08:20:16.432627 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vtxbq\" (UniqueName: \"kubernetes.io/projected/8d38c463-603e-4e77-9af8-d9329875a9a3-kube-api-access-vtxbq\") pod \"cinder-api-0\" (UID: \"8d38c463-603e-4e77-9af8-d9329875a9a3\") " pod="openstack/cinder-api-0" Oct 05 08:20:16 crc kubenswrapper[4846]: I1005 08:20:16.432670 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8d38c463-603e-4e77-9af8-d9329875a9a3-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"8d38c463-603e-4e77-9af8-d9329875a9a3\") " pod="openstack/cinder-api-0" Oct 05 08:20:16 crc kubenswrapper[4846]: I1005 08:20:16.432711 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8d38c463-603e-4e77-9af8-d9329875a9a3-public-tls-certs\") pod \"cinder-api-0\" (UID: \"8d38c463-603e-4e77-9af8-d9329875a9a3\") " pod="openstack/cinder-api-0" Oct 05 08:20:16 crc kubenswrapper[4846]: I1005 08:20:16.432747 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8d38c463-603e-4e77-9af8-d9329875a9a3-config-data\") pod \"cinder-api-0\" (UID: \"8d38c463-603e-4e77-9af8-d9329875a9a3\") " pod="openstack/cinder-api-0" Oct 05 08:20:16 crc kubenswrapper[4846]: I1005 08:20:16.432794 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8d38c463-603e-4e77-9af8-d9329875a9a3-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"8d38c463-603e-4e77-9af8-d9329875a9a3\") " pod="openstack/cinder-api-0" Oct 05 08:20:16 crc kubenswrapper[4846]: I1005 08:20:16.432843 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8d38c463-603e-4e77-9af8-d9329875a9a3-config-data-custom\") pod \"cinder-api-0\" (UID: \"8d38c463-603e-4e77-9af8-d9329875a9a3\") " pod="openstack/cinder-api-0" Oct 05 08:20:16 crc kubenswrapper[4846]: I1005 08:20:16.432880 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8d38c463-603e-4e77-9af8-d9329875a9a3-scripts\") pod \"cinder-api-0\" (UID: \"8d38c463-603e-4e77-9af8-d9329875a9a3\") " pod="openstack/cinder-api-0" Oct 05 08:20:16 crc kubenswrapper[4846]: I1005 08:20:16.519293 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="410aa2e5-a827-496c-a66c-836ad471f4ef" path="/var/lib/kubelet/pods/410aa2e5-a827-496c-a66c-836ad471f4ef/volumes" Oct 05 08:20:16 crc kubenswrapper[4846]: I1005 08:20:16.535637 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vtxbq\" (UniqueName: \"kubernetes.io/projected/8d38c463-603e-4e77-9af8-d9329875a9a3-kube-api-access-vtxbq\") pod \"cinder-api-0\" (UID: \"8d38c463-603e-4e77-9af8-d9329875a9a3\") " pod="openstack/cinder-api-0" Oct 05 08:20:16 crc kubenswrapper[4846]: I1005 08:20:16.535736 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8d38c463-603e-4e77-9af8-d9329875a9a3-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"8d38c463-603e-4e77-9af8-d9329875a9a3\") " pod="openstack/cinder-api-0" Oct 05 08:20:16 crc kubenswrapper[4846]: I1005 08:20:16.535792 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8d38c463-603e-4e77-9af8-d9329875a9a3-public-tls-certs\") pod \"cinder-api-0\" (UID: \"8d38c463-603e-4e77-9af8-d9329875a9a3\") " pod="openstack/cinder-api-0" Oct 05 08:20:16 crc kubenswrapper[4846]: I1005 08:20:16.535840 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8d38c463-603e-4e77-9af8-d9329875a9a3-config-data\") pod \"cinder-api-0\" (UID: \"8d38c463-603e-4e77-9af8-d9329875a9a3\") " pod="openstack/cinder-api-0" Oct 05 08:20:16 crc kubenswrapper[4846]: I1005 08:20:16.535904 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8d38c463-603e-4e77-9af8-d9329875a9a3-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"8d38c463-603e-4e77-9af8-d9329875a9a3\") " pod="openstack/cinder-api-0" Oct 05 08:20:16 crc kubenswrapper[4846]: I1005 08:20:16.535963 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8d38c463-603e-4e77-9af8-d9329875a9a3-config-data-custom\") pod \"cinder-api-0\" (UID: \"8d38c463-603e-4e77-9af8-d9329875a9a3\") " pod="openstack/cinder-api-0" Oct 05 08:20:16 crc kubenswrapper[4846]: I1005 08:20:16.536001 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8d38c463-603e-4e77-9af8-d9329875a9a3-scripts\") pod \"cinder-api-0\" (UID: \"8d38c463-603e-4e77-9af8-d9329875a9a3\") " pod="openstack/cinder-api-0" Oct 05 08:20:16 crc kubenswrapper[4846]: I1005 08:20:16.536085 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8d38c463-603e-4e77-9af8-d9329875a9a3-logs\") pod \"cinder-api-0\" (UID: \"8d38c463-603e-4e77-9af8-d9329875a9a3\") " pod="openstack/cinder-api-0" Oct 05 08:20:16 crc kubenswrapper[4846]: I1005 08:20:16.536163 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/8d38c463-603e-4e77-9af8-d9329875a9a3-etc-machine-id\") pod \"cinder-api-0\" (UID: \"8d38c463-603e-4e77-9af8-d9329875a9a3\") " pod="openstack/cinder-api-0" Oct 05 08:20:16 crc kubenswrapper[4846]: I1005 08:20:16.536348 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/8d38c463-603e-4e77-9af8-d9329875a9a3-etc-machine-id\") pod \"cinder-api-0\" (UID: \"8d38c463-603e-4e77-9af8-d9329875a9a3\") " pod="openstack/cinder-api-0" Oct 05 08:20:16 crc kubenswrapper[4846]: I1005 08:20:16.537430 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8d38c463-603e-4e77-9af8-d9329875a9a3-logs\") pod \"cinder-api-0\" (UID: \"8d38c463-603e-4e77-9af8-d9329875a9a3\") " pod="openstack/cinder-api-0" Oct 05 08:20:16 crc kubenswrapper[4846]: I1005 08:20:16.545043 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8d38c463-603e-4e77-9af8-d9329875a9a3-public-tls-certs\") pod \"cinder-api-0\" (UID: \"8d38c463-603e-4e77-9af8-d9329875a9a3\") " pod="openstack/cinder-api-0" Oct 05 08:20:16 crc kubenswrapper[4846]: I1005 08:20:16.545468 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8d38c463-603e-4e77-9af8-d9329875a9a3-scripts\") pod \"cinder-api-0\" (UID: \"8d38c463-603e-4e77-9af8-d9329875a9a3\") " pod="openstack/cinder-api-0" Oct 05 08:20:16 crc kubenswrapper[4846]: I1005 08:20:16.545941 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8d38c463-603e-4e77-9af8-d9329875a9a3-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"8d38c463-603e-4e77-9af8-d9329875a9a3\") " pod="openstack/cinder-api-0" Oct 05 08:20:16 crc kubenswrapper[4846]: I1005 08:20:16.546398 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8d38c463-603e-4e77-9af8-d9329875a9a3-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"8d38c463-603e-4e77-9af8-d9329875a9a3\") " pod="openstack/cinder-api-0" Oct 05 08:20:16 crc kubenswrapper[4846]: I1005 08:20:16.547054 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8d38c463-603e-4e77-9af8-d9329875a9a3-config-data-custom\") pod \"cinder-api-0\" (UID: \"8d38c463-603e-4e77-9af8-d9329875a9a3\") " pod="openstack/cinder-api-0" Oct 05 08:20:16 crc kubenswrapper[4846]: I1005 08:20:16.550158 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8d38c463-603e-4e77-9af8-d9329875a9a3-config-data\") pod \"cinder-api-0\" (UID: \"8d38c463-603e-4e77-9af8-d9329875a9a3\") " pod="openstack/cinder-api-0" Oct 05 08:20:16 crc kubenswrapper[4846]: I1005 08:20:16.570099 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vtxbq\" (UniqueName: \"kubernetes.io/projected/8d38c463-603e-4e77-9af8-d9329875a9a3-kube-api-access-vtxbq\") pod \"cinder-api-0\" (UID: \"8d38c463-603e-4e77-9af8-d9329875a9a3\") " pod="openstack/cinder-api-0" Oct 05 08:20:16 crc kubenswrapper[4846]: I1005 08:20:16.864818 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 05 08:20:17 crc kubenswrapper[4846]: I1005 08:20:17.358135 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 05 08:20:18 crc kubenswrapper[4846]: I1005 08:20:18.153958 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"8d38c463-603e-4e77-9af8-d9329875a9a3","Type":"ContainerStarted","Data":"56340812c7f026e6dcc78095df8817fc42ef7a7061e92795d70975e8e6b1734a"} Oct 05 08:20:18 crc kubenswrapper[4846]: I1005 08:20:18.154130 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"8d38c463-603e-4e77-9af8-d9329875a9a3","Type":"ContainerStarted","Data":"ea6938035d58dffe093f0d1ee1970d32ac7facc2966cea7b6c84e3a9a35b6069"} Oct 05 08:20:19 crc kubenswrapper[4846]: I1005 08:20:19.176467 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"8d38c463-603e-4e77-9af8-d9329875a9a3","Type":"ContainerStarted","Data":"47d7639384f933df4dfd330d12d0d27f16c43e9b68d0017c6a1c12aa399bbb4f"} Oct 05 08:20:19 crc kubenswrapper[4846]: I1005 08:20:19.176941 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Oct 05 08:20:19 crc kubenswrapper[4846]: I1005 08:20:19.201498 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=3.201470864 podStartE2EDuration="3.201470864s" podCreationTimestamp="2025-10-05 08:20:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 08:20:19.191034874 +0000 UTC m=+5541.431887669" watchObservedRunningTime="2025-10-05 08:20:19.201470864 +0000 UTC m=+5541.442323639" Oct 05 08:20:20 crc kubenswrapper[4846]: I1005 08:20:20.848941 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Oct 05 08:20:20 crc kubenswrapper[4846]: I1005 08:20:20.945844 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 05 08:20:21 crc kubenswrapper[4846]: I1005 08:20:21.198821 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="ae8c941d-3995-4aaf-9432-ce7cd855aee0" containerName="cinder-scheduler" containerID="cri-o://b54f3fed06b5591e3efb0cac561abc5ff051c6c99e83a2b219268934ba8d206a" gracePeriod=30 Oct 05 08:20:21 crc kubenswrapper[4846]: I1005 08:20:21.198935 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="ae8c941d-3995-4aaf-9432-ce7cd855aee0" containerName="probe" containerID="cri-o://866f37b3bdf278a6f121b83f945737bf3eba536da595e56befea6f55cb60c478" gracePeriod=30 Oct 05 08:20:22 crc kubenswrapper[4846]: I1005 08:20:22.211053 4846 generic.go:334] "Generic (PLEG): container finished" podID="ae8c941d-3995-4aaf-9432-ce7cd855aee0" containerID="866f37b3bdf278a6f121b83f945737bf3eba536da595e56befea6f55cb60c478" exitCode=0 Oct 05 08:20:22 crc kubenswrapper[4846]: I1005 08:20:22.211149 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"ae8c941d-3995-4aaf-9432-ce7cd855aee0","Type":"ContainerDied","Data":"866f37b3bdf278a6f121b83f945737bf3eba536da595e56befea6f55cb60c478"} Oct 05 08:20:22 crc kubenswrapper[4846]: I1005 08:20:22.991420 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 05 08:20:23 crc kubenswrapper[4846]: I1005 08:20:23.190491 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ae8c941d-3995-4aaf-9432-ce7cd855aee0-config-data\") pod \"ae8c941d-3995-4aaf-9432-ce7cd855aee0\" (UID: \"ae8c941d-3995-4aaf-9432-ce7cd855aee0\") " Oct 05 08:20:23 crc kubenswrapper[4846]: I1005 08:20:23.190578 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ae8c941d-3995-4aaf-9432-ce7cd855aee0-scripts\") pod \"ae8c941d-3995-4aaf-9432-ce7cd855aee0\" (UID: \"ae8c941d-3995-4aaf-9432-ce7cd855aee0\") " Oct 05 08:20:23 crc kubenswrapper[4846]: I1005 08:20:23.190631 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s8mtv\" (UniqueName: \"kubernetes.io/projected/ae8c941d-3995-4aaf-9432-ce7cd855aee0-kube-api-access-s8mtv\") pod \"ae8c941d-3995-4aaf-9432-ce7cd855aee0\" (UID: \"ae8c941d-3995-4aaf-9432-ce7cd855aee0\") " Oct 05 08:20:23 crc kubenswrapper[4846]: I1005 08:20:23.190703 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ae8c941d-3995-4aaf-9432-ce7cd855aee0-combined-ca-bundle\") pod \"ae8c941d-3995-4aaf-9432-ce7cd855aee0\" (UID: \"ae8c941d-3995-4aaf-9432-ce7cd855aee0\") " Oct 05 08:20:23 crc kubenswrapper[4846]: I1005 08:20:23.190807 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ae8c941d-3995-4aaf-9432-ce7cd855aee0-etc-machine-id\") pod \"ae8c941d-3995-4aaf-9432-ce7cd855aee0\" (UID: \"ae8c941d-3995-4aaf-9432-ce7cd855aee0\") " Oct 05 08:20:23 crc kubenswrapper[4846]: I1005 08:20:23.190825 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ae8c941d-3995-4aaf-9432-ce7cd855aee0-config-data-custom\") pod \"ae8c941d-3995-4aaf-9432-ce7cd855aee0\" (UID: \"ae8c941d-3995-4aaf-9432-ce7cd855aee0\") " Oct 05 08:20:23 crc kubenswrapper[4846]: I1005 08:20:23.193068 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ae8c941d-3995-4aaf-9432-ce7cd855aee0-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "ae8c941d-3995-4aaf-9432-ce7cd855aee0" (UID: "ae8c941d-3995-4aaf-9432-ce7cd855aee0"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 05 08:20:23 crc kubenswrapper[4846]: I1005 08:20:23.221669 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ae8c941d-3995-4aaf-9432-ce7cd855aee0-kube-api-access-s8mtv" (OuterVolumeSpecName: "kube-api-access-s8mtv") pod "ae8c941d-3995-4aaf-9432-ce7cd855aee0" (UID: "ae8c941d-3995-4aaf-9432-ce7cd855aee0"). InnerVolumeSpecName "kube-api-access-s8mtv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:20:23 crc kubenswrapper[4846]: I1005 08:20:23.221960 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ae8c941d-3995-4aaf-9432-ce7cd855aee0-scripts" (OuterVolumeSpecName: "scripts") pod "ae8c941d-3995-4aaf-9432-ce7cd855aee0" (UID: "ae8c941d-3995-4aaf-9432-ce7cd855aee0"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:20:23 crc kubenswrapper[4846]: I1005 08:20:23.224126 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ae8c941d-3995-4aaf-9432-ce7cd855aee0-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "ae8c941d-3995-4aaf-9432-ce7cd855aee0" (UID: "ae8c941d-3995-4aaf-9432-ce7cd855aee0"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:20:23 crc kubenswrapper[4846]: I1005 08:20:23.230228 4846 generic.go:334] "Generic (PLEG): container finished" podID="ae8c941d-3995-4aaf-9432-ce7cd855aee0" containerID="b54f3fed06b5591e3efb0cac561abc5ff051c6c99e83a2b219268934ba8d206a" exitCode=0 Oct 05 08:20:23 crc kubenswrapper[4846]: I1005 08:20:23.230277 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"ae8c941d-3995-4aaf-9432-ce7cd855aee0","Type":"ContainerDied","Data":"b54f3fed06b5591e3efb0cac561abc5ff051c6c99e83a2b219268934ba8d206a"} Oct 05 08:20:23 crc kubenswrapper[4846]: I1005 08:20:23.230311 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"ae8c941d-3995-4aaf-9432-ce7cd855aee0","Type":"ContainerDied","Data":"c881f7015006f72755a8251f93e87e60973c0eecf408bd44f5bdf4012fbdb18f"} Oct 05 08:20:23 crc kubenswrapper[4846]: I1005 08:20:23.230331 4846 scope.go:117] "RemoveContainer" containerID="866f37b3bdf278a6f121b83f945737bf3eba536da595e56befea6f55cb60c478" Oct 05 08:20:23 crc kubenswrapper[4846]: I1005 08:20:23.230442 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 05 08:20:23 crc kubenswrapper[4846]: I1005 08:20:23.259809 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ae8c941d-3995-4aaf-9432-ce7cd855aee0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ae8c941d-3995-4aaf-9432-ce7cd855aee0" (UID: "ae8c941d-3995-4aaf-9432-ce7cd855aee0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:20:23 crc kubenswrapper[4846]: I1005 08:20:23.292768 4846 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ae8c941d-3995-4aaf-9432-ce7cd855aee0-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 05 08:20:23 crc kubenswrapper[4846]: I1005 08:20:23.292796 4846 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ae8c941d-3995-4aaf-9432-ce7cd855aee0-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 05 08:20:23 crc kubenswrapper[4846]: I1005 08:20:23.292806 4846 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ae8c941d-3995-4aaf-9432-ce7cd855aee0-scripts\") on node \"crc\" DevicePath \"\"" Oct 05 08:20:23 crc kubenswrapper[4846]: I1005 08:20:23.292815 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s8mtv\" (UniqueName: \"kubernetes.io/projected/ae8c941d-3995-4aaf-9432-ce7cd855aee0-kube-api-access-s8mtv\") on node \"crc\" DevicePath \"\"" Oct 05 08:20:23 crc kubenswrapper[4846]: I1005 08:20:23.292824 4846 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ae8c941d-3995-4aaf-9432-ce7cd855aee0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 08:20:23 crc kubenswrapper[4846]: I1005 08:20:23.311328 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ae8c941d-3995-4aaf-9432-ce7cd855aee0-config-data" (OuterVolumeSpecName: "config-data") pod "ae8c941d-3995-4aaf-9432-ce7cd855aee0" (UID: "ae8c941d-3995-4aaf-9432-ce7cd855aee0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:20:23 crc kubenswrapper[4846]: I1005 08:20:23.345244 4846 scope.go:117] "RemoveContainer" containerID="b54f3fed06b5591e3efb0cac561abc5ff051c6c99e83a2b219268934ba8d206a" Oct 05 08:20:23 crc kubenswrapper[4846]: I1005 08:20:23.366571 4846 scope.go:117] "RemoveContainer" containerID="866f37b3bdf278a6f121b83f945737bf3eba536da595e56befea6f55cb60c478" Oct 05 08:20:23 crc kubenswrapper[4846]: E1005 08:20:23.368057 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"866f37b3bdf278a6f121b83f945737bf3eba536da595e56befea6f55cb60c478\": container with ID starting with 866f37b3bdf278a6f121b83f945737bf3eba536da595e56befea6f55cb60c478 not found: ID does not exist" containerID="866f37b3bdf278a6f121b83f945737bf3eba536da595e56befea6f55cb60c478" Oct 05 08:20:23 crc kubenswrapper[4846]: I1005 08:20:23.368109 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"866f37b3bdf278a6f121b83f945737bf3eba536da595e56befea6f55cb60c478"} err="failed to get container status \"866f37b3bdf278a6f121b83f945737bf3eba536da595e56befea6f55cb60c478\": rpc error: code = NotFound desc = could not find container \"866f37b3bdf278a6f121b83f945737bf3eba536da595e56befea6f55cb60c478\": container with ID starting with 866f37b3bdf278a6f121b83f945737bf3eba536da595e56befea6f55cb60c478 not found: ID does not exist" Oct 05 08:20:23 crc kubenswrapper[4846]: I1005 08:20:23.368140 4846 scope.go:117] "RemoveContainer" containerID="b54f3fed06b5591e3efb0cac561abc5ff051c6c99e83a2b219268934ba8d206a" Oct 05 08:20:23 crc kubenswrapper[4846]: E1005 08:20:23.378775 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b54f3fed06b5591e3efb0cac561abc5ff051c6c99e83a2b219268934ba8d206a\": container with ID starting with b54f3fed06b5591e3efb0cac561abc5ff051c6c99e83a2b219268934ba8d206a not found: ID does not exist" containerID="b54f3fed06b5591e3efb0cac561abc5ff051c6c99e83a2b219268934ba8d206a" Oct 05 08:20:23 crc kubenswrapper[4846]: I1005 08:20:23.378825 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b54f3fed06b5591e3efb0cac561abc5ff051c6c99e83a2b219268934ba8d206a"} err="failed to get container status \"b54f3fed06b5591e3efb0cac561abc5ff051c6c99e83a2b219268934ba8d206a\": rpc error: code = NotFound desc = could not find container \"b54f3fed06b5591e3efb0cac561abc5ff051c6c99e83a2b219268934ba8d206a\": container with ID starting with b54f3fed06b5591e3efb0cac561abc5ff051c6c99e83a2b219268934ba8d206a not found: ID does not exist" Oct 05 08:20:23 crc kubenswrapper[4846]: I1005 08:20:23.395678 4846 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ae8c941d-3995-4aaf-9432-ce7cd855aee0-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 08:20:23 crc kubenswrapper[4846]: I1005 08:20:23.580513 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 05 08:20:23 crc kubenswrapper[4846]: I1005 08:20:23.589458 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 05 08:20:23 crc kubenswrapper[4846]: I1005 08:20:23.606085 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Oct 05 08:20:23 crc kubenswrapper[4846]: E1005 08:20:23.606464 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ae8c941d-3995-4aaf-9432-ce7cd855aee0" containerName="probe" Oct 05 08:20:23 crc kubenswrapper[4846]: I1005 08:20:23.606483 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="ae8c941d-3995-4aaf-9432-ce7cd855aee0" containerName="probe" Oct 05 08:20:23 crc kubenswrapper[4846]: E1005 08:20:23.606505 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ae8c941d-3995-4aaf-9432-ce7cd855aee0" containerName="cinder-scheduler" Oct 05 08:20:23 crc kubenswrapper[4846]: I1005 08:20:23.606514 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="ae8c941d-3995-4aaf-9432-ce7cd855aee0" containerName="cinder-scheduler" Oct 05 08:20:23 crc kubenswrapper[4846]: I1005 08:20:23.606719 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="ae8c941d-3995-4aaf-9432-ce7cd855aee0" containerName="probe" Oct 05 08:20:23 crc kubenswrapper[4846]: I1005 08:20:23.606752 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="ae8c941d-3995-4aaf-9432-ce7cd855aee0" containerName="cinder-scheduler" Oct 05 08:20:23 crc kubenswrapper[4846]: I1005 08:20:23.608066 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 05 08:20:23 crc kubenswrapper[4846]: I1005 08:20:23.610082 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Oct 05 08:20:23 crc kubenswrapper[4846]: I1005 08:20:23.625354 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 05 08:20:23 crc kubenswrapper[4846]: I1005 08:20:23.803271 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ff8cs\" (UniqueName: \"kubernetes.io/projected/70e68e86-4002-41af-ac0c-7426ae1b96d3-kube-api-access-ff8cs\") pod \"cinder-scheduler-0\" (UID: \"70e68e86-4002-41af-ac0c-7426ae1b96d3\") " pod="openstack/cinder-scheduler-0" Oct 05 08:20:23 crc kubenswrapper[4846]: I1005 08:20:23.803435 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/70e68e86-4002-41af-ac0c-7426ae1b96d3-scripts\") pod \"cinder-scheduler-0\" (UID: \"70e68e86-4002-41af-ac0c-7426ae1b96d3\") " pod="openstack/cinder-scheduler-0" Oct 05 08:20:23 crc kubenswrapper[4846]: I1005 08:20:23.803930 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/70e68e86-4002-41af-ac0c-7426ae1b96d3-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"70e68e86-4002-41af-ac0c-7426ae1b96d3\") " pod="openstack/cinder-scheduler-0" Oct 05 08:20:23 crc kubenswrapper[4846]: I1005 08:20:23.804338 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/70e68e86-4002-41af-ac0c-7426ae1b96d3-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"70e68e86-4002-41af-ac0c-7426ae1b96d3\") " pod="openstack/cinder-scheduler-0" Oct 05 08:20:23 crc kubenswrapper[4846]: I1005 08:20:23.804499 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70e68e86-4002-41af-ac0c-7426ae1b96d3-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"70e68e86-4002-41af-ac0c-7426ae1b96d3\") " pod="openstack/cinder-scheduler-0" Oct 05 08:20:23 crc kubenswrapper[4846]: I1005 08:20:23.804723 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/70e68e86-4002-41af-ac0c-7426ae1b96d3-config-data\") pod \"cinder-scheduler-0\" (UID: \"70e68e86-4002-41af-ac0c-7426ae1b96d3\") " pod="openstack/cinder-scheduler-0" Oct 05 08:20:23 crc kubenswrapper[4846]: I1005 08:20:23.906240 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/70e68e86-4002-41af-ac0c-7426ae1b96d3-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"70e68e86-4002-41af-ac0c-7426ae1b96d3\") " pod="openstack/cinder-scheduler-0" Oct 05 08:20:23 crc kubenswrapper[4846]: I1005 08:20:23.906303 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70e68e86-4002-41af-ac0c-7426ae1b96d3-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"70e68e86-4002-41af-ac0c-7426ae1b96d3\") " pod="openstack/cinder-scheduler-0" Oct 05 08:20:23 crc kubenswrapper[4846]: I1005 08:20:23.906353 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/70e68e86-4002-41af-ac0c-7426ae1b96d3-config-data\") pod \"cinder-scheduler-0\" (UID: \"70e68e86-4002-41af-ac0c-7426ae1b96d3\") " pod="openstack/cinder-scheduler-0" Oct 05 08:20:23 crc kubenswrapper[4846]: I1005 08:20:23.906409 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ff8cs\" (UniqueName: \"kubernetes.io/projected/70e68e86-4002-41af-ac0c-7426ae1b96d3-kube-api-access-ff8cs\") pod \"cinder-scheduler-0\" (UID: \"70e68e86-4002-41af-ac0c-7426ae1b96d3\") " pod="openstack/cinder-scheduler-0" Oct 05 08:20:23 crc kubenswrapper[4846]: I1005 08:20:23.906448 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/70e68e86-4002-41af-ac0c-7426ae1b96d3-scripts\") pod \"cinder-scheduler-0\" (UID: \"70e68e86-4002-41af-ac0c-7426ae1b96d3\") " pod="openstack/cinder-scheduler-0" Oct 05 08:20:23 crc kubenswrapper[4846]: I1005 08:20:23.906502 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/70e68e86-4002-41af-ac0c-7426ae1b96d3-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"70e68e86-4002-41af-ac0c-7426ae1b96d3\") " pod="openstack/cinder-scheduler-0" Oct 05 08:20:23 crc kubenswrapper[4846]: I1005 08:20:23.906605 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/70e68e86-4002-41af-ac0c-7426ae1b96d3-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"70e68e86-4002-41af-ac0c-7426ae1b96d3\") " pod="openstack/cinder-scheduler-0" Oct 05 08:20:23 crc kubenswrapper[4846]: I1005 08:20:23.911263 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/70e68e86-4002-41af-ac0c-7426ae1b96d3-scripts\") pod \"cinder-scheduler-0\" (UID: \"70e68e86-4002-41af-ac0c-7426ae1b96d3\") " pod="openstack/cinder-scheduler-0" Oct 05 08:20:23 crc kubenswrapper[4846]: I1005 08:20:23.913252 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70e68e86-4002-41af-ac0c-7426ae1b96d3-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"70e68e86-4002-41af-ac0c-7426ae1b96d3\") " pod="openstack/cinder-scheduler-0" Oct 05 08:20:23 crc kubenswrapper[4846]: I1005 08:20:23.913313 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/70e68e86-4002-41af-ac0c-7426ae1b96d3-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"70e68e86-4002-41af-ac0c-7426ae1b96d3\") " pod="openstack/cinder-scheduler-0" Oct 05 08:20:23 crc kubenswrapper[4846]: I1005 08:20:23.915378 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/70e68e86-4002-41af-ac0c-7426ae1b96d3-config-data\") pod \"cinder-scheduler-0\" (UID: \"70e68e86-4002-41af-ac0c-7426ae1b96d3\") " pod="openstack/cinder-scheduler-0" Oct 05 08:20:23 crc kubenswrapper[4846]: I1005 08:20:23.931225 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ff8cs\" (UniqueName: \"kubernetes.io/projected/70e68e86-4002-41af-ac0c-7426ae1b96d3-kube-api-access-ff8cs\") pod \"cinder-scheduler-0\" (UID: \"70e68e86-4002-41af-ac0c-7426ae1b96d3\") " pod="openstack/cinder-scheduler-0" Oct 05 08:20:23 crc kubenswrapper[4846]: I1005 08:20:23.942840 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 05 08:20:24 crc kubenswrapper[4846]: I1005 08:20:24.478497 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 05 08:20:24 crc kubenswrapper[4846]: W1005 08:20:24.483235 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod70e68e86_4002_41af_ac0c_7426ae1b96d3.slice/crio-8746908198d3b1d0e432711ae07778efb0b0a5f20118cc4502d266b8e472b2b9 WatchSource:0}: Error finding container 8746908198d3b1d0e432711ae07778efb0b0a5f20118cc4502d266b8e472b2b9: Status 404 returned error can't find the container with id 8746908198d3b1d0e432711ae07778efb0b0a5f20118cc4502d266b8e472b2b9 Oct 05 08:20:24 crc kubenswrapper[4846]: I1005 08:20:24.515915 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ae8c941d-3995-4aaf-9432-ce7cd855aee0" path="/var/lib/kubelet/pods/ae8c941d-3995-4aaf-9432-ce7cd855aee0/volumes" Oct 05 08:20:25 crc kubenswrapper[4846]: I1005 08:20:25.259327 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"70e68e86-4002-41af-ac0c-7426ae1b96d3","Type":"ContainerStarted","Data":"a926cb898eb31f4d849edd1eef2a4c285bd94b45d8dd69dfa4423e3d0dcdf58b"} Oct 05 08:20:25 crc kubenswrapper[4846]: I1005 08:20:25.259960 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"70e68e86-4002-41af-ac0c-7426ae1b96d3","Type":"ContainerStarted","Data":"8746908198d3b1d0e432711ae07778efb0b0a5f20118cc4502d266b8e472b2b9"} Oct 05 08:20:26 crc kubenswrapper[4846]: I1005 08:20:26.271605 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"70e68e86-4002-41af-ac0c-7426ae1b96d3","Type":"ContainerStarted","Data":"f70052c497212e67712ce9463b681500608624046443bc60e8f2553614d7fc39"} Oct 05 08:20:26 crc kubenswrapper[4846]: I1005 08:20:26.292645 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=3.292623344 podStartE2EDuration="3.292623344s" podCreationTimestamp="2025-10-05 08:20:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 08:20:26.291068442 +0000 UTC m=+5548.531921227" watchObservedRunningTime="2025-10-05 08:20:26.292623344 +0000 UTC m=+5548.533476119" Oct 05 08:20:28 crc kubenswrapper[4846]: I1005 08:20:28.679345 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Oct 05 08:20:28 crc kubenswrapper[4846]: I1005 08:20:28.943017 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Oct 05 08:20:34 crc kubenswrapper[4846]: I1005 08:20:34.181860 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Oct 05 08:20:36 crc kubenswrapper[4846]: I1005 08:20:36.884023 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-create-8ddb7"] Oct 05 08:20:36 crc kubenswrapper[4846]: I1005 08:20:36.886388 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-8ddb7" Oct 05 08:20:36 crc kubenswrapper[4846]: I1005 08:20:36.896323 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-8ddb7"] Oct 05 08:20:36 crc kubenswrapper[4846]: I1005 08:20:36.999207 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sx8kj\" (UniqueName: \"kubernetes.io/projected/e4b32f67-df27-4951-8798-2a026ca73183-kube-api-access-sx8kj\") pod \"glance-db-create-8ddb7\" (UID: \"e4b32f67-df27-4951-8798-2a026ca73183\") " pod="openstack/glance-db-create-8ddb7" Oct 05 08:20:37 crc kubenswrapper[4846]: I1005 08:20:37.102028 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sx8kj\" (UniqueName: \"kubernetes.io/projected/e4b32f67-df27-4951-8798-2a026ca73183-kube-api-access-sx8kj\") pod \"glance-db-create-8ddb7\" (UID: \"e4b32f67-df27-4951-8798-2a026ca73183\") " pod="openstack/glance-db-create-8ddb7" Oct 05 08:20:37 crc kubenswrapper[4846]: I1005 08:20:37.125569 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sx8kj\" (UniqueName: \"kubernetes.io/projected/e4b32f67-df27-4951-8798-2a026ca73183-kube-api-access-sx8kj\") pod \"glance-db-create-8ddb7\" (UID: \"e4b32f67-df27-4951-8798-2a026ca73183\") " pod="openstack/glance-db-create-8ddb7" Oct 05 08:20:37 crc kubenswrapper[4846]: I1005 08:20:37.228500 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-8ddb7" Oct 05 08:20:37 crc kubenswrapper[4846]: I1005 08:20:37.758487 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-8ddb7"] Oct 05 08:20:38 crc kubenswrapper[4846]: I1005 08:20:38.423888 4846 generic.go:334] "Generic (PLEG): container finished" podID="e4b32f67-df27-4951-8798-2a026ca73183" containerID="a83b3ccc5c15e129aa70a4da820d9dfab416efa3f815e11f8169a82b8b16290c" exitCode=0 Oct 05 08:20:38 crc kubenswrapper[4846]: I1005 08:20:38.423993 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-8ddb7" event={"ID":"e4b32f67-df27-4951-8798-2a026ca73183","Type":"ContainerDied","Data":"a83b3ccc5c15e129aa70a4da820d9dfab416efa3f815e11f8169a82b8b16290c"} Oct 05 08:20:38 crc kubenswrapper[4846]: I1005 08:20:38.424282 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-8ddb7" event={"ID":"e4b32f67-df27-4951-8798-2a026ca73183","Type":"ContainerStarted","Data":"e13f4a23860af5dd9b1594a61ca7e7ec4a56c75ebe7387b180bdfcfbdb6f724e"} Oct 05 08:20:39 crc kubenswrapper[4846]: I1005 08:20:39.923159 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-8ddb7" Oct 05 08:20:39 crc kubenswrapper[4846]: I1005 08:20:39.969362 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sx8kj\" (UniqueName: \"kubernetes.io/projected/e4b32f67-df27-4951-8798-2a026ca73183-kube-api-access-sx8kj\") pod \"e4b32f67-df27-4951-8798-2a026ca73183\" (UID: \"e4b32f67-df27-4951-8798-2a026ca73183\") " Oct 05 08:20:39 crc kubenswrapper[4846]: I1005 08:20:39.979317 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e4b32f67-df27-4951-8798-2a026ca73183-kube-api-access-sx8kj" (OuterVolumeSpecName: "kube-api-access-sx8kj") pod "e4b32f67-df27-4951-8798-2a026ca73183" (UID: "e4b32f67-df27-4951-8798-2a026ca73183"). InnerVolumeSpecName "kube-api-access-sx8kj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:20:40 crc kubenswrapper[4846]: I1005 08:20:40.071343 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sx8kj\" (UniqueName: \"kubernetes.io/projected/e4b32f67-df27-4951-8798-2a026ca73183-kube-api-access-sx8kj\") on node \"crc\" DevicePath \"\"" Oct 05 08:20:40 crc kubenswrapper[4846]: I1005 08:20:40.451923 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-8ddb7" event={"ID":"e4b32f67-df27-4951-8798-2a026ca73183","Type":"ContainerDied","Data":"e13f4a23860af5dd9b1594a61ca7e7ec4a56c75ebe7387b180bdfcfbdb6f724e"} Oct 05 08:20:40 crc kubenswrapper[4846]: I1005 08:20:40.452384 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e13f4a23860af5dd9b1594a61ca7e7ec4a56c75ebe7387b180bdfcfbdb6f724e" Oct 05 08:20:40 crc kubenswrapper[4846]: I1005 08:20:40.451989 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-8ddb7" Oct 05 08:20:47 crc kubenswrapper[4846]: I1005 08:20:47.014809 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-38f8-account-create-6qwhv"] Oct 05 08:20:47 crc kubenswrapper[4846]: E1005 08:20:47.015957 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e4b32f67-df27-4951-8798-2a026ca73183" containerName="mariadb-database-create" Oct 05 08:20:47 crc kubenswrapper[4846]: I1005 08:20:47.015983 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="e4b32f67-df27-4951-8798-2a026ca73183" containerName="mariadb-database-create" Oct 05 08:20:47 crc kubenswrapper[4846]: I1005 08:20:47.016368 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="e4b32f67-df27-4951-8798-2a026ca73183" containerName="mariadb-database-create" Oct 05 08:20:47 crc kubenswrapper[4846]: I1005 08:20:47.017371 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-38f8-account-create-6qwhv" Oct 05 08:20:47 crc kubenswrapper[4846]: I1005 08:20:47.021305 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-db-secret" Oct 05 08:20:47 crc kubenswrapper[4846]: I1005 08:20:47.029663 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-38f8-account-create-6qwhv"] Oct 05 08:20:47 crc kubenswrapper[4846]: I1005 08:20:47.112403 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5md99\" (UniqueName: \"kubernetes.io/projected/d9aa2e92-ea0b-45e4-bf48-e02ccacf73a6-kube-api-access-5md99\") pod \"glance-38f8-account-create-6qwhv\" (UID: \"d9aa2e92-ea0b-45e4-bf48-e02ccacf73a6\") " pod="openstack/glance-38f8-account-create-6qwhv" Oct 05 08:20:47 crc kubenswrapper[4846]: I1005 08:20:47.215554 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5md99\" (UniqueName: \"kubernetes.io/projected/d9aa2e92-ea0b-45e4-bf48-e02ccacf73a6-kube-api-access-5md99\") pod \"glance-38f8-account-create-6qwhv\" (UID: \"d9aa2e92-ea0b-45e4-bf48-e02ccacf73a6\") " pod="openstack/glance-38f8-account-create-6qwhv" Oct 05 08:20:47 crc kubenswrapper[4846]: I1005 08:20:47.244713 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5md99\" (UniqueName: \"kubernetes.io/projected/d9aa2e92-ea0b-45e4-bf48-e02ccacf73a6-kube-api-access-5md99\") pod \"glance-38f8-account-create-6qwhv\" (UID: \"d9aa2e92-ea0b-45e4-bf48-e02ccacf73a6\") " pod="openstack/glance-38f8-account-create-6qwhv" Oct 05 08:20:47 crc kubenswrapper[4846]: I1005 08:20:47.360616 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-38f8-account-create-6qwhv" Oct 05 08:20:47 crc kubenswrapper[4846]: I1005 08:20:47.872277 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-38f8-account-create-6qwhv"] Oct 05 08:20:48 crc kubenswrapper[4846]: I1005 08:20:48.547294 4846 generic.go:334] "Generic (PLEG): container finished" podID="d9aa2e92-ea0b-45e4-bf48-e02ccacf73a6" containerID="1bb972c8ef5b143675bd5127e7485b4a1e4461c95e56cabc05b7cc4303674150" exitCode=0 Oct 05 08:20:48 crc kubenswrapper[4846]: I1005 08:20:48.547444 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-38f8-account-create-6qwhv" event={"ID":"d9aa2e92-ea0b-45e4-bf48-e02ccacf73a6","Type":"ContainerDied","Data":"1bb972c8ef5b143675bd5127e7485b4a1e4461c95e56cabc05b7cc4303674150"} Oct 05 08:20:48 crc kubenswrapper[4846]: I1005 08:20:48.547680 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-38f8-account-create-6qwhv" event={"ID":"d9aa2e92-ea0b-45e4-bf48-e02ccacf73a6","Type":"ContainerStarted","Data":"db6e96f52ca4b11b44b6e51d5f7f89e19eea57d5fc76fa2acced42adc1d5f190"} Oct 05 08:20:49 crc kubenswrapper[4846]: I1005 08:20:49.945999 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-38f8-account-create-6qwhv" Oct 05 08:20:50 crc kubenswrapper[4846]: I1005 08:20:50.075146 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5md99\" (UniqueName: \"kubernetes.io/projected/d9aa2e92-ea0b-45e4-bf48-e02ccacf73a6-kube-api-access-5md99\") pod \"d9aa2e92-ea0b-45e4-bf48-e02ccacf73a6\" (UID: \"d9aa2e92-ea0b-45e4-bf48-e02ccacf73a6\") " Oct 05 08:20:50 crc kubenswrapper[4846]: I1005 08:20:50.083339 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d9aa2e92-ea0b-45e4-bf48-e02ccacf73a6-kube-api-access-5md99" (OuterVolumeSpecName: "kube-api-access-5md99") pod "d9aa2e92-ea0b-45e4-bf48-e02ccacf73a6" (UID: "d9aa2e92-ea0b-45e4-bf48-e02ccacf73a6"). InnerVolumeSpecName "kube-api-access-5md99". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:20:50 crc kubenswrapper[4846]: I1005 08:20:50.177406 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5md99\" (UniqueName: \"kubernetes.io/projected/d9aa2e92-ea0b-45e4-bf48-e02ccacf73a6-kube-api-access-5md99\") on node \"crc\" DevicePath \"\"" Oct 05 08:20:50 crc kubenswrapper[4846]: I1005 08:20:50.573774 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-38f8-account-create-6qwhv" event={"ID":"d9aa2e92-ea0b-45e4-bf48-e02ccacf73a6","Type":"ContainerDied","Data":"db6e96f52ca4b11b44b6e51d5f7f89e19eea57d5fc76fa2acced42adc1d5f190"} Oct 05 08:20:50 crc kubenswrapper[4846]: I1005 08:20:50.573825 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="db6e96f52ca4b11b44b6e51d5f7f89e19eea57d5fc76fa2acced42adc1d5f190" Oct 05 08:20:50 crc kubenswrapper[4846]: I1005 08:20:50.573830 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-38f8-account-create-6qwhv" Oct 05 08:20:52 crc kubenswrapper[4846]: I1005 08:20:52.161802 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-sync-frwh2"] Oct 05 08:20:52 crc kubenswrapper[4846]: E1005 08:20:52.162610 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d9aa2e92-ea0b-45e4-bf48-e02ccacf73a6" containerName="mariadb-account-create" Oct 05 08:20:52 crc kubenswrapper[4846]: I1005 08:20:52.162626 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="d9aa2e92-ea0b-45e4-bf48-e02ccacf73a6" containerName="mariadb-account-create" Oct 05 08:20:52 crc kubenswrapper[4846]: I1005 08:20:52.162837 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="d9aa2e92-ea0b-45e4-bf48-e02ccacf73a6" containerName="mariadb-account-create" Oct 05 08:20:52 crc kubenswrapper[4846]: I1005 08:20:52.163601 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-frwh2" Oct 05 08:20:52 crc kubenswrapper[4846]: I1005 08:20:52.172986 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-4zskb" Oct 05 08:20:52 crc kubenswrapper[4846]: I1005 08:20:52.173785 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-config-data" Oct 05 08:20:52 crc kubenswrapper[4846]: I1005 08:20:52.186142 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-frwh2"] Oct 05 08:20:52 crc kubenswrapper[4846]: I1005 08:20:52.316513 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9jqxx\" (UniqueName: \"kubernetes.io/projected/46c68460-594a-4ab4-b41a-268676e6b950-kube-api-access-9jqxx\") pod \"glance-db-sync-frwh2\" (UID: \"46c68460-594a-4ab4-b41a-268676e6b950\") " pod="openstack/glance-db-sync-frwh2" Oct 05 08:20:52 crc kubenswrapper[4846]: I1005 08:20:52.316630 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/46c68460-594a-4ab4-b41a-268676e6b950-db-sync-config-data\") pod \"glance-db-sync-frwh2\" (UID: \"46c68460-594a-4ab4-b41a-268676e6b950\") " pod="openstack/glance-db-sync-frwh2" Oct 05 08:20:52 crc kubenswrapper[4846]: I1005 08:20:52.316660 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/46c68460-594a-4ab4-b41a-268676e6b950-combined-ca-bundle\") pod \"glance-db-sync-frwh2\" (UID: \"46c68460-594a-4ab4-b41a-268676e6b950\") " pod="openstack/glance-db-sync-frwh2" Oct 05 08:20:52 crc kubenswrapper[4846]: I1005 08:20:52.316686 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/46c68460-594a-4ab4-b41a-268676e6b950-config-data\") pod \"glance-db-sync-frwh2\" (UID: \"46c68460-594a-4ab4-b41a-268676e6b950\") " pod="openstack/glance-db-sync-frwh2" Oct 05 08:20:52 crc kubenswrapper[4846]: I1005 08:20:52.418475 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/46c68460-594a-4ab4-b41a-268676e6b950-combined-ca-bundle\") pod \"glance-db-sync-frwh2\" (UID: \"46c68460-594a-4ab4-b41a-268676e6b950\") " pod="openstack/glance-db-sync-frwh2" Oct 05 08:20:52 crc kubenswrapper[4846]: I1005 08:20:52.418538 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/46c68460-594a-4ab4-b41a-268676e6b950-config-data\") pod \"glance-db-sync-frwh2\" (UID: \"46c68460-594a-4ab4-b41a-268676e6b950\") " pod="openstack/glance-db-sync-frwh2" Oct 05 08:20:52 crc kubenswrapper[4846]: I1005 08:20:52.418598 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9jqxx\" (UniqueName: \"kubernetes.io/projected/46c68460-594a-4ab4-b41a-268676e6b950-kube-api-access-9jqxx\") pod \"glance-db-sync-frwh2\" (UID: \"46c68460-594a-4ab4-b41a-268676e6b950\") " pod="openstack/glance-db-sync-frwh2" Oct 05 08:20:52 crc kubenswrapper[4846]: I1005 08:20:52.418683 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/46c68460-594a-4ab4-b41a-268676e6b950-db-sync-config-data\") pod \"glance-db-sync-frwh2\" (UID: \"46c68460-594a-4ab4-b41a-268676e6b950\") " pod="openstack/glance-db-sync-frwh2" Oct 05 08:20:52 crc kubenswrapper[4846]: I1005 08:20:52.424522 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/46c68460-594a-4ab4-b41a-268676e6b950-config-data\") pod \"glance-db-sync-frwh2\" (UID: \"46c68460-594a-4ab4-b41a-268676e6b950\") " pod="openstack/glance-db-sync-frwh2" Oct 05 08:20:52 crc kubenswrapper[4846]: I1005 08:20:52.427355 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/46c68460-594a-4ab4-b41a-268676e6b950-db-sync-config-data\") pod \"glance-db-sync-frwh2\" (UID: \"46c68460-594a-4ab4-b41a-268676e6b950\") " pod="openstack/glance-db-sync-frwh2" Oct 05 08:20:52 crc kubenswrapper[4846]: I1005 08:20:52.431153 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/46c68460-594a-4ab4-b41a-268676e6b950-combined-ca-bundle\") pod \"glance-db-sync-frwh2\" (UID: \"46c68460-594a-4ab4-b41a-268676e6b950\") " pod="openstack/glance-db-sync-frwh2" Oct 05 08:20:52 crc kubenswrapper[4846]: I1005 08:20:52.449491 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9jqxx\" (UniqueName: \"kubernetes.io/projected/46c68460-594a-4ab4-b41a-268676e6b950-kube-api-access-9jqxx\") pod \"glance-db-sync-frwh2\" (UID: \"46c68460-594a-4ab4-b41a-268676e6b950\") " pod="openstack/glance-db-sync-frwh2" Oct 05 08:20:52 crc kubenswrapper[4846]: I1005 08:20:52.486742 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-frwh2" Oct 05 08:20:53 crc kubenswrapper[4846]: I1005 08:20:53.084319 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-frwh2"] Oct 05 08:20:53 crc kubenswrapper[4846]: I1005 08:20:53.325990 4846 patch_prober.go:28] interesting pod/machine-config-daemon-fscvf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 08:20:53 crc kubenswrapper[4846]: I1005 08:20:53.326074 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 08:20:53 crc kubenswrapper[4846]: I1005 08:20:53.602068 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-frwh2" event={"ID":"46c68460-594a-4ab4-b41a-268676e6b950","Type":"ContainerStarted","Data":"4725be1f314615b6bbb5f5c0486536e344391e37d93d55ac79ac1ffe2e2b4b7a"} Oct 05 08:21:09 crc kubenswrapper[4846]: I1005 08:21:09.750220 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-frwh2" event={"ID":"46c68460-594a-4ab4-b41a-268676e6b950","Type":"ContainerStarted","Data":"329b7338c5617bcc40d190152a65d70ede4f8a7fe7014f8dec828b9861bbf7fc"} Oct 05 08:21:09 crc kubenswrapper[4846]: I1005 08:21:09.783154 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-sync-frwh2" podStartSLOduration=2.662828897 podStartE2EDuration="17.783125871s" podCreationTimestamp="2025-10-05 08:20:52 +0000 UTC" firstStartedPulling="2025-10-05 08:20:53.084759718 +0000 UTC m=+5575.325612493" lastFinishedPulling="2025-10-05 08:21:08.205056652 +0000 UTC m=+5590.445909467" observedRunningTime="2025-10-05 08:21:09.768057477 +0000 UTC m=+5592.008910302" watchObservedRunningTime="2025-10-05 08:21:09.783125871 +0000 UTC m=+5592.023978656" Oct 05 08:21:12 crc kubenswrapper[4846]: I1005 08:21:12.783398 4846 generic.go:334] "Generic (PLEG): container finished" podID="46c68460-594a-4ab4-b41a-268676e6b950" containerID="329b7338c5617bcc40d190152a65d70ede4f8a7fe7014f8dec828b9861bbf7fc" exitCode=0 Oct 05 08:21:12 crc kubenswrapper[4846]: I1005 08:21:12.783529 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-frwh2" event={"ID":"46c68460-594a-4ab4-b41a-268676e6b950","Type":"ContainerDied","Data":"329b7338c5617bcc40d190152a65d70ede4f8a7fe7014f8dec828b9861bbf7fc"} Oct 05 08:21:14 crc kubenswrapper[4846]: I1005 08:21:14.305895 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-frwh2" Oct 05 08:21:14 crc kubenswrapper[4846]: I1005 08:21:14.446135 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/46c68460-594a-4ab4-b41a-268676e6b950-combined-ca-bundle\") pod \"46c68460-594a-4ab4-b41a-268676e6b950\" (UID: \"46c68460-594a-4ab4-b41a-268676e6b950\") " Oct 05 08:21:14 crc kubenswrapper[4846]: I1005 08:21:14.446249 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9jqxx\" (UniqueName: \"kubernetes.io/projected/46c68460-594a-4ab4-b41a-268676e6b950-kube-api-access-9jqxx\") pod \"46c68460-594a-4ab4-b41a-268676e6b950\" (UID: \"46c68460-594a-4ab4-b41a-268676e6b950\") " Oct 05 08:21:14 crc kubenswrapper[4846]: I1005 08:21:14.446307 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/46c68460-594a-4ab4-b41a-268676e6b950-db-sync-config-data\") pod \"46c68460-594a-4ab4-b41a-268676e6b950\" (UID: \"46c68460-594a-4ab4-b41a-268676e6b950\") " Oct 05 08:21:14 crc kubenswrapper[4846]: I1005 08:21:14.446378 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/46c68460-594a-4ab4-b41a-268676e6b950-config-data\") pod \"46c68460-594a-4ab4-b41a-268676e6b950\" (UID: \"46c68460-594a-4ab4-b41a-268676e6b950\") " Oct 05 08:21:14 crc kubenswrapper[4846]: I1005 08:21:14.452839 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/46c68460-594a-4ab4-b41a-268676e6b950-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "46c68460-594a-4ab4-b41a-268676e6b950" (UID: "46c68460-594a-4ab4-b41a-268676e6b950"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:21:14 crc kubenswrapper[4846]: I1005 08:21:14.453104 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/46c68460-594a-4ab4-b41a-268676e6b950-kube-api-access-9jqxx" (OuterVolumeSpecName: "kube-api-access-9jqxx") pod "46c68460-594a-4ab4-b41a-268676e6b950" (UID: "46c68460-594a-4ab4-b41a-268676e6b950"). InnerVolumeSpecName "kube-api-access-9jqxx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:21:14 crc kubenswrapper[4846]: I1005 08:21:14.477332 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/46c68460-594a-4ab4-b41a-268676e6b950-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "46c68460-594a-4ab4-b41a-268676e6b950" (UID: "46c68460-594a-4ab4-b41a-268676e6b950"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:21:14 crc kubenswrapper[4846]: I1005 08:21:14.516475 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/46c68460-594a-4ab4-b41a-268676e6b950-config-data" (OuterVolumeSpecName: "config-data") pod "46c68460-594a-4ab4-b41a-268676e6b950" (UID: "46c68460-594a-4ab4-b41a-268676e6b950"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:21:14 crc kubenswrapper[4846]: I1005 08:21:14.549453 4846 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/46c68460-594a-4ab4-b41a-268676e6b950-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 08:21:14 crc kubenswrapper[4846]: I1005 08:21:14.549504 4846 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/46c68460-594a-4ab4-b41a-268676e6b950-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 08:21:14 crc kubenswrapper[4846]: I1005 08:21:14.549525 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9jqxx\" (UniqueName: \"kubernetes.io/projected/46c68460-594a-4ab4-b41a-268676e6b950-kube-api-access-9jqxx\") on node \"crc\" DevicePath \"\"" Oct 05 08:21:14 crc kubenswrapper[4846]: I1005 08:21:14.549541 4846 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/46c68460-594a-4ab4-b41a-268676e6b950-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 08:21:14 crc kubenswrapper[4846]: I1005 08:21:14.809983 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-frwh2" event={"ID":"46c68460-594a-4ab4-b41a-268676e6b950","Type":"ContainerDied","Data":"4725be1f314615b6bbb5f5c0486536e344391e37d93d55ac79ac1ffe2e2b4b7a"} Oct 05 08:21:14 crc kubenswrapper[4846]: I1005 08:21:14.810021 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4725be1f314615b6bbb5f5c0486536e344391e37d93d55ac79ac1ffe2e2b4b7a" Oct 05 08:21:14 crc kubenswrapper[4846]: I1005 08:21:14.810582 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-frwh2" Oct 05 08:21:15 crc kubenswrapper[4846]: I1005 08:21:15.146601 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Oct 05 08:21:15 crc kubenswrapper[4846]: E1005 08:21:15.147771 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="46c68460-594a-4ab4-b41a-268676e6b950" containerName="glance-db-sync" Oct 05 08:21:15 crc kubenswrapper[4846]: I1005 08:21:15.147836 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="46c68460-594a-4ab4-b41a-268676e6b950" containerName="glance-db-sync" Oct 05 08:21:15 crc kubenswrapper[4846]: I1005 08:21:15.148724 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="46c68460-594a-4ab4-b41a-268676e6b950" containerName="glance-db-sync" Oct 05 08:21:15 crc kubenswrapper[4846]: I1005 08:21:15.172457 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 05 08:21:15 crc kubenswrapper[4846]: I1005 08:21:15.172555 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 05 08:21:15 crc kubenswrapper[4846]: I1005 08:21:15.176638 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Oct 05 08:21:15 crc kubenswrapper[4846]: I1005 08:21:15.176912 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-4zskb" Oct 05 08:21:15 crc kubenswrapper[4846]: I1005 08:21:15.176974 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Oct 05 08:21:15 crc kubenswrapper[4846]: I1005 08:21:15.263255 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-59bf5dd889-89l2d"] Oct 05 08:21:15 crc kubenswrapper[4846]: I1005 08:21:15.265089 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-59bf5dd889-89l2d" Oct 05 08:21:15 crc kubenswrapper[4846]: I1005 08:21:15.265485 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2059932a-7f7d-45ff-8a75-333a6711afdf-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"2059932a-7f7d-45ff-8a75-333a6711afdf\") " pod="openstack/glance-default-external-api-0" Oct 05 08:21:15 crc kubenswrapper[4846]: I1005 08:21:15.265608 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2059932a-7f7d-45ff-8a75-333a6711afdf-config-data\") pod \"glance-default-external-api-0\" (UID: \"2059932a-7f7d-45ff-8a75-333a6711afdf\") " pod="openstack/glance-default-external-api-0" Oct 05 08:21:15 crc kubenswrapper[4846]: I1005 08:21:15.265700 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2059932a-7f7d-45ff-8a75-333a6711afdf-logs\") pod \"glance-default-external-api-0\" (UID: \"2059932a-7f7d-45ff-8a75-333a6711afdf\") " pod="openstack/glance-default-external-api-0" Oct 05 08:21:15 crc kubenswrapper[4846]: I1005 08:21:15.265762 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/2059932a-7f7d-45ff-8a75-333a6711afdf-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"2059932a-7f7d-45ff-8a75-333a6711afdf\") " pod="openstack/glance-default-external-api-0" Oct 05 08:21:15 crc kubenswrapper[4846]: I1005 08:21:15.265791 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8xlmx\" (UniqueName: \"kubernetes.io/projected/2059932a-7f7d-45ff-8a75-333a6711afdf-kube-api-access-8xlmx\") pod \"glance-default-external-api-0\" (UID: \"2059932a-7f7d-45ff-8a75-333a6711afdf\") " pod="openstack/glance-default-external-api-0" Oct 05 08:21:15 crc kubenswrapper[4846]: I1005 08:21:15.265828 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2059932a-7f7d-45ff-8a75-333a6711afdf-scripts\") pod \"glance-default-external-api-0\" (UID: \"2059932a-7f7d-45ff-8a75-333a6711afdf\") " pod="openstack/glance-default-external-api-0" Oct 05 08:21:15 crc kubenswrapper[4846]: I1005 08:21:15.282483 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-59bf5dd889-89l2d"] Oct 05 08:21:15 crc kubenswrapper[4846]: I1005 08:21:15.367288 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2059932a-7f7d-45ff-8a75-333a6711afdf-logs\") pod \"glance-default-external-api-0\" (UID: \"2059932a-7f7d-45ff-8a75-333a6711afdf\") " pod="openstack/glance-default-external-api-0" Oct 05 08:21:15 crc kubenswrapper[4846]: I1005 08:21:15.367350 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/2059932a-7f7d-45ff-8a75-333a6711afdf-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"2059932a-7f7d-45ff-8a75-333a6711afdf\") " pod="openstack/glance-default-external-api-0" Oct 05 08:21:15 crc kubenswrapper[4846]: I1005 08:21:15.367382 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8xlmx\" (UniqueName: \"kubernetes.io/projected/2059932a-7f7d-45ff-8a75-333a6711afdf-kube-api-access-8xlmx\") pod \"glance-default-external-api-0\" (UID: \"2059932a-7f7d-45ff-8a75-333a6711afdf\") " pod="openstack/glance-default-external-api-0" Oct 05 08:21:15 crc kubenswrapper[4846]: I1005 08:21:15.367426 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/363aa87a-56da-4c6b-b6f6-82e66ed412ce-dns-svc\") pod \"dnsmasq-dns-59bf5dd889-89l2d\" (UID: \"363aa87a-56da-4c6b-b6f6-82e66ed412ce\") " pod="openstack/dnsmasq-dns-59bf5dd889-89l2d" Oct 05 08:21:15 crc kubenswrapper[4846]: I1005 08:21:15.367453 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2059932a-7f7d-45ff-8a75-333a6711afdf-scripts\") pod \"glance-default-external-api-0\" (UID: \"2059932a-7f7d-45ff-8a75-333a6711afdf\") " pod="openstack/glance-default-external-api-0" Oct 05 08:21:15 crc kubenswrapper[4846]: I1005 08:21:15.367530 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2059932a-7f7d-45ff-8a75-333a6711afdf-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"2059932a-7f7d-45ff-8a75-333a6711afdf\") " pod="openstack/glance-default-external-api-0" Oct 05 08:21:15 crc kubenswrapper[4846]: I1005 08:21:15.367561 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/363aa87a-56da-4c6b-b6f6-82e66ed412ce-config\") pod \"dnsmasq-dns-59bf5dd889-89l2d\" (UID: \"363aa87a-56da-4c6b-b6f6-82e66ed412ce\") " pod="openstack/dnsmasq-dns-59bf5dd889-89l2d" Oct 05 08:21:15 crc kubenswrapper[4846]: I1005 08:21:15.367583 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4tfjp\" (UniqueName: \"kubernetes.io/projected/363aa87a-56da-4c6b-b6f6-82e66ed412ce-kube-api-access-4tfjp\") pod \"dnsmasq-dns-59bf5dd889-89l2d\" (UID: \"363aa87a-56da-4c6b-b6f6-82e66ed412ce\") " pod="openstack/dnsmasq-dns-59bf5dd889-89l2d" Oct 05 08:21:15 crc kubenswrapper[4846]: I1005 08:21:15.367644 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/363aa87a-56da-4c6b-b6f6-82e66ed412ce-ovsdbserver-sb\") pod \"dnsmasq-dns-59bf5dd889-89l2d\" (UID: \"363aa87a-56da-4c6b-b6f6-82e66ed412ce\") " pod="openstack/dnsmasq-dns-59bf5dd889-89l2d" Oct 05 08:21:15 crc kubenswrapper[4846]: I1005 08:21:15.367678 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/363aa87a-56da-4c6b-b6f6-82e66ed412ce-ovsdbserver-nb\") pod \"dnsmasq-dns-59bf5dd889-89l2d\" (UID: \"363aa87a-56da-4c6b-b6f6-82e66ed412ce\") " pod="openstack/dnsmasq-dns-59bf5dd889-89l2d" Oct 05 08:21:15 crc kubenswrapper[4846]: I1005 08:21:15.367703 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2059932a-7f7d-45ff-8a75-333a6711afdf-config-data\") pod \"glance-default-external-api-0\" (UID: \"2059932a-7f7d-45ff-8a75-333a6711afdf\") " pod="openstack/glance-default-external-api-0" Oct 05 08:21:15 crc kubenswrapper[4846]: I1005 08:21:15.367853 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2059932a-7f7d-45ff-8a75-333a6711afdf-logs\") pod \"glance-default-external-api-0\" (UID: \"2059932a-7f7d-45ff-8a75-333a6711afdf\") " pod="openstack/glance-default-external-api-0" Oct 05 08:21:15 crc kubenswrapper[4846]: I1005 08:21:15.371288 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/2059932a-7f7d-45ff-8a75-333a6711afdf-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"2059932a-7f7d-45ff-8a75-333a6711afdf\") " pod="openstack/glance-default-external-api-0" Oct 05 08:21:15 crc kubenswrapper[4846]: I1005 08:21:15.375367 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2059932a-7f7d-45ff-8a75-333a6711afdf-scripts\") pod \"glance-default-external-api-0\" (UID: \"2059932a-7f7d-45ff-8a75-333a6711afdf\") " pod="openstack/glance-default-external-api-0" Oct 05 08:21:15 crc kubenswrapper[4846]: I1005 08:21:15.387223 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2059932a-7f7d-45ff-8a75-333a6711afdf-config-data\") pod \"glance-default-external-api-0\" (UID: \"2059932a-7f7d-45ff-8a75-333a6711afdf\") " pod="openstack/glance-default-external-api-0" Oct 05 08:21:15 crc kubenswrapper[4846]: I1005 08:21:15.398034 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2059932a-7f7d-45ff-8a75-333a6711afdf-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"2059932a-7f7d-45ff-8a75-333a6711afdf\") " pod="openstack/glance-default-external-api-0" Oct 05 08:21:15 crc kubenswrapper[4846]: I1005 08:21:15.409969 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8xlmx\" (UniqueName: \"kubernetes.io/projected/2059932a-7f7d-45ff-8a75-333a6711afdf-kube-api-access-8xlmx\") pod \"glance-default-external-api-0\" (UID: \"2059932a-7f7d-45ff-8a75-333a6711afdf\") " pod="openstack/glance-default-external-api-0" Oct 05 08:21:15 crc kubenswrapper[4846]: I1005 08:21:15.464220 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 05 08:21:15 crc kubenswrapper[4846]: I1005 08:21:15.465845 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 05 08:21:15 crc kubenswrapper[4846]: I1005 08:21:15.471301 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/363aa87a-56da-4c6b-b6f6-82e66ed412ce-dns-svc\") pod \"dnsmasq-dns-59bf5dd889-89l2d\" (UID: \"363aa87a-56da-4c6b-b6f6-82e66ed412ce\") " pod="openstack/dnsmasq-dns-59bf5dd889-89l2d" Oct 05 08:21:15 crc kubenswrapper[4846]: I1005 08:21:15.471708 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/363aa87a-56da-4c6b-b6f6-82e66ed412ce-config\") pod \"dnsmasq-dns-59bf5dd889-89l2d\" (UID: \"363aa87a-56da-4c6b-b6f6-82e66ed412ce\") " pod="openstack/dnsmasq-dns-59bf5dd889-89l2d" Oct 05 08:21:15 crc kubenswrapper[4846]: I1005 08:21:15.471730 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4tfjp\" (UniqueName: \"kubernetes.io/projected/363aa87a-56da-4c6b-b6f6-82e66ed412ce-kube-api-access-4tfjp\") pod \"dnsmasq-dns-59bf5dd889-89l2d\" (UID: \"363aa87a-56da-4c6b-b6f6-82e66ed412ce\") " pod="openstack/dnsmasq-dns-59bf5dd889-89l2d" Oct 05 08:21:15 crc kubenswrapper[4846]: I1005 08:21:15.471775 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/363aa87a-56da-4c6b-b6f6-82e66ed412ce-ovsdbserver-sb\") pod \"dnsmasq-dns-59bf5dd889-89l2d\" (UID: \"363aa87a-56da-4c6b-b6f6-82e66ed412ce\") " pod="openstack/dnsmasq-dns-59bf5dd889-89l2d" Oct 05 08:21:15 crc kubenswrapper[4846]: I1005 08:21:15.471801 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/363aa87a-56da-4c6b-b6f6-82e66ed412ce-ovsdbserver-nb\") pod \"dnsmasq-dns-59bf5dd889-89l2d\" (UID: \"363aa87a-56da-4c6b-b6f6-82e66ed412ce\") " pod="openstack/dnsmasq-dns-59bf5dd889-89l2d" Oct 05 08:21:15 crc kubenswrapper[4846]: I1005 08:21:15.472794 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/363aa87a-56da-4c6b-b6f6-82e66ed412ce-ovsdbserver-nb\") pod \"dnsmasq-dns-59bf5dd889-89l2d\" (UID: \"363aa87a-56da-4c6b-b6f6-82e66ed412ce\") " pod="openstack/dnsmasq-dns-59bf5dd889-89l2d" Oct 05 08:21:15 crc kubenswrapper[4846]: I1005 08:21:15.473197 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/363aa87a-56da-4c6b-b6f6-82e66ed412ce-dns-svc\") pod \"dnsmasq-dns-59bf5dd889-89l2d\" (UID: \"363aa87a-56da-4c6b-b6f6-82e66ed412ce\") " pod="openstack/dnsmasq-dns-59bf5dd889-89l2d" Oct 05 08:21:15 crc kubenswrapper[4846]: I1005 08:21:15.473348 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/363aa87a-56da-4c6b-b6f6-82e66ed412ce-config\") pod \"dnsmasq-dns-59bf5dd889-89l2d\" (UID: \"363aa87a-56da-4c6b-b6f6-82e66ed412ce\") " pod="openstack/dnsmasq-dns-59bf5dd889-89l2d" Oct 05 08:21:15 crc kubenswrapper[4846]: I1005 08:21:15.473462 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Oct 05 08:21:15 crc kubenswrapper[4846]: I1005 08:21:15.474297 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/363aa87a-56da-4c6b-b6f6-82e66ed412ce-ovsdbserver-sb\") pod \"dnsmasq-dns-59bf5dd889-89l2d\" (UID: \"363aa87a-56da-4c6b-b6f6-82e66ed412ce\") " pod="openstack/dnsmasq-dns-59bf5dd889-89l2d" Oct 05 08:21:15 crc kubenswrapper[4846]: I1005 08:21:15.475634 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 05 08:21:15 crc kubenswrapper[4846]: I1005 08:21:15.503830 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4tfjp\" (UniqueName: \"kubernetes.io/projected/363aa87a-56da-4c6b-b6f6-82e66ed412ce-kube-api-access-4tfjp\") pod \"dnsmasq-dns-59bf5dd889-89l2d\" (UID: \"363aa87a-56da-4c6b-b6f6-82e66ed412ce\") " pod="openstack/dnsmasq-dns-59bf5dd889-89l2d" Oct 05 08:21:15 crc kubenswrapper[4846]: I1005 08:21:15.511242 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 05 08:21:15 crc kubenswrapper[4846]: I1005 08:21:15.573249 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/976cec21-fc6f-440f-b2ea-a181f52fd624-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"976cec21-fc6f-440f-b2ea-a181f52fd624\") " pod="openstack/glance-default-internal-api-0" Oct 05 08:21:15 crc kubenswrapper[4846]: I1005 08:21:15.573557 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/976cec21-fc6f-440f-b2ea-a181f52fd624-config-data\") pod \"glance-default-internal-api-0\" (UID: \"976cec21-fc6f-440f-b2ea-a181f52fd624\") " pod="openstack/glance-default-internal-api-0" Oct 05 08:21:15 crc kubenswrapper[4846]: I1005 08:21:15.573647 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/976cec21-fc6f-440f-b2ea-a181f52fd624-scripts\") pod \"glance-default-internal-api-0\" (UID: \"976cec21-fc6f-440f-b2ea-a181f52fd624\") " pod="openstack/glance-default-internal-api-0" Oct 05 08:21:15 crc kubenswrapper[4846]: I1005 08:21:15.573755 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6n879\" (UniqueName: \"kubernetes.io/projected/976cec21-fc6f-440f-b2ea-a181f52fd624-kube-api-access-6n879\") pod \"glance-default-internal-api-0\" (UID: \"976cec21-fc6f-440f-b2ea-a181f52fd624\") " pod="openstack/glance-default-internal-api-0" Oct 05 08:21:15 crc kubenswrapper[4846]: I1005 08:21:15.573845 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/976cec21-fc6f-440f-b2ea-a181f52fd624-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"976cec21-fc6f-440f-b2ea-a181f52fd624\") " pod="openstack/glance-default-internal-api-0" Oct 05 08:21:15 crc kubenswrapper[4846]: I1005 08:21:15.573926 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/976cec21-fc6f-440f-b2ea-a181f52fd624-logs\") pod \"glance-default-internal-api-0\" (UID: \"976cec21-fc6f-440f-b2ea-a181f52fd624\") " pod="openstack/glance-default-internal-api-0" Oct 05 08:21:15 crc kubenswrapper[4846]: I1005 08:21:15.583583 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-59bf5dd889-89l2d" Oct 05 08:21:15 crc kubenswrapper[4846]: I1005 08:21:15.676322 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6n879\" (UniqueName: \"kubernetes.io/projected/976cec21-fc6f-440f-b2ea-a181f52fd624-kube-api-access-6n879\") pod \"glance-default-internal-api-0\" (UID: \"976cec21-fc6f-440f-b2ea-a181f52fd624\") " pod="openstack/glance-default-internal-api-0" Oct 05 08:21:15 crc kubenswrapper[4846]: I1005 08:21:15.676399 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/976cec21-fc6f-440f-b2ea-a181f52fd624-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"976cec21-fc6f-440f-b2ea-a181f52fd624\") " pod="openstack/glance-default-internal-api-0" Oct 05 08:21:15 crc kubenswrapper[4846]: I1005 08:21:15.676434 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/976cec21-fc6f-440f-b2ea-a181f52fd624-logs\") pod \"glance-default-internal-api-0\" (UID: \"976cec21-fc6f-440f-b2ea-a181f52fd624\") " pod="openstack/glance-default-internal-api-0" Oct 05 08:21:15 crc kubenswrapper[4846]: I1005 08:21:15.676467 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/976cec21-fc6f-440f-b2ea-a181f52fd624-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"976cec21-fc6f-440f-b2ea-a181f52fd624\") " pod="openstack/glance-default-internal-api-0" Oct 05 08:21:15 crc kubenswrapper[4846]: I1005 08:21:15.676514 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/976cec21-fc6f-440f-b2ea-a181f52fd624-config-data\") pod \"glance-default-internal-api-0\" (UID: \"976cec21-fc6f-440f-b2ea-a181f52fd624\") " pod="openstack/glance-default-internal-api-0" Oct 05 08:21:15 crc kubenswrapper[4846]: I1005 08:21:15.676548 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/976cec21-fc6f-440f-b2ea-a181f52fd624-scripts\") pod \"glance-default-internal-api-0\" (UID: \"976cec21-fc6f-440f-b2ea-a181f52fd624\") " pod="openstack/glance-default-internal-api-0" Oct 05 08:21:15 crc kubenswrapper[4846]: I1005 08:21:15.677377 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/976cec21-fc6f-440f-b2ea-a181f52fd624-logs\") pod \"glance-default-internal-api-0\" (UID: \"976cec21-fc6f-440f-b2ea-a181f52fd624\") " pod="openstack/glance-default-internal-api-0" Oct 05 08:21:15 crc kubenswrapper[4846]: I1005 08:21:15.677447 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/976cec21-fc6f-440f-b2ea-a181f52fd624-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"976cec21-fc6f-440f-b2ea-a181f52fd624\") " pod="openstack/glance-default-internal-api-0" Oct 05 08:21:15 crc kubenswrapper[4846]: I1005 08:21:15.680818 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/976cec21-fc6f-440f-b2ea-a181f52fd624-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"976cec21-fc6f-440f-b2ea-a181f52fd624\") " pod="openstack/glance-default-internal-api-0" Oct 05 08:21:15 crc kubenswrapper[4846]: I1005 08:21:15.681211 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/976cec21-fc6f-440f-b2ea-a181f52fd624-config-data\") pod \"glance-default-internal-api-0\" (UID: \"976cec21-fc6f-440f-b2ea-a181f52fd624\") " pod="openstack/glance-default-internal-api-0" Oct 05 08:21:15 crc kubenswrapper[4846]: I1005 08:21:15.685837 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/976cec21-fc6f-440f-b2ea-a181f52fd624-scripts\") pod \"glance-default-internal-api-0\" (UID: \"976cec21-fc6f-440f-b2ea-a181f52fd624\") " pod="openstack/glance-default-internal-api-0" Oct 05 08:21:15 crc kubenswrapper[4846]: I1005 08:21:15.708642 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6n879\" (UniqueName: \"kubernetes.io/projected/976cec21-fc6f-440f-b2ea-a181f52fd624-kube-api-access-6n879\") pod \"glance-default-internal-api-0\" (UID: \"976cec21-fc6f-440f-b2ea-a181f52fd624\") " pod="openstack/glance-default-internal-api-0" Oct 05 08:21:15 crc kubenswrapper[4846]: I1005 08:21:15.890134 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 05 08:21:16 crc kubenswrapper[4846]: I1005 08:21:16.164817 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-59bf5dd889-89l2d"] Oct 05 08:21:16 crc kubenswrapper[4846]: I1005 08:21:16.176236 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 05 08:21:16 crc kubenswrapper[4846]: I1005 08:21:16.350896 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 05 08:21:16 crc kubenswrapper[4846]: I1005 08:21:16.466217 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 05 08:21:16 crc kubenswrapper[4846]: W1005 08:21:16.481101 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod976cec21_fc6f_440f_b2ea_a181f52fd624.slice/crio-eb016c833e42626c935f4e2b8ac0adbcb01597a1e856ac9c420af07bbd81cbb2 WatchSource:0}: Error finding container eb016c833e42626c935f4e2b8ac0adbcb01597a1e856ac9c420af07bbd81cbb2: Status 404 returned error can't find the container with id eb016c833e42626c935f4e2b8ac0adbcb01597a1e856ac9c420af07bbd81cbb2 Oct 05 08:21:16 crc kubenswrapper[4846]: I1005 08:21:16.848340 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"2059932a-7f7d-45ff-8a75-333a6711afdf","Type":"ContainerStarted","Data":"8ef8c322f6da2990f202fe71373fc3bd9b047f0d55a012c8b32032780822b211"} Oct 05 08:21:16 crc kubenswrapper[4846]: I1005 08:21:16.848745 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"2059932a-7f7d-45ff-8a75-333a6711afdf","Type":"ContainerStarted","Data":"6b404493c6db8bac188f9631c29896f27f3a3358dd283b65b95d1c37548a4200"} Oct 05 08:21:16 crc kubenswrapper[4846]: I1005 08:21:16.852247 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"976cec21-fc6f-440f-b2ea-a181f52fd624","Type":"ContainerStarted","Data":"eb016c833e42626c935f4e2b8ac0adbcb01597a1e856ac9c420af07bbd81cbb2"} Oct 05 08:21:16 crc kubenswrapper[4846]: I1005 08:21:16.855291 4846 generic.go:334] "Generic (PLEG): container finished" podID="363aa87a-56da-4c6b-b6f6-82e66ed412ce" containerID="20a4c275f58000d4bd34403062d8b3fcc48b6c65b849273bf11d495d67cfe3ba" exitCode=0 Oct 05 08:21:16 crc kubenswrapper[4846]: I1005 08:21:16.855334 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59bf5dd889-89l2d" event={"ID":"363aa87a-56da-4c6b-b6f6-82e66ed412ce","Type":"ContainerDied","Data":"20a4c275f58000d4bd34403062d8b3fcc48b6c65b849273bf11d495d67cfe3ba"} Oct 05 08:21:16 crc kubenswrapper[4846]: I1005 08:21:16.855357 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59bf5dd889-89l2d" event={"ID":"363aa87a-56da-4c6b-b6f6-82e66ed412ce","Type":"ContainerStarted","Data":"f1a42bc1fb2486eb2d7c653376a2277aa7553d2f80858473427c9bc5c170a990"} Oct 05 08:21:17 crc kubenswrapper[4846]: I1005 08:21:17.308100 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 05 08:21:17 crc kubenswrapper[4846]: I1005 08:21:17.868673 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"2059932a-7f7d-45ff-8a75-333a6711afdf","Type":"ContainerStarted","Data":"473b4ffc35fbaac4fee2179df1d2e7ee30f434ac41d6e8f1645ea47428180b34"} Oct 05 08:21:17 crc kubenswrapper[4846]: I1005 08:21:17.868769 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="2059932a-7f7d-45ff-8a75-333a6711afdf" containerName="glance-log" containerID="cri-o://8ef8c322f6da2990f202fe71373fc3bd9b047f0d55a012c8b32032780822b211" gracePeriod=30 Oct 05 08:21:17 crc kubenswrapper[4846]: I1005 08:21:17.868842 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="2059932a-7f7d-45ff-8a75-333a6711afdf" containerName="glance-httpd" containerID="cri-o://473b4ffc35fbaac4fee2179df1d2e7ee30f434ac41d6e8f1645ea47428180b34" gracePeriod=30 Oct 05 08:21:17 crc kubenswrapper[4846]: I1005 08:21:17.871449 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"976cec21-fc6f-440f-b2ea-a181f52fd624","Type":"ContainerStarted","Data":"c4440fe126de02dd5b2a1f044ad7cdfd08c87e2f809ed07e76700fa7299a811e"} Oct 05 08:21:17 crc kubenswrapper[4846]: I1005 08:21:17.871499 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"976cec21-fc6f-440f-b2ea-a181f52fd624","Type":"ContainerStarted","Data":"03f5a7ce92d7760417b7debfdc2b9a753f6e29e6da93d51623b51a65bab02dd9"} Oct 05 08:21:17 crc kubenswrapper[4846]: I1005 08:21:17.871669 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="976cec21-fc6f-440f-b2ea-a181f52fd624" containerName="glance-log" containerID="cri-o://03f5a7ce92d7760417b7debfdc2b9a753f6e29e6da93d51623b51a65bab02dd9" gracePeriod=30 Oct 05 08:21:17 crc kubenswrapper[4846]: I1005 08:21:17.871773 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="976cec21-fc6f-440f-b2ea-a181f52fd624" containerName="glance-httpd" containerID="cri-o://c4440fe126de02dd5b2a1f044ad7cdfd08c87e2f809ed07e76700fa7299a811e" gracePeriod=30 Oct 05 08:21:17 crc kubenswrapper[4846]: I1005 08:21:17.879136 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59bf5dd889-89l2d" event={"ID":"363aa87a-56da-4c6b-b6f6-82e66ed412ce","Type":"ContainerStarted","Data":"5fe82ad5b1b763d4dac8451b390a7ca0634e2c75837c1e2773e60ca5e6dae11c"} Oct 05 08:21:17 crc kubenswrapper[4846]: I1005 08:21:17.879257 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-59bf5dd889-89l2d" Oct 05 08:21:17 crc kubenswrapper[4846]: I1005 08:21:17.922359 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=2.922335853 podStartE2EDuration="2.922335853s" podCreationTimestamp="2025-10-05 08:21:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 08:21:17.919533808 +0000 UTC m=+5600.160386603" watchObservedRunningTime="2025-10-05 08:21:17.922335853 +0000 UTC m=+5600.163188628" Oct 05 08:21:17 crc kubenswrapper[4846]: I1005 08:21:17.925361 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=2.9253430639999998 podStartE2EDuration="2.925343064s" podCreationTimestamp="2025-10-05 08:21:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 08:21:17.903643833 +0000 UTC m=+5600.144496598" watchObservedRunningTime="2025-10-05 08:21:17.925343064 +0000 UTC m=+5600.166195839" Oct 05 08:21:17 crc kubenswrapper[4846]: I1005 08:21:17.942538 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-59bf5dd889-89l2d" podStartSLOduration=2.942519184 podStartE2EDuration="2.942519184s" podCreationTimestamp="2025-10-05 08:21:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 08:21:17.93601425 +0000 UTC m=+5600.176867025" watchObservedRunningTime="2025-10-05 08:21:17.942519184 +0000 UTC m=+5600.183371959" Oct 05 08:21:18 crc kubenswrapper[4846]: I1005 08:21:18.485981 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 05 08:21:18 crc kubenswrapper[4846]: I1005 08:21:18.549653 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6n879\" (UniqueName: \"kubernetes.io/projected/976cec21-fc6f-440f-b2ea-a181f52fd624-kube-api-access-6n879\") pod \"976cec21-fc6f-440f-b2ea-a181f52fd624\" (UID: \"976cec21-fc6f-440f-b2ea-a181f52fd624\") " Oct 05 08:21:18 crc kubenswrapper[4846]: I1005 08:21:18.549786 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/976cec21-fc6f-440f-b2ea-a181f52fd624-config-data\") pod \"976cec21-fc6f-440f-b2ea-a181f52fd624\" (UID: \"976cec21-fc6f-440f-b2ea-a181f52fd624\") " Oct 05 08:21:18 crc kubenswrapper[4846]: I1005 08:21:18.549808 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/976cec21-fc6f-440f-b2ea-a181f52fd624-scripts\") pod \"976cec21-fc6f-440f-b2ea-a181f52fd624\" (UID: \"976cec21-fc6f-440f-b2ea-a181f52fd624\") " Oct 05 08:21:18 crc kubenswrapper[4846]: I1005 08:21:18.549850 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/976cec21-fc6f-440f-b2ea-a181f52fd624-httpd-run\") pod \"976cec21-fc6f-440f-b2ea-a181f52fd624\" (UID: \"976cec21-fc6f-440f-b2ea-a181f52fd624\") " Oct 05 08:21:18 crc kubenswrapper[4846]: I1005 08:21:18.549885 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/976cec21-fc6f-440f-b2ea-a181f52fd624-logs\") pod \"976cec21-fc6f-440f-b2ea-a181f52fd624\" (UID: \"976cec21-fc6f-440f-b2ea-a181f52fd624\") " Oct 05 08:21:18 crc kubenswrapper[4846]: I1005 08:21:18.549969 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/976cec21-fc6f-440f-b2ea-a181f52fd624-combined-ca-bundle\") pod \"976cec21-fc6f-440f-b2ea-a181f52fd624\" (UID: \"976cec21-fc6f-440f-b2ea-a181f52fd624\") " Oct 05 08:21:18 crc kubenswrapper[4846]: I1005 08:21:18.550391 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/976cec21-fc6f-440f-b2ea-a181f52fd624-logs" (OuterVolumeSpecName: "logs") pod "976cec21-fc6f-440f-b2ea-a181f52fd624" (UID: "976cec21-fc6f-440f-b2ea-a181f52fd624"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 08:21:18 crc kubenswrapper[4846]: I1005 08:21:18.551070 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/976cec21-fc6f-440f-b2ea-a181f52fd624-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "976cec21-fc6f-440f-b2ea-a181f52fd624" (UID: "976cec21-fc6f-440f-b2ea-a181f52fd624"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 08:21:18 crc kubenswrapper[4846]: I1005 08:21:18.551675 4846 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/976cec21-fc6f-440f-b2ea-a181f52fd624-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 05 08:21:18 crc kubenswrapper[4846]: I1005 08:21:18.551706 4846 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/976cec21-fc6f-440f-b2ea-a181f52fd624-logs\") on node \"crc\" DevicePath \"\"" Oct 05 08:21:18 crc kubenswrapper[4846]: I1005 08:21:18.557731 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/976cec21-fc6f-440f-b2ea-a181f52fd624-scripts" (OuterVolumeSpecName: "scripts") pod "976cec21-fc6f-440f-b2ea-a181f52fd624" (UID: "976cec21-fc6f-440f-b2ea-a181f52fd624"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:21:18 crc kubenswrapper[4846]: I1005 08:21:18.557778 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/976cec21-fc6f-440f-b2ea-a181f52fd624-kube-api-access-6n879" (OuterVolumeSpecName: "kube-api-access-6n879") pod "976cec21-fc6f-440f-b2ea-a181f52fd624" (UID: "976cec21-fc6f-440f-b2ea-a181f52fd624"). InnerVolumeSpecName "kube-api-access-6n879". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:21:18 crc kubenswrapper[4846]: I1005 08:21:18.581247 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/976cec21-fc6f-440f-b2ea-a181f52fd624-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "976cec21-fc6f-440f-b2ea-a181f52fd624" (UID: "976cec21-fc6f-440f-b2ea-a181f52fd624"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:21:18 crc kubenswrapper[4846]: I1005 08:21:18.607440 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/976cec21-fc6f-440f-b2ea-a181f52fd624-config-data" (OuterVolumeSpecName: "config-data") pod "976cec21-fc6f-440f-b2ea-a181f52fd624" (UID: "976cec21-fc6f-440f-b2ea-a181f52fd624"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:21:18 crc kubenswrapper[4846]: I1005 08:21:18.639819 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 05 08:21:18 crc kubenswrapper[4846]: I1005 08:21:18.653402 4846 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/976cec21-fc6f-440f-b2ea-a181f52fd624-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 08:21:18 crc kubenswrapper[4846]: I1005 08:21:18.653432 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6n879\" (UniqueName: \"kubernetes.io/projected/976cec21-fc6f-440f-b2ea-a181f52fd624-kube-api-access-6n879\") on node \"crc\" DevicePath \"\"" Oct 05 08:21:18 crc kubenswrapper[4846]: I1005 08:21:18.653441 4846 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/976cec21-fc6f-440f-b2ea-a181f52fd624-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 08:21:18 crc kubenswrapper[4846]: I1005 08:21:18.653452 4846 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/976cec21-fc6f-440f-b2ea-a181f52fd624-scripts\") on node \"crc\" DevicePath \"\"" Oct 05 08:21:18 crc kubenswrapper[4846]: I1005 08:21:18.754114 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/2059932a-7f7d-45ff-8a75-333a6711afdf-httpd-run\") pod \"2059932a-7f7d-45ff-8a75-333a6711afdf\" (UID: \"2059932a-7f7d-45ff-8a75-333a6711afdf\") " Oct 05 08:21:18 crc kubenswrapper[4846]: I1005 08:21:18.754210 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2059932a-7f7d-45ff-8a75-333a6711afdf-config-data\") pod \"2059932a-7f7d-45ff-8a75-333a6711afdf\" (UID: \"2059932a-7f7d-45ff-8a75-333a6711afdf\") " Oct 05 08:21:18 crc kubenswrapper[4846]: I1005 08:21:18.754305 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2059932a-7f7d-45ff-8a75-333a6711afdf-combined-ca-bundle\") pod \"2059932a-7f7d-45ff-8a75-333a6711afdf\" (UID: \"2059932a-7f7d-45ff-8a75-333a6711afdf\") " Oct 05 08:21:18 crc kubenswrapper[4846]: I1005 08:21:18.754333 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8xlmx\" (UniqueName: \"kubernetes.io/projected/2059932a-7f7d-45ff-8a75-333a6711afdf-kube-api-access-8xlmx\") pod \"2059932a-7f7d-45ff-8a75-333a6711afdf\" (UID: \"2059932a-7f7d-45ff-8a75-333a6711afdf\") " Oct 05 08:21:18 crc kubenswrapper[4846]: I1005 08:21:18.754373 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2059932a-7f7d-45ff-8a75-333a6711afdf-logs\") pod \"2059932a-7f7d-45ff-8a75-333a6711afdf\" (UID: \"2059932a-7f7d-45ff-8a75-333a6711afdf\") " Oct 05 08:21:18 crc kubenswrapper[4846]: I1005 08:21:18.754600 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2059932a-7f7d-45ff-8a75-333a6711afdf-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "2059932a-7f7d-45ff-8a75-333a6711afdf" (UID: "2059932a-7f7d-45ff-8a75-333a6711afdf"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 08:21:18 crc kubenswrapper[4846]: I1005 08:21:18.754771 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2059932a-7f7d-45ff-8a75-333a6711afdf-logs" (OuterVolumeSpecName: "logs") pod "2059932a-7f7d-45ff-8a75-333a6711afdf" (UID: "2059932a-7f7d-45ff-8a75-333a6711afdf"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 08:21:18 crc kubenswrapper[4846]: I1005 08:21:18.754896 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2059932a-7f7d-45ff-8a75-333a6711afdf-scripts\") pod \"2059932a-7f7d-45ff-8a75-333a6711afdf\" (UID: \"2059932a-7f7d-45ff-8a75-333a6711afdf\") " Oct 05 08:21:18 crc kubenswrapper[4846]: I1005 08:21:18.755610 4846 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2059932a-7f7d-45ff-8a75-333a6711afdf-logs\") on node \"crc\" DevicePath \"\"" Oct 05 08:21:18 crc kubenswrapper[4846]: I1005 08:21:18.755626 4846 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/2059932a-7f7d-45ff-8a75-333a6711afdf-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 05 08:21:18 crc kubenswrapper[4846]: I1005 08:21:18.758283 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2059932a-7f7d-45ff-8a75-333a6711afdf-scripts" (OuterVolumeSpecName: "scripts") pod "2059932a-7f7d-45ff-8a75-333a6711afdf" (UID: "2059932a-7f7d-45ff-8a75-333a6711afdf"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:21:18 crc kubenswrapper[4846]: I1005 08:21:18.759496 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2059932a-7f7d-45ff-8a75-333a6711afdf-kube-api-access-8xlmx" (OuterVolumeSpecName: "kube-api-access-8xlmx") pod "2059932a-7f7d-45ff-8a75-333a6711afdf" (UID: "2059932a-7f7d-45ff-8a75-333a6711afdf"). InnerVolumeSpecName "kube-api-access-8xlmx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:21:18 crc kubenswrapper[4846]: I1005 08:21:18.782335 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2059932a-7f7d-45ff-8a75-333a6711afdf-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2059932a-7f7d-45ff-8a75-333a6711afdf" (UID: "2059932a-7f7d-45ff-8a75-333a6711afdf"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:21:18 crc kubenswrapper[4846]: I1005 08:21:18.797569 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2059932a-7f7d-45ff-8a75-333a6711afdf-config-data" (OuterVolumeSpecName: "config-data") pod "2059932a-7f7d-45ff-8a75-333a6711afdf" (UID: "2059932a-7f7d-45ff-8a75-333a6711afdf"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:21:18 crc kubenswrapper[4846]: I1005 08:21:18.856687 4846 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2059932a-7f7d-45ff-8a75-333a6711afdf-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 08:21:18 crc kubenswrapper[4846]: I1005 08:21:18.856715 4846 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2059932a-7f7d-45ff-8a75-333a6711afdf-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 08:21:18 crc kubenswrapper[4846]: I1005 08:21:18.856728 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8xlmx\" (UniqueName: \"kubernetes.io/projected/2059932a-7f7d-45ff-8a75-333a6711afdf-kube-api-access-8xlmx\") on node \"crc\" DevicePath \"\"" Oct 05 08:21:18 crc kubenswrapper[4846]: I1005 08:21:18.856738 4846 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2059932a-7f7d-45ff-8a75-333a6711afdf-scripts\") on node \"crc\" DevicePath \"\"" Oct 05 08:21:18 crc kubenswrapper[4846]: I1005 08:21:18.889218 4846 generic.go:334] "Generic (PLEG): container finished" podID="976cec21-fc6f-440f-b2ea-a181f52fd624" containerID="c4440fe126de02dd5b2a1f044ad7cdfd08c87e2f809ed07e76700fa7299a811e" exitCode=143 Oct 05 08:21:18 crc kubenswrapper[4846]: I1005 08:21:18.889267 4846 generic.go:334] "Generic (PLEG): container finished" podID="976cec21-fc6f-440f-b2ea-a181f52fd624" containerID="03f5a7ce92d7760417b7debfdc2b9a753f6e29e6da93d51623b51a65bab02dd9" exitCode=143 Oct 05 08:21:18 crc kubenswrapper[4846]: I1005 08:21:18.889313 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"976cec21-fc6f-440f-b2ea-a181f52fd624","Type":"ContainerDied","Data":"c4440fe126de02dd5b2a1f044ad7cdfd08c87e2f809ed07e76700fa7299a811e"} Oct 05 08:21:18 crc kubenswrapper[4846]: I1005 08:21:18.889384 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"976cec21-fc6f-440f-b2ea-a181f52fd624","Type":"ContainerDied","Data":"03f5a7ce92d7760417b7debfdc2b9a753f6e29e6da93d51623b51a65bab02dd9"} Oct 05 08:21:18 crc kubenswrapper[4846]: I1005 08:21:18.889410 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"976cec21-fc6f-440f-b2ea-a181f52fd624","Type":"ContainerDied","Data":"eb016c833e42626c935f4e2b8ac0adbcb01597a1e856ac9c420af07bbd81cbb2"} Oct 05 08:21:18 crc kubenswrapper[4846]: I1005 08:21:18.889431 4846 scope.go:117] "RemoveContainer" containerID="c4440fe126de02dd5b2a1f044ad7cdfd08c87e2f809ed07e76700fa7299a811e" Oct 05 08:21:18 crc kubenswrapper[4846]: I1005 08:21:18.889455 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 05 08:21:18 crc kubenswrapper[4846]: I1005 08:21:18.892061 4846 generic.go:334] "Generic (PLEG): container finished" podID="2059932a-7f7d-45ff-8a75-333a6711afdf" containerID="473b4ffc35fbaac4fee2179df1d2e7ee30f434ac41d6e8f1645ea47428180b34" exitCode=0 Oct 05 08:21:18 crc kubenswrapper[4846]: I1005 08:21:18.892089 4846 generic.go:334] "Generic (PLEG): container finished" podID="2059932a-7f7d-45ff-8a75-333a6711afdf" containerID="8ef8c322f6da2990f202fe71373fc3bd9b047f0d55a012c8b32032780822b211" exitCode=143 Oct 05 08:21:18 crc kubenswrapper[4846]: I1005 08:21:18.892450 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 05 08:21:18 crc kubenswrapper[4846]: I1005 08:21:18.892444 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"2059932a-7f7d-45ff-8a75-333a6711afdf","Type":"ContainerDied","Data":"473b4ffc35fbaac4fee2179df1d2e7ee30f434ac41d6e8f1645ea47428180b34"} Oct 05 08:21:18 crc kubenswrapper[4846]: I1005 08:21:18.892639 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"2059932a-7f7d-45ff-8a75-333a6711afdf","Type":"ContainerDied","Data":"8ef8c322f6da2990f202fe71373fc3bd9b047f0d55a012c8b32032780822b211"} Oct 05 08:21:18 crc kubenswrapper[4846]: I1005 08:21:18.892657 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"2059932a-7f7d-45ff-8a75-333a6711afdf","Type":"ContainerDied","Data":"6b404493c6db8bac188f9631c29896f27f3a3358dd283b65b95d1c37548a4200"} Oct 05 08:21:18 crc kubenswrapper[4846]: I1005 08:21:18.923318 4846 scope.go:117] "RemoveContainer" containerID="03f5a7ce92d7760417b7debfdc2b9a753f6e29e6da93d51623b51a65bab02dd9" Oct 05 08:21:18 crc kubenswrapper[4846]: I1005 08:21:18.941201 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 05 08:21:18 crc kubenswrapper[4846]: I1005 08:21:18.958746 4846 scope.go:117] "RemoveContainer" containerID="c4440fe126de02dd5b2a1f044ad7cdfd08c87e2f809ed07e76700fa7299a811e" Oct 05 08:21:18 crc kubenswrapper[4846]: E1005 08:21:18.959573 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c4440fe126de02dd5b2a1f044ad7cdfd08c87e2f809ed07e76700fa7299a811e\": container with ID starting with c4440fe126de02dd5b2a1f044ad7cdfd08c87e2f809ed07e76700fa7299a811e not found: ID does not exist" containerID="c4440fe126de02dd5b2a1f044ad7cdfd08c87e2f809ed07e76700fa7299a811e" Oct 05 08:21:18 crc kubenswrapper[4846]: I1005 08:21:18.959615 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c4440fe126de02dd5b2a1f044ad7cdfd08c87e2f809ed07e76700fa7299a811e"} err="failed to get container status \"c4440fe126de02dd5b2a1f044ad7cdfd08c87e2f809ed07e76700fa7299a811e\": rpc error: code = NotFound desc = could not find container \"c4440fe126de02dd5b2a1f044ad7cdfd08c87e2f809ed07e76700fa7299a811e\": container with ID starting with c4440fe126de02dd5b2a1f044ad7cdfd08c87e2f809ed07e76700fa7299a811e not found: ID does not exist" Oct 05 08:21:18 crc kubenswrapper[4846]: I1005 08:21:18.959652 4846 scope.go:117] "RemoveContainer" containerID="03f5a7ce92d7760417b7debfdc2b9a753f6e29e6da93d51623b51a65bab02dd9" Oct 05 08:21:18 crc kubenswrapper[4846]: E1005 08:21:18.961467 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"03f5a7ce92d7760417b7debfdc2b9a753f6e29e6da93d51623b51a65bab02dd9\": container with ID starting with 03f5a7ce92d7760417b7debfdc2b9a753f6e29e6da93d51623b51a65bab02dd9 not found: ID does not exist" containerID="03f5a7ce92d7760417b7debfdc2b9a753f6e29e6da93d51623b51a65bab02dd9" Oct 05 08:21:18 crc kubenswrapper[4846]: I1005 08:21:18.961508 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"03f5a7ce92d7760417b7debfdc2b9a753f6e29e6da93d51623b51a65bab02dd9"} err="failed to get container status \"03f5a7ce92d7760417b7debfdc2b9a753f6e29e6da93d51623b51a65bab02dd9\": rpc error: code = NotFound desc = could not find container \"03f5a7ce92d7760417b7debfdc2b9a753f6e29e6da93d51623b51a65bab02dd9\": container with ID starting with 03f5a7ce92d7760417b7debfdc2b9a753f6e29e6da93d51623b51a65bab02dd9 not found: ID does not exist" Oct 05 08:21:18 crc kubenswrapper[4846]: I1005 08:21:18.961534 4846 scope.go:117] "RemoveContainer" containerID="c4440fe126de02dd5b2a1f044ad7cdfd08c87e2f809ed07e76700fa7299a811e" Oct 05 08:21:18 crc kubenswrapper[4846]: I1005 08:21:18.961853 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c4440fe126de02dd5b2a1f044ad7cdfd08c87e2f809ed07e76700fa7299a811e"} err="failed to get container status \"c4440fe126de02dd5b2a1f044ad7cdfd08c87e2f809ed07e76700fa7299a811e\": rpc error: code = NotFound desc = could not find container \"c4440fe126de02dd5b2a1f044ad7cdfd08c87e2f809ed07e76700fa7299a811e\": container with ID starting with c4440fe126de02dd5b2a1f044ad7cdfd08c87e2f809ed07e76700fa7299a811e not found: ID does not exist" Oct 05 08:21:18 crc kubenswrapper[4846]: I1005 08:21:18.961875 4846 scope.go:117] "RemoveContainer" containerID="03f5a7ce92d7760417b7debfdc2b9a753f6e29e6da93d51623b51a65bab02dd9" Oct 05 08:21:18 crc kubenswrapper[4846]: I1005 08:21:18.962032 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"03f5a7ce92d7760417b7debfdc2b9a753f6e29e6da93d51623b51a65bab02dd9"} err="failed to get container status \"03f5a7ce92d7760417b7debfdc2b9a753f6e29e6da93d51623b51a65bab02dd9\": rpc error: code = NotFound desc = could not find container \"03f5a7ce92d7760417b7debfdc2b9a753f6e29e6da93d51623b51a65bab02dd9\": container with ID starting with 03f5a7ce92d7760417b7debfdc2b9a753f6e29e6da93d51623b51a65bab02dd9 not found: ID does not exist" Oct 05 08:21:18 crc kubenswrapper[4846]: I1005 08:21:18.962050 4846 scope.go:117] "RemoveContainer" containerID="473b4ffc35fbaac4fee2179df1d2e7ee30f434ac41d6e8f1645ea47428180b34" Oct 05 08:21:18 crc kubenswrapper[4846]: I1005 08:21:18.963070 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 05 08:21:18 crc kubenswrapper[4846]: I1005 08:21:18.980696 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 05 08:21:18 crc kubenswrapper[4846]: I1005 08:21:18.991283 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 05 08:21:18 crc kubenswrapper[4846]: I1005 08:21:18.992695 4846 scope.go:117] "RemoveContainer" containerID="8ef8c322f6da2990f202fe71373fc3bd9b047f0d55a012c8b32032780822b211" Oct 05 08:21:18 crc kubenswrapper[4846]: I1005 08:21:18.998579 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Oct 05 08:21:18 crc kubenswrapper[4846]: E1005 08:21:18.999057 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2059932a-7f7d-45ff-8a75-333a6711afdf" containerName="glance-httpd" Oct 05 08:21:18 crc kubenswrapper[4846]: I1005 08:21:18.999079 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="2059932a-7f7d-45ff-8a75-333a6711afdf" containerName="glance-httpd" Oct 05 08:21:18 crc kubenswrapper[4846]: E1005 08:21:18.999095 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="976cec21-fc6f-440f-b2ea-a181f52fd624" containerName="glance-httpd" Oct 05 08:21:18 crc kubenswrapper[4846]: I1005 08:21:18.999105 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="976cec21-fc6f-440f-b2ea-a181f52fd624" containerName="glance-httpd" Oct 05 08:21:18 crc kubenswrapper[4846]: E1005 08:21:18.999123 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2059932a-7f7d-45ff-8a75-333a6711afdf" containerName="glance-log" Oct 05 08:21:18 crc kubenswrapper[4846]: I1005 08:21:18.999131 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="2059932a-7f7d-45ff-8a75-333a6711afdf" containerName="glance-log" Oct 05 08:21:18 crc kubenswrapper[4846]: E1005 08:21:18.999160 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="976cec21-fc6f-440f-b2ea-a181f52fd624" containerName="glance-log" Oct 05 08:21:18 crc kubenswrapper[4846]: I1005 08:21:18.999167 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="976cec21-fc6f-440f-b2ea-a181f52fd624" containerName="glance-log" Oct 05 08:21:18 crc kubenswrapper[4846]: I1005 08:21:18.999401 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="976cec21-fc6f-440f-b2ea-a181f52fd624" containerName="glance-httpd" Oct 05 08:21:18 crc kubenswrapper[4846]: I1005 08:21:18.999429 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="2059932a-7f7d-45ff-8a75-333a6711afdf" containerName="glance-httpd" Oct 05 08:21:18 crc kubenswrapper[4846]: I1005 08:21:18.999444 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="976cec21-fc6f-440f-b2ea-a181f52fd624" containerName="glance-log" Oct 05 08:21:18 crc kubenswrapper[4846]: I1005 08:21:18.999466 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="2059932a-7f7d-45ff-8a75-333a6711afdf" containerName="glance-log" Oct 05 08:21:19 crc kubenswrapper[4846]: I1005 08:21:19.000496 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 05 08:21:19 crc kubenswrapper[4846]: I1005 08:21:19.006614 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Oct 05 08:21:19 crc kubenswrapper[4846]: I1005 08:21:19.006818 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Oct 05 08:21:19 crc kubenswrapper[4846]: I1005 08:21:19.006816 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Oct 05 08:21:19 crc kubenswrapper[4846]: I1005 08:21:19.006890 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-4zskb" Oct 05 08:21:19 crc kubenswrapper[4846]: I1005 08:21:19.006943 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 05 08:21:19 crc kubenswrapper[4846]: I1005 08:21:19.008223 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 05 08:21:19 crc kubenswrapper[4846]: I1005 08:21:19.012944 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Oct 05 08:21:19 crc kubenswrapper[4846]: I1005 08:21:19.013108 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Oct 05 08:21:19 crc kubenswrapper[4846]: I1005 08:21:19.017649 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 05 08:21:19 crc kubenswrapper[4846]: I1005 08:21:19.025459 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 05 08:21:19 crc kubenswrapper[4846]: I1005 08:21:19.033015 4846 scope.go:117] "RemoveContainer" containerID="473b4ffc35fbaac4fee2179df1d2e7ee30f434ac41d6e8f1645ea47428180b34" Oct 05 08:21:19 crc kubenswrapper[4846]: E1005 08:21:19.033652 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"473b4ffc35fbaac4fee2179df1d2e7ee30f434ac41d6e8f1645ea47428180b34\": container with ID starting with 473b4ffc35fbaac4fee2179df1d2e7ee30f434ac41d6e8f1645ea47428180b34 not found: ID does not exist" containerID="473b4ffc35fbaac4fee2179df1d2e7ee30f434ac41d6e8f1645ea47428180b34" Oct 05 08:21:19 crc kubenswrapper[4846]: I1005 08:21:19.033681 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"473b4ffc35fbaac4fee2179df1d2e7ee30f434ac41d6e8f1645ea47428180b34"} err="failed to get container status \"473b4ffc35fbaac4fee2179df1d2e7ee30f434ac41d6e8f1645ea47428180b34\": rpc error: code = NotFound desc = could not find container \"473b4ffc35fbaac4fee2179df1d2e7ee30f434ac41d6e8f1645ea47428180b34\": container with ID starting with 473b4ffc35fbaac4fee2179df1d2e7ee30f434ac41d6e8f1645ea47428180b34 not found: ID does not exist" Oct 05 08:21:19 crc kubenswrapper[4846]: I1005 08:21:19.033698 4846 scope.go:117] "RemoveContainer" containerID="8ef8c322f6da2990f202fe71373fc3bd9b047f0d55a012c8b32032780822b211" Oct 05 08:21:19 crc kubenswrapper[4846]: E1005 08:21:19.035284 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8ef8c322f6da2990f202fe71373fc3bd9b047f0d55a012c8b32032780822b211\": container with ID starting with 8ef8c322f6da2990f202fe71373fc3bd9b047f0d55a012c8b32032780822b211 not found: ID does not exist" containerID="8ef8c322f6da2990f202fe71373fc3bd9b047f0d55a012c8b32032780822b211" Oct 05 08:21:19 crc kubenswrapper[4846]: I1005 08:21:19.035319 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8ef8c322f6da2990f202fe71373fc3bd9b047f0d55a012c8b32032780822b211"} err="failed to get container status \"8ef8c322f6da2990f202fe71373fc3bd9b047f0d55a012c8b32032780822b211\": rpc error: code = NotFound desc = could not find container \"8ef8c322f6da2990f202fe71373fc3bd9b047f0d55a012c8b32032780822b211\": container with ID starting with 8ef8c322f6da2990f202fe71373fc3bd9b047f0d55a012c8b32032780822b211 not found: ID does not exist" Oct 05 08:21:19 crc kubenswrapper[4846]: I1005 08:21:19.035339 4846 scope.go:117] "RemoveContainer" containerID="473b4ffc35fbaac4fee2179df1d2e7ee30f434ac41d6e8f1645ea47428180b34" Oct 05 08:21:19 crc kubenswrapper[4846]: I1005 08:21:19.036598 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"473b4ffc35fbaac4fee2179df1d2e7ee30f434ac41d6e8f1645ea47428180b34"} err="failed to get container status \"473b4ffc35fbaac4fee2179df1d2e7ee30f434ac41d6e8f1645ea47428180b34\": rpc error: code = NotFound desc = could not find container \"473b4ffc35fbaac4fee2179df1d2e7ee30f434ac41d6e8f1645ea47428180b34\": container with ID starting with 473b4ffc35fbaac4fee2179df1d2e7ee30f434ac41d6e8f1645ea47428180b34 not found: ID does not exist" Oct 05 08:21:19 crc kubenswrapper[4846]: I1005 08:21:19.036619 4846 scope.go:117] "RemoveContainer" containerID="8ef8c322f6da2990f202fe71373fc3bd9b047f0d55a012c8b32032780822b211" Oct 05 08:21:19 crc kubenswrapper[4846]: I1005 08:21:19.036824 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8ef8c322f6da2990f202fe71373fc3bd9b047f0d55a012c8b32032780822b211"} err="failed to get container status \"8ef8c322f6da2990f202fe71373fc3bd9b047f0d55a012c8b32032780822b211\": rpc error: code = NotFound desc = could not find container \"8ef8c322f6da2990f202fe71373fc3bd9b047f0d55a012c8b32032780822b211\": container with ID starting with 8ef8c322f6da2990f202fe71373fc3bd9b047f0d55a012c8b32032780822b211 not found: ID does not exist" Oct 05 08:21:19 crc kubenswrapper[4846]: I1005 08:21:19.059535 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c77ecef6-3a1d-43d5-90b1-dd3de33d7de2-config-data\") pod \"glance-default-external-api-0\" (UID: \"c77ecef6-3a1d-43d5-90b1-dd3de33d7de2\") " pod="openstack/glance-default-external-api-0" Oct 05 08:21:19 crc kubenswrapper[4846]: I1005 08:21:19.059608 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c77ecef6-3a1d-43d5-90b1-dd3de33d7de2-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"c77ecef6-3a1d-43d5-90b1-dd3de33d7de2\") " pod="openstack/glance-default-external-api-0" Oct 05 08:21:19 crc kubenswrapper[4846]: I1005 08:21:19.059650 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9bceecf1-6148-4b88-9604-4a37330b8211-scripts\") pod \"glance-default-internal-api-0\" (UID: \"9bceecf1-6148-4b88-9604-4a37330b8211\") " pod="openstack/glance-default-internal-api-0" Oct 05 08:21:19 crc kubenswrapper[4846]: I1005 08:21:19.059837 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c77ecef6-3a1d-43d5-90b1-dd3de33d7de2-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"c77ecef6-3a1d-43d5-90b1-dd3de33d7de2\") " pod="openstack/glance-default-external-api-0" Oct 05 08:21:19 crc kubenswrapper[4846]: I1005 08:21:19.059918 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w7255\" (UniqueName: \"kubernetes.io/projected/c77ecef6-3a1d-43d5-90b1-dd3de33d7de2-kube-api-access-w7255\") pod \"glance-default-external-api-0\" (UID: \"c77ecef6-3a1d-43d5-90b1-dd3de33d7de2\") " pod="openstack/glance-default-external-api-0" Oct 05 08:21:19 crc kubenswrapper[4846]: I1005 08:21:19.060007 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9bceecf1-6148-4b88-9604-4a37330b8211-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"9bceecf1-6148-4b88-9604-4a37330b8211\") " pod="openstack/glance-default-internal-api-0" Oct 05 08:21:19 crc kubenswrapper[4846]: I1005 08:21:19.060088 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9bceecf1-6148-4b88-9604-4a37330b8211-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"9bceecf1-6148-4b88-9604-4a37330b8211\") " pod="openstack/glance-default-internal-api-0" Oct 05 08:21:19 crc kubenswrapper[4846]: I1005 08:21:19.060145 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9bceecf1-6148-4b88-9604-4a37330b8211-config-data\") pod \"glance-default-internal-api-0\" (UID: \"9bceecf1-6148-4b88-9604-4a37330b8211\") " pod="openstack/glance-default-internal-api-0" Oct 05 08:21:19 crc kubenswrapper[4846]: I1005 08:21:19.060249 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/c77ecef6-3a1d-43d5-90b1-dd3de33d7de2-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"c77ecef6-3a1d-43d5-90b1-dd3de33d7de2\") " pod="openstack/glance-default-external-api-0" Oct 05 08:21:19 crc kubenswrapper[4846]: I1005 08:21:19.060386 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kc8fp\" (UniqueName: \"kubernetes.io/projected/9bceecf1-6148-4b88-9604-4a37330b8211-kube-api-access-kc8fp\") pod \"glance-default-internal-api-0\" (UID: \"9bceecf1-6148-4b88-9604-4a37330b8211\") " pod="openstack/glance-default-internal-api-0" Oct 05 08:21:19 crc kubenswrapper[4846]: I1005 08:21:19.060495 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9bceecf1-6148-4b88-9604-4a37330b8211-logs\") pod \"glance-default-internal-api-0\" (UID: \"9bceecf1-6148-4b88-9604-4a37330b8211\") " pod="openstack/glance-default-internal-api-0" Oct 05 08:21:19 crc kubenswrapper[4846]: I1005 08:21:19.060595 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c77ecef6-3a1d-43d5-90b1-dd3de33d7de2-scripts\") pod \"glance-default-external-api-0\" (UID: \"c77ecef6-3a1d-43d5-90b1-dd3de33d7de2\") " pod="openstack/glance-default-external-api-0" Oct 05 08:21:19 crc kubenswrapper[4846]: I1005 08:21:19.060701 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/9bceecf1-6148-4b88-9604-4a37330b8211-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"9bceecf1-6148-4b88-9604-4a37330b8211\") " pod="openstack/glance-default-internal-api-0" Oct 05 08:21:19 crc kubenswrapper[4846]: I1005 08:21:19.060737 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c77ecef6-3a1d-43d5-90b1-dd3de33d7de2-logs\") pod \"glance-default-external-api-0\" (UID: \"c77ecef6-3a1d-43d5-90b1-dd3de33d7de2\") " pod="openstack/glance-default-external-api-0" Oct 05 08:21:19 crc kubenswrapper[4846]: I1005 08:21:19.162172 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kc8fp\" (UniqueName: \"kubernetes.io/projected/9bceecf1-6148-4b88-9604-4a37330b8211-kube-api-access-kc8fp\") pod \"glance-default-internal-api-0\" (UID: \"9bceecf1-6148-4b88-9604-4a37330b8211\") " pod="openstack/glance-default-internal-api-0" Oct 05 08:21:19 crc kubenswrapper[4846]: I1005 08:21:19.162259 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9bceecf1-6148-4b88-9604-4a37330b8211-logs\") pod \"glance-default-internal-api-0\" (UID: \"9bceecf1-6148-4b88-9604-4a37330b8211\") " pod="openstack/glance-default-internal-api-0" Oct 05 08:21:19 crc kubenswrapper[4846]: I1005 08:21:19.162313 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c77ecef6-3a1d-43d5-90b1-dd3de33d7de2-scripts\") pod \"glance-default-external-api-0\" (UID: \"c77ecef6-3a1d-43d5-90b1-dd3de33d7de2\") " pod="openstack/glance-default-external-api-0" Oct 05 08:21:19 crc kubenswrapper[4846]: I1005 08:21:19.162353 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/9bceecf1-6148-4b88-9604-4a37330b8211-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"9bceecf1-6148-4b88-9604-4a37330b8211\") " pod="openstack/glance-default-internal-api-0" Oct 05 08:21:19 crc kubenswrapper[4846]: I1005 08:21:19.162379 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c77ecef6-3a1d-43d5-90b1-dd3de33d7de2-logs\") pod \"glance-default-external-api-0\" (UID: \"c77ecef6-3a1d-43d5-90b1-dd3de33d7de2\") " pod="openstack/glance-default-external-api-0" Oct 05 08:21:19 crc kubenswrapper[4846]: I1005 08:21:19.162417 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c77ecef6-3a1d-43d5-90b1-dd3de33d7de2-config-data\") pod \"glance-default-external-api-0\" (UID: \"c77ecef6-3a1d-43d5-90b1-dd3de33d7de2\") " pod="openstack/glance-default-external-api-0" Oct 05 08:21:19 crc kubenswrapper[4846]: I1005 08:21:19.162460 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c77ecef6-3a1d-43d5-90b1-dd3de33d7de2-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"c77ecef6-3a1d-43d5-90b1-dd3de33d7de2\") " pod="openstack/glance-default-external-api-0" Oct 05 08:21:19 crc kubenswrapper[4846]: I1005 08:21:19.162489 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9bceecf1-6148-4b88-9604-4a37330b8211-scripts\") pod \"glance-default-internal-api-0\" (UID: \"9bceecf1-6148-4b88-9604-4a37330b8211\") " pod="openstack/glance-default-internal-api-0" Oct 05 08:21:19 crc kubenswrapper[4846]: I1005 08:21:19.162533 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c77ecef6-3a1d-43d5-90b1-dd3de33d7de2-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"c77ecef6-3a1d-43d5-90b1-dd3de33d7de2\") " pod="openstack/glance-default-external-api-0" Oct 05 08:21:19 crc kubenswrapper[4846]: I1005 08:21:19.162562 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w7255\" (UniqueName: \"kubernetes.io/projected/c77ecef6-3a1d-43d5-90b1-dd3de33d7de2-kube-api-access-w7255\") pod \"glance-default-external-api-0\" (UID: \"c77ecef6-3a1d-43d5-90b1-dd3de33d7de2\") " pod="openstack/glance-default-external-api-0" Oct 05 08:21:19 crc kubenswrapper[4846]: I1005 08:21:19.162590 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9bceecf1-6148-4b88-9604-4a37330b8211-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"9bceecf1-6148-4b88-9604-4a37330b8211\") " pod="openstack/glance-default-internal-api-0" Oct 05 08:21:19 crc kubenswrapper[4846]: I1005 08:21:19.162635 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9bceecf1-6148-4b88-9604-4a37330b8211-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"9bceecf1-6148-4b88-9604-4a37330b8211\") " pod="openstack/glance-default-internal-api-0" Oct 05 08:21:19 crc kubenswrapper[4846]: I1005 08:21:19.162666 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9bceecf1-6148-4b88-9604-4a37330b8211-config-data\") pod \"glance-default-internal-api-0\" (UID: \"9bceecf1-6148-4b88-9604-4a37330b8211\") " pod="openstack/glance-default-internal-api-0" Oct 05 08:21:19 crc kubenswrapper[4846]: I1005 08:21:19.162697 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/c77ecef6-3a1d-43d5-90b1-dd3de33d7de2-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"c77ecef6-3a1d-43d5-90b1-dd3de33d7de2\") " pod="openstack/glance-default-external-api-0" Oct 05 08:21:19 crc kubenswrapper[4846]: I1005 08:21:19.164672 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/c77ecef6-3a1d-43d5-90b1-dd3de33d7de2-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"c77ecef6-3a1d-43d5-90b1-dd3de33d7de2\") " pod="openstack/glance-default-external-api-0" Oct 05 08:21:19 crc kubenswrapper[4846]: I1005 08:21:19.166536 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c77ecef6-3a1d-43d5-90b1-dd3de33d7de2-logs\") pod \"glance-default-external-api-0\" (UID: \"c77ecef6-3a1d-43d5-90b1-dd3de33d7de2\") " pod="openstack/glance-default-external-api-0" Oct 05 08:21:19 crc kubenswrapper[4846]: I1005 08:21:19.166922 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/9bceecf1-6148-4b88-9604-4a37330b8211-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"9bceecf1-6148-4b88-9604-4a37330b8211\") " pod="openstack/glance-default-internal-api-0" Oct 05 08:21:19 crc kubenswrapper[4846]: I1005 08:21:19.167599 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9bceecf1-6148-4b88-9604-4a37330b8211-logs\") pod \"glance-default-internal-api-0\" (UID: \"9bceecf1-6148-4b88-9604-4a37330b8211\") " pod="openstack/glance-default-internal-api-0" Oct 05 08:21:19 crc kubenswrapper[4846]: I1005 08:21:19.168558 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9bceecf1-6148-4b88-9604-4a37330b8211-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"9bceecf1-6148-4b88-9604-4a37330b8211\") " pod="openstack/glance-default-internal-api-0" Oct 05 08:21:19 crc kubenswrapper[4846]: I1005 08:21:19.169359 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9bceecf1-6148-4b88-9604-4a37330b8211-scripts\") pod \"glance-default-internal-api-0\" (UID: \"9bceecf1-6148-4b88-9604-4a37330b8211\") " pod="openstack/glance-default-internal-api-0" Oct 05 08:21:19 crc kubenswrapper[4846]: I1005 08:21:19.169674 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c77ecef6-3a1d-43d5-90b1-dd3de33d7de2-scripts\") pod \"glance-default-external-api-0\" (UID: \"c77ecef6-3a1d-43d5-90b1-dd3de33d7de2\") " pod="openstack/glance-default-external-api-0" Oct 05 08:21:19 crc kubenswrapper[4846]: I1005 08:21:19.170014 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c77ecef6-3a1d-43d5-90b1-dd3de33d7de2-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"c77ecef6-3a1d-43d5-90b1-dd3de33d7de2\") " pod="openstack/glance-default-external-api-0" Oct 05 08:21:19 crc kubenswrapper[4846]: I1005 08:21:19.171838 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c77ecef6-3a1d-43d5-90b1-dd3de33d7de2-config-data\") pod \"glance-default-external-api-0\" (UID: \"c77ecef6-3a1d-43d5-90b1-dd3de33d7de2\") " pod="openstack/glance-default-external-api-0" Oct 05 08:21:19 crc kubenswrapper[4846]: I1005 08:21:19.173633 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9bceecf1-6148-4b88-9604-4a37330b8211-config-data\") pod \"glance-default-internal-api-0\" (UID: \"9bceecf1-6148-4b88-9604-4a37330b8211\") " pod="openstack/glance-default-internal-api-0" Oct 05 08:21:19 crc kubenswrapper[4846]: I1005 08:21:19.174123 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9bceecf1-6148-4b88-9604-4a37330b8211-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"9bceecf1-6148-4b88-9604-4a37330b8211\") " pod="openstack/glance-default-internal-api-0" Oct 05 08:21:19 crc kubenswrapper[4846]: I1005 08:21:19.174255 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c77ecef6-3a1d-43d5-90b1-dd3de33d7de2-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"c77ecef6-3a1d-43d5-90b1-dd3de33d7de2\") " pod="openstack/glance-default-external-api-0" Oct 05 08:21:19 crc kubenswrapper[4846]: I1005 08:21:19.178325 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w7255\" (UniqueName: \"kubernetes.io/projected/c77ecef6-3a1d-43d5-90b1-dd3de33d7de2-kube-api-access-w7255\") pod \"glance-default-external-api-0\" (UID: \"c77ecef6-3a1d-43d5-90b1-dd3de33d7de2\") " pod="openstack/glance-default-external-api-0" Oct 05 08:21:19 crc kubenswrapper[4846]: I1005 08:21:19.183490 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kc8fp\" (UniqueName: \"kubernetes.io/projected/9bceecf1-6148-4b88-9604-4a37330b8211-kube-api-access-kc8fp\") pod \"glance-default-internal-api-0\" (UID: \"9bceecf1-6148-4b88-9604-4a37330b8211\") " pod="openstack/glance-default-internal-api-0" Oct 05 08:21:19 crc kubenswrapper[4846]: I1005 08:21:19.333937 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 05 08:21:19 crc kubenswrapper[4846]: I1005 08:21:19.346956 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 05 08:21:19 crc kubenswrapper[4846]: I1005 08:21:19.957217 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 05 08:21:20 crc kubenswrapper[4846]: I1005 08:21:20.036843 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 05 08:21:20 crc kubenswrapper[4846]: I1005 08:21:20.508618 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2059932a-7f7d-45ff-8a75-333a6711afdf" path="/var/lib/kubelet/pods/2059932a-7f7d-45ff-8a75-333a6711afdf/volumes" Oct 05 08:21:20 crc kubenswrapper[4846]: I1005 08:21:20.511580 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="976cec21-fc6f-440f-b2ea-a181f52fd624" path="/var/lib/kubelet/pods/976cec21-fc6f-440f-b2ea-a181f52fd624/volumes" Oct 05 08:21:20 crc kubenswrapper[4846]: I1005 08:21:20.918757 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"9bceecf1-6148-4b88-9604-4a37330b8211","Type":"ContainerStarted","Data":"2afe0c9e24d73820fd0e74d9e701dd3e0b5e5424e84fa2d6087a65639080a151"} Oct 05 08:21:20 crc kubenswrapper[4846]: I1005 08:21:20.919106 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"9bceecf1-6148-4b88-9604-4a37330b8211","Type":"ContainerStarted","Data":"1f45371f108557ad3137c9f316725d9f0083b908606dcace3367c337cd76bcbf"} Oct 05 08:21:20 crc kubenswrapper[4846]: I1005 08:21:20.920732 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"c77ecef6-3a1d-43d5-90b1-dd3de33d7de2","Type":"ContainerStarted","Data":"d0b2a95a1c72c58f215d678122d3d9a70d119c082300731b85c1b4d15fcb5ab0"} Oct 05 08:21:20 crc kubenswrapper[4846]: I1005 08:21:20.920756 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"c77ecef6-3a1d-43d5-90b1-dd3de33d7de2","Type":"ContainerStarted","Data":"0ea80ef9da8ba621e1ee9170929b011f536b8e7d9fe7573206557e238d515dc9"} Oct 05 08:21:21 crc kubenswrapper[4846]: I1005 08:21:21.933304 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"9bceecf1-6148-4b88-9604-4a37330b8211","Type":"ContainerStarted","Data":"a00db97f5fd0530b607658f46a3c6536ad9be6f121de426f95c1f017bccad02e"} Oct 05 08:21:21 crc kubenswrapper[4846]: I1005 08:21:21.939130 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"c77ecef6-3a1d-43d5-90b1-dd3de33d7de2","Type":"ContainerStarted","Data":"8e3d24f7a9c614dd76a5fe862718bcd0dda3735120671933705a9bef54139d4f"} Oct 05 08:21:21 crc kubenswrapper[4846]: I1005 08:21:21.959239 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=3.959219283 podStartE2EDuration="3.959219283s" podCreationTimestamp="2025-10-05 08:21:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 08:21:21.956972173 +0000 UTC m=+5604.197825008" watchObservedRunningTime="2025-10-05 08:21:21.959219283 +0000 UTC m=+5604.200072058" Oct 05 08:21:21 crc kubenswrapper[4846]: I1005 08:21:21.991594 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=3.99157297 podStartE2EDuration="3.99157297s" podCreationTimestamp="2025-10-05 08:21:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 08:21:21.987005607 +0000 UTC m=+5604.227858422" watchObservedRunningTime="2025-10-05 08:21:21.99157297 +0000 UTC m=+5604.232425755" Oct 05 08:21:23 crc kubenswrapper[4846]: I1005 08:21:23.325355 4846 patch_prober.go:28] interesting pod/machine-config-daemon-fscvf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 08:21:23 crc kubenswrapper[4846]: I1005 08:21:23.325717 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 08:21:25 crc kubenswrapper[4846]: I1005 08:21:25.586379 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-59bf5dd889-89l2d" Oct 05 08:21:25 crc kubenswrapper[4846]: I1005 08:21:25.653491 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-84896b68fc-njlvz"] Oct 05 08:21:25 crc kubenswrapper[4846]: I1005 08:21:25.653761 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-84896b68fc-njlvz" podUID="5eccdabd-f2e9-4ff9-8fee-37db337983b5" containerName="dnsmasq-dns" containerID="cri-o://6306c230a0140fbc7b9d73056688b0174f8dade5645ebb85c57ab843ae00807e" gracePeriod=10 Oct 05 08:21:25 crc kubenswrapper[4846]: I1005 08:21:25.987855 4846 generic.go:334] "Generic (PLEG): container finished" podID="5eccdabd-f2e9-4ff9-8fee-37db337983b5" containerID="6306c230a0140fbc7b9d73056688b0174f8dade5645ebb85c57ab843ae00807e" exitCode=0 Oct 05 08:21:25 crc kubenswrapper[4846]: I1005 08:21:25.988008 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-84896b68fc-njlvz" event={"ID":"5eccdabd-f2e9-4ff9-8fee-37db337983b5","Type":"ContainerDied","Data":"6306c230a0140fbc7b9d73056688b0174f8dade5645ebb85c57ab843ae00807e"} Oct 05 08:21:26 crc kubenswrapper[4846]: I1005 08:21:26.121369 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-84896b68fc-njlvz" Oct 05 08:21:26 crc kubenswrapper[4846]: I1005 08:21:26.210115 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5eccdabd-f2e9-4ff9-8fee-37db337983b5-config\") pod \"5eccdabd-f2e9-4ff9-8fee-37db337983b5\" (UID: \"5eccdabd-f2e9-4ff9-8fee-37db337983b5\") " Oct 05 08:21:26 crc kubenswrapper[4846]: I1005 08:21:26.210188 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jmfrz\" (UniqueName: \"kubernetes.io/projected/5eccdabd-f2e9-4ff9-8fee-37db337983b5-kube-api-access-jmfrz\") pod \"5eccdabd-f2e9-4ff9-8fee-37db337983b5\" (UID: \"5eccdabd-f2e9-4ff9-8fee-37db337983b5\") " Oct 05 08:21:26 crc kubenswrapper[4846]: I1005 08:21:26.210221 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5eccdabd-f2e9-4ff9-8fee-37db337983b5-ovsdbserver-nb\") pod \"5eccdabd-f2e9-4ff9-8fee-37db337983b5\" (UID: \"5eccdabd-f2e9-4ff9-8fee-37db337983b5\") " Oct 05 08:21:26 crc kubenswrapper[4846]: I1005 08:21:26.210238 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5eccdabd-f2e9-4ff9-8fee-37db337983b5-ovsdbserver-sb\") pod \"5eccdabd-f2e9-4ff9-8fee-37db337983b5\" (UID: \"5eccdabd-f2e9-4ff9-8fee-37db337983b5\") " Oct 05 08:21:26 crc kubenswrapper[4846]: I1005 08:21:26.210320 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5eccdabd-f2e9-4ff9-8fee-37db337983b5-dns-svc\") pod \"5eccdabd-f2e9-4ff9-8fee-37db337983b5\" (UID: \"5eccdabd-f2e9-4ff9-8fee-37db337983b5\") " Oct 05 08:21:26 crc kubenswrapper[4846]: I1005 08:21:26.230389 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5eccdabd-f2e9-4ff9-8fee-37db337983b5-kube-api-access-jmfrz" (OuterVolumeSpecName: "kube-api-access-jmfrz") pod "5eccdabd-f2e9-4ff9-8fee-37db337983b5" (UID: "5eccdabd-f2e9-4ff9-8fee-37db337983b5"). InnerVolumeSpecName "kube-api-access-jmfrz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:21:26 crc kubenswrapper[4846]: I1005 08:21:26.257775 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5eccdabd-f2e9-4ff9-8fee-37db337983b5-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "5eccdabd-f2e9-4ff9-8fee-37db337983b5" (UID: "5eccdabd-f2e9-4ff9-8fee-37db337983b5"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 08:21:26 crc kubenswrapper[4846]: I1005 08:21:26.258760 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5eccdabd-f2e9-4ff9-8fee-37db337983b5-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "5eccdabd-f2e9-4ff9-8fee-37db337983b5" (UID: "5eccdabd-f2e9-4ff9-8fee-37db337983b5"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 08:21:26 crc kubenswrapper[4846]: I1005 08:21:26.268350 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5eccdabd-f2e9-4ff9-8fee-37db337983b5-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "5eccdabd-f2e9-4ff9-8fee-37db337983b5" (UID: "5eccdabd-f2e9-4ff9-8fee-37db337983b5"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 08:21:26 crc kubenswrapper[4846]: I1005 08:21:26.295676 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5eccdabd-f2e9-4ff9-8fee-37db337983b5-config" (OuterVolumeSpecName: "config") pod "5eccdabd-f2e9-4ff9-8fee-37db337983b5" (UID: "5eccdabd-f2e9-4ff9-8fee-37db337983b5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 08:21:26 crc kubenswrapper[4846]: I1005 08:21:26.312979 4846 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5eccdabd-f2e9-4ff9-8fee-37db337983b5-config\") on node \"crc\" DevicePath \"\"" Oct 05 08:21:26 crc kubenswrapper[4846]: I1005 08:21:26.313026 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jmfrz\" (UniqueName: \"kubernetes.io/projected/5eccdabd-f2e9-4ff9-8fee-37db337983b5-kube-api-access-jmfrz\") on node \"crc\" DevicePath \"\"" Oct 05 08:21:26 crc kubenswrapper[4846]: I1005 08:21:26.313043 4846 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5eccdabd-f2e9-4ff9-8fee-37db337983b5-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 05 08:21:26 crc kubenswrapper[4846]: I1005 08:21:26.313055 4846 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5eccdabd-f2e9-4ff9-8fee-37db337983b5-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 05 08:21:26 crc kubenswrapper[4846]: I1005 08:21:26.313066 4846 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5eccdabd-f2e9-4ff9-8fee-37db337983b5-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 05 08:21:27 crc kubenswrapper[4846]: I1005 08:21:27.004369 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-84896b68fc-njlvz" event={"ID":"5eccdabd-f2e9-4ff9-8fee-37db337983b5","Type":"ContainerDied","Data":"c57603f6c6b243b53a81d00e89ff2d595bbf3338ef1cb015b90121f5d1870893"} Oct 05 08:21:27 crc kubenswrapper[4846]: I1005 08:21:27.004428 4846 scope.go:117] "RemoveContainer" containerID="6306c230a0140fbc7b9d73056688b0174f8dade5645ebb85c57ab843ae00807e" Oct 05 08:21:27 crc kubenswrapper[4846]: I1005 08:21:27.004469 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-84896b68fc-njlvz" Oct 05 08:21:27 crc kubenswrapper[4846]: I1005 08:21:27.043430 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-84896b68fc-njlvz"] Oct 05 08:21:27 crc kubenswrapper[4846]: I1005 08:21:27.050561 4846 scope.go:117] "RemoveContainer" containerID="e6f91d0ac3643b5c00c030c3cd42dbdcfb6aa3bce679bf16f630b96882ef39b2" Oct 05 08:21:27 crc kubenswrapper[4846]: I1005 08:21:27.059500 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-84896b68fc-njlvz"] Oct 05 08:21:28 crc kubenswrapper[4846]: I1005 08:21:28.511413 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5eccdabd-f2e9-4ff9-8fee-37db337983b5" path="/var/lib/kubelet/pods/5eccdabd-f2e9-4ff9-8fee-37db337983b5/volumes" Oct 05 08:21:29 crc kubenswrapper[4846]: I1005 08:21:29.335133 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Oct 05 08:21:29 crc kubenswrapper[4846]: I1005 08:21:29.335256 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Oct 05 08:21:29 crc kubenswrapper[4846]: I1005 08:21:29.347112 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Oct 05 08:21:29 crc kubenswrapper[4846]: I1005 08:21:29.347156 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Oct 05 08:21:29 crc kubenswrapper[4846]: I1005 08:21:29.380096 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Oct 05 08:21:29 crc kubenswrapper[4846]: I1005 08:21:29.389457 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Oct 05 08:21:29 crc kubenswrapper[4846]: I1005 08:21:29.391249 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Oct 05 08:21:29 crc kubenswrapper[4846]: I1005 08:21:29.426261 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Oct 05 08:21:30 crc kubenswrapper[4846]: I1005 08:21:30.042510 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Oct 05 08:21:30 crc kubenswrapper[4846]: I1005 08:21:30.042862 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Oct 05 08:21:30 crc kubenswrapper[4846]: I1005 08:21:30.042891 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Oct 05 08:21:30 crc kubenswrapper[4846]: I1005 08:21:30.042910 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Oct 05 08:21:32 crc kubenswrapper[4846]: I1005 08:21:32.002120 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Oct 05 08:21:32 crc kubenswrapper[4846]: I1005 08:21:32.003941 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Oct 05 08:21:32 crc kubenswrapper[4846]: I1005 08:21:32.018970 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Oct 05 08:21:32 crc kubenswrapper[4846]: I1005 08:21:32.062704 4846 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 05 08:21:32 crc kubenswrapper[4846]: I1005 08:21:32.260837 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Oct 05 08:21:32 crc kubenswrapper[4846]: I1005 08:21:32.353675 4846 scope.go:117] "RemoveContainer" containerID="6daf0df1174036065afa31dbfba00b9815003c98ae9df135d26c50cca12908d9" Oct 05 08:21:32 crc kubenswrapper[4846]: I1005 08:21:32.382073 4846 scope.go:117] "RemoveContainer" containerID="b8ae166929fded5140c26d7b2278fb033be3a4e71c6e06c59c2bb77f63a71dd6" Oct 05 08:21:32 crc kubenswrapper[4846]: I1005 08:21:32.424895 4846 scope.go:117] "RemoveContainer" containerID="2b4f4484952a6da87c239627a030a3cfcd91dc71eda2fd217654b8974c743dc5" Oct 05 08:21:40 crc kubenswrapper[4846]: I1005 08:21:40.456539 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-create-ft5wq"] Oct 05 08:21:40 crc kubenswrapper[4846]: E1005 08:21:40.457298 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5eccdabd-f2e9-4ff9-8fee-37db337983b5" containerName="init" Oct 05 08:21:40 crc kubenswrapper[4846]: I1005 08:21:40.457309 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="5eccdabd-f2e9-4ff9-8fee-37db337983b5" containerName="init" Oct 05 08:21:40 crc kubenswrapper[4846]: E1005 08:21:40.457324 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5eccdabd-f2e9-4ff9-8fee-37db337983b5" containerName="dnsmasq-dns" Oct 05 08:21:40 crc kubenswrapper[4846]: I1005 08:21:40.457331 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="5eccdabd-f2e9-4ff9-8fee-37db337983b5" containerName="dnsmasq-dns" Oct 05 08:21:40 crc kubenswrapper[4846]: I1005 08:21:40.457509 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="5eccdabd-f2e9-4ff9-8fee-37db337983b5" containerName="dnsmasq-dns" Oct 05 08:21:40 crc kubenswrapper[4846]: I1005 08:21:40.458071 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-ft5wq" Oct 05 08:21:40 crc kubenswrapper[4846]: I1005 08:21:40.474140 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-ft5wq"] Oct 05 08:21:40 crc kubenswrapper[4846]: I1005 08:21:40.549888 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xvgtj\" (UniqueName: \"kubernetes.io/projected/c8c439e7-0384-49d4-bdbf-f2fd0cc28118-kube-api-access-xvgtj\") pod \"placement-db-create-ft5wq\" (UID: \"c8c439e7-0384-49d4-bdbf-f2fd0cc28118\") " pod="openstack/placement-db-create-ft5wq" Oct 05 08:21:40 crc kubenswrapper[4846]: I1005 08:21:40.652385 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xvgtj\" (UniqueName: \"kubernetes.io/projected/c8c439e7-0384-49d4-bdbf-f2fd0cc28118-kube-api-access-xvgtj\") pod \"placement-db-create-ft5wq\" (UID: \"c8c439e7-0384-49d4-bdbf-f2fd0cc28118\") " pod="openstack/placement-db-create-ft5wq" Oct 05 08:21:40 crc kubenswrapper[4846]: I1005 08:21:40.679405 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xvgtj\" (UniqueName: \"kubernetes.io/projected/c8c439e7-0384-49d4-bdbf-f2fd0cc28118-kube-api-access-xvgtj\") pod \"placement-db-create-ft5wq\" (UID: \"c8c439e7-0384-49d4-bdbf-f2fd0cc28118\") " pod="openstack/placement-db-create-ft5wq" Oct 05 08:21:40 crc kubenswrapper[4846]: I1005 08:21:40.785062 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-ft5wq" Oct 05 08:21:41 crc kubenswrapper[4846]: I1005 08:21:41.257043 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-ft5wq"] Oct 05 08:21:41 crc kubenswrapper[4846]: W1005 08:21:41.262716 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc8c439e7_0384_49d4_bdbf_f2fd0cc28118.slice/crio-6f210c2d7df61ac70a0ed408868f3bbfcf5e76cac38d3021c3baf59c09e58388 WatchSource:0}: Error finding container 6f210c2d7df61ac70a0ed408868f3bbfcf5e76cac38d3021c3baf59c09e58388: Status 404 returned error can't find the container with id 6f210c2d7df61ac70a0ed408868f3bbfcf5e76cac38d3021c3baf59c09e58388 Oct 05 08:21:41 crc kubenswrapper[4846]: E1005 08:21:41.742096 4846 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc8c439e7_0384_49d4_bdbf_f2fd0cc28118.slice/crio-conmon-9283ee8b199a6e8c0dfeb0966df9bc7ada80e4f5b05fa35bf7f11a12ca52a6e8.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc8c439e7_0384_49d4_bdbf_f2fd0cc28118.slice/crio-9283ee8b199a6e8c0dfeb0966df9bc7ada80e4f5b05fa35bf7f11a12ca52a6e8.scope\": RecentStats: unable to find data in memory cache]" Oct 05 08:21:42 crc kubenswrapper[4846]: I1005 08:21:42.164982 4846 generic.go:334] "Generic (PLEG): container finished" podID="c8c439e7-0384-49d4-bdbf-f2fd0cc28118" containerID="9283ee8b199a6e8c0dfeb0966df9bc7ada80e4f5b05fa35bf7f11a12ca52a6e8" exitCode=0 Oct 05 08:21:42 crc kubenswrapper[4846]: I1005 08:21:42.165061 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-ft5wq" event={"ID":"c8c439e7-0384-49d4-bdbf-f2fd0cc28118","Type":"ContainerDied","Data":"9283ee8b199a6e8c0dfeb0966df9bc7ada80e4f5b05fa35bf7f11a12ca52a6e8"} Oct 05 08:21:42 crc kubenswrapper[4846]: I1005 08:21:42.165089 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-ft5wq" event={"ID":"c8c439e7-0384-49d4-bdbf-f2fd0cc28118","Type":"ContainerStarted","Data":"6f210c2d7df61ac70a0ed408868f3bbfcf5e76cac38d3021c3baf59c09e58388"} Oct 05 08:21:43 crc kubenswrapper[4846]: I1005 08:21:43.578284 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-ft5wq" Oct 05 08:21:43 crc kubenswrapper[4846]: I1005 08:21:43.710125 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xvgtj\" (UniqueName: \"kubernetes.io/projected/c8c439e7-0384-49d4-bdbf-f2fd0cc28118-kube-api-access-xvgtj\") pod \"c8c439e7-0384-49d4-bdbf-f2fd0cc28118\" (UID: \"c8c439e7-0384-49d4-bdbf-f2fd0cc28118\") " Oct 05 08:21:43 crc kubenswrapper[4846]: I1005 08:21:43.716324 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c8c439e7-0384-49d4-bdbf-f2fd0cc28118-kube-api-access-xvgtj" (OuterVolumeSpecName: "kube-api-access-xvgtj") pod "c8c439e7-0384-49d4-bdbf-f2fd0cc28118" (UID: "c8c439e7-0384-49d4-bdbf-f2fd0cc28118"). InnerVolumeSpecName "kube-api-access-xvgtj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:21:43 crc kubenswrapper[4846]: I1005 08:21:43.812318 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xvgtj\" (UniqueName: \"kubernetes.io/projected/c8c439e7-0384-49d4-bdbf-f2fd0cc28118-kube-api-access-xvgtj\") on node \"crc\" DevicePath \"\"" Oct 05 08:21:44 crc kubenswrapper[4846]: I1005 08:21:44.196256 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-ft5wq" event={"ID":"c8c439e7-0384-49d4-bdbf-f2fd0cc28118","Type":"ContainerDied","Data":"6f210c2d7df61ac70a0ed408868f3bbfcf5e76cac38d3021c3baf59c09e58388"} Oct 05 08:21:44 crc kubenswrapper[4846]: I1005 08:21:44.196737 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6f210c2d7df61ac70a0ed408868f3bbfcf5e76cac38d3021c3baf59c09e58388" Oct 05 08:21:44 crc kubenswrapper[4846]: I1005 08:21:44.196852 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-ft5wq" Oct 05 08:21:50 crc kubenswrapper[4846]: I1005 08:21:50.481243 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-c6ce-account-create-26v82"] Oct 05 08:21:50 crc kubenswrapper[4846]: E1005 08:21:50.482069 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c8c439e7-0384-49d4-bdbf-f2fd0cc28118" containerName="mariadb-database-create" Oct 05 08:21:50 crc kubenswrapper[4846]: I1005 08:21:50.482082 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="c8c439e7-0384-49d4-bdbf-f2fd0cc28118" containerName="mariadb-database-create" Oct 05 08:21:50 crc kubenswrapper[4846]: I1005 08:21:50.482304 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="c8c439e7-0384-49d4-bdbf-f2fd0cc28118" containerName="mariadb-database-create" Oct 05 08:21:50 crc kubenswrapper[4846]: I1005 08:21:50.483002 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-c6ce-account-create-26v82" Oct 05 08:21:50 crc kubenswrapper[4846]: I1005 08:21:50.485766 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-db-secret" Oct 05 08:21:50 crc kubenswrapper[4846]: I1005 08:21:50.492285 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-c6ce-account-create-26v82"] Oct 05 08:21:50 crc kubenswrapper[4846]: I1005 08:21:50.554622 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b295r\" (UniqueName: \"kubernetes.io/projected/5aed0561-41c0-47d2-95c2-a23f79b6f8f8-kube-api-access-b295r\") pod \"placement-c6ce-account-create-26v82\" (UID: \"5aed0561-41c0-47d2-95c2-a23f79b6f8f8\") " pod="openstack/placement-c6ce-account-create-26v82" Oct 05 08:21:50 crc kubenswrapper[4846]: I1005 08:21:50.656610 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b295r\" (UniqueName: \"kubernetes.io/projected/5aed0561-41c0-47d2-95c2-a23f79b6f8f8-kube-api-access-b295r\") pod \"placement-c6ce-account-create-26v82\" (UID: \"5aed0561-41c0-47d2-95c2-a23f79b6f8f8\") " pod="openstack/placement-c6ce-account-create-26v82" Oct 05 08:21:50 crc kubenswrapper[4846]: I1005 08:21:50.686609 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b295r\" (UniqueName: \"kubernetes.io/projected/5aed0561-41c0-47d2-95c2-a23f79b6f8f8-kube-api-access-b295r\") pod \"placement-c6ce-account-create-26v82\" (UID: \"5aed0561-41c0-47d2-95c2-a23f79b6f8f8\") " pod="openstack/placement-c6ce-account-create-26v82" Oct 05 08:21:50 crc kubenswrapper[4846]: I1005 08:21:50.812445 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-c6ce-account-create-26v82" Oct 05 08:21:51 crc kubenswrapper[4846]: I1005 08:21:51.419698 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-c6ce-account-create-26v82"] Oct 05 08:21:52 crc kubenswrapper[4846]: I1005 08:21:52.286151 4846 generic.go:334] "Generic (PLEG): container finished" podID="5aed0561-41c0-47d2-95c2-a23f79b6f8f8" containerID="bf0445b72a12e1f9d4fcf1a8742dc0aa8e65169aaeed48321b63545b19175f45" exitCode=0 Oct 05 08:21:52 crc kubenswrapper[4846]: I1005 08:21:52.286220 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-c6ce-account-create-26v82" event={"ID":"5aed0561-41c0-47d2-95c2-a23f79b6f8f8","Type":"ContainerDied","Data":"bf0445b72a12e1f9d4fcf1a8742dc0aa8e65169aaeed48321b63545b19175f45"} Oct 05 08:21:52 crc kubenswrapper[4846]: I1005 08:21:52.286278 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-c6ce-account-create-26v82" event={"ID":"5aed0561-41c0-47d2-95c2-a23f79b6f8f8","Type":"ContainerStarted","Data":"dd17d3f16bc8856235deb45ab1fa2bfe00325d43d456bfa9a94f52010c7dcf83"} Oct 05 08:21:53 crc kubenswrapper[4846]: I1005 08:21:53.325401 4846 patch_prober.go:28] interesting pod/machine-config-daemon-fscvf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 08:21:53 crc kubenswrapper[4846]: I1005 08:21:53.325742 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 08:21:53 crc kubenswrapper[4846]: I1005 08:21:53.325815 4846 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" Oct 05 08:21:53 crc kubenswrapper[4846]: I1005 08:21:53.326897 4846 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"fdafa27da24296f9087c33b6b12a6598d2a8d61fc985b06ad680401828c9feb6"} pod="openshift-machine-config-operator/machine-config-daemon-fscvf" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 05 08:21:53 crc kubenswrapper[4846]: I1005 08:21:53.327007 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" containerName="machine-config-daemon" containerID="cri-o://fdafa27da24296f9087c33b6b12a6598d2a8d61fc985b06ad680401828c9feb6" gracePeriod=600 Oct 05 08:21:53 crc kubenswrapper[4846]: E1005 08:21:53.452150 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 08:21:53 crc kubenswrapper[4846]: I1005 08:21:53.707603 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-c6ce-account-create-26v82" Oct 05 08:21:53 crc kubenswrapper[4846]: I1005 08:21:53.838856 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b295r\" (UniqueName: \"kubernetes.io/projected/5aed0561-41c0-47d2-95c2-a23f79b6f8f8-kube-api-access-b295r\") pod \"5aed0561-41c0-47d2-95c2-a23f79b6f8f8\" (UID: \"5aed0561-41c0-47d2-95c2-a23f79b6f8f8\") " Oct 05 08:21:53 crc kubenswrapper[4846]: I1005 08:21:53.845794 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5aed0561-41c0-47d2-95c2-a23f79b6f8f8-kube-api-access-b295r" (OuterVolumeSpecName: "kube-api-access-b295r") pod "5aed0561-41c0-47d2-95c2-a23f79b6f8f8" (UID: "5aed0561-41c0-47d2-95c2-a23f79b6f8f8"). InnerVolumeSpecName "kube-api-access-b295r". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:21:53 crc kubenswrapper[4846]: I1005 08:21:53.941162 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b295r\" (UniqueName: \"kubernetes.io/projected/5aed0561-41c0-47d2-95c2-a23f79b6f8f8-kube-api-access-b295r\") on node \"crc\" DevicePath \"\"" Oct 05 08:21:54 crc kubenswrapper[4846]: I1005 08:21:54.312704 4846 generic.go:334] "Generic (PLEG): container finished" podID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" containerID="fdafa27da24296f9087c33b6b12a6598d2a8d61fc985b06ad680401828c9feb6" exitCode=0 Oct 05 08:21:54 crc kubenswrapper[4846]: I1005 08:21:54.312958 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" event={"ID":"ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7","Type":"ContainerDied","Data":"fdafa27da24296f9087c33b6b12a6598d2a8d61fc985b06ad680401828c9feb6"} Oct 05 08:21:54 crc kubenswrapper[4846]: I1005 08:21:54.313367 4846 scope.go:117] "RemoveContainer" containerID="f8ac1a187331f3fff703acc696060a82d728f8e154fe3e11ba6c956d3d906397" Oct 05 08:21:54 crc kubenswrapper[4846]: I1005 08:21:54.314237 4846 scope.go:117] "RemoveContainer" containerID="fdafa27da24296f9087c33b6b12a6598d2a8d61fc985b06ad680401828c9feb6" Oct 05 08:21:54 crc kubenswrapper[4846]: E1005 08:21:54.315104 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 08:21:54 crc kubenswrapper[4846]: I1005 08:21:54.316963 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-c6ce-account-create-26v82" event={"ID":"5aed0561-41c0-47d2-95c2-a23f79b6f8f8","Type":"ContainerDied","Data":"dd17d3f16bc8856235deb45ab1fa2bfe00325d43d456bfa9a94f52010c7dcf83"} Oct 05 08:21:54 crc kubenswrapper[4846]: I1005 08:21:54.317001 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="dd17d3f16bc8856235deb45ab1fa2bfe00325d43d456bfa9a94f52010c7dcf83" Oct 05 08:21:54 crc kubenswrapper[4846]: I1005 08:21:54.317078 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-c6ce-account-create-26v82" Oct 05 08:21:55 crc kubenswrapper[4846]: I1005 08:21:55.742654 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-77bbb5966f-sbz7q"] Oct 05 08:21:55 crc kubenswrapper[4846]: E1005 08:21:55.743346 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5aed0561-41c0-47d2-95c2-a23f79b6f8f8" containerName="mariadb-account-create" Oct 05 08:21:55 crc kubenswrapper[4846]: I1005 08:21:55.743370 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="5aed0561-41c0-47d2-95c2-a23f79b6f8f8" containerName="mariadb-account-create" Oct 05 08:21:55 crc kubenswrapper[4846]: I1005 08:21:55.743750 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="5aed0561-41c0-47d2-95c2-a23f79b6f8f8" containerName="mariadb-account-create" Oct 05 08:21:55 crc kubenswrapper[4846]: I1005 08:21:55.745040 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-77bbb5966f-sbz7q" Oct 05 08:21:55 crc kubenswrapper[4846]: I1005 08:21:55.756595 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-sync-bsrnx"] Oct 05 08:21:55 crc kubenswrapper[4846]: I1005 08:21:55.758028 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-bsrnx" Oct 05 08:21:55 crc kubenswrapper[4846]: I1005 08:21:55.763403 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Oct 05 08:21:55 crc kubenswrapper[4846]: I1005 08:21:55.763454 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Oct 05 08:21:55 crc kubenswrapper[4846]: I1005 08:21:55.763484 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-5xd6t" Oct 05 08:21:55 crc kubenswrapper[4846]: I1005 08:21:55.769161 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-bsrnx"] Oct 05 08:21:55 crc kubenswrapper[4846]: I1005 08:21:55.778933 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-77bbb5966f-sbz7q"] Oct 05 08:21:55 crc kubenswrapper[4846]: I1005 08:21:55.884523 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d61014b-9453-4db8-8ae5-226d01d076a6-combined-ca-bundle\") pod \"placement-db-sync-bsrnx\" (UID: \"7d61014b-9453-4db8-8ae5-226d01d076a6\") " pod="openstack/placement-db-sync-bsrnx" Oct 05 08:21:55 crc kubenswrapper[4846]: I1005 08:21:55.884589 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sb4br\" (UniqueName: \"kubernetes.io/projected/7d61014b-9453-4db8-8ae5-226d01d076a6-kube-api-access-sb4br\") pod \"placement-db-sync-bsrnx\" (UID: \"7d61014b-9453-4db8-8ae5-226d01d076a6\") " pod="openstack/placement-db-sync-bsrnx" Oct 05 08:21:55 crc kubenswrapper[4846]: I1005 08:21:55.884645 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7d61014b-9453-4db8-8ae5-226d01d076a6-logs\") pod \"placement-db-sync-bsrnx\" (UID: \"7d61014b-9453-4db8-8ae5-226d01d076a6\") " pod="openstack/placement-db-sync-bsrnx" Oct 05 08:21:55 crc kubenswrapper[4846]: I1005 08:21:55.884897 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nd7ts\" (UniqueName: \"kubernetes.io/projected/b051b1a4-96d9-442e-964e-b8d08275caa8-kube-api-access-nd7ts\") pod \"dnsmasq-dns-77bbb5966f-sbz7q\" (UID: \"b051b1a4-96d9-442e-964e-b8d08275caa8\") " pod="openstack/dnsmasq-dns-77bbb5966f-sbz7q" Oct 05 08:21:55 crc kubenswrapper[4846]: I1005 08:21:55.884977 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b051b1a4-96d9-442e-964e-b8d08275caa8-dns-svc\") pod \"dnsmasq-dns-77bbb5966f-sbz7q\" (UID: \"b051b1a4-96d9-442e-964e-b8d08275caa8\") " pod="openstack/dnsmasq-dns-77bbb5966f-sbz7q" Oct 05 08:21:55 crc kubenswrapper[4846]: I1005 08:21:55.885015 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7d61014b-9453-4db8-8ae5-226d01d076a6-config-data\") pod \"placement-db-sync-bsrnx\" (UID: \"7d61014b-9453-4db8-8ae5-226d01d076a6\") " pod="openstack/placement-db-sync-bsrnx" Oct 05 08:21:55 crc kubenswrapper[4846]: I1005 08:21:55.885271 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b051b1a4-96d9-442e-964e-b8d08275caa8-config\") pod \"dnsmasq-dns-77bbb5966f-sbz7q\" (UID: \"b051b1a4-96d9-442e-964e-b8d08275caa8\") " pod="openstack/dnsmasq-dns-77bbb5966f-sbz7q" Oct 05 08:21:55 crc kubenswrapper[4846]: I1005 08:21:55.885415 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7d61014b-9453-4db8-8ae5-226d01d076a6-scripts\") pod \"placement-db-sync-bsrnx\" (UID: \"7d61014b-9453-4db8-8ae5-226d01d076a6\") " pod="openstack/placement-db-sync-bsrnx" Oct 05 08:21:55 crc kubenswrapper[4846]: I1005 08:21:55.885504 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b051b1a4-96d9-442e-964e-b8d08275caa8-ovsdbserver-nb\") pod \"dnsmasq-dns-77bbb5966f-sbz7q\" (UID: \"b051b1a4-96d9-442e-964e-b8d08275caa8\") " pod="openstack/dnsmasq-dns-77bbb5966f-sbz7q" Oct 05 08:21:55 crc kubenswrapper[4846]: I1005 08:21:55.885601 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b051b1a4-96d9-442e-964e-b8d08275caa8-ovsdbserver-sb\") pod \"dnsmasq-dns-77bbb5966f-sbz7q\" (UID: \"b051b1a4-96d9-442e-964e-b8d08275caa8\") " pod="openstack/dnsmasq-dns-77bbb5966f-sbz7q" Oct 05 08:21:55 crc kubenswrapper[4846]: I1005 08:21:55.987024 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7d61014b-9453-4db8-8ae5-226d01d076a6-scripts\") pod \"placement-db-sync-bsrnx\" (UID: \"7d61014b-9453-4db8-8ae5-226d01d076a6\") " pod="openstack/placement-db-sync-bsrnx" Oct 05 08:21:55 crc kubenswrapper[4846]: I1005 08:21:55.987092 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b051b1a4-96d9-442e-964e-b8d08275caa8-ovsdbserver-nb\") pod \"dnsmasq-dns-77bbb5966f-sbz7q\" (UID: \"b051b1a4-96d9-442e-964e-b8d08275caa8\") " pod="openstack/dnsmasq-dns-77bbb5966f-sbz7q" Oct 05 08:21:55 crc kubenswrapper[4846]: I1005 08:21:55.987136 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b051b1a4-96d9-442e-964e-b8d08275caa8-ovsdbserver-sb\") pod \"dnsmasq-dns-77bbb5966f-sbz7q\" (UID: \"b051b1a4-96d9-442e-964e-b8d08275caa8\") " pod="openstack/dnsmasq-dns-77bbb5966f-sbz7q" Oct 05 08:21:55 crc kubenswrapper[4846]: I1005 08:21:55.987159 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d61014b-9453-4db8-8ae5-226d01d076a6-combined-ca-bundle\") pod \"placement-db-sync-bsrnx\" (UID: \"7d61014b-9453-4db8-8ae5-226d01d076a6\") " pod="openstack/placement-db-sync-bsrnx" Oct 05 08:21:55 crc kubenswrapper[4846]: I1005 08:21:55.987187 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sb4br\" (UniqueName: \"kubernetes.io/projected/7d61014b-9453-4db8-8ae5-226d01d076a6-kube-api-access-sb4br\") pod \"placement-db-sync-bsrnx\" (UID: \"7d61014b-9453-4db8-8ae5-226d01d076a6\") " pod="openstack/placement-db-sync-bsrnx" Oct 05 08:21:55 crc kubenswrapper[4846]: I1005 08:21:55.987228 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7d61014b-9453-4db8-8ae5-226d01d076a6-logs\") pod \"placement-db-sync-bsrnx\" (UID: \"7d61014b-9453-4db8-8ae5-226d01d076a6\") " pod="openstack/placement-db-sync-bsrnx" Oct 05 08:21:55 crc kubenswrapper[4846]: I1005 08:21:55.987264 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nd7ts\" (UniqueName: \"kubernetes.io/projected/b051b1a4-96d9-442e-964e-b8d08275caa8-kube-api-access-nd7ts\") pod \"dnsmasq-dns-77bbb5966f-sbz7q\" (UID: \"b051b1a4-96d9-442e-964e-b8d08275caa8\") " pod="openstack/dnsmasq-dns-77bbb5966f-sbz7q" Oct 05 08:21:55 crc kubenswrapper[4846]: I1005 08:21:55.987285 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b051b1a4-96d9-442e-964e-b8d08275caa8-dns-svc\") pod \"dnsmasq-dns-77bbb5966f-sbz7q\" (UID: \"b051b1a4-96d9-442e-964e-b8d08275caa8\") " pod="openstack/dnsmasq-dns-77bbb5966f-sbz7q" Oct 05 08:21:55 crc kubenswrapper[4846]: I1005 08:21:55.987300 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7d61014b-9453-4db8-8ae5-226d01d076a6-config-data\") pod \"placement-db-sync-bsrnx\" (UID: \"7d61014b-9453-4db8-8ae5-226d01d076a6\") " pod="openstack/placement-db-sync-bsrnx" Oct 05 08:21:55 crc kubenswrapper[4846]: I1005 08:21:55.987347 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b051b1a4-96d9-442e-964e-b8d08275caa8-config\") pod \"dnsmasq-dns-77bbb5966f-sbz7q\" (UID: \"b051b1a4-96d9-442e-964e-b8d08275caa8\") " pod="openstack/dnsmasq-dns-77bbb5966f-sbz7q" Oct 05 08:21:55 crc kubenswrapper[4846]: I1005 08:21:55.988007 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7d61014b-9453-4db8-8ae5-226d01d076a6-logs\") pod \"placement-db-sync-bsrnx\" (UID: \"7d61014b-9453-4db8-8ae5-226d01d076a6\") " pod="openstack/placement-db-sync-bsrnx" Oct 05 08:21:55 crc kubenswrapper[4846]: I1005 08:21:55.988124 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b051b1a4-96d9-442e-964e-b8d08275caa8-ovsdbserver-sb\") pod \"dnsmasq-dns-77bbb5966f-sbz7q\" (UID: \"b051b1a4-96d9-442e-964e-b8d08275caa8\") " pod="openstack/dnsmasq-dns-77bbb5966f-sbz7q" Oct 05 08:21:55 crc kubenswrapper[4846]: I1005 08:21:55.988610 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b051b1a4-96d9-442e-964e-b8d08275caa8-config\") pod \"dnsmasq-dns-77bbb5966f-sbz7q\" (UID: \"b051b1a4-96d9-442e-964e-b8d08275caa8\") " pod="openstack/dnsmasq-dns-77bbb5966f-sbz7q" Oct 05 08:21:55 crc kubenswrapper[4846]: I1005 08:21:55.988766 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b051b1a4-96d9-442e-964e-b8d08275caa8-ovsdbserver-nb\") pod \"dnsmasq-dns-77bbb5966f-sbz7q\" (UID: \"b051b1a4-96d9-442e-964e-b8d08275caa8\") " pod="openstack/dnsmasq-dns-77bbb5966f-sbz7q" Oct 05 08:21:55 crc kubenswrapper[4846]: I1005 08:21:55.988948 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b051b1a4-96d9-442e-964e-b8d08275caa8-dns-svc\") pod \"dnsmasq-dns-77bbb5966f-sbz7q\" (UID: \"b051b1a4-96d9-442e-964e-b8d08275caa8\") " pod="openstack/dnsmasq-dns-77bbb5966f-sbz7q" Oct 05 08:21:55 crc kubenswrapper[4846]: I1005 08:21:55.991252 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7d61014b-9453-4db8-8ae5-226d01d076a6-scripts\") pod \"placement-db-sync-bsrnx\" (UID: \"7d61014b-9453-4db8-8ae5-226d01d076a6\") " pod="openstack/placement-db-sync-bsrnx" Oct 05 08:21:55 crc kubenswrapper[4846]: I1005 08:21:55.991864 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7d61014b-9453-4db8-8ae5-226d01d076a6-config-data\") pod \"placement-db-sync-bsrnx\" (UID: \"7d61014b-9453-4db8-8ae5-226d01d076a6\") " pod="openstack/placement-db-sync-bsrnx" Oct 05 08:21:55 crc kubenswrapper[4846]: I1005 08:21:55.992604 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d61014b-9453-4db8-8ae5-226d01d076a6-combined-ca-bundle\") pod \"placement-db-sync-bsrnx\" (UID: \"7d61014b-9453-4db8-8ae5-226d01d076a6\") " pod="openstack/placement-db-sync-bsrnx" Oct 05 08:21:56 crc kubenswrapper[4846]: I1005 08:21:56.006918 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sb4br\" (UniqueName: \"kubernetes.io/projected/7d61014b-9453-4db8-8ae5-226d01d076a6-kube-api-access-sb4br\") pod \"placement-db-sync-bsrnx\" (UID: \"7d61014b-9453-4db8-8ae5-226d01d076a6\") " pod="openstack/placement-db-sync-bsrnx" Oct 05 08:21:56 crc kubenswrapper[4846]: I1005 08:21:56.013004 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nd7ts\" (UniqueName: \"kubernetes.io/projected/b051b1a4-96d9-442e-964e-b8d08275caa8-kube-api-access-nd7ts\") pod \"dnsmasq-dns-77bbb5966f-sbz7q\" (UID: \"b051b1a4-96d9-442e-964e-b8d08275caa8\") " pod="openstack/dnsmasq-dns-77bbb5966f-sbz7q" Oct 05 08:21:56 crc kubenswrapper[4846]: I1005 08:21:56.065216 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-77bbb5966f-sbz7q" Oct 05 08:21:56 crc kubenswrapper[4846]: I1005 08:21:56.081881 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-bsrnx" Oct 05 08:21:56 crc kubenswrapper[4846]: I1005 08:21:56.521378 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-77bbb5966f-sbz7q"] Oct 05 08:21:56 crc kubenswrapper[4846]: I1005 08:21:56.531012 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-bsrnx"] Oct 05 08:21:56 crc kubenswrapper[4846]: W1005 08:21:56.531100 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb051b1a4_96d9_442e_964e_b8d08275caa8.slice/crio-7aa5f13a54200698ef8ab86faf1817aa7f433aa16e6a8a846eb2a272900fede0 WatchSource:0}: Error finding container 7aa5f13a54200698ef8ab86faf1817aa7f433aa16e6a8a846eb2a272900fede0: Status 404 returned error can't find the container with id 7aa5f13a54200698ef8ab86faf1817aa7f433aa16e6a8a846eb2a272900fede0 Oct 05 08:21:57 crc kubenswrapper[4846]: I1005 08:21:57.365847 4846 generic.go:334] "Generic (PLEG): container finished" podID="b051b1a4-96d9-442e-964e-b8d08275caa8" containerID="c9c740790070d806894467efe2d16ac8d06520c7277b300d876b021acda86422" exitCode=0 Oct 05 08:21:57 crc kubenswrapper[4846]: I1005 08:21:57.365916 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77bbb5966f-sbz7q" event={"ID":"b051b1a4-96d9-442e-964e-b8d08275caa8","Type":"ContainerDied","Data":"c9c740790070d806894467efe2d16ac8d06520c7277b300d876b021acda86422"} Oct 05 08:21:57 crc kubenswrapper[4846]: I1005 08:21:57.365948 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77bbb5966f-sbz7q" event={"ID":"b051b1a4-96d9-442e-964e-b8d08275caa8","Type":"ContainerStarted","Data":"7aa5f13a54200698ef8ab86faf1817aa7f433aa16e6a8a846eb2a272900fede0"} Oct 05 08:21:57 crc kubenswrapper[4846]: I1005 08:21:57.371862 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-bsrnx" event={"ID":"7d61014b-9453-4db8-8ae5-226d01d076a6","Type":"ContainerStarted","Data":"e8ce6310658f662e3fa41c202f713d7734fe09e58f11506c693359c86c85ccc6"} Oct 05 08:21:58 crc kubenswrapper[4846]: I1005 08:21:58.384749 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77bbb5966f-sbz7q" event={"ID":"b051b1a4-96d9-442e-964e-b8d08275caa8","Type":"ContainerStarted","Data":"e82e7b7fc0cc2840f78eca7217d9dac879df1726b4e407d90ddf0068193d427d"} Oct 05 08:21:58 crc kubenswrapper[4846]: I1005 08:21:58.385455 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-77bbb5966f-sbz7q" Oct 05 08:21:58 crc kubenswrapper[4846]: I1005 08:21:58.413612 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-77bbb5966f-sbz7q" podStartSLOduration=3.4135903069999998 podStartE2EDuration="3.413590307s" podCreationTimestamp="2025-10-05 08:21:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 08:21:58.406367584 +0000 UTC m=+5640.647220359" watchObservedRunningTime="2025-10-05 08:21:58.413590307 +0000 UTC m=+5640.654443082" Oct 05 08:22:00 crc kubenswrapper[4846]: I1005 08:22:00.410529 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-bsrnx" event={"ID":"7d61014b-9453-4db8-8ae5-226d01d076a6","Type":"ContainerStarted","Data":"c93489b784c6b4122970b363d7fff9e7a3a2208c64021f94bd3ca2ab12d03f6b"} Oct 05 08:22:00 crc kubenswrapper[4846]: I1005 08:22:00.437397 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-sync-bsrnx" podStartSLOduration=1.945728569 podStartE2EDuration="5.437379355s" podCreationTimestamp="2025-10-05 08:21:55 +0000 UTC" firstStartedPulling="2025-10-05 08:21:56.532923922 +0000 UTC m=+5638.773776717" lastFinishedPulling="2025-10-05 08:22:00.024574728 +0000 UTC m=+5642.265427503" observedRunningTime="2025-10-05 08:22:00.431869167 +0000 UTC m=+5642.672721962" watchObservedRunningTime="2025-10-05 08:22:00.437379355 +0000 UTC m=+5642.678232130" Oct 05 08:22:01 crc kubenswrapper[4846]: I1005 08:22:01.426583 4846 generic.go:334] "Generic (PLEG): container finished" podID="7d61014b-9453-4db8-8ae5-226d01d076a6" containerID="c93489b784c6b4122970b363d7fff9e7a3a2208c64021f94bd3ca2ab12d03f6b" exitCode=0 Oct 05 08:22:01 crc kubenswrapper[4846]: I1005 08:22:01.426665 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-bsrnx" event={"ID":"7d61014b-9453-4db8-8ae5-226d01d076a6","Type":"ContainerDied","Data":"c93489b784c6b4122970b363d7fff9e7a3a2208c64021f94bd3ca2ab12d03f6b"} Oct 05 08:22:02 crc kubenswrapper[4846]: I1005 08:22:02.907423 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-bsrnx" Oct 05 08:22:03 crc kubenswrapper[4846]: I1005 08:22:03.042433 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d61014b-9453-4db8-8ae5-226d01d076a6-combined-ca-bundle\") pod \"7d61014b-9453-4db8-8ae5-226d01d076a6\" (UID: \"7d61014b-9453-4db8-8ae5-226d01d076a6\") " Oct 05 08:22:03 crc kubenswrapper[4846]: I1005 08:22:03.042586 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb4br\" (UniqueName: \"kubernetes.io/projected/7d61014b-9453-4db8-8ae5-226d01d076a6-kube-api-access-sb4br\") pod \"7d61014b-9453-4db8-8ae5-226d01d076a6\" (UID: \"7d61014b-9453-4db8-8ae5-226d01d076a6\") " Oct 05 08:22:03 crc kubenswrapper[4846]: I1005 08:22:03.042631 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7d61014b-9453-4db8-8ae5-226d01d076a6-config-data\") pod \"7d61014b-9453-4db8-8ae5-226d01d076a6\" (UID: \"7d61014b-9453-4db8-8ae5-226d01d076a6\") " Oct 05 08:22:03 crc kubenswrapper[4846]: I1005 08:22:03.042827 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7d61014b-9453-4db8-8ae5-226d01d076a6-scripts\") pod \"7d61014b-9453-4db8-8ae5-226d01d076a6\" (UID: \"7d61014b-9453-4db8-8ae5-226d01d076a6\") " Oct 05 08:22:03 crc kubenswrapper[4846]: I1005 08:22:03.042893 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7d61014b-9453-4db8-8ae5-226d01d076a6-logs\") pod \"7d61014b-9453-4db8-8ae5-226d01d076a6\" (UID: \"7d61014b-9453-4db8-8ae5-226d01d076a6\") " Oct 05 08:22:03 crc kubenswrapper[4846]: I1005 08:22:03.043513 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7d61014b-9453-4db8-8ae5-226d01d076a6-logs" (OuterVolumeSpecName: "logs") pod "7d61014b-9453-4db8-8ae5-226d01d076a6" (UID: "7d61014b-9453-4db8-8ae5-226d01d076a6"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 08:22:03 crc kubenswrapper[4846]: I1005 08:22:03.047839 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7d61014b-9453-4db8-8ae5-226d01d076a6-scripts" (OuterVolumeSpecName: "scripts") pod "7d61014b-9453-4db8-8ae5-226d01d076a6" (UID: "7d61014b-9453-4db8-8ae5-226d01d076a6"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:22:03 crc kubenswrapper[4846]: I1005 08:22:03.049540 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7d61014b-9453-4db8-8ae5-226d01d076a6-kube-api-access-sb4br" (OuterVolumeSpecName: "kube-api-access-sb4br") pod "7d61014b-9453-4db8-8ae5-226d01d076a6" (UID: "7d61014b-9453-4db8-8ae5-226d01d076a6"). InnerVolumeSpecName "kube-api-access-sb4br". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:22:03 crc kubenswrapper[4846]: I1005 08:22:03.068310 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7d61014b-9453-4db8-8ae5-226d01d076a6-config-data" (OuterVolumeSpecName: "config-data") pod "7d61014b-9453-4db8-8ae5-226d01d076a6" (UID: "7d61014b-9453-4db8-8ae5-226d01d076a6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:22:03 crc kubenswrapper[4846]: I1005 08:22:03.086686 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7d61014b-9453-4db8-8ae5-226d01d076a6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7d61014b-9453-4db8-8ae5-226d01d076a6" (UID: "7d61014b-9453-4db8-8ae5-226d01d076a6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:22:03 crc kubenswrapper[4846]: I1005 08:22:03.146543 4846 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7d61014b-9453-4db8-8ae5-226d01d076a6-scripts\") on node \"crc\" DevicePath \"\"" Oct 05 08:22:03 crc kubenswrapper[4846]: I1005 08:22:03.146826 4846 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7d61014b-9453-4db8-8ae5-226d01d076a6-logs\") on node \"crc\" DevicePath \"\"" Oct 05 08:22:03 crc kubenswrapper[4846]: I1005 08:22:03.146955 4846 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d61014b-9453-4db8-8ae5-226d01d076a6-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 08:22:03 crc kubenswrapper[4846]: I1005 08:22:03.147214 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb4br\" (UniqueName: \"kubernetes.io/projected/7d61014b-9453-4db8-8ae5-226d01d076a6-kube-api-access-sb4br\") on node \"crc\" DevicePath \"\"" Oct 05 08:22:03 crc kubenswrapper[4846]: I1005 08:22:03.147422 4846 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7d61014b-9453-4db8-8ae5-226d01d076a6-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 08:22:03 crc kubenswrapper[4846]: I1005 08:22:03.448154 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-bsrnx" event={"ID":"7d61014b-9453-4db8-8ae5-226d01d076a6","Type":"ContainerDied","Data":"e8ce6310658f662e3fa41c202f713d7734fe09e58f11506c693359c86c85ccc6"} Oct 05 08:22:03 crc kubenswrapper[4846]: I1005 08:22:03.448221 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e8ce6310658f662e3fa41c202f713d7734fe09e58f11506c693359c86c85ccc6" Oct 05 08:22:03 crc kubenswrapper[4846]: I1005 08:22:03.448291 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-bsrnx" Oct 05 08:22:03 crc kubenswrapper[4846]: I1005 08:22:03.556112 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-6fb6f8b89d-dvr8k"] Oct 05 08:22:03 crc kubenswrapper[4846]: E1005 08:22:03.556615 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7d61014b-9453-4db8-8ae5-226d01d076a6" containerName="placement-db-sync" Oct 05 08:22:03 crc kubenswrapper[4846]: I1005 08:22:03.556636 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="7d61014b-9453-4db8-8ae5-226d01d076a6" containerName="placement-db-sync" Oct 05 08:22:03 crc kubenswrapper[4846]: I1005 08:22:03.556866 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="7d61014b-9453-4db8-8ae5-226d01d076a6" containerName="placement-db-sync" Oct 05 08:22:03 crc kubenswrapper[4846]: I1005 08:22:03.558092 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-6fb6f8b89d-dvr8k" Oct 05 08:22:03 crc kubenswrapper[4846]: I1005 08:22:03.561042 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Oct 05 08:22:03 crc kubenswrapper[4846]: I1005 08:22:03.561204 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-5xd6t" Oct 05 08:22:03 crc kubenswrapper[4846]: I1005 08:22:03.561686 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Oct 05 08:22:03 crc kubenswrapper[4846]: I1005 08:22:03.562700 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-public-svc" Oct 05 08:22:03 crc kubenswrapper[4846]: I1005 08:22:03.562740 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-internal-svc" Oct 05 08:22:03 crc kubenswrapper[4846]: I1005 08:22:03.583831 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-6fb6f8b89d-dvr8k"] Oct 05 08:22:03 crc kubenswrapper[4846]: I1005 08:22:03.659018 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7qjjm\" (UniqueName: \"kubernetes.io/projected/fbf245a7-f5e8-4ddf-89eb-737454886513-kube-api-access-7qjjm\") pod \"placement-6fb6f8b89d-dvr8k\" (UID: \"fbf245a7-f5e8-4ddf-89eb-737454886513\") " pod="openstack/placement-6fb6f8b89d-dvr8k" Oct 05 08:22:03 crc kubenswrapper[4846]: I1005 08:22:03.659143 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fbf245a7-f5e8-4ddf-89eb-737454886513-combined-ca-bundle\") pod \"placement-6fb6f8b89d-dvr8k\" (UID: \"fbf245a7-f5e8-4ddf-89eb-737454886513\") " pod="openstack/placement-6fb6f8b89d-dvr8k" Oct 05 08:22:03 crc kubenswrapper[4846]: I1005 08:22:03.659210 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/fbf245a7-f5e8-4ddf-89eb-737454886513-public-tls-certs\") pod \"placement-6fb6f8b89d-dvr8k\" (UID: \"fbf245a7-f5e8-4ddf-89eb-737454886513\") " pod="openstack/placement-6fb6f8b89d-dvr8k" Oct 05 08:22:03 crc kubenswrapper[4846]: I1005 08:22:03.659254 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fbf245a7-f5e8-4ddf-89eb-737454886513-scripts\") pod \"placement-6fb6f8b89d-dvr8k\" (UID: \"fbf245a7-f5e8-4ddf-89eb-737454886513\") " pod="openstack/placement-6fb6f8b89d-dvr8k" Oct 05 08:22:03 crc kubenswrapper[4846]: I1005 08:22:03.659300 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fbf245a7-f5e8-4ddf-89eb-737454886513-config-data\") pod \"placement-6fb6f8b89d-dvr8k\" (UID: \"fbf245a7-f5e8-4ddf-89eb-737454886513\") " pod="openstack/placement-6fb6f8b89d-dvr8k" Oct 05 08:22:03 crc kubenswrapper[4846]: I1005 08:22:03.659343 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/fbf245a7-f5e8-4ddf-89eb-737454886513-internal-tls-certs\") pod \"placement-6fb6f8b89d-dvr8k\" (UID: \"fbf245a7-f5e8-4ddf-89eb-737454886513\") " pod="openstack/placement-6fb6f8b89d-dvr8k" Oct 05 08:22:03 crc kubenswrapper[4846]: I1005 08:22:03.659376 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fbf245a7-f5e8-4ddf-89eb-737454886513-logs\") pod \"placement-6fb6f8b89d-dvr8k\" (UID: \"fbf245a7-f5e8-4ddf-89eb-737454886513\") " pod="openstack/placement-6fb6f8b89d-dvr8k" Oct 05 08:22:03 crc kubenswrapper[4846]: I1005 08:22:03.761008 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fbf245a7-f5e8-4ddf-89eb-737454886513-combined-ca-bundle\") pod \"placement-6fb6f8b89d-dvr8k\" (UID: \"fbf245a7-f5e8-4ddf-89eb-737454886513\") " pod="openstack/placement-6fb6f8b89d-dvr8k" Oct 05 08:22:03 crc kubenswrapper[4846]: I1005 08:22:03.761324 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/fbf245a7-f5e8-4ddf-89eb-737454886513-public-tls-certs\") pod \"placement-6fb6f8b89d-dvr8k\" (UID: \"fbf245a7-f5e8-4ddf-89eb-737454886513\") " pod="openstack/placement-6fb6f8b89d-dvr8k" Oct 05 08:22:03 crc kubenswrapper[4846]: I1005 08:22:03.761360 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fbf245a7-f5e8-4ddf-89eb-737454886513-scripts\") pod \"placement-6fb6f8b89d-dvr8k\" (UID: \"fbf245a7-f5e8-4ddf-89eb-737454886513\") " pod="openstack/placement-6fb6f8b89d-dvr8k" Oct 05 08:22:03 crc kubenswrapper[4846]: I1005 08:22:03.761422 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fbf245a7-f5e8-4ddf-89eb-737454886513-config-data\") pod \"placement-6fb6f8b89d-dvr8k\" (UID: \"fbf245a7-f5e8-4ddf-89eb-737454886513\") " pod="openstack/placement-6fb6f8b89d-dvr8k" Oct 05 08:22:03 crc kubenswrapper[4846]: I1005 08:22:03.761460 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/fbf245a7-f5e8-4ddf-89eb-737454886513-internal-tls-certs\") pod \"placement-6fb6f8b89d-dvr8k\" (UID: \"fbf245a7-f5e8-4ddf-89eb-737454886513\") " pod="openstack/placement-6fb6f8b89d-dvr8k" Oct 05 08:22:03 crc kubenswrapper[4846]: I1005 08:22:03.761493 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fbf245a7-f5e8-4ddf-89eb-737454886513-logs\") pod \"placement-6fb6f8b89d-dvr8k\" (UID: \"fbf245a7-f5e8-4ddf-89eb-737454886513\") " pod="openstack/placement-6fb6f8b89d-dvr8k" Oct 05 08:22:03 crc kubenswrapper[4846]: I1005 08:22:03.761536 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7qjjm\" (UniqueName: \"kubernetes.io/projected/fbf245a7-f5e8-4ddf-89eb-737454886513-kube-api-access-7qjjm\") pod \"placement-6fb6f8b89d-dvr8k\" (UID: \"fbf245a7-f5e8-4ddf-89eb-737454886513\") " pod="openstack/placement-6fb6f8b89d-dvr8k" Oct 05 08:22:03 crc kubenswrapper[4846]: I1005 08:22:03.762208 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fbf245a7-f5e8-4ddf-89eb-737454886513-logs\") pod \"placement-6fb6f8b89d-dvr8k\" (UID: \"fbf245a7-f5e8-4ddf-89eb-737454886513\") " pod="openstack/placement-6fb6f8b89d-dvr8k" Oct 05 08:22:03 crc kubenswrapper[4846]: I1005 08:22:03.765796 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fbf245a7-f5e8-4ddf-89eb-737454886513-scripts\") pod \"placement-6fb6f8b89d-dvr8k\" (UID: \"fbf245a7-f5e8-4ddf-89eb-737454886513\") " pod="openstack/placement-6fb6f8b89d-dvr8k" Oct 05 08:22:03 crc kubenswrapper[4846]: I1005 08:22:03.766842 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fbf245a7-f5e8-4ddf-89eb-737454886513-combined-ca-bundle\") pod \"placement-6fb6f8b89d-dvr8k\" (UID: \"fbf245a7-f5e8-4ddf-89eb-737454886513\") " pod="openstack/placement-6fb6f8b89d-dvr8k" Oct 05 08:22:03 crc kubenswrapper[4846]: I1005 08:22:03.772676 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fbf245a7-f5e8-4ddf-89eb-737454886513-config-data\") pod \"placement-6fb6f8b89d-dvr8k\" (UID: \"fbf245a7-f5e8-4ddf-89eb-737454886513\") " pod="openstack/placement-6fb6f8b89d-dvr8k" Oct 05 08:22:03 crc kubenswrapper[4846]: I1005 08:22:03.773691 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/fbf245a7-f5e8-4ddf-89eb-737454886513-internal-tls-certs\") pod \"placement-6fb6f8b89d-dvr8k\" (UID: \"fbf245a7-f5e8-4ddf-89eb-737454886513\") " pod="openstack/placement-6fb6f8b89d-dvr8k" Oct 05 08:22:03 crc kubenswrapper[4846]: I1005 08:22:03.774695 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/fbf245a7-f5e8-4ddf-89eb-737454886513-public-tls-certs\") pod \"placement-6fb6f8b89d-dvr8k\" (UID: \"fbf245a7-f5e8-4ddf-89eb-737454886513\") " pod="openstack/placement-6fb6f8b89d-dvr8k" Oct 05 08:22:03 crc kubenswrapper[4846]: I1005 08:22:03.775676 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7qjjm\" (UniqueName: \"kubernetes.io/projected/fbf245a7-f5e8-4ddf-89eb-737454886513-kube-api-access-7qjjm\") pod \"placement-6fb6f8b89d-dvr8k\" (UID: \"fbf245a7-f5e8-4ddf-89eb-737454886513\") " pod="openstack/placement-6fb6f8b89d-dvr8k" Oct 05 08:22:03 crc kubenswrapper[4846]: I1005 08:22:03.884049 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-6fb6f8b89d-dvr8k" Oct 05 08:22:04 crc kubenswrapper[4846]: I1005 08:22:04.388787 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-6fb6f8b89d-dvr8k"] Oct 05 08:22:04 crc kubenswrapper[4846]: W1005 08:22:04.391876 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfbf245a7_f5e8_4ddf_89eb_737454886513.slice/crio-f18d8f56fa1ed18e6d713687c204ad9e8964b950b585e3c77f698c136d4c03fa WatchSource:0}: Error finding container f18d8f56fa1ed18e6d713687c204ad9e8964b950b585e3c77f698c136d4c03fa: Status 404 returned error can't find the container with id f18d8f56fa1ed18e6d713687c204ad9e8964b950b585e3c77f698c136d4c03fa Oct 05 08:22:04 crc kubenswrapper[4846]: I1005 08:22:04.466255 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-6fb6f8b89d-dvr8k" event={"ID":"fbf245a7-f5e8-4ddf-89eb-737454886513","Type":"ContainerStarted","Data":"f18d8f56fa1ed18e6d713687c204ad9e8964b950b585e3c77f698c136d4c03fa"} Oct 05 08:22:05 crc kubenswrapper[4846]: I1005 08:22:05.480551 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-6fb6f8b89d-dvr8k" event={"ID":"fbf245a7-f5e8-4ddf-89eb-737454886513","Type":"ContainerStarted","Data":"e6ea6dddbe9e081edb3f76357a230352d8ee94c4b9b0f9fa5791846f81177e0e"} Oct 05 08:22:05 crc kubenswrapper[4846]: I1005 08:22:05.481059 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-6fb6f8b89d-dvr8k" event={"ID":"fbf245a7-f5e8-4ddf-89eb-737454886513","Type":"ContainerStarted","Data":"abd184c9ccac41b10afa586c11b7dcc9bbad400b2bcd85e0de4d750f6ded9ea9"} Oct 05 08:22:05 crc kubenswrapper[4846]: I1005 08:22:05.483269 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-6fb6f8b89d-dvr8k" Oct 05 08:22:05 crc kubenswrapper[4846]: I1005 08:22:05.483477 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-6fb6f8b89d-dvr8k" Oct 05 08:22:05 crc kubenswrapper[4846]: I1005 08:22:05.515975 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-6fb6f8b89d-dvr8k" podStartSLOduration=2.515941746 podStartE2EDuration="2.515941746s" podCreationTimestamp="2025-10-05 08:22:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 08:22:05.50712405 +0000 UTC m=+5647.747976855" watchObservedRunningTime="2025-10-05 08:22:05.515941746 +0000 UTC m=+5647.756794571" Oct 05 08:22:06 crc kubenswrapper[4846]: I1005 08:22:06.068230 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-77bbb5966f-sbz7q" Oct 05 08:22:06 crc kubenswrapper[4846]: I1005 08:22:06.136676 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-59bf5dd889-89l2d"] Oct 05 08:22:06 crc kubenswrapper[4846]: I1005 08:22:06.136947 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-59bf5dd889-89l2d" podUID="363aa87a-56da-4c6b-b6f6-82e66ed412ce" containerName="dnsmasq-dns" containerID="cri-o://5fe82ad5b1b763d4dac8451b390a7ca0634e2c75837c1e2773e60ca5e6dae11c" gracePeriod=10 Oct 05 08:22:06 crc kubenswrapper[4846]: I1005 08:22:06.527402 4846 generic.go:334] "Generic (PLEG): container finished" podID="363aa87a-56da-4c6b-b6f6-82e66ed412ce" containerID="5fe82ad5b1b763d4dac8451b390a7ca0634e2c75837c1e2773e60ca5e6dae11c" exitCode=0 Oct 05 08:22:06 crc kubenswrapper[4846]: I1005 08:22:06.540022 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59bf5dd889-89l2d" event={"ID":"363aa87a-56da-4c6b-b6f6-82e66ed412ce","Type":"ContainerDied","Data":"5fe82ad5b1b763d4dac8451b390a7ca0634e2c75837c1e2773e60ca5e6dae11c"} Oct 05 08:22:06 crc kubenswrapper[4846]: I1005 08:22:06.637974 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-59bf5dd889-89l2d" Oct 05 08:22:06 crc kubenswrapper[4846]: I1005 08:22:06.723241 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/363aa87a-56da-4c6b-b6f6-82e66ed412ce-ovsdbserver-nb\") pod \"363aa87a-56da-4c6b-b6f6-82e66ed412ce\" (UID: \"363aa87a-56da-4c6b-b6f6-82e66ed412ce\") " Oct 05 08:22:06 crc kubenswrapper[4846]: I1005 08:22:06.723326 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/363aa87a-56da-4c6b-b6f6-82e66ed412ce-dns-svc\") pod \"363aa87a-56da-4c6b-b6f6-82e66ed412ce\" (UID: \"363aa87a-56da-4c6b-b6f6-82e66ed412ce\") " Oct 05 08:22:06 crc kubenswrapper[4846]: I1005 08:22:06.723378 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/363aa87a-56da-4c6b-b6f6-82e66ed412ce-ovsdbserver-sb\") pod \"363aa87a-56da-4c6b-b6f6-82e66ed412ce\" (UID: \"363aa87a-56da-4c6b-b6f6-82e66ed412ce\") " Oct 05 08:22:06 crc kubenswrapper[4846]: I1005 08:22:06.723397 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4tfjp\" (UniqueName: \"kubernetes.io/projected/363aa87a-56da-4c6b-b6f6-82e66ed412ce-kube-api-access-4tfjp\") pod \"363aa87a-56da-4c6b-b6f6-82e66ed412ce\" (UID: \"363aa87a-56da-4c6b-b6f6-82e66ed412ce\") " Oct 05 08:22:06 crc kubenswrapper[4846]: I1005 08:22:06.723465 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/363aa87a-56da-4c6b-b6f6-82e66ed412ce-config\") pod \"363aa87a-56da-4c6b-b6f6-82e66ed412ce\" (UID: \"363aa87a-56da-4c6b-b6f6-82e66ed412ce\") " Oct 05 08:22:06 crc kubenswrapper[4846]: I1005 08:22:06.735411 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/363aa87a-56da-4c6b-b6f6-82e66ed412ce-kube-api-access-4tfjp" (OuterVolumeSpecName: "kube-api-access-4tfjp") pod "363aa87a-56da-4c6b-b6f6-82e66ed412ce" (UID: "363aa87a-56da-4c6b-b6f6-82e66ed412ce"). InnerVolumeSpecName "kube-api-access-4tfjp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:22:06 crc kubenswrapper[4846]: I1005 08:22:06.770791 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/363aa87a-56da-4c6b-b6f6-82e66ed412ce-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "363aa87a-56da-4c6b-b6f6-82e66ed412ce" (UID: "363aa87a-56da-4c6b-b6f6-82e66ed412ce"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 08:22:06 crc kubenswrapper[4846]: I1005 08:22:06.777524 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/363aa87a-56da-4c6b-b6f6-82e66ed412ce-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "363aa87a-56da-4c6b-b6f6-82e66ed412ce" (UID: "363aa87a-56da-4c6b-b6f6-82e66ed412ce"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 08:22:06 crc kubenswrapper[4846]: I1005 08:22:06.777847 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/363aa87a-56da-4c6b-b6f6-82e66ed412ce-config" (OuterVolumeSpecName: "config") pod "363aa87a-56da-4c6b-b6f6-82e66ed412ce" (UID: "363aa87a-56da-4c6b-b6f6-82e66ed412ce"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 08:22:06 crc kubenswrapper[4846]: I1005 08:22:06.778626 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/363aa87a-56da-4c6b-b6f6-82e66ed412ce-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "363aa87a-56da-4c6b-b6f6-82e66ed412ce" (UID: "363aa87a-56da-4c6b-b6f6-82e66ed412ce"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 08:22:06 crc kubenswrapper[4846]: I1005 08:22:06.826609 4846 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/363aa87a-56da-4c6b-b6f6-82e66ed412ce-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 05 08:22:06 crc kubenswrapper[4846]: I1005 08:22:06.826644 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4tfjp\" (UniqueName: \"kubernetes.io/projected/363aa87a-56da-4c6b-b6f6-82e66ed412ce-kube-api-access-4tfjp\") on node \"crc\" DevicePath \"\"" Oct 05 08:22:06 crc kubenswrapper[4846]: I1005 08:22:06.826656 4846 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/363aa87a-56da-4c6b-b6f6-82e66ed412ce-config\") on node \"crc\" DevicePath \"\"" Oct 05 08:22:06 crc kubenswrapper[4846]: I1005 08:22:06.826664 4846 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/363aa87a-56da-4c6b-b6f6-82e66ed412ce-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 05 08:22:06 crc kubenswrapper[4846]: I1005 08:22:06.826673 4846 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/363aa87a-56da-4c6b-b6f6-82e66ed412ce-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 05 08:22:07 crc kubenswrapper[4846]: I1005 08:22:07.498002 4846 scope.go:117] "RemoveContainer" containerID="fdafa27da24296f9087c33b6b12a6598d2a8d61fc985b06ad680401828c9feb6" Oct 05 08:22:07 crc kubenswrapper[4846]: E1005 08:22:07.498483 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 08:22:07 crc kubenswrapper[4846]: I1005 08:22:07.543173 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59bf5dd889-89l2d" event={"ID":"363aa87a-56da-4c6b-b6f6-82e66ed412ce","Type":"ContainerDied","Data":"f1a42bc1fb2486eb2d7c653376a2277aa7553d2f80858473427c9bc5c170a990"} Oct 05 08:22:07 crc kubenswrapper[4846]: I1005 08:22:07.543280 4846 scope.go:117] "RemoveContainer" containerID="5fe82ad5b1b763d4dac8451b390a7ca0634e2c75837c1e2773e60ca5e6dae11c" Oct 05 08:22:07 crc kubenswrapper[4846]: I1005 08:22:07.543280 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-59bf5dd889-89l2d" Oct 05 08:22:07 crc kubenswrapper[4846]: I1005 08:22:07.574866 4846 scope.go:117] "RemoveContainer" containerID="20a4c275f58000d4bd34403062d8b3fcc48b6c65b849273bf11d495d67cfe3ba" Oct 05 08:22:07 crc kubenswrapper[4846]: I1005 08:22:07.602742 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-59bf5dd889-89l2d"] Oct 05 08:22:07 crc kubenswrapper[4846]: I1005 08:22:07.617144 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-59bf5dd889-89l2d"] Oct 05 08:22:08 crc kubenswrapper[4846]: I1005 08:22:08.512783 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="363aa87a-56da-4c6b-b6f6-82e66ed412ce" path="/var/lib/kubelet/pods/363aa87a-56da-4c6b-b6f6-82e66ed412ce/volumes" Oct 05 08:22:20 crc kubenswrapper[4846]: I1005 08:22:20.498275 4846 scope.go:117] "RemoveContainer" containerID="fdafa27da24296f9087c33b6b12a6598d2a8d61fc985b06ad680401828c9feb6" Oct 05 08:22:20 crc kubenswrapper[4846]: E1005 08:22:20.499979 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 08:22:31 crc kubenswrapper[4846]: I1005 08:22:31.498559 4846 scope.go:117] "RemoveContainer" containerID="fdafa27da24296f9087c33b6b12a6598d2a8d61fc985b06ad680401828c9feb6" Oct 05 08:22:31 crc kubenswrapper[4846]: E1005 08:22:31.499606 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 08:22:32 crc kubenswrapper[4846]: I1005 08:22:32.537600 4846 scope.go:117] "RemoveContainer" containerID="7832c0fde700ceb61df332f26acd0cb00cb3c3565d3185c78b3864e1dabf2b88" Oct 05 08:22:32 crc kubenswrapper[4846]: I1005 08:22:32.560739 4846 scope.go:117] "RemoveContainer" containerID="b5a670ff8fd08b29599244922b266773ed660c53c819ce0149d0e95ceabb3973" Oct 05 08:22:34 crc kubenswrapper[4846]: I1005 08:22:34.888670 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-6fb6f8b89d-dvr8k" Oct 05 08:22:34 crc kubenswrapper[4846]: I1005 08:22:34.889083 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-6fb6f8b89d-dvr8k" Oct 05 08:22:42 crc kubenswrapper[4846]: I1005 08:22:42.498517 4846 scope.go:117] "RemoveContainer" containerID="fdafa27da24296f9087c33b6b12a6598d2a8d61fc985b06ad680401828c9feb6" Oct 05 08:22:42 crc kubenswrapper[4846]: E1005 08:22:42.499743 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 08:22:55 crc kubenswrapper[4846]: I1005 08:22:55.497520 4846 scope.go:117] "RemoveContainer" containerID="fdafa27da24296f9087c33b6b12a6598d2a8d61fc985b06ad680401828c9feb6" Oct 05 08:22:55 crc kubenswrapper[4846]: E1005 08:22:55.498738 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 08:22:57 crc kubenswrapper[4846]: I1005 08:22:57.837532 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-db-create-62kls"] Oct 05 08:22:57 crc kubenswrapper[4846]: E1005 08:22:57.838235 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="363aa87a-56da-4c6b-b6f6-82e66ed412ce" containerName="init" Oct 05 08:22:57 crc kubenswrapper[4846]: I1005 08:22:57.838251 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="363aa87a-56da-4c6b-b6f6-82e66ed412ce" containerName="init" Oct 05 08:22:57 crc kubenswrapper[4846]: E1005 08:22:57.838281 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="363aa87a-56da-4c6b-b6f6-82e66ed412ce" containerName="dnsmasq-dns" Oct 05 08:22:57 crc kubenswrapper[4846]: I1005 08:22:57.838289 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="363aa87a-56da-4c6b-b6f6-82e66ed412ce" containerName="dnsmasq-dns" Oct 05 08:22:57 crc kubenswrapper[4846]: I1005 08:22:57.838490 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="363aa87a-56da-4c6b-b6f6-82e66ed412ce" containerName="dnsmasq-dns" Oct 05 08:22:57 crc kubenswrapper[4846]: I1005 08:22:57.839147 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-62kls" Oct 05 08:22:57 crc kubenswrapper[4846]: I1005 08:22:57.850912 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-62kls"] Oct 05 08:22:57 crc kubenswrapper[4846]: I1005 08:22:57.899805 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xkx9b\" (UniqueName: \"kubernetes.io/projected/1c3333df-ae72-4159-a51d-9e11f0a0669a-kube-api-access-xkx9b\") pod \"nova-api-db-create-62kls\" (UID: \"1c3333df-ae72-4159-a51d-9e11f0a0669a\") " pod="openstack/nova-api-db-create-62kls" Oct 05 08:22:57 crc kubenswrapper[4846]: I1005 08:22:57.936607 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-db-create-q8mtl"] Oct 05 08:22:57 crc kubenswrapper[4846]: I1005 08:22:57.937702 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-q8mtl" Oct 05 08:22:57 crc kubenswrapper[4846]: I1005 08:22:57.970053 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-q8mtl"] Oct 05 08:22:58 crc kubenswrapper[4846]: I1005 08:22:58.001762 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xkx9b\" (UniqueName: \"kubernetes.io/projected/1c3333df-ae72-4159-a51d-9e11f0a0669a-kube-api-access-xkx9b\") pod \"nova-api-db-create-62kls\" (UID: \"1c3333df-ae72-4159-a51d-9e11f0a0669a\") " pod="openstack/nova-api-db-create-62kls" Oct 05 08:22:58 crc kubenswrapper[4846]: I1005 08:22:58.001922 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dfqt2\" (UniqueName: \"kubernetes.io/projected/d720d2a2-9841-47d1-9d55-d5e41d35d947-kube-api-access-dfqt2\") pod \"nova-cell0-db-create-q8mtl\" (UID: \"d720d2a2-9841-47d1-9d55-d5e41d35d947\") " pod="openstack/nova-cell0-db-create-q8mtl" Oct 05 08:22:58 crc kubenswrapper[4846]: I1005 08:22:58.033163 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xkx9b\" (UniqueName: \"kubernetes.io/projected/1c3333df-ae72-4159-a51d-9e11f0a0669a-kube-api-access-xkx9b\") pod \"nova-api-db-create-62kls\" (UID: \"1c3333df-ae72-4159-a51d-9e11f0a0669a\") " pod="openstack/nova-api-db-create-62kls" Oct 05 08:22:58 crc kubenswrapper[4846]: I1005 08:22:58.104665 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dfqt2\" (UniqueName: \"kubernetes.io/projected/d720d2a2-9841-47d1-9d55-d5e41d35d947-kube-api-access-dfqt2\") pod \"nova-cell0-db-create-q8mtl\" (UID: \"d720d2a2-9841-47d1-9d55-d5e41d35d947\") " pod="openstack/nova-cell0-db-create-q8mtl" Oct 05 08:22:58 crc kubenswrapper[4846]: I1005 08:22:58.138770 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dfqt2\" (UniqueName: \"kubernetes.io/projected/d720d2a2-9841-47d1-9d55-d5e41d35d947-kube-api-access-dfqt2\") pod \"nova-cell0-db-create-q8mtl\" (UID: \"d720d2a2-9841-47d1-9d55-d5e41d35d947\") " pod="openstack/nova-cell0-db-create-q8mtl" Oct 05 08:22:58 crc kubenswrapper[4846]: I1005 08:22:58.142102 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-db-create-rc6qh"] Oct 05 08:22:58 crc kubenswrapper[4846]: I1005 08:22:58.144133 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-rc6qh" Oct 05 08:22:58 crc kubenswrapper[4846]: I1005 08:22:58.153604 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-rc6qh"] Oct 05 08:22:58 crc kubenswrapper[4846]: I1005 08:22:58.209729 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mq86k\" (UniqueName: \"kubernetes.io/projected/b0df8f79-c188-4014-9e9e-9e0f692c0357-kube-api-access-mq86k\") pod \"nova-cell1-db-create-rc6qh\" (UID: \"b0df8f79-c188-4014-9e9e-9e0f692c0357\") " pod="openstack/nova-cell1-db-create-rc6qh" Oct 05 08:22:58 crc kubenswrapper[4846]: I1005 08:22:58.219845 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-62kls" Oct 05 08:22:58 crc kubenswrapper[4846]: I1005 08:22:58.254318 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-q8mtl" Oct 05 08:22:58 crc kubenswrapper[4846]: I1005 08:22:58.312776 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mq86k\" (UniqueName: \"kubernetes.io/projected/b0df8f79-c188-4014-9e9e-9e0f692c0357-kube-api-access-mq86k\") pod \"nova-cell1-db-create-rc6qh\" (UID: \"b0df8f79-c188-4014-9e9e-9e0f692c0357\") " pod="openstack/nova-cell1-db-create-rc6qh" Oct 05 08:22:58 crc kubenswrapper[4846]: I1005 08:22:58.343262 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mq86k\" (UniqueName: \"kubernetes.io/projected/b0df8f79-c188-4014-9e9e-9e0f692c0357-kube-api-access-mq86k\") pod \"nova-cell1-db-create-rc6qh\" (UID: \"b0df8f79-c188-4014-9e9e-9e0f692c0357\") " pod="openstack/nova-cell1-db-create-rc6qh" Oct 05 08:22:58 crc kubenswrapper[4846]: I1005 08:22:58.493028 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-rc6qh" Oct 05 08:22:58 crc kubenswrapper[4846]: I1005 08:22:58.716071 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-62kls"] Oct 05 08:22:58 crc kubenswrapper[4846]: W1005 08:22:58.720467 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1c3333df_ae72_4159_a51d_9e11f0a0669a.slice/crio-b300c51fdf9ae1f737d577fbe9b0334e69bccdffecd44fe87fc12647bd9c0dac WatchSource:0}: Error finding container b300c51fdf9ae1f737d577fbe9b0334e69bccdffecd44fe87fc12647bd9c0dac: Status 404 returned error can't find the container with id b300c51fdf9ae1f737d577fbe9b0334e69bccdffecd44fe87fc12647bd9c0dac Oct 05 08:22:58 crc kubenswrapper[4846]: I1005 08:22:58.760777 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-q8mtl"] Oct 05 08:22:58 crc kubenswrapper[4846]: W1005 08:22:58.766995 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd720d2a2_9841_47d1_9d55_d5e41d35d947.slice/crio-53ab95c1efb070d3a0a073be89fed6b0ca0007e9276e1ce5bdc0f1172d887d10 WatchSource:0}: Error finding container 53ab95c1efb070d3a0a073be89fed6b0ca0007e9276e1ce5bdc0f1172d887d10: Status 404 returned error can't find the container with id 53ab95c1efb070d3a0a073be89fed6b0ca0007e9276e1ce5bdc0f1172d887d10 Oct 05 08:22:58 crc kubenswrapper[4846]: I1005 08:22:58.953746 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-rc6qh"] Oct 05 08:22:58 crc kubenswrapper[4846]: W1005 08:22:58.955331 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb0df8f79_c188_4014_9e9e_9e0f692c0357.slice/crio-73797afb4ce549524802dca73c2ebeee6ac6e19e13657b1d110c3e25f2f041a6 WatchSource:0}: Error finding container 73797afb4ce549524802dca73c2ebeee6ac6e19e13657b1d110c3e25f2f041a6: Status 404 returned error can't find the container with id 73797afb4ce549524802dca73c2ebeee6ac6e19e13657b1d110c3e25f2f041a6 Oct 05 08:22:59 crc kubenswrapper[4846]: I1005 08:22:59.056480 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-create-mblxr"] Oct 05 08:22:59 crc kubenswrapper[4846]: I1005 08:22:59.063395 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-create-mblxr"] Oct 05 08:22:59 crc kubenswrapper[4846]: I1005 08:22:59.106205 4846 generic.go:334] "Generic (PLEG): container finished" podID="d720d2a2-9841-47d1-9d55-d5e41d35d947" containerID="1cd3859419831ec9f4d7c87785196b007bc00e0dbe13b32b853243a4dd9589fc" exitCode=0 Oct 05 08:22:59 crc kubenswrapper[4846]: I1005 08:22:59.106278 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-q8mtl" event={"ID":"d720d2a2-9841-47d1-9d55-d5e41d35d947","Type":"ContainerDied","Data":"1cd3859419831ec9f4d7c87785196b007bc00e0dbe13b32b853243a4dd9589fc"} Oct 05 08:22:59 crc kubenswrapper[4846]: I1005 08:22:59.106305 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-q8mtl" event={"ID":"d720d2a2-9841-47d1-9d55-d5e41d35d947","Type":"ContainerStarted","Data":"53ab95c1efb070d3a0a073be89fed6b0ca0007e9276e1ce5bdc0f1172d887d10"} Oct 05 08:22:59 crc kubenswrapper[4846]: I1005 08:22:59.108803 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-rc6qh" event={"ID":"b0df8f79-c188-4014-9e9e-9e0f692c0357","Type":"ContainerStarted","Data":"73797afb4ce549524802dca73c2ebeee6ac6e19e13657b1d110c3e25f2f041a6"} Oct 05 08:22:59 crc kubenswrapper[4846]: I1005 08:22:59.110561 4846 generic.go:334] "Generic (PLEG): container finished" podID="1c3333df-ae72-4159-a51d-9e11f0a0669a" containerID="3aada75a88067f157714735eac38160174336f81486c01c86f742d2382264842" exitCode=0 Oct 05 08:22:59 crc kubenswrapper[4846]: I1005 08:22:59.110606 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-62kls" event={"ID":"1c3333df-ae72-4159-a51d-9e11f0a0669a","Type":"ContainerDied","Data":"3aada75a88067f157714735eac38160174336f81486c01c86f742d2382264842"} Oct 05 08:22:59 crc kubenswrapper[4846]: I1005 08:22:59.110621 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-62kls" event={"ID":"1c3333df-ae72-4159-a51d-9e11f0a0669a","Type":"ContainerStarted","Data":"b300c51fdf9ae1f737d577fbe9b0334e69bccdffecd44fe87fc12647bd9c0dac"} Oct 05 08:23:00 crc kubenswrapper[4846]: I1005 08:23:00.123476 4846 generic.go:334] "Generic (PLEG): container finished" podID="b0df8f79-c188-4014-9e9e-9e0f692c0357" containerID="633f2300149ab2e249a1f8037040539ca61cfa678dd1ab445ac0e86432b15b34" exitCode=0 Oct 05 08:23:00 crc kubenswrapper[4846]: I1005 08:23:00.123638 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-rc6qh" event={"ID":"b0df8f79-c188-4014-9e9e-9e0f692c0357","Type":"ContainerDied","Data":"633f2300149ab2e249a1f8037040539ca61cfa678dd1ab445ac0e86432b15b34"} Oct 05 08:23:00 crc kubenswrapper[4846]: I1005 08:23:00.507194 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd37756c-8442-4121-8c0d-fa534d95edab" path="/var/lib/kubelet/pods/bd37756c-8442-4121-8c0d-fa534d95edab/volumes" Oct 05 08:23:00 crc kubenswrapper[4846]: I1005 08:23:00.556381 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-q8mtl" Oct 05 08:23:00 crc kubenswrapper[4846]: I1005 08:23:00.564140 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-62kls" Oct 05 08:23:00 crc kubenswrapper[4846]: I1005 08:23:00.671673 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dfqt2\" (UniqueName: \"kubernetes.io/projected/d720d2a2-9841-47d1-9d55-d5e41d35d947-kube-api-access-dfqt2\") pod \"d720d2a2-9841-47d1-9d55-d5e41d35d947\" (UID: \"d720d2a2-9841-47d1-9d55-d5e41d35d947\") " Oct 05 08:23:00 crc kubenswrapper[4846]: I1005 08:23:00.671990 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xkx9b\" (UniqueName: \"kubernetes.io/projected/1c3333df-ae72-4159-a51d-9e11f0a0669a-kube-api-access-xkx9b\") pod \"1c3333df-ae72-4159-a51d-9e11f0a0669a\" (UID: \"1c3333df-ae72-4159-a51d-9e11f0a0669a\") " Oct 05 08:23:00 crc kubenswrapper[4846]: I1005 08:23:00.677508 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d720d2a2-9841-47d1-9d55-d5e41d35d947-kube-api-access-dfqt2" (OuterVolumeSpecName: "kube-api-access-dfqt2") pod "d720d2a2-9841-47d1-9d55-d5e41d35d947" (UID: "d720d2a2-9841-47d1-9d55-d5e41d35d947"). InnerVolumeSpecName "kube-api-access-dfqt2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:23:00 crc kubenswrapper[4846]: I1005 08:23:00.679400 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1c3333df-ae72-4159-a51d-9e11f0a0669a-kube-api-access-xkx9b" (OuterVolumeSpecName: "kube-api-access-xkx9b") pod "1c3333df-ae72-4159-a51d-9e11f0a0669a" (UID: "1c3333df-ae72-4159-a51d-9e11f0a0669a"). InnerVolumeSpecName "kube-api-access-xkx9b". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:23:00 crc kubenswrapper[4846]: I1005 08:23:00.774232 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xkx9b\" (UniqueName: \"kubernetes.io/projected/1c3333df-ae72-4159-a51d-9e11f0a0669a-kube-api-access-xkx9b\") on node \"crc\" DevicePath \"\"" Oct 05 08:23:00 crc kubenswrapper[4846]: I1005 08:23:00.774262 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dfqt2\" (UniqueName: \"kubernetes.io/projected/d720d2a2-9841-47d1-9d55-d5e41d35d947-kube-api-access-dfqt2\") on node \"crc\" DevicePath \"\"" Oct 05 08:23:01 crc kubenswrapper[4846]: I1005 08:23:01.140748 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-62kls" Oct 05 08:23:01 crc kubenswrapper[4846]: I1005 08:23:01.140741 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-62kls" event={"ID":"1c3333df-ae72-4159-a51d-9e11f0a0669a","Type":"ContainerDied","Data":"b300c51fdf9ae1f737d577fbe9b0334e69bccdffecd44fe87fc12647bd9c0dac"} Oct 05 08:23:01 crc kubenswrapper[4846]: I1005 08:23:01.140939 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b300c51fdf9ae1f737d577fbe9b0334e69bccdffecd44fe87fc12647bd9c0dac" Oct 05 08:23:01 crc kubenswrapper[4846]: I1005 08:23:01.144398 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-q8mtl" Oct 05 08:23:01 crc kubenswrapper[4846]: I1005 08:23:01.146376 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-q8mtl" event={"ID":"d720d2a2-9841-47d1-9d55-d5e41d35d947","Type":"ContainerDied","Data":"53ab95c1efb070d3a0a073be89fed6b0ca0007e9276e1ce5bdc0f1172d887d10"} Oct 05 08:23:01 crc kubenswrapper[4846]: I1005 08:23:01.146470 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="53ab95c1efb070d3a0a073be89fed6b0ca0007e9276e1ce5bdc0f1172d887d10" Oct 05 08:23:01 crc kubenswrapper[4846]: I1005 08:23:01.528195 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-rc6qh" Oct 05 08:23:01 crc kubenswrapper[4846]: I1005 08:23:01.590682 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mq86k\" (UniqueName: \"kubernetes.io/projected/b0df8f79-c188-4014-9e9e-9e0f692c0357-kube-api-access-mq86k\") pod \"b0df8f79-c188-4014-9e9e-9e0f692c0357\" (UID: \"b0df8f79-c188-4014-9e9e-9e0f692c0357\") " Oct 05 08:23:01 crc kubenswrapper[4846]: I1005 08:23:01.594654 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b0df8f79-c188-4014-9e9e-9e0f692c0357-kube-api-access-mq86k" (OuterVolumeSpecName: "kube-api-access-mq86k") pod "b0df8f79-c188-4014-9e9e-9e0f692c0357" (UID: "b0df8f79-c188-4014-9e9e-9e0f692c0357"). InnerVolumeSpecName "kube-api-access-mq86k". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:23:01 crc kubenswrapper[4846]: I1005 08:23:01.693343 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mq86k\" (UniqueName: \"kubernetes.io/projected/b0df8f79-c188-4014-9e9e-9e0f692c0357-kube-api-access-mq86k\") on node \"crc\" DevicePath \"\"" Oct 05 08:23:02 crc kubenswrapper[4846]: I1005 08:23:02.178060 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-rc6qh" event={"ID":"b0df8f79-c188-4014-9e9e-9e0f692c0357","Type":"ContainerDied","Data":"73797afb4ce549524802dca73c2ebeee6ac6e19e13657b1d110c3e25f2f041a6"} Oct 05 08:23:02 crc kubenswrapper[4846]: I1005 08:23:02.178106 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="73797afb4ce549524802dca73c2ebeee6ac6e19e13657b1d110c3e25f2f041a6" Oct 05 08:23:02 crc kubenswrapper[4846]: I1005 08:23:02.178234 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-rc6qh" Oct 05 08:23:08 crc kubenswrapper[4846]: I1005 08:23:08.094117 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-b599-account-create-b7c5d"] Oct 05 08:23:08 crc kubenswrapper[4846]: E1005 08:23:08.094879 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1c3333df-ae72-4159-a51d-9e11f0a0669a" containerName="mariadb-database-create" Oct 05 08:23:08 crc kubenswrapper[4846]: I1005 08:23:08.094890 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="1c3333df-ae72-4159-a51d-9e11f0a0669a" containerName="mariadb-database-create" Oct 05 08:23:08 crc kubenswrapper[4846]: E1005 08:23:08.094918 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b0df8f79-c188-4014-9e9e-9e0f692c0357" containerName="mariadb-database-create" Oct 05 08:23:08 crc kubenswrapper[4846]: I1005 08:23:08.094925 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="b0df8f79-c188-4014-9e9e-9e0f692c0357" containerName="mariadb-database-create" Oct 05 08:23:08 crc kubenswrapper[4846]: E1005 08:23:08.094940 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d720d2a2-9841-47d1-9d55-d5e41d35d947" containerName="mariadb-database-create" Oct 05 08:23:08 crc kubenswrapper[4846]: I1005 08:23:08.094945 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="d720d2a2-9841-47d1-9d55-d5e41d35d947" containerName="mariadb-database-create" Oct 05 08:23:08 crc kubenswrapper[4846]: I1005 08:23:08.095137 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="1c3333df-ae72-4159-a51d-9e11f0a0669a" containerName="mariadb-database-create" Oct 05 08:23:08 crc kubenswrapper[4846]: I1005 08:23:08.095158 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="d720d2a2-9841-47d1-9d55-d5e41d35d947" containerName="mariadb-database-create" Oct 05 08:23:08 crc kubenswrapper[4846]: I1005 08:23:08.095174 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="b0df8f79-c188-4014-9e9e-9e0f692c0357" containerName="mariadb-database-create" Oct 05 08:23:08 crc kubenswrapper[4846]: I1005 08:23:08.095755 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-b599-account-create-b7c5d" Oct 05 08:23:08 crc kubenswrapper[4846]: I1005 08:23:08.099455 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-db-secret" Oct 05 08:23:08 crc kubenswrapper[4846]: I1005 08:23:08.108884 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-b599-account-create-b7c5d"] Oct 05 08:23:08 crc kubenswrapper[4846]: I1005 08:23:08.224524 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cmt69\" (UniqueName: \"kubernetes.io/projected/5dedc799-d3d4-4fe0-a2de-317784fd1db3-kube-api-access-cmt69\") pod \"nova-api-b599-account-create-b7c5d\" (UID: \"5dedc799-d3d4-4fe0-a2de-317784fd1db3\") " pod="openstack/nova-api-b599-account-create-b7c5d" Oct 05 08:23:08 crc kubenswrapper[4846]: I1005 08:23:08.283930 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-fc2b-account-create-bnrgd"] Oct 05 08:23:08 crc kubenswrapper[4846]: I1005 08:23:08.285193 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-fc2b-account-create-bnrgd" Oct 05 08:23:08 crc kubenswrapper[4846]: I1005 08:23:08.289845 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-db-secret" Oct 05 08:23:08 crc kubenswrapper[4846]: I1005 08:23:08.292368 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-fc2b-account-create-bnrgd"] Oct 05 08:23:08 crc kubenswrapper[4846]: I1005 08:23:08.326124 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cmt69\" (UniqueName: \"kubernetes.io/projected/5dedc799-d3d4-4fe0-a2de-317784fd1db3-kube-api-access-cmt69\") pod \"nova-api-b599-account-create-b7c5d\" (UID: \"5dedc799-d3d4-4fe0-a2de-317784fd1db3\") " pod="openstack/nova-api-b599-account-create-b7c5d" Oct 05 08:23:08 crc kubenswrapper[4846]: I1005 08:23:08.349058 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cmt69\" (UniqueName: \"kubernetes.io/projected/5dedc799-d3d4-4fe0-a2de-317784fd1db3-kube-api-access-cmt69\") pod \"nova-api-b599-account-create-b7c5d\" (UID: \"5dedc799-d3d4-4fe0-a2de-317784fd1db3\") " pod="openstack/nova-api-b599-account-create-b7c5d" Oct 05 08:23:08 crc kubenswrapper[4846]: I1005 08:23:08.418646 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-b599-account-create-b7c5d" Oct 05 08:23:08 crc kubenswrapper[4846]: I1005 08:23:08.427138 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zh22f\" (UniqueName: \"kubernetes.io/projected/3d0781ed-2c7e-452d-8d71-26c62c1c00e1-kube-api-access-zh22f\") pod \"nova-cell0-fc2b-account-create-bnrgd\" (UID: \"3d0781ed-2c7e-452d-8d71-26c62c1c00e1\") " pod="openstack/nova-cell0-fc2b-account-create-bnrgd" Oct 05 08:23:08 crc kubenswrapper[4846]: I1005 08:23:08.485868 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-27ec-account-create-5gjp8"] Oct 05 08:23:08 crc kubenswrapper[4846]: I1005 08:23:08.487026 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-27ec-account-create-5gjp8" Oct 05 08:23:08 crc kubenswrapper[4846]: I1005 08:23:08.489742 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-db-secret" Oct 05 08:23:08 crc kubenswrapper[4846]: I1005 08:23:08.523173 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-27ec-account-create-5gjp8"] Oct 05 08:23:08 crc kubenswrapper[4846]: I1005 08:23:08.529334 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-769zm\" (UniqueName: \"kubernetes.io/projected/54b580fc-9fa1-4871-bfa8-832910ffffec-kube-api-access-769zm\") pod \"nova-cell1-27ec-account-create-5gjp8\" (UID: \"54b580fc-9fa1-4871-bfa8-832910ffffec\") " pod="openstack/nova-cell1-27ec-account-create-5gjp8" Oct 05 08:23:08 crc kubenswrapper[4846]: I1005 08:23:08.529386 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zh22f\" (UniqueName: \"kubernetes.io/projected/3d0781ed-2c7e-452d-8d71-26c62c1c00e1-kube-api-access-zh22f\") pod \"nova-cell0-fc2b-account-create-bnrgd\" (UID: \"3d0781ed-2c7e-452d-8d71-26c62c1c00e1\") " pod="openstack/nova-cell0-fc2b-account-create-bnrgd" Oct 05 08:23:08 crc kubenswrapper[4846]: I1005 08:23:08.548736 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zh22f\" (UniqueName: \"kubernetes.io/projected/3d0781ed-2c7e-452d-8d71-26c62c1c00e1-kube-api-access-zh22f\") pod \"nova-cell0-fc2b-account-create-bnrgd\" (UID: \"3d0781ed-2c7e-452d-8d71-26c62c1c00e1\") " pod="openstack/nova-cell0-fc2b-account-create-bnrgd" Oct 05 08:23:08 crc kubenswrapper[4846]: I1005 08:23:08.601333 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-fc2b-account-create-bnrgd" Oct 05 08:23:08 crc kubenswrapper[4846]: I1005 08:23:08.636422 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-769zm\" (UniqueName: \"kubernetes.io/projected/54b580fc-9fa1-4871-bfa8-832910ffffec-kube-api-access-769zm\") pod \"nova-cell1-27ec-account-create-5gjp8\" (UID: \"54b580fc-9fa1-4871-bfa8-832910ffffec\") " pod="openstack/nova-cell1-27ec-account-create-5gjp8" Oct 05 08:23:08 crc kubenswrapper[4846]: I1005 08:23:08.659870 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-769zm\" (UniqueName: \"kubernetes.io/projected/54b580fc-9fa1-4871-bfa8-832910ffffec-kube-api-access-769zm\") pod \"nova-cell1-27ec-account-create-5gjp8\" (UID: \"54b580fc-9fa1-4871-bfa8-832910ffffec\") " pod="openstack/nova-cell1-27ec-account-create-5gjp8" Oct 05 08:23:08 crc kubenswrapper[4846]: I1005 08:23:08.897809 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-27ec-account-create-5gjp8" Oct 05 08:23:08 crc kubenswrapper[4846]: I1005 08:23:08.901549 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-b599-account-create-b7c5d"] Oct 05 08:23:08 crc kubenswrapper[4846]: W1005 08:23:08.905516 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5dedc799_d3d4_4fe0_a2de_317784fd1db3.slice/crio-76c544f9dc3ae76d1296ca9cc4f780af9ebd0aec142a2c39094b67a7e884a5a0 WatchSource:0}: Error finding container 76c544f9dc3ae76d1296ca9cc4f780af9ebd0aec142a2c39094b67a7e884a5a0: Status 404 returned error can't find the container with id 76c544f9dc3ae76d1296ca9cc4f780af9ebd0aec142a2c39094b67a7e884a5a0 Oct 05 08:23:09 crc kubenswrapper[4846]: I1005 08:23:09.029300 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-d973-account-create-nt7ql"] Oct 05 08:23:09 crc kubenswrapper[4846]: I1005 08:23:09.038339 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-d973-account-create-nt7ql"] Oct 05 08:23:09 crc kubenswrapper[4846]: W1005 08:23:09.060531 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3d0781ed_2c7e_452d_8d71_26c62c1c00e1.slice/crio-b7c36a7fd20b4e5c78876088b653bf835bbe0ff1672e5e7fffd8e1e90245da1e WatchSource:0}: Error finding container b7c36a7fd20b4e5c78876088b653bf835bbe0ff1672e5e7fffd8e1e90245da1e: Status 404 returned error can't find the container with id b7c36a7fd20b4e5c78876088b653bf835bbe0ff1672e5e7fffd8e1e90245da1e Oct 05 08:23:09 crc kubenswrapper[4846]: I1005 08:23:09.061938 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-fc2b-account-create-bnrgd"] Oct 05 08:23:09 crc kubenswrapper[4846]: I1005 08:23:09.252141 4846 generic.go:334] "Generic (PLEG): container finished" podID="5dedc799-d3d4-4fe0-a2de-317784fd1db3" containerID="c44383b11f0e5106d7bda217ca484922769bfcc65aa3687d5dc72e75b0293816" exitCode=0 Oct 05 08:23:09 crc kubenswrapper[4846]: I1005 08:23:09.252244 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-b599-account-create-b7c5d" event={"ID":"5dedc799-d3d4-4fe0-a2de-317784fd1db3","Type":"ContainerDied","Data":"c44383b11f0e5106d7bda217ca484922769bfcc65aa3687d5dc72e75b0293816"} Oct 05 08:23:09 crc kubenswrapper[4846]: I1005 08:23:09.252277 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-b599-account-create-b7c5d" event={"ID":"5dedc799-d3d4-4fe0-a2de-317784fd1db3","Type":"ContainerStarted","Data":"76c544f9dc3ae76d1296ca9cc4f780af9ebd0aec142a2c39094b67a7e884a5a0"} Oct 05 08:23:09 crc kubenswrapper[4846]: I1005 08:23:09.254101 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-fc2b-account-create-bnrgd" event={"ID":"3d0781ed-2c7e-452d-8d71-26c62c1c00e1","Type":"ContainerStarted","Data":"a6ff819f08387f88aee4ad007ec1517ce487f455ff69578da57596e2835b80ff"} Oct 05 08:23:09 crc kubenswrapper[4846]: I1005 08:23:09.254136 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-fc2b-account-create-bnrgd" event={"ID":"3d0781ed-2c7e-452d-8d71-26c62c1c00e1","Type":"ContainerStarted","Data":"b7c36a7fd20b4e5c78876088b653bf835bbe0ff1672e5e7fffd8e1e90245da1e"} Oct 05 08:23:09 crc kubenswrapper[4846]: I1005 08:23:09.280391 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-fc2b-account-create-bnrgd" podStartSLOduration=1.280371486 podStartE2EDuration="1.280371486s" podCreationTimestamp="2025-10-05 08:23:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 08:23:09.279525264 +0000 UTC m=+5711.520378049" watchObservedRunningTime="2025-10-05 08:23:09.280371486 +0000 UTC m=+5711.521224271" Oct 05 08:23:09 crc kubenswrapper[4846]: I1005 08:23:09.345767 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-27ec-account-create-5gjp8"] Oct 05 08:23:10 crc kubenswrapper[4846]: I1005 08:23:10.267278 4846 generic.go:334] "Generic (PLEG): container finished" podID="54b580fc-9fa1-4871-bfa8-832910ffffec" containerID="d27f160f9bbb9185928b4f60c6cd30913c1c0c697e9f2f40ec11972b0a5bbbd3" exitCode=0 Oct 05 08:23:10 crc kubenswrapper[4846]: I1005 08:23:10.267375 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-27ec-account-create-5gjp8" event={"ID":"54b580fc-9fa1-4871-bfa8-832910ffffec","Type":"ContainerDied","Data":"d27f160f9bbb9185928b4f60c6cd30913c1c0c697e9f2f40ec11972b0a5bbbd3"} Oct 05 08:23:10 crc kubenswrapper[4846]: I1005 08:23:10.267431 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-27ec-account-create-5gjp8" event={"ID":"54b580fc-9fa1-4871-bfa8-832910ffffec","Type":"ContainerStarted","Data":"95a919f29ab96d0cdc465540bb5388cccb773902c02b194155600eca737556fc"} Oct 05 08:23:10 crc kubenswrapper[4846]: I1005 08:23:10.269642 4846 generic.go:334] "Generic (PLEG): container finished" podID="3d0781ed-2c7e-452d-8d71-26c62c1c00e1" containerID="a6ff819f08387f88aee4ad007ec1517ce487f455ff69578da57596e2835b80ff" exitCode=0 Oct 05 08:23:10 crc kubenswrapper[4846]: I1005 08:23:10.269730 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-fc2b-account-create-bnrgd" event={"ID":"3d0781ed-2c7e-452d-8d71-26c62c1c00e1","Type":"ContainerDied","Data":"a6ff819f08387f88aee4ad007ec1517ce487f455ff69578da57596e2835b80ff"} Oct 05 08:23:10 crc kubenswrapper[4846]: I1005 08:23:10.501962 4846 scope.go:117] "RemoveContainer" containerID="fdafa27da24296f9087c33b6b12a6598d2a8d61fc985b06ad680401828c9feb6" Oct 05 08:23:10 crc kubenswrapper[4846]: E1005 08:23:10.502651 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 08:23:10 crc kubenswrapper[4846]: I1005 08:23:10.509965 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6f59c11b-1a78-464d-b6c5-2367019b1140" path="/var/lib/kubelet/pods/6f59c11b-1a78-464d-b6c5-2367019b1140/volumes" Oct 05 08:23:10 crc kubenswrapper[4846]: I1005 08:23:10.654962 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-b599-account-create-b7c5d" Oct 05 08:23:10 crc kubenswrapper[4846]: I1005 08:23:10.778941 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cmt69\" (UniqueName: \"kubernetes.io/projected/5dedc799-d3d4-4fe0-a2de-317784fd1db3-kube-api-access-cmt69\") pod \"5dedc799-d3d4-4fe0-a2de-317784fd1db3\" (UID: \"5dedc799-d3d4-4fe0-a2de-317784fd1db3\") " Oct 05 08:23:10 crc kubenswrapper[4846]: I1005 08:23:10.784489 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5dedc799-d3d4-4fe0-a2de-317784fd1db3-kube-api-access-cmt69" (OuterVolumeSpecName: "kube-api-access-cmt69") pod "5dedc799-d3d4-4fe0-a2de-317784fd1db3" (UID: "5dedc799-d3d4-4fe0-a2de-317784fd1db3"). InnerVolumeSpecName "kube-api-access-cmt69". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:23:10 crc kubenswrapper[4846]: I1005 08:23:10.881098 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cmt69\" (UniqueName: \"kubernetes.io/projected/5dedc799-d3d4-4fe0-a2de-317784fd1db3-kube-api-access-cmt69\") on node \"crc\" DevicePath \"\"" Oct 05 08:23:11 crc kubenswrapper[4846]: I1005 08:23:11.282047 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-b599-account-create-b7c5d" event={"ID":"5dedc799-d3d4-4fe0-a2de-317784fd1db3","Type":"ContainerDied","Data":"76c544f9dc3ae76d1296ca9cc4f780af9ebd0aec142a2c39094b67a7e884a5a0"} Oct 05 08:23:11 crc kubenswrapper[4846]: I1005 08:23:11.282125 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="76c544f9dc3ae76d1296ca9cc4f780af9ebd0aec142a2c39094b67a7e884a5a0" Oct 05 08:23:11 crc kubenswrapper[4846]: I1005 08:23:11.282222 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-b599-account-create-b7c5d" Oct 05 08:23:11 crc kubenswrapper[4846]: I1005 08:23:11.736048 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-fc2b-account-create-bnrgd" Oct 05 08:23:11 crc kubenswrapper[4846]: I1005 08:23:11.745892 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-27ec-account-create-5gjp8" Oct 05 08:23:11 crc kubenswrapper[4846]: I1005 08:23:11.801909 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-769zm\" (UniqueName: \"kubernetes.io/projected/54b580fc-9fa1-4871-bfa8-832910ffffec-kube-api-access-769zm\") pod \"54b580fc-9fa1-4871-bfa8-832910ffffec\" (UID: \"54b580fc-9fa1-4871-bfa8-832910ffffec\") " Oct 05 08:23:11 crc kubenswrapper[4846]: I1005 08:23:11.802014 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zh22f\" (UniqueName: \"kubernetes.io/projected/3d0781ed-2c7e-452d-8d71-26c62c1c00e1-kube-api-access-zh22f\") pod \"3d0781ed-2c7e-452d-8d71-26c62c1c00e1\" (UID: \"3d0781ed-2c7e-452d-8d71-26c62c1c00e1\") " Oct 05 08:23:11 crc kubenswrapper[4846]: I1005 08:23:11.805926 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/54b580fc-9fa1-4871-bfa8-832910ffffec-kube-api-access-769zm" (OuterVolumeSpecName: "kube-api-access-769zm") pod "54b580fc-9fa1-4871-bfa8-832910ffffec" (UID: "54b580fc-9fa1-4871-bfa8-832910ffffec"). InnerVolumeSpecName "kube-api-access-769zm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:23:11 crc kubenswrapper[4846]: I1005 08:23:11.813448 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3d0781ed-2c7e-452d-8d71-26c62c1c00e1-kube-api-access-zh22f" (OuterVolumeSpecName: "kube-api-access-zh22f") pod "3d0781ed-2c7e-452d-8d71-26c62c1c00e1" (UID: "3d0781ed-2c7e-452d-8d71-26c62c1c00e1"). InnerVolumeSpecName "kube-api-access-zh22f". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:23:11 crc kubenswrapper[4846]: I1005 08:23:11.903970 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zh22f\" (UniqueName: \"kubernetes.io/projected/3d0781ed-2c7e-452d-8d71-26c62c1c00e1-kube-api-access-zh22f\") on node \"crc\" DevicePath \"\"" Oct 05 08:23:11 crc kubenswrapper[4846]: I1005 08:23:11.904007 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-769zm\" (UniqueName: \"kubernetes.io/projected/54b580fc-9fa1-4871-bfa8-832910ffffec-kube-api-access-769zm\") on node \"crc\" DevicePath \"\"" Oct 05 08:23:12 crc kubenswrapper[4846]: I1005 08:23:12.291906 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-fc2b-account-create-bnrgd" event={"ID":"3d0781ed-2c7e-452d-8d71-26c62c1c00e1","Type":"ContainerDied","Data":"b7c36a7fd20b4e5c78876088b653bf835bbe0ff1672e5e7fffd8e1e90245da1e"} Oct 05 08:23:12 crc kubenswrapper[4846]: I1005 08:23:12.291942 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b7c36a7fd20b4e5c78876088b653bf835bbe0ff1672e5e7fffd8e1e90245da1e" Oct 05 08:23:12 crc kubenswrapper[4846]: I1005 08:23:12.291943 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-fc2b-account-create-bnrgd" Oct 05 08:23:12 crc kubenswrapper[4846]: I1005 08:23:12.294275 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-27ec-account-create-5gjp8" event={"ID":"54b580fc-9fa1-4871-bfa8-832910ffffec","Type":"ContainerDied","Data":"95a919f29ab96d0cdc465540bb5388cccb773902c02b194155600eca737556fc"} Oct 05 08:23:12 crc kubenswrapper[4846]: I1005 08:23:12.294323 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="95a919f29ab96d0cdc465540bb5388cccb773902c02b194155600eca737556fc" Oct 05 08:23:12 crc kubenswrapper[4846]: I1005 08:23:12.294336 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-27ec-account-create-5gjp8" Oct 05 08:23:13 crc kubenswrapper[4846]: I1005 08:23:13.513441 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-db-sync-sskzq"] Oct 05 08:23:13 crc kubenswrapper[4846]: E1005 08:23:13.514365 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3d0781ed-2c7e-452d-8d71-26c62c1c00e1" containerName="mariadb-account-create" Oct 05 08:23:13 crc kubenswrapper[4846]: I1005 08:23:13.514380 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="3d0781ed-2c7e-452d-8d71-26c62c1c00e1" containerName="mariadb-account-create" Oct 05 08:23:13 crc kubenswrapper[4846]: E1005 08:23:13.514399 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="54b580fc-9fa1-4871-bfa8-832910ffffec" containerName="mariadb-account-create" Oct 05 08:23:13 crc kubenswrapper[4846]: I1005 08:23:13.514405 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="54b580fc-9fa1-4871-bfa8-832910ffffec" containerName="mariadb-account-create" Oct 05 08:23:13 crc kubenswrapper[4846]: E1005 08:23:13.514428 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5dedc799-d3d4-4fe0-a2de-317784fd1db3" containerName="mariadb-account-create" Oct 05 08:23:13 crc kubenswrapper[4846]: I1005 08:23:13.514436 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="5dedc799-d3d4-4fe0-a2de-317784fd1db3" containerName="mariadb-account-create" Oct 05 08:23:13 crc kubenswrapper[4846]: I1005 08:23:13.514608 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="3d0781ed-2c7e-452d-8d71-26c62c1c00e1" containerName="mariadb-account-create" Oct 05 08:23:13 crc kubenswrapper[4846]: I1005 08:23:13.514620 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="5dedc799-d3d4-4fe0-a2de-317784fd1db3" containerName="mariadb-account-create" Oct 05 08:23:13 crc kubenswrapper[4846]: I1005 08:23:13.514638 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="54b580fc-9fa1-4871-bfa8-832910ffffec" containerName="mariadb-account-create" Oct 05 08:23:13 crc kubenswrapper[4846]: I1005 08:23:13.515289 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-sskzq" Oct 05 08:23:13 crc kubenswrapper[4846]: I1005 08:23:13.517068 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-scripts" Oct 05 08:23:13 crc kubenswrapper[4846]: I1005 08:23:13.517380 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-hhf6x" Oct 05 08:23:13 crc kubenswrapper[4846]: I1005 08:23:13.518222 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Oct 05 08:23:13 crc kubenswrapper[4846]: I1005 08:23:13.533341 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-sskzq"] Oct 05 08:23:13 crc kubenswrapper[4846]: I1005 08:23:13.635073 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a7e835d9-77a2-4066-b151-42b30db24e78-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-sskzq\" (UID: \"a7e835d9-77a2-4066-b151-42b30db24e78\") " pod="openstack/nova-cell0-conductor-db-sync-sskzq" Oct 05 08:23:13 crc kubenswrapper[4846]: I1005 08:23:13.635297 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a7e835d9-77a2-4066-b151-42b30db24e78-scripts\") pod \"nova-cell0-conductor-db-sync-sskzq\" (UID: \"a7e835d9-77a2-4066-b151-42b30db24e78\") " pod="openstack/nova-cell0-conductor-db-sync-sskzq" Oct 05 08:23:13 crc kubenswrapper[4846]: I1005 08:23:13.635497 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a7e835d9-77a2-4066-b151-42b30db24e78-config-data\") pod \"nova-cell0-conductor-db-sync-sskzq\" (UID: \"a7e835d9-77a2-4066-b151-42b30db24e78\") " pod="openstack/nova-cell0-conductor-db-sync-sskzq" Oct 05 08:23:13 crc kubenswrapper[4846]: I1005 08:23:13.635619 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4dmks\" (UniqueName: \"kubernetes.io/projected/a7e835d9-77a2-4066-b151-42b30db24e78-kube-api-access-4dmks\") pod \"nova-cell0-conductor-db-sync-sskzq\" (UID: \"a7e835d9-77a2-4066-b151-42b30db24e78\") " pod="openstack/nova-cell0-conductor-db-sync-sskzq" Oct 05 08:23:13 crc kubenswrapper[4846]: I1005 08:23:13.738349 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a7e835d9-77a2-4066-b151-42b30db24e78-scripts\") pod \"nova-cell0-conductor-db-sync-sskzq\" (UID: \"a7e835d9-77a2-4066-b151-42b30db24e78\") " pod="openstack/nova-cell0-conductor-db-sync-sskzq" Oct 05 08:23:13 crc kubenswrapper[4846]: I1005 08:23:13.738414 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a7e835d9-77a2-4066-b151-42b30db24e78-config-data\") pod \"nova-cell0-conductor-db-sync-sskzq\" (UID: \"a7e835d9-77a2-4066-b151-42b30db24e78\") " pod="openstack/nova-cell0-conductor-db-sync-sskzq" Oct 05 08:23:13 crc kubenswrapper[4846]: I1005 08:23:13.738447 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4dmks\" (UniqueName: \"kubernetes.io/projected/a7e835d9-77a2-4066-b151-42b30db24e78-kube-api-access-4dmks\") pod \"nova-cell0-conductor-db-sync-sskzq\" (UID: \"a7e835d9-77a2-4066-b151-42b30db24e78\") " pod="openstack/nova-cell0-conductor-db-sync-sskzq" Oct 05 08:23:13 crc kubenswrapper[4846]: I1005 08:23:13.738520 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a7e835d9-77a2-4066-b151-42b30db24e78-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-sskzq\" (UID: \"a7e835d9-77a2-4066-b151-42b30db24e78\") " pod="openstack/nova-cell0-conductor-db-sync-sskzq" Oct 05 08:23:13 crc kubenswrapper[4846]: I1005 08:23:13.747719 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a7e835d9-77a2-4066-b151-42b30db24e78-scripts\") pod \"nova-cell0-conductor-db-sync-sskzq\" (UID: \"a7e835d9-77a2-4066-b151-42b30db24e78\") " pod="openstack/nova-cell0-conductor-db-sync-sskzq" Oct 05 08:23:13 crc kubenswrapper[4846]: I1005 08:23:13.753865 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a7e835d9-77a2-4066-b151-42b30db24e78-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-sskzq\" (UID: \"a7e835d9-77a2-4066-b151-42b30db24e78\") " pod="openstack/nova-cell0-conductor-db-sync-sskzq" Oct 05 08:23:13 crc kubenswrapper[4846]: I1005 08:23:13.753972 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a7e835d9-77a2-4066-b151-42b30db24e78-config-data\") pod \"nova-cell0-conductor-db-sync-sskzq\" (UID: \"a7e835d9-77a2-4066-b151-42b30db24e78\") " pod="openstack/nova-cell0-conductor-db-sync-sskzq" Oct 05 08:23:13 crc kubenswrapper[4846]: I1005 08:23:13.764646 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4dmks\" (UniqueName: \"kubernetes.io/projected/a7e835d9-77a2-4066-b151-42b30db24e78-kube-api-access-4dmks\") pod \"nova-cell0-conductor-db-sync-sskzq\" (UID: \"a7e835d9-77a2-4066-b151-42b30db24e78\") " pod="openstack/nova-cell0-conductor-db-sync-sskzq" Oct 05 08:23:13 crc kubenswrapper[4846]: I1005 08:23:13.833946 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-sskzq" Oct 05 08:23:14 crc kubenswrapper[4846]: W1005 08:23:14.135890 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda7e835d9_77a2_4066_b151_42b30db24e78.slice/crio-713dbc7379ff906211c93467f89e35cfc9565d9a38920d3c258ff20afa4d4912 WatchSource:0}: Error finding container 713dbc7379ff906211c93467f89e35cfc9565d9a38920d3c258ff20afa4d4912: Status 404 returned error can't find the container with id 713dbc7379ff906211c93467f89e35cfc9565d9a38920d3c258ff20afa4d4912 Oct 05 08:23:14 crc kubenswrapper[4846]: I1005 08:23:14.138082 4846 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 05 08:23:14 crc kubenswrapper[4846]: I1005 08:23:14.139449 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-sskzq"] Oct 05 08:23:14 crc kubenswrapper[4846]: I1005 08:23:14.315568 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-sskzq" event={"ID":"a7e835d9-77a2-4066-b151-42b30db24e78","Type":"ContainerStarted","Data":"713dbc7379ff906211c93467f89e35cfc9565d9a38920d3c258ff20afa4d4912"} Oct 05 08:23:21 crc kubenswrapper[4846]: I1005 08:23:21.026103 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-sync-g76dd"] Oct 05 08:23:21 crc kubenswrapper[4846]: I1005 08:23:21.033151 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-sync-g76dd"] Oct 05 08:23:22 crc kubenswrapper[4846]: I1005 08:23:22.509372 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8548783c-915e-4230-8b94-04cffa8ae2d9" path="/var/lib/kubelet/pods/8548783c-915e-4230-8b94-04cffa8ae2d9/volumes" Oct 05 08:23:23 crc kubenswrapper[4846]: I1005 08:23:23.426190 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-sskzq" event={"ID":"a7e835d9-77a2-4066-b151-42b30db24e78","Type":"ContainerStarted","Data":"0f757a5a9d9b747a9815a75c8d03fd44c128374cc6e796440a93564d5f796b8d"} Oct 05 08:23:23 crc kubenswrapper[4846]: I1005 08:23:23.450133 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-db-sync-sskzq" podStartSLOduration=1.4724957920000001 podStartE2EDuration="10.450114582s" podCreationTimestamp="2025-10-05 08:23:13 +0000 UTC" firstStartedPulling="2025-10-05 08:23:14.137860206 +0000 UTC m=+5716.378712981" lastFinishedPulling="2025-10-05 08:23:23.115478986 +0000 UTC m=+5725.356331771" observedRunningTime="2025-10-05 08:23:23.447042111 +0000 UTC m=+5725.687894906" watchObservedRunningTime="2025-10-05 08:23:23.450114582 +0000 UTC m=+5725.690967347" Oct 05 08:23:25 crc kubenswrapper[4846]: I1005 08:23:25.498422 4846 scope.go:117] "RemoveContainer" containerID="fdafa27da24296f9087c33b6b12a6598d2a8d61fc985b06ad680401828c9feb6" Oct 05 08:23:25 crc kubenswrapper[4846]: E1005 08:23:25.499151 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 08:23:28 crc kubenswrapper[4846]: I1005 08:23:28.471712 4846 generic.go:334] "Generic (PLEG): container finished" podID="a7e835d9-77a2-4066-b151-42b30db24e78" containerID="0f757a5a9d9b747a9815a75c8d03fd44c128374cc6e796440a93564d5f796b8d" exitCode=0 Oct 05 08:23:28 crc kubenswrapper[4846]: I1005 08:23:28.471762 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-sskzq" event={"ID":"a7e835d9-77a2-4066-b151-42b30db24e78","Type":"ContainerDied","Data":"0f757a5a9d9b747a9815a75c8d03fd44c128374cc6e796440a93564d5f796b8d"} Oct 05 08:23:29 crc kubenswrapper[4846]: I1005 08:23:29.877398 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-sskzq" Oct 05 08:23:30 crc kubenswrapper[4846]: I1005 08:23:30.005242 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a7e835d9-77a2-4066-b151-42b30db24e78-scripts\") pod \"a7e835d9-77a2-4066-b151-42b30db24e78\" (UID: \"a7e835d9-77a2-4066-b151-42b30db24e78\") " Oct 05 08:23:30 crc kubenswrapper[4846]: I1005 08:23:30.005402 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a7e835d9-77a2-4066-b151-42b30db24e78-config-data\") pod \"a7e835d9-77a2-4066-b151-42b30db24e78\" (UID: \"a7e835d9-77a2-4066-b151-42b30db24e78\") " Oct 05 08:23:30 crc kubenswrapper[4846]: I1005 08:23:30.005573 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a7e835d9-77a2-4066-b151-42b30db24e78-combined-ca-bundle\") pod \"a7e835d9-77a2-4066-b151-42b30db24e78\" (UID: \"a7e835d9-77a2-4066-b151-42b30db24e78\") " Oct 05 08:23:30 crc kubenswrapper[4846]: I1005 08:23:30.005671 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4dmks\" (UniqueName: \"kubernetes.io/projected/a7e835d9-77a2-4066-b151-42b30db24e78-kube-api-access-4dmks\") pod \"a7e835d9-77a2-4066-b151-42b30db24e78\" (UID: \"a7e835d9-77a2-4066-b151-42b30db24e78\") " Oct 05 08:23:30 crc kubenswrapper[4846]: I1005 08:23:30.015071 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a7e835d9-77a2-4066-b151-42b30db24e78-scripts" (OuterVolumeSpecName: "scripts") pod "a7e835d9-77a2-4066-b151-42b30db24e78" (UID: "a7e835d9-77a2-4066-b151-42b30db24e78"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:23:30 crc kubenswrapper[4846]: I1005 08:23:30.020448 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a7e835d9-77a2-4066-b151-42b30db24e78-kube-api-access-4dmks" (OuterVolumeSpecName: "kube-api-access-4dmks") pod "a7e835d9-77a2-4066-b151-42b30db24e78" (UID: "a7e835d9-77a2-4066-b151-42b30db24e78"). InnerVolumeSpecName "kube-api-access-4dmks". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:23:30 crc kubenswrapper[4846]: I1005 08:23:30.037667 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a7e835d9-77a2-4066-b151-42b30db24e78-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a7e835d9-77a2-4066-b151-42b30db24e78" (UID: "a7e835d9-77a2-4066-b151-42b30db24e78"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:23:30 crc kubenswrapper[4846]: I1005 08:23:30.047698 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a7e835d9-77a2-4066-b151-42b30db24e78-config-data" (OuterVolumeSpecName: "config-data") pod "a7e835d9-77a2-4066-b151-42b30db24e78" (UID: "a7e835d9-77a2-4066-b151-42b30db24e78"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:23:30 crc kubenswrapper[4846]: I1005 08:23:30.107702 4846 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a7e835d9-77a2-4066-b151-42b30db24e78-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 08:23:30 crc kubenswrapper[4846]: I1005 08:23:30.107742 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4dmks\" (UniqueName: \"kubernetes.io/projected/a7e835d9-77a2-4066-b151-42b30db24e78-kube-api-access-4dmks\") on node \"crc\" DevicePath \"\"" Oct 05 08:23:30 crc kubenswrapper[4846]: I1005 08:23:30.107753 4846 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a7e835d9-77a2-4066-b151-42b30db24e78-scripts\") on node \"crc\" DevicePath \"\"" Oct 05 08:23:30 crc kubenswrapper[4846]: I1005 08:23:30.107762 4846 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a7e835d9-77a2-4066-b151-42b30db24e78-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 08:23:30 crc kubenswrapper[4846]: I1005 08:23:30.493809 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-sskzq" event={"ID":"a7e835d9-77a2-4066-b151-42b30db24e78","Type":"ContainerDied","Data":"713dbc7379ff906211c93467f89e35cfc9565d9a38920d3c258ff20afa4d4912"} Oct 05 08:23:30 crc kubenswrapper[4846]: I1005 08:23:30.493851 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="713dbc7379ff906211c93467f89e35cfc9565d9a38920d3c258ff20afa4d4912" Oct 05 08:23:30 crc kubenswrapper[4846]: I1005 08:23:30.493914 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-sskzq" Oct 05 08:23:30 crc kubenswrapper[4846]: I1005 08:23:30.581708 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 05 08:23:30 crc kubenswrapper[4846]: E1005 08:23:30.582300 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a7e835d9-77a2-4066-b151-42b30db24e78" containerName="nova-cell0-conductor-db-sync" Oct 05 08:23:30 crc kubenswrapper[4846]: I1005 08:23:30.582381 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="a7e835d9-77a2-4066-b151-42b30db24e78" containerName="nova-cell0-conductor-db-sync" Oct 05 08:23:30 crc kubenswrapper[4846]: I1005 08:23:30.591765 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="a7e835d9-77a2-4066-b151-42b30db24e78" containerName="nova-cell0-conductor-db-sync" Oct 05 08:23:30 crc kubenswrapper[4846]: I1005 08:23:30.604021 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Oct 05 08:23:30 crc kubenswrapper[4846]: I1005 08:23:30.608464 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-hhf6x" Oct 05 08:23:30 crc kubenswrapper[4846]: I1005 08:23:30.611008 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Oct 05 08:23:30 crc kubenswrapper[4846]: I1005 08:23:30.625115 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-746n7\" (UniqueName: \"kubernetes.io/projected/672ba7aa-607e-4872-9083-cb2847606b7c-kube-api-access-746n7\") pod \"nova-cell0-conductor-0\" (UID: \"672ba7aa-607e-4872-9083-cb2847606b7c\") " pod="openstack/nova-cell0-conductor-0" Oct 05 08:23:30 crc kubenswrapper[4846]: I1005 08:23:30.625285 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/672ba7aa-607e-4872-9083-cb2847606b7c-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"672ba7aa-607e-4872-9083-cb2847606b7c\") " pod="openstack/nova-cell0-conductor-0" Oct 05 08:23:30 crc kubenswrapper[4846]: I1005 08:23:30.625353 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/672ba7aa-607e-4872-9083-cb2847606b7c-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"672ba7aa-607e-4872-9083-cb2847606b7c\") " pod="openstack/nova-cell0-conductor-0" Oct 05 08:23:30 crc kubenswrapper[4846]: I1005 08:23:30.637341 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 05 08:23:30 crc kubenswrapper[4846]: I1005 08:23:30.726146 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/672ba7aa-607e-4872-9083-cb2847606b7c-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"672ba7aa-607e-4872-9083-cb2847606b7c\") " pod="openstack/nova-cell0-conductor-0" Oct 05 08:23:30 crc kubenswrapper[4846]: I1005 08:23:30.726515 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-746n7\" (UniqueName: \"kubernetes.io/projected/672ba7aa-607e-4872-9083-cb2847606b7c-kube-api-access-746n7\") pod \"nova-cell0-conductor-0\" (UID: \"672ba7aa-607e-4872-9083-cb2847606b7c\") " pod="openstack/nova-cell0-conductor-0" Oct 05 08:23:30 crc kubenswrapper[4846]: I1005 08:23:30.726906 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/672ba7aa-607e-4872-9083-cb2847606b7c-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"672ba7aa-607e-4872-9083-cb2847606b7c\") " pod="openstack/nova-cell0-conductor-0" Oct 05 08:23:30 crc kubenswrapper[4846]: I1005 08:23:30.734260 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/672ba7aa-607e-4872-9083-cb2847606b7c-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"672ba7aa-607e-4872-9083-cb2847606b7c\") " pod="openstack/nova-cell0-conductor-0" Oct 05 08:23:30 crc kubenswrapper[4846]: I1005 08:23:30.737900 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/672ba7aa-607e-4872-9083-cb2847606b7c-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"672ba7aa-607e-4872-9083-cb2847606b7c\") " pod="openstack/nova-cell0-conductor-0" Oct 05 08:23:30 crc kubenswrapper[4846]: I1005 08:23:30.744998 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-746n7\" (UniqueName: \"kubernetes.io/projected/672ba7aa-607e-4872-9083-cb2847606b7c-kube-api-access-746n7\") pod \"nova-cell0-conductor-0\" (UID: \"672ba7aa-607e-4872-9083-cb2847606b7c\") " pod="openstack/nova-cell0-conductor-0" Oct 05 08:23:30 crc kubenswrapper[4846]: I1005 08:23:30.926995 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Oct 05 08:23:31 crc kubenswrapper[4846]: I1005 08:23:31.386554 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 05 08:23:31 crc kubenswrapper[4846]: W1005 08:23:31.395199 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod672ba7aa_607e_4872_9083_cb2847606b7c.slice/crio-16e9d57bae311cbe11b32d992f81e5b45f1c88394da60fc86883e372c4962abc WatchSource:0}: Error finding container 16e9d57bae311cbe11b32d992f81e5b45f1c88394da60fc86883e372c4962abc: Status 404 returned error can't find the container with id 16e9d57bae311cbe11b32d992f81e5b45f1c88394da60fc86883e372c4962abc Oct 05 08:23:31 crc kubenswrapper[4846]: I1005 08:23:31.502641 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"672ba7aa-607e-4872-9083-cb2847606b7c","Type":"ContainerStarted","Data":"16e9d57bae311cbe11b32d992f81e5b45f1c88394da60fc86883e372c4962abc"} Oct 05 08:23:32 crc kubenswrapper[4846]: I1005 08:23:32.510613 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"672ba7aa-607e-4872-9083-cb2847606b7c","Type":"ContainerStarted","Data":"7411f369dc75ad75e553fd68df2151115121338637f010473f2d659c8276b45e"} Oct 05 08:23:32 crc kubenswrapper[4846]: I1005 08:23:32.512164 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Oct 05 08:23:32 crc kubenswrapper[4846]: I1005 08:23:32.660536 4846 scope.go:117] "RemoveContainer" containerID="c721c574769f3dc4ebcc372911f37c071866043013c026ac0fc7582fa24ddb00" Oct 05 08:23:32 crc kubenswrapper[4846]: I1005 08:23:32.700149 4846 scope.go:117] "RemoveContainer" containerID="b27004a16006eacec2ac7eb7962288a705d7cb1acb5d972dac5454fc0759e990" Oct 05 08:23:32 crc kubenswrapper[4846]: I1005 08:23:32.743523 4846 scope.go:117] "RemoveContainer" containerID="5d0729fccbe964bd9f23bb56ceb1eac397bcd850556d8691a91d1c009adb5c87" Oct 05 08:23:35 crc kubenswrapper[4846]: I1005 08:23:35.058415 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=5.05838888 podStartE2EDuration="5.05838888s" podCreationTimestamp="2025-10-05 08:23:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 08:23:32.529812024 +0000 UTC m=+5734.770664799" watchObservedRunningTime="2025-10-05 08:23:35.05838888 +0000 UTC m=+5737.299241675" Oct 05 08:23:35 crc kubenswrapper[4846]: I1005 08:23:35.060891 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-fw275"] Oct 05 08:23:35 crc kubenswrapper[4846]: I1005 08:23:35.070755 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-fw275"] Oct 05 08:23:36 crc kubenswrapper[4846]: I1005 08:23:36.509584 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f49efe05-7f4e-4fb1-b523-ccdc75791838" path="/var/lib/kubelet/pods/f49efe05-7f4e-4fb1-b523-ccdc75791838/volumes" Oct 05 08:23:37 crc kubenswrapper[4846]: I1005 08:23:37.497496 4846 scope.go:117] "RemoveContainer" containerID="fdafa27da24296f9087c33b6b12a6598d2a8d61fc985b06ad680401828c9feb6" Oct 05 08:23:37 crc kubenswrapper[4846]: E1005 08:23:37.497942 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 08:23:40 crc kubenswrapper[4846]: I1005 08:23:40.969357 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Oct 05 08:23:41 crc kubenswrapper[4846]: I1005 08:23:41.605435 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-cell-mapping-4kt89"] Oct 05 08:23:41 crc kubenswrapper[4846]: I1005 08:23:41.606926 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-4kt89" Oct 05 08:23:41 crc kubenswrapper[4846]: I1005 08:23:41.608764 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-scripts" Oct 05 08:23:41 crc kubenswrapper[4846]: I1005 08:23:41.608923 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-config-data" Oct 05 08:23:41 crc kubenswrapper[4846]: I1005 08:23:41.624475 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-4kt89"] Oct 05 08:23:41 crc kubenswrapper[4846]: I1005 08:23:41.659648 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/285fade3-a17c-4cfc-937c-096ebdb15848-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-4kt89\" (UID: \"285fade3-a17c-4cfc-937c-096ebdb15848\") " pod="openstack/nova-cell0-cell-mapping-4kt89" Oct 05 08:23:41 crc kubenswrapper[4846]: I1005 08:23:41.659777 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/285fade3-a17c-4cfc-937c-096ebdb15848-scripts\") pod \"nova-cell0-cell-mapping-4kt89\" (UID: \"285fade3-a17c-4cfc-937c-096ebdb15848\") " pod="openstack/nova-cell0-cell-mapping-4kt89" Oct 05 08:23:41 crc kubenswrapper[4846]: I1005 08:23:41.659822 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/285fade3-a17c-4cfc-937c-096ebdb15848-config-data\") pod \"nova-cell0-cell-mapping-4kt89\" (UID: \"285fade3-a17c-4cfc-937c-096ebdb15848\") " pod="openstack/nova-cell0-cell-mapping-4kt89" Oct 05 08:23:41 crc kubenswrapper[4846]: I1005 08:23:41.659863 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mgqpx\" (UniqueName: \"kubernetes.io/projected/285fade3-a17c-4cfc-937c-096ebdb15848-kube-api-access-mgqpx\") pod \"nova-cell0-cell-mapping-4kt89\" (UID: \"285fade3-a17c-4cfc-937c-096ebdb15848\") " pod="openstack/nova-cell0-cell-mapping-4kt89" Oct 05 08:23:41 crc kubenswrapper[4846]: I1005 08:23:41.741860 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Oct 05 08:23:41 crc kubenswrapper[4846]: I1005 08:23:41.743043 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 05 08:23:41 crc kubenswrapper[4846]: I1005 08:23:41.747508 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Oct 05 08:23:41 crc kubenswrapper[4846]: I1005 08:23:41.760737 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 05 08:23:41 crc kubenswrapper[4846]: I1005 08:23:41.761994 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/285fade3-a17c-4cfc-937c-096ebdb15848-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-4kt89\" (UID: \"285fade3-a17c-4cfc-937c-096ebdb15848\") " pod="openstack/nova-cell0-cell-mapping-4kt89" Oct 05 08:23:41 crc kubenswrapper[4846]: I1005 08:23:41.762077 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/285fade3-a17c-4cfc-937c-096ebdb15848-scripts\") pod \"nova-cell0-cell-mapping-4kt89\" (UID: \"285fade3-a17c-4cfc-937c-096ebdb15848\") " pod="openstack/nova-cell0-cell-mapping-4kt89" Oct 05 08:23:41 crc kubenswrapper[4846]: I1005 08:23:41.762114 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/285fade3-a17c-4cfc-937c-096ebdb15848-config-data\") pod \"nova-cell0-cell-mapping-4kt89\" (UID: \"285fade3-a17c-4cfc-937c-096ebdb15848\") " pod="openstack/nova-cell0-cell-mapping-4kt89" Oct 05 08:23:41 crc kubenswrapper[4846]: I1005 08:23:41.762149 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mgqpx\" (UniqueName: \"kubernetes.io/projected/285fade3-a17c-4cfc-937c-096ebdb15848-kube-api-access-mgqpx\") pod \"nova-cell0-cell-mapping-4kt89\" (UID: \"285fade3-a17c-4cfc-937c-096ebdb15848\") " pod="openstack/nova-cell0-cell-mapping-4kt89" Oct 05 08:23:41 crc kubenswrapper[4846]: I1005 08:23:41.768834 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/285fade3-a17c-4cfc-937c-096ebdb15848-scripts\") pod \"nova-cell0-cell-mapping-4kt89\" (UID: \"285fade3-a17c-4cfc-937c-096ebdb15848\") " pod="openstack/nova-cell0-cell-mapping-4kt89" Oct 05 08:23:41 crc kubenswrapper[4846]: I1005 08:23:41.769253 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/285fade3-a17c-4cfc-937c-096ebdb15848-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-4kt89\" (UID: \"285fade3-a17c-4cfc-937c-096ebdb15848\") " pod="openstack/nova-cell0-cell-mapping-4kt89" Oct 05 08:23:41 crc kubenswrapper[4846]: I1005 08:23:41.774001 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/285fade3-a17c-4cfc-937c-096ebdb15848-config-data\") pod \"nova-cell0-cell-mapping-4kt89\" (UID: \"285fade3-a17c-4cfc-937c-096ebdb15848\") " pod="openstack/nova-cell0-cell-mapping-4kt89" Oct 05 08:23:41 crc kubenswrapper[4846]: I1005 08:23:41.804936 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mgqpx\" (UniqueName: \"kubernetes.io/projected/285fade3-a17c-4cfc-937c-096ebdb15848-kube-api-access-mgqpx\") pod \"nova-cell0-cell-mapping-4kt89\" (UID: \"285fade3-a17c-4cfc-937c-096ebdb15848\") " pod="openstack/nova-cell0-cell-mapping-4kt89" Oct 05 08:23:41 crc kubenswrapper[4846]: I1005 08:23:41.832241 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 05 08:23:41 crc kubenswrapper[4846]: I1005 08:23:41.833556 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 05 08:23:41 crc kubenswrapper[4846]: I1005 08:23:41.840287 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Oct 05 08:23:41 crc kubenswrapper[4846]: I1005 08:23:41.864391 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5b32c235-5879-482b-a396-6d0d7e98816d-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"5b32c235-5879-482b-a396-6d0d7e98816d\") " pod="openstack/nova-scheduler-0" Oct 05 08:23:41 crc kubenswrapper[4846]: I1005 08:23:41.864500 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9cxbt\" (UniqueName: \"kubernetes.io/projected/5b32c235-5879-482b-a396-6d0d7e98816d-kube-api-access-9cxbt\") pod \"nova-scheduler-0\" (UID: \"5b32c235-5879-482b-a396-6d0d7e98816d\") " pod="openstack/nova-scheduler-0" Oct 05 08:23:41 crc kubenswrapper[4846]: I1005 08:23:41.864538 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5b32c235-5879-482b-a396-6d0d7e98816d-config-data\") pod \"nova-scheduler-0\" (UID: \"5b32c235-5879-482b-a396-6d0d7e98816d\") " pod="openstack/nova-scheduler-0" Oct 05 08:23:41 crc kubenswrapper[4846]: I1005 08:23:41.867724 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 05 08:23:41 crc kubenswrapper[4846]: I1005 08:23:41.917836 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 05 08:23:41 crc kubenswrapper[4846]: I1005 08:23:41.920813 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 05 08:23:41 crc kubenswrapper[4846]: I1005 08:23:41.924703 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 05 08:23:41 crc kubenswrapper[4846]: I1005 08:23:41.931132 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 05 08:23:41 crc kubenswrapper[4846]: I1005 08:23:41.945490 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-4kt89" Oct 05 08:23:41 crc kubenswrapper[4846]: I1005 08:23:41.975395 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8jhmk\" (UniqueName: \"kubernetes.io/projected/fec088ae-74f0-42bb-8225-428c7562f4c0-kube-api-access-8jhmk\") pod \"nova-cell1-novncproxy-0\" (UID: \"fec088ae-74f0-42bb-8225-428c7562f4c0\") " pod="openstack/nova-cell1-novncproxy-0" Oct 05 08:23:41 crc kubenswrapper[4846]: I1005 08:23:41.975460 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5b32c235-5879-482b-a396-6d0d7e98816d-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"5b32c235-5879-482b-a396-6d0d7e98816d\") " pod="openstack/nova-scheduler-0" Oct 05 08:23:41 crc kubenswrapper[4846]: I1005 08:23:41.975505 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fec088ae-74f0-42bb-8225-428c7562f4c0-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"fec088ae-74f0-42bb-8225-428c7562f4c0\") " pod="openstack/nova-cell1-novncproxy-0" Oct 05 08:23:41 crc kubenswrapper[4846]: I1005 08:23:41.975534 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fec088ae-74f0-42bb-8225-428c7562f4c0-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"fec088ae-74f0-42bb-8225-428c7562f4c0\") " pod="openstack/nova-cell1-novncproxy-0" Oct 05 08:23:41 crc kubenswrapper[4846]: I1005 08:23:41.975557 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9cxbt\" (UniqueName: \"kubernetes.io/projected/5b32c235-5879-482b-a396-6d0d7e98816d-kube-api-access-9cxbt\") pod \"nova-scheduler-0\" (UID: \"5b32c235-5879-482b-a396-6d0d7e98816d\") " pod="openstack/nova-scheduler-0" Oct 05 08:23:41 crc kubenswrapper[4846]: I1005 08:23:41.975583 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5b32c235-5879-482b-a396-6d0d7e98816d-config-data\") pod \"nova-scheduler-0\" (UID: \"5b32c235-5879-482b-a396-6d0d7e98816d\") " pod="openstack/nova-scheduler-0" Oct 05 08:23:41 crc kubenswrapper[4846]: I1005 08:23:41.987901 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5b32c235-5879-482b-a396-6d0d7e98816d-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"5b32c235-5879-482b-a396-6d0d7e98816d\") " pod="openstack/nova-scheduler-0" Oct 05 08:23:41 crc kubenswrapper[4846]: I1005 08:23:41.988359 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5b32c235-5879-482b-a396-6d0d7e98816d-config-data\") pod \"nova-scheduler-0\" (UID: \"5b32c235-5879-482b-a396-6d0d7e98816d\") " pod="openstack/nova-scheduler-0" Oct 05 08:23:42 crc kubenswrapper[4846]: I1005 08:23:42.008242 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 05 08:23:42 crc kubenswrapper[4846]: I1005 08:23:42.009724 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 05 08:23:42 crc kubenswrapper[4846]: I1005 08:23:42.019655 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 05 08:23:42 crc kubenswrapper[4846]: I1005 08:23:42.032618 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 05 08:23:42 crc kubenswrapper[4846]: I1005 08:23:42.052788 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9cxbt\" (UniqueName: \"kubernetes.io/projected/5b32c235-5879-482b-a396-6d0d7e98816d-kube-api-access-9cxbt\") pod \"nova-scheduler-0\" (UID: \"5b32c235-5879-482b-a396-6d0d7e98816d\") " pod="openstack/nova-scheduler-0" Oct 05 08:23:42 crc kubenswrapper[4846]: I1005 08:23:42.077357 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7dd897db-45ea-491c-ace4-9e14fd8104dc-logs\") pod \"nova-api-0\" (UID: \"7dd897db-45ea-491c-ace4-9e14fd8104dc\") " pod="openstack/nova-api-0" Oct 05 08:23:42 crc kubenswrapper[4846]: I1005 08:23:42.077460 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fec088ae-74f0-42bb-8225-428c7562f4c0-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"fec088ae-74f0-42bb-8225-428c7562f4c0\") " pod="openstack/nova-cell1-novncproxy-0" Oct 05 08:23:42 crc kubenswrapper[4846]: I1005 08:23:42.078106 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7dd897db-45ea-491c-ace4-9e14fd8104dc-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"7dd897db-45ea-491c-ace4-9e14fd8104dc\") " pod="openstack/nova-api-0" Oct 05 08:23:42 crc kubenswrapper[4846]: I1005 08:23:42.078143 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fec088ae-74f0-42bb-8225-428c7562f4c0-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"fec088ae-74f0-42bb-8225-428c7562f4c0\") " pod="openstack/nova-cell1-novncproxy-0" Oct 05 08:23:42 crc kubenswrapper[4846]: I1005 08:23:42.078223 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gkhvb\" (UniqueName: \"kubernetes.io/projected/7dd897db-45ea-491c-ace4-9e14fd8104dc-kube-api-access-gkhvb\") pod \"nova-api-0\" (UID: \"7dd897db-45ea-491c-ace4-9e14fd8104dc\") " pod="openstack/nova-api-0" Oct 05 08:23:42 crc kubenswrapper[4846]: I1005 08:23:42.078282 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7dd897db-45ea-491c-ace4-9e14fd8104dc-config-data\") pod \"nova-api-0\" (UID: \"7dd897db-45ea-491c-ace4-9e14fd8104dc\") " pod="openstack/nova-api-0" Oct 05 08:23:42 crc kubenswrapper[4846]: I1005 08:23:42.078340 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8jhmk\" (UniqueName: \"kubernetes.io/projected/fec088ae-74f0-42bb-8225-428c7562f4c0-kube-api-access-8jhmk\") pod \"nova-cell1-novncproxy-0\" (UID: \"fec088ae-74f0-42bb-8225-428c7562f4c0\") " pod="openstack/nova-cell1-novncproxy-0" Oct 05 08:23:42 crc kubenswrapper[4846]: I1005 08:23:42.085904 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fec088ae-74f0-42bb-8225-428c7562f4c0-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"fec088ae-74f0-42bb-8225-428c7562f4c0\") " pod="openstack/nova-cell1-novncproxy-0" Oct 05 08:23:42 crc kubenswrapper[4846]: I1005 08:23:42.085911 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fec088ae-74f0-42bb-8225-428c7562f4c0-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"fec088ae-74f0-42bb-8225-428c7562f4c0\") " pod="openstack/nova-cell1-novncproxy-0" Oct 05 08:23:42 crc kubenswrapper[4846]: I1005 08:23:42.099797 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8jhmk\" (UniqueName: \"kubernetes.io/projected/fec088ae-74f0-42bb-8225-428c7562f4c0-kube-api-access-8jhmk\") pod \"nova-cell1-novncproxy-0\" (UID: \"fec088ae-74f0-42bb-8225-428c7562f4c0\") " pod="openstack/nova-cell1-novncproxy-0" Oct 05 08:23:42 crc kubenswrapper[4846]: I1005 08:23:42.145980 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-774686bfb5-gk9ch"] Oct 05 08:23:42 crc kubenswrapper[4846]: I1005 08:23:42.147885 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-774686bfb5-gk9ch" Oct 05 08:23:42 crc kubenswrapper[4846]: I1005 08:23:42.169016 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 05 08:23:42 crc kubenswrapper[4846]: I1005 08:23:42.174269 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-774686bfb5-gk9ch"] Oct 05 08:23:42 crc kubenswrapper[4846]: I1005 08:23:42.189796 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7dd897db-45ea-491c-ace4-9e14fd8104dc-config-data\") pod \"nova-api-0\" (UID: \"7dd897db-45ea-491c-ace4-9e14fd8104dc\") " pod="openstack/nova-api-0" Oct 05 08:23:42 crc kubenswrapper[4846]: I1005 08:23:42.189950 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ae74e41e-d09c-4666-85bf-998ff2834598-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"ae74e41e-d09c-4666-85bf-998ff2834598\") " pod="openstack/nova-metadata-0" Oct 05 08:23:42 crc kubenswrapper[4846]: I1005 08:23:42.190130 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7dd897db-45ea-491c-ace4-9e14fd8104dc-logs\") pod \"nova-api-0\" (UID: \"7dd897db-45ea-491c-ace4-9e14fd8104dc\") " pod="openstack/nova-api-0" Oct 05 08:23:42 crc kubenswrapper[4846]: I1005 08:23:42.190318 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dktkp\" (UniqueName: \"kubernetes.io/projected/ae74e41e-d09c-4666-85bf-998ff2834598-kube-api-access-dktkp\") pod \"nova-metadata-0\" (UID: \"ae74e41e-d09c-4666-85bf-998ff2834598\") " pod="openstack/nova-metadata-0" Oct 05 08:23:42 crc kubenswrapper[4846]: I1005 08:23:42.190352 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ae74e41e-d09c-4666-85bf-998ff2834598-logs\") pod \"nova-metadata-0\" (UID: \"ae74e41e-d09c-4666-85bf-998ff2834598\") " pod="openstack/nova-metadata-0" Oct 05 08:23:42 crc kubenswrapper[4846]: I1005 08:23:42.190401 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7dd897db-45ea-491c-ace4-9e14fd8104dc-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"7dd897db-45ea-491c-ace4-9e14fd8104dc\") " pod="openstack/nova-api-0" Oct 05 08:23:42 crc kubenswrapper[4846]: I1005 08:23:42.190482 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ae74e41e-d09c-4666-85bf-998ff2834598-config-data\") pod \"nova-metadata-0\" (UID: \"ae74e41e-d09c-4666-85bf-998ff2834598\") " pod="openstack/nova-metadata-0" Oct 05 08:23:42 crc kubenswrapper[4846]: I1005 08:23:42.190536 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gkhvb\" (UniqueName: \"kubernetes.io/projected/7dd897db-45ea-491c-ace4-9e14fd8104dc-kube-api-access-gkhvb\") pod \"nova-api-0\" (UID: \"7dd897db-45ea-491c-ace4-9e14fd8104dc\") " pod="openstack/nova-api-0" Oct 05 08:23:42 crc kubenswrapper[4846]: I1005 08:23:42.190870 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7dd897db-45ea-491c-ace4-9e14fd8104dc-logs\") pod \"nova-api-0\" (UID: \"7dd897db-45ea-491c-ace4-9e14fd8104dc\") " pod="openstack/nova-api-0" Oct 05 08:23:42 crc kubenswrapper[4846]: I1005 08:23:42.210278 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7dd897db-45ea-491c-ace4-9e14fd8104dc-config-data\") pod \"nova-api-0\" (UID: \"7dd897db-45ea-491c-ace4-9e14fd8104dc\") " pod="openstack/nova-api-0" Oct 05 08:23:42 crc kubenswrapper[4846]: I1005 08:23:42.211893 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7dd897db-45ea-491c-ace4-9e14fd8104dc-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"7dd897db-45ea-491c-ace4-9e14fd8104dc\") " pod="openstack/nova-api-0" Oct 05 08:23:42 crc kubenswrapper[4846]: I1005 08:23:42.222944 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 05 08:23:42 crc kubenswrapper[4846]: I1005 08:23:42.231589 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gkhvb\" (UniqueName: \"kubernetes.io/projected/7dd897db-45ea-491c-ace4-9e14fd8104dc-kube-api-access-gkhvb\") pod \"nova-api-0\" (UID: \"7dd897db-45ea-491c-ace4-9e14fd8104dc\") " pod="openstack/nova-api-0" Oct 05 08:23:42 crc kubenswrapper[4846]: I1005 08:23:42.277812 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 05 08:23:42 crc kubenswrapper[4846]: I1005 08:23:42.304209 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ae74e41e-d09c-4666-85bf-998ff2834598-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"ae74e41e-d09c-4666-85bf-998ff2834598\") " pod="openstack/nova-metadata-0" Oct 05 08:23:42 crc kubenswrapper[4846]: I1005 08:23:42.304294 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fe918116-df47-4b09-977a-0f16a9cd5e75-dns-svc\") pod \"dnsmasq-dns-774686bfb5-gk9ch\" (UID: \"fe918116-df47-4b09-977a-0f16a9cd5e75\") " pod="openstack/dnsmasq-dns-774686bfb5-gk9ch" Oct 05 08:23:42 crc kubenswrapper[4846]: I1005 08:23:42.304355 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fe918116-df47-4b09-977a-0f16a9cd5e75-config\") pod \"dnsmasq-dns-774686bfb5-gk9ch\" (UID: \"fe918116-df47-4b09-977a-0f16a9cd5e75\") " pod="openstack/dnsmasq-dns-774686bfb5-gk9ch" Oct 05 08:23:42 crc kubenswrapper[4846]: I1005 08:23:42.304398 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/fe918116-df47-4b09-977a-0f16a9cd5e75-ovsdbserver-nb\") pod \"dnsmasq-dns-774686bfb5-gk9ch\" (UID: \"fe918116-df47-4b09-977a-0f16a9cd5e75\") " pod="openstack/dnsmasq-dns-774686bfb5-gk9ch" Oct 05 08:23:42 crc kubenswrapper[4846]: I1005 08:23:42.304680 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j2q2j\" (UniqueName: \"kubernetes.io/projected/fe918116-df47-4b09-977a-0f16a9cd5e75-kube-api-access-j2q2j\") pod \"dnsmasq-dns-774686bfb5-gk9ch\" (UID: \"fe918116-df47-4b09-977a-0f16a9cd5e75\") " pod="openstack/dnsmasq-dns-774686bfb5-gk9ch" Oct 05 08:23:42 crc kubenswrapper[4846]: I1005 08:23:42.304869 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dktkp\" (UniqueName: \"kubernetes.io/projected/ae74e41e-d09c-4666-85bf-998ff2834598-kube-api-access-dktkp\") pod \"nova-metadata-0\" (UID: \"ae74e41e-d09c-4666-85bf-998ff2834598\") " pod="openstack/nova-metadata-0" Oct 05 08:23:42 crc kubenswrapper[4846]: I1005 08:23:42.304917 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ae74e41e-d09c-4666-85bf-998ff2834598-logs\") pod \"nova-metadata-0\" (UID: \"ae74e41e-d09c-4666-85bf-998ff2834598\") " pod="openstack/nova-metadata-0" Oct 05 08:23:42 crc kubenswrapper[4846]: I1005 08:23:42.305099 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/fe918116-df47-4b09-977a-0f16a9cd5e75-ovsdbserver-sb\") pod \"dnsmasq-dns-774686bfb5-gk9ch\" (UID: \"fe918116-df47-4b09-977a-0f16a9cd5e75\") " pod="openstack/dnsmasq-dns-774686bfb5-gk9ch" Oct 05 08:23:42 crc kubenswrapper[4846]: I1005 08:23:42.305171 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ae74e41e-d09c-4666-85bf-998ff2834598-config-data\") pod \"nova-metadata-0\" (UID: \"ae74e41e-d09c-4666-85bf-998ff2834598\") " pod="openstack/nova-metadata-0" Oct 05 08:23:42 crc kubenswrapper[4846]: I1005 08:23:42.306360 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ae74e41e-d09c-4666-85bf-998ff2834598-logs\") pod \"nova-metadata-0\" (UID: \"ae74e41e-d09c-4666-85bf-998ff2834598\") " pod="openstack/nova-metadata-0" Oct 05 08:23:42 crc kubenswrapper[4846]: I1005 08:23:42.311515 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ae74e41e-d09c-4666-85bf-998ff2834598-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"ae74e41e-d09c-4666-85bf-998ff2834598\") " pod="openstack/nova-metadata-0" Oct 05 08:23:42 crc kubenswrapper[4846]: I1005 08:23:42.313067 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ae74e41e-d09c-4666-85bf-998ff2834598-config-data\") pod \"nova-metadata-0\" (UID: \"ae74e41e-d09c-4666-85bf-998ff2834598\") " pod="openstack/nova-metadata-0" Oct 05 08:23:42 crc kubenswrapper[4846]: I1005 08:23:42.333626 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dktkp\" (UniqueName: \"kubernetes.io/projected/ae74e41e-d09c-4666-85bf-998ff2834598-kube-api-access-dktkp\") pod \"nova-metadata-0\" (UID: \"ae74e41e-d09c-4666-85bf-998ff2834598\") " pod="openstack/nova-metadata-0" Oct 05 08:23:42 crc kubenswrapper[4846]: I1005 08:23:42.409460 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/fe918116-df47-4b09-977a-0f16a9cd5e75-ovsdbserver-sb\") pod \"dnsmasq-dns-774686bfb5-gk9ch\" (UID: \"fe918116-df47-4b09-977a-0f16a9cd5e75\") " pod="openstack/dnsmasq-dns-774686bfb5-gk9ch" Oct 05 08:23:42 crc kubenswrapper[4846]: I1005 08:23:42.409560 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fe918116-df47-4b09-977a-0f16a9cd5e75-dns-svc\") pod \"dnsmasq-dns-774686bfb5-gk9ch\" (UID: \"fe918116-df47-4b09-977a-0f16a9cd5e75\") " pod="openstack/dnsmasq-dns-774686bfb5-gk9ch" Oct 05 08:23:42 crc kubenswrapper[4846]: I1005 08:23:42.409592 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fe918116-df47-4b09-977a-0f16a9cd5e75-config\") pod \"dnsmasq-dns-774686bfb5-gk9ch\" (UID: \"fe918116-df47-4b09-977a-0f16a9cd5e75\") " pod="openstack/dnsmasq-dns-774686bfb5-gk9ch" Oct 05 08:23:42 crc kubenswrapper[4846]: I1005 08:23:42.409617 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/fe918116-df47-4b09-977a-0f16a9cd5e75-ovsdbserver-nb\") pod \"dnsmasq-dns-774686bfb5-gk9ch\" (UID: \"fe918116-df47-4b09-977a-0f16a9cd5e75\") " pod="openstack/dnsmasq-dns-774686bfb5-gk9ch" Oct 05 08:23:42 crc kubenswrapper[4846]: I1005 08:23:42.409649 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j2q2j\" (UniqueName: \"kubernetes.io/projected/fe918116-df47-4b09-977a-0f16a9cd5e75-kube-api-access-j2q2j\") pod \"dnsmasq-dns-774686bfb5-gk9ch\" (UID: \"fe918116-df47-4b09-977a-0f16a9cd5e75\") " pod="openstack/dnsmasq-dns-774686bfb5-gk9ch" Oct 05 08:23:42 crc kubenswrapper[4846]: I1005 08:23:42.410777 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/fe918116-df47-4b09-977a-0f16a9cd5e75-ovsdbserver-sb\") pod \"dnsmasq-dns-774686bfb5-gk9ch\" (UID: \"fe918116-df47-4b09-977a-0f16a9cd5e75\") " pod="openstack/dnsmasq-dns-774686bfb5-gk9ch" Oct 05 08:23:42 crc kubenswrapper[4846]: I1005 08:23:42.411274 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fe918116-df47-4b09-977a-0f16a9cd5e75-dns-svc\") pod \"dnsmasq-dns-774686bfb5-gk9ch\" (UID: \"fe918116-df47-4b09-977a-0f16a9cd5e75\") " pod="openstack/dnsmasq-dns-774686bfb5-gk9ch" Oct 05 08:23:42 crc kubenswrapper[4846]: I1005 08:23:42.411762 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fe918116-df47-4b09-977a-0f16a9cd5e75-config\") pod \"dnsmasq-dns-774686bfb5-gk9ch\" (UID: \"fe918116-df47-4b09-977a-0f16a9cd5e75\") " pod="openstack/dnsmasq-dns-774686bfb5-gk9ch" Oct 05 08:23:42 crc kubenswrapper[4846]: I1005 08:23:42.414781 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/fe918116-df47-4b09-977a-0f16a9cd5e75-ovsdbserver-nb\") pod \"dnsmasq-dns-774686bfb5-gk9ch\" (UID: \"fe918116-df47-4b09-977a-0f16a9cd5e75\") " pod="openstack/dnsmasq-dns-774686bfb5-gk9ch" Oct 05 08:23:42 crc kubenswrapper[4846]: I1005 08:23:42.425650 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j2q2j\" (UniqueName: \"kubernetes.io/projected/fe918116-df47-4b09-977a-0f16a9cd5e75-kube-api-access-j2q2j\") pod \"dnsmasq-dns-774686bfb5-gk9ch\" (UID: \"fe918116-df47-4b09-977a-0f16a9cd5e75\") " pod="openstack/dnsmasq-dns-774686bfb5-gk9ch" Oct 05 08:23:42 crc kubenswrapper[4846]: I1005 08:23:42.439517 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 05 08:23:42 crc kubenswrapper[4846]: I1005 08:23:42.491230 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-774686bfb5-gk9ch" Oct 05 08:23:42 crc kubenswrapper[4846]: I1005 08:23:42.588657 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 05 08:23:42 crc kubenswrapper[4846]: W1005 08:23:42.645204 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod285fade3_a17c_4cfc_937c_096ebdb15848.slice/crio-eefea35ba5b4d42e228e8f87cb6bd6d7498e7203231f2f7e800bdce93d3f65b5 WatchSource:0}: Error finding container eefea35ba5b4d42e228e8f87cb6bd6d7498e7203231f2f7e800bdce93d3f65b5: Status 404 returned error can't find the container with id eefea35ba5b4d42e228e8f87cb6bd6d7498e7203231f2f7e800bdce93d3f65b5 Oct 05 08:23:42 crc kubenswrapper[4846]: I1005 08:23:42.650729 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-4kt89"] Oct 05 08:23:42 crc kubenswrapper[4846]: I1005 08:23:42.690192 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"5b32c235-5879-482b-a396-6d0d7e98816d","Type":"ContainerStarted","Data":"4b4e22560041b1fab9bdf5260cec89a79ca2e629ee57dbf8ed157093cccfa3c8"} Oct 05 08:23:42 crc kubenswrapper[4846]: I1005 08:23:42.698666 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 05 08:23:42 crc kubenswrapper[4846]: I1005 08:23:42.923738 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 05 08:23:42 crc kubenswrapper[4846]: W1005 08:23:42.937456 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7dd897db_45ea_491c_ace4_9e14fd8104dc.slice/crio-fc84a8f57da4592101f0d6afaa1cbb65bc48560c9ef1796fedf75a030ef1d362 WatchSource:0}: Error finding container fc84a8f57da4592101f0d6afaa1cbb65bc48560c9ef1796fedf75a030ef1d362: Status 404 returned error can't find the container with id fc84a8f57da4592101f0d6afaa1cbb65bc48560c9ef1796fedf75a030ef1d362 Oct 05 08:23:43 crc kubenswrapper[4846]: I1005 08:23:43.012691 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-db-sync-p9qj9"] Oct 05 08:23:43 crc kubenswrapper[4846]: I1005 08:23:43.014102 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-p9qj9" Oct 05 08:23:43 crc kubenswrapper[4846]: I1005 08:23:43.015843 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Oct 05 08:23:43 crc kubenswrapper[4846]: I1005 08:23:43.019590 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-scripts" Oct 05 08:23:43 crc kubenswrapper[4846]: I1005 08:23:43.025145 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5098a05b-cc49-4af9-b3cb-7d6ff17f38f2-config-data\") pod \"nova-cell1-conductor-db-sync-p9qj9\" (UID: \"5098a05b-cc49-4af9-b3cb-7d6ff17f38f2\") " pod="openstack/nova-cell1-conductor-db-sync-p9qj9" Oct 05 08:23:43 crc kubenswrapper[4846]: I1005 08:23:43.025268 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lq8n2\" (UniqueName: \"kubernetes.io/projected/5098a05b-cc49-4af9-b3cb-7d6ff17f38f2-kube-api-access-lq8n2\") pod \"nova-cell1-conductor-db-sync-p9qj9\" (UID: \"5098a05b-cc49-4af9-b3cb-7d6ff17f38f2\") " pod="openstack/nova-cell1-conductor-db-sync-p9qj9" Oct 05 08:23:43 crc kubenswrapper[4846]: I1005 08:23:43.025316 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5098a05b-cc49-4af9-b3cb-7d6ff17f38f2-scripts\") pod \"nova-cell1-conductor-db-sync-p9qj9\" (UID: \"5098a05b-cc49-4af9-b3cb-7d6ff17f38f2\") " pod="openstack/nova-cell1-conductor-db-sync-p9qj9" Oct 05 08:23:43 crc kubenswrapper[4846]: I1005 08:23:43.025336 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5098a05b-cc49-4af9-b3cb-7d6ff17f38f2-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-p9qj9\" (UID: \"5098a05b-cc49-4af9-b3cb-7d6ff17f38f2\") " pod="openstack/nova-cell1-conductor-db-sync-p9qj9" Oct 05 08:23:43 crc kubenswrapper[4846]: I1005 08:23:43.026596 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-p9qj9"] Oct 05 08:23:43 crc kubenswrapper[4846]: I1005 08:23:43.070458 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 05 08:23:43 crc kubenswrapper[4846]: W1005 08:23:43.087479 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podae74e41e_d09c_4666_85bf_998ff2834598.slice/crio-d9e4d03db3e6e0101686e86b30baf3bde1e3129a3608358255025b6449b17d0f WatchSource:0}: Error finding container d9e4d03db3e6e0101686e86b30baf3bde1e3129a3608358255025b6449b17d0f: Status 404 returned error can't find the container with id d9e4d03db3e6e0101686e86b30baf3bde1e3129a3608358255025b6449b17d0f Oct 05 08:23:43 crc kubenswrapper[4846]: I1005 08:23:43.125147 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-774686bfb5-gk9ch"] Oct 05 08:23:43 crc kubenswrapper[4846]: I1005 08:23:43.126640 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5098a05b-cc49-4af9-b3cb-7d6ff17f38f2-config-data\") pod \"nova-cell1-conductor-db-sync-p9qj9\" (UID: \"5098a05b-cc49-4af9-b3cb-7d6ff17f38f2\") " pod="openstack/nova-cell1-conductor-db-sync-p9qj9" Oct 05 08:23:43 crc kubenswrapper[4846]: I1005 08:23:43.126695 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lq8n2\" (UniqueName: \"kubernetes.io/projected/5098a05b-cc49-4af9-b3cb-7d6ff17f38f2-kube-api-access-lq8n2\") pod \"nova-cell1-conductor-db-sync-p9qj9\" (UID: \"5098a05b-cc49-4af9-b3cb-7d6ff17f38f2\") " pod="openstack/nova-cell1-conductor-db-sync-p9qj9" Oct 05 08:23:43 crc kubenswrapper[4846]: I1005 08:23:43.126740 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5098a05b-cc49-4af9-b3cb-7d6ff17f38f2-scripts\") pod \"nova-cell1-conductor-db-sync-p9qj9\" (UID: \"5098a05b-cc49-4af9-b3cb-7d6ff17f38f2\") " pod="openstack/nova-cell1-conductor-db-sync-p9qj9" Oct 05 08:23:43 crc kubenswrapper[4846]: I1005 08:23:43.126761 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5098a05b-cc49-4af9-b3cb-7d6ff17f38f2-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-p9qj9\" (UID: \"5098a05b-cc49-4af9-b3cb-7d6ff17f38f2\") " pod="openstack/nova-cell1-conductor-db-sync-p9qj9" Oct 05 08:23:43 crc kubenswrapper[4846]: W1005 08:23:43.131950 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfe918116_df47_4b09_977a_0f16a9cd5e75.slice/crio-9fe4b5bce5a8984f034964dd685870e3615e59fd43751e92fb3594aa4cf58cbe WatchSource:0}: Error finding container 9fe4b5bce5a8984f034964dd685870e3615e59fd43751e92fb3594aa4cf58cbe: Status 404 returned error can't find the container with id 9fe4b5bce5a8984f034964dd685870e3615e59fd43751e92fb3594aa4cf58cbe Oct 05 08:23:43 crc kubenswrapper[4846]: I1005 08:23:43.132497 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5098a05b-cc49-4af9-b3cb-7d6ff17f38f2-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-p9qj9\" (UID: \"5098a05b-cc49-4af9-b3cb-7d6ff17f38f2\") " pod="openstack/nova-cell1-conductor-db-sync-p9qj9" Oct 05 08:23:43 crc kubenswrapper[4846]: I1005 08:23:43.132659 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5098a05b-cc49-4af9-b3cb-7d6ff17f38f2-config-data\") pod \"nova-cell1-conductor-db-sync-p9qj9\" (UID: \"5098a05b-cc49-4af9-b3cb-7d6ff17f38f2\") " pod="openstack/nova-cell1-conductor-db-sync-p9qj9" Oct 05 08:23:43 crc kubenswrapper[4846]: I1005 08:23:43.133525 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5098a05b-cc49-4af9-b3cb-7d6ff17f38f2-scripts\") pod \"nova-cell1-conductor-db-sync-p9qj9\" (UID: \"5098a05b-cc49-4af9-b3cb-7d6ff17f38f2\") " pod="openstack/nova-cell1-conductor-db-sync-p9qj9" Oct 05 08:23:43 crc kubenswrapper[4846]: I1005 08:23:43.143697 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lq8n2\" (UniqueName: \"kubernetes.io/projected/5098a05b-cc49-4af9-b3cb-7d6ff17f38f2-kube-api-access-lq8n2\") pod \"nova-cell1-conductor-db-sync-p9qj9\" (UID: \"5098a05b-cc49-4af9-b3cb-7d6ff17f38f2\") " pod="openstack/nova-cell1-conductor-db-sync-p9qj9" Oct 05 08:23:43 crc kubenswrapper[4846]: I1005 08:23:43.347834 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-p9qj9" Oct 05 08:23:43 crc kubenswrapper[4846]: I1005 08:23:43.708838 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"7dd897db-45ea-491c-ace4-9e14fd8104dc","Type":"ContainerStarted","Data":"fc84a8f57da4592101f0d6afaa1cbb65bc48560c9ef1796fedf75a030ef1d362"} Oct 05 08:23:43 crc kubenswrapper[4846]: I1005 08:23:43.715858 4846 generic.go:334] "Generic (PLEG): container finished" podID="fe918116-df47-4b09-977a-0f16a9cd5e75" containerID="a5603d53b137d2ae2b7e74a3aec98074f8b5806e3c1a208442e8b69426282119" exitCode=0 Oct 05 08:23:43 crc kubenswrapper[4846]: I1005 08:23:43.715960 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-774686bfb5-gk9ch" event={"ID":"fe918116-df47-4b09-977a-0f16a9cd5e75","Type":"ContainerDied","Data":"a5603d53b137d2ae2b7e74a3aec98074f8b5806e3c1a208442e8b69426282119"} Oct 05 08:23:43 crc kubenswrapper[4846]: I1005 08:23:43.716159 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-774686bfb5-gk9ch" event={"ID":"fe918116-df47-4b09-977a-0f16a9cd5e75","Type":"ContainerStarted","Data":"9fe4b5bce5a8984f034964dd685870e3615e59fd43751e92fb3594aa4cf58cbe"} Oct 05 08:23:43 crc kubenswrapper[4846]: I1005 08:23:43.722814 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-4kt89" event={"ID":"285fade3-a17c-4cfc-937c-096ebdb15848","Type":"ContainerStarted","Data":"3fdb7b36842986265b48883b87d5734a37d51381cf8ee9b9978d76d1754d5762"} Oct 05 08:23:43 crc kubenswrapper[4846]: I1005 08:23:43.722851 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-4kt89" event={"ID":"285fade3-a17c-4cfc-937c-096ebdb15848","Type":"ContainerStarted","Data":"eefea35ba5b4d42e228e8f87cb6bd6d7498e7203231f2f7e800bdce93d3f65b5"} Oct 05 08:23:43 crc kubenswrapper[4846]: I1005 08:23:43.743031 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"ae74e41e-d09c-4666-85bf-998ff2834598","Type":"ContainerStarted","Data":"d9e4d03db3e6e0101686e86b30baf3bde1e3129a3608358255025b6449b17d0f"} Oct 05 08:23:43 crc kubenswrapper[4846]: I1005 08:23:43.750312 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"fec088ae-74f0-42bb-8225-428c7562f4c0","Type":"ContainerStarted","Data":"ab4c57ed069012578f10a551343b61429804dda258dda111020b3dcb40073360"} Oct 05 08:23:43 crc kubenswrapper[4846]: I1005 08:23:43.768187 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-cell-mapping-4kt89" podStartSLOduration=2.768152577 podStartE2EDuration="2.768152577s" podCreationTimestamp="2025-10-05 08:23:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 08:23:43.756631171 +0000 UTC m=+5745.997483946" watchObservedRunningTime="2025-10-05 08:23:43.768152577 +0000 UTC m=+5746.009005352" Oct 05 08:23:43 crc kubenswrapper[4846]: I1005 08:23:43.911608 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-p9qj9"] Oct 05 08:23:43 crc kubenswrapper[4846]: W1005 08:23:43.927312 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5098a05b_cc49_4af9_b3cb_7d6ff17f38f2.slice/crio-70a502663dc211eb6cf627ca08d54240bc973046a9b917021c909a6011772ee6 WatchSource:0}: Error finding container 70a502663dc211eb6cf627ca08d54240bc973046a9b917021c909a6011772ee6: Status 404 returned error can't find the container with id 70a502663dc211eb6cf627ca08d54240bc973046a9b917021c909a6011772ee6 Oct 05 08:23:44 crc kubenswrapper[4846]: I1005 08:23:44.759780 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-774686bfb5-gk9ch" event={"ID":"fe918116-df47-4b09-977a-0f16a9cd5e75","Type":"ContainerStarted","Data":"613fdc9574b6e854ad675725a919d6f0abac840992d9db4ceffa46ccef85db43"} Oct 05 08:23:44 crc kubenswrapper[4846]: I1005 08:23:44.760245 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-774686bfb5-gk9ch" Oct 05 08:23:44 crc kubenswrapper[4846]: I1005 08:23:44.766899 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-p9qj9" event={"ID":"5098a05b-cc49-4af9-b3cb-7d6ff17f38f2","Type":"ContainerStarted","Data":"70a502663dc211eb6cf627ca08d54240bc973046a9b917021c909a6011772ee6"} Oct 05 08:23:44 crc kubenswrapper[4846]: I1005 08:23:44.785247 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-774686bfb5-gk9ch" podStartSLOduration=2.785228155 podStartE2EDuration="2.785228155s" podCreationTimestamp="2025-10-05 08:23:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 08:23:44.781822925 +0000 UTC m=+5747.022675710" watchObservedRunningTime="2025-10-05 08:23:44.785228155 +0000 UTC m=+5747.026080930" Oct 05 08:23:46 crc kubenswrapper[4846]: I1005 08:23:46.037070 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 05 08:23:46 crc kubenswrapper[4846]: I1005 08:23:46.052831 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 05 08:23:46 crc kubenswrapper[4846]: I1005 08:23:46.785516 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-p9qj9" event={"ID":"5098a05b-cc49-4af9-b3cb-7d6ff17f38f2","Type":"ContainerStarted","Data":"3d4a8fd72c0b7509378ec345a1a94dba89ea5a6e23491caf00bbff90fdab2edb"} Oct 05 08:23:46 crc kubenswrapper[4846]: I1005 08:23:46.805122 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-db-sync-p9qj9" podStartSLOduration=4.805103154 podStartE2EDuration="4.805103154s" podCreationTimestamp="2025-10-05 08:23:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 08:23:46.803641575 +0000 UTC m=+5749.044494350" watchObservedRunningTime="2025-10-05 08:23:46.805103154 +0000 UTC m=+5749.045955929" Oct 05 08:23:47 crc kubenswrapper[4846]: I1005 08:23:47.794809 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"ae74e41e-d09c-4666-85bf-998ff2834598","Type":"ContainerStarted","Data":"8a0404d21b974b925b073efe33f608deed3c290d2b7d7c46636142d299f72b85"} Oct 05 08:23:47 crc kubenswrapper[4846]: I1005 08:23:47.795156 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"ae74e41e-d09c-4666-85bf-998ff2834598","Type":"ContainerStarted","Data":"cb0a188943d876625d7cba5eaa5a2442291eb423189efba4da3beba6f0a6d7f9"} Oct 05 08:23:47 crc kubenswrapper[4846]: I1005 08:23:47.794933 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="ae74e41e-d09c-4666-85bf-998ff2834598" containerName="nova-metadata-metadata" containerID="cri-o://8a0404d21b974b925b073efe33f608deed3c290d2b7d7c46636142d299f72b85" gracePeriod=30 Oct 05 08:23:47 crc kubenswrapper[4846]: I1005 08:23:47.794861 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="ae74e41e-d09c-4666-85bf-998ff2834598" containerName="nova-metadata-log" containerID="cri-o://cb0a188943d876625d7cba5eaa5a2442291eb423189efba4da3beba6f0a6d7f9" gracePeriod=30 Oct 05 08:23:47 crc kubenswrapper[4846]: I1005 08:23:47.800768 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"fec088ae-74f0-42bb-8225-428c7562f4c0","Type":"ContainerStarted","Data":"7da5d1791e7d851f68e870544b931c04cce1e5bdfa1fb2d9e4c7e9c64e5d6b7a"} Oct 05 08:23:47 crc kubenswrapper[4846]: I1005 08:23:47.800839 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="fec088ae-74f0-42bb-8225-428c7562f4c0" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://7da5d1791e7d851f68e870544b931c04cce1e5bdfa1fb2d9e4c7e9c64e5d6b7a" gracePeriod=30 Oct 05 08:23:47 crc kubenswrapper[4846]: I1005 08:23:47.803048 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"7dd897db-45ea-491c-ace4-9e14fd8104dc","Type":"ContainerStarted","Data":"a9b2b33f2bcfd4bcf52586879c9e4a052e012f883cf43a7a91236b907b130f38"} Oct 05 08:23:47 crc kubenswrapper[4846]: I1005 08:23:47.803234 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"7dd897db-45ea-491c-ace4-9e14fd8104dc","Type":"ContainerStarted","Data":"55def4cf6a9df17558b786604340d80a748c419979728eed68a3e3a1093b52d0"} Oct 05 08:23:47 crc kubenswrapper[4846]: I1005 08:23:47.807987 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"5b32c235-5879-482b-a396-6d0d7e98816d","Type":"ContainerStarted","Data":"500743e9d035ab2c1788d5f88dc1507f2bc7aa4912a9f386cf8d555dbdec9935"} Oct 05 08:23:47 crc kubenswrapper[4846]: I1005 08:23:47.819920 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.898018325 podStartE2EDuration="6.819898111s" podCreationTimestamp="2025-10-05 08:23:41 +0000 UTC" firstStartedPulling="2025-10-05 08:23:43.091633082 +0000 UTC m=+5745.332485867" lastFinishedPulling="2025-10-05 08:23:47.013512838 +0000 UTC m=+5749.254365653" observedRunningTime="2025-10-05 08:23:47.814381424 +0000 UTC m=+5750.055234199" watchObservedRunningTime="2025-10-05 08:23:47.819898111 +0000 UTC m=+5750.060750886" Oct 05 08:23:47 crc kubenswrapper[4846]: I1005 08:23:47.857750 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.468308054 podStartE2EDuration="6.857730056s" podCreationTimestamp="2025-10-05 08:23:41 +0000 UTC" firstStartedPulling="2025-10-05 08:23:42.622157205 +0000 UTC m=+5744.863009980" lastFinishedPulling="2025-10-05 08:23:47.011579207 +0000 UTC m=+5749.252431982" observedRunningTime="2025-10-05 08:23:47.839693477 +0000 UTC m=+5750.080546252" watchObservedRunningTime="2025-10-05 08:23:47.857730056 +0000 UTC m=+5750.098582831" Oct 05 08:23:47 crc kubenswrapper[4846]: I1005 08:23:47.860663 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.787236653 podStartE2EDuration="6.860655563s" podCreationTimestamp="2025-10-05 08:23:41 +0000 UTC" firstStartedPulling="2025-10-05 08:23:42.939909683 +0000 UTC m=+5745.180762458" lastFinishedPulling="2025-10-05 08:23:47.013328593 +0000 UTC m=+5749.254181368" observedRunningTime="2025-10-05 08:23:47.85452105 +0000 UTC m=+5750.095373825" watchObservedRunningTime="2025-10-05 08:23:47.860655563 +0000 UTC m=+5750.101508338" Oct 05 08:23:47 crc kubenswrapper[4846]: I1005 08:23:47.873339 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.61950971 podStartE2EDuration="6.87332246s" podCreationTimestamp="2025-10-05 08:23:41 +0000 UTC" firstStartedPulling="2025-10-05 08:23:42.757765297 +0000 UTC m=+5744.998618072" lastFinishedPulling="2025-10-05 08:23:47.011578047 +0000 UTC m=+5749.252430822" observedRunningTime="2025-10-05 08:23:47.868635345 +0000 UTC m=+5750.109488120" watchObservedRunningTime="2025-10-05 08:23:47.87332246 +0000 UTC m=+5750.114175235" Oct 05 08:23:48 crc kubenswrapper[4846]: I1005 08:23:48.418293 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 05 08:23:48 crc kubenswrapper[4846]: I1005 08:23:48.452230 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ae74e41e-d09c-4666-85bf-998ff2834598-logs\") pod \"ae74e41e-d09c-4666-85bf-998ff2834598\" (UID: \"ae74e41e-d09c-4666-85bf-998ff2834598\") " Oct 05 08:23:48 crc kubenswrapper[4846]: I1005 08:23:48.452333 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dktkp\" (UniqueName: \"kubernetes.io/projected/ae74e41e-d09c-4666-85bf-998ff2834598-kube-api-access-dktkp\") pod \"ae74e41e-d09c-4666-85bf-998ff2834598\" (UID: \"ae74e41e-d09c-4666-85bf-998ff2834598\") " Oct 05 08:23:48 crc kubenswrapper[4846]: I1005 08:23:48.452451 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ae74e41e-d09c-4666-85bf-998ff2834598-combined-ca-bundle\") pod \"ae74e41e-d09c-4666-85bf-998ff2834598\" (UID: \"ae74e41e-d09c-4666-85bf-998ff2834598\") " Oct 05 08:23:48 crc kubenswrapper[4846]: I1005 08:23:48.452566 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ae74e41e-d09c-4666-85bf-998ff2834598-config-data\") pod \"ae74e41e-d09c-4666-85bf-998ff2834598\" (UID: \"ae74e41e-d09c-4666-85bf-998ff2834598\") " Oct 05 08:23:48 crc kubenswrapper[4846]: I1005 08:23:48.452689 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ae74e41e-d09c-4666-85bf-998ff2834598-logs" (OuterVolumeSpecName: "logs") pod "ae74e41e-d09c-4666-85bf-998ff2834598" (UID: "ae74e41e-d09c-4666-85bf-998ff2834598"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 08:23:48 crc kubenswrapper[4846]: I1005 08:23:48.453365 4846 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ae74e41e-d09c-4666-85bf-998ff2834598-logs\") on node \"crc\" DevicePath \"\"" Oct 05 08:23:48 crc kubenswrapper[4846]: I1005 08:23:48.467763 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ae74e41e-d09c-4666-85bf-998ff2834598-kube-api-access-dktkp" (OuterVolumeSpecName: "kube-api-access-dktkp") pod "ae74e41e-d09c-4666-85bf-998ff2834598" (UID: "ae74e41e-d09c-4666-85bf-998ff2834598"). InnerVolumeSpecName "kube-api-access-dktkp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:23:48 crc kubenswrapper[4846]: I1005 08:23:48.486362 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ae74e41e-d09c-4666-85bf-998ff2834598-config-data" (OuterVolumeSpecName: "config-data") pod "ae74e41e-d09c-4666-85bf-998ff2834598" (UID: "ae74e41e-d09c-4666-85bf-998ff2834598"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:23:48 crc kubenswrapper[4846]: I1005 08:23:48.488267 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ae74e41e-d09c-4666-85bf-998ff2834598-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ae74e41e-d09c-4666-85bf-998ff2834598" (UID: "ae74e41e-d09c-4666-85bf-998ff2834598"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:23:48 crc kubenswrapper[4846]: I1005 08:23:48.554770 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dktkp\" (UniqueName: \"kubernetes.io/projected/ae74e41e-d09c-4666-85bf-998ff2834598-kube-api-access-dktkp\") on node \"crc\" DevicePath \"\"" Oct 05 08:23:48 crc kubenswrapper[4846]: I1005 08:23:48.554846 4846 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ae74e41e-d09c-4666-85bf-998ff2834598-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 08:23:48 crc kubenswrapper[4846]: I1005 08:23:48.554883 4846 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ae74e41e-d09c-4666-85bf-998ff2834598-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 08:23:48 crc kubenswrapper[4846]: I1005 08:23:48.821574 4846 generic.go:334] "Generic (PLEG): container finished" podID="ae74e41e-d09c-4666-85bf-998ff2834598" containerID="8a0404d21b974b925b073efe33f608deed3c290d2b7d7c46636142d299f72b85" exitCode=0 Oct 05 08:23:48 crc kubenswrapper[4846]: I1005 08:23:48.823123 4846 generic.go:334] "Generic (PLEG): container finished" podID="ae74e41e-d09c-4666-85bf-998ff2834598" containerID="cb0a188943d876625d7cba5eaa5a2442291eb423189efba4da3beba6f0a6d7f9" exitCode=143 Oct 05 08:23:48 crc kubenswrapper[4846]: I1005 08:23:48.823352 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"ae74e41e-d09c-4666-85bf-998ff2834598","Type":"ContainerDied","Data":"8a0404d21b974b925b073efe33f608deed3c290d2b7d7c46636142d299f72b85"} Oct 05 08:23:48 crc kubenswrapper[4846]: I1005 08:23:48.823683 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"ae74e41e-d09c-4666-85bf-998ff2834598","Type":"ContainerDied","Data":"cb0a188943d876625d7cba5eaa5a2442291eb423189efba4da3beba6f0a6d7f9"} Oct 05 08:23:48 crc kubenswrapper[4846]: I1005 08:23:48.823805 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"ae74e41e-d09c-4666-85bf-998ff2834598","Type":"ContainerDied","Data":"d9e4d03db3e6e0101686e86b30baf3bde1e3129a3608358255025b6449b17d0f"} Oct 05 08:23:48 crc kubenswrapper[4846]: I1005 08:23:48.823954 4846 scope.go:117] "RemoveContainer" containerID="8a0404d21b974b925b073efe33f608deed3c290d2b7d7c46636142d299f72b85" Oct 05 08:23:48 crc kubenswrapper[4846]: I1005 08:23:48.824299 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 05 08:23:48 crc kubenswrapper[4846]: I1005 08:23:48.827711 4846 generic.go:334] "Generic (PLEG): container finished" podID="285fade3-a17c-4cfc-937c-096ebdb15848" containerID="3fdb7b36842986265b48883b87d5734a37d51381cf8ee9b9978d76d1754d5762" exitCode=0 Oct 05 08:23:48 crc kubenswrapper[4846]: I1005 08:23:48.828455 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-4kt89" event={"ID":"285fade3-a17c-4cfc-937c-096ebdb15848","Type":"ContainerDied","Data":"3fdb7b36842986265b48883b87d5734a37d51381cf8ee9b9978d76d1754d5762"} Oct 05 08:23:48 crc kubenswrapper[4846]: I1005 08:23:48.866951 4846 scope.go:117] "RemoveContainer" containerID="cb0a188943d876625d7cba5eaa5a2442291eb423189efba4da3beba6f0a6d7f9" Oct 05 08:23:48 crc kubenswrapper[4846]: I1005 08:23:48.880771 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 05 08:23:48 crc kubenswrapper[4846]: I1005 08:23:48.905351 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Oct 05 08:23:48 crc kubenswrapper[4846]: I1005 08:23:48.913060 4846 scope.go:117] "RemoveContainer" containerID="8a0404d21b974b925b073efe33f608deed3c290d2b7d7c46636142d299f72b85" Oct 05 08:23:48 crc kubenswrapper[4846]: E1005 08:23:48.913792 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8a0404d21b974b925b073efe33f608deed3c290d2b7d7c46636142d299f72b85\": container with ID starting with 8a0404d21b974b925b073efe33f608deed3c290d2b7d7c46636142d299f72b85 not found: ID does not exist" containerID="8a0404d21b974b925b073efe33f608deed3c290d2b7d7c46636142d299f72b85" Oct 05 08:23:48 crc kubenswrapper[4846]: I1005 08:23:48.913845 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8a0404d21b974b925b073efe33f608deed3c290d2b7d7c46636142d299f72b85"} err="failed to get container status \"8a0404d21b974b925b073efe33f608deed3c290d2b7d7c46636142d299f72b85\": rpc error: code = NotFound desc = could not find container \"8a0404d21b974b925b073efe33f608deed3c290d2b7d7c46636142d299f72b85\": container with ID starting with 8a0404d21b974b925b073efe33f608deed3c290d2b7d7c46636142d299f72b85 not found: ID does not exist" Oct 05 08:23:48 crc kubenswrapper[4846]: I1005 08:23:48.913886 4846 scope.go:117] "RemoveContainer" containerID="cb0a188943d876625d7cba5eaa5a2442291eb423189efba4da3beba6f0a6d7f9" Oct 05 08:23:48 crc kubenswrapper[4846]: E1005 08:23:48.914455 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cb0a188943d876625d7cba5eaa5a2442291eb423189efba4da3beba6f0a6d7f9\": container with ID starting with cb0a188943d876625d7cba5eaa5a2442291eb423189efba4da3beba6f0a6d7f9 not found: ID does not exist" containerID="cb0a188943d876625d7cba5eaa5a2442291eb423189efba4da3beba6f0a6d7f9" Oct 05 08:23:48 crc kubenswrapper[4846]: I1005 08:23:48.914480 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cb0a188943d876625d7cba5eaa5a2442291eb423189efba4da3beba6f0a6d7f9"} err="failed to get container status \"cb0a188943d876625d7cba5eaa5a2442291eb423189efba4da3beba6f0a6d7f9\": rpc error: code = NotFound desc = could not find container \"cb0a188943d876625d7cba5eaa5a2442291eb423189efba4da3beba6f0a6d7f9\": container with ID starting with cb0a188943d876625d7cba5eaa5a2442291eb423189efba4da3beba6f0a6d7f9 not found: ID does not exist" Oct 05 08:23:48 crc kubenswrapper[4846]: I1005 08:23:48.914497 4846 scope.go:117] "RemoveContainer" containerID="8a0404d21b974b925b073efe33f608deed3c290d2b7d7c46636142d299f72b85" Oct 05 08:23:48 crc kubenswrapper[4846]: I1005 08:23:48.914973 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8a0404d21b974b925b073efe33f608deed3c290d2b7d7c46636142d299f72b85"} err="failed to get container status \"8a0404d21b974b925b073efe33f608deed3c290d2b7d7c46636142d299f72b85\": rpc error: code = NotFound desc = could not find container \"8a0404d21b974b925b073efe33f608deed3c290d2b7d7c46636142d299f72b85\": container with ID starting with 8a0404d21b974b925b073efe33f608deed3c290d2b7d7c46636142d299f72b85 not found: ID does not exist" Oct 05 08:23:48 crc kubenswrapper[4846]: I1005 08:23:48.914995 4846 scope.go:117] "RemoveContainer" containerID="cb0a188943d876625d7cba5eaa5a2442291eb423189efba4da3beba6f0a6d7f9" Oct 05 08:23:48 crc kubenswrapper[4846]: I1005 08:23:48.916746 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cb0a188943d876625d7cba5eaa5a2442291eb423189efba4da3beba6f0a6d7f9"} err="failed to get container status \"cb0a188943d876625d7cba5eaa5a2442291eb423189efba4da3beba6f0a6d7f9\": rpc error: code = NotFound desc = could not find container \"cb0a188943d876625d7cba5eaa5a2442291eb423189efba4da3beba6f0a6d7f9\": container with ID starting with cb0a188943d876625d7cba5eaa5a2442291eb423189efba4da3beba6f0a6d7f9 not found: ID does not exist" Oct 05 08:23:48 crc kubenswrapper[4846]: I1005 08:23:48.925574 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 05 08:23:48 crc kubenswrapper[4846]: E1005 08:23:48.926147 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ae74e41e-d09c-4666-85bf-998ff2834598" containerName="nova-metadata-log" Oct 05 08:23:48 crc kubenswrapper[4846]: I1005 08:23:48.926197 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="ae74e41e-d09c-4666-85bf-998ff2834598" containerName="nova-metadata-log" Oct 05 08:23:48 crc kubenswrapper[4846]: E1005 08:23:48.926229 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ae74e41e-d09c-4666-85bf-998ff2834598" containerName="nova-metadata-metadata" Oct 05 08:23:48 crc kubenswrapper[4846]: I1005 08:23:48.926241 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="ae74e41e-d09c-4666-85bf-998ff2834598" containerName="nova-metadata-metadata" Oct 05 08:23:48 crc kubenswrapper[4846]: I1005 08:23:48.926588 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="ae74e41e-d09c-4666-85bf-998ff2834598" containerName="nova-metadata-metadata" Oct 05 08:23:48 crc kubenswrapper[4846]: I1005 08:23:48.926617 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="ae74e41e-d09c-4666-85bf-998ff2834598" containerName="nova-metadata-log" Oct 05 08:23:48 crc kubenswrapper[4846]: I1005 08:23:48.928937 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 05 08:23:48 crc kubenswrapper[4846]: I1005 08:23:48.931358 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 05 08:23:48 crc kubenswrapper[4846]: I1005 08:23:48.932087 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 05 08:23:48 crc kubenswrapper[4846]: I1005 08:23:48.932595 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Oct 05 08:23:48 crc kubenswrapper[4846]: I1005 08:23:48.978108 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6b299eca-df14-4479-90a1-41f096335519-config-data\") pod \"nova-metadata-0\" (UID: \"6b299eca-df14-4479-90a1-41f096335519\") " pod="openstack/nova-metadata-0" Oct 05 08:23:48 crc kubenswrapper[4846]: I1005 08:23:48.978171 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6b299eca-df14-4479-90a1-41f096335519-logs\") pod \"nova-metadata-0\" (UID: \"6b299eca-df14-4479-90a1-41f096335519\") " pod="openstack/nova-metadata-0" Oct 05 08:23:48 crc kubenswrapper[4846]: I1005 08:23:48.978235 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sf4rd\" (UniqueName: \"kubernetes.io/projected/6b299eca-df14-4479-90a1-41f096335519-kube-api-access-sf4rd\") pod \"nova-metadata-0\" (UID: \"6b299eca-df14-4479-90a1-41f096335519\") " pod="openstack/nova-metadata-0" Oct 05 08:23:48 crc kubenswrapper[4846]: I1005 08:23:48.978486 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b299eca-df14-4479-90a1-41f096335519-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"6b299eca-df14-4479-90a1-41f096335519\") " pod="openstack/nova-metadata-0" Oct 05 08:23:48 crc kubenswrapper[4846]: I1005 08:23:48.978703 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/6b299eca-df14-4479-90a1-41f096335519-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"6b299eca-df14-4479-90a1-41f096335519\") " pod="openstack/nova-metadata-0" Oct 05 08:23:49 crc kubenswrapper[4846]: I1005 08:23:49.082970 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b299eca-df14-4479-90a1-41f096335519-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"6b299eca-df14-4479-90a1-41f096335519\") " pod="openstack/nova-metadata-0" Oct 05 08:23:49 crc kubenswrapper[4846]: I1005 08:23:49.083075 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/6b299eca-df14-4479-90a1-41f096335519-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"6b299eca-df14-4479-90a1-41f096335519\") " pod="openstack/nova-metadata-0" Oct 05 08:23:49 crc kubenswrapper[4846]: I1005 08:23:49.083113 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6b299eca-df14-4479-90a1-41f096335519-config-data\") pod \"nova-metadata-0\" (UID: \"6b299eca-df14-4479-90a1-41f096335519\") " pod="openstack/nova-metadata-0" Oct 05 08:23:49 crc kubenswrapper[4846]: I1005 08:23:49.083151 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6b299eca-df14-4479-90a1-41f096335519-logs\") pod \"nova-metadata-0\" (UID: \"6b299eca-df14-4479-90a1-41f096335519\") " pod="openstack/nova-metadata-0" Oct 05 08:23:49 crc kubenswrapper[4846]: I1005 08:23:49.083219 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sf4rd\" (UniqueName: \"kubernetes.io/projected/6b299eca-df14-4479-90a1-41f096335519-kube-api-access-sf4rd\") pod \"nova-metadata-0\" (UID: \"6b299eca-df14-4479-90a1-41f096335519\") " pod="openstack/nova-metadata-0" Oct 05 08:23:49 crc kubenswrapper[4846]: I1005 08:23:49.083676 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6b299eca-df14-4479-90a1-41f096335519-logs\") pod \"nova-metadata-0\" (UID: \"6b299eca-df14-4479-90a1-41f096335519\") " pod="openstack/nova-metadata-0" Oct 05 08:23:49 crc kubenswrapper[4846]: I1005 08:23:49.086691 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b299eca-df14-4479-90a1-41f096335519-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"6b299eca-df14-4479-90a1-41f096335519\") " pod="openstack/nova-metadata-0" Oct 05 08:23:49 crc kubenswrapper[4846]: I1005 08:23:49.087799 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6b299eca-df14-4479-90a1-41f096335519-config-data\") pod \"nova-metadata-0\" (UID: \"6b299eca-df14-4479-90a1-41f096335519\") " pod="openstack/nova-metadata-0" Oct 05 08:23:49 crc kubenswrapper[4846]: I1005 08:23:49.095035 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/6b299eca-df14-4479-90a1-41f096335519-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"6b299eca-df14-4479-90a1-41f096335519\") " pod="openstack/nova-metadata-0" Oct 05 08:23:49 crc kubenswrapper[4846]: I1005 08:23:49.101016 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sf4rd\" (UniqueName: \"kubernetes.io/projected/6b299eca-df14-4479-90a1-41f096335519-kube-api-access-sf4rd\") pod \"nova-metadata-0\" (UID: \"6b299eca-df14-4479-90a1-41f096335519\") " pod="openstack/nova-metadata-0" Oct 05 08:23:49 crc kubenswrapper[4846]: I1005 08:23:49.256139 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 05 08:23:49 crc kubenswrapper[4846]: I1005 08:23:49.714034 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 05 08:23:49 crc kubenswrapper[4846]: I1005 08:23:49.842811 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"6b299eca-df14-4479-90a1-41f096335519","Type":"ContainerStarted","Data":"8aae06db8d81d026c500e08a3b4a727ae019e4b4bf609e42eb162bac0c5afc2a"} Oct 05 08:23:49 crc kubenswrapper[4846]: I1005 08:23:49.845271 4846 generic.go:334] "Generic (PLEG): container finished" podID="5098a05b-cc49-4af9-b3cb-7d6ff17f38f2" containerID="3d4a8fd72c0b7509378ec345a1a94dba89ea5a6e23491caf00bbff90fdab2edb" exitCode=0 Oct 05 08:23:49 crc kubenswrapper[4846]: I1005 08:23:49.845347 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-p9qj9" event={"ID":"5098a05b-cc49-4af9-b3cb-7d6ff17f38f2","Type":"ContainerDied","Data":"3d4a8fd72c0b7509378ec345a1a94dba89ea5a6e23491caf00bbff90fdab2edb"} Oct 05 08:23:50 crc kubenswrapper[4846]: I1005 08:23:50.127535 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-4kt89" Oct 05 08:23:50 crc kubenswrapper[4846]: I1005 08:23:50.203078 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mgqpx\" (UniqueName: \"kubernetes.io/projected/285fade3-a17c-4cfc-937c-096ebdb15848-kube-api-access-mgqpx\") pod \"285fade3-a17c-4cfc-937c-096ebdb15848\" (UID: \"285fade3-a17c-4cfc-937c-096ebdb15848\") " Oct 05 08:23:50 crc kubenswrapper[4846]: I1005 08:23:50.203238 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/285fade3-a17c-4cfc-937c-096ebdb15848-scripts\") pod \"285fade3-a17c-4cfc-937c-096ebdb15848\" (UID: \"285fade3-a17c-4cfc-937c-096ebdb15848\") " Oct 05 08:23:50 crc kubenswrapper[4846]: I1005 08:23:50.203333 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/285fade3-a17c-4cfc-937c-096ebdb15848-config-data\") pod \"285fade3-a17c-4cfc-937c-096ebdb15848\" (UID: \"285fade3-a17c-4cfc-937c-096ebdb15848\") " Oct 05 08:23:50 crc kubenswrapper[4846]: I1005 08:23:50.203415 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/285fade3-a17c-4cfc-937c-096ebdb15848-combined-ca-bundle\") pod \"285fade3-a17c-4cfc-937c-096ebdb15848\" (UID: \"285fade3-a17c-4cfc-937c-096ebdb15848\") " Oct 05 08:23:50 crc kubenswrapper[4846]: I1005 08:23:50.209145 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/285fade3-a17c-4cfc-937c-096ebdb15848-kube-api-access-mgqpx" (OuterVolumeSpecName: "kube-api-access-mgqpx") pod "285fade3-a17c-4cfc-937c-096ebdb15848" (UID: "285fade3-a17c-4cfc-937c-096ebdb15848"). InnerVolumeSpecName "kube-api-access-mgqpx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:23:50 crc kubenswrapper[4846]: I1005 08:23:50.209817 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/285fade3-a17c-4cfc-937c-096ebdb15848-scripts" (OuterVolumeSpecName: "scripts") pod "285fade3-a17c-4cfc-937c-096ebdb15848" (UID: "285fade3-a17c-4cfc-937c-096ebdb15848"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:23:50 crc kubenswrapper[4846]: I1005 08:23:50.234466 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/285fade3-a17c-4cfc-937c-096ebdb15848-config-data" (OuterVolumeSpecName: "config-data") pod "285fade3-a17c-4cfc-937c-096ebdb15848" (UID: "285fade3-a17c-4cfc-937c-096ebdb15848"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:23:50 crc kubenswrapper[4846]: I1005 08:23:50.235442 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/285fade3-a17c-4cfc-937c-096ebdb15848-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "285fade3-a17c-4cfc-937c-096ebdb15848" (UID: "285fade3-a17c-4cfc-937c-096ebdb15848"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:23:50 crc kubenswrapper[4846]: I1005 08:23:50.306154 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mgqpx\" (UniqueName: \"kubernetes.io/projected/285fade3-a17c-4cfc-937c-096ebdb15848-kube-api-access-mgqpx\") on node \"crc\" DevicePath \"\"" Oct 05 08:23:50 crc kubenswrapper[4846]: I1005 08:23:50.306202 4846 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/285fade3-a17c-4cfc-937c-096ebdb15848-scripts\") on node \"crc\" DevicePath \"\"" Oct 05 08:23:50 crc kubenswrapper[4846]: I1005 08:23:50.306214 4846 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/285fade3-a17c-4cfc-937c-096ebdb15848-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 08:23:50 crc kubenswrapper[4846]: I1005 08:23:50.306227 4846 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/285fade3-a17c-4cfc-937c-096ebdb15848-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 08:23:50 crc kubenswrapper[4846]: I1005 08:23:50.516095 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ae74e41e-d09c-4666-85bf-998ff2834598" path="/var/lib/kubelet/pods/ae74e41e-d09c-4666-85bf-998ff2834598/volumes" Oct 05 08:23:50 crc kubenswrapper[4846]: I1005 08:23:50.894255 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"6b299eca-df14-4479-90a1-41f096335519","Type":"ContainerStarted","Data":"bd9218496be534953fa5dc544c2ec447b71d71f3bd829bb398620a64311688aa"} Oct 05 08:23:50 crc kubenswrapper[4846]: I1005 08:23:50.894347 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"6b299eca-df14-4479-90a1-41f096335519","Type":"ContainerStarted","Data":"dd5f26ecb5ebeacdc0dfdf9dee15f21dafa24300d2187a2cb65da7c58895d7b7"} Oct 05 08:23:50 crc kubenswrapper[4846]: I1005 08:23:50.903383 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-4kt89" Oct 05 08:23:50 crc kubenswrapper[4846]: I1005 08:23:50.903435 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-4kt89" event={"ID":"285fade3-a17c-4cfc-937c-096ebdb15848","Type":"ContainerDied","Data":"eefea35ba5b4d42e228e8f87cb6bd6d7498e7203231f2f7e800bdce93d3f65b5"} Oct 05 08:23:50 crc kubenswrapper[4846]: I1005 08:23:50.903557 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="eefea35ba5b4d42e228e8f87cb6bd6d7498e7203231f2f7e800bdce93d3f65b5" Oct 05 08:23:50 crc kubenswrapper[4846]: I1005 08:23:50.929902 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.929886917 podStartE2EDuration="2.929886917s" podCreationTimestamp="2025-10-05 08:23:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 08:23:50.928914521 +0000 UTC m=+5753.169767366" watchObservedRunningTime="2025-10-05 08:23:50.929886917 +0000 UTC m=+5753.170739702" Oct 05 08:23:51 crc kubenswrapper[4846]: I1005 08:23:51.058491 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 05 08:23:51 crc kubenswrapper[4846]: I1005 08:23:51.058787 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="7dd897db-45ea-491c-ace4-9e14fd8104dc" containerName="nova-api-log" containerID="cri-o://55def4cf6a9df17558b786604340d80a748c419979728eed68a3e3a1093b52d0" gracePeriod=30 Oct 05 08:23:51 crc kubenswrapper[4846]: I1005 08:23:51.059350 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="7dd897db-45ea-491c-ace4-9e14fd8104dc" containerName="nova-api-api" containerID="cri-o://a9b2b33f2bcfd4bcf52586879c9e4a052e012f883cf43a7a91236b907b130f38" gracePeriod=30 Oct 05 08:23:51 crc kubenswrapper[4846]: I1005 08:23:51.070916 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 05 08:23:51 crc kubenswrapper[4846]: I1005 08:23:51.071139 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="5b32c235-5879-482b-a396-6d0d7e98816d" containerName="nova-scheduler-scheduler" containerID="cri-o://500743e9d035ab2c1788d5f88dc1507f2bc7aa4912a9f386cf8d555dbdec9935" gracePeriod=30 Oct 05 08:23:51 crc kubenswrapper[4846]: I1005 08:23:51.149090 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 05 08:23:51 crc kubenswrapper[4846]: I1005 08:23:51.365672 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-p9qj9" Oct 05 08:23:51 crc kubenswrapper[4846]: I1005 08:23:51.447882 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5098a05b-cc49-4af9-b3cb-7d6ff17f38f2-config-data\") pod \"5098a05b-cc49-4af9-b3cb-7d6ff17f38f2\" (UID: \"5098a05b-cc49-4af9-b3cb-7d6ff17f38f2\") " Oct 05 08:23:51 crc kubenswrapper[4846]: I1005 08:23:51.448002 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lq8n2\" (UniqueName: \"kubernetes.io/projected/5098a05b-cc49-4af9-b3cb-7d6ff17f38f2-kube-api-access-lq8n2\") pod \"5098a05b-cc49-4af9-b3cb-7d6ff17f38f2\" (UID: \"5098a05b-cc49-4af9-b3cb-7d6ff17f38f2\") " Oct 05 08:23:51 crc kubenswrapper[4846]: I1005 08:23:51.448083 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5098a05b-cc49-4af9-b3cb-7d6ff17f38f2-combined-ca-bundle\") pod \"5098a05b-cc49-4af9-b3cb-7d6ff17f38f2\" (UID: \"5098a05b-cc49-4af9-b3cb-7d6ff17f38f2\") " Oct 05 08:23:51 crc kubenswrapper[4846]: I1005 08:23:51.448104 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5098a05b-cc49-4af9-b3cb-7d6ff17f38f2-scripts\") pod \"5098a05b-cc49-4af9-b3cb-7d6ff17f38f2\" (UID: \"5098a05b-cc49-4af9-b3cb-7d6ff17f38f2\") " Oct 05 08:23:51 crc kubenswrapper[4846]: I1005 08:23:51.454663 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5098a05b-cc49-4af9-b3cb-7d6ff17f38f2-kube-api-access-lq8n2" (OuterVolumeSpecName: "kube-api-access-lq8n2") pod "5098a05b-cc49-4af9-b3cb-7d6ff17f38f2" (UID: "5098a05b-cc49-4af9-b3cb-7d6ff17f38f2"). InnerVolumeSpecName "kube-api-access-lq8n2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:23:51 crc kubenswrapper[4846]: I1005 08:23:51.495091 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5098a05b-cc49-4af9-b3cb-7d6ff17f38f2-scripts" (OuterVolumeSpecName: "scripts") pod "5098a05b-cc49-4af9-b3cb-7d6ff17f38f2" (UID: "5098a05b-cc49-4af9-b3cb-7d6ff17f38f2"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:23:51 crc kubenswrapper[4846]: I1005 08:23:51.500969 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5098a05b-cc49-4af9-b3cb-7d6ff17f38f2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5098a05b-cc49-4af9-b3cb-7d6ff17f38f2" (UID: "5098a05b-cc49-4af9-b3cb-7d6ff17f38f2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:23:51 crc kubenswrapper[4846]: I1005 08:23:51.503694 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5098a05b-cc49-4af9-b3cb-7d6ff17f38f2-config-data" (OuterVolumeSpecName: "config-data") pod "5098a05b-cc49-4af9-b3cb-7d6ff17f38f2" (UID: "5098a05b-cc49-4af9-b3cb-7d6ff17f38f2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:23:51 crc kubenswrapper[4846]: I1005 08:23:51.550650 4846 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5098a05b-cc49-4af9-b3cb-7d6ff17f38f2-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 08:23:51 crc kubenswrapper[4846]: I1005 08:23:51.550687 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lq8n2\" (UniqueName: \"kubernetes.io/projected/5098a05b-cc49-4af9-b3cb-7d6ff17f38f2-kube-api-access-lq8n2\") on node \"crc\" DevicePath \"\"" Oct 05 08:23:51 crc kubenswrapper[4846]: I1005 08:23:51.550697 4846 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5098a05b-cc49-4af9-b3cb-7d6ff17f38f2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 08:23:51 crc kubenswrapper[4846]: I1005 08:23:51.550707 4846 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5098a05b-cc49-4af9-b3cb-7d6ff17f38f2-scripts\") on node \"crc\" DevicePath \"\"" Oct 05 08:23:51 crc kubenswrapper[4846]: I1005 08:23:51.580863 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 05 08:23:51 crc kubenswrapper[4846]: I1005 08:23:51.651844 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7dd897db-45ea-491c-ace4-9e14fd8104dc-logs\") pod \"7dd897db-45ea-491c-ace4-9e14fd8104dc\" (UID: \"7dd897db-45ea-491c-ace4-9e14fd8104dc\") " Oct 05 08:23:51 crc kubenswrapper[4846]: I1005 08:23:51.651903 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7dd897db-45ea-491c-ace4-9e14fd8104dc-combined-ca-bundle\") pod \"7dd897db-45ea-491c-ace4-9e14fd8104dc\" (UID: \"7dd897db-45ea-491c-ace4-9e14fd8104dc\") " Oct 05 08:23:51 crc kubenswrapper[4846]: I1005 08:23:51.651946 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gkhvb\" (UniqueName: \"kubernetes.io/projected/7dd897db-45ea-491c-ace4-9e14fd8104dc-kube-api-access-gkhvb\") pod \"7dd897db-45ea-491c-ace4-9e14fd8104dc\" (UID: \"7dd897db-45ea-491c-ace4-9e14fd8104dc\") " Oct 05 08:23:51 crc kubenswrapper[4846]: I1005 08:23:51.652002 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7dd897db-45ea-491c-ace4-9e14fd8104dc-config-data\") pod \"7dd897db-45ea-491c-ace4-9e14fd8104dc\" (UID: \"7dd897db-45ea-491c-ace4-9e14fd8104dc\") " Oct 05 08:23:51 crc kubenswrapper[4846]: I1005 08:23:51.652347 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7dd897db-45ea-491c-ace4-9e14fd8104dc-logs" (OuterVolumeSpecName: "logs") pod "7dd897db-45ea-491c-ace4-9e14fd8104dc" (UID: "7dd897db-45ea-491c-ace4-9e14fd8104dc"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 08:23:51 crc kubenswrapper[4846]: I1005 08:23:51.655022 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7dd897db-45ea-491c-ace4-9e14fd8104dc-kube-api-access-gkhvb" (OuterVolumeSpecName: "kube-api-access-gkhvb") pod "7dd897db-45ea-491c-ace4-9e14fd8104dc" (UID: "7dd897db-45ea-491c-ace4-9e14fd8104dc"). InnerVolumeSpecName "kube-api-access-gkhvb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:23:51 crc kubenswrapper[4846]: I1005 08:23:51.676336 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7dd897db-45ea-491c-ace4-9e14fd8104dc-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7dd897db-45ea-491c-ace4-9e14fd8104dc" (UID: "7dd897db-45ea-491c-ace4-9e14fd8104dc"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:23:51 crc kubenswrapper[4846]: I1005 08:23:51.678270 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7dd897db-45ea-491c-ace4-9e14fd8104dc-config-data" (OuterVolumeSpecName: "config-data") pod "7dd897db-45ea-491c-ace4-9e14fd8104dc" (UID: "7dd897db-45ea-491c-ace4-9e14fd8104dc"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:23:51 crc kubenswrapper[4846]: I1005 08:23:51.755339 4846 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7dd897db-45ea-491c-ace4-9e14fd8104dc-logs\") on node \"crc\" DevicePath \"\"" Oct 05 08:23:51 crc kubenswrapper[4846]: I1005 08:23:51.755389 4846 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7dd897db-45ea-491c-ace4-9e14fd8104dc-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 08:23:51 crc kubenswrapper[4846]: I1005 08:23:51.755402 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gkhvb\" (UniqueName: \"kubernetes.io/projected/7dd897db-45ea-491c-ace4-9e14fd8104dc-kube-api-access-gkhvb\") on node \"crc\" DevicePath \"\"" Oct 05 08:23:51 crc kubenswrapper[4846]: I1005 08:23:51.755413 4846 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7dd897db-45ea-491c-ace4-9e14fd8104dc-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 08:23:51 crc kubenswrapper[4846]: I1005 08:23:51.914684 4846 generic.go:334] "Generic (PLEG): container finished" podID="7dd897db-45ea-491c-ace4-9e14fd8104dc" containerID="a9b2b33f2bcfd4bcf52586879c9e4a052e012f883cf43a7a91236b907b130f38" exitCode=0 Oct 05 08:23:51 crc kubenswrapper[4846]: I1005 08:23:51.914729 4846 generic.go:334] "Generic (PLEG): container finished" podID="7dd897db-45ea-491c-ace4-9e14fd8104dc" containerID="55def4cf6a9df17558b786604340d80a748c419979728eed68a3e3a1093b52d0" exitCode=143 Oct 05 08:23:51 crc kubenswrapper[4846]: I1005 08:23:51.914769 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 05 08:23:51 crc kubenswrapper[4846]: I1005 08:23:51.914768 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"7dd897db-45ea-491c-ace4-9e14fd8104dc","Type":"ContainerDied","Data":"a9b2b33f2bcfd4bcf52586879c9e4a052e012f883cf43a7a91236b907b130f38"} Oct 05 08:23:51 crc kubenswrapper[4846]: I1005 08:23:51.914908 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"7dd897db-45ea-491c-ace4-9e14fd8104dc","Type":"ContainerDied","Data":"55def4cf6a9df17558b786604340d80a748c419979728eed68a3e3a1093b52d0"} Oct 05 08:23:51 crc kubenswrapper[4846]: I1005 08:23:51.914928 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"7dd897db-45ea-491c-ace4-9e14fd8104dc","Type":"ContainerDied","Data":"fc84a8f57da4592101f0d6afaa1cbb65bc48560c9ef1796fedf75a030ef1d362"} Oct 05 08:23:51 crc kubenswrapper[4846]: I1005 08:23:51.914948 4846 scope.go:117] "RemoveContainer" containerID="a9b2b33f2bcfd4bcf52586879c9e4a052e012f883cf43a7a91236b907b130f38" Oct 05 08:23:51 crc kubenswrapper[4846]: I1005 08:23:51.917365 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-p9qj9" event={"ID":"5098a05b-cc49-4af9-b3cb-7d6ff17f38f2","Type":"ContainerDied","Data":"70a502663dc211eb6cf627ca08d54240bc973046a9b917021c909a6011772ee6"} Oct 05 08:23:51 crc kubenswrapper[4846]: I1005 08:23:51.917411 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="70a502663dc211eb6cf627ca08d54240bc973046a9b917021c909a6011772ee6" Oct 05 08:23:51 crc kubenswrapper[4846]: I1005 08:23:51.917438 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-p9qj9" Oct 05 08:23:51 crc kubenswrapper[4846]: I1005 08:23:51.932522 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 05 08:23:51 crc kubenswrapper[4846]: E1005 08:23:51.932949 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7dd897db-45ea-491c-ace4-9e14fd8104dc" containerName="nova-api-api" Oct 05 08:23:51 crc kubenswrapper[4846]: I1005 08:23:51.932971 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="7dd897db-45ea-491c-ace4-9e14fd8104dc" containerName="nova-api-api" Oct 05 08:23:51 crc kubenswrapper[4846]: E1005 08:23:51.932992 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7dd897db-45ea-491c-ace4-9e14fd8104dc" containerName="nova-api-log" Oct 05 08:23:51 crc kubenswrapper[4846]: I1005 08:23:51.933002 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="7dd897db-45ea-491c-ace4-9e14fd8104dc" containerName="nova-api-log" Oct 05 08:23:51 crc kubenswrapper[4846]: E1005 08:23:51.933036 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5098a05b-cc49-4af9-b3cb-7d6ff17f38f2" containerName="nova-cell1-conductor-db-sync" Oct 05 08:23:51 crc kubenswrapper[4846]: I1005 08:23:51.933046 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="5098a05b-cc49-4af9-b3cb-7d6ff17f38f2" containerName="nova-cell1-conductor-db-sync" Oct 05 08:23:51 crc kubenswrapper[4846]: E1005 08:23:51.933072 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="285fade3-a17c-4cfc-937c-096ebdb15848" containerName="nova-manage" Oct 05 08:23:51 crc kubenswrapper[4846]: I1005 08:23:51.933079 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="285fade3-a17c-4cfc-937c-096ebdb15848" containerName="nova-manage" Oct 05 08:23:51 crc kubenswrapper[4846]: I1005 08:23:51.933260 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="5098a05b-cc49-4af9-b3cb-7d6ff17f38f2" containerName="nova-cell1-conductor-db-sync" Oct 05 08:23:51 crc kubenswrapper[4846]: I1005 08:23:51.933273 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="7dd897db-45ea-491c-ace4-9e14fd8104dc" containerName="nova-api-api" Oct 05 08:23:51 crc kubenswrapper[4846]: I1005 08:23:51.933288 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="7dd897db-45ea-491c-ace4-9e14fd8104dc" containerName="nova-api-log" Oct 05 08:23:51 crc kubenswrapper[4846]: I1005 08:23:51.933310 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="285fade3-a17c-4cfc-937c-096ebdb15848" containerName="nova-manage" Oct 05 08:23:51 crc kubenswrapper[4846]: I1005 08:23:51.933978 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Oct 05 08:23:51 crc kubenswrapper[4846]: I1005 08:23:51.935798 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Oct 05 08:23:51 crc kubenswrapper[4846]: I1005 08:23:51.945583 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 05 08:23:51 crc kubenswrapper[4846]: I1005 08:23:51.965420 4846 scope.go:117] "RemoveContainer" containerID="55def4cf6a9df17558b786604340d80a748c419979728eed68a3e3a1093b52d0" Oct 05 08:23:51 crc kubenswrapper[4846]: I1005 08:23:51.986585 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 05 08:23:51 crc kubenswrapper[4846]: I1005 08:23:51.997541 4846 scope.go:117] "RemoveContainer" containerID="a9b2b33f2bcfd4bcf52586879c9e4a052e012f883cf43a7a91236b907b130f38" Oct 05 08:23:52 crc kubenswrapper[4846]: I1005 08:23:51.999207 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Oct 05 08:23:52 crc kubenswrapper[4846]: E1005 08:23:51.999247 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a9b2b33f2bcfd4bcf52586879c9e4a052e012f883cf43a7a91236b907b130f38\": container with ID starting with a9b2b33f2bcfd4bcf52586879c9e4a052e012f883cf43a7a91236b907b130f38 not found: ID does not exist" containerID="a9b2b33f2bcfd4bcf52586879c9e4a052e012f883cf43a7a91236b907b130f38" Oct 05 08:23:52 crc kubenswrapper[4846]: I1005 08:23:51.999281 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a9b2b33f2bcfd4bcf52586879c9e4a052e012f883cf43a7a91236b907b130f38"} err="failed to get container status \"a9b2b33f2bcfd4bcf52586879c9e4a052e012f883cf43a7a91236b907b130f38\": rpc error: code = NotFound desc = could not find container \"a9b2b33f2bcfd4bcf52586879c9e4a052e012f883cf43a7a91236b907b130f38\": container with ID starting with a9b2b33f2bcfd4bcf52586879c9e4a052e012f883cf43a7a91236b907b130f38 not found: ID does not exist" Oct 05 08:23:52 crc kubenswrapper[4846]: I1005 08:23:51.999308 4846 scope.go:117] "RemoveContainer" containerID="55def4cf6a9df17558b786604340d80a748c419979728eed68a3e3a1093b52d0" Oct 05 08:23:52 crc kubenswrapper[4846]: E1005 08:23:51.999567 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"55def4cf6a9df17558b786604340d80a748c419979728eed68a3e3a1093b52d0\": container with ID starting with 55def4cf6a9df17558b786604340d80a748c419979728eed68a3e3a1093b52d0 not found: ID does not exist" containerID="55def4cf6a9df17558b786604340d80a748c419979728eed68a3e3a1093b52d0" Oct 05 08:23:52 crc kubenswrapper[4846]: I1005 08:23:51.999589 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"55def4cf6a9df17558b786604340d80a748c419979728eed68a3e3a1093b52d0"} err="failed to get container status \"55def4cf6a9df17558b786604340d80a748c419979728eed68a3e3a1093b52d0\": rpc error: code = NotFound desc = could not find container \"55def4cf6a9df17558b786604340d80a748c419979728eed68a3e3a1093b52d0\": container with ID starting with 55def4cf6a9df17558b786604340d80a748c419979728eed68a3e3a1093b52d0 not found: ID does not exist" Oct 05 08:23:52 crc kubenswrapper[4846]: I1005 08:23:51.999608 4846 scope.go:117] "RemoveContainer" containerID="a9b2b33f2bcfd4bcf52586879c9e4a052e012f883cf43a7a91236b907b130f38" Oct 05 08:23:52 crc kubenswrapper[4846]: I1005 08:23:51.999820 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a9b2b33f2bcfd4bcf52586879c9e4a052e012f883cf43a7a91236b907b130f38"} err="failed to get container status \"a9b2b33f2bcfd4bcf52586879c9e4a052e012f883cf43a7a91236b907b130f38\": rpc error: code = NotFound desc = could not find container \"a9b2b33f2bcfd4bcf52586879c9e4a052e012f883cf43a7a91236b907b130f38\": container with ID starting with a9b2b33f2bcfd4bcf52586879c9e4a052e012f883cf43a7a91236b907b130f38 not found: ID does not exist" Oct 05 08:23:52 crc kubenswrapper[4846]: I1005 08:23:51.999838 4846 scope.go:117] "RemoveContainer" containerID="55def4cf6a9df17558b786604340d80a748c419979728eed68a3e3a1093b52d0" Oct 05 08:23:52 crc kubenswrapper[4846]: I1005 08:23:52.000057 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"55def4cf6a9df17558b786604340d80a748c419979728eed68a3e3a1093b52d0"} err="failed to get container status \"55def4cf6a9df17558b786604340d80a748c419979728eed68a3e3a1093b52d0\": rpc error: code = NotFound desc = could not find container \"55def4cf6a9df17558b786604340d80a748c419979728eed68a3e3a1093b52d0\": container with ID starting with 55def4cf6a9df17558b786604340d80a748c419979728eed68a3e3a1093b52d0 not found: ID does not exist" Oct 05 08:23:52 crc kubenswrapper[4846]: I1005 08:23:52.014334 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 05 08:23:52 crc kubenswrapper[4846]: I1005 08:23:52.015969 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 05 08:23:52 crc kubenswrapper[4846]: I1005 08:23:52.018150 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 05 08:23:52 crc kubenswrapper[4846]: I1005 08:23:52.038640 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 05 08:23:52 crc kubenswrapper[4846]: I1005 08:23:52.064818 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/188d67ae-21a9-43c8-a33f-44084cbdc3d8-logs\") pod \"nova-api-0\" (UID: \"188d67ae-21a9-43c8-a33f-44084cbdc3d8\") " pod="openstack/nova-api-0" Oct 05 08:23:52 crc kubenswrapper[4846]: I1005 08:23:52.064916 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4950f233-520f-43d7-bba1-6695f817b029-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"4950f233-520f-43d7-bba1-6695f817b029\") " pod="openstack/nova-cell1-conductor-0" Oct 05 08:23:52 crc kubenswrapper[4846]: I1005 08:23:52.065599 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/188d67ae-21a9-43c8-a33f-44084cbdc3d8-config-data\") pod \"nova-api-0\" (UID: \"188d67ae-21a9-43c8-a33f-44084cbdc3d8\") " pod="openstack/nova-api-0" Oct 05 08:23:52 crc kubenswrapper[4846]: I1005 08:23:52.065730 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/188d67ae-21a9-43c8-a33f-44084cbdc3d8-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"188d67ae-21a9-43c8-a33f-44084cbdc3d8\") " pod="openstack/nova-api-0" Oct 05 08:23:52 crc kubenswrapper[4846]: I1005 08:23:52.065793 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4950f233-520f-43d7-bba1-6695f817b029-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"4950f233-520f-43d7-bba1-6695f817b029\") " pod="openstack/nova-cell1-conductor-0" Oct 05 08:23:52 crc kubenswrapper[4846]: I1005 08:23:52.065894 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rmftc\" (UniqueName: \"kubernetes.io/projected/188d67ae-21a9-43c8-a33f-44084cbdc3d8-kube-api-access-rmftc\") pod \"nova-api-0\" (UID: \"188d67ae-21a9-43c8-a33f-44084cbdc3d8\") " pod="openstack/nova-api-0" Oct 05 08:23:52 crc kubenswrapper[4846]: I1005 08:23:52.065978 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7ng45\" (UniqueName: \"kubernetes.io/projected/4950f233-520f-43d7-bba1-6695f817b029-kube-api-access-7ng45\") pod \"nova-cell1-conductor-0\" (UID: \"4950f233-520f-43d7-bba1-6695f817b029\") " pod="openstack/nova-cell1-conductor-0" Oct 05 08:23:52 crc kubenswrapper[4846]: I1005 08:23:52.168341 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/188d67ae-21a9-43c8-a33f-44084cbdc3d8-config-data\") pod \"nova-api-0\" (UID: \"188d67ae-21a9-43c8-a33f-44084cbdc3d8\") " pod="openstack/nova-api-0" Oct 05 08:23:52 crc kubenswrapper[4846]: I1005 08:23:52.169396 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/188d67ae-21a9-43c8-a33f-44084cbdc3d8-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"188d67ae-21a9-43c8-a33f-44084cbdc3d8\") " pod="openstack/nova-api-0" Oct 05 08:23:52 crc kubenswrapper[4846]: I1005 08:23:52.169458 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4950f233-520f-43d7-bba1-6695f817b029-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"4950f233-520f-43d7-bba1-6695f817b029\") " pod="openstack/nova-cell1-conductor-0" Oct 05 08:23:52 crc kubenswrapper[4846]: I1005 08:23:52.169548 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rmftc\" (UniqueName: \"kubernetes.io/projected/188d67ae-21a9-43c8-a33f-44084cbdc3d8-kube-api-access-rmftc\") pod \"nova-api-0\" (UID: \"188d67ae-21a9-43c8-a33f-44084cbdc3d8\") " pod="openstack/nova-api-0" Oct 05 08:23:52 crc kubenswrapper[4846]: I1005 08:23:52.169597 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7ng45\" (UniqueName: \"kubernetes.io/projected/4950f233-520f-43d7-bba1-6695f817b029-kube-api-access-7ng45\") pod \"nova-cell1-conductor-0\" (UID: \"4950f233-520f-43d7-bba1-6695f817b029\") " pod="openstack/nova-cell1-conductor-0" Oct 05 08:23:52 crc kubenswrapper[4846]: I1005 08:23:52.169828 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/188d67ae-21a9-43c8-a33f-44084cbdc3d8-logs\") pod \"nova-api-0\" (UID: \"188d67ae-21a9-43c8-a33f-44084cbdc3d8\") " pod="openstack/nova-api-0" Oct 05 08:23:52 crc kubenswrapper[4846]: I1005 08:23:52.169991 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4950f233-520f-43d7-bba1-6695f817b029-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"4950f233-520f-43d7-bba1-6695f817b029\") " pod="openstack/nova-cell1-conductor-0" Oct 05 08:23:52 crc kubenswrapper[4846]: I1005 08:23:52.170000 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Oct 05 08:23:52 crc kubenswrapper[4846]: I1005 08:23:52.170844 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/188d67ae-21a9-43c8-a33f-44084cbdc3d8-logs\") pod \"nova-api-0\" (UID: \"188d67ae-21a9-43c8-a33f-44084cbdc3d8\") " pod="openstack/nova-api-0" Oct 05 08:23:52 crc kubenswrapper[4846]: I1005 08:23:52.176830 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4950f233-520f-43d7-bba1-6695f817b029-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"4950f233-520f-43d7-bba1-6695f817b029\") " pod="openstack/nova-cell1-conductor-0" Oct 05 08:23:52 crc kubenswrapper[4846]: I1005 08:23:52.176917 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/188d67ae-21a9-43c8-a33f-44084cbdc3d8-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"188d67ae-21a9-43c8-a33f-44084cbdc3d8\") " pod="openstack/nova-api-0" Oct 05 08:23:52 crc kubenswrapper[4846]: I1005 08:23:52.177067 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/188d67ae-21a9-43c8-a33f-44084cbdc3d8-config-data\") pod \"nova-api-0\" (UID: \"188d67ae-21a9-43c8-a33f-44084cbdc3d8\") " pod="openstack/nova-api-0" Oct 05 08:23:52 crc kubenswrapper[4846]: I1005 08:23:52.177428 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4950f233-520f-43d7-bba1-6695f817b029-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"4950f233-520f-43d7-bba1-6695f817b029\") " pod="openstack/nova-cell1-conductor-0" Oct 05 08:23:52 crc kubenswrapper[4846]: I1005 08:23:52.190053 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7ng45\" (UniqueName: \"kubernetes.io/projected/4950f233-520f-43d7-bba1-6695f817b029-kube-api-access-7ng45\") pod \"nova-cell1-conductor-0\" (UID: \"4950f233-520f-43d7-bba1-6695f817b029\") " pod="openstack/nova-cell1-conductor-0" Oct 05 08:23:52 crc kubenswrapper[4846]: I1005 08:23:52.190402 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rmftc\" (UniqueName: \"kubernetes.io/projected/188d67ae-21a9-43c8-a33f-44084cbdc3d8-kube-api-access-rmftc\") pod \"nova-api-0\" (UID: \"188d67ae-21a9-43c8-a33f-44084cbdc3d8\") " pod="openstack/nova-api-0" Oct 05 08:23:52 crc kubenswrapper[4846]: I1005 08:23:52.224105 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Oct 05 08:23:52 crc kubenswrapper[4846]: I1005 08:23:52.272739 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Oct 05 08:23:52 crc kubenswrapper[4846]: I1005 08:23:52.343565 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 05 08:23:52 crc kubenswrapper[4846]: I1005 08:23:52.493470 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-774686bfb5-gk9ch" Oct 05 08:23:52 crc kubenswrapper[4846]: I1005 08:23:52.501366 4846 scope.go:117] "RemoveContainer" containerID="fdafa27da24296f9087c33b6b12a6598d2a8d61fc985b06ad680401828c9feb6" Oct 05 08:23:52 crc kubenswrapper[4846]: E1005 08:23:52.501650 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 08:23:52 crc kubenswrapper[4846]: I1005 08:23:52.519616 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7dd897db-45ea-491c-ace4-9e14fd8104dc" path="/var/lib/kubelet/pods/7dd897db-45ea-491c-ace4-9e14fd8104dc/volumes" Oct 05 08:23:52 crc kubenswrapper[4846]: I1005 08:23:52.563452 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-77bbb5966f-sbz7q"] Oct 05 08:23:52 crc kubenswrapper[4846]: I1005 08:23:52.563723 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-77bbb5966f-sbz7q" podUID="b051b1a4-96d9-442e-964e-b8d08275caa8" containerName="dnsmasq-dns" containerID="cri-o://e82e7b7fc0cc2840f78eca7217d9dac879df1726b4e407d90ddf0068193d427d" gracePeriod=10 Oct 05 08:23:52 crc kubenswrapper[4846]: I1005 08:23:52.738541 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 05 08:23:52 crc kubenswrapper[4846]: W1005 08:23:52.743535 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4950f233_520f_43d7_bba1_6695f817b029.slice/crio-70fc77f261d8108bb745ceb205f4867a778c0f57ffaea865933e28a3ed7a491f WatchSource:0}: Error finding container 70fc77f261d8108bb745ceb205f4867a778c0f57ffaea865933e28a3ed7a491f: Status 404 returned error can't find the container with id 70fc77f261d8108bb745ceb205f4867a778c0f57ffaea865933e28a3ed7a491f Oct 05 08:23:52 crc kubenswrapper[4846]: I1005 08:23:52.902241 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 05 08:23:52 crc kubenswrapper[4846]: W1005 08:23:52.906689 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod188d67ae_21a9_43c8_a33f_44084cbdc3d8.slice/crio-e0dc1f1ba68b0cb985b9b59098683e590693e2e6b8b14fbfdf79ffe6b4894fc9 WatchSource:0}: Error finding container e0dc1f1ba68b0cb985b9b59098683e590693e2e6b8b14fbfdf79ffe6b4894fc9: Status 404 returned error can't find the container with id e0dc1f1ba68b0cb985b9b59098683e590693e2e6b8b14fbfdf79ffe6b4894fc9 Oct 05 08:23:52 crc kubenswrapper[4846]: I1005 08:23:52.937847 4846 generic.go:334] "Generic (PLEG): container finished" podID="b051b1a4-96d9-442e-964e-b8d08275caa8" containerID="e82e7b7fc0cc2840f78eca7217d9dac879df1726b4e407d90ddf0068193d427d" exitCode=0 Oct 05 08:23:52 crc kubenswrapper[4846]: I1005 08:23:52.937947 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77bbb5966f-sbz7q" event={"ID":"b051b1a4-96d9-442e-964e-b8d08275caa8","Type":"ContainerDied","Data":"e82e7b7fc0cc2840f78eca7217d9dac879df1726b4e407d90ddf0068193d427d"} Oct 05 08:23:52 crc kubenswrapper[4846]: I1005 08:23:52.940774 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"4950f233-520f-43d7-bba1-6695f817b029","Type":"ContainerStarted","Data":"55353d85045e1540708f470739a6fef7985bc1caeacccdc149c6f115c318e396"} Oct 05 08:23:52 crc kubenswrapper[4846]: I1005 08:23:52.940811 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"4950f233-520f-43d7-bba1-6695f817b029","Type":"ContainerStarted","Data":"70fc77f261d8108bb745ceb205f4867a778c0f57ffaea865933e28a3ed7a491f"} Oct 05 08:23:52 crc kubenswrapper[4846]: I1005 08:23:52.942365 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Oct 05 08:23:52 crc kubenswrapper[4846]: I1005 08:23:52.945431 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="6b299eca-df14-4479-90a1-41f096335519" containerName="nova-metadata-log" containerID="cri-o://dd5f26ecb5ebeacdc0dfdf9dee15f21dafa24300d2187a2cb65da7c58895d7b7" gracePeriod=30 Oct 05 08:23:52 crc kubenswrapper[4846]: I1005 08:23:52.945859 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"188d67ae-21a9-43c8-a33f-44084cbdc3d8","Type":"ContainerStarted","Data":"e0dc1f1ba68b0cb985b9b59098683e590693e2e6b8b14fbfdf79ffe6b4894fc9"} Oct 05 08:23:52 crc kubenswrapper[4846]: I1005 08:23:52.945934 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="6b299eca-df14-4479-90a1-41f096335519" containerName="nova-metadata-metadata" containerID="cri-o://bd9218496be534953fa5dc544c2ec447b71d71f3bd829bb398620a64311688aa" gracePeriod=30 Oct 05 08:23:52 crc kubenswrapper[4846]: I1005 08:23:52.963297 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=1.963254542 podStartE2EDuration="1.963254542s" podCreationTimestamp="2025-10-05 08:23:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 08:23:52.962155372 +0000 UTC m=+5755.203008147" watchObservedRunningTime="2025-10-05 08:23:52.963254542 +0000 UTC m=+5755.204107317" Oct 05 08:23:53 crc kubenswrapper[4846]: I1005 08:23:53.063463 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-77bbb5966f-sbz7q" Oct 05 08:23:53 crc kubenswrapper[4846]: I1005 08:23:53.204623 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b051b1a4-96d9-442e-964e-b8d08275caa8-ovsdbserver-sb\") pod \"b051b1a4-96d9-442e-964e-b8d08275caa8\" (UID: \"b051b1a4-96d9-442e-964e-b8d08275caa8\") " Oct 05 08:23:53 crc kubenswrapper[4846]: I1005 08:23:53.204678 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b051b1a4-96d9-442e-964e-b8d08275caa8-config\") pod \"b051b1a4-96d9-442e-964e-b8d08275caa8\" (UID: \"b051b1a4-96d9-442e-964e-b8d08275caa8\") " Oct 05 08:23:53 crc kubenswrapper[4846]: I1005 08:23:53.204705 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b051b1a4-96d9-442e-964e-b8d08275caa8-dns-svc\") pod \"b051b1a4-96d9-442e-964e-b8d08275caa8\" (UID: \"b051b1a4-96d9-442e-964e-b8d08275caa8\") " Oct 05 08:23:53 crc kubenswrapper[4846]: I1005 08:23:53.204751 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nd7ts\" (UniqueName: \"kubernetes.io/projected/b051b1a4-96d9-442e-964e-b8d08275caa8-kube-api-access-nd7ts\") pod \"b051b1a4-96d9-442e-964e-b8d08275caa8\" (UID: \"b051b1a4-96d9-442e-964e-b8d08275caa8\") " Oct 05 08:23:53 crc kubenswrapper[4846]: I1005 08:23:53.204825 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b051b1a4-96d9-442e-964e-b8d08275caa8-ovsdbserver-nb\") pod \"b051b1a4-96d9-442e-964e-b8d08275caa8\" (UID: \"b051b1a4-96d9-442e-964e-b8d08275caa8\") " Oct 05 08:23:53 crc kubenswrapper[4846]: I1005 08:23:53.209219 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b051b1a4-96d9-442e-964e-b8d08275caa8-kube-api-access-nd7ts" (OuterVolumeSpecName: "kube-api-access-nd7ts") pod "b051b1a4-96d9-442e-964e-b8d08275caa8" (UID: "b051b1a4-96d9-442e-964e-b8d08275caa8"). InnerVolumeSpecName "kube-api-access-nd7ts". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:23:53 crc kubenswrapper[4846]: I1005 08:23:53.266621 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b051b1a4-96d9-442e-964e-b8d08275caa8-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "b051b1a4-96d9-442e-964e-b8d08275caa8" (UID: "b051b1a4-96d9-442e-964e-b8d08275caa8"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 08:23:53 crc kubenswrapper[4846]: I1005 08:23:53.267172 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b051b1a4-96d9-442e-964e-b8d08275caa8-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "b051b1a4-96d9-442e-964e-b8d08275caa8" (UID: "b051b1a4-96d9-442e-964e-b8d08275caa8"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 08:23:53 crc kubenswrapper[4846]: I1005 08:23:53.271050 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b051b1a4-96d9-442e-964e-b8d08275caa8-config" (OuterVolumeSpecName: "config") pod "b051b1a4-96d9-442e-964e-b8d08275caa8" (UID: "b051b1a4-96d9-442e-964e-b8d08275caa8"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 08:23:53 crc kubenswrapper[4846]: I1005 08:23:53.271843 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b051b1a4-96d9-442e-964e-b8d08275caa8-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "b051b1a4-96d9-442e-964e-b8d08275caa8" (UID: "b051b1a4-96d9-442e-964e-b8d08275caa8"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 08:23:53 crc kubenswrapper[4846]: I1005 08:23:53.307379 4846 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b051b1a4-96d9-442e-964e-b8d08275caa8-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 05 08:23:53 crc kubenswrapper[4846]: I1005 08:23:53.307415 4846 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b051b1a4-96d9-442e-964e-b8d08275caa8-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 05 08:23:53 crc kubenswrapper[4846]: I1005 08:23:53.307427 4846 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b051b1a4-96d9-442e-964e-b8d08275caa8-config\") on node \"crc\" DevicePath \"\"" Oct 05 08:23:53 crc kubenswrapper[4846]: I1005 08:23:53.307438 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nd7ts\" (UniqueName: \"kubernetes.io/projected/b051b1a4-96d9-442e-964e-b8d08275caa8-kube-api-access-nd7ts\") on node \"crc\" DevicePath \"\"" Oct 05 08:23:53 crc kubenswrapper[4846]: I1005 08:23:53.307450 4846 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b051b1a4-96d9-442e-964e-b8d08275caa8-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 05 08:23:53 crc kubenswrapper[4846]: I1005 08:23:53.532065 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 05 08:23:53 crc kubenswrapper[4846]: I1005 08:23:53.613257 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6b299eca-df14-4479-90a1-41f096335519-logs\") pod \"6b299eca-df14-4479-90a1-41f096335519\" (UID: \"6b299eca-df14-4479-90a1-41f096335519\") " Oct 05 08:23:53 crc kubenswrapper[4846]: I1005 08:23:53.613343 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/6b299eca-df14-4479-90a1-41f096335519-nova-metadata-tls-certs\") pod \"6b299eca-df14-4479-90a1-41f096335519\" (UID: \"6b299eca-df14-4479-90a1-41f096335519\") " Oct 05 08:23:53 crc kubenswrapper[4846]: I1005 08:23:53.613444 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sf4rd\" (UniqueName: \"kubernetes.io/projected/6b299eca-df14-4479-90a1-41f096335519-kube-api-access-sf4rd\") pod \"6b299eca-df14-4479-90a1-41f096335519\" (UID: \"6b299eca-df14-4479-90a1-41f096335519\") " Oct 05 08:23:53 crc kubenswrapper[4846]: I1005 08:23:53.613466 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b299eca-df14-4479-90a1-41f096335519-combined-ca-bundle\") pod \"6b299eca-df14-4479-90a1-41f096335519\" (UID: \"6b299eca-df14-4479-90a1-41f096335519\") " Oct 05 08:23:53 crc kubenswrapper[4846]: I1005 08:23:53.613490 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6b299eca-df14-4479-90a1-41f096335519-config-data\") pod \"6b299eca-df14-4479-90a1-41f096335519\" (UID: \"6b299eca-df14-4479-90a1-41f096335519\") " Oct 05 08:23:53 crc kubenswrapper[4846]: I1005 08:23:53.613641 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6b299eca-df14-4479-90a1-41f096335519-logs" (OuterVolumeSpecName: "logs") pod "6b299eca-df14-4479-90a1-41f096335519" (UID: "6b299eca-df14-4479-90a1-41f096335519"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 08:23:53 crc kubenswrapper[4846]: I1005 08:23:53.614782 4846 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6b299eca-df14-4479-90a1-41f096335519-logs\") on node \"crc\" DevicePath \"\"" Oct 05 08:23:53 crc kubenswrapper[4846]: I1005 08:23:53.617123 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6b299eca-df14-4479-90a1-41f096335519-kube-api-access-sf4rd" (OuterVolumeSpecName: "kube-api-access-sf4rd") pod "6b299eca-df14-4479-90a1-41f096335519" (UID: "6b299eca-df14-4479-90a1-41f096335519"). InnerVolumeSpecName "kube-api-access-sf4rd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:23:53 crc kubenswrapper[4846]: I1005 08:23:53.635614 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6b299eca-df14-4479-90a1-41f096335519-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6b299eca-df14-4479-90a1-41f096335519" (UID: "6b299eca-df14-4479-90a1-41f096335519"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:23:53 crc kubenswrapper[4846]: I1005 08:23:53.636073 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6b299eca-df14-4479-90a1-41f096335519-config-data" (OuterVolumeSpecName: "config-data") pod "6b299eca-df14-4479-90a1-41f096335519" (UID: "6b299eca-df14-4479-90a1-41f096335519"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:23:53 crc kubenswrapper[4846]: I1005 08:23:53.661426 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6b299eca-df14-4479-90a1-41f096335519-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "6b299eca-df14-4479-90a1-41f096335519" (UID: "6b299eca-df14-4479-90a1-41f096335519"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:23:53 crc kubenswrapper[4846]: I1005 08:23:53.716505 4846 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/6b299eca-df14-4479-90a1-41f096335519-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 05 08:23:53 crc kubenswrapper[4846]: I1005 08:23:53.716570 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sf4rd\" (UniqueName: \"kubernetes.io/projected/6b299eca-df14-4479-90a1-41f096335519-kube-api-access-sf4rd\") on node \"crc\" DevicePath \"\"" Oct 05 08:23:53 crc kubenswrapper[4846]: I1005 08:23:53.716594 4846 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b299eca-df14-4479-90a1-41f096335519-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 08:23:53 crc kubenswrapper[4846]: I1005 08:23:53.716613 4846 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6b299eca-df14-4479-90a1-41f096335519-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 08:23:53 crc kubenswrapper[4846]: I1005 08:23:53.960117 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"188d67ae-21a9-43c8-a33f-44084cbdc3d8","Type":"ContainerStarted","Data":"006e3a4d809a2526a062908765859aa3087dc6d5e6f8d19c45d2849dc9615a2f"} Oct 05 08:23:53 crc kubenswrapper[4846]: I1005 08:23:53.960176 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"188d67ae-21a9-43c8-a33f-44084cbdc3d8","Type":"ContainerStarted","Data":"5c4be27156878d0b803aae47db8a307662fe8dd8ff93e8dd174a04950cc25827"} Oct 05 08:23:53 crc kubenswrapper[4846]: I1005 08:23:53.964686 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-77bbb5966f-sbz7q" Oct 05 08:23:53 crc kubenswrapper[4846]: I1005 08:23:53.965848 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77bbb5966f-sbz7q" event={"ID":"b051b1a4-96d9-442e-964e-b8d08275caa8","Type":"ContainerDied","Data":"7aa5f13a54200698ef8ab86faf1817aa7f433aa16e6a8a846eb2a272900fede0"} Oct 05 08:23:53 crc kubenswrapper[4846]: I1005 08:23:53.965948 4846 scope.go:117] "RemoveContainer" containerID="e82e7b7fc0cc2840f78eca7217d9dac879df1726b4e407d90ddf0068193d427d" Oct 05 08:23:53 crc kubenswrapper[4846]: I1005 08:23:53.980858 4846 generic.go:334] "Generic (PLEG): container finished" podID="6b299eca-df14-4479-90a1-41f096335519" containerID="bd9218496be534953fa5dc544c2ec447b71d71f3bd829bb398620a64311688aa" exitCode=0 Oct 05 08:23:53 crc kubenswrapper[4846]: I1005 08:23:53.980923 4846 generic.go:334] "Generic (PLEG): container finished" podID="6b299eca-df14-4479-90a1-41f096335519" containerID="dd5f26ecb5ebeacdc0dfdf9dee15f21dafa24300d2187a2cb65da7c58895d7b7" exitCode=143 Oct 05 08:23:53 crc kubenswrapper[4846]: I1005 08:23:53.980964 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 05 08:23:53 crc kubenswrapper[4846]: I1005 08:23:53.981063 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"6b299eca-df14-4479-90a1-41f096335519","Type":"ContainerDied","Data":"bd9218496be534953fa5dc544c2ec447b71d71f3bd829bb398620a64311688aa"} Oct 05 08:23:53 crc kubenswrapper[4846]: I1005 08:23:53.981136 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"6b299eca-df14-4479-90a1-41f096335519","Type":"ContainerDied","Data":"dd5f26ecb5ebeacdc0dfdf9dee15f21dafa24300d2187a2cb65da7c58895d7b7"} Oct 05 08:23:53 crc kubenswrapper[4846]: I1005 08:23:53.981169 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"6b299eca-df14-4479-90a1-41f096335519","Type":"ContainerDied","Data":"8aae06db8d81d026c500e08a3b4a727ae019e4b4bf609e42eb162bac0c5afc2a"} Oct 05 08:23:53 crc kubenswrapper[4846]: I1005 08:23:53.992382 4846 scope.go:117] "RemoveContainer" containerID="c9c740790070d806894467efe2d16ac8d06520c7277b300d876b021acda86422" Oct 05 08:23:53 crc kubenswrapper[4846]: I1005 08:23:53.994088 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.994076845 podStartE2EDuration="2.994076845s" podCreationTimestamp="2025-10-05 08:23:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 08:23:53.992997437 +0000 UTC m=+5756.233850222" watchObservedRunningTime="2025-10-05 08:23:53.994076845 +0000 UTC m=+5756.234929620" Oct 05 08:23:54 crc kubenswrapper[4846]: I1005 08:23:54.039312 4846 scope.go:117] "RemoveContainer" containerID="bd9218496be534953fa5dc544c2ec447b71d71f3bd829bb398620a64311688aa" Oct 05 08:23:54 crc kubenswrapper[4846]: I1005 08:23:54.063398 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-77bbb5966f-sbz7q"] Oct 05 08:23:54 crc kubenswrapper[4846]: I1005 08:23:54.085384 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-77bbb5966f-sbz7q"] Oct 05 08:23:54 crc kubenswrapper[4846]: I1005 08:23:54.089549 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 05 08:23:54 crc kubenswrapper[4846]: I1005 08:23:54.095374 4846 scope.go:117] "RemoveContainer" containerID="dd5f26ecb5ebeacdc0dfdf9dee15f21dafa24300d2187a2cb65da7c58895d7b7" Oct 05 08:23:54 crc kubenswrapper[4846]: I1005 08:23:54.102825 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Oct 05 08:23:54 crc kubenswrapper[4846]: I1005 08:23:54.111148 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 05 08:23:54 crc kubenswrapper[4846]: E1005 08:23:54.111619 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b051b1a4-96d9-442e-964e-b8d08275caa8" containerName="init" Oct 05 08:23:54 crc kubenswrapper[4846]: I1005 08:23:54.111642 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="b051b1a4-96d9-442e-964e-b8d08275caa8" containerName="init" Oct 05 08:23:54 crc kubenswrapper[4846]: E1005 08:23:54.111660 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b051b1a4-96d9-442e-964e-b8d08275caa8" containerName="dnsmasq-dns" Oct 05 08:23:54 crc kubenswrapper[4846]: I1005 08:23:54.111667 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="b051b1a4-96d9-442e-964e-b8d08275caa8" containerName="dnsmasq-dns" Oct 05 08:23:54 crc kubenswrapper[4846]: E1005 08:23:54.111686 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6b299eca-df14-4479-90a1-41f096335519" containerName="nova-metadata-log" Oct 05 08:23:54 crc kubenswrapper[4846]: I1005 08:23:54.111693 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="6b299eca-df14-4479-90a1-41f096335519" containerName="nova-metadata-log" Oct 05 08:23:54 crc kubenswrapper[4846]: E1005 08:23:54.111717 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6b299eca-df14-4479-90a1-41f096335519" containerName="nova-metadata-metadata" Oct 05 08:23:54 crc kubenswrapper[4846]: I1005 08:23:54.111723 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="6b299eca-df14-4479-90a1-41f096335519" containerName="nova-metadata-metadata" Oct 05 08:23:54 crc kubenswrapper[4846]: I1005 08:23:54.111889 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="b051b1a4-96d9-442e-964e-b8d08275caa8" containerName="dnsmasq-dns" Oct 05 08:23:54 crc kubenswrapper[4846]: I1005 08:23:54.111903 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="6b299eca-df14-4479-90a1-41f096335519" containerName="nova-metadata-log" Oct 05 08:23:54 crc kubenswrapper[4846]: I1005 08:23:54.111910 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="6b299eca-df14-4479-90a1-41f096335519" containerName="nova-metadata-metadata" Oct 05 08:23:54 crc kubenswrapper[4846]: I1005 08:23:54.112885 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 05 08:23:54 crc kubenswrapper[4846]: I1005 08:23:54.114040 4846 scope.go:117] "RemoveContainer" containerID="bd9218496be534953fa5dc544c2ec447b71d71f3bd829bb398620a64311688aa" Oct 05 08:23:54 crc kubenswrapper[4846]: I1005 08:23:54.114529 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 05 08:23:54 crc kubenswrapper[4846]: I1005 08:23:54.114716 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Oct 05 08:23:54 crc kubenswrapper[4846]: E1005 08:23:54.114742 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bd9218496be534953fa5dc544c2ec447b71d71f3bd829bb398620a64311688aa\": container with ID starting with bd9218496be534953fa5dc544c2ec447b71d71f3bd829bb398620a64311688aa not found: ID does not exist" containerID="bd9218496be534953fa5dc544c2ec447b71d71f3bd829bb398620a64311688aa" Oct 05 08:23:54 crc kubenswrapper[4846]: I1005 08:23:54.114802 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bd9218496be534953fa5dc544c2ec447b71d71f3bd829bb398620a64311688aa"} err="failed to get container status \"bd9218496be534953fa5dc544c2ec447b71d71f3bd829bb398620a64311688aa\": rpc error: code = NotFound desc = could not find container \"bd9218496be534953fa5dc544c2ec447b71d71f3bd829bb398620a64311688aa\": container with ID starting with bd9218496be534953fa5dc544c2ec447b71d71f3bd829bb398620a64311688aa not found: ID does not exist" Oct 05 08:23:54 crc kubenswrapper[4846]: I1005 08:23:54.114827 4846 scope.go:117] "RemoveContainer" containerID="dd5f26ecb5ebeacdc0dfdf9dee15f21dafa24300d2187a2cb65da7c58895d7b7" Oct 05 08:23:54 crc kubenswrapper[4846]: E1005 08:23:54.115343 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dd5f26ecb5ebeacdc0dfdf9dee15f21dafa24300d2187a2cb65da7c58895d7b7\": container with ID starting with dd5f26ecb5ebeacdc0dfdf9dee15f21dafa24300d2187a2cb65da7c58895d7b7 not found: ID does not exist" containerID="dd5f26ecb5ebeacdc0dfdf9dee15f21dafa24300d2187a2cb65da7c58895d7b7" Oct 05 08:23:54 crc kubenswrapper[4846]: I1005 08:23:54.115406 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dd5f26ecb5ebeacdc0dfdf9dee15f21dafa24300d2187a2cb65da7c58895d7b7"} err="failed to get container status \"dd5f26ecb5ebeacdc0dfdf9dee15f21dafa24300d2187a2cb65da7c58895d7b7\": rpc error: code = NotFound desc = could not find container \"dd5f26ecb5ebeacdc0dfdf9dee15f21dafa24300d2187a2cb65da7c58895d7b7\": container with ID starting with dd5f26ecb5ebeacdc0dfdf9dee15f21dafa24300d2187a2cb65da7c58895d7b7 not found: ID does not exist" Oct 05 08:23:54 crc kubenswrapper[4846]: I1005 08:23:54.115425 4846 scope.go:117] "RemoveContainer" containerID="bd9218496be534953fa5dc544c2ec447b71d71f3bd829bb398620a64311688aa" Oct 05 08:23:54 crc kubenswrapper[4846]: I1005 08:23:54.118657 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 05 08:23:54 crc kubenswrapper[4846]: I1005 08:23:54.118860 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bd9218496be534953fa5dc544c2ec447b71d71f3bd829bb398620a64311688aa"} err="failed to get container status \"bd9218496be534953fa5dc544c2ec447b71d71f3bd829bb398620a64311688aa\": rpc error: code = NotFound desc = could not find container \"bd9218496be534953fa5dc544c2ec447b71d71f3bd829bb398620a64311688aa\": container with ID starting with bd9218496be534953fa5dc544c2ec447b71d71f3bd829bb398620a64311688aa not found: ID does not exist" Oct 05 08:23:54 crc kubenswrapper[4846]: I1005 08:23:54.118895 4846 scope.go:117] "RemoveContainer" containerID="dd5f26ecb5ebeacdc0dfdf9dee15f21dafa24300d2187a2cb65da7c58895d7b7" Oct 05 08:23:54 crc kubenswrapper[4846]: I1005 08:23:54.119143 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dd5f26ecb5ebeacdc0dfdf9dee15f21dafa24300d2187a2cb65da7c58895d7b7"} err="failed to get container status \"dd5f26ecb5ebeacdc0dfdf9dee15f21dafa24300d2187a2cb65da7c58895d7b7\": rpc error: code = NotFound desc = could not find container \"dd5f26ecb5ebeacdc0dfdf9dee15f21dafa24300d2187a2cb65da7c58895d7b7\": container with ID starting with dd5f26ecb5ebeacdc0dfdf9dee15f21dafa24300d2187a2cb65da7c58895d7b7 not found: ID does not exist" Oct 05 08:23:54 crc kubenswrapper[4846]: I1005 08:23:54.238156 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8d409b1e-a91c-49c0-96f2-3df8b6679199-logs\") pod \"nova-metadata-0\" (UID: \"8d409b1e-a91c-49c0-96f2-3df8b6679199\") " pod="openstack/nova-metadata-0" Oct 05 08:23:54 crc kubenswrapper[4846]: I1005 08:23:54.238250 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8d409b1e-a91c-49c0-96f2-3df8b6679199-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"8d409b1e-a91c-49c0-96f2-3df8b6679199\") " pod="openstack/nova-metadata-0" Oct 05 08:23:54 crc kubenswrapper[4846]: I1005 08:23:54.238433 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/8d409b1e-a91c-49c0-96f2-3df8b6679199-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"8d409b1e-a91c-49c0-96f2-3df8b6679199\") " pod="openstack/nova-metadata-0" Oct 05 08:23:54 crc kubenswrapper[4846]: I1005 08:23:54.238506 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8d409b1e-a91c-49c0-96f2-3df8b6679199-config-data\") pod \"nova-metadata-0\" (UID: \"8d409b1e-a91c-49c0-96f2-3df8b6679199\") " pod="openstack/nova-metadata-0" Oct 05 08:23:54 crc kubenswrapper[4846]: I1005 08:23:54.238531 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vdnc4\" (UniqueName: \"kubernetes.io/projected/8d409b1e-a91c-49c0-96f2-3df8b6679199-kube-api-access-vdnc4\") pod \"nova-metadata-0\" (UID: \"8d409b1e-a91c-49c0-96f2-3df8b6679199\") " pod="openstack/nova-metadata-0" Oct 05 08:23:54 crc kubenswrapper[4846]: I1005 08:23:54.340201 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8d409b1e-a91c-49c0-96f2-3df8b6679199-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"8d409b1e-a91c-49c0-96f2-3df8b6679199\") " pod="openstack/nova-metadata-0" Oct 05 08:23:54 crc kubenswrapper[4846]: I1005 08:23:54.340533 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/8d409b1e-a91c-49c0-96f2-3df8b6679199-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"8d409b1e-a91c-49c0-96f2-3df8b6679199\") " pod="openstack/nova-metadata-0" Oct 05 08:23:54 crc kubenswrapper[4846]: I1005 08:23:54.340593 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8d409b1e-a91c-49c0-96f2-3df8b6679199-config-data\") pod \"nova-metadata-0\" (UID: \"8d409b1e-a91c-49c0-96f2-3df8b6679199\") " pod="openstack/nova-metadata-0" Oct 05 08:23:54 crc kubenswrapper[4846]: I1005 08:23:54.340660 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vdnc4\" (UniqueName: \"kubernetes.io/projected/8d409b1e-a91c-49c0-96f2-3df8b6679199-kube-api-access-vdnc4\") pod \"nova-metadata-0\" (UID: \"8d409b1e-a91c-49c0-96f2-3df8b6679199\") " pod="openstack/nova-metadata-0" Oct 05 08:23:54 crc kubenswrapper[4846]: I1005 08:23:54.340845 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8d409b1e-a91c-49c0-96f2-3df8b6679199-logs\") pod \"nova-metadata-0\" (UID: \"8d409b1e-a91c-49c0-96f2-3df8b6679199\") " pod="openstack/nova-metadata-0" Oct 05 08:23:54 crc kubenswrapper[4846]: I1005 08:23:54.341808 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8d409b1e-a91c-49c0-96f2-3df8b6679199-logs\") pod \"nova-metadata-0\" (UID: \"8d409b1e-a91c-49c0-96f2-3df8b6679199\") " pod="openstack/nova-metadata-0" Oct 05 08:23:54 crc kubenswrapper[4846]: I1005 08:23:54.348456 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8d409b1e-a91c-49c0-96f2-3df8b6679199-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"8d409b1e-a91c-49c0-96f2-3df8b6679199\") " pod="openstack/nova-metadata-0" Oct 05 08:23:54 crc kubenswrapper[4846]: I1005 08:23:54.349003 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8d409b1e-a91c-49c0-96f2-3df8b6679199-config-data\") pod \"nova-metadata-0\" (UID: \"8d409b1e-a91c-49c0-96f2-3df8b6679199\") " pod="openstack/nova-metadata-0" Oct 05 08:23:54 crc kubenswrapper[4846]: I1005 08:23:54.358452 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/8d409b1e-a91c-49c0-96f2-3df8b6679199-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"8d409b1e-a91c-49c0-96f2-3df8b6679199\") " pod="openstack/nova-metadata-0" Oct 05 08:23:54 crc kubenswrapper[4846]: I1005 08:23:54.362982 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vdnc4\" (UniqueName: \"kubernetes.io/projected/8d409b1e-a91c-49c0-96f2-3df8b6679199-kube-api-access-vdnc4\") pod \"nova-metadata-0\" (UID: \"8d409b1e-a91c-49c0-96f2-3df8b6679199\") " pod="openstack/nova-metadata-0" Oct 05 08:23:54 crc kubenswrapper[4846]: I1005 08:23:54.428800 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 05 08:23:54 crc kubenswrapper[4846]: I1005 08:23:54.520463 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6b299eca-df14-4479-90a1-41f096335519" path="/var/lib/kubelet/pods/6b299eca-df14-4479-90a1-41f096335519/volumes" Oct 05 08:23:54 crc kubenswrapper[4846]: I1005 08:23:54.521441 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b051b1a4-96d9-442e-964e-b8d08275caa8" path="/var/lib/kubelet/pods/b051b1a4-96d9-442e-964e-b8d08275caa8/volumes" Oct 05 08:23:54 crc kubenswrapper[4846]: I1005 08:23:54.840951 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 05 08:23:54 crc kubenswrapper[4846]: I1005 08:23:54.999067 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"8d409b1e-a91c-49c0-96f2-3df8b6679199","Type":"ContainerStarted","Data":"4ae0ff76b3396458eac202d4dd90f675b7f8b27f03adb01bafff3609184319b6"} Oct 05 08:23:56 crc kubenswrapper[4846]: I1005 08:23:56.018945 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"8d409b1e-a91c-49c0-96f2-3df8b6679199","Type":"ContainerStarted","Data":"63fe74b1e29ffea6657ebecf1c49c55957af6382c195a72b7cc62916280cf64d"} Oct 05 08:23:56 crc kubenswrapper[4846]: I1005 08:23:56.020282 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"8d409b1e-a91c-49c0-96f2-3df8b6679199","Type":"ContainerStarted","Data":"bbb7fcf74c7b4ccfb8c58ccda5d6ab76c1a1670b40113965e1dd4965a295a344"} Oct 05 08:23:56 crc kubenswrapper[4846]: I1005 08:23:56.053139 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.053111243 podStartE2EDuration="2.053111243s" podCreationTimestamp="2025-10-05 08:23:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 08:23:56.045341447 +0000 UTC m=+5758.286194252" watchObservedRunningTime="2025-10-05 08:23:56.053111243 +0000 UTC m=+5758.293964038" Oct 05 08:23:57 crc kubenswrapper[4846]: I1005 08:23:57.329871 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Oct 05 08:23:59 crc kubenswrapper[4846]: I1005 08:23:59.429302 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 05 08:23:59 crc kubenswrapper[4846]: I1005 08:23:59.430425 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 05 08:24:02 crc kubenswrapper[4846]: I1005 08:24:02.345242 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 05 08:24:02 crc kubenswrapper[4846]: I1005 08:24:02.345922 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 05 08:24:03 crc kubenswrapper[4846]: I1005 08:24:03.428425 4846 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="188d67ae-21a9-43c8-a33f-44084cbdc3d8" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.1.90:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 05 08:24:03 crc kubenswrapper[4846]: I1005 08:24:03.428618 4846 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="188d67ae-21a9-43c8-a33f-44084cbdc3d8" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.1.90:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 05 08:24:04 crc kubenswrapper[4846]: I1005 08:24:04.429485 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 05 08:24:04 crc kubenswrapper[4846]: I1005 08:24:04.429777 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 05 08:24:05 crc kubenswrapper[4846]: I1005 08:24:05.444345 4846 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="8d409b1e-a91c-49c0-96f2-3df8b6679199" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.1.91:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 05 08:24:05 crc kubenswrapper[4846]: I1005 08:24:05.444407 4846 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="8d409b1e-a91c-49c0-96f2-3df8b6679199" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.1.91:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 05 08:24:07 crc kubenswrapper[4846]: I1005 08:24:07.498753 4846 scope.go:117] "RemoveContainer" containerID="fdafa27da24296f9087c33b6b12a6598d2a8d61fc985b06ad680401828c9feb6" Oct 05 08:24:07 crc kubenswrapper[4846]: E1005 08:24:07.499624 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 08:24:13 crc kubenswrapper[4846]: I1005 08:24:13.386417 4846 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="188d67ae-21a9-43c8-a33f-44084cbdc3d8" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.1.90:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 05 08:24:13 crc kubenswrapper[4846]: I1005 08:24:13.427477 4846 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="188d67ae-21a9-43c8-a33f-44084cbdc3d8" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.1.90:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 05 08:24:15 crc kubenswrapper[4846]: I1005 08:24:15.439400 4846 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="8d409b1e-a91c-49c0-96f2-3df8b6679199" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.1.91:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 05 08:24:15 crc kubenswrapper[4846]: I1005 08:24:15.439428 4846 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="8d409b1e-a91c-49c0-96f2-3df8b6679199" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.1.91:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 05 08:24:18 crc kubenswrapper[4846]: I1005 08:24:18.256378 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 05 08:24:18 crc kubenswrapper[4846]: I1005 08:24:18.292568 4846 generic.go:334] "Generic (PLEG): container finished" podID="fec088ae-74f0-42bb-8225-428c7562f4c0" containerID="7da5d1791e7d851f68e870544b931c04cce1e5bdfa1fb2d9e4c7e9c64e5d6b7a" exitCode=137 Oct 05 08:24:18 crc kubenswrapper[4846]: I1005 08:24:18.292953 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"fec088ae-74f0-42bb-8225-428c7562f4c0","Type":"ContainerDied","Data":"7da5d1791e7d851f68e870544b931c04cce1e5bdfa1fb2d9e4c7e9c64e5d6b7a"} Oct 05 08:24:18 crc kubenswrapper[4846]: I1005 08:24:18.293112 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"fec088ae-74f0-42bb-8225-428c7562f4c0","Type":"ContainerDied","Data":"ab4c57ed069012578f10a551343b61429804dda258dda111020b3dcb40073360"} Oct 05 08:24:18 crc kubenswrapper[4846]: I1005 08:24:18.293255 4846 scope.go:117] "RemoveContainer" containerID="7da5d1791e7d851f68e870544b931c04cce1e5bdfa1fb2d9e4c7e9c64e5d6b7a" Oct 05 08:24:18 crc kubenswrapper[4846]: I1005 08:24:18.293427 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 05 08:24:18 crc kubenswrapper[4846]: I1005 08:24:18.316355 4846 scope.go:117] "RemoveContainer" containerID="7da5d1791e7d851f68e870544b931c04cce1e5bdfa1fb2d9e4c7e9c64e5d6b7a" Oct 05 08:24:18 crc kubenswrapper[4846]: E1005 08:24:18.317599 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7da5d1791e7d851f68e870544b931c04cce1e5bdfa1fb2d9e4c7e9c64e5d6b7a\": container with ID starting with 7da5d1791e7d851f68e870544b931c04cce1e5bdfa1fb2d9e4c7e9c64e5d6b7a not found: ID does not exist" containerID="7da5d1791e7d851f68e870544b931c04cce1e5bdfa1fb2d9e4c7e9c64e5d6b7a" Oct 05 08:24:18 crc kubenswrapper[4846]: I1005 08:24:18.317635 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7da5d1791e7d851f68e870544b931c04cce1e5bdfa1fb2d9e4c7e9c64e5d6b7a"} err="failed to get container status \"7da5d1791e7d851f68e870544b931c04cce1e5bdfa1fb2d9e4c7e9c64e5d6b7a\": rpc error: code = NotFound desc = could not find container \"7da5d1791e7d851f68e870544b931c04cce1e5bdfa1fb2d9e4c7e9c64e5d6b7a\": container with ID starting with 7da5d1791e7d851f68e870544b931c04cce1e5bdfa1fb2d9e4c7e9c64e5d6b7a not found: ID does not exist" Oct 05 08:24:18 crc kubenswrapper[4846]: I1005 08:24:18.353738 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8jhmk\" (UniqueName: \"kubernetes.io/projected/fec088ae-74f0-42bb-8225-428c7562f4c0-kube-api-access-8jhmk\") pod \"fec088ae-74f0-42bb-8225-428c7562f4c0\" (UID: \"fec088ae-74f0-42bb-8225-428c7562f4c0\") " Oct 05 08:24:18 crc kubenswrapper[4846]: I1005 08:24:18.353904 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fec088ae-74f0-42bb-8225-428c7562f4c0-config-data\") pod \"fec088ae-74f0-42bb-8225-428c7562f4c0\" (UID: \"fec088ae-74f0-42bb-8225-428c7562f4c0\") " Oct 05 08:24:18 crc kubenswrapper[4846]: I1005 08:24:18.353980 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fec088ae-74f0-42bb-8225-428c7562f4c0-combined-ca-bundle\") pod \"fec088ae-74f0-42bb-8225-428c7562f4c0\" (UID: \"fec088ae-74f0-42bb-8225-428c7562f4c0\") " Oct 05 08:24:18 crc kubenswrapper[4846]: I1005 08:24:18.370056 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fec088ae-74f0-42bb-8225-428c7562f4c0-kube-api-access-8jhmk" (OuterVolumeSpecName: "kube-api-access-8jhmk") pod "fec088ae-74f0-42bb-8225-428c7562f4c0" (UID: "fec088ae-74f0-42bb-8225-428c7562f4c0"). InnerVolumeSpecName "kube-api-access-8jhmk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:24:18 crc kubenswrapper[4846]: I1005 08:24:18.385402 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fec088ae-74f0-42bb-8225-428c7562f4c0-config-data" (OuterVolumeSpecName: "config-data") pod "fec088ae-74f0-42bb-8225-428c7562f4c0" (UID: "fec088ae-74f0-42bb-8225-428c7562f4c0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:24:18 crc kubenswrapper[4846]: I1005 08:24:18.387852 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fec088ae-74f0-42bb-8225-428c7562f4c0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "fec088ae-74f0-42bb-8225-428c7562f4c0" (UID: "fec088ae-74f0-42bb-8225-428c7562f4c0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:24:18 crc kubenswrapper[4846]: I1005 08:24:18.455965 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8jhmk\" (UniqueName: \"kubernetes.io/projected/fec088ae-74f0-42bb-8225-428c7562f4c0-kube-api-access-8jhmk\") on node \"crc\" DevicePath \"\"" Oct 05 08:24:18 crc kubenswrapper[4846]: I1005 08:24:18.456005 4846 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fec088ae-74f0-42bb-8225-428c7562f4c0-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 08:24:18 crc kubenswrapper[4846]: I1005 08:24:18.456016 4846 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fec088ae-74f0-42bb-8225-428c7562f4c0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 08:24:18 crc kubenswrapper[4846]: I1005 08:24:18.625657 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 05 08:24:18 crc kubenswrapper[4846]: I1005 08:24:18.635436 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 05 08:24:18 crc kubenswrapper[4846]: I1005 08:24:18.654917 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 05 08:24:18 crc kubenswrapper[4846]: E1005 08:24:18.655366 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fec088ae-74f0-42bb-8225-428c7562f4c0" containerName="nova-cell1-novncproxy-novncproxy" Oct 05 08:24:18 crc kubenswrapper[4846]: I1005 08:24:18.655387 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="fec088ae-74f0-42bb-8225-428c7562f4c0" containerName="nova-cell1-novncproxy-novncproxy" Oct 05 08:24:18 crc kubenswrapper[4846]: I1005 08:24:18.655576 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="fec088ae-74f0-42bb-8225-428c7562f4c0" containerName="nova-cell1-novncproxy-novncproxy" Oct 05 08:24:18 crc kubenswrapper[4846]: I1005 08:24:18.656263 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 05 08:24:18 crc kubenswrapper[4846]: I1005 08:24:18.659792 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-public-svc" Oct 05 08:24:18 crc kubenswrapper[4846]: I1005 08:24:18.659984 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-vencrypt" Oct 05 08:24:18 crc kubenswrapper[4846]: I1005 08:24:18.660009 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Oct 05 08:24:18 crc kubenswrapper[4846]: I1005 08:24:18.667020 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 05 08:24:18 crc kubenswrapper[4846]: I1005 08:24:18.761867 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/78d644fc-c8ab-4b0f-bdf4-772bdca9adf3-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"78d644fc-c8ab-4b0f-bdf4-772bdca9adf3\") " pod="openstack/nova-cell1-novncproxy-0" Oct 05 08:24:18 crc kubenswrapper[4846]: I1005 08:24:18.761925 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/78d644fc-c8ab-4b0f-bdf4-772bdca9adf3-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"78d644fc-c8ab-4b0f-bdf4-772bdca9adf3\") " pod="openstack/nova-cell1-novncproxy-0" Oct 05 08:24:18 crc kubenswrapper[4846]: I1005 08:24:18.761950 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/78d644fc-c8ab-4b0f-bdf4-772bdca9adf3-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"78d644fc-c8ab-4b0f-bdf4-772bdca9adf3\") " pod="openstack/nova-cell1-novncproxy-0" Oct 05 08:24:18 crc kubenswrapper[4846]: I1005 08:24:18.762057 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nnsq4\" (UniqueName: \"kubernetes.io/projected/78d644fc-c8ab-4b0f-bdf4-772bdca9adf3-kube-api-access-nnsq4\") pod \"nova-cell1-novncproxy-0\" (UID: \"78d644fc-c8ab-4b0f-bdf4-772bdca9adf3\") " pod="openstack/nova-cell1-novncproxy-0" Oct 05 08:24:18 crc kubenswrapper[4846]: I1005 08:24:18.762136 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/78d644fc-c8ab-4b0f-bdf4-772bdca9adf3-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"78d644fc-c8ab-4b0f-bdf4-772bdca9adf3\") " pod="openstack/nova-cell1-novncproxy-0" Oct 05 08:24:18 crc kubenswrapper[4846]: I1005 08:24:18.863629 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/78d644fc-c8ab-4b0f-bdf4-772bdca9adf3-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"78d644fc-c8ab-4b0f-bdf4-772bdca9adf3\") " pod="openstack/nova-cell1-novncproxy-0" Oct 05 08:24:18 crc kubenswrapper[4846]: I1005 08:24:18.863747 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/78d644fc-c8ab-4b0f-bdf4-772bdca9adf3-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"78d644fc-c8ab-4b0f-bdf4-772bdca9adf3\") " pod="openstack/nova-cell1-novncproxy-0" Oct 05 08:24:18 crc kubenswrapper[4846]: I1005 08:24:18.863783 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/78d644fc-c8ab-4b0f-bdf4-772bdca9adf3-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"78d644fc-c8ab-4b0f-bdf4-772bdca9adf3\") " pod="openstack/nova-cell1-novncproxy-0" Oct 05 08:24:18 crc kubenswrapper[4846]: I1005 08:24:18.864605 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/78d644fc-c8ab-4b0f-bdf4-772bdca9adf3-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"78d644fc-c8ab-4b0f-bdf4-772bdca9adf3\") " pod="openstack/nova-cell1-novncproxy-0" Oct 05 08:24:18 crc kubenswrapper[4846]: I1005 08:24:18.864665 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nnsq4\" (UniqueName: \"kubernetes.io/projected/78d644fc-c8ab-4b0f-bdf4-772bdca9adf3-kube-api-access-nnsq4\") pod \"nova-cell1-novncproxy-0\" (UID: \"78d644fc-c8ab-4b0f-bdf4-772bdca9adf3\") " pod="openstack/nova-cell1-novncproxy-0" Oct 05 08:24:18 crc kubenswrapper[4846]: I1005 08:24:18.867587 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/78d644fc-c8ab-4b0f-bdf4-772bdca9adf3-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"78d644fc-c8ab-4b0f-bdf4-772bdca9adf3\") " pod="openstack/nova-cell1-novncproxy-0" Oct 05 08:24:18 crc kubenswrapper[4846]: I1005 08:24:18.868096 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/78d644fc-c8ab-4b0f-bdf4-772bdca9adf3-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"78d644fc-c8ab-4b0f-bdf4-772bdca9adf3\") " pod="openstack/nova-cell1-novncproxy-0" Oct 05 08:24:18 crc kubenswrapper[4846]: I1005 08:24:18.870652 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/78d644fc-c8ab-4b0f-bdf4-772bdca9adf3-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"78d644fc-c8ab-4b0f-bdf4-772bdca9adf3\") " pod="openstack/nova-cell1-novncproxy-0" Oct 05 08:24:18 crc kubenswrapper[4846]: I1005 08:24:18.872631 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/78d644fc-c8ab-4b0f-bdf4-772bdca9adf3-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"78d644fc-c8ab-4b0f-bdf4-772bdca9adf3\") " pod="openstack/nova-cell1-novncproxy-0" Oct 05 08:24:18 crc kubenswrapper[4846]: I1005 08:24:18.897495 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nnsq4\" (UniqueName: \"kubernetes.io/projected/78d644fc-c8ab-4b0f-bdf4-772bdca9adf3-kube-api-access-nnsq4\") pod \"nova-cell1-novncproxy-0\" (UID: \"78d644fc-c8ab-4b0f-bdf4-772bdca9adf3\") " pod="openstack/nova-cell1-novncproxy-0" Oct 05 08:24:18 crc kubenswrapper[4846]: I1005 08:24:18.980172 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 05 08:24:19 crc kubenswrapper[4846]: I1005 08:24:19.495091 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 05 08:24:19 crc kubenswrapper[4846]: I1005 08:24:19.497968 4846 scope.go:117] "RemoveContainer" containerID="fdafa27da24296f9087c33b6b12a6598d2a8d61fc985b06ad680401828c9feb6" Oct 05 08:24:19 crc kubenswrapper[4846]: E1005 08:24:19.498282 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 08:24:19 crc kubenswrapper[4846]: W1005 08:24:19.513131 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod78d644fc_c8ab_4b0f_bdf4_772bdca9adf3.slice/crio-0be791d594d06208cc66849c19aefe87bef79f3e6eeebe7b0cfbf1a3bde4d214 WatchSource:0}: Error finding container 0be791d594d06208cc66849c19aefe87bef79f3e6eeebe7b0cfbf1a3bde4d214: Status 404 returned error can't find the container with id 0be791d594d06208cc66849c19aefe87bef79f3e6eeebe7b0cfbf1a3bde4d214 Oct 05 08:24:20 crc kubenswrapper[4846]: I1005 08:24:20.313343 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"78d644fc-c8ab-4b0f-bdf4-772bdca9adf3","Type":"ContainerStarted","Data":"d47b6e318e6731e9b3e6fb387ebef5d14e1c30c172bd17141712bb3de315d254"} Oct 05 08:24:20 crc kubenswrapper[4846]: I1005 08:24:20.313628 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"78d644fc-c8ab-4b0f-bdf4-772bdca9adf3","Type":"ContainerStarted","Data":"0be791d594d06208cc66849c19aefe87bef79f3e6eeebe7b0cfbf1a3bde4d214"} Oct 05 08:24:20 crc kubenswrapper[4846]: I1005 08:24:20.369694 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.369674659 podStartE2EDuration="2.369674659s" podCreationTimestamp="2025-10-05 08:24:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 08:24:20.354523816 +0000 UTC m=+5782.595376591" watchObservedRunningTime="2025-10-05 08:24:20.369674659 +0000 UTC m=+5782.610527434" Oct 05 08:24:20 crc kubenswrapper[4846]: I1005 08:24:20.511376 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fec088ae-74f0-42bb-8225-428c7562f4c0" path="/var/lib/kubelet/pods/fec088ae-74f0-42bb-8225-428c7562f4c0/volumes" Oct 05 08:24:21 crc kubenswrapper[4846]: I1005 08:24:21.323119 4846 generic.go:334] "Generic (PLEG): container finished" podID="5b32c235-5879-482b-a396-6d0d7e98816d" containerID="500743e9d035ab2c1788d5f88dc1507f2bc7aa4912a9f386cf8d555dbdec9935" exitCode=137 Oct 05 08:24:21 crc kubenswrapper[4846]: I1005 08:24:21.323223 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"5b32c235-5879-482b-a396-6d0d7e98816d","Type":"ContainerDied","Data":"500743e9d035ab2c1788d5f88dc1507f2bc7aa4912a9f386cf8d555dbdec9935"} Oct 05 08:24:21 crc kubenswrapper[4846]: I1005 08:24:21.573977 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 05 08:24:21 crc kubenswrapper[4846]: I1005 08:24:21.623384 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9cxbt\" (UniqueName: \"kubernetes.io/projected/5b32c235-5879-482b-a396-6d0d7e98816d-kube-api-access-9cxbt\") pod \"5b32c235-5879-482b-a396-6d0d7e98816d\" (UID: \"5b32c235-5879-482b-a396-6d0d7e98816d\") " Oct 05 08:24:21 crc kubenswrapper[4846]: I1005 08:24:21.623475 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5b32c235-5879-482b-a396-6d0d7e98816d-combined-ca-bundle\") pod \"5b32c235-5879-482b-a396-6d0d7e98816d\" (UID: \"5b32c235-5879-482b-a396-6d0d7e98816d\") " Oct 05 08:24:21 crc kubenswrapper[4846]: I1005 08:24:21.623603 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5b32c235-5879-482b-a396-6d0d7e98816d-config-data\") pod \"5b32c235-5879-482b-a396-6d0d7e98816d\" (UID: \"5b32c235-5879-482b-a396-6d0d7e98816d\") " Oct 05 08:24:21 crc kubenswrapper[4846]: I1005 08:24:21.632895 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b32c235-5879-482b-a396-6d0d7e98816d-kube-api-access-9cxbt" (OuterVolumeSpecName: "kube-api-access-9cxbt") pod "5b32c235-5879-482b-a396-6d0d7e98816d" (UID: "5b32c235-5879-482b-a396-6d0d7e98816d"). InnerVolumeSpecName "kube-api-access-9cxbt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:24:21 crc kubenswrapper[4846]: I1005 08:24:21.658301 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b32c235-5879-482b-a396-6d0d7e98816d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5b32c235-5879-482b-a396-6d0d7e98816d" (UID: "5b32c235-5879-482b-a396-6d0d7e98816d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:24:21 crc kubenswrapper[4846]: I1005 08:24:21.665518 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b32c235-5879-482b-a396-6d0d7e98816d-config-data" (OuterVolumeSpecName: "config-data") pod "5b32c235-5879-482b-a396-6d0d7e98816d" (UID: "5b32c235-5879-482b-a396-6d0d7e98816d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:24:21 crc kubenswrapper[4846]: I1005 08:24:21.726131 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9cxbt\" (UniqueName: \"kubernetes.io/projected/5b32c235-5879-482b-a396-6d0d7e98816d-kube-api-access-9cxbt\") on node \"crc\" DevicePath \"\"" Oct 05 08:24:21 crc kubenswrapper[4846]: I1005 08:24:21.726160 4846 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5b32c235-5879-482b-a396-6d0d7e98816d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 08:24:21 crc kubenswrapper[4846]: I1005 08:24:21.726169 4846 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5b32c235-5879-482b-a396-6d0d7e98816d-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 08:24:22 crc kubenswrapper[4846]: I1005 08:24:22.336658 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"5b32c235-5879-482b-a396-6d0d7e98816d","Type":"ContainerDied","Data":"4b4e22560041b1fab9bdf5260cec89a79ca2e629ee57dbf8ed157093cccfa3c8"} Oct 05 08:24:22 crc kubenswrapper[4846]: I1005 08:24:22.336723 4846 scope.go:117] "RemoveContainer" containerID="500743e9d035ab2c1788d5f88dc1507f2bc7aa4912a9f386cf8d555dbdec9935" Oct 05 08:24:22 crc kubenswrapper[4846]: I1005 08:24:22.336748 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 05 08:24:22 crc kubenswrapper[4846]: I1005 08:24:22.344728 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 05 08:24:22 crc kubenswrapper[4846]: I1005 08:24:22.344848 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 05 08:24:22 crc kubenswrapper[4846]: I1005 08:24:22.386881 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 05 08:24:22 crc kubenswrapper[4846]: I1005 08:24:22.395891 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Oct 05 08:24:22 crc kubenswrapper[4846]: I1005 08:24:22.424973 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Oct 05 08:24:22 crc kubenswrapper[4846]: E1005 08:24:22.425745 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5b32c235-5879-482b-a396-6d0d7e98816d" containerName="nova-scheduler-scheduler" Oct 05 08:24:22 crc kubenswrapper[4846]: I1005 08:24:22.425792 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="5b32c235-5879-482b-a396-6d0d7e98816d" containerName="nova-scheduler-scheduler" Oct 05 08:24:22 crc kubenswrapper[4846]: I1005 08:24:22.426168 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="5b32c235-5879-482b-a396-6d0d7e98816d" containerName="nova-scheduler-scheduler" Oct 05 08:24:22 crc kubenswrapper[4846]: I1005 08:24:22.427713 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 05 08:24:22 crc kubenswrapper[4846]: I1005 08:24:22.442915 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 05 08:24:22 crc kubenswrapper[4846]: I1005 08:24:22.466088 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Oct 05 08:24:22 crc kubenswrapper[4846]: I1005 08:24:22.512795 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b32c235-5879-482b-a396-6d0d7e98816d" path="/var/lib/kubelet/pods/5b32c235-5879-482b-a396-6d0d7e98816d/volumes" Oct 05 08:24:22 crc kubenswrapper[4846]: I1005 08:24:22.570054 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z8mtg\" (UniqueName: \"kubernetes.io/projected/76dec6e0-c589-4072-b65f-42ee496e0c61-kube-api-access-z8mtg\") pod \"nova-scheduler-0\" (UID: \"76dec6e0-c589-4072-b65f-42ee496e0c61\") " pod="openstack/nova-scheduler-0" Oct 05 08:24:22 crc kubenswrapper[4846]: I1005 08:24:22.570229 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/76dec6e0-c589-4072-b65f-42ee496e0c61-config-data\") pod \"nova-scheduler-0\" (UID: \"76dec6e0-c589-4072-b65f-42ee496e0c61\") " pod="openstack/nova-scheduler-0" Oct 05 08:24:22 crc kubenswrapper[4846]: I1005 08:24:22.570284 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76dec6e0-c589-4072-b65f-42ee496e0c61-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"76dec6e0-c589-4072-b65f-42ee496e0c61\") " pod="openstack/nova-scheduler-0" Oct 05 08:24:22 crc kubenswrapper[4846]: I1005 08:24:22.671982 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z8mtg\" (UniqueName: \"kubernetes.io/projected/76dec6e0-c589-4072-b65f-42ee496e0c61-kube-api-access-z8mtg\") pod \"nova-scheduler-0\" (UID: \"76dec6e0-c589-4072-b65f-42ee496e0c61\") " pod="openstack/nova-scheduler-0" Oct 05 08:24:22 crc kubenswrapper[4846]: I1005 08:24:22.672062 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/76dec6e0-c589-4072-b65f-42ee496e0c61-config-data\") pod \"nova-scheduler-0\" (UID: \"76dec6e0-c589-4072-b65f-42ee496e0c61\") " pod="openstack/nova-scheduler-0" Oct 05 08:24:22 crc kubenswrapper[4846]: I1005 08:24:22.672114 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76dec6e0-c589-4072-b65f-42ee496e0c61-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"76dec6e0-c589-4072-b65f-42ee496e0c61\") " pod="openstack/nova-scheduler-0" Oct 05 08:24:22 crc kubenswrapper[4846]: I1005 08:24:22.683963 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76dec6e0-c589-4072-b65f-42ee496e0c61-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"76dec6e0-c589-4072-b65f-42ee496e0c61\") " pod="openstack/nova-scheduler-0" Oct 05 08:24:22 crc kubenswrapper[4846]: I1005 08:24:22.699440 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/76dec6e0-c589-4072-b65f-42ee496e0c61-config-data\") pod \"nova-scheduler-0\" (UID: \"76dec6e0-c589-4072-b65f-42ee496e0c61\") " pod="openstack/nova-scheduler-0" Oct 05 08:24:22 crc kubenswrapper[4846]: I1005 08:24:22.704761 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z8mtg\" (UniqueName: \"kubernetes.io/projected/76dec6e0-c589-4072-b65f-42ee496e0c61-kube-api-access-z8mtg\") pod \"nova-scheduler-0\" (UID: \"76dec6e0-c589-4072-b65f-42ee496e0c61\") " pod="openstack/nova-scheduler-0" Oct 05 08:24:22 crc kubenswrapper[4846]: I1005 08:24:22.787505 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 05 08:24:23 crc kubenswrapper[4846]: I1005 08:24:23.308551 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 05 08:24:23 crc kubenswrapper[4846]: W1005 08:24:23.322231 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod76dec6e0_c589_4072_b65f_42ee496e0c61.slice/crio-f08055fbb0d11e2641506b70081764f971a0afb10ac647ddc76c8d378de12d30 WatchSource:0}: Error finding container f08055fbb0d11e2641506b70081764f971a0afb10ac647ddc76c8d378de12d30: Status 404 returned error can't find the container with id f08055fbb0d11e2641506b70081764f971a0afb10ac647ddc76c8d378de12d30 Oct 05 08:24:23 crc kubenswrapper[4846]: I1005 08:24:23.351136 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"76dec6e0-c589-4072-b65f-42ee496e0c61","Type":"ContainerStarted","Data":"f08055fbb0d11e2641506b70081764f971a0afb10ac647ddc76c8d378de12d30"} Oct 05 08:24:23 crc kubenswrapper[4846]: I1005 08:24:23.428556 4846 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="188d67ae-21a9-43c8-a33f-44084cbdc3d8" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.1.90:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 05 08:24:23 crc kubenswrapper[4846]: I1005 08:24:23.428883 4846 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="188d67ae-21a9-43c8-a33f-44084cbdc3d8" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.1.90:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 05 08:24:23 crc kubenswrapper[4846]: I1005 08:24:23.980701 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Oct 05 08:24:24 crc kubenswrapper[4846]: I1005 08:24:24.374004 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"76dec6e0-c589-4072-b65f-42ee496e0c61","Type":"ContainerStarted","Data":"7692194d2e1f5715282a95ce181ed64987bc7fa067bb95c2b1a7ff9e6b878221"} Oct 05 08:24:24 crc kubenswrapper[4846]: I1005 08:24:24.399202 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.399153341 podStartE2EDuration="2.399153341s" podCreationTimestamp="2025-10-05 08:24:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 08:24:24.38931377 +0000 UTC m=+5786.630166585" watchObservedRunningTime="2025-10-05 08:24:24.399153341 +0000 UTC m=+5786.640006146" Oct 05 08:24:25 crc kubenswrapper[4846]: I1005 08:24:25.440381 4846 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="8d409b1e-a91c-49c0-96f2-3df8b6679199" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.1.91:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 05 08:24:25 crc kubenswrapper[4846]: I1005 08:24:25.440435 4846 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="8d409b1e-a91c-49c0-96f2-3df8b6679199" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.1.91:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 05 08:24:27 crc kubenswrapper[4846]: I1005 08:24:27.788288 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Oct 05 08:24:28 crc kubenswrapper[4846]: I1005 08:24:28.981204 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Oct 05 08:24:29 crc kubenswrapper[4846]: I1005 08:24:29.012391 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Oct 05 08:24:29 crc kubenswrapper[4846]: I1005 08:24:29.464605 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Oct 05 08:24:29 crc kubenswrapper[4846]: I1005 08:24:29.671364 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-cell-mapping-gdt57"] Oct 05 08:24:29 crc kubenswrapper[4846]: I1005 08:24:29.672981 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-gdt57" Oct 05 08:24:29 crc kubenswrapper[4846]: I1005 08:24:29.675696 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-config-data" Oct 05 08:24:29 crc kubenswrapper[4846]: I1005 08:24:29.676114 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-scripts" Oct 05 08:24:29 crc kubenswrapper[4846]: I1005 08:24:29.688475 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-gdt57"] Oct 05 08:24:29 crc kubenswrapper[4846]: I1005 08:24:29.730669 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f8888283-b66d-4993-80a0-a7b6b1bac02b-scripts\") pod \"nova-cell1-cell-mapping-gdt57\" (UID: \"f8888283-b66d-4993-80a0-a7b6b1bac02b\") " pod="openstack/nova-cell1-cell-mapping-gdt57" Oct 05 08:24:29 crc kubenswrapper[4846]: I1005 08:24:29.730739 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f8888283-b66d-4993-80a0-a7b6b1bac02b-config-data\") pod \"nova-cell1-cell-mapping-gdt57\" (UID: \"f8888283-b66d-4993-80a0-a7b6b1bac02b\") " pod="openstack/nova-cell1-cell-mapping-gdt57" Oct 05 08:24:29 crc kubenswrapper[4846]: I1005 08:24:29.730936 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f8888283-b66d-4993-80a0-a7b6b1bac02b-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-gdt57\" (UID: \"f8888283-b66d-4993-80a0-a7b6b1bac02b\") " pod="openstack/nova-cell1-cell-mapping-gdt57" Oct 05 08:24:29 crc kubenswrapper[4846]: I1005 08:24:29.730990 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2fnkk\" (UniqueName: \"kubernetes.io/projected/f8888283-b66d-4993-80a0-a7b6b1bac02b-kube-api-access-2fnkk\") pod \"nova-cell1-cell-mapping-gdt57\" (UID: \"f8888283-b66d-4993-80a0-a7b6b1bac02b\") " pod="openstack/nova-cell1-cell-mapping-gdt57" Oct 05 08:24:29 crc kubenswrapper[4846]: I1005 08:24:29.832940 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f8888283-b66d-4993-80a0-a7b6b1bac02b-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-gdt57\" (UID: \"f8888283-b66d-4993-80a0-a7b6b1bac02b\") " pod="openstack/nova-cell1-cell-mapping-gdt57" Oct 05 08:24:29 crc kubenswrapper[4846]: I1005 08:24:29.833003 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2fnkk\" (UniqueName: \"kubernetes.io/projected/f8888283-b66d-4993-80a0-a7b6b1bac02b-kube-api-access-2fnkk\") pod \"nova-cell1-cell-mapping-gdt57\" (UID: \"f8888283-b66d-4993-80a0-a7b6b1bac02b\") " pod="openstack/nova-cell1-cell-mapping-gdt57" Oct 05 08:24:29 crc kubenswrapper[4846]: I1005 08:24:29.833032 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f8888283-b66d-4993-80a0-a7b6b1bac02b-scripts\") pod \"nova-cell1-cell-mapping-gdt57\" (UID: \"f8888283-b66d-4993-80a0-a7b6b1bac02b\") " pod="openstack/nova-cell1-cell-mapping-gdt57" Oct 05 08:24:29 crc kubenswrapper[4846]: I1005 08:24:29.833060 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f8888283-b66d-4993-80a0-a7b6b1bac02b-config-data\") pod \"nova-cell1-cell-mapping-gdt57\" (UID: \"f8888283-b66d-4993-80a0-a7b6b1bac02b\") " pod="openstack/nova-cell1-cell-mapping-gdt57" Oct 05 08:24:29 crc kubenswrapper[4846]: I1005 08:24:29.839326 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f8888283-b66d-4993-80a0-a7b6b1bac02b-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-gdt57\" (UID: \"f8888283-b66d-4993-80a0-a7b6b1bac02b\") " pod="openstack/nova-cell1-cell-mapping-gdt57" Oct 05 08:24:29 crc kubenswrapper[4846]: I1005 08:24:29.839411 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f8888283-b66d-4993-80a0-a7b6b1bac02b-config-data\") pod \"nova-cell1-cell-mapping-gdt57\" (UID: \"f8888283-b66d-4993-80a0-a7b6b1bac02b\") " pod="openstack/nova-cell1-cell-mapping-gdt57" Oct 05 08:24:29 crc kubenswrapper[4846]: I1005 08:24:29.839769 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f8888283-b66d-4993-80a0-a7b6b1bac02b-scripts\") pod \"nova-cell1-cell-mapping-gdt57\" (UID: \"f8888283-b66d-4993-80a0-a7b6b1bac02b\") " pod="openstack/nova-cell1-cell-mapping-gdt57" Oct 05 08:24:29 crc kubenswrapper[4846]: I1005 08:24:29.848139 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2fnkk\" (UniqueName: \"kubernetes.io/projected/f8888283-b66d-4993-80a0-a7b6b1bac02b-kube-api-access-2fnkk\") pod \"nova-cell1-cell-mapping-gdt57\" (UID: \"f8888283-b66d-4993-80a0-a7b6b1bac02b\") " pod="openstack/nova-cell1-cell-mapping-gdt57" Oct 05 08:24:29 crc kubenswrapper[4846]: I1005 08:24:29.995684 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-gdt57" Oct 05 08:24:30 crc kubenswrapper[4846]: I1005 08:24:30.524772 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-gdt57"] Oct 05 08:24:30 crc kubenswrapper[4846]: W1005 08:24:30.529760 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf8888283_b66d_4993_80a0_a7b6b1bac02b.slice/crio-52463a38fde6374ed55979c20f9197c4d02c22a1c853a1ac0ddcc93b56768fce WatchSource:0}: Error finding container 52463a38fde6374ed55979c20f9197c4d02c22a1c853a1ac0ddcc93b56768fce: Status 404 returned error can't find the container with id 52463a38fde6374ed55979c20f9197c4d02c22a1c853a1ac0ddcc93b56768fce Oct 05 08:24:31 crc kubenswrapper[4846]: I1005 08:24:31.457532 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-gdt57" event={"ID":"f8888283-b66d-4993-80a0-a7b6b1bac02b","Type":"ContainerStarted","Data":"717a4de0219e85cadcebc33059340186dbabb2f100ca30cb91d92d6000152b08"} Oct 05 08:24:31 crc kubenswrapper[4846]: I1005 08:24:31.457911 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-gdt57" event={"ID":"f8888283-b66d-4993-80a0-a7b6b1bac02b","Type":"ContainerStarted","Data":"52463a38fde6374ed55979c20f9197c4d02c22a1c853a1ac0ddcc93b56768fce"} Oct 05 08:24:31 crc kubenswrapper[4846]: I1005 08:24:31.484812 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-cell-mapping-gdt57" podStartSLOduration=2.484783819 podStartE2EDuration="2.484783819s" podCreationTimestamp="2025-10-05 08:24:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 08:24:31.4806464 +0000 UTC m=+5793.721499175" watchObservedRunningTime="2025-10-05 08:24:31.484783819 +0000 UTC m=+5793.725636634" Oct 05 08:24:31 crc kubenswrapper[4846]: I1005 08:24:31.498088 4846 scope.go:117] "RemoveContainer" containerID="fdafa27da24296f9087c33b6b12a6598d2a8d61fc985b06ad680401828c9feb6" Oct 05 08:24:31 crc kubenswrapper[4846]: E1005 08:24:31.498632 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 08:24:32 crc kubenswrapper[4846]: I1005 08:24:32.788903 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Oct 05 08:24:32 crc kubenswrapper[4846]: I1005 08:24:32.834119 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Oct 05 08:24:32 crc kubenswrapper[4846]: I1005 08:24:32.872342 4846 scope.go:117] "RemoveContainer" containerID="beaae9624261c763dbf004d0473c97d2be8428df0aeae1a54589c73d0a02db3f" Oct 05 08:24:33 crc kubenswrapper[4846]: I1005 08:24:33.427457 4846 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="188d67ae-21a9-43c8-a33f-44084cbdc3d8" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.1.90:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 05 08:24:33 crc kubenswrapper[4846]: I1005 08:24:33.427814 4846 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="188d67ae-21a9-43c8-a33f-44084cbdc3d8" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.1.90:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 05 08:24:33 crc kubenswrapper[4846]: I1005 08:24:33.523512 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Oct 05 08:24:35 crc kubenswrapper[4846]: I1005 08:24:35.437396 4846 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="8d409b1e-a91c-49c0-96f2-3df8b6679199" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.1.91:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 05 08:24:35 crc kubenswrapper[4846]: I1005 08:24:35.438210 4846 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="8d409b1e-a91c-49c0-96f2-3df8b6679199" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.1.91:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 05 08:24:35 crc kubenswrapper[4846]: I1005 08:24:35.510868 4846 generic.go:334] "Generic (PLEG): container finished" podID="f8888283-b66d-4993-80a0-a7b6b1bac02b" containerID="717a4de0219e85cadcebc33059340186dbabb2f100ca30cb91d92d6000152b08" exitCode=0 Oct 05 08:24:35 crc kubenswrapper[4846]: I1005 08:24:35.510929 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-gdt57" event={"ID":"f8888283-b66d-4993-80a0-a7b6b1bac02b","Type":"ContainerDied","Data":"717a4de0219e85cadcebc33059340186dbabb2f100ca30cb91d92d6000152b08"} Oct 05 08:24:36 crc kubenswrapper[4846]: I1005 08:24:36.927140 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-gdt57" Oct 05 08:24:37 crc kubenswrapper[4846]: I1005 08:24:37.087142 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f8888283-b66d-4993-80a0-a7b6b1bac02b-combined-ca-bundle\") pod \"f8888283-b66d-4993-80a0-a7b6b1bac02b\" (UID: \"f8888283-b66d-4993-80a0-a7b6b1bac02b\") " Oct 05 08:24:37 crc kubenswrapper[4846]: I1005 08:24:37.087265 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f8888283-b66d-4993-80a0-a7b6b1bac02b-scripts\") pod \"f8888283-b66d-4993-80a0-a7b6b1bac02b\" (UID: \"f8888283-b66d-4993-80a0-a7b6b1bac02b\") " Oct 05 08:24:37 crc kubenswrapper[4846]: I1005 08:24:37.087962 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2fnkk\" (UniqueName: \"kubernetes.io/projected/f8888283-b66d-4993-80a0-a7b6b1bac02b-kube-api-access-2fnkk\") pod \"f8888283-b66d-4993-80a0-a7b6b1bac02b\" (UID: \"f8888283-b66d-4993-80a0-a7b6b1bac02b\") " Oct 05 08:24:37 crc kubenswrapper[4846]: I1005 08:24:37.088003 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f8888283-b66d-4993-80a0-a7b6b1bac02b-config-data\") pod \"f8888283-b66d-4993-80a0-a7b6b1bac02b\" (UID: \"f8888283-b66d-4993-80a0-a7b6b1bac02b\") " Oct 05 08:24:37 crc kubenswrapper[4846]: I1005 08:24:37.092421 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f8888283-b66d-4993-80a0-a7b6b1bac02b-scripts" (OuterVolumeSpecName: "scripts") pod "f8888283-b66d-4993-80a0-a7b6b1bac02b" (UID: "f8888283-b66d-4993-80a0-a7b6b1bac02b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:24:37 crc kubenswrapper[4846]: I1005 08:24:37.093747 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f8888283-b66d-4993-80a0-a7b6b1bac02b-kube-api-access-2fnkk" (OuterVolumeSpecName: "kube-api-access-2fnkk") pod "f8888283-b66d-4993-80a0-a7b6b1bac02b" (UID: "f8888283-b66d-4993-80a0-a7b6b1bac02b"). InnerVolumeSpecName "kube-api-access-2fnkk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:24:37 crc kubenswrapper[4846]: I1005 08:24:37.116340 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f8888283-b66d-4993-80a0-a7b6b1bac02b-config-data" (OuterVolumeSpecName: "config-data") pod "f8888283-b66d-4993-80a0-a7b6b1bac02b" (UID: "f8888283-b66d-4993-80a0-a7b6b1bac02b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:24:37 crc kubenswrapper[4846]: I1005 08:24:37.127757 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f8888283-b66d-4993-80a0-a7b6b1bac02b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f8888283-b66d-4993-80a0-a7b6b1bac02b" (UID: "f8888283-b66d-4993-80a0-a7b6b1bac02b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:24:37 crc kubenswrapper[4846]: I1005 08:24:37.190278 4846 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f8888283-b66d-4993-80a0-a7b6b1bac02b-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 08:24:37 crc kubenswrapper[4846]: I1005 08:24:37.190314 4846 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f8888283-b66d-4993-80a0-a7b6b1bac02b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 08:24:37 crc kubenswrapper[4846]: I1005 08:24:37.190330 4846 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f8888283-b66d-4993-80a0-a7b6b1bac02b-scripts\") on node \"crc\" DevicePath \"\"" Oct 05 08:24:37 crc kubenswrapper[4846]: I1005 08:24:37.190344 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2fnkk\" (UniqueName: \"kubernetes.io/projected/f8888283-b66d-4993-80a0-a7b6b1bac02b-kube-api-access-2fnkk\") on node \"crc\" DevicePath \"\"" Oct 05 08:24:37 crc kubenswrapper[4846]: I1005 08:24:37.536955 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-gdt57" event={"ID":"f8888283-b66d-4993-80a0-a7b6b1bac02b","Type":"ContainerDied","Data":"52463a38fde6374ed55979c20f9197c4d02c22a1c853a1ac0ddcc93b56768fce"} Oct 05 08:24:37 crc kubenswrapper[4846]: I1005 08:24:37.537006 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="52463a38fde6374ed55979c20f9197c4d02c22a1c853a1ac0ddcc93b56768fce" Oct 05 08:24:37 crc kubenswrapper[4846]: I1005 08:24:37.536978 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-gdt57" Oct 05 08:24:37 crc kubenswrapper[4846]: I1005 08:24:37.719549 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 05 08:24:37 crc kubenswrapper[4846]: I1005 08:24:37.719902 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="188d67ae-21a9-43c8-a33f-44084cbdc3d8" containerName="nova-api-log" containerID="cri-o://5c4be27156878d0b803aae47db8a307662fe8dd8ff93e8dd174a04950cc25827" gracePeriod=30 Oct 05 08:24:37 crc kubenswrapper[4846]: I1005 08:24:37.719990 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="188d67ae-21a9-43c8-a33f-44084cbdc3d8" containerName="nova-api-api" containerID="cri-o://006e3a4d809a2526a062908765859aa3087dc6d5e6f8d19c45d2849dc9615a2f" gracePeriod=30 Oct 05 08:24:37 crc kubenswrapper[4846]: I1005 08:24:37.726704 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 05 08:24:37 crc kubenswrapper[4846]: I1005 08:24:37.726908 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="76dec6e0-c589-4072-b65f-42ee496e0c61" containerName="nova-scheduler-scheduler" containerID="cri-o://7692194d2e1f5715282a95ce181ed64987bc7fa067bb95c2b1a7ff9e6b878221" gracePeriod=30 Oct 05 08:24:37 crc kubenswrapper[4846]: I1005 08:24:37.773276 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 05 08:24:37 crc kubenswrapper[4846]: I1005 08:24:37.773570 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="8d409b1e-a91c-49c0-96f2-3df8b6679199" containerName="nova-metadata-log" containerID="cri-o://bbb7fcf74c7b4ccfb8c58ccda5d6ab76c1a1670b40113965e1dd4965a295a344" gracePeriod=30 Oct 05 08:24:37 crc kubenswrapper[4846]: I1005 08:24:37.773675 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="8d409b1e-a91c-49c0-96f2-3df8b6679199" containerName="nova-metadata-metadata" containerID="cri-o://63fe74b1e29ffea6657ebecf1c49c55957af6382c195a72b7cc62916280cf64d" gracePeriod=30 Oct 05 08:24:37 crc kubenswrapper[4846]: E1005 08:24:37.790033 4846 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="7692194d2e1f5715282a95ce181ed64987bc7fa067bb95c2b1a7ff9e6b878221" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 05 08:24:37 crc kubenswrapper[4846]: E1005 08:24:37.791364 4846 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="7692194d2e1f5715282a95ce181ed64987bc7fa067bb95c2b1a7ff9e6b878221" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 05 08:24:37 crc kubenswrapper[4846]: E1005 08:24:37.792978 4846 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="7692194d2e1f5715282a95ce181ed64987bc7fa067bb95c2b1a7ff9e6b878221" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 05 08:24:37 crc kubenswrapper[4846]: E1005 08:24:37.793047 4846 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="76dec6e0-c589-4072-b65f-42ee496e0c61" containerName="nova-scheduler-scheduler" Oct 05 08:24:38 crc kubenswrapper[4846]: I1005 08:24:38.549002 4846 generic.go:334] "Generic (PLEG): container finished" podID="188d67ae-21a9-43c8-a33f-44084cbdc3d8" containerID="5c4be27156878d0b803aae47db8a307662fe8dd8ff93e8dd174a04950cc25827" exitCode=143 Oct 05 08:24:38 crc kubenswrapper[4846]: I1005 08:24:38.549068 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"188d67ae-21a9-43c8-a33f-44084cbdc3d8","Type":"ContainerDied","Data":"5c4be27156878d0b803aae47db8a307662fe8dd8ff93e8dd174a04950cc25827"} Oct 05 08:24:38 crc kubenswrapper[4846]: I1005 08:24:38.551951 4846 generic.go:334] "Generic (PLEG): container finished" podID="8d409b1e-a91c-49c0-96f2-3df8b6679199" containerID="bbb7fcf74c7b4ccfb8c58ccda5d6ab76c1a1670b40113965e1dd4965a295a344" exitCode=143 Oct 05 08:24:38 crc kubenswrapper[4846]: I1005 08:24:38.551986 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"8d409b1e-a91c-49c0-96f2-3df8b6679199","Type":"ContainerDied","Data":"bbb7fcf74c7b4ccfb8c58ccda5d6ab76c1a1670b40113965e1dd4965a295a344"} Oct 05 08:24:42 crc kubenswrapper[4846]: I1005 08:24:42.497009 4846 scope.go:117] "RemoveContainer" containerID="fdafa27da24296f9087c33b6b12a6598d2a8d61fc985b06ad680401828c9feb6" Oct 05 08:24:42 crc kubenswrapper[4846]: E1005 08:24:42.498618 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 08:24:42 crc kubenswrapper[4846]: E1005 08:24:42.791142 4846 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="7692194d2e1f5715282a95ce181ed64987bc7fa067bb95c2b1a7ff9e6b878221" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 05 08:24:42 crc kubenswrapper[4846]: E1005 08:24:42.793377 4846 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="7692194d2e1f5715282a95ce181ed64987bc7fa067bb95c2b1a7ff9e6b878221" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 05 08:24:42 crc kubenswrapper[4846]: E1005 08:24:42.795997 4846 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="7692194d2e1f5715282a95ce181ed64987bc7fa067bb95c2b1a7ff9e6b878221" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 05 08:24:42 crc kubenswrapper[4846]: E1005 08:24:42.796044 4846 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="76dec6e0-c589-4072-b65f-42ee496e0c61" containerName="nova-scheduler-scheduler" Oct 05 08:24:47 crc kubenswrapper[4846]: E1005 08:24:47.790996 4846 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="7692194d2e1f5715282a95ce181ed64987bc7fa067bb95c2b1a7ff9e6b878221" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 05 08:24:47 crc kubenswrapper[4846]: E1005 08:24:47.793237 4846 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="7692194d2e1f5715282a95ce181ed64987bc7fa067bb95c2b1a7ff9e6b878221" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 05 08:24:47 crc kubenswrapper[4846]: E1005 08:24:47.795172 4846 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="7692194d2e1f5715282a95ce181ed64987bc7fa067bb95c2b1a7ff9e6b878221" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 05 08:24:47 crc kubenswrapper[4846]: E1005 08:24:47.795243 4846 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="76dec6e0-c589-4072-b65f-42ee496e0c61" containerName="nova-scheduler-scheduler" Oct 05 08:24:51 crc kubenswrapper[4846]: I1005 08:24:51.690064 4846 generic.go:334] "Generic (PLEG): container finished" podID="8d409b1e-a91c-49c0-96f2-3df8b6679199" containerID="63fe74b1e29ffea6657ebecf1c49c55957af6382c195a72b7cc62916280cf64d" exitCode=0 Oct 05 08:24:51 crc kubenswrapper[4846]: I1005 08:24:51.690655 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"8d409b1e-a91c-49c0-96f2-3df8b6679199","Type":"ContainerDied","Data":"63fe74b1e29ffea6657ebecf1c49c55957af6382c195a72b7cc62916280cf64d"} Oct 05 08:24:51 crc kubenswrapper[4846]: I1005 08:24:51.690690 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"8d409b1e-a91c-49c0-96f2-3df8b6679199","Type":"ContainerDied","Data":"4ae0ff76b3396458eac202d4dd90f675b7f8b27f03adb01bafff3609184319b6"} Oct 05 08:24:51 crc kubenswrapper[4846]: I1005 08:24:51.690706 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4ae0ff76b3396458eac202d4dd90f675b7f8b27f03adb01bafff3609184319b6" Oct 05 08:24:51 crc kubenswrapper[4846]: I1005 08:24:51.694311 4846 generic.go:334] "Generic (PLEG): container finished" podID="188d67ae-21a9-43c8-a33f-44084cbdc3d8" containerID="006e3a4d809a2526a062908765859aa3087dc6d5e6f8d19c45d2849dc9615a2f" exitCode=0 Oct 05 08:24:51 crc kubenswrapper[4846]: I1005 08:24:51.694343 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"188d67ae-21a9-43c8-a33f-44084cbdc3d8","Type":"ContainerDied","Data":"006e3a4d809a2526a062908765859aa3087dc6d5e6f8d19c45d2849dc9615a2f"} Oct 05 08:24:51 crc kubenswrapper[4846]: I1005 08:24:51.694363 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"188d67ae-21a9-43c8-a33f-44084cbdc3d8","Type":"ContainerDied","Data":"e0dc1f1ba68b0cb985b9b59098683e590693e2e6b8b14fbfdf79ffe6b4894fc9"} Oct 05 08:24:51 crc kubenswrapper[4846]: I1005 08:24:51.694375 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e0dc1f1ba68b0cb985b9b59098683e590693e2e6b8b14fbfdf79ffe6b4894fc9" Oct 05 08:24:51 crc kubenswrapper[4846]: I1005 08:24:51.718402 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 05 08:24:51 crc kubenswrapper[4846]: I1005 08:24:51.728912 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 05 08:24:51 crc kubenswrapper[4846]: I1005 08:24:51.797010 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/188d67ae-21a9-43c8-a33f-44084cbdc3d8-combined-ca-bundle\") pod \"188d67ae-21a9-43c8-a33f-44084cbdc3d8\" (UID: \"188d67ae-21a9-43c8-a33f-44084cbdc3d8\") " Oct 05 08:24:51 crc kubenswrapper[4846]: I1005 08:24:51.797168 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/188d67ae-21a9-43c8-a33f-44084cbdc3d8-config-data\") pod \"188d67ae-21a9-43c8-a33f-44084cbdc3d8\" (UID: \"188d67ae-21a9-43c8-a33f-44084cbdc3d8\") " Oct 05 08:24:51 crc kubenswrapper[4846]: I1005 08:24:51.797222 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/188d67ae-21a9-43c8-a33f-44084cbdc3d8-logs\") pod \"188d67ae-21a9-43c8-a33f-44084cbdc3d8\" (UID: \"188d67ae-21a9-43c8-a33f-44084cbdc3d8\") " Oct 05 08:24:51 crc kubenswrapper[4846]: I1005 08:24:51.797324 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rmftc\" (UniqueName: \"kubernetes.io/projected/188d67ae-21a9-43c8-a33f-44084cbdc3d8-kube-api-access-rmftc\") pod \"188d67ae-21a9-43c8-a33f-44084cbdc3d8\" (UID: \"188d67ae-21a9-43c8-a33f-44084cbdc3d8\") " Oct 05 08:24:51 crc kubenswrapper[4846]: I1005 08:24:51.797774 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/188d67ae-21a9-43c8-a33f-44084cbdc3d8-logs" (OuterVolumeSpecName: "logs") pod "188d67ae-21a9-43c8-a33f-44084cbdc3d8" (UID: "188d67ae-21a9-43c8-a33f-44084cbdc3d8"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 08:24:51 crc kubenswrapper[4846]: I1005 08:24:51.803826 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/188d67ae-21a9-43c8-a33f-44084cbdc3d8-kube-api-access-rmftc" (OuterVolumeSpecName: "kube-api-access-rmftc") pod "188d67ae-21a9-43c8-a33f-44084cbdc3d8" (UID: "188d67ae-21a9-43c8-a33f-44084cbdc3d8"). InnerVolumeSpecName "kube-api-access-rmftc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:24:51 crc kubenswrapper[4846]: I1005 08:24:51.823648 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/188d67ae-21a9-43c8-a33f-44084cbdc3d8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "188d67ae-21a9-43c8-a33f-44084cbdc3d8" (UID: "188d67ae-21a9-43c8-a33f-44084cbdc3d8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:24:51 crc kubenswrapper[4846]: I1005 08:24:51.826665 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/188d67ae-21a9-43c8-a33f-44084cbdc3d8-config-data" (OuterVolumeSpecName: "config-data") pod "188d67ae-21a9-43c8-a33f-44084cbdc3d8" (UID: "188d67ae-21a9-43c8-a33f-44084cbdc3d8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:24:51 crc kubenswrapper[4846]: I1005 08:24:51.899440 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/8d409b1e-a91c-49c0-96f2-3df8b6679199-nova-metadata-tls-certs\") pod \"8d409b1e-a91c-49c0-96f2-3df8b6679199\" (UID: \"8d409b1e-a91c-49c0-96f2-3df8b6679199\") " Oct 05 08:24:51 crc kubenswrapper[4846]: I1005 08:24:51.899647 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vdnc4\" (UniqueName: \"kubernetes.io/projected/8d409b1e-a91c-49c0-96f2-3df8b6679199-kube-api-access-vdnc4\") pod \"8d409b1e-a91c-49c0-96f2-3df8b6679199\" (UID: \"8d409b1e-a91c-49c0-96f2-3df8b6679199\") " Oct 05 08:24:51 crc kubenswrapper[4846]: I1005 08:24:51.899693 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8d409b1e-a91c-49c0-96f2-3df8b6679199-logs\") pod \"8d409b1e-a91c-49c0-96f2-3df8b6679199\" (UID: \"8d409b1e-a91c-49c0-96f2-3df8b6679199\") " Oct 05 08:24:51 crc kubenswrapper[4846]: I1005 08:24:51.899733 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8d409b1e-a91c-49c0-96f2-3df8b6679199-config-data\") pod \"8d409b1e-a91c-49c0-96f2-3df8b6679199\" (UID: \"8d409b1e-a91c-49c0-96f2-3df8b6679199\") " Oct 05 08:24:51 crc kubenswrapper[4846]: I1005 08:24:51.899897 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8d409b1e-a91c-49c0-96f2-3df8b6679199-combined-ca-bundle\") pod \"8d409b1e-a91c-49c0-96f2-3df8b6679199\" (UID: \"8d409b1e-a91c-49c0-96f2-3df8b6679199\") " Oct 05 08:24:51 crc kubenswrapper[4846]: I1005 08:24:51.900441 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rmftc\" (UniqueName: \"kubernetes.io/projected/188d67ae-21a9-43c8-a33f-44084cbdc3d8-kube-api-access-rmftc\") on node \"crc\" DevicePath \"\"" Oct 05 08:24:51 crc kubenswrapper[4846]: I1005 08:24:51.900464 4846 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/188d67ae-21a9-43c8-a33f-44084cbdc3d8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 08:24:51 crc kubenswrapper[4846]: I1005 08:24:51.900478 4846 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/188d67ae-21a9-43c8-a33f-44084cbdc3d8-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 08:24:51 crc kubenswrapper[4846]: I1005 08:24:51.900475 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8d409b1e-a91c-49c0-96f2-3df8b6679199-logs" (OuterVolumeSpecName: "logs") pod "8d409b1e-a91c-49c0-96f2-3df8b6679199" (UID: "8d409b1e-a91c-49c0-96f2-3df8b6679199"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 08:24:51 crc kubenswrapper[4846]: I1005 08:24:51.900490 4846 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/188d67ae-21a9-43c8-a33f-44084cbdc3d8-logs\") on node \"crc\" DevicePath \"\"" Oct 05 08:24:51 crc kubenswrapper[4846]: I1005 08:24:51.905845 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8d409b1e-a91c-49c0-96f2-3df8b6679199-kube-api-access-vdnc4" (OuterVolumeSpecName: "kube-api-access-vdnc4") pod "8d409b1e-a91c-49c0-96f2-3df8b6679199" (UID: "8d409b1e-a91c-49c0-96f2-3df8b6679199"). InnerVolumeSpecName "kube-api-access-vdnc4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:24:51 crc kubenswrapper[4846]: I1005 08:24:51.924423 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8d409b1e-a91c-49c0-96f2-3df8b6679199-config-data" (OuterVolumeSpecName: "config-data") pod "8d409b1e-a91c-49c0-96f2-3df8b6679199" (UID: "8d409b1e-a91c-49c0-96f2-3df8b6679199"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:24:51 crc kubenswrapper[4846]: I1005 08:24:51.944457 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8d409b1e-a91c-49c0-96f2-3df8b6679199-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8d409b1e-a91c-49c0-96f2-3df8b6679199" (UID: "8d409b1e-a91c-49c0-96f2-3df8b6679199"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:24:51 crc kubenswrapper[4846]: I1005 08:24:51.949969 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8d409b1e-a91c-49c0-96f2-3df8b6679199-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "8d409b1e-a91c-49c0-96f2-3df8b6679199" (UID: "8d409b1e-a91c-49c0-96f2-3df8b6679199"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:24:52 crc kubenswrapper[4846]: I1005 08:24:52.002435 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vdnc4\" (UniqueName: \"kubernetes.io/projected/8d409b1e-a91c-49c0-96f2-3df8b6679199-kube-api-access-vdnc4\") on node \"crc\" DevicePath \"\"" Oct 05 08:24:52 crc kubenswrapper[4846]: I1005 08:24:52.002501 4846 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8d409b1e-a91c-49c0-96f2-3df8b6679199-logs\") on node \"crc\" DevicePath \"\"" Oct 05 08:24:52 crc kubenswrapper[4846]: I1005 08:24:52.002522 4846 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8d409b1e-a91c-49c0-96f2-3df8b6679199-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 08:24:52 crc kubenswrapper[4846]: I1005 08:24:52.002540 4846 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8d409b1e-a91c-49c0-96f2-3df8b6679199-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 08:24:52 crc kubenswrapper[4846]: I1005 08:24:52.002557 4846 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/8d409b1e-a91c-49c0-96f2-3df8b6679199-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 05 08:24:52 crc kubenswrapper[4846]: I1005 08:24:52.707447 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 05 08:24:52 crc kubenswrapper[4846]: I1005 08:24:52.707447 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 05 08:24:52 crc kubenswrapper[4846]: I1005 08:24:52.751028 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 05 08:24:52 crc kubenswrapper[4846]: I1005 08:24:52.773768 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Oct 05 08:24:52 crc kubenswrapper[4846]: E1005 08:24:52.791383 4846 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="7692194d2e1f5715282a95ce181ed64987bc7fa067bb95c2b1a7ff9e6b878221" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 05 08:24:52 crc kubenswrapper[4846]: I1005 08:24:52.798362 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 05 08:24:52 crc kubenswrapper[4846]: E1005 08:24:52.801322 4846 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="7692194d2e1f5715282a95ce181ed64987bc7fa067bb95c2b1a7ff9e6b878221" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 05 08:24:52 crc kubenswrapper[4846]: E1005 08:24:52.811254 4846 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="7692194d2e1f5715282a95ce181ed64987bc7fa067bb95c2b1a7ff9e6b878221" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 05 08:24:52 crc kubenswrapper[4846]: E1005 08:24:52.811380 4846 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="76dec6e0-c589-4072-b65f-42ee496e0c61" containerName="nova-scheduler-scheduler" Oct 05 08:24:52 crc kubenswrapper[4846]: I1005 08:24:52.813578 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Oct 05 08:24:52 crc kubenswrapper[4846]: I1005 08:24:52.839617 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 05 08:24:52 crc kubenswrapper[4846]: E1005 08:24:52.840317 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8d409b1e-a91c-49c0-96f2-3df8b6679199" containerName="nova-metadata-log" Oct 05 08:24:52 crc kubenswrapper[4846]: I1005 08:24:52.840377 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="8d409b1e-a91c-49c0-96f2-3df8b6679199" containerName="nova-metadata-log" Oct 05 08:24:52 crc kubenswrapper[4846]: E1005 08:24:52.840411 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f8888283-b66d-4993-80a0-a7b6b1bac02b" containerName="nova-manage" Oct 05 08:24:52 crc kubenswrapper[4846]: I1005 08:24:52.840424 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="f8888283-b66d-4993-80a0-a7b6b1bac02b" containerName="nova-manage" Oct 05 08:24:52 crc kubenswrapper[4846]: E1005 08:24:52.840457 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="188d67ae-21a9-43c8-a33f-44084cbdc3d8" containerName="nova-api-api" Oct 05 08:24:52 crc kubenswrapper[4846]: I1005 08:24:52.840472 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="188d67ae-21a9-43c8-a33f-44084cbdc3d8" containerName="nova-api-api" Oct 05 08:24:52 crc kubenswrapper[4846]: E1005 08:24:52.840498 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="188d67ae-21a9-43c8-a33f-44084cbdc3d8" containerName="nova-api-log" Oct 05 08:24:52 crc kubenswrapper[4846]: I1005 08:24:52.840512 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="188d67ae-21a9-43c8-a33f-44084cbdc3d8" containerName="nova-api-log" Oct 05 08:24:52 crc kubenswrapper[4846]: E1005 08:24:52.840548 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8d409b1e-a91c-49c0-96f2-3df8b6679199" containerName="nova-metadata-metadata" Oct 05 08:24:52 crc kubenswrapper[4846]: I1005 08:24:52.840562 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="8d409b1e-a91c-49c0-96f2-3df8b6679199" containerName="nova-metadata-metadata" Oct 05 08:24:52 crc kubenswrapper[4846]: I1005 08:24:52.840798 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="188d67ae-21a9-43c8-a33f-44084cbdc3d8" containerName="nova-api-api" Oct 05 08:24:52 crc kubenswrapper[4846]: I1005 08:24:52.840812 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="8d409b1e-a91c-49c0-96f2-3df8b6679199" containerName="nova-metadata-metadata" Oct 05 08:24:52 crc kubenswrapper[4846]: I1005 08:24:52.840829 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="8d409b1e-a91c-49c0-96f2-3df8b6679199" containerName="nova-metadata-log" Oct 05 08:24:52 crc kubenswrapper[4846]: I1005 08:24:52.840858 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="f8888283-b66d-4993-80a0-a7b6b1bac02b" containerName="nova-manage" Oct 05 08:24:52 crc kubenswrapper[4846]: I1005 08:24:52.840872 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="188d67ae-21a9-43c8-a33f-44084cbdc3d8" containerName="nova-api-log" Oct 05 08:24:52 crc kubenswrapper[4846]: I1005 08:24:52.842053 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 05 08:24:52 crc kubenswrapper[4846]: I1005 08:24:52.844314 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 05 08:24:52 crc kubenswrapper[4846]: I1005 08:24:52.851527 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 05 08:24:52 crc kubenswrapper[4846]: I1005 08:24:52.855833 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 05 08:24:52 crc kubenswrapper[4846]: I1005 08:24:52.862680 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 05 08:24:52 crc kubenswrapper[4846]: I1005 08:24:52.866342 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Oct 05 08:24:52 crc kubenswrapper[4846]: I1005 08:24:52.869421 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 05 08:24:52 crc kubenswrapper[4846]: I1005 08:24:52.877574 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 05 08:24:52 crc kubenswrapper[4846]: I1005 08:24:52.922054 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d94b0d13-b52b-4fcb-b016-7d5c91f72600-logs\") pod \"nova-api-0\" (UID: \"d94b0d13-b52b-4fcb-b016-7d5c91f72600\") " pod="openstack/nova-api-0" Oct 05 08:24:52 crc kubenswrapper[4846]: I1005 08:24:52.922114 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d94b0d13-b52b-4fcb-b016-7d5c91f72600-config-data\") pod \"nova-api-0\" (UID: \"d94b0d13-b52b-4fcb-b016-7d5c91f72600\") " pod="openstack/nova-api-0" Oct 05 08:24:52 crc kubenswrapper[4846]: I1005 08:24:52.922406 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d94b0d13-b52b-4fcb-b016-7d5c91f72600-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"d94b0d13-b52b-4fcb-b016-7d5c91f72600\") " pod="openstack/nova-api-0" Oct 05 08:24:52 crc kubenswrapper[4846]: I1005 08:24:52.922516 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4v2x9\" (UniqueName: \"kubernetes.io/projected/d94b0d13-b52b-4fcb-b016-7d5c91f72600-kube-api-access-4v2x9\") pod \"nova-api-0\" (UID: \"d94b0d13-b52b-4fcb-b016-7d5c91f72600\") " pod="openstack/nova-api-0" Oct 05 08:24:53 crc kubenswrapper[4846]: I1005 08:24:53.024067 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a0ed1ebb-faf2-48bb-8fe4-efcecaa2bb05-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"a0ed1ebb-faf2-48bb-8fe4-efcecaa2bb05\") " pod="openstack/nova-metadata-0" Oct 05 08:24:53 crc kubenswrapper[4846]: I1005 08:24:53.024328 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d94b0d13-b52b-4fcb-b016-7d5c91f72600-logs\") pod \"nova-api-0\" (UID: \"d94b0d13-b52b-4fcb-b016-7d5c91f72600\") " pod="openstack/nova-api-0" Oct 05 08:24:53 crc kubenswrapper[4846]: I1005 08:24:53.024511 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d94b0d13-b52b-4fcb-b016-7d5c91f72600-config-data\") pod \"nova-api-0\" (UID: \"d94b0d13-b52b-4fcb-b016-7d5c91f72600\") " pod="openstack/nova-api-0" Oct 05 08:24:53 crc kubenswrapper[4846]: I1005 08:24:53.024601 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a0ed1ebb-faf2-48bb-8fe4-efcecaa2bb05-config-data\") pod \"nova-metadata-0\" (UID: \"a0ed1ebb-faf2-48bb-8fe4-efcecaa2bb05\") " pod="openstack/nova-metadata-0" Oct 05 08:24:53 crc kubenswrapper[4846]: I1005 08:24:53.024695 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/a0ed1ebb-faf2-48bb-8fe4-efcecaa2bb05-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"a0ed1ebb-faf2-48bb-8fe4-efcecaa2bb05\") " pod="openstack/nova-metadata-0" Oct 05 08:24:53 crc kubenswrapper[4846]: I1005 08:24:53.024714 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d94b0d13-b52b-4fcb-b016-7d5c91f72600-logs\") pod \"nova-api-0\" (UID: \"d94b0d13-b52b-4fcb-b016-7d5c91f72600\") " pod="openstack/nova-api-0" Oct 05 08:24:53 crc kubenswrapper[4846]: I1005 08:24:53.024859 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nf499\" (UniqueName: \"kubernetes.io/projected/a0ed1ebb-faf2-48bb-8fe4-efcecaa2bb05-kube-api-access-nf499\") pod \"nova-metadata-0\" (UID: \"a0ed1ebb-faf2-48bb-8fe4-efcecaa2bb05\") " pod="openstack/nova-metadata-0" Oct 05 08:24:53 crc kubenswrapper[4846]: I1005 08:24:53.024974 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d94b0d13-b52b-4fcb-b016-7d5c91f72600-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"d94b0d13-b52b-4fcb-b016-7d5c91f72600\") " pod="openstack/nova-api-0" Oct 05 08:24:53 crc kubenswrapper[4846]: I1005 08:24:53.025077 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4v2x9\" (UniqueName: \"kubernetes.io/projected/d94b0d13-b52b-4fcb-b016-7d5c91f72600-kube-api-access-4v2x9\") pod \"nova-api-0\" (UID: \"d94b0d13-b52b-4fcb-b016-7d5c91f72600\") " pod="openstack/nova-api-0" Oct 05 08:24:53 crc kubenswrapper[4846]: I1005 08:24:53.025121 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a0ed1ebb-faf2-48bb-8fe4-efcecaa2bb05-logs\") pod \"nova-metadata-0\" (UID: \"a0ed1ebb-faf2-48bb-8fe4-efcecaa2bb05\") " pod="openstack/nova-metadata-0" Oct 05 08:24:53 crc kubenswrapper[4846]: I1005 08:24:53.030150 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d94b0d13-b52b-4fcb-b016-7d5c91f72600-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"d94b0d13-b52b-4fcb-b016-7d5c91f72600\") " pod="openstack/nova-api-0" Oct 05 08:24:53 crc kubenswrapper[4846]: I1005 08:24:53.030467 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d94b0d13-b52b-4fcb-b016-7d5c91f72600-config-data\") pod \"nova-api-0\" (UID: \"d94b0d13-b52b-4fcb-b016-7d5c91f72600\") " pod="openstack/nova-api-0" Oct 05 08:24:53 crc kubenswrapper[4846]: I1005 08:24:53.053064 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4v2x9\" (UniqueName: \"kubernetes.io/projected/d94b0d13-b52b-4fcb-b016-7d5c91f72600-kube-api-access-4v2x9\") pod \"nova-api-0\" (UID: \"d94b0d13-b52b-4fcb-b016-7d5c91f72600\") " pod="openstack/nova-api-0" Oct 05 08:24:53 crc kubenswrapper[4846]: I1005 08:24:53.126697 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a0ed1ebb-faf2-48bb-8fe4-efcecaa2bb05-config-data\") pod \"nova-metadata-0\" (UID: \"a0ed1ebb-faf2-48bb-8fe4-efcecaa2bb05\") " pod="openstack/nova-metadata-0" Oct 05 08:24:53 crc kubenswrapper[4846]: I1005 08:24:53.126750 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/a0ed1ebb-faf2-48bb-8fe4-efcecaa2bb05-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"a0ed1ebb-faf2-48bb-8fe4-efcecaa2bb05\") " pod="openstack/nova-metadata-0" Oct 05 08:24:53 crc kubenswrapper[4846]: I1005 08:24:53.126791 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nf499\" (UniqueName: \"kubernetes.io/projected/a0ed1ebb-faf2-48bb-8fe4-efcecaa2bb05-kube-api-access-nf499\") pod \"nova-metadata-0\" (UID: \"a0ed1ebb-faf2-48bb-8fe4-efcecaa2bb05\") " pod="openstack/nova-metadata-0" Oct 05 08:24:53 crc kubenswrapper[4846]: I1005 08:24:53.126842 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a0ed1ebb-faf2-48bb-8fe4-efcecaa2bb05-logs\") pod \"nova-metadata-0\" (UID: \"a0ed1ebb-faf2-48bb-8fe4-efcecaa2bb05\") " pod="openstack/nova-metadata-0" Oct 05 08:24:53 crc kubenswrapper[4846]: I1005 08:24:53.126866 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a0ed1ebb-faf2-48bb-8fe4-efcecaa2bb05-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"a0ed1ebb-faf2-48bb-8fe4-efcecaa2bb05\") " pod="openstack/nova-metadata-0" Oct 05 08:24:53 crc kubenswrapper[4846]: I1005 08:24:53.127777 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a0ed1ebb-faf2-48bb-8fe4-efcecaa2bb05-logs\") pod \"nova-metadata-0\" (UID: \"a0ed1ebb-faf2-48bb-8fe4-efcecaa2bb05\") " pod="openstack/nova-metadata-0" Oct 05 08:24:53 crc kubenswrapper[4846]: I1005 08:24:53.131120 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a0ed1ebb-faf2-48bb-8fe4-efcecaa2bb05-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"a0ed1ebb-faf2-48bb-8fe4-efcecaa2bb05\") " pod="openstack/nova-metadata-0" Oct 05 08:24:53 crc kubenswrapper[4846]: I1005 08:24:53.131827 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a0ed1ebb-faf2-48bb-8fe4-efcecaa2bb05-config-data\") pod \"nova-metadata-0\" (UID: \"a0ed1ebb-faf2-48bb-8fe4-efcecaa2bb05\") " pod="openstack/nova-metadata-0" Oct 05 08:24:53 crc kubenswrapper[4846]: I1005 08:24:53.135598 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/a0ed1ebb-faf2-48bb-8fe4-efcecaa2bb05-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"a0ed1ebb-faf2-48bb-8fe4-efcecaa2bb05\") " pod="openstack/nova-metadata-0" Oct 05 08:24:53 crc kubenswrapper[4846]: I1005 08:24:53.155681 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nf499\" (UniqueName: \"kubernetes.io/projected/a0ed1ebb-faf2-48bb-8fe4-efcecaa2bb05-kube-api-access-nf499\") pod \"nova-metadata-0\" (UID: \"a0ed1ebb-faf2-48bb-8fe4-efcecaa2bb05\") " pod="openstack/nova-metadata-0" Oct 05 08:24:53 crc kubenswrapper[4846]: I1005 08:24:53.171924 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 05 08:24:53 crc kubenswrapper[4846]: I1005 08:24:53.182521 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 05 08:24:53 crc kubenswrapper[4846]: I1005 08:24:53.712836 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 05 08:24:53 crc kubenswrapper[4846]: I1005 08:24:53.792824 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 05 08:24:53 crc kubenswrapper[4846]: W1005 08:24:53.792896 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd94b0d13_b52b_4fcb_b016_7d5c91f72600.slice/crio-8efde6ebdd901237d1986d669796f82020e9be1f7925321118c661e34fcff0ed WatchSource:0}: Error finding container 8efde6ebdd901237d1986d669796f82020e9be1f7925321118c661e34fcff0ed: Status 404 returned error can't find the container with id 8efde6ebdd901237d1986d669796f82020e9be1f7925321118c661e34fcff0ed Oct 05 08:24:54 crc kubenswrapper[4846]: I1005 08:24:54.513400 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="188d67ae-21a9-43c8-a33f-44084cbdc3d8" path="/var/lib/kubelet/pods/188d67ae-21a9-43c8-a33f-44084cbdc3d8/volumes" Oct 05 08:24:54 crc kubenswrapper[4846]: I1005 08:24:54.514999 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8d409b1e-a91c-49c0-96f2-3df8b6679199" path="/var/lib/kubelet/pods/8d409b1e-a91c-49c0-96f2-3df8b6679199/volumes" Oct 05 08:24:54 crc kubenswrapper[4846]: I1005 08:24:54.736411 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"d94b0d13-b52b-4fcb-b016-7d5c91f72600","Type":"ContainerStarted","Data":"6f2689b0dec92601169c548cffb2d10afbe1f569a214735f18107ae080a3e631"} Oct 05 08:24:54 crc kubenswrapper[4846]: I1005 08:24:54.736536 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"d94b0d13-b52b-4fcb-b016-7d5c91f72600","Type":"ContainerStarted","Data":"a98d5bd27987abf87bfcba4469cec38ec4b051938f756192a742e3b0b24e5589"} Oct 05 08:24:54 crc kubenswrapper[4846]: I1005 08:24:54.736568 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"d94b0d13-b52b-4fcb-b016-7d5c91f72600","Type":"ContainerStarted","Data":"8efde6ebdd901237d1986d669796f82020e9be1f7925321118c661e34fcff0ed"} Oct 05 08:24:54 crc kubenswrapper[4846]: I1005 08:24:54.739821 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"a0ed1ebb-faf2-48bb-8fe4-efcecaa2bb05","Type":"ContainerStarted","Data":"426817ac622bb93bbce4d37648f9e92d5a85879c645c126fde022f827fbd461a"} Oct 05 08:24:54 crc kubenswrapper[4846]: I1005 08:24:54.739909 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"a0ed1ebb-faf2-48bb-8fe4-efcecaa2bb05","Type":"ContainerStarted","Data":"61f00a73cfd19cacff2ab04863a43cce8b74f56a2ee186d97ecc5edaa8983aec"} Oct 05 08:24:54 crc kubenswrapper[4846]: I1005 08:24:54.739930 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"a0ed1ebb-faf2-48bb-8fe4-efcecaa2bb05","Type":"ContainerStarted","Data":"c49958b7cdfe3ce6aceb97827968e9db73b3479550a08d910e4c331710a10cce"} Oct 05 08:24:54 crc kubenswrapper[4846]: I1005 08:24:54.764779 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.764749849 podStartE2EDuration="2.764749849s" podCreationTimestamp="2025-10-05 08:24:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 08:24:54.760434444 +0000 UTC m=+5817.001287259" watchObservedRunningTime="2025-10-05 08:24:54.764749849 +0000 UTC m=+5817.005602664" Oct 05 08:24:54 crc kubenswrapper[4846]: I1005 08:24:54.797761 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.797734715 podStartE2EDuration="2.797734715s" podCreationTimestamp="2025-10-05 08:24:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 08:24:54.786690361 +0000 UTC m=+5817.027543146" watchObservedRunningTime="2025-10-05 08:24:54.797734715 +0000 UTC m=+5817.038587530" Oct 05 08:24:55 crc kubenswrapper[4846]: I1005 08:24:55.498542 4846 scope.go:117] "RemoveContainer" containerID="fdafa27da24296f9087c33b6b12a6598d2a8d61fc985b06ad680401828c9feb6" Oct 05 08:24:55 crc kubenswrapper[4846]: E1005 08:24:55.499101 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 08:24:57 crc kubenswrapper[4846]: E1005 08:24:57.791071 4846 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="7692194d2e1f5715282a95ce181ed64987bc7fa067bb95c2b1a7ff9e6b878221" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 05 08:24:57 crc kubenswrapper[4846]: E1005 08:24:57.793667 4846 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="7692194d2e1f5715282a95ce181ed64987bc7fa067bb95c2b1a7ff9e6b878221" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 05 08:24:57 crc kubenswrapper[4846]: E1005 08:24:57.795402 4846 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="7692194d2e1f5715282a95ce181ed64987bc7fa067bb95c2b1a7ff9e6b878221" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 05 08:24:57 crc kubenswrapper[4846]: E1005 08:24:57.795443 4846 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="76dec6e0-c589-4072-b65f-42ee496e0c61" containerName="nova-scheduler-scheduler" Oct 05 08:24:58 crc kubenswrapper[4846]: I1005 08:24:58.183617 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 05 08:24:58 crc kubenswrapper[4846]: I1005 08:24:58.183707 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 05 08:25:02 crc kubenswrapper[4846]: E1005 08:25:02.792996 4846 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="7692194d2e1f5715282a95ce181ed64987bc7fa067bb95c2b1a7ff9e6b878221" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 05 08:25:02 crc kubenswrapper[4846]: E1005 08:25:02.795849 4846 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="7692194d2e1f5715282a95ce181ed64987bc7fa067bb95c2b1a7ff9e6b878221" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 05 08:25:02 crc kubenswrapper[4846]: E1005 08:25:02.801836 4846 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="7692194d2e1f5715282a95ce181ed64987bc7fa067bb95c2b1a7ff9e6b878221" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 05 08:25:02 crc kubenswrapper[4846]: E1005 08:25:02.801953 4846 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="76dec6e0-c589-4072-b65f-42ee496e0c61" containerName="nova-scheduler-scheduler" Oct 05 08:25:03 crc kubenswrapper[4846]: I1005 08:25:03.173340 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 05 08:25:03 crc kubenswrapper[4846]: I1005 08:25:03.173432 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 05 08:25:03 crc kubenswrapper[4846]: I1005 08:25:03.183722 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 05 08:25:03 crc kubenswrapper[4846]: I1005 08:25:03.183816 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 05 08:25:04 crc kubenswrapper[4846]: I1005 08:25:04.271361 4846 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="d94b0d13-b52b-4fcb-b016-7d5c91f72600" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.1.95:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 05 08:25:04 crc kubenswrapper[4846]: I1005 08:25:04.271419 4846 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="a0ed1ebb-faf2-48bb-8fe4-efcecaa2bb05" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.1.96:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 05 08:25:04 crc kubenswrapper[4846]: I1005 08:25:04.271419 4846 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="d94b0d13-b52b-4fcb-b016-7d5c91f72600" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.1.95:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 05 08:25:04 crc kubenswrapper[4846]: I1005 08:25:04.271372 4846 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="a0ed1ebb-faf2-48bb-8fe4-efcecaa2bb05" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.1.96:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 05 08:25:07 crc kubenswrapper[4846]: E1005 08:25:07.789275 4846 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 7692194d2e1f5715282a95ce181ed64987bc7fa067bb95c2b1a7ff9e6b878221 is running failed: container process not found" containerID="7692194d2e1f5715282a95ce181ed64987bc7fa067bb95c2b1a7ff9e6b878221" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 05 08:25:07 crc kubenswrapper[4846]: E1005 08:25:07.790925 4846 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 7692194d2e1f5715282a95ce181ed64987bc7fa067bb95c2b1a7ff9e6b878221 is running failed: container process not found" containerID="7692194d2e1f5715282a95ce181ed64987bc7fa067bb95c2b1a7ff9e6b878221" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 05 08:25:07 crc kubenswrapper[4846]: E1005 08:25:07.791771 4846 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 7692194d2e1f5715282a95ce181ed64987bc7fa067bb95c2b1a7ff9e6b878221 is running failed: container process not found" containerID="7692194d2e1f5715282a95ce181ed64987bc7fa067bb95c2b1a7ff9e6b878221" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 05 08:25:07 crc kubenswrapper[4846]: E1005 08:25:07.792042 4846 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 7692194d2e1f5715282a95ce181ed64987bc7fa067bb95c2b1a7ff9e6b878221 is running failed: container process not found" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="76dec6e0-c589-4072-b65f-42ee496e0c61" containerName="nova-scheduler-scheduler" Oct 05 08:25:07 crc kubenswrapper[4846]: I1005 08:25:07.886942 4846 generic.go:334] "Generic (PLEG): container finished" podID="76dec6e0-c589-4072-b65f-42ee496e0c61" containerID="7692194d2e1f5715282a95ce181ed64987bc7fa067bb95c2b1a7ff9e6b878221" exitCode=137 Oct 05 08:25:07 crc kubenswrapper[4846]: I1005 08:25:07.887003 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"76dec6e0-c589-4072-b65f-42ee496e0c61","Type":"ContainerDied","Data":"7692194d2e1f5715282a95ce181ed64987bc7fa067bb95c2b1a7ff9e6b878221"} Oct 05 08:25:08 crc kubenswrapper[4846]: I1005 08:25:08.179695 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 05 08:25:08 crc kubenswrapper[4846]: I1005 08:25:08.267118 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/76dec6e0-c589-4072-b65f-42ee496e0c61-config-data\") pod \"76dec6e0-c589-4072-b65f-42ee496e0c61\" (UID: \"76dec6e0-c589-4072-b65f-42ee496e0c61\") " Oct 05 08:25:08 crc kubenswrapper[4846]: I1005 08:25:08.267252 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z8mtg\" (UniqueName: \"kubernetes.io/projected/76dec6e0-c589-4072-b65f-42ee496e0c61-kube-api-access-z8mtg\") pod \"76dec6e0-c589-4072-b65f-42ee496e0c61\" (UID: \"76dec6e0-c589-4072-b65f-42ee496e0c61\") " Oct 05 08:25:08 crc kubenswrapper[4846]: I1005 08:25:08.267287 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76dec6e0-c589-4072-b65f-42ee496e0c61-combined-ca-bundle\") pod \"76dec6e0-c589-4072-b65f-42ee496e0c61\" (UID: \"76dec6e0-c589-4072-b65f-42ee496e0c61\") " Oct 05 08:25:08 crc kubenswrapper[4846]: I1005 08:25:08.276051 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/76dec6e0-c589-4072-b65f-42ee496e0c61-kube-api-access-z8mtg" (OuterVolumeSpecName: "kube-api-access-z8mtg") pod "76dec6e0-c589-4072-b65f-42ee496e0c61" (UID: "76dec6e0-c589-4072-b65f-42ee496e0c61"). InnerVolumeSpecName "kube-api-access-z8mtg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:25:08 crc kubenswrapper[4846]: I1005 08:25:08.296955 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/76dec6e0-c589-4072-b65f-42ee496e0c61-config-data" (OuterVolumeSpecName: "config-data") pod "76dec6e0-c589-4072-b65f-42ee496e0c61" (UID: "76dec6e0-c589-4072-b65f-42ee496e0c61"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:25:08 crc kubenswrapper[4846]: I1005 08:25:08.312324 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/76dec6e0-c589-4072-b65f-42ee496e0c61-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "76dec6e0-c589-4072-b65f-42ee496e0c61" (UID: "76dec6e0-c589-4072-b65f-42ee496e0c61"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:25:08 crc kubenswrapper[4846]: I1005 08:25:08.369436 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z8mtg\" (UniqueName: \"kubernetes.io/projected/76dec6e0-c589-4072-b65f-42ee496e0c61-kube-api-access-z8mtg\") on node \"crc\" DevicePath \"\"" Oct 05 08:25:08 crc kubenswrapper[4846]: I1005 08:25:08.369477 4846 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76dec6e0-c589-4072-b65f-42ee496e0c61-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 08:25:08 crc kubenswrapper[4846]: I1005 08:25:08.369491 4846 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/76dec6e0-c589-4072-b65f-42ee496e0c61-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 08:25:08 crc kubenswrapper[4846]: I1005 08:25:08.897579 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"76dec6e0-c589-4072-b65f-42ee496e0c61","Type":"ContainerDied","Data":"f08055fbb0d11e2641506b70081764f971a0afb10ac647ddc76c8d378de12d30"} Oct 05 08:25:08 crc kubenswrapper[4846]: I1005 08:25:08.897636 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 05 08:25:08 crc kubenswrapper[4846]: I1005 08:25:08.897646 4846 scope.go:117] "RemoveContainer" containerID="7692194d2e1f5715282a95ce181ed64987bc7fa067bb95c2b1a7ff9e6b878221" Oct 05 08:25:08 crc kubenswrapper[4846]: I1005 08:25:08.928453 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 05 08:25:08 crc kubenswrapper[4846]: I1005 08:25:08.954460 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Oct 05 08:25:08 crc kubenswrapper[4846]: I1005 08:25:08.970093 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Oct 05 08:25:08 crc kubenswrapper[4846]: E1005 08:25:08.970502 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="76dec6e0-c589-4072-b65f-42ee496e0c61" containerName="nova-scheduler-scheduler" Oct 05 08:25:08 crc kubenswrapper[4846]: I1005 08:25:08.970514 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="76dec6e0-c589-4072-b65f-42ee496e0c61" containerName="nova-scheduler-scheduler" Oct 05 08:25:08 crc kubenswrapper[4846]: I1005 08:25:08.970706 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="76dec6e0-c589-4072-b65f-42ee496e0c61" containerName="nova-scheduler-scheduler" Oct 05 08:25:08 crc kubenswrapper[4846]: I1005 08:25:08.971300 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 05 08:25:08 crc kubenswrapper[4846]: I1005 08:25:08.981123 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 05 08:25:08 crc kubenswrapper[4846]: I1005 08:25:08.995631 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Oct 05 08:25:09 crc kubenswrapper[4846]: I1005 08:25:09.083818 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3309d2cf-a492-4302-9aaf-30c3c1cd3f57-config-data\") pod \"nova-scheduler-0\" (UID: \"3309d2cf-a492-4302-9aaf-30c3c1cd3f57\") " pod="openstack/nova-scheduler-0" Oct 05 08:25:09 crc kubenswrapper[4846]: I1005 08:25:09.083878 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3309d2cf-a492-4302-9aaf-30c3c1cd3f57-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"3309d2cf-a492-4302-9aaf-30c3c1cd3f57\") " pod="openstack/nova-scheduler-0" Oct 05 08:25:09 crc kubenswrapper[4846]: I1005 08:25:09.084355 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7qkkk\" (UniqueName: \"kubernetes.io/projected/3309d2cf-a492-4302-9aaf-30c3c1cd3f57-kube-api-access-7qkkk\") pod \"nova-scheduler-0\" (UID: \"3309d2cf-a492-4302-9aaf-30c3c1cd3f57\") " pod="openstack/nova-scheduler-0" Oct 05 08:25:09 crc kubenswrapper[4846]: I1005 08:25:09.186813 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7qkkk\" (UniqueName: \"kubernetes.io/projected/3309d2cf-a492-4302-9aaf-30c3c1cd3f57-kube-api-access-7qkkk\") pod \"nova-scheduler-0\" (UID: \"3309d2cf-a492-4302-9aaf-30c3c1cd3f57\") " pod="openstack/nova-scheduler-0" Oct 05 08:25:09 crc kubenswrapper[4846]: I1005 08:25:09.187009 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3309d2cf-a492-4302-9aaf-30c3c1cd3f57-config-data\") pod \"nova-scheduler-0\" (UID: \"3309d2cf-a492-4302-9aaf-30c3c1cd3f57\") " pod="openstack/nova-scheduler-0" Oct 05 08:25:09 crc kubenswrapper[4846]: I1005 08:25:09.187051 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3309d2cf-a492-4302-9aaf-30c3c1cd3f57-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"3309d2cf-a492-4302-9aaf-30c3c1cd3f57\") " pod="openstack/nova-scheduler-0" Oct 05 08:25:09 crc kubenswrapper[4846]: I1005 08:25:09.192331 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3309d2cf-a492-4302-9aaf-30c3c1cd3f57-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"3309d2cf-a492-4302-9aaf-30c3c1cd3f57\") " pod="openstack/nova-scheduler-0" Oct 05 08:25:09 crc kubenswrapper[4846]: I1005 08:25:09.197857 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3309d2cf-a492-4302-9aaf-30c3c1cd3f57-config-data\") pod \"nova-scheduler-0\" (UID: \"3309d2cf-a492-4302-9aaf-30c3c1cd3f57\") " pod="openstack/nova-scheduler-0" Oct 05 08:25:09 crc kubenswrapper[4846]: I1005 08:25:09.213195 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7qkkk\" (UniqueName: \"kubernetes.io/projected/3309d2cf-a492-4302-9aaf-30c3c1cd3f57-kube-api-access-7qkkk\") pod \"nova-scheduler-0\" (UID: \"3309d2cf-a492-4302-9aaf-30c3c1cd3f57\") " pod="openstack/nova-scheduler-0" Oct 05 08:25:09 crc kubenswrapper[4846]: I1005 08:25:09.321447 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 05 08:25:09 crc kubenswrapper[4846]: I1005 08:25:09.830488 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 05 08:25:09 crc kubenswrapper[4846]: W1005 08:25:09.831926 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3309d2cf_a492_4302_9aaf_30c3c1cd3f57.slice/crio-95bee5b37d68db33c99d15b6d7af39ba404887ee4e81b2bf1e2cafb643e3ae38 WatchSource:0}: Error finding container 95bee5b37d68db33c99d15b6d7af39ba404887ee4e81b2bf1e2cafb643e3ae38: Status 404 returned error can't find the container with id 95bee5b37d68db33c99d15b6d7af39ba404887ee4e81b2bf1e2cafb643e3ae38 Oct 05 08:25:09 crc kubenswrapper[4846]: I1005 08:25:09.915433 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"3309d2cf-a492-4302-9aaf-30c3c1cd3f57","Type":"ContainerStarted","Data":"95bee5b37d68db33c99d15b6d7af39ba404887ee4e81b2bf1e2cafb643e3ae38"} Oct 05 08:25:10 crc kubenswrapper[4846]: I1005 08:25:10.504654 4846 scope.go:117] "RemoveContainer" containerID="fdafa27da24296f9087c33b6b12a6598d2a8d61fc985b06ad680401828c9feb6" Oct 05 08:25:10 crc kubenswrapper[4846]: E1005 08:25:10.505563 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 08:25:10 crc kubenswrapper[4846]: I1005 08:25:10.530646 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="76dec6e0-c589-4072-b65f-42ee496e0c61" path="/var/lib/kubelet/pods/76dec6e0-c589-4072-b65f-42ee496e0c61/volumes" Oct 05 08:25:10 crc kubenswrapper[4846]: I1005 08:25:10.957768 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"3309d2cf-a492-4302-9aaf-30c3c1cd3f57","Type":"ContainerStarted","Data":"be07c0f8ff32f38239cb958ae582e85c424e467a057d0930617ab06cf7454997"} Oct 05 08:25:10 crc kubenswrapper[4846]: I1005 08:25:10.993460 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.99343361 podStartE2EDuration="2.99343361s" podCreationTimestamp="2025-10-05 08:25:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 08:25:10.98402267 +0000 UTC m=+5833.224875485" watchObservedRunningTime="2025-10-05 08:25:10.99343361 +0000 UTC m=+5833.234286395" Oct 05 08:25:13 crc kubenswrapper[4846]: I1005 08:25:13.179565 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 05 08:25:13 crc kubenswrapper[4846]: I1005 08:25:13.180525 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 05 08:25:13 crc kubenswrapper[4846]: I1005 08:25:13.185591 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 05 08:25:13 crc kubenswrapper[4846]: I1005 08:25:13.188629 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 05 08:25:13 crc kubenswrapper[4846]: I1005 08:25:13.191985 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 05 08:25:13 crc kubenswrapper[4846]: I1005 08:25:13.201870 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 05 08:25:13 crc kubenswrapper[4846]: I1005 08:25:13.209642 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 05 08:25:13 crc kubenswrapper[4846]: I1005 08:25:13.993268 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 05 08:25:13 crc kubenswrapper[4846]: I1005 08:25:13.998520 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 05 08:25:14 crc kubenswrapper[4846]: I1005 08:25:14.013591 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 05 08:25:14 crc kubenswrapper[4846]: I1005 08:25:14.212444 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-75b55fd8f5-x2t6q"] Oct 05 08:25:14 crc kubenswrapper[4846]: I1005 08:25:14.218958 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-75b55fd8f5-x2t6q" Oct 05 08:25:14 crc kubenswrapper[4846]: I1005 08:25:14.227821 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-75b55fd8f5-x2t6q"] Oct 05 08:25:14 crc kubenswrapper[4846]: I1005 08:25:14.315163 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fgjtm\" (UniqueName: \"kubernetes.io/projected/b0deda53-5806-4d7a-b5a6-0796d87a73e3-kube-api-access-fgjtm\") pod \"dnsmasq-dns-75b55fd8f5-x2t6q\" (UID: \"b0deda53-5806-4d7a-b5a6-0796d87a73e3\") " pod="openstack/dnsmasq-dns-75b55fd8f5-x2t6q" Oct 05 08:25:14 crc kubenswrapper[4846]: I1005 08:25:14.315298 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b0deda53-5806-4d7a-b5a6-0796d87a73e3-dns-svc\") pod \"dnsmasq-dns-75b55fd8f5-x2t6q\" (UID: \"b0deda53-5806-4d7a-b5a6-0796d87a73e3\") " pod="openstack/dnsmasq-dns-75b55fd8f5-x2t6q" Oct 05 08:25:14 crc kubenswrapper[4846]: I1005 08:25:14.315431 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b0deda53-5806-4d7a-b5a6-0796d87a73e3-config\") pod \"dnsmasq-dns-75b55fd8f5-x2t6q\" (UID: \"b0deda53-5806-4d7a-b5a6-0796d87a73e3\") " pod="openstack/dnsmasq-dns-75b55fd8f5-x2t6q" Oct 05 08:25:14 crc kubenswrapper[4846]: I1005 08:25:14.315471 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b0deda53-5806-4d7a-b5a6-0796d87a73e3-ovsdbserver-nb\") pod \"dnsmasq-dns-75b55fd8f5-x2t6q\" (UID: \"b0deda53-5806-4d7a-b5a6-0796d87a73e3\") " pod="openstack/dnsmasq-dns-75b55fd8f5-x2t6q" Oct 05 08:25:14 crc kubenswrapper[4846]: I1005 08:25:14.315501 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b0deda53-5806-4d7a-b5a6-0796d87a73e3-ovsdbserver-sb\") pod \"dnsmasq-dns-75b55fd8f5-x2t6q\" (UID: \"b0deda53-5806-4d7a-b5a6-0796d87a73e3\") " pod="openstack/dnsmasq-dns-75b55fd8f5-x2t6q" Oct 05 08:25:14 crc kubenswrapper[4846]: I1005 08:25:14.322562 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Oct 05 08:25:14 crc kubenswrapper[4846]: I1005 08:25:14.419053 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b0deda53-5806-4d7a-b5a6-0796d87a73e3-config\") pod \"dnsmasq-dns-75b55fd8f5-x2t6q\" (UID: \"b0deda53-5806-4d7a-b5a6-0796d87a73e3\") " pod="openstack/dnsmasq-dns-75b55fd8f5-x2t6q" Oct 05 08:25:14 crc kubenswrapper[4846]: I1005 08:25:14.419140 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b0deda53-5806-4d7a-b5a6-0796d87a73e3-ovsdbserver-nb\") pod \"dnsmasq-dns-75b55fd8f5-x2t6q\" (UID: \"b0deda53-5806-4d7a-b5a6-0796d87a73e3\") " pod="openstack/dnsmasq-dns-75b55fd8f5-x2t6q" Oct 05 08:25:14 crc kubenswrapper[4846]: I1005 08:25:14.419173 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b0deda53-5806-4d7a-b5a6-0796d87a73e3-ovsdbserver-sb\") pod \"dnsmasq-dns-75b55fd8f5-x2t6q\" (UID: \"b0deda53-5806-4d7a-b5a6-0796d87a73e3\") " pod="openstack/dnsmasq-dns-75b55fd8f5-x2t6q" Oct 05 08:25:14 crc kubenswrapper[4846]: I1005 08:25:14.419219 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fgjtm\" (UniqueName: \"kubernetes.io/projected/b0deda53-5806-4d7a-b5a6-0796d87a73e3-kube-api-access-fgjtm\") pod \"dnsmasq-dns-75b55fd8f5-x2t6q\" (UID: \"b0deda53-5806-4d7a-b5a6-0796d87a73e3\") " pod="openstack/dnsmasq-dns-75b55fd8f5-x2t6q" Oct 05 08:25:14 crc kubenswrapper[4846]: I1005 08:25:14.419241 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b0deda53-5806-4d7a-b5a6-0796d87a73e3-dns-svc\") pod \"dnsmasq-dns-75b55fd8f5-x2t6q\" (UID: \"b0deda53-5806-4d7a-b5a6-0796d87a73e3\") " pod="openstack/dnsmasq-dns-75b55fd8f5-x2t6q" Oct 05 08:25:14 crc kubenswrapper[4846]: I1005 08:25:14.420446 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b0deda53-5806-4d7a-b5a6-0796d87a73e3-ovsdbserver-nb\") pod \"dnsmasq-dns-75b55fd8f5-x2t6q\" (UID: \"b0deda53-5806-4d7a-b5a6-0796d87a73e3\") " pod="openstack/dnsmasq-dns-75b55fd8f5-x2t6q" Oct 05 08:25:14 crc kubenswrapper[4846]: I1005 08:25:14.420454 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b0deda53-5806-4d7a-b5a6-0796d87a73e3-dns-svc\") pod \"dnsmasq-dns-75b55fd8f5-x2t6q\" (UID: \"b0deda53-5806-4d7a-b5a6-0796d87a73e3\") " pod="openstack/dnsmasq-dns-75b55fd8f5-x2t6q" Oct 05 08:25:14 crc kubenswrapper[4846]: I1005 08:25:14.420698 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b0deda53-5806-4d7a-b5a6-0796d87a73e3-ovsdbserver-sb\") pod \"dnsmasq-dns-75b55fd8f5-x2t6q\" (UID: \"b0deda53-5806-4d7a-b5a6-0796d87a73e3\") " pod="openstack/dnsmasq-dns-75b55fd8f5-x2t6q" Oct 05 08:25:14 crc kubenswrapper[4846]: I1005 08:25:14.420910 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b0deda53-5806-4d7a-b5a6-0796d87a73e3-config\") pod \"dnsmasq-dns-75b55fd8f5-x2t6q\" (UID: \"b0deda53-5806-4d7a-b5a6-0796d87a73e3\") " pod="openstack/dnsmasq-dns-75b55fd8f5-x2t6q" Oct 05 08:25:14 crc kubenswrapper[4846]: I1005 08:25:14.439282 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fgjtm\" (UniqueName: \"kubernetes.io/projected/b0deda53-5806-4d7a-b5a6-0796d87a73e3-kube-api-access-fgjtm\") pod \"dnsmasq-dns-75b55fd8f5-x2t6q\" (UID: \"b0deda53-5806-4d7a-b5a6-0796d87a73e3\") " pod="openstack/dnsmasq-dns-75b55fd8f5-x2t6q" Oct 05 08:25:14 crc kubenswrapper[4846]: I1005 08:25:14.550065 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-75b55fd8f5-x2t6q" Oct 05 08:25:15 crc kubenswrapper[4846]: I1005 08:25:15.094885 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-75b55fd8f5-x2t6q"] Oct 05 08:25:16 crc kubenswrapper[4846]: I1005 08:25:16.010303 4846 generic.go:334] "Generic (PLEG): container finished" podID="b0deda53-5806-4d7a-b5a6-0796d87a73e3" containerID="ed5b9f4aa40a4911065311eb03e48f6bfa47c1f5a7e4b941da19575205491ce8" exitCode=0 Oct 05 08:25:16 crc kubenswrapper[4846]: I1005 08:25:16.011374 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-75b55fd8f5-x2t6q" event={"ID":"b0deda53-5806-4d7a-b5a6-0796d87a73e3","Type":"ContainerDied","Data":"ed5b9f4aa40a4911065311eb03e48f6bfa47c1f5a7e4b941da19575205491ce8"} Oct 05 08:25:16 crc kubenswrapper[4846]: I1005 08:25:16.011719 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-75b55fd8f5-x2t6q" event={"ID":"b0deda53-5806-4d7a-b5a6-0796d87a73e3","Type":"ContainerStarted","Data":"23e29b503c3189c910e436dde2125f0b73676ee92e98d1a0b876d441ba20ea4d"} Oct 05 08:25:17 crc kubenswrapper[4846]: I1005 08:25:17.026632 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-75b55fd8f5-x2t6q" event={"ID":"b0deda53-5806-4d7a-b5a6-0796d87a73e3","Type":"ContainerStarted","Data":"246752f532c50fb2a246b0125c3f3caa7558862206b89fd92ac714fe9b916ab2"} Oct 05 08:25:17 crc kubenswrapper[4846]: I1005 08:25:17.028051 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-75b55fd8f5-x2t6q" Oct 05 08:25:17 crc kubenswrapper[4846]: I1005 08:25:17.063957 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-75b55fd8f5-x2t6q" podStartSLOduration=3.063930381 podStartE2EDuration="3.063930381s" podCreationTimestamp="2025-10-05 08:25:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 08:25:17.054428789 +0000 UTC m=+5839.295281624" watchObservedRunningTime="2025-10-05 08:25:17.063930381 +0000 UTC m=+5839.304783216" Oct 05 08:25:17 crc kubenswrapper[4846]: I1005 08:25:17.092477 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 05 08:25:17 crc kubenswrapper[4846]: I1005 08:25:17.092735 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="d94b0d13-b52b-4fcb-b016-7d5c91f72600" containerName="nova-api-log" containerID="cri-o://a98d5bd27987abf87bfcba4469cec38ec4b051938f756192a742e3b0b24e5589" gracePeriod=30 Oct 05 08:25:17 crc kubenswrapper[4846]: I1005 08:25:17.092925 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="d94b0d13-b52b-4fcb-b016-7d5c91f72600" containerName="nova-api-api" containerID="cri-o://6f2689b0dec92601169c548cffb2d10afbe1f569a214735f18107ae080a3e631" gracePeriod=30 Oct 05 08:25:18 crc kubenswrapper[4846]: I1005 08:25:18.036631 4846 generic.go:334] "Generic (PLEG): container finished" podID="d94b0d13-b52b-4fcb-b016-7d5c91f72600" containerID="a98d5bd27987abf87bfcba4469cec38ec4b051938f756192a742e3b0b24e5589" exitCode=143 Oct 05 08:25:18 crc kubenswrapper[4846]: I1005 08:25:18.036725 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"d94b0d13-b52b-4fcb-b016-7d5c91f72600","Type":"ContainerDied","Data":"a98d5bd27987abf87bfcba4469cec38ec4b051938f756192a742e3b0b24e5589"} Oct 05 08:25:19 crc kubenswrapper[4846]: I1005 08:25:19.322170 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Oct 05 08:25:19 crc kubenswrapper[4846]: I1005 08:25:19.374976 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Oct 05 08:25:20 crc kubenswrapper[4846]: I1005 08:25:20.108837 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Oct 05 08:25:20 crc kubenswrapper[4846]: I1005 08:25:20.695865 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 05 08:25:20 crc kubenswrapper[4846]: I1005 08:25:20.852627 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d94b0d13-b52b-4fcb-b016-7d5c91f72600-combined-ca-bundle\") pod \"d94b0d13-b52b-4fcb-b016-7d5c91f72600\" (UID: \"d94b0d13-b52b-4fcb-b016-7d5c91f72600\") " Oct 05 08:25:20 crc kubenswrapper[4846]: I1005 08:25:20.852800 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d94b0d13-b52b-4fcb-b016-7d5c91f72600-config-data\") pod \"d94b0d13-b52b-4fcb-b016-7d5c91f72600\" (UID: \"d94b0d13-b52b-4fcb-b016-7d5c91f72600\") " Oct 05 08:25:20 crc kubenswrapper[4846]: I1005 08:25:20.852921 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d94b0d13-b52b-4fcb-b016-7d5c91f72600-logs\") pod \"d94b0d13-b52b-4fcb-b016-7d5c91f72600\" (UID: \"d94b0d13-b52b-4fcb-b016-7d5c91f72600\") " Oct 05 08:25:20 crc kubenswrapper[4846]: I1005 08:25:20.852982 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4v2x9\" (UniqueName: \"kubernetes.io/projected/d94b0d13-b52b-4fcb-b016-7d5c91f72600-kube-api-access-4v2x9\") pod \"d94b0d13-b52b-4fcb-b016-7d5c91f72600\" (UID: \"d94b0d13-b52b-4fcb-b016-7d5c91f72600\") " Oct 05 08:25:20 crc kubenswrapper[4846]: I1005 08:25:20.853735 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d94b0d13-b52b-4fcb-b016-7d5c91f72600-logs" (OuterVolumeSpecName: "logs") pod "d94b0d13-b52b-4fcb-b016-7d5c91f72600" (UID: "d94b0d13-b52b-4fcb-b016-7d5c91f72600"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 08:25:20 crc kubenswrapper[4846]: I1005 08:25:20.871378 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d94b0d13-b52b-4fcb-b016-7d5c91f72600-kube-api-access-4v2x9" (OuterVolumeSpecName: "kube-api-access-4v2x9") pod "d94b0d13-b52b-4fcb-b016-7d5c91f72600" (UID: "d94b0d13-b52b-4fcb-b016-7d5c91f72600"). InnerVolumeSpecName "kube-api-access-4v2x9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:25:20 crc kubenswrapper[4846]: I1005 08:25:20.896447 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d94b0d13-b52b-4fcb-b016-7d5c91f72600-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d94b0d13-b52b-4fcb-b016-7d5c91f72600" (UID: "d94b0d13-b52b-4fcb-b016-7d5c91f72600"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:25:20 crc kubenswrapper[4846]: I1005 08:25:20.902041 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d94b0d13-b52b-4fcb-b016-7d5c91f72600-config-data" (OuterVolumeSpecName: "config-data") pod "d94b0d13-b52b-4fcb-b016-7d5c91f72600" (UID: "d94b0d13-b52b-4fcb-b016-7d5c91f72600"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:25:20 crc kubenswrapper[4846]: I1005 08:25:20.955927 4846 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d94b0d13-b52b-4fcb-b016-7d5c91f72600-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 08:25:20 crc kubenswrapper[4846]: I1005 08:25:20.956281 4846 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d94b0d13-b52b-4fcb-b016-7d5c91f72600-logs\") on node \"crc\" DevicePath \"\"" Oct 05 08:25:20 crc kubenswrapper[4846]: I1005 08:25:20.956292 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4v2x9\" (UniqueName: \"kubernetes.io/projected/d94b0d13-b52b-4fcb-b016-7d5c91f72600-kube-api-access-4v2x9\") on node \"crc\" DevicePath \"\"" Oct 05 08:25:20 crc kubenswrapper[4846]: I1005 08:25:20.956302 4846 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d94b0d13-b52b-4fcb-b016-7d5c91f72600-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 08:25:21 crc kubenswrapper[4846]: I1005 08:25:21.072379 4846 generic.go:334] "Generic (PLEG): container finished" podID="d94b0d13-b52b-4fcb-b016-7d5c91f72600" containerID="6f2689b0dec92601169c548cffb2d10afbe1f569a214735f18107ae080a3e631" exitCode=0 Oct 05 08:25:21 crc kubenswrapper[4846]: I1005 08:25:21.072443 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"d94b0d13-b52b-4fcb-b016-7d5c91f72600","Type":"ContainerDied","Data":"6f2689b0dec92601169c548cffb2d10afbe1f569a214735f18107ae080a3e631"} Oct 05 08:25:21 crc kubenswrapper[4846]: I1005 08:25:21.072494 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 05 08:25:21 crc kubenswrapper[4846]: I1005 08:25:21.072510 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"d94b0d13-b52b-4fcb-b016-7d5c91f72600","Type":"ContainerDied","Data":"8efde6ebdd901237d1986d669796f82020e9be1f7925321118c661e34fcff0ed"} Oct 05 08:25:21 crc kubenswrapper[4846]: I1005 08:25:21.072549 4846 scope.go:117] "RemoveContainer" containerID="6f2689b0dec92601169c548cffb2d10afbe1f569a214735f18107ae080a3e631" Oct 05 08:25:21 crc kubenswrapper[4846]: I1005 08:25:21.105378 4846 scope.go:117] "RemoveContainer" containerID="a98d5bd27987abf87bfcba4469cec38ec4b051938f756192a742e3b0b24e5589" Oct 05 08:25:21 crc kubenswrapper[4846]: I1005 08:25:21.115310 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 05 08:25:21 crc kubenswrapper[4846]: I1005 08:25:21.124062 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Oct 05 08:25:21 crc kubenswrapper[4846]: I1005 08:25:21.140973 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 05 08:25:21 crc kubenswrapper[4846]: E1005 08:25:21.141378 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d94b0d13-b52b-4fcb-b016-7d5c91f72600" containerName="nova-api-log" Oct 05 08:25:21 crc kubenswrapper[4846]: I1005 08:25:21.141391 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="d94b0d13-b52b-4fcb-b016-7d5c91f72600" containerName="nova-api-log" Oct 05 08:25:21 crc kubenswrapper[4846]: E1005 08:25:21.141403 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d94b0d13-b52b-4fcb-b016-7d5c91f72600" containerName="nova-api-api" Oct 05 08:25:21 crc kubenswrapper[4846]: I1005 08:25:21.141409 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="d94b0d13-b52b-4fcb-b016-7d5c91f72600" containerName="nova-api-api" Oct 05 08:25:21 crc kubenswrapper[4846]: I1005 08:25:21.141583 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="d94b0d13-b52b-4fcb-b016-7d5c91f72600" containerName="nova-api-log" Oct 05 08:25:21 crc kubenswrapper[4846]: I1005 08:25:21.141605 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="d94b0d13-b52b-4fcb-b016-7d5c91f72600" containerName="nova-api-api" Oct 05 08:25:21 crc kubenswrapper[4846]: I1005 08:25:21.142906 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 05 08:25:21 crc kubenswrapper[4846]: I1005 08:25:21.144433 4846 scope.go:117] "RemoveContainer" containerID="6f2689b0dec92601169c548cffb2d10afbe1f569a214735f18107ae080a3e631" Oct 05 08:25:21 crc kubenswrapper[4846]: I1005 08:25:21.145393 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 05 08:25:21 crc kubenswrapper[4846]: I1005 08:25:21.145542 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Oct 05 08:25:21 crc kubenswrapper[4846]: I1005 08:25:21.145825 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Oct 05 08:25:21 crc kubenswrapper[4846]: E1005 08:25:21.146806 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6f2689b0dec92601169c548cffb2d10afbe1f569a214735f18107ae080a3e631\": container with ID starting with 6f2689b0dec92601169c548cffb2d10afbe1f569a214735f18107ae080a3e631 not found: ID does not exist" containerID="6f2689b0dec92601169c548cffb2d10afbe1f569a214735f18107ae080a3e631" Oct 05 08:25:21 crc kubenswrapper[4846]: I1005 08:25:21.146834 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6f2689b0dec92601169c548cffb2d10afbe1f569a214735f18107ae080a3e631"} err="failed to get container status \"6f2689b0dec92601169c548cffb2d10afbe1f569a214735f18107ae080a3e631\": rpc error: code = NotFound desc = could not find container \"6f2689b0dec92601169c548cffb2d10afbe1f569a214735f18107ae080a3e631\": container with ID starting with 6f2689b0dec92601169c548cffb2d10afbe1f569a214735f18107ae080a3e631 not found: ID does not exist" Oct 05 08:25:21 crc kubenswrapper[4846]: I1005 08:25:21.146852 4846 scope.go:117] "RemoveContainer" containerID="a98d5bd27987abf87bfcba4469cec38ec4b051938f756192a742e3b0b24e5589" Oct 05 08:25:21 crc kubenswrapper[4846]: E1005 08:25:21.147225 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a98d5bd27987abf87bfcba4469cec38ec4b051938f756192a742e3b0b24e5589\": container with ID starting with a98d5bd27987abf87bfcba4469cec38ec4b051938f756192a742e3b0b24e5589 not found: ID does not exist" containerID="a98d5bd27987abf87bfcba4469cec38ec4b051938f756192a742e3b0b24e5589" Oct 05 08:25:21 crc kubenswrapper[4846]: I1005 08:25:21.147244 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a98d5bd27987abf87bfcba4469cec38ec4b051938f756192a742e3b0b24e5589"} err="failed to get container status \"a98d5bd27987abf87bfcba4469cec38ec4b051938f756192a742e3b0b24e5589\": rpc error: code = NotFound desc = could not find container \"a98d5bd27987abf87bfcba4469cec38ec4b051938f756192a742e3b0b24e5589\": container with ID starting with a98d5bd27987abf87bfcba4469cec38ec4b051938f756192a742e3b0b24e5589 not found: ID does not exist" Oct 05 08:25:21 crc kubenswrapper[4846]: I1005 08:25:21.184956 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 05 08:25:21 crc kubenswrapper[4846]: I1005 08:25:21.261670 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/29bec123-d7d4-47d6-ae64-68b2aab524e4-config-data\") pod \"nova-api-0\" (UID: \"29bec123-d7d4-47d6-ae64-68b2aab524e4\") " pod="openstack/nova-api-0" Oct 05 08:25:21 crc kubenswrapper[4846]: I1005 08:25:21.261772 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/29bec123-d7d4-47d6-ae64-68b2aab524e4-public-tls-certs\") pod \"nova-api-0\" (UID: \"29bec123-d7d4-47d6-ae64-68b2aab524e4\") " pod="openstack/nova-api-0" Oct 05 08:25:21 crc kubenswrapper[4846]: I1005 08:25:21.261807 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rnbd4\" (UniqueName: \"kubernetes.io/projected/29bec123-d7d4-47d6-ae64-68b2aab524e4-kube-api-access-rnbd4\") pod \"nova-api-0\" (UID: \"29bec123-d7d4-47d6-ae64-68b2aab524e4\") " pod="openstack/nova-api-0" Oct 05 08:25:21 crc kubenswrapper[4846]: I1005 08:25:21.261856 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/29bec123-d7d4-47d6-ae64-68b2aab524e4-internal-tls-certs\") pod \"nova-api-0\" (UID: \"29bec123-d7d4-47d6-ae64-68b2aab524e4\") " pod="openstack/nova-api-0" Oct 05 08:25:21 crc kubenswrapper[4846]: I1005 08:25:21.261880 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/29bec123-d7d4-47d6-ae64-68b2aab524e4-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"29bec123-d7d4-47d6-ae64-68b2aab524e4\") " pod="openstack/nova-api-0" Oct 05 08:25:21 crc kubenswrapper[4846]: I1005 08:25:21.261902 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/29bec123-d7d4-47d6-ae64-68b2aab524e4-logs\") pod \"nova-api-0\" (UID: \"29bec123-d7d4-47d6-ae64-68b2aab524e4\") " pod="openstack/nova-api-0" Oct 05 08:25:21 crc kubenswrapper[4846]: I1005 08:25:21.364047 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/29bec123-d7d4-47d6-ae64-68b2aab524e4-public-tls-certs\") pod \"nova-api-0\" (UID: \"29bec123-d7d4-47d6-ae64-68b2aab524e4\") " pod="openstack/nova-api-0" Oct 05 08:25:21 crc kubenswrapper[4846]: I1005 08:25:21.364110 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rnbd4\" (UniqueName: \"kubernetes.io/projected/29bec123-d7d4-47d6-ae64-68b2aab524e4-kube-api-access-rnbd4\") pod \"nova-api-0\" (UID: \"29bec123-d7d4-47d6-ae64-68b2aab524e4\") " pod="openstack/nova-api-0" Oct 05 08:25:21 crc kubenswrapper[4846]: I1005 08:25:21.364167 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/29bec123-d7d4-47d6-ae64-68b2aab524e4-internal-tls-certs\") pod \"nova-api-0\" (UID: \"29bec123-d7d4-47d6-ae64-68b2aab524e4\") " pod="openstack/nova-api-0" Oct 05 08:25:21 crc kubenswrapper[4846]: I1005 08:25:21.364215 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/29bec123-d7d4-47d6-ae64-68b2aab524e4-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"29bec123-d7d4-47d6-ae64-68b2aab524e4\") " pod="openstack/nova-api-0" Oct 05 08:25:21 crc kubenswrapper[4846]: I1005 08:25:21.364243 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/29bec123-d7d4-47d6-ae64-68b2aab524e4-logs\") pod \"nova-api-0\" (UID: \"29bec123-d7d4-47d6-ae64-68b2aab524e4\") " pod="openstack/nova-api-0" Oct 05 08:25:21 crc kubenswrapper[4846]: I1005 08:25:21.364311 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/29bec123-d7d4-47d6-ae64-68b2aab524e4-config-data\") pod \"nova-api-0\" (UID: \"29bec123-d7d4-47d6-ae64-68b2aab524e4\") " pod="openstack/nova-api-0" Oct 05 08:25:21 crc kubenswrapper[4846]: I1005 08:25:21.365843 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/29bec123-d7d4-47d6-ae64-68b2aab524e4-logs\") pod \"nova-api-0\" (UID: \"29bec123-d7d4-47d6-ae64-68b2aab524e4\") " pod="openstack/nova-api-0" Oct 05 08:25:21 crc kubenswrapper[4846]: I1005 08:25:21.370132 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/29bec123-d7d4-47d6-ae64-68b2aab524e4-config-data\") pod \"nova-api-0\" (UID: \"29bec123-d7d4-47d6-ae64-68b2aab524e4\") " pod="openstack/nova-api-0" Oct 05 08:25:21 crc kubenswrapper[4846]: I1005 08:25:21.385453 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/29bec123-d7d4-47d6-ae64-68b2aab524e4-public-tls-certs\") pod \"nova-api-0\" (UID: \"29bec123-d7d4-47d6-ae64-68b2aab524e4\") " pod="openstack/nova-api-0" Oct 05 08:25:21 crc kubenswrapper[4846]: I1005 08:25:21.387829 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/29bec123-d7d4-47d6-ae64-68b2aab524e4-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"29bec123-d7d4-47d6-ae64-68b2aab524e4\") " pod="openstack/nova-api-0" Oct 05 08:25:21 crc kubenswrapper[4846]: I1005 08:25:21.389754 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rnbd4\" (UniqueName: \"kubernetes.io/projected/29bec123-d7d4-47d6-ae64-68b2aab524e4-kube-api-access-rnbd4\") pod \"nova-api-0\" (UID: \"29bec123-d7d4-47d6-ae64-68b2aab524e4\") " pod="openstack/nova-api-0" Oct 05 08:25:21 crc kubenswrapper[4846]: I1005 08:25:21.397030 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/29bec123-d7d4-47d6-ae64-68b2aab524e4-internal-tls-certs\") pod \"nova-api-0\" (UID: \"29bec123-d7d4-47d6-ae64-68b2aab524e4\") " pod="openstack/nova-api-0" Oct 05 08:25:21 crc kubenswrapper[4846]: I1005 08:25:21.471401 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 05 08:25:21 crc kubenswrapper[4846]: I1005 08:25:21.498395 4846 scope.go:117] "RemoveContainer" containerID="fdafa27da24296f9087c33b6b12a6598d2a8d61fc985b06ad680401828c9feb6" Oct 05 08:25:21 crc kubenswrapper[4846]: E1005 08:25:21.498676 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 08:25:21 crc kubenswrapper[4846]: I1005 08:25:21.994878 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 05 08:25:22 crc kubenswrapper[4846]: I1005 08:25:22.084315 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"29bec123-d7d4-47d6-ae64-68b2aab524e4","Type":"ContainerStarted","Data":"969d76578245005e846856a5ed7bfb101ce7d29d780cc01d9b4d53ae750adb14"} Oct 05 08:25:22 crc kubenswrapper[4846]: I1005 08:25:22.510483 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d94b0d13-b52b-4fcb-b016-7d5c91f72600" path="/var/lib/kubelet/pods/d94b0d13-b52b-4fcb-b016-7d5c91f72600/volumes" Oct 05 08:25:23 crc kubenswrapper[4846]: I1005 08:25:23.098338 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"29bec123-d7d4-47d6-ae64-68b2aab524e4","Type":"ContainerStarted","Data":"0b1b66903a1ae5452af5f6f5c676e5e72227f1509e9438272a22b31171c8814c"} Oct 05 08:25:23 crc kubenswrapper[4846]: I1005 08:25:23.098412 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"29bec123-d7d4-47d6-ae64-68b2aab524e4","Type":"ContainerStarted","Data":"a472b6104863e77453086f959d88f81c611f09276e87a825918364149a86c5b2"} Oct 05 08:25:24 crc kubenswrapper[4846]: I1005 08:25:24.552559 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-75b55fd8f5-x2t6q" Oct 05 08:25:24 crc kubenswrapper[4846]: I1005 08:25:24.587310 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=3.587283894 podStartE2EDuration="3.587283894s" podCreationTimestamp="2025-10-05 08:25:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 08:25:23.125721422 +0000 UTC m=+5845.366574267" watchObservedRunningTime="2025-10-05 08:25:24.587283894 +0000 UTC m=+5846.828136709" Oct 05 08:25:24 crc kubenswrapper[4846]: I1005 08:25:24.631857 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-774686bfb5-gk9ch"] Oct 05 08:25:24 crc kubenswrapper[4846]: I1005 08:25:24.636251 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-774686bfb5-gk9ch" podUID="fe918116-df47-4b09-977a-0f16a9cd5e75" containerName="dnsmasq-dns" containerID="cri-o://613fdc9574b6e854ad675725a919d6f0abac840992d9db4ceffa46ccef85db43" gracePeriod=10 Oct 05 08:25:25 crc kubenswrapper[4846]: I1005 08:25:25.092755 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-774686bfb5-gk9ch" Oct 05 08:25:25 crc kubenswrapper[4846]: I1005 08:25:25.126587 4846 generic.go:334] "Generic (PLEG): container finished" podID="fe918116-df47-4b09-977a-0f16a9cd5e75" containerID="613fdc9574b6e854ad675725a919d6f0abac840992d9db4ceffa46ccef85db43" exitCode=0 Oct 05 08:25:25 crc kubenswrapper[4846]: I1005 08:25:25.126630 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-774686bfb5-gk9ch" event={"ID":"fe918116-df47-4b09-977a-0f16a9cd5e75","Type":"ContainerDied","Data":"613fdc9574b6e854ad675725a919d6f0abac840992d9db4ceffa46ccef85db43"} Oct 05 08:25:25 crc kubenswrapper[4846]: I1005 08:25:25.126656 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-774686bfb5-gk9ch" event={"ID":"fe918116-df47-4b09-977a-0f16a9cd5e75","Type":"ContainerDied","Data":"9fe4b5bce5a8984f034964dd685870e3615e59fd43751e92fb3594aa4cf58cbe"} Oct 05 08:25:25 crc kubenswrapper[4846]: I1005 08:25:25.126672 4846 scope.go:117] "RemoveContainer" containerID="613fdc9574b6e854ad675725a919d6f0abac840992d9db4ceffa46ccef85db43" Oct 05 08:25:25 crc kubenswrapper[4846]: I1005 08:25:25.126809 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-774686bfb5-gk9ch" Oct 05 08:25:25 crc kubenswrapper[4846]: I1005 08:25:25.150339 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fe918116-df47-4b09-977a-0f16a9cd5e75-dns-svc\") pod \"fe918116-df47-4b09-977a-0f16a9cd5e75\" (UID: \"fe918116-df47-4b09-977a-0f16a9cd5e75\") " Oct 05 08:25:25 crc kubenswrapper[4846]: I1005 08:25:25.150411 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/fe918116-df47-4b09-977a-0f16a9cd5e75-ovsdbserver-nb\") pod \"fe918116-df47-4b09-977a-0f16a9cd5e75\" (UID: \"fe918116-df47-4b09-977a-0f16a9cd5e75\") " Oct 05 08:25:25 crc kubenswrapper[4846]: I1005 08:25:25.150437 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j2q2j\" (UniqueName: \"kubernetes.io/projected/fe918116-df47-4b09-977a-0f16a9cd5e75-kube-api-access-j2q2j\") pod \"fe918116-df47-4b09-977a-0f16a9cd5e75\" (UID: \"fe918116-df47-4b09-977a-0f16a9cd5e75\") " Oct 05 08:25:25 crc kubenswrapper[4846]: I1005 08:25:25.150464 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/fe918116-df47-4b09-977a-0f16a9cd5e75-ovsdbserver-sb\") pod \"fe918116-df47-4b09-977a-0f16a9cd5e75\" (UID: \"fe918116-df47-4b09-977a-0f16a9cd5e75\") " Oct 05 08:25:25 crc kubenswrapper[4846]: I1005 08:25:25.150694 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fe918116-df47-4b09-977a-0f16a9cd5e75-config\") pod \"fe918116-df47-4b09-977a-0f16a9cd5e75\" (UID: \"fe918116-df47-4b09-977a-0f16a9cd5e75\") " Oct 05 08:25:25 crc kubenswrapper[4846]: I1005 08:25:25.161441 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fe918116-df47-4b09-977a-0f16a9cd5e75-kube-api-access-j2q2j" (OuterVolumeSpecName: "kube-api-access-j2q2j") pod "fe918116-df47-4b09-977a-0f16a9cd5e75" (UID: "fe918116-df47-4b09-977a-0f16a9cd5e75"). InnerVolumeSpecName "kube-api-access-j2q2j". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:25:25 crc kubenswrapper[4846]: I1005 08:25:25.179724 4846 scope.go:117] "RemoveContainer" containerID="a5603d53b137d2ae2b7e74a3aec98074f8b5806e3c1a208442e8b69426282119" Oct 05 08:25:25 crc kubenswrapper[4846]: I1005 08:25:25.218894 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fe918116-df47-4b09-977a-0f16a9cd5e75-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "fe918116-df47-4b09-977a-0f16a9cd5e75" (UID: "fe918116-df47-4b09-977a-0f16a9cd5e75"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 08:25:25 crc kubenswrapper[4846]: I1005 08:25:25.220582 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fe918116-df47-4b09-977a-0f16a9cd5e75-config" (OuterVolumeSpecName: "config") pod "fe918116-df47-4b09-977a-0f16a9cd5e75" (UID: "fe918116-df47-4b09-977a-0f16a9cd5e75"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 08:25:25 crc kubenswrapper[4846]: I1005 08:25:25.222303 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fe918116-df47-4b09-977a-0f16a9cd5e75-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "fe918116-df47-4b09-977a-0f16a9cd5e75" (UID: "fe918116-df47-4b09-977a-0f16a9cd5e75"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 08:25:25 crc kubenswrapper[4846]: I1005 08:25:25.233921 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fe918116-df47-4b09-977a-0f16a9cd5e75-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "fe918116-df47-4b09-977a-0f16a9cd5e75" (UID: "fe918116-df47-4b09-977a-0f16a9cd5e75"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 08:25:25 crc kubenswrapper[4846]: I1005 08:25:25.255853 4846 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/fe918116-df47-4b09-977a-0f16a9cd5e75-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 05 08:25:25 crc kubenswrapper[4846]: I1005 08:25:25.255894 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j2q2j\" (UniqueName: \"kubernetes.io/projected/fe918116-df47-4b09-977a-0f16a9cd5e75-kube-api-access-j2q2j\") on node \"crc\" DevicePath \"\"" Oct 05 08:25:25 crc kubenswrapper[4846]: I1005 08:25:25.255907 4846 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/fe918116-df47-4b09-977a-0f16a9cd5e75-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 05 08:25:25 crc kubenswrapper[4846]: I1005 08:25:25.255919 4846 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fe918116-df47-4b09-977a-0f16a9cd5e75-config\") on node \"crc\" DevicePath \"\"" Oct 05 08:25:25 crc kubenswrapper[4846]: I1005 08:25:25.255933 4846 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fe918116-df47-4b09-977a-0f16a9cd5e75-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 05 08:25:25 crc kubenswrapper[4846]: I1005 08:25:25.261110 4846 scope.go:117] "RemoveContainer" containerID="613fdc9574b6e854ad675725a919d6f0abac840992d9db4ceffa46ccef85db43" Oct 05 08:25:25 crc kubenswrapper[4846]: E1005 08:25:25.261610 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"613fdc9574b6e854ad675725a919d6f0abac840992d9db4ceffa46ccef85db43\": container with ID starting with 613fdc9574b6e854ad675725a919d6f0abac840992d9db4ceffa46ccef85db43 not found: ID does not exist" containerID="613fdc9574b6e854ad675725a919d6f0abac840992d9db4ceffa46ccef85db43" Oct 05 08:25:25 crc kubenswrapper[4846]: I1005 08:25:25.261720 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"613fdc9574b6e854ad675725a919d6f0abac840992d9db4ceffa46ccef85db43"} err="failed to get container status \"613fdc9574b6e854ad675725a919d6f0abac840992d9db4ceffa46ccef85db43\": rpc error: code = NotFound desc = could not find container \"613fdc9574b6e854ad675725a919d6f0abac840992d9db4ceffa46ccef85db43\": container with ID starting with 613fdc9574b6e854ad675725a919d6f0abac840992d9db4ceffa46ccef85db43 not found: ID does not exist" Oct 05 08:25:25 crc kubenswrapper[4846]: I1005 08:25:25.261806 4846 scope.go:117] "RemoveContainer" containerID="a5603d53b137d2ae2b7e74a3aec98074f8b5806e3c1a208442e8b69426282119" Oct 05 08:25:25 crc kubenswrapper[4846]: E1005 08:25:25.262095 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a5603d53b137d2ae2b7e74a3aec98074f8b5806e3c1a208442e8b69426282119\": container with ID starting with a5603d53b137d2ae2b7e74a3aec98074f8b5806e3c1a208442e8b69426282119 not found: ID does not exist" containerID="a5603d53b137d2ae2b7e74a3aec98074f8b5806e3c1a208442e8b69426282119" Oct 05 08:25:25 crc kubenswrapper[4846]: I1005 08:25:25.262210 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a5603d53b137d2ae2b7e74a3aec98074f8b5806e3c1a208442e8b69426282119"} err="failed to get container status \"a5603d53b137d2ae2b7e74a3aec98074f8b5806e3c1a208442e8b69426282119\": rpc error: code = NotFound desc = could not find container \"a5603d53b137d2ae2b7e74a3aec98074f8b5806e3c1a208442e8b69426282119\": container with ID starting with a5603d53b137d2ae2b7e74a3aec98074f8b5806e3c1a208442e8b69426282119 not found: ID does not exist" Oct 05 08:25:25 crc kubenswrapper[4846]: I1005 08:25:25.463657 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-774686bfb5-gk9ch"] Oct 05 08:25:25 crc kubenswrapper[4846]: I1005 08:25:25.470460 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-774686bfb5-gk9ch"] Oct 05 08:25:26 crc kubenswrapper[4846]: I1005 08:25:26.519561 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fe918116-df47-4b09-977a-0f16a9cd5e75" path="/var/lib/kubelet/pods/fe918116-df47-4b09-977a-0f16a9cd5e75/volumes" Oct 05 08:25:31 crc kubenswrapper[4846]: I1005 08:25:31.471680 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 05 08:25:31 crc kubenswrapper[4846]: I1005 08:25:31.472228 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 05 08:25:32 crc kubenswrapper[4846]: I1005 08:25:32.492515 4846 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="29bec123-d7d4-47d6-ae64-68b2aab524e4" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.1.99:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 05 08:25:32 crc kubenswrapper[4846]: I1005 08:25:32.492542 4846 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="29bec123-d7d4-47d6-ae64-68b2aab524e4" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.1.99:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 05 08:25:33 crc kubenswrapper[4846]: I1005 08:25:33.497905 4846 scope.go:117] "RemoveContainer" containerID="fdafa27da24296f9087c33b6b12a6598d2a8d61fc985b06ad680401828c9feb6" Oct 05 08:25:33 crc kubenswrapper[4846]: E1005 08:25:33.498424 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 08:25:41 crc kubenswrapper[4846]: I1005 08:25:41.482848 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 05 08:25:41 crc kubenswrapper[4846]: I1005 08:25:41.484804 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 05 08:25:41 crc kubenswrapper[4846]: I1005 08:25:41.485003 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 05 08:25:41 crc kubenswrapper[4846]: I1005 08:25:41.493607 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 05 08:25:42 crc kubenswrapper[4846]: I1005 08:25:42.345172 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 05 08:25:42 crc kubenswrapper[4846]: I1005 08:25:42.352091 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 05 08:25:46 crc kubenswrapper[4846]: I1005 08:25:46.498690 4846 scope.go:117] "RemoveContainer" containerID="fdafa27da24296f9087c33b6b12a6598d2a8d61fc985b06ad680401828c9feb6" Oct 05 08:25:46 crc kubenswrapper[4846]: E1005 08:25:46.499801 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 08:25:53 crc kubenswrapper[4846]: I1005 08:25:53.039114 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-create-c2xrh"] Oct 05 08:25:53 crc kubenswrapper[4846]: I1005 08:25:53.047752 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-create-c2xrh"] Oct 05 08:25:53 crc kubenswrapper[4846]: I1005 08:25:53.937323 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-57d69df789-t686p"] Oct 05 08:25:53 crc kubenswrapper[4846]: E1005 08:25:53.958901 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fe918116-df47-4b09-977a-0f16a9cd5e75" containerName="init" Oct 05 08:25:53 crc kubenswrapper[4846]: I1005 08:25:53.958944 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="fe918116-df47-4b09-977a-0f16a9cd5e75" containerName="init" Oct 05 08:25:53 crc kubenswrapper[4846]: E1005 08:25:53.967758 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fe918116-df47-4b09-977a-0f16a9cd5e75" containerName="dnsmasq-dns" Oct 05 08:25:53 crc kubenswrapper[4846]: I1005 08:25:53.967792 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="fe918116-df47-4b09-977a-0f16a9cd5e75" containerName="dnsmasq-dns" Oct 05 08:25:53 crc kubenswrapper[4846]: I1005 08:25:53.978850 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="fe918116-df47-4b09-977a-0f16a9cd5e75" containerName="dnsmasq-dns" Oct 05 08:25:53 crc kubenswrapper[4846]: I1005 08:25:53.989270 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-57d69df789-t686p"] Oct 05 08:25:53 crc kubenswrapper[4846]: I1005 08:25:53.989429 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-57d69df789-t686p" Oct 05 08:25:53 crc kubenswrapper[4846]: I1005 08:25:53.991419 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon" Oct 05 08:25:53 crc kubenswrapper[4846]: I1005 08:25:53.992243 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon-horizon-dockercfg-nzd2c" Oct 05 08:25:53 crc kubenswrapper[4846]: I1005 08:25:53.992814 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-config-data" Oct 05 08:25:53 crc kubenswrapper[4846]: I1005 08:25:53.993627 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-scripts" Oct 05 08:25:54 crc kubenswrapper[4846]: I1005 08:25:54.016554 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 05 08:25:54 crc kubenswrapper[4846]: I1005 08:25:54.017172 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="c77ecef6-3a1d-43d5-90b1-dd3de33d7de2" containerName="glance-log" containerID="cri-o://d0b2a95a1c72c58f215d678122d3d9a70d119c082300731b85c1b4d15fcb5ab0" gracePeriod=30 Oct 05 08:25:54 crc kubenswrapper[4846]: I1005 08:25:54.017389 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="c77ecef6-3a1d-43d5-90b1-dd3de33d7de2" containerName="glance-httpd" containerID="cri-o://8e3d24f7a9c614dd76a5fe862718bcd0dda3735120671933705a9bef54139d4f" gracePeriod=30 Oct 05 08:25:54 crc kubenswrapper[4846]: I1005 08:25:54.067620 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-896999bb9-9s4tk"] Oct 05 08:25:54 crc kubenswrapper[4846]: I1005 08:25:54.069252 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-896999bb9-9s4tk" Oct 05 08:25:54 crc kubenswrapper[4846]: I1005 08:25:54.093030 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-896999bb9-9s4tk"] Oct 05 08:25:54 crc kubenswrapper[4846]: I1005 08:25:54.107109 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 05 08:25:54 crc kubenswrapper[4846]: I1005 08:25:54.107428 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="9bceecf1-6148-4b88-9604-4a37330b8211" containerName="glance-log" containerID="cri-o://2afe0c9e24d73820fd0e74d9e701dd3e0b5e5424e84fa2d6087a65639080a151" gracePeriod=30 Oct 05 08:25:54 crc kubenswrapper[4846]: I1005 08:25:54.107504 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="9bceecf1-6148-4b88-9604-4a37330b8211" containerName="glance-httpd" containerID="cri-o://a00db97f5fd0530b607658f46a3c6536ad9be6f121de426f95c1f017bccad02e" gracePeriod=30 Oct 05 08:25:54 crc kubenswrapper[4846]: I1005 08:25:54.121574 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a7850a96-3d5e-4b6e-938b-91cfa2d7d3cd-config-data\") pod \"horizon-57d69df789-t686p\" (UID: \"a7850a96-3d5e-4b6e-938b-91cfa2d7d3cd\") " pod="openstack/horizon-57d69df789-t686p" Oct 05 08:25:54 crc kubenswrapper[4846]: I1005 08:25:54.121614 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cb5xb\" (UniqueName: \"kubernetes.io/projected/a7850a96-3d5e-4b6e-938b-91cfa2d7d3cd-kube-api-access-cb5xb\") pod \"horizon-57d69df789-t686p\" (UID: \"a7850a96-3d5e-4b6e-938b-91cfa2d7d3cd\") " pod="openstack/horizon-57d69df789-t686p" Oct 05 08:25:54 crc kubenswrapper[4846]: I1005 08:25:54.121651 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a7850a96-3d5e-4b6e-938b-91cfa2d7d3cd-scripts\") pod \"horizon-57d69df789-t686p\" (UID: \"a7850a96-3d5e-4b6e-938b-91cfa2d7d3cd\") " pod="openstack/horizon-57d69df789-t686p" Oct 05 08:25:54 crc kubenswrapper[4846]: I1005 08:25:54.121710 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/a7850a96-3d5e-4b6e-938b-91cfa2d7d3cd-horizon-secret-key\") pod \"horizon-57d69df789-t686p\" (UID: \"a7850a96-3d5e-4b6e-938b-91cfa2d7d3cd\") " pod="openstack/horizon-57d69df789-t686p" Oct 05 08:25:54 crc kubenswrapper[4846]: I1005 08:25:54.121741 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a7850a96-3d5e-4b6e-938b-91cfa2d7d3cd-logs\") pod \"horizon-57d69df789-t686p\" (UID: \"a7850a96-3d5e-4b6e-938b-91cfa2d7d3cd\") " pod="openstack/horizon-57d69df789-t686p" Oct 05 08:25:54 crc kubenswrapper[4846]: I1005 08:25:54.225527 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/a7850a96-3d5e-4b6e-938b-91cfa2d7d3cd-horizon-secret-key\") pod \"horizon-57d69df789-t686p\" (UID: \"a7850a96-3d5e-4b6e-938b-91cfa2d7d3cd\") " pod="openstack/horizon-57d69df789-t686p" Oct 05 08:25:54 crc kubenswrapper[4846]: I1005 08:25:54.225687 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/60577687-a955-4db0-a8a3-4a3bb3592d3d-logs\") pod \"horizon-896999bb9-9s4tk\" (UID: \"60577687-a955-4db0-a8a3-4a3bb3592d3d\") " pod="openstack/horizon-896999bb9-9s4tk" Oct 05 08:25:54 crc kubenswrapper[4846]: I1005 08:25:54.225740 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a7850a96-3d5e-4b6e-938b-91cfa2d7d3cd-logs\") pod \"horizon-57d69df789-t686p\" (UID: \"a7850a96-3d5e-4b6e-938b-91cfa2d7d3cd\") " pod="openstack/horizon-57d69df789-t686p" Oct 05 08:25:54 crc kubenswrapper[4846]: I1005 08:25:54.225765 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/60577687-a955-4db0-a8a3-4a3bb3592d3d-horizon-secret-key\") pod \"horizon-896999bb9-9s4tk\" (UID: \"60577687-a955-4db0-a8a3-4a3bb3592d3d\") " pod="openstack/horizon-896999bb9-9s4tk" Oct 05 08:25:54 crc kubenswrapper[4846]: I1005 08:25:54.225872 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/60577687-a955-4db0-a8a3-4a3bb3592d3d-scripts\") pod \"horizon-896999bb9-9s4tk\" (UID: \"60577687-a955-4db0-a8a3-4a3bb3592d3d\") " pod="openstack/horizon-896999bb9-9s4tk" Oct 05 08:25:54 crc kubenswrapper[4846]: I1005 08:25:54.225980 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/60577687-a955-4db0-a8a3-4a3bb3592d3d-config-data\") pod \"horizon-896999bb9-9s4tk\" (UID: \"60577687-a955-4db0-a8a3-4a3bb3592d3d\") " pod="openstack/horizon-896999bb9-9s4tk" Oct 05 08:25:54 crc kubenswrapper[4846]: I1005 08:25:54.226080 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zl5mn\" (UniqueName: \"kubernetes.io/projected/60577687-a955-4db0-a8a3-4a3bb3592d3d-kube-api-access-zl5mn\") pod \"horizon-896999bb9-9s4tk\" (UID: \"60577687-a955-4db0-a8a3-4a3bb3592d3d\") " pod="openstack/horizon-896999bb9-9s4tk" Oct 05 08:25:54 crc kubenswrapper[4846]: I1005 08:25:54.226282 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a7850a96-3d5e-4b6e-938b-91cfa2d7d3cd-config-data\") pod \"horizon-57d69df789-t686p\" (UID: \"a7850a96-3d5e-4b6e-938b-91cfa2d7d3cd\") " pod="openstack/horizon-57d69df789-t686p" Oct 05 08:25:54 crc kubenswrapper[4846]: I1005 08:25:54.226331 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cb5xb\" (UniqueName: \"kubernetes.io/projected/a7850a96-3d5e-4b6e-938b-91cfa2d7d3cd-kube-api-access-cb5xb\") pod \"horizon-57d69df789-t686p\" (UID: \"a7850a96-3d5e-4b6e-938b-91cfa2d7d3cd\") " pod="openstack/horizon-57d69df789-t686p" Oct 05 08:25:54 crc kubenswrapper[4846]: I1005 08:25:54.226399 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a7850a96-3d5e-4b6e-938b-91cfa2d7d3cd-scripts\") pod \"horizon-57d69df789-t686p\" (UID: \"a7850a96-3d5e-4b6e-938b-91cfa2d7d3cd\") " pod="openstack/horizon-57d69df789-t686p" Oct 05 08:25:54 crc kubenswrapper[4846]: I1005 08:25:54.227391 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a7850a96-3d5e-4b6e-938b-91cfa2d7d3cd-logs\") pod \"horizon-57d69df789-t686p\" (UID: \"a7850a96-3d5e-4b6e-938b-91cfa2d7d3cd\") " pod="openstack/horizon-57d69df789-t686p" Oct 05 08:25:54 crc kubenswrapper[4846]: I1005 08:25:54.227581 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a7850a96-3d5e-4b6e-938b-91cfa2d7d3cd-scripts\") pod \"horizon-57d69df789-t686p\" (UID: \"a7850a96-3d5e-4b6e-938b-91cfa2d7d3cd\") " pod="openstack/horizon-57d69df789-t686p" Oct 05 08:25:54 crc kubenswrapper[4846]: I1005 08:25:54.229958 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a7850a96-3d5e-4b6e-938b-91cfa2d7d3cd-config-data\") pod \"horizon-57d69df789-t686p\" (UID: \"a7850a96-3d5e-4b6e-938b-91cfa2d7d3cd\") " pod="openstack/horizon-57d69df789-t686p" Oct 05 08:25:54 crc kubenswrapper[4846]: I1005 08:25:54.232447 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/a7850a96-3d5e-4b6e-938b-91cfa2d7d3cd-horizon-secret-key\") pod \"horizon-57d69df789-t686p\" (UID: \"a7850a96-3d5e-4b6e-938b-91cfa2d7d3cd\") " pod="openstack/horizon-57d69df789-t686p" Oct 05 08:25:54 crc kubenswrapper[4846]: I1005 08:25:54.243790 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cb5xb\" (UniqueName: \"kubernetes.io/projected/a7850a96-3d5e-4b6e-938b-91cfa2d7d3cd-kube-api-access-cb5xb\") pod \"horizon-57d69df789-t686p\" (UID: \"a7850a96-3d5e-4b6e-938b-91cfa2d7d3cd\") " pod="openstack/horizon-57d69df789-t686p" Oct 05 08:25:54 crc kubenswrapper[4846]: I1005 08:25:54.327708 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/60577687-a955-4db0-a8a3-4a3bb3592d3d-logs\") pod \"horizon-896999bb9-9s4tk\" (UID: \"60577687-a955-4db0-a8a3-4a3bb3592d3d\") " pod="openstack/horizon-896999bb9-9s4tk" Oct 05 08:25:54 crc kubenswrapper[4846]: I1005 08:25:54.327824 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/60577687-a955-4db0-a8a3-4a3bb3592d3d-horizon-secret-key\") pod \"horizon-896999bb9-9s4tk\" (UID: \"60577687-a955-4db0-a8a3-4a3bb3592d3d\") " pod="openstack/horizon-896999bb9-9s4tk" Oct 05 08:25:54 crc kubenswrapper[4846]: I1005 08:25:54.327876 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/60577687-a955-4db0-a8a3-4a3bb3592d3d-scripts\") pod \"horizon-896999bb9-9s4tk\" (UID: \"60577687-a955-4db0-a8a3-4a3bb3592d3d\") " pod="openstack/horizon-896999bb9-9s4tk" Oct 05 08:25:54 crc kubenswrapper[4846]: I1005 08:25:54.327915 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/60577687-a955-4db0-a8a3-4a3bb3592d3d-config-data\") pod \"horizon-896999bb9-9s4tk\" (UID: \"60577687-a955-4db0-a8a3-4a3bb3592d3d\") " pod="openstack/horizon-896999bb9-9s4tk" Oct 05 08:25:54 crc kubenswrapper[4846]: I1005 08:25:54.327950 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zl5mn\" (UniqueName: \"kubernetes.io/projected/60577687-a955-4db0-a8a3-4a3bb3592d3d-kube-api-access-zl5mn\") pod \"horizon-896999bb9-9s4tk\" (UID: \"60577687-a955-4db0-a8a3-4a3bb3592d3d\") " pod="openstack/horizon-896999bb9-9s4tk" Oct 05 08:25:54 crc kubenswrapper[4846]: I1005 08:25:54.328856 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/60577687-a955-4db0-a8a3-4a3bb3592d3d-logs\") pod \"horizon-896999bb9-9s4tk\" (UID: \"60577687-a955-4db0-a8a3-4a3bb3592d3d\") " pod="openstack/horizon-896999bb9-9s4tk" Oct 05 08:25:54 crc kubenswrapper[4846]: I1005 08:25:54.329916 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/60577687-a955-4db0-a8a3-4a3bb3592d3d-scripts\") pod \"horizon-896999bb9-9s4tk\" (UID: \"60577687-a955-4db0-a8a3-4a3bb3592d3d\") " pod="openstack/horizon-896999bb9-9s4tk" Oct 05 08:25:54 crc kubenswrapper[4846]: I1005 08:25:54.330377 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/60577687-a955-4db0-a8a3-4a3bb3592d3d-config-data\") pod \"horizon-896999bb9-9s4tk\" (UID: \"60577687-a955-4db0-a8a3-4a3bb3592d3d\") " pod="openstack/horizon-896999bb9-9s4tk" Oct 05 08:25:54 crc kubenswrapper[4846]: I1005 08:25:54.337697 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/60577687-a955-4db0-a8a3-4a3bb3592d3d-horizon-secret-key\") pod \"horizon-896999bb9-9s4tk\" (UID: \"60577687-a955-4db0-a8a3-4a3bb3592d3d\") " pod="openstack/horizon-896999bb9-9s4tk" Oct 05 08:25:54 crc kubenswrapper[4846]: I1005 08:25:54.341475 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-57d69df789-t686p" Oct 05 08:25:54 crc kubenswrapper[4846]: I1005 08:25:54.345152 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zl5mn\" (UniqueName: \"kubernetes.io/projected/60577687-a955-4db0-a8a3-4a3bb3592d3d-kube-api-access-zl5mn\") pod \"horizon-896999bb9-9s4tk\" (UID: \"60577687-a955-4db0-a8a3-4a3bb3592d3d\") " pod="openstack/horizon-896999bb9-9s4tk" Oct 05 08:25:54 crc kubenswrapper[4846]: I1005 08:25:54.395905 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-896999bb9-9s4tk" Oct 05 08:25:54 crc kubenswrapper[4846]: I1005 08:25:54.495162 4846 generic.go:334] "Generic (PLEG): container finished" podID="9bceecf1-6148-4b88-9604-4a37330b8211" containerID="2afe0c9e24d73820fd0e74d9e701dd3e0b5e5424e84fa2d6087a65639080a151" exitCode=143 Oct 05 08:25:54 crc kubenswrapper[4846]: I1005 08:25:54.495263 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"9bceecf1-6148-4b88-9604-4a37330b8211","Type":"ContainerDied","Data":"2afe0c9e24d73820fd0e74d9e701dd3e0b5e5424e84fa2d6087a65639080a151"} Oct 05 08:25:54 crc kubenswrapper[4846]: I1005 08:25:54.496632 4846 generic.go:334] "Generic (PLEG): container finished" podID="c77ecef6-3a1d-43d5-90b1-dd3de33d7de2" containerID="d0b2a95a1c72c58f215d678122d3d9a70d119c082300731b85c1b4d15fcb5ab0" exitCode=143 Oct 05 08:25:54 crc kubenswrapper[4846]: I1005 08:25:54.496651 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"c77ecef6-3a1d-43d5-90b1-dd3de33d7de2","Type":"ContainerDied","Data":"d0b2a95a1c72c58f215d678122d3d9a70d119c082300731b85c1b4d15fcb5ab0"} Oct 05 08:25:54 crc kubenswrapper[4846]: I1005 08:25:54.511068 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e39a19ff-3716-4940-9945-3602c8cbcaf2" path="/var/lib/kubelet/pods/e39a19ff-3716-4940-9945-3602c8cbcaf2/volumes" Oct 05 08:25:54 crc kubenswrapper[4846]: I1005 08:25:54.817734 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-57d69df789-t686p"] Oct 05 08:25:54 crc kubenswrapper[4846]: I1005 08:25:54.899536 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-896999bb9-9s4tk"] Oct 05 08:25:54 crc kubenswrapper[4846]: W1005 08:25:54.906097 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod60577687_a955_4db0_a8a3_4a3bb3592d3d.slice/crio-c36ee474c579c405b3f96953dda49a0a00393e3aaeb619ea5e0d70b49e9fea6f WatchSource:0}: Error finding container c36ee474c579c405b3f96953dda49a0a00393e3aaeb619ea5e0d70b49e9fea6f: Status 404 returned error can't find the container with id c36ee474c579c405b3f96953dda49a0a00393e3aaeb619ea5e0d70b49e9fea6f Oct 05 08:25:55 crc kubenswrapper[4846]: I1005 08:25:55.511994 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-896999bb9-9s4tk" event={"ID":"60577687-a955-4db0-a8a3-4a3bb3592d3d","Type":"ContainerStarted","Data":"c36ee474c579c405b3f96953dda49a0a00393e3aaeb619ea5e0d70b49e9fea6f"} Oct 05 08:25:55 crc kubenswrapper[4846]: I1005 08:25:55.513964 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-57d69df789-t686p" event={"ID":"a7850a96-3d5e-4b6e-938b-91cfa2d7d3cd","Type":"ContainerStarted","Data":"1b1ad7c1c452928e64b3e179da694962dd87843ce67331603eecf625afd06c2a"} Oct 05 08:25:55 crc kubenswrapper[4846]: I1005 08:25:55.785023 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-57d69df789-t686p"] Oct 05 08:25:55 crc kubenswrapper[4846]: I1005 08:25:55.833266 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-d98bd9cc6-4s784"] Oct 05 08:25:55 crc kubenswrapper[4846]: I1005 08:25:55.834788 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-d98bd9cc6-4s784" Oct 05 08:25:55 crc kubenswrapper[4846]: I1005 08:25:55.837876 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-horizon-svc" Oct 05 08:25:55 crc kubenswrapper[4846]: I1005 08:25:55.888255 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-d98bd9cc6-4s784"] Oct 05 08:25:55 crc kubenswrapper[4846]: I1005 08:25:55.910049 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-896999bb9-9s4tk"] Oct 05 08:25:55 crc kubenswrapper[4846]: I1005 08:25:55.921402 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-f8ddb658d-6wlz2"] Oct 05 08:25:55 crc kubenswrapper[4846]: I1005 08:25:55.928420 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-f8ddb658d-6wlz2" Oct 05 08:25:55 crc kubenswrapper[4846]: I1005 08:25:55.940443 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-f8ddb658d-6wlz2"] Oct 05 08:25:55 crc kubenswrapper[4846]: I1005 08:25:55.982088 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p4xpr\" (UniqueName: \"kubernetes.io/projected/c8df1242-e04a-450b-a0b4-9d408f4b3862-kube-api-access-p4xpr\") pod \"horizon-d98bd9cc6-4s784\" (UID: \"c8df1242-e04a-450b-a0b4-9d408f4b3862\") " pod="openstack/horizon-d98bd9cc6-4s784" Oct 05 08:25:55 crc kubenswrapper[4846]: I1005 08:25:55.982157 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c8df1242-e04a-450b-a0b4-9d408f4b3862-logs\") pod \"horizon-d98bd9cc6-4s784\" (UID: \"c8df1242-e04a-450b-a0b4-9d408f4b3862\") " pod="openstack/horizon-d98bd9cc6-4s784" Oct 05 08:25:55 crc kubenswrapper[4846]: I1005 08:25:55.983485 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/c8df1242-e04a-450b-a0b4-9d408f4b3862-horizon-tls-certs\") pod \"horizon-d98bd9cc6-4s784\" (UID: \"c8df1242-e04a-450b-a0b4-9d408f4b3862\") " pod="openstack/horizon-d98bd9cc6-4s784" Oct 05 08:25:55 crc kubenswrapper[4846]: I1005 08:25:55.983528 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c8df1242-e04a-450b-a0b4-9d408f4b3862-scripts\") pod \"horizon-d98bd9cc6-4s784\" (UID: \"c8df1242-e04a-450b-a0b4-9d408f4b3862\") " pod="openstack/horizon-d98bd9cc6-4s784" Oct 05 08:25:55 crc kubenswrapper[4846]: I1005 08:25:55.983586 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c8df1242-e04a-450b-a0b4-9d408f4b3862-config-data\") pod \"horizon-d98bd9cc6-4s784\" (UID: \"c8df1242-e04a-450b-a0b4-9d408f4b3862\") " pod="openstack/horizon-d98bd9cc6-4s784" Oct 05 08:25:55 crc kubenswrapper[4846]: I1005 08:25:55.983620 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c8df1242-e04a-450b-a0b4-9d408f4b3862-combined-ca-bundle\") pod \"horizon-d98bd9cc6-4s784\" (UID: \"c8df1242-e04a-450b-a0b4-9d408f4b3862\") " pod="openstack/horizon-d98bd9cc6-4s784" Oct 05 08:25:55 crc kubenswrapper[4846]: I1005 08:25:55.983831 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/c8df1242-e04a-450b-a0b4-9d408f4b3862-horizon-secret-key\") pod \"horizon-d98bd9cc6-4s784\" (UID: \"c8df1242-e04a-450b-a0b4-9d408f4b3862\") " pod="openstack/horizon-d98bd9cc6-4s784" Oct 05 08:25:56 crc kubenswrapper[4846]: I1005 08:25:56.085434 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/c8df1242-e04a-450b-a0b4-9d408f4b3862-horizon-tls-certs\") pod \"horizon-d98bd9cc6-4s784\" (UID: \"c8df1242-e04a-450b-a0b4-9d408f4b3862\") " pod="openstack/horizon-d98bd9cc6-4s784" Oct 05 08:25:56 crc kubenswrapper[4846]: I1005 08:25:56.085499 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/8329aafa-c3d7-474c-a31b-a56e0d23ddce-horizon-secret-key\") pod \"horizon-f8ddb658d-6wlz2\" (UID: \"8329aafa-c3d7-474c-a31b-a56e0d23ddce\") " pod="openstack/horizon-f8ddb658d-6wlz2" Oct 05 08:25:56 crc kubenswrapper[4846]: I1005 08:25:56.085521 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c8df1242-e04a-450b-a0b4-9d408f4b3862-scripts\") pod \"horizon-d98bd9cc6-4s784\" (UID: \"c8df1242-e04a-450b-a0b4-9d408f4b3862\") " pod="openstack/horizon-d98bd9cc6-4s784" Oct 05 08:25:56 crc kubenswrapper[4846]: I1005 08:25:56.085544 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8329aafa-c3d7-474c-a31b-a56e0d23ddce-scripts\") pod \"horizon-f8ddb658d-6wlz2\" (UID: \"8329aafa-c3d7-474c-a31b-a56e0d23ddce\") " pod="openstack/horizon-f8ddb658d-6wlz2" Oct 05 08:25:56 crc kubenswrapper[4846]: I1005 08:25:56.085600 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8329aafa-c3d7-474c-a31b-a56e0d23ddce-config-data\") pod \"horizon-f8ddb658d-6wlz2\" (UID: \"8329aafa-c3d7-474c-a31b-a56e0d23ddce\") " pod="openstack/horizon-f8ddb658d-6wlz2" Oct 05 08:25:56 crc kubenswrapper[4846]: I1005 08:25:56.085633 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c8df1242-e04a-450b-a0b4-9d408f4b3862-config-data\") pod \"horizon-d98bd9cc6-4s784\" (UID: \"c8df1242-e04a-450b-a0b4-9d408f4b3862\") " pod="openstack/horizon-d98bd9cc6-4s784" Oct 05 08:25:56 crc kubenswrapper[4846]: I1005 08:25:56.085653 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/8329aafa-c3d7-474c-a31b-a56e0d23ddce-horizon-tls-certs\") pod \"horizon-f8ddb658d-6wlz2\" (UID: \"8329aafa-c3d7-474c-a31b-a56e0d23ddce\") " pod="openstack/horizon-f8ddb658d-6wlz2" Oct 05 08:25:56 crc kubenswrapper[4846]: I1005 08:25:56.085684 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c8df1242-e04a-450b-a0b4-9d408f4b3862-combined-ca-bundle\") pod \"horizon-d98bd9cc6-4s784\" (UID: \"c8df1242-e04a-450b-a0b4-9d408f4b3862\") " pod="openstack/horizon-d98bd9cc6-4s784" Oct 05 08:25:56 crc kubenswrapper[4846]: I1005 08:25:56.085705 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8329aafa-c3d7-474c-a31b-a56e0d23ddce-combined-ca-bundle\") pod \"horizon-f8ddb658d-6wlz2\" (UID: \"8329aafa-c3d7-474c-a31b-a56e0d23ddce\") " pod="openstack/horizon-f8ddb658d-6wlz2" Oct 05 08:25:56 crc kubenswrapper[4846]: I1005 08:25:56.085895 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8329aafa-c3d7-474c-a31b-a56e0d23ddce-logs\") pod \"horizon-f8ddb658d-6wlz2\" (UID: \"8329aafa-c3d7-474c-a31b-a56e0d23ddce\") " pod="openstack/horizon-f8ddb658d-6wlz2" Oct 05 08:25:56 crc kubenswrapper[4846]: I1005 08:25:56.085943 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/c8df1242-e04a-450b-a0b4-9d408f4b3862-horizon-secret-key\") pod \"horizon-d98bd9cc6-4s784\" (UID: \"c8df1242-e04a-450b-a0b4-9d408f4b3862\") " pod="openstack/horizon-d98bd9cc6-4s784" Oct 05 08:25:56 crc kubenswrapper[4846]: I1005 08:25:56.085990 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p4xpr\" (UniqueName: \"kubernetes.io/projected/c8df1242-e04a-450b-a0b4-9d408f4b3862-kube-api-access-p4xpr\") pod \"horizon-d98bd9cc6-4s784\" (UID: \"c8df1242-e04a-450b-a0b4-9d408f4b3862\") " pod="openstack/horizon-d98bd9cc6-4s784" Oct 05 08:25:56 crc kubenswrapper[4846]: I1005 08:25:56.086020 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wddvf\" (UniqueName: \"kubernetes.io/projected/8329aafa-c3d7-474c-a31b-a56e0d23ddce-kube-api-access-wddvf\") pod \"horizon-f8ddb658d-6wlz2\" (UID: \"8329aafa-c3d7-474c-a31b-a56e0d23ddce\") " pod="openstack/horizon-f8ddb658d-6wlz2" Oct 05 08:25:56 crc kubenswrapper[4846]: I1005 08:25:56.086658 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c8df1242-e04a-450b-a0b4-9d408f4b3862-logs\") pod \"horizon-d98bd9cc6-4s784\" (UID: \"c8df1242-e04a-450b-a0b4-9d408f4b3862\") " pod="openstack/horizon-d98bd9cc6-4s784" Oct 05 08:25:56 crc kubenswrapper[4846]: I1005 08:25:56.086658 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c8df1242-e04a-450b-a0b4-9d408f4b3862-scripts\") pod \"horizon-d98bd9cc6-4s784\" (UID: \"c8df1242-e04a-450b-a0b4-9d408f4b3862\") " pod="openstack/horizon-d98bd9cc6-4s784" Oct 05 08:25:56 crc kubenswrapper[4846]: I1005 08:25:56.086373 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c8df1242-e04a-450b-a0b4-9d408f4b3862-logs\") pod \"horizon-d98bd9cc6-4s784\" (UID: \"c8df1242-e04a-450b-a0b4-9d408f4b3862\") " pod="openstack/horizon-d98bd9cc6-4s784" Oct 05 08:25:56 crc kubenswrapper[4846]: I1005 08:25:56.087304 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c8df1242-e04a-450b-a0b4-9d408f4b3862-config-data\") pod \"horizon-d98bd9cc6-4s784\" (UID: \"c8df1242-e04a-450b-a0b4-9d408f4b3862\") " pod="openstack/horizon-d98bd9cc6-4s784" Oct 05 08:25:56 crc kubenswrapper[4846]: I1005 08:25:56.090655 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/c8df1242-e04a-450b-a0b4-9d408f4b3862-horizon-secret-key\") pod \"horizon-d98bd9cc6-4s784\" (UID: \"c8df1242-e04a-450b-a0b4-9d408f4b3862\") " pod="openstack/horizon-d98bd9cc6-4s784" Oct 05 08:25:56 crc kubenswrapper[4846]: I1005 08:25:56.091073 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c8df1242-e04a-450b-a0b4-9d408f4b3862-combined-ca-bundle\") pod \"horizon-d98bd9cc6-4s784\" (UID: \"c8df1242-e04a-450b-a0b4-9d408f4b3862\") " pod="openstack/horizon-d98bd9cc6-4s784" Oct 05 08:25:56 crc kubenswrapper[4846]: I1005 08:25:56.091934 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/c8df1242-e04a-450b-a0b4-9d408f4b3862-horizon-tls-certs\") pod \"horizon-d98bd9cc6-4s784\" (UID: \"c8df1242-e04a-450b-a0b4-9d408f4b3862\") " pod="openstack/horizon-d98bd9cc6-4s784" Oct 05 08:25:56 crc kubenswrapper[4846]: I1005 08:25:56.110128 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p4xpr\" (UniqueName: \"kubernetes.io/projected/c8df1242-e04a-450b-a0b4-9d408f4b3862-kube-api-access-p4xpr\") pod \"horizon-d98bd9cc6-4s784\" (UID: \"c8df1242-e04a-450b-a0b4-9d408f4b3862\") " pod="openstack/horizon-d98bd9cc6-4s784" Oct 05 08:25:56 crc kubenswrapper[4846]: I1005 08:25:56.188587 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/8329aafa-c3d7-474c-a31b-a56e0d23ddce-horizon-secret-key\") pod \"horizon-f8ddb658d-6wlz2\" (UID: \"8329aafa-c3d7-474c-a31b-a56e0d23ddce\") " pod="openstack/horizon-f8ddb658d-6wlz2" Oct 05 08:25:56 crc kubenswrapper[4846]: I1005 08:25:56.188628 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8329aafa-c3d7-474c-a31b-a56e0d23ddce-scripts\") pod \"horizon-f8ddb658d-6wlz2\" (UID: \"8329aafa-c3d7-474c-a31b-a56e0d23ddce\") " pod="openstack/horizon-f8ddb658d-6wlz2" Oct 05 08:25:56 crc kubenswrapper[4846]: I1005 08:25:56.188663 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8329aafa-c3d7-474c-a31b-a56e0d23ddce-config-data\") pod \"horizon-f8ddb658d-6wlz2\" (UID: \"8329aafa-c3d7-474c-a31b-a56e0d23ddce\") " pod="openstack/horizon-f8ddb658d-6wlz2" Oct 05 08:25:56 crc kubenswrapper[4846]: I1005 08:25:56.188695 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/8329aafa-c3d7-474c-a31b-a56e0d23ddce-horizon-tls-certs\") pod \"horizon-f8ddb658d-6wlz2\" (UID: \"8329aafa-c3d7-474c-a31b-a56e0d23ddce\") " pod="openstack/horizon-f8ddb658d-6wlz2" Oct 05 08:25:56 crc kubenswrapper[4846]: I1005 08:25:56.188727 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8329aafa-c3d7-474c-a31b-a56e0d23ddce-combined-ca-bundle\") pod \"horizon-f8ddb658d-6wlz2\" (UID: \"8329aafa-c3d7-474c-a31b-a56e0d23ddce\") " pod="openstack/horizon-f8ddb658d-6wlz2" Oct 05 08:25:56 crc kubenswrapper[4846]: I1005 08:25:56.188751 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8329aafa-c3d7-474c-a31b-a56e0d23ddce-logs\") pod \"horizon-f8ddb658d-6wlz2\" (UID: \"8329aafa-c3d7-474c-a31b-a56e0d23ddce\") " pod="openstack/horizon-f8ddb658d-6wlz2" Oct 05 08:25:56 crc kubenswrapper[4846]: I1005 08:25:56.188821 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wddvf\" (UniqueName: \"kubernetes.io/projected/8329aafa-c3d7-474c-a31b-a56e0d23ddce-kube-api-access-wddvf\") pod \"horizon-f8ddb658d-6wlz2\" (UID: \"8329aafa-c3d7-474c-a31b-a56e0d23ddce\") " pod="openstack/horizon-f8ddb658d-6wlz2" Oct 05 08:25:56 crc kubenswrapper[4846]: I1005 08:25:56.189481 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8329aafa-c3d7-474c-a31b-a56e0d23ddce-scripts\") pod \"horizon-f8ddb658d-6wlz2\" (UID: \"8329aafa-c3d7-474c-a31b-a56e0d23ddce\") " pod="openstack/horizon-f8ddb658d-6wlz2" Oct 05 08:25:56 crc kubenswrapper[4846]: I1005 08:25:56.189558 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8329aafa-c3d7-474c-a31b-a56e0d23ddce-logs\") pod \"horizon-f8ddb658d-6wlz2\" (UID: \"8329aafa-c3d7-474c-a31b-a56e0d23ddce\") " pod="openstack/horizon-f8ddb658d-6wlz2" Oct 05 08:25:56 crc kubenswrapper[4846]: I1005 08:25:56.190478 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8329aafa-c3d7-474c-a31b-a56e0d23ddce-config-data\") pod \"horizon-f8ddb658d-6wlz2\" (UID: \"8329aafa-c3d7-474c-a31b-a56e0d23ddce\") " pod="openstack/horizon-f8ddb658d-6wlz2" Oct 05 08:25:56 crc kubenswrapper[4846]: I1005 08:25:56.192509 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8329aafa-c3d7-474c-a31b-a56e0d23ddce-combined-ca-bundle\") pod \"horizon-f8ddb658d-6wlz2\" (UID: \"8329aafa-c3d7-474c-a31b-a56e0d23ddce\") " pod="openstack/horizon-f8ddb658d-6wlz2" Oct 05 08:25:56 crc kubenswrapper[4846]: I1005 08:25:56.192906 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/8329aafa-c3d7-474c-a31b-a56e0d23ddce-horizon-secret-key\") pod \"horizon-f8ddb658d-6wlz2\" (UID: \"8329aafa-c3d7-474c-a31b-a56e0d23ddce\") " pod="openstack/horizon-f8ddb658d-6wlz2" Oct 05 08:25:56 crc kubenswrapper[4846]: I1005 08:25:56.194334 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/8329aafa-c3d7-474c-a31b-a56e0d23ddce-horizon-tls-certs\") pod \"horizon-f8ddb658d-6wlz2\" (UID: \"8329aafa-c3d7-474c-a31b-a56e0d23ddce\") " pod="openstack/horizon-f8ddb658d-6wlz2" Oct 05 08:25:56 crc kubenswrapper[4846]: I1005 08:25:56.194447 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-d98bd9cc6-4s784" Oct 05 08:25:56 crc kubenswrapper[4846]: I1005 08:25:56.227166 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wddvf\" (UniqueName: \"kubernetes.io/projected/8329aafa-c3d7-474c-a31b-a56e0d23ddce-kube-api-access-wddvf\") pod \"horizon-f8ddb658d-6wlz2\" (UID: \"8329aafa-c3d7-474c-a31b-a56e0d23ddce\") " pod="openstack/horizon-f8ddb658d-6wlz2" Oct 05 08:25:56 crc kubenswrapper[4846]: I1005 08:25:56.262457 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-f8ddb658d-6wlz2" Oct 05 08:25:56 crc kubenswrapper[4846]: I1005 08:25:56.682190 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-d98bd9cc6-4s784"] Oct 05 08:25:56 crc kubenswrapper[4846]: W1005 08:25:56.691502 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc8df1242_e04a_450b_a0b4_9d408f4b3862.slice/crio-788762c5cb0dbb71578577fb3fdf613e547c15cd66bf4837d8e776c8bc4c02e7 WatchSource:0}: Error finding container 788762c5cb0dbb71578577fb3fdf613e547c15cd66bf4837d8e776c8bc4c02e7: Status 404 returned error can't find the container with id 788762c5cb0dbb71578577fb3fdf613e547c15cd66bf4837d8e776c8bc4c02e7 Oct 05 08:25:56 crc kubenswrapper[4846]: I1005 08:25:56.779473 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-f8ddb658d-6wlz2"] Oct 05 08:25:56 crc kubenswrapper[4846]: W1005 08:25:56.784306 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8329aafa_c3d7_474c_a31b_a56e0d23ddce.slice/crio-93e436d5d46990653a1327c7c30e1872329b7faf59be26dd23f40a69830d10f3 WatchSource:0}: Error finding container 93e436d5d46990653a1327c7c30e1872329b7faf59be26dd23f40a69830d10f3: Status 404 returned error can't find the container with id 93e436d5d46990653a1327c7c30e1872329b7faf59be26dd23f40a69830d10f3 Oct 05 08:25:57 crc kubenswrapper[4846]: I1005 08:25:57.541708 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-d98bd9cc6-4s784" event={"ID":"c8df1242-e04a-450b-a0b4-9d408f4b3862","Type":"ContainerStarted","Data":"788762c5cb0dbb71578577fb3fdf613e547c15cd66bf4837d8e776c8bc4c02e7"} Oct 05 08:25:57 crc kubenswrapper[4846]: I1005 08:25:57.546708 4846 generic.go:334] "Generic (PLEG): container finished" podID="9bceecf1-6148-4b88-9604-4a37330b8211" containerID="a00db97f5fd0530b607658f46a3c6536ad9be6f121de426f95c1f017bccad02e" exitCode=0 Oct 05 08:25:57 crc kubenswrapper[4846]: I1005 08:25:57.546758 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"9bceecf1-6148-4b88-9604-4a37330b8211","Type":"ContainerDied","Data":"a00db97f5fd0530b607658f46a3c6536ad9be6f121de426f95c1f017bccad02e"} Oct 05 08:25:57 crc kubenswrapper[4846]: I1005 08:25:57.548456 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-f8ddb658d-6wlz2" event={"ID":"8329aafa-c3d7-474c-a31b-a56e0d23ddce","Type":"ContainerStarted","Data":"93e436d5d46990653a1327c7c30e1872329b7faf59be26dd23f40a69830d10f3"} Oct 05 08:25:57 crc kubenswrapper[4846]: I1005 08:25:57.551668 4846 generic.go:334] "Generic (PLEG): container finished" podID="c77ecef6-3a1d-43d5-90b1-dd3de33d7de2" containerID="8e3d24f7a9c614dd76a5fe862718bcd0dda3735120671933705a9bef54139d4f" exitCode=0 Oct 05 08:25:57 crc kubenswrapper[4846]: I1005 08:25:57.551692 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"c77ecef6-3a1d-43d5-90b1-dd3de33d7de2","Type":"ContainerDied","Data":"8e3d24f7a9c614dd76a5fe862718bcd0dda3735120671933705a9bef54139d4f"} Oct 05 08:25:57 crc kubenswrapper[4846]: I1005 08:25:57.769615 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 05 08:25:57 crc kubenswrapper[4846]: I1005 08:25:57.842886 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 05 08:25:57 crc kubenswrapper[4846]: I1005 08:25:57.939754 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7255\" (UniqueName: \"kubernetes.io/projected/c77ecef6-3a1d-43d5-90b1-dd3de33d7de2-kube-api-access-w7255\") pod \"c77ecef6-3a1d-43d5-90b1-dd3de33d7de2\" (UID: \"c77ecef6-3a1d-43d5-90b1-dd3de33d7de2\") " Oct 05 08:25:57 crc kubenswrapper[4846]: I1005 08:25:57.939805 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c77ecef6-3a1d-43d5-90b1-dd3de33d7de2-public-tls-certs\") pod \"c77ecef6-3a1d-43d5-90b1-dd3de33d7de2\" (UID: \"c77ecef6-3a1d-43d5-90b1-dd3de33d7de2\") " Oct 05 08:25:57 crc kubenswrapper[4846]: I1005 08:25:57.939832 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9bceecf1-6148-4b88-9604-4a37330b8211-config-data\") pod \"9bceecf1-6148-4b88-9604-4a37330b8211\" (UID: \"9bceecf1-6148-4b88-9604-4a37330b8211\") " Oct 05 08:25:57 crc kubenswrapper[4846]: I1005 08:25:57.939900 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c77ecef6-3a1d-43d5-90b1-dd3de33d7de2-scripts\") pod \"c77ecef6-3a1d-43d5-90b1-dd3de33d7de2\" (UID: \"c77ecef6-3a1d-43d5-90b1-dd3de33d7de2\") " Oct 05 08:25:57 crc kubenswrapper[4846]: I1005 08:25:57.939916 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c77ecef6-3a1d-43d5-90b1-dd3de33d7de2-config-data\") pod \"c77ecef6-3a1d-43d5-90b1-dd3de33d7de2\" (UID: \"c77ecef6-3a1d-43d5-90b1-dd3de33d7de2\") " Oct 05 08:25:57 crc kubenswrapper[4846]: I1005 08:25:57.939969 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c77ecef6-3a1d-43d5-90b1-dd3de33d7de2-combined-ca-bundle\") pod \"c77ecef6-3a1d-43d5-90b1-dd3de33d7de2\" (UID: \"c77ecef6-3a1d-43d5-90b1-dd3de33d7de2\") " Oct 05 08:25:57 crc kubenswrapper[4846]: I1005 08:25:57.940014 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/9bceecf1-6148-4b88-9604-4a37330b8211-httpd-run\") pod \"9bceecf1-6148-4b88-9604-4a37330b8211\" (UID: \"9bceecf1-6148-4b88-9604-4a37330b8211\") " Oct 05 08:25:57 crc kubenswrapper[4846]: I1005 08:25:57.940093 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9bceecf1-6148-4b88-9604-4a37330b8211-internal-tls-certs\") pod \"9bceecf1-6148-4b88-9604-4a37330b8211\" (UID: \"9bceecf1-6148-4b88-9604-4a37330b8211\") " Oct 05 08:25:57 crc kubenswrapper[4846]: I1005 08:25:57.940119 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/c77ecef6-3a1d-43d5-90b1-dd3de33d7de2-httpd-run\") pod \"c77ecef6-3a1d-43d5-90b1-dd3de33d7de2\" (UID: \"c77ecef6-3a1d-43d5-90b1-dd3de33d7de2\") " Oct 05 08:25:57 crc kubenswrapper[4846]: I1005 08:25:57.940142 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9bceecf1-6148-4b88-9604-4a37330b8211-scripts\") pod \"9bceecf1-6148-4b88-9604-4a37330b8211\" (UID: \"9bceecf1-6148-4b88-9604-4a37330b8211\") " Oct 05 08:25:57 crc kubenswrapper[4846]: I1005 08:25:57.940175 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c77ecef6-3a1d-43d5-90b1-dd3de33d7de2-logs\") pod \"c77ecef6-3a1d-43d5-90b1-dd3de33d7de2\" (UID: \"c77ecef6-3a1d-43d5-90b1-dd3de33d7de2\") " Oct 05 08:25:57 crc kubenswrapper[4846]: I1005 08:25:57.940213 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9bceecf1-6148-4b88-9604-4a37330b8211-combined-ca-bundle\") pod \"9bceecf1-6148-4b88-9604-4a37330b8211\" (UID: \"9bceecf1-6148-4b88-9604-4a37330b8211\") " Oct 05 08:25:57 crc kubenswrapper[4846]: I1005 08:25:57.940232 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9bceecf1-6148-4b88-9604-4a37330b8211-logs\") pod \"9bceecf1-6148-4b88-9604-4a37330b8211\" (UID: \"9bceecf1-6148-4b88-9604-4a37330b8211\") " Oct 05 08:25:57 crc kubenswrapper[4846]: I1005 08:25:57.940267 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kc8fp\" (UniqueName: \"kubernetes.io/projected/9bceecf1-6148-4b88-9604-4a37330b8211-kube-api-access-kc8fp\") pod \"9bceecf1-6148-4b88-9604-4a37330b8211\" (UID: \"9bceecf1-6148-4b88-9604-4a37330b8211\") " Oct 05 08:25:57 crc kubenswrapper[4846]: I1005 08:25:57.942743 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9bceecf1-6148-4b88-9604-4a37330b8211-logs" (OuterVolumeSpecName: "logs") pod "9bceecf1-6148-4b88-9604-4a37330b8211" (UID: "9bceecf1-6148-4b88-9604-4a37330b8211"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 08:25:57 crc kubenswrapper[4846]: I1005 08:25:57.943815 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c77ecef6-3a1d-43d5-90b1-dd3de33d7de2-logs" (OuterVolumeSpecName: "logs") pod "c77ecef6-3a1d-43d5-90b1-dd3de33d7de2" (UID: "c77ecef6-3a1d-43d5-90b1-dd3de33d7de2"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 08:25:57 crc kubenswrapper[4846]: I1005 08:25:57.944085 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c77ecef6-3a1d-43d5-90b1-dd3de33d7de2-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "c77ecef6-3a1d-43d5-90b1-dd3de33d7de2" (UID: "c77ecef6-3a1d-43d5-90b1-dd3de33d7de2"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 08:25:57 crc kubenswrapper[4846]: I1005 08:25:57.944619 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9bceecf1-6148-4b88-9604-4a37330b8211-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "9bceecf1-6148-4b88-9604-4a37330b8211" (UID: "9bceecf1-6148-4b88-9604-4a37330b8211"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 08:25:57 crc kubenswrapper[4846]: I1005 08:25:57.945333 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c77ecef6-3a1d-43d5-90b1-dd3de33d7de2-kube-api-access-w7255" (OuterVolumeSpecName: "kube-api-access-w7255") pod "c77ecef6-3a1d-43d5-90b1-dd3de33d7de2" (UID: "c77ecef6-3a1d-43d5-90b1-dd3de33d7de2"). InnerVolumeSpecName "kube-api-access-w7255". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:25:57 crc kubenswrapper[4846]: I1005 08:25:57.947619 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c77ecef6-3a1d-43d5-90b1-dd3de33d7de2-scripts" (OuterVolumeSpecName: "scripts") pod "c77ecef6-3a1d-43d5-90b1-dd3de33d7de2" (UID: "c77ecef6-3a1d-43d5-90b1-dd3de33d7de2"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:25:57 crc kubenswrapper[4846]: I1005 08:25:57.948531 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9bceecf1-6148-4b88-9604-4a37330b8211-kube-api-access-kc8fp" (OuterVolumeSpecName: "kube-api-access-kc8fp") pod "9bceecf1-6148-4b88-9604-4a37330b8211" (UID: "9bceecf1-6148-4b88-9604-4a37330b8211"). InnerVolumeSpecName "kube-api-access-kc8fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:25:57 crc kubenswrapper[4846]: I1005 08:25:57.950213 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9bceecf1-6148-4b88-9604-4a37330b8211-scripts" (OuterVolumeSpecName: "scripts") pod "9bceecf1-6148-4b88-9604-4a37330b8211" (UID: "9bceecf1-6148-4b88-9604-4a37330b8211"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:25:57 crc kubenswrapper[4846]: I1005 08:25:57.967754 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9bceecf1-6148-4b88-9604-4a37330b8211-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9bceecf1-6148-4b88-9604-4a37330b8211" (UID: "9bceecf1-6148-4b88-9604-4a37330b8211"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:25:57 crc kubenswrapper[4846]: I1005 08:25:57.972381 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c77ecef6-3a1d-43d5-90b1-dd3de33d7de2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c77ecef6-3a1d-43d5-90b1-dd3de33d7de2" (UID: "c77ecef6-3a1d-43d5-90b1-dd3de33d7de2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:25:58 crc kubenswrapper[4846]: I1005 08:25:58.001892 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c77ecef6-3a1d-43d5-90b1-dd3de33d7de2-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "c77ecef6-3a1d-43d5-90b1-dd3de33d7de2" (UID: "c77ecef6-3a1d-43d5-90b1-dd3de33d7de2"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:25:58 crc kubenswrapper[4846]: I1005 08:25:58.007511 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9bceecf1-6148-4b88-9604-4a37330b8211-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "9bceecf1-6148-4b88-9604-4a37330b8211" (UID: "9bceecf1-6148-4b88-9604-4a37330b8211"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:25:58 crc kubenswrapper[4846]: I1005 08:25:58.016254 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c77ecef6-3a1d-43d5-90b1-dd3de33d7de2-config-data" (OuterVolumeSpecName: "config-data") pod "c77ecef6-3a1d-43d5-90b1-dd3de33d7de2" (UID: "c77ecef6-3a1d-43d5-90b1-dd3de33d7de2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:25:58 crc kubenswrapper[4846]: I1005 08:25:58.030884 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9bceecf1-6148-4b88-9604-4a37330b8211-config-data" (OuterVolumeSpecName: "config-data") pod "9bceecf1-6148-4b88-9604-4a37330b8211" (UID: "9bceecf1-6148-4b88-9604-4a37330b8211"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:25:58 crc kubenswrapper[4846]: I1005 08:25:58.042205 4846 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c77ecef6-3a1d-43d5-90b1-dd3de33d7de2-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 08:25:58 crc kubenswrapper[4846]: I1005 08:25:58.042240 4846 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c77ecef6-3a1d-43d5-90b1-dd3de33d7de2-scripts\") on node \"crc\" DevicePath \"\"" Oct 05 08:25:58 crc kubenswrapper[4846]: I1005 08:25:58.042252 4846 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c77ecef6-3a1d-43d5-90b1-dd3de33d7de2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 08:25:58 crc kubenswrapper[4846]: I1005 08:25:58.042266 4846 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/9bceecf1-6148-4b88-9604-4a37330b8211-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 05 08:25:58 crc kubenswrapper[4846]: I1005 08:25:58.042278 4846 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9bceecf1-6148-4b88-9604-4a37330b8211-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 05 08:25:58 crc kubenswrapper[4846]: I1005 08:25:58.042291 4846 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/c77ecef6-3a1d-43d5-90b1-dd3de33d7de2-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 05 08:25:58 crc kubenswrapper[4846]: I1005 08:25:58.042301 4846 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9bceecf1-6148-4b88-9604-4a37330b8211-scripts\") on node \"crc\" DevicePath \"\"" Oct 05 08:25:58 crc kubenswrapper[4846]: I1005 08:25:58.042311 4846 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c77ecef6-3a1d-43d5-90b1-dd3de33d7de2-logs\") on node \"crc\" DevicePath \"\"" Oct 05 08:25:58 crc kubenswrapper[4846]: I1005 08:25:58.042320 4846 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9bceecf1-6148-4b88-9604-4a37330b8211-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 08:25:58 crc kubenswrapper[4846]: I1005 08:25:58.042328 4846 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9bceecf1-6148-4b88-9604-4a37330b8211-logs\") on node \"crc\" DevicePath \"\"" Oct 05 08:25:58 crc kubenswrapper[4846]: I1005 08:25:58.042336 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kc8fp\" (UniqueName: \"kubernetes.io/projected/9bceecf1-6148-4b88-9604-4a37330b8211-kube-api-access-kc8fp\") on node \"crc\" DevicePath \"\"" Oct 05 08:25:58 crc kubenswrapper[4846]: I1005 08:25:58.042346 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7255\" (UniqueName: \"kubernetes.io/projected/c77ecef6-3a1d-43d5-90b1-dd3de33d7de2-kube-api-access-w7255\") on node \"crc\" DevicePath \"\"" Oct 05 08:25:58 crc kubenswrapper[4846]: I1005 08:25:58.042353 4846 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c77ecef6-3a1d-43d5-90b1-dd3de33d7de2-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 05 08:25:58 crc kubenswrapper[4846]: I1005 08:25:58.042363 4846 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9bceecf1-6148-4b88-9604-4a37330b8211-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 08:25:58 crc kubenswrapper[4846]: I1005 08:25:58.566298 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"9bceecf1-6148-4b88-9604-4a37330b8211","Type":"ContainerDied","Data":"1f45371f108557ad3137c9f316725d9f0083b908606dcace3367c337cd76bcbf"} Oct 05 08:25:58 crc kubenswrapper[4846]: I1005 08:25:58.566742 4846 scope.go:117] "RemoveContainer" containerID="a00db97f5fd0530b607658f46a3c6536ad9be6f121de426f95c1f017bccad02e" Oct 05 08:25:58 crc kubenswrapper[4846]: I1005 08:25:58.566317 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 05 08:25:58 crc kubenswrapper[4846]: I1005 08:25:58.572129 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"c77ecef6-3a1d-43d5-90b1-dd3de33d7de2","Type":"ContainerDied","Data":"0ea80ef9da8ba621e1ee9170929b011f536b8e7d9fe7573206557e238d515dc9"} Oct 05 08:25:58 crc kubenswrapper[4846]: I1005 08:25:58.572265 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 05 08:25:58 crc kubenswrapper[4846]: I1005 08:25:58.619344 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 05 08:25:58 crc kubenswrapper[4846]: I1005 08:25:58.633426 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 05 08:25:58 crc kubenswrapper[4846]: I1005 08:25:58.646236 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 05 08:25:58 crc kubenswrapper[4846]: I1005 08:25:58.655489 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 05 08:25:58 crc kubenswrapper[4846]: I1005 08:25:58.661159 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 05 08:25:58 crc kubenswrapper[4846]: E1005 08:25:58.661618 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9bceecf1-6148-4b88-9604-4a37330b8211" containerName="glance-httpd" Oct 05 08:25:58 crc kubenswrapper[4846]: I1005 08:25:58.661638 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="9bceecf1-6148-4b88-9604-4a37330b8211" containerName="glance-httpd" Oct 05 08:25:58 crc kubenswrapper[4846]: E1005 08:25:58.661657 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9bceecf1-6148-4b88-9604-4a37330b8211" containerName="glance-log" Oct 05 08:25:58 crc kubenswrapper[4846]: I1005 08:25:58.661695 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="9bceecf1-6148-4b88-9604-4a37330b8211" containerName="glance-log" Oct 05 08:25:58 crc kubenswrapper[4846]: E1005 08:25:58.661716 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c77ecef6-3a1d-43d5-90b1-dd3de33d7de2" containerName="glance-httpd" Oct 05 08:25:58 crc kubenswrapper[4846]: I1005 08:25:58.661723 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="c77ecef6-3a1d-43d5-90b1-dd3de33d7de2" containerName="glance-httpd" Oct 05 08:25:58 crc kubenswrapper[4846]: E1005 08:25:58.661751 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c77ecef6-3a1d-43d5-90b1-dd3de33d7de2" containerName="glance-log" Oct 05 08:25:58 crc kubenswrapper[4846]: I1005 08:25:58.661757 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="c77ecef6-3a1d-43d5-90b1-dd3de33d7de2" containerName="glance-log" Oct 05 08:25:58 crc kubenswrapper[4846]: I1005 08:25:58.661929 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="c77ecef6-3a1d-43d5-90b1-dd3de33d7de2" containerName="glance-httpd" Oct 05 08:25:58 crc kubenswrapper[4846]: I1005 08:25:58.661948 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="9bceecf1-6148-4b88-9604-4a37330b8211" containerName="glance-httpd" Oct 05 08:25:58 crc kubenswrapper[4846]: I1005 08:25:58.661967 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="9bceecf1-6148-4b88-9604-4a37330b8211" containerName="glance-log" Oct 05 08:25:58 crc kubenswrapper[4846]: I1005 08:25:58.661977 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="c77ecef6-3a1d-43d5-90b1-dd3de33d7de2" containerName="glance-log" Oct 05 08:25:58 crc kubenswrapper[4846]: I1005 08:25:58.663051 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 05 08:25:58 crc kubenswrapper[4846]: I1005 08:25:58.668308 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 05 08:25:58 crc kubenswrapper[4846]: I1005 08:25:58.669580 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Oct 05 08:25:58 crc kubenswrapper[4846]: I1005 08:25:58.669849 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Oct 05 08:25:58 crc kubenswrapper[4846]: I1005 08:25:58.670005 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-4zskb" Oct 05 08:25:58 crc kubenswrapper[4846]: I1005 08:25:58.670172 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Oct 05 08:25:58 crc kubenswrapper[4846]: I1005 08:25:58.675421 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Oct 05 08:25:58 crc kubenswrapper[4846]: I1005 08:25:58.677559 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 05 08:25:58 crc kubenswrapper[4846]: I1005 08:25:58.680389 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Oct 05 08:25:58 crc kubenswrapper[4846]: I1005 08:25:58.680918 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Oct 05 08:25:58 crc kubenswrapper[4846]: I1005 08:25:58.683964 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 05 08:25:58 crc kubenswrapper[4846]: I1005 08:25:58.763841 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/25d587cf-9f44-4c05-9ff9-c427f8337de5-logs\") pod \"glance-default-internal-api-0\" (UID: \"25d587cf-9f44-4c05-9ff9-c427f8337de5\") " pod="openstack/glance-default-internal-api-0" Oct 05 08:25:58 crc kubenswrapper[4846]: I1005 08:25:58.763922 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/25d587cf-9f44-4c05-9ff9-c427f8337de5-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"25d587cf-9f44-4c05-9ff9-c427f8337de5\") " pod="openstack/glance-default-internal-api-0" Oct 05 08:25:58 crc kubenswrapper[4846]: I1005 08:25:58.763976 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/25d587cf-9f44-4c05-9ff9-c427f8337de5-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"25d587cf-9f44-4c05-9ff9-c427f8337de5\") " pod="openstack/glance-default-internal-api-0" Oct 05 08:25:58 crc kubenswrapper[4846]: I1005 08:25:58.763997 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/25d587cf-9f44-4c05-9ff9-c427f8337de5-config-data\") pod \"glance-default-internal-api-0\" (UID: \"25d587cf-9f44-4c05-9ff9-c427f8337de5\") " pod="openstack/glance-default-internal-api-0" Oct 05 08:25:58 crc kubenswrapper[4846]: I1005 08:25:58.764015 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/25d587cf-9f44-4c05-9ff9-c427f8337de5-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"25d587cf-9f44-4c05-9ff9-c427f8337de5\") " pod="openstack/glance-default-internal-api-0" Oct 05 08:25:58 crc kubenswrapper[4846]: I1005 08:25:58.764060 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mw7rh\" (UniqueName: \"kubernetes.io/projected/25d587cf-9f44-4c05-9ff9-c427f8337de5-kube-api-access-mw7rh\") pod \"glance-default-internal-api-0\" (UID: \"25d587cf-9f44-4c05-9ff9-c427f8337de5\") " pod="openstack/glance-default-internal-api-0" Oct 05 08:25:58 crc kubenswrapper[4846]: I1005 08:25:58.764075 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/25d587cf-9f44-4c05-9ff9-c427f8337de5-scripts\") pod \"glance-default-internal-api-0\" (UID: \"25d587cf-9f44-4c05-9ff9-c427f8337de5\") " pod="openstack/glance-default-internal-api-0" Oct 05 08:25:58 crc kubenswrapper[4846]: I1005 08:25:58.867312 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a913b914-fe8a-4d4a-9969-7f4df6313f02-scripts\") pod \"glance-default-external-api-0\" (UID: \"a913b914-fe8a-4d4a-9969-7f4df6313f02\") " pod="openstack/glance-default-external-api-0" Oct 05 08:25:58 crc kubenswrapper[4846]: I1005 08:25:58.867377 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/25d587cf-9f44-4c05-9ff9-c427f8337de5-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"25d587cf-9f44-4c05-9ff9-c427f8337de5\") " pod="openstack/glance-default-internal-api-0" Oct 05 08:25:58 crc kubenswrapper[4846]: I1005 08:25:58.867415 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/25d587cf-9f44-4c05-9ff9-c427f8337de5-config-data\") pod \"glance-default-internal-api-0\" (UID: \"25d587cf-9f44-4c05-9ff9-c427f8337de5\") " pod="openstack/glance-default-internal-api-0" Oct 05 08:25:58 crc kubenswrapper[4846]: I1005 08:25:58.867511 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/25d587cf-9f44-4c05-9ff9-c427f8337de5-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"25d587cf-9f44-4c05-9ff9-c427f8337de5\") " pod="openstack/glance-default-internal-api-0" Oct 05 08:25:58 crc kubenswrapper[4846]: I1005 08:25:58.867609 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a913b914-fe8a-4d4a-9969-7f4df6313f02-config-data\") pod \"glance-default-external-api-0\" (UID: \"a913b914-fe8a-4d4a-9969-7f4df6313f02\") " pod="openstack/glance-default-external-api-0" Oct 05 08:25:58 crc kubenswrapper[4846]: I1005 08:25:58.867648 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a913b914-fe8a-4d4a-9969-7f4df6313f02-logs\") pod \"glance-default-external-api-0\" (UID: \"a913b914-fe8a-4d4a-9969-7f4df6313f02\") " pod="openstack/glance-default-external-api-0" Oct 05 08:25:58 crc kubenswrapper[4846]: I1005 08:25:58.867733 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/25d587cf-9f44-4c05-9ff9-c427f8337de5-scripts\") pod \"glance-default-internal-api-0\" (UID: \"25d587cf-9f44-4c05-9ff9-c427f8337de5\") " pod="openstack/glance-default-internal-api-0" Oct 05 08:25:58 crc kubenswrapper[4846]: I1005 08:25:58.867757 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mw7rh\" (UniqueName: \"kubernetes.io/projected/25d587cf-9f44-4c05-9ff9-c427f8337de5-kube-api-access-mw7rh\") pod \"glance-default-internal-api-0\" (UID: \"25d587cf-9f44-4c05-9ff9-c427f8337de5\") " pod="openstack/glance-default-internal-api-0" Oct 05 08:25:58 crc kubenswrapper[4846]: I1005 08:25:58.867868 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tnm56\" (UniqueName: \"kubernetes.io/projected/a913b914-fe8a-4d4a-9969-7f4df6313f02-kube-api-access-tnm56\") pod \"glance-default-external-api-0\" (UID: \"a913b914-fe8a-4d4a-9969-7f4df6313f02\") " pod="openstack/glance-default-external-api-0" Oct 05 08:25:58 crc kubenswrapper[4846]: I1005 08:25:58.867898 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/a913b914-fe8a-4d4a-9969-7f4df6313f02-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"a913b914-fe8a-4d4a-9969-7f4df6313f02\") " pod="openstack/glance-default-external-api-0" Oct 05 08:25:58 crc kubenswrapper[4846]: I1005 08:25:58.868040 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/25d587cf-9f44-4c05-9ff9-c427f8337de5-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"25d587cf-9f44-4c05-9ff9-c427f8337de5\") " pod="openstack/glance-default-internal-api-0" Oct 05 08:25:58 crc kubenswrapper[4846]: I1005 08:25:58.868572 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/25d587cf-9f44-4c05-9ff9-c427f8337de5-logs\") pod \"glance-default-internal-api-0\" (UID: \"25d587cf-9f44-4c05-9ff9-c427f8337de5\") " pod="openstack/glance-default-internal-api-0" Oct 05 08:25:58 crc kubenswrapper[4846]: I1005 08:25:58.868674 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a913b914-fe8a-4d4a-9969-7f4df6313f02-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"a913b914-fe8a-4d4a-9969-7f4df6313f02\") " pod="openstack/glance-default-external-api-0" Oct 05 08:25:58 crc kubenswrapper[4846]: I1005 08:25:58.868764 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/25d587cf-9f44-4c05-9ff9-c427f8337de5-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"25d587cf-9f44-4c05-9ff9-c427f8337de5\") " pod="openstack/glance-default-internal-api-0" Oct 05 08:25:58 crc kubenswrapper[4846]: I1005 08:25:58.868837 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a913b914-fe8a-4d4a-9969-7f4df6313f02-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"a913b914-fe8a-4d4a-9969-7f4df6313f02\") " pod="openstack/glance-default-external-api-0" Oct 05 08:25:58 crc kubenswrapper[4846]: I1005 08:25:58.869352 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/25d587cf-9f44-4c05-9ff9-c427f8337de5-logs\") pod \"glance-default-internal-api-0\" (UID: \"25d587cf-9f44-4c05-9ff9-c427f8337de5\") " pod="openstack/glance-default-internal-api-0" Oct 05 08:25:58 crc kubenswrapper[4846]: I1005 08:25:58.885641 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/25d587cf-9f44-4c05-9ff9-c427f8337de5-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"25d587cf-9f44-4c05-9ff9-c427f8337de5\") " pod="openstack/glance-default-internal-api-0" Oct 05 08:25:58 crc kubenswrapper[4846]: I1005 08:25:58.885801 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/25d587cf-9f44-4c05-9ff9-c427f8337de5-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"25d587cf-9f44-4c05-9ff9-c427f8337de5\") " pod="openstack/glance-default-internal-api-0" Oct 05 08:25:58 crc kubenswrapper[4846]: I1005 08:25:58.887043 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/25d587cf-9f44-4c05-9ff9-c427f8337de5-config-data\") pod \"glance-default-internal-api-0\" (UID: \"25d587cf-9f44-4c05-9ff9-c427f8337de5\") " pod="openstack/glance-default-internal-api-0" Oct 05 08:25:58 crc kubenswrapper[4846]: I1005 08:25:58.887726 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/25d587cf-9f44-4c05-9ff9-c427f8337de5-scripts\") pod \"glance-default-internal-api-0\" (UID: \"25d587cf-9f44-4c05-9ff9-c427f8337de5\") " pod="openstack/glance-default-internal-api-0" Oct 05 08:25:58 crc kubenswrapper[4846]: I1005 08:25:58.891069 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mw7rh\" (UniqueName: \"kubernetes.io/projected/25d587cf-9f44-4c05-9ff9-c427f8337de5-kube-api-access-mw7rh\") pod \"glance-default-internal-api-0\" (UID: \"25d587cf-9f44-4c05-9ff9-c427f8337de5\") " pod="openstack/glance-default-internal-api-0" Oct 05 08:25:58 crc kubenswrapper[4846]: I1005 08:25:58.970300 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a913b914-fe8a-4d4a-9969-7f4df6313f02-config-data\") pod \"glance-default-external-api-0\" (UID: \"a913b914-fe8a-4d4a-9969-7f4df6313f02\") " pod="openstack/glance-default-external-api-0" Oct 05 08:25:58 crc kubenswrapper[4846]: I1005 08:25:58.970681 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a913b914-fe8a-4d4a-9969-7f4df6313f02-logs\") pod \"glance-default-external-api-0\" (UID: \"a913b914-fe8a-4d4a-9969-7f4df6313f02\") " pod="openstack/glance-default-external-api-0" Oct 05 08:25:58 crc kubenswrapper[4846]: I1005 08:25:58.971164 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a913b914-fe8a-4d4a-9969-7f4df6313f02-logs\") pod \"glance-default-external-api-0\" (UID: \"a913b914-fe8a-4d4a-9969-7f4df6313f02\") " pod="openstack/glance-default-external-api-0" Oct 05 08:25:58 crc kubenswrapper[4846]: I1005 08:25:58.971308 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tnm56\" (UniqueName: \"kubernetes.io/projected/a913b914-fe8a-4d4a-9969-7f4df6313f02-kube-api-access-tnm56\") pod \"glance-default-external-api-0\" (UID: \"a913b914-fe8a-4d4a-9969-7f4df6313f02\") " pod="openstack/glance-default-external-api-0" Oct 05 08:25:58 crc kubenswrapper[4846]: I1005 08:25:58.971820 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/a913b914-fe8a-4d4a-9969-7f4df6313f02-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"a913b914-fe8a-4d4a-9969-7f4df6313f02\") " pod="openstack/glance-default-external-api-0" Oct 05 08:25:58 crc kubenswrapper[4846]: I1005 08:25:58.973345 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/a913b914-fe8a-4d4a-9969-7f4df6313f02-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"a913b914-fe8a-4d4a-9969-7f4df6313f02\") " pod="openstack/glance-default-external-api-0" Oct 05 08:25:58 crc kubenswrapper[4846]: I1005 08:25:58.973503 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a913b914-fe8a-4d4a-9969-7f4df6313f02-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"a913b914-fe8a-4d4a-9969-7f4df6313f02\") " pod="openstack/glance-default-external-api-0" Oct 05 08:25:58 crc kubenswrapper[4846]: I1005 08:25:58.974939 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a913b914-fe8a-4d4a-9969-7f4df6313f02-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"a913b914-fe8a-4d4a-9969-7f4df6313f02\") " pod="openstack/glance-default-external-api-0" Oct 05 08:25:58 crc kubenswrapper[4846]: I1005 08:25:58.975001 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a913b914-fe8a-4d4a-9969-7f4df6313f02-config-data\") pod \"glance-default-external-api-0\" (UID: \"a913b914-fe8a-4d4a-9969-7f4df6313f02\") " pod="openstack/glance-default-external-api-0" Oct 05 08:25:58 crc kubenswrapper[4846]: I1005 08:25:58.975005 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a913b914-fe8a-4d4a-9969-7f4df6313f02-scripts\") pod \"glance-default-external-api-0\" (UID: \"a913b914-fe8a-4d4a-9969-7f4df6313f02\") " pod="openstack/glance-default-external-api-0" Oct 05 08:25:58 crc kubenswrapper[4846]: I1005 08:25:58.977481 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a913b914-fe8a-4d4a-9969-7f4df6313f02-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"a913b914-fe8a-4d4a-9969-7f4df6313f02\") " pod="openstack/glance-default-external-api-0" Oct 05 08:25:58 crc kubenswrapper[4846]: I1005 08:25:58.979801 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a913b914-fe8a-4d4a-9969-7f4df6313f02-scripts\") pod \"glance-default-external-api-0\" (UID: \"a913b914-fe8a-4d4a-9969-7f4df6313f02\") " pod="openstack/glance-default-external-api-0" Oct 05 08:25:58 crc kubenswrapper[4846]: I1005 08:25:58.985462 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 05 08:25:58 crc kubenswrapper[4846]: I1005 08:25:58.989237 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tnm56\" (UniqueName: \"kubernetes.io/projected/a913b914-fe8a-4d4a-9969-7f4df6313f02-kube-api-access-tnm56\") pod \"glance-default-external-api-0\" (UID: \"a913b914-fe8a-4d4a-9969-7f4df6313f02\") " pod="openstack/glance-default-external-api-0" Oct 05 08:25:58 crc kubenswrapper[4846]: I1005 08:25:58.995432 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a913b914-fe8a-4d4a-9969-7f4df6313f02-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"a913b914-fe8a-4d4a-9969-7f4df6313f02\") " pod="openstack/glance-default-external-api-0" Oct 05 08:25:58 crc kubenswrapper[4846]: I1005 08:25:58.995974 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 05 08:26:00 crc kubenswrapper[4846]: I1005 08:26:00.514592 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9bceecf1-6148-4b88-9604-4a37330b8211" path="/var/lib/kubelet/pods/9bceecf1-6148-4b88-9604-4a37330b8211/volumes" Oct 05 08:26:00 crc kubenswrapper[4846]: I1005 08:26:00.516363 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c77ecef6-3a1d-43d5-90b1-dd3de33d7de2" path="/var/lib/kubelet/pods/c77ecef6-3a1d-43d5-90b1-dd3de33d7de2/volumes" Oct 05 08:26:01 crc kubenswrapper[4846]: I1005 08:26:01.033998 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-pcspm"] Oct 05 08:26:01 crc kubenswrapper[4846]: I1005 08:26:01.037358 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-pcspm" Oct 05 08:26:01 crc kubenswrapper[4846]: I1005 08:26:01.048738 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-pcspm"] Oct 05 08:26:01 crc kubenswrapper[4846]: I1005 08:26:01.227138 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9d267364-8423-4287-b7b7-31750ffb9a30-utilities\") pod \"community-operators-pcspm\" (UID: \"9d267364-8423-4287-b7b7-31750ffb9a30\") " pod="openshift-marketplace/community-operators-pcspm" Oct 05 08:26:01 crc kubenswrapper[4846]: I1005 08:26:01.227199 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9d267364-8423-4287-b7b7-31750ffb9a30-catalog-content\") pod \"community-operators-pcspm\" (UID: \"9d267364-8423-4287-b7b7-31750ffb9a30\") " pod="openshift-marketplace/community-operators-pcspm" Oct 05 08:26:01 crc kubenswrapper[4846]: I1005 08:26:01.227414 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rprsd\" (UniqueName: \"kubernetes.io/projected/9d267364-8423-4287-b7b7-31750ffb9a30-kube-api-access-rprsd\") pod \"community-operators-pcspm\" (UID: \"9d267364-8423-4287-b7b7-31750ffb9a30\") " pod="openshift-marketplace/community-operators-pcspm" Oct 05 08:26:01 crc kubenswrapper[4846]: I1005 08:26:01.328812 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rprsd\" (UniqueName: \"kubernetes.io/projected/9d267364-8423-4287-b7b7-31750ffb9a30-kube-api-access-rprsd\") pod \"community-operators-pcspm\" (UID: \"9d267364-8423-4287-b7b7-31750ffb9a30\") " pod="openshift-marketplace/community-operators-pcspm" Oct 05 08:26:01 crc kubenswrapper[4846]: I1005 08:26:01.328958 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9d267364-8423-4287-b7b7-31750ffb9a30-utilities\") pod \"community-operators-pcspm\" (UID: \"9d267364-8423-4287-b7b7-31750ffb9a30\") " pod="openshift-marketplace/community-operators-pcspm" Oct 05 08:26:01 crc kubenswrapper[4846]: I1005 08:26:01.328986 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9d267364-8423-4287-b7b7-31750ffb9a30-catalog-content\") pod \"community-operators-pcspm\" (UID: \"9d267364-8423-4287-b7b7-31750ffb9a30\") " pod="openshift-marketplace/community-operators-pcspm" Oct 05 08:26:01 crc kubenswrapper[4846]: I1005 08:26:01.329591 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9d267364-8423-4287-b7b7-31750ffb9a30-catalog-content\") pod \"community-operators-pcspm\" (UID: \"9d267364-8423-4287-b7b7-31750ffb9a30\") " pod="openshift-marketplace/community-operators-pcspm" Oct 05 08:26:01 crc kubenswrapper[4846]: I1005 08:26:01.330083 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9d267364-8423-4287-b7b7-31750ffb9a30-utilities\") pod \"community-operators-pcspm\" (UID: \"9d267364-8423-4287-b7b7-31750ffb9a30\") " pod="openshift-marketplace/community-operators-pcspm" Oct 05 08:26:01 crc kubenswrapper[4846]: I1005 08:26:01.361998 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rprsd\" (UniqueName: \"kubernetes.io/projected/9d267364-8423-4287-b7b7-31750ffb9a30-kube-api-access-rprsd\") pod \"community-operators-pcspm\" (UID: \"9d267364-8423-4287-b7b7-31750ffb9a30\") " pod="openshift-marketplace/community-operators-pcspm" Oct 05 08:26:01 crc kubenswrapper[4846]: I1005 08:26:01.497511 4846 scope.go:117] "RemoveContainer" containerID="fdafa27da24296f9087c33b6b12a6598d2a8d61fc985b06ad680401828c9feb6" Oct 05 08:26:01 crc kubenswrapper[4846]: E1005 08:26:01.497831 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 08:26:01 crc kubenswrapper[4846]: I1005 08:26:01.658733 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-pcspm" Oct 05 08:26:03 crc kubenswrapper[4846]: I1005 08:26:03.048375 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-57a8-account-create-tjbkx"] Oct 05 08:26:03 crc kubenswrapper[4846]: I1005 08:26:03.086341 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-57a8-account-create-tjbkx"] Oct 05 08:26:03 crc kubenswrapper[4846]: I1005 08:26:03.497614 4846 scope.go:117] "RemoveContainer" containerID="2afe0c9e24d73820fd0e74d9e701dd3e0b5e5424e84fa2d6087a65639080a151" Oct 05 08:26:03 crc kubenswrapper[4846]: I1005 08:26:03.622093 4846 scope.go:117] "RemoveContainer" containerID="8e3d24f7a9c614dd76a5fe862718bcd0dda3735120671933705a9bef54139d4f" Oct 05 08:26:03 crc kubenswrapper[4846]: I1005 08:26:03.723487 4846 scope.go:117] "RemoveContainer" containerID="d0b2a95a1c72c58f215d678122d3d9a70d119c082300731b85c1b4d15fcb5ab0" Oct 05 08:26:04 crc kubenswrapper[4846]: I1005 08:26:04.190422 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 05 08:26:04 crc kubenswrapper[4846]: I1005 08:26:04.220313 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-pcspm"] Oct 05 08:26:04 crc kubenswrapper[4846]: I1005 08:26:04.285886 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 05 08:26:04 crc kubenswrapper[4846]: W1005 08:26:04.295965 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda913b914_fe8a_4d4a_9969_7f4df6313f02.slice/crio-39bb0122a4520de21a4e0e2eacac39794b2c08124de020fc7a83c72ec0e75082 WatchSource:0}: Error finding container 39bb0122a4520de21a4e0e2eacac39794b2c08124de020fc7a83c72ec0e75082: Status 404 returned error can't find the container with id 39bb0122a4520de21a4e0e2eacac39794b2c08124de020fc7a83c72ec0e75082 Oct 05 08:26:04 crc kubenswrapper[4846]: I1005 08:26:04.512878 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="674af39b-f67b-4e1c-872b-4115a622dc8c" path="/var/lib/kubelet/pods/674af39b-f67b-4e1c-872b-4115a622dc8c/volumes" Oct 05 08:26:04 crc kubenswrapper[4846]: I1005 08:26:04.653492 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-896999bb9-9s4tk" event={"ID":"60577687-a955-4db0-a8a3-4a3bb3592d3d","Type":"ContainerStarted","Data":"a46b538411ffdc6325323b2290356fa188c5541bc8d61e4f704531ed0afa5218"} Oct 05 08:26:04 crc kubenswrapper[4846]: I1005 08:26:04.653550 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-896999bb9-9s4tk" event={"ID":"60577687-a955-4db0-a8a3-4a3bb3592d3d","Type":"ContainerStarted","Data":"4cd4dba45e425213b6db3d4755355db87fb9bb6161c24db4e6122856a4867458"} Oct 05 08:26:04 crc kubenswrapper[4846]: I1005 08:26:04.653616 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-896999bb9-9s4tk" podUID="60577687-a955-4db0-a8a3-4a3bb3592d3d" containerName="horizon-log" containerID="cri-o://4cd4dba45e425213b6db3d4755355db87fb9bb6161c24db4e6122856a4867458" gracePeriod=30 Oct 05 08:26:04 crc kubenswrapper[4846]: I1005 08:26:04.653736 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-896999bb9-9s4tk" podUID="60577687-a955-4db0-a8a3-4a3bb3592d3d" containerName="horizon" containerID="cri-o://a46b538411ffdc6325323b2290356fa188c5541bc8d61e4f704531ed0afa5218" gracePeriod=30 Oct 05 08:26:04 crc kubenswrapper[4846]: I1005 08:26:04.656986 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-57d69df789-t686p" event={"ID":"a7850a96-3d5e-4b6e-938b-91cfa2d7d3cd","Type":"ContainerStarted","Data":"8d11d5b70809900c4244838b21da99802cdb2c4ccda42e7a52e6f6870375b13c"} Oct 05 08:26:04 crc kubenswrapper[4846]: I1005 08:26:04.657019 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-57d69df789-t686p" event={"ID":"a7850a96-3d5e-4b6e-938b-91cfa2d7d3cd","Type":"ContainerStarted","Data":"e4992853fa52b5bb3a31d5a9acfce433dfb842a99e551f0bcd14839c7aabb16b"} Oct 05 08:26:04 crc kubenswrapper[4846]: I1005 08:26:04.657111 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-57d69df789-t686p" podUID="a7850a96-3d5e-4b6e-938b-91cfa2d7d3cd" containerName="horizon-log" containerID="cri-o://e4992853fa52b5bb3a31d5a9acfce433dfb842a99e551f0bcd14839c7aabb16b" gracePeriod=30 Oct 05 08:26:04 crc kubenswrapper[4846]: I1005 08:26:04.657375 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-57d69df789-t686p" podUID="a7850a96-3d5e-4b6e-938b-91cfa2d7d3cd" containerName="horizon" containerID="cri-o://8d11d5b70809900c4244838b21da99802cdb2c4ccda42e7a52e6f6870375b13c" gracePeriod=30 Oct 05 08:26:04 crc kubenswrapper[4846]: I1005 08:26:04.660327 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-f8ddb658d-6wlz2" event={"ID":"8329aafa-c3d7-474c-a31b-a56e0d23ddce","Type":"ContainerStarted","Data":"d5ef0f6c75a94d51341d689bb7cb65065aad64a24ca9ea981f961495ba170d35"} Oct 05 08:26:04 crc kubenswrapper[4846]: I1005 08:26:04.660375 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-f8ddb658d-6wlz2" event={"ID":"8329aafa-c3d7-474c-a31b-a56e0d23ddce","Type":"ContainerStarted","Data":"22c179cdba0416e23837204edd3d2a434931018a7d897c19b9aa1476bc215022"} Oct 05 08:26:04 crc kubenswrapper[4846]: I1005 08:26:04.666678 4846 generic.go:334] "Generic (PLEG): container finished" podID="9d267364-8423-4287-b7b7-31750ffb9a30" containerID="962c3597ef6571c263bf8a2c037026f9ac6f415079a4dc795fa4fa6d4de0ad43" exitCode=0 Oct 05 08:26:04 crc kubenswrapper[4846]: I1005 08:26:04.666748 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pcspm" event={"ID":"9d267364-8423-4287-b7b7-31750ffb9a30","Type":"ContainerDied","Data":"962c3597ef6571c263bf8a2c037026f9ac6f415079a4dc795fa4fa6d4de0ad43"} Oct 05 08:26:04 crc kubenswrapper[4846]: I1005 08:26:04.666776 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pcspm" event={"ID":"9d267364-8423-4287-b7b7-31750ffb9a30","Type":"ContainerStarted","Data":"507893ea8a861add1a79af8403b4ee0177dcc339043fc96160d46ec64b18f7de"} Oct 05 08:26:04 crc kubenswrapper[4846]: I1005 08:26:04.669614 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-d98bd9cc6-4s784" event={"ID":"c8df1242-e04a-450b-a0b4-9d408f4b3862","Type":"ContainerStarted","Data":"ef3afe5c2b471b1257316a3a5ac93be12891e742390679711548a50b322dfd7c"} Oct 05 08:26:04 crc kubenswrapper[4846]: I1005 08:26:04.669642 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-d98bd9cc6-4s784" event={"ID":"c8df1242-e04a-450b-a0b4-9d408f4b3862","Type":"ContainerStarted","Data":"fa6050e2b9830faaa192d66869bdf6af4ade7688b8f5126c21c8162a7872705e"} Oct 05 08:26:04 crc kubenswrapper[4846]: I1005 08:26:04.672540 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"a913b914-fe8a-4d4a-9969-7f4df6313f02","Type":"ContainerStarted","Data":"39bb0122a4520de21a4e0e2eacac39794b2c08124de020fc7a83c72ec0e75082"} Oct 05 08:26:04 crc kubenswrapper[4846]: I1005 08:26:04.674097 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"25d587cf-9f44-4c05-9ff9-c427f8337de5","Type":"ContainerStarted","Data":"9dcab1dce57109149aca7118d4bf0b835d181fead255212a26c897e03927970f"} Oct 05 08:26:04 crc kubenswrapper[4846]: I1005 08:26:04.681529 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-896999bb9-9s4tk" podStartSLOduration=1.935441166 podStartE2EDuration="10.681507516s" podCreationTimestamp="2025-10-05 08:25:54 +0000 UTC" firstStartedPulling="2025-10-05 08:25:54.908466444 +0000 UTC m=+5877.149319219" lastFinishedPulling="2025-10-05 08:26:03.654532794 +0000 UTC m=+5885.895385569" observedRunningTime="2025-10-05 08:26:04.67524412 +0000 UTC m=+5886.916096895" watchObservedRunningTime="2025-10-05 08:26:04.681507516 +0000 UTC m=+5886.922360291" Oct 05 08:26:04 crc kubenswrapper[4846]: I1005 08:26:04.703927 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-57d69df789-t686p" podStartSLOduration=2.817680148 podStartE2EDuration="11.703911731s" podCreationTimestamp="2025-10-05 08:25:53 +0000 UTC" firstStartedPulling="2025-10-05 08:25:54.839508312 +0000 UTC m=+5877.080361077" lastFinishedPulling="2025-10-05 08:26:03.725739855 +0000 UTC m=+5885.966592660" observedRunningTime="2025-10-05 08:26:04.697828039 +0000 UTC m=+5886.938680824" watchObservedRunningTime="2025-10-05 08:26:04.703911731 +0000 UTC m=+5886.944764506" Oct 05 08:26:04 crc kubenswrapper[4846]: I1005 08:26:04.734979 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-d98bd9cc6-4s784" podStartSLOduration=2.703371183 podStartE2EDuration="9.734963746s" podCreationTimestamp="2025-10-05 08:25:55 +0000 UTC" firstStartedPulling="2025-10-05 08:25:56.693286019 +0000 UTC m=+5878.934138794" lastFinishedPulling="2025-10-05 08:26:03.724878572 +0000 UTC m=+5885.965731357" observedRunningTime="2025-10-05 08:26:04.729498321 +0000 UTC m=+5886.970351096" watchObservedRunningTime="2025-10-05 08:26:04.734963746 +0000 UTC m=+5886.975816521" Oct 05 08:26:04 crc kubenswrapper[4846]: I1005 08:26:04.758488 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-f8ddb658d-6wlz2" podStartSLOduration=2.818911431 podStartE2EDuration="9.75846425s" podCreationTimestamp="2025-10-05 08:25:55 +0000 UTC" firstStartedPulling="2025-10-05 08:25:56.788710023 +0000 UTC m=+5879.029562808" lastFinishedPulling="2025-10-05 08:26:03.728262842 +0000 UTC m=+5885.969115627" observedRunningTime="2025-10-05 08:26:04.749953454 +0000 UTC m=+5886.990806229" watchObservedRunningTime="2025-10-05 08:26:04.75846425 +0000 UTC m=+5886.999317025" Oct 05 08:26:05 crc kubenswrapper[4846]: I1005 08:26:05.689634 4846 generic.go:334] "Generic (PLEG): container finished" podID="9d267364-8423-4287-b7b7-31750ffb9a30" containerID="ee9d85161c089dc4f7a056cbee82fbe82d7c8fe4732482f0c8dd8e862ad02070" exitCode=0 Oct 05 08:26:05 crc kubenswrapper[4846]: I1005 08:26:05.689731 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pcspm" event={"ID":"9d267364-8423-4287-b7b7-31750ffb9a30","Type":"ContainerDied","Data":"ee9d85161c089dc4f7a056cbee82fbe82d7c8fe4732482f0c8dd8e862ad02070"} Oct 05 08:26:05 crc kubenswrapper[4846]: I1005 08:26:05.695424 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"a913b914-fe8a-4d4a-9969-7f4df6313f02","Type":"ContainerStarted","Data":"5651150daf8487d41a91296197914f79e7b4d0b8b5ec7be0d3a576523941020e"} Oct 05 08:26:05 crc kubenswrapper[4846]: I1005 08:26:05.695467 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"a913b914-fe8a-4d4a-9969-7f4df6313f02","Type":"ContainerStarted","Data":"2dca798a0f87e108d0f48f308c4170830aea5c2e2dddf62564aca6108b92368b"} Oct 05 08:26:05 crc kubenswrapper[4846]: I1005 08:26:05.701356 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"25d587cf-9f44-4c05-9ff9-c427f8337de5","Type":"ContainerStarted","Data":"2f6e0c33003210a71876fa86bd53dcc7fbc969768a1e4df2dcccd06e263fc00e"} Oct 05 08:26:05 crc kubenswrapper[4846]: I1005 08:26:05.701391 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"25d587cf-9f44-4c05-9ff9-c427f8337de5","Type":"ContainerStarted","Data":"136d0d0fd37a38b317a66b15b1d336883c8c1f1823d42e822eae58ca82b67f9a"} Oct 05 08:26:05 crc kubenswrapper[4846]: I1005 08:26:05.743578 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=7.7435585289999995 podStartE2EDuration="7.743558529s" podCreationTimestamp="2025-10-05 08:25:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 08:26:05.734655883 +0000 UTC m=+5887.975508668" watchObservedRunningTime="2025-10-05 08:26:05.743558529 +0000 UTC m=+5887.984411304" Oct 05 08:26:05 crc kubenswrapper[4846]: I1005 08:26:05.756604 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=7.7565837250000005 podStartE2EDuration="7.756583725s" podCreationTimestamp="2025-10-05 08:25:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 08:26:05.756247886 +0000 UTC m=+5887.997100671" watchObservedRunningTime="2025-10-05 08:26:05.756583725 +0000 UTC m=+5887.997436500" Oct 05 08:26:06 crc kubenswrapper[4846]: I1005 08:26:06.194907 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-d98bd9cc6-4s784" Oct 05 08:26:06 crc kubenswrapper[4846]: I1005 08:26:06.194962 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-d98bd9cc6-4s784" Oct 05 08:26:06 crc kubenswrapper[4846]: I1005 08:26:06.263760 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-f8ddb658d-6wlz2" Oct 05 08:26:06 crc kubenswrapper[4846]: I1005 08:26:06.264088 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-f8ddb658d-6wlz2" Oct 05 08:26:06 crc kubenswrapper[4846]: I1005 08:26:06.716744 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pcspm" event={"ID":"9d267364-8423-4287-b7b7-31750ffb9a30","Type":"ContainerStarted","Data":"2538b291eba770e3f6b3b160f084726ea9df586a442533aa06f4dafb8fbe1d64"} Oct 05 08:26:06 crc kubenswrapper[4846]: I1005 08:26:06.740315 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-pcspm" podStartSLOduration=4.292364077 podStartE2EDuration="5.740290287s" podCreationTimestamp="2025-10-05 08:26:01 +0000 UTC" firstStartedPulling="2025-10-05 08:26:04.668063299 +0000 UTC m=+5886.908916074" lastFinishedPulling="2025-10-05 08:26:06.115989469 +0000 UTC m=+5888.356842284" observedRunningTime="2025-10-05 08:26:06.740098732 +0000 UTC m=+5888.980951547" watchObservedRunningTime="2025-10-05 08:26:06.740290287 +0000 UTC m=+5888.981143072" Oct 05 08:26:08 crc kubenswrapper[4846]: I1005 08:26:08.980802 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Oct 05 08:26:08 crc kubenswrapper[4846]: I1005 08:26:08.986639 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Oct 05 08:26:08 crc kubenswrapper[4846]: I1005 08:26:08.996801 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Oct 05 08:26:08 crc kubenswrapper[4846]: I1005 08:26:08.996857 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Oct 05 08:26:09 crc kubenswrapper[4846]: I1005 08:26:09.028924 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Oct 05 08:26:09 crc kubenswrapper[4846]: I1005 08:26:09.029493 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Oct 05 08:26:09 crc kubenswrapper[4846]: I1005 08:26:09.030432 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Oct 05 08:26:09 crc kubenswrapper[4846]: I1005 08:26:09.050131 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Oct 05 08:26:09 crc kubenswrapper[4846]: I1005 08:26:09.767998 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Oct 05 08:26:09 crc kubenswrapper[4846]: I1005 08:26:09.769102 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Oct 05 08:26:09 crc kubenswrapper[4846]: I1005 08:26:09.769142 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Oct 05 08:26:09 crc kubenswrapper[4846]: I1005 08:26:09.769163 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Oct 05 08:26:11 crc kubenswrapper[4846]: I1005 08:26:11.638778 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Oct 05 08:26:11 crc kubenswrapper[4846]: I1005 08:26:11.659488 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-pcspm" Oct 05 08:26:11 crc kubenswrapper[4846]: I1005 08:26:11.659530 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-pcspm" Oct 05 08:26:11 crc kubenswrapper[4846]: I1005 08:26:11.749920 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-pcspm" Oct 05 08:26:11 crc kubenswrapper[4846]: I1005 08:26:11.758339 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Oct 05 08:26:11 crc kubenswrapper[4846]: I1005 08:26:11.764072 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Oct 05 08:26:11 crc kubenswrapper[4846]: I1005 08:26:11.916143 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-pcspm" Oct 05 08:26:12 crc kubenswrapper[4846]: I1005 08:26:12.006434 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-pcspm"] Oct 05 08:26:12 crc kubenswrapper[4846]: I1005 08:26:12.801650 4846 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 05 08:26:12 crc kubenswrapper[4846]: I1005 08:26:12.911900 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Oct 05 08:26:13 crc kubenswrapper[4846]: I1005 08:26:13.812491 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-pcspm" podUID="9d267364-8423-4287-b7b7-31750ffb9a30" containerName="registry-server" containerID="cri-o://2538b291eba770e3f6b3b160f084726ea9df586a442533aa06f4dafb8fbe1d64" gracePeriod=2 Oct 05 08:26:14 crc kubenswrapper[4846]: I1005 08:26:14.027839 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-sync-n84rx"] Oct 05 08:26:14 crc kubenswrapper[4846]: I1005 08:26:14.046534 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-sync-n84rx"] Oct 05 08:26:14 crc kubenswrapper[4846]: I1005 08:26:14.276277 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-pcspm" Oct 05 08:26:14 crc kubenswrapper[4846]: I1005 08:26:14.342250 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-57d69df789-t686p" Oct 05 08:26:14 crc kubenswrapper[4846]: I1005 08:26:14.396795 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-896999bb9-9s4tk" Oct 05 08:26:14 crc kubenswrapper[4846]: I1005 08:26:14.437638 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9d267364-8423-4287-b7b7-31750ffb9a30-utilities\") pod \"9d267364-8423-4287-b7b7-31750ffb9a30\" (UID: \"9d267364-8423-4287-b7b7-31750ffb9a30\") " Oct 05 08:26:14 crc kubenswrapper[4846]: I1005 08:26:14.437720 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rprsd\" (UniqueName: \"kubernetes.io/projected/9d267364-8423-4287-b7b7-31750ffb9a30-kube-api-access-rprsd\") pod \"9d267364-8423-4287-b7b7-31750ffb9a30\" (UID: \"9d267364-8423-4287-b7b7-31750ffb9a30\") " Oct 05 08:26:14 crc kubenswrapper[4846]: I1005 08:26:14.437762 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9d267364-8423-4287-b7b7-31750ffb9a30-catalog-content\") pod \"9d267364-8423-4287-b7b7-31750ffb9a30\" (UID: \"9d267364-8423-4287-b7b7-31750ffb9a30\") " Oct 05 08:26:14 crc kubenswrapper[4846]: I1005 08:26:14.457971 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9d267364-8423-4287-b7b7-31750ffb9a30-utilities" (OuterVolumeSpecName: "utilities") pod "9d267364-8423-4287-b7b7-31750ffb9a30" (UID: "9d267364-8423-4287-b7b7-31750ffb9a30"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 08:26:14 crc kubenswrapper[4846]: I1005 08:26:14.464366 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d267364-8423-4287-b7b7-31750ffb9a30-kube-api-access-rprsd" (OuterVolumeSpecName: "kube-api-access-rprsd") pod "9d267364-8423-4287-b7b7-31750ffb9a30" (UID: "9d267364-8423-4287-b7b7-31750ffb9a30"). InnerVolumeSpecName "kube-api-access-rprsd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:26:14 crc kubenswrapper[4846]: I1005 08:26:14.479148 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9d267364-8423-4287-b7b7-31750ffb9a30-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9d267364-8423-4287-b7b7-31750ffb9a30" (UID: "9d267364-8423-4287-b7b7-31750ffb9a30"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 08:26:14 crc kubenswrapper[4846]: I1005 08:26:14.517297 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f92090a2-228b-41f7-b97b-11df5415877a" path="/var/lib/kubelet/pods/f92090a2-228b-41f7-b97b-11df5415877a/volumes" Oct 05 08:26:14 crc kubenswrapper[4846]: I1005 08:26:14.539790 4846 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9d267364-8423-4287-b7b7-31750ffb9a30-utilities\") on node \"crc\" DevicePath \"\"" Oct 05 08:26:14 crc kubenswrapper[4846]: I1005 08:26:14.539832 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rprsd\" (UniqueName: \"kubernetes.io/projected/9d267364-8423-4287-b7b7-31750ffb9a30-kube-api-access-rprsd\") on node \"crc\" DevicePath \"\"" Oct 05 08:26:14 crc kubenswrapper[4846]: I1005 08:26:14.539842 4846 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9d267364-8423-4287-b7b7-31750ffb9a30-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 05 08:26:14 crc kubenswrapper[4846]: I1005 08:26:14.828590 4846 generic.go:334] "Generic (PLEG): container finished" podID="9d267364-8423-4287-b7b7-31750ffb9a30" containerID="2538b291eba770e3f6b3b160f084726ea9df586a442533aa06f4dafb8fbe1d64" exitCode=0 Oct 05 08:26:14 crc kubenswrapper[4846]: I1005 08:26:14.828639 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pcspm" event={"ID":"9d267364-8423-4287-b7b7-31750ffb9a30","Type":"ContainerDied","Data":"2538b291eba770e3f6b3b160f084726ea9df586a442533aa06f4dafb8fbe1d64"} Oct 05 08:26:14 crc kubenswrapper[4846]: I1005 08:26:14.828673 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pcspm" event={"ID":"9d267364-8423-4287-b7b7-31750ffb9a30","Type":"ContainerDied","Data":"507893ea8a861add1a79af8403b4ee0177dcc339043fc96160d46ec64b18f7de"} Oct 05 08:26:14 crc kubenswrapper[4846]: I1005 08:26:14.828691 4846 scope.go:117] "RemoveContainer" containerID="2538b291eba770e3f6b3b160f084726ea9df586a442533aa06f4dafb8fbe1d64" Oct 05 08:26:14 crc kubenswrapper[4846]: I1005 08:26:14.828938 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-pcspm" Oct 05 08:26:14 crc kubenswrapper[4846]: I1005 08:26:14.865440 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-pcspm"] Oct 05 08:26:14 crc kubenswrapper[4846]: I1005 08:26:14.874145 4846 scope.go:117] "RemoveContainer" containerID="ee9d85161c089dc4f7a056cbee82fbe82d7c8fe4732482f0c8dd8e862ad02070" Oct 05 08:26:14 crc kubenswrapper[4846]: I1005 08:26:14.874424 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-pcspm"] Oct 05 08:26:14 crc kubenswrapper[4846]: I1005 08:26:14.906701 4846 scope.go:117] "RemoveContainer" containerID="962c3597ef6571c263bf8a2c037026f9ac6f415079a4dc795fa4fa6d4de0ad43" Oct 05 08:26:14 crc kubenswrapper[4846]: I1005 08:26:14.956126 4846 scope.go:117] "RemoveContainer" containerID="2538b291eba770e3f6b3b160f084726ea9df586a442533aa06f4dafb8fbe1d64" Oct 05 08:26:14 crc kubenswrapper[4846]: E1005 08:26:14.956750 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2538b291eba770e3f6b3b160f084726ea9df586a442533aa06f4dafb8fbe1d64\": container with ID starting with 2538b291eba770e3f6b3b160f084726ea9df586a442533aa06f4dafb8fbe1d64 not found: ID does not exist" containerID="2538b291eba770e3f6b3b160f084726ea9df586a442533aa06f4dafb8fbe1d64" Oct 05 08:26:14 crc kubenswrapper[4846]: I1005 08:26:14.956800 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2538b291eba770e3f6b3b160f084726ea9df586a442533aa06f4dafb8fbe1d64"} err="failed to get container status \"2538b291eba770e3f6b3b160f084726ea9df586a442533aa06f4dafb8fbe1d64\": rpc error: code = NotFound desc = could not find container \"2538b291eba770e3f6b3b160f084726ea9df586a442533aa06f4dafb8fbe1d64\": container with ID starting with 2538b291eba770e3f6b3b160f084726ea9df586a442533aa06f4dafb8fbe1d64 not found: ID does not exist" Oct 05 08:26:14 crc kubenswrapper[4846]: I1005 08:26:14.956941 4846 scope.go:117] "RemoveContainer" containerID="ee9d85161c089dc4f7a056cbee82fbe82d7c8fe4732482f0c8dd8e862ad02070" Oct 05 08:26:14 crc kubenswrapper[4846]: E1005 08:26:14.957424 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ee9d85161c089dc4f7a056cbee82fbe82d7c8fe4732482f0c8dd8e862ad02070\": container with ID starting with ee9d85161c089dc4f7a056cbee82fbe82d7c8fe4732482f0c8dd8e862ad02070 not found: ID does not exist" containerID="ee9d85161c089dc4f7a056cbee82fbe82d7c8fe4732482f0c8dd8e862ad02070" Oct 05 08:26:14 crc kubenswrapper[4846]: I1005 08:26:14.957504 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ee9d85161c089dc4f7a056cbee82fbe82d7c8fe4732482f0c8dd8e862ad02070"} err="failed to get container status \"ee9d85161c089dc4f7a056cbee82fbe82d7c8fe4732482f0c8dd8e862ad02070\": rpc error: code = NotFound desc = could not find container \"ee9d85161c089dc4f7a056cbee82fbe82d7c8fe4732482f0c8dd8e862ad02070\": container with ID starting with ee9d85161c089dc4f7a056cbee82fbe82d7c8fe4732482f0c8dd8e862ad02070 not found: ID does not exist" Oct 05 08:26:14 crc kubenswrapper[4846]: I1005 08:26:14.957530 4846 scope.go:117] "RemoveContainer" containerID="962c3597ef6571c263bf8a2c037026f9ac6f415079a4dc795fa4fa6d4de0ad43" Oct 05 08:26:14 crc kubenswrapper[4846]: E1005 08:26:14.957838 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"962c3597ef6571c263bf8a2c037026f9ac6f415079a4dc795fa4fa6d4de0ad43\": container with ID starting with 962c3597ef6571c263bf8a2c037026f9ac6f415079a4dc795fa4fa6d4de0ad43 not found: ID does not exist" containerID="962c3597ef6571c263bf8a2c037026f9ac6f415079a4dc795fa4fa6d4de0ad43" Oct 05 08:26:14 crc kubenswrapper[4846]: I1005 08:26:14.957892 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"962c3597ef6571c263bf8a2c037026f9ac6f415079a4dc795fa4fa6d4de0ad43"} err="failed to get container status \"962c3597ef6571c263bf8a2c037026f9ac6f415079a4dc795fa4fa6d4de0ad43\": rpc error: code = NotFound desc = could not find container \"962c3597ef6571c263bf8a2c037026f9ac6f415079a4dc795fa4fa6d4de0ad43\": container with ID starting with 962c3597ef6571c263bf8a2c037026f9ac6f415079a4dc795fa4fa6d4de0ad43 not found: ID does not exist" Oct 05 08:26:15 crc kubenswrapper[4846]: I1005 08:26:15.497572 4846 scope.go:117] "RemoveContainer" containerID="fdafa27da24296f9087c33b6b12a6598d2a8d61fc985b06ad680401828c9feb6" Oct 05 08:26:15 crc kubenswrapper[4846]: E1005 08:26:15.497868 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 08:26:16 crc kubenswrapper[4846]: I1005 08:26:16.196932 4846 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-d98bd9cc6-4s784" podUID="c8df1242-e04a-450b-a0b4-9d408f4b3862" containerName="horizon" probeResult="failure" output="Get \"https://10.217.1.102:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.102:8443: connect: connection refused" Oct 05 08:26:16 crc kubenswrapper[4846]: I1005 08:26:16.266277 4846 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-f8ddb658d-6wlz2" podUID="8329aafa-c3d7-474c-a31b-a56e0d23ddce" containerName="horizon" probeResult="failure" output="Get \"https://10.217.1.103:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.103:8443: connect: connection refused" Oct 05 08:26:16 crc kubenswrapper[4846]: I1005 08:26:16.508872 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d267364-8423-4287-b7b7-31750ffb9a30" path="/var/lib/kubelet/pods/9d267364-8423-4287-b7b7-31750ffb9a30/volumes" Oct 05 08:26:28 crc kubenswrapper[4846]: I1005 08:26:28.079060 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-d98bd9cc6-4s784" Oct 05 08:26:28 crc kubenswrapper[4846]: I1005 08:26:28.083446 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-f8ddb658d-6wlz2" Oct 05 08:26:29 crc kubenswrapper[4846]: I1005 08:26:29.694588 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-f8ddb658d-6wlz2" Oct 05 08:26:29 crc kubenswrapper[4846]: I1005 08:26:29.751385 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-d98bd9cc6-4s784" Oct 05 08:26:29 crc kubenswrapper[4846]: I1005 08:26:29.782972 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-d98bd9cc6-4s784"] Oct 05 08:26:30 crc kubenswrapper[4846]: I1005 08:26:30.006080 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-d98bd9cc6-4s784" podUID="c8df1242-e04a-450b-a0b4-9d408f4b3862" containerName="horizon-log" containerID="cri-o://fa6050e2b9830faaa192d66869bdf6af4ade7688b8f5126c21c8162a7872705e" gracePeriod=30 Oct 05 08:26:30 crc kubenswrapper[4846]: I1005 08:26:30.006137 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-d98bd9cc6-4s784" podUID="c8df1242-e04a-450b-a0b4-9d408f4b3862" containerName="horizon" containerID="cri-o://ef3afe5c2b471b1257316a3a5ac93be12891e742390679711548a50b322dfd7c" gracePeriod=30 Oct 05 08:26:30 crc kubenswrapper[4846]: I1005 08:26:30.499695 4846 scope.go:117] "RemoveContainer" containerID="fdafa27da24296f9087c33b6b12a6598d2a8d61fc985b06ad680401828c9feb6" Oct 05 08:26:30 crc kubenswrapper[4846]: E1005 08:26:30.499952 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 08:26:33 crc kubenswrapper[4846]: I1005 08:26:33.173954 4846 scope.go:117] "RemoveContainer" containerID="0b0a481e09ec59c44725cebd2f7c6a88a11897b33401732318eb9b88a948f74f" Oct 05 08:26:33 crc kubenswrapper[4846]: I1005 08:26:33.206889 4846 scope.go:117] "RemoveContainer" containerID="786aa4b04ab3c851cfcfe099c1b123e134328bd98f3a0fb361907803d7ff58ab" Oct 05 08:26:33 crc kubenswrapper[4846]: I1005 08:26:33.314565 4846 scope.go:117] "RemoveContainer" containerID="05037c12b243f1732aa2a993b48998c6aa0359cbab8104a684ee66ccdb12ea92" Oct 05 08:26:34 crc kubenswrapper[4846]: I1005 08:26:34.053152 4846 generic.go:334] "Generic (PLEG): container finished" podID="c8df1242-e04a-450b-a0b4-9d408f4b3862" containerID="ef3afe5c2b471b1257316a3a5ac93be12891e742390679711548a50b322dfd7c" exitCode=0 Oct 05 08:26:34 crc kubenswrapper[4846]: I1005 08:26:34.053247 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-d98bd9cc6-4s784" event={"ID":"c8df1242-e04a-450b-a0b4-9d408f4b3862","Type":"ContainerDied","Data":"ef3afe5c2b471b1257316a3a5ac93be12891e742390679711548a50b322dfd7c"} Oct 05 08:26:35 crc kubenswrapper[4846]: I1005 08:26:35.078927 4846 generic.go:334] "Generic (PLEG): container finished" podID="a7850a96-3d5e-4b6e-938b-91cfa2d7d3cd" containerID="8d11d5b70809900c4244838b21da99802cdb2c4ccda42e7a52e6f6870375b13c" exitCode=137 Oct 05 08:26:35 crc kubenswrapper[4846]: I1005 08:26:35.079300 4846 generic.go:334] "Generic (PLEG): container finished" podID="a7850a96-3d5e-4b6e-938b-91cfa2d7d3cd" containerID="e4992853fa52b5bb3a31d5a9acfce433dfb842a99e551f0bcd14839c7aabb16b" exitCode=137 Oct 05 08:26:35 crc kubenswrapper[4846]: I1005 08:26:35.079003 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-57d69df789-t686p" event={"ID":"a7850a96-3d5e-4b6e-938b-91cfa2d7d3cd","Type":"ContainerDied","Data":"8d11d5b70809900c4244838b21da99802cdb2c4ccda42e7a52e6f6870375b13c"} Oct 05 08:26:35 crc kubenswrapper[4846]: I1005 08:26:35.079677 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-57d69df789-t686p" event={"ID":"a7850a96-3d5e-4b6e-938b-91cfa2d7d3cd","Type":"ContainerDied","Data":"e4992853fa52b5bb3a31d5a9acfce433dfb842a99e551f0bcd14839c7aabb16b"} Oct 05 08:26:35 crc kubenswrapper[4846]: I1005 08:26:35.082095 4846 generic.go:334] "Generic (PLEG): container finished" podID="60577687-a955-4db0-a8a3-4a3bb3592d3d" containerID="a46b538411ffdc6325323b2290356fa188c5541bc8d61e4f704531ed0afa5218" exitCode=137 Oct 05 08:26:35 crc kubenswrapper[4846]: I1005 08:26:35.082118 4846 generic.go:334] "Generic (PLEG): container finished" podID="60577687-a955-4db0-a8a3-4a3bb3592d3d" containerID="4cd4dba45e425213b6db3d4755355db87fb9bb6161c24db4e6122856a4867458" exitCode=137 Oct 05 08:26:35 crc kubenswrapper[4846]: I1005 08:26:35.082100 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-896999bb9-9s4tk" event={"ID":"60577687-a955-4db0-a8a3-4a3bb3592d3d","Type":"ContainerDied","Data":"a46b538411ffdc6325323b2290356fa188c5541bc8d61e4f704531ed0afa5218"} Oct 05 08:26:35 crc kubenswrapper[4846]: I1005 08:26:35.082164 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-896999bb9-9s4tk" event={"ID":"60577687-a955-4db0-a8a3-4a3bb3592d3d","Type":"ContainerDied","Data":"4cd4dba45e425213b6db3d4755355db87fb9bb6161c24db4e6122856a4867458"} Oct 05 08:26:35 crc kubenswrapper[4846]: I1005 08:26:35.259237 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-896999bb9-9s4tk" Oct 05 08:26:35 crc kubenswrapper[4846]: I1005 08:26:35.272884 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-57d69df789-t686p" Oct 05 08:26:35 crc kubenswrapper[4846]: I1005 08:26:35.306722 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/60577687-a955-4db0-a8a3-4a3bb3592d3d-logs\") pod \"60577687-a955-4db0-a8a3-4a3bb3592d3d\" (UID: \"60577687-a955-4db0-a8a3-4a3bb3592d3d\") " Oct 05 08:26:35 crc kubenswrapper[4846]: I1005 08:26:35.306870 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/60577687-a955-4db0-a8a3-4a3bb3592d3d-horizon-secret-key\") pod \"60577687-a955-4db0-a8a3-4a3bb3592d3d\" (UID: \"60577687-a955-4db0-a8a3-4a3bb3592d3d\") " Oct 05 08:26:35 crc kubenswrapper[4846]: I1005 08:26:35.307033 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a7850a96-3d5e-4b6e-938b-91cfa2d7d3cd-scripts\") pod \"a7850a96-3d5e-4b6e-938b-91cfa2d7d3cd\" (UID: \"a7850a96-3d5e-4b6e-938b-91cfa2d7d3cd\") " Oct 05 08:26:35 crc kubenswrapper[4846]: I1005 08:26:35.307131 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a7850a96-3d5e-4b6e-938b-91cfa2d7d3cd-config-data\") pod \"a7850a96-3d5e-4b6e-938b-91cfa2d7d3cd\" (UID: \"a7850a96-3d5e-4b6e-938b-91cfa2d7d3cd\") " Oct 05 08:26:35 crc kubenswrapper[4846]: I1005 08:26:35.307171 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cb5xb\" (UniqueName: \"kubernetes.io/projected/a7850a96-3d5e-4b6e-938b-91cfa2d7d3cd-kube-api-access-cb5xb\") pod \"a7850a96-3d5e-4b6e-938b-91cfa2d7d3cd\" (UID: \"a7850a96-3d5e-4b6e-938b-91cfa2d7d3cd\") " Oct 05 08:26:35 crc kubenswrapper[4846]: I1005 08:26:35.307219 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zl5mn\" (UniqueName: \"kubernetes.io/projected/60577687-a955-4db0-a8a3-4a3bb3592d3d-kube-api-access-zl5mn\") pod \"60577687-a955-4db0-a8a3-4a3bb3592d3d\" (UID: \"60577687-a955-4db0-a8a3-4a3bb3592d3d\") " Oct 05 08:26:35 crc kubenswrapper[4846]: I1005 08:26:35.307250 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/a7850a96-3d5e-4b6e-938b-91cfa2d7d3cd-horizon-secret-key\") pod \"a7850a96-3d5e-4b6e-938b-91cfa2d7d3cd\" (UID: \"a7850a96-3d5e-4b6e-938b-91cfa2d7d3cd\") " Oct 05 08:26:35 crc kubenswrapper[4846]: I1005 08:26:35.307303 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/60577687-a955-4db0-a8a3-4a3bb3592d3d-config-data\") pod \"60577687-a955-4db0-a8a3-4a3bb3592d3d\" (UID: \"60577687-a955-4db0-a8a3-4a3bb3592d3d\") " Oct 05 08:26:35 crc kubenswrapper[4846]: I1005 08:26:35.307382 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a7850a96-3d5e-4b6e-938b-91cfa2d7d3cd-logs\") pod \"a7850a96-3d5e-4b6e-938b-91cfa2d7d3cd\" (UID: \"a7850a96-3d5e-4b6e-938b-91cfa2d7d3cd\") " Oct 05 08:26:35 crc kubenswrapper[4846]: I1005 08:26:35.308862 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/60577687-a955-4db0-a8a3-4a3bb3592d3d-scripts\") pod \"60577687-a955-4db0-a8a3-4a3bb3592d3d\" (UID: \"60577687-a955-4db0-a8a3-4a3bb3592d3d\") " Oct 05 08:26:35 crc kubenswrapper[4846]: I1005 08:26:35.308023 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/60577687-a955-4db0-a8a3-4a3bb3592d3d-logs" (OuterVolumeSpecName: "logs") pod "60577687-a955-4db0-a8a3-4a3bb3592d3d" (UID: "60577687-a955-4db0-a8a3-4a3bb3592d3d"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 08:26:35 crc kubenswrapper[4846]: I1005 08:26:35.310632 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a7850a96-3d5e-4b6e-938b-91cfa2d7d3cd-logs" (OuterVolumeSpecName: "logs") pod "a7850a96-3d5e-4b6e-938b-91cfa2d7d3cd" (UID: "a7850a96-3d5e-4b6e-938b-91cfa2d7d3cd"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 08:26:35 crc kubenswrapper[4846]: I1005 08:26:35.314361 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a7850a96-3d5e-4b6e-938b-91cfa2d7d3cd-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "a7850a96-3d5e-4b6e-938b-91cfa2d7d3cd" (UID: "a7850a96-3d5e-4b6e-938b-91cfa2d7d3cd"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:26:35 crc kubenswrapper[4846]: I1005 08:26:35.314675 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a7850a96-3d5e-4b6e-938b-91cfa2d7d3cd-kube-api-access-cb5xb" (OuterVolumeSpecName: "kube-api-access-cb5xb") pod "a7850a96-3d5e-4b6e-938b-91cfa2d7d3cd" (UID: "a7850a96-3d5e-4b6e-938b-91cfa2d7d3cd"). InnerVolumeSpecName "kube-api-access-cb5xb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:26:35 crc kubenswrapper[4846]: I1005 08:26:35.314838 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/60577687-a955-4db0-a8a3-4a3bb3592d3d-kube-api-access-zl5mn" (OuterVolumeSpecName: "kube-api-access-zl5mn") pod "60577687-a955-4db0-a8a3-4a3bb3592d3d" (UID: "60577687-a955-4db0-a8a3-4a3bb3592d3d"). InnerVolumeSpecName "kube-api-access-zl5mn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:26:35 crc kubenswrapper[4846]: I1005 08:26:35.320684 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/60577687-a955-4db0-a8a3-4a3bb3592d3d-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "60577687-a955-4db0-a8a3-4a3bb3592d3d" (UID: "60577687-a955-4db0-a8a3-4a3bb3592d3d"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:26:35 crc kubenswrapper[4846]: I1005 08:26:35.343909 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/60577687-a955-4db0-a8a3-4a3bb3592d3d-config-data" (OuterVolumeSpecName: "config-data") pod "60577687-a955-4db0-a8a3-4a3bb3592d3d" (UID: "60577687-a955-4db0-a8a3-4a3bb3592d3d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 08:26:35 crc kubenswrapper[4846]: I1005 08:26:35.347484 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a7850a96-3d5e-4b6e-938b-91cfa2d7d3cd-scripts" (OuterVolumeSpecName: "scripts") pod "a7850a96-3d5e-4b6e-938b-91cfa2d7d3cd" (UID: "a7850a96-3d5e-4b6e-938b-91cfa2d7d3cd"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 08:26:35 crc kubenswrapper[4846]: I1005 08:26:35.350688 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a7850a96-3d5e-4b6e-938b-91cfa2d7d3cd-config-data" (OuterVolumeSpecName: "config-data") pod "a7850a96-3d5e-4b6e-938b-91cfa2d7d3cd" (UID: "a7850a96-3d5e-4b6e-938b-91cfa2d7d3cd"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 08:26:35 crc kubenswrapper[4846]: I1005 08:26:35.351076 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/60577687-a955-4db0-a8a3-4a3bb3592d3d-scripts" (OuterVolumeSpecName: "scripts") pod "60577687-a955-4db0-a8a3-4a3bb3592d3d" (UID: "60577687-a955-4db0-a8a3-4a3bb3592d3d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 08:26:35 crc kubenswrapper[4846]: I1005 08:26:35.412025 4846 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/60577687-a955-4db0-a8a3-4a3bb3592d3d-scripts\") on node \"crc\" DevicePath \"\"" Oct 05 08:26:35 crc kubenswrapper[4846]: I1005 08:26:35.412258 4846 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/60577687-a955-4db0-a8a3-4a3bb3592d3d-logs\") on node \"crc\" DevicePath \"\"" Oct 05 08:26:35 crc kubenswrapper[4846]: I1005 08:26:35.412386 4846 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/60577687-a955-4db0-a8a3-4a3bb3592d3d-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Oct 05 08:26:35 crc kubenswrapper[4846]: I1005 08:26:35.412467 4846 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a7850a96-3d5e-4b6e-938b-91cfa2d7d3cd-scripts\") on node \"crc\" DevicePath \"\"" Oct 05 08:26:35 crc kubenswrapper[4846]: I1005 08:26:35.412548 4846 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a7850a96-3d5e-4b6e-938b-91cfa2d7d3cd-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 08:26:35 crc kubenswrapper[4846]: I1005 08:26:35.412625 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cb5xb\" (UniqueName: \"kubernetes.io/projected/a7850a96-3d5e-4b6e-938b-91cfa2d7d3cd-kube-api-access-cb5xb\") on node \"crc\" DevicePath \"\"" Oct 05 08:26:35 crc kubenswrapper[4846]: I1005 08:26:35.412695 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zl5mn\" (UniqueName: \"kubernetes.io/projected/60577687-a955-4db0-a8a3-4a3bb3592d3d-kube-api-access-zl5mn\") on node \"crc\" DevicePath \"\"" Oct 05 08:26:35 crc kubenswrapper[4846]: I1005 08:26:35.412760 4846 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/a7850a96-3d5e-4b6e-938b-91cfa2d7d3cd-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Oct 05 08:26:35 crc kubenswrapper[4846]: I1005 08:26:35.412840 4846 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/60577687-a955-4db0-a8a3-4a3bb3592d3d-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 08:26:35 crc kubenswrapper[4846]: I1005 08:26:35.412909 4846 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a7850a96-3d5e-4b6e-938b-91cfa2d7d3cd-logs\") on node \"crc\" DevicePath \"\"" Oct 05 08:26:36 crc kubenswrapper[4846]: I1005 08:26:36.096396 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-896999bb9-9s4tk" event={"ID":"60577687-a955-4db0-a8a3-4a3bb3592d3d","Type":"ContainerDied","Data":"c36ee474c579c405b3f96953dda49a0a00393e3aaeb619ea5e0d70b49e9fea6f"} Oct 05 08:26:36 crc kubenswrapper[4846]: I1005 08:26:36.096462 4846 scope.go:117] "RemoveContainer" containerID="a46b538411ffdc6325323b2290356fa188c5541bc8d61e4f704531ed0afa5218" Oct 05 08:26:36 crc kubenswrapper[4846]: I1005 08:26:36.097472 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-896999bb9-9s4tk" Oct 05 08:26:36 crc kubenswrapper[4846]: I1005 08:26:36.100020 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-57d69df789-t686p" event={"ID":"a7850a96-3d5e-4b6e-938b-91cfa2d7d3cd","Type":"ContainerDied","Data":"1b1ad7c1c452928e64b3e179da694962dd87843ce67331603eecf625afd06c2a"} Oct 05 08:26:36 crc kubenswrapper[4846]: I1005 08:26:36.100083 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-57d69df789-t686p" Oct 05 08:26:36 crc kubenswrapper[4846]: I1005 08:26:36.159714 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-896999bb9-9s4tk"] Oct 05 08:26:36 crc kubenswrapper[4846]: I1005 08:26:36.171649 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-896999bb9-9s4tk"] Oct 05 08:26:36 crc kubenswrapper[4846]: I1005 08:26:36.181415 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-57d69df789-t686p"] Oct 05 08:26:36 crc kubenswrapper[4846]: I1005 08:26:36.189103 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-57d69df789-t686p"] Oct 05 08:26:36 crc kubenswrapper[4846]: I1005 08:26:36.195215 4846 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-d98bd9cc6-4s784" podUID="c8df1242-e04a-450b-a0b4-9d408f4b3862" containerName="horizon" probeResult="failure" output="Get \"https://10.217.1.102:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.102:8443: connect: connection refused" Oct 05 08:26:36 crc kubenswrapper[4846]: I1005 08:26:36.294889 4846 scope.go:117] "RemoveContainer" containerID="4cd4dba45e425213b6db3d4755355db87fb9bb6161c24db4e6122856a4867458" Oct 05 08:26:36 crc kubenswrapper[4846]: I1005 08:26:36.317715 4846 scope.go:117] "RemoveContainer" containerID="8d11d5b70809900c4244838b21da99802cdb2c4ccda42e7a52e6f6870375b13c" Oct 05 08:26:36 crc kubenswrapper[4846]: I1005 08:26:36.543340 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="60577687-a955-4db0-a8a3-4a3bb3592d3d" path="/var/lib/kubelet/pods/60577687-a955-4db0-a8a3-4a3bb3592d3d/volumes" Oct 05 08:26:36 crc kubenswrapper[4846]: I1005 08:26:36.544633 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a7850a96-3d5e-4b6e-938b-91cfa2d7d3cd" path="/var/lib/kubelet/pods/a7850a96-3d5e-4b6e-938b-91cfa2d7d3cd/volumes" Oct 05 08:26:36 crc kubenswrapper[4846]: I1005 08:26:36.545617 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-hnd8d"] Oct 05 08:26:36 crc kubenswrapper[4846]: E1005 08:26:36.546065 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9d267364-8423-4287-b7b7-31750ffb9a30" containerName="registry-server" Oct 05 08:26:36 crc kubenswrapper[4846]: I1005 08:26:36.546096 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="9d267364-8423-4287-b7b7-31750ffb9a30" containerName="registry-server" Oct 05 08:26:36 crc kubenswrapper[4846]: E1005 08:26:36.546120 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="60577687-a955-4db0-a8a3-4a3bb3592d3d" containerName="horizon" Oct 05 08:26:36 crc kubenswrapper[4846]: I1005 08:26:36.546131 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="60577687-a955-4db0-a8a3-4a3bb3592d3d" containerName="horizon" Oct 05 08:26:36 crc kubenswrapper[4846]: E1005 08:26:36.546154 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9d267364-8423-4287-b7b7-31750ffb9a30" containerName="extract-utilities" Oct 05 08:26:36 crc kubenswrapper[4846]: I1005 08:26:36.546165 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="9d267364-8423-4287-b7b7-31750ffb9a30" containerName="extract-utilities" Oct 05 08:26:36 crc kubenswrapper[4846]: E1005 08:26:36.546269 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a7850a96-3d5e-4b6e-938b-91cfa2d7d3cd" containerName="horizon-log" Oct 05 08:26:36 crc kubenswrapper[4846]: I1005 08:26:36.546283 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="a7850a96-3d5e-4b6e-938b-91cfa2d7d3cd" containerName="horizon-log" Oct 05 08:26:36 crc kubenswrapper[4846]: E1005 08:26:36.546305 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="60577687-a955-4db0-a8a3-4a3bb3592d3d" containerName="horizon-log" Oct 05 08:26:36 crc kubenswrapper[4846]: I1005 08:26:36.546316 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="60577687-a955-4db0-a8a3-4a3bb3592d3d" containerName="horizon-log" Oct 05 08:26:36 crc kubenswrapper[4846]: E1005 08:26:36.546336 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a7850a96-3d5e-4b6e-938b-91cfa2d7d3cd" containerName="horizon" Oct 05 08:26:36 crc kubenswrapper[4846]: I1005 08:26:36.546349 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="a7850a96-3d5e-4b6e-938b-91cfa2d7d3cd" containerName="horizon" Oct 05 08:26:36 crc kubenswrapper[4846]: E1005 08:26:36.546380 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9d267364-8423-4287-b7b7-31750ffb9a30" containerName="extract-content" Oct 05 08:26:36 crc kubenswrapper[4846]: I1005 08:26:36.546391 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="9d267364-8423-4287-b7b7-31750ffb9a30" containerName="extract-content" Oct 05 08:26:36 crc kubenswrapper[4846]: I1005 08:26:36.546674 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="60577687-a955-4db0-a8a3-4a3bb3592d3d" containerName="horizon-log" Oct 05 08:26:36 crc kubenswrapper[4846]: I1005 08:26:36.546696 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="a7850a96-3d5e-4b6e-938b-91cfa2d7d3cd" containerName="horizon" Oct 05 08:26:36 crc kubenswrapper[4846]: I1005 08:26:36.546739 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="9d267364-8423-4287-b7b7-31750ffb9a30" containerName="registry-server" Oct 05 08:26:36 crc kubenswrapper[4846]: I1005 08:26:36.546754 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="60577687-a955-4db0-a8a3-4a3bb3592d3d" containerName="horizon" Oct 05 08:26:36 crc kubenswrapper[4846]: I1005 08:26:36.546781 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="a7850a96-3d5e-4b6e-938b-91cfa2d7d3cd" containerName="horizon-log" Oct 05 08:26:36 crc kubenswrapper[4846]: I1005 08:26:36.550910 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-hnd8d"] Oct 05 08:26:36 crc kubenswrapper[4846]: I1005 08:26:36.551147 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hnd8d" Oct 05 08:26:36 crc kubenswrapper[4846]: I1005 08:26:36.605747 4846 scope.go:117] "RemoveContainer" containerID="e4992853fa52b5bb3a31d5a9acfce433dfb842a99e551f0bcd14839c7aabb16b" Oct 05 08:26:36 crc kubenswrapper[4846]: I1005 08:26:36.642289 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bf1f32be-7e26-400a-b8cb-4360f56d5542-catalog-content\") pod \"redhat-operators-hnd8d\" (UID: \"bf1f32be-7e26-400a-b8cb-4360f56d5542\") " pod="openshift-marketplace/redhat-operators-hnd8d" Oct 05 08:26:36 crc kubenswrapper[4846]: I1005 08:26:36.642518 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tlt56\" (UniqueName: \"kubernetes.io/projected/bf1f32be-7e26-400a-b8cb-4360f56d5542-kube-api-access-tlt56\") pod \"redhat-operators-hnd8d\" (UID: \"bf1f32be-7e26-400a-b8cb-4360f56d5542\") " pod="openshift-marketplace/redhat-operators-hnd8d" Oct 05 08:26:36 crc kubenswrapper[4846]: I1005 08:26:36.642877 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bf1f32be-7e26-400a-b8cb-4360f56d5542-utilities\") pod \"redhat-operators-hnd8d\" (UID: \"bf1f32be-7e26-400a-b8cb-4360f56d5542\") " pod="openshift-marketplace/redhat-operators-hnd8d" Oct 05 08:26:36 crc kubenswrapper[4846]: I1005 08:26:36.744843 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bf1f32be-7e26-400a-b8cb-4360f56d5542-utilities\") pod \"redhat-operators-hnd8d\" (UID: \"bf1f32be-7e26-400a-b8cb-4360f56d5542\") " pod="openshift-marketplace/redhat-operators-hnd8d" Oct 05 08:26:36 crc kubenswrapper[4846]: I1005 08:26:36.745055 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bf1f32be-7e26-400a-b8cb-4360f56d5542-catalog-content\") pod \"redhat-operators-hnd8d\" (UID: \"bf1f32be-7e26-400a-b8cb-4360f56d5542\") " pod="openshift-marketplace/redhat-operators-hnd8d" Oct 05 08:26:36 crc kubenswrapper[4846]: I1005 08:26:36.745133 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tlt56\" (UniqueName: \"kubernetes.io/projected/bf1f32be-7e26-400a-b8cb-4360f56d5542-kube-api-access-tlt56\") pod \"redhat-operators-hnd8d\" (UID: \"bf1f32be-7e26-400a-b8cb-4360f56d5542\") " pod="openshift-marketplace/redhat-operators-hnd8d" Oct 05 08:26:36 crc kubenswrapper[4846]: I1005 08:26:36.745448 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bf1f32be-7e26-400a-b8cb-4360f56d5542-utilities\") pod \"redhat-operators-hnd8d\" (UID: \"bf1f32be-7e26-400a-b8cb-4360f56d5542\") " pod="openshift-marketplace/redhat-operators-hnd8d" Oct 05 08:26:36 crc kubenswrapper[4846]: I1005 08:26:36.745747 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bf1f32be-7e26-400a-b8cb-4360f56d5542-catalog-content\") pod \"redhat-operators-hnd8d\" (UID: \"bf1f32be-7e26-400a-b8cb-4360f56d5542\") " pod="openshift-marketplace/redhat-operators-hnd8d" Oct 05 08:26:36 crc kubenswrapper[4846]: I1005 08:26:36.766283 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tlt56\" (UniqueName: \"kubernetes.io/projected/bf1f32be-7e26-400a-b8cb-4360f56d5542-kube-api-access-tlt56\") pod \"redhat-operators-hnd8d\" (UID: \"bf1f32be-7e26-400a-b8cb-4360f56d5542\") " pod="openshift-marketplace/redhat-operators-hnd8d" Oct 05 08:26:36 crc kubenswrapper[4846]: I1005 08:26:36.899895 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hnd8d" Oct 05 08:26:37 crc kubenswrapper[4846]: I1005 08:26:37.348048 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-hnd8d"] Oct 05 08:26:38 crc kubenswrapper[4846]: I1005 08:26:38.125112 4846 generic.go:334] "Generic (PLEG): container finished" podID="bf1f32be-7e26-400a-b8cb-4360f56d5542" containerID="16ae5ad5b874726e59e3c5fb361696064b98c8ca7e865df8c79ce543fe69004c" exitCode=0 Oct 05 08:26:38 crc kubenswrapper[4846]: I1005 08:26:38.125236 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hnd8d" event={"ID":"bf1f32be-7e26-400a-b8cb-4360f56d5542","Type":"ContainerDied","Data":"16ae5ad5b874726e59e3c5fb361696064b98c8ca7e865df8c79ce543fe69004c"} Oct 05 08:26:38 crc kubenswrapper[4846]: I1005 08:26:38.125417 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hnd8d" event={"ID":"bf1f32be-7e26-400a-b8cb-4360f56d5542","Type":"ContainerStarted","Data":"659309a855cab462e5254dcc6bbdc3ff5f6d834fd769ece049bcfc913277e199"} Oct 05 08:26:39 crc kubenswrapper[4846]: I1005 08:26:39.138207 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hnd8d" event={"ID":"bf1f32be-7e26-400a-b8cb-4360f56d5542","Type":"ContainerStarted","Data":"82d7d0ee985def23842bfed6bc57b0784e46730024042a4fff260b26fe1fc8e1"} Oct 05 08:26:40 crc kubenswrapper[4846]: I1005 08:26:40.154889 4846 generic.go:334] "Generic (PLEG): container finished" podID="bf1f32be-7e26-400a-b8cb-4360f56d5542" containerID="82d7d0ee985def23842bfed6bc57b0784e46730024042a4fff260b26fe1fc8e1" exitCode=0 Oct 05 08:26:40 crc kubenswrapper[4846]: I1005 08:26:40.155015 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hnd8d" event={"ID":"bf1f32be-7e26-400a-b8cb-4360f56d5542","Type":"ContainerDied","Data":"82d7d0ee985def23842bfed6bc57b0784e46730024042a4fff260b26fe1fc8e1"} Oct 05 08:26:41 crc kubenswrapper[4846]: I1005 08:26:41.171852 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hnd8d" event={"ID":"bf1f32be-7e26-400a-b8cb-4360f56d5542","Type":"ContainerStarted","Data":"1a3c276c30a2634c1ddb8c7293c2b8de58c7e4d4a815f0a0427b2e1617cab7bd"} Oct 05 08:26:41 crc kubenswrapper[4846]: I1005 08:26:41.199498 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-hnd8d" podStartSLOduration=2.634059184 podStartE2EDuration="5.199464489s" podCreationTimestamp="2025-10-05 08:26:36 +0000 UTC" firstStartedPulling="2025-10-05 08:26:38.127696048 +0000 UTC m=+5920.368548833" lastFinishedPulling="2025-10-05 08:26:40.693101323 +0000 UTC m=+5922.933954138" observedRunningTime="2025-10-05 08:26:41.198815032 +0000 UTC m=+5923.439667867" watchObservedRunningTime="2025-10-05 08:26:41.199464489 +0000 UTC m=+5923.440317304" Oct 05 08:26:45 crc kubenswrapper[4846]: I1005 08:26:45.498140 4846 scope.go:117] "RemoveContainer" containerID="fdafa27da24296f9087c33b6b12a6598d2a8d61fc985b06ad680401828c9feb6" Oct 05 08:26:45 crc kubenswrapper[4846]: E1005 08:26:45.498764 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 08:26:46 crc kubenswrapper[4846]: I1005 08:26:46.195616 4846 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-d98bd9cc6-4s784" podUID="c8df1242-e04a-450b-a0b4-9d408f4b3862" containerName="horizon" probeResult="failure" output="Get \"https://10.217.1.102:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.102:8443: connect: connection refused" Oct 05 08:26:46 crc kubenswrapper[4846]: I1005 08:26:46.901117 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-hnd8d" Oct 05 08:26:46 crc kubenswrapper[4846]: I1005 08:26:46.901482 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-hnd8d" Oct 05 08:26:46 crc kubenswrapper[4846]: I1005 08:26:46.960730 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-hnd8d" Oct 05 08:26:47 crc kubenswrapper[4846]: I1005 08:26:47.315311 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-hnd8d" Oct 05 08:26:47 crc kubenswrapper[4846]: I1005 08:26:47.386798 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-hnd8d"] Oct 05 08:26:49 crc kubenswrapper[4846]: I1005 08:26:49.258292 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-hnd8d" podUID="bf1f32be-7e26-400a-b8cb-4360f56d5542" containerName="registry-server" containerID="cri-o://1a3c276c30a2634c1ddb8c7293c2b8de58c7e4d4a815f0a0427b2e1617cab7bd" gracePeriod=2 Oct 05 08:26:49 crc kubenswrapper[4846]: I1005 08:26:49.764746 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hnd8d" Oct 05 08:26:49 crc kubenswrapper[4846]: I1005 08:26:49.962604 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bf1f32be-7e26-400a-b8cb-4360f56d5542-utilities\") pod \"bf1f32be-7e26-400a-b8cb-4360f56d5542\" (UID: \"bf1f32be-7e26-400a-b8cb-4360f56d5542\") " Oct 05 08:26:49 crc kubenswrapper[4846]: I1005 08:26:49.962961 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tlt56\" (UniqueName: \"kubernetes.io/projected/bf1f32be-7e26-400a-b8cb-4360f56d5542-kube-api-access-tlt56\") pod \"bf1f32be-7e26-400a-b8cb-4360f56d5542\" (UID: \"bf1f32be-7e26-400a-b8cb-4360f56d5542\") " Oct 05 08:26:49 crc kubenswrapper[4846]: I1005 08:26:49.963060 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bf1f32be-7e26-400a-b8cb-4360f56d5542-catalog-content\") pod \"bf1f32be-7e26-400a-b8cb-4360f56d5542\" (UID: \"bf1f32be-7e26-400a-b8cb-4360f56d5542\") " Oct 05 08:26:49 crc kubenswrapper[4846]: I1005 08:26:49.965296 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bf1f32be-7e26-400a-b8cb-4360f56d5542-utilities" (OuterVolumeSpecName: "utilities") pod "bf1f32be-7e26-400a-b8cb-4360f56d5542" (UID: "bf1f32be-7e26-400a-b8cb-4360f56d5542"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 08:26:49 crc kubenswrapper[4846]: I1005 08:26:49.972532 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf1f32be-7e26-400a-b8cb-4360f56d5542-kube-api-access-tlt56" (OuterVolumeSpecName: "kube-api-access-tlt56") pod "bf1f32be-7e26-400a-b8cb-4360f56d5542" (UID: "bf1f32be-7e26-400a-b8cb-4360f56d5542"). InnerVolumeSpecName "kube-api-access-tlt56". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:26:50 crc kubenswrapper[4846]: I1005 08:26:50.065716 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tlt56\" (UniqueName: \"kubernetes.io/projected/bf1f32be-7e26-400a-b8cb-4360f56d5542-kube-api-access-tlt56\") on node \"crc\" DevicePath \"\"" Oct 05 08:26:50 crc kubenswrapper[4846]: I1005 08:26:50.065768 4846 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bf1f32be-7e26-400a-b8cb-4360f56d5542-utilities\") on node \"crc\" DevicePath \"\"" Oct 05 08:26:50 crc kubenswrapper[4846]: I1005 08:26:50.272355 4846 generic.go:334] "Generic (PLEG): container finished" podID="bf1f32be-7e26-400a-b8cb-4360f56d5542" containerID="1a3c276c30a2634c1ddb8c7293c2b8de58c7e4d4a815f0a0427b2e1617cab7bd" exitCode=0 Oct 05 08:26:50 crc kubenswrapper[4846]: I1005 08:26:50.272427 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hnd8d" event={"ID":"bf1f32be-7e26-400a-b8cb-4360f56d5542","Type":"ContainerDied","Data":"1a3c276c30a2634c1ddb8c7293c2b8de58c7e4d4a815f0a0427b2e1617cab7bd"} Oct 05 08:26:50 crc kubenswrapper[4846]: I1005 08:26:50.272518 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hnd8d" event={"ID":"bf1f32be-7e26-400a-b8cb-4360f56d5542","Type":"ContainerDied","Data":"659309a855cab462e5254dcc6bbdc3ff5f6d834fd769ece049bcfc913277e199"} Oct 05 08:26:50 crc kubenswrapper[4846]: I1005 08:26:50.272554 4846 scope.go:117] "RemoveContainer" containerID="1a3c276c30a2634c1ddb8c7293c2b8de58c7e4d4a815f0a0427b2e1617cab7bd" Oct 05 08:26:50 crc kubenswrapper[4846]: I1005 08:26:50.272454 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hnd8d" Oct 05 08:26:50 crc kubenswrapper[4846]: I1005 08:26:50.313198 4846 scope.go:117] "RemoveContainer" containerID="82d7d0ee985def23842bfed6bc57b0784e46730024042a4fff260b26fe1fc8e1" Oct 05 08:26:50 crc kubenswrapper[4846]: I1005 08:26:50.348804 4846 scope.go:117] "RemoveContainer" containerID="16ae5ad5b874726e59e3c5fb361696064b98c8ca7e865df8c79ce543fe69004c" Oct 05 08:26:50 crc kubenswrapper[4846]: I1005 08:26:50.416615 4846 scope.go:117] "RemoveContainer" containerID="1a3c276c30a2634c1ddb8c7293c2b8de58c7e4d4a815f0a0427b2e1617cab7bd" Oct 05 08:26:50 crc kubenswrapper[4846]: E1005 08:26:50.417160 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1a3c276c30a2634c1ddb8c7293c2b8de58c7e4d4a815f0a0427b2e1617cab7bd\": container with ID starting with 1a3c276c30a2634c1ddb8c7293c2b8de58c7e4d4a815f0a0427b2e1617cab7bd not found: ID does not exist" containerID="1a3c276c30a2634c1ddb8c7293c2b8de58c7e4d4a815f0a0427b2e1617cab7bd" Oct 05 08:26:50 crc kubenswrapper[4846]: I1005 08:26:50.417215 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1a3c276c30a2634c1ddb8c7293c2b8de58c7e4d4a815f0a0427b2e1617cab7bd"} err="failed to get container status \"1a3c276c30a2634c1ddb8c7293c2b8de58c7e4d4a815f0a0427b2e1617cab7bd\": rpc error: code = NotFound desc = could not find container \"1a3c276c30a2634c1ddb8c7293c2b8de58c7e4d4a815f0a0427b2e1617cab7bd\": container with ID starting with 1a3c276c30a2634c1ddb8c7293c2b8de58c7e4d4a815f0a0427b2e1617cab7bd not found: ID does not exist" Oct 05 08:26:50 crc kubenswrapper[4846]: I1005 08:26:50.417243 4846 scope.go:117] "RemoveContainer" containerID="82d7d0ee985def23842bfed6bc57b0784e46730024042a4fff260b26fe1fc8e1" Oct 05 08:26:50 crc kubenswrapper[4846]: E1005 08:26:50.417688 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"82d7d0ee985def23842bfed6bc57b0784e46730024042a4fff260b26fe1fc8e1\": container with ID starting with 82d7d0ee985def23842bfed6bc57b0784e46730024042a4fff260b26fe1fc8e1 not found: ID does not exist" containerID="82d7d0ee985def23842bfed6bc57b0784e46730024042a4fff260b26fe1fc8e1" Oct 05 08:26:50 crc kubenswrapper[4846]: I1005 08:26:50.417727 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"82d7d0ee985def23842bfed6bc57b0784e46730024042a4fff260b26fe1fc8e1"} err="failed to get container status \"82d7d0ee985def23842bfed6bc57b0784e46730024042a4fff260b26fe1fc8e1\": rpc error: code = NotFound desc = could not find container \"82d7d0ee985def23842bfed6bc57b0784e46730024042a4fff260b26fe1fc8e1\": container with ID starting with 82d7d0ee985def23842bfed6bc57b0784e46730024042a4fff260b26fe1fc8e1 not found: ID does not exist" Oct 05 08:26:50 crc kubenswrapper[4846]: I1005 08:26:50.417757 4846 scope.go:117] "RemoveContainer" containerID="16ae5ad5b874726e59e3c5fb361696064b98c8ca7e865df8c79ce543fe69004c" Oct 05 08:26:50 crc kubenswrapper[4846]: E1005 08:26:50.421405 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"16ae5ad5b874726e59e3c5fb361696064b98c8ca7e865df8c79ce543fe69004c\": container with ID starting with 16ae5ad5b874726e59e3c5fb361696064b98c8ca7e865df8c79ce543fe69004c not found: ID does not exist" containerID="16ae5ad5b874726e59e3c5fb361696064b98c8ca7e865df8c79ce543fe69004c" Oct 05 08:26:50 crc kubenswrapper[4846]: I1005 08:26:50.421450 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"16ae5ad5b874726e59e3c5fb361696064b98c8ca7e865df8c79ce543fe69004c"} err="failed to get container status \"16ae5ad5b874726e59e3c5fb361696064b98c8ca7e865df8c79ce543fe69004c\": rpc error: code = NotFound desc = could not find container \"16ae5ad5b874726e59e3c5fb361696064b98c8ca7e865df8c79ce543fe69004c\": container with ID starting with 16ae5ad5b874726e59e3c5fb361696064b98c8ca7e865df8c79ce543fe69004c not found: ID does not exist" Oct 05 08:26:50 crc kubenswrapper[4846]: I1005 08:26:50.938935 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bf1f32be-7e26-400a-b8cb-4360f56d5542-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "bf1f32be-7e26-400a-b8cb-4360f56d5542" (UID: "bf1f32be-7e26-400a-b8cb-4360f56d5542"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 08:26:50 crc kubenswrapper[4846]: I1005 08:26:50.984133 4846 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bf1f32be-7e26-400a-b8cb-4360f56d5542-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 05 08:26:51 crc kubenswrapper[4846]: I1005 08:26:51.234500 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-hnd8d"] Oct 05 08:26:51 crc kubenswrapper[4846]: I1005 08:26:51.253822 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-hnd8d"] Oct 05 08:26:52 crc kubenswrapper[4846]: I1005 08:26:52.514798 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf1f32be-7e26-400a-b8cb-4360f56d5542" path="/var/lib/kubelet/pods/bf1f32be-7e26-400a-b8cb-4360f56d5542/volumes" Oct 05 08:26:56 crc kubenswrapper[4846]: I1005 08:26:56.195303 4846 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-d98bd9cc6-4s784" podUID="c8df1242-e04a-450b-a0b4-9d408f4b3862" containerName="horizon" probeResult="failure" output="Get \"https://10.217.1.102:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.102:8443: connect: connection refused" Oct 05 08:26:56 crc kubenswrapper[4846]: I1005 08:26:56.195965 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-d98bd9cc6-4s784" Oct 05 08:26:56 crc kubenswrapper[4846]: I1005 08:26:56.498403 4846 scope.go:117] "RemoveContainer" containerID="fdafa27da24296f9087c33b6b12a6598d2a8d61fc985b06ad680401828c9feb6" Oct 05 08:26:57 crc kubenswrapper[4846]: I1005 08:26:57.379067 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" event={"ID":"ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7","Type":"ContainerStarted","Data":"5901a64fb8754c398c6b2b246ea8952e5e7f49cedcb4e34d861ec3f14a1f9619"} Oct 05 08:27:00 crc kubenswrapper[4846]: I1005 08:27:00.408573 4846 generic.go:334] "Generic (PLEG): container finished" podID="c8df1242-e04a-450b-a0b4-9d408f4b3862" containerID="fa6050e2b9830faaa192d66869bdf6af4ade7688b8f5126c21c8162a7872705e" exitCode=137 Oct 05 08:27:00 crc kubenswrapper[4846]: I1005 08:27:00.409247 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-d98bd9cc6-4s784" event={"ID":"c8df1242-e04a-450b-a0b4-9d408f4b3862","Type":"ContainerDied","Data":"fa6050e2b9830faaa192d66869bdf6af4ade7688b8f5126c21c8162a7872705e"} Oct 05 08:27:00 crc kubenswrapper[4846]: I1005 08:27:00.409288 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-d98bd9cc6-4s784" event={"ID":"c8df1242-e04a-450b-a0b4-9d408f4b3862","Type":"ContainerDied","Data":"788762c5cb0dbb71578577fb3fdf613e547c15cd66bf4837d8e776c8bc4c02e7"} Oct 05 08:27:00 crc kubenswrapper[4846]: I1005 08:27:00.409303 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="788762c5cb0dbb71578577fb3fdf613e547c15cd66bf4837d8e776c8bc4c02e7" Oct 05 08:27:00 crc kubenswrapper[4846]: I1005 08:27:00.434297 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-d98bd9cc6-4s784" Oct 05 08:27:00 crc kubenswrapper[4846]: I1005 08:27:00.553066 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p4xpr\" (UniqueName: \"kubernetes.io/projected/c8df1242-e04a-450b-a0b4-9d408f4b3862-kube-api-access-p4xpr\") pod \"c8df1242-e04a-450b-a0b4-9d408f4b3862\" (UID: \"c8df1242-e04a-450b-a0b4-9d408f4b3862\") " Oct 05 08:27:00 crc kubenswrapper[4846]: I1005 08:27:00.553136 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c8df1242-e04a-450b-a0b4-9d408f4b3862-combined-ca-bundle\") pod \"c8df1242-e04a-450b-a0b4-9d408f4b3862\" (UID: \"c8df1242-e04a-450b-a0b4-9d408f4b3862\") " Oct 05 08:27:00 crc kubenswrapper[4846]: I1005 08:27:00.553300 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/c8df1242-e04a-450b-a0b4-9d408f4b3862-horizon-secret-key\") pod \"c8df1242-e04a-450b-a0b4-9d408f4b3862\" (UID: \"c8df1242-e04a-450b-a0b4-9d408f4b3862\") " Oct 05 08:27:00 crc kubenswrapper[4846]: I1005 08:27:00.553350 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c8df1242-e04a-450b-a0b4-9d408f4b3862-config-data\") pod \"c8df1242-e04a-450b-a0b4-9d408f4b3862\" (UID: \"c8df1242-e04a-450b-a0b4-9d408f4b3862\") " Oct 05 08:27:00 crc kubenswrapper[4846]: I1005 08:27:00.553387 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c8df1242-e04a-450b-a0b4-9d408f4b3862-logs\") pod \"c8df1242-e04a-450b-a0b4-9d408f4b3862\" (UID: \"c8df1242-e04a-450b-a0b4-9d408f4b3862\") " Oct 05 08:27:00 crc kubenswrapper[4846]: I1005 08:27:00.553433 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c8df1242-e04a-450b-a0b4-9d408f4b3862-scripts\") pod \"c8df1242-e04a-450b-a0b4-9d408f4b3862\" (UID: \"c8df1242-e04a-450b-a0b4-9d408f4b3862\") " Oct 05 08:27:00 crc kubenswrapper[4846]: I1005 08:27:00.553507 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/c8df1242-e04a-450b-a0b4-9d408f4b3862-horizon-tls-certs\") pod \"c8df1242-e04a-450b-a0b4-9d408f4b3862\" (UID: \"c8df1242-e04a-450b-a0b4-9d408f4b3862\") " Oct 05 08:27:00 crc kubenswrapper[4846]: I1005 08:27:00.554022 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c8df1242-e04a-450b-a0b4-9d408f4b3862-logs" (OuterVolumeSpecName: "logs") pod "c8df1242-e04a-450b-a0b4-9d408f4b3862" (UID: "c8df1242-e04a-450b-a0b4-9d408f4b3862"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 08:27:00 crc kubenswrapper[4846]: I1005 08:27:00.568995 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c8df1242-e04a-450b-a0b4-9d408f4b3862-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "c8df1242-e04a-450b-a0b4-9d408f4b3862" (UID: "c8df1242-e04a-450b-a0b4-9d408f4b3862"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:27:00 crc kubenswrapper[4846]: I1005 08:27:00.571501 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c8df1242-e04a-450b-a0b4-9d408f4b3862-kube-api-access-p4xpr" (OuterVolumeSpecName: "kube-api-access-p4xpr") pod "c8df1242-e04a-450b-a0b4-9d408f4b3862" (UID: "c8df1242-e04a-450b-a0b4-9d408f4b3862"). InnerVolumeSpecName "kube-api-access-p4xpr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:27:00 crc kubenswrapper[4846]: I1005 08:27:00.587704 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c8df1242-e04a-450b-a0b4-9d408f4b3862-config-data" (OuterVolumeSpecName: "config-data") pod "c8df1242-e04a-450b-a0b4-9d408f4b3862" (UID: "c8df1242-e04a-450b-a0b4-9d408f4b3862"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 08:27:00 crc kubenswrapper[4846]: I1005 08:27:00.592974 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c8df1242-e04a-450b-a0b4-9d408f4b3862-scripts" (OuterVolumeSpecName: "scripts") pod "c8df1242-e04a-450b-a0b4-9d408f4b3862" (UID: "c8df1242-e04a-450b-a0b4-9d408f4b3862"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 08:27:00 crc kubenswrapper[4846]: I1005 08:27:00.600758 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c8df1242-e04a-450b-a0b4-9d408f4b3862-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c8df1242-e04a-450b-a0b4-9d408f4b3862" (UID: "c8df1242-e04a-450b-a0b4-9d408f4b3862"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:27:00 crc kubenswrapper[4846]: I1005 08:27:00.616530 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c8df1242-e04a-450b-a0b4-9d408f4b3862-horizon-tls-certs" (OuterVolumeSpecName: "horizon-tls-certs") pod "c8df1242-e04a-450b-a0b4-9d408f4b3862" (UID: "c8df1242-e04a-450b-a0b4-9d408f4b3862"). InnerVolumeSpecName "horizon-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:27:00 crc kubenswrapper[4846]: I1005 08:27:00.656246 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p4xpr\" (UniqueName: \"kubernetes.io/projected/c8df1242-e04a-450b-a0b4-9d408f4b3862-kube-api-access-p4xpr\") on node \"crc\" DevicePath \"\"" Oct 05 08:27:00 crc kubenswrapper[4846]: I1005 08:27:00.656575 4846 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c8df1242-e04a-450b-a0b4-9d408f4b3862-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 08:27:00 crc kubenswrapper[4846]: I1005 08:27:00.658101 4846 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/c8df1242-e04a-450b-a0b4-9d408f4b3862-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Oct 05 08:27:00 crc kubenswrapper[4846]: I1005 08:27:00.658126 4846 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c8df1242-e04a-450b-a0b4-9d408f4b3862-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 08:27:00 crc kubenswrapper[4846]: I1005 08:27:00.658138 4846 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c8df1242-e04a-450b-a0b4-9d408f4b3862-logs\") on node \"crc\" DevicePath \"\"" Oct 05 08:27:00 crc kubenswrapper[4846]: I1005 08:27:00.658151 4846 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c8df1242-e04a-450b-a0b4-9d408f4b3862-scripts\") on node \"crc\" DevicePath \"\"" Oct 05 08:27:00 crc kubenswrapper[4846]: I1005 08:27:00.658161 4846 reconciler_common.go:293] "Volume detached for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/c8df1242-e04a-450b-a0b4-9d408f4b3862-horizon-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 05 08:27:01 crc kubenswrapper[4846]: I1005 08:27:01.419020 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-d98bd9cc6-4s784" Oct 05 08:27:01 crc kubenswrapper[4846]: I1005 08:27:01.474724 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-d98bd9cc6-4s784"] Oct 05 08:27:01 crc kubenswrapper[4846]: I1005 08:27:01.482962 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-d98bd9cc6-4s784"] Oct 05 08:27:02 crc kubenswrapper[4846]: I1005 08:27:02.521590 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c8df1242-e04a-450b-a0b4-9d408f4b3862" path="/var/lib/kubelet/pods/c8df1242-e04a-450b-a0b4-9d408f4b3862/volumes" Oct 05 08:27:06 crc kubenswrapper[4846]: I1005 08:27:06.052369 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-create-p9zcj"] Oct 05 08:27:06 crc kubenswrapper[4846]: I1005 08:27:06.088202 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-create-p9zcj"] Oct 05 08:27:06 crc kubenswrapper[4846]: I1005 08:27:06.511849 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0cc5829c-be80-4bf4-adec-b18b8dd4aca5" path="/var/lib/kubelet/pods/0cc5829c-be80-4bf4-adec-b18b8dd4aca5/volumes" Oct 05 08:27:16 crc kubenswrapper[4846]: I1005 08:27:16.050521 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-e233-account-create-h5tj4"] Oct 05 08:27:16 crc kubenswrapper[4846]: I1005 08:27:16.062472 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-e233-account-create-h5tj4"] Oct 05 08:27:16 crc kubenswrapper[4846]: I1005 08:27:16.516640 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9af60fac-0c55-45b4-9752-955f8df372d5" path="/var/lib/kubelet/pods/9af60fac-0c55-45b4-9752-955f8df372d5/volumes" Oct 05 08:27:23 crc kubenswrapper[4846]: I1005 08:27:23.111042 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-47zxt"] Oct 05 08:27:23 crc kubenswrapper[4846]: E1005 08:27:23.113172 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bf1f32be-7e26-400a-b8cb-4360f56d5542" containerName="registry-server" Oct 05 08:27:23 crc kubenswrapper[4846]: I1005 08:27:23.113232 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="bf1f32be-7e26-400a-b8cb-4360f56d5542" containerName="registry-server" Oct 05 08:27:23 crc kubenswrapper[4846]: E1005 08:27:23.113284 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c8df1242-e04a-450b-a0b4-9d408f4b3862" containerName="horizon-log" Oct 05 08:27:23 crc kubenswrapper[4846]: I1005 08:27:23.113297 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="c8df1242-e04a-450b-a0b4-9d408f4b3862" containerName="horizon-log" Oct 05 08:27:23 crc kubenswrapper[4846]: E1005 08:27:23.113328 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bf1f32be-7e26-400a-b8cb-4360f56d5542" containerName="extract-content" Oct 05 08:27:23 crc kubenswrapper[4846]: I1005 08:27:23.113343 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="bf1f32be-7e26-400a-b8cb-4360f56d5542" containerName="extract-content" Oct 05 08:27:23 crc kubenswrapper[4846]: E1005 08:27:23.113368 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c8df1242-e04a-450b-a0b4-9d408f4b3862" containerName="horizon" Oct 05 08:27:23 crc kubenswrapper[4846]: I1005 08:27:23.113382 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="c8df1242-e04a-450b-a0b4-9d408f4b3862" containerName="horizon" Oct 05 08:27:23 crc kubenswrapper[4846]: E1005 08:27:23.115669 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bf1f32be-7e26-400a-b8cb-4360f56d5542" containerName="extract-utilities" Oct 05 08:27:23 crc kubenswrapper[4846]: I1005 08:27:23.115693 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="bf1f32be-7e26-400a-b8cb-4360f56d5542" containerName="extract-utilities" Oct 05 08:27:23 crc kubenswrapper[4846]: I1005 08:27:23.116229 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="c8df1242-e04a-450b-a0b4-9d408f4b3862" containerName="horizon-log" Oct 05 08:27:23 crc kubenswrapper[4846]: I1005 08:27:23.116293 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="bf1f32be-7e26-400a-b8cb-4360f56d5542" containerName="registry-server" Oct 05 08:27:23 crc kubenswrapper[4846]: I1005 08:27:23.116315 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="c8df1242-e04a-450b-a0b4-9d408f4b3862" containerName="horizon" Oct 05 08:27:23 crc kubenswrapper[4846]: I1005 08:27:23.119595 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-47zxt" Oct 05 08:27:23 crc kubenswrapper[4846]: I1005 08:27:23.135519 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-47zxt"] Oct 05 08:27:23 crc kubenswrapper[4846]: I1005 08:27:23.168243 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-srfh2\" (UniqueName: \"kubernetes.io/projected/b978eedb-81df-4394-a22f-2c9fb7981791-kube-api-access-srfh2\") pod \"redhat-marketplace-47zxt\" (UID: \"b978eedb-81df-4394-a22f-2c9fb7981791\") " pod="openshift-marketplace/redhat-marketplace-47zxt" Oct 05 08:27:23 crc kubenswrapper[4846]: I1005 08:27:23.168743 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b978eedb-81df-4394-a22f-2c9fb7981791-catalog-content\") pod \"redhat-marketplace-47zxt\" (UID: \"b978eedb-81df-4394-a22f-2c9fb7981791\") " pod="openshift-marketplace/redhat-marketplace-47zxt" Oct 05 08:27:23 crc kubenswrapper[4846]: I1005 08:27:23.169049 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b978eedb-81df-4394-a22f-2c9fb7981791-utilities\") pod \"redhat-marketplace-47zxt\" (UID: \"b978eedb-81df-4394-a22f-2c9fb7981791\") " pod="openshift-marketplace/redhat-marketplace-47zxt" Oct 05 08:27:23 crc kubenswrapper[4846]: I1005 08:27:23.270957 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b978eedb-81df-4394-a22f-2c9fb7981791-catalog-content\") pod \"redhat-marketplace-47zxt\" (UID: \"b978eedb-81df-4394-a22f-2c9fb7981791\") " pod="openshift-marketplace/redhat-marketplace-47zxt" Oct 05 08:27:23 crc kubenswrapper[4846]: I1005 08:27:23.271385 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b978eedb-81df-4394-a22f-2c9fb7981791-utilities\") pod \"redhat-marketplace-47zxt\" (UID: \"b978eedb-81df-4394-a22f-2c9fb7981791\") " pod="openshift-marketplace/redhat-marketplace-47zxt" Oct 05 08:27:23 crc kubenswrapper[4846]: I1005 08:27:23.271449 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-srfh2\" (UniqueName: \"kubernetes.io/projected/b978eedb-81df-4394-a22f-2c9fb7981791-kube-api-access-srfh2\") pod \"redhat-marketplace-47zxt\" (UID: \"b978eedb-81df-4394-a22f-2c9fb7981791\") " pod="openshift-marketplace/redhat-marketplace-47zxt" Oct 05 08:27:23 crc kubenswrapper[4846]: I1005 08:27:23.271628 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b978eedb-81df-4394-a22f-2c9fb7981791-catalog-content\") pod \"redhat-marketplace-47zxt\" (UID: \"b978eedb-81df-4394-a22f-2c9fb7981791\") " pod="openshift-marketplace/redhat-marketplace-47zxt" Oct 05 08:27:23 crc kubenswrapper[4846]: I1005 08:27:23.271912 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b978eedb-81df-4394-a22f-2c9fb7981791-utilities\") pod \"redhat-marketplace-47zxt\" (UID: \"b978eedb-81df-4394-a22f-2c9fb7981791\") " pod="openshift-marketplace/redhat-marketplace-47zxt" Oct 05 08:27:23 crc kubenswrapper[4846]: I1005 08:27:23.289635 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-srfh2\" (UniqueName: \"kubernetes.io/projected/b978eedb-81df-4394-a22f-2c9fb7981791-kube-api-access-srfh2\") pod \"redhat-marketplace-47zxt\" (UID: \"b978eedb-81df-4394-a22f-2c9fb7981791\") " pod="openshift-marketplace/redhat-marketplace-47zxt" Oct 05 08:27:23 crc kubenswrapper[4846]: I1005 08:27:23.470939 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-47zxt" Oct 05 08:27:23 crc kubenswrapper[4846]: I1005 08:27:23.964878 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-47zxt"] Oct 05 08:27:23 crc kubenswrapper[4846]: W1005 08:27:23.970372 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb978eedb_81df_4394_a22f_2c9fb7981791.slice/crio-57f4b46f0647cbdc6880162b8d6f58f87914a38ffbfad898ea1fe557b80323c4 WatchSource:0}: Error finding container 57f4b46f0647cbdc6880162b8d6f58f87914a38ffbfad898ea1fe557b80323c4: Status 404 returned error can't find the container with id 57f4b46f0647cbdc6880162b8d6f58f87914a38ffbfad898ea1fe557b80323c4 Oct 05 08:27:24 crc kubenswrapper[4846]: I1005 08:27:24.703715 4846 generic.go:334] "Generic (PLEG): container finished" podID="b978eedb-81df-4394-a22f-2c9fb7981791" containerID="3cbe3ecc81ad4981a422438847fac4d03bbddb0d3395f191b12bc68c3d326ca6" exitCode=0 Oct 05 08:27:24 crc kubenswrapper[4846]: I1005 08:27:24.703982 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-47zxt" event={"ID":"b978eedb-81df-4394-a22f-2c9fb7981791","Type":"ContainerDied","Data":"3cbe3ecc81ad4981a422438847fac4d03bbddb0d3395f191b12bc68c3d326ca6"} Oct 05 08:27:24 crc kubenswrapper[4846]: I1005 08:27:24.704271 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-47zxt" event={"ID":"b978eedb-81df-4394-a22f-2c9fb7981791","Type":"ContainerStarted","Data":"57f4b46f0647cbdc6880162b8d6f58f87914a38ffbfad898ea1fe557b80323c4"} Oct 05 08:27:25 crc kubenswrapper[4846]: I1005 08:27:25.054958 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-sync-42dlq"] Oct 05 08:27:25 crc kubenswrapper[4846]: I1005 08:27:25.064484 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-sync-42dlq"] Oct 05 08:27:25 crc kubenswrapper[4846]: I1005 08:27:25.716628 4846 generic.go:334] "Generic (PLEG): container finished" podID="b978eedb-81df-4394-a22f-2c9fb7981791" containerID="e8e633461859c97dcbab0582a35632f1a349a5e289fc1adc43c45c28faebc1ff" exitCode=0 Oct 05 08:27:25 crc kubenswrapper[4846]: I1005 08:27:25.716715 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-47zxt" event={"ID":"b978eedb-81df-4394-a22f-2c9fb7981791","Type":"ContainerDied","Data":"e8e633461859c97dcbab0582a35632f1a349a5e289fc1adc43c45c28faebc1ff"} Oct 05 08:27:26 crc kubenswrapper[4846]: I1005 08:27:26.517100 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e2077da3-902b-48e2-93ef-be196e85b88d" path="/var/lib/kubelet/pods/e2077da3-902b-48e2-93ef-be196e85b88d/volumes" Oct 05 08:27:26 crc kubenswrapper[4846]: I1005 08:27:26.740005 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-47zxt" event={"ID":"b978eedb-81df-4394-a22f-2c9fb7981791","Type":"ContainerStarted","Data":"169729410d49818f7ed40e4e7a1637b7ef9098728b8544ae83f8b1a66751771e"} Oct 05 08:27:26 crc kubenswrapper[4846]: I1005 08:27:26.765939 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-47zxt" podStartSLOduration=2.304806266 podStartE2EDuration="3.765918995s" podCreationTimestamp="2025-10-05 08:27:23 +0000 UTC" firstStartedPulling="2025-10-05 08:27:24.70621655 +0000 UTC m=+5966.947069365" lastFinishedPulling="2025-10-05 08:27:26.167329279 +0000 UTC m=+5968.408182094" observedRunningTime="2025-10-05 08:27:26.763933322 +0000 UTC m=+5969.004786137" watchObservedRunningTime="2025-10-05 08:27:26.765918995 +0000 UTC m=+5969.006771770" Oct 05 08:27:33 crc kubenswrapper[4846]: I1005 08:27:33.471202 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-47zxt" Oct 05 08:27:33 crc kubenswrapper[4846]: I1005 08:27:33.471709 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-47zxt" Oct 05 08:27:33 crc kubenswrapper[4846]: I1005 08:27:33.501872 4846 scope.go:117] "RemoveContainer" containerID="21f9d172a502fb5a7fb911386591365ff76f834fae51cb1409391d9b9c1ced56" Oct 05 08:27:33 crc kubenswrapper[4846]: I1005 08:27:33.527630 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-47zxt" Oct 05 08:27:33 crc kubenswrapper[4846]: I1005 08:27:33.563947 4846 scope.go:117] "RemoveContainer" containerID="dfa0b5dfa965bb1c9af8af1b32b0e6a7337a6350d305c54ed95ea8a8ad9ee9c6" Oct 05 08:27:33 crc kubenswrapper[4846]: I1005 08:27:33.592522 4846 scope.go:117] "RemoveContainer" containerID="16eb4369807860d0dae16b3f48adac19c01ac1eea08e5434c521d62557752894" Oct 05 08:27:33 crc kubenswrapper[4846]: I1005 08:27:33.870799 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-47zxt" Oct 05 08:27:33 crc kubenswrapper[4846]: I1005 08:27:33.935826 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-47zxt"] Oct 05 08:27:35 crc kubenswrapper[4846]: I1005 08:27:35.385363 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-85d5597856-frbmx"] Oct 05 08:27:35 crc kubenswrapper[4846]: I1005 08:27:35.387098 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-85d5597856-frbmx" Oct 05 08:27:35 crc kubenswrapper[4846]: I1005 08:27:35.402005 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-85d5597856-frbmx"] Oct 05 08:27:35 crc kubenswrapper[4846]: I1005 08:27:35.541949 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/82fe0928-8008-48ce-8993-41ff90d89bcd-combined-ca-bundle\") pod \"horizon-85d5597856-frbmx\" (UID: \"82fe0928-8008-48ce-8993-41ff90d89bcd\") " pod="openstack/horizon-85d5597856-frbmx" Oct 05 08:27:35 crc kubenswrapper[4846]: I1005 08:27:35.542031 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/82fe0928-8008-48ce-8993-41ff90d89bcd-scripts\") pod \"horizon-85d5597856-frbmx\" (UID: \"82fe0928-8008-48ce-8993-41ff90d89bcd\") " pod="openstack/horizon-85d5597856-frbmx" Oct 05 08:27:35 crc kubenswrapper[4846]: I1005 08:27:35.542126 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/82fe0928-8008-48ce-8993-41ff90d89bcd-config-data\") pod \"horizon-85d5597856-frbmx\" (UID: \"82fe0928-8008-48ce-8993-41ff90d89bcd\") " pod="openstack/horizon-85d5597856-frbmx" Oct 05 08:27:35 crc kubenswrapper[4846]: I1005 08:27:35.542193 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/82fe0928-8008-48ce-8993-41ff90d89bcd-horizon-secret-key\") pod \"horizon-85d5597856-frbmx\" (UID: \"82fe0928-8008-48ce-8993-41ff90d89bcd\") " pod="openstack/horizon-85d5597856-frbmx" Oct 05 08:27:35 crc kubenswrapper[4846]: I1005 08:27:35.542654 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/82fe0928-8008-48ce-8993-41ff90d89bcd-horizon-tls-certs\") pod \"horizon-85d5597856-frbmx\" (UID: \"82fe0928-8008-48ce-8993-41ff90d89bcd\") " pod="openstack/horizon-85d5597856-frbmx" Oct 05 08:27:35 crc kubenswrapper[4846]: I1005 08:27:35.542695 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/82fe0928-8008-48ce-8993-41ff90d89bcd-logs\") pod \"horizon-85d5597856-frbmx\" (UID: \"82fe0928-8008-48ce-8993-41ff90d89bcd\") " pod="openstack/horizon-85d5597856-frbmx" Oct 05 08:27:35 crc kubenswrapper[4846]: I1005 08:27:35.542725 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5nh6l\" (UniqueName: \"kubernetes.io/projected/82fe0928-8008-48ce-8993-41ff90d89bcd-kube-api-access-5nh6l\") pod \"horizon-85d5597856-frbmx\" (UID: \"82fe0928-8008-48ce-8993-41ff90d89bcd\") " pod="openstack/horizon-85d5597856-frbmx" Oct 05 08:27:35 crc kubenswrapper[4846]: I1005 08:27:35.645287 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/82fe0928-8008-48ce-8993-41ff90d89bcd-horizon-tls-certs\") pod \"horizon-85d5597856-frbmx\" (UID: \"82fe0928-8008-48ce-8993-41ff90d89bcd\") " pod="openstack/horizon-85d5597856-frbmx" Oct 05 08:27:35 crc kubenswrapper[4846]: I1005 08:27:35.645339 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/82fe0928-8008-48ce-8993-41ff90d89bcd-logs\") pod \"horizon-85d5597856-frbmx\" (UID: \"82fe0928-8008-48ce-8993-41ff90d89bcd\") " pod="openstack/horizon-85d5597856-frbmx" Oct 05 08:27:35 crc kubenswrapper[4846]: I1005 08:27:35.645402 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5nh6l\" (UniqueName: \"kubernetes.io/projected/82fe0928-8008-48ce-8993-41ff90d89bcd-kube-api-access-5nh6l\") pod \"horizon-85d5597856-frbmx\" (UID: \"82fe0928-8008-48ce-8993-41ff90d89bcd\") " pod="openstack/horizon-85d5597856-frbmx" Oct 05 08:27:35 crc kubenswrapper[4846]: I1005 08:27:35.645486 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/82fe0928-8008-48ce-8993-41ff90d89bcd-combined-ca-bundle\") pod \"horizon-85d5597856-frbmx\" (UID: \"82fe0928-8008-48ce-8993-41ff90d89bcd\") " pod="openstack/horizon-85d5597856-frbmx" Oct 05 08:27:35 crc kubenswrapper[4846]: I1005 08:27:35.645535 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/82fe0928-8008-48ce-8993-41ff90d89bcd-scripts\") pod \"horizon-85d5597856-frbmx\" (UID: \"82fe0928-8008-48ce-8993-41ff90d89bcd\") " pod="openstack/horizon-85d5597856-frbmx" Oct 05 08:27:35 crc kubenswrapper[4846]: I1005 08:27:35.645569 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/82fe0928-8008-48ce-8993-41ff90d89bcd-config-data\") pod \"horizon-85d5597856-frbmx\" (UID: \"82fe0928-8008-48ce-8993-41ff90d89bcd\") " pod="openstack/horizon-85d5597856-frbmx" Oct 05 08:27:35 crc kubenswrapper[4846]: I1005 08:27:35.645632 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/82fe0928-8008-48ce-8993-41ff90d89bcd-horizon-secret-key\") pod \"horizon-85d5597856-frbmx\" (UID: \"82fe0928-8008-48ce-8993-41ff90d89bcd\") " pod="openstack/horizon-85d5597856-frbmx" Oct 05 08:27:35 crc kubenswrapper[4846]: I1005 08:27:35.648286 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/82fe0928-8008-48ce-8993-41ff90d89bcd-logs\") pod \"horizon-85d5597856-frbmx\" (UID: \"82fe0928-8008-48ce-8993-41ff90d89bcd\") " pod="openstack/horizon-85d5597856-frbmx" Oct 05 08:27:35 crc kubenswrapper[4846]: I1005 08:27:35.648724 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/82fe0928-8008-48ce-8993-41ff90d89bcd-scripts\") pod \"horizon-85d5597856-frbmx\" (UID: \"82fe0928-8008-48ce-8993-41ff90d89bcd\") " pod="openstack/horizon-85d5597856-frbmx" Oct 05 08:27:35 crc kubenswrapper[4846]: I1005 08:27:35.648956 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/82fe0928-8008-48ce-8993-41ff90d89bcd-config-data\") pod \"horizon-85d5597856-frbmx\" (UID: \"82fe0928-8008-48ce-8993-41ff90d89bcd\") " pod="openstack/horizon-85d5597856-frbmx" Oct 05 08:27:35 crc kubenswrapper[4846]: I1005 08:27:35.653280 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/82fe0928-8008-48ce-8993-41ff90d89bcd-horizon-secret-key\") pod \"horizon-85d5597856-frbmx\" (UID: \"82fe0928-8008-48ce-8993-41ff90d89bcd\") " pod="openstack/horizon-85d5597856-frbmx" Oct 05 08:27:35 crc kubenswrapper[4846]: I1005 08:27:35.653951 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/82fe0928-8008-48ce-8993-41ff90d89bcd-combined-ca-bundle\") pod \"horizon-85d5597856-frbmx\" (UID: \"82fe0928-8008-48ce-8993-41ff90d89bcd\") " pod="openstack/horizon-85d5597856-frbmx" Oct 05 08:27:35 crc kubenswrapper[4846]: I1005 08:27:35.654734 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/82fe0928-8008-48ce-8993-41ff90d89bcd-horizon-tls-certs\") pod \"horizon-85d5597856-frbmx\" (UID: \"82fe0928-8008-48ce-8993-41ff90d89bcd\") " pod="openstack/horizon-85d5597856-frbmx" Oct 05 08:27:35 crc kubenswrapper[4846]: I1005 08:27:35.672381 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5nh6l\" (UniqueName: \"kubernetes.io/projected/82fe0928-8008-48ce-8993-41ff90d89bcd-kube-api-access-5nh6l\") pod \"horizon-85d5597856-frbmx\" (UID: \"82fe0928-8008-48ce-8993-41ff90d89bcd\") " pod="openstack/horizon-85d5597856-frbmx" Oct 05 08:27:35 crc kubenswrapper[4846]: I1005 08:27:35.729997 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-85d5597856-frbmx" Oct 05 08:27:35 crc kubenswrapper[4846]: I1005 08:27:35.841102 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-47zxt" podUID="b978eedb-81df-4394-a22f-2c9fb7981791" containerName="registry-server" containerID="cri-o://169729410d49818f7ed40e4e7a1637b7ef9098728b8544ae83f8b1a66751771e" gracePeriod=2 Oct 05 08:27:36 crc kubenswrapper[4846]: I1005 08:27:36.190400 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-47zxt" Oct 05 08:27:36 crc kubenswrapper[4846]: I1005 08:27:36.266314 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-srfh2\" (UniqueName: \"kubernetes.io/projected/b978eedb-81df-4394-a22f-2c9fb7981791-kube-api-access-srfh2\") pod \"b978eedb-81df-4394-a22f-2c9fb7981791\" (UID: \"b978eedb-81df-4394-a22f-2c9fb7981791\") " Oct 05 08:27:36 crc kubenswrapper[4846]: I1005 08:27:36.266402 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b978eedb-81df-4394-a22f-2c9fb7981791-utilities\") pod \"b978eedb-81df-4394-a22f-2c9fb7981791\" (UID: \"b978eedb-81df-4394-a22f-2c9fb7981791\") " Oct 05 08:27:36 crc kubenswrapper[4846]: I1005 08:27:36.266461 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b978eedb-81df-4394-a22f-2c9fb7981791-catalog-content\") pod \"b978eedb-81df-4394-a22f-2c9fb7981791\" (UID: \"b978eedb-81df-4394-a22f-2c9fb7981791\") " Oct 05 08:27:36 crc kubenswrapper[4846]: I1005 08:27:36.268952 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b978eedb-81df-4394-a22f-2c9fb7981791-utilities" (OuterVolumeSpecName: "utilities") pod "b978eedb-81df-4394-a22f-2c9fb7981791" (UID: "b978eedb-81df-4394-a22f-2c9fb7981791"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 08:27:36 crc kubenswrapper[4846]: I1005 08:27:36.276387 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-85d5597856-frbmx"] Oct 05 08:27:36 crc kubenswrapper[4846]: I1005 08:27:36.279679 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b978eedb-81df-4394-a22f-2c9fb7981791-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b978eedb-81df-4394-a22f-2c9fb7981791" (UID: "b978eedb-81df-4394-a22f-2c9fb7981791"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 08:27:36 crc kubenswrapper[4846]: I1005 08:27:36.285381 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b978eedb-81df-4394-a22f-2c9fb7981791-kube-api-access-srfh2" (OuterVolumeSpecName: "kube-api-access-srfh2") pod "b978eedb-81df-4394-a22f-2c9fb7981791" (UID: "b978eedb-81df-4394-a22f-2c9fb7981791"). InnerVolumeSpecName "kube-api-access-srfh2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:27:36 crc kubenswrapper[4846]: I1005 08:27:36.369366 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-srfh2\" (UniqueName: \"kubernetes.io/projected/b978eedb-81df-4394-a22f-2c9fb7981791-kube-api-access-srfh2\") on node \"crc\" DevicePath \"\"" Oct 05 08:27:36 crc kubenswrapper[4846]: I1005 08:27:36.369753 4846 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b978eedb-81df-4394-a22f-2c9fb7981791-utilities\") on node \"crc\" DevicePath \"\"" Oct 05 08:27:36 crc kubenswrapper[4846]: I1005 08:27:36.369769 4846 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b978eedb-81df-4394-a22f-2c9fb7981791-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 05 08:27:36 crc kubenswrapper[4846]: I1005 08:27:36.677783 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-db-create-qd22b"] Oct 05 08:27:36 crc kubenswrapper[4846]: E1005 08:27:36.678343 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b978eedb-81df-4394-a22f-2c9fb7981791" containerName="registry-server" Oct 05 08:27:36 crc kubenswrapper[4846]: I1005 08:27:36.678356 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="b978eedb-81df-4394-a22f-2c9fb7981791" containerName="registry-server" Oct 05 08:27:36 crc kubenswrapper[4846]: E1005 08:27:36.678372 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b978eedb-81df-4394-a22f-2c9fb7981791" containerName="extract-content" Oct 05 08:27:36 crc kubenswrapper[4846]: I1005 08:27:36.678379 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="b978eedb-81df-4394-a22f-2c9fb7981791" containerName="extract-content" Oct 05 08:27:36 crc kubenswrapper[4846]: E1005 08:27:36.678399 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b978eedb-81df-4394-a22f-2c9fb7981791" containerName="extract-utilities" Oct 05 08:27:36 crc kubenswrapper[4846]: I1005 08:27:36.678405 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="b978eedb-81df-4394-a22f-2c9fb7981791" containerName="extract-utilities" Oct 05 08:27:36 crc kubenswrapper[4846]: I1005 08:27:36.678561 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="b978eedb-81df-4394-a22f-2c9fb7981791" containerName="registry-server" Oct 05 08:27:36 crc kubenswrapper[4846]: I1005 08:27:36.682219 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-create-qd22b" Oct 05 08:27:36 crc kubenswrapper[4846]: I1005 08:27:36.696689 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-create-qd22b"] Oct 05 08:27:36 crc kubenswrapper[4846]: I1005 08:27:36.781052 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wvnfq\" (UniqueName: \"kubernetes.io/projected/c651e1a1-e229-40f6-8c40-76b40a301d78-kube-api-access-wvnfq\") pod \"heat-db-create-qd22b\" (UID: \"c651e1a1-e229-40f6-8c40-76b40a301d78\") " pod="openstack/heat-db-create-qd22b" Oct 05 08:27:36 crc kubenswrapper[4846]: I1005 08:27:36.849574 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-85d5597856-frbmx" event={"ID":"82fe0928-8008-48ce-8993-41ff90d89bcd","Type":"ContainerStarted","Data":"20f15c90190784e168b9f6b41e0849954e289a0b2037e6c34e139bfc9dfaa56c"} Oct 05 08:27:36 crc kubenswrapper[4846]: I1005 08:27:36.849623 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-85d5597856-frbmx" event={"ID":"82fe0928-8008-48ce-8993-41ff90d89bcd","Type":"ContainerStarted","Data":"c00448c429ff43173b3fb98843c02ad7165cb4d6b77219d2455b6505023706bd"} Oct 05 08:27:36 crc kubenswrapper[4846]: I1005 08:27:36.849636 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-85d5597856-frbmx" event={"ID":"82fe0928-8008-48ce-8993-41ff90d89bcd","Type":"ContainerStarted","Data":"9dce2ab2a0123db8756e9f40d55b9817ba0cad752498f52ca7ccd64b6ed1d5d9"} Oct 05 08:27:36 crc kubenswrapper[4846]: I1005 08:27:36.854030 4846 generic.go:334] "Generic (PLEG): container finished" podID="b978eedb-81df-4394-a22f-2c9fb7981791" containerID="169729410d49818f7ed40e4e7a1637b7ef9098728b8544ae83f8b1a66751771e" exitCode=0 Oct 05 08:27:36 crc kubenswrapper[4846]: I1005 08:27:36.854067 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-47zxt" event={"ID":"b978eedb-81df-4394-a22f-2c9fb7981791","Type":"ContainerDied","Data":"169729410d49818f7ed40e4e7a1637b7ef9098728b8544ae83f8b1a66751771e"} Oct 05 08:27:36 crc kubenswrapper[4846]: I1005 08:27:36.854088 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-47zxt" event={"ID":"b978eedb-81df-4394-a22f-2c9fb7981791","Type":"ContainerDied","Data":"57f4b46f0647cbdc6880162b8d6f58f87914a38ffbfad898ea1fe557b80323c4"} Oct 05 08:27:36 crc kubenswrapper[4846]: I1005 08:27:36.854109 4846 scope.go:117] "RemoveContainer" containerID="169729410d49818f7ed40e4e7a1637b7ef9098728b8544ae83f8b1a66751771e" Oct 05 08:27:36 crc kubenswrapper[4846]: I1005 08:27:36.854257 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-47zxt" Oct 05 08:27:36 crc kubenswrapper[4846]: I1005 08:27:36.884266 4846 scope.go:117] "RemoveContainer" containerID="e8e633461859c97dcbab0582a35632f1a349a5e289fc1adc43c45c28faebc1ff" Oct 05 08:27:36 crc kubenswrapper[4846]: I1005 08:27:36.888526 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-85d5597856-frbmx" podStartSLOduration=1.888503289 podStartE2EDuration="1.888503289s" podCreationTimestamp="2025-10-05 08:27:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 08:27:36.877003454 +0000 UTC m=+5979.117856229" watchObservedRunningTime="2025-10-05 08:27:36.888503289 +0000 UTC m=+5979.129356074" Oct 05 08:27:36 crc kubenswrapper[4846]: I1005 08:27:36.889485 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wvnfq\" (UniqueName: \"kubernetes.io/projected/c651e1a1-e229-40f6-8c40-76b40a301d78-kube-api-access-wvnfq\") pod \"heat-db-create-qd22b\" (UID: \"c651e1a1-e229-40f6-8c40-76b40a301d78\") " pod="openstack/heat-db-create-qd22b" Oct 05 08:27:36 crc kubenswrapper[4846]: I1005 08:27:36.918013 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wvnfq\" (UniqueName: \"kubernetes.io/projected/c651e1a1-e229-40f6-8c40-76b40a301d78-kube-api-access-wvnfq\") pod \"heat-db-create-qd22b\" (UID: \"c651e1a1-e229-40f6-8c40-76b40a301d78\") " pod="openstack/heat-db-create-qd22b" Oct 05 08:27:36 crc kubenswrapper[4846]: I1005 08:27:36.923705 4846 scope.go:117] "RemoveContainer" containerID="3cbe3ecc81ad4981a422438847fac4d03bbddb0d3395f191b12bc68c3d326ca6" Oct 05 08:27:36 crc kubenswrapper[4846]: I1005 08:27:36.932128 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-47zxt"] Oct 05 08:27:36 crc kubenswrapper[4846]: I1005 08:27:36.939567 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-47zxt"] Oct 05 08:27:37 crc kubenswrapper[4846]: I1005 08:27:37.020642 4846 scope.go:117] "RemoveContainer" containerID="169729410d49818f7ed40e4e7a1637b7ef9098728b8544ae83f8b1a66751771e" Oct 05 08:27:37 crc kubenswrapper[4846]: E1005 08:27:37.021170 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"169729410d49818f7ed40e4e7a1637b7ef9098728b8544ae83f8b1a66751771e\": container with ID starting with 169729410d49818f7ed40e4e7a1637b7ef9098728b8544ae83f8b1a66751771e not found: ID does not exist" containerID="169729410d49818f7ed40e4e7a1637b7ef9098728b8544ae83f8b1a66751771e" Oct 05 08:27:37 crc kubenswrapper[4846]: I1005 08:27:37.021496 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"169729410d49818f7ed40e4e7a1637b7ef9098728b8544ae83f8b1a66751771e"} err="failed to get container status \"169729410d49818f7ed40e4e7a1637b7ef9098728b8544ae83f8b1a66751771e\": rpc error: code = NotFound desc = could not find container \"169729410d49818f7ed40e4e7a1637b7ef9098728b8544ae83f8b1a66751771e\": container with ID starting with 169729410d49818f7ed40e4e7a1637b7ef9098728b8544ae83f8b1a66751771e not found: ID does not exist" Oct 05 08:27:37 crc kubenswrapper[4846]: I1005 08:27:37.021676 4846 scope.go:117] "RemoveContainer" containerID="e8e633461859c97dcbab0582a35632f1a349a5e289fc1adc43c45c28faebc1ff" Oct 05 08:27:37 crc kubenswrapper[4846]: E1005 08:27:37.022326 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e8e633461859c97dcbab0582a35632f1a349a5e289fc1adc43c45c28faebc1ff\": container with ID starting with e8e633461859c97dcbab0582a35632f1a349a5e289fc1adc43c45c28faebc1ff not found: ID does not exist" containerID="e8e633461859c97dcbab0582a35632f1a349a5e289fc1adc43c45c28faebc1ff" Oct 05 08:27:37 crc kubenswrapper[4846]: I1005 08:27:37.022500 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e8e633461859c97dcbab0582a35632f1a349a5e289fc1adc43c45c28faebc1ff"} err="failed to get container status \"e8e633461859c97dcbab0582a35632f1a349a5e289fc1adc43c45c28faebc1ff\": rpc error: code = NotFound desc = could not find container \"e8e633461859c97dcbab0582a35632f1a349a5e289fc1adc43c45c28faebc1ff\": container with ID starting with e8e633461859c97dcbab0582a35632f1a349a5e289fc1adc43c45c28faebc1ff not found: ID does not exist" Oct 05 08:27:37 crc kubenswrapper[4846]: I1005 08:27:37.022670 4846 scope.go:117] "RemoveContainer" containerID="3cbe3ecc81ad4981a422438847fac4d03bbddb0d3395f191b12bc68c3d326ca6" Oct 05 08:27:37 crc kubenswrapper[4846]: E1005 08:27:37.023164 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3cbe3ecc81ad4981a422438847fac4d03bbddb0d3395f191b12bc68c3d326ca6\": container with ID starting with 3cbe3ecc81ad4981a422438847fac4d03bbddb0d3395f191b12bc68c3d326ca6 not found: ID does not exist" containerID="3cbe3ecc81ad4981a422438847fac4d03bbddb0d3395f191b12bc68c3d326ca6" Oct 05 08:27:37 crc kubenswrapper[4846]: I1005 08:27:37.023240 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3cbe3ecc81ad4981a422438847fac4d03bbddb0d3395f191b12bc68c3d326ca6"} err="failed to get container status \"3cbe3ecc81ad4981a422438847fac4d03bbddb0d3395f191b12bc68c3d326ca6\": rpc error: code = NotFound desc = could not find container \"3cbe3ecc81ad4981a422438847fac4d03bbddb0d3395f191b12bc68c3d326ca6\": container with ID starting with 3cbe3ecc81ad4981a422438847fac4d03bbddb0d3395f191b12bc68c3d326ca6 not found: ID does not exist" Oct 05 08:27:37 crc kubenswrapper[4846]: I1005 08:27:37.043605 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-create-qd22b" Oct 05 08:27:37 crc kubenswrapper[4846]: I1005 08:27:37.557747 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-create-qd22b"] Oct 05 08:27:37 crc kubenswrapper[4846]: I1005 08:27:37.869470 4846 generic.go:334] "Generic (PLEG): container finished" podID="c651e1a1-e229-40f6-8c40-76b40a301d78" containerID="4e0247457600bd829ebc59711475f72260f4bcf0734779b1ccf5eeb7694a54df" exitCode=0 Oct 05 08:27:37 crc kubenswrapper[4846]: I1005 08:27:37.869562 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-create-qd22b" event={"ID":"c651e1a1-e229-40f6-8c40-76b40a301d78","Type":"ContainerDied","Data":"4e0247457600bd829ebc59711475f72260f4bcf0734779b1ccf5eeb7694a54df"} Oct 05 08:27:37 crc kubenswrapper[4846]: I1005 08:27:37.869623 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-create-qd22b" event={"ID":"c651e1a1-e229-40f6-8c40-76b40a301d78","Type":"ContainerStarted","Data":"67a408031c1019e3d4fca53250fbc5ab3dbb11e9eee4d6de00cc85d084afde44"} Oct 05 08:27:38 crc kubenswrapper[4846]: I1005 08:27:38.518820 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b978eedb-81df-4394-a22f-2c9fb7981791" path="/var/lib/kubelet/pods/b978eedb-81df-4394-a22f-2c9fb7981791/volumes" Oct 05 08:27:39 crc kubenswrapper[4846]: I1005 08:27:39.298726 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-create-qd22b" Oct 05 08:27:39 crc kubenswrapper[4846]: I1005 08:27:39.450784 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wvnfq\" (UniqueName: \"kubernetes.io/projected/c651e1a1-e229-40f6-8c40-76b40a301d78-kube-api-access-wvnfq\") pod \"c651e1a1-e229-40f6-8c40-76b40a301d78\" (UID: \"c651e1a1-e229-40f6-8c40-76b40a301d78\") " Oct 05 08:27:39 crc kubenswrapper[4846]: I1005 08:27:39.456905 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c651e1a1-e229-40f6-8c40-76b40a301d78-kube-api-access-wvnfq" (OuterVolumeSpecName: "kube-api-access-wvnfq") pod "c651e1a1-e229-40f6-8c40-76b40a301d78" (UID: "c651e1a1-e229-40f6-8c40-76b40a301d78"). InnerVolumeSpecName "kube-api-access-wvnfq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:27:39 crc kubenswrapper[4846]: I1005 08:27:39.552970 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wvnfq\" (UniqueName: \"kubernetes.io/projected/c651e1a1-e229-40f6-8c40-76b40a301d78-kube-api-access-wvnfq\") on node \"crc\" DevicePath \"\"" Oct 05 08:27:39 crc kubenswrapper[4846]: I1005 08:27:39.898778 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-create-qd22b" event={"ID":"c651e1a1-e229-40f6-8c40-76b40a301d78","Type":"ContainerDied","Data":"67a408031c1019e3d4fca53250fbc5ab3dbb11e9eee4d6de00cc85d084afde44"} Oct 05 08:27:39 crc kubenswrapper[4846]: I1005 08:27:39.898823 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="67a408031c1019e3d4fca53250fbc5ab3dbb11e9eee4d6de00cc85d084afde44" Oct 05 08:27:39 crc kubenswrapper[4846]: I1005 08:27:39.898844 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-create-qd22b" Oct 05 08:27:45 crc kubenswrapper[4846]: I1005 08:27:45.730666 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-85d5597856-frbmx" Oct 05 08:27:45 crc kubenswrapper[4846]: I1005 08:27:45.731322 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-85d5597856-frbmx" Oct 05 08:27:46 crc kubenswrapper[4846]: I1005 08:27:46.933843 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-1991-account-create-bp8jt"] Oct 05 08:27:46 crc kubenswrapper[4846]: E1005 08:27:46.934911 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c651e1a1-e229-40f6-8c40-76b40a301d78" containerName="mariadb-database-create" Oct 05 08:27:46 crc kubenswrapper[4846]: I1005 08:27:46.934936 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="c651e1a1-e229-40f6-8c40-76b40a301d78" containerName="mariadb-database-create" Oct 05 08:27:46 crc kubenswrapper[4846]: I1005 08:27:46.935356 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="c651e1a1-e229-40f6-8c40-76b40a301d78" containerName="mariadb-database-create" Oct 05 08:27:46 crc kubenswrapper[4846]: I1005 08:27:46.936486 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-1991-account-create-bp8jt" Oct 05 08:27:46 crc kubenswrapper[4846]: I1005 08:27:46.938587 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-db-secret" Oct 05 08:27:46 crc kubenswrapper[4846]: I1005 08:27:46.944778 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-1991-account-create-bp8jt"] Oct 05 08:27:47 crc kubenswrapper[4846]: I1005 08:27:47.079571 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4dq8d\" (UniqueName: \"kubernetes.io/projected/8307ddbf-1547-41e5-a501-b2d68e00b333-kube-api-access-4dq8d\") pod \"heat-1991-account-create-bp8jt\" (UID: \"8307ddbf-1547-41e5-a501-b2d68e00b333\") " pod="openstack/heat-1991-account-create-bp8jt" Oct 05 08:27:47 crc kubenswrapper[4846]: I1005 08:27:47.182367 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4dq8d\" (UniqueName: \"kubernetes.io/projected/8307ddbf-1547-41e5-a501-b2d68e00b333-kube-api-access-4dq8d\") pod \"heat-1991-account-create-bp8jt\" (UID: \"8307ddbf-1547-41e5-a501-b2d68e00b333\") " pod="openstack/heat-1991-account-create-bp8jt" Oct 05 08:27:47 crc kubenswrapper[4846]: I1005 08:27:47.215174 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4dq8d\" (UniqueName: \"kubernetes.io/projected/8307ddbf-1547-41e5-a501-b2d68e00b333-kube-api-access-4dq8d\") pod \"heat-1991-account-create-bp8jt\" (UID: \"8307ddbf-1547-41e5-a501-b2d68e00b333\") " pod="openstack/heat-1991-account-create-bp8jt" Oct 05 08:27:47 crc kubenswrapper[4846]: I1005 08:27:47.293352 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-1991-account-create-bp8jt" Oct 05 08:27:47 crc kubenswrapper[4846]: I1005 08:27:47.758107 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-1991-account-create-bp8jt"] Oct 05 08:27:48 crc kubenswrapper[4846]: I1005 08:27:48.002290 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-1991-account-create-bp8jt" event={"ID":"8307ddbf-1547-41e5-a501-b2d68e00b333","Type":"ContainerStarted","Data":"de8921f6e6acd8fc6cff9fa00d10c8d1df5dff3a6e7973952551590a8641423b"} Oct 05 08:27:49 crc kubenswrapper[4846]: I1005 08:27:49.018859 4846 generic.go:334] "Generic (PLEG): container finished" podID="8307ddbf-1547-41e5-a501-b2d68e00b333" containerID="660a8e9f27bfd3ab4802ae5d8e70d3d15734e986878cce2205c5ef39c55111ca" exitCode=0 Oct 05 08:27:49 crc kubenswrapper[4846]: I1005 08:27:49.018926 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-1991-account-create-bp8jt" event={"ID":"8307ddbf-1547-41e5-a501-b2d68e00b333","Type":"ContainerDied","Data":"660a8e9f27bfd3ab4802ae5d8e70d3d15734e986878cce2205c5ef39c55111ca"} Oct 05 08:27:50 crc kubenswrapper[4846]: I1005 08:27:50.395982 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-1991-account-create-bp8jt" Oct 05 08:27:50 crc kubenswrapper[4846]: I1005 08:27:50.561645 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4dq8d\" (UniqueName: \"kubernetes.io/projected/8307ddbf-1547-41e5-a501-b2d68e00b333-kube-api-access-4dq8d\") pod \"8307ddbf-1547-41e5-a501-b2d68e00b333\" (UID: \"8307ddbf-1547-41e5-a501-b2d68e00b333\") " Oct 05 08:27:50 crc kubenswrapper[4846]: I1005 08:27:50.567493 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8307ddbf-1547-41e5-a501-b2d68e00b333-kube-api-access-4dq8d" (OuterVolumeSpecName: "kube-api-access-4dq8d") pod "8307ddbf-1547-41e5-a501-b2d68e00b333" (UID: "8307ddbf-1547-41e5-a501-b2d68e00b333"). InnerVolumeSpecName "kube-api-access-4dq8d". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:27:50 crc kubenswrapper[4846]: I1005 08:27:50.664715 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4dq8d\" (UniqueName: \"kubernetes.io/projected/8307ddbf-1547-41e5-a501-b2d68e00b333-kube-api-access-4dq8d\") on node \"crc\" DevicePath \"\"" Oct 05 08:27:51 crc kubenswrapper[4846]: I1005 08:27:51.081238 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-1991-account-create-bp8jt" event={"ID":"8307ddbf-1547-41e5-a501-b2d68e00b333","Type":"ContainerDied","Data":"de8921f6e6acd8fc6cff9fa00d10c8d1df5dff3a6e7973952551590a8641423b"} Oct 05 08:27:51 crc kubenswrapper[4846]: I1005 08:27:51.081310 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="de8921f6e6acd8fc6cff9fa00d10c8d1df5dff3a6e7973952551590a8641423b" Oct 05 08:27:51 crc kubenswrapper[4846]: I1005 08:27:51.081398 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-1991-account-create-bp8jt" Oct 05 08:27:52 crc kubenswrapper[4846]: I1005 08:27:52.007330 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-db-sync-cxjgw"] Oct 05 08:27:52 crc kubenswrapper[4846]: E1005 08:27:52.007783 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8307ddbf-1547-41e5-a501-b2d68e00b333" containerName="mariadb-account-create" Oct 05 08:27:52 crc kubenswrapper[4846]: I1005 08:27:52.007795 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="8307ddbf-1547-41e5-a501-b2d68e00b333" containerName="mariadb-account-create" Oct 05 08:27:52 crc kubenswrapper[4846]: I1005 08:27:52.008001 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="8307ddbf-1547-41e5-a501-b2d68e00b333" containerName="mariadb-account-create" Oct 05 08:27:52 crc kubenswrapper[4846]: I1005 08:27:52.008737 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-cxjgw" Oct 05 08:27:52 crc kubenswrapper[4846]: I1005 08:27:52.010821 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-heat-dockercfg-ksrp6" Oct 05 08:27:52 crc kubenswrapper[4846]: I1005 08:27:52.010854 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-config-data" Oct 05 08:27:52 crc kubenswrapper[4846]: I1005 08:27:52.036765 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-sync-cxjgw"] Oct 05 08:27:52 crc kubenswrapper[4846]: I1005 08:27:52.205712 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d95e7d3b-24b0-4200-af1a-72c6dc45ef56-combined-ca-bundle\") pod \"heat-db-sync-cxjgw\" (UID: \"d95e7d3b-24b0-4200-af1a-72c6dc45ef56\") " pod="openstack/heat-db-sync-cxjgw" Oct 05 08:27:52 crc kubenswrapper[4846]: I1005 08:27:52.206830 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5jb7f\" (UniqueName: \"kubernetes.io/projected/d95e7d3b-24b0-4200-af1a-72c6dc45ef56-kube-api-access-5jb7f\") pod \"heat-db-sync-cxjgw\" (UID: \"d95e7d3b-24b0-4200-af1a-72c6dc45ef56\") " pod="openstack/heat-db-sync-cxjgw" Oct 05 08:27:52 crc kubenswrapper[4846]: I1005 08:27:52.206937 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d95e7d3b-24b0-4200-af1a-72c6dc45ef56-config-data\") pod \"heat-db-sync-cxjgw\" (UID: \"d95e7d3b-24b0-4200-af1a-72c6dc45ef56\") " pod="openstack/heat-db-sync-cxjgw" Oct 05 08:27:52 crc kubenswrapper[4846]: I1005 08:27:52.308486 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5jb7f\" (UniqueName: \"kubernetes.io/projected/d95e7d3b-24b0-4200-af1a-72c6dc45ef56-kube-api-access-5jb7f\") pod \"heat-db-sync-cxjgw\" (UID: \"d95e7d3b-24b0-4200-af1a-72c6dc45ef56\") " pod="openstack/heat-db-sync-cxjgw" Oct 05 08:27:52 crc kubenswrapper[4846]: I1005 08:27:52.308538 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d95e7d3b-24b0-4200-af1a-72c6dc45ef56-config-data\") pod \"heat-db-sync-cxjgw\" (UID: \"d95e7d3b-24b0-4200-af1a-72c6dc45ef56\") " pod="openstack/heat-db-sync-cxjgw" Oct 05 08:27:52 crc kubenswrapper[4846]: I1005 08:27:52.308603 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d95e7d3b-24b0-4200-af1a-72c6dc45ef56-combined-ca-bundle\") pod \"heat-db-sync-cxjgw\" (UID: \"d95e7d3b-24b0-4200-af1a-72c6dc45ef56\") " pod="openstack/heat-db-sync-cxjgw" Oct 05 08:27:52 crc kubenswrapper[4846]: I1005 08:27:52.326003 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d95e7d3b-24b0-4200-af1a-72c6dc45ef56-config-data\") pod \"heat-db-sync-cxjgw\" (UID: \"d95e7d3b-24b0-4200-af1a-72c6dc45ef56\") " pod="openstack/heat-db-sync-cxjgw" Oct 05 08:27:52 crc kubenswrapper[4846]: I1005 08:27:52.326516 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d95e7d3b-24b0-4200-af1a-72c6dc45ef56-combined-ca-bundle\") pod \"heat-db-sync-cxjgw\" (UID: \"d95e7d3b-24b0-4200-af1a-72c6dc45ef56\") " pod="openstack/heat-db-sync-cxjgw" Oct 05 08:27:52 crc kubenswrapper[4846]: I1005 08:27:52.327761 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5jb7f\" (UniqueName: \"kubernetes.io/projected/d95e7d3b-24b0-4200-af1a-72c6dc45ef56-kube-api-access-5jb7f\") pod \"heat-db-sync-cxjgw\" (UID: \"d95e7d3b-24b0-4200-af1a-72c6dc45ef56\") " pod="openstack/heat-db-sync-cxjgw" Oct 05 08:27:52 crc kubenswrapper[4846]: I1005 08:27:52.626842 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-cxjgw" Oct 05 08:27:53 crc kubenswrapper[4846]: I1005 08:27:53.163586 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-sync-cxjgw"] Oct 05 08:27:54 crc kubenswrapper[4846]: I1005 08:27:54.116300 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-cxjgw" event={"ID":"d95e7d3b-24b0-4200-af1a-72c6dc45ef56","Type":"ContainerStarted","Data":"ad72690f7521cf6b5723567979fa7898600c551bc177381a3c36be432cd0bd20"} Oct 05 08:27:57 crc kubenswrapper[4846]: I1005 08:27:57.673326 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-85d5597856-frbmx" Oct 05 08:27:59 crc kubenswrapper[4846]: I1005 08:27:59.311932 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-85d5597856-frbmx" Oct 05 08:27:59 crc kubenswrapper[4846]: I1005 08:27:59.387799 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-f8ddb658d-6wlz2"] Oct 05 08:27:59 crc kubenswrapper[4846]: I1005 08:27:59.388081 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-f8ddb658d-6wlz2" podUID="8329aafa-c3d7-474c-a31b-a56e0d23ddce" containerName="horizon" containerID="cri-o://d5ef0f6c75a94d51341d689bb7cb65065aad64a24ca9ea981f961495ba170d35" gracePeriod=30 Oct 05 08:27:59 crc kubenswrapper[4846]: I1005 08:27:59.388273 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-f8ddb658d-6wlz2" podUID="8329aafa-c3d7-474c-a31b-a56e0d23ddce" containerName="horizon-log" containerID="cri-o://22c179cdba0416e23837204edd3d2a434931018a7d897c19b9aa1476bc215022" gracePeriod=30 Oct 05 08:28:02 crc kubenswrapper[4846]: I1005 08:28:02.230991 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-cxjgw" event={"ID":"d95e7d3b-24b0-4200-af1a-72c6dc45ef56","Type":"ContainerStarted","Data":"91cd2e2593dec9c125d5555846c0eb484079601ab70ea8aa9be280027c2d6110"} Oct 05 08:28:02 crc kubenswrapper[4846]: I1005 08:28:02.251832 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-db-sync-cxjgw" podStartSLOduration=2.738980303 podStartE2EDuration="11.25181419s" podCreationTimestamp="2025-10-05 08:27:51 +0000 UTC" firstStartedPulling="2025-10-05 08:27:53.138426975 +0000 UTC m=+5995.379279760" lastFinishedPulling="2025-10-05 08:28:01.651260842 +0000 UTC m=+6003.892113647" observedRunningTime="2025-10-05 08:28:02.251334807 +0000 UTC m=+6004.492187602" watchObservedRunningTime="2025-10-05 08:28:02.25181419 +0000 UTC m=+6004.492666985" Oct 05 08:28:02 crc kubenswrapper[4846]: E1005 08:28:02.734550 4846 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8329aafa_c3d7_474c_a31b_a56e0d23ddce.slice/crio-conmon-d5ef0f6c75a94d51341d689bb7cb65065aad64a24ca9ea981f961495ba170d35.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8329aafa_c3d7_474c_a31b_a56e0d23ddce.slice/crio-d5ef0f6c75a94d51341d689bb7cb65065aad64a24ca9ea981f961495ba170d35.scope\": RecentStats: unable to find data in memory cache]" Oct 05 08:28:03 crc kubenswrapper[4846]: I1005 08:28:03.249676 4846 generic.go:334] "Generic (PLEG): container finished" podID="8329aafa-c3d7-474c-a31b-a56e0d23ddce" containerID="d5ef0f6c75a94d51341d689bb7cb65065aad64a24ca9ea981f961495ba170d35" exitCode=0 Oct 05 08:28:03 crc kubenswrapper[4846]: I1005 08:28:03.249765 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-f8ddb658d-6wlz2" event={"ID":"8329aafa-c3d7-474c-a31b-a56e0d23ddce","Type":"ContainerDied","Data":"d5ef0f6c75a94d51341d689bb7cb65065aad64a24ca9ea981f961495ba170d35"} Oct 05 08:28:04 crc kubenswrapper[4846]: I1005 08:28:04.259453 4846 generic.go:334] "Generic (PLEG): container finished" podID="d95e7d3b-24b0-4200-af1a-72c6dc45ef56" containerID="91cd2e2593dec9c125d5555846c0eb484079601ab70ea8aa9be280027c2d6110" exitCode=0 Oct 05 08:28:04 crc kubenswrapper[4846]: I1005 08:28:04.259640 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-cxjgw" event={"ID":"d95e7d3b-24b0-4200-af1a-72c6dc45ef56","Type":"ContainerDied","Data":"91cd2e2593dec9c125d5555846c0eb484079601ab70ea8aa9be280027c2d6110"} Oct 05 08:28:05 crc kubenswrapper[4846]: I1005 08:28:05.643166 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-cxjgw" Oct 05 08:28:05 crc kubenswrapper[4846]: I1005 08:28:05.772550 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d95e7d3b-24b0-4200-af1a-72c6dc45ef56-config-data\") pod \"d95e7d3b-24b0-4200-af1a-72c6dc45ef56\" (UID: \"d95e7d3b-24b0-4200-af1a-72c6dc45ef56\") " Oct 05 08:28:05 crc kubenswrapper[4846]: I1005 08:28:05.772611 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5jb7f\" (UniqueName: \"kubernetes.io/projected/d95e7d3b-24b0-4200-af1a-72c6dc45ef56-kube-api-access-5jb7f\") pod \"d95e7d3b-24b0-4200-af1a-72c6dc45ef56\" (UID: \"d95e7d3b-24b0-4200-af1a-72c6dc45ef56\") " Oct 05 08:28:05 crc kubenswrapper[4846]: I1005 08:28:05.772705 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d95e7d3b-24b0-4200-af1a-72c6dc45ef56-combined-ca-bundle\") pod \"d95e7d3b-24b0-4200-af1a-72c6dc45ef56\" (UID: \"d95e7d3b-24b0-4200-af1a-72c6dc45ef56\") " Oct 05 08:28:05 crc kubenswrapper[4846]: I1005 08:28:05.779419 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d95e7d3b-24b0-4200-af1a-72c6dc45ef56-kube-api-access-5jb7f" (OuterVolumeSpecName: "kube-api-access-5jb7f") pod "d95e7d3b-24b0-4200-af1a-72c6dc45ef56" (UID: "d95e7d3b-24b0-4200-af1a-72c6dc45ef56"). InnerVolumeSpecName "kube-api-access-5jb7f". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:28:05 crc kubenswrapper[4846]: I1005 08:28:05.801380 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d95e7d3b-24b0-4200-af1a-72c6dc45ef56-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d95e7d3b-24b0-4200-af1a-72c6dc45ef56" (UID: "d95e7d3b-24b0-4200-af1a-72c6dc45ef56"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:28:05 crc kubenswrapper[4846]: I1005 08:28:05.857387 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d95e7d3b-24b0-4200-af1a-72c6dc45ef56-config-data" (OuterVolumeSpecName: "config-data") pod "d95e7d3b-24b0-4200-af1a-72c6dc45ef56" (UID: "d95e7d3b-24b0-4200-af1a-72c6dc45ef56"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:28:05 crc kubenswrapper[4846]: I1005 08:28:05.882742 4846 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d95e7d3b-24b0-4200-af1a-72c6dc45ef56-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 08:28:05 crc kubenswrapper[4846]: I1005 08:28:05.882805 4846 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d95e7d3b-24b0-4200-af1a-72c6dc45ef56-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 08:28:05 crc kubenswrapper[4846]: I1005 08:28:05.882825 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5jb7f\" (UniqueName: \"kubernetes.io/projected/d95e7d3b-24b0-4200-af1a-72c6dc45ef56-kube-api-access-5jb7f\") on node \"crc\" DevicePath \"\"" Oct 05 08:28:06 crc kubenswrapper[4846]: I1005 08:28:06.263586 4846 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-f8ddb658d-6wlz2" podUID="8329aafa-c3d7-474c-a31b-a56e0d23ddce" containerName="horizon" probeResult="failure" output="Get \"https://10.217.1.103:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.103:8443: connect: connection refused" Oct 05 08:28:06 crc kubenswrapper[4846]: I1005 08:28:06.283880 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-cxjgw" event={"ID":"d95e7d3b-24b0-4200-af1a-72c6dc45ef56","Type":"ContainerDied","Data":"ad72690f7521cf6b5723567979fa7898600c551bc177381a3c36be432cd0bd20"} Oct 05 08:28:06 crc kubenswrapper[4846]: I1005 08:28:06.283928 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ad72690f7521cf6b5723567979fa7898600c551bc177381a3c36be432cd0bd20" Oct 05 08:28:06 crc kubenswrapper[4846]: I1005 08:28:06.284006 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-cxjgw" Oct 05 08:28:07 crc kubenswrapper[4846]: I1005 08:28:07.475144 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-engine-6b557c77dd-fbhjb"] Oct 05 08:28:07 crc kubenswrapper[4846]: E1005 08:28:07.475673 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d95e7d3b-24b0-4200-af1a-72c6dc45ef56" containerName="heat-db-sync" Oct 05 08:28:07 crc kubenswrapper[4846]: I1005 08:28:07.475693 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="d95e7d3b-24b0-4200-af1a-72c6dc45ef56" containerName="heat-db-sync" Oct 05 08:28:07 crc kubenswrapper[4846]: I1005 08:28:07.475936 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="d95e7d3b-24b0-4200-af1a-72c6dc45ef56" containerName="heat-db-sync" Oct 05 08:28:07 crc kubenswrapper[4846]: I1005 08:28:07.476742 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-6b557c77dd-fbhjb" Oct 05 08:28:07 crc kubenswrapper[4846]: I1005 08:28:07.478880 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-config-data" Oct 05 08:28:07 crc kubenswrapper[4846]: I1005 08:28:07.479233 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-engine-config-data" Oct 05 08:28:07 crc kubenswrapper[4846]: I1005 08:28:07.479293 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-heat-dockercfg-ksrp6" Oct 05 08:28:07 crc kubenswrapper[4846]: I1005 08:28:07.500966 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-engine-6b557c77dd-fbhjb"] Oct 05 08:28:07 crc kubenswrapper[4846]: I1005 08:28:07.626679 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2bfbe6d7-d95b-4ac0-a8f6-52f89ede6ed2-config-data\") pod \"heat-engine-6b557c77dd-fbhjb\" (UID: \"2bfbe6d7-d95b-4ac0-a8f6-52f89ede6ed2\") " pod="openstack/heat-engine-6b557c77dd-fbhjb" Oct 05 08:28:07 crc kubenswrapper[4846]: I1005 08:28:07.626902 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2bfbe6d7-d95b-4ac0-a8f6-52f89ede6ed2-config-data-custom\") pod \"heat-engine-6b557c77dd-fbhjb\" (UID: \"2bfbe6d7-d95b-4ac0-a8f6-52f89ede6ed2\") " pod="openstack/heat-engine-6b557c77dd-fbhjb" Oct 05 08:28:07 crc kubenswrapper[4846]: I1005 08:28:07.627133 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2bfbe6d7-d95b-4ac0-a8f6-52f89ede6ed2-combined-ca-bundle\") pod \"heat-engine-6b557c77dd-fbhjb\" (UID: \"2bfbe6d7-d95b-4ac0-a8f6-52f89ede6ed2\") " pod="openstack/heat-engine-6b557c77dd-fbhjb" Oct 05 08:28:07 crc kubenswrapper[4846]: I1005 08:28:07.627213 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dfvvj\" (UniqueName: \"kubernetes.io/projected/2bfbe6d7-d95b-4ac0-a8f6-52f89ede6ed2-kube-api-access-dfvvj\") pod \"heat-engine-6b557c77dd-fbhjb\" (UID: \"2bfbe6d7-d95b-4ac0-a8f6-52f89ede6ed2\") " pod="openstack/heat-engine-6b557c77dd-fbhjb" Oct 05 08:28:07 crc kubenswrapper[4846]: I1005 08:28:07.635584 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-api-846d578b5-97wdg"] Oct 05 08:28:07 crc kubenswrapper[4846]: I1005 08:28:07.636764 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-846d578b5-97wdg" Oct 05 08:28:07 crc kubenswrapper[4846]: I1005 08:28:07.638654 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-api-config-data" Oct 05 08:28:07 crc kubenswrapper[4846]: I1005 08:28:07.652968 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-846d578b5-97wdg"] Oct 05 08:28:07 crc kubenswrapper[4846]: I1005 08:28:07.701744 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-cfnapi-547d5fb769-fllds"] Oct 05 08:28:07 crc kubenswrapper[4846]: I1005 08:28:07.703213 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-547d5fb769-fllds" Oct 05 08:28:07 crc kubenswrapper[4846]: I1005 08:28:07.705404 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-cfnapi-config-data" Oct 05 08:28:07 crc kubenswrapper[4846]: I1005 08:28:07.710193 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-547d5fb769-fllds"] Oct 05 08:28:07 crc kubenswrapper[4846]: I1005 08:28:07.728533 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2eaaa848-5b1c-402c-af31-23dc2d3eaef1-combined-ca-bundle\") pod \"heat-api-846d578b5-97wdg\" (UID: \"2eaaa848-5b1c-402c-af31-23dc2d3eaef1\") " pod="openstack/heat-api-846d578b5-97wdg" Oct 05 08:28:07 crc kubenswrapper[4846]: I1005 08:28:07.728575 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2eaaa848-5b1c-402c-af31-23dc2d3eaef1-config-data-custom\") pod \"heat-api-846d578b5-97wdg\" (UID: \"2eaaa848-5b1c-402c-af31-23dc2d3eaef1\") " pod="openstack/heat-api-846d578b5-97wdg" Oct 05 08:28:07 crc kubenswrapper[4846]: I1005 08:28:07.728610 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2bfbe6d7-d95b-4ac0-a8f6-52f89ede6ed2-config-data\") pod \"heat-engine-6b557c77dd-fbhjb\" (UID: \"2bfbe6d7-d95b-4ac0-a8f6-52f89ede6ed2\") " pod="openstack/heat-engine-6b557c77dd-fbhjb" Oct 05 08:28:07 crc kubenswrapper[4846]: I1005 08:28:07.728660 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2eaaa848-5b1c-402c-af31-23dc2d3eaef1-config-data\") pod \"heat-api-846d578b5-97wdg\" (UID: \"2eaaa848-5b1c-402c-af31-23dc2d3eaef1\") " pod="openstack/heat-api-846d578b5-97wdg" Oct 05 08:28:07 crc kubenswrapper[4846]: I1005 08:28:07.728691 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2bfbe6d7-d95b-4ac0-a8f6-52f89ede6ed2-config-data-custom\") pod \"heat-engine-6b557c77dd-fbhjb\" (UID: \"2bfbe6d7-d95b-4ac0-a8f6-52f89ede6ed2\") " pod="openstack/heat-engine-6b557c77dd-fbhjb" Oct 05 08:28:07 crc kubenswrapper[4846]: I1005 08:28:07.728720 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9lq27\" (UniqueName: \"kubernetes.io/projected/9fff0fb3-5365-420b-afc2-c997c3b98b6a-kube-api-access-9lq27\") pod \"heat-cfnapi-547d5fb769-fllds\" (UID: \"9fff0fb3-5365-420b-afc2-c997c3b98b6a\") " pod="openstack/heat-cfnapi-547d5fb769-fllds" Oct 05 08:28:07 crc kubenswrapper[4846]: I1005 08:28:07.728737 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p5m2r\" (UniqueName: \"kubernetes.io/projected/2eaaa848-5b1c-402c-af31-23dc2d3eaef1-kube-api-access-p5m2r\") pod \"heat-api-846d578b5-97wdg\" (UID: \"2eaaa848-5b1c-402c-af31-23dc2d3eaef1\") " pod="openstack/heat-api-846d578b5-97wdg" Oct 05 08:28:07 crc kubenswrapper[4846]: I1005 08:28:07.728774 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9fff0fb3-5365-420b-afc2-c997c3b98b6a-config-data\") pod \"heat-cfnapi-547d5fb769-fllds\" (UID: \"9fff0fb3-5365-420b-afc2-c997c3b98b6a\") " pod="openstack/heat-cfnapi-547d5fb769-fllds" Oct 05 08:28:07 crc kubenswrapper[4846]: I1005 08:28:07.728812 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2bfbe6d7-d95b-4ac0-a8f6-52f89ede6ed2-combined-ca-bundle\") pod \"heat-engine-6b557c77dd-fbhjb\" (UID: \"2bfbe6d7-d95b-4ac0-a8f6-52f89ede6ed2\") " pod="openstack/heat-engine-6b557c77dd-fbhjb" Oct 05 08:28:07 crc kubenswrapper[4846]: I1005 08:28:07.728841 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9fff0fb3-5365-420b-afc2-c997c3b98b6a-combined-ca-bundle\") pod \"heat-cfnapi-547d5fb769-fllds\" (UID: \"9fff0fb3-5365-420b-afc2-c997c3b98b6a\") " pod="openstack/heat-cfnapi-547d5fb769-fllds" Oct 05 08:28:07 crc kubenswrapper[4846]: I1005 08:28:07.728866 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9fff0fb3-5365-420b-afc2-c997c3b98b6a-config-data-custom\") pod \"heat-cfnapi-547d5fb769-fllds\" (UID: \"9fff0fb3-5365-420b-afc2-c997c3b98b6a\") " pod="openstack/heat-cfnapi-547d5fb769-fllds" Oct 05 08:28:07 crc kubenswrapper[4846]: I1005 08:28:07.728885 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dfvvj\" (UniqueName: \"kubernetes.io/projected/2bfbe6d7-d95b-4ac0-a8f6-52f89ede6ed2-kube-api-access-dfvvj\") pod \"heat-engine-6b557c77dd-fbhjb\" (UID: \"2bfbe6d7-d95b-4ac0-a8f6-52f89ede6ed2\") " pod="openstack/heat-engine-6b557c77dd-fbhjb" Oct 05 08:28:07 crc kubenswrapper[4846]: I1005 08:28:07.735733 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2bfbe6d7-d95b-4ac0-a8f6-52f89ede6ed2-config-data\") pod \"heat-engine-6b557c77dd-fbhjb\" (UID: \"2bfbe6d7-d95b-4ac0-a8f6-52f89ede6ed2\") " pod="openstack/heat-engine-6b557c77dd-fbhjb" Oct 05 08:28:07 crc kubenswrapper[4846]: I1005 08:28:07.736485 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2bfbe6d7-d95b-4ac0-a8f6-52f89ede6ed2-config-data-custom\") pod \"heat-engine-6b557c77dd-fbhjb\" (UID: \"2bfbe6d7-d95b-4ac0-a8f6-52f89ede6ed2\") " pod="openstack/heat-engine-6b557c77dd-fbhjb" Oct 05 08:28:07 crc kubenswrapper[4846]: I1005 08:28:07.739510 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2bfbe6d7-d95b-4ac0-a8f6-52f89ede6ed2-combined-ca-bundle\") pod \"heat-engine-6b557c77dd-fbhjb\" (UID: \"2bfbe6d7-d95b-4ac0-a8f6-52f89ede6ed2\") " pod="openstack/heat-engine-6b557c77dd-fbhjb" Oct 05 08:28:07 crc kubenswrapper[4846]: I1005 08:28:07.749913 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dfvvj\" (UniqueName: \"kubernetes.io/projected/2bfbe6d7-d95b-4ac0-a8f6-52f89ede6ed2-kube-api-access-dfvvj\") pod \"heat-engine-6b557c77dd-fbhjb\" (UID: \"2bfbe6d7-d95b-4ac0-a8f6-52f89ede6ed2\") " pod="openstack/heat-engine-6b557c77dd-fbhjb" Oct 05 08:28:07 crc kubenswrapper[4846]: I1005 08:28:07.804073 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-6b557c77dd-fbhjb" Oct 05 08:28:07 crc kubenswrapper[4846]: I1005 08:28:07.830368 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9fff0fb3-5365-420b-afc2-c997c3b98b6a-combined-ca-bundle\") pod \"heat-cfnapi-547d5fb769-fllds\" (UID: \"9fff0fb3-5365-420b-afc2-c997c3b98b6a\") " pod="openstack/heat-cfnapi-547d5fb769-fllds" Oct 05 08:28:07 crc kubenswrapper[4846]: I1005 08:28:07.830425 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9fff0fb3-5365-420b-afc2-c997c3b98b6a-config-data-custom\") pod \"heat-cfnapi-547d5fb769-fllds\" (UID: \"9fff0fb3-5365-420b-afc2-c997c3b98b6a\") " pod="openstack/heat-cfnapi-547d5fb769-fllds" Oct 05 08:28:07 crc kubenswrapper[4846]: I1005 08:28:07.830484 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2eaaa848-5b1c-402c-af31-23dc2d3eaef1-combined-ca-bundle\") pod \"heat-api-846d578b5-97wdg\" (UID: \"2eaaa848-5b1c-402c-af31-23dc2d3eaef1\") " pod="openstack/heat-api-846d578b5-97wdg" Oct 05 08:28:07 crc kubenswrapper[4846]: I1005 08:28:07.830509 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2eaaa848-5b1c-402c-af31-23dc2d3eaef1-config-data-custom\") pod \"heat-api-846d578b5-97wdg\" (UID: \"2eaaa848-5b1c-402c-af31-23dc2d3eaef1\") " pod="openstack/heat-api-846d578b5-97wdg" Oct 05 08:28:07 crc kubenswrapper[4846]: I1005 08:28:07.830588 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2eaaa848-5b1c-402c-af31-23dc2d3eaef1-config-data\") pod \"heat-api-846d578b5-97wdg\" (UID: \"2eaaa848-5b1c-402c-af31-23dc2d3eaef1\") " pod="openstack/heat-api-846d578b5-97wdg" Oct 05 08:28:07 crc kubenswrapper[4846]: I1005 08:28:07.830644 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9lq27\" (UniqueName: \"kubernetes.io/projected/9fff0fb3-5365-420b-afc2-c997c3b98b6a-kube-api-access-9lq27\") pod \"heat-cfnapi-547d5fb769-fllds\" (UID: \"9fff0fb3-5365-420b-afc2-c997c3b98b6a\") " pod="openstack/heat-cfnapi-547d5fb769-fllds" Oct 05 08:28:07 crc kubenswrapper[4846]: I1005 08:28:07.830671 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p5m2r\" (UniqueName: \"kubernetes.io/projected/2eaaa848-5b1c-402c-af31-23dc2d3eaef1-kube-api-access-p5m2r\") pod \"heat-api-846d578b5-97wdg\" (UID: \"2eaaa848-5b1c-402c-af31-23dc2d3eaef1\") " pod="openstack/heat-api-846d578b5-97wdg" Oct 05 08:28:07 crc kubenswrapper[4846]: I1005 08:28:07.830722 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9fff0fb3-5365-420b-afc2-c997c3b98b6a-config-data\") pod \"heat-cfnapi-547d5fb769-fllds\" (UID: \"9fff0fb3-5365-420b-afc2-c997c3b98b6a\") " pod="openstack/heat-cfnapi-547d5fb769-fllds" Oct 05 08:28:07 crc kubenswrapper[4846]: I1005 08:28:07.836483 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9fff0fb3-5365-420b-afc2-c997c3b98b6a-config-data-custom\") pod \"heat-cfnapi-547d5fb769-fllds\" (UID: \"9fff0fb3-5365-420b-afc2-c997c3b98b6a\") " pod="openstack/heat-cfnapi-547d5fb769-fllds" Oct 05 08:28:07 crc kubenswrapper[4846]: I1005 08:28:07.838538 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2eaaa848-5b1c-402c-af31-23dc2d3eaef1-config-data\") pod \"heat-api-846d578b5-97wdg\" (UID: \"2eaaa848-5b1c-402c-af31-23dc2d3eaef1\") " pod="openstack/heat-api-846d578b5-97wdg" Oct 05 08:28:07 crc kubenswrapper[4846]: I1005 08:28:07.839845 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2eaaa848-5b1c-402c-af31-23dc2d3eaef1-config-data-custom\") pod \"heat-api-846d578b5-97wdg\" (UID: \"2eaaa848-5b1c-402c-af31-23dc2d3eaef1\") " pod="openstack/heat-api-846d578b5-97wdg" Oct 05 08:28:07 crc kubenswrapper[4846]: I1005 08:28:07.844872 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9fff0fb3-5365-420b-afc2-c997c3b98b6a-combined-ca-bundle\") pod \"heat-cfnapi-547d5fb769-fllds\" (UID: \"9fff0fb3-5365-420b-afc2-c997c3b98b6a\") " pod="openstack/heat-cfnapi-547d5fb769-fllds" Oct 05 08:28:07 crc kubenswrapper[4846]: I1005 08:28:07.847414 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2eaaa848-5b1c-402c-af31-23dc2d3eaef1-combined-ca-bundle\") pod \"heat-api-846d578b5-97wdg\" (UID: \"2eaaa848-5b1c-402c-af31-23dc2d3eaef1\") " pod="openstack/heat-api-846d578b5-97wdg" Oct 05 08:28:07 crc kubenswrapper[4846]: I1005 08:28:07.848440 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9fff0fb3-5365-420b-afc2-c997c3b98b6a-config-data\") pod \"heat-cfnapi-547d5fb769-fllds\" (UID: \"9fff0fb3-5365-420b-afc2-c997c3b98b6a\") " pod="openstack/heat-cfnapi-547d5fb769-fllds" Oct 05 08:28:07 crc kubenswrapper[4846]: I1005 08:28:07.853382 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p5m2r\" (UniqueName: \"kubernetes.io/projected/2eaaa848-5b1c-402c-af31-23dc2d3eaef1-kube-api-access-p5m2r\") pod \"heat-api-846d578b5-97wdg\" (UID: \"2eaaa848-5b1c-402c-af31-23dc2d3eaef1\") " pod="openstack/heat-api-846d578b5-97wdg" Oct 05 08:28:07 crc kubenswrapper[4846]: I1005 08:28:07.854040 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9lq27\" (UniqueName: \"kubernetes.io/projected/9fff0fb3-5365-420b-afc2-c997c3b98b6a-kube-api-access-9lq27\") pod \"heat-cfnapi-547d5fb769-fllds\" (UID: \"9fff0fb3-5365-420b-afc2-c997c3b98b6a\") " pod="openstack/heat-cfnapi-547d5fb769-fllds" Oct 05 08:28:07 crc kubenswrapper[4846]: I1005 08:28:07.983631 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-846d578b5-97wdg" Oct 05 08:28:08 crc kubenswrapper[4846]: I1005 08:28:08.021463 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-547d5fb769-fllds" Oct 05 08:28:08 crc kubenswrapper[4846]: I1005 08:28:08.325683 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-engine-6b557c77dd-fbhjb"] Oct 05 08:28:08 crc kubenswrapper[4846]: W1005 08:28:08.478682 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2eaaa848_5b1c_402c_af31_23dc2d3eaef1.slice/crio-528c59b17ff7fdfd24529a3fba5f0444086a5eaa5bdd27ed995cc60acce68d1e WatchSource:0}: Error finding container 528c59b17ff7fdfd24529a3fba5f0444086a5eaa5bdd27ed995cc60acce68d1e: Status 404 returned error can't find the container with id 528c59b17ff7fdfd24529a3fba5f0444086a5eaa5bdd27ed995cc60acce68d1e Oct 05 08:28:08 crc kubenswrapper[4846]: I1005 08:28:08.491777 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-846d578b5-97wdg"] Oct 05 08:28:08 crc kubenswrapper[4846]: I1005 08:28:08.545705 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-547d5fb769-fllds"] Oct 05 08:28:09 crc kubenswrapper[4846]: I1005 08:28:09.352426 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-6b557c77dd-fbhjb" event={"ID":"2bfbe6d7-d95b-4ac0-a8f6-52f89ede6ed2","Type":"ContainerStarted","Data":"7a898116d7abb553b82d6daafdb6c659629f2308f48ba7d6ed48fe80c60b8ef5"} Oct 05 08:28:09 crc kubenswrapper[4846]: I1005 08:28:09.352687 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-engine-6b557c77dd-fbhjb" Oct 05 08:28:09 crc kubenswrapper[4846]: I1005 08:28:09.352698 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-6b557c77dd-fbhjb" event={"ID":"2bfbe6d7-d95b-4ac0-a8f6-52f89ede6ed2","Type":"ContainerStarted","Data":"9ac5aab6c23b2260ae33b2e74e29f510606142eb021382c77a67fd5196fc26cd"} Oct 05 08:28:09 crc kubenswrapper[4846]: I1005 08:28:09.357618 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-846d578b5-97wdg" event={"ID":"2eaaa848-5b1c-402c-af31-23dc2d3eaef1","Type":"ContainerStarted","Data":"528c59b17ff7fdfd24529a3fba5f0444086a5eaa5bdd27ed995cc60acce68d1e"} Oct 05 08:28:09 crc kubenswrapper[4846]: I1005 08:28:09.360896 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-547d5fb769-fllds" event={"ID":"9fff0fb3-5365-420b-afc2-c997c3b98b6a","Type":"ContainerStarted","Data":"bc5607bc256d552d4767b57e06574d82ff727c0b4aca1bdd885c10c902f0acd2"} Oct 05 08:28:09 crc kubenswrapper[4846]: I1005 08:28:09.380801 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-engine-6b557c77dd-fbhjb" podStartSLOduration=2.38078089 podStartE2EDuration="2.38078089s" podCreationTimestamp="2025-10-05 08:28:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 08:28:09.370688122 +0000 UTC m=+6011.611540917" watchObservedRunningTime="2025-10-05 08:28:09.38078089 +0000 UTC m=+6011.621633665" Oct 05 08:28:11 crc kubenswrapper[4846]: I1005 08:28:11.395772 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-846d578b5-97wdg" event={"ID":"2eaaa848-5b1c-402c-af31-23dc2d3eaef1","Type":"ContainerStarted","Data":"15f2420fd1410113564e93430e6af24b281cb5a1db66770570c42107ebadba74"} Oct 05 08:28:11 crc kubenswrapper[4846]: I1005 08:28:11.397513 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-api-846d578b5-97wdg" Oct 05 08:28:11 crc kubenswrapper[4846]: I1005 08:28:11.398642 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-547d5fb769-fllds" event={"ID":"9fff0fb3-5365-420b-afc2-c997c3b98b6a","Type":"ContainerStarted","Data":"bf30db2f6b7764098249a02cd5401433aff89b58d2a0867c25f3e93fd461a127"} Oct 05 08:28:11 crc kubenswrapper[4846]: I1005 08:28:11.398979 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-cfnapi-547d5fb769-fllds" Oct 05 08:28:11 crc kubenswrapper[4846]: I1005 08:28:11.420716 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-api-846d578b5-97wdg" podStartSLOduration=2.627303276 podStartE2EDuration="4.42070011s" podCreationTimestamp="2025-10-05 08:28:07 +0000 UTC" firstStartedPulling="2025-10-05 08:28:08.481456438 +0000 UTC m=+6010.722309213" lastFinishedPulling="2025-10-05 08:28:10.274853262 +0000 UTC m=+6012.515706047" observedRunningTime="2025-10-05 08:28:11.415887783 +0000 UTC m=+6013.656740568" watchObservedRunningTime="2025-10-05 08:28:11.42070011 +0000 UTC m=+6013.661552885" Oct 05 08:28:11 crc kubenswrapper[4846]: I1005 08:28:11.444635 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-cfnapi-547d5fb769-fllds" podStartSLOduration=2.720836529 podStartE2EDuration="4.444608815s" podCreationTimestamp="2025-10-05 08:28:07 +0000 UTC" firstStartedPulling="2025-10-05 08:28:08.5527193 +0000 UTC m=+6010.793572075" lastFinishedPulling="2025-10-05 08:28:10.276491586 +0000 UTC m=+6012.517344361" observedRunningTime="2025-10-05 08:28:11.43914598 +0000 UTC m=+6013.679998775" watchObservedRunningTime="2025-10-05 08:28:11.444608815 +0000 UTC m=+6013.685461590" Oct 05 08:28:14 crc kubenswrapper[4846]: I1005 08:28:14.512845 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-engine-54b8fb86bb-q8qps"] Oct 05 08:28:14 crc kubenswrapper[4846]: I1005 08:28:14.518497 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-54b8fb86bb-q8qps" Oct 05 08:28:14 crc kubenswrapper[4846]: I1005 08:28:14.533694 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-engine-54b8fb86bb-q8qps"] Oct 05 08:28:14 crc kubenswrapper[4846]: I1005 08:28:14.546327 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-api-755575957b-g5bfm"] Oct 05 08:28:14 crc kubenswrapper[4846]: I1005 08:28:14.547584 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-755575957b-g5bfm" Oct 05 08:28:14 crc kubenswrapper[4846]: I1005 08:28:14.562057 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-755575957b-g5bfm"] Oct 05 08:28:14 crc kubenswrapper[4846]: I1005 08:28:14.619921 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-cfnapi-86b55f6c-pznr6"] Oct 05 08:28:14 crc kubenswrapper[4846]: I1005 08:28:14.621227 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-86b55f6c-pznr6" Oct 05 08:28:14 crc kubenswrapper[4846]: I1005 08:28:14.637102 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-86b55f6c-pznr6"] Oct 05 08:28:14 crc kubenswrapper[4846]: I1005 08:28:14.674459 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1bfe1edf-d89c-4d9e-b4fa-896c161c965f-config-data\") pod \"heat-api-755575957b-g5bfm\" (UID: \"1bfe1edf-d89c-4d9e-b4fa-896c161c965f\") " pod="openstack/heat-api-755575957b-g5bfm" Oct 05 08:28:14 crc kubenswrapper[4846]: I1005 08:28:14.674519 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1bfe1edf-d89c-4d9e-b4fa-896c161c965f-config-data-custom\") pod \"heat-api-755575957b-g5bfm\" (UID: \"1bfe1edf-d89c-4d9e-b4fa-896c161c965f\") " pod="openstack/heat-api-755575957b-g5bfm" Oct 05 08:28:14 crc kubenswrapper[4846]: I1005 08:28:14.674603 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6cz2d\" (UniqueName: \"kubernetes.io/projected/28512aa9-3230-44b3-a1f2-55247d128483-kube-api-access-6cz2d\") pod \"heat-engine-54b8fb86bb-q8qps\" (UID: \"28512aa9-3230-44b3-a1f2-55247d128483\") " pod="openstack/heat-engine-54b8fb86bb-q8qps" Oct 05 08:28:14 crc kubenswrapper[4846]: I1005 08:28:14.674645 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/28512aa9-3230-44b3-a1f2-55247d128483-config-data-custom\") pod \"heat-engine-54b8fb86bb-q8qps\" (UID: \"28512aa9-3230-44b3-a1f2-55247d128483\") " pod="openstack/heat-engine-54b8fb86bb-q8qps" Oct 05 08:28:14 crc kubenswrapper[4846]: I1005 08:28:14.674704 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/28512aa9-3230-44b3-a1f2-55247d128483-combined-ca-bundle\") pod \"heat-engine-54b8fb86bb-q8qps\" (UID: \"28512aa9-3230-44b3-a1f2-55247d128483\") " pod="openstack/heat-engine-54b8fb86bb-q8qps" Oct 05 08:28:14 crc kubenswrapper[4846]: I1005 08:28:14.674934 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bfv5r\" (UniqueName: \"kubernetes.io/projected/1bfe1edf-d89c-4d9e-b4fa-896c161c965f-kube-api-access-bfv5r\") pod \"heat-api-755575957b-g5bfm\" (UID: \"1bfe1edf-d89c-4d9e-b4fa-896c161c965f\") " pod="openstack/heat-api-755575957b-g5bfm" Oct 05 08:28:14 crc kubenswrapper[4846]: I1005 08:28:14.675028 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1bfe1edf-d89c-4d9e-b4fa-896c161c965f-combined-ca-bundle\") pod \"heat-api-755575957b-g5bfm\" (UID: \"1bfe1edf-d89c-4d9e-b4fa-896c161c965f\") " pod="openstack/heat-api-755575957b-g5bfm" Oct 05 08:28:14 crc kubenswrapper[4846]: I1005 08:28:14.675068 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/28512aa9-3230-44b3-a1f2-55247d128483-config-data\") pod \"heat-engine-54b8fb86bb-q8qps\" (UID: \"28512aa9-3230-44b3-a1f2-55247d128483\") " pod="openstack/heat-engine-54b8fb86bb-q8qps" Oct 05 08:28:14 crc kubenswrapper[4846]: I1005 08:28:14.776251 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ftf8x\" (UniqueName: \"kubernetes.io/projected/530d7f87-9f71-4d89-8520-ae77ca8a901d-kube-api-access-ftf8x\") pod \"heat-cfnapi-86b55f6c-pznr6\" (UID: \"530d7f87-9f71-4d89-8520-ae77ca8a901d\") " pod="openstack/heat-cfnapi-86b55f6c-pznr6" Oct 05 08:28:14 crc kubenswrapper[4846]: I1005 08:28:14.776300 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bfv5r\" (UniqueName: \"kubernetes.io/projected/1bfe1edf-d89c-4d9e-b4fa-896c161c965f-kube-api-access-bfv5r\") pod \"heat-api-755575957b-g5bfm\" (UID: \"1bfe1edf-d89c-4d9e-b4fa-896c161c965f\") " pod="openstack/heat-api-755575957b-g5bfm" Oct 05 08:28:14 crc kubenswrapper[4846]: I1005 08:28:14.776348 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1bfe1edf-d89c-4d9e-b4fa-896c161c965f-combined-ca-bundle\") pod \"heat-api-755575957b-g5bfm\" (UID: \"1bfe1edf-d89c-4d9e-b4fa-896c161c965f\") " pod="openstack/heat-api-755575957b-g5bfm" Oct 05 08:28:14 crc kubenswrapper[4846]: I1005 08:28:14.776535 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/28512aa9-3230-44b3-a1f2-55247d128483-config-data\") pod \"heat-engine-54b8fb86bb-q8qps\" (UID: \"28512aa9-3230-44b3-a1f2-55247d128483\") " pod="openstack/heat-engine-54b8fb86bb-q8qps" Oct 05 08:28:14 crc kubenswrapper[4846]: I1005 08:28:14.776602 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1bfe1edf-d89c-4d9e-b4fa-896c161c965f-config-data\") pod \"heat-api-755575957b-g5bfm\" (UID: \"1bfe1edf-d89c-4d9e-b4fa-896c161c965f\") " pod="openstack/heat-api-755575957b-g5bfm" Oct 05 08:28:14 crc kubenswrapper[4846]: I1005 08:28:14.776636 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/530d7f87-9f71-4d89-8520-ae77ca8a901d-config-data\") pod \"heat-cfnapi-86b55f6c-pznr6\" (UID: \"530d7f87-9f71-4d89-8520-ae77ca8a901d\") " pod="openstack/heat-cfnapi-86b55f6c-pznr6" Oct 05 08:28:14 crc kubenswrapper[4846]: I1005 08:28:14.776673 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/530d7f87-9f71-4d89-8520-ae77ca8a901d-config-data-custom\") pod \"heat-cfnapi-86b55f6c-pznr6\" (UID: \"530d7f87-9f71-4d89-8520-ae77ca8a901d\") " pod="openstack/heat-cfnapi-86b55f6c-pznr6" Oct 05 08:28:14 crc kubenswrapper[4846]: I1005 08:28:14.776702 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1bfe1edf-d89c-4d9e-b4fa-896c161c965f-config-data-custom\") pod \"heat-api-755575957b-g5bfm\" (UID: \"1bfe1edf-d89c-4d9e-b4fa-896c161c965f\") " pod="openstack/heat-api-755575957b-g5bfm" Oct 05 08:28:14 crc kubenswrapper[4846]: I1005 08:28:14.776813 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/530d7f87-9f71-4d89-8520-ae77ca8a901d-combined-ca-bundle\") pod \"heat-cfnapi-86b55f6c-pznr6\" (UID: \"530d7f87-9f71-4d89-8520-ae77ca8a901d\") " pod="openstack/heat-cfnapi-86b55f6c-pznr6" Oct 05 08:28:14 crc kubenswrapper[4846]: I1005 08:28:14.776874 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6cz2d\" (UniqueName: \"kubernetes.io/projected/28512aa9-3230-44b3-a1f2-55247d128483-kube-api-access-6cz2d\") pod \"heat-engine-54b8fb86bb-q8qps\" (UID: \"28512aa9-3230-44b3-a1f2-55247d128483\") " pod="openstack/heat-engine-54b8fb86bb-q8qps" Oct 05 08:28:14 crc kubenswrapper[4846]: I1005 08:28:14.776909 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/28512aa9-3230-44b3-a1f2-55247d128483-config-data-custom\") pod \"heat-engine-54b8fb86bb-q8qps\" (UID: \"28512aa9-3230-44b3-a1f2-55247d128483\") " pod="openstack/heat-engine-54b8fb86bb-q8qps" Oct 05 08:28:14 crc kubenswrapper[4846]: I1005 08:28:14.776963 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/28512aa9-3230-44b3-a1f2-55247d128483-combined-ca-bundle\") pod \"heat-engine-54b8fb86bb-q8qps\" (UID: \"28512aa9-3230-44b3-a1f2-55247d128483\") " pod="openstack/heat-engine-54b8fb86bb-q8qps" Oct 05 08:28:14 crc kubenswrapper[4846]: I1005 08:28:14.783381 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/28512aa9-3230-44b3-a1f2-55247d128483-config-data\") pod \"heat-engine-54b8fb86bb-q8qps\" (UID: \"28512aa9-3230-44b3-a1f2-55247d128483\") " pod="openstack/heat-engine-54b8fb86bb-q8qps" Oct 05 08:28:14 crc kubenswrapper[4846]: I1005 08:28:14.787688 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1bfe1edf-d89c-4d9e-b4fa-896c161c965f-config-data-custom\") pod \"heat-api-755575957b-g5bfm\" (UID: \"1bfe1edf-d89c-4d9e-b4fa-896c161c965f\") " pod="openstack/heat-api-755575957b-g5bfm" Oct 05 08:28:14 crc kubenswrapper[4846]: I1005 08:28:14.789925 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/28512aa9-3230-44b3-a1f2-55247d128483-config-data-custom\") pod \"heat-engine-54b8fb86bb-q8qps\" (UID: \"28512aa9-3230-44b3-a1f2-55247d128483\") " pod="openstack/heat-engine-54b8fb86bb-q8qps" Oct 05 08:28:14 crc kubenswrapper[4846]: I1005 08:28:14.790820 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1bfe1edf-d89c-4d9e-b4fa-896c161c965f-combined-ca-bundle\") pod \"heat-api-755575957b-g5bfm\" (UID: \"1bfe1edf-d89c-4d9e-b4fa-896c161c965f\") " pod="openstack/heat-api-755575957b-g5bfm" Oct 05 08:28:14 crc kubenswrapper[4846]: I1005 08:28:14.791815 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1bfe1edf-d89c-4d9e-b4fa-896c161c965f-config-data\") pod \"heat-api-755575957b-g5bfm\" (UID: \"1bfe1edf-d89c-4d9e-b4fa-896c161c965f\") " pod="openstack/heat-api-755575957b-g5bfm" Oct 05 08:28:14 crc kubenswrapper[4846]: I1005 08:28:14.799348 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6cz2d\" (UniqueName: \"kubernetes.io/projected/28512aa9-3230-44b3-a1f2-55247d128483-kube-api-access-6cz2d\") pod \"heat-engine-54b8fb86bb-q8qps\" (UID: \"28512aa9-3230-44b3-a1f2-55247d128483\") " pod="openstack/heat-engine-54b8fb86bb-q8qps" Oct 05 08:28:14 crc kubenswrapper[4846]: I1005 08:28:14.804215 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/28512aa9-3230-44b3-a1f2-55247d128483-combined-ca-bundle\") pod \"heat-engine-54b8fb86bb-q8qps\" (UID: \"28512aa9-3230-44b3-a1f2-55247d128483\") " pod="openstack/heat-engine-54b8fb86bb-q8qps" Oct 05 08:28:14 crc kubenswrapper[4846]: I1005 08:28:14.808868 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bfv5r\" (UniqueName: \"kubernetes.io/projected/1bfe1edf-d89c-4d9e-b4fa-896c161c965f-kube-api-access-bfv5r\") pod \"heat-api-755575957b-g5bfm\" (UID: \"1bfe1edf-d89c-4d9e-b4fa-896c161c965f\") " pod="openstack/heat-api-755575957b-g5bfm" Oct 05 08:28:14 crc kubenswrapper[4846]: I1005 08:28:14.841020 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-54b8fb86bb-q8qps" Oct 05 08:28:14 crc kubenswrapper[4846]: I1005 08:28:14.872509 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-755575957b-g5bfm" Oct 05 08:28:14 crc kubenswrapper[4846]: I1005 08:28:14.878396 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ftf8x\" (UniqueName: \"kubernetes.io/projected/530d7f87-9f71-4d89-8520-ae77ca8a901d-kube-api-access-ftf8x\") pod \"heat-cfnapi-86b55f6c-pznr6\" (UID: \"530d7f87-9f71-4d89-8520-ae77ca8a901d\") " pod="openstack/heat-cfnapi-86b55f6c-pznr6" Oct 05 08:28:14 crc kubenswrapper[4846]: I1005 08:28:14.878523 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/530d7f87-9f71-4d89-8520-ae77ca8a901d-config-data\") pod \"heat-cfnapi-86b55f6c-pznr6\" (UID: \"530d7f87-9f71-4d89-8520-ae77ca8a901d\") " pod="openstack/heat-cfnapi-86b55f6c-pznr6" Oct 05 08:28:14 crc kubenswrapper[4846]: I1005 08:28:14.878556 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/530d7f87-9f71-4d89-8520-ae77ca8a901d-config-data-custom\") pod \"heat-cfnapi-86b55f6c-pznr6\" (UID: \"530d7f87-9f71-4d89-8520-ae77ca8a901d\") " pod="openstack/heat-cfnapi-86b55f6c-pznr6" Oct 05 08:28:14 crc kubenswrapper[4846]: I1005 08:28:14.878631 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/530d7f87-9f71-4d89-8520-ae77ca8a901d-combined-ca-bundle\") pod \"heat-cfnapi-86b55f6c-pznr6\" (UID: \"530d7f87-9f71-4d89-8520-ae77ca8a901d\") " pod="openstack/heat-cfnapi-86b55f6c-pznr6" Oct 05 08:28:14 crc kubenswrapper[4846]: I1005 08:28:14.884557 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/530d7f87-9f71-4d89-8520-ae77ca8a901d-combined-ca-bundle\") pod \"heat-cfnapi-86b55f6c-pznr6\" (UID: \"530d7f87-9f71-4d89-8520-ae77ca8a901d\") " pod="openstack/heat-cfnapi-86b55f6c-pznr6" Oct 05 08:28:14 crc kubenswrapper[4846]: I1005 08:28:14.887268 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/530d7f87-9f71-4d89-8520-ae77ca8a901d-config-data\") pod \"heat-cfnapi-86b55f6c-pznr6\" (UID: \"530d7f87-9f71-4d89-8520-ae77ca8a901d\") " pod="openstack/heat-cfnapi-86b55f6c-pznr6" Oct 05 08:28:14 crc kubenswrapper[4846]: I1005 08:28:14.889014 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/530d7f87-9f71-4d89-8520-ae77ca8a901d-config-data-custom\") pod \"heat-cfnapi-86b55f6c-pznr6\" (UID: \"530d7f87-9f71-4d89-8520-ae77ca8a901d\") " pod="openstack/heat-cfnapi-86b55f6c-pznr6" Oct 05 08:28:14 crc kubenswrapper[4846]: I1005 08:28:14.902441 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ftf8x\" (UniqueName: \"kubernetes.io/projected/530d7f87-9f71-4d89-8520-ae77ca8a901d-kube-api-access-ftf8x\") pod \"heat-cfnapi-86b55f6c-pznr6\" (UID: \"530d7f87-9f71-4d89-8520-ae77ca8a901d\") " pod="openstack/heat-cfnapi-86b55f6c-pznr6" Oct 05 08:28:14 crc kubenswrapper[4846]: I1005 08:28:14.944981 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-86b55f6c-pznr6" Oct 05 08:28:15 crc kubenswrapper[4846]: I1005 08:28:15.355745 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-engine-54b8fb86bb-q8qps"] Oct 05 08:28:15 crc kubenswrapper[4846]: W1005 08:28:15.360550 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod28512aa9_3230_44b3_a1f2_55247d128483.slice/crio-8dd5b8d92f3570ea9d572393da2f32fd3bf76eda095346a3439667a0e8eb1030 WatchSource:0}: Error finding container 8dd5b8d92f3570ea9d572393da2f32fd3bf76eda095346a3439667a0e8eb1030: Status 404 returned error can't find the container with id 8dd5b8d92f3570ea9d572393da2f32fd3bf76eda095346a3439667a0e8eb1030 Oct 05 08:28:15 crc kubenswrapper[4846]: I1005 08:28:15.430741 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-755575957b-g5bfm"] Oct 05 08:28:15 crc kubenswrapper[4846]: I1005 08:28:15.434802 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-54b8fb86bb-q8qps" event={"ID":"28512aa9-3230-44b3-a1f2-55247d128483","Type":"ContainerStarted","Data":"8dd5b8d92f3570ea9d572393da2f32fd3bf76eda095346a3439667a0e8eb1030"} Oct 05 08:28:15 crc kubenswrapper[4846]: I1005 08:28:15.538076 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-86b55f6c-pznr6"] Oct 05 08:28:15 crc kubenswrapper[4846]: W1005 08:28:15.569465 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod530d7f87_9f71_4d89_8520_ae77ca8a901d.slice/crio-34b986c098f4251ffba18a3428f624de40febcbe5eb6a874b7fd2a08c809554b WatchSource:0}: Error finding container 34b986c098f4251ffba18a3428f624de40febcbe5eb6a874b7fd2a08c809554b: Status 404 returned error can't find the container with id 34b986c098f4251ffba18a3428f624de40febcbe5eb6a874b7fd2a08c809554b Oct 05 08:28:15 crc kubenswrapper[4846]: I1005 08:28:15.701590 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-api-846d578b5-97wdg"] Oct 05 08:28:15 crc kubenswrapper[4846]: I1005 08:28:15.702380 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/heat-api-846d578b5-97wdg" podUID="2eaaa848-5b1c-402c-af31-23dc2d3eaef1" containerName="heat-api" containerID="cri-o://15f2420fd1410113564e93430e6af24b281cb5a1db66770570c42107ebadba74" gracePeriod=60 Oct 05 08:28:15 crc kubenswrapper[4846]: I1005 08:28:15.748065 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-cfnapi-547d5fb769-fllds"] Oct 05 08:28:15 crc kubenswrapper[4846]: I1005 08:28:15.748281 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/heat-cfnapi-547d5fb769-fllds" podUID="9fff0fb3-5365-420b-afc2-c997c3b98b6a" containerName="heat-cfnapi" containerID="cri-o://bf30db2f6b7764098249a02cd5401433aff89b58d2a0867c25f3e93fd461a127" gracePeriod=60 Oct 05 08:28:15 crc kubenswrapper[4846]: I1005 08:28:15.757807 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-api-69bf66cb7b-c4q2z"] Oct 05 08:28:15 crc kubenswrapper[4846]: I1005 08:28:15.759069 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-69bf66cb7b-c4q2z" Oct 05 08:28:15 crc kubenswrapper[4846]: I1005 08:28:15.761108 4846 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/heat-cfnapi-547d5fb769-fllds" podUID="9fff0fb3-5365-420b-afc2-c997c3b98b6a" containerName="heat-cfnapi" probeResult="failure" output="Get \"http://10.217.1.115:8000/healthcheck\": EOF" Oct 05 08:28:15 crc kubenswrapper[4846]: I1005 08:28:15.764246 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-heat-api-public-svc" Oct 05 08:28:15 crc kubenswrapper[4846]: I1005 08:28:15.771877 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-heat-api-internal-svc" Oct 05 08:28:15 crc kubenswrapper[4846]: I1005 08:28:15.775919 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-cfnapi-77cc459545-h5rt6"] Oct 05 08:28:15 crc kubenswrapper[4846]: I1005 08:28:15.777227 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-77cc459545-h5rt6" Oct 05 08:28:15 crc kubenswrapper[4846]: I1005 08:28:15.779693 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-heat-cfnapi-internal-svc" Oct 05 08:28:15 crc kubenswrapper[4846]: I1005 08:28:15.779830 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-heat-cfnapi-public-svc" Oct 05 08:28:15 crc kubenswrapper[4846]: I1005 08:28:15.805370 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-69bf66cb7b-c4q2z"] Oct 05 08:28:15 crc kubenswrapper[4846]: I1005 08:28:15.846460 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-77cc459545-h5rt6"] Oct 05 08:28:15 crc kubenswrapper[4846]: I1005 08:28:15.897728 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2db59028-72ef-4d73-84b1-d6bc6af749a5-combined-ca-bundle\") pod \"heat-api-69bf66cb7b-c4q2z\" (UID: \"2db59028-72ef-4d73-84b1-d6bc6af749a5\") " pod="openstack/heat-api-69bf66cb7b-c4q2z" Oct 05 08:28:15 crc kubenswrapper[4846]: I1005 08:28:15.897780 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xrpgm\" (UniqueName: \"kubernetes.io/projected/2db59028-72ef-4d73-84b1-d6bc6af749a5-kube-api-access-xrpgm\") pod \"heat-api-69bf66cb7b-c4q2z\" (UID: \"2db59028-72ef-4d73-84b1-d6bc6af749a5\") " pod="openstack/heat-api-69bf66cb7b-c4q2z" Oct 05 08:28:15 crc kubenswrapper[4846]: I1005 08:28:15.897807 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2db59028-72ef-4d73-84b1-d6bc6af749a5-config-data-custom\") pod \"heat-api-69bf66cb7b-c4q2z\" (UID: \"2db59028-72ef-4d73-84b1-d6bc6af749a5\") " pod="openstack/heat-api-69bf66cb7b-c4q2z" Oct 05 08:28:15 crc kubenswrapper[4846]: I1005 08:28:15.897852 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2db59028-72ef-4d73-84b1-d6bc6af749a5-internal-tls-certs\") pod \"heat-api-69bf66cb7b-c4q2z\" (UID: \"2db59028-72ef-4d73-84b1-d6bc6af749a5\") " pod="openstack/heat-api-69bf66cb7b-c4q2z" Oct 05 08:28:15 crc kubenswrapper[4846]: I1005 08:28:15.897875 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ed7a6ddf-cae5-48a3-931f-85c7287d77d8-public-tls-certs\") pod \"heat-cfnapi-77cc459545-h5rt6\" (UID: \"ed7a6ddf-cae5-48a3-931f-85c7287d77d8\") " pod="openstack/heat-cfnapi-77cc459545-h5rt6" Oct 05 08:28:15 crc kubenswrapper[4846]: I1005 08:28:15.897926 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2db59028-72ef-4d73-84b1-d6bc6af749a5-public-tls-certs\") pod \"heat-api-69bf66cb7b-c4q2z\" (UID: \"2db59028-72ef-4d73-84b1-d6bc6af749a5\") " pod="openstack/heat-api-69bf66cb7b-c4q2z" Oct 05 08:28:15 crc kubenswrapper[4846]: I1005 08:28:15.897943 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ed7a6ddf-cae5-48a3-931f-85c7287d77d8-config-data-custom\") pod \"heat-cfnapi-77cc459545-h5rt6\" (UID: \"ed7a6ddf-cae5-48a3-931f-85c7287d77d8\") " pod="openstack/heat-cfnapi-77cc459545-h5rt6" Oct 05 08:28:15 crc kubenswrapper[4846]: I1005 08:28:15.897975 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-glhtk\" (UniqueName: \"kubernetes.io/projected/ed7a6ddf-cae5-48a3-931f-85c7287d77d8-kube-api-access-glhtk\") pod \"heat-cfnapi-77cc459545-h5rt6\" (UID: \"ed7a6ddf-cae5-48a3-931f-85c7287d77d8\") " pod="openstack/heat-cfnapi-77cc459545-h5rt6" Oct 05 08:28:15 crc kubenswrapper[4846]: I1005 08:28:15.898006 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ed7a6ddf-cae5-48a3-931f-85c7287d77d8-config-data\") pod \"heat-cfnapi-77cc459545-h5rt6\" (UID: \"ed7a6ddf-cae5-48a3-931f-85c7287d77d8\") " pod="openstack/heat-cfnapi-77cc459545-h5rt6" Oct 05 08:28:15 crc kubenswrapper[4846]: I1005 08:28:15.898033 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ed7a6ddf-cae5-48a3-931f-85c7287d77d8-combined-ca-bundle\") pod \"heat-cfnapi-77cc459545-h5rt6\" (UID: \"ed7a6ddf-cae5-48a3-931f-85c7287d77d8\") " pod="openstack/heat-cfnapi-77cc459545-h5rt6" Oct 05 08:28:15 crc kubenswrapper[4846]: I1005 08:28:15.898058 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2db59028-72ef-4d73-84b1-d6bc6af749a5-config-data\") pod \"heat-api-69bf66cb7b-c4q2z\" (UID: \"2db59028-72ef-4d73-84b1-d6bc6af749a5\") " pod="openstack/heat-api-69bf66cb7b-c4q2z" Oct 05 08:28:15 crc kubenswrapper[4846]: I1005 08:28:15.898091 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ed7a6ddf-cae5-48a3-931f-85c7287d77d8-internal-tls-certs\") pod \"heat-cfnapi-77cc459545-h5rt6\" (UID: \"ed7a6ddf-cae5-48a3-931f-85c7287d77d8\") " pod="openstack/heat-cfnapi-77cc459545-h5rt6" Oct 05 08:28:15 crc kubenswrapper[4846]: I1005 08:28:15.999456 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2db59028-72ef-4d73-84b1-d6bc6af749a5-internal-tls-certs\") pod \"heat-api-69bf66cb7b-c4q2z\" (UID: \"2db59028-72ef-4d73-84b1-d6bc6af749a5\") " pod="openstack/heat-api-69bf66cb7b-c4q2z" Oct 05 08:28:15 crc kubenswrapper[4846]: I1005 08:28:15.999697 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ed7a6ddf-cae5-48a3-931f-85c7287d77d8-public-tls-certs\") pod \"heat-cfnapi-77cc459545-h5rt6\" (UID: \"ed7a6ddf-cae5-48a3-931f-85c7287d77d8\") " pod="openstack/heat-cfnapi-77cc459545-h5rt6" Oct 05 08:28:15 crc kubenswrapper[4846]: I1005 08:28:15.999752 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2db59028-72ef-4d73-84b1-d6bc6af749a5-public-tls-certs\") pod \"heat-api-69bf66cb7b-c4q2z\" (UID: \"2db59028-72ef-4d73-84b1-d6bc6af749a5\") " pod="openstack/heat-api-69bf66cb7b-c4q2z" Oct 05 08:28:16 crc kubenswrapper[4846]: I1005 08:28:15.999767 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ed7a6ddf-cae5-48a3-931f-85c7287d77d8-config-data-custom\") pod \"heat-cfnapi-77cc459545-h5rt6\" (UID: \"ed7a6ddf-cae5-48a3-931f-85c7287d77d8\") " pod="openstack/heat-cfnapi-77cc459545-h5rt6" Oct 05 08:28:16 crc kubenswrapper[4846]: I1005 08:28:15.999800 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-glhtk\" (UniqueName: \"kubernetes.io/projected/ed7a6ddf-cae5-48a3-931f-85c7287d77d8-kube-api-access-glhtk\") pod \"heat-cfnapi-77cc459545-h5rt6\" (UID: \"ed7a6ddf-cae5-48a3-931f-85c7287d77d8\") " pod="openstack/heat-cfnapi-77cc459545-h5rt6" Oct 05 08:28:16 crc kubenswrapper[4846]: I1005 08:28:15.999834 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ed7a6ddf-cae5-48a3-931f-85c7287d77d8-config-data\") pod \"heat-cfnapi-77cc459545-h5rt6\" (UID: \"ed7a6ddf-cae5-48a3-931f-85c7287d77d8\") " pod="openstack/heat-cfnapi-77cc459545-h5rt6" Oct 05 08:28:16 crc kubenswrapper[4846]: I1005 08:28:15.999854 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ed7a6ddf-cae5-48a3-931f-85c7287d77d8-combined-ca-bundle\") pod \"heat-cfnapi-77cc459545-h5rt6\" (UID: \"ed7a6ddf-cae5-48a3-931f-85c7287d77d8\") " pod="openstack/heat-cfnapi-77cc459545-h5rt6" Oct 05 08:28:16 crc kubenswrapper[4846]: I1005 08:28:15.999881 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2db59028-72ef-4d73-84b1-d6bc6af749a5-config-data\") pod \"heat-api-69bf66cb7b-c4q2z\" (UID: \"2db59028-72ef-4d73-84b1-d6bc6af749a5\") " pod="openstack/heat-api-69bf66cb7b-c4q2z" Oct 05 08:28:16 crc kubenswrapper[4846]: I1005 08:28:16.000001 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ed7a6ddf-cae5-48a3-931f-85c7287d77d8-internal-tls-certs\") pod \"heat-cfnapi-77cc459545-h5rt6\" (UID: \"ed7a6ddf-cae5-48a3-931f-85c7287d77d8\") " pod="openstack/heat-cfnapi-77cc459545-h5rt6" Oct 05 08:28:16 crc kubenswrapper[4846]: I1005 08:28:16.000271 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2db59028-72ef-4d73-84b1-d6bc6af749a5-combined-ca-bundle\") pod \"heat-api-69bf66cb7b-c4q2z\" (UID: \"2db59028-72ef-4d73-84b1-d6bc6af749a5\") " pod="openstack/heat-api-69bf66cb7b-c4q2z" Oct 05 08:28:16 crc kubenswrapper[4846]: I1005 08:28:16.000305 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xrpgm\" (UniqueName: \"kubernetes.io/projected/2db59028-72ef-4d73-84b1-d6bc6af749a5-kube-api-access-xrpgm\") pod \"heat-api-69bf66cb7b-c4q2z\" (UID: \"2db59028-72ef-4d73-84b1-d6bc6af749a5\") " pod="openstack/heat-api-69bf66cb7b-c4q2z" Oct 05 08:28:16 crc kubenswrapper[4846]: I1005 08:28:16.000331 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2db59028-72ef-4d73-84b1-d6bc6af749a5-config-data-custom\") pod \"heat-api-69bf66cb7b-c4q2z\" (UID: \"2db59028-72ef-4d73-84b1-d6bc6af749a5\") " pod="openstack/heat-api-69bf66cb7b-c4q2z" Oct 05 08:28:16 crc kubenswrapper[4846]: I1005 08:28:16.004689 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ed7a6ddf-cae5-48a3-931f-85c7287d77d8-config-data-custom\") pod \"heat-cfnapi-77cc459545-h5rt6\" (UID: \"ed7a6ddf-cae5-48a3-931f-85c7287d77d8\") " pod="openstack/heat-cfnapi-77cc459545-h5rt6" Oct 05 08:28:16 crc kubenswrapper[4846]: I1005 08:28:16.006338 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ed7a6ddf-cae5-48a3-931f-85c7287d77d8-combined-ca-bundle\") pod \"heat-cfnapi-77cc459545-h5rt6\" (UID: \"ed7a6ddf-cae5-48a3-931f-85c7287d77d8\") " pod="openstack/heat-cfnapi-77cc459545-h5rt6" Oct 05 08:28:16 crc kubenswrapper[4846]: I1005 08:28:16.006826 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2db59028-72ef-4d73-84b1-d6bc6af749a5-config-data-custom\") pod \"heat-api-69bf66cb7b-c4q2z\" (UID: \"2db59028-72ef-4d73-84b1-d6bc6af749a5\") " pod="openstack/heat-api-69bf66cb7b-c4q2z" Oct 05 08:28:16 crc kubenswrapper[4846]: I1005 08:28:16.007775 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2db59028-72ef-4d73-84b1-d6bc6af749a5-public-tls-certs\") pod \"heat-api-69bf66cb7b-c4q2z\" (UID: \"2db59028-72ef-4d73-84b1-d6bc6af749a5\") " pod="openstack/heat-api-69bf66cb7b-c4q2z" Oct 05 08:28:16 crc kubenswrapper[4846]: I1005 08:28:16.011224 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2db59028-72ef-4d73-84b1-d6bc6af749a5-config-data\") pod \"heat-api-69bf66cb7b-c4q2z\" (UID: \"2db59028-72ef-4d73-84b1-d6bc6af749a5\") " pod="openstack/heat-api-69bf66cb7b-c4q2z" Oct 05 08:28:16 crc kubenswrapper[4846]: I1005 08:28:16.015030 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2db59028-72ef-4d73-84b1-d6bc6af749a5-combined-ca-bundle\") pod \"heat-api-69bf66cb7b-c4q2z\" (UID: \"2db59028-72ef-4d73-84b1-d6bc6af749a5\") " pod="openstack/heat-api-69bf66cb7b-c4q2z" Oct 05 08:28:16 crc kubenswrapper[4846]: I1005 08:28:16.015995 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ed7a6ddf-cae5-48a3-931f-85c7287d77d8-internal-tls-certs\") pod \"heat-cfnapi-77cc459545-h5rt6\" (UID: \"ed7a6ddf-cae5-48a3-931f-85c7287d77d8\") " pod="openstack/heat-cfnapi-77cc459545-h5rt6" Oct 05 08:28:16 crc kubenswrapper[4846]: I1005 08:28:16.023860 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ed7a6ddf-cae5-48a3-931f-85c7287d77d8-config-data\") pod \"heat-cfnapi-77cc459545-h5rt6\" (UID: \"ed7a6ddf-cae5-48a3-931f-85c7287d77d8\") " pod="openstack/heat-cfnapi-77cc459545-h5rt6" Oct 05 08:28:16 crc kubenswrapper[4846]: I1005 08:28:16.024138 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ed7a6ddf-cae5-48a3-931f-85c7287d77d8-public-tls-certs\") pod \"heat-cfnapi-77cc459545-h5rt6\" (UID: \"ed7a6ddf-cae5-48a3-931f-85c7287d77d8\") " pod="openstack/heat-cfnapi-77cc459545-h5rt6" Oct 05 08:28:16 crc kubenswrapper[4846]: I1005 08:28:16.024172 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xrpgm\" (UniqueName: \"kubernetes.io/projected/2db59028-72ef-4d73-84b1-d6bc6af749a5-kube-api-access-xrpgm\") pod \"heat-api-69bf66cb7b-c4q2z\" (UID: \"2db59028-72ef-4d73-84b1-d6bc6af749a5\") " pod="openstack/heat-api-69bf66cb7b-c4q2z" Oct 05 08:28:16 crc kubenswrapper[4846]: I1005 08:28:16.024806 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-glhtk\" (UniqueName: \"kubernetes.io/projected/ed7a6ddf-cae5-48a3-931f-85c7287d77d8-kube-api-access-glhtk\") pod \"heat-cfnapi-77cc459545-h5rt6\" (UID: \"ed7a6ddf-cae5-48a3-931f-85c7287d77d8\") " pod="openstack/heat-cfnapi-77cc459545-h5rt6" Oct 05 08:28:16 crc kubenswrapper[4846]: I1005 08:28:16.029903 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2db59028-72ef-4d73-84b1-d6bc6af749a5-internal-tls-certs\") pod \"heat-api-69bf66cb7b-c4q2z\" (UID: \"2db59028-72ef-4d73-84b1-d6bc6af749a5\") " pod="openstack/heat-api-69bf66cb7b-c4q2z" Oct 05 08:28:16 crc kubenswrapper[4846]: I1005 08:28:16.075662 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-69bf66cb7b-c4q2z" Oct 05 08:28:16 crc kubenswrapper[4846]: I1005 08:28:16.205233 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-77cc459545-h5rt6" Oct 05 08:28:16 crc kubenswrapper[4846]: I1005 08:28:16.264358 4846 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-f8ddb658d-6wlz2" podUID="8329aafa-c3d7-474c-a31b-a56e0d23ddce" containerName="horizon" probeResult="failure" output="Get \"https://10.217.1.103:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.103:8443: connect: connection refused" Oct 05 08:28:16 crc kubenswrapper[4846]: I1005 08:28:16.445687 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-755575957b-g5bfm" event={"ID":"1bfe1edf-d89c-4d9e-b4fa-896c161c965f","Type":"ContainerStarted","Data":"09a155e42fc2f443417fb3b73124d87160a4548b0461d3474325d478c535190f"} Oct 05 08:28:16 crc kubenswrapper[4846]: I1005 08:28:16.449413 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-86b55f6c-pznr6" event={"ID":"530d7f87-9f71-4d89-8520-ae77ca8a901d","Type":"ContainerStarted","Data":"34b986c098f4251ffba18a3428f624de40febcbe5eb6a874b7fd2a08c809554b"} Oct 05 08:28:16 crc kubenswrapper[4846]: I1005 08:28:16.569280 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-69bf66cb7b-c4q2z"] Oct 05 08:28:16 crc kubenswrapper[4846]: W1005 08:28:16.573723 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2db59028_72ef_4d73_84b1_d6bc6af749a5.slice/crio-9ab074da63991e32427f825a4b6753da435cbea61b05d0bfea40c31e8785d208 WatchSource:0}: Error finding container 9ab074da63991e32427f825a4b6753da435cbea61b05d0bfea40c31e8785d208: Status 404 returned error can't find the container with id 9ab074da63991e32427f825a4b6753da435cbea61b05d0bfea40c31e8785d208 Oct 05 08:28:16 crc kubenswrapper[4846]: I1005 08:28:16.696107 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-77cc459545-h5rt6"] Oct 05 08:28:17 crc kubenswrapper[4846]: I1005 08:28:17.415696 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-api-846d578b5-97wdg" Oct 05 08:28:17 crc kubenswrapper[4846]: I1005 08:28:17.469870 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-54b8fb86bb-q8qps" event={"ID":"28512aa9-3230-44b3-a1f2-55247d128483","Type":"ContainerStarted","Data":"94dce17c256777bfe24c771f99a397baa677eecef09b776a0bad6efb3f14722d"} Oct 05 08:28:17 crc kubenswrapper[4846]: I1005 08:28:17.471268 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-engine-54b8fb86bb-q8qps" Oct 05 08:28:17 crc kubenswrapper[4846]: I1005 08:28:17.480364 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-77cc459545-h5rt6" event={"ID":"ed7a6ddf-cae5-48a3-931f-85c7287d77d8","Type":"ContainerStarted","Data":"6454fa24cf5ef92c879bd324b1edd9b00a033d4cb491f001f8925f8d35618dab"} Oct 05 08:28:17 crc kubenswrapper[4846]: I1005 08:28:17.480419 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-77cc459545-h5rt6" event={"ID":"ed7a6ddf-cae5-48a3-931f-85c7287d77d8","Type":"ContainerStarted","Data":"ec53cd3da18d7273769d45745c84bc1b957d58ffeabd0978709d62a4130dca3f"} Oct 05 08:28:17 crc kubenswrapper[4846]: I1005 08:28:17.481262 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-cfnapi-77cc459545-h5rt6" Oct 05 08:28:17 crc kubenswrapper[4846]: I1005 08:28:17.482759 4846 generic.go:334] "Generic (PLEG): container finished" podID="1bfe1edf-d89c-4d9e-b4fa-896c161c965f" containerID="a59d3e51fe6db4ac6ea0162a8379d91c95c939fdb291897ad8d2c18be9c247c2" exitCode=1 Oct 05 08:28:17 crc kubenswrapper[4846]: I1005 08:28:17.482815 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-755575957b-g5bfm" event={"ID":"1bfe1edf-d89c-4d9e-b4fa-896c161c965f","Type":"ContainerDied","Data":"a59d3e51fe6db4ac6ea0162a8379d91c95c939fdb291897ad8d2c18be9c247c2"} Oct 05 08:28:17 crc kubenswrapper[4846]: I1005 08:28:17.483119 4846 scope.go:117] "RemoveContainer" containerID="a59d3e51fe6db4ac6ea0162a8379d91c95c939fdb291897ad8d2c18be9c247c2" Oct 05 08:28:17 crc kubenswrapper[4846]: I1005 08:28:17.490590 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-engine-54b8fb86bb-q8qps" podStartSLOduration=3.490574275 podStartE2EDuration="3.490574275s" podCreationTimestamp="2025-10-05 08:28:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 08:28:17.488040888 +0000 UTC m=+6019.728893663" watchObservedRunningTime="2025-10-05 08:28:17.490574275 +0000 UTC m=+6019.731427050" Oct 05 08:28:17 crc kubenswrapper[4846]: I1005 08:28:17.495144 4846 generic.go:334] "Generic (PLEG): container finished" podID="530d7f87-9f71-4d89-8520-ae77ca8a901d" containerID="e1ae734829fae84b719a9406b39877ac1a6947ec0b62a04c0a906f327c988174" exitCode=1 Oct 05 08:28:17 crc kubenswrapper[4846]: I1005 08:28:17.495254 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-86b55f6c-pznr6" event={"ID":"530d7f87-9f71-4d89-8520-ae77ca8a901d","Type":"ContainerDied","Data":"e1ae734829fae84b719a9406b39877ac1a6947ec0b62a04c0a906f327c988174"} Oct 05 08:28:17 crc kubenswrapper[4846]: I1005 08:28:17.495856 4846 scope.go:117] "RemoveContainer" containerID="e1ae734829fae84b719a9406b39877ac1a6947ec0b62a04c0a906f327c988174" Oct 05 08:28:17 crc kubenswrapper[4846]: I1005 08:28:17.500426 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-69bf66cb7b-c4q2z" event={"ID":"2db59028-72ef-4d73-84b1-d6bc6af749a5","Type":"ContainerStarted","Data":"80d8f210da4d8c4f9ca407f8526380eba67ed20e56fd60af7f51aefcf88da00a"} Oct 05 08:28:17 crc kubenswrapper[4846]: I1005 08:28:17.500460 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-69bf66cb7b-c4q2z" event={"ID":"2db59028-72ef-4d73-84b1-d6bc6af749a5","Type":"ContainerStarted","Data":"9ab074da63991e32427f825a4b6753da435cbea61b05d0bfea40c31e8785d208"} Oct 05 08:28:17 crc kubenswrapper[4846]: I1005 08:28:17.500606 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-api-69bf66cb7b-c4q2z" Oct 05 08:28:17 crc kubenswrapper[4846]: I1005 08:28:17.506590 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-cfnapi-77cc459545-h5rt6" podStartSLOduration=2.5065755899999997 podStartE2EDuration="2.50657559s" podCreationTimestamp="2025-10-05 08:28:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 08:28:17.504782562 +0000 UTC m=+6019.745635357" watchObservedRunningTime="2025-10-05 08:28:17.50657559 +0000 UTC m=+6019.747428365" Oct 05 08:28:17 crc kubenswrapper[4846]: I1005 08:28:17.573379 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-api-69bf66cb7b-c4q2z" podStartSLOduration=2.573361403 podStartE2EDuration="2.573361403s" podCreationTimestamp="2025-10-05 08:28:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 08:28:17.569367877 +0000 UTC m=+6019.810220652" watchObservedRunningTime="2025-10-05 08:28:17.573361403 +0000 UTC m=+6019.814214178" Oct 05 08:28:18 crc kubenswrapper[4846]: I1005 08:28:18.512586 4846 generic.go:334] "Generic (PLEG): container finished" podID="1bfe1edf-d89c-4d9e-b4fa-896c161c965f" containerID="58966a1871f676fc0f8cf0db57eddd59ffc87a81cbc865e4737a18455c414414" exitCode=1 Oct 05 08:28:18 crc kubenswrapper[4846]: I1005 08:28:18.512847 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-755575957b-g5bfm" event={"ID":"1bfe1edf-d89c-4d9e-b4fa-896c161c965f","Type":"ContainerDied","Data":"58966a1871f676fc0f8cf0db57eddd59ffc87a81cbc865e4737a18455c414414"} Oct 05 08:28:18 crc kubenswrapper[4846]: I1005 08:28:18.512876 4846 scope.go:117] "RemoveContainer" containerID="a59d3e51fe6db4ac6ea0162a8379d91c95c939fdb291897ad8d2c18be9c247c2" Oct 05 08:28:18 crc kubenswrapper[4846]: I1005 08:28:18.513502 4846 scope.go:117] "RemoveContainer" containerID="58966a1871f676fc0f8cf0db57eddd59ffc87a81cbc865e4737a18455c414414" Oct 05 08:28:18 crc kubenswrapper[4846]: E1005 08:28:18.513712 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-api\" with CrashLoopBackOff: \"back-off 10s restarting failed container=heat-api pod=heat-api-755575957b-g5bfm_openstack(1bfe1edf-d89c-4d9e-b4fa-896c161c965f)\"" pod="openstack/heat-api-755575957b-g5bfm" podUID="1bfe1edf-d89c-4d9e-b4fa-896c161c965f" Oct 05 08:28:18 crc kubenswrapper[4846]: I1005 08:28:18.518115 4846 generic.go:334] "Generic (PLEG): container finished" podID="530d7f87-9f71-4d89-8520-ae77ca8a901d" containerID="f3ae09967d484aad4f24024e136b94726a3970ba09d86410546e022ead58d468" exitCode=1 Oct 05 08:28:18 crc kubenswrapper[4846]: I1005 08:28:18.518163 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-86b55f6c-pznr6" event={"ID":"530d7f87-9f71-4d89-8520-ae77ca8a901d","Type":"ContainerDied","Data":"f3ae09967d484aad4f24024e136b94726a3970ba09d86410546e022ead58d468"} Oct 05 08:28:18 crc kubenswrapper[4846]: I1005 08:28:18.518813 4846 scope.go:117] "RemoveContainer" containerID="f3ae09967d484aad4f24024e136b94726a3970ba09d86410546e022ead58d468" Oct 05 08:28:18 crc kubenswrapper[4846]: E1005 08:28:18.519157 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-cfnapi\" with CrashLoopBackOff: \"back-off 10s restarting failed container=heat-cfnapi pod=heat-cfnapi-86b55f6c-pznr6_openstack(530d7f87-9f71-4d89-8520-ae77ca8a901d)\"" pod="openstack/heat-cfnapi-86b55f6c-pznr6" podUID="530d7f87-9f71-4d89-8520-ae77ca8a901d" Oct 05 08:28:18 crc kubenswrapper[4846]: I1005 08:28:18.572872 4846 scope.go:117] "RemoveContainer" containerID="e1ae734829fae84b719a9406b39877ac1a6947ec0b62a04c0a906f327c988174" Oct 05 08:28:19 crc kubenswrapper[4846]: I1005 08:28:19.530296 4846 scope.go:117] "RemoveContainer" containerID="58966a1871f676fc0f8cf0db57eddd59ffc87a81cbc865e4737a18455c414414" Oct 05 08:28:19 crc kubenswrapper[4846]: E1005 08:28:19.530872 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-api\" with CrashLoopBackOff: \"back-off 10s restarting failed container=heat-api pod=heat-api-755575957b-g5bfm_openstack(1bfe1edf-d89c-4d9e-b4fa-896c161c965f)\"" pod="openstack/heat-api-755575957b-g5bfm" podUID="1bfe1edf-d89c-4d9e-b4fa-896c161c965f" Oct 05 08:28:19 crc kubenswrapper[4846]: I1005 08:28:19.534637 4846 scope.go:117] "RemoveContainer" containerID="f3ae09967d484aad4f24024e136b94726a3970ba09d86410546e022ead58d468" Oct 05 08:28:19 crc kubenswrapper[4846]: E1005 08:28:19.535050 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-cfnapi\" with CrashLoopBackOff: \"back-off 10s restarting failed container=heat-cfnapi pod=heat-cfnapi-86b55f6c-pznr6_openstack(530d7f87-9f71-4d89-8520-ae77ca8a901d)\"" pod="openstack/heat-cfnapi-86b55f6c-pznr6" podUID="530d7f87-9f71-4d89-8520-ae77ca8a901d" Oct 05 08:28:19 crc kubenswrapper[4846]: I1005 08:28:19.872714 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-api-755575957b-g5bfm" Oct 05 08:28:19 crc kubenswrapper[4846]: I1005 08:28:19.872783 4846 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack/heat-api-755575957b-g5bfm" Oct 05 08:28:19 crc kubenswrapper[4846]: I1005 08:28:19.946359 4846 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack/heat-cfnapi-86b55f6c-pznr6" Oct 05 08:28:19 crc kubenswrapper[4846]: I1005 08:28:19.946411 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-cfnapi-86b55f6c-pznr6" Oct 05 08:28:20 crc kubenswrapper[4846]: I1005 08:28:20.544070 4846 scope.go:117] "RemoveContainer" containerID="58966a1871f676fc0f8cf0db57eddd59ffc87a81cbc865e4737a18455c414414" Oct 05 08:28:20 crc kubenswrapper[4846]: E1005 08:28:20.544328 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-api\" with CrashLoopBackOff: \"back-off 10s restarting failed container=heat-api pod=heat-api-755575957b-g5bfm_openstack(1bfe1edf-d89c-4d9e-b4fa-896c161c965f)\"" pod="openstack/heat-api-755575957b-g5bfm" podUID="1bfe1edf-d89c-4d9e-b4fa-896c161c965f" Oct 05 08:28:20 crc kubenswrapper[4846]: I1005 08:28:20.544343 4846 scope.go:117] "RemoveContainer" containerID="f3ae09967d484aad4f24024e136b94726a3970ba09d86410546e022ead58d468" Oct 05 08:28:20 crc kubenswrapper[4846]: E1005 08:28:20.544634 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-cfnapi\" with CrashLoopBackOff: \"back-off 10s restarting failed container=heat-cfnapi pod=heat-cfnapi-86b55f6c-pznr6_openstack(530d7f87-9f71-4d89-8520-ae77ca8a901d)\"" pod="openstack/heat-cfnapi-86b55f6c-pznr6" podUID="530d7f87-9f71-4d89-8520-ae77ca8a901d" Oct 05 08:28:21 crc kubenswrapper[4846]: I1005 08:28:21.134166 4846 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/heat-api-846d578b5-97wdg" podUID="2eaaa848-5b1c-402c-af31-23dc2d3eaef1" containerName="heat-api" probeResult="failure" output="Get \"http://10.217.1.114:8004/healthcheck\": read tcp 10.217.0.2:60504->10.217.1.114:8004: read: connection reset by peer" Oct 05 08:28:21 crc kubenswrapper[4846]: I1005 08:28:21.159588 4846 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/heat-cfnapi-547d5fb769-fllds" podUID="9fff0fb3-5365-420b-afc2-c997c3b98b6a" containerName="heat-cfnapi" probeResult="failure" output="Get \"http://10.217.1.115:8000/healthcheck\": read tcp 10.217.0.2:56270->10.217.1.115:8000: read: connection reset by peer" Oct 05 08:28:21 crc kubenswrapper[4846]: I1005 08:28:21.554257 4846 generic.go:334] "Generic (PLEG): container finished" podID="2eaaa848-5b1c-402c-af31-23dc2d3eaef1" containerID="15f2420fd1410113564e93430e6af24b281cb5a1db66770570c42107ebadba74" exitCode=0 Oct 05 08:28:21 crc kubenswrapper[4846]: I1005 08:28:21.554835 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-846d578b5-97wdg" event={"ID":"2eaaa848-5b1c-402c-af31-23dc2d3eaef1","Type":"ContainerDied","Data":"15f2420fd1410113564e93430e6af24b281cb5a1db66770570c42107ebadba74"} Oct 05 08:28:22 crc kubenswrapper[4846]: I1005 08:28:21.556746 4846 generic.go:334] "Generic (PLEG): container finished" podID="9fff0fb3-5365-420b-afc2-c997c3b98b6a" containerID="bf30db2f6b7764098249a02cd5401433aff89b58d2a0867c25f3e93fd461a127" exitCode=0 Oct 05 08:28:22 crc kubenswrapper[4846]: I1005 08:28:21.557243 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-547d5fb769-fllds" event={"ID":"9fff0fb3-5365-420b-afc2-c997c3b98b6a","Type":"ContainerDied","Data":"bf30db2f6b7764098249a02cd5401433aff89b58d2a0867c25f3e93fd461a127"} Oct 05 08:28:22 crc kubenswrapper[4846]: I1005 08:28:21.558060 4846 scope.go:117] "RemoveContainer" containerID="58966a1871f676fc0f8cf0db57eddd59ffc87a81cbc865e4737a18455c414414" Oct 05 08:28:22 crc kubenswrapper[4846]: E1005 08:28:21.558545 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-api\" with CrashLoopBackOff: \"back-off 10s restarting failed container=heat-api pod=heat-api-755575957b-g5bfm_openstack(1bfe1edf-d89c-4d9e-b4fa-896c161c965f)\"" pod="openstack/heat-api-755575957b-g5bfm" podUID="1bfe1edf-d89c-4d9e-b4fa-896c161c965f" Oct 05 08:28:22 crc kubenswrapper[4846]: I1005 08:28:21.763452 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-846d578b5-97wdg" Oct 05 08:28:22 crc kubenswrapper[4846]: I1005 08:28:21.773980 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-547d5fb769-fllds" Oct 05 08:28:22 crc kubenswrapper[4846]: I1005 08:28:21.816986 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9fff0fb3-5365-420b-afc2-c997c3b98b6a-combined-ca-bundle\") pod \"9fff0fb3-5365-420b-afc2-c997c3b98b6a\" (UID: \"9fff0fb3-5365-420b-afc2-c997c3b98b6a\") " Oct 05 08:28:22 crc kubenswrapper[4846]: I1005 08:28:21.817030 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2eaaa848-5b1c-402c-af31-23dc2d3eaef1-combined-ca-bundle\") pod \"2eaaa848-5b1c-402c-af31-23dc2d3eaef1\" (UID: \"2eaaa848-5b1c-402c-af31-23dc2d3eaef1\") " Oct 05 08:28:22 crc kubenswrapper[4846]: I1005 08:28:21.817076 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9fff0fb3-5365-420b-afc2-c997c3b98b6a-config-data-custom\") pod \"9fff0fb3-5365-420b-afc2-c997c3b98b6a\" (UID: \"9fff0fb3-5365-420b-afc2-c997c3b98b6a\") " Oct 05 08:28:22 crc kubenswrapper[4846]: I1005 08:28:21.817137 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9fff0fb3-5365-420b-afc2-c997c3b98b6a-config-data\") pod \"9fff0fb3-5365-420b-afc2-c997c3b98b6a\" (UID: \"9fff0fb3-5365-420b-afc2-c997c3b98b6a\") " Oct 05 08:28:22 crc kubenswrapper[4846]: I1005 08:28:21.817197 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9lq27\" (UniqueName: \"kubernetes.io/projected/9fff0fb3-5365-420b-afc2-c997c3b98b6a-kube-api-access-9lq27\") pod \"9fff0fb3-5365-420b-afc2-c997c3b98b6a\" (UID: \"9fff0fb3-5365-420b-afc2-c997c3b98b6a\") " Oct 05 08:28:22 crc kubenswrapper[4846]: I1005 08:28:21.817304 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p5m2r\" (UniqueName: \"kubernetes.io/projected/2eaaa848-5b1c-402c-af31-23dc2d3eaef1-kube-api-access-p5m2r\") pod \"2eaaa848-5b1c-402c-af31-23dc2d3eaef1\" (UID: \"2eaaa848-5b1c-402c-af31-23dc2d3eaef1\") " Oct 05 08:28:22 crc kubenswrapper[4846]: I1005 08:28:21.817349 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2eaaa848-5b1c-402c-af31-23dc2d3eaef1-config-data-custom\") pod \"2eaaa848-5b1c-402c-af31-23dc2d3eaef1\" (UID: \"2eaaa848-5b1c-402c-af31-23dc2d3eaef1\") " Oct 05 08:28:22 crc kubenswrapper[4846]: I1005 08:28:21.817376 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2eaaa848-5b1c-402c-af31-23dc2d3eaef1-config-data\") pod \"2eaaa848-5b1c-402c-af31-23dc2d3eaef1\" (UID: \"2eaaa848-5b1c-402c-af31-23dc2d3eaef1\") " Oct 05 08:28:22 crc kubenswrapper[4846]: I1005 08:28:21.821936 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9fff0fb3-5365-420b-afc2-c997c3b98b6a-kube-api-access-9lq27" (OuterVolumeSpecName: "kube-api-access-9lq27") pod "9fff0fb3-5365-420b-afc2-c997c3b98b6a" (UID: "9fff0fb3-5365-420b-afc2-c997c3b98b6a"). InnerVolumeSpecName "kube-api-access-9lq27". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:28:22 crc kubenswrapper[4846]: I1005 08:28:21.822968 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2eaaa848-5b1c-402c-af31-23dc2d3eaef1-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "2eaaa848-5b1c-402c-af31-23dc2d3eaef1" (UID: "2eaaa848-5b1c-402c-af31-23dc2d3eaef1"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:28:22 crc kubenswrapper[4846]: I1005 08:28:21.824101 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9fff0fb3-5365-420b-afc2-c997c3b98b6a-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "9fff0fb3-5365-420b-afc2-c997c3b98b6a" (UID: "9fff0fb3-5365-420b-afc2-c997c3b98b6a"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:28:22 crc kubenswrapper[4846]: I1005 08:28:21.842747 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2eaaa848-5b1c-402c-af31-23dc2d3eaef1-kube-api-access-p5m2r" (OuterVolumeSpecName: "kube-api-access-p5m2r") pod "2eaaa848-5b1c-402c-af31-23dc2d3eaef1" (UID: "2eaaa848-5b1c-402c-af31-23dc2d3eaef1"). InnerVolumeSpecName "kube-api-access-p5m2r". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:28:22 crc kubenswrapper[4846]: I1005 08:28:21.846279 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2eaaa848-5b1c-402c-af31-23dc2d3eaef1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2eaaa848-5b1c-402c-af31-23dc2d3eaef1" (UID: "2eaaa848-5b1c-402c-af31-23dc2d3eaef1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:28:22 crc kubenswrapper[4846]: I1005 08:28:21.874008 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9fff0fb3-5365-420b-afc2-c997c3b98b6a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9fff0fb3-5365-420b-afc2-c997c3b98b6a" (UID: "9fff0fb3-5365-420b-afc2-c997c3b98b6a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:28:22 crc kubenswrapper[4846]: I1005 08:28:21.874829 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2eaaa848-5b1c-402c-af31-23dc2d3eaef1-config-data" (OuterVolumeSpecName: "config-data") pod "2eaaa848-5b1c-402c-af31-23dc2d3eaef1" (UID: "2eaaa848-5b1c-402c-af31-23dc2d3eaef1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:28:22 crc kubenswrapper[4846]: I1005 08:28:21.876585 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9fff0fb3-5365-420b-afc2-c997c3b98b6a-config-data" (OuterVolumeSpecName: "config-data") pod "9fff0fb3-5365-420b-afc2-c997c3b98b6a" (UID: "9fff0fb3-5365-420b-afc2-c997c3b98b6a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:28:22 crc kubenswrapper[4846]: I1005 08:28:21.919027 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p5m2r\" (UniqueName: \"kubernetes.io/projected/2eaaa848-5b1c-402c-af31-23dc2d3eaef1-kube-api-access-p5m2r\") on node \"crc\" DevicePath \"\"" Oct 05 08:28:22 crc kubenswrapper[4846]: I1005 08:28:21.919050 4846 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2eaaa848-5b1c-402c-af31-23dc2d3eaef1-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 05 08:28:22 crc kubenswrapper[4846]: I1005 08:28:21.919059 4846 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2eaaa848-5b1c-402c-af31-23dc2d3eaef1-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 08:28:22 crc kubenswrapper[4846]: I1005 08:28:21.919067 4846 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9fff0fb3-5365-420b-afc2-c997c3b98b6a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 08:28:22 crc kubenswrapper[4846]: I1005 08:28:21.919075 4846 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2eaaa848-5b1c-402c-af31-23dc2d3eaef1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 08:28:22 crc kubenswrapper[4846]: I1005 08:28:21.919083 4846 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9fff0fb3-5365-420b-afc2-c997c3b98b6a-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 05 08:28:22 crc kubenswrapper[4846]: I1005 08:28:21.919091 4846 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9fff0fb3-5365-420b-afc2-c997c3b98b6a-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 08:28:22 crc kubenswrapper[4846]: I1005 08:28:21.919099 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9lq27\" (UniqueName: \"kubernetes.io/projected/9fff0fb3-5365-420b-afc2-c997c3b98b6a-kube-api-access-9lq27\") on node \"crc\" DevicePath \"\"" Oct 05 08:28:22 crc kubenswrapper[4846]: I1005 08:28:22.474847 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-api-69bf66cb7b-c4q2z" Oct 05 08:28:22 crc kubenswrapper[4846]: I1005 08:28:22.513519 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-cfnapi-77cc459545-h5rt6" Oct 05 08:28:22 crc kubenswrapper[4846]: I1005 08:28:22.530428 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-api-755575957b-g5bfm"] Oct 05 08:28:22 crc kubenswrapper[4846]: I1005 08:28:22.578779 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-846d578b5-97wdg" Oct 05 08:28:22 crc kubenswrapper[4846]: I1005 08:28:22.586046 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-846d578b5-97wdg" event={"ID":"2eaaa848-5b1c-402c-af31-23dc2d3eaef1","Type":"ContainerDied","Data":"528c59b17ff7fdfd24529a3fba5f0444086a5eaa5bdd27ed995cc60acce68d1e"} Oct 05 08:28:22 crc kubenswrapper[4846]: I1005 08:28:22.586121 4846 scope.go:117] "RemoveContainer" containerID="15f2420fd1410113564e93430e6af24b281cb5a1db66770570c42107ebadba74" Oct 05 08:28:22 crc kubenswrapper[4846]: I1005 08:28:22.594254 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-547d5fb769-fllds" Oct 05 08:28:22 crc kubenswrapper[4846]: I1005 08:28:22.594339 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-547d5fb769-fllds" event={"ID":"9fff0fb3-5365-420b-afc2-c997c3b98b6a","Type":"ContainerDied","Data":"bc5607bc256d552d4767b57e06574d82ff727c0b4aca1bdd885c10c902f0acd2"} Oct 05 08:28:22 crc kubenswrapper[4846]: I1005 08:28:22.656921 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-cfnapi-86b55f6c-pznr6"] Oct 05 08:28:22 crc kubenswrapper[4846]: I1005 08:28:22.681114 4846 scope.go:117] "RemoveContainer" containerID="bf30db2f6b7764098249a02cd5401433aff89b58d2a0867c25f3e93fd461a127" Oct 05 08:28:22 crc kubenswrapper[4846]: I1005 08:28:22.711948 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-api-846d578b5-97wdg"] Oct 05 08:28:22 crc kubenswrapper[4846]: I1005 08:28:22.721051 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-api-846d578b5-97wdg"] Oct 05 08:28:22 crc kubenswrapper[4846]: I1005 08:28:22.730454 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-cfnapi-547d5fb769-fllds"] Oct 05 08:28:22 crc kubenswrapper[4846]: I1005 08:28:22.741206 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-cfnapi-547d5fb769-fllds"] Oct 05 08:28:23 crc kubenswrapper[4846]: I1005 08:28:23.147762 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-755575957b-g5bfm" Oct 05 08:28:23 crc kubenswrapper[4846]: I1005 08:28:23.153964 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-86b55f6c-pznr6" Oct 05 08:28:23 crc kubenswrapper[4846]: I1005 08:28:23.253866 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1bfe1edf-d89c-4d9e-b4fa-896c161c965f-config-data\") pod \"1bfe1edf-d89c-4d9e-b4fa-896c161c965f\" (UID: \"1bfe1edf-d89c-4d9e-b4fa-896c161c965f\") " Oct 05 08:28:23 crc kubenswrapper[4846]: I1005 08:28:23.253963 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1bfe1edf-d89c-4d9e-b4fa-896c161c965f-config-data-custom\") pod \"1bfe1edf-d89c-4d9e-b4fa-896c161c965f\" (UID: \"1bfe1edf-d89c-4d9e-b4fa-896c161c965f\") " Oct 05 08:28:23 crc kubenswrapper[4846]: I1005 08:28:23.254041 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ftf8x\" (UniqueName: \"kubernetes.io/projected/530d7f87-9f71-4d89-8520-ae77ca8a901d-kube-api-access-ftf8x\") pod \"530d7f87-9f71-4d89-8520-ae77ca8a901d\" (UID: \"530d7f87-9f71-4d89-8520-ae77ca8a901d\") " Oct 05 08:28:23 crc kubenswrapper[4846]: I1005 08:28:23.254230 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/530d7f87-9f71-4d89-8520-ae77ca8a901d-config-data\") pod \"530d7f87-9f71-4d89-8520-ae77ca8a901d\" (UID: \"530d7f87-9f71-4d89-8520-ae77ca8a901d\") " Oct 05 08:28:23 crc kubenswrapper[4846]: I1005 08:28:23.254431 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/530d7f87-9f71-4d89-8520-ae77ca8a901d-config-data-custom\") pod \"530d7f87-9f71-4d89-8520-ae77ca8a901d\" (UID: \"530d7f87-9f71-4d89-8520-ae77ca8a901d\") " Oct 05 08:28:23 crc kubenswrapper[4846]: I1005 08:28:23.254513 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1bfe1edf-d89c-4d9e-b4fa-896c161c965f-combined-ca-bundle\") pod \"1bfe1edf-d89c-4d9e-b4fa-896c161c965f\" (UID: \"1bfe1edf-d89c-4d9e-b4fa-896c161c965f\") " Oct 05 08:28:23 crc kubenswrapper[4846]: I1005 08:28:23.254560 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bfv5r\" (UniqueName: \"kubernetes.io/projected/1bfe1edf-d89c-4d9e-b4fa-896c161c965f-kube-api-access-bfv5r\") pod \"1bfe1edf-d89c-4d9e-b4fa-896c161c965f\" (UID: \"1bfe1edf-d89c-4d9e-b4fa-896c161c965f\") " Oct 05 08:28:23 crc kubenswrapper[4846]: I1005 08:28:23.254591 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/530d7f87-9f71-4d89-8520-ae77ca8a901d-combined-ca-bundle\") pod \"530d7f87-9f71-4d89-8520-ae77ca8a901d\" (UID: \"530d7f87-9f71-4d89-8520-ae77ca8a901d\") " Oct 05 08:28:23 crc kubenswrapper[4846]: I1005 08:28:23.262046 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bfe1edf-d89c-4d9e-b4fa-896c161c965f-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "1bfe1edf-d89c-4d9e-b4fa-896c161c965f" (UID: "1bfe1edf-d89c-4d9e-b4fa-896c161c965f"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:28:23 crc kubenswrapper[4846]: I1005 08:28:23.262594 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bfe1edf-d89c-4d9e-b4fa-896c161c965f-kube-api-access-bfv5r" (OuterVolumeSpecName: "kube-api-access-bfv5r") pod "1bfe1edf-d89c-4d9e-b4fa-896c161c965f" (UID: "1bfe1edf-d89c-4d9e-b4fa-896c161c965f"). InnerVolumeSpecName "kube-api-access-bfv5r". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:28:23 crc kubenswrapper[4846]: I1005 08:28:23.263141 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/530d7f87-9f71-4d89-8520-ae77ca8a901d-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "530d7f87-9f71-4d89-8520-ae77ca8a901d" (UID: "530d7f87-9f71-4d89-8520-ae77ca8a901d"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:28:23 crc kubenswrapper[4846]: I1005 08:28:23.263660 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/530d7f87-9f71-4d89-8520-ae77ca8a901d-kube-api-access-ftf8x" (OuterVolumeSpecName: "kube-api-access-ftf8x") pod "530d7f87-9f71-4d89-8520-ae77ca8a901d" (UID: "530d7f87-9f71-4d89-8520-ae77ca8a901d"). InnerVolumeSpecName "kube-api-access-ftf8x". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:28:23 crc kubenswrapper[4846]: I1005 08:28:23.291228 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bfe1edf-d89c-4d9e-b4fa-896c161c965f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1bfe1edf-d89c-4d9e-b4fa-896c161c965f" (UID: "1bfe1edf-d89c-4d9e-b4fa-896c161c965f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:28:23 crc kubenswrapper[4846]: I1005 08:28:23.303150 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/530d7f87-9f71-4d89-8520-ae77ca8a901d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "530d7f87-9f71-4d89-8520-ae77ca8a901d" (UID: "530d7f87-9f71-4d89-8520-ae77ca8a901d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:28:23 crc kubenswrapper[4846]: I1005 08:28:23.323098 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/530d7f87-9f71-4d89-8520-ae77ca8a901d-config-data" (OuterVolumeSpecName: "config-data") pod "530d7f87-9f71-4d89-8520-ae77ca8a901d" (UID: "530d7f87-9f71-4d89-8520-ae77ca8a901d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:28:23 crc kubenswrapper[4846]: I1005 08:28:23.331005 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bfe1edf-d89c-4d9e-b4fa-896c161c965f-config-data" (OuterVolumeSpecName: "config-data") pod "1bfe1edf-d89c-4d9e-b4fa-896c161c965f" (UID: "1bfe1edf-d89c-4d9e-b4fa-896c161c965f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:28:23 crc kubenswrapper[4846]: I1005 08:28:23.356620 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ftf8x\" (UniqueName: \"kubernetes.io/projected/530d7f87-9f71-4d89-8520-ae77ca8a901d-kube-api-access-ftf8x\") on node \"crc\" DevicePath \"\"" Oct 05 08:28:23 crc kubenswrapper[4846]: I1005 08:28:23.356658 4846 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/530d7f87-9f71-4d89-8520-ae77ca8a901d-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 08:28:23 crc kubenswrapper[4846]: I1005 08:28:23.356671 4846 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/530d7f87-9f71-4d89-8520-ae77ca8a901d-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 05 08:28:23 crc kubenswrapper[4846]: I1005 08:28:23.356683 4846 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1bfe1edf-d89c-4d9e-b4fa-896c161c965f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 08:28:23 crc kubenswrapper[4846]: I1005 08:28:23.356698 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bfv5r\" (UniqueName: \"kubernetes.io/projected/1bfe1edf-d89c-4d9e-b4fa-896c161c965f-kube-api-access-bfv5r\") on node \"crc\" DevicePath \"\"" Oct 05 08:28:23 crc kubenswrapper[4846]: I1005 08:28:23.356710 4846 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/530d7f87-9f71-4d89-8520-ae77ca8a901d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 08:28:23 crc kubenswrapper[4846]: I1005 08:28:23.356723 4846 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1bfe1edf-d89c-4d9e-b4fa-896c161c965f-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 08:28:23 crc kubenswrapper[4846]: I1005 08:28:23.356735 4846 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1bfe1edf-d89c-4d9e-b4fa-896c161c965f-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 05 08:28:23 crc kubenswrapper[4846]: I1005 08:28:23.605864 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-86b55f6c-pznr6" event={"ID":"530d7f87-9f71-4d89-8520-ae77ca8a901d","Type":"ContainerDied","Data":"34b986c098f4251ffba18a3428f624de40febcbe5eb6a874b7fd2a08c809554b"} Oct 05 08:28:23 crc kubenswrapper[4846]: I1005 08:28:23.605933 4846 scope.go:117] "RemoveContainer" containerID="f3ae09967d484aad4f24024e136b94726a3970ba09d86410546e022ead58d468" Oct 05 08:28:23 crc kubenswrapper[4846]: I1005 08:28:23.605927 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-86b55f6c-pznr6" Oct 05 08:28:23 crc kubenswrapper[4846]: I1005 08:28:23.613234 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-755575957b-g5bfm" event={"ID":"1bfe1edf-d89c-4d9e-b4fa-896c161c965f","Type":"ContainerDied","Data":"09a155e42fc2f443417fb3b73124d87160a4548b0461d3474325d478c535190f"} Oct 05 08:28:23 crc kubenswrapper[4846]: I1005 08:28:23.613269 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-755575957b-g5bfm" Oct 05 08:28:23 crc kubenswrapper[4846]: I1005 08:28:23.661562 4846 scope.go:117] "RemoveContainer" containerID="58966a1871f676fc0f8cf0db57eddd59ffc87a81cbc865e4737a18455c414414" Oct 05 08:28:23 crc kubenswrapper[4846]: I1005 08:28:23.766997 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-api-755575957b-g5bfm"] Oct 05 08:28:23 crc kubenswrapper[4846]: I1005 08:28:23.775243 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-api-755575957b-g5bfm"] Oct 05 08:28:23 crc kubenswrapper[4846]: I1005 08:28:23.783285 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-cfnapi-86b55f6c-pznr6"] Oct 05 08:28:23 crc kubenswrapper[4846]: I1005 08:28:23.790493 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-cfnapi-86b55f6c-pznr6"] Oct 05 08:28:24 crc kubenswrapper[4846]: I1005 08:28:24.509375 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bfe1edf-d89c-4d9e-b4fa-896c161c965f" path="/var/lib/kubelet/pods/1bfe1edf-d89c-4d9e-b4fa-896c161c965f/volumes" Oct 05 08:28:24 crc kubenswrapper[4846]: I1005 08:28:24.510439 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2eaaa848-5b1c-402c-af31-23dc2d3eaef1" path="/var/lib/kubelet/pods/2eaaa848-5b1c-402c-af31-23dc2d3eaef1/volumes" Oct 05 08:28:24 crc kubenswrapper[4846]: I1005 08:28:24.511055 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="530d7f87-9f71-4d89-8520-ae77ca8a901d" path="/var/lib/kubelet/pods/530d7f87-9f71-4d89-8520-ae77ca8a901d/volumes" Oct 05 08:28:24 crc kubenswrapper[4846]: I1005 08:28:24.512289 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9fff0fb3-5365-420b-afc2-c997c3b98b6a" path="/var/lib/kubelet/pods/9fff0fb3-5365-420b-afc2-c997c3b98b6a/volumes" Oct 05 08:28:26 crc kubenswrapper[4846]: I1005 08:28:26.264016 4846 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-f8ddb658d-6wlz2" podUID="8329aafa-c3d7-474c-a31b-a56e0d23ddce" containerName="horizon" probeResult="failure" output="Get \"https://10.217.1.103:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.103:8443: connect: connection refused" Oct 05 08:28:26 crc kubenswrapper[4846]: I1005 08:28:26.264170 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-f8ddb658d-6wlz2" Oct 05 08:28:27 crc kubenswrapper[4846]: I1005 08:28:27.849477 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-engine-6b557c77dd-fbhjb" Oct 05 08:28:29 crc kubenswrapper[4846]: I1005 08:28:29.684064 4846 generic.go:334] "Generic (PLEG): container finished" podID="8329aafa-c3d7-474c-a31b-a56e0d23ddce" containerID="22c179cdba0416e23837204edd3d2a434931018a7d897c19b9aa1476bc215022" exitCode=137 Oct 05 08:28:29 crc kubenswrapper[4846]: I1005 08:28:29.684296 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-f8ddb658d-6wlz2" event={"ID":"8329aafa-c3d7-474c-a31b-a56e0d23ddce","Type":"ContainerDied","Data":"22c179cdba0416e23837204edd3d2a434931018a7d897c19b9aa1476bc215022"} Oct 05 08:28:29 crc kubenswrapper[4846]: I1005 08:28:29.849551 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-f8ddb658d-6wlz2" Oct 05 08:28:30 crc kubenswrapper[4846]: I1005 08:28:30.017801 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/8329aafa-c3d7-474c-a31b-a56e0d23ddce-horizon-secret-key\") pod \"8329aafa-c3d7-474c-a31b-a56e0d23ddce\" (UID: \"8329aafa-c3d7-474c-a31b-a56e0d23ddce\") " Oct 05 08:28:30 crc kubenswrapper[4846]: I1005 08:28:30.017854 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/8329aafa-c3d7-474c-a31b-a56e0d23ddce-horizon-tls-certs\") pod \"8329aafa-c3d7-474c-a31b-a56e0d23ddce\" (UID: \"8329aafa-c3d7-474c-a31b-a56e0d23ddce\") " Oct 05 08:28:30 crc kubenswrapper[4846]: I1005 08:28:30.017903 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8329aafa-c3d7-474c-a31b-a56e0d23ddce-combined-ca-bundle\") pod \"8329aafa-c3d7-474c-a31b-a56e0d23ddce\" (UID: \"8329aafa-c3d7-474c-a31b-a56e0d23ddce\") " Oct 05 08:28:30 crc kubenswrapper[4846]: I1005 08:28:30.018091 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wddvf\" (UniqueName: \"kubernetes.io/projected/8329aafa-c3d7-474c-a31b-a56e0d23ddce-kube-api-access-wddvf\") pod \"8329aafa-c3d7-474c-a31b-a56e0d23ddce\" (UID: \"8329aafa-c3d7-474c-a31b-a56e0d23ddce\") " Oct 05 08:28:30 crc kubenswrapper[4846]: I1005 08:28:30.018123 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8329aafa-c3d7-474c-a31b-a56e0d23ddce-config-data\") pod \"8329aafa-c3d7-474c-a31b-a56e0d23ddce\" (UID: \"8329aafa-c3d7-474c-a31b-a56e0d23ddce\") " Oct 05 08:28:30 crc kubenswrapper[4846]: I1005 08:28:30.018210 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8329aafa-c3d7-474c-a31b-a56e0d23ddce-logs\") pod \"8329aafa-c3d7-474c-a31b-a56e0d23ddce\" (UID: \"8329aafa-c3d7-474c-a31b-a56e0d23ddce\") " Oct 05 08:28:30 crc kubenswrapper[4846]: I1005 08:28:30.018271 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8329aafa-c3d7-474c-a31b-a56e0d23ddce-scripts\") pod \"8329aafa-c3d7-474c-a31b-a56e0d23ddce\" (UID: \"8329aafa-c3d7-474c-a31b-a56e0d23ddce\") " Oct 05 08:28:30 crc kubenswrapper[4846]: I1005 08:28:30.019268 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8329aafa-c3d7-474c-a31b-a56e0d23ddce-logs" (OuterVolumeSpecName: "logs") pod "8329aafa-c3d7-474c-a31b-a56e0d23ddce" (UID: "8329aafa-c3d7-474c-a31b-a56e0d23ddce"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 08:28:30 crc kubenswrapper[4846]: I1005 08:28:30.023565 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8329aafa-c3d7-474c-a31b-a56e0d23ddce-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "8329aafa-c3d7-474c-a31b-a56e0d23ddce" (UID: "8329aafa-c3d7-474c-a31b-a56e0d23ddce"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:28:30 crc kubenswrapper[4846]: I1005 08:28:30.036911 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8329aafa-c3d7-474c-a31b-a56e0d23ddce-kube-api-access-wddvf" (OuterVolumeSpecName: "kube-api-access-wddvf") pod "8329aafa-c3d7-474c-a31b-a56e0d23ddce" (UID: "8329aafa-c3d7-474c-a31b-a56e0d23ddce"). InnerVolumeSpecName "kube-api-access-wddvf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:28:30 crc kubenswrapper[4846]: I1005 08:28:30.046473 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8329aafa-c3d7-474c-a31b-a56e0d23ddce-config-data" (OuterVolumeSpecName: "config-data") pod "8329aafa-c3d7-474c-a31b-a56e0d23ddce" (UID: "8329aafa-c3d7-474c-a31b-a56e0d23ddce"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 08:28:30 crc kubenswrapper[4846]: I1005 08:28:30.062214 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8329aafa-c3d7-474c-a31b-a56e0d23ddce-scripts" (OuterVolumeSpecName: "scripts") pod "8329aafa-c3d7-474c-a31b-a56e0d23ddce" (UID: "8329aafa-c3d7-474c-a31b-a56e0d23ddce"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 08:28:30 crc kubenswrapper[4846]: I1005 08:28:30.065193 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8329aafa-c3d7-474c-a31b-a56e0d23ddce-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8329aafa-c3d7-474c-a31b-a56e0d23ddce" (UID: "8329aafa-c3d7-474c-a31b-a56e0d23ddce"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:28:30 crc kubenswrapper[4846]: I1005 08:28:30.090001 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8329aafa-c3d7-474c-a31b-a56e0d23ddce-horizon-tls-certs" (OuterVolumeSpecName: "horizon-tls-certs") pod "8329aafa-c3d7-474c-a31b-a56e0d23ddce" (UID: "8329aafa-c3d7-474c-a31b-a56e0d23ddce"). InnerVolumeSpecName "horizon-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:28:30 crc kubenswrapper[4846]: I1005 08:28:30.120578 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wddvf\" (UniqueName: \"kubernetes.io/projected/8329aafa-c3d7-474c-a31b-a56e0d23ddce-kube-api-access-wddvf\") on node \"crc\" DevicePath \"\"" Oct 05 08:28:30 crc kubenswrapper[4846]: I1005 08:28:30.120619 4846 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8329aafa-c3d7-474c-a31b-a56e0d23ddce-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 08:28:30 crc kubenswrapper[4846]: I1005 08:28:30.120632 4846 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8329aafa-c3d7-474c-a31b-a56e0d23ddce-logs\") on node \"crc\" DevicePath \"\"" Oct 05 08:28:30 crc kubenswrapper[4846]: I1005 08:28:30.120645 4846 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8329aafa-c3d7-474c-a31b-a56e0d23ddce-scripts\") on node \"crc\" DevicePath \"\"" Oct 05 08:28:30 crc kubenswrapper[4846]: I1005 08:28:30.120656 4846 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/8329aafa-c3d7-474c-a31b-a56e0d23ddce-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Oct 05 08:28:30 crc kubenswrapper[4846]: I1005 08:28:30.120666 4846 reconciler_common.go:293] "Volume detached for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/8329aafa-c3d7-474c-a31b-a56e0d23ddce-horizon-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 05 08:28:30 crc kubenswrapper[4846]: I1005 08:28:30.120676 4846 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8329aafa-c3d7-474c-a31b-a56e0d23ddce-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 08:28:30 crc kubenswrapper[4846]: I1005 08:28:30.707226 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-f8ddb658d-6wlz2" event={"ID":"8329aafa-c3d7-474c-a31b-a56e0d23ddce","Type":"ContainerDied","Data":"93e436d5d46990653a1327c7c30e1872329b7faf59be26dd23f40a69830d10f3"} Oct 05 08:28:30 crc kubenswrapper[4846]: I1005 08:28:30.707744 4846 scope.go:117] "RemoveContainer" containerID="d5ef0f6c75a94d51341d689bb7cb65065aad64a24ca9ea981f961495ba170d35" Oct 05 08:28:30 crc kubenswrapper[4846]: I1005 08:28:30.707286 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-f8ddb658d-6wlz2" Oct 05 08:28:30 crc kubenswrapper[4846]: I1005 08:28:30.757798 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-f8ddb658d-6wlz2"] Oct 05 08:28:30 crc kubenswrapper[4846]: I1005 08:28:30.769306 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-f8ddb658d-6wlz2"] Oct 05 08:28:30 crc kubenswrapper[4846]: I1005 08:28:30.921520 4846 scope.go:117] "RemoveContainer" containerID="22c179cdba0416e23837204edd3d2a434931018a7d897c19b9aa1476bc215022" Oct 05 08:28:32 crc kubenswrapper[4846]: I1005 08:28:32.514506 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8329aafa-c3d7-474c-a31b-a56e0d23ddce" path="/var/lib/kubelet/pods/8329aafa-c3d7-474c-a31b-a56e0d23ddce/volumes" Oct 05 08:28:34 crc kubenswrapper[4846]: I1005 08:28:34.891298 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-engine-54b8fb86bb-q8qps" Oct 05 08:28:34 crc kubenswrapper[4846]: I1005 08:28:34.952999 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-engine-6b557c77dd-fbhjb"] Oct 05 08:28:34 crc kubenswrapper[4846]: I1005 08:28:34.953272 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/heat-engine-6b557c77dd-fbhjb" podUID="2bfbe6d7-d95b-4ac0-a8f6-52f89ede6ed2" containerName="heat-engine" containerID="cri-o://7a898116d7abb553b82d6daafdb6c659629f2308f48ba7d6ed48fe80c60b8ef5" gracePeriod=60 Oct 05 08:28:37 crc kubenswrapper[4846]: E1005 08:28:37.806588 4846 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="7a898116d7abb553b82d6daafdb6c659629f2308f48ba7d6ed48fe80c60b8ef5" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Oct 05 08:28:37 crc kubenswrapper[4846]: E1005 08:28:37.808049 4846 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="7a898116d7abb553b82d6daafdb6c659629f2308f48ba7d6ed48fe80c60b8ef5" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Oct 05 08:28:37 crc kubenswrapper[4846]: E1005 08:28:37.809284 4846 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="7a898116d7abb553b82d6daafdb6c659629f2308f48ba7d6ed48fe80c60b8ef5" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Oct 05 08:28:37 crc kubenswrapper[4846]: E1005 08:28:37.809320 4846 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/heat-engine-6b557c77dd-fbhjb" podUID="2bfbe6d7-d95b-4ac0-a8f6-52f89ede6ed2" containerName="heat-engine" Oct 05 08:28:46 crc kubenswrapper[4846]: I1005 08:28:46.716255 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-6b557c77dd-fbhjb" Oct 05 08:28:46 crc kubenswrapper[4846]: I1005 08:28:46.775492 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2bfbe6d7-d95b-4ac0-a8f6-52f89ede6ed2-config-data\") pod \"2bfbe6d7-d95b-4ac0-a8f6-52f89ede6ed2\" (UID: \"2bfbe6d7-d95b-4ac0-a8f6-52f89ede6ed2\") " Oct 05 08:28:46 crc kubenswrapper[4846]: I1005 08:28:46.775541 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2bfbe6d7-d95b-4ac0-a8f6-52f89ede6ed2-config-data-custom\") pod \"2bfbe6d7-d95b-4ac0-a8f6-52f89ede6ed2\" (UID: \"2bfbe6d7-d95b-4ac0-a8f6-52f89ede6ed2\") " Oct 05 08:28:46 crc kubenswrapper[4846]: I1005 08:28:46.775647 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dfvvj\" (UniqueName: \"kubernetes.io/projected/2bfbe6d7-d95b-4ac0-a8f6-52f89ede6ed2-kube-api-access-dfvvj\") pod \"2bfbe6d7-d95b-4ac0-a8f6-52f89ede6ed2\" (UID: \"2bfbe6d7-d95b-4ac0-a8f6-52f89ede6ed2\") " Oct 05 08:28:46 crc kubenswrapper[4846]: I1005 08:28:46.775875 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2bfbe6d7-d95b-4ac0-a8f6-52f89ede6ed2-combined-ca-bundle\") pod \"2bfbe6d7-d95b-4ac0-a8f6-52f89ede6ed2\" (UID: \"2bfbe6d7-d95b-4ac0-a8f6-52f89ede6ed2\") " Oct 05 08:28:46 crc kubenswrapper[4846]: I1005 08:28:46.782483 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2bfbe6d7-d95b-4ac0-a8f6-52f89ede6ed2-kube-api-access-dfvvj" (OuterVolumeSpecName: "kube-api-access-dfvvj") pod "2bfbe6d7-d95b-4ac0-a8f6-52f89ede6ed2" (UID: "2bfbe6d7-d95b-4ac0-a8f6-52f89ede6ed2"). InnerVolumeSpecName "kube-api-access-dfvvj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:28:46 crc kubenswrapper[4846]: I1005 08:28:46.790468 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2bfbe6d7-d95b-4ac0-a8f6-52f89ede6ed2-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "2bfbe6d7-d95b-4ac0-a8f6-52f89ede6ed2" (UID: "2bfbe6d7-d95b-4ac0-a8f6-52f89ede6ed2"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:28:46 crc kubenswrapper[4846]: I1005 08:28:46.805688 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2bfbe6d7-d95b-4ac0-a8f6-52f89ede6ed2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2bfbe6d7-d95b-4ac0-a8f6-52f89ede6ed2" (UID: "2bfbe6d7-d95b-4ac0-a8f6-52f89ede6ed2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:28:46 crc kubenswrapper[4846]: I1005 08:28:46.853507 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2bfbe6d7-d95b-4ac0-a8f6-52f89ede6ed2-config-data" (OuterVolumeSpecName: "config-data") pod "2bfbe6d7-d95b-4ac0-a8f6-52f89ede6ed2" (UID: "2bfbe6d7-d95b-4ac0-a8f6-52f89ede6ed2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:28:46 crc kubenswrapper[4846]: I1005 08:28:46.878809 4846 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2bfbe6d7-d95b-4ac0-a8f6-52f89ede6ed2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 08:28:46 crc kubenswrapper[4846]: I1005 08:28:46.878858 4846 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2bfbe6d7-d95b-4ac0-a8f6-52f89ede6ed2-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 08:28:46 crc kubenswrapper[4846]: I1005 08:28:46.878875 4846 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2bfbe6d7-d95b-4ac0-a8f6-52f89ede6ed2-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 05 08:28:46 crc kubenswrapper[4846]: I1005 08:28:46.878895 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dfvvj\" (UniqueName: \"kubernetes.io/projected/2bfbe6d7-d95b-4ac0-a8f6-52f89ede6ed2-kube-api-access-dfvvj\") on node \"crc\" DevicePath \"\"" Oct 05 08:28:46 crc kubenswrapper[4846]: I1005 08:28:46.936913 4846 generic.go:334] "Generic (PLEG): container finished" podID="2bfbe6d7-d95b-4ac0-a8f6-52f89ede6ed2" containerID="7a898116d7abb553b82d6daafdb6c659629f2308f48ba7d6ed48fe80c60b8ef5" exitCode=0 Oct 05 08:28:46 crc kubenswrapper[4846]: I1005 08:28:46.936996 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-6b557c77dd-fbhjb" event={"ID":"2bfbe6d7-d95b-4ac0-a8f6-52f89ede6ed2","Type":"ContainerDied","Data":"7a898116d7abb553b82d6daafdb6c659629f2308f48ba7d6ed48fe80c60b8ef5"} Oct 05 08:28:46 crc kubenswrapper[4846]: I1005 08:28:46.937021 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-6b557c77dd-fbhjb" Oct 05 08:28:46 crc kubenswrapper[4846]: I1005 08:28:46.937079 4846 scope.go:117] "RemoveContainer" containerID="7a898116d7abb553b82d6daafdb6c659629f2308f48ba7d6ed48fe80c60b8ef5" Oct 05 08:28:46 crc kubenswrapper[4846]: I1005 08:28:46.937063 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-6b557c77dd-fbhjb" event={"ID":"2bfbe6d7-d95b-4ac0-a8f6-52f89ede6ed2","Type":"ContainerDied","Data":"9ac5aab6c23b2260ae33b2e74e29f510606142eb021382c77a67fd5196fc26cd"} Oct 05 08:28:46 crc kubenswrapper[4846]: I1005 08:28:46.980715 4846 scope.go:117] "RemoveContainer" containerID="7a898116d7abb553b82d6daafdb6c659629f2308f48ba7d6ed48fe80c60b8ef5" Oct 05 08:28:46 crc kubenswrapper[4846]: E1005 08:28:46.981957 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7a898116d7abb553b82d6daafdb6c659629f2308f48ba7d6ed48fe80c60b8ef5\": container with ID starting with 7a898116d7abb553b82d6daafdb6c659629f2308f48ba7d6ed48fe80c60b8ef5 not found: ID does not exist" containerID="7a898116d7abb553b82d6daafdb6c659629f2308f48ba7d6ed48fe80c60b8ef5" Oct 05 08:28:46 crc kubenswrapper[4846]: I1005 08:28:46.982053 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7a898116d7abb553b82d6daafdb6c659629f2308f48ba7d6ed48fe80c60b8ef5"} err="failed to get container status \"7a898116d7abb553b82d6daafdb6c659629f2308f48ba7d6ed48fe80c60b8ef5\": rpc error: code = NotFound desc = could not find container \"7a898116d7abb553b82d6daafdb6c659629f2308f48ba7d6ed48fe80c60b8ef5\": container with ID starting with 7a898116d7abb553b82d6daafdb6c659629f2308f48ba7d6ed48fe80c60b8ef5 not found: ID does not exist" Oct 05 08:28:47 crc kubenswrapper[4846]: I1005 08:28:47.011193 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-engine-6b557c77dd-fbhjb"] Oct 05 08:28:47 crc kubenswrapper[4846]: I1005 08:28:47.028007 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-engine-6b557c77dd-fbhjb"] Oct 05 08:28:48 crc kubenswrapper[4846]: I1005 08:28:48.507113 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2bfbe6d7-d95b-4ac0-a8f6-52f89ede6ed2" path="/var/lib/kubelet/pods/2bfbe6d7-d95b-4ac0-a8f6-52f89ede6ed2/volumes" Oct 05 08:28:49 crc kubenswrapper[4846]: I1005 08:28:49.093264 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dpmkhb"] Oct 05 08:28:49 crc kubenswrapper[4846]: E1005 08:28:49.093677 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2eaaa848-5b1c-402c-af31-23dc2d3eaef1" containerName="heat-api" Oct 05 08:28:49 crc kubenswrapper[4846]: I1005 08:28:49.093693 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="2eaaa848-5b1c-402c-af31-23dc2d3eaef1" containerName="heat-api" Oct 05 08:28:49 crc kubenswrapper[4846]: E1005 08:28:49.093709 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8329aafa-c3d7-474c-a31b-a56e0d23ddce" containerName="horizon" Oct 05 08:28:49 crc kubenswrapper[4846]: I1005 08:28:49.093715 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="8329aafa-c3d7-474c-a31b-a56e0d23ddce" containerName="horizon" Oct 05 08:28:49 crc kubenswrapper[4846]: E1005 08:28:49.093727 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="530d7f87-9f71-4d89-8520-ae77ca8a901d" containerName="heat-cfnapi" Oct 05 08:28:49 crc kubenswrapper[4846]: I1005 08:28:49.093733 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="530d7f87-9f71-4d89-8520-ae77ca8a901d" containerName="heat-cfnapi" Oct 05 08:28:49 crc kubenswrapper[4846]: E1005 08:28:49.093742 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2bfbe6d7-d95b-4ac0-a8f6-52f89ede6ed2" containerName="heat-engine" Oct 05 08:28:49 crc kubenswrapper[4846]: I1005 08:28:49.093747 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="2bfbe6d7-d95b-4ac0-a8f6-52f89ede6ed2" containerName="heat-engine" Oct 05 08:28:49 crc kubenswrapper[4846]: E1005 08:28:49.093757 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1bfe1edf-d89c-4d9e-b4fa-896c161c965f" containerName="heat-api" Oct 05 08:28:49 crc kubenswrapper[4846]: I1005 08:28:49.093764 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="1bfe1edf-d89c-4d9e-b4fa-896c161c965f" containerName="heat-api" Oct 05 08:28:49 crc kubenswrapper[4846]: E1005 08:28:49.093800 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8329aafa-c3d7-474c-a31b-a56e0d23ddce" containerName="horizon-log" Oct 05 08:28:49 crc kubenswrapper[4846]: I1005 08:28:49.093807 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="8329aafa-c3d7-474c-a31b-a56e0d23ddce" containerName="horizon-log" Oct 05 08:28:49 crc kubenswrapper[4846]: E1005 08:28:49.093819 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1bfe1edf-d89c-4d9e-b4fa-896c161c965f" containerName="heat-api" Oct 05 08:28:49 crc kubenswrapper[4846]: I1005 08:28:49.093825 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="1bfe1edf-d89c-4d9e-b4fa-896c161c965f" containerName="heat-api" Oct 05 08:28:49 crc kubenswrapper[4846]: E1005 08:28:49.093832 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9fff0fb3-5365-420b-afc2-c997c3b98b6a" containerName="heat-cfnapi" Oct 05 08:28:49 crc kubenswrapper[4846]: I1005 08:28:49.093837 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="9fff0fb3-5365-420b-afc2-c997c3b98b6a" containerName="heat-cfnapi" Oct 05 08:28:49 crc kubenswrapper[4846]: I1005 08:28:49.094007 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="1bfe1edf-d89c-4d9e-b4fa-896c161c965f" containerName="heat-api" Oct 05 08:28:49 crc kubenswrapper[4846]: I1005 08:28:49.094020 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="8329aafa-c3d7-474c-a31b-a56e0d23ddce" containerName="horizon" Oct 05 08:28:49 crc kubenswrapper[4846]: I1005 08:28:49.094028 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="8329aafa-c3d7-474c-a31b-a56e0d23ddce" containerName="horizon-log" Oct 05 08:28:49 crc kubenswrapper[4846]: I1005 08:28:49.094044 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="530d7f87-9f71-4d89-8520-ae77ca8a901d" containerName="heat-cfnapi" Oct 05 08:28:49 crc kubenswrapper[4846]: I1005 08:28:49.094055 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="1bfe1edf-d89c-4d9e-b4fa-896c161c965f" containerName="heat-api" Oct 05 08:28:49 crc kubenswrapper[4846]: I1005 08:28:49.094064 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="2eaaa848-5b1c-402c-af31-23dc2d3eaef1" containerName="heat-api" Oct 05 08:28:49 crc kubenswrapper[4846]: I1005 08:28:49.094073 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="9fff0fb3-5365-420b-afc2-c997c3b98b6a" containerName="heat-cfnapi" Oct 05 08:28:49 crc kubenswrapper[4846]: I1005 08:28:49.094081 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="530d7f87-9f71-4d89-8520-ae77ca8a901d" containerName="heat-cfnapi" Oct 05 08:28:49 crc kubenswrapper[4846]: I1005 08:28:49.094093 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="2bfbe6d7-d95b-4ac0-a8f6-52f89ede6ed2" containerName="heat-engine" Oct 05 08:28:49 crc kubenswrapper[4846]: E1005 08:28:49.094339 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="530d7f87-9f71-4d89-8520-ae77ca8a901d" containerName="heat-cfnapi" Oct 05 08:28:49 crc kubenswrapper[4846]: I1005 08:28:49.094352 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="530d7f87-9f71-4d89-8520-ae77ca8a901d" containerName="heat-cfnapi" Oct 05 08:28:49 crc kubenswrapper[4846]: I1005 08:28:49.095645 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dpmkhb" Oct 05 08:28:49 crc kubenswrapper[4846]: I1005 08:28:49.098874 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Oct 05 08:28:49 crc kubenswrapper[4846]: I1005 08:28:49.107541 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dpmkhb"] Oct 05 08:28:49 crc kubenswrapper[4846]: I1005 08:28:49.120930 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a5383419-49af-4680-9793-d361430cd80a-bundle\") pod \"a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dpmkhb\" (UID: \"a5383419-49af-4680-9793-d361430cd80a\") " pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dpmkhb" Oct 05 08:28:49 crc kubenswrapper[4846]: I1005 08:28:49.121063 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a5383419-49af-4680-9793-d361430cd80a-util\") pod \"a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dpmkhb\" (UID: \"a5383419-49af-4680-9793-d361430cd80a\") " pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dpmkhb" Oct 05 08:28:49 crc kubenswrapper[4846]: I1005 08:28:49.121288 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zthgs\" (UniqueName: \"kubernetes.io/projected/a5383419-49af-4680-9793-d361430cd80a-kube-api-access-zthgs\") pod \"a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dpmkhb\" (UID: \"a5383419-49af-4680-9793-d361430cd80a\") " pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dpmkhb" Oct 05 08:28:49 crc kubenswrapper[4846]: I1005 08:28:49.222452 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zthgs\" (UniqueName: \"kubernetes.io/projected/a5383419-49af-4680-9793-d361430cd80a-kube-api-access-zthgs\") pod \"a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dpmkhb\" (UID: \"a5383419-49af-4680-9793-d361430cd80a\") " pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dpmkhb" Oct 05 08:28:49 crc kubenswrapper[4846]: I1005 08:28:49.222562 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a5383419-49af-4680-9793-d361430cd80a-bundle\") pod \"a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dpmkhb\" (UID: \"a5383419-49af-4680-9793-d361430cd80a\") " pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dpmkhb" Oct 05 08:28:49 crc kubenswrapper[4846]: I1005 08:28:49.222621 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a5383419-49af-4680-9793-d361430cd80a-util\") pod \"a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dpmkhb\" (UID: \"a5383419-49af-4680-9793-d361430cd80a\") " pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dpmkhb" Oct 05 08:28:49 crc kubenswrapper[4846]: I1005 08:28:49.223291 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a5383419-49af-4680-9793-d361430cd80a-util\") pod \"a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dpmkhb\" (UID: \"a5383419-49af-4680-9793-d361430cd80a\") " pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dpmkhb" Oct 05 08:28:49 crc kubenswrapper[4846]: I1005 08:28:49.223448 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a5383419-49af-4680-9793-d361430cd80a-bundle\") pod \"a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dpmkhb\" (UID: \"a5383419-49af-4680-9793-d361430cd80a\") " pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dpmkhb" Oct 05 08:28:49 crc kubenswrapper[4846]: I1005 08:28:49.247132 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zthgs\" (UniqueName: \"kubernetes.io/projected/a5383419-49af-4680-9793-d361430cd80a-kube-api-access-zthgs\") pod \"a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dpmkhb\" (UID: \"a5383419-49af-4680-9793-d361430cd80a\") " pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dpmkhb" Oct 05 08:28:49 crc kubenswrapper[4846]: I1005 08:28:49.425841 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dpmkhb" Oct 05 08:28:49 crc kubenswrapper[4846]: I1005 08:28:49.898107 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dpmkhb"] Oct 05 08:28:49 crc kubenswrapper[4846]: W1005 08:28:49.910042 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda5383419_49af_4680_9793_d361430cd80a.slice/crio-1cd730829343b66bd1a3da694c222b2d4ac19bcc8d0ef403d6ccae9fc0e95e3c WatchSource:0}: Error finding container 1cd730829343b66bd1a3da694c222b2d4ac19bcc8d0ef403d6ccae9fc0e95e3c: Status 404 returned error can't find the container with id 1cd730829343b66bd1a3da694c222b2d4ac19bcc8d0ef403d6ccae9fc0e95e3c Oct 05 08:28:49 crc kubenswrapper[4846]: I1005 08:28:49.968970 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dpmkhb" event={"ID":"a5383419-49af-4680-9793-d361430cd80a","Type":"ContainerStarted","Data":"1cd730829343b66bd1a3da694c222b2d4ac19bcc8d0ef403d6ccae9fc0e95e3c"} Oct 05 08:28:50 crc kubenswrapper[4846]: I1005 08:28:50.981420 4846 generic.go:334] "Generic (PLEG): container finished" podID="a5383419-49af-4680-9793-d361430cd80a" containerID="fb187d08f8dfa1127dcbc4a7ddc05b0afb47c4c9826f20727072e17e3d9ecdd8" exitCode=0 Oct 05 08:28:50 crc kubenswrapper[4846]: I1005 08:28:50.981497 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dpmkhb" event={"ID":"a5383419-49af-4680-9793-d361430cd80a","Type":"ContainerDied","Data":"fb187d08f8dfa1127dcbc4a7ddc05b0afb47c4c9826f20727072e17e3d9ecdd8"} Oct 05 08:28:50 crc kubenswrapper[4846]: I1005 08:28:50.984044 4846 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 05 08:28:53 crc kubenswrapper[4846]: I1005 08:28:53.020860 4846 generic.go:334] "Generic (PLEG): container finished" podID="a5383419-49af-4680-9793-d361430cd80a" containerID="c8c1bac6b5c7a2f929cb3ad41200c27b53fb284d71aaef5ba5cf6804564ca6c1" exitCode=0 Oct 05 08:28:53 crc kubenswrapper[4846]: I1005 08:28:53.020958 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dpmkhb" event={"ID":"a5383419-49af-4680-9793-d361430cd80a","Type":"ContainerDied","Data":"c8c1bac6b5c7a2f929cb3ad41200c27b53fb284d71aaef5ba5cf6804564ca6c1"} Oct 05 08:28:54 crc kubenswrapper[4846]: I1005 08:28:54.041645 4846 generic.go:334] "Generic (PLEG): container finished" podID="a5383419-49af-4680-9793-d361430cd80a" containerID="6ae8e712a2b3c3e2f4d0ba6e767098f2990eb46e3440c6a821e117c70dcbac90" exitCode=0 Oct 05 08:28:54 crc kubenswrapper[4846]: I1005 08:28:54.041776 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dpmkhb" event={"ID":"a5383419-49af-4680-9793-d361430cd80a","Type":"ContainerDied","Data":"6ae8e712a2b3c3e2f4d0ba6e767098f2990eb46e3440c6a821e117c70dcbac90"} Oct 05 08:28:55 crc kubenswrapper[4846]: I1005 08:28:55.537889 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dpmkhb" Oct 05 08:28:55 crc kubenswrapper[4846]: I1005 08:28:55.591205 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a5383419-49af-4680-9793-d361430cd80a-bundle\") pod \"a5383419-49af-4680-9793-d361430cd80a\" (UID: \"a5383419-49af-4680-9793-d361430cd80a\") " Oct 05 08:28:55 crc kubenswrapper[4846]: I1005 08:28:55.591288 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a5383419-49af-4680-9793-d361430cd80a-util\") pod \"a5383419-49af-4680-9793-d361430cd80a\" (UID: \"a5383419-49af-4680-9793-d361430cd80a\") " Oct 05 08:28:55 crc kubenswrapper[4846]: I1005 08:28:55.591403 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zthgs\" (UniqueName: \"kubernetes.io/projected/a5383419-49af-4680-9793-d361430cd80a-kube-api-access-zthgs\") pod \"a5383419-49af-4680-9793-d361430cd80a\" (UID: \"a5383419-49af-4680-9793-d361430cd80a\") " Oct 05 08:28:55 crc kubenswrapper[4846]: I1005 08:28:55.593932 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a5383419-49af-4680-9793-d361430cd80a-bundle" (OuterVolumeSpecName: "bundle") pod "a5383419-49af-4680-9793-d361430cd80a" (UID: "a5383419-49af-4680-9793-d361430cd80a"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 08:28:55 crc kubenswrapper[4846]: I1005 08:28:55.600890 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a5383419-49af-4680-9793-d361430cd80a-kube-api-access-zthgs" (OuterVolumeSpecName: "kube-api-access-zthgs") pod "a5383419-49af-4680-9793-d361430cd80a" (UID: "a5383419-49af-4680-9793-d361430cd80a"). InnerVolumeSpecName "kube-api-access-zthgs". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:28:55 crc kubenswrapper[4846]: I1005 08:28:55.692988 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zthgs\" (UniqueName: \"kubernetes.io/projected/a5383419-49af-4680-9793-d361430cd80a-kube-api-access-zthgs\") on node \"crc\" DevicePath \"\"" Oct 05 08:28:55 crc kubenswrapper[4846]: I1005 08:28:55.693443 4846 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a5383419-49af-4680-9793-d361430cd80a-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 08:28:55 crc kubenswrapper[4846]: I1005 08:28:55.770421 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a5383419-49af-4680-9793-d361430cd80a-util" (OuterVolumeSpecName: "util") pod "a5383419-49af-4680-9793-d361430cd80a" (UID: "a5383419-49af-4680-9793-d361430cd80a"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 08:28:55 crc kubenswrapper[4846]: I1005 08:28:55.795890 4846 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a5383419-49af-4680-9793-d361430cd80a-util\") on node \"crc\" DevicePath \"\"" Oct 05 08:28:56 crc kubenswrapper[4846]: I1005 08:28:56.089488 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dpmkhb" event={"ID":"a5383419-49af-4680-9793-d361430cd80a","Type":"ContainerDied","Data":"1cd730829343b66bd1a3da694c222b2d4ac19bcc8d0ef403d6ccae9fc0e95e3c"} Oct 05 08:28:56 crc kubenswrapper[4846]: I1005 08:28:56.089569 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1cd730829343b66bd1a3da694c222b2d4ac19bcc8d0ef403d6ccae9fc0e95e3c" Oct 05 08:28:56 crc kubenswrapper[4846]: I1005 08:28:56.089590 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dpmkhb" Oct 05 08:28:59 crc kubenswrapper[4846]: I1005 08:28:59.067862 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-create-4ccwk"] Oct 05 08:28:59 crc kubenswrapper[4846]: I1005 08:28:59.078807 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-create-4ccwk"] Oct 05 08:29:00 crc kubenswrapper[4846]: I1005 08:29:00.513124 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf32b0d-81b1-4774-bf03-995ba2dcfe2d" path="/var/lib/kubelet/pods/1bf32b0d-81b1-4774-bf03-995ba2dcfe2d/volumes" Oct 05 08:29:06 crc kubenswrapper[4846]: I1005 08:29:06.545649 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-7c8cf85677-gjc7q"] Oct 05 08:29:06 crc kubenswrapper[4846]: E1005 08:29:06.546454 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a5383419-49af-4680-9793-d361430cd80a" containerName="pull" Oct 05 08:29:06 crc kubenswrapper[4846]: I1005 08:29:06.546466 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="a5383419-49af-4680-9793-d361430cd80a" containerName="pull" Oct 05 08:29:06 crc kubenswrapper[4846]: E1005 08:29:06.546481 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a5383419-49af-4680-9793-d361430cd80a" containerName="util" Oct 05 08:29:06 crc kubenswrapper[4846]: I1005 08:29:06.546487 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="a5383419-49af-4680-9793-d361430cd80a" containerName="util" Oct 05 08:29:06 crc kubenswrapper[4846]: E1005 08:29:06.546499 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a5383419-49af-4680-9793-d361430cd80a" containerName="extract" Oct 05 08:29:06 crc kubenswrapper[4846]: I1005 08:29:06.546504 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="a5383419-49af-4680-9793-d361430cd80a" containerName="extract" Oct 05 08:29:06 crc kubenswrapper[4846]: I1005 08:29:06.546671 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="a5383419-49af-4680-9793-d361430cd80a" containerName="extract" Oct 05 08:29:06 crc kubenswrapper[4846]: I1005 08:29:06.547327 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-7c8cf85677-gjc7q" Oct 05 08:29:06 crc kubenswrapper[4846]: I1005 08:29:06.549204 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-dockercfg-jndr4" Oct 05 08:29:06 crc kubenswrapper[4846]: I1005 08:29:06.549946 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators"/"kube-root-ca.crt" Oct 05 08:29:06 crc kubenswrapper[4846]: I1005 08:29:06.553388 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators"/"openshift-service-ca.crt" Oct 05 08:29:06 crc kubenswrapper[4846]: I1005 08:29:06.559731 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-7c8cf85677-gjc7q"] Oct 05 08:29:06 crc kubenswrapper[4846]: I1005 08:29:06.674392 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-59bdf9d4c4-mrtq4"] Oct 05 08:29:06 crc kubenswrapper[4846]: I1005 08:29:06.675906 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-59bdf9d4c4-mrtq4" Oct 05 08:29:06 crc kubenswrapper[4846]: I1005 08:29:06.682145 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-admission-webhook-service-cert" Oct 05 08:29:06 crc kubenswrapper[4846]: I1005 08:29:06.682357 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-admission-webhook-dockercfg-xwp2m" Oct 05 08:29:06 crc kubenswrapper[4846]: I1005 08:29:06.686133 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-59bdf9d4c4-44zmx"] Oct 05 08:29:06 crc kubenswrapper[4846]: I1005 08:29:06.687469 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-59bdf9d4c4-44zmx" Oct 05 08:29:06 crc kubenswrapper[4846]: I1005 08:29:06.698563 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-59bdf9d4c4-mrtq4"] Oct 05 08:29:06 crc kubenswrapper[4846]: I1005 08:29:06.717188 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-59bdf9d4c4-44zmx"] Oct 05 08:29:06 crc kubenswrapper[4846]: I1005 08:29:06.733914 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4h5zh\" (UniqueName: \"kubernetes.io/projected/bb1e8f48-88f8-4623-9503-3589e855967a-kube-api-access-4h5zh\") pod \"obo-prometheus-operator-7c8cf85677-gjc7q\" (UID: \"bb1e8f48-88f8-4623-9503-3589e855967a\") " pod="openshift-operators/obo-prometheus-operator-7c8cf85677-gjc7q" Oct 05 08:29:06 crc kubenswrapper[4846]: I1005 08:29:06.836293 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/3f591385-e1df-4dbd-a42d-a96f6a866cad-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-59bdf9d4c4-44zmx\" (UID: \"3f591385-e1df-4dbd-a42d-a96f6a866cad\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-59bdf9d4c4-44zmx" Oct 05 08:29:06 crc kubenswrapper[4846]: I1005 08:29:06.836348 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/3f591385-e1df-4dbd-a42d-a96f6a866cad-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-59bdf9d4c4-44zmx\" (UID: \"3f591385-e1df-4dbd-a42d-a96f6a866cad\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-59bdf9d4c4-44zmx" Oct 05 08:29:06 crc kubenswrapper[4846]: I1005 08:29:06.836464 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/d7e575d9-c148-487e-ad29-10c42e593221-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-59bdf9d4c4-mrtq4\" (UID: \"d7e575d9-c148-487e-ad29-10c42e593221\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-59bdf9d4c4-mrtq4" Oct 05 08:29:06 crc kubenswrapper[4846]: I1005 08:29:06.836523 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4h5zh\" (UniqueName: \"kubernetes.io/projected/bb1e8f48-88f8-4623-9503-3589e855967a-kube-api-access-4h5zh\") pod \"obo-prometheus-operator-7c8cf85677-gjc7q\" (UID: \"bb1e8f48-88f8-4623-9503-3589e855967a\") " pod="openshift-operators/obo-prometheus-operator-7c8cf85677-gjc7q" Oct 05 08:29:06 crc kubenswrapper[4846]: I1005 08:29:06.836551 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/d7e575d9-c148-487e-ad29-10c42e593221-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-59bdf9d4c4-mrtq4\" (UID: \"d7e575d9-c148-487e-ad29-10c42e593221\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-59bdf9d4c4-mrtq4" Oct 05 08:29:06 crc kubenswrapper[4846]: I1005 08:29:06.857937 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4h5zh\" (UniqueName: \"kubernetes.io/projected/bb1e8f48-88f8-4623-9503-3589e855967a-kube-api-access-4h5zh\") pod \"obo-prometheus-operator-7c8cf85677-gjc7q\" (UID: \"bb1e8f48-88f8-4623-9503-3589e855967a\") " pod="openshift-operators/obo-prometheus-operator-7c8cf85677-gjc7q" Oct 05 08:29:06 crc kubenswrapper[4846]: I1005 08:29:06.867284 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/observability-operator-cc5f78dfc-95rb8"] Oct 05 08:29:06 crc kubenswrapper[4846]: I1005 08:29:06.868521 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-cc5f78dfc-95rb8" Oct 05 08:29:06 crc kubenswrapper[4846]: I1005 08:29:06.868853 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-7c8cf85677-gjc7q" Oct 05 08:29:06 crc kubenswrapper[4846]: I1005 08:29:06.872027 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"observability-operator-sa-dockercfg-dmmk6" Oct 05 08:29:06 crc kubenswrapper[4846]: I1005 08:29:06.872095 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"observability-operator-tls" Oct 05 08:29:06 crc kubenswrapper[4846]: I1005 08:29:06.893242 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/observability-operator-cc5f78dfc-95rb8"] Oct 05 08:29:06 crc kubenswrapper[4846]: I1005 08:29:06.938895 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/d7e575d9-c148-487e-ad29-10c42e593221-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-59bdf9d4c4-mrtq4\" (UID: \"d7e575d9-c148-487e-ad29-10c42e593221\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-59bdf9d4c4-mrtq4" Oct 05 08:29:06 crc kubenswrapper[4846]: I1005 08:29:06.938964 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/d7e575d9-c148-487e-ad29-10c42e593221-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-59bdf9d4c4-mrtq4\" (UID: \"d7e575d9-c148-487e-ad29-10c42e593221\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-59bdf9d4c4-mrtq4" Oct 05 08:29:06 crc kubenswrapper[4846]: I1005 08:29:06.939010 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/3f591385-e1df-4dbd-a42d-a96f6a866cad-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-59bdf9d4c4-44zmx\" (UID: \"3f591385-e1df-4dbd-a42d-a96f6a866cad\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-59bdf9d4c4-44zmx" Oct 05 08:29:06 crc kubenswrapper[4846]: I1005 08:29:06.939042 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/3f591385-e1df-4dbd-a42d-a96f6a866cad-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-59bdf9d4c4-44zmx\" (UID: \"3f591385-e1df-4dbd-a42d-a96f6a866cad\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-59bdf9d4c4-44zmx" Oct 05 08:29:06 crc kubenswrapper[4846]: I1005 08:29:06.944108 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/d7e575d9-c148-487e-ad29-10c42e593221-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-59bdf9d4c4-mrtq4\" (UID: \"d7e575d9-c148-487e-ad29-10c42e593221\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-59bdf9d4c4-mrtq4" Oct 05 08:29:06 crc kubenswrapper[4846]: I1005 08:29:06.945747 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/3f591385-e1df-4dbd-a42d-a96f6a866cad-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-59bdf9d4c4-44zmx\" (UID: \"3f591385-e1df-4dbd-a42d-a96f6a866cad\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-59bdf9d4c4-44zmx" Oct 05 08:29:06 crc kubenswrapper[4846]: I1005 08:29:06.946691 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/3f591385-e1df-4dbd-a42d-a96f6a866cad-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-59bdf9d4c4-44zmx\" (UID: \"3f591385-e1df-4dbd-a42d-a96f6a866cad\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-59bdf9d4c4-44zmx" Oct 05 08:29:06 crc kubenswrapper[4846]: I1005 08:29:06.947975 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/d7e575d9-c148-487e-ad29-10c42e593221-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-59bdf9d4c4-mrtq4\" (UID: \"d7e575d9-c148-487e-ad29-10c42e593221\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-59bdf9d4c4-mrtq4" Oct 05 08:29:07 crc kubenswrapper[4846]: I1005 08:29:07.005917 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/perses-operator-54bc95c9fb-llcsj"] Oct 05 08:29:07 crc kubenswrapper[4846]: I1005 08:29:07.007677 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-54bc95c9fb-llcsj" Oct 05 08:29:07 crc kubenswrapper[4846]: I1005 08:29:07.011601 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"perses-operator-dockercfg-f2pfl" Oct 05 08:29:07 crc kubenswrapper[4846]: I1005 08:29:07.018331 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/perses-operator-54bc95c9fb-llcsj"] Oct 05 08:29:07 crc kubenswrapper[4846]: I1005 08:29:07.018734 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-59bdf9d4c4-mrtq4" Oct 05 08:29:07 crc kubenswrapper[4846]: I1005 08:29:07.033998 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-59bdf9d4c4-44zmx" Oct 05 08:29:07 crc kubenswrapper[4846]: I1005 08:29:07.041015 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-flj5b\" (UniqueName: \"kubernetes.io/projected/c96f8bbb-d1ef-4345-8ce7-8e72d543e47e-kube-api-access-flj5b\") pod \"observability-operator-cc5f78dfc-95rb8\" (UID: \"c96f8bbb-d1ef-4345-8ce7-8e72d543e47e\") " pod="openshift-operators/observability-operator-cc5f78dfc-95rb8" Oct 05 08:29:07 crc kubenswrapper[4846]: I1005 08:29:07.041050 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/c96f8bbb-d1ef-4345-8ce7-8e72d543e47e-observability-operator-tls\") pod \"observability-operator-cc5f78dfc-95rb8\" (UID: \"c96f8bbb-d1ef-4345-8ce7-8e72d543e47e\") " pod="openshift-operators/observability-operator-cc5f78dfc-95rb8" Oct 05 08:29:07 crc kubenswrapper[4846]: I1005 08:29:07.146400 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5d686\" (UniqueName: \"kubernetes.io/projected/a17876c6-28e5-4ac8-9ebd-ed1be29bb964-kube-api-access-5d686\") pod \"perses-operator-54bc95c9fb-llcsj\" (UID: \"a17876c6-28e5-4ac8-9ebd-ed1be29bb964\") " pod="openshift-operators/perses-operator-54bc95c9fb-llcsj" Oct 05 08:29:07 crc kubenswrapper[4846]: I1005 08:29:07.146505 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/a17876c6-28e5-4ac8-9ebd-ed1be29bb964-openshift-service-ca\") pod \"perses-operator-54bc95c9fb-llcsj\" (UID: \"a17876c6-28e5-4ac8-9ebd-ed1be29bb964\") " pod="openshift-operators/perses-operator-54bc95c9fb-llcsj" Oct 05 08:29:07 crc kubenswrapper[4846]: I1005 08:29:07.146568 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-flj5b\" (UniqueName: \"kubernetes.io/projected/c96f8bbb-d1ef-4345-8ce7-8e72d543e47e-kube-api-access-flj5b\") pod \"observability-operator-cc5f78dfc-95rb8\" (UID: \"c96f8bbb-d1ef-4345-8ce7-8e72d543e47e\") " pod="openshift-operators/observability-operator-cc5f78dfc-95rb8" Oct 05 08:29:07 crc kubenswrapper[4846]: I1005 08:29:07.146607 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/c96f8bbb-d1ef-4345-8ce7-8e72d543e47e-observability-operator-tls\") pod \"observability-operator-cc5f78dfc-95rb8\" (UID: \"c96f8bbb-d1ef-4345-8ce7-8e72d543e47e\") " pod="openshift-operators/observability-operator-cc5f78dfc-95rb8" Oct 05 08:29:07 crc kubenswrapper[4846]: I1005 08:29:07.157399 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/c96f8bbb-d1ef-4345-8ce7-8e72d543e47e-observability-operator-tls\") pod \"observability-operator-cc5f78dfc-95rb8\" (UID: \"c96f8bbb-d1ef-4345-8ce7-8e72d543e47e\") " pod="openshift-operators/observability-operator-cc5f78dfc-95rb8" Oct 05 08:29:07 crc kubenswrapper[4846]: I1005 08:29:07.180266 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-flj5b\" (UniqueName: \"kubernetes.io/projected/c96f8bbb-d1ef-4345-8ce7-8e72d543e47e-kube-api-access-flj5b\") pod \"observability-operator-cc5f78dfc-95rb8\" (UID: \"c96f8bbb-d1ef-4345-8ce7-8e72d543e47e\") " pod="openshift-operators/observability-operator-cc5f78dfc-95rb8" Oct 05 08:29:07 crc kubenswrapper[4846]: I1005 08:29:07.248438 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5d686\" (UniqueName: \"kubernetes.io/projected/a17876c6-28e5-4ac8-9ebd-ed1be29bb964-kube-api-access-5d686\") pod \"perses-operator-54bc95c9fb-llcsj\" (UID: \"a17876c6-28e5-4ac8-9ebd-ed1be29bb964\") " pod="openshift-operators/perses-operator-54bc95c9fb-llcsj" Oct 05 08:29:07 crc kubenswrapper[4846]: I1005 08:29:07.248521 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/a17876c6-28e5-4ac8-9ebd-ed1be29bb964-openshift-service-ca\") pod \"perses-operator-54bc95c9fb-llcsj\" (UID: \"a17876c6-28e5-4ac8-9ebd-ed1be29bb964\") " pod="openshift-operators/perses-operator-54bc95c9fb-llcsj" Oct 05 08:29:07 crc kubenswrapper[4846]: I1005 08:29:07.249514 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/a17876c6-28e5-4ac8-9ebd-ed1be29bb964-openshift-service-ca\") pod \"perses-operator-54bc95c9fb-llcsj\" (UID: \"a17876c6-28e5-4ac8-9ebd-ed1be29bb964\") " pod="openshift-operators/perses-operator-54bc95c9fb-llcsj" Oct 05 08:29:07 crc kubenswrapper[4846]: I1005 08:29:07.263999 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5d686\" (UniqueName: \"kubernetes.io/projected/a17876c6-28e5-4ac8-9ebd-ed1be29bb964-kube-api-access-5d686\") pod \"perses-operator-54bc95c9fb-llcsj\" (UID: \"a17876c6-28e5-4ac8-9ebd-ed1be29bb964\") " pod="openshift-operators/perses-operator-54bc95c9fb-llcsj" Oct 05 08:29:07 crc kubenswrapper[4846]: I1005 08:29:07.354187 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-cc5f78dfc-95rb8" Oct 05 08:29:07 crc kubenswrapper[4846]: I1005 08:29:07.451208 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-7c8cf85677-gjc7q"] Oct 05 08:29:07 crc kubenswrapper[4846]: I1005 08:29:07.468397 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-54bc95c9fb-llcsj" Oct 05 08:29:07 crc kubenswrapper[4846]: I1005 08:29:07.705620 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-59bdf9d4c4-mrtq4"] Oct 05 08:29:07 crc kubenswrapper[4846]: W1005 08:29:07.706629 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd7e575d9_c148_487e_ad29_10c42e593221.slice/crio-cd6d53c47b88aa7efbbaab1e6d9499bc645e147f2dde5cbf6d5b439fe79bb2cb WatchSource:0}: Error finding container cd6d53c47b88aa7efbbaab1e6d9499bc645e147f2dde5cbf6d5b439fe79bb2cb: Status 404 returned error can't find the container with id cd6d53c47b88aa7efbbaab1e6d9499bc645e147f2dde5cbf6d5b439fe79bb2cb Oct 05 08:29:07 crc kubenswrapper[4846]: W1005 08:29:07.769626 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3f591385_e1df_4dbd_a42d_a96f6a866cad.slice/crio-b8a8d3a9bbf7f5684b718c0077175ec60260f274497b91ebe6534686fb31031d WatchSource:0}: Error finding container b8a8d3a9bbf7f5684b718c0077175ec60260f274497b91ebe6534686fb31031d: Status 404 returned error can't find the container with id b8a8d3a9bbf7f5684b718c0077175ec60260f274497b91ebe6534686fb31031d Oct 05 08:29:07 crc kubenswrapper[4846]: I1005 08:29:07.784394 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-59bdf9d4c4-44zmx"] Oct 05 08:29:07 crc kubenswrapper[4846]: I1005 08:29:07.940813 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/observability-operator-cc5f78dfc-95rb8"] Oct 05 08:29:08 crc kubenswrapper[4846]: I1005 08:29:08.047461 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/perses-operator-54bc95c9fb-llcsj"] Oct 05 08:29:08 crc kubenswrapper[4846]: I1005 08:29:08.239306 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/perses-operator-54bc95c9fb-llcsj" event={"ID":"a17876c6-28e5-4ac8-9ebd-ed1be29bb964","Type":"ContainerStarted","Data":"70cc177017050d1d0f9bf21909ae23de2ec778212ee9d3e1ab164c3068ae119d"} Oct 05 08:29:08 crc kubenswrapper[4846]: I1005 08:29:08.241105 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-59bdf9d4c4-mrtq4" event={"ID":"d7e575d9-c148-487e-ad29-10c42e593221","Type":"ContainerStarted","Data":"cd6d53c47b88aa7efbbaab1e6d9499bc645e147f2dde5cbf6d5b439fe79bb2cb"} Oct 05 08:29:08 crc kubenswrapper[4846]: I1005 08:29:08.242386 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/observability-operator-cc5f78dfc-95rb8" event={"ID":"c96f8bbb-d1ef-4345-8ce7-8e72d543e47e","Type":"ContainerStarted","Data":"bda23f612df0a0754458d907f067253a0a744a1718889bff2dae4ebf49a50f6e"} Oct 05 08:29:08 crc kubenswrapper[4846]: I1005 08:29:08.245048 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-59bdf9d4c4-44zmx" event={"ID":"3f591385-e1df-4dbd-a42d-a96f6a866cad","Type":"ContainerStarted","Data":"b8a8d3a9bbf7f5684b718c0077175ec60260f274497b91ebe6534686fb31031d"} Oct 05 08:29:08 crc kubenswrapper[4846]: I1005 08:29:08.247049 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-7c8cf85677-gjc7q" event={"ID":"bb1e8f48-88f8-4623-9503-3589e855967a","Type":"ContainerStarted","Data":"f386dd51f0f9d30bf7f9d66910937ecf4a518e24e7f2b406688ad8d1ab2e249e"} Oct 05 08:29:09 crc kubenswrapper[4846]: I1005 08:29:09.038279 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-dac8-account-create-gt5sl"] Oct 05 08:29:09 crc kubenswrapper[4846]: I1005 08:29:09.047786 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-dac8-account-create-gt5sl"] Oct 05 08:29:10 crc kubenswrapper[4846]: I1005 08:29:10.513867 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="778049cd-b935-404f-887f-63cb4e7d1054" path="/var/lib/kubelet/pods/778049cd-b935-404f-887f-63cb4e7d1054/volumes" Oct 05 08:29:13 crc kubenswrapper[4846]: I1005 08:29:13.324052 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/perses-operator-54bc95c9fb-llcsj" event={"ID":"a17876c6-28e5-4ac8-9ebd-ed1be29bb964","Type":"ContainerStarted","Data":"08fccfe72f1b624a50cc103c53186a5fa26130d6c290668c871adc859afe4e25"} Oct 05 08:29:13 crc kubenswrapper[4846]: I1005 08:29:13.325824 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operators/perses-operator-54bc95c9fb-llcsj" Oct 05 08:29:13 crc kubenswrapper[4846]: I1005 08:29:13.329700 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-59bdf9d4c4-mrtq4" event={"ID":"d7e575d9-c148-487e-ad29-10c42e593221","Type":"ContainerStarted","Data":"eb7511ea5d6b97a2413e3d4bc705a91186a5eeeb018518fa2d868cefbad7160b"} Oct 05 08:29:13 crc kubenswrapper[4846]: I1005 08:29:13.354622 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-59bdf9d4c4-44zmx" event={"ID":"3f591385-e1df-4dbd-a42d-a96f6a866cad","Type":"ContainerStarted","Data":"b6131efd6df3f10c4b796896089a25d506cbbd3ba9d5848e1dd1772cd8c3af0e"} Oct 05 08:29:13 crc kubenswrapper[4846]: I1005 08:29:13.363355 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/perses-operator-54bc95c9fb-llcsj" podStartSLOduration=3.355956607 podStartE2EDuration="7.363336953s" podCreationTimestamp="2025-10-05 08:29:06 +0000 UTC" firstStartedPulling="2025-10-05 08:29:08.054242871 +0000 UTC m=+6070.295095646" lastFinishedPulling="2025-10-05 08:29:12.061623227 +0000 UTC m=+6074.302475992" observedRunningTime="2025-10-05 08:29:13.35756124 +0000 UTC m=+6075.598414005" watchObservedRunningTime="2025-10-05 08:29:13.363336953 +0000 UTC m=+6075.604189728" Oct 05 08:29:13 crc kubenswrapper[4846]: I1005 08:29:13.363964 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-7c8cf85677-gjc7q" event={"ID":"bb1e8f48-88f8-4623-9503-3589e855967a","Type":"ContainerStarted","Data":"ff32aab1fdf6eb58627fa77a893b2cd1bd062892cdaabb0913897b55decddbe0"} Oct 05 08:29:13 crc kubenswrapper[4846]: I1005 08:29:13.382657 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-admission-webhook-59bdf9d4c4-44zmx" podStartSLOduration=3.111348512 podStartE2EDuration="7.382640936s" podCreationTimestamp="2025-10-05 08:29:06 +0000 UTC" firstStartedPulling="2025-10-05 08:29:07.777529693 +0000 UTC m=+6070.018382468" lastFinishedPulling="2025-10-05 08:29:12.048822117 +0000 UTC m=+6074.289674892" observedRunningTime="2025-10-05 08:29:13.381252429 +0000 UTC m=+6075.622105194" watchObservedRunningTime="2025-10-05 08:29:13.382640936 +0000 UTC m=+6075.623493711" Oct 05 08:29:13 crc kubenswrapper[4846]: I1005 08:29:13.423558 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-admission-webhook-59bdf9d4c4-mrtq4" podStartSLOduration=3.077777301 podStartE2EDuration="7.423539212s" podCreationTimestamp="2025-10-05 08:29:06 +0000 UTC" firstStartedPulling="2025-10-05 08:29:07.709667051 +0000 UTC m=+6069.950519826" lastFinishedPulling="2025-10-05 08:29:12.055428962 +0000 UTC m=+6074.296281737" observedRunningTime="2025-10-05 08:29:13.421760515 +0000 UTC m=+6075.662613290" watchObservedRunningTime="2025-10-05 08:29:13.423539212 +0000 UTC m=+6075.664391987" Oct 05 08:29:13 crc kubenswrapper[4846]: I1005 08:29:13.472754 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-7c8cf85677-gjc7q" podStartSLOduration=2.910844746 podStartE2EDuration="7.472733928s" podCreationTimestamp="2025-10-05 08:29:06 +0000 UTC" firstStartedPulling="2025-10-05 08:29:07.497460335 +0000 UTC m=+6069.738313110" lastFinishedPulling="2025-10-05 08:29:12.059349517 +0000 UTC m=+6074.300202292" observedRunningTime="2025-10-05 08:29:13.472448721 +0000 UTC m=+6075.713301516" watchObservedRunningTime="2025-10-05 08:29:13.472733928 +0000 UTC m=+6075.713586703" Oct 05 08:29:17 crc kubenswrapper[4846]: I1005 08:29:17.472782 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operators/perses-operator-54bc95c9fb-llcsj" Oct 05 08:29:18 crc kubenswrapper[4846]: I1005 08:29:18.440600 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/observability-operator-cc5f78dfc-95rb8" event={"ID":"c96f8bbb-d1ef-4345-8ce7-8e72d543e47e","Type":"ContainerStarted","Data":"e08e88285e12412007fb70195e43426cd6c9f06be683e0f5f52a20da5fbd529b"} Oct 05 08:29:18 crc kubenswrapper[4846]: I1005 08:29:18.441112 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operators/observability-operator-cc5f78dfc-95rb8" Oct 05 08:29:18 crc kubenswrapper[4846]: I1005 08:29:18.459899 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/observability-operator-cc5f78dfc-95rb8" podStartSLOduration=2.839090242 podStartE2EDuration="12.459882282s" podCreationTimestamp="2025-10-05 08:29:06 +0000 UTC" firstStartedPulling="2025-10-05 08:29:07.955244862 +0000 UTC m=+6070.196097637" lastFinishedPulling="2025-10-05 08:29:17.576036902 +0000 UTC m=+6079.816889677" observedRunningTime="2025-10-05 08:29:18.458544947 +0000 UTC m=+6080.699397712" watchObservedRunningTime="2025-10-05 08:29:18.459882282 +0000 UTC m=+6080.700735047" Oct 05 08:29:18 crc kubenswrapper[4846]: I1005 08:29:18.509766 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operators/observability-operator-cc5f78dfc-95rb8" Oct 05 08:29:21 crc kubenswrapper[4846]: I1005 08:29:21.260532 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstackclient"] Oct 05 08:29:21 crc kubenswrapper[4846]: I1005 08:29:21.261088 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/openstackclient" podUID="ef74de1f-e213-4ea6-8d5c-9ba3d8416f0b" containerName="openstackclient" containerID="cri-o://d6bf0f646ef44b67def03d565e027cf316116ed0fa05b1e4122c3b048ee80e62" gracePeriod=2 Oct 05 08:29:21 crc kubenswrapper[4846]: I1005 08:29:21.269651 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstackclient"] Oct 05 08:29:21 crc kubenswrapper[4846]: I1005 08:29:21.301770 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Oct 05 08:29:21 crc kubenswrapper[4846]: E1005 08:29:21.302340 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ef74de1f-e213-4ea6-8d5c-9ba3d8416f0b" containerName="openstackclient" Oct 05 08:29:21 crc kubenswrapper[4846]: I1005 08:29:21.302358 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="ef74de1f-e213-4ea6-8d5c-9ba3d8416f0b" containerName="openstackclient" Oct 05 08:29:21 crc kubenswrapper[4846]: I1005 08:29:21.302691 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="ef74de1f-e213-4ea6-8d5c-9ba3d8416f0b" containerName="openstackclient" Oct 05 08:29:21 crc kubenswrapper[4846]: I1005 08:29:21.303421 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 05 08:29:21 crc kubenswrapper[4846]: I1005 08:29:21.309446 4846 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="ef74de1f-e213-4ea6-8d5c-9ba3d8416f0b" podUID="6a40eb6f-d0a6-40c0-a48c-f461d052bb21" Oct 05 08:29:21 crc kubenswrapper[4846]: I1005 08:29:21.330228 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5ttpz\" (UniqueName: \"kubernetes.io/projected/6a40eb6f-d0a6-40c0-a48c-f461d052bb21-kube-api-access-5ttpz\") pod \"openstackclient\" (UID: \"6a40eb6f-d0a6-40c0-a48c-f461d052bb21\") " pod="openstack/openstackclient" Oct 05 08:29:21 crc kubenswrapper[4846]: I1005 08:29:21.332451 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/6a40eb6f-d0a6-40c0-a48c-f461d052bb21-openstack-config-secret\") pod \"openstackclient\" (UID: \"6a40eb6f-d0a6-40c0-a48c-f461d052bb21\") " pod="openstack/openstackclient" Oct 05 08:29:21 crc kubenswrapper[4846]: I1005 08:29:21.332697 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6a40eb6f-d0a6-40c0-a48c-f461d052bb21-combined-ca-bundle\") pod \"openstackclient\" (UID: \"6a40eb6f-d0a6-40c0-a48c-f461d052bb21\") " pod="openstack/openstackclient" Oct 05 08:29:21 crc kubenswrapper[4846]: I1005 08:29:21.332795 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/6a40eb6f-d0a6-40c0-a48c-f461d052bb21-openstack-config\") pod \"openstackclient\" (UID: \"6a40eb6f-d0a6-40c0-a48c-f461d052bb21\") " pod="openstack/openstackclient" Oct 05 08:29:21 crc kubenswrapper[4846]: I1005 08:29:21.330688 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Oct 05 08:29:21 crc kubenswrapper[4846]: I1005 08:29:21.434789 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/6a40eb6f-d0a6-40c0-a48c-f461d052bb21-openstack-config-secret\") pod \"openstackclient\" (UID: \"6a40eb6f-d0a6-40c0-a48c-f461d052bb21\") " pod="openstack/openstackclient" Oct 05 08:29:21 crc kubenswrapper[4846]: I1005 08:29:21.434907 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6a40eb6f-d0a6-40c0-a48c-f461d052bb21-combined-ca-bundle\") pod \"openstackclient\" (UID: \"6a40eb6f-d0a6-40c0-a48c-f461d052bb21\") " pod="openstack/openstackclient" Oct 05 08:29:21 crc kubenswrapper[4846]: I1005 08:29:21.434939 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/6a40eb6f-d0a6-40c0-a48c-f461d052bb21-openstack-config\") pod \"openstackclient\" (UID: \"6a40eb6f-d0a6-40c0-a48c-f461d052bb21\") " pod="openstack/openstackclient" Oct 05 08:29:21 crc kubenswrapper[4846]: I1005 08:29:21.434988 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5ttpz\" (UniqueName: \"kubernetes.io/projected/6a40eb6f-d0a6-40c0-a48c-f461d052bb21-kube-api-access-5ttpz\") pod \"openstackclient\" (UID: \"6a40eb6f-d0a6-40c0-a48c-f461d052bb21\") " pod="openstack/openstackclient" Oct 05 08:29:21 crc kubenswrapper[4846]: I1005 08:29:21.435842 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/6a40eb6f-d0a6-40c0-a48c-f461d052bb21-openstack-config\") pod \"openstackclient\" (UID: \"6a40eb6f-d0a6-40c0-a48c-f461d052bb21\") " pod="openstack/openstackclient" Oct 05 08:29:21 crc kubenswrapper[4846]: I1005 08:29:21.458618 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/6a40eb6f-d0a6-40c0-a48c-f461d052bb21-openstack-config-secret\") pod \"openstackclient\" (UID: \"6a40eb6f-d0a6-40c0-a48c-f461d052bb21\") " pod="openstack/openstackclient" Oct 05 08:29:21 crc kubenswrapper[4846]: I1005 08:29:21.463320 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6a40eb6f-d0a6-40c0-a48c-f461d052bb21-combined-ca-bundle\") pod \"openstackclient\" (UID: \"6a40eb6f-d0a6-40c0-a48c-f461d052bb21\") " pod="openstack/openstackclient" Oct 05 08:29:21 crc kubenswrapper[4846]: I1005 08:29:21.464741 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5ttpz\" (UniqueName: \"kubernetes.io/projected/6a40eb6f-d0a6-40c0-a48c-f461d052bb21-kube-api-access-5ttpz\") pod \"openstackclient\" (UID: \"6a40eb6f-d0a6-40c0-a48c-f461d052bb21\") " pod="openstack/openstackclient" Oct 05 08:29:21 crc kubenswrapper[4846]: I1005 08:29:21.580627 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Oct 05 08:29:21 crc kubenswrapper[4846]: I1005 08:29:21.581864 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 05 08:29:21 crc kubenswrapper[4846]: I1005 08:29:21.583916 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-ceilometer-dockercfg-5lwbk" Oct 05 08:29:21 crc kubenswrapper[4846]: I1005 08:29:21.597809 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 05 08:29:21 crc kubenswrapper[4846]: I1005 08:29:21.655713 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 05 08:29:21 crc kubenswrapper[4846]: I1005 08:29:21.746437 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bvf9c\" (UniqueName: \"kubernetes.io/projected/502e1afd-2dc4-49f5-9c20-a73834426de8-kube-api-access-bvf9c\") pod \"kube-state-metrics-0\" (UID: \"502e1afd-2dc4-49f5-9c20-a73834426de8\") " pod="openstack/kube-state-metrics-0" Oct 05 08:29:21 crc kubenswrapper[4846]: I1005 08:29:21.848513 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bvf9c\" (UniqueName: \"kubernetes.io/projected/502e1afd-2dc4-49f5-9c20-a73834426de8-kube-api-access-bvf9c\") pod \"kube-state-metrics-0\" (UID: \"502e1afd-2dc4-49f5-9c20-a73834426de8\") " pod="openstack/kube-state-metrics-0" Oct 05 08:29:21 crc kubenswrapper[4846]: I1005 08:29:21.908960 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bvf9c\" (UniqueName: \"kubernetes.io/projected/502e1afd-2dc4-49f5-9c20-a73834426de8-kube-api-access-bvf9c\") pod \"kube-state-metrics-0\" (UID: \"502e1afd-2dc4-49f5-9c20-a73834426de8\") " pod="openstack/kube-state-metrics-0" Oct 05 08:29:22 crc kubenswrapper[4846]: I1005 08:29:22.200692 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 05 08:29:22 crc kubenswrapper[4846]: I1005 08:29:22.378119 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/alertmanager-metric-storage-0"] Oct 05 08:29:22 crc kubenswrapper[4846]: I1005 08:29:22.397628 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/alertmanager-metric-storage-0" Oct 05 08:29:22 crc kubenswrapper[4846]: I1005 08:29:22.422327 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"alertmanager-metric-storage-generated" Oct 05 08:29:22 crc kubenswrapper[4846]: I1005 08:29:22.422887 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"alertmanager-metric-storage-tls-assets-0" Oct 05 08:29:22 crc kubenswrapper[4846]: I1005 08:29:22.423026 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"alertmanager-metric-storage-web-config" Oct 05 08:29:22 crc kubenswrapper[4846]: I1005 08:29:22.457665 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"metric-storage-alertmanager-dockercfg-f9z2c" Oct 05 08:29:22 crc kubenswrapper[4846]: I1005 08:29:22.471319 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/e459c893-d9e7-44de-a985-49b29f59f459-tls-assets\") pod \"alertmanager-metric-storage-0\" (UID: \"e459c893-d9e7-44de-a985-49b29f59f459\") " pod="openstack/alertmanager-metric-storage-0" Oct 05 08:29:22 crc kubenswrapper[4846]: I1005 08:29:22.471380 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/e459c893-d9e7-44de-a985-49b29f59f459-config-volume\") pod \"alertmanager-metric-storage-0\" (UID: \"e459c893-d9e7-44de-a985-49b29f59f459\") " pod="openstack/alertmanager-metric-storage-0" Oct 05 08:29:22 crc kubenswrapper[4846]: I1005 08:29:22.471421 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"alertmanager-metric-storage-db\" (UniqueName: \"kubernetes.io/empty-dir/e459c893-d9e7-44de-a985-49b29f59f459-alertmanager-metric-storage-db\") pod \"alertmanager-metric-storage-0\" (UID: \"e459c893-d9e7-44de-a985-49b29f59f459\") " pod="openstack/alertmanager-metric-storage-0" Oct 05 08:29:22 crc kubenswrapper[4846]: I1005 08:29:22.471453 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/e459c893-d9e7-44de-a985-49b29f59f459-web-config\") pod \"alertmanager-metric-storage-0\" (UID: \"e459c893-d9e7-44de-a985-49b29f59f459\") " pod="openstack/alertmanager-metric-storage-0" Oct 05 08:29:22 crc kubenswrapper[4846]: I1005 08:29:22.471473 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/e459c893-d9e7-44de-a985-49b29f59f459-config-out\") pod \"alertmanager-metric-storage-0\" (UID: \"e459c893-d9e7-44de-a985-49b29f59f459\") " pod="openstack/alertmanager-metric-storage-0" Oct 05 08:29:22 crc kubenswrapper[4846]: I1005 08:29:22.471488 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qr5xz\" (UniqueName: \"kubernetes.io/projected/e459c893-d9e7-44de-a985-49b29f59f459-kube-api-access-qr5xz\") pod \"alertmanager-metric-storage-0\" (UID: \"e459c893-d9e7-44de-a985-49b29f59f459\") " pod="openstack/alertmanager-metric-storage-0" Oct 05 08:29:22 crc kubenswrapper[4846]: I1005 08:29:22.472895 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/alertmanager-metric-storage-0"] Oct 05 08:29:22 crc kubenswrapper[4846]: I1005 08:29:22.574301 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/e459c893-d9e7-44de-a985-49b29f59f459-web-config\") pod \"alertmanager-metric-storage-0\" (UID: \"e459c893-d9e7-44de-a985-49b29f59f459\") " pod="openstack/alertmanager-metric-storage-0" Oct 05 08:29:22 crc kubenswrapper[4846]: I1005 08:29:22.574346 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/e459c893-d9e7-44de-a985-49b29f59f459-config-out\") pod \"alertmanager-metric-storage-0\" (UID: \"e459c893-d9e7-44de-a985-49b29f59f459\") " pod="openstack/alertmanager-metric-storage-0" Oct 05 08:29:22 crc kubenswrapper[4846]: I1005 08:29:22.574375 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qr5xz\" (UniqueName: \"kubernetes.io/projected/e459c893-d9e7-44de-a985-49b29f59f459-kube-api-access-qr5xz\") pod \"alertmanager-metric-storage-0\" (UID: \"e459c893-d9e7-44de-a985-49b29f59f459\") " pod="openstack/alertmanager-metric-storage-0" Oct 05 08:29:22 crc kubenswrapper[4846]: I1005 08:29:22.574566 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/e459c893-d9e7-44de-a985-49b29f59f459-tls-assets\") pod \"alertmanager-metric-storage-0\" (UID: \"e459c893-d9e7-44de-a985-49b29f59f459\") " pod="openstack/alertmanager-metric-storage-0" Oct 05 08:29:22 crc kubenswrapper[4846]: I1005 08:29:22.574625 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/e459c893-d9e7-44de-a985-49b29f59f459-config-volume\") pod \"alertmanager-metric-storage-0\" (UID: \"e459c893-d9e7-44de-a985-49b29f59f459\") " pod="openstack/alertmanager-metric-storage-0" Oct 05 08:29:22 crc kubenswrapper[4846]: I1005 08:29:22.574663 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"alertmanager-metric-storage-db\" (UniqueName: \"kubernetes.io/empty-dir/e459c893-d9e7-44de-a985-49b29f59f459-alertmanager-metric-storage-db\") pod \"alertmanager-metric-storage-0\" (UID: \"e459c893-d9e7-44de-a985-49b29f59f459\") " pod="openstack/alertmanager-metric-storage-0" Oct 05 08:29:22 crc kubenswrapper[4846]: I1005 08:29:22.575476 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"alertmanager-metric-storage-db\" (UniqueName: \"kubernetes.io/empty-dir/e459c893-d9e7-44de-a985-49b29f59f459-alertmanager-metric-storage-db\") pod \"alertmanager-metric-storage-0\" (UID: \"e459c893-d9e7-44de-a985-49b29f59f459\") " pod="openstack/alertmanager-metric-storage-0" Oct 05 08:29:22 crc kubenswrapper[4846]: I1005 08:29:22.600944 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/e459c893-d9e7-44de-a985-49b29f59f459-web-config\") pod \"alertmanager-metric-storage-0\" (UID: \"e459c893-d9e7-44de-a985-49b29f59f459\") " pod="openstack/alertmanager-metric-storage-0" Oct 05 08:29:22 crc kubenswrapper[4846]: I1005 08:29:22.607264 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/e459c893-d9e7-44de-a985-49b29f59f459-config-out\") pod \"alertmanager-metric-storage-0\" (UID: \"e459c893-d9e7-44de-a985-49b29f59f459\") " pod="openstack/alertmanager-metric-storage-0" Oct 05 08:29:22 crc kubenswrapper[4846]: I1005 08:29:22.622216 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/e459c893-d9e7-44de-a985-49b29f59f459-tls-assets\") pod \"alertmanager-metric-storage-0\" (UID: \"e459c893-d9e7-44de-a985-49b29f59f459\") " pod="openstack/alertmanager-metric-storage-0" Oct 05 08:29:22 crc kubenswrapper[4846]: I1005 08:29:22.622737 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/e459c893-d9e7-44de-a985-49b29f59f459-config-volume\") pod \"alertmanager-metric-storage-0\" (UID: \"e459c893-d9e7-44de-a985-49b29f59f459\") " pod="openstack/alertmanager-metric-storage-0" Oct 05 08:29:22 crc kubenswrapper[4846]: I1005 08:29:22.642574 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qr5xz\" (UniqueName: \"kubernetes.io/projected/e459c893-d9e7-44de-a985-49b29f59f459-kube-api-access-qr5xz\") pod \"alertmanager-metric-storage-0\" (UID: \"e459c893-d9e7-44de-a985-49b29f59f459\") " pod="openstack/alertmanager-metric-storage-0" Oct 05 08:29:22 crc kubenswrapper[4846]: I1005 08:29:22.813547 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/alertmanager-metric-storage-0" Oct 05 08:29:22 crc kubenswrapper[4846]: I1005 08:29:22.913057 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/prometheus-metric-storage-0"] Oct 05 08:29:22 crc kubenswrapper[4846]: I1005 08:29:22.932405 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Oct 05 08:29:22 crc kubenswrapper[4846]: I1005 08:29:22.936836 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage" Oct 05 08:29:22 crc kubenswrapper[4846]: I1005 08:29:22.937015 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"prometheus-metric-storage-rulefiles-0" Oct 05 08:29:22 crc kubenswrapper[4846]: I1005 08:29:22.941698 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-tls-assets-0" Oct 05 08:29:22 crc kubenswrapper[4846]: I1005 08:29:22.941945 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"metric-storage-prometheus-dockercfg-mkgcg" Oct 05 08:29:22 crc kubenswrapper[4846]: I1005 08:29:22.942049 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-web-config" Oct 05 08:29:22 crc kubenswrapper[4846]: I1005 08:29:22.942158 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-thanos-prometheus-http-client-file" Oct 05 08:29:22 crc kubenswrapper[4846]: I1005 08:29:22.957472 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Oct 05 08:29:23 crc kubenswrapper[4846]: I1005 08:29:22.999449 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Oct 05 08:29:23 crc kubenswrapper[4846]: I1005 08:29:23.095801 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/9eeca81b-6fda-4c22-a221-3adcb9900077-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"9eeca81b-6fda-4c22-a221-3adcb9900077\") " pod="openstack/prometheus-metric-storage-0" Oct 05 08:29:23 crc kubenswrapper[4846]: I1005 08:29:23.095894 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m8pt2\" (UniqueName: \"kubernetes.io/projected/9eeca81b-6fda-4c22-a221-3adcb9900077-kube-api-access-m8pt2\") pod \"prometheus-metric-storage-0\" (UID: \"9eeca81b-6fda-4c22-a221-3adcb9900077\") " pod="openstack/prometheus-metric-storage-0" Oct 05 08:29:23 crc kubenswrapper[4846]: I1005 08:29:23.095926 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/9eeca81b-6fda-4c22-a221-3adcb9900077-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"9eeca81b-6fda-4c22-a221-3adcb9900077\") " pod="openstack/prometheus-metric-storage-0" Oct 05 08:29:23 crc kubenswrapper[4846]: I1005 08:29:23.095957 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/9eeca81b-6fda-4c22-a221-3adcb9900077-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"9eeca81b-6fda-4c22-a221-3adcb9900077\") " pod="openstack/prometheus-metric-storage-0" Oct 05 08:29:23 crc kubenswrapper[4846]: I1005 08:29:23.096032 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/9eeca81b-6fda-4c22-a221-3adcb9900077-config\") pod \"prometheus-metric-storage-0\" (UID: \"9eeca81b-6fda-4c22-a221-3adcb9900077\") " pod="openstack/prometheus-metric-storage-0" Oct 05 08:29:23 crc kubenswrapper[4846]: I1005 08:29:23.096055 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/9eeca81b-6fda-4c22-a221-3adcb9900077-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"9eeca81b-6fda-4c22-a221-3adcb9900077\") " pod="openstack/prometheus-metric-storage-0" Oct 05 08:29:23 crc kubenswrapper[4846]: I1005 08:29:23.096134 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-1cc7951e-b4e7-48bb-b9fd-de1baa2910d8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-1cc7951e-b4e7-48bb-b9fd-de1baa2910d8\") pod \"prometheus-metric-storage-0\" (UID: \"9eeca81b-6fda-4c22-a221-3adcb9900077\") " pod="openstack/prometheus-metric-storage-0" Oct 05 08:29:23 crc kubenswrapper[4846]: I1005 08:29:23.096158 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/9eeca81b-6fda-4c22-a221-3adcb9900077-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"9eeca81b-6fda-4c22-a221-3adcb9900077\") " pod="openstack/prometheus-metric-storage-0" Oct 05 08:29:23 crc kubenswrapper[4846]: I1005 08:29:23.172065 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 05 08:29:23 crc kubenswrapper[4846]: I1005 08:29:23.199787 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/9eeca81b-6fda-4c22-a221-3adcb9900077-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"9eeca81b-6fda-4c22-a221-3adcb9900077\") " pod="openstack/prometheus-metric-storage-0" Oct 05 08:29:23 crc kubenswrapper[4846]: I1005 08:29:23.199875 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m8pt2\" (UniqueName: \"kubernetes.io/projected/9eeca81b-6fda-4c22-a221-3adcb9900077-kube-api-access-m8pt2\") pod \"prometheus-metric-storage-0\" (UID: \"9eeca81b-6fda-4c22-a221-3adcb9900077\") " pod="openstack/prometheus-metric-storage-0" Oct 05 08:29:23 crc kubenswrapper[4846]: I1005 08:29:23.199894 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/9eeca81b-6fda-4c22-a221-3adcb9900077-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"9eeca81b-6fda-4c22-a221-3adcb9900077\") " pod="openstack/prometheus-metric-storage-0" Oct 05 08:29:23 crc kubenswrapper[4846]: I1005 08:29:23.199918 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/9eeca81b-6fda-4c22-a221-3adcb9900077-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"9eeca81b-6fda-4c22-a221-3adcb9900077\") " pod="openstack/prometheus-metric-storage-0" Oct 05 08:29:23 crc kubenswrapper[4846]: I1005 08:29:23.199991 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/9eeca81b-6fda-4c22-a221-3adcb9900077-config\") pod \"prometheus-metric-storage-0\" (UID: \"9eeca81b-6fda-4c22-a221-3adcb9900077\") " pod="openstack/prometheus-metric-storage-0" Oct 05 08:29:23 crc kubenswrapper[4846]: I1005 08:29:23.200013 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/9eeca81b-6fda-4c22-a221-3adcb9900077-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"9eeca81b-6fda-4c22-a221-3adcb9900077\") " pod="openstack/prometheus-metric-storage-0" Oct 05 08:29:23 crc kubenswrapper[4846]: I1005 08:29:23.200095 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-1cc7951e-b4e7-48bb-b9fd-de1baa2910d8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-1cc7951e-b4e7-48bb-b9fd-de1baa2910d8\") pod \"prometheus-metric-storage-0\" (UID: \"9eeca81b-6fda-4c22-a221-3adcb9900077\") " pod="openstack/prometheus-metric-storage-0" Oct 05 08:29:23 crc kubenswrapper[4846]: I1005 08:29:23.200114 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/9eeca81b-6fda-4c22-a221-3adcb9900077-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"9eeca81b-6fda-4c22-a221-3adcb9900077\") " pod="openstack/prometheus-metric-storage-0" Oct 05 08:29:23 crc kubenswrapper[4846]: I1005 08:29:23.202760 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/9eeca81b-6fda-4c22-a221-3adcb9900077-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"9eeca81b-6fda-4c22-a221-3adcb9900077\") " pod="openstack/prometheus-metric-storage-0" Oct 05 08:29:23 crc kubenswrapper[4846]: I1005 08:29:23.222711 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/9eeca81b-6fda-4c22-a221-3adcb9900077-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"9eeca81b-6fda-4c22-a221-3adcb9900077\") " pod="openstack/prometheus-metric-storage-0" Oct 05 08:29:23 crc kubenswrapper[4846]: I1005 08:29:23.227370 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/9eeca81b-6fda-4c22-a221-3adcb9900077-config\") pod \"prometheus-metric-storage-0\" (UID: \"9eeca81b-6fda-4c22-a221-3adcb9900077\") " pod="openstack/prometheus-metric-storage-0" Oct 05 08:29:23 crc kubenswrapper[4846]: I1005 08:29:23.231893 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/9eeca81b-6fda-4c22-a221-3adcb9900077-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"9eeca81b-6fda-4c22-a221-3adcb9900077\") " pod="openstack/prometheus-metric-storage-0" Oct 05 08:29:23 crc kubenswrapper[4846]: I1005 08:29:23.236243 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/9eeca81b-6fda-4c22-a221-3adcb9900077-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"9eeca81b-6fda-4c22-a221-3adcb9900077\") " pod="openstack/prometheus-metric-storage-0" Oct 05 08:29:23 crc kubenswrapper[4846]: I1005 08:29:23.237808 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/9eeca81b-6fda-4c22-a221-3adcb9900077-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"9eeca81b-6fda-4c22-a221-3adcb9900077\") " pod="openstack/prometheus-metric-storage-0" Oct 05 08:29:23 crc kubenswrapper[4846]: I1005 08:29:23.262017 4846 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 05 08:29:23 crc kubenswrapper[4846]: I1005 08:29:23.262069 4846 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-1cc7951e-b4e7-48bb-b9fd-de1baa2910d8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-1cc7951e-b4e7-48bb-b9fd-de1baa2910d8\") pod \"prometheus-metric-storage-0\" (UID: \"9eeca81b-6fda-4c22-a221-3adcb9900077\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/0ec9240d9ee399c4d378df6ba11f50328a91e5e421b3e1fde811ca7b96ace4eb/globalmount\"" pod="openstack/prometheus-metric-storage-0" Oct 05 08:29:23 crc kubenswrapper[4846]: I1005 08:29:23.262686 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m8pt2\" (UniqueName: \"kubernetes.io/projected/9eeca81b-6fda-4c22-a221-3adcb9900077-kube-api-access-m8pt2\") pod \"prometheus-metric-storage-0\" (UID: \"9eeca81b-6fda-4c22-a221-3adcb9900077\") " pod="openstack/prometheus-metric-storage-0" Oct 05 08:29:23 crc kubenswrapper[4846]: I1005 08:29:23.332661 4846 patch_prober.go:28] interesting pod/machine-config-daemon-fscvf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 08:29:23 crc kubenswrapper[4846]: I1005 08:29:23.333053 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 08:29:23 crc kubenswrapper[4846]: I1005 08:29:23.499571 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-1cc7951e-b4e7-48bb-b9fd-de1baa2910d8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-1cc7951e-b4e7-48bb-b9fd-de1baa2910d8\") pod \"prometheus-metric-storage-0\" (UID: \"9eeca81b-6fda-4c22-a221-3adcb9900077\") " pod="openstack/prometheus-metric-storage-0" Oct 05 08:29:23 crc kubenswrapper[4846]: I1005 08:29:23.551889 4846 generic.go:334] "Generic (PLEG): container finished" podID="ef74de1f-e213-4ea6-8d5c-9ba3d8416f0b" containerID="d6bf0f646ef44b67def03d565e027cf316116ed0fa05b1e4122c3b048ee80e62" exitCode=137 Oct 05 08:29:23 crc kubenswrapper[4846]: I1005 08:29:23.557927 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"6a40eb6f-d0a6-40c0-a48c-f461d052bb21","Type":"ContainerStarted","Data":"943fa265e3fd9c54b0b5f2ea6ddc2579681c1ef36e9f3b1078f0edcbe9f75c9f"} Oct 05 08:29:23 crc kubenswrapper[4846]: I1005 08:29:23.559527 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"502e1afd-2dc4-49f5-9c20-a73834426de8","Type":"ContainerStarted","Data":"8e3301baf37fa9917c8d8f20a2aad6fcaa7c19dd723cf3e6ba184816ffd05cb5"} Oct 05 08:29:23 crc kubenswrapper[4846]: I1005 08:29:23.621390 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Oct 05 08:29:23 crc kubenswrapper[4846]: I1005 08:29:23.672985 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/alertmanager-metric-storage-0"] Oct 05 08:29:23 crc kubenswrapper[4846]: I1005 08:29:23.983490 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 05 08:29:24 crc kubenswrapper[4846]: I1005 08:29:24.032907 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef74de1f-e213-4ea6-8d5c-9ba3d8416f0b-combined-ca-bundle\") pod \"ef74de1f-e213-4ea6-8d5c-9ba3d8416f0b\" (UID: \"ef74de1f-e213-4ea6-8d5c-9ba3d8416f0b\") " Oct 05 08:29:24 crc kubenswrapper[4846]: I1005 08:29:24.033110 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kjvfz\" (UniqueName: \"kubernetes.io/projected/ef74de1f-e213-4ea6-8d5c-9ba3d8416f0b-kube-api-access-kjvfz\") pod \"ef74de1f-e213-4ea6-8d5c-9ba3d8416f0b\" (UID: \"ef74de1f-e213-4ea6-8d5c-9ba3d8416f0b\") " Oct 05 08:29:24 crc kubenswrapper[4846]: I1005 08:29:24.033240 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/ef74de1f-e213-4ea6-8d5c-9ba3d8416f0b-openstack-config-secret\") pod \"ef74de1f-e213-4ea6-8d5c-9ba3d8416f0b\" (UID: \"ef74de1f-e213-4ea6-8d5c-9ba3d8416f0b\") " Oct 05 08:29:24 crc kubenswrapper[4846]: I1005 08:29:24.033805 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/ef74de1f-e213-4ea6-8d5c-9ba3d8416f0b-openstack-config\") pod \"ef74de1f-e213-4ea6-8d5c-9ba3d8416f0b\" (UID: \"ef74de1f-e213-4ea6-8d5c-9ba3d8416f0b\") " Oct 05 08:29:24 crc kubenswrapper[4846]: I1005 08:29:24.042689 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ef74de1f-e213-4ea6-8d5c-9ba3d8416f0b-kube-api-access-kjvfz" (OuterVolumeSpecName: "kube-api-access-kjvfz") pod "ef74de1f-e213-4ea6-8d5c-9ba3d8416f0b" (UID: "ef74de1f-e213-4ea6-8d5c-9ba3d8416f0b"). InnerVolumeSpecName "kube-api-access-kjvfz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:29:24 crc kubenswrapper[4846]: I1005 08:29:24.074990 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ef74de1f-e213-4ea6-8d5c-9ba3d8416f0b-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "ef74de1f-e213-4ea6-8d5c-9ba3d8416f0b" (UID: "ef74de1f-e213-4ea6-8d5c-9ba3d8416f0b"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 08:29:24 crc kubenswrapper[4846]: I1005 08:29:24.077153 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ef74de1f-e213-4ea6-8d5c-9ba3d8416f0b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ef74de1f-e213-4ea6-8d5c-9ba3d8416f0b" (UID: "ef74de1f-e213-4ea6-8d5c-9ba3d8416f0b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:29:24 crc kubenswrapper[4846]: I1005 08:29:24.126353 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ef74de1f-e213-4ea6-8d5c-9ba3d8416f0b-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "ef74de1f-e213-4ea6-8d5c-9ba3d8416f0b" (UID: "ef74de1f-e213-4ea6-8d5c-9ba3d8416f0b"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:29:24 crc kubenswrapper[4846]: I1005 08:29:24.136063 4846 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/ef74de1f-e213-4ea6-8d5c-9ba3d8416f0b-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Oct 05 08:29:24 crc kubenswrapper[4846]: I1005 08:29:24.136109 4846 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/ef74de1f-e213-4ea6-8d5c-9ba3d8416f0b-openstack-config\") on node \"crc\" DevicePath \"\"" Oct 05 08:29:24 crc kubenswrapper[4846]: I1005 08:29:24.136119 4846 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef74de1f-e213-4ea6-8d5c-9ba3d8416f0b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 08:29:24 crc kubenswrapper[4846]: I1005 08:29:24.136128 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kjvfz\" (UniqueName: \"kubernetes.io/projected/ef74de1f-e213-4ea6-8d5c-9ba3d8416f0b-kube-api-access-kjvfz\") on node \"crc\" DevicePath \"\"" Oct 05 08:29:24 crc kubenswrapper[4846]: I1005 08:29:24.364033 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Oct 05 08:29:24 crc kubenswrapper[4846]: I1005 08:29:24.529756 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ef74de1f-e213-4ea6-8d5c-9ba3d8416f0b" path="/var/lib/kubelet/pods/ef74de1f-e213-4ea6-8d5c-9ba3d8416f0b/volumes" Oct 05 08:29:24 crc kubenswrapper[4846]: I1005 08:29:24.583365 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/alertmanager-metric-storage-0" event={"ID":"e459c893-d9e7-44de-a985-49b29f59f459","Type":"ContainerStarted","Data":"a20bb10ecbe2f40215699852679873a8f70197aec506516408aaf64616e9a33a"} Oct 05 08:29:24 crc kubenswrapper[4846]: I1005 08:29:24.585291 4846 scope.go:117] "RemoveContainer" containerID="d6bf0f646ef44b67def03d565e027cf316116ed0fa05b1e4122c3b048ee80e62" Oct 05 08:29:24 crc kubenswrapper[4846]: I1005 08:29:24.585300 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 05 08:29:24 crc kubenswrapper[4846]: I1005 08:29:24.593641 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"6a40eb6f-d0a6-40c0-a48c-f461d052bb21","Type":"ContainerStarted","Data":"f5d45d11f84632d3e6cb961ad6e25fff9c7da4c7fbcfbf87f4f57154de66649e"} Oct 05 08:29:24 crc kubenswrapper[4846]: I1005 08:29:24.602021 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"9eeca81b-6fda-4c22-a221-3adcb9900077","Type":"ContainerStarted","Data":"e8538609494690d52f4d579e3ae6523d65701c417f2551cb9f7b3cad911741a5"} Oct 05 08:29:24 crc kubenswrapper[4846]: I1005 08:29:24.603183 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"502e1afd-2dc4-49f5-9c20-a73834426de8","Type":"ContainerStarted","Data":"584a6806d2a36bfd7832c05c23cd2ef062e16cae388ebc0095dfa18e6efd8aae"} Oct 05 08:29:24 crc kubenswrapper[4846]: I1005 08:29:24.604129 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Oct 05 08:29:24 crc kubenswrapper[4846]: I1005 08:29:24.618161 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=3.618143105 podStartE2EDuration="3.618143105s" podCreationTimestamp="2025-10-05 08:29:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 08:29:24.611109388 +0000 UTC m=+6086.851962153" watchObservedRunningTime="2025-10-05 08:29:24.618143105 +0000 UTC m=+6086.858995880" Oct 05 08:29:24 crc kubenswrapper[4846]: I1005 08:29:24.656328 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=3.183511844 podStartE2EDuration="3.656302808s" podCreationTimestamp="2025-10-05 08:29:21 +0000 UTC" firstStartedPulling="2025-10-05 08:29:23.252971473 +0000 UTC m=+6085.493824248" lastFinishedPulling="2025-10-05 08:29:23.725762447 +0000 UTC m=+6085.966615212" observedRunningTime="2025-10-05 08:29:24.645876521 +0000 UTC m=+6086.886729296" watchObservedRunningTime="2025-10-05 08:29:24.656302808 +0000 UTC m=+6086.897155583" Oct 05 08:29:31 crc kubenswrapper[4846]: I1005 08:29:31.706263 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/alertmanager-metric-storage-0" event={"ID":"e459c893-d9e7-44de-a985-49b29f59f459","Type":"ContainerStarted","Data":"320d6ae2e0691e4a7e888b0fd3e41b2475ed3a147eb870219c076f5b2019f777"} Oct 05 08:29:31 crc kubenswrapper[4846]: I1005 08:29:31.708255 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"9eeca81b-6fda-4c22-a221-3adcb9900077","Type":"ContainerStarted","Data":"387ea30ec891b63656dcd74018a0d5d1f86fea2b6f29a8bbc972f659e551aa35"} Oct 05 08:29:32 crc kubenswrapper[4846]: I1005 08:29:32.206224 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Oct 05 08:29:33 crc kubenswrapper[4846]: I1005 08:29:33.875738 4846 scope.go:117] "RemoveContainer" containerID="0bfea0204492bee3feb5c3ffb076b60bd6f751b3c236782fcfad5d9ec9d3c52f" Oct 05 08:29:33 crc kubenswrapper[4846]: I1005 08:29:33.920450 4846 scope.go:117] "RemoveContainer" containerID="64a7de2a593c540a6acafc8f361280fb2d3e1a2d97fb1d267443feeb2ea268ae" Oct 05 08:29:36 crc kubenswrapper[4846]: I1005 08:29:36.032056 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-sync-4dvh4"] Oct 05 08:29:36 crc kubenswrapper[4846]: I1005 08:29:36.041369 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-sync-4dvh4"] Oct 05 08:29:36 crc kubenswrapper[4846]: I1005 08:29:36.513485 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d91d2d9a-0d5c-4bab-bae1-5b428da00095" path="/var/lib/kubelet/pods/d91d2d9a-0d5c-4bab-bae1-5b428da00095/volumes" Oct 05 08:29:38 crc kubenswrapper[4846]: I1005 08:29:38.799544 4846 generic.go:334] "Generic (PLEG): container finished" podID="9eeca81b-6fda-4c22-a221-3adcb9900077" containerID="387ea30ec891b63656dcd74018a0d5d1f86fea2b6f29a8bbc972f659e551aa35" exitCode=0 Oct 05 08:29:38 crc kubenswrapper[4846]: I1005 08:29:38.799627 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"9eeca81b-6fda-4c22-a221-3adcb9900077","Type":"ContainerDied","Data":"387ea30ec891b63656dcd74018a0d5d1f86fea2b6f29a8bbc972f659e551aa35"} Oct 05 08:29:39 crc kubenswrapper[4846]: I1005 08:29:39.815408 4846 generic.go:334] "Generic (PLEG): container finished" podID="e459c893-d9e7-44de-a985-49b29f59f459" containerID="320d6ae2e0691e4a7e888b0fd3e41b2475ed3a147eb870219c076f5b2019f777" exitCode=0 Oct 05 08:29:39 crc kubenswrapper[4846]: I1005 08:29:39.815467 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/alertmanager-metric-storage-0" event={"ID":"e459c893-d9e7-44de-a985-49b29f59f459","Type":"ContainerDied","Data":"320d6ae2e0691e4a7e888b0fd3e41b2475ed3a147eb870219c076f5b2019f777"} Oct 05 08:29:43 crc kubenswrapper[4846]: I1005 08:29:43.524368 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-9lkxx"] Oct 05 08:29:43 crc kubenswrapper[4846]: I1005 08:29:43.550442 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-9lkxx"] Oct 05 08:29:43 crc kubenswrapper[4846]: I1005 08:29:43.550548 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9lkxx" Oct 05 08:29:43 crc kubenswrapper[4846]: I1005 08:29:43.583723 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7bc1dff8-1996-4278-89f2-1004c67562c3-utilities\") pod \"certified-operators-9lkxx\" (UID: \"7bc1dff8-1996-4278-89f2-1004c67562c3\") " pod="openshift-marketplace/certified-operators-9lkxx" Oct 05 08:29:43 crc kubenswrapper[4846]: I1005 08:29:43.583780 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vbzrq\" (UniqueName: \"kubernetes.io/projected/7bc1dff8-1996-4278-89f2-1004c67562c3-kube-api-access-vbzrq\") pod \"certified-operators-9lkxx\" (UID: \"7bc1dff8-1996-4278-89f2-1004c67562c3\") " pod="openshift-marketplace/certified-operators-9lkxx" Oct 05 08:29:43 crc kubenswrapper[4846]: I1005 08:29:43.584037 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7bc1dff8-1996-4278-89f2-1004c67562c3-catalog-content\") pod \"certified-operators-9lkxx\" (UID: \"7bc1dff8-1996-4278-89f2-1004c67562c3\") " pod="openshift-marketplace/certified-operators-9lkxx" Oct 05 08:29:43 crc kubenswrapper[4846]: I1005 08:29:43.686338 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7bc1dff8-1996-4278-89f2-1004c67562c3-catalog-content\") pod \"certified-operators-9lkxx\" (UID: \"7bc1dff8-1996-4278-89f2-1004c67562c3\") " pod="openshift-marketplace/certified-operators-9lkxx" Oct 05 08:29:43 crc kubenswrapper[4846]: I1005 08:29:43.686418 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7bc1dff8-1996-4278-89f2-1004c67562c3-utilities\") pod \"certified-operators-9lkxx\" (UID: \"7bc1dff8-1996-4278-89f2-1004c67562c3\") " pod="openshift-marketplace/certified-operators-9lkxx" Oct 05 08:29:43 crc kubenswrapper[4846]: I1005 08:29:43.686600 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vbzrq\" (UniqueName: \"kubernetes.io/projected/7bc1dff8-1996-4278-89f2-1004c67562c3-kube-api-access-vbzrq\") pod \"certified-operators-9lkxx\" (UID: \"7bc1dff8-1996-4278-89f2-1004c67562c3\") " pod="openshift-marketplace/certified-operators-9lkxx" Oct 05 08:29:43 crc kubenswrapper[4846]: I1005 08:29:43.687505 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7bc1dff8-1996-4278-89f2-1004c67562c3-utilities\") pod \"certified-operators-9lkxx\" (UID: \"7bc1dff8-1996-4278-89f2-1004c67562c3\") " pod="openshift-marketplace/certified-operators-9lkxx" Oct 05 08:29:43 crc kubenswrapper[4846]: I1005 08:29:43.688256 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7bc1dff8-1996-4278-89f2-1004c67562c3-catalog-content\") pod \"certified-operators-9lkxx\" (UID: \"7bc1dff8-1996-4278-89f2-1004c67562c3\") " pod="openshift-marketplace/certified-operators-9lkxx" Oct 05 08:29:43 crc kubenswrapper[4846]: I1005 08:29:43.713765 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vbzrq\" (UniqueName: \"kubernetes.io/projected/7bc1dff8-1996-4278-89f2-1004c67562c3-kube-api-access-vbzrq\") pod \"certified-operators-9lkxx\" (UID: \"7bc1dff8-1996-4278-89f2-1004c67562c3\") " pod="openshift-marketplace/certified-operators-9lkxx" Oct 05 08:29:43 crc kubenswrapper[4846]: I1005 08:29:43.872789 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9lkxx" Oct 05 08:29:45 crc kubenswrapper[4846]: I1005 08:29:45.888673 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-9lkxx"] Oct 05 08:29:45 crc kubenswrapper[4846]: I1005 08:29:45.931486 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"9eeca81b-6fda-4c22-a221-3adcb9900077","Type":"ContainerStarted","Data":"35860887d251721dfd419731dae2306db24d39a9d93d8f1543f1fe5315075bf9"} Oct 05 08:29:45 crc kubenswrapper[4846]: I1005 08:29:45.949564 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/alertmanager-metric-storage-0" event={"ID":"e459c893-d9e7-44de-a985-49b29f59f459","Type":"ContainerStarted","Data":"03e6518628b70361027b4721ae8596aff8b612198ebbdf427ad5d16057b15598"} Oct 05 08:29:46 crc kubenswrapper[4846]: I1005 08:29:46.963866 4846 generic.go:334] "Generic (PLEG): container finished" podID="7bc1dff8-1996-4278-89f2-1004c67562c3" containerID="99ae8d8df0f28c4cd815a55eb43c6ca1128f1d885aa854b5752ba4327d2fdb63" exitCode=0 Oct 05 08:29:46 crc kubenswrapper[4846]: I1005 08:29:46.963932 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9lkxx" event={"ID":"7bc1dff8-1996-4278-89f2-1004c67562c3","Type":"ContainerDied","Data":"99ae8d8df0f28c4cd815a55eb43c6ca1128f1d885aa854b5752ba4327d2fdb63"} Oct 05 08:29:46 crc kubenswrapper[4846]: I1005 08:29:46.964285 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9lkxx" event={"ID":"7bc1dff8-1996-4278-89f2-1004c67562c3","Type":"ContainerStarted","Data":"ad318fab82dac6c13b92049619676842c54b0e99733b9328891c31101c44fe33"} Oct 05 08:29:48 crc kubenswrapper[4846]: I1005 08:29:48.991891 4846 generic.go:334] "Generic (PLEG): container finished" podID="7bc1dff8-1996-4278-89f2-1004c67562c3" containerID="26e9a6b7fb2f28262ee8b8451f374c4fdba5c3a9bc40446b7bb2355af2dfdc7c" exitCode=0 Oct 05 08:29:48 crc kubenswrapper[4846]: I1005 08:29:48.991998 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9lkxx" event={"ID":"7bc1dff8-1996-4278-89f2-1004c67562c3","Type":"ContainerDied","Data":"26e9a6b7fb2f28262ee8b8451f374c4fdba5c3a9bc40446b7bb2355af2dfdc7c"} Oct 05 08:29:51 crc kubenswrapper[4846]: I1005 08:29:51.013718 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/alertmanager-metric-storage-0" event={"ID":"e459c893-d9e7-44de-a985-49b29f59f459","Type":"ContainerStarted","Data":"aa0a8256d860c7552842f529ed2158ccb5371c68340f15c924edb3e55cbc0952"} Oct 05 08:29:51 crc kubenswrapper[4846]: I1005 08:29:51.014281 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/alertmanager-metric-storage-0" Oct 05 08:29:51 crc kubenswrapper[4846]: I1005 08:29:51.017430 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/alertmanager-metric-storage-0" Oct 05 08:29:51 crc kubenswrapper[4846]: I1005 08:29:51.018131 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9lkxx" event={"ID":"7bc1dff8-1996-4278-89f2-1004c67562c3","Type":"ContainerStarted","Data":"4fa9df62f6f1fdf926b083d1d0191ae3233a17658d3b18a6621d988223c7343b"} Oct 05 08:29:51 crc kubenswrapper[4846]: I1005 08:29:51.020567 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"9eeca81b-6fda-4c22-a221-3adcb9900077","Type":"ContainerStarted","Data":"5cb1120abfdcdf858d882f7c6f47445cb49f210041a4a9486cda3b94e9ad8438"} Oct 05 08:29:51 crc kubenswrapper[4846]: I1005 08:29:51.051339 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/alertmanager-metric-storage-0" podStartSLOduration=7.458920142 podStartE2EDuration="29.051319275s" podCreationTimestamp="2025-10-05 08:29:22 +0000 UTC" firstStartedPulling="2025-10-05 08:29:23.723558879 +0000 UTC m=+6085.964411654" lastFinishedPulling="2025-10-05 08:29:45.315957992 +0000 UTC m=+6107.556810787" observedRunningTime="2025-10-05 08:29:51.044424122 +0000 UTC m=+6113.285276917" watchObservedRunningTime="2025-10-05 08:29:51.051319275 +0000 UTC m=+6113.292172050" Oct 05 08:29:51 crc kubenswrapper[4846]: I1005 08:29:51.070714 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-9lkxx" podStartSLOduration=4.995205041 podStartE2EDuration="8.07069454s" podCreationTimestamp="2025-10-05 08:29:43 +0000 UTC" firstStartedPulling="2025-10-05 08:29:46.966521644 +0000 UTC m=+6109.207374429" lastFinishedPulling="2025-10-05 08:29:50.042011143 +0000 UTC m=+6112.282863928" observedRunningTime="2025-10-05 08:29:51.064191447 +0000 UTC m=+6113.305044222" watchObservedRunningTime="2025-10-05 08:29:51.07069454 +0000 UTC m=+6113.311547315" Oct 05 08:29:53 crc kubenswrapper[4846]: I1005 08:29:53.046325 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"9eeca81b-6fda-4c22-a221-3adcb9900077","Type":"ContainerStarted","Data":"855c10a9b75c2c9926f5dd593debd83762c84203e420e5b91cfe42e7db8b6559"} Oct 05 08:29:53 crc kubenswrapper[4846]: I1005 08:29:53.085016 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/prometheus-metric-storage-0" podStartSLOduration=3.838919448 podStartE2EDuration="32.084987389s" podCreationTimestamp="2025-10-05 08:29:21 +0000 UTC" firstStartedPulling="2025-10-05 08:29:24.448392477 +0000 UTC m=+6086.689245252" lastFinishedPulling="2025-10-05 08:29:52.694460418 +0000 UTC m=+6114.935313193" observedRunningTime="2025-10-05 08:29:53.077733146 +0000 UTC m=+6115.318585921" watchObservedRunningTime="2025-10-05 08:29:53.084987389 +0000 UTC m=+6115.325840184" Oct 05 08:29:53 crc kubenswrapper[4846]: I1005 08:29:53.325443 4846 patch_prober.go:28] interesting pod/machine-config-daemon-fscvf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 08:29:53 crc kubenswrapper[4846]: I1005 08:29:53.325501 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 08:29:53 crc kubenswrapper[4846]: I1005 08:29:53.622424 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/prometheus-metric-storage-0" Oct 05 08:29:53 crc kubenswrapper[4846]: I1005 08:29:53.624790 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/prometheus-metric-storage-0" Oct 05 08:29:53 crc kubenswrapper[4846]: I1005 08:29:53.626049 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/prometheus-metric-storage-0" Oct 05 08:29:53 crc kubenswrapper[4846]: I1005 08:29:53.874440 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-9lkxx" Oct 05 08:29:53 crc kubenswrapper[4846]: I1005 08:29:53.874517 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-9lkxx" Oct 05 08:29:53 crc kubenswrapper[4846]: I1005 08:29:53.954616 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-9lkxx" Oct 05 08:29:54 crc kubenswrapper[4846]: I1005 08:29:54.057410 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/prometheus-metric-storage-0" Oct 05 08:29:55 crc kubenswrapper[4846]: I1005 08:29:55.455638 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstackclient"] Oct 05 08:29:55 crc kubenswrapper[4846]: I1005 08:29:55.456215 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/openstackclient" podUID="6a40eb6f-d0a6-40c0-a48c-f461d052bb21" containerName="openstackclient" containerID="cri-o://f5d45d11f84632d3e6cb961ad6e25fff9c7da4c7fbcfbf87f4f57154de66649e" gracePeriod=2 Oct 05 08:29:55 crc kubenswrapper[4846]: I1005 08:29:55.483136 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstackclient"] Oct 05 08:29:55 crc kubenswrapper[4846]: I1005 08:29:55.505425 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Oct 05 08:29:55 crc kubenswrapper[4846]: E1005 08:29:55.506070 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6a40eb6f-d0a6-40c0-a48c-f461d052bb21" containerName="openstackclient" Oct 05 08:29:55 crc kubenswrapper[4846]: I1005 08:29:55.506116 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="6a40eb6f-d0a6-40c0-a48c-f461d052bb21" containerName="openstackclient" Oct 05 08:29:55 crc kubenswrapper[4846]: I1005 08:29:55.507034 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="6a40eb6f-d0a6-40c0-a48c-f461d052bb21" containerName="openstackclient" Oct 05 08:29:55 crc kubenswrapper[4846]: I1005 08:29:55.509212 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 05 08:29:55 crc kubenswrapper[4846]: I1005 08:29:55.512360 4846 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="6a40eb6f-d0a6-40c0-a48c-f461d052bb21" podUID="e45f7b43-2271-4c62-a0b5-a66462317e84" Oct 05 08:29:55 crc kubenswrapper[4846]: I1005 08:29:55.520764 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Oct 05 08:29:55 crc kubenswrapper[4846]: I1005 08:29:55.690872 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5gszq\" (UniqueName: \"kubernetes.io/projected/e45f7b43-2271-4c62-a0b5-a66462317e84-kube-api-access-5gszq\") pod \"openstackclient\" (UID: \"e45f7b43-2271-4c62-a0b5-a66462317e84\") " pod="openstack/openstackclient" Oct 05 08:29:55 crc kubenswrapper[4846]: I1005 08:29:55.690916 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/e45f7b43-2271-4c62-a0b5-a66462317e84-openstack-config-secret\") pod \"openstackclient\" (UID: \"e45f7b43-2271-4c62-a0b5-a66462317e84\") " pod="openstack/openstackclient" Oct 05 08:29:55 crc kubenswrapper[4846]: I1005 08:29:55.690940 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e45f7b43-2271-4c62-a0b5-a66462317e84-combined-ca-bundle\") pod \"openstackclient\" (UID: \"e45f7b43-2271-4c62-a0b5-a66462317e84\") " pod="openstack/openstackclient" Oct 05 08:29:55 crc kubenswrapper[4846]: I1005 08:29:55.690992 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/e45f7b43-2271-4c62-a0b5-a66462317e84-openstack-config\") pod \"openstackclient\" (UID: \"e45f7b43-2271-4c62-a0b5-a66462317e84\") " pod="openstack/openstackclient" Oct 05 08:29:55 crc kubenswrapper[4846]: I1005 08:29:55.793101 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5gszq\" (UniqueName: \"kubernetes.io/projected/e45f7b43-2271-4c62-a0b5-a66462317e84-kube-api-access-5gszq\") pod \"openstackclient\" (UID: \"e45f7b43-2271-4c62-a0b5-a66462317e84\") " pod="openstack/openstackclient" Oct 05 08:29:55 crc kubenswrapper[4846]: I1005 08:29:55.793371 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/e45f7b43-2271-4c62-a0b5-a66462317e84-openstack-config-secret\") pod \"openstackclient\" (UID: \"e45f7b43-2271-4c62-a0b5-a66462317e84\") " pod="openstack/openstackclient" Oct 05 08:29:55 crc kubenswrapper[4846]: I1005 08:29:55.793457 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e45f7b43-2271-4c62-a0b5-a66462317e84-combined-ca-bundle\") pod \"openstackclient\" (UID: \"e45f7b43-2271-4c62-a0b5-a66462317e84\") " pod="openstack/openstackclient" Oct 05 08:29:55 crc kubenswrapper[4846]: I1005 08:29:55.793605 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/e45f7b43-2271-4c62-a0b5-a66462317e84-openstack-config\") pod \"openstackclient\" (UID: \"e45f7b43-2271-4c62-a0b5-a66462317e84\") " pod="openstack/openstackclient" Oct 05 08:29:55 crc kubenswrapper[4846]: I1005 08:29:55.794480 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/e45f7b43-2271-4c62-a0b5-a66462317e84-openstack-config\") pod \"openstackclient\" (UID: \"e45f7b43-2271-4c62-a0b5-a66462317e84\") " pod="openstack/openstackclient" Oct 05 08:29:55 crc kubenswrapper[4846]: I1005 08:29:55.798705 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e45f7b43-2271-4c62-a0b5-a66462317e84-combined-ca-bundle\") pod \"openstackclient\" (UID: \"e45f7b43-2271-4c62-a0b5-a66462317e84\") " pod="openstack/openstackclient" Oct 05 08:29:55 crc kubenswrapper[4846]: I1005 08:29:55.806883 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/e45f7b43-2271-4c62-a0b5-a66462317e84-openstack-config-secret\") pod \"openstackclient\" (UID: \"e45f7b43-2271-4c62-a0b5-a66462317e84\") " pod="openstack/openstackclient" Oct 05 08:29:55 crc kubenswrapper[4846]: I1005 08:29:55.824918 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5gszq\" (UniqueName: \"kubernetes.io/projected/e45f7b43-2271-4c62-a0b5-a66462317e84-kube-api-access-5gszq\") pod \"openstackclient\" (UID: \"e45f7b43-2271-4c62-a0b5-a66462317e84\") " pod="openstack/openstackclient" Oct 05 08:29:55 crc kubenswrapper[4846]: I1005 08:29:55.852477 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 05 08:29:56 crc kubenswrapper[4846]: I1005 08:29:56.400974 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Oct 05 08:29:56 crc kubenswrapper[4846]: I1005 08:29:56.744093 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/prometheus-metric-storage-0"] Oct 05 08:29:57 crc kubenswrapper[4846]: I1005 08:29:57.086303 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"e45f7b43-2271-4c62-a0b5-a66462317e84","Type":"ContainerStarted","Data":"48c0e6261e6775e82ed1eebce447b5186f9073502c3bbb523878c6cb123175ca"} Oct 05 08:29:57 crc kubenswrapper[4846]: I1005 08:29:57.086368 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"e45f7b43-2271-4c62-a0b5-a66462317e84","Type":"ContainerStarted","Data":"332c831896a8a58b34883d743738f06ba64f94e6254bed3b0e52e290e586afdb"} Oct 05 08:29:57 crc kubenswrapper[4846]: I1005 08:29:57.086606 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/prometheus-metric-storage-0" podUID="9eeca81b-6fda-4c22-a221-3adcb9900077" containerName="prometheus" containerID="cri-o://35860887d251721dfd419731dae2306db24d39a9d93d8f1543f1fe5315075bf9" gracePeriod=600 Oct 05 08:29:57 crc kubenswrapper[4846]: I1005 08:29:57.086670 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/prometheus-metric-storage-0" podUID="9eeca81b-6fda-4c22-a221-3adcb9900077" containerName="config-reloader" containerID="cri-o://5cb1120abfdcdf858d882f7c6f47445cb49f210041a4a9486cda3b94e9ad8438" gracePeriod=600 Oct 05 08:29:57 crc kubenswrapper[4846]: I1005 08:29:57.086681 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/prometheus-metric-storage-0" podUID="9eeca81b-6fda-4c22-a221-3adcb9900077" containerName="thanos-sidecar" containerID="cri-o://855c10a9b75c2c9926f5dd593debd83762c84203e420e5b91cfe42e7db8b6559" gracePeriod=600 Oct 05 08:29:58 crc kubenswrapper[4846]: I1005 08:29:58.100032 4846 generic.go:334] "Generic (PLEG): container finished" podID="6a40eb6f-d0a6-40c0-a48c-f461d052bb21" containerID="f5d45d11f84632d3e6cb961ad6e25fff9c7da4c7fbcfbf87f4f57154de66649e" exitCode=137 Oct 05 08:29:58 crc kubenswrapper[4846]: I1005 08:29:58.104474 4846 generic.go:334] "Generic (PLEG): container finished" podID="9eeca81b-6fda-4c22-a221-3adcb9900077" containerID="855c10a9b75c2c9926f5dd593debd83762c84203e420e5b91cfe42e7db8b6559" exitCode=0 Oct 05 08:29:58 crc kubenswrapper[4846]: I1005 08:29:58.104496 4846 generic.go:334] "Generic (PLEG): container finished" podID="9eeca81b-6fda-4c22-a221-3adcb9900077" containerID="5cb1120abfdcdf858d882f7c6f47445cb49f210041a4a9486cda3b94e9ad8438" exitCode=0 Oct 05 08:29:58 crc kubenswrapper[4846]: I1005 08:29:58.104503 4846 generic.go:334] "Generic (PLEG): container finished" podID="9eeca81b-6fda-4c22-a221-3adcb9900077" containerID="35860887d251721dfd419731dae2306db24d39a9d93d8f1543f1fe5315075bf9" exitCode=0 Oct 05 08:29:58 crc kubenswrapper[4846]: I1005 08:29:58.104687 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"9eeca81b-6fda-4c22-a221-3adcb9900077","Type":"ContainerDied","Data":"855c10a9b75c2c9926f5dd593debd83762c84203e420e5b91cfe42e7db8b6559"} Oct 05 08:29:58 crc kubenswrapper[4846]: I1005 08:29:58.104731 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"9eeca81b-6fda-4c22-a221-3adcb9900077","Type":"ContainerDied","Data":"5cb1120abfdcdf858d882f7c6f47445cb49f210041a4a9486cda3b94e9ad8438"} Oct 05 08:29:58 crc kubenswrapper[4846]: I1005 08:29:58.104743 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"9eeca81b-6fda-4c22-a221-3adcb9900077","Type":"ContainerDied","Data":"35860887d251721dfd419731dae2306db24d39a9d93d8f1543f1fe5315075bf9"} Oct 05 08:29:58 crc kubenswrapper[4846]: I1005 08:29:58.104753 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"9eeca81b-6fda-4c22-a221-3adcb9900077","Type":"ContainerDied","Data":"e8538609494690d52f4d579e3ae6523d65701c417f2551cb9f7b3cad911741a5"} Oct 05 08:29:58 crc kubenswrapper[4846]: I1005 08:29:58.104765 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e8538609494690d52f4d579e3ae6523d65701c417f2551cb9f7b3cad911741a5" Oct 05 08:29:58 crc kubenswrapper[4846]: I1005 08:29:58.106716 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Oct 05 08:29:58 crc kubenswrapper[4846]: I1005 08:29:58.133506 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=3.133486932 podStartE2EDuration="3.133486932s" podCreationTimestamp="2025-10-05 08:29:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 08:29:57.1061048 +0000 UTC m=+6119.346957585" watchObservedRunningTime="2025-10-05 08:29:58.133486932 +0000 UTC m=+6120.374339707" Oct 05 08:29:58 crc kubenswrapper[4846]: I1005 08:29:58.250030 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m8pt2\" (UniqueName: \"kubernetes.io/projected/9eeca81b-6fda-4c22-a221-3adcb9900077-kube-api-access-m8pt2\") pod \"9eeca81b-6fda-4c22-a221-3adcb9900077\" (UID: \"9eeca81b-6fda-4c22-a221-3adcb9900077\") " Oct 05 08:29:58 crc kubenswrapper[4846]: I1005 08:29:58.250326 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"prometheus-metric-storage-db\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-1cc7951e-b4e7-48bb-b9fd-de1baa2910d8\") pod \"9eeca81b-6fda-4c22-a221-3adcb9900077\" (UID: \"9eeca81b-6fda-4c22-a221-3adcb9900077\") " Oct 05 08:29:58 crc kubenswrapper[4846]: I1005 08:29:58.250405 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/9eeca81b-6fda-4c22-a221-3adcb9900077-config\") pod \"9eeca81b-6fda-4c22-a221-3adcb9900077\" (UID: \"9eeca81b-6fda-4c22-a221-3adcb9900077\") " Oct 05 08:29:58 crc kubenswrapper[4846]: I1005 08:29:58.250452 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/9eeca81b-6fda-4c22-a221-3adcb9900077-tls-assets\") pod \"9eeca81b-6fda-4c22-a221-3adcb9900077\" (UID: \"9eeca81b-6fda-4c22-a221-3adcb9900077\") " Oct 05 08:29:58 crc kubenswrapper[4846]: I1005 08:29:58.250508 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/9eeca81b-6fda-4c22-a221-3adcb9900077-thanos-prometheus-http-client-file\") pod \"9eeca81b-6fda-4c22-a221-3adcb9900077\" (UID: \"9eeca81b-6fda-4c22-a221-3adcb9900077\") " Oct 05 08:29:58 crc kubenswrapper[4846]: I1005 08:29:58.250555 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/9eeca81b-6fda-4c22-a221-3adcb9900077-web-config\") pod \"9eeca81b-6fda-4c22-a221-3adcb9900077\" (UID: \"9eeca81b-6fda-4c22-a221-3adcb9900077\") " Oct 05 08:29:58 crc kubenswrapper[4846]: I1005 08:29:58.250601 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/9eeca81b-6fda-4c22-a221-3adcb9900077-prometheus-metric-storage-rulefiles-0\") pod \"9eeca81b-6fda-4c22-a221-3adcb9900077\" (UID: \"9eeca81b-6fda-4c22-a221-3adcb9900077\") " Oct 05 08:29:58 crc kubenswrapper[4846]: I1005 08:29:58.250619 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/9eeca81b-6fda-4c22-a221-3adcb9900077-config-out\") pod \"9eeca81b-6fda-4c22-a221-3adcb9900077\" (UID: \"9eeca81b-6fda-4c22-a221-3adcb9900077\") " Oct 05 08:29:58 crc kubenswrapper[4846]: I1005 08:29:58.256148 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9eeca81b-6fda-4c22-a221-3adcb9900077-prometheus-metric-storage-rulefiles-0" (OuterVolumeSpecName: "prometheus-metric-storage-rulefiles-0") pod "9eeca81b-6fda-4c22-a221-3adcb9900077" (UID: "9eeca81b-6fda-4c22-a221-3adcb9900077"). InnerVolumeSpecName "prometheus-metric-storage-rulefiles-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 08:29:58 crc kubenswrapper[4846]: I1005 08:29:58.256434 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9eeca81b-6fda-4c22-a221-3adcb9900077-kube-api-access-m8pt2" (OuterVolumeSpecName: "kube-api-access-m8pt2") pod "9eeca81b-6fda-4c22-a221-3adcb9900077" (UID: "9eeca81b-6fda-4c22-a221-3adcb9900077"). InnerVolumeSpecName "kube-api-access-m8pt2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:29:58 crc kubenswrapper[4846]: I1005 08:29:58.256887 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9eeca81b-6fda-4c22-a221-3adcb9900077-tls-assets" (OuterVolumeSpecName: "tls-assets") pod "9eeca81b-6fda-4c22-a221-3adcb9900077" (UID: "9eeca81b-6fda-4c22-a221-3adcb9900077"). InnerVolumeSpecName "tls-assets". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:29:58 crc kubenswrapper[4846]: I1005 08:29:58.257685 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9eeca81b-6fda-4c22-a221-3adcb9900077-config-out" (OuterVolumeSpecName: "config-out") pod "9eeca81b-6fda-4c22-a221-3adcb9900077" (UID: "9eeca81b-6fda-4c22-a221-3adcb9900077"). InnerVolumeSpecName "config-out". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 08:29:58 crc kubenswrapper[4846]: I1005 08:29:58.258627 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9eeca81b-6fda-4c22-a221-3adcb9900077-thanos-prometheus-http-client-file" (OuterVolumeSpecName: "thanos-prometheus-http-client-file") pod "9eeca81b-6fda-4c22-a221-3adcb9900077" (UID: "9eeca81b-6fda-4c22-a221-3adcb9900077"). InnerVolumeSpecName "thanos-prometheus-http-client-file". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:29:58 crc kubenswrapper[4846]: I1005 08:29:58.259704 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9eeca81b-6fda-4c22-a221-3adcb9900077-config" (OuterVolumeSpecName: "config") pod "9eeca81b-6fda-4c22-a221-3adcb9900077" (UID: "9eeca81b-6fda-4c22-a221-3adcb9900077"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:29:58 crc kubenswrapper[4846]: I1005 08:29:58.276193 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-1cc7951e-b4e7-48bb-b9fd-de1baa2910d8" (OuterVolumeSpecName: "prometheus-metric-storage-db") pod "9eeca81b-6fda-4c22-a221-3adcb9900077" (UID: "9eeca81b-6fda-4c22-a221-3adcb9900077"). InnerVolumeSpecName "pvc-1cc7951e-b4e7-48bb-b9fd-de1baa2910d8". PluginName "kubernetes.io/csi", VolumeGidValue "" Oct 05 08:29:58 crc kubenswrapper[4846]: I1005 08:29:58.297438 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9eeca81b-6fda-4c22-a221-3adcb9900077-web-config" (OuterVolumeSpecName: "web-config") pod "9eeca81b-6fda-4c22-a221-3adcb9900077" (UID: "9eeca81b-6fda-4c22-a221-3adcb9900077"). InnerVolumeSpecName "web-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:29:58 crc kubenswrapper[4846]: I1005 08:29:58.352697 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m8pt2\" (UniqueName: \"kubernetes.io/projected/9eeca81b-6fda-4c22-a221-3adcb9900077-kube-api-access-m8pt2\") on node \"crc\" DevicePath \"\"" Oct 05 08:29:58 crc kubenswrapper[4846]: I1005 08:29:58.352764 4846 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-1cc7951e-b4e7-48bb-b9fd-de1baa2910d8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-1cc7951e-b4e7-48bb-b9fd-de1baa2910d8\") on node \"crc\" " Oct 05 08:29:58 crc kubenswrapper[4846]: I1005 08:29:58.352777 4846 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/9eeca81b-6fda-4c22-a221-3adcb9900077-config\") on node \"crc\" DevicePath \"\"" Oct 05 08:29:58 crc kubenswrapper[4846]: I1005 08:29:58.352786 4846 reconciler_common.go:293] "Volume detached for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/9eeca81b-6fda-4c22-a221-3adcb9900077-tls-assets\") on node \"crc\" DevicePath \"\"" Oct 05 08:29:58 crc kubenswrapper[4846]: I1005 08:29:58.352799 4846 reconciler_common.go:293] "Volume detached for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/9eeca81b-6fda-4c22-a221-3adcb9900077-thanos-prometheus-http-client-file\") on node \"crc\" DevicePath \"\"" Oct 05 08:29:58 crc kubenswrapper[4846]: I1005 08:29:58.377990 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 05 08:29:58 crc kubenswrapper[4846]: I1005 08:29:58.378355 4846 reconciler_common.go:293] "Volume detached for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/9eeca81b-6fda-4c22-a221-3adcb9900077-web-config\") on node \"crc\" DevicePath \"\"" Oct 05 08:29:58 crc kubenswrapper[4846]: I1005 08:29:58.378401 4846 reconciler_common.go:293] "Volume detached for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/9eeca81b-6fda-4c22-a221-3adcb9900077-prometheus-metric-storage-rulefiles-0\") on node \"crc\" DevicePath \"\"" Oct 05 08:29:58 crc kubenswrapper[4846]: I1005 08:29:58.378413 4846 reconciler_common.go:293] "Volume detached for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/9eeca81b-6fda-4c22-a221-3adcb9900077-config-out\") on node \"crc\" DevicePath \"\"" Oct 05 08:29:58 crc kubenswrapper[4846]: I1005 08:29:58.391033 4846 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Oct 05 08:29:58 crc kubenswrapper[4846]: I1005 08:29:58.391268 4846 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-1cc7951e-b4e7-48bb-b9fd-de1baa2910d8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-1cc7951e-b4e7-48bb-b9fd-de1baa2910d8") on node "crc" Oct 05 08:29:58 crc kubenswrapper[4846]: I1005 08:29:58.481668 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5ttpz\" (UniqueName: \"kubernetes.io/projected/6a40eb6f-d0a6-40c0-a48c-f461d052bb21-kube-api-access-5ttpz\") pod \"6a40eb6f-d0a6-40c0-a48c-f461d052bb21\" (UID: \"6a40eb6f-d0a6-40c0-a48c-f461d052bb21\") " Oct 05 08:29:58 crc kubenswrapper[4846]: I1005 08:29:58.481742 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/6a40eb6f-d0a6-40c0-a48c-f461d052bb21-openstack-config\") pod \"6a40eb6f-d0a6-40c0-a48c-f461d052bb21\" (UID: \"6a40eb6f-d0a6-40c0-a48c-f461d052bb21\") " Oct 05 08:29:58 crc kubenswrapper[4846]: I1005 08:29:58.481819 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6a40eb6f-d0a6-40c0-a48c-f461d052bb21-combined-ca-bundle\") pod \"6a40eb6f-d0a6-40c0-a48c-f461d052bb21\" (UID: \"6a40eb6f-d0a6-40c0-a48c-f461d052bb21\") " Oct 05 08:29:58 crc kubenswrapper[4846]: I1005 08:29:58.481920 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/6a40eb6f-d0a6-40c0-a48c-f461d052bb21-openstack-config-secret\") pod \"6a40eb6f-d0a6-40c0-a48c-f461d052bb21\" (UID: \"6a40eb6f-d0a6-40c0-a48c-f461d052bb21\") " Oct 05 08:29:58 crc kubenswrapper[4846]: I1005 08:29:58.483562 4846 reconciler_common.go:293] "Volume detached for volume \"pvc-1cc7951e-b4e7-48bb-b9fd-de1baa2910d8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-1cc7951e-b4e7-48bb-b9fd-de1baa2910d8\") on node \"crc\" DevicePath \"\"" Oct 05 08:29:58 crc kubenswrapper[4846]: I1005 08:29:58.526649 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6a40eb6f-d0a6-40c0-a48c-f461d052bb21-kube-api-access-5ttpz" (OuterVolumeSpecName: "kube-api-access-5ttpz") pod "6a40eb6f-d0a6-40c0-a48c-f461d052bb21" (UID: "6a40eb6f-d0a6-40c0-a48c-f461d052bb21"). InnerVolumeSpecName "kube-api-access-5ttpz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:29:58 crc kubenswrapper[4846]: I1005 08:29:58.530820 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6a40eb6f-d0a6-40c0-a48c-f461d052bb21-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6a40eb6f-d0a6-40c0-a48c-f461d052bb21" (UID: "6a40eb6f-d0a6-40c0-a48c-f461d052bb21"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:29:58 crc kubenswrapper[4846]: I1005 08:29:58.556007 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6a40eb6f-d0a6-40c0-a48c-f461d052bb21-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "6a40eb6f-d0a6-40c0-a48c-f461d052bb21" (UID: "6a40eb6f-d0a6-40c0-a48c-f461d052bb21"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 08:29:58 crc kubenswrapper[4846]: I1005 08:29:58.586091 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5ttpz\" (UniqueName: \"kubernetes.io/projected/6a40eb6f-d0a6-40c0-a48c-f461d052bb21-kube-api-access-5ttpz\") on node \"crc\" DevicePath \"\"" Oct 05 08:29:58 crc kubenswrapper[4846]: I1005 08:29:58.586117 4846 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/6a40eb6f-d0a6-40c0-a48c-f461d052bb21-openstack-config\") on node \"crc\" DevicePath \"\"" Oct 05 08:29:58 crc kubenswrapper[4846]: I1005 08:29:58.586126 4846 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6a40eb6f-d0a6-40c0-a48c-f461d052bb21-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 08:29:58 crc kubenswrapper[4846]: I1005 08:29:58.587976 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6a40eb6f-d0a6-40c0-a48c-f461d052bb21-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "6a40eb6f-d0a6-40c0-a48c-f461d052bb21" (UID: "6a40eb6f-d0a6-40c0-a48c-f461d052bb21"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:29:58 crc kubenswrapper[4846]: I1005 08:29:58.687934 4846 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/6a40eb6f-d0a6-40c0-a48c-f461d052bb21-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Oct 05 08:29:59 crc kubenswrapper[4846]: I1005 08:29:59.048942 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 05 08:29:59 crc kubenswrapper[4846]: E1005 08:29:59.049701 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9eeca81b-6fda-4c22-a221-3adcb9900077" containerName="prometheus" Oct 05 08:29:59 crc kubenswrapper[4846]: I1005 08:29:59.049724 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="9eeca81b-6fda-4c22-a221-3adcb9900077" containerName="prometheus" Oct 05 08:29:59 crc kubenswrapper[4846]: E1005 08:29:59.049748 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9eeca81b-6fda-4c22-a221-3adcb9900077" containerName="thanos-sidecar" Oct 05 08:29:59 crc kubenswrapper[4846]: I1005 08:29:59.049756 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="9eeca81b-6fda-4c22-a221-3adcb9900077" containerName="thanos-sidecar" Oct 05 08:29:59 crc kubenswrapper[4846]: E1005 08:29:59.049784 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9eeca81b-6fda-4c22-a221-3adcb9900077" containerName="config-reloader" Oct 05 08:29:59 crc kubenswrapper[4846]: I1005 08:29:59.049792 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="9eeca81b-6fda-4c22-a221-3adcb9900077" containerName="config-reloader" Oct 05 08:29:59 crc kubenswrapper[4846]: E1005 08:29:59.049803 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9eeca81b-6fda-4c22-a221-3adcb9900077" containerName="init-config-reloader" Oct 05 08:29:59 crc kubenswrapper[4846]: I1005 08:29:59.049811 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="9eeca81b-6fda-4c22-a221-3adcb9900077" containerName="init-config-reloader" Oct 05 08:29:59 crc kubenswrapper[4846]: I1005 08:29:59.050058 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="9eeca81b-6fda-4c22-a221-3adcb9900077" containerName="thanos-sidecar" Oct 05 08:29:59 crc kubenswrapper[4846]: I1005 08:29:59.050085 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="9eeca81b-6fda-4c22-a221-3adcb9900077" containerName="prometheus" Oct 05 08:29:59 crc kubenswrapper[4846]: I1005 08:29:59.050107 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="9eeca81b-6fda-4c22-a221-3adcb9900077" containerName="config-reloader" Oct 05 08:29:59 crc kubenswrapper[4846]: I1005 08:29:59.052294 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 05 08:29:59 crc kubenswrapper[4846]: I1005 08:29:59.055649 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 05 08:29:59 crc kubenswrapper[4846]: I1005 08:29:59.055805 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 05 08:29:59 crc kubenswrapper[4846]: I1005 08:29:59.060874 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 05 08:29:59 crc kubenswrapper[4846]: I1005 08:29:59.097395 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/317b95c9-9b6d-4ba8-a96c-f69bcc8bb888-log-httpd\") pod \"ceilometer-0\" (UID: \"317b95c9-9b6d-4ba8-a96c-f69bcc8bb888\") " pod="openstack/ceilometer-0" Oct 05 08:29:59 crc kubenswrapper[4846]: I1005 08:29:59.097465 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pfttb\" (UniqueName: \"kubernetes.io/projected/317b95c9-9b6d-4ba8-a96c-f69bcc8bb888-kube-api-access-pfttb\") pod \"ceilometer-0\" (UID: \"317b95c9-9b6d-4ba8-a96c-f69bcc8bb888\") " pod="openstack/ceilometer-0" Oct 05 08:29:59 crc kubenswrapper[4846]: I1005 08:29:59.097531 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/317b95c9-9b6d-4ba8-a96c-f69bcc8bb888-run-httpd\") pod \"ceilometer-0\" (UID: \"317b95c9-9b6d-4ba8-a96c-f69bcc8bb888\") " pod="openstack/ceilometer-0" Oct 05 08:29:59 crc kubenswrapper[4846]: I1005 08:29:59.097626 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/317b95c9-9b6d-4ba8-a96c-f69bcc8bb888-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"317b95c9-9b6d-4ba8-a96c-f69bcc8bb888\") " pod="openstack/ceilometer-0" Oct 05 08:29:59 crc kubenswrapper[4846]: I1005 08:29:59.097659 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/317b95c9-9b6d-4ba8-a96c-f69bcc8bb888-scripts\") pod \"ceilometer-0\" (UID: \"317b95c9-9b6d-4ba8-a96c-f69bcc8bb888\") " pod="openstack/ceilometer-0" Oct 05 08:29:59 crc kubenswrapper[4846]: I1005 08:29:59.097731 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/317b95c9-9b6d-4ba8-a96c-f69bcc8bb888-config-data\") pod \"ceilometer-0\" (UID: \"317b95c9-9b6d-4ba8-a96c-f69bcc8bb888\") " pod="openstack/ceilometer-0" Oct 05 08:29:59 crc kubenswrapper[4846]: I1005 08:29:59.097757 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/317b95c9-9b6d-4ba8-a96c-f69bcc8bb888-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"317b95c9-9b6d-4ba8-a96c-f69bcc8bb888\") " pod="openstack/ceilometer-0" Oct 05 08:29:59 crc kubenswrapper[4846]: I1005 08:29:59.128501 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Oct 05 08:29:59 crc kubenswrapper[4846]: I1005 08:29:59.129465 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 05 08:29:59 crc kubenswrapper[4846]: I1005 08:29:59.129497 4846 scope.go:117] "RemoveContainer" containerID="f5d45d11f84632d3e6cb961ad6e25fff9c7da4c7fbcfbf87f4f57154de66649e" Oct 05 08:29:59 crc kubenswrapper[4846]: I1005 08:29:59.153644 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/prometheus-metric-storage-0"] Oct 05 08:29:59 crc kubenswrapper[4846]: I1005 08:29:59.169746 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/prometheus-metric-storage-0"] Oct 05 08:29:59 crc kubenswrapper[4846]: I1005 08:29:59.177978 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/prometheus-metric-storage-0"] Oct 05 08:29:59 crc kubenswrapper[4846]: I1005 08:29:59.181221 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Oct 05 08:29:59 crc kubenswrapper[4846]: I1005 08:29:59.185341 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-web-config" Oct 05 08:29:59 crc kubenswrapper[4846]: I1005 08:29:59.186486 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage" Oct 05 08:29:59 crc kubenswrapper[4846]: I1005 08:29:59.187852 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-thanos-prometheus-http-client-file" Oct 05 08:29:59 crc kubenswrapper[4846]: I1005 08:29:59.188783 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"prometheus-metric-storage-rulefiles-0" Oct 05 08:29:59 crc kubenswrapper[4846]: I1005 08:29:59.188903 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-metric-storage-prometheus-svc" Oct 05 08:29:59 crc kubenswrapper[4846]: I1005 08:29:59.189325 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"metric-storage-prometheus-dockercfg-mkgcg" Oct 05 08:29:59 crc kubenswrapper[4846]: I1005 08:29:59.196201 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-tls-assets-0" Oct 05 08:29:59 crc kubenswrapper[4846]: I1005 08:29:59.199060 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/317b95c9-9b6d-4ba8-a96c-f69bcc8bb888-log-httpd\") pod \"ceilometer-0\" (UID: \"317b95c9-9b6d-4ba8-a96c-f69bcc8bb888\") " pod="openstack/ceilometer-0" Oct 05 08:29:59 crc kubenswrapper[4846]: I1005 08:29:59.199112 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pfttb\" (UniqueName: \"kubernetes.io/projected/317b95c9-9b6d-4ba8-a96c-f69bcc8bb888-kube-api-access-pfttb\") pod \"ceilometer-0\" (UID: \"317b95c9-9b6d-4ba8-a96c-f69bcc8bb888\") " pod="openstack/ceilometer-0" Oct 05 08:29:59 crc kubenswrapper[4846]: I1005 08:29:59.199164 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/317b95c9-9b6d-4ba8-a96c-f69bcc8bb888-run-httpd\") pod \"ceilometer-0\" (UID: \"317b95c9-9b6d-4ba8-a96c-f69bcc8bb888\") " pod="openstack/ceilometer-0" Oct 05 08:29:59 crc kubenswrapper[4846]: I1005 08:29:59.199246 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/317b95c9-9b6d-4ba8-a96c-f69bcc8bb888-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"317b95c9-9b6d-4ba8-a96c-f69bcc8bb888\") " pod="openstack/ceilometer-0" Oct 05 08:29:59 crc kubenswrapper[4846]: I1005 08:29:59.199272 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/317b95c9-9b6d-4ba8-a96c-f69bcc8bb888-scripts\") pod \"ceilometer-0\" (UID: \"317b95c9-9b6d-4ba8-a96c-f69bcc8bb888\") " pod="openstack/ceilometer-0" Oct 05 08:29:59 crc kubenswrapper[4846]: I1005 08:29:59.199309 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/317b95c9-9b6d-4ba8-a96c-f69bcc8bb888-config-data\") pod \"ceilometer-0\" (UID: \"317b95c9-9b6d-4ba8-a96c-f69bcc8bb888\") " pod="openstack/ceilometer-0" Oct 05 08:29:59 crc kubenswrapper[4846]: I1005 08:29:59.199329 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/317b95c9-9b6d-4ba8-a96c-f69bcc8bb888-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"317b95c9-9b6d-4ba8-a96c-f69bcc8bb888\") " pod="openstack/ceilometer-0" Oct 05 08:29:59 crc kubenswrapper[4846]: I1005 08:29:59.215206 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/317b95c9-9b6d-4ba8-a96c-f69bcc8bb888-run-httpd\") pod \"ceilometer-0\" (UID: \"317b95c9-9b6d-4ba8-a96c-f69bcc8bb888\") " pod="openstack/ceilometer-0" Oct 05 08:29:59 crc kubenswrapper[4846]: I1005 08:29:59.215422 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/317b95c9-9b6d-4ba8-a96c-f69bcc8bb888-log-httpd\") pod \"ceilometer-0\" (UID: \"317b95c9-9b6d-4ba8-a96c-f69bcc8bb888\") " pod="openstack/ceilometer-0" Oct 05 08:29:59 crc kubenswrapper[4846]: I1005 08:29:59.219429 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/317b95c9-9b6d-4ba8-a96c-f69bcc8bb888-scripts\") pod \"ceilometer-0\" (UID: \"317b95c9-9b6d-4ba8-a96c-f69bcc8bb888\") " pod="openstack/ceilometer-0" Oct 05 08:29:59 crc kubenswrapper[4846]: I1005 08:29:59.221013 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/317b95c9-9b6d-4ba8-a96c-f69bcc8bb888-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"317b95c9-9b6d-4ba8-a96c-f69bcc8bb888\") " pod="openstack/ceilometer-0" Oct 05 08:29:59 crc kubenswrapper[4846]: I1005 08:29:59.222807 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/317b95c9-9b6d-4ba8-a96c-f69bcc8bb888-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"317b95c9-9b6d-4ba8-a96c-f69bcc8bb888\") " pod="openstack/ceilometer-0" Oct 05 08:29:59 crc kubenswrapper[4846]: I1005 08:29:59.226742 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/317b95c9-9b6d-4ba8-a96c-f69bcc8bb888-config-data\") pod \"ceilometer-0\" (UID: \"317b95c9-9b6d-4ba8-a96c-f69bcc8bb888\") " pod="openstack/ceilometer-0" Oct 05 08:29:59 crc kubenswrapper[4846]: I1005 08:29:59.240283 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Oct 05 08:29:59 crc kubenswrapper[4846]: I1005 08:29:59.258836 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pfttb\" (UniqueName: \"kubernetes.io/projected/317b95c9-9b6d-4ba8-a96c-f69bcc8bb888-kube-api-access-pfttb\") pod \"ceilometer-0\" (UID: \"317b95c9-9b6d-4ba8-a96c-f69bcc8bb888\") " pod="openstack/ceilometer-0" Oct 05 08:29:59 crc kubenswrapper[4846]: I1005 08:29:59.302226 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/8e051aea-4a36-4dd1-9d60-60842891f011-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"8e051aea-4a36-4dd1-9d60-60842891f011\") " pod="openstack/prometheus-metric-storage-0" Oct 05 08:29:59 crc kubenswrapper[4846]: I1005 08:29:59.302274 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/8e051aea-4a36-4dd1-9d60-60842891f011-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"8e051aea-4a36-4dd1-9d60-60842891f011\") " pod="openstack/prometheus-metric-storage-0" Oct 05 08:29:59 crc kubenswrapper[4846]: I1005 08:29:59.302327 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-95kpc\" (UniqueName: \"kubernetes.io/projected/8e051aea-4a36-4dd1-9d60-60842891f011-kube-api-access-95kpc\") pod \"prometheus-metric-storage-0\" (UID: \"8e051aea-4a36-4dd1-9d60-60842891f011\") " pod="openstack/prometheus-metric-storage-0" Oct 05 08:29:59 crc kubenswrapper[4846]: I1005 08:29:59.302606 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/8e051aea-4a36-4dd1-9d60-60842891f011-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"8e051aea-4a36-4dd1-9d60-60842891f011\") " pod="openstack/prometheus-metric-storage-0" Oct 05 08:29:59 crc kubenswrapper[4846]: I1005 08:29:59.302669 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/8e051aea-4a36-4dd1-9d60-60842891f011-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"8e051aea-4a36-4dd1-9d60-60842891f011\") " pod="openstack/prometheus-metric-storage-0" Oct 05 08:29:59 crc kubenswrapper[4846]: I1005 08:29:59.302773 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/8e051aea-4a36-4dd1-9d60-60842891f011-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"8e051aea-4a36-4dd1-9d60-60842891f011\") " pod="openstack/prometheus-metric-storage-0" Oct 05 08:29:59 crc kubenswrapper[4846]: I1005 08:29:59.302877 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-1cc7951e-b4e7-48bb-b9fd-de1baa2910d8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-1cc7951e-b4e7-48bb-b9fd-de1baa2910d8\") pod \"prometheus-metric-storage-0\" (UID: \"8e051aea-4a36-4dd1-9d60-60842891f011\") " pod="openstack/prometheus-metric-storage-0" Oct 05 08:29:59 crc kubenswrapper[4846]: I1005 08:29:59.302987 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/8e051aea-4a36-4dd1-9d60-60842891f011-config\") pod \"prometheus-metric-storage-0\" (UID: \"8e051aea-4a36-4dd1-9d60-60842891f011\") " pod="openstack/prometheus-metric-storage-0" Oct 05 08:29:59 crc kubenswrapper[4846]: I1005 08:29:59.303011 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/8e051aea-4a36-4dd1-9d60-60842891f011-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"8e051aea-4a36-4dd1-9d60-60842891f011\") " pod="openstack/prometheus-metric-storage-0" Oct 05 08:29:59 crc kubenswrapper[4846]: I1005 08:29:59.303078 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8e051aea-4a36-4dd1-9d60-60842891f011-secret-combined-ca-bundle\") pod \"prometheus-metric-storage-0\" (UID: \"8e051aea-4a36-4dd1-9d60-60842891f011\") " pod="openstack/prometheus-metric-storage-0" Oct 05 08:29:59 crc kubenswrapper[4846]: I1005 08:29:59.303173 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/8e051aea-4a36-4dd1-9d60-60842891f011-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"8e051aea-4a36-4dd1-9d60-60842891f011\") " pod="openstack/prometheus-metric-storage-0" Oct 05 08:29:59 crc kubenswrapper[4846]: I1005 08:29:59.405313 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/8e051aea-4a36-4dd1-9d60-60842891f011-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"8e051aea-4a36-4dd1-9d60-60842891f011\") " pod="openstack/prometheus-metric-storage-0" Oct 05 08:29:59 crc kubenswrapper[4846]: I1005 08:29:59.405576 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-1cc7951e-b4e7-48bb-b9fd-de1baa2910d8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-1cc7951e-b4e7-48bb-b9fd-de1baa2910d8\") pod \"prometheus-metric-storage-0\" (UID: \"8e051aea-4a36-4dd1-9d60-60842891f011\") " pod="openstack/prometheus-metric-storage-0" Oct 05 08:29:59 crc kubenswrapper[4846]: I1005 08:29:59.405622 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/8e051aea-4a36-4dd1-9d60-60842891f011-config\") pod \"prometheus-metric-storage-0\" (UID: \"8e051aea-4a36-4dd1-9d60-60842891f011\") " pod="openstack/prometheus-metric-storage-0" Oct 05 08:29:59 crc kubenswrapper[4846]: I1005 08:29:59.405641 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/8e051aea-4a36-4dd1-9d60-60842891f011-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"8e051aea-4a36-4dd1-9d60-60842891f011\") " pod="openstack/prometheus-metric-storage-0" Oct 05 08:29:59 crc kubenswrapper[4846]: I1005 08:29:59.405657 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8e051aea-4a36-4dd1-9d60-60842891f011-secret-combined-ca-bundle\") pod \"prometheus-metric-storage-0\" (UID: \"8e051aea-4a36-4dd1-9d60-60842891f011\") " pod="openstack/prometheus-metric-storage-0" Oct 05 08:29:59 crc kubenswrapper[4846]: I1005 08:29:59.405688 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/8e051aea-4a36-4dd1-9d60-60842891f011-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"8e051aea-4a36-4dd1-9d60-60842891f011\") " pod="openstack/prometheus-metric-storage-0" Oct 05 08:29:59 crc kubenswrapper[4846]: I1005 08:29:59.405723 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/8e051aea-4a36-4dd1-9d60-60842891f011-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"8e051aea-4a36-4dd1-9d60-60842891f011\") " pod="openstack/prometheus-metric-storage-0" Oct 05 08:29:59 crc kubenswrapper[4846]: I1005 08:29:59.405747 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/8e051aea-4a36-4dd1-9d60-60842891f011-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"8e051aea-4a36-4dd1-9d60-60842891f011\") " pod="openstack/prometheus-metric-storage-0" Oct 05 08:29:59 crc kubenswrapper[4846]: I1005 08:29:59.405780 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-95kpc\" (UniqueName: \"kubernetes.io/projected/8e051aea-4a36-4dd1-9d60-60842891f011-kube-api-access-95kpc\") pod \"prometheus-metric-storage-0\" (UID: \"8e051aea-4a36-4dd1-9d60-60842891f011\") " pod="openstack/prometheus-metric-storage-0" Oct 05 08:29:59 crc kubenswrapper[4846]: I1005 08:29:59.405847 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/8e051aea-4a36-4dd1-9d60-60842891f011-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"8e051aea-4a36-4dd1-9d60-60842891f011\") " pod="openstack/prometheus-metric-storage-0" Oct 05 08:29:59 crc kubenswrapper[4846]: I1005 08:29:59.405870 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/8e051aea-4a36-4dd1-9d60-60842891f011-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"8e051aea-4a36-4dd1-9d60-60842891f011\") " pod="openstack/prometheus-metric-storage-0" Oct 05 08:29:59 crc kubenswrapper[4846]: I1005 08:29:59.407906 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/8e051aea-4a36-4dd1-9d60-60842891f011-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"8e051aea-4a36-4dd1-9d60-60842891f011\") " pod="openstack/prometheus-metric-storage-0" Oct 05 08:29:59 crc kubenswrapper[4846]: I1005 08:29:59.409446 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 05 08:29:59 crc kubenswrapper[4846]: I1005 08:29:59.413958 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/8e051aea-4a36-4dd1-9d60-60842891f011-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"8e051aea-4a36-4dd1-9d60-60842891f011\") " pod="openstack/prometheus-metric-storage-0" Oct 05 08:29:59 crc kubenswrapper[4846]: I1005 08:29:59.419897 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8e051aea-4a36-4dd1-9d60-60842891f011-secret-combined-ca-bundle\") pod \"prometheus-metric-storage-0\" (UID: \"8e051aea-4a36-4dd1-9d60-60842891f011\") " pod="openstack/prometheus-metric-storage-0" Oct 05 08:29:59 crc kubenswrapper[4846]: I1005 08:29:59.419915 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/8e051aea-4a36-4dd1-9d60-60842891f011-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"8e051aea-4a36-4dd1-9d60-60842891f011\") " pod="openstack/prometheus-metric-storage-0" Oct 05 08:29:59 crc kubenswrapper[4846]: I1005 08:29:59.430912 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/8e051aea-4a36-4dd1-9d60-60842891f011-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"8e051aea-4a36-4dd1-9d60-60842891f011\") " pod="openstack/prometheus-metric-storage-0" Oct 05 08:29:59 crc kubenswrapper[4846]: I1005 08:29:59.431416 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/8e051aea-4a36-4dd1-9d60-60842891f011-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"8e051aea-4a36-4dd1-9d60-60842891f011\") " pod="openstack/prometheus-metric-storage-0" Oct 05 08:29:59 crc kubenswrapper[4846]: I1005 08:29:59.431424 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/8e051aea-4a36-4dd1-9d60-60842891f011-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"8e051aea-4a36-4dd1-9d60-60842891f011\") " pod="openstack/prometheus-metric-storage-0" Oct 05 08:29:59 crc kubenswrapper[4846]: I1005 08:29:59.431984 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/8e051aea-4a36-4dd1-9d60-60842891f011-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"8e051aea-4a36-4dd1-9d60-60842891f011\") " pod="openstack/prometheus-metric-storage-0" Oct 05 08:29:59 crc kubenswrapper[4846]: I1005 08:29:59.433286 4846 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 05 08:29:59 crc kubenswrapper[4846]: I1005 08:29:59.433339 4846 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-1cc7951e-b4e7-48bb-b9fd-de1baa2910d8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-1cc7951e-b4e7-48bb-b9fd-de1baa2910d8\") pod \"prometheus-metric-storage-0\" (UID: \"8e051aea-4a36-4dd1-9d60-60842891f011\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/0ec9240d9ee399c4d378df6ba11f50328a91e5e421b3e1fde811ca7b96ace4eb/globalmount\"" pod="openstack/prometheus-metric-storage-0" Oct 05 08:29:59 crc kubenswrapper[4846]: I1005 08:29:59.434808 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-95kpc\" (UniqueName: \"kubernetes.io/projected/8e051aea-4a36-4dd1-9d60-60842891f011-kube-api-access-95kpc\") pod \"prometheus-metric-storage-0\" (UID: \"8e051aea-4a36-4dd1-9d60-60842891f011\") " pod="openstack/prometheus-metric-storage-0" Oct 05 08:29:59 crc kubenswrapper[4846]: I1005 08:29:59.446328 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/8e051aea-4a36-4dd1-9d60-60842891f011-config\") pod \"prometheus-metric-storage-0\" (UID: \"8e051aea-4a36-4dd1-9d60-60842891f011\") " pod="openstack/prometheus-metric-storage-0" Oct 05 08:29:59 crc kubenswrapper[4846]: I1005 08:29:59.531453 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-1cc7951e-b4e7-48bb-b9fd-de1baa2910d8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-1cc7951e-b4e7-48bb-b9fd-de1baa2910d8\") pod \"prometheus-metric-storage-0\" (UID: \"8e051aea-4a36-4dd1-9d60-60842891f011\") " pod="openstack/prometheus-metric-storage-0" Oct 05 08:29:59 crc kubenswrapper[4846]: I1005 08:29:59.796628 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Oct 05 08:29:59 crc kubenswrapper[4846]: I1005 08:29:59.931912 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 05 08:30:00 crc kubenswrapper[4846]: I1005 08:30:00.136992 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29327550-2445r"] Oct 05 08:30:00 crc kubenswrapper[4846]: I1005 08:30:00.138448 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29327550-2445r" Oct 05 08:30:00 crc kubenswrapper[4846]: I1005 08:30:00.139476 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"317b95c9-9b6d-4ba8-a96c-f69bcc8bb888","Type":"ContainerStarted","Data":"48c49b6801bb84bfafc7f2724a02df14c68cef9977d32f2c97732cc1a07505ce"} Oct 05 08:30:00 crc kubenswrapper[4846]: I1005 08:30:00.141437 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 05 08:30:00 crc kubenswrapper[4846]: I1005 08:30:00.141823 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 05 08:30:00 crc kubenswrapper[4846]: I1005 08:30:00.146544 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29327550-2445r"] Oct 05 08:30:00 crc kubenswrapper[4846]: I1005 08:30:00.263356 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Oct 05 08:30:00 crc kubenswrapper[4846]: I1005 08:30:00.322358 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c4867d81-361b-4e63-9860-2099edac282c-config-volume\") pod \"collect-profiles-29327550-2445r\" (UID: \"c4867d81-361b-4e63-9860-2099edac282c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327550-2445r" Oct 05 08:30:00 crc kubenswrapper[4846]: I1005 08:30:00.322443 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j4mqj\" (UniqueName: \"kubernetes.io/projected/c4867d81-361b-4e63-9860-2099edac282c-kube-api-access-j4mqj\") pod \"collect-profiles-29327550-2445r\" (UID: \"c4867d81-361b-4e63-9860-2099edac282c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327550-2445r" Oct 05 08:30:00 crc kubenswrapper[4846]: I1005 08:30:00.322559 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c4867d81-361b-4e63-9860-2099edac282c-secret-volume\") pod \"collect-profiles-29327550-2445r\" (UID: \"c4867d81-361b-4e63-9860-2099edac282c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327550-2445r" Oct 05 08:30:00 crc kubenswrapper[4846]: I1005 08:30:00.424265 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c4867d81-361b-4e63-9860-2099edac282c-secret-volume\") pod \"collect-profiles-29327550-2445r\" (UID: \"c4867d81-361b-4e63-9860-2099edac282c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327550-2445r" Oct 05 08:30:00 crc kubenswrapper[4846]: I1005 08:30:00.424386 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c4867d81-361b-4e63-9860-2099edac282c-config-volume\") pod \"collect-profiles-29327550-2445r\" (UID: \"c4867d81-361b-4e63-9860-2099edac282c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327550-2445r" Oct 05 08:30:00 crc kubenswrapper[4846]: I1005 08:30:00.424423 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j4mqj\" (UniqueName: \"kubernetes.io/projected/c4867d81-361b-4e63-9860-2099edac282c-kube-api-access-j4mqj\") pod \"collect-profiles-29327550-2445r\" (UID: \"c4867d81-361b-4e63-9860-2099edac282c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327550-2445r" Oct 05 08:30:00 crc kubenswrapper[4846]: I1005 08:30:00.428409 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c4867d81-361b-4e63-9860-2099edac282c-config-volume\") pod \"collect-profiles-29327550-2445r\" (UID: \"c4867d81-361b-4e63-9860-2099edac282c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327550-2445r" Oct 05 08:30:00 crc kubenswrapper[4846]: I1005 08:30:00.430735 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c4867d81-361b-4e63-9860-2099edac282c-secret-volume\") pod \"collect-profiles-29327550-2445r\" (UID: \"c4867d81-361b-4e63-9860-2099edac282c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327550-2445r" Oct 05 08:30:00 crc kubenswrapper[4846]: I1005 08:30:00.455149 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j4mqj\" (UniqueName: \"kubernetes.io/projected/c4867d81-361b-4e63-9860-2099edac282c-kube-api-access-j4mqj\") pod \"collect-profiles-29327550-2445r\" (UID: \"c4867d81-361b-4e63-9860-2099edac282c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327550-2445r" Oct 05 08:30:00 crc kubenswrapper[4846]: I1005 08:30:00.463329 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29327550-2445r" Oct 05 08:30:00 crc kubenswrapper[4846]: I1005 08:30:00.529922 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6a40eb6f-d0a6-40c0-a48c-f461d052bb21" path="/var/lib/kubelet/pods/6a40eb6f-d0a6-40c0-a48c-f461d052bb21/volumes" Oct 05 08:30:00 crc kubenswrapper[4846]: I1005 08:30:00.530986 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9eeca81b-6fda-4c22-a221-3adcb9900077" path="/var/lib/kubelet/pods/9eeca81b-6fda-4c22-a221-3adcb9900077/volumes" Oct 05 08:30:00 crc kubenswrapper[4846]: I1005 08:30:00.992992 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29327550-2445r"] Oct 05 08:30:01 crc kubenswrapper[4846]: I1005 08:30:01.160470 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29327550-2445r" event={"ID":"c4867d81-361b-4e63-9860-2099edac282c","Type":"ContainerStarted","Data":"9139e002c3f22a38d61d22a4ab0bbad57cdb0dff772497fba018bbdf0a30af58"} Oct 05 08:30:01 crc kubenswrapper[4846]: I1005 08:30:01.165054 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"8e051aea-4a36-4dd1-9d60-60842891f011","Type":"ContainerStarted","Data":"b3908739e128c2d069b70620ac4f59713f05221f5b5a94e113fb30370da2accf"} Oct 05 08:30:02 crc kubenswrapper[4846]: I1005 08:30:02.181607 4846 generic.go:334] "Generic (PLEG): container finished" podID="c4867d81-361b-4e63-9860-2099edac282c" containerID="c4ac4f425d1a512844e1cd0c566e532368ed403a9743d1c0c2073c2db29e0862" exitCode=0 Oct 05 08:30:02 crc kubenswrapper[4846]: I1005 08:30:02.181858 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29327550-2445r" event={"ID":"c4867d81-361b-4e63-9860-2099edac282c","Type":"ContainerDied","Data":"c4ac4f425d1a512844e1cd0c566e532368ed403a9743d1c0c2073c2db29e0862"} Oct 05 08:30:03 crc kubenswrapper[4846]: I1005 08:30:03.939653 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-9lkxx" Oct 05 08:30:03 crc kubenswrapper[4846]: I1005 08:30:03.996537 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-9lkxx"] Oct 05 08:30:04 crc kubenswrapper[4846]: I1005 08:30:04.216045 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-9lkxx" podUID="7bc1dff8-1996-4278-89f2-1004c67562c3" containerName="registry-server" containerID="cri-o://4fa9df62f6f1fdf926b083d1d0191ae3233a17658d3b18a6621d988223c7343b" gracePeriod=2 Oct 05 08:30:04 crc kubenswrapper[4846]: I1005 08:30:04.216810 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"8e051aea-4a36-4dd1-9d60-60842891f011","Type":"ContainerStarted","Data":"aafb87ef488436a2880ab6c338f076cbca094992c2e73780b2ea225e5301306c"} Oct 05 08:30:04 crc kubenswrapper[4846]: I1005 08:30:04.463242 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29327550-2445r" Oct 05 08:30:04 crc kubenswrapper[4846]: I1005 08:30:04.541307 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j4mqj\" (UniqueName: \"kubernetes.io/projected/c4867d81-361b-4e63-9860-2099edac282c-kube-api-access-j4mqj\") pod \"c4867d81-361b-4e63-9860-2099edac282c\" (UID: \"c4867d81-361b-4e63-9860-2099edac282c\") " Oct 05 08:30:04 crc kubenswrapper[4846]: I1005 08:30:04.541448 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c4867d81-361b-4e63-9860-2099edac282c-secret-volume\") pod \"c4867d81-361b-4e63-9860-2099edac282c\" (UID: \"c4867d81-361b-4e63-9860-2099edac282c\") " Oct 05 08:30:04 crc kubenswrapper[4846]: I1005 08:30:04.541565 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c4867d81-361b-4e63-9860-2099edac282c-config-volume\") pod \"c4867d81-361b-4e63-9860-2099edac282c\" (UID: \"c4867d81-361b-4e63-9860-2099edac282c\") " Oct 05 08:30:04 crc kubenswrapper[4846]: I1005 08:30:04.543969 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c4867d81-361b-4e63-9860-2099edac282c-config-volume" (OuterVolumeSpecName: "config-volume") pod "c4867d81-361b-4e63-9860-2099edac282c" (UID: "c4867d81-361b-4e63-9860-2099edac282c"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 08:30:04 crc kubenswrapper[4846]: I1005 08:30:04.562723 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c4867d81-361b-4e63-9860-2099edac282c-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "c4867d81-361b-4e63-9860-2099edac282c" (UID: "c4867d81-361b-4e63-9860-2099edac282c"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:30:04 crc kubenswrapper[4846]: I1005 08:30:04.576409 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c4867d81-361b-4e63-9860-2099edac282c-kube-api-access-j4mqj" (OuterVolumeSpecName: "kube-api-access-j4mqj") pod "c4867d81-361b-4e63-9860-2099edac282c" (UID: "c4867d81-361b-4e63-9860-2099edac282c"). InnerVolumeSpecName "kube-api-access-j4mqj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:30:04 crc kubenswrapper[4846]: I1005 08:30:04.645265 4846 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c4867d81-361b-4e63-9860-2099edac282c-config-volume\") on node \"crc\" DevicePath \"\"" Oct 05 08:30:04 crc kubenswrapper[4846]: I1005 08:30:04.645301 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j4mqj\" (UniqueName: \"kubernetes.io/projected/c4867d81-361b-4e63-9860-2099edac282c-kube-api-access-j4mqj\") on node \"crc\" DevicePath \"\"" Oct 05 08:30:04 crc kubenswrapper[4846]: I1005 08:30:04.645316 4846 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c4867d81-361b-4e63-9860-2099edac282c-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 05 08:30:05 crc kubenswrapper[4846]: I1005 08:30:05.235818 4846 generic.go:334] "Generic (PLEG): container finished" podID="7bc1dff8-1996-4278-89f2-1004c67562c3" containerID="4fa9df62f6f1fdf926b083d1d0191ae3233a17658d3b18a6621d988223c7343b" exitCode=0 Oct 05 08:30:05 crc kubenswrapper[4846]: I1005 08:30:05.235919 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9lkxx" event={"ID":"7bc1dff8-1996-4278-89f2-1004c67562c3","Type":"ContainerDied","Data":"4fa9df62f6f1fdf926b083d1d0191ae3233a17658d3b18a6621d988223c7343b"} Oct 05 08:30:05 crc kubenswrapper[4846]: I1005 08:30:05.240052 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29327550-2445r" Oct 05 08:30:05 crc kubenswrapper[4846]: I1005 08:30:05.241260 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29327550-2445r" event={"ID":"c4867d81-361b-4e63-9860-2099edac282c","Type":"ContainerDied","Data":"9139e002c3f22a38d61d22a4ab0bbad57cdb0dff772497fba018bbdf0a30af58"} Oct 05 08:30:05 crc kubenswrapper[4846]: I1005 08:30:05.241312 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9139e002c3f22a38d61d22a4ab0bbad57cdb0dff772497fba018bbdf0a30af58" Oct 05 08:30:05 crc kubenswrapper[4846]: I1005 08:30:05.560628 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29327505-gczm6"] Oct 05 08:30:05 crc kubenswrapper[4846]: I1005 08:30:05.572583 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29327505-gczm6"] Oct 05 08:30:06 crc kubenswrapper[4846]: I1005 08:30:06.021719 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9lkxx" Oct 05 08:30:06 crc kubenswrapper[4846]: I1005 08:30:06.078037 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7bc1dff8-1996-4278-89f2-1004c67562c3-utilities\") pod \"7bc1dff8-1996-4278-89f2-1004c67562c3\" (UID: \"7bc1dff8-1996-4278-89f2-1004c67562c3\") " Oct 05 08:30:06 crc kubenswrapper[4846]: I1005 08:30:06.078108 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7bc1dff8-1996-4278-89f2-1004c67562c3-catalog-content\") pod \"7bc1dff8-1996-4278-89f2-1004c67562c3\" (UID: \"7bc1dff8-1996-4278-89f2-1004c67562c3\") " Oct 05 08:30:06 crc kubenswrapper[4846]: I1005 08:30:06.078292 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vbzrq\" (UniqueName: \"kubernetes.io/projected/7bc1dff8-1996-4278-89f2-1004c67562c3-kube-api-access-vbzrq\") pod \"7bc1dff8-1996-4278-89f2-1004c67562c3\" (UID: \"7bc1dff8-1996-4278-89f2-1004c67562c3\") " Oct 05 08:30:06 crc kubenswrapper[4846]: I1005 08:30:06.080066 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7bc1dff8-1996-4278-89f2-1004c67562c3-utilities" (OuterVolumeSpecName: "utilities") pod "7bc1dff8-1996-4278-89f2-1004c67562c3" (UID: "7bc1dff8-1996-4278-89f2-1004c67562c3"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 08:30:06 crc kubenswrapper[4846]: I1005 08:30:06.080922 4846 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7bc1dff8-1996-4278-89f2-1004c67562c3-utilities\") on node \"crc\" DevicePath \"\"" Oct 05 08:30:06 crc kubenswrapper[4846]: I1005 08:30:06.085710 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bc1dff8-1996-4278-89f2-1004c67562c3-kube-api-access-vbzrq" (OuterVolumeSpecName: "kube-api-access-vbzrq") pod "7bc1dff8-1996-4278-89f2-1004c67562c3" (UID: "7bc1dff8-1996-4278-89f2-1004c67562c3"). InnerVolumeSpecName "kube-api-access-vbzrq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:30:06 crc kubenswrapper[4846]: I1005 08:30:06.126127 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7bc1dff8-1996-4278-89f2-1004c67562c3-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7bc1dff8-1996-4278-89f2-1004c67562c3" (UID: "7bc1dff8-1996-4278-89f2-1004c67562c3"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 08:30:06 crc kubenswrapper[4846]: I1005 08:30:06.183263 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vbzrq\" (UniqueName: \"kubernetes.io/projected/7bc1dff8-1996-4278-89f2-1004c67562c3-kube-api-access-vbzrq\") on node \"crc\" DevicePath \"\"" Oct 05 08:30:06 crc kubenswrapper[4846]: I1005 08:30:06.183501 4846 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7bc1dff8-1996-4278-89f2-1004c67562c3-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 05 08:30:06 crc kubenswrapper[4846]: I1005 08:30:06.250814 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9lkxx" event={"ID":"7bc1dff8-1996-4278-89f2-1004c67562c3","Type":"ContainerDied","Data":"ad318fab82dac6c13b92049619676842c54b0e99733b9328891c31101c44fe33"} Oct 05 08:30:06 crc kubenswrapper[4846]: I1005 08:30:06.250862 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9lkxx" Oct 05 08:30:06 crc kubenswrapper[4846]: I1005 08:30:06.250881 4846 scope.go:117] "RemoveContainer" containerID="4fa9df62f6f1fdf926b083d1d0191ae3233a17658d3b18a6621d988223c7343b" Oct 05 08:30:06 crc kubenswrapper[4846]: I1005 08:30:06.261752 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"317b95c9-9b6d-4ba8-a96c-f69bcc8bb888","Type":"ContainerStarted","Data":"7a6e0f2c3730c396e8eafaf048394a6415c61b772ba967597d899bca0f544e7b"} Oct 05 08:30:06 crc kubenswrapper[4846]: I1005 08:30:06.300299 4846 scope.go:117] "RemoveContainer" containerID="26e9a6b7fb2f28262ee8b8451f374c4fdba5c3a9bc40446b7bb2355af2dfdc7c" Oct 05 08:30:06 crc kubenswrapper[4846]: I1005 08:30:06.304891 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-9lkxx"] Oct 05 08:30:06 crc kubenswrapper[4846]: I1005 08:30:06.314573 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-9lkxx"] Oct 05 08:30:06 crc kubenswrapper[4846]: I1005 08:30:06.321737 4846 scope.go:117] "RemoveContainer" containerID="99ae8d8df0f28c4cd815a55eb43c6ca1128f1d885aa854b5752ba4327d2fdb63" Oct 05 08:30:06 crc kubenswrapper[4846]: I1005 08:30:06.511024 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bc1dff8-1996-4278-89f2-1004c67562c3" path="/var/lib/kubelet/pods/7bc1dff8-1996-4278-89f2-1004c67562c3/volumes" Oct 05 08:30:06 crc kubenswrapper[4846]: I1005 08:30:06.512115 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a4d2ee77-b77c-4ff4-a27a-164efbf89c7a" path="/var/lib/kubelet/pods/a4d2ee77-b77c-4ff4-a27a-164efbf89c7a/volumes" Oct 05 08:30:06 crc kubenswrapper[4846]: E1005 08:30:06.512154 4846 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7bc1dff8_1996_4278_89f2_1004c67562c3.slice/crio-ad318fab82dac6c13b92049619676842c54b0e99733b9328891c31101c44fe33\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9eeca81b_6fda_4c22_a221_3adcb9900077.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9eeca81b_6fda_4c22_a221_3adcb9900077.slice/crio-e8538609494690d52f4d579e3ae6523d65701c417f2551cb9f7b3cad911741a5\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7bc1dff8_1996_4278_89f2_1004c67562c3.slice\": RecentStats: unable to find data in memory cache]" Oct 05 08:30:07 crc kubenswrapper[4846]: I1005 08:30:07.277902 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"317b95c9-9b6d-4ba8-a96c-f69bcc8bb888","Type":"ContainerStarted","Data":"da73a53a1c8eae6d9d176006d4fdf8e62b33d6c8c9090f1c593a8bd0ea79f501"} Oct 05 08:30:08 crc kubenswrapper[4846]: I1005 08:30:08.291876 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"317b95c9-9b6d-4ba8-a96c-f69bcc8bb888","Type":"ContainerStarted","Data":"4c0b66909bef91bfd67e7042c5bc22a2b196bd7f4f03736a48f63caa951e0b5d"} Oct 05 08:30:10 crc kubenswrapper[4846]: I1005 08:30:10.336709 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"317b95c9-9b6d-4ba8-a96c-f69bcc8bb888","Type":"ContainerStarted","Data":"7af5c616692c82ac6dd7f74d046c5062dd9b70877997a1f20373ab3124c53a42"} Oct 05 08:30:10 crc kubenswrapper[4846]: I1005 08:30:10.337394 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 05 08:30:10 crc kubenswrapper[4846]: I1005 08:30:10.383076 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.8268272030000001 podStartE2EDuration="11.383047838s" podCreationTimestamp="2025-10-05 08:29:59 +0000 UTC" firstStartedPulling="2025-10-05 08:29:59.937818346 +0000 UTC m=+6122.178671131" lastFinishedPulling="2025-10-05 08:30:09.494038961 +0000 UTC m=+6131.734891766" observedRunningTime="2025-10-05 08:30:10.363627362 +0000 UTC m=+6132.604480177" watchObservedRunningTime="2025-10-05 08:30:10.383047838 +0000 UTC m=+6132.623900643" Oct 05 08:30:12 crc kubenswrapper[4846]: I1005 08:30:12.357547 4846 generic.go:334] "Generic (PLEG): container finished" podID="8e051aea-4a36-4dd1-9d60-60842891f011" containerID="aafb87ef488436a2880ab6c338f076cbca094992c2e73780b2ea225e5301306c" exitCode=0 Oct 05 08:30:12 crc kubenswrapper[4846]: I1005 08:30:12.357628 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"8e051aea-4a36-4dd1-9d60-60842891f011","Type":"ContainerDied","Data":"aafb87ef488436a2880ab6c338f076cbca094992c2e73780b2ea225e5301306c"} Oct 05 08:30:13 crc kubenswrapper[4846]: I1005 08:30:13.368344 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"8e051aea-4a36-4dd1-9d60-60842891f011","Type":"ContainerStarted","Data":"0d2aaa28e919906c4191c8c95e57c140ca3775ac294af75a7d5fb93b147c32f3"} Oct 05 08:30:13 crc kubenswrapper[4846]: I1005 08:30:13.578476 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-db-create-7sdkh"] Oct 05 08:30:13 crc kubenswrapper[4846]: E1005 08:30:13.579006 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7bc1dff8-1996-4278-89f2-1004c67562c3" containerName="extract-content" Oct 05 08:30:13 crc kubenswrapper[4846]: I1005 08:30:13.579027 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="7bc1dff8-1996-4278-89f2-1004c67562c3" containerName="extract-content" Oct 05 08:30:13 crc kubenswrapper[4846]: E1005 08:30:13.579047 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7bc1dff8-1996-4278-89f2-1004c67562c3" containerName="registry-server" Oct 05 08:30:13 crc kubenswrapper[4846]: I1005 08:30:13.579056 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="7bc1dff8-1996-4278-89f2-1004c67562c3" containerName="registry-server" Oct 05 08:30:13 crc kubenswrapper[4846]: E1005 08:30:13.579080 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c4867d81-361b-4e63-9860-2099edac282c" containerName="collect-profiles" Oct 05 08:30:13 crc kubenswrapper[4846]: I1005 08:30:13.579088 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="c4867d81-361b-4e63-9860-2099edac282c" containerName="collect-profiles" Oct 05 08:30:13 crc kubenswrapper[4846]: E1005 08:30:13.579127 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7bc1dff8-1996-4278-89f2-1004c67562c3" containerName="extract-utilities" Oct 05 08:30:13 crc kubenswrapper[4846]: I1005 08:30:13.579135 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="7bc1dff8-1996-4278-89f2-1004c67562c3" containerName="extract-utilities" Oct 05 08:30:13 crc kubenswrapper[4846]: I1005 08:30:13.579394 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="c4867d81-361b-4e63-9860-2099edac282c" containerName="collect-profiles" Oct 05 08:30:13 crc kubenswrapper[4846]: I1005 08:30:13.579420 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="7bc1dff8-1996-4278-89f2-1004c67562c3" containerName="registry-server" Oct 05 08:30:13 crc kubenswrapper[4846]: I1005 08:30:13.580142 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-create-7sdkh" Oct 05 08:30:13 crc kubenswrapper[4846]: I1005 08:30:13.591050 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-db-create-7sdkh"] Oct 05 08:30:13 crc kubenswrapper[4846]: I1005 08:30:13.674418 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lmv4t\" (UniqueName: \"kubernetes.io/projected/e48f4506-d1b2-4dca-b3d8-cd1693af6c01-kube-api-access-lmv4t\") pod \"aodh-db-create-7sdkh\" (UID: \"e48f4506-d1b2-4dca-b3d8-cd1693af6c01\") " pod="openstack/aodh-db-create-7sdkh" Oct 05 08:30:13 crc kubenswrapper[4846]: I1005 08:30:13.776058 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lmv4t\" (UniqueName: \"kubernetes.io/projected/e48f4506-d1b2-4dca-b3d8-cd1693af6c01-kube-api-access-lmv4t\") pod \"aodh-db-create-7sdkh\" (UID: \"e48f4506-d1b2-4dca-b3d8-cd1693af6c01\") " pod="openstack/aodh-db-create-7sdkh" Oct 05 08:30:13 crc kubenswrapper[4846]: I1005 08:30:13.792324 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lmv4t\" (UniqueName: \"kubernetes.io/projected/e48f4506-d1b2-4dca-b3d8-cd1693af6c01-kube-api-access-lmv4t\") pod \"aodh-db-create-7sdkh\" (UID: \"e48f4506-d1b2-4dca-b3d8-cd1693af6c01\") " pod="openstack/aodh-db-create-7sdkh" Oct 05 08:30:13 crc kubenswrapper[4846]: I1005 08:30:13.903199 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-create-7sdkh" Oct 05 08:30:14 crc kubenswrapper[4846]: I1005 08:30:14.413650 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-db-create-7sdkh"] Oct 05 08:30:15 crc kubenswrapper[4846]: I1005 08:30:15.393066 4846 generic.go:334] "Generic (PLEG): container finished" podID="e48f4506-d1b2-4dca-b3d8-cd1693af6c01" containerID="c579311872b97cc6e9253bc2a70b0c9d987477a88ca327922851f37cd7e6722e" exitCode=0 Oct 05 08:30:15 crc kubenswrapper[4846]: I1005 08:30:15.393450 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-create-7sdkh" event={"ID":"e48f4506-d1b2-4dca-b3d8-cd1693af6c01","Type":"ContainerDied","Data":"c579311872b97cc6e9253bc2a70b0c9d987477a88ca327922851f37cd7e6722e"} Oct 05 08:30:15 crc kubenswrapper[4846]: I1005 08:30:15.393492 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-create-7sdkh" event={"ID":"e48f4506-d1b2-4dca-b3d8-cd1693af6c01","Type":"ContainerStarted","Data":"893c613a32588182aba2558987bece7a4b2622b8c157762cca12d8d9b0be24b4"} Oct 05 08:30:16 crc kubenswrapper[4846]: E1005 08:30:16.799533 4846 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9eeca81b_6fda_4c22_a221_3adcb9900077.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9eeca81b_6fda_4c22_a221_3adcb9900077.slice/crio-e8538609494690d52f4d579e3ae6523d65701c417f2551cb9f7b3cad911741a5\": RecentStats: unable to find data in memory cache]" Oct 05 08:30:16 crc kubenswrapper[4846]: I1005 08:30:16.870314 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-create-7sdkh" Oct 05 08:30:16 crc kubenswrapper[4846]: I1005 08:30:16.967211 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lmv4t\" (UniqueName: \"kubernetes.io/projected/e48f4506-d1b2-4dca-b3d8-cd1693af6c01-kube-api-access-lmv4t\") pod \"e48f4506-d1b2-4dca-b3d8-cd1693af6c01\" (UID: \"e48f4506-d1b2-4dca-b3d8-cd1693af6c01\") " Oct 05 08:30:16 crc kubenswrapper[4846]: I1005 08:30:16.985875 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e48f4506-d1b2-4dca-b3d8-cd1693af6c01-kube-api-access-lmv4t" (OuterVolumeSpecName: "kube-api-access-lmv4t") pod "e48f4506-d1b2-4dca-b3d8-cd1693af6c01" (UID: "e48f4506-d1b2-4dca-b3d8-cd1693af6c01"). InnerVolumeSpecName "kube-api-access-lmv4t". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:30:17 crc kubenswrapper[4846]: I1005 08:30:17.069680 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lmv4t\" (UniqueName: \"kubernetes.io/projected/e48f4506-d1b2-4dca-b3d8-cd1693af6c01-kube-api-access-lmv4t\") on node \"crc\" DevicePath \"\"" Oct 05 08:30:17 crc kubenswrapper[4846]: I1005 08:30:17.429793 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-create-7sdkh" event={"ID":"e48f4506-d1b2-4dca-b3d8-cd1693af6c01","Type":"ContainerDied","Data":"893c613a32588182aba2558987bece7a4b2622b8c157762cca12d8d9b0be24b4"} Oct 05 08:30:17 crc kubenswrapper[4846]: I1005 08:30:17.430122 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="893c613a32588182aba2558987bece7a4b2622b8c157762cca12d8d9b0be24b4" Oct 05 08:30:17 crc kubenswrapper[4846]: I1005 08:30:17.429837 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-create-7sdkh" Oct 05 08:30:17 crc kubenswrapper[4846]: I1005 08:30:17.436619 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"8e051aea-4a36-4dd1-9d60-60842891f011","Type":"ContainerStarted","Data":"252edbfce9d6ce9ee35a53594f220b367ade9d2b2236b4c2bcf481b7e9464b37"} Oct 05 08:30:17 crc kubenswrapper[4846]: I1005 08:30:17.436659 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"8e051aea-4a36-4dd1-9d60-60842891f011","Type":"ContainerStarted","Data":"59b86f13efc420e29b6db1f95aea8f6062795ec1686a7cd10b9c7ab41d1f17e0"} Oct 05 08:30:17 crc kubenswrapper[4846]: I1005 08:30:17.479161 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/prometheus-metric-storage-0" podStartSLOduration=18.479141644 podStartE2EDuration="18.479141644s" podCreationTimestamp="2025-10-05 08:29:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 08:30:17.47446953 +0000 UTC m=+6139.715322305" watchObservedRunningTime="2025-10-05 08:30:17.479141644 +0000 UTC m=+6139.719994429" Oct 05 08:30:19 crc kubenswrapper[4846]: I1005 08:30:19.797245 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/prometheus-metric-storage-0" Oct 05 08:30:23 crc kubenswrapper[4846]: I1005 08:30:23.325006 4846 patch_prober.go:28] interesting pod/machine-config-daemon-fscvf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 08:30:23 crc kubenswrapper[4846]: I1005 08:30:23.325574 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 08:30:23 crc kubenswrapper[4846]: I1005 08:30:23.325668 4846 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" Oct 05 08:30:23 crc kubenswrapper[4846]: I1005 08:30:23.326707 4846 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"5901a64fb8754c398c6b2b246ea8952e5e7f49cedcb4e34d861ec3f14a1f9619"} pod="openshift-machine-config-operator/machine-config-daemon-fscvf" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 05 08:30:23 crc kubenswrapper[4846]: I1005 08:30:23.326815 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" containerName="machine-config-daemon" containerID="cri-o://5901a64fb8754c398c6b2b246ea8952e5e7f49cedcb4e34d861ec3f14a1f9619" gracePeriod=600 Oct 05 08:30:23 crc kubenswrapper[4846]: I1005 08:30:23.524262 4846 generic.go:334] "Generic (PLEG): container finished" podID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" containerID="5901a64fb8754c398c6b2b246ea8952e5e7f49cedcb4e34d861ec3f14a1f9619" exitCode=0 Oct 05 08:30:23 crc kubenswrapper[4846]: I1005 08:30:23.524473 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" event={"ID":"ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7","Type":"ContainerDied","Data":"5901a64fb8754c398c6b2b246ea8952e5e7f49cedcb4e34d861ec3f14a1f9619"} Oct 05 08:30:23 crc kubenswrapper[4846]: I1005 08:30:23.524609 4846 scope.go:117] "RemoveContainer" containerID="fdafa27da24296f9087c33b6b12a6598d2a8d61fc985b06ad680401828c9feb6" Oct 05 08:30:23 crc kubenswrapper[4846]: I1005 08:30:23.725438 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-9865-account-create-fwqsc"] Oct 05 08:30:23 crc kubenswrapper[4846]: E1005 08:30:23.726251 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e48f4506-d1b2-4dca-b3d8-cd1693af6c01" containerName="mariadb-database-create" Oct 05 08:30:23 crc kubenswrapper[4846]: I1005 08:30:23.726276 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="e48f4506-d1b2-4dca-b3d8-cd1693af6c01" containerName="mariadb-database-create" Oct 05 08:30:23 crc kubenswrapper[4846]: I1005 08:30:23.726489 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="e48f4506-d1b2-4dca-b3d8-cd1693af6c01" containerName="mariadb-database-create" Oct 05 08:30:23 crc kubenswrapper[4846]: I1005 08:30:23.728660 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-9865-account-create-fwqsc" Oct 05 08:30:23 crc kubenswrapper[4846]: I1005 08:30:23.731670 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-db-secret" Oct 05 08:30:23 crc kubenswrapper[4846]: I1005 08:30:23.785321 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-9865-account-create-fwqsc"] Oct 05 08:30:23 crc kubenswrapper[4846]: I1005 08:30:23.832549 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nkdk9\" (UniqueName: \"kubernetes.io/projected/331c68dd-4707-4e6c-b6cf-9d9686ef9e7e-kube-api-access-nkdk9\") pod \"aodh-9865-account-create-fwqsc\" (UID: \"331c68dd-4707-4e6c-b6cf-9d9686ef9e7e\") " pod="openstack/aodh-9865-account-create-fwqsc" Oct 05 08:30:23 crc kubenswrapper[4846]: I1005 08:30:23.934756 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nkdk9\" (UniqueName: \"kubernetes.io/projected/331c68dd-4707-4e6c-b6cf-9d9686ef9e7e-kube-api-access-nkdk9\") pod \"aodh-9865-account-create-fwqsc\" (UID: \"331c68dd-4707-4e6c-b6cf-9d9686ef9e7e\") " pod="openstack/aodh-9865-account-create-fwqsc" Oct 05 08:30:23 crc kubenswrapper[4846]: I1005 08:30:23.956525 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nkdk9\" (UniqueName: \"kubernetes.io/projected/331c68dd-4707-4e6c-b6cf-9d9686ef9e7e-kube-api-access-nkdk9\") pod \"aodh-9865-account-create-fwqsc\" (UID: \"331c68dd-4707-4e6c-b6cf-9d9686ef9e7e\") " pod="openstack/aodh-9865-account-create-fwqsc" Oct 05 08:30:24 crc kubenswrapper[4846]: I1005 08:30:24.050945 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-9865-account-create-fwqsc" Oct 05 08:30:24 crc kubenswrapper[4846]: I1005 08:30:24.540129 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" event={"ID":"ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7","Type":"ContainerStarted","Data":"ad5522212e6e288d74747979bc3c3cc6cbe3871d7b66a9bb6a2f8a05f575f23e"} Oct 05 08:30:24 crc kubenswrapper[4846]: I1005 08:30:24.560590 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-9865-account-create-fwqsc"] Oct 05 08:30:24 crc kubenswrapper[4846]: W1005 08:30:24.563430 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod331c68dd_4707_4e6c_b6cf_9d9686ef9e7e.slice/crio-08034c146918d627881f06f77b7903135f34ff3ad1fafbd6e42dca83e45d9211 WatchSource:0}: Error finding container 08034c146918d627881f06f77b7903135f34ff3ad1fafbd6e42dca83e45d9211: Status 404 returned error can't find the container with id 08034c146918d627881f06f77b7903135f34ff3ad1fafbd6e42dca83e45d9211 Oct 05 08:30:25 crc kubenswrapper[4846]: I1005 08:30:25.553529 4846 generic.go:334] "Generic (PLEG): container finished" podID="331c68dd-4707-4e6c-b6cf-9d9686ef9e7e" containerID="9baf95086c181c1170151d82782e7e4354fee9158e2a84f52fa098572c09c14d" exitCode=0 Oct 05 08:30:25 crc kubenswrapper[4846]: I1005 08:30:25.553636 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-9865-account-create-fwqsc" event={"ID":"331c68dd-4707-4e6c-b6cf-9d9686ef9e7e","Type":"ContainerDied","Data":"9baf95086c181c1170151d82782e7e4354fee9158e2a84f52fa098572c09c14d"} Oct 05 08:30:25 crc kubenswrapper[4846]: I1005 08:30:25.553943 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-9865-account-create-fwqsc" event={"ID":"331c68dd-4707-4e6c-b6cf-9d9686ef9e7e","Type":"ContainerStarted","Data":"08034c146918d627881f06f77b7903135f34ff3ad1fafbd6e42dca83e45d9211"} Oct 05 08:30:27 crc kubenswrapper[4846]: I1005 08:30:27.059641 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-9865-account-create-fwqsc" Oct 05 08:30:27 crc kubenswrapper[4846]: I1005 08:30:27.099079 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nkdk9\" (UniqueName: \"kubernetes.io/projected/331c68dd-4707-4e6c-b6cf-9d9686ef9e7e-kube-api-access-nkdk9\") pod \"331c68dd-4707-4e6c-b6cf-9d9686ef9e7e\" (UID: \"331c68dd-4707-4e6c-b6cf-9d9686ef9e7e\") " Oct 05 08:30:27 crc kubenswrapper[4846]: I1005 08:30:27.105557 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/331c68dd-4707-4e6c-b6cf-9d9686ef9e7e-kube-api-access-nkdk9" (OuterVolumeSpecName: "kube-api-access-nkdk9") pod "331c68dd-4707-4e6c-b6cf-9d9686ef9e7e" (UID: "331c68dd-4707-4e6c-b6cf-9d9686ef9e7e"). InnerVolumeSpecName "kube-api-access-nkdk9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:30:27 crc kubenswrapper[4846]: E1005 08:30:27.128505 4846 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9eeca81b_6fda_4c22_a221_3adcb9900077.slice/crio-e8538609494690d52f4d579e3ae6523d65701c417f2551cb9f7b3cad911741a5\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9eeca81b_6fda_4c22_a221_3adcb9900077.slice\": RecentStats: unable to find data in memory cache]" Oct 05 08:30:27 crc kubenswrapper[4846]: I1005 08:30:27.201877 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nkdk9\" (UniqueName: \"kubernetes.io/projected/331c68dd-4707-4e6c-b6cf-9d9686ef9e7e-kube-api-access-nkdk9\") on node \"crc\" DevicePath \"\"" Oct 05 08:30:27 crc kubenswrapper[4846]: I1005 08:30:27.574476 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-9865-account-create-fwqsc" event={"ID":"331c68dd-4707-4e6c-b6cf-9d9686ef9e7e","Type":"ContainerDied","Data":"08034c146918d627881f06f77b7903135f34ff3ad1fafbd6e42dca83e45d9211"} Oct 05 08:30:27 crc kubenswrapper[4846]: I1005 08:30:27.574693 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="08034c146918d627881f06f77b7903135f34ff3ad1fafbd6e42dca83e45d9211" Oct 05 08:30:27 crc kubenswrapper[4846]: I1005 08:30:27.574540 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-9865-account-create-fwqsc" Oct 05 08:30:29 crc kubenswrapper[4846]: I1005 08:30:29.150650 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-db-sync-5tnzg"] Oct 05 08:30:29 crc kubenswrapper[4846]: E1005 08:30:29.152827 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="331c68dd-4707-4e6c-b6cf-9d9686ef9e7e" containerName="mariadb-account-create" Oct 05 08:30:29 crc kubenswrapper[4846]: I1005 08:30:29.152842 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="331c68dd-4707-4e6c-b6cf-9d9686ef9e7e" containerName="mariadb-account-create" Oct 05 08:30:29 crc kubenswrapper[4846]: I1005 08:30:29.153050 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="331c68dd-4707-4e6c-b6cf-9d9686ef9e7e" containerName="mariadb-account-create" Oct 05 08:30:29 crc kubenswrapper[4846]: I1005 08:30:29.153970 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-5tnzg" Oct 05 08:30:29 crc kubenswrapper[4846]: I1005 08:30:29.160585 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-scripts" Oct 05 08:30:29 crc kubenswrapper[4846]: I1005 08:30:29.160704 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-autoscaling-dockercfg-666qr" Oct 05 08:30:29 crc kubenswrapper[4846]: I1005 08:30:29.160906 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-config-data" Oct 05 08:30:29 crc kubenswrapper[4846]: I1005 08:30:29.171849 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-db-sync-5tnzg"] Oct 05 08:30:29 crc kubenswrapper[4846]: I1005 08:30:29.247411 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/19856b5e-44a2-4b5e-bee6-2cde2af29996-config-data\") pod \"aodh-db-sync-5tnzg\" (UID: \"19856b5e-44a2-4b5e-bee6-2cde2af29996\") " pod="openstack/aodh-db-sync-5tnzg" Oct 05 08:30:29 crc kubenswrapper[4846]: I1005 08:30:29.247599 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/19856b5e-44a2-4b5e-bee6-2cde2af29996-scripts\") pod \"aodh-db-sync-5tnzg\" (UID: \"19856b5e-44a2-4b5e-bee6-2cde2af29996\") " pod="openstack/aodh-db-sync-5tnzg" Oct 05 08:30:29 crc kubenswrapper[4846]: I1005 08:30:29.247863 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/19856b5e-44a2-4b5e-bee6-2cde2af29996-combined-ca-bundle\") pod \"aodh-db-sync-5tnzg\" (UID: \"19856b5e-44a2-4b5e-bee6-2cde2af29996\") " pod="openstack/aodh-db-sync-5tnzg" Oct 05 08:30:29 crc kubenswrapper[4846]: I1005 08:30:29.247975 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zsgfq\" (UniqueName: \"kubernetes.io/projected/19856b5e-44a2-4b5e-bee6-2cde2af29996-kube-api-access-zsgfq\") pod \"aodh-db-sync-5tnzg\" (UID: \"19856b5e-44a2-4b5e-bee6-2cde2af29996\") " pod="openstack/aodh-db-sync-5tnzg" Oct 05 08:30:29 crc kubenswrapper[4846]: I1005 08:30:29.348739 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zsgfq\" (UniqueName: \"kubernetes.io/projected/19856b5e-44a2-4b5e-bee6-2cde2af29996-kube-api-access-zsgfq\") pod \"aodh-db-sync-5tnzg\" (UID: \"19856b5e-44a2-4b5e-bee6-2cde2af29996\") " pod="openstack/aodh-db-sync-5tnzg" Oct 05 08:30:29 crc kubenswrapper[4846]: I1005 08:30:29.348798 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/19856b5e-44a2-4b5e-bee6-2cde2af29996-config-data\") pod \"aodh-db-sync-5tnzg\" (UID: \"19856b5e-44a2-4b5e-bee6-2cde2af29996\") " pod="openstack/aodh-db-sync-5tnzg" Oct 05 08:30:29 crc kubenswrapper[4846]: I1005 08:30:29.348872 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/19856b5e-44a2-4b5e-bee6-2cde2af29996-scripts\") pod \"aodh-db-sync-5tnzg\" (UID: \"19856b5e-44a2-4b5e-bee6-2cde2af29996\") " pod="openstack/aodh-db-sync-5tnzg" Oct 05 08:30:29 crc kubenswrapper[4846]: I1005 08:30:29.348942 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/19856b5e-44a2-4b5e-bee6-2cde2af29996-combined-ca-bundle\") pod \"aodh-db-sync-5tnzg\" (UID: \"19856b5e-44a2-4b5e-bee6-2cde2af29996\") " pod="openstack/aodh-db-sync-5tnzg" Oct 05 08:30:29 crc kubenswrapper[4846]: I1005 08:30:29.355049 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/19856b5e-44a2-4b5e-bee6-2cde2af29996-config-data\") pod \"aodh-db-sync-5tnzg\" (UID: \"19856b5e-44a2-4b5e-bee6-2cde2af29996\") " pod="openstack/aodh-db-sync-5tnzg" Oct 05 08:30:29 crc kubenswrapper[4846]: I1005 08:30:29.355995 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/19856b5e-44a2-4b5e-bee6-2cde2af29996-scripts\") pod \"aodh-db-sync-5tnzg\" (UID: \"19856b5e-44a2-4b5e-bee6-2cde2af29996\") " pod="openstack/aodh-db-sync-5tnzg" Oct 05 08:30:29 crc kubenswrapper[4846]: I1005 08:30:29.356211 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/19856b5e-44a2-4b5e-bee6-2cde2af29996-combined-ca-bundle\") pod \"aodh-db-sync-5tnzg\" (UID: \"19856b5e-44a2-4b5e-bee6-2cde2af29996\") " pod="openstack/aodh-db-sync-5tnzg" Oct 05 08:30:29 crc kubenswrapper[4846]: I1005 08:30:29.368376 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zsgfq\" (UniqueName: \"kubernetes.io/projected/19856b5e-44a2-4b5e-bee6-2cde2af29996-kube-api-access-zsgfq\") pod \"aodh-db-sync-5tnzg\" (UID: \"19856b5e-44a2-4b5e-bee6-2cde2af29996\") " pod="openstack/aodh-db-sync-5tnzg" Oct 05 08:30:29 crc kubenswrapper[4846]: I1005 08:30:29.431450 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Oct 05 08:30:29 crc kubenswrapper[4846]: I1005 08:30:29.478735 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-5tnzg" Oct 05 08:30:29 crc kubenswrapper[4846]: I1005 08:30:29.799400 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/prometheus-metric-storage-0" Oct 05 08:30:29 crc kubenswrapper[4846]: I1005 08:30:29.805165 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/prometheus-metric-storage-0" Oct 05 08:30:29 crc kubenswrapper[4846]: I1005 08:30:29.968321 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-db-sync-5tnzg"] Oct 05 08:30:30 crc kubenswrapper[4846]: W1005 08:30:30.059578 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod19856b5e_44a2_4b5e_bee6_2cde2af29996.slice/crio-ce61faaadbeef1c1169438e93fbe81bb0f29b20ff25652e5fe80d75cf1c677f3 WatchSource:0}: Error finding container ce61faaadbeef1c1169438e93fbe81bb0f29b20ff25652e5fe80d75cf1c677f3: Status 404 returned error can't find the container with id ce61faaadbeef1c1169438e93fbe81bb0f29b20ff25652e5fe80d75cf1c677f3 Oct 05 08:30:30 crc kubenswrapper[4846]: I1005 08:30:30.652234 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-5tnzg" event={"ID":"19856b5e-44a2-4b5e-bee6-2cde2af29996","Type":"ContainerStarted","Data":"ce61faaadbeef1c1169438e93fbe81bb0f29b20ff25652e5fe80d75cf1c677f3"} Oct 05 08:30:30 crc kubenswrapper[4846]: I1005 08:30:30.656959 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/prometheus-metric-storage-0" Oct 05 08:30:34 crc kubenswrapper[4846]: I1005 08:30:34.076450 4846 scope.go:117] "RemoveContainer" containerID="5c4be27156878d0b803aae47db8a307662fe8dd8ff93e8dd174a04950cc25827" Oct 05 08:30:34 crc kubenswrapper[4846]: I1005 08:30:34.419222 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 05 08:30:34 crc kubenswrapper[4846]: I1005 08:30:34.419496 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="502e1afd-2dc4-49f5-9c20-a73834426de8" containerName="kube-state-metrics" containerID="cri-o://584a6806d2a36bfd7832c05c23cd2ef062e16cae388ebc0095dfa18e6efd8aae" gracePeriod=30 Oct 05 08:30:34 crc kubenswrapper[4846]: I1005 08:30:34.638233 4846 scope.go:117] "RemoveContainer" containerID="488875f124d0a3c38ab627e4c84ddc6477d5a7ae2ce09b96bd4b553918976a69" Oct 05 08:30:34 crc kubenswrapper[4846]: I1005 08:30:34.699679 4846 generic.go:334] "Generic (PLEG): container finished" podID="502e1afd-2dc4-49f5-9c20-a73834426de8" containerID="584a6806d2a36bfd7832c05c23cd2ef062e16cae388ebc0095dfa18e6efd8aae" exitCode=2 Oct 05 08:30:34 crc kubenswrapper[4846]: I1005 08:30:34.699752 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"502e1afd-2dc4-49f5-9c20-a73834426de8","Type":"ContainerDied","Data":"584a6806d2a36bfd7832c05c23cd2ef062e16cae388ebc0095dfa18e6efd8aae"} Oct 05 08:30:34 crc kubenswrapper[4846]: I1005 08:30:34.700256 4846 scope.go:117] "RemoveContainer" containerID="bbb7fcf74c7b4ccfb8c58ccda5d6ab76c1a1670b40113965e1dd4965a295a344" Oct 05 08:30:34 crc kubenswrapper[4846]: I1005 08:30:34.932507 4846 scope.go:117] "RemoveContainer" containerID="2aaed7a87773121e7ebff2f4276b7f6b86cd9bf885d338e114304462977b8737" Oct 05 08:30:34 crc kubenswrapper[4846]: I1005 08:30:34.974092 4846 scope.go:117] "RemoveContainer" containerID="63fe74b1e29ffea6657ebecf1c49c55957af6382c195a72b7cc62916280cf64d" Oct 05 08:30:34 crc kubenswrapper[4846]: I1005 08:30:34.995566 4846 scope.go:117] "RemoveContainer" containerID="006e3a4d809a2526a062908765859aa3087dc6d5e6f8d19c45d2849dc9615a2f" Oct 05 08:30:35 crc kubenswrapper[4846]: I1005 08:30:35.025307 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 05 08:30:35 crc kubenswrapper[4846]: I1005 08:30:35.165777 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bvf9c\" (UniqueName: \"kubernetes.io/projected/502e1afd-2dc4-49f5-9c20-a73834426de8-kube-api-access-bvf9c\") pod \"502e1afd-2dc4-49f5-9c20-a73834426de8\" (UID: \"502e1afd-2dc4-49f5-9c20-a73834426de8\") " Oct 05 08:30:35 crc kubenswrapper[4846]: I1005 08:30:35.174290 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/502e1afd-2dc4-49f5-9c20-a73834426de8-kube-api-access-bvf9c" (OuterVolumeSpecName: "kube-api-access-bvf9c") pod "502e1afd-2dc4-49f5-9c20-a73834426de8" (UID: "502e1afd-2dc4-49f5-9c20-a73834426de8"). InnerVolumeSpecName "kube-api-access-bvf9c". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:30:35 crc kubenswrapper[4846]: I1005 08:30:35.268688 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bvf9c\" (UniqueName: \"kubernetes.io/projected/502e1afd-2dc4-49f5-9c20-a73834426de8-kube-api-access-bvf9c\") on node \"crc\" DevicePath \"\"" Oct 05 08:30:35 crc kubenswrapper[4846]: I1005 08:30:35.729900 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-5tnzg" event={"ID":"19856b5e-44a2-4b5e-bee6-2cde2af29996","Type":"ContainerStarted","Data":"2dc7314cc10ac3e793d12d8c243041065e139b7ece7dc8cca1794cec5b2f85f5"} Oct 05 08:30:35 crc kubenswrapper[4846]: I1005 08:30:35.732000 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"502e1afd-2dc4-49f5-9c20-a73834426de8","Type":"ContainerDied","Data":"8e3301baf37fa9917c8d8f20a2aad6fcaa7c19dd723cf3e6ba184816ffd05cb5"} Oct 05 08:30:35 crc kubenswrapper[4846]: I1005 08:30:35.732043 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 05 08:30:35 crc kubenswrapper[4846]: I1005 08:30:35.732061 4846 scope.go:117] "RemoveContainer" containerID="584a6806d2a36bfd7832c05c23cd2ef062e16cae388ebc0095dfa18e6efd8aae" Oct 05 08:30:35 crc kubenswrapper[4846]: I1005 08:30:35.782216 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/aodh-db-sync-5tnzg" podStartSLOduration=2.102153443 podStartE2EDuration="6.782170591s" podCreationTimestamp="2025-10-05 08:30:29 +0000 UTC" firstStartedPulling="2025-10-05 08:30:30.061212861 +0000 UTC m=+6152.302065636" lastFinishedPulling="2025-10-05 08:30:34.741229999 +0000 UTC m=+6156.982082784" observedRunningTime="2025-10-05 08:30:35.747673745 +0000 UTC m=+6157.988526520" watchObservedRunningTime="2025-10-05 08:30:35.782170591 +0000 UTC m=+6158.023023376" Oct 05 08:30:35 crc kubenswrapper[4846]: I1005 08:30:35.805194 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 05 08:30:35 crc kubenswrapper[4846]: I1005 08:30:35.818391 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 05 08:30:35 crc kubenswrapper[4846]: I1005 08:30:35.829709 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Oct 05 08:30:35 crc kubenswrapper[4846]: E1005 08:30:35.830233 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="502e1afd-2dc4-49f5-9c20-a73834426de8" containerName="kube-state-metrics" Oct 05 08:30:35 crc kubenswrapper[4846]: I1005 08:30:35.830250 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="502e1afd-2dc4-49f5-9c20-a73834426de8" containerName="kube-state-metrics" Oct 05 08:30:35 crc kubenswrapper[4846]: I1005 08:30:35.830455 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="502e1afd-2dc4-49f5-9c20-a73834426de8" containerName="kube-state-metrics" Oct 05 08:30:35 crc kubenswrapper[4846]: I1005 08:30:35.831297 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 05 08:30:35 crc kubenswrapper[4846]: I1005 08:30:35.834136 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-kube-state-metrics-svc" Oct 05 08:30:35 crc kubenswrapper[4846]: I1005 08:30:35.834368 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"kube-state-metrics-tls-config" Oct 05 08:30:35 crc kubenswrapper[4846]: I1005 08:30:35.846344 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 05 08:30:35 crc kubenswrapper[4846]: I1005 08:30:35.986740 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/f89bcf16-c362-44a0-9869-551cea0beacc-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"f89bcf16-c362-44a0-9869-551cea0beacc\") " pod="openstack/kube-state-metrics-0" Oct 05 08:30:35 crc kubenswrapper[4846]: I1005 08:30:35.986852 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/f89bcf16-c362-44a0-9869-551cea0beacc-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"f89bcf16-c362-44a0-9869-551cea0beacc\") " pod="openstack/kube-state-metrics-0" Oct 05 08:30:35 crc kubenswrapper[4846]: I1005 08:30:35.986886 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f89bcf16-c362-44a0-9869-551cea0beacc-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"f89bcf16-c362-44a0-9869-551cea0beacc\") " pod="openstack/kube-state-metrics-0" Oct 05 08:30:35 crc kubenswrapper[4846]: I1005 08:30:35.987124 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j95sj\" (UniqueName: \"kubernetes.io/projected/f89bcf16-c362-44a0-9869-551cea0beacc-kube-api-access-j95sj\") pod \"kube-state-metrics-0\" (UID: \"f89bcf16-c362-44a0-9869-551cea0beacc\") " pod="openstack/kube-state-metrics-0" Oct 05 08:30:36 crc kubenswrapper[4846]: I1005 08:30:36.089503 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/f89bcf16-c362-44a0-9869-551cea0beacc-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"f89bcf16-c362-44a0-9869-551cea0beacc\") " pod="openstack/kube-state-metrics-0" Oct 05 08:30:36 crc kubenswrapper[4846]: I1005 08:30:36.089590 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f89bcf16-c362-44a0-9869-551cea0beacc-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"f89bcf16-c362-44a0-9869-551cea0beacc\") " pod="openstack/kube-state-metrics-0" Oct 05 08:30:36 crc kubenswrapper[4846]: I1005 08:30:36.089682 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j95sj\" (UniqueName: \"kubernetes.io/projected/f89bcf16-c362-44a0-9869-551cea0beacc-kube-api-access-j95sj\") pod \"kube-state-metrics-0\" (UID: \"f89bcf16-c362-44a0-9869-551cea0beacc\") " pod="openstack/kube-state-metrics-0" Oct 05 08:30:36 crc kubenswrapper[4846]: I1005 08:30:36.089792 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/f89bcf16-c362-44a0-9869-551cea0beacc-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"f89bcf16-c362-44a0-9869-551cea0beacc\") " pod="openstack/kube-state-metrics-0" Oct 05 08:30:36 crc kubenswrapper[4846]: I1005 08:30:36.099799 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/f89bcf16-c362-44a0-9869-551cea0beacc-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"f89bcf16-c362-44a0-9869-551cea0beacc\") " pod="openstack/kube-state-metrics-0" Oct 05 08:30:36 crc kubenswrapper[4846]: I1005 08:30:36.101928 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f89bcf16-c362-44a0-9869-551cea0beacc-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"f89bcf16-c362-44a0-9869-551cea0beacc\") " pod="openstack/kube-state-metrics-0" Oct 05 08:30:36 crc kubenswrapper[4846]: I1005 08:30:36.106216 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/f89bcf16-c362-44a0-9869-551cea0beacc-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"f89bcf16-c362-44a0-9869-551cea0beacc\") " pod="openstack/kube-state-metrics-0" Oct 05 08:30:36 crc kubenswrapper[4846]: I1005 08:30:36.121932 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j95sj\" (UniqueName: \"kubernetes.io/projected/f89bcf16-c362-44a0-9869-551cea0beacc-kube-api-access-j95sj\") pod \"kube-state-metrics-0\" (UID: \"f89bcf16-c362-44a0-9869-551cea0beacc\") " pod="openstack/kube-state-metrics-0" Oct 05 08:30:36 crc kubenswrapper[4846]: I1005 08:30:36.148654 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 05 08:30:36 crc kubenswrapper[4846]: I1005 08:30:36.425104 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 05 08:30:36 crc kubenswrapper[4846]: I1005 08:30:36.425650 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="317b95c9-9b6d-4ba8-a96c-f69bcc8bb888" containerName="ceilometer-central-agent" containerID="cri-o://7a6e0f2c3730c396e8eafaf048394a6415c61b772ba967597d899bca0f544e7b" gracePeriod=30 Oct 05 08:30:36 crc kubenswrapper[4846]: I1005 08:30:36.426067 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="317b95c9-9b6d-4ba8-a96c-f69bcc8bb888" containerName="proxy-httpd" containerID="cri-o://7af5c616692c82ac6dd7f74d046c5062dd9b70877997a1f20373ab3124c53a42" gracePeriod=30 Oct 05 08:30:36 crc kubenswrapper[4846]: I1005 08:30:36.426115 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="317b95c9-9b6d-4ba8-a96c-f69bcc8bb888" containerName="sg-core" containerID="cri-o://4c0b66909bef91bfd67e7042c5bc22a2b196bd7f4f03736a48f63caa951e0b5d" gracePeriod=30 Oct 05 08:30:36 crc kubenswrapper[4846]: I1005 08:30:36.426150 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="317b95c9-9b6d-4ba8-a96c-f69bcc8bb888" containerName="ceilometer-notification-agent" containerID="cri-o://da73a53a1c8eae6d9d176006d4fdf8e62b33d6c8c9090f1c593a8bd0ea79f501" gracePeriod=30 Oct 05 08:30:36 crc kubenswrapper[4846]: I1005 08:30:36.507992 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="502e1afd-2dc4-49f5-9c20-a73834426de8" path="/var/lib/kubelet/pods/502e1afd-2dc4-49f5-9c20-a73834426de8/volumes" Oct 05 08:30:36 crc kubenswrapper[4846]: I1005 08:30:36.635916 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 05 08:30:36 crc kubenswrapper[4846]: I1005 08:30:36.743559 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"f89bcf16-c362-44a0-9869-551cea0beacc","Type":"ContainerStarted","Data":"8a490c9269ff1f5c853437005c439c55307ee2402525b70dc5e3cc8074b27421"} Oct 05 08:30:36 crc kubenswrapper[4846]: I1005 08:30:36.746769 4846 generic.go:334] "Generic (PLEG): container finished" podID="317b95c9-9b6d-4ba8-a96c-f69bcc8bb888" containerID="7af5c616692c82ac6dd7f74d046c5062dd9b70877997a1f20373ab3124c53a42" exitCode=0 Oct 05 08:30:36 crc kubenswrapper[4846]: I1005 08:30:36.746799 4846 generic.go:334] "Generic (PLEG): container finished" podID="317b95c9-9b6d-4ba8-a96c-f69bcc8bb888" containerID="4c0b66909bef91bfd67e7042c5bc22a2b196bd7f4f03736a48f63caa951e0b5d" exitCode=2 Oct 05 08:30:36 crc kubenswrapper[4846]: I1005 08:30:36.747061 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"317b95c9-9b6d-4ba8-a96c-f69bcc8bb888","Type":"ContainerDied","Data":"7af5c616692c82ac6dd7f74d046c5062dd9b70877997a1f20373ab3124c53a42"} Oct 05 08:30:36 crc kubenswrapper[4846]: I1005 08:30:36.747106 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"317b95c9-9b6d-4ba8-a96c-f69bcc8bb888","Type":"ContainerDied","Data":"4c0b66909bef91bfd67e7042c5bc22a2b196bd7f4f03736a48f63caa951e0b5d"} Oct 05 08:30:37 crc kubenswrapper[4846]: E1005 08:30:37.420748 4846 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod19856b5e_44a2_4b5e_bee6_2cde2af29996.slice/crio-conmon-2dc7314cc10ac3e793d12d8c243041065e139b7ece7dc8cca1794cec5b2f85f5.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9eeca81b_6fda_4c22_a221_3adcb9900077.slice/crio-e8538609494690d52f4d579e3ae6523d65701c417f2551cb9f7b3cad911741a5\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9eeca81b_6fda_4c22_a221_3adcb9900077.slice\": RecentStats: unable to find data in memory cache]" Oct 05 08:30:37 crc kubenswrapper[4846]: I1005 08:30:37.759571 4846 generic.go:334] "Generic (PLEG): container finished" podID="317b95c9-9b6d-4ba8-a96c-f69bcc8bb888" containerID="7a6e0f2c3730c396e8eafaf048394a6415c61b772ba967597d899bca0f544e7b" exitCode=0 Oct 05 08:30:37 crc kubenswrapper[4846]: I1005 08:30:37.759627 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"317b95c9-9b6d-4ba8-a96c-f69bcc8bb888","Type":"ContainerDied","Data":"7a6e0f2c3730c396e8eafaf048394a6415c61b772ba967597d899bca0f544e7b"} Oct 05 08:30:37 crc kubenswrapper[4846]: I1005 08:30:37.761304 4846 generic.go:334] "Generic (PLEG): container finished" podID="19856b5e-44a2-4b5e-bee6-2cde2af29996" containerID="2dc7314cc10ac3e793d12d8c243041065e139b7ece7dc8cca1794cec5b2f85f5" exitCode=0 Oct 05 08:30:37 crc kubenswrapper[4846]: I1005 08:30:37.761344 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-5tnzg" event={"ID":"19856b5e-44a2-4b5e-bee6-2cde2af29996","Type":"ContainerDied","Data":"2dc7314cc10ac3e793d12d8c243041065e139b7ece7dc8cca1794cec5b2f85f5"} Oct 05 08:30:37 crc kubenswrapper[4846]: I1005 08:30:37.763174 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"f89bcf16-c362-44a0-9869-551cea0beacc","Type":"ContainerStarted","Data":"2ee5ab87878953ed41b080109d4497225f512ee25635c7210906ab156238642a"} Oct 05 08:30:37 crc kubenswrapper[4846]: I1005 08:30:37.763934 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Oct 05 08:30:37 crc kubenswrapper[4846]: I1005 08:30:37.818475 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=2.467401513 podStartE2EDuration="2.818448595s" podCreationTimestamp="2025-10-05 08:30:35 +0000 UTC" firstStartedPulling="2025-10-05 08:30:36.640781592 +0000 UTC m=+6158.881634367" lastFinishedPulling="2025-10-05 08:30:36.991828674 +0000 UTC m=+6159.232681449" observedRunningTime="2025-10-05 08:30:37.803558929 +0000 UTC m=+6160.044411714" watchObservedRunningTime="2025-10-05 08:30:37.818448595 +0000 UTC m=+6160.059301370" Oct 05 08:30:39 crc kubenswrapper[4846]: I1005 08:30:39.258733 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-5tnzg" Oct 05 08:30:39 crc kubenswrapper[4846]: I1005 08:30:39.366293 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/19856b5e-44a2-4b5e-bee6-2cde2af29996-config-data\") pod \"19856b5e-44a2-4b5e-bee6-2cde2af29996\" (UID: \"19856b5e-44a2-4b5e-bee6-2cde2af29996\") " Oct 05 08:30:39 crc kubenswrapper[4846]: I1005 08:30:39.366606 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zsgfq\" (UniqueName: \"kubernetes.io/projected/19856b5e-44a2-4b5e-bee6-2cde2af29996-kube-api-access-zsgfq\") pod \"19856b5e-44a2-4b5e-bee6-2cde2af29996\" (UID: \"19856b5e-44a2-4b5e-bee6-2cde2af29996\") " Oct 05 08:30:39 crc kubenswrapper[4846]: I1005 08:30:39.366647 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/19856b5e-44a2-4b5e-bee6-2cde2af29996-scripts\") pod \"19856b5e-44a2-4b5e-bee6-2cde2af29996\" (UID: \"19856b5e-44a2-4b5e-bee6-2cde2af29996\") " Oct 05 08:30:39 crc kubenswrapper[4846]: I1005 08:30:39.366692 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/19856b5e-44a2-4b5e-bee6-2cde2af29996-combined-ca-bundle\") pod \"19856b5e-44a2-4b5e-bee6-2cde2af29996\" (UID: \"19856b5e-44a2-4b5e-bee6-2cde2af29996\") " Oct 05 08:30:39 crc kubenswrapper[4846]: I1005 08:30:39.380742 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/19856b5e-44a2-4b5e-bee6-2cde2af29996-scripts" (OuterVolumeSpecName: "scripts") pod "19856b5e-44a2-4b5e-bee6-2cde2af29996" (UID: "19856b5e-44a2-4b5e-bee6-2cde2af29996"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:30:39 crc kubenswrapper[4846]: I1005 08:30:39.387331 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/19856b5e-44a2-4b5e-bee6-2cde2af29996-kube-api-access-zsgfq" (OuterVolumeSpecName: "kube-api-access-zsgfq") pod "19856b5e-44a2-4b5e-bee6-2cde2af29996" (UID: "19856b5e-44a2-4b5e-bee6-2cde2af29996"). InnerVolumeSpecName "kube-api-access-zsgfq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:30:39 crc kubenswrapper[4846]: I1005 08:30:39.428579 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/19856b5e-44a2-4b5e-bee6-2cde2af29996-config-data" (OuterVolumeSpecName: "config-data") pod "19856b5e-44a2-4b5e-bee6-2cde2af29996" (UID: "19856b5e-44a2-4b5e-bee6-2cde2af29996"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:30:39 crc kubenswrapper[4846]: I1005 08:30:39.434134 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/19856b5e-44a2-4b5e-bee6-2cde2af29996-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "19856b5e-44a2-4b5e-bee6-2cde2af29996" (UID: "19856b5e-44a2-4b5e-bee6-2cde2af29996"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:30:39 crc kubenswrapper[4846]: I1005 08:30:39.469669 4846 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/19856b5e-44a2-4b5e-bee6-2cde2af29996-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 08:30:39 crc kubenswrapper[4846]: I1005 08:30:39.469709 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zsgfq\" (UniqueName: \"kubernetes.io/projected/19856b5e-44a2-4b5e-bee6-2cde2af29996-kube-api-access-zsgfq\") on node \"crc\" DevicePath \"\"" Oct 05 08:30:39 crc kubenswrapper[4846]: I1005 08:30:39.469720 4846 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/19856b5e-44a2-4b5e-bee6-2cde2af29996-scripts\") on node \"crc\" DevicePath \"\"" Oct 05 08:30:39 crc kubenswrapper[4846]: I1005 08:30:39.469729 4846 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/19856b5e-44a2-4b5e-bee6-2cde2af29996-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 08:30:39 crc kubenswrapper[4846]: I1005 08:30:39.531136 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 05 08:30:39 crc kubenswrapper[4846]: I1005 08:30:39.671713 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/317b95c9-9b6d-4ba8-a96c-f69bcc8bb888-combined-ca-bundle\") pod \"317b95c9-9b6d-4ba8-a96c-f69bcc8bb888\" (UID: \"317b95c9-9b6d-4ba8-a96c-f69bcc8bb888\") " Oct 05 08:30:39 crc kubenswrapper[4846]: I1005 08:30:39.671802 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/317b95c9-9b6d-4ba8-a96c-f69bcc8bb888-scripts\") pod \"317b95c9-9b6d-4ba8-a96c-f69bcc8bb888\" (UID: \"317b95c9-9b6d-4ba8-a96c-f69bcc8bb888\") " Oct 05 08:30:39 crc kubenswrapper[4846]: I1005 08:30:39.672697 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/317b95c9-9b6d-4ba8-a96c-f69bcc8bb888-run-httpd\") pod \"317b95c9-9b6d-4ba8-a96c-f69bcc8bb888\" (UID: \"317b95c9-9b6d-4ba8-a96c-f69bcc8bb888\") " Oct 05 08:30:39 crc kubenswrapper[4846]: I1005 08:30:39.672748 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pfttb\" (UniqueName: \"kubernetes.io/projected/317b95c9-9b6d-4ba8-a96c-f69bcc8bb888-kube-api-access-pfttb\") pod \"317b95c9-9b6d-4ba8-a96c-f69bcc8bb888\" (UID: \"317b95c9-9b6d-4ba8-a96c-f69bcc8bb888\") " Oct 05 08:30:39 crc kubenswrapper[4846]: I1005 08:30:39.672958 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/317b95c9-9b6d-4ba8-a96c-f69bcc8bb888-log-httpd\") pod \"317b95c9-9b6d-4ba8-a96c-f69bcc8bb888\" (UID: \"317b95c9-9b6d-4ba8-a96c-f69bcc8bb888\") " Oct 05 08:30:39 crc kubenswrapper[4846]: I1005 08:30:39.673065 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/317b95c9-9b6d-4ba8-a96c-f69bcc8bb888-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "317b95c9-9b6d-4ba8-a96c-f69bcc8bb888" (UID: "317b95c9-9b6d-4ba8-a96c-f69bcc8bb888"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 08:30:39 crc kubenswrapper[4846]: I1005 08:30:39.673083 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/317b95c9-9b6d-4ba8-a96c-f69bcc8bb888-config-data\") pod \"317b95c9-9b6d-4ba8-a96c-f69bcc8bb888\" (UID: \"317b95c9-9b6d-4ba8-a96c-f69bcc8bb888\") " Oct 05 08:30:39 crc kubenswrapper[4846]: I1005 08:30:39.673131 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/317b95c9-9b6d-4ba8-a96c-f69bcc8bb888-sg-core-conf-yaml\") pod \"317b95c9-9b6d-4ba8-a96c-f69bcc8bb888\" (UID: \"317b95c9-9b6d-4ba8-a96c-f69bcc8bb888\") " Oct 05 08:30:39 crc kubenswrapper[4846]: I1005 08:30:39.673847 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/317b95c9-9b6d-4ba8-a96c-f69bcc8bb888-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "317b95c9-9b6d-4ba8-a96c-f69bcc8bb888" (UID: "317b95c9-9b6d-4ba8-a96c-f69bcc8bb888"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 08:30:39 crc kubenswrapper[4846]: I1005 08:30:39.674274 4846 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/317b95c9-9b6d-4ba8-a96c-f69bcc8bb888-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 05 08:30:39 crc kubenswrapper[4846]: I1005 08:30:39.675415 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/317b95c9-9b6d-4ba8-a96c-f69bcc8bb888-scripts" (OuterVolumeSpecName: "scripts") pod "317b95c9-9b6d-4ba8-a96c-f69bcc8bb888" (UID: "317b95c9-9b6d-4ba8-a96c-f69bcc8bb888"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:30:39 crc kubenswrapper[4846]: I1005 08:30:39.676044 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/317b95c9-9b6d-4ba8-a96c-f69bcc8bb888-kube-api-access-pfttb" (OuterVolumeSpecName: "kube-api-access-pfttb") pod "317b95c9-9b6d-4ba8-a96c-f69bcc8bb888" (UID: "317b95c9-9b6d-4ba8-a96c-f69bcc8bb888"). InnerVolumeSpecName "kube-api-access-pfttb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:30:39 crc kubenswrapper[4846]: I1005 08:30:39.696842 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/317b95c9-9b6d-4ba8-a96c-f69bcc8bb888-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "317b95c9-9b6d-4ba8-a96c-f69bcc8bb888" (UID: "317b95c9-9b6d-4ba8-a96c-f69bcc8bb888"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:30:39 crc kubenswrapper[4846]: I1005 08:30:39.776029 4846 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/317b95c9-9b6d-4ba8-a96c-f69bcc8bb888-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 05 08:30:39 crc kubenswrapper[4846]: I1005 08:30:39.776072 4846 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/317b95c9-9b6d-4ba8-a96c-f69bcc8bb888-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 05 08:30:39 crc kubenswrapper[4846]: I1005 08:30:39.776093 4846 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/317b95c9-9b6d-4ba8-a96c-f69bcc8bb888-scripts\") on node \"crc\" DevicePath \"\"" Oct 05 08:30:39 crc kubenswrapper[4846]: I1005 08:30:39.776109 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pfttb\" (UniqueName: \"kubernetes.io/projected/317b95c9-9b6d-4ba8-a96c-f69bcc8bb888-kube-api-access-pfttb\") on node \"crc\" DevicePath \"\"" Oct 05 08:30:39 crc kubenswrapper[4846]: I1005 08:30:39.789763 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/317b95c9-9b6d-4ba8-a96c-f69bcc8bb888-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "317b95c9-9b6d-4ba8-a96c-f69bcc8bb888" (UID: "317b95c9-9b6d-4ba8-a96c-f69bcc8bb888"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:30:39 crc kubenswrapper[4846]: I1005 08:30:39.790522 4846 generic.go:334] "Generic (PLEG): container finished" podID="317b95c9-9b6d-4ba8-a96c-f69bcc8bb888" containerID="da73a53a1c8eae6d9d176006d4fdf8e62b33d6c8c9090f1c593a8bd0ea79f501" exitCode=0 Oct 05 08:30:39 crc kubenswrapper[4846]: I1005 08:30:39.790593 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 05 08:30:39 crc kubenswrapper[4846]: I1005 08:30:39.790630 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"317b95c9-9b6d-4ba8-a96c-f69bcc8bb888","Type":"ContainerDied","Data":"da73a53a1c8eae6d9d176006d4fdf8e62b33d6c8c9090f1c593a8bd0ea79f501"} Oct 05 08:30:39 crc kubenswrapper[4846]: I1005 08:30:39.790689 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"317b95c9-9b6d-4ba8-a96c-f69bcc8bb888","Type":"ContainerDied","Data":"48c49b6801bb84bfafc7f2724a02df14c68cef9977d32f2c97732cc1a07505ce"} Oct 05 08:30:39 crc kubenswrapper[4846]: I1005 08:30:39.790718 4846 scope.go:117] "RemoveContainer" containerID="7af5c616692c82ac6dd7f74d046c5062dd9b70877997a1f20373ab3124c53a42" Oct 05 08:30:39 crc kubenswrapper[4846]: I1005 08:30:39.792938 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-5tnzg" event={"ID":"19856b5e-44a2-4b5e-bee6-2cde2af29996","Type":"ContainerDied","Data":"ce61faaadbeef1c1169438e93fbe81bb0f29b20ff25652e5fe80d75cf1c677f3"} Oct 05 08:30:39 crc kubenswrapper[4846]: I1005 08:30:39.792971 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ce61faaadbeef1c1169438e93fbe81bb0f29b20ff25652e5fe80d75cf1c677f3" Oct 05 08:30:39 crc kubenswrapper[4846]: I1005 08:30:39.792945 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-5tnzg" Oct 05 08:30:39 crc kubenswrapper[4846]: I1005 08:30:39.794586 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/317b95c9-9b6d-4ba8-a96c-f69bcc8bb888-config-data" (OuterVolumeSpecName: "config-data") pod "317b95c9-9b6d-4ba8-a96c-f69bcc8bb888" (UID: "317b95c9-9b6d-4ba8-a96c-f69bcc8bb888"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:30:39 crc kubenswrapper[4846]: I1005 08:30:39.830046 4846 scope.go:117] "RemoveContainer" containerID="4c0b66909bef91bfd67e7042c5bc22a2b196bd7f4f03736a48f63caa951e0b5d" Oct 05 08:30:39 crc kubenswrapper[4846]: I1005 08:30:39.877592 4846 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/317b95c9-9b6d-4ba8-a96c-f69bcc8bb888-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 08:30:39 crc kubenswrapper[4846]: I1005 08:30:39.877633 4846 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/317b95c9-9b6d-4ba8-a96c-f69bcc8bb888-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 08:30:39 crc kubenswrapper[4846]: I1005 08:30:39.927460 4846 scope.go:117] "RemoveContainer" containerID="da73a53a1c8eae6d9d176006d4fdf8e62b33d6c8c9090f1c593a8bd0ea79f501" Oct 05 08:30:39 crc kubenswrapper[4846]: I1005 08:30:39.952518 4846 scope.go:117] "RemoveContainer" containerID="7a6e0f2c3730c396e8eafaf048394a6415c61b772ba967597d899bca0f544e7b" Oct 05 08:30:39 crc kubenswrapper[4846]: I1005 08:30:39.973911 4846 scope.go:117] "RemoveContainer" containerID="7af5c616692c82ac6dd7f74d046c5062dd9b70877997a1f20373ab3124c53a42" Oct 05 08:30:39 crc kubenswrapper[4846]: E1005 08:30:39.974577 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7af5c616692c82ac6dd7f74d046c5062dd9b70877997a1f20373ab3124c53a42\": container with ID starting with 7af5c616692c82ac6dd7f74d046c5062dd9b70877997a1f20373ab3124c53a42 not found: ID does not exist" containerID="7af5c616692c82ac6dd7f74d046c5062dd9b70877997a1f20373ab3124c53a42" Oct 05 08:30:39 crc kubenswrapper[4846]: I1005 08:30:39.974610 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7af5c616692c82ac6dd7f74d046c5062dd9b70877997a1f20373ab3124c53a42"} err="failed to get container status \"7af5c616692c82ac6dd7f74d046c5062dd9b70877997a1f20373ab3124c53a42\": rpc error: code = NotFound desc = could not find container \"7af5c616692c82ac6dd7f74d046c5062dd9b70877997a1f20373ab3124c53a42\": container with ID starting with 7af5c616692c82ac6dd7f74d046c5062dd9b70877997a1f20373ab3124c53a42 not found: ID does not exist" Oct 05 08:30:39 crc kubenswrapper[4846]: I1005 08:30:39.974639 4846 scope.go:117] "RemoveContainer" containerID="4c0b66909bef91bfd67e7042c5bc22a2b196bd7f4f03736a48f63caa951e0b5d" Oct 05 08:30:39 crc kubenswrapper[4846]: E1005 08:30:39.974930 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4c0b66909bef91bfd67e7042c5bc22a2b196bd7f4f03736a48f63caa951e0b5d\": container with ID starting with 4c0b66909bef91bfd67e7042c5bc22a2b196bd7f4f03736a48f63caa951e0b5d not found: ID does not exist" containerID="4c0b66909bef91bfd67e7042c5bc22a2b196bd7f4f03736a48f63caa951e0b5d" Oct 05 08:30:39 crc kubenswrapper[4846]: I1005 08:30:39.974959 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4c0b66909bef91bfd67e7042c5bc22a2b196bd7f4f03736a48f63caa951e0b5d"} err="failed to get container status \"4c0b66909bef91bfd67e7042c5bc22a2b196bd7f4f03736a48f63caa951e0b5d\": rpc error: code = NotFound desc = could not find container \"4c0b66909bef91bfd67e7042c5bc22a2b196bd7f4f03736a48f63caa951e0b5d\": container with ID starting with 4c0b66909bef91bfd67e7042c5bc22a2b196bd7f4f03736a48f63caa951e0b5d not found: ID does not exist" Oct 05 08:30:39 crc kubenswrapper[4846]: I1005 08:30:39.974978 4846 scope.go:117] "RemoveContainer" containerID="da73a53a1c8eae6d9d176006d4fdf8e62b33d6c8c9090f1c593a8bd0ea79f501" Oct 05 08:30:39 crc kubenswrapper[4846]: E1005 08:30:39.975216 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"da73a53a1c8eae6d9d176006d4fdf8e62b33d6c8c9090f1c593a8bd0ea79f501\": container with ID starting with da73a53a1c8eae6d9d176006d4fdf8e62b33d6c8c9090f1c593a8bd0ea79f501 not found: ID does not exist" containerID="da73a53a1c8eae6d9d176006d4fdf8e62b33d6c8c9090f1c593a8bd0ea79f501" Oct 05 08:30:39 crc kubenswrapper[4846]: I1005 08:30:39.975248 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"da73a53a1c8eae6d9d176006d4fdf8e62b33d6c8c9090f1c593a8bd0ea79f501"} err="failed to get container status \"da73a53a1c8eae6d9d176006d4fdf8e62b33d6c8c9090f1c593a8bd0ea79f501\": rpc error: code = NotFound desc = could not find container \"da73a53a1c8eae6d9d176006d4fdf8e62b33d6c8c9090f1c593a8bd0ea79f501\": container with ID starting with da73a53a1c8eae6d9d176006d4fdf8e62b33d6c8c9090f1c593a8bd0ea79f501 not found: ID does not exist" Oct 05 08:30:39 crc kubenswrapper[4846]: I1005 08:30:39.975267 4846 scope.go:117] "RemoveContainer" containerID="7a6e0f2c3730c396e8eafaf048394a6415c61b772ba967597d899bca0f544e7b" Oct 05 08:30:39 crc kubenswrapper[4846]: E1005 08:30:39.975655 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7a6e0f2c3730c396e8eafaf048394a6415c61b772ba967597d899bca0f544e7b\": container with ID starting with 7a6e0f2c3730c396e8eafaf048394a6415c61b772ba967597d899bca0f544e7b not found: ID does not exist" containerID="7a6e0f2c3730c396e8eafaf048394a6415c61b772ba967597d899bca0f544e7b" Oct 05 08:30:39 crc kubenswrapper[4846]: I1005 08:30:39.975682 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7a6e0f2c3730c396e8eafaf048394a6415c61b772ba967597d899bca0f544e7b"} err="failed to get container status \"7a6e0f2c3730c396e8eafaf048394a6415c61b772ba967597d899bca0f544e7b\": rpc error: code = NotFound desc = could not find container \"7a6e0f2c3730c396e8eafaf048394a6415c61b772ba967597d899bca0f544e7b\": container with ID starting with 7a6e0f2c3730c396e8eafaf048394a6415c61b772ba967597d899bca0f544e7b not found: ID does not exist" Oct 05 08:30:40 crc kubenswrapper[4846]: I1005 08:30:40.040112 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-create-8ddb7"] Oct 05 08:30:40 crc kubenswrapper[4846]: I1005 08:30:40.049849 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-create-8ddb7"] Oct 05 08:30:40 crc kubenswrapper[4846]: I1005 08:30:40.129155 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 05 08:30:40 crc kubenswrapper[4846]: I1005 08:30:40.140681 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 05 08:30:40 crc kubenswrapper[4846]: I1005 08:30:40.167067 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 05 08:30:40 crc kubenswrapper[4846]: E1005 08:30:40.167597 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="317b95c9-9b6d-4ba8-a96c-f69bcc8bb888" containerName="ceilometer-central-agent" Oct 05 08:30:40 crc kubenswrapper[4846]: I1005 08:30:40.167637 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="317b95c9-9b6d-4ba8-a96c-f69bcc8bb888" containerName="ceilometer-central-agent" Oct 05 08:30:40 crc kubenswrapper[4846]: E1005 08:30:40.167655 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="317b95c9-9b6d-4ba8-a96c-f69bcc8bb888" containerName="sg-core" Oct 05 08:30:40 crc kubenswrapper[4846]: I1005 08:30:40.167662 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="317b95c9-9b6d-4ba8-a96c-f69bcc8bb888" containerName="sg-core" Oct 05 08:30:40 crc kubenswrapper[4846]: E1005 08:30:40.167687 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="317b95c9-9b6d-4ba8-a96c-f69bcc8bb888" containerName="ceilometer-notification-agent" Oct 05 08:30:40 crc kubenswrapper[4846]: I1005 08:30:40.167693 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="317b95c9-9b6d-4ba8-a96c-f69bcc8bb888" containerName="ceilometer-notification-agent" Oct 05 08:30:40 crc kubenswrapper[4846]: E1005 08:30:40.167706 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="317b95c9-9b6d-4ba8-a96c-f69bcc8bb888" containerName="proxy-httpd" Oct 05 08:30:40 crc kubenswrapper[4846]: I1005 08:30:40.167713 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="317b95c9-9b6d-4ba8-a96c-f69bcc8bb888" containerName="proxy-httpd" Oct 05 08:30:40 crc kubenswrapper[4846]: E1005 08:30:40.167732 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="19856b5e-44a2-4b5e-bee6-2cde2af29996" containerName="aodh-db-sync" Oct 05 08:30:40 crc kubenswrapper[4846]: I1005 08:30:40.167741 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="19856b5e-44a2-4b5e-bee6-2cde2af29996" containerName="aodh-db-sync" Oct 05 08:30:40 crc kubenswrapper[4846]: I1005 08:30:40.167975 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="317b95c9-9b6d-4ba8-a96c-f69bcc8bb888" containerName="sg-core" Oct 05 08:30:40 crc kubenswrapper[4846]: I1005 08:30:40.168003 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="317b95c9-9b6d-4ba8-a96c-f69bcc8bb888" containerName="ceilometer-central-agent" Oct 05 08:30:40 crc kubenswrapper[4846]: I1005 08:30:40.168015 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="317b95c9-9b6d-4ba8-a96c-f69bcc8bb888" containerName="ceilometer-notification-agent" Oct 05 08:30:40 crc kubenswrapper[4846]: I1005 08:30:40.168037 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="317b95c9-9b6d-4ba8-a96c-f69bcc8bb888" containerName="proxy-httpd" Oct 05 08:30:40 crc kubenswrapper[4846]: I1005 08:30:40.168046 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="19856b5e-44a2-4b5e-bee6-2cde2af29996" containerName="aodh-db-sync" Oct 05 08:30:40 crc kubenswrapper[4846]: I1005 08:30:40.170442 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 05 08:30:40 crc kubenswrapper[4846]: I1005 08:30:40.172402 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Oct 05 08:30:40 crc kubenswrapper[4846]: I1005 08:30:40.172845 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 05 08:30:40 crc kubenswrapper[4846]: I1005 08:30:40.173041 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 05 08:30:40 crc kubenswrapper[4846]: I1005 08:30:40.179629 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 05 08:30:40 crc kubenswrapper[4846]: I1005 08:30:40.285135 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c7a796f-58ed-436b-aef9-5a3af4161c99-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"8c7a796f-58ed-436b-aef9-5a3af4161c99\") " pod="openstack/ceilometer-0" Oct 05 08:30:40 crc kubenswrapper[4846]: I1005 08:30:40.285472 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8c7a796f-58ed-436b-aef9-5a3af4161c99-log-httpd\") pod \"ceilometer-0\" (UID: \"8c7a796f-58ed-436b-aef9-5a3af4161c99\") " pod="openstack/ceilometer-0" Oct 05 08:30:40 crc kubenswrapper[4846]: I1005 08:30:40.285511 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/8c7a796f-58ed-436b-aef9-5a3af4161c99-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"8c7a796f-58ed-436b-aef9-5a3af4161c99\") " pod="openstack/ceilometer-0" Oct 05 08:30:40 crc kubenswrapper[4846]: I1005 08:30:40.285558 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hrlv5\" (UniqueName: \"kubernetes.io/projected/8c7a796f-58ed-436b-aef9-5a3af4161c99-kube-api-access-hrlv5\") pod \"ceilometer-0\" (UID: \"8c7a796f-58ed-436b-aef9-5a3af4161c99\") " pod="openstack/ceilometer-0" Oct 05 08:30:40 crc kubenswrapper[4846]: I1005 08:30:40.285602 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8c7a796f-58ed-436b-aef9-5a3af4161c99-run-httpd\") pod \"ceilometer-0\" (UID: \"8c7a796f-58ed-436b-aef9-5a3af4161c99\") " pod="openstack/ceilometer-0" Oct 05 08:30:40 crc kubenswrapper[4846]: I1005 08:30:40.285630 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/8c7a796f-58ed-436b-aef9-5a3af4161c99-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"8c7a796f-58ed-436b-aef9-5a3af4161c99\") " pod="openstack/ceilometer-0" Oct 05 08:30:40 crc kubenswrapper[4846]: I1005 08:30:40.285687 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8c7a796f-58ed-436b-aef9-5a3af4161c99-scripts\") pod \"ceilometer-0\" (UID: \"8c7a796f-58ed-436b-aef9-5a3af4161c99\") " pod="openstack/ceilometer-0" Oct 05 08:30:40 crc kubenswrapper[4846]: I1005 08:30:40.285709 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8c7a796f-58ed-436b-aef9-5a3af4161c99-config-data\") pod \"ceilometer-0\" (UID: \"8c7a796f-58ed-436b-aef9-5a3af4161c99\") " pod="openstack/ceilometer-0" Oct 05 08:30:40 crc kubenswrapper[4846]: I1005 08:30:40.387852 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8c7a796f-58ed-436b-aef9-5a3af4161c99-scripts\") pod \"ceilometer-0\" (UID: \"8c7a796f-58ed-436b-aef9-5a3af4161c99\") " pod="openstack/ceilometer-0" Oct 05 08:30:40 crc kubenswrapper[4846]: I1005 08:30:40.387946 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8c7a796f-58ed-436b-aef9-5a3af4161c99-config-data\") pod \"ceilometer-0\" (UID: \"8c7a796f-58ed-436b-aef9-5a3af4161c99\") " pod="openstack/ceilometer-0" Oct 05 08:30:40 crc kubenswrapper[4846]: I1005 08:30:40.388036 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c7a796f-58ed-436b-aef9-5a3af4161c99-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"8c7a796f-58ed-436b-aef9-5a3af4161c99\") " pod="openstack/ceilometer-0" Oct 05 08:30:40 crc kubenswrapper[4846]: I1005 08:30:40.388121 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8c7a796f-58ed-436b-aef9-5a3af4161c99-log-httpd\") pod \"ceilometer-0\" (UID: \"8c7a796f-58ed-436b-aef9-5a3af4161c99\") " pod="openstack/ceilometer-0" Oct 05 08:30:40 crc kubenswrapper[4846]: I1005 08:30:40.389084 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8c7a796f-58ed-436b-aef9-5a3af4161c99-log-httpd\") pod \"ceilometer-0\" (UID: \"8c7a796f-58ed-436b-aef9-5a3af4161c99\") " pod="openstack/ceilometer-0" Oct 05 08:30:40 crc kubenswrapper[4846]: I1005 08:30:40.389419 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/8c7a796f-58ed-436b-aef9-5a3af4161c99-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"8c7a796f-58ed-436b-aef9-5a3af4161c99\") " pod="openstack/ceilometer-0" Oct 05 08:30:40 crc kubenswrapper[4846]: I1005 08:30:40.389618 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hrlv5\" (UniqueName: \"kubernetes.io/projected/8c7a796f-58ed-436b-aef9-5a3af4161c99-kube-api-access-hrlv5\") pod \"ceilometer-0\" (UID: \"8c7a796f-58ed-436b-aef9-5a3af4161c99\") " pod="openstack/ceilometer-0" Oct 05 08:30:40 crc kubenswrapper[4846]: I1005 08:30:40.390324 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8c7a796f-58ed-436b-aef9-5a3af4161c99-run-httpd\") pod \"ceilometer-0\" (UID: \"8c7a796f-58ed-436b-aef9-5a3af4161c99\") " pod="openstack/ceilometer-0" Oct 05 08:30:40 crc kubenswrapper[4846]: I1005 08:30:40.390886 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8c7a796f-58ed-436b-aef9-5a3af4161c99-run-httpd\") pod \"ceilometer-0\" (UID: \"8c7a796f-58ed-436b-aef9-5a3af4161c99\") " pod="openstack/ceilometer-0" Oct 05 08:30:40 crc kubenswrapper[4846]: I1005 08:30:40.391644 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/8c7a796f-58ed-436b-aef9-5a3af4161c99-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"8c7a796f-58ed-436b-aef9-5a3af4161c99\") " pod="openstack/ceilometer-0" Oct 05 08:30:40 crc kubenswrapper[4846]: I1005 08:30:40.392478 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8c7a796f-58ed-436b-aef9-5a3af4161c99-config-data\") pod \"ceilometer-0\" (UID: \"8c7a796f-58ed-436b-aef9-5a3af4161c99\") " pod="openstack/ceilometer-0" Oct 05 08:30:40 crc kubenswrapper[4846]: I1005 08:30:40.393576 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8c7a796f-58ed-436b-aef9-5a3af4161c99-scripts\") pod \"ceilometer-0\" (UID: \"8c7a796f-58ed-436b-aef9-5a3af4161c99\") " pod="openstack/ceilometer-0" Oct 05 08:30:40 crc kubenswrapper[4846]: I1005 08:30:40.397321 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/8c7a796f-58ed-436b-aef9-5a3af4161c99-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"8c7a796f-58ed-436b-aef9-5a3af4161c99\") " pod="openstack/ceilometer-0" Oct 05 08:30:40 crc kubenswrapper[4846]: I1005 08:30:40.410643 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/8c7a796f-58ed-436b-aef9-5a3af4161c99-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"8c7a796f-58ed-436b-aef9-5a3af4161c99\") " pod="openstack/ceilometer-0" Oct 05 08:30:40 crc kubenswrapper[4846]: I1005 08:30:40.412687 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hrlv5\" (UniqueName: \"kubernetes.io/projected/8c7a796f-58ed-436b-aef9-5a3af4161c99-kube-api-access-hrlv5\") pod \"ceilometer-0\" (UID: \"8c7a796f-58ed-436b-aef9-5a3af4161c99\") " pod="openstack/ceilometer-0" Oct 05 08:30:40 crc kubenswrapper[4846]: I1005 08:30:40.420263 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c7a796f-58ed-436b-aef9-5a3af4161c99-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"8c7a796f-58ed-436b-aef9-5a3af4161c99\") " pod="openstack/ceilometer-0" Oct 05 08:30:40 crc kubenswrapper[4846]: I1005 08:30:40.512458 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="317b95c9-9b6d-4ba8-a96c-f69bcc8bb888" path="/var/lib/kubelet/pods/317b95c9-9b6d-4ba8-a96c-f69bcc8bb888/volumes" Oct 05 08:30:40 crc kubenswrapper[4846]: I1005 08:30:40.513702 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e4b32f67-df27-4951-8798-2a026ca73183" path="/var/lib/kubelet/pods/e4b32f67-df27-4951-8798-2a026ca73183/volumes" Oct 05 08:30:40 crc kubenswrapper[4846]: I1005 08:30:40.513728 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 05 08:30:40 crc kubenswrapper[4846]: I1005 08:30:40.998547 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 05 08:30:41 crc kubenswrapper[4846]: W1005 08:30:41.009584 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8c7a796f_58ed_436b_aef9_5a3af4161c99.slice/crio-ad494b73e87da769a8a1b2152c0de7248861780dba43caf35f3f2c6b06988e47 WatchSource:0}: Error finding container ad494b73e87da769a8a1b2152c0de7248861780dba43caf35f3f2c6b06988e47: Status 404 returned error can't find the container with id ad494b73e87da769a8a1b2152c0de7248861780dba43caf35f3f2c6b06988e47 Oct 05 08:30:41 crc kubenswrapper[4846]: I1005 08:30:41.823386 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8c7a796f-58ed-436b-aef9-5a3af4161c99","Type":"ContainerStarted","Data":"2c803c74afd6a83d53fac7c8eda7624c53a6da308d32c58796661464b251b0c1"} Oct 05 08:30:41 crc kubenswrapper[4846]: I1005 08:30:41.824226 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8c7a796f-58ed-436b-aef9-5a3af4161c99","Type":"ContainerStarted","Data":"ad494b73e87da769a8a1b2152c0de7248861780dba43caf35f3f2c6b06988e47"} Oct 05 08:30:42 crc kubenswrapper[4846]: I1005 08:30:42.841057 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8c7a796f-58ed-436b-aef9-5a3af4161c99","Type":"ContainerStarted","Data":"88776f196cb8122e8a327c2ad8e8e609afdf0184de66732201683b73784627c9"} Oct 05 08:30:42 crc kubenswrapper[4846]: I1005 08:30:42.841457 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8c7a796f-58ed-436b-aef9-5a3af4161c99","Type":"ContainerStarted","Data":"8e74fbdd446dcfc85adc4504e1df3064933b2a956c9c05ac76d77b7781fbdf43"} Oct 05 08:30:43 crc kubenswrapper[4846]: I1005 08:30:43.874783 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-0"] Oct 05 08:30:43 crc kubenswrapper[4846]: I1005 08:30:43.879852 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Oct 05 08:30:43 crc kubenswrapper[4846]: I1005 08:30:43.883615 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-autoscaling-dockercfg-666qr" Oct 05 08:30:43 crc kubenswrapper[4846]: I1005 08:30:43.883928 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-config-data" Oct 05 08:30:43 crc kubenswrapper[4846]: I1005 08:30:43.883948 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-scripts" Oct 05 08:30:43 crc kubenswrapper[4846]: I1005 08:30:43.890499 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-0"] Oct 05 08:30:43 crc kubenswrapper[4846]: I1005 08:30:43.984736 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/05c8753d-a522-467c-9a13-93961e9061c1-scripts\") pod \"aodh-0\" (UID: \"05c8753d-a522-467c-9a13-93961e9061c1\") " pod="openstack/aodh-0" Oct 05 08:30:43 crc kubenswrapper[4846]: I1005 08:30:43.984988 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/05c8753d-a522-467c-9a13-93961e9061c1-config-data\") pod \"aodh-0\" (UID: \"05c8753d-a522-467c-9a13-93961e9061c1\") " pod="openstack/aodh-0" Oct 05 08:30:43 crc kubenswrapper[4846]: I1005 08:30:43.985169 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cszgp\" (UniqueName: \"kubernetes.io/projected/05c8753d-a522-467c-9a13-93961e9061c1-kube-api-access-cszgp\") pod \"aodh-0\" (UID: \"05c8753d-a522-467c-9a13-93961e9061c1\") " pod="openstack/aodh-0" Oct 05 08:30:43 crc kubenswrapper[4846]: I1005 08:30:43.985413 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/05c8753d-a522-467c-9a13-93961e9061c1-combined-ca-bundle\") pod \"aodh-0\" (UID: \"05c8753d-a522-467c-9a13-93961e9061c1\") " pod="openstack/aodh-0" Oct 05 08:30:44 crc kubenswrapper[4846]: I1005 08:30:44.087496 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cszgp\" (UniqueName: \"kubernetes.io/projected/05c8753d-a522-467c-9a13-93961e9061c1-kube-api-access-cszgp\") pod \"aodh-0\" (UID: \"05c8753d-a522-467c-9a13-93961e9061c1\") " pod="openstack/aodh-0" Oct 05 08:30:44 crc kubenswrapper[4846]: I1005 08:30:44.087651 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/05c8753d-a522-467c-9a13-93961e9061c1-combined-ca-bundle\") pod \"aodh-0\" (UID: \"05c8753d-a522-467c-9a13-93961e9061c1\") " pod="openstack/aodh-0" Oct 05 08:30:44 crc kubenswrapper[4846]: I1005 08:30:44.087725 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/05c8753d-a522-467c-9a13-93961e9061c1-scripts\") pod \"aodh-0\" (UID: \"05c8753d-a522-467c-9a13-93961e9061c1\") " pod="openstack/aodh-0" Oct 05 08:30:44 crc kubenswrapper[4846]: I1005 08:30:44.087782 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/05c8753d-a522-467c-9a13-93961e9061c1-config-data\") pod \"aodh-0\" (UID: \"05c8753d-a522-467c-9a13-93961e9061c1\") " pod="openstack/aodh-0" Oct 05 08:30:44 crc kubenswrapper[4846]: I1005 08:30:44.091620 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/05c8753d-a522-467c-9a13-93961e9061c1-scripts\") pod \"aodh-0\" (UID: \"05c8753d-a522-467c-9a13-93961e9061c1\") " pod="openstack/aodh-0" Oct 05 08:30:44 crc kubenswrapper[4846]: I1005 08:30:44.098813 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/05c8753d-a522-467c-9a13-93961e9061c1-combined-ca-bundle\") pod \"aodh-0\" (UID: \"05c8753d-a522-467c-9a13-93961e9061c1\") " pod="openstack/aodh-0" Oct 05 08:30:44 crc kubenswrapper[4846]: I1005 08:30:44.102164 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/05c8753d-a522-467c-9a13-93961e9061c1-config-data\") pod \"aodh-0\" (UID: \"05c8753d-a522-467c-9a13-93961e9061c1\") " pod="openstack/aodh-0" Oct 05 08:30:44 crc kubenswrapper[4846]: I1005 08:30:44.111661 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cszgp\" (UniqueName: \"kubernetes.io/projected/05c8753d-a522-467c-9a13-93961e9061c1-kube-api-access-cszgp\") pod \"aodh-0\" (UID: \"05c8753d-a522-467c-9a13-93961e9061c1\") " pod="openstack/aodh-0" Oct 05 08:30:44 crc kubenswrapper[4846]: I1005 08:30:44.198770 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Oct 05 08:30:44 crc kubenswrapper[4846]: I1005 08:30:44.668948 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-0"] Oct 05 08:30:44 crc kubenswrapper[4846]: W1005 08:30:44.670681 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod05c8753d_a522_467c_9a13_93961e9061c1.slice/crio-a2898f58c9ffbfb291ec354ffa6a14431dfda6302c79da544e90036914b95c4d WatchSource:0}: Error finding container a2898f58c9ffbfb291ec354ffa6a14431dfda6302c79da544e90036914b95c4d: Status 404 returned error can't find the container with id a2898f58c9ffbfb291ec354ffa6a14431dfda6302c79da544e90036914b95c4d Oct 05 08:30:44 crc kubenswrapper[4846]: I1005 08:30:44.879089 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8c7a796f-58ed-436b-aef9-5a3af4161c99","Type":"ContainerStarted","Data":"1257dcf06640a3d51ed6a0d23dcee6a8666095312a504ce29cb7966a45415217"} Oct 05 08:30:44 crc kubenswrapper[4846]: I1005 08:30:44.879496 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 05 08:30:44 crc kubenswrapper[4846]: I1005 08:30:44.889792 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"05c8753d-a522-467c-9a13-93961e9061c1","Type":"ContainerStarted","Data":"a2898f58c9ffbfb291ec354ffa6a14431dfda6302c79da544e90036914b95c4d"} Oct 05 08:30:44 crc kubenswrapper[4846]: I1005 08:30:44.914265 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.236035844 podStartE2EDuration="4.914241413s" podCreationTimestamp="2025-10-05 08:30:40 +0000 UTC" firstStartedPulling="2025-10-05 08:30:41.013225072 +0000 UTC m=+6163.254077837" lastFinishedPulling="2025-10-05 08:30:43.691430621 +0000 UTC m=+6165.932283406" observedRunningTime="2025-10-05 08:30:44.906246431 +0000 UTC m=+6167.147099206" watchObservedRunningTime="2025-10-05 08:30:44.914241413 +0000 UTC m=+6167.155094188" Oct 05 08:30:45 crc kubenswrapper[4846]: I1005 08:30:45.765494 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 05 08:30:45 crc kubenswrapper[4846]: I1005 08:30:45.907019 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"05c8753d-a522-467c-9a13-93961e9061c1","Type":"ContainerStarted","Data":"0fc4a5a1699764227e218cb40055ffedcc4d7218376607ac97664dfd27ff5b30"} Oct 05 08:30:46 crc kubenswrapper[4846]: I1005 08:30:46.161496 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Oct 05 08:30:46 crc kubenswrapper[4846]: I1005 08:30:46.918146 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"05c8753d-a522-467c-9a13-93961e9061c1","Type":"ContainerStarted","Data":"f07446ef7642b96e7c794af68bca50a7f991d2fb618f2c6a493a0d9e86eb1a43"} Oct 05 08:30:46 crc kubenswrapper[4846]: I1005 08:30:46.918373 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="8c7a796f-58ed-436b-aef9-5a3af4161c99" containerName="ceilometer-central-agent" containerID="cri-o://2c803c74afd6a83d53fac7c8eda7624c53a6da308d32c58796661464b251b0c1" gracePeriod=30 Oct 05 08:30:46 crc kubenswrapper[4846]: I1005 08:30:46.918411 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="8c7a796f-58ed-436b-aef9-5a3af4161c99" containerName="proxy-httpd" containerID="cri-o://1257dcf06640a3d51ed6a0d23dcee6a8666095312a504ce29cb7966a45415217" gracePeriod=30 Oct 05 08:30:46 crc kubenswrapper[4846]: I1005 08:30:46.918411 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="8c7a796f-58ed-436b-aef9-5a3af4161c99" containerName="sg-core" containerID="cri-o://88776f196cb8122e8a327c2ad8e8e609afdf0184de66732201683b73784627c9" gracePeriod=30 Oct 05 08:30:46 crc kubenswrapper[4846]: I1005 08:30:46.918438 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="8c7a796f-58ed-436b-aef9-5a3af4161c99" containerName="ceilometer-notification-agent" containerID="cri-o://8e74fbdd446dcfc85adc4504e1df3064933b2a956c9c05ac76d77b7781fbdf43" gracePeriod=30 Oct 05 08:30:47 crc kubenswrapper[4846]: I1005 08:30:47.381193 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-0"] Oct 05 08:30:47 crc kubenswrapper[4846]: E1005 08:30:47.675573 4846 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9eeca81b_6fda_4c22_a221_3adcb9900077.slice/crio-e8538609494690d52f4d579e3ae6523d65701c417f2551cb9f7b3cad911741a5\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9eeca81b_6fda_4c22_a221_3adcb9900077.slice\": RecentStats: unable to find data in memory cache]" Oct 05 08:30:47 crc kubenswrapper[4846]: I1005 08:30:47.936472 4846 generic.go:334] "Generic (PLEG): container finished" podID="8c7a796f-58ed-436b-aef9-5a3af4161c99" containerID="1257dcf06640a3d51ed6a0d23dcee6a8666095312a504ce29cb7966a45415217" exitCode=0 Oct 05 08:30:47 crc kubenswrapper[4846]: I1005 08:30:47.936507 4846 generic.go:334] "Generic (PLEG): container finished" podID="8c7a796f-58ed-436b-aef9-5a3af4161c99" containerID="88776f196cb8122e8a327c2ad8e8e609afdf0184de66732201683b73784627c9" exitCode=2 Oct 05 08:30:47 crc kubenswrapper[4846]: I1005 08:30:47.936514 4846 generic.go:334] "Generic (PLEG): container finished" podID="8c7a796f-58ed-436b-aef9-5a3af4161c99" containerID="8e74fbdd446dcfc85adc4504e1df3064933b2a956c9c05ac76d77b7781fbdf43" exitCode=0 Oct 05 08:30:47 crc kubenswrapper[4846]: I1005 08:30:47.936565 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8c7a796f-58ed-436b-aef9-5a3af4161c99","Type":"ContainerDied","Data":"1257dcf06640a3d51ed6a0d23dcee6a8666095312a504ce29cb7966a45415217"} Oct 05 08:30:47 crc kubenswrapper[4846]: I1005 08:30:47.936639 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8c7a796f-58ed-436b-aef9-5a3af4161c99","Type":"ContainerDied","Data":"88776f196cb8122e8a327c2ad8e8e609afdf0184de66732201683b73784627c9"} Oct 05 08:30:47 crc kubenswrapper[4846]: I1005 08:30:47.936651 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8c7a796f-58ed-436b-aef9-5a3af4161c99","Type":"ContainerDied","Data":"8e74fbdd446dcfc85adc4504e1df3064933b2a956c9c05ac76d77b7781fbdf43"} Oct 05 08:30:48 crc kubenswrapper[4846]: I1005 08:30:48.973482 4846 generic.go:334] "Generic (PLEG): container finished" podID="8c7a796f-58ed-436b-aef9-5a3af4161c99" containerID="2c803c74afd6a83d53fac7c8eda7624c53a6da308d32c58796661464b251b0c1" exitCode=0 Oct 05 08:30:48 crc kubenswrapper[4846]: I1005 08:30:48.973589 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8c7a796f-58ed-436b-aef9-5a3af4161c99","Type":"ContainerDied","Data":"2c803c74afd6a83d53fac7c8eda7624c53a6da308d32c58796661464b251b0c1"} Oct 05 08:30:48 crc kubenswrapper[4846]: I1005 08:30:48.973929 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8c7a796f-58ed-436b-aef9-5a3af4161c99","Type":"ContainerDied","Data":"ad494b73e87da769a8a1b2152c0de7248861780dba43caf35f3f2c6b06988e47"} Oct 05 08:30:48 crc kubenswrapper[4846]: I1005 08:30:48.973942 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ad494b73e87da769a8a1b2152c0de7248861780dba43caf35f3f2c6b06988e47" Oct 05 08:30:48 crc kubenswrapper[4846]: I1005 08:30:48.976747 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"05c8753d-a522-467c-9a13-93961e9061c1","Type":"ContainerStarted","Data":"ddf97cb37c806fcba6e84c283ff3c7442d458018d05d266877c7bc25b8ad7687"} Oct 05 08:30:49 crc kubenswrapper[4846]: I1005 08:30:49.021677 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 05 08:30:49 crc kubenswrapper[4846]: I1005 08:30:49.191903 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hrlv5\" (UniqueName: \"kubernetes.io/projected/8c7a796f-58ed-436b-aef9-5a3af4161c99-kube-api-access-hrlv5\") pod \"8c7a796f-58ed-436b-aef9-5a3af4161c99\" (UID: \"8c7a796f-58ed-436b-aef9-5a3af4161c99\") " Oct 05 08:30:49 crc kubenswrapper[4846]: I1005 08:30:49.192015 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8c7a796f-58ed-436b-aef9-5a3af4161c99-run-httpd\") pod \"8c7a796f-58ed-436b-aef9-5a3af4161c99\" (UID: \"8c7a796f-58ed-436b-aef9-5a3af4161c99\") " Oct 05 08:30:49 crc kubenswrapper[4846]: I1005 08:30:49.192076 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c7a796f-58ed-436b-aef9-5a3af4161c99-combined-ca-bundle\") pod \"8c7a796f-58ed-436b-aef9-5a3af4161c99\" (UID: \"8c7a796f-58ed-436b-aef9-5a3af4161c99\") " Oct 05 08:30:49 crc kubenswrapper[4846]: I1005 08:30:49.192132 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8c7a796f-58ed-436b-aef9-5a3af4161c99-log-httpd\") pod \"8c7a796f-58ed-436b-aef9-5a3af4161c99\" (UID: \"8c7a796f-58ed-436b-aef9-5a3af4161c99\") " Oct 05 08:30:49 crc kubenswrapper[4846]: I1005 08:30:49.192214 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/8c7a796f-58ed-436b-aef9-5a3af4161c99-sg-core-conf-yaml\") pod \"8c7a796f-58ed-436b-aef9-5a3af4161c99\" (UID: \"8c7a796f-58ed-436b-aef9-5a3af4161c99\") " Oct 05 08:30:49 crc kubenswrapper[4846]: I1005 08:30:49.192247 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8c7a796f-58ed-436b-aef9-5a3af4161c99-config-data\") pod \"8c7a796f-58ed-436b-aef9-5a3af4161c99\" (UID: \"8c7a796f-58ed-436b-aef9-5a3af4161c99\") " Oct 05 08:30:49 crc kubenswrapper[4846]: I1005 08:30:49.192289 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/8c7a796f-58ed-436b-aef9-5a3af4161c99-ceilometer-tls-certs\") pod \"8c7a796f-58ed-436b-aef9-5a3af4161c99\" (UID: \"8c7a796f-58ed-436b-aef9-5a3af4161c99\") " Oct 05 08:30:49 crc kubenswrapper[4846]: I1005 08:30:49.192322 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8c7a796f-58ed-436b-aef9-5a3af4161c99-scripts\") pod \"8c7a796f-58ed-436b-aef9-5a3af4161c99\" (UID: \"8c7a796f-58ed-436b-aef9-5a3af4161c99\") " Oct 05 08:30:49 crc kubenswrapper[4846]: I1005 08:30:49.193701 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8c7a796f-58ed-436b-aef9-5a3af4161c99-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "8c7a796f-58ed-436b-aef9-5a3af4161c99" (UID: "8c7a796f-58ed-436b-aef9-5a3af4161c99"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 08:30:49 crc kubenswrapper[4846]: I1005 08:30:49.195364 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8c7a796f-58ed-436b-aef9-5a3af4161c99-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "8c7a796f-58ed-436b-aef9-5a3af4161c99" (UID: "8c7a796f-58ed-436b-aef9-5a3af4161c99"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 08:30:49 crc kubenswrapper[4846]: I1005 08:30:49.199544 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8c7a796f-58ed-436b-aef9-5a3af4161c99-kube-api-access-hrlv5" (OuterVolumeSpecName: "kube-api-access-hrlv5") pod "8c7a796f-58ed-436b-aef9-5a3af4161c99" (UID: "8c7a796f-58ed-436b-aef9-5a3af4161c99"). InnerVolumeSpecName "kube-api-access-hrlv5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:30:49 crc kubenswrapper[4846]: I1005 08:30:49.211860 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8c7a796f-58ed-436b-aef9-5a3af4161c99-scripts" (OuterVolumeSpecName: "scripts") pod "8c7a796f-58ed-436b-aef9-5a3af4161c99" (UID: "8c7a796f-58ed-436b-aef9-5a3af4161c99"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:30:49 crc kubenswrapper[4846]: I1005 08:30:49.224986 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8c7a796f-58ed-436b-aef9-5a3af4161c99-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "8c7a796f-58ed-436b-aef9-5a3af4161c99" (UID: "8c7a796f-58ed-436b-aef9-5a3af4161c99"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:30:49 crc kubenswrapper[4846]: I1005 08:30:49.290368 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8c7a796f-58ed-436b-aef9-5a3af4161c99-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "8c7a796f-58ed-436b-aef9-5a3af4161c99" (UID: "8c7a796f-58ed-436b-aef9-5a3af4161c99"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:30:49 crc kubenswrapper[4846]: I1005 08:30:49.298798 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8c7a796f-58ed-436b-aef9-5a3af4161c99-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8c7a796f-58ed-436b-aef9-5a3af4161c99" (UID: "8c7a796f-58ed-436b-aef9-5a3af4161c99"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:30:49 crc kubenswrapper[4846]: I1005 08:30:49.302395 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c7a796f-58ed-436b-aef9-5a3af4161c99-combined-ca-bundle\") pod \"8c7a796f-58ed-436b-aef9-5a3af4161c99\" (UID: \"8c7a796f-58ed-436b-aef9-5a3af4161c99\") " Oct 05 08:30:49 crc kubenswrapper[4846]: W1005 08:30:49.302703 4846 empty_dir.go:500] Warning: Unmount skipped because path does not exist: /var/lib/kubelet/pods/8c7a796f-58ed-436b-aef9-5a3af4161c99/volumes/kubernetes.io~secret/combined-ca-bundle Oct 05 08:30:49 crc kubenswrapper[4846]: I1005 08:30:49.303046 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8c7a796f-58ed-436b-aef9-5a3af4161c99-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8c7a796f-58ed-436b-aef9-5a3af4161c99" (UID: "8c7a796f-58ed-436b-aef9-5a3af4161c99"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:30:49 crc kubenswrapper[4846]: I1005 08:30:49.303658 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hrlv5\" (UniqueName: \"kubernetes.io/projected/8c7a796f-58ed-436b-aef9-5a3af4161c99-kube-api-access-hrlv5\") on node \"crc\" DevicePath \"\"" Oct 05 08:30:49 crc kubenswrapper[4846]: I1005 08:30:49.303778 4846 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8c7a796f-58ed-436b-aef9-5a3af4161c99-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 05 08:30:49 crc kubenswrapper[4846]: I1005 08:30:49.303835 4846 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c7a796f-58ed-436b-aef9-5a3af4161c99-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 08:30:49 crc kubenswrapper[4846]: I1005 08:30:49.303914 4846 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8c7a796f-58ed-436b-aef9-5a3af4161c99-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 05 08:30:49 crc kubenswrapper[4846]: I1005 08:30:49.303990 4846 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/8c7a796f-58ed-436b-aef9-5a3af4161c99-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 05 08:30:49 crc kubenswrapper[4846]: I1005 08:30:49.304052 4846 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/8c7a796f-58ed-436b-aef9-5a3af4161c99-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 05 08:30:49 crc kubenswrapper[4846]: I1005 08:30:49.304130 4846 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8c7a796f-58ed-436b-aef9-5a3af4161c99-scripts\") on node \"crc\" DevicePath \"\"" Oct 05 08:30:49 crc kubenswrapper[4846]: I1005 08:30:49.364304 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8c7a796f-58ed-436b-aef9-5a3af4161c99-config-data" (OuterVolumeSpecName: "config-data") pod "8c7a796f-58ed-436b-aef9-5a3af4161c99" (UID: "8c7a796f-58ed-436b-aef9-5a3af4161c99"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:30:49 crc kubenswrapper[4846]: I1005 08:30:49.406278 4846 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8c7a796f-58ed-436b-aef9-5a3af4161c99-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 08:30:49 crc kubenswrapper[4846]: I1005 08:30:49.994404 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"05c8753d-a522-467c-9a13-93961e9061c1","Type":"ContainerStarted","Data":"d7f1352d94a9be8d08e2468ed0cfea65a2bee3fb2ca2a4d8f522015dc3840aae"} Oct 05 08:30:49 crc kubenswrapper[4846]: I1005 08:30:49.994418 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 05 08:30:49 crc kubenswrapper[4846]: I1005 08:30:49.995275 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="05c8753d-a522-467c-9a13-93961e9061c1" containerName="aodh-listener" containerID="cri-o://d7f1352d94a9be8d08e2468ed0cfea65a2bee3fb2ca2a4d8f522015dc3840aae" gracePeriod=30 Oct 05 08:30:49 crc kubenswrapper[4846]: I1005 08:30:49.995402 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="05c8753d-a522-467c-9a13-93961e9061c1" containerName="aodh-notifier" containerID="cri-o://ddf97cb37c806fcba6e84c283ff3c7442d458018d05d266877c7bc25b8ad7687" gracePeriod=30 Oct 05 08:30:49 crc kubenswrapper[4846]: I1005 08:30:49.995476 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="05c8753d-a522-467c-9a13-93961e9061c1" containerName="aodh-evaluator" containerID="cri-o://f07446ef7642b96e7c794af68bca50a7f991d2fb618f2c6a493a0d9e86eb1a43" gracePeriod=30 Oct 05 08:30:49 crc kubenswrapper[4846]: I1005 08:30:49.994577 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="05c8753d-a522-467c-9a13-93961e9061c1" containerName="aodh-api" containerID="cri-o://0fc4a5a1699764227e218cb40055ffedcc4d7218376607ac97664dfd27ff5b30" gracePeriod=30 Oct 05 08:30:50 crc kubenswrapper[4846]: I1005 08:30:50.024938 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/aodh-0" podStartSLOduration=2.185358163 podStartE2EDuration="7.024917646s" podCreationTimestamp="2025-10-05 08:30:43 +0000 UTC" firstStartedPulling="2025-10-05 08:30:44.674312782 +0000 UTC m=+6166.915165557" lastFinishedPulling="2025-10-05 08:30:49.513872265 +0000 UTC m=+6171.754725040" observedRunningTime="2025-10-05 08:30:50.021041273 +0000 UTC m=+6172.261894048" watchObservedRunningTime="2025-10-05 08:30:50.024917646 +0000 UTC m=+6172.265770431" Oct 05 08:30:50 crc kubenswrapper[4846]: I1005 08:30:50.059744 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 05 08:30:50 crc kubenswrapper[4846]: I1005 08:30:50.069393 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 05 08:30:50 crc kubenswrapper[4846]: I1005 08:30:50.077404 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-38f8-account-create-6qwhv"] Oct 05 08:30:50 crc kubenswrapper[4846]: I1005 08:30:50.089331 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-38f8-account-create-6qwhv"] Oct 05 08:30:50 crc kubenswrapper[4846]: I1005 08:30:50.100768 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 05 08:30:50 crc kubenswrapper[4846]: E1005 08:30:50.101410 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8c7a796f-58ed-436b-aef9-5a3af4161c99" containerName="ceilometer-central-agent" Oct 05 08:30:50 crc kubenswrapper[4846]: I1005 08:30:50.101433 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="8c7a796f-58ed-436b-aef9-5a3af4161c99" containerName="ceilometer-central-agent" Oct 05 08:30:50 crc kubenswrapper[4846]: E1005 08:30:50.101453 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8c7a796f-58ed-436b-aef9-5a3af4161c99" containerName="ceilometer-notification-agent" Oct 05 08:30:50 crc kubenswrapper[4846]: I1005 08:30:50.101463 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="8c7a796f-58ed-436b-aef9-5a3af4161c99" containerName="ceilometer-notification-agent" Oct 05 08:30:50 crc kubenswrapper[4846]: E1005 08:30:50.101490 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8c7a796f-58ed-436b-aef9-5a3af4161c99" containerName="sg-core" Oct 05 08:30:50 crc kubenswrapper[4846]: I1005 08:30:50.101500 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="8c7a796f-58ed-436b-aef9-5a3af4161c99" containerName="sg-core" Oct 05 08:30:50 crc kubenswrapper[4846]: E1005 08:30:50.101531 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8c7a796f-58ed-436b-aef9-5a3af4161c99" containerName="proxy-httpd" Oct 05 08:30:50 crc kubenswrapper[4846]: I1005 08:30:50.101539 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="8c7a796f-58ed-436b-aef9-5a3af4161c99" containerName="proxy-httpd" Oct 05 08:30:50 crc kubenswrapper[4846]: I1005 08:30:50.101852 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="8c7a796f-58ed-436b-aef9-5a3af4161c99" containerName="proxy-httpd" Oct 05 08:30:50 crc kubenswrapper[4846]: I1005 08:30:50.101874 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="8c7a796f-58ed-436b-aef9-5a3af4161c99" containerName="ceilometer-central-agent" Oct 05 08:30:50 crc kubenswrapper[4846]: I1005 08:30:50.101899 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="8c7a796f-58ed-436b-aef9-5a3af4161c99" containerName="sg-core" Oct 05 08:30:50 crc kubenswrapper[4846]: I1005 08:30:50.101917 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="8c7a796f-58ed-436b-aef9-5a3af4161c99" containerName="ceilometer-notification-agent" Oct 05 08:30:50 crc kubenswrapper[4846]: I1005 08:30:50.104353 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 05 08:30:50 crc kubenswrapper[4846]: I1005 08:30:50.108782 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 05 08:30:50 crc kubenswrapper[4846]: I1005 08:30:50.109051 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 05 08:30:50 crc kubenswrapper[4846]: I1005 08:30:50.111111 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Oct 05 08:30:50 crc kubenswrapper[4846]: I1005 08:30:50.111782 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 05 08:30:50 crc kubenswrapper[4846]: I1005 08:30:50.223426 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0e52324c-bd19-4988-8053-ebe656cc00ba-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"0e52324c-bd19-4988-8053-ebe656cc00ba\") " pod="openstack/ceilometer-0" Oct 05 08:30:50 crc kubenswrapper[4846]: I1005 08:30:50.223463 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0e52324c-bd19-4988-8053-ebe656cc00ba-scripts\") pod \"ceilometer-0\" (UID: \"0e52324c-bd19-4988-8053-ebe656cc00ba\") " pod="openstack/ceilometer-0" Oct 05 08:30:50 crc kubenswrapper[4846]: I1005 08:30:50.223664 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0e52324c-bd19-4988-8053-ebe656cc00ba-log-httpd\") pod \"ceilometer-0\" (UID: \"0e52324c-bd19-4988-8053-ebe656cc00ba\") " pod="openstack/ceilometer-0" Oct 05 08:30:50 crc kubenswrapper[4846]: I1005 08:30:50.223683 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0e52324c-bd19-4988-8053-ebe656cc00ba-run-httpd\") pod \"ceilometer-0\" (UID: \"0e52324c-bd19-4988-8053-ebe656cc00ba\") " pod="openstack/ceilometer-0" Oct 05 08:30:50 crc kubenswrapper[4846]: I1005 08:30:50.223734 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/0e52324c-bd19-4988-8053-ebe656cc00ba-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"0e52324c-bd19-4988-8053-ebe656cc00ba\") " pod="openstack/ceilometer-0" Oct 05 08:30:50 crc kubenswrapper[4846]: I1005 08:30:50.223756 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0e52324c-bd19-4988-8053-ebe656cc00ba-config-data\") pod \"ceilometer-0\" (UID: \"0e52324c-bd19-4988-8053-ebe656cc00ba\") " pod="openstack/ceilometer-0" Oct 05 08:30:50 crc kubenswrapper[4846]: I1005 08:30:50.223780 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0e52324c-bd19-4988-8053-ebe656cc00ba-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"0e52324c-bd19-4988-8053-ebe656cc00ba\") " pod="openstack/ceilometer-0" Oct 05 08:30:50 crc kubenswrapper[4846]: I1005 08:30:50.223813 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8dv2p\" (UniqueName: \"kubernetes.io/projected/0e52324c-bd19-4988-8053-ebe656cc00ba-kube-api-access-8dv2p\") pod \"ceilometer-0\" (UID: \"0e52324c-bd19-4988-8053-ebe656cc00ba\") " pod="openstack/ceilometer-0" Oct 05 08:30:50 crc kubenswrapper[4846]: I1005 08:30:50.258923 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 05 08:30:50 crc kubenswrapper[4846]: E1005 08:30:50.259687 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[ceilometer-tls-certs combined-ca-bundle config-data kube-api-access-8dv2p log-httpd run-httpd scripts sg-core-conf-yaml], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openstack/ceilometer-0" podUID="0e52324c-bd19-4988-8053-ebe656cc00ba" Oct 05 08:30:50 crc kubenswrapper[4846]: I1005 08:30:50.325820 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0e52324c-bd19-4988-8053-ebe656cc00ba-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"0e52324c-bd19-4988-8053-ebe656cc00ba\") " pod="openstack/ceilometer-0" Oct 05 08:30:50 crc kubenswrapper[4846]: I1005 08:30:50.325858 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0e52324c-bd19-4988-8053-ebe656cc00ba-scripts\") pod \"ceilometer-0\" (UID: \"0e52324c-bd19-4988-8053-ebe656cc00ba\") " pod="openstack/ceilometer-0" Oct 05 08:30:50 crc kubenswrapper[4846]: I1005 08:30:50.325939 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0e52324c-bd19-4988-8053-ebe656cc00ba-log-httpd\") pod \"ceilometer-0\" (UID: \"0e52324c-bd19-4988-8053-ebe656cc00ba\") " pod="openstack/ceilometer-0" Oct 05 08:30:50 crc kubenswrapper[4846]: I1005 08:30:50.325955 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0e52324c-bd19-4988-8053-ebe656cc00ba-run-httpd\") pod \"ceilometer-0\" (UID: \"0e52324c-bd19-4988-8053-ebe656cc00ba\") " pod="openstack/ceilometer-0" Oct 05 08:30:50 crc kubenswrapper[4846]: I1005 08:30:50.326006 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/0e52324c-bd19-4988-8053-ebe656cc00ba-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"0e52324c-bd19-4988-8053-ebe656cc00ba\") " pod="openstack/ceilometer-0" Oct 05 08:30:50 crc kubenswrapper[4846]: I1005 08:30:50.326025 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0e52324c-bd19-4988-8053-ebe656cc00ba-config-data\") pod \"ceilometer-0\" (UID: \"0e52324c-bd19-4988-8053-ebe656cc00ba\") " pod="openstack/ceilometer-0" Oct 05 08:30:50 crc kubenswrapper[4846]: I1005 08:30:50.326046 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0e52324c-bd19-4988-8053-ebe656cc00ba-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"0e52324c-bd19-4988-8053-ebe656cc00ba\") " pod="openstack/ceilometer-0" Oct 05 08:30:50 crc kubenswrapper[4846]: I1005 08:30:50.326075 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8dv2p\" (UniqueName: \"kubernetes.io/projected/0e52324c-bd19-4988-8053-ebe656cc00ba-kube-api-access-8dv2p\") pod \"ceilometer-0\" (UID: \"0e52324c-bd19-4988-8053-ebe656cc00ba\") " pod="openstack/ceilometer-0" Oct 05 08:30:50 crc kubenswrapper[4846]: I1005 08:30:50.327224 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0e52324c-bd19-4988-8053-ebe656cc00ba-log-httpd\") pod \"ceilometer-0\" (UID: \"0e52324c-bd19-4988-8053-ebe656cc00ba\") " pod="openstack/ceilometer-0" Oct 05 08:30:50 crc kubenswrapper[4846]: I1005 08:30:50.327328 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0e52324c-bd19-4988-8053-ebe656cc00ba-run-httpd\") pod \"ceilometer-0\" (UID: \"0e52324c-bd19-4988-8053-ebe656cc00ba\") " pod="openstack/ceilometer-0" Oct 05 08:30:50 crc kubenswrapper[4846]: I1005 08:30:50.332913 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0e52324c-bd19-4988-8053-ebe656cc00ba-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"0e52324c-bd19-4988-8053-ebe656cc00ba\") " pod="openstack/ceilometer-0" Oct 05 08:30:50 crc kubenswrapper[4846]: I1005 08:30:50.338631 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0e52324c-bd19-4988-8053-ebe656cc00ba-config-data\") pod \"ceilometer-0\" (UID: \"0e52324c-bd19-4988-8053-ebe656cc00ba\") " pod="openstack/ceilometer-0" Oct 05 08:30:50 crc kubenswrapper[4846]: I1005 08:30:50.339646 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/0e52324c-bd19-4988-8053-ebe656cc00ba-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"0e52324c-bd19-4988-8053-ebe656cc00ba\") " pod="openstack/ceilometer-0" Oct 05 08:30:50 crc kubenswrapper[4846]: I1005 08:30:50.347015 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0e52324c-bd19-4988-8053-ebe656cc00ba-scripts\") pod \"ceilometer-0\" (UID: \"0e52324c-bd19-4988-8053-ebe656cc00ba\") " pod="openstack/ceilometer-0" Oct 05 08:30:50 crc kubenswrapper[4846]: I1005 08:30:50.347576 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0e52324c-bd19-4988-8053-ebe656cc00ba-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"0e52324c-bd19-4988-8053-ebe656cc00ba\") " pod="openstack/ceilometer-0" Oct 05 08:30:50 crc kubenswrapper[4846]: I1005 08:30:50.356793 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8dv2p\" (UniqueName: \"kubernetes.io/projected/0e52324c-bd19-4988-8053-ebe656cc00ba-kube-api-access-8dv2p\") pod \"ceilometer-0\" (UID: \"0e52324c-bd19-4988-8053-ebe656cc00ba\") " pod="openstack/ceilometer-0" Oct 05 08:30:50 crc kubenswrapper[4846]: I1005 08:30:50.511854 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8c7a796f-58ed-436b-aef9-5a3af4161c99" path="/var/lib/kubelet/pods/8c7a796f-58ed-436b-aef9-5a3af4161c99/volumes" Oct 05 08:30:50 crc kubenswrapper[4846]: I1005 08:30:50.512712 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d9aa2e92-ea0b-45e4-bf48-e02ccacf73a6" path="/var/lib/kubelet/pods/d9aa2e92-ea0b-45e4-bf48-e02ccacf73a6/volumes" Oct 05 08:30:51 crc kubenswrapper[4846]: I1005 08:30:51.007486 4846 generic.go:334] "Generic (PLEG): container finished" podID="05c8753d-a522-467c-9a13-93961e9061c1" containerID="f07446ef7642b96e7c794af68bca50a7f991d2fb618f2c6a493a0d9e86eb1a43" exitCode=0 Oct 05 08:30:51 crc kubenswrapper[4846]: I1005 08:30:51.007781 4846 generic.go:334] "Generic (PLEG): container finished" podID="05c8753d-a522-467c-9a13-93961e9061c1" containerID="0fc4a5a1699764227e218cb40055ffedcc4d7218376607ac97664dfd27ff5b30" exitCode=0 Oct 05 08:30:51 crc kubenswrapper[4846]: I1005 08:30:51.007835 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 05 08:30:51 crc kubenswrapper[4846]: I1005 08:30:51.007566 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"05c8753d-a522-467c-9a13-93961e9061c1","Type":"ContainerDied","Data":"f07446ef7642b96e7c794af68bca50a7f991d2fb618f2c6a493a0d9e86eb1a43"} Oct 05 08:30:51 crc kubenswrapper[4846]: I1005 08:30:51.008630 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"05c8753d-a522-467c-9a13-93961e9061c1","Type":"ContainerDied","Data":"0fc4a5a1699764227e218cb40055ffedcc4d7218376607ac97664dfd27ff5b30"} Oct 05 08:30:51 crc kubenswrapper[4846]: I1005 08:30:51.020722 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 05 08:30:51 crc kubenswrapper[4846]: I1005 08:30:51.141316 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0e52324c-bd19-4988-8053-ebe656cc00ba-run-httpd\") pod \"0e52324c-bd19-4988-8053-ebe656cc00ba\" (UID: \"0e52324c-bd19-4988-8053-ebe656cc00ba\") " Oct 05 08:30:51 crc kubenswrapper[4846]: I1005 08:30:51.141388 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0e52324c-bd19-4988-8053-ebe656cc00ba-sg-core-conf-yaml\") pod \"0e52324c-bd19-4988-8053-ebe656cc00ba\" (UID: \"0e52324c-bd19-4988-8053-ebe656cc00ba\") " Oct 05 08:30:51 crc kubenswrapper[4846]: I1005 08:30:51.141440 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0e52324c-bd19-4988-8053-ebe656cc00ba-scripts\") pod \"0e52324c-bd19-4988-8053-ebe656cc00ba\" (UID: \"0e52324c-bd19-4988-8053-ebe656cc00ba\") " Oct 05 08:30:51 crc kubenswrapper[4846]: I1005 08:30:51.141495 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8dv2p\" (UniqueName: \"kubernetes.io/projected/0e52324c-bd19-4988-8053-ebe656cc00ba-kube-api-access-8dv2p\") pod \"0e52324c-bd19-4988-8053-ebe656cc00ba\" (UID: \"0e52324c-bd19-4988-8053-ebe656cc00ba\") " Oct 05 08:30:51 crc kubenswrapper[4846]: I1005 08:30:51.141518 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0e52324c-bd19-4988-8053-ebe656cc00ba-config-data\") pod \"0e52324c-bd19-4988-8053-ebe656cc00ba\" (UID: \"0e52324c-bd19-4988-8053-ebe656cc00ba\") " Oct 05 08:30:51 crc kubenswrapper[4846]: I1005 08:30:51.141550 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/0e52324c-bd19-4988-8053-ebe656cc00ba-ceilometer-tls-certs\") pod \"0e52324c-bd19-4988-8053-ebe656cc00ba\" (UID: \"0e52324c-bd19-4988-8053-ebe656cc00ba\") " Oct 05 08:30:51 crc kubenswrapper[4846]: I1005 08:30:51.141595 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0e52324c-bd19-4988-8053-ebe656cc00ba-combined-ca-bundle\") pod \"0e52324c-bd19-4988-8053-ebe656cc00ba\" (UID: \"0e52324c-bd19-4988-8053-ebe656cc00ba\") " Oct 05 08:30:51 crc kubenswrapper[4846]: I1005 08:30:51.141619 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0e52324c-bd19-4988-8053-ebe656cc00ba-log-httpd\") pod \"0e52324c-bd19-4988-8053-ebe656cc00ba\" (UID: \"0e52324c-bd19-4988-8053-ebe656cc00ba\") " Oct 05 08:30:51 crc kubenswrapper[4846]: I1005 08:30:51.141686 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0e52324c-bd19-4988-8053-ebe656cc00ba-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "0e52324c-bd19-4988-8053-ebe656cc00ba" (UID: "0e52324c-bd19-4988-8053-ebe656cc00ba"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 08:30:51 crc kubenswrapper[4846]: I1005 08:30:51.142119 4846 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0e52324c-bd19-4988-8053-ebe656cc00ba-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 05 08:30:51 crc kubenswrapper[4846]: I1005 08:30:51.142439 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0e52324c-bd19-4988-8053-ebe656cc00ba-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "0e52324c-bd19-4988-8053-ebe656cc00ba" (UID: "0e52324c-bd19-4988-8053-ebe656cc00ba"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 08:30:51 crc kubenswrapper[4846]: I1005 08:30:51.146631 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0e52324c-bd19-4988-8053-ebe656cc00ba-scripts" (OuterVolumeSpecName: "scripts") pod "0e52324c-bd19-4988-8053-ebe656cc00ba" (UID: "0e52324c-bd19-4988-8053-ebe656cc00ba"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:30:51 crc kubenswrapper[4846]: I1005 08:30:51.146646 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0e52324c-bd19-4988-8053-ebe656cc00ba-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "0e52324c-bd19-4988-8053-ebe656cc00ba" (UID: "0e52324c-bd19-4988-8053-ebe656cc00ba"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:30:51 crc kubenswrapper[4846]: I1005 08:30:51.147487 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0e52324c-bd19-4988-8053-ebe656cc00ba-config-data" (OuterVolumeSpecName: "config-data") pod "0e52324c-bd19-4988-8053-ebe656cc00ba" (UID: "0e52324c-bd19-4988-8053-ebe656cc00ba"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:30:51 crc kubenswrapper[4846]: I1005 08:30:51.147533 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0e52324c-bd19-4988-8053-ebe656cc00ba-kube-api-access-8dv2p" (OuterVolumeSpecName: "kube-api-access-8dv2p") pod "0e52324c-bd19-4988-8053-ebe656cc00ba" (UID: "0e52324c-bd19-4988-8053-ebe656cc00ba"). InnerVolumeSpecName "kube-api-access-8dv2p". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:30:51 crc kubenswrapper[4846]: I1005 08:30:51.147671 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0e52324c-bd19-4988-8053-ebe656cc00ba-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0e52324c-bd19-4988-8053-ebe656cc00ba" (UID: "0e52324c-bd19-4988-8053-ebe656cc00ba"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:30:51 crc kubenswrapper[4846]: I1005 08:30:51.169668 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0e52324c-bd19-4988-8053-ebe656cc00ba-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "0e52324c-bd19-4988-8053-ebe656cc00ba" (UID: "0e52324c-bd19-4988-8053-ebe656cc00ba"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:30:51 crc kubenswrapper[4846]: I1005 08:30:51.244061 4846 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0e52324c-bd19-4988-8053-ebe656cc00ba-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 05 08:30:51 crc kubenswrapper[4846]: I1005 08:30:51.244099 4846 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0e52324c-bd19-4988-8053-ebe656cc00ba-scripts\") on node \"crc\" DevicePath \"\"" Oct 05 08:30:51 crc kubenswrapper[4846]: I1005 08:30:51.244109 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8dv2p\" (UniqueName: \"kubernetes.io/projected/0e52324c-bd19-4988-8053-ebe656cc00ba-kube-api-access-8dv2p\") on node \"crc\" DevicePath \"\"" Oct 05 08:30:51 crc kubenswrapper[4846]: I1005 08:30:51.244119 4846 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0e52324c-bd19-4988-8053-ebe656cc00ba-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 08:30:51 crc kubenswrapper[4846]: I1005 08:30:51.244129 4846 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/0e52324c-bd19-4988-8053-ebe656cc00ba-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 05 08:30:51 crc kubenswrapper[4846]: I1005 08:30:51.244138 4846 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0e52324c-bd19-4988-8053-ebe656cc00ba-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 08:30:51 crc kubenswrapper[4846]: I1005 08:30:51.244145 4846 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0e52324c-bd19-4988-8053-ebe656cc00ba-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 05 08:30:52 crc kubenswrapper[4846]: I1005 08:30:52.017108 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 05 08:30:52 crc kubenswrapper[4846]: I1005 08:30:52.111272 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 05 08:30:52 crc kubenswrapper[4846]: I1005 08:30:52.172740 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 05 08:30:52 crc kubenswrapper[4846]: I1005 08:30:52.201566 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 05 08:30:52 crc kubenswrapper[4846]: I1005 08:30:52.208333 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 05 08:30:52 crc kubenswrapper[4846]: I1005 08:30:52.215144 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 05 08:30:52 crc kubenswrapper[4846]: I1005 08:30:52.215404 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Oct 05 08:30:52 crc kubenswrapper[4846]: I1005 08:30:52.215487 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 05 08:30:52 crc kubenswrapper[4846]: I1005 08:30:52.230967 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 05 08:30:52 crc kubenswrapper[4846]: I1005 08:30:52.395127 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zfrbv\" (UniqueName: \"kubernetes.io/projected/31f6d8fa-326e-46cb-83ca-9294b1e20903-kube-api-access-zfrbv\") pod \"ceilometer-0\" (UID: \"31f6d8fa-326e-46cb-83ca-9294b1e20903\") " pod="openstack/ceilometer-0" Oct 05 08:30:52 crc kubenswrapper[4846]: I1005 08:30:52.395228 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/31f6d8fa-326e-46cb-83ca-9294b1e20903-config-data\") pod \"ceilometer-0\" (UID: \"31f6d8fa-326e-46cb-83ca-9294b1e20903\") " pod="openstack/ceilometer-0" Oct 05 08:30:52 crc kubenswrapper[4846]: I1005 08:30:52.395264 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/31f6d8fa-326e-46cb-83ca-9294b1e20903-log-httpd\") pod \"ceilometer-0\" (UID: \"31f6d8fa-326e-46cb-83ca-9294b1e20903\") " pod="openstack/ceilometer-0" Oct 05 08:30:52 crc kubenswrapper[4846]: I1005 08:30:52.395317 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/31f6d8fa-326e-46cb-83ca-9294b1e20903-run-httpd\") pod \"ceilometer-0\" (UID: \"31f6d8fa-326e-46cb-83ca-9294b1e20903\") " pod="openstack/ceilometer-0" Oct 05 08:30:52 crc kubenswrapper[4846]: I1005 08:30:52.395335 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/31f6d8fa-326e-46cb-83ca-9294b1e20903-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"31f6d8fa-326e-46cb-83ca-9294b1e20903\") " pod="openstack/ceilometer-0" Oct 05 08:30:52 crc kubenswrapper[4846]: I1005 08:30:52.395361 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/31f6d8fa-326e-46cb-83ca-9294b1e20903-scripts\") pod \"ceilometer-0\" (UID: \"31f6d8fa-326e-46cb-83ca-9294b1e20903\") " pod="openstack/ceilometer-0" Oct 05 08:30:52 crc kubenswrapper[4846]: I1005 08:30:52.395379 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/31f6d8fa-326e-46cb-83ca-9294b1e20903-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"31f6d8fa-326e-46cb-83ca-9294b1e20903\") " pod="openstack/ceilometer-0" Oct 05 08:30:52 crc kubenswrapper[4846]: I1005 08:30:52.395447 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/31f6d8fa-326e-46cb-83ca-9294b1e20903-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"31f6d8fa-326e-46cb-83ca-9294b1e20903\") " pod="openstack/ceilometer-0" Oct 05 08:30:52 crc kubenswrapper[4846]: I1005 08:30:52.499161 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zfrbv\" (UniqueName: \"kubernetes.io/projected/31f6d8fa-326e-46cb-83ca-9294b1e20903-kube-api-access-zfrbv\") pod \"ceilometer-0\" (UID: \"31f6d8fa-326e-46cb-83ca-9294b1e20903\") " pod="openstack/ceilometer-0" Oct 05 08:30:52 crc kubenswrapper[4846]: I1005 08:30:52.499870 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/31f6d8fa-326e-46cb-83ca-9294b1e20903-config-data\") pod \"ceilometer-0\" (UID: \"31f6d8fa-326e-46cb-83ca-9294b1e20903\") " pod="openstack/ceilometer-0" Oct 05 08:30:52 crc kubenswrapper[4846]: I1005 08:30:52.499905 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/31f6d8fa-326e-46cb-83ca-9294b1e20903-log-httpd\") pod \"ceilometer-0\" (UID: \"31f6d8fa-326e-46cb-83ca-9294b1e20903\") " pod="openstack/ceilometer-0" Oct 05 08:30:52 crc kubenswrapper[4846]: I1005 08:30:52.500560 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/31f6d8fa-326e-46cb-83ca-9294b1e20903-log-httpd\") pod \"ceilometer-0\" (UID: \"31f6d8fa-326e-46cb-83ca-9294b1e20903\") " pod="openstack/ceilometer-0" Oct 05 08:30:52 crc kubenswrapper[4846]: I1005 08:30:52.500653 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/31f6d8fa-326e-46cb-83ca-9294b1e20903-run-httpd\") pod \"ceilometer-0\" (UID: \"31f6d8fa-326e-46cb-83ca-9294b1e20903\") " pod="openstack/ceilometer-0" Oct 05 08:30:52 crc kubenswrapper[4846]: I1005 08:30:52.500705 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/31f6d8fa-326e-46cb-83ca-9294b1e20903-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"31f6d8fa-326e-46cb-83ca-9294b1e20903\") " pod="openstack/ceilometer-0" Oct 05 08:30:52 crc kubenswrapper[4846]: I1005 08:30:52.500759 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/31f6d8fa-326e-46cb-83ca-9294b1e20903-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"31f6d8fa-326e-46cb-83ca-9294b1e20903\") " pod="openstack/ceilometer-0" Oct 05 08:30:52 crc kubenswrapper[4846]: I1005 08:30:52.500787 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/31f6d8fa-326e-46cb-83ca-9294b1e20903-scripts\") pod \"ceilometer-0\" (UID: \"31f6d8fa-326e-46cb-83ca-9294b1e20903\") " pod="openstack/ceilometer-0" Oct 05 08:30:52 crc kubenswrapper[4846]: I1005 08:30:52.500945 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/31f6d8fa-326e-46cb-83ca-9294b1e20903-run-httpd\") pod \"ceilometer-0\" (UID: \"31f6d8fa-326e-46cb-83ca-9294b1e20903\") " pod="openstack/ceilometer-0" Oct 05 08:30:52 crc kubenswrapper[4846]: I1005 08:30:52.500962 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/31f6d8fa-326e-46cb-83ca-9294b1e20903-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"31f6d8fa-326e-46cb-83ca-9294b1e20903\") " pod="openstack/ceilometer-0" Oct 05 08:30:52 crc kubenswrapper[4846]: I1005 08:30:52.503762 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/31f6d8fa-326e-46cb-83ca-9294b1e20903-config-data\") pod \"ceilometer-0\" (UID: \"31f6d8fa-326e-46cb-83ca-9294b1e20903\") " pod="openstack/ceilometer-0" Oct 05 08:30:52 crc kubenswrapper[4846]: I1005 08:30:52.504778 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/31f6d8fa-326e-46cb-83ca-9294b1e20903-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"31f6d8fa-326e-46cb-83ca-9294b1e20903\") " pod="openstack/ceilometer-0" Oct 05 08:30:52 crc kubenswrapper[4846]: I1005 08:30:52.505194 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/31f6d8fa-326e-46cb-83ca-9294b1e20903-scripts\") pod \"ceilometer-0\" (UID: \"31f6d8fa-326e-46cb-83ca-9294b1e20903\") " pod="openstack/ceilometer-0" Oct 05 08:30:52 crc kubenswrapper[4846]: I1005 08:30:52.506550 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/31f6d8fa-326e-46cb-83ca-9294b1e20903-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"31f6d8fa-326e-46cb-83ca-9294b1e20903\") " pod="openstack/ceilometer-0" Oct 05 08:30:52 crc kubenswrapper[4846]: I1005 08:30:52.507503 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/31f6d8fa-326e-46cb-83ca-9294b1e20903-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"31f6d8fa-326e-46cb-83ca-9294b1e20903\") " pod="openstack/ceilometer-0" Oct 05 08:30:52 crc kubenswrapper[4846]: I1005 08:30:52.516674 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0e52324c-bd19-4988-8053-ebe656cc00ba" path="/var/lib/kubelet/pods/0e52324c-bd19-4988-8053-ebe656cc00ba/volumes" Oct 05 08:30:52 crc kubenswrapper[4846]: I1005 08:30:52.518124 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zfrbv\" (UniqueName: \"kubernetes.io/projected/31f6d8fa-326e-46cb-83ca-9294b1e20903-kube-api-access-zfrbv\") pod \"ceilometer-0\" (UID: \"31f6d8fa-326e-46cb-83ca-9294b1e20903\") " pod="openstack/ceilometer-0" Oct 05 08:30:52 crc kubenswrapper[4846]: I1005 08:30:52.531083 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 05 08:30:52 crc kubenswrapper[4846]: I1005 08:30:52.990957 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 05 08:30:52 crc kubenswrapper[4846]: W1005 08:30:52.996287 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod31f6d8fa_326e_46cb_83ca_9294b1e20903.slice/crio-65610295aa71dbfacf3dbb0f9cce2540e5cddb47a56e77bdb6784d1fb362754a WatchSource:0}: Error finding container 65610295aa71dbfacf3dbb0f9cce2540e5cddb47a56e77bdb6784d1fb362754a: Status 404 returned error can't find the container with id 65610295aa71dbfacf3dbb0f9cce2540e5cddb47a56e77bdb6784d1fb362754a Oct 05 08:30:53 crc kubenswrapper[4846]: I1005 08:30:53.025926 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"31f6d8fa-326e-46cb-83ca-9294b1e20903","Type":"ContainerStarted","Data":"65610295aa71dbfacf3dbb0f9cce2540e5cddb47a56e77bdb6784d1fb362754a"} Oct 05 08:30:54 crc kubenswrapper[4846]: I1005 08:30:54.036478 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"31f6d8fa-326e-46cb-83ca-9294b1e20903","Type":"ContainerStarted","Data":"45b6108f7630b83747ea9dcfd3c48f135b517f5a089d8f7b0eb375c23e78ba95"} Oct 05 08:30:54 crc kubenswrapper[4846]: I1005 08:30:54.036909 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"31f6d8fa-326e-46cb-83ca-9294b1e20903","Type":"ContainerStarted","Data":"86e309660ac1e5fa9e7ac0c8f32d7f83c9c2cb54098bc4b72a7b3d8052fd7563"} Oct 05 08:30:55 crc kubenswrapper[4846]: I1005 08:30:55.048805 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"31f6d8fa-326e-46cb-83ca-9294b1e20903","Type":"ContainerStarted","Data":"56c2ab99e3608fe5da78a5fe0b166df497b171f0b4618a145be0cc065e38e49f"} Oct 05 08:30:56 crc kubenswrapper[4846]: I1005 08:30:56.069235 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"31f6d8fa-326e-46cb-83ca-9294b1e20903","Type":"ContainerStarted","Data":"5d7c625b281dccbd4025e94a37db50bb03a885abdee7679eb352c914b036a347"} Oct 05 08:30:56 crc kubenswrapper[4846]: I1005 08:30:56.069921 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 05 08:30:56 crc kubenswrapper[4846]: I1005 08:30:56.104206 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.657730685 podStartE2EDuration="4.104162801s" podCreationTimestamp="2025-10-05 08:30:52 +0000 UTC" firstStartedPulling="2025-10-05 08:30:53.00002326 +0000 UTC m=+6175.240876075" lastFinishedPulling="2025-10-05 08:30:55.446455426 +0000 UTC m=+6177.687308191" observedRunningTime="2025-10-05 08:30:56.095322717 +0000 UTC m=+6178.336175502" watchObservedRunningTime="2025-10-05 08:30:56.104162801 +0000 UTC m=+6178.345015576" Oct 05 08:30:57 crc kubenswrapper[4846]: E1005 08:30:57.949999 4846 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9eeca81b_6fda_4c22_a221_3adcb9900077.slice/crio-e8538609494690d52f4d579e3ae6523d65701c417f2551cb9f7b3cad911741a5\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9eeca81b_6fda_4c22_a221_3adcb9900077.slice\": RecentStats: unable to find data in memory cache]" Oct 05 08:31:14 crc kubenswrapper[4846]: I1005 08:31:14.056571 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-sync-frwh2"] Oct 05 08:31:14 crc kubenswrapper[4846]: I1005 08:31:14.064115 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-sync-frwh2"] Oct 05 08:31:14 crc kubenswrapper[4846]: I1005 08:31:14.512054 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="46c68460-594a-4ab4-b41a-268676e6b950" path="/var/lib/kubelet/pods/46c68460-594a-4ab4-b41a-268676e6b950/volumes" Oct 05 08:31:20 crc kubenswrapper[4846]: E1005 08:31:20.375304 4846 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod05c8753d_a522_467c_9a13_93961e9061c1.slice/crio-conmon-d7f1352d94a9be8d08e2468ed0cfea65a2bee3fb2ca2a4d8f522015dc3840aae.scope\": RecentStats: unable to find data in memory cache]" Oct 05 08:31:20 crc kubenswrapper[4846]: I1005 08:31:20.382088 4846 generic.go:334] "Generic (PLEG): container finished" podID="05c8753d-a522-467c-9a13-93961e9061c1" containerID="d7f1352d94a9be8d08e2468ed0cfea65a2bee3fb2ca2a4d8f522015dc3840aae" exitCode=137 Oct 05 08:31:20 crc kubenswrapper[4846]: I1005 08:31:20.382118 4846 generic.go:334] "Generic (PLEG): container finished" podID="05c8753d-a522-467c-9a13-93961e9061c1" containerID="ddf97cb37c806fcba6e84c283ff3c7442d458018d05d266877c7bc25b8ad7687" exitCode=137 Oct 05 08:31:20 crc kubenswrapper[4846]: I1005 08:31:20.382138 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"05c8753d-a522-467c-9a13-93961e9061c1","Type":"ContainerDied","Data":"d7f1352d94a9be8d08e2468ed0cfea65a2bee3fb2ca2a4d8f522015dc3840aae"} Oct 05 08:31:20 crc kubenswrapper[4846]: I1005 08:31:20.382164 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"05c8753d-a522-467c-9a13-93961e9061c1","Type":"ContainerDied","Data":"ddf97cb37c806fcba6e84c283ff3c7442d458018d05d266877c7bc25b8ad7687"} Oct 05 08:31:20 crc kubenswrapper[4846]: I1005 08:31:20.502774 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Oct 05 08:31:20 crc kubenswrapper[4846]: I1005 08:31:20.535395 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/05c8753d-a522-467c-9a13-93961e9061c1-scripts\") pod \"05c8753d-a522-467c-9a13-93961e9061c1\" (UID: \"05c8753d-a522-467c-9a13-93961e9061c1\") " Oct 05 08:31:20 crc kubenswrapper[4846]: I1005 08:31:20.535467 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cszgp\" (UniqueName: \"kubernetes.io/projected/05c8753d-a522-467c-9a13-93961e9061c1-kube-api-access-cszgp\") pod \"05c8753d-a522-467c-9a13-93961e9061c1\" (UID: \"05c8753d-a522-467c-9a13-93961e9061c1\") " Oct 05 08:31:20 crc kubenswrapper[4846]: I1005 08:31:20.535492 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/05c8753d-a522-467c-9a13-93961e9061c1-combined-ca-bundle\") pod \"05c8753d-a522-467c-9a13-93961e9061c1\" (UID: \"05c8753d-a522-467c-9a13-93961e9061c1\") " Oct 05 08:31:20 crc kubenswrapper[4846]: I1005 08:31:20.535558 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/05c8753d-a522-467c-9a13-93961e9061c1-config-data\") pod \"05c8753d-a522-467c-9a13-93961e9061c1\" (UID: \"05c8753d-a522-467c-9a13-93961e9061c1\") " Oct 05 08:31:20 crc kubenswrapper[4846]: I1005 08:31:20.542416 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/05c8753d-a522-467c-9a13-93961e9061c1-scripts" (OuterVolumeSpecName: "scripts") pod "05c8753d-a522-467c-9a13-93961e9061c1" (UID: "05c8753d-a522-467c-9a13-93961e9061c1"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:31:20 crc kubenswrapper[4846]: I1005 08:31:20.550490 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/05c8753d-a522-467c-9a13-93961e9061c1-kube-api-access-cszgp" (OuterVolumeSpecName: "kube-api-access-cszgp") pod "05c8753d-a522-467c-9a13-93961e9061c1" (UID: "05c8753d-a522-467c-9a13-93961e9061c1"). InnerVolumeSpecName "kube-api-access-cszgp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:31:20 crc kubenswrapper[4846]: I1005 08:31:20.637986 4846 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/05c8753d-a522-467c-9a13-93961e9061c1-scripts\") on node \"crc\" DevicePath \"\"" Oct 05 08:31:20 crc kubenswrapper[4846]: I1005 08:31:20.638284 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cszgp\" (UniqueName: \"kubernetes.io/projected/05c8753d-a522-467c-9a13-93961e9061c1-kube-api-access-cszgp\") on node \"crc\" DevicePath \"\"" Oct 05 08:31:20 crc kubenswrapper[4846]: I1005 08:31:20.654117 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/05c8753d-a522-467c-9a13-93961e9061c1-config-data" (OuterVolumeSpecName: "config-data") pod "05c8753d-a522-467c-9a13-93961e9061c1" (UID: "05c8753d-a522-467c-9a13-93961e9061c1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:31:20 crc kubenswrapper[4846]: I1005 08:31:20.656936 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/05c8753d-a522-467c-9a13-93961e9061c1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "05c8753d-a522-467c-9a13-93961e9061c1" (UID: "05c8753d-a522-467c-9a13-93961e9061c1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:31:20 crc kubenswrapper[4846]: I1005 08:31:20.740296 4846 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/05c8753d-a522-467c-9a13-93961e9061c1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 08:31:20 crc kubenswrapper[4846]: I1005 08:31:20.740331 4846 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/05c8753d-a522-467c-9a13-93961e9061c1-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 08:31:21 crc kubenswrapper[4846]: I1005 08:31:21.397084 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"05c8753d-a522-467c-9a13-93961e9061c1","Type":"ContainerDied","Data":"a2898f58c9ffbfb291ec354ffa6a14431dfda6302c79da544e90036914b95c4d"} Oct 05 08:31:21 crc kubenswrapper[4846]: I1005 08:31:21.397135 4846 scope.go:117] "RemoveContainer" containerID="d7f1352d94a9be8d08e2468ed0cfea65a2bee3fb2ca2a4d8f522015dc3840aae" Oct 05 08:31:21 crc kubenswrapper[4846]: I1005 08:31:21.397289 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Oct 05 08:31:21 crc kubenswrapper[4846]: I1005 08:31:21.441577 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-0"] Oct 05 08:31:21 crc kubenswrapper[4846]: I1005 08:31:21.446479 4846 scope.go:117] "RemoveContainer" containerID="ddf97cb37c806fcba6e84c283ff3c7442d458018d05d266877c7bc25b8ad7687" Oct 05 08:31:21 crc kubenswrapper[4846]: I1005 08:31:21.450686 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/aodh-0"] Oct 05 08:31:21 crc kubenswrapper[4846]: I1005 08:31:21.482712 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-0"] Oct 05 08:31:21 crc kubenswrapper[4846]: E1005 08:31:21.483212 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="05c8753d-a522-467c-9a13-93961e9061c1" containerName="aodh-listener" Oct 05 08:31:21 crc kubenswrapper[4846]: I1005 08:31:21.483232 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="05c8753d-a522-467c-9a13-93961e9061c1" containerName="aodh-listener" Oct 05 08:31:21 crc kubenswrapper[4846]: E1005 08:31:21.483252 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="05c8753d-a522-467c-9a13-93961e9061c1" containerName="aodh-api" Oct 05 08:31:21 crc kubenswrapper[4846]: I1005 08:31:21.483260 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="05c8753d-a522-467c-9a13-93961e9061c1" containerName="aodh-api" Oct 05 08:31:21 crc kubenswrapper[4846]: E1005 08:31:21.483275 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="05c8753d-a522-467c-9a13-93961e9061c1" containerName="aodh-evaluator" Oct 05 08:31:21 crc kubenswrapper[4846]: I1005 08:31:21.483286 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="05c8753d-a522-467c-9a13-93961e9061c1" containerName="aodh-evaluator" Oct 05 08:31:21 crc kubenswrapper[4846]: E1005 08:31:21.483322 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="05c8753d-a522-467c-9a13-93961e9061c1" containerName="aodh-notifier" Oct 05 08:31:21 crc kubenswrapper[4846]: I1005 08:31:21.483331 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="05c8753d-a522-467c-9a13-93961e9061c1" containerName="aodh-notifier" Oct 05 08:31:21 crc kubenswrapper[4846]: I1005 08:31:21.483581 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="05c8753d-a522-467c-9a13-93961e9061c1" containerName="aodh-evaluator" Oct 05 08:31:21 crc kubenswrapper[4846]: I1005 08:31:21.483613 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="05c8753d-a522-467c-9a13-93961e9061c1" containerName="aodh-listener" Oct 05 08:31:21 crc kubenswrapper[4846]: I1005 08:31:21.483629 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="05c8753d-a522-467c-9a13-93961e9061c1" containerName="aodh-api" Oct 05 08:31:21 crc kubenswrapper[4846]: I1005 08:31:21.483649 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="05c8753d-a522-467c-9a13-93961e9061c1" containerName="aodh-notifier" Oct 05 08:31:21 crc kubenswrapper[4846]: I1005 08:31:21.486328 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Oct 05 08:31:21 crc kubenswrapper[4846]: I1005 08:31:21.488692 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-scripts" Oct 05 08:31:21 crc kubenswrapper[4846]: I1005 08:31:21.488797 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-config-data" Oct 05 08:31:21 crc kubenswrapper[4846]: I1005 08:31:21.489012 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-autoscaling-dockercfg-666qr" Oct 05 08:31:21 crc kubenswrapper[4846]: I1005 08:31:21.489204 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-aodh-public-svc" Oct 05 08:31:21 crc kubenswrapper[4846]: I1005 08:31:21.491861 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-aodh-internal-svc" Oct 05 08:31:21 crc kubenswrapper[4846]: I1005 08:31:21.496278 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-0"] Oct 05 08:31:21 crc kubenswrapper[4846]: I1005 08:31:21.536080 4846 scope.go:117] "RemoveContainer" containerID="f07446ef7642b96e7c794af68bca50a7f991d2fb618f2c6a493a0d9e86eb1a43" Oct 05 08:31:21 crc kubenswrapper[4846]: I1005 08:31:21.555559 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c2926ee6-0f75-41e1-8140-cd2a2bfbad2b-public-tls-certs\") pod \"aodh-0\" (UID: \"c2926ee6-0f75-41e1-8140-cd2a2bfbad2b\") " pod="openstack/aodh-0" Oct 05 08:31:21 crc kubenswrapper[4846]: I1005 08:31:21.555696 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2926ee6-0f75-41e1-8140-cd2a2bfbad2b-combined-ca-bundle\") pod \"aodh-0\" (UID: \"c2926ee6-0f75-41e1-8140-cd2a2bfbad2b\") " pod="openstack/aodh-0" Oct 05 08:31:21 crc kubenswrapper[4846]: I1005 08:31:21.556011 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c2926ee6-0f75-41e1-8140-cd2a2bfbad2b-config-data\") pod \"aodh-0\" (UID: \"c2926ee6-0f75-41e1-8140-cd2a2bfbad2b\") " pod="openstack/aodh-0" Oct 05 08:31:21 crc kubenswrapper[4846]: I1005 08:31:21.556045 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c2926ee6-0f75-41e1-8140-cd2a2bfbad2b-scripts\") pod \"aodh-0\" (UID: \"c2926ee6-0f75-41e1-8140-cd2a2bfbad2b\") " pod="openstack/aodh-0" Oct 05 08:31:21 crc kubenswrapper[4846]: I1005 08:31:21.556066 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c2926ee6-0f75-41e1-8140-cd2a2bfbad2b-internal-tls-certs\") pod \"aodh-0\" (UID: \"c2926ee6-0f75-41e1-8140-cd2a2bfbad2b\") " pod="openstack/aodh-0" Oct 05 08:31:21 crc kubenswrapper[4846]: I1005 08:31:21.556086 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gbxmj\" (UniqueName: \"kubernetes.io/projected/c2926ee6-0f75-41e1-8140-cd2a2bfbad2b-kube-api-access-gbxmj\") pod \"aodh-0\" (UID: \"c2926ee6-0f75-41e1-8140-cd2a2bfbad2b\") " pod="openstack/aodh-0" Oct 05 08:31:21 crc kubenswrapper[4846]: I1005 08:31:21.559413 4846 scope.go:117] "RemoveContainer" containerID="0fc4a5a1699764227e218cb40055ffedcc4d7218376607ac97664dfd27ff5b30" Oct 05 08:31:21 crc kubenswrapper[4846]: I1005 08:31:21.657050 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c2926ee6-0f75-41e1-8140-cd2a2bfbad2b-public-tls-certs\") pod \"aodh-0\" (UID: \"c2926ee6-0f75-41e1-8140-cd2a2bfbad2b\") " pod="openstack/aodh-0" Oct 05 08:31:21 crc kubenswrapper[4846]: I1005 08:31:21.657142 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2926ee6-0f75-41e1-8140-cd2a2bfbad2b-combined-ca-bundle\") pod \"aodh-0\" (UID: \"c2926ee6-0f75-41e1-8140-cd2a2bfbad2b\") " pod="openstack/aodh-0" Oct 05 08:31:21 crc kubenswrapper[4846]: I1005 08:31:21.657198 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c2926ee6-0f75-41e1-8140-cd2a2bfbad2b-config-data\") pod \"aodh-0\" (UID: \"c2926ee6-0f75-41e1-8140-cd2a2bfbad2b\") " pod="openstack/aodh-0" Oct 05 08:31:21 crc kubenswrapper[4846]: I1005 08:31:21.657222 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c2926ee6-0f75-41e1-8140-cd2a2bfbad2b-scripts\") pod \"aodh-0\" (UID: \"c2926ee6-0f75-41e1-8140-cd2a2bfbad2b\") " pod="openstack/aodh-0" Oct 05 08:31:21 crc kubenswrapper[4846]: I1005 08:31:21.657242 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c2926ee6-0f75-41e1-8140-cd2a2bfbad2b-internal-tls-certs\") pod \"aodh-0\" (UID: \"c2926ee6-0f75-41e1-8140-cd2a2bfbad2b\") " pod="openstack/aodh-0" Oct 05 08:31:21 crc kubenswrapper[4846]: I1005 08:31:21.657265 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gbxmj\" (UniqueName: \"kubernetes.io/projected/c2926ee6-0f75-41e1-8140-cd2a2bfbad2b-kube-api-access-gbxmj\") pod \"aodh-0\" (UID: \"c2926ee6-0f75-41e1-8140-cd2a2bfbad2b\") " pod="openstack/aodh-0" Oct 05 08:31:21 crc kubenswrapper[4846]: I1005 08:31:21.662956 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c2926ee6-0f75-41e1-8140-cd2a2bfbad2b-public-tls-certs\") pod \"aodh-0\" (UID: \"c2926ee6-0f75-41e1-8140-cd2a2bfbad2b\") " pod="openstack/aodh-0" Oct 05 08:31:21 crc kubenswrapper[4846]: I1005 08:31:21.663656 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c2926ee6-0f75-41e1-8140-cd2a2bfbad2b-scripts\") pod \"aodh-0\" (UID: \"c2926ee6-0f75-41e1-8140-cd2a2bfbad2b\") " pod="openstack/aodh-0" Oct 05 08:31:21 crc kubenswrapper[4846]: I1005 08:31:21.663944 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c2926ee6-0f75-41e1-8140-cd2a2bfbad2b-config-data\") pod \"aodh-0\" (UID: \"c2926ee6-0f75-41e1-8140-cd2a2bfbad2b\") " pod="openstack/aodh-0" Oct 05 08:31:21 crc kubenswrapper[4846]: I1005 08:31:21.670464 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c2926ee6-0f75-41e1-8140-cd2a2bfbad2b-internal-tls-certs\") pod \"aodh-0\" (UID: \"c2926ee6-0f75-41e1-8140-cd2a2bfbad2b\") " pod="openstack/aodh-0" Oct 05 08:31:21 crc kubenswrapper[4846]: I1005 08:31:21.672546 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2926ee6-0f75-41e1-8140-cd2a2bfbad2b-combined-ca-bundle\") pod \"aodh-0\" (UID: \"c2926ee6-0f75-41e1-8140-cd2a2bfbad2b\") " pod="openstack/aodh-0" Oct 05 08:31:21 crc kubenswrapper[4846]: I1005 08:31:21.684209 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gbxmj\" (UniqueName: \"kubernetes.io/projected/c2926ee6-0f75-41e1-8140-cd2a2bfbad2b-kube-api-access-gbxmj\") pod \"aodh-0\" (UID: \"c2926ee6-0f75-41e1-8140-cd2a2bfbad2b\") " pod="openstack/aodh-0" Oct 05 08:31:21 crc kubenswrapper[4846]: I1005 08:31:21.813350 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Oct 05 08:31:22 crc kubenswrapper[4846]: I1005 08:31:22.309659 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-0"] Oct 05 08:31:22 crc kubenswrapper[4846]: I1005 08:31:22.408660 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"c2926ee6-0f75-41e1-8140-cd2a2bfbad2b","Type":"ContainerStarted","Data":"0657bbd866be7aa7dede5d239afa1db7e3b21f7c41f1636f31a642ac1bd40abd"} Oct 05 08:31:22 crc kubenswrapper[4846]: I1005 08:31:22.508117 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="05c8753d-a522-467c-9a13-93961e9061c1" path="/var/lib/kubelet/pods/05c8753d-a522-467c-9a13-93961e9061c1/volumes" Oct 05 08:31:22 crc kubenswrapper[4846]: I1005 08:31:22.561310 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Oct 05 08:31:23 crc kubenswrapper[4846]: I1005 08:31:23.421274 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"c2926ee6-0f75-41e1-8140-cd2a2bfbad2b","Type":"ContainerStarted","Data":"9e82b8bc93809bb9c8d3d202901eb37fcf00f5b729cceab353c98b1a159c1c41"} Oct 05 08:31:23 crc kubenswrapper[4846]: I1005 08:31:23.422101 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"c2926ee6-0f75-41e1-8140-cd2a2bfbad2b","Type":"ContainerStarted","Data":"d60c24a4c26607cba6e5d0099d063548df7e59b7f227b9d9bf07cfcec4cfd346"} Oct 05 08:31:24 crc kubenswrapper[4846]: I1005 08:31:24.432878 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"c2926ee6-0f75-41e1-8140-cd2a2bfbad2b","Type":"ContainerStarted","Data":"235ee3e3d676d4fecc6a90ce1a35000bb102f9f7b8a74c021e7aaec7dffb95a6"} Oct 05 08:31:24 crc kubenswrapper[4846]: I1005 08:31:24.433540 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"c2926ee6-0f75-41e1-8140-cd2a2bfbad2b","Type":"ContainerStarted","Data":"59d270936b742ddd2450058e1447885cc343af7a0c5d30b1c7a24735f1e05cac"} Oct 05 08:31:24 crc kubenswrapper[4846]: I1005 08:31:24.470443 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/aodh-0" podStartSLOduration=2.057795304 podStartE2EDuration="3.470416816s" podCreationTimestamp="2025-10-05 08:31:21 +0000 UTC" firstStartedPulling="2025-10-05 08:31:22.329298288 +0000 UTC m=+6204.570151063" lastFinishedPulling="2025-10-05 08:31:23.7419198 +0000 UTC m=+6205.982772575" observedRunningTime="2025-10-05 08:31:24.454790181 +0000 UTC m=+6206.695642966" watchObservedRunningTime="2025-10-05 08:31:24.470416816 +0000 UTC m=+6206.711269591" Oct 05 08:31:29 crc kubenswrapper[4846]: I1005 08:31:29.736786 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-84596854c9-s7r6q"] Oct 05 08:31:29 crc kubenswrapper[4846]: I1005 08:31:29.738944 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-84596854c9-s7r6q" Oct 05 08:31:29 crc kubenswrapper[4846]: I1005 08:31:29.741547 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1" Oct 05 08:31:29 crc kubenswrapper[4846]: I1005 08:31:29.758210 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-84596854c9-s7r6q"] Oct 05 08:31:29 crc kubenswrapper[4846]: I1005 08:31:29.874607 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/28de9f78-4022-487b-a1e5-f2d54c294a6f-ovsdbserver-nb\") pod \"dnsmasq-dns-84596854c9-s7r6q\" (UID: \"28de9f78-4022-487b-a1e5-f2d54c294a6f\") " pod="openstack/dnsmasq-dns-84596854c9-s7r6q" Oct 05 08:31:29 crc kubenswrapper[4846]: I1005 08:31:29.874649 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/28de9f78-4022-487b-a1e5-f2d54c294a6f-ovsdbserver-sb\") pod \"dnsmasq-dns-84596854c9-s7r6q\" (UID: \"28de9f78-4022-487b-a1e5-f2d54c294a6f\") " pod="openstack/dnsmasq-dns-84596854c9-s7r6q" Oct 05 08:31:29 crc kubenswrapper[4846]: I1005 08:31:29.874692 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4zjdx\" (UniqueName: \"kubernetes.io/projected/28de9f78-4022-487b-a1e5-f2d54c294a6f-kube-api-access-4zjdx\") pod \"dnsmasq-dns-84596854c9-s7r6q\" (UID: \"28de9f78-4022-487b-a1e5-f2d54c294a6f\") " pod="openstack/dnsmasq-dns-84596854c9-s7r6q" Oct 05 08:31:29 crc kubenswrapper[4846]: I1005 08:31:29.874829 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/28de9f78-4022-487b-a1e5-f2d54c294a6f-dns-svc\") pod \"dnsmasq-dns-84596854c9-s7r6q\" (UID: \"28de9f78-4022-487b-a1e5-f2d54c294a6f\") " pod="openstack/dnsmasq-dns-84596854c9-s7r6q" Oct 05 08:31:29 crc kubenswrapper[4846]: I1005 08:31:29.874871 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/28de9f78-4022-487b-a1e5-f2d54c294a6f-openstack-cell1\") pod \"dnsmasq-dns-84596854c9-s7r6q\" (UID: \"28de9f78-4022-487b-a1e5-f2d54c294a6f\") " pod="openstack/dnsmasq-dns-84596854c9-s7r6q" Oct 05 08:31:29 crc kubenswrapper[4846]: I1005 08:31:29.874944 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/28de9f78-4022-487b-a1e5-f2d54c294a6f-config\") pod \"dnsmasq-dns-84596854c9-s7r6q\" (UID: \"28de9f78-4022-487b-a1e5-f2d54c294a6f\") " pod="openstack/dnsmasq-dns-84596854c9-s7r6q" Oct 05 08:31:29 crc kubenswrapper[4846]: I1005 08:31:29.977441 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/28de9f78-4022-487b-a1e5-f2d54c294a6f-ovsdbserver-nb\") pod \"dnsmasq-dns-84596854c9-s7r6q\" (UID: \"28de9f78-4022-487b-a1e5-f2d54c294a6f\") " pod="openstack/dnsmasq-dns-84596854c9-s7r6q" Oct 05 08:31:29 crc kubenswrapper[4846]: I1005 08:31:29.977724 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/28de9f78-4022-487b-a1e5-f2d54c294a6f-ovsdbserver-sb\") pod \"dnsmasq-dns-84596854c9-s7r6q\" (UID: \"28de9f78-4022-487b-a1e5-f2d54c294a6f\") " pod="openstack/dnsmasq-dns-84596854c9-s7r6q" Oct 05 08:31:29 crc kubenswrapper[4846]: I1005 08:31:29.977757 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4zjdx\" (UniqueName: \"kubernetes.io/projected/28de9f78-4022-487b-a1e5-f2d54c294a6f-kube-api-access-4zjdx\") pod \"dnsmasq-dns-84596854c9-s7r6q\" (UID: \"28de9f78-4022-487b-a1e5-f2d54c294a6f\") " pod="openstack/dnsmasq-dns-84596854c9-s7r6q" Oct 05 08:31:29 crc kubenswrapper[4846]: I1005 08:31:29.977832 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/28de9f78-4022-487b-a1e5-f2d54c294a6f-dns-svc\") pod \"dnsmasq-dns-84596854c9-s7r6q\" (UID: \"28de9f78-4022-487b-a1e5-f2d54c294a6f\") " pod="openstack/dnsmasq-dns-84596854c9-s7r6q" Oct 05 08:31:29 crc kubenswrapper[4846]: I1005 08:31:29.977857 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/28de9f78-4022-487b-a1e5-f2d54c294a6f-openstack-cell1\") pod \"dnsmasq-dns-84596854c9-s7r6q\" (UID: \"28de9f78-4022-487b-a1e5-f2d54c294a6f\") " pod="openstack/dnsmasq-dns-84596854c9-s7r6q" Oct 05 08:31:29 crc kubenswrapper[4846]: I1005 08:31:29.977897 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/28de9f78-4022-487b-a1e5-f2d54c294a6f-config\") pod \"dnsmasq-dns-84596854c9-s7r6q\" (UID: \"28de9f78-4022-487b-a1e5-f2d54c294a6f\") " pod="openstack/dnsmasq-dns-84596854c9-s7r6q" Oct 05 08:31:29 crc kubenswrapper[4846]: I1005 08:31:29.978346 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/28de9f78-4022-487b-a1e5-f2d54c294a6f-ovsdbserver-nb\") pod \"dnsmasq-dns-84596854c9-s7r6q\" (UID: \"28de9f78-4022-487b-a1e5-f2d54c294a6f\") " pod="openstack/dnsmasq-dns-84596854c9-s7r6q" Oct 05 08:31:29 crc kubenswrapper[4846]: I1005 08:31:29.979125 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/28de9f78-4022-487b-a1e5-f2d54c294a6f-openstack-cell1\") pod \"dnsmasq-dns-84596854c9-s7r6q\" (UID: \"28de9f78-4022-487b-a1e5-f2d54c294a6f\") " pod="openstack/dnsmasq-dns-84596854c9-s7r6q" Oct 05 08:31:29 crc kubenswrapper[4846]: I1005 08:31:29.979141 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/28de9f78-4022-487b-a1e5-f2d54c294a6f-ovsdbserver-sb\") pod \"dnsmasq-dns-84596854c9-s7r6q\" (UID: \"28de9f78-4022-487b-a1e5-f2d54c294a6f\") " pod="openstack/dnsmasq-dns-84596854c9-s7r6q" Oct 05 08:31:29 crc kubenswrapper[4846]: I1005 08:31:29.979251 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/28de9f78-4022-487b-a1e5-f2d54c294a6f-config\") pod \"dnsmasq-dns-84596854c9-s7r6q\" (UID: \"28de9f78-4022-487b-a1e5-f2d54c294a6f\") " pod="openstack/dnsmasq-dns-84596854c9-s7r6q" Oct 05 08:31:29 crc kubenswrapper[4846]: I1005 08:31:29.979451 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/28de9f78-4022-487b-a1e5-f2d54c294a6f-dns-svc\") pod \"dnsmasq-dns-84596854c9-s7r6q\" (UID: \"28de9f78-4022-487b-a1e5-f2d54c294a6f\") " pod="openstack/dnsmasq-dns-84596854c9-s7r6q" Oct 05 08:31:30 crc kubenswrapper[4846]: I1005 08:31:30.001316 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4zjdx\" (UniqueName: \"kubernetes.io/projected/28de9f78-4022-487b-a1e5-f2d54c294a6f-kube-api-access-4zjdx\") pod \"dnsmasq-dns-84596854c9-s7r6q\" (UID: \"28de9f78-4022-487b-a1e5-f2d54c294a6f\") " pod="openstack/dnsmasq-dns-84596854c9-s7r6q" Oct 05 08:31:30 crc kubenswrapper[4846]: I1005 08:31:30.056456 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-84596854c9-s7r6q" Oct 05 08:31:30 crc kubenswrapper[4846]: I1005 08:31:30.608564 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-84596854c9-s7r6q"] Oct 05 08:31:30 crc kubenswrapper[4846]: W1005 08:31:30.614744 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod28de9f78_4022_487b_a1e5_f2d54c294a6f.slice/crio-6b0478157a94577a4b5ebb528d982d43b83c606f7e09e877f3fb478f02428d2c WatchSource:0}: Error finding container 6b0478157a94577a4b5ebb528d982d43b83c606f7e09e877f3fb478f02428d2c: Status 404 returned error can't find the container with id 6b0478157a94577a4b5ebb528d982d43b83c606f7e09e877f3fb478f02428d2c Oct 05 08:31:31 crc kubenswrapper[4846]: I1005 08:31:31.512044 4846 generic.go:334] "Generic (PLEG): container finished" podID="28de9f78-4022-487b-a1e5-f2d54c294a6f" containerID="6a560279a860c4d2c7239e7931b8b2f4630250e19746c6a388beec17e8e3a6ff" exitCode=0 Oct 05 08:31:31 crc kubenswrapper[4846]: I1005 08:31:31.512246 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-84596854c9-s7r6q" event={"ID":"28de9f78-4022-487b-a1e5-f2d54c294a6f","Type":"ContainerDied","Data":"6a560279a860c4d2c7239e7931b8b2f4630250e19746c6a388beec17e8e3a6ff"} Oct 05 08:31:31 crc kubenswrapper[4846]: I1005 08:31:31.512743 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-84596854c9-s7r6q" event={"ID":"28de9f78-4022-487b-a1e5-f2d54c294a6f","Type":"ContainerStarted","Data":"6b0478157a94577a4b5ebb528d982d43b83c606f7e09e877f3fb478f02428d2c"} Oct 05 08:31:32 crc kubenswrapper[4846]: I1005 08:31:32.528903 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-84596854c9-s7r6q" event={"ID":"28de9f78-4022-487b-a1e5-f2d54c294a6f","Type":"ContainerStarted","Data":"acf0d070b927f3eb0bf7c28c76861fd6b2bdd37cd1ba7cc79709f44218842f8d"} Oct 05 08:31:32 crc kubenswrapper[4846]: I1005 08:31:32.529251 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-84596854c9-s7r6q" Oct 05 08:31:32 crc kubenswrapper[4846]: I1005 08:31:32.568575 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-84596854c9-s7r6q" podStartSLOduration=3.568536819 podStartE2EDuration="3.568536819s" podCreationTimestamp="2025-10-05 08:31:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 08:31:32.555328049 +0000 UTC m=+6214.796180884" watchObservedRunningTime="2025-10-05 08:31:32.568536819 +0000 UTC m=+6214.809389634" Oct 05 08:31:35 crc kubenswrapper[4846]: I1005 08:31:35.187652 4846 scope.go:117] "RemoveContainer" containerID="a83b3ccc5c15e129aa70a4da820d9dfab416efa3f815e11f8169a82b8b16290c" Oct 05 08:31:35 crc kubenswrapper[4846]: I1005 08:31:35.231220 4846 scope.go:117] "RemoveContainer" containerID="329b7338c5617bcc40d190152a65d70ede4f8a7fe7014f8dec828b9861bbf7fc" Oct 05 08:31:35 crc kubenswrapper[4846]: I1005 08:31:35.329667 4846 scope.go:117] "RemoveContainer" containerID="1bb972c8ef5b143675bd5127e7485b4a1e4461c95e56cabc05b7cc4303674150" Oct 05 08:31:40 crc kubenswrapper[4846]: I1005 08:31:40.059252 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-84596854c9-s7r6q" Oct 05 08:31:40 crc kubenswrapper[4846]: I1005 08:31:40.181897 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-75b55fd8f5-x2t6q"] Oct 05 08:31:40 crc kubenswrapper[4846]: I1005 08:31:40.182214 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-75b55fd8f5-x2t6q" podUID="b0deda53-5806-4d7a-b5a6-0796d87a73e3" containerName="dnsmasq-dns" containerID="cri-o://246752f532c50fb2a246b0125c3f3caa7558862206b89fd92ac714fe9b916ab2" gracePeriod=10 Oct 05 08:31:40 crc kubenswrapper[4846]: I1005 08:31:40.350309 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-8869d4b55-9z82r"] Oct 05 08:31:40 crc kubenswrapper[4846]: I1005 08:31:40.352069 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8869d4b55-9z82r" Oct 05 08:31:40 crc kubenswrapper[4846]: I1005 08:31:40.361902 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-8869d4b55-9z82r"] Oct 05 08:31:40 crc kubenswrapper[4846]: I1005 08:31:40.454558 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a17b7828-72b0-4715-9aa6-484f29bc3125-ovsdbserver-sb\") pod \"dnsmasq-dns-8869d4b55-9z82r\" (UID: \"a17b7828-72b0-4715-9aa6-484f29bc3125\") " pod="openstack/dnsmasq-dns-8869d4b55-9z82r" Oct 05 08:31:40 crc kubenswrapper[4846]: I1005 08:31:40.454606 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a17b7828-72b0-4715-9aa6-484f29bc3125-dns-svc\") pod \"dnsmasq-dns-8869d4b55-9z82r\" (UID: \"a17b7828-72b0-4715-9aa6-484f29bc3125\") " pod="openstack/dnsmasq-dns-8869d4b55-9z82r" Oct 05 08:31:40 crc kubenswrapper[4846]: I1005 08:31:40.454651 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/a17b7828-72b0-4715-9aa6-484f29bc3125-openstack-cell1\") pod \"dnsmasq-dns-8869d4b55-9z82r\" (UID: \"a17b7828-72b0-4715-9aa6-484f29bc3125\") " pod="openstack/dnsmasq-dns-8869d4b55-9z82r" Oct 05 08:31:40 crc kubenswrapper[4846]: I1005 08:31:40.454693 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a17b7828-72b0-4715-9aa6-484f29bc3125-config\") pod \"dnsmasq-dns-8869d4b55-9z82r\" (UID: \"a17b7828-72b0-4715-9aa6-484f29bc3125\") " pod="openstack/dnsmasq-dns-8869d4b55-9z82r" Oct 05 08:31:40 crc kubenswrapper[4846]: I1005 08:31:40.454708 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7v7vx\" (UniqueName: \"kubernetes.io/projected/a17b7828-72b0-4715-9aa6-484f29bc3125-kube-api-access-7v7vx\") pod \"dnsmasq-dns-8869d4b55-9z82r\" (UID: \"a17b7828-72b0-4715-9aa6-484f29bc3125\") " pod="openstack/dnsmasq-dns-8869d4b55-9z82r" Oct 05 08:31:40 crc kubenswrapper[4846]: I1005 08:31:40.454745 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a17b7828-72b0-4715-9aa6-484f29bc3125-ovsdbserver-nb\") pod \"dnsmasq-dns-8869d4b55-9z82r\" (UID: \"a17b7828-72b0-4715-9aa6-484f29bc3125\") " pod="openstack/dnsmasq-dns-8869d4b55-9z82r" Oct 05 08:31:40 crc kubenswrapper[4846]: I1005 08:31:40.557319 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a17b7828-72b0-4715-9aa6-484f29bc3125-ovsdbserver-sb\") pod \"dnsmasq-dns-8869d4b55-9z82r\" (UID: \"a17b7828-72b0-4715-9aa6-484f29bc3125\") " pod="openstack/dnsmasq-dns-8869d4b55-9z82r" Oct 05 08:31:40 crc kubenswrapper[4846]: I1005 08:31:40.557640 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a17b7828-72b0-4715-9aa6-484f29bc3125-dns-svc\") pod \"dnsmasq-dns-8869d4b55-9z82r\" (UID: \"a17b7828-72b0-4715-9aa6-484f29bc3125\") " pod="openstack/dnsmasq-dns-8869d4b55-9z82r" Oct 05 08:31:40 crc kubenswrapper[4846]: I1005 08:31:40.557692 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/a17b7828-72b0-4715-9aa6-484f29bc3125-openstack-cell1\") pod \"dnsmasq-dns-8869d4b55-9z82r\" (UID: \"a17b7828-72b0-4715-9aa6-484f29bc3125\") " pod="openstack/dnsmasq-dns-8869d4b55-9z82r" Oct 05 08:31:40 crc kubenswrapper[4846]: I1005 08:31:40.557730 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a17b7828-72b0-4715-9aa6-484f29bc3125-config\") pod \"dnsmasq-dns-8869d4b55-9z82r\" (UID: \"a17b7828-72b0-4715-9aa6-484f29bc3125\") " pod="openstack/dnsmasq-dns-8869d4b55-9z82r" Oct 05 08:31:40 crc kubenswrapper[4846]: I1005 08:31:40.557750 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7v7vx\" (UniqueName: \"kubernetes.io/projected/a17b7828-72b0-4715-9aa6-484f29bc3125-kube-api-access-7v7vx\") pod \"dnsmasq-dns-8869d4b55-9z82r\" (UID: \"a17b7828-72b0-4715-9aa6-484f29bc3125\") " pod="openstack/dnsmasq-dns-8869d4b55-9z82r" Oct 05 08:31:40 crc kubenswrapper[4846]: I1005 08:31:40.557793 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a17b7828-72b0-4715-9aa6-484f29bc3125-ovsdbserver-nb\") pod \"dnsmasq-dns-8869d4b55-9z82r\" (UID: \"a17b7828-72b0-4715-9aa6-484f29bc3125\") " pod="openstack/dnsmasq-dns-8869d4b55-9z82r" Oct 05 08:31:40 crc kubenswrapper[4846]: I1005 08:31:40.561479 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a17b7828-72b0-4715-9aa6-484f29bc3125-ovsdbserver-sb\") pod \"dnsmasq-dns-8869d4b55-9z82r\" (UID: \"a17b7828-72b0-4715-9aa6-484f29bc3125\") " pod="openstack/dnsmasq-dns-8869d4b55-9z82r" Oct 05 08:31:40 crc kubenswrapper[4846]: I1005 08:31:40.562381 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a17b7828-72b0-4715-9aa6-484f29bc3125-config\") pod \"dnsmasq-dns-8869d4b55-9z82r\" (UID: \"a17b7828-72b0-4715-9aa6-484f29bc3125\") " pod="openstack/dnsmasq-dns-8869d4b55-9z82r" Oct 05 08:31:40 crc kubenswrapper[4846]: I1005 08:31:40.562530 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/a17b7828-72b0-4715-9aa6-484f29bc3125-openstack-cell1\") pod \"dnsmasq-dns-8869d4b55-9z82r\" (UID: \"a17b7828-72b0-4715-9aa6-484f29bc3125\") " pod="openstack/dnsmasq-dns-8869d4b55-9z82r" Oct 05 08:31:40 crc kubenswrapper[4846]: I1005 08:31:40.572442 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a17b7828-72b0-4715-9aa6-484f29bc3125-ovsdbserver-nb\") pod \"dnsmasq-dns-8869d4b55-9z82r\" (UID: \"a17b7828-72b0-4715-9aa6-484f29bc3125\") " pod="openstack/dnsmasq-dns-8869d4b55-9z82r" Oct 05 08:31:40 crc kubenswrapper[4846]: I1005 08:31:40.573999 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a17b7828-72b0-4715-9aa6-484f29bc3125-dns-svc\") pod \"dnsmasq-dns-8869d4b55-9z82r\" (UID: \"a17b7828-72b0-4715-9aa6-484f29bc3125\") " pod="openstack/dnsmasq-dns-8869d4b55-9z82r" Oct 05 08:31:40 crc kubenswrapper[4846]: I1005 08:31:40.587795 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7v7vx\" (UniqueName: \"kubernetes.io/projected/a17b7828-72b0-4715-9aa6-484f29bc3125-kube-api-access-7v7vx\") pod \"dnsmasq-dns-8869d4b55-9z82r\" (UID: \"a17b7828-72b0-4715-9aa6-484f29bc3125\") " pod="openstack/dnsmasq-dns-8869d4b55-9z82r" Oct 05 08:31:40 crc kubenswrapper[4846]: I1005 08:31:40.637777 4846 generic.go:334] "Generic (PLEG): container finished" podID="b0deda53-5806-4d7a-b5a6-0796d87a73e3" containerID="246752f532c50fb2a246b0125c3f3caa7558862206b89fd92ac714fe9b916ab2" exitCode=0 Oct 05 08:31:40 crc kubenswrapper[4846]: I1005 08:31:40.637826 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-75b55fd8f5-x2t6q" event={"ID":"b0deda53-5806-4d7a-b5a6-0796d87a73e3","Type":"ContainerDied","Data":"246752f532c50fb2a246b0125c3f3caa7558862206b89fd92ac714fe9b916ab2"} Oct 05 08:31:40 crc kubenswrapper[4846]: I1005 08:31:40.685919 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8869d4b55-9z82r" Oct 05 08:31:40 crc kubenswrapper[4846]: I1005 08:31:40.836168 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-75b55fd8f5-x2t6q" Oct 05 08:31:40 crc kubenswrapper[4846]: I1005 08:31:40.967046 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fgjtm\" (UniqueName: \"kubernetes.io/projected/b0deda53-5806-4d7a-b5a6-0796d87a73e3-kube-api-access-fgjtm\") pod \"b0deda53-5806-4d7a-b5a6-0796d87a73e3\" (UID: \"b0deda53-5806-4d7a-b5a6-0796d87a73e3\") " Oct 05 08:31:40 crc kubenswrapper[4846]: I1005 08:31:40.967445 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b0deda53-5806-4d7a-b5a6-0796d87a73e3-dns-svc\") pod \"b0deda53-5806-4d7a-b5a6-0796d87a73e3\" (UID: \"b0deda53-5806-4d7a-b5a6-0796d87a73e3\") " Oct 05 08:31:40 crc kubenswrapper[4846]: I1005 08:31:40.967594 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b0deda53-5806-4d7a-b5a6-0796d87a73e3-ovsdbserver-nb\") pod \"b0deda53-5806-4d7a-b5a6-0796d87a73e3\" (UID: \"b0deda53-5806-4d7a-b5a6-0796d87a73e3\") " Oct 05 08:31:40 crc kubenswrapper[4846]: I1005 08:31:40.969497 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b0deda53-5806-4d7a-b5a6-0796d87a73e3-config\") pod \"b0deda53-5806-4d7a-b5a6-0796d87a73e3\" (UID: \"b0deda53-5806-4d7a-b5a6-0796d87a73e3\") " Oct 05 08:31:40 crc kubenswrapper[4846]: I1005 08:31:40.969532 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b0deda53-5806-4d7a-b5a6-0796d87a73e3-ovsdbserver-sb\") pod \"b0deda53-5806-4d7a-b5a6-0796d87a73e3\" (UID: \"b0deda53-5806-4d7a-b5a6-0796d87a73e3\") " Oct 05 08:31:41 crc kubenswrapper[4846]: I1005 08:31:40.975958 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b0deda53-5806-4d7a-b5a6-0796d87a73e3-kube-api-access-fgjtm" (OuterVolumeSpecName: "kube-api-access-fgjtm") pod "b0deda53-5806-4d7a-b5a6-0796d87a73e3" (UID: "b0deda53-5806-4d7a-b5a6-0796d87a73e3"). InnerVolumeSpecName "kube-api-access-fgjtm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:31:41 crc kubenswrapper[4846]: I1005 08:31:41.038821 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b0deda53-5806-4d7a-b5a6-0796d87a73e3-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "b0deda53-5806-4d7a-b5a6-0796d87a73e3" (UID: "b0deda53-5806-4d7a-b5a6-0796d87a73e3"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 08:31:41 crc kubenswrapper[4846]: I1005 08:31:41.044723 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b0deda53-5806-4d7a-b5a6-0796d87a73e3-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "b0deda53-5806-4d7a-b5a6-0796d87a73e3" (UID: "b0deda53-5806-4d7a-b5a6-0796d87a73e3"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 08:31:41 crc kubenswrapper[4846]: I1005 08:31:41.056587 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b0deda53-5806-4d7a-b5a6-0796d87a73e3-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "b0deda53-5806-4d7a-b5a6-0796d87a73e3" (UID: "b0deda53-5806-4d7a-b5a6-0796d87a73e3"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 08:31:41 crc kubenswrapper[4846]: I1005 08:31:41.059869 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b0deda53-5806-4d7a-b5a6-0796d87a73e3-config" (OuterVolumeSpecName: "config") pod "b0deda53-5806-4d7a-b5a6-0796d87a73e3" (UID: "b0deda53-5806-4d7a-b5a6-0796d87a73e3"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 08:31:41 crc kubenswrapper[4846]: I1005 08:31:41.072804 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fgjtm\" (UniqueName: \"kubernetes.io/projected/b0deda53-5806-4d7a-b5a6-0796d87a73e3-kube-api-access-fgjtm\") on node \"crc\" DevicePath \"\"" Oct 05 08:31:41 crc kubenswrapper[4846]: I1005 08:31:41.072907 4846 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b0deda53-5806-4d7a-b5a6-0796d87a73e3-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 05 08:31:41 crc kubenswrapper[4846]: I1005 08:31:41.073074 4846 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b0deda53-5806-4d7a-b5a6-0796d87a73e3-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 05 08:31:41 crc kubenswrapper[4846]: I1005 08:31:41.073146 4846 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b0deda53-5806-4d7a-b5a6-0796d87a73e3-config\") on node \"crc\" DevicePath \"\"" Oct 05 08:31:41 crc kubenswrapper[4846]: I1005 08:31:41.073232 4846 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b0deda53-5806-4d7a-b5a6-0796d87a73e3-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 05 08:31:41 crc kubenswrapper[4846]: I1005 08:31:41.267388 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-8869d4b55-9z82r"] Oct 05 08:31:41 crc kubenswrapper[4846]: I1005 08:31:41.648538 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-75b55fd8f5-x2t6q" event={"ID":"b0deda53-5806-4d7a-b5a6-0796d87a73e3","Type":"ContainerDied","Data":"23e29b503c3189c910e436dde2125f0b73676ee92e98d1a0b876d441ba20ea4d"} Oct 05 08:31:41 crc kubenswrapper[4846]: I1005 08:31:41.648862 4846 scope.go:117] "RemoveContainer" containerID="246752f532c50fb2a246b0125c3f3caa7558862206b89fd92ac714fe9b916ab2" Oct 05 08:31:41 crc kubenswrapper[4846]: I1005 08:31:41.648588 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-75b55fd8f5-x2t6q" Oct 05 08:31:41 crc kubenswrapper[4846]: I1005 08:31:41.650497 4846 generic.go:334] "Generic (PLEG): container finished" podID="a17b7828-72b0-4715-9aa6-484f29bc3125" containerID="f3eca9e23e2eaab6ec1ca7d55e4e8e8ab33d98dd72d46206e11fcf3e17959583" exitCode=0 Oct 05 08:31:41 crc kubenswrapper[4846]: I1005 08:31:41.650552 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8869d4b55-9z82r" event={"ID":"a17b7828-72b0-4715-9aa6-484f29bc3125","Type":"ContainerDied","Data":"f3eca9e23e2eaab6ec1ca7d55e4e8e8ab33d98dd72d46206e11fcf3e17959583"} Oct 05 08:31:41 crc kubenswrapper[4846]: I1005 08:31:41.650592 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8869d4b55-9z82r" event={"ID":"a17b7828-72b0-4715-9aa6-484f29bc3125","Type":"ContainerStarted","Data":"6564550dbfb6923da80da2dacdc5815ddb43e9628f614eea13ea222447cd4a08"} Oct 05 08:31:41 crc kubenswrapper[4846]: I1005 08:31:41.684430 4846 scope.go:117] "RemoveContainer" containerID="ed5b9f4aa40a4911065311eb03e48f6bfa47c1f5a7e4b941da19575205491ce8" Oct 05 08:31:41 crc kubenswrapper[4846]: I1005 08:31:41.866724 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-75b55fd8f5-x2t6q"] Oct 05 08:31:41 crc kubenswrapper[4846]: I1005 08:31:41.876609 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-75b55fd8f5-x2t6q"] Oct 05 08:31:42 crc kubenswrapper[4846]: I1005 08:31:42.508862 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b0deda53-5806-4d7a-b5a6-0796d87a73e3" path="/var/lib/kubelet/pods/b0deda53-5806-4d7a-b5a6-0796d87a73e3/volumes" Oct 05 08:31:42 crc kubenswrapper[4846]: I1005 08:31:42.661911 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8869d4b55-9z82r" event={"ID":"a17b7828-72b0-4715-9aa6-484f29bc3125","Type":"ContainerStarted","Data":"94eca01c581e37c91d260ae499a88454404994f28fbf0caabe2d3630e0179d79"} Oct 05 08:31:42 crc kubenswrapper[4846]: I1005 08:31:42.662092 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-8869d4b55-9z82r" Oct 05 08:31:42 crc kubenswrapper[4846]: I1005 08:31:42.689056 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-8869d4b55-9z82r" podStartSLOduration=2.689027887 podStartE2EDuration="2.689027887s" podCreationTimestamp="2025-10-05 08:31:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 08:31:42.678430248 +0000 UTC m=+6224.919283023" watchObservedRunningTime="2025-10-05 08:31:42.689027887 +0000 UTC m=+6224.929880692" Oct 05 08:31:44 crc kubenswrapper[4846]: I1005 08:31:44.056598 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-create-ft5wq"] Oct 05 08:31:44 crc kubenswrapper[4846]: I1005 08:31:44.073799 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-create-ft5wq"] Oct 05 08:31:44 crc kubenswrapper[4846]: I1005 08:31:44.509845 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c8c439e7-0384-49d4-bdbf-f2fd0cc28118" path="/var/lib/kubelet/pods/c8c439e7-0384-49d4-bdbf-f2fd0cc28118/volumes" Oct 05 08:31:50 crc kubenswrapper[4846]: I1005 08:31:50.688510 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-8869d4b55-9z82r" Oct 05 08:31:50 crc kubenswrapper[4846]: I1005 08:31:50.793918 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-84596854c9-s7r6q"] Oct 05 08:31:50 crc kubenswrapper[4846]: I1005 08:31:50.794227 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-84596854c9-s7r6q" podUID="28de9f78-4022-487b-a1e5-f2d54c294a6f" containerName="dnsmasq-dns" containerID="cri-o://acf0d070b927f3eb0bf7c28c76861fd6b2bdd37cd1ba7cc79709f44218842f8d" gracePeriod=10 Oct 05 08:31:51 crc kubenswrapper[4846]: I1005 08:31:51.395170 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-84596854c9-s7r6q" Oct 05 08:31:51 crc kubenswrapper[4846]: I1005 08:31:51.426948 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/28de9f78-4022-487b-a1e5-f2d54c294a6f-ovsdbserver-sb\") pod \"28de9f78-4022-487b-a1e5-f2d54c294a6f\" (UID: \"28de9f78-4022-487b-a1e5-f2d54c294a6f\") " Oct 05 08:31:51 crc kubenswrapper[4846]: I1005 08:31:51.427064 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/28de9f78-4022-487b-a1e5-f2d54c294a6f-ovsdbserver-nb\") pod \"28de9f78-4022-487b-a1e5-f2d54c294a6f\" (UID: \"28de9f78-4022-487b-a1e5-f2d54c294a6f\") " Oct 05 08:31:51 crc kubenswrapper[4846]: I1005 08:31:51.427139 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/28de9f78-4022-487b-a1e5-f2d54c294a6f-dns-svc\") pod \"28de9f78-4022-487b-a1e5-f2d54c294a6f\" (UID: \"28de9f78-4022-487b-a1e5-f2d54c294a6f\") " Oct 05 08:31:51 crc kubenswrapper[4846]: I1005 08:31:51.427307 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/28de9f78-4022-487b-a1e5-f2d54c294a6f-openstack-cell1\") pod \"28de9f78-4022-487b-a1e5-f2d54c294a6f\" (UID: \"28de9f78-4022-487b-a1e5-f2d54c294a6f\") " Oct 05 08:31:51 crc kubenswrapper[4846]: I1005 08:31:51.427335 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4zjdx\" (UniqueName: \"kubernetes.io/projected/28de9f78-4022-487b-a1e5-f2d54c294a6f-kube-api-access-4zjdx\") pod \"28de9f78-4022-487b-a1e5-f2d54c294a6f\" (UID: \"28de9f78-4022-487b-a1e5-f2d54c294a6f\") " Oct 05 08:31:51 crc kubenswrapper[4846]: I1005 08:31:51.427413 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/28de9f78-4022-487b-a1e5-f2d54c294a6f-config\") pod \"28de9f78-4022-487b-a1e5-f2d54c294a6f\" (UID: \"28de9f78-4022-487b-a1e5-f2d54c294a6f\") " Oct 05 08:31:51 crc kubenswrapper[4846]: I1005 08:31:51.433268 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/28de9f78-4022-487b-a1e5-f2d54c294a6f-kube-api-access-4zjdx" (OuterVolumeSpecName: "kube-api-access-4zjdx") pod "28de9f78-4022-487b-a1e5-f2d54c294a6f" (UID: "28de9f78-4022-487b-a1e5-f2d54c294a6f"). InnerVolumeSpecName "kube-api-access-4zjdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:31:51 crc kubenswrapper[4846]: I1005 08:31:51.484020 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/28de9f78-4022-487b-a1e5-f2d54c294a6f-openstack-cell1" (OuterVolumeSpecName: "openstack-cell1") pod "28de9f78-4022-487b-a1e5-f2d54c294a6f" (UID: "28de9f78-4022-487b-a1e5-f2d54c294a6f"). InnerVolumeSpecName "openstack-cell1". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 08:31:51 crc kubenswrapper[4846]: I1005 08:31:51.494060 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/28de9f78-4022-487b-a1e5-f2d54c294a6f-config" (OuterVolumeSpecName: "config") pod "28de9f78-4022-487b-a1e5-f2d54c294a6f" (UID: "28de9f78-4022-487b-a1e5-f2d54c294a6f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 08:31:51 crc kubenswrapper[4846]: I1005 08:31:51.494744 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/28de9f78-4022-487b-a1e5-f2d54c294a6f-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "28de9f78-4022-487b-a1e5-f2d54c294a6f" (UID: "28de9f78-4022-487b-a1e5-f2d54c294a6f"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 08:31:51 crc kubenswrapper[4846]: I1005 08:31:51.500808 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/28de9f78-4022-487b-a1e5-f2d54c294a6f-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "28de9f78-4022-487b-a1e5-f2d54c294a6f" (UID: "28de9f78-4022-487b-a1e5-f2d54c294a6f"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 08:31:51 crc kubenswrapper[4846]: I1005 08:31:51.506276 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/28de9f78-4022-487b-a1e5-f2d54c294a6f-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "28de9f78-4022-487b-a1e5-f2d54c294a6f" (UID: "28de9f78-4022-487b-a1e5-f2d54c294a6f"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 08:31:51 crc kubenswrapper[4846]: I1005 08:31:51.530539 4846 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/28de9f78-4022-487b-a1e5-f2d54c294a6f-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 05 08:31:51 crc kubenswrapper[4846]: I1005 08:31:51.530579 4846 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/28de9f78-4022-487b-a1e5-f2d54c294a6f-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 05 08:31:51 crc kubenswrapper[4846]: I1005 08:31:51.530596 4846 reconciler_common.go:293] "Volume detached for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/28de9f78-4022-487b-a1e5-f2d54c294a6f-openstack-cell1\") on node \"crc\" DevicePath \"\"" Oct 05 08:31:51 crc kubenswrapper[4846]: I1005 08:31:51.530608 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4zjdx\" (UniqueName: \"kubernetes.io/projected/28de9f78-4022-487b-a1e5-f2d54c294a6f-kube-api-access-4zjdx\") on node \"crc\" DevicePath \"\"" Oct 05 08:31:51 crc kubenswrapper[4846]: I1005 08:31:51.530650 4846 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/28de9f78-4022-487b-a1e5-f2d54c294a6f-config\") on node \"crc\" DevicePath \"\"" Oct 05 08:31:51 crc kubenswrapper[4846]: I1005 08:31:51.530665 4846 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/28de9f78-4022-487b-a1e5-f2d54c294a6f-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 05 08:31:51 crc kubenswrapper[4846]: I1005 08:31:51.790968 4846 generic.go:334] "Generic (PLEG): container finished" podID="28de9f78-4022-487b-a1e5-f2d54c294a6f" containerID="acf0d070b927f3eb0bf7c28c76861fd6b2bdd37cd1ba7cc79709f44218842f8d" exitCode=0 Oct 05 08:31:51 crc kubenswrapper[4846]: I1005 08:31:51.791022 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-84596854c9-s7r6q" Oct 05 08:31:51 crc kubenswrapper[4846]: I1005 08:31:51.791041 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-84596854c9-s7r6q" event={"ID":"28de9f78-4022-487b-a1e5-f2d54c294a6f","Type":"ContainerDied","Data":"acf0d070b927f3eb0bf7c28c76861fd6b2bdd37cd1ba7cc79709f44218842f8d"} Oct 05 08:31:51 crc kubenswrapper[4846]: I1005 08:31:51.791447 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-84596854c9-s7r6q" event={"ID":"28de9f78-4022-487b-a1e5-f2d54c294a6f","Type":"ContainerDied","Data":"6b0478157a94577a4b5ebb528d982d43b83c606f7e09e877f3fb478f02428d2c"} Oct 05 08:31:51 crc kubenswrapper[4846]: I1005 08:31:51.791469 4846 scope.go:117] "RemoveContainer" containerID="acf0d070b927f3eb0bf7c28c76861fd6b2bdd37cd1ba7cc79709f44218842f8d" Oct 05 08:31:51 crc kubenswrapper[4846]: I1005 08:31:51.829707 4846 scope.go:117] "RemoveContainer" containerID="6a560279a860c4d2c7239e7931b8b2f4630250e19746c6a388beec17e8e3a6ff" Oct 05 08:31:51 crc kubenswrapper[4846]: I1005 08:31:51.832429 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-84596854c9-s7r6q"] Oct 05 08:31:51 crc kubenswrapper[4846]: I1005 08:31:51.848248 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-84596854c9-s7r6q"] Oct 05 08:31:51 crc kubenswrapper[4846]: I1005 08:31:51.862026 4846 scope.go:117] "RemoveContainer" containerID="acf0d070b927f3eb0bf7c28c76861fd6b2bdd37cd1ba7cc79709f44218842f8d" Oct 05 08:31:51 crc kubenswrapper[4846]: E1005 08:31:51.862716 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"acf0d070b927f3eb0bf7c28c76861fd6b2bdd37cd1ba7cc79709f44218842f8d\": container with ID starting with acf0d070b927f3eb0bf7c28c76861fd6b2bdd37cd1ba7cc79709f44218842f8d not found: ID does not exist" containerID="acf0d070b927f3eb0bf7c28c76861fd6b2bdd37cd1ba7cc79709f44218842f8d" Oct 05 08:31:51 crc kubenswrapper[4846]: I1005 08:31:51.862792 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"acf0d070b927f3eb0bf7c28c76861fd6b2bdd37cd1ba7cc79709f44218842f8d"} err="failed to get container status \"acf0d070b927f3eb0bf7c28c76861fd6b2bdd37cd1ba7cc79709f44218842f8d\": rpc error: code = NotFound desc = could not find container \"acf0d070b927f3eb0bf7c28c76861fd6b2bdd37cd1ba7cc79709f44218842f8d\": container with ID starting with acf0d070b927f3eb0bf7c28c76861fd6b2bdd37cd1ba7cc79709f44218842f8d not found: ID does not exist" Oct 05 08:31:51 crc kubenswrapper[4846]: I1005 08:31:51.862836 4846 scope.go:117] "RemoveContainer" containerID="6a560279a860c4d2c7239e7931b8b2f4630250e19746c6a388beec17e8e3a6ff" Oct 05 08:31:51 crc kubenswrapper[4846]: E1005 08:31:51.863542 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6a560279a860c4d2c7239e7931b8b2f4630250e19746c6a388beec17e8e3a6ff\": container with ID starting with 6a560279a860c4d2c7239e7931b8b2f4630250e19746c6a388beec17e8e3a6ff not found: ID does not exist" containerID="6a560279a860c4d2c7239e7931b8b2f4630250e19746c6a388beec17e8e3a6ff" Oct 05 08:31:51 crc kubenswrapper[4846]: I1005 08:31:51.863611 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6a560279a860c4d2c7239e7931b8b2f4630250e19746c6a388beec17e8e3a6ff"} err="failed to get container status \"6a560279a860c4d2c7239e7931b8b2f4630250e19746c6a388beec17e8e3a6ff\": rpc error: code = NotFound desc = could not find container \"6a560279a860c4d2c7239e7931b8b2f4630250e19746c6a388beec17e8e3a6ff\": container with ID starting with 6a560279a860c4d2c7239e7931b8b2f4630250e19746c6a388beec17e8e3a6ff not found: ID does not exist" Oct 05 08:31:52 crc kubenswrapper[4846]: I1005 08:31:52.512036 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="28de9f78-4022-487b-a1e5-f2d54c294a6f" path="/var/lib/kubelet/pods/28de9f78-4022-487b-a1e5-f2d54c294a6f/volumes" Oct 05 08:31:54 crc kubenswrapper[4846]: I1005 08:31:54.067990 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-c6ce-account-create-26v82"] Oct 05 08:31:54 crc kubenswrapper[4846]: I1005 08:31:54.080685 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-c6ce-account-create-26v82"] Oct 05 08:31:54 crc kubenswrapper[4846]: I1005 08:31:54.514294 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5aed0561-41c0-47d2-95c2-a23f79b6f8f8" path="/var/lib/kubelet/pods/5aed0561-41c0-47d2-95c2-a23f79b6f8f8/volumes" Oct 05 08:32:01 crc kubenswrapper[4846]: I1005 08:32:01.900169 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/pre-adoption-validation-openstack-pre-adoption-openstack-c672lz"] Oct 05 08:32:01 crc kubenswrapper[4846]: E1005 08:32:01.901430 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="28de9f78-4022-487b-a1e5-f2d54c294a6f" containerName="dnsmasq-dns" Oct 05 08:32:01 crc kubenswrapper[4846]: I1005 08:32:01.901451 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="28de9f78-4022-487b-a1e5-f2d54c294a6f" containerName="dnsmasq-dns" Oct 05 08:32:01 crc kubenswrapper[4846]: E1005 08:32:01.901482 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b0deda53-5806-4d7a-b5a6-0796d87a73e3" containerName="dnsmasq-dns" Oct 05 08:32:01 crc kubenswrapper[4846]: I1005 08:32:01.901495 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="b0deda53-5806-4d7a-b5a6-0796d87a73e3" containerName="dnsmasq-dns" Oct 05 08:32:01 crc kubenswrapper[4846]: E1005 08:32:01.901533 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b0deda53-5806-4d7a-b5a6-0796d87a73e3" containerName="init" Oct 05 08:32:01 crc kubenswrapper[4846]: I1005 08:32:01.901547 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="b0deda53-5806-4d7a-b5a6-0796d87a73e3" containerName="init" Oct 05 08:32:01 crc kubenswrapper[4846]: E1005 08:32:01.901588 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="28de9f78-4022-487b-a1e5-f2d54c294a6f" containerName="init" Oct 05 08:32:01 crc kubenswrapper[4846]: I1005 08:32:01.901600 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="28de9f78-4022-487b-a1e5-f2d54c294a6f" containerName="init" Oct 05 08:32:01 crc kubenswrapper[4846]: I1005 08:32:01.901921 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="b0deda53-5806-4d7a-b5a6-0796d87a73e3" containerName="dnsmasq-dns" Oct 05 08:32:01 crc kubenswrapper[4846]: I1005 08:32:01.901970 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="28de9f78-4022-487b-a1e5-f2d54c294a6f" containerName="dnsmasq-dns" Oct 05 08:32:01 crc kubenswrapper[4846]: I1005 08:32:01.903372 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-c672lz" Oct 05 08:32:01 crc kubenswrapper[4846]: I1005 08:32:01.907905 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Oct 05 08:32:01 crc kubenswrapper[4846]: I1005 08:32:01.908140 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-7687s" Oct 05 08:32:01 crc kubenswrapper[4846]: I1005 08:32:01.909878 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 05 08:32:01 crc kubenswrapper[4846]: I1005 08:32:01.916172 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Oct 05 08:32:01 crc kubenswrapper[4846]: I1005 08:32:01.922926 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/pre-adoption-validation-openstack-pre-adoption-openstack-c672lz"] Oct 05 08:32:01 crc kubenswrapper[4846]: I1005 08:32:01.985211 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5sznp\" (UniqueName: \"kubernetes.io/projected/4ca8fa61-f7ee-4f10-b047-589a5558c05b-kube-api-access-5sznp\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-c672lz\" (UID: \"4ca8fa61-f7ee-4f10-b047-589a5558c05b\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-c672lz" Oct 05 08:32:01 crc kubenswrapper[4846]: I1005 08:32:01.985623 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pre-adoption-validation-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4ca8fa61-f7ee-4f10-b047-589a5558c05b-pre-adoption-validation-combined-ca-bundle\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-c672lz\" (UID: \"4ca8fa61-f7ee-4f10-b047-589a5558c05b\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-c672lz" Oct 05 08:32:01 crc kubenswrapper[4846]: I1005 08:32:01.985833 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4ca8fa61-f7ee-4f10-b047-589a5558c05b-inventory\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-c672lz\" (UID: \"4ca8fa61-f7ee-4f10-b047-589a5558c05b\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-c672lz" Oct 05 08:32:01 crc kubenswrapper[4846]: I1005 08:32:01.985986 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4ca8fa61-f7ee-4f10-b047-589a5558c05b-ssh-key\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-c672lz\" (UID: \"4ca8fa61-f7ee-4f10-b047-589a5558c05b\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-c672lz" Oct 05 08:32:02 crc kubenswrapper[4846]: I1005 08:32:02.087430 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4ca8fa61-f7ee-4f10-b047-589a5558c05b-inventory\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-c672lz\" (UID: \"4ca8fa61-f7ee-4f10-b047-589a5558c05b\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-c672lz" Oct 05 08:32:02 crc kubenswrapper[4846]: I1005 08:32:02.087526 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4ca8fa61-f7ee-4f10-b047-589a5558c05b-ssh-key\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-c672lz\" (UID: \"4ca8fa61-f7ee-4f10-b047-589a5558c05b\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-c672lz" Oct 05 08:32:02 crc kubenswrapper[4846]: I1005 08:32:02.087593 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5sznp\" (UniqueName: \"kubernetes.io/projected/4ca8fa61-f7ee-4f10-b047-589a5558c05b-kube-api-access-5sznp\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-c672lz\" (UID: \"4ca8fa61-f7ee-4f10-b047-589a5558c05b\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-c672lz" Oct 05 08:32:02 crc kubenswrapper[4846]: I1005 08:32:02.087661 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pre-adoption-validation-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4ca8fa61-f7ee-4f10-b047-589a5558c05b-pre-adoption-validation-combined-ca-bundle\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-c672lz\" (UID: \"4ca8fa61-f7ee-4f10-b047-589a5558c05b\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-c672lz" Oct 05 08:32:02 crc kubenswrapper[4846]: I1005 08:32:02.095814 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4ca8fa61-f7ee-4f10-b047-589a5558c05b-inventory\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-c672lz\" (UID: \"4ca8fa61-f7ee-4f10-b047-589a5558c05b\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-c672lz" Oct 05 08:32:02 crc kubenswrapper[4846]: I1005 08:32:02.095991 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pre-adoption-validation-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4ca8fa61-f7ee-4f10-b047-589a5558c05b-pre-adoption-validation-combined-ca-bundle\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-c672lz\" (UID: \"4ca8fa61-f7ee-4f10-b047-589a5558c05b\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-c672lz" Oct 05 08:32:02 crc kubenswrapper[4846]: I1005 08:32:02.096305 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4ca8fa61-f7ee-4f10-b047-589a5558c05b-ssh-key\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-c672lz\" (UID: \"4ca8fa61-f7ee-4f10-b047-589a5558c05b\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-c672lz" Oct 05 08:32:02 crc kubenswrapper[4846]: I1005 08:32:02.109428 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5sznp\" (UniqueName: \"kubernetes.io/projected/4ca8fa61-f7ee-4f10-b047-589a5558c05b-kube-api-access-5sznp\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-c672lz\" (UID: \"4ca8fa61-f7ee-4f10-b047-589a5558c05b\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-c672lz" Oct 05 08:32:02 crc kubenswrapper[4846]: I1005 08:32:02.233927 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-c672lz" Oct 05 08:32:02 crc kubenswrapper[4846]: I1005 08:32:02.891706 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/pre-adoption-validation-openstack-pre-adoption-openstack-c672lz"] Oct 05 08:32:02 crc kubenswrapper[4846]: I1005 08:32:02.941854 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-c672lz" event={"ID":"4ca8fa61-f7ee-4f10-b047-589a5558c05b","Type":"ContainerStarted","Data":"891848a05a25d5b1d52f13d8ede63a3f1ec8a3207881fb7f10a4d3170950337c"} Oct 05 08:32:03 crc kubenswrapper[4846]: I1005 08:32:03.035177 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-sync-bsrnx"] Oct 05 08:32:03 crc kubenswrapper[4846]: I1005 08:32:03.045471 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-sync-bsrnx"] Oct 05 08:32:04 crc kubenswrapper[4846]: I1005 08:32:04.510239 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7d61014b-9453-4db8-8ae5-226d01d076a6" path="/var/lib/kubelet/pods/7d61014b-9453-4db8-8ae5-226d01d076a6/volumes" Oct 05 08:32:12 crc kubenswrapper[4846]: I1005 08:32:12.067133 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-c672lz" event={"ID":"4ca8fa61-f7ee-4f10-b047-589a5558c05b","Type":"ContainerStarted","Data":"e0dbecc00872335f728a00b6b8f66f206cae1eacee637ad93f50bc7b8fc0f42d"} Oct 05 08:32:12 crc kubenswrapper[4846]: I1005 08:32:12.104040 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-c672lz" podStartSLOduration=2.715795063 podStartE2EDuration="11.104003479s" podCreationTimestamp="2025-10-05 08:32:01 +0000 UTC" firstStartedPulling="2025-10-05 08:32:02.895163492 +0000 UTC m=+6245.136016267" lastFinishedPulling="2025-10-05 08:32:11.283371908 +0000 UTC m=+6253.524224683" observedRunningTime="2025-10-05 08:32:12.093529203 +0000 UTC m=+6254.334381978" watchObservedRunningTime="2025-10-05 08:32:12.104003479 +0000 UTC m=+6254.344856304" Oct 05 08:32:23 crc kubenswrapper[4846]: I1005 08:32:23.324934 4846 patch_prober.go:28] interesting pod/machine-config-daemon-fscvf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 08:32:23 crc kubenswrapper[4846]: I1005 08:32:23.325642 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 08:32:25 crc kubenswrapper[4846]: I1005 08:32:25.262893 4846 generic.go:334] "Generic (PLEG): container finished" podID="4ca8fa61-f7ee-4f10-b047-589a5558c05b" containerID="e0dbecc00872335f728a00b6b8f66f206cae1eacee637ad93f50bc7b8fc0f42d" exitCode=0 Oct 05 08:32:25 crc kubenswrapper[4846]: I1005 08:32:25.263042 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-c672lz" event={"ID":"4ca8fa61-f7ee-4f10-b047-589a5558c05b","Type":"ContainerDied","Data":"e0dbecc00872335f728a00b6b8f66f206cae1eacee637ad93f50bc7b8fc0f42d"} Oct 05 08:32:26 crc kubenswrapper[4846]: I1005 08:32:26.842251 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-c672lz" Oct 05 08:32:26 crc kubenswrapper[4846]: I1005 08:32:26.974773 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5sznp\" (UniqueName: \"kubernetes.io/projected/4ca8fa61-f7ee-4f10-b047-589a5558c05b-kube-api-access-5sznp\") pod \"4ca8fa61-f7ee-4f10-b047-589a5558c05b\" (UID: \"4ca8fa61-f7ee-4f10-b047-589a5558c05b\") " Oct 05 08:32:26 crc kubenswrapper[4846]: I1005 08:32:26.974829 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4ca8fa61-f7ee-4f10-b047-589a5558c05b-inventory\") pod \"4ca8fa61-f7ee-4f10-b047-589a5558c05b\" (UID: \"4ca8fa61-f7ee-4f10-b047-589a5558c05b\") " Oct 05 08:32:26 crc kubenswrapper[4846]: I1005 08:32:26.974853 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pre-adoption-validation-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4ca8fa61-f7ee-4f10-b047-589a5558c05b-pre-adoption-validation-combined-ca-bundle\") pod \"4ca8fa61-f7ee-4f10-b047-589a5558c05b\" (UID: \"4ca8fa61-f7ee-4f10-b047-589a5558c05b\") " Oct 05 08:32:26 crc kubenswrapper[4846]: I1005 08:32:26.974916 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4ca8fa61-f7ee-4f10-b047-589a5558c05b-ssh-key\") pod \"4ca8fa61-f7ee-4f10-b047-589a5558c05b\" (UID: \"4ca8fa61-f7ee-4f10-b047-589a5558c05b\") " Oct 05 08:32:26 crc kubenswrapper[4846]: I1005 08:32:26.981604 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4ca8fa61-f7ee-4f10-b047-589a5558c05b-pre-adoption-validation-combined-ca-bundle" (OuterVolumeSpecName: "pre-adoption-validation-combined-ca-bundle") pod "4ca8fa61-f7ee-4f10-b047-589a5558c05b" (UID: "4ca8fa61-f7ee-4f10-b047-589a5558c05b"). InnerVolumeSpecName "pre-adoption-validation-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:32:26 crc kubenswrapper[4846]: I1005 08:32:26.982638 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4ca8fa61-f7ee-4f10-b047-589a5558c05b-kube-api-access-5sznp" (OuterVolumeSpecName: "kube-api-access-5sznp") pod "4ca8fa61-f7ee-4f10-b047-589a5558c05b" (UID: "4ca8fa61-f7ee-4f10-b047-589a5558c05b"). InnerVolumeSpecName "kube-api-access-5sznp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:32:27 crc kubenswrapper[4846]: I1005 08:32:27.004999 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4ca8fa61-f7ee-4f10-b047-589a5558c05b-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "4ca8fa61-f7ee-4f10-b047-589a5558c05b" (UID: "4ca8fa61-f7ee-4f10-b047-589a5558c05b"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:32:27 crc kubenswrapper[4846]: I1005 08:32:27.006640 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4ca8fa61-f7ee-4f10-b047-589a5558c05b-inventory" (OuterVolumeSpecName: "inventory") pod "4ca8fa61-f7ee-4f10-b047-589a5558c05b" (UID: "4ca8fa61-f7ee-4f10-b047-589a5558c05b"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:32:27 crc kubenswrapper[4846]: I1005 08:32:27.079360 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5sznp\" (UniqueName: \"kubernetes.io/projected/4ca8fa61-f7ee-4f10-b047-589a5558c05b-kube-api-access-5sznp\") on node \"crc\" DevicePath \"\"" Oct 05 08:32:27 crc kubenswrapper[4846]: I1005 08:32:27.079392 4846 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4ca8fa61-f7ee-4f10-b047-589a5558c05b-inventory\") on node \"crc\" DevicePath \"\"" Oct 05 08:32:27 crc kubenswrapper[4846]: I1005 08:32:27.079405 4846 reconciler_common.go:293] "Volume detached for volume \"pre-adoption-validation-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4ca8fa61-f7ee-4f10-b047-589a5558c05b-pre-adoption-validation-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 08:32:27 crc kubenswrapper[4846]: I1005 08:32:27.079416 4846 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4ca8fa61-f7ee-4f10-b047-589a5558c05b-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 05 08:32:27 crc kubenswrapper[4846]: I1005 08:32:27.286993 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-c672lz" event={"ID":"4ca8fa61-f7ee-4f10-b047-589a5558c05b","Type":"ContainerDied","Data":"891848a05a25d5b1d52f13d8ede63a3f1ec8a3207881fb7f10a4d3170950337c"} Oct 05 08:32:27 crc kubenswrapper[4846]: I1005 08:32:27.287035 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="891848a05a25d5b1d52f13d8ede63a3f1ec8a3207881fb7f10a4d3170950337c" Oct 05 08:32:27 crc kubenswrapper[4846]: I1005 08:32:27.287094 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-c672lz" Oct 05 08:32:35 crc kubenswrapper[4846]: I1005 08:32:35.261527 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-ptzpk"] Oct 05 08:32:35 crc kubenswrapper[4846]: E1005 08:32:35.262779 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4ca8fa61-f7ee-4f10-b047-589a5558c05b" containerName="pre-adoption-validation-openstack-pre-adoption-openstack-cell1" Oct 05 08:32:35 crc kubenswrapper[4846]: I1005 08:32:35.262796 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="4ca8fa61-f7ee-4f10-b047-589a5558c05b" containerName="pre-adoption-validation-openstack-pre-adoption-openstack-cell1" Oct 05 08:32:35 crc kubenswrapper[4846]: I1005 08:32:35.263078 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="4ca8fa61-f7ee-4f10-b047-589a5558c05b" containerName="pre-adoption-validation-openstack-pre-adoption-openstack-cell1" Oct 05 08:32:35 crc kubenswrapper[4846]: I1005 08:32:35.263991 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-ptzpk" Oct 05 08:32:35 crc kubenswrapper[4846]: I1005 08:32:35.266378 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Oct 05 08:32:35 crc kubenswrapper[4846]: I1005 08:32:35.267515 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 05 08:32:35 crc kubenswrapper[4846]: I1005 08:32:35.267947 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Oct 05 08:32:35 crc kubenswrapper[4846]: I1005 08:32:35.272438 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-ptzpk"] Oct 05 08:32:35 crc kubenswrapper[4846]: I1005 08:32:35.277553 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-7687s" Oct 05 08:32:35 crc kubenswrapper[4846]: I1005 08:32:35.374501 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/19baf6c0-a301-49fe-be7b-6682904106b9-ssh-key\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-ptzpk\" (UID: \"19baf6c0-a301-49fe-be7b-6682904106b9\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-ptzpk" Oct 05 08:32:35 crc kubenswrapper[4846]: I1005 08:32:35.374561 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/19baf6c0-a301-49fe-be7b-6682904106b9-inventory\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-ptzpk\" (UID: \"19baf6c0-a301-49fe-be7b-6682904106b9\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-ptzpk" Oct 05 08:32:35 crc kubenswrapper[4846]: I1005 08:32:35.374586 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-brqj8\" (UniqueName: \"kubernetes.io/projected/19baf6c0-a301-49fe-be7b-6682904106b9-kube-api-access-brqj8\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-ptzpk\" (UID: \"19baf6c0-a301-49fe-be7b-6682904106b9\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-ptzpk" Oct 05 08:32:35 crc kubenswrapper[4846]: I1005 08:32:35.374607 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tripleo-cleanup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/19baf6c0-a301-49fe-be7b-6682904106b9-tripleo-cleanup-combined-ca-bundle\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-ptzpk\" (UID: \"19baf6c0-a301-49fe-be7b-6682904106b9\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-ptzpk" Oct 05 08:32:35 crc kubenswrapper[4846]: I1005 08:32:35.476902 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/19baf6c0-a301-49fe-be7b-6682904106b9-ssh-key\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-ptzpk\" (UID: \"19baf6c0-a301-49fe-be7b-6682904106b9\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-ptzpk" Oct 05 08:32:35 crc kubenswrapper[4846]: I1005 08:32:35.477008 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/19baf6c0-a301-49fe-be7b-6682904106b9-inventory\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-ptzpk\" (UID: \"19baf6c0-a301-49fe-be7b-6682904106b9\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-ptzpk" Oct 05 08:32:35 crc kubenswrapper[4846]: I1005 08:32:35.477057 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-brqj8\" (UniqueName: \"kubernetes.io/projected/19baf6c0-a301-49fe-be7b-6682904106b9-kube-api-access-brqj8\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-ptzpk\" (UID: \"19baf6c0-a301-49fe-be7b-6682904106b9\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-ptzpk" Oct 05 08:32:35 crc kubenswrapper[4846]: I1005 08:32:35.477105 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tripleo-cleanup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/19baf6c0-a301-49fe-be7b-6682904106b9-tripleo-cleanup-combined-ca-bundle\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-ptzpk\" (UID: \"19baf6c0-a301-49fe-be7b-6682904106b9\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-ptzpk" Oct 05 08:32:35 crc kubenswrapper[4846]: I1005 08:32:35.486171 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/19baf6c0-a301-49fe-be7b-6682904106b9-ssh-key\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-ptzpk\" (UID: \"19baf6c0-a301-49fe-be7b-6682904106b9\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-ptzpk" Oct 05 08:32:35 crc kubenswrapper[4846]: I1005 08:32:35.486461 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/19baf6c0-a301-49fe-be7b-6682904106b9-inventory\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-ptzpk\" (UID: \"19baf6c0-a301-49fe-be7b-6682904106b9\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-ptzpk" Oct 05 08:32:35 crc kubenswrapper[4846]: I1005 08:32:35.486867 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tripleo-cleanup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/19baf6c0-a301-49fe-be7b-6682904106b9-tripleo-cleanup-combined-ca-bundle\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-ptzpk\" (UID: \"19baf6c0-a301-49fe-be7b-6682904106b9\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-ptzpk" Oct 05 08:32:35 crc kubenswrapper[4846]: I1005 08:32:35.495439 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-brqj8\" (UniqueName: \"kubernetes.io/projected/19baf6c0-a301-49fe-be7b-6682904106b9-kube-api-access-brqj8\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-ptzpk\" (UID: \"19baf6c0-a301-49fe-be7b-6682904106b9\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-ptzpk" Oct 05 08:32:35 crc kubenswrapper[4846]: I1005 08:32:35.521029 4846 scope.go:117] "RemoveContainer" containerID="c93489b784c6b4122970b363d7fff9e7a3a2208c64021f94bd3ca2ab12d03f6b" Oct 05 08:32:35 crc kubenswrapper[4846]: I1005 08:32:35.634120 4846 scope.go:117] "RemoveContainer" containerID="fa6050e2b9830faaa192d66869bdf6af4ade7688b8f5126c21c8162a7872705e" Oct 05 08:32:35 crc kubenswrapper[4846]: I1005 08:32:35.647411 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-ptzpk" Oct 05 08:32:35 crc kubenswrapper[4846]: I1005 08:32:35.659533 4846 scope.go:117] "RemoveContainer" containerID="ef3afe5c2b471b1257316a3a5ac93be12891e742390679711548a50b322dfd7c" Oct 05 08:32:35 crc kubenswrapper[4846]: I1005 08:32:35.866998 4846 scope.go:117] "RemoveContainer" containerID="9283ee8b199a6e8c0dfeb0966df9bc7ada80e4f5b05fa35bf7f11a12ca52a6e8" Oct 05 08:32:35 crc kubenswrapper[4846]: I1005 08:32:35.910716 4846 scope.go:117] "RemoveContainer" containerID="bf0445b72a12e1f9d4fcf1a8742dc0aa8e65169aaeed48321b63545b19175f45" Oct 05 08:32:36 crc kubenswrapper[4846]: I1005 08:32:36.197702 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-ptzpk"] Oct 05 08:32:36 crc kubenswrapper[4846]: W1005 08:32:36.204777 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod19baf6c0_a301_49fe_be7b_6682904106b9.slice/crio-6727932025eab23e57fab3c11862dbb665b4f88404d0c9381ae8b575e93100c2 WatchSource:0}: Error finding container 6727932025eab23e57fab3c11862dbb665b4f88404d0c9381ae8b575e93100c2: Status 404 returned error can't find the container with id 6727932025eab23e57fab3c11862dbb665b4f88404d0c9381ae8b575e93100c2 Oct 05 08:32:36 crc kubenswrapper[4846]: I1005 08:32:36.434284 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-ptzpk" event={"ID":"19baf6c0-a301-49fe-be7b-6682904106b9","Type":"ContainerStarted","Data":"6727932025eab23e57fab3c11862dbb665b4f88404d0c9381ae8b575e93100c2"} Oct 05 08:32:37 crc kubenswrapper[4846]: I1005 08:32:37.444514 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-ptzpk" event={"ID":"19baf6c0-a301-49fe-be7b-6682904106b9","Type":"ContainerStarted","Data":"5ba7013871b15a45e9c5ea364c64a9710a10033aaa75a7c468cacbde3028d3f9"} Oct 05 08:32:37 crc kubenswrapper[4846]: I1005 08:32:37.473993 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-ptzpk" podStartSLOduration=2.031248957 podStartE2EDuration="2.473976266s" podCreationTimestamp="2025-10-05 08:32:35 +0000 UTC" firstStartedPulling="2025-10-05 08:32:36.207276188 +0000 UTC m=+6278.448128963" lastFinishedPulling="2025-10-05 08:32:36.650003457 +0000 UTC m=+6278.890856272" observedRunningTime="2025-10-05 08:32:37.466333865 +0000 UTC m=+6279.707186640" watchObservedRunningTime="2025-10-05 08:32:37.473976266 +0000 UTC m=+6279.714829041" Oct 05 08:32:53 crc kubenswrapper[4846]: I1005 08:32:53.325806 4846 patch_prober.go:28] interesting pod/machine-config-daemon-fscvf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 08:32:53 crc kubenswrapper[4846]: I1005 08:32:53.326330 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 08:33:01 crc kubenswrapper[4846]: I1005 08:33:01.046944 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-db-create-62kls"] Oct 05 08:33:01 crc kubenswrapper[4846]: I1005 08:33:01.061946 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-db-create-q8mtl"] Oct 05 08:33:01 crc kubenswrapper[4846]: I1005 08:33:01.075307 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-db-create-q8mtl"] Oct 05 08:33:01 crc kubenswrapper[4846]: I1005 08:33:01.082051 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-db-create-62kls"] Oct 05 08:33:02 crc kubenswrapper[4846]: I1005 08:33:02.048055 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-db-create-rc6qh"] Oct 05 08:33:02 crc kubenswrapper[4846]: I1005 08:33:02.067552 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-db-create-rc6qh"] Oct 05 08:33:02 crc kubenswrapper[4846]: I1005 08:33:02.516358 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1c3333df-ae72-4159-a51d-9e11f0a0669a" path="/var/lib/kubelet/pods/1c3333df-ae72-4159-a51d-9e11f0a0669a/volumes" Oct 05 08:33:02 crc kubenswrapper[4846]: I1005 08:33:02.517386 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b0df8f79-c188-4014-9e9e-9e0f692c0357" path="/var/lib/kubelet/pods/b0df8f79-c188-4014-9e9e-9e0f692c0357/volumes" Oct 05 08:33:02 crc kubenswrapper[4846]: I1005 08:33:02.518491 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d720d2a2-9841-47d1-9d55-d5e41d35d947" path="/var/lib/kubelet/pods/d720d2a2-9841-47d1-9d55-d5e41d35d947/volumes" Oct 05 08:33:11 crc kubenswrapper[4846]: I1005 08:33:11.051089 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-b599-account-create-b7c5d"] Oct 05 08:33:11 crc kubenswrapper[4846]: I1005 08:33:11.067280 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-b599-account-create-b7c5d"] Oct 05 08:33:12 crc kubenswrapper[4846]: I1005 08:33:12.074562 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-fc2b-account-create-bnrgd"] Oct 05 08:33:12 crc kubenswrapper[4846]: I1005 08:33:12.087434 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-fc2b-account-create-bnrgd"] Oct 05 08:33:12 crc kubenswrapper[4846]: I1005 08:33:12.104692 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-27ec-account-create-5gjp8"] Oct 05 08:33:12 crc kubenswrapper[4846]: I1005 08:33:12.114322 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-27ec-account-create-5gjp8"] Oct 05 08:33:12 crc kubenswrapper[4846]: I1005 08:33:12.521172 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3d0781ed-2c7e-452d-8d71-26c62c1c00e1" path="/var/lib/kubelet/pods/3d0781ed-2c7e-452d-8d71-26c62c1c00e1/volumes" Oct 05 08:33:12 crc kubenswrapper[4846]: I1005 08:33:12.522951 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="54b580fc-9fa1-4871-bfa8-832910ffffec" path="/var/lib/kubelet/pods/54b580fc-9fa1-4871-bfa8-832910ffffec/volumes" Oct 05 08:33:12 crc kubenswrapper[4846]: I1005 08:33:12.524352 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5dedc799-d3d4-4fe0-a2de-317784fd1db3" path="/var/lib/kubelet/pods/5dedc799-d3d4-4fe0-a2de-317784fd1db3/volumes" Oct 05 08:33:23 crc kubenswrapper[4846]: I1005 08:33:23.325995 4846 patch_prober.go:28] interesting pod/machine-config-daemon-fscvf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 08:33:23 crc kubenswrapper[4846]: I1005 08:33:23.326821 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 08:33:23 crc kubenswrapper[4846]: I1005 08:33:23.326926 4846 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" Oct 05 08:33:23 crc kubenswrapper[4846]: I1005 08:33:23.329168 4846 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"ad5522212e6e288d74747979bc3c3cc6cbe3871d7b66a9bb6a2f8a05f575f23e"} pod="openshift-machine-config-operator/machine-config-daemon-fscvf" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 05 08:33:23 crc kubenswrapper[4846]: I1005 08:33:23.329331 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" containerName="machine-config-daemon" containerID="cri-o://ad5522212e6e288d74747979bc3c3cc6cbe3871d7b66a9bb6a2f8a05f575f23e" gracePeriod=600 Oct 05 08:33:23 crc kubenswrapper[4846]: E1005 08:33:23.452034 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 08:33:24 crc kubenswrapper[4846]: I1005 08:33:24.013911 4846 generic.go:334] "Generic (PLEG): container finished" podID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" containerID="ad5522212e6e288d74747979bc3c3cc6cbe3871d7b66a9bb6a2f8a05f575f23e" exitCode=0 Oct 05 08:33:24 crc kubenswrapper[4846]: I1005 08:33:24.013961 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" event={"ID":"ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7","Type":"ContainerDied","Data":"ad5522212e6e288d74747979bc3c3cc6cbe3871d7b66a9bb6a2f8a05f575f23e"} Oct 05 08:33:24 crc kubenswrapper[4846]: I1005 08:33:24.013995 4846 scope.go:117] "RemoveContainer" containerID="5901a64fb8754c398c6b2b246ea8952e5e7f49cedcb4e34d861ec3f14a1f9619" Oct 05 08:33:24 crc kubenswrapper[4846]: I1005 08:33:24.014969 4846 scope.go:117] "RemoveContainer" containerID="ad5522212e6e288d74747979bc3c3cc6cbe3871d7b66a9bb6a2f8a05f575f23e" Oct 05 08:33:24 crc kubenswrapper[4846]: E1005 08:33:24.015735 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 08:33:30 crc kubenswrapper[4846]: I1005 08:33:30.047339 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-sskzq"] Oct 05 08:33:30 crc kubenswrapper[4846]: I1005 08:33:30.055050 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-sskzq"] Oct 05 08:33:30 crc kubenswrapper[4846]: I1005 08:33:30.515305 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a7e835d9-77a2-4066-b151-42b30db24e78" path="/var/lib/kubelet/pods/a7e835d9-77a2-4066-b151-42b30db24e78/volumes" Oct 05 08:33:36 crc kubenswrapper[4846]: I1005 08:33:36.061373 4846 scope.go:117] "RemoveContainer" containerID="a6ff819f08387f88aee4ad007ec1517ce487f455ff69578da57596e2835b80ff" Oct 05 08:33:36 crc kubenswrapper[4846]: I1005 08:33:36.087339 4846 scope.go:117] "RemoveContainer" containerID="c44383b11f0e5106d7bda217ca484922769bfcc65aa3687d5dc72e75b0293816" Oct 05 08:33:36 crc kubenswrapper[4846]: I1005 08:33:36.148583 4846 scope.go:117] "RemoveContainer" containerID="633f2300149ab2e249a1f8037040539ca61cfa678dd1ab445ac0e86432b15b34" Oct 05 08:33:36 crc kubenswrapper[4846]: I1005 08:33:36.227690 4846 scope.go:117] "RemoveContainer" containerID="3aada75a88067f157714735eac38160174336f81486c01c86f742d2382264842" Oct 05 08:33:36 crc kubenswrapper[4846]: I1005 08:33:36.277548 4846 scope.go:117] "RemoveContainer" containerID="0f757a5a9d9b747a9815a75c8d03fd44c128374cc6e796440a93564d5f796b8d" Oct 05 08:33:36 crc kubenswrapper[4846]: I1005 08:33:36.354319 4846 scope.go:117] "RemoveContainer" containerID="1cd3859419831ec9f4d7c87785196b007bc00e0dbe13b32b853243a4dd9589fc" Oct 05 08:33:36 crc kubenswrapper[4846]: I1005 08:33:36.378479 4846 scope.go:117] "RemoveContainer" containerID="d27f160f9bbb9185928b4f60c6cd30913c1c0c697e9f2f40ec11972b0a5bbbd3" Oct 05 08:33:36 crc kubenswrapper[4846]: I1005 08:33:36.498040 4846 scope.go:117] "RemoveContainer" containerID="ad5522212e6e288d74747979bc3c3cc6cbe3871d7b66a9bb6a2f8a05f575f23e" Oct 05 08:33:36 crc kubenswrapper[4846]: E1005 08:33:36.498498 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 08:33:50 crc kubenswrapper[4846]: I1005 08:33:50.047228 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-cell-mapping-4kt89"] Oct 05 08:33:50 crc kubenswrapper[4846]: I1005 08:33:50.062469 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-cell-mapping-4kt89"] Oct 05 08:33:50 crc kubenswrapper[4846]: I1005 08:33:50.498059 4846 scope.go:117] "RemoveContainer" containerID="ad5522212e6e288d74747979bc3c3cc6cbe3871d7b66a9bb6a2f8a05f575f23e" Oct 05 08:33:50 crc kubenswrapper[4846]: E1005 08:33:50.498560 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 08:33:50 crc kubenswrapper[4846]: I1005 08:33:50.516786 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="285fade3-a17c-4cfc-937c-096ebdb15848" path="/var/lib/kubelet/pods/285fade3-a17c-4cfc-937c-096ebdb15848/volumes" Oct 05 08:33:51 crc kubenswrapper[4846]: I1005 08:33:51.055378 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-p9qj9"] Oct 05 08:33:51 crc kubenswrapper[4846]: I1005 08:33:51.069403 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-p9qj9"] Oct 05 08:33:52 crc kubenswrapper[4846]: I1005 08:33:52.541011 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5098a05b-cc49-4af9-b3cb-7d6ff17f38f2" path="/var/lib/kubelet/pods/5098a05b-cc49-4af9-b3cb-7d6ff17f38f2/volumes" Oct 05 08:34:02 crc kubenswrapper[4846]: I1005 08:34:02.498353 4846 scope.go:117] "RemoveContainer" containerID="ad5522212e6e288d74747979bc3c3cc6cbe3871d7b66a9bb6a2f8a05f575f23e" Oct 05 08:34:02 crc kubenswrapper[4846]: E1005 08:34:02.499261 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 08:34:16 crc kubenswrapper[4846]: I1005 08:34:16.498115 4846 scope.go:117] "RemoveContainer" containerID="ad5522212e6e288d74747979bc3c3cc6cbe3871d7b66a9bb6a2f8a05f575f23e" Oct 05 08:34:16 crc kubenswrapper[4846]: E1005 08:34:16.499411 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 08:34:31 crc kubenswrapper[4846]: I1005 08:34:31.500074 4846 scope.go:117] "RemoveContainer" containerID="ad5522212e6e288d74747979bc3c3cc6cbe3871d7b66a9bb6a2f8a05f575f23e" Oct 05 08:34:31 crc kubenswrapper[4846]: E1005 08:34:31.501571 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 08:34:36 crc kubenswrapper[4846]: I1005 08:34:36.555808 4846 scope.go:117] "RemoveContainer" containerID="3d4a8fd72c0b7509378ec345a1a94dba89ea5a6e23491caf00bbff90fdab2edb" Oct 05 08:34:36 crc kubenswrapper[4846]: I1005 08:34:36.654536 4846 scope.go:117] "RemoveContainer" containerID="3fdb7b36842986265b48883b87d5734a37d51381cf8ee9b9978d76d1754d5762" Oct 05 08:34:37 crc kubenswrapper[4846]: I1005 08:34:37.049067 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-cell-mapping-gdt57"] Oct 05 08:34:37 crc kubenswrapper[4846]: I1005 08:34:37.057030 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-cell-mapping-gdt57"] Oct 05 08:34:38 crc kubenswrapper[4846]: I1005 08:34:38.520077 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f8888283-b66d-4993-80a0-a7b6b1bac02b" path="/var/lib/kubelet/pods/f8888283-b66d-4993-80a0-a7b6b1bac02b/volumes" Oct 05 08:34:45 crc kubenswrapper[4846]: I1005 08:34:45.498522 4846 scope.go:117] "RemoveContainer" containerID="ad5522212e6e288d74747979bc3c3cc6cbe3871d7b66a9bb6a2f8a05f575f23e" Oct 05 08:34:45 crc kubenswrapper[4846]: E1005 08:34:45.499323 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 08:35:00 crc kubenswrapper[4846]: I1005 08:35:00.498389 4846 scope.go:117] "RemoveContainer" containerID="ad5522212e6e288d74747979bc3c3cc6cbe3871d7b66a9bb6a2f8a05f575f23e" Oct 05 08:35:00 crc kubenswrapper[4846]: E1005 08:35:00.499431 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 08:35:12 crc kubenswrapper[4846]: I1005 08:35:12.497829 4846 scope.go:117] "RemoveContainer" containerID="ad5522212e6e288d74747979bc3c3cc6cbe3871d7b66a9bb6a2f8a05f575f23e" Oct 05 08:35:12 crc kubenswrapper[4846]: E1005 08:35:12.498893 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 08:35:27 crc kubenswrapper[4846]: I1005 08:35:27.499742 4846 scope.go:117] "RemoveContainer" containerID="ad5522212e6e288d74747979bc3c3cc6cbe3871d7b66a9bb6a2f8a05f575f23e" Oct 05 08:35:27 crc kubenswrapper[4846]: E1005 08:35:27.501507 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 08:35:36 crc kubenswrapper[4846]: I1005 08:35:36.853399 4846 scope.go:117] "RemoveContainer" containerID="717a4de0219e85cadcebc33059340186dbabb2f100ca30cb91d92d6000152b08" Oct 05 08:35:36 crc kubenswrapper[4846]: I1005 08:35:36.903034 4846 scope.go:117] "RemoveContainer" containerID="387ea30ec891b63656dcd74018a0d5d1f86fea2b6f29a8bbc972f659e551aa35" Oct 05 08:35:42 crc kubenswrapper[4846]: I1005 08:35:42.498603 4846 scope.go:117] "RemoveContainer" containerID="ad5522212e6e288d74747979bc3c3cc6cbe3871d7b66a9bb6a2f8a05f575f23e" Oct 05 08:35:42 crc kubenswrapper[4846]: E1005 08:35:42.500411 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 08:35:56 crc kubenswrapper[4846]: I1005 08:35:56.498121 4846 scope.go:117] "RemoveContainer" containerID="ad5522212e6e288d74747979bc3c3cc6cbe3871d7b66a9bb6a2f8a05f575f23e" Oct 05 08:35:56 crc kubenswrapper[4846]: E1005 08:35:56.499415 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 08:36:10 crc kubenswrapper[4846]: I1005 08:36:10.498099 4846 scope.go:117] "RemoveContainer" containerID="ad5522212e6e288d74747979bc3c3cc6cbe3871d7b66a9bb6a2f8a05f575f23e" Oct 05 08:36:10 crc kubenswrapper[4846]: E1005 08:36:10.499286 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 08:36:25 crc kubenswrapper[4846]: I1005 08:36:25.499054 4846 scope.go:117] "RemoveContainer" containerID="ad5522212e6e288d74747979bc3c3cc6cbe3871d7b66a9bb6a2f8a05f575f23e" Oct 05 08:36:25 crc kubenswrapper[4846]: E1005 08:36:25.500244 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 08:36:36 crc kubenswrapper[4846]: I1005 08:36:36.500939 4846 scope.go:117] "RemoveContainer" containerID="ad5522212e6e288d74747979bc3c3cc6cbe3871d7b66a9bb6a2f8a05f575f23e" Oct 05 08:36:36 crc kubenswrapper[4846]: E1005 08:36:36.501968 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 08:36:37 crc kubenswrapper[4846]: I1005 08:36:37.004218 4846 scope.go:117] "RemoveContainer" containerID="855c10a9b75c2c9926f5dd593debd83762c84203e420e5b91cfe42e7db8b6559" Oct 05 08:36:37 crc kubenswrapper[4846]: I1005 08:36:37.042565 4846 scope.go:117] "RemoveContainer" containerID="35860887d251721dfd419731dae2306db24d39a9d93d8f1543f1fe5315075bf9" Oct 05 08:36:37 crc kubenswrapper[4846]: I1005 08:36:37.076302 4846 scope.go:117] "RemoveContainer" containerID="5cb1120abfdcdf858d882f7c6f47445cb49f210041a4a9486cda3b94e9ad8438" Oct 05 08:36:49 crc kubenswrapper[4846]: I1005 08:36:49.499326 4846 scope.go:117] "RemoveContainer" containerID="ad5522212e6e288d74747979bc3c3cc6cbe3871d7b66a9bb6a2f8a05f575f23e" Oct 05 08:36:49 crc kubenswrapper[4846]: E1005 08:36:49.500857 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 08:36:54 crc kubenswrapper[4846]: I1005 08:36:54.229265 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-rdh4n"] Oct 05 08:36:54 crc kubenswrapper[4846]: I1005 08:36:54.236218 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rdh4n" Oct 05 08:36:54 crc kubenswrapper[4846]: I1005 08:36:54.243063 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-rdh4n"] Oct 05 08:36:54 crc kubenswrapper[4846]: I1005 08:36:54.416020 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hvn8l\" (UniqueName: \"kubernetes.io/projected/12c63dbb-761d-40ba-b5cd-3407b8b3ff10-kube-api-access-hvn8l\") pod \"community-operators-rdh4n\" (UID: \"12c63dbb-761d-40ba-b5cd-3407b8b3ff10\") " pod="openshift-marketplace/community-operators-rdh4n" Oct 05 08:36:54 crc kubenswrapper[4846]: I1005 08:36:54.416073 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/12c63dbb-761d-40ba-b5cd-3407b8b3ff10-catalog-content\") pod \"community-operators-rdh4n\" (UID: \"12c63dbb-761d-40ba-b5cd-3407b8b3ff10\") " pod="openshift-marketplace/community-operators-rdh4n" Oct 05 08:36:54 crc kubenswrapper[4846]: I1005 08:36:54.416102 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/12c63dbb-761d-40ba-b5cd-3407b8b3ff10-utilities\") pod \"community-operators-rdh4n\" (UID: \"12c63dbb-761d-40ba-b5cd-3407b8b3ff10\") " pod="openshift-marketplace/community-operators-rdh4n" Oct 05 08:36:54 crc kubenswrapper[4846]: I1005 08:36:54.517776 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hvn8l\" (UniqueName: \"kubernetes.io/projected/12c63dbb-761d-40ba-b5cd-3407b8b3ff10-kube-api-access-hvn8l\") pod \"community-operators-rdh4n\" (UID: \"12c63dbb-761d-40ba-b5cd-3407b8b3ff10\") " pod="openshift-marketplace/community-operators-rdh4n" Oct 05 08:36:54 crc kubenswrapper[4846]: I1005 08:36:54.517827 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/12c63dbb-761d-40ba-b5cd-3407b8b3ff10-catalog-content\") pod \"community-operators-rdh4n\" (UID: \"12c63dbb-761d-40ba-b5cd-3407b8b3ff10\") " pod="openshift-marketplace/community-operators-rdh4n" Oct 05 08:36:54 crc kubenswrapper[4846]: I1005 08:36:54.517849 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/12c63dbb-761d-40ba-b5cd-3407b8b3ff10-utilities\") pod \"community-operators-rdh4n\" (UID: \"12c63dbb-761d-40ba-b5cd-3407b8b3ff10\") " pod="openshift-marketplace/community-operators-rdh4n" Oct 05 08:36:54 crc kubenswrapper[4846]: I1005 08:36:54.519003 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/12c63dbb-761d-40ba-b5cd-3407b8b3ff10-catalog-content\") pod \"community-operators-rdh4n\" (UID: \"12c63dbb-761d-40ba-b5cd-3407b8b3ff10\") " pod="openshift-marketplace/community-operators-rdh4n" Oct 05 08:36:54 crc kubenswrapper[4846]: I1005 08:36:54.519348 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/12c63dbb-761d-40ba-b5cd-3407b8b3ff10-utilities\") pod \"community-operators-rdh4n\" (UID: \"12c63dbb-761d-40ba-b5cd-3407b8b3ff10\") " pod="openshift-marketplace/community-operators-rdh4n" Oct 05 08:36:54 crc kubenswrapper[4846]: I1005 08:36:54.557247 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hvn8l\" (UniqueName: \"kubernetes.io/projected/12c63dbb-761d-40ba-b5cd-3407b8b3ff10-kube-api-access-hvn8l\") pod \"community-operators-rdh4n\" (UID: \"12c63dbb-761d-40ba-b5cd-3407b8b3ff10\") " pod="openshift-marketplace/community-operators-rdh4n" Oct 05 08:36:54 crc kubenswrapper[4846]: I1005 08:36:54.574545 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rdh4n" Oct 05 08:36:55 crc kubenswrapper[4846]: I1005 08:36:55.124597 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-rdh4n"] Oct 05 08:36:55 crc kubenswrapper[4846]: I1005 08:36:55.779660 4846 generic.go:334] "Generic (PLEG): container finished" podID="12c63dbb-761d-40ba-b5cd-3407b8b3ff10" containerID="21e24f7429ce27406685c16462e1bcdfb6b0c3d4b9c963ed32d992047493e7e4" exitCode=0 Oct 05 08:36:55 crc kubenswrapper[4846]: I1005 08:36:55.779720 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rdh4n" event={"ID":"12c63dbb-761d-40ba-b5cd-3407b8b3ff10","Type":"ContainerDied","Data":"21e24f7429ce27406685c16462e1bcdfb6b0c3d4b9c963ed32d992047493e7e4"} Oct 05 08:36:55 crc kubenswrapper[4846]: I1005 08:36:55.780047 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rdh4n" event={"ID":"12c63dbb-761d-40ba-b5cd-3407b8b3ff10","Type":"ContainerStarted","Data":"34863ac31850da2e47a3d9ba4bc54c6ade2976b01fcf6d0864251e7a3f94d8e9"} Oct 05 08:36:55 crc kubenswrapper[4846]: I1005 08:36:55.783324 4846 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 05 08:36:56 crc kubenswrapper[4846]: I1005 08:36:56.790727 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rdh4n" event={"ID":"12c63dbb-761d-40ba-b5cd-3407b8b3ff10","Type":"ContainerStarted","Data":"7bee4128a22582828c0ffa96f85fb33165f690c68642cc81ff53ec2105b2e1dd"} Oct 05 08:36:58 crc kubenswrapper[4846]: I1005 08:36:58.817754 4846 generic.go:334] "Generic (PLEG): container finished" podID="12c63dbb-761d-40ba-b5cd-3407b8b3ff10" containerID="7bee4128a22582828c0ffa96f85fb33165f690c68642cc81ff53ec2105b2e1dd" exitCode=0 Oct 05 08:36:58 crc kubenswrapper[4846]: I1005 08:36:58.817883 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rdh4n" event={"ID":"12c63dbb-761d-40ba-b5cd-3407b8b3ff10","Type":"ContainerDied","Data":"7bee4128a22582828c0ffa96f85fb33165f690c68642cc81ff53ec2105b2e1dd"} Oct 05 08:36:59 crc kubenswrapper[4846]: I1005 08:36:59.830412 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rdh4n" event={"ID":"12c63dbb-761d-40ba-b5cd-3407b8b3ff10","Type":"ContainerStarted","Data":"c195a527ff2a920f99ac7ee871ab454a16eec78b65232e1e5a7bf3c313329371"} Oct 05 08:36:59 crc kubenswrapper[4846]: I1005 08:36:59.860207 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-rdh4n" podStartSLOduration=2.362742286 podStartE2EDuration="5.860170062s" podCreationTimestamp="2025-10-05 08:36:54 +0000 UTC" firstStartedPulling="2025-10-05 08:36:55.783069179 +0000 UTC m=+6538.023921954" lastFinishedPulling="2025-10-05 08:36:59.280496955 +0000 UTC m=+6541.521349730" observedRunningTime="2025-10-05 08:36:59.856665579 +0000 UTC m=+6542.097518364" watchObservedRunningTime="2025-10-05 08:36:59.860170062 +0000 UTC m=+6542.101022847" Oct 05 08:37:03 crc kubenswrapper[4846]: I1005 08:37:03.498170 4846 scope.go:117] "RemoveContainer" containerID="ad5522212e6e288d74747979bc3c3cc6cbe3871d7b66a9bb6a2f8a05f575f23e" Oct 05 08:37:03 crc kubenswrapper[4846]: E1005 08:37:03.499053 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 08:37:04 crc kubenswrapper[4846]: I1005 08:37:04.575730 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-rdh4n" Oct 05 08:37:04 crc kubenswrapper[4846]: I1005 08:37:04.575857 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-rdh4n" Oct 05 08:37:04 crc kubenswrapper[4846]: I1005 08:37:04.666354 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-rdh4n" Oct 05 08:37:04 crc kubenswrapper[4846]: I1005 08:37:04.937799 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-rdh4n" Oct 05 08:37:05 crc kubenswrapper[4846]: I1005 08:37:05.008054 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-rdh4n"] Oct 05 08:37:06 crc kubenswrapper[4846]: I1005 08:37:06.911839 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-rdh4n" podUID="12c63dbb-761d-40ba-b5cd-3407b8b3ff10" containerName="registry-server" containerID="cri-o://c195a527ff2a920f99ac7ee871ab454a16eec78b65232e1e5a7bf3c313329371" gracePeriod=2 Oct 05 08:37:07 crc kubenswrapper[4846]: I1005 08:37:07.911749 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rdh4n" Oct 05 08:37:07 crc kubenswrapper[4846]: I1005 08:37:07.942268 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/12c63dbb-761d-40ba-b5cd-3407b8b3ff10-catalog-content\") pod \"12c63dbb-761d-40ba-b5cd-3407b8b3ff10\" (UID: \"12c63dbb-761d-40ba-b5cd-3407b8b3ff10\") " Oct 05 08:37:07 crc kubenswrapper[4846]: I1005 08:37:07.942366 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hvn8l\" (UniqueName: \"kubernetes.io/projected/12c63dbb-761d-40ba-b5cd-3407b8b3ff10-kube-api-access-hvn8l\") pod \"12c63dbb-761d-40ba-b5cd-3407b8b3ff10\" (UID: \"12c63dbb-761d-40ba-b5cd-3407b8b3ff10\") " Oct 05 08:37:07 crc kubenswrapper[4846]: I1005 08:37:07.942438 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/12c63dbb-761d-40ba-b5cd-3407b8b3ff10-utilities\") pod \"12c63dbb-761d-40ba-b5cd-3407b8b3ff10\" (UID: \"12c63dbb-761d-40ba-b5cd-3407b8b3ff10\") " Oct 05 08:37:07 crc kubenswrapper[4846]: I1005 08:37:07.942863 4846 generic.go:334] "Generic (PLEG): container finished" podID="12c63dbb-761d-40ba-b5cd-3407b8b3ff10" containerID="c195a527ff2a920f99ac7ee871ab454a16eec78b65232e1e5a7bf3c313329371" exitCode=0 Oct 05 08:37:07 crc kubenswrapper[4846]: I1005 08:37:07.942944 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rdh4n" event={"ID":"12c63dbb-761d-40ba-b5cd-3407b8b3ff10","Type":"ContainerDied","Data":"c195a527ff2a920f99ac7ee871ab454a16eec78b65232e1e5a7bf3c313329371"} Oct 05 08:37:07 crc kubenswrapper[4846]: I1005 08:37:07.942987 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rdh4n" event={"ID":"12c63dbb-761d-40ba-b5cd-3407b8b3ff10","Type":"ContainerDied","Data":"34863ac31850da2e47a3d9ba4bc54c6ade2976b01fcf6d0864251e7a3f94d8e9"} Oct 05 08:37:07 crc kubenswrapper[4846]: I1005 08:37:07.943017 4846 scope.go:117] "RemoveContainer" containerID="c195a527ff2a920f99ac7ee871ab454a16eec78b65232e1e5a7bf3c313329371" Oct 05 08:37:07 crc kubenswrapper[4846]: I1005 08:37:07.943457 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rdh4n" Oct 05 08:37:07 crc kubenswrapper[4846]: I1005 08:37:07.944537 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/12c63dbb-761d-40ba-b5cd-3407b8b3ff10-utilities" (OuterVolumeSpecName: "utilities") pod "12c63dbb-761d-40ba-b5cd-3407b8b3ff10" (UID: "12c63dbb-761d-40ba-b5cd-3407b8b3ff10"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 08:37:07 crc kubenswrapper[4846]: I1005 08:37:07.953580 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/12c63dbb-761d-40ba-b5cd-3407b8b3ff10-kube-api-access-hvn8l" (OuterVolumeSpecName: "kube-api-access-hvn8l") pod "12c63dbb-761d-40ba-b5cd-3407b8b3ff10" (UID: "12c63dbb-761d-40ba-b5cd-3407b8b3ff10"). InnerVolumeSpecName "kube-api-access-hvn8l". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:37:07 crc kubenswrapper[4846]: I1005 08:37:07.990489 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/12c63dbb-761d-40ba-b5cd-3407b8b3ff10-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "12c63dbb-761d-40ba-b5cd-3407b8b3ff10" (UID: "12c63dbb-761d-40ba-b5cd-3407b8b3ff10"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 08:37:08 crc kubenswrapper[4846]: I1005 08:37:08.022257 4846 scope.go:117] "RemoveContainer" containerID="7bee4128a22582828c0ffa96f85fb33165f690c68642cc81ff53ec2105b2e1dd" Oct 05 08:37:08 crc kubenswrapper[4846]: I1005 08:37:08.045754 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hvn8l\" (UniqueName: \"kubernetes.io/projected/12c63dbb-761d-40ba-b5cd-3407b8b3ff10-kube-api-access-hvn8l\") on node \"crc\" DevicePath \"\"" Oct 05 08:37:08 crc kubenswrapper[4846]: I1005 08:37:08.045804 4846 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/12c63dbb-761d-40ba-b5cd-3407b8b3ff10-utilities\") on node \"crc\" DevicePath \"\"" Oct 05 08:37:08 crc kubenswrapper[4846]: I1005 08:37:08.045815 4846 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/12c63dbb-761d-40ba-b5cd-3407b8b3ff10-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 05 08:37:08 crc kubenswrapper[4846]: I1005 08:37:08.050412 4846 scope.go:117] "RemoveContainer" containerID="21e24f7429ce27406685c16462e1bcdfb6b0c3d4b9c963ed32d992047493e7e4" Oct 05 08:37:08 crc kubenswrapper[4846]: I1005 08:37:08.095938 4846 scope.go:117] "RemoveContainer" containerID="c195a527ff2a920f99ac7ee871ab454a16eec78b65232e1e5a7bf3c313329371" Oct 05 08:37:08 crc kubenswrapper[4846]: E1005 08:37:08.096618 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c195a527ff2a920f99ac7ee871ab454a16eec78b65232e1e5a7bf3c313329371\": container with ID starting with c195a527ff2a920f99ac7ee871ab454a16eec78b65232e1e5a7bf3c313329371 not found: ID does not exist" containerID="c195a527ff2a920f99ac7ee871ab454a16eec78b65232e1e5a7bf3c313329371" Oct 05 08:37:08 crc kubenswrapper[4846]: I1005 08:37:08.096679 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c195a527ff2a920f99ac7ee871ab454a16eec78b65232e1e5a7bf3c313329371"} err="failed to get container status \"c195a527ff2a920f99ac7ee871ab454a16eec78b65232e1e5a7bf3c313329371\": rpc error: code = NotFound desc = could not find container \"c195a527ff2a920f99ac7ee871ab454a16eec78b65232e1e5a7bf3c313329371\": container with ID starting with c195a527ff2a920f99ac7ee871ab454a16eec78b65232e1e5a7bf3c313329371 not found: ID does not exist" Oct 05 08:37:08 crc kubenswrapper[4846]: I1005 08:37:08.096713 4846 scope.go:117] "RemoveContainer" containerID="7bee4128a22582828c0ffa96f85fb33165f690c68642cc81ff53ec2105b2e1dd" Oct 05 08:37:08 crc kubenswrapper[4846]: E1005 08:37:08.097254 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7bee4128a22582828c0ffa96f85fb33165f690c68642cc81ff53ec2105b2e1dd\": container with ID starting with 7bee4128a22582828c0ffa96f85fb33165f690c68642cc81ff53ec2105b2e1dd not found: ID does not exist" containerID="7bee4128a22582828c0ffa96f85fb33165f690c68642cc81ff53ec2105b2e1dd" Oct 05 08:37:08 crc kubenswrapper[4846]: I1005 08:37:08.097305 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7bee4128a22582828c0ffa96f85fb33165f690c68642cc81ff53ec2105b2e1dd"} err="failed to get container status \"7bee4128a22582828c0ffa96f85fb33165f690c68642cc81ff53ec2105b2e1dd\": rpc error: code = NotFound desc = could not find container \"7bee4128a22582828c0ffa96f85fb33165f690c68642cc81ff53ec2105b2e1dd\": container with ID starting with 7bee4128a22582828c0ffa96f85fb33165f690c68642cc81ff53ec2105b2e1dd not found: ID does not exist" Oct 05 08:37:08 crc kubenswrapper[4846]: I1005 08:37:08.097325 4846 scope.go:117] "RemoveContainer" containerID="21e24f7429ce27406685c16462e1bcdfb6b0c3d4b9c963ed32d992047493e7e4" Oct 05 08:37:08 crc kubenswrapper[4846]: E1005 08:37:08.097823 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"21e24f7429ce27406685c16462e1bcdfb6b0c3d4b9c963ed32d992047493e7e4\": container with ID starting with 21e24f7429ce27406685c16462e1bcdfb6b0c3d4b9c963ed32d992047493e7e4 not found: ID does not exist" containerID="21e24f7429ce27406685c16462e1bcdfb6b0c3d4b9c963ed32d992047493e7e4" Oct 05 08:37:08 crc kubenswrapper[4846]: I1005 08:37:08.097885 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"21e24f7429ce27406685c16462e1bcdfb6b0c3d4b9c963ed32d992047493e7e4"} err="failed to get container status \"21e24f7429ce27406685c16462e1bcdfb6b0c3d4b9c963ed32d992047493e7e4\": rpc error: code = NotFound desc = could not find container \"21e24f7429ce27406685c16462e1bcdfb6b0c3d4b9c963ed32d992047493e7e4\": container with ID starting with 21e24f7429ce27406685c16462e1bcdfb6b0c3d4b9c963ed32d992047493e7e4 not found: ID does not exist" Oct 05 08:37:08 crc kubenswrapper[4846]: I1005 08:37:08.298594 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-rdh4n"] Oct 05 08:37:08 crc kubenswrapper[4846]: I1005 08:37:08.312048 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-rdh4n"] Oct 05 08:37:08 crc kubenswrapper[4846]: I1005 08:37:08.521721 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="12c63dbb-761d-40ba-b5cd-3407b8b3ff10" path="/var/lib/kubelet/pods/12c63dbb-761d-40ba-b5cd-3407b8b3ff10/volumes" Oct 05 08:37:16 crc kubenswrapper[4846]: I1005 08:37:16.499114 4846 scope.go:117] "RemoveContainer" containerID="ad5522212e6e288d74747979bc3c3cc6cbe3871d7b66a9bb6a2f8a05f575f23e" Oct 05 08:37:16 crc kubenswrapper[4846]: E1005 08:37:16.500496 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 08:37:28 crc kubenswrapper[4846]: I1005 08:37:28.510726 4846 scope.go:117] "RemoveContainer" containerID="ad5522212e6e288d74747979bc3c3cc6cbe3871d7b66a9bb6a2f8a05f575f23e" Oct 05 08:37:28 crc kubenswrapper[4846]: E1005 08:37:28.512642 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 08:37:31 crc kubenswrapper[4846]: I1005 08:37:31.443438 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-vm5lm"] Oct 05 08:37:31 crc kubenswrapper[4846]: E1005 08:37:31.444350 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="12c63dbb-761d-40ba-b5cd-3407b8b3ff10" containerName="extract-content" Oct 05 08:37:31 crc kubenswrapper[4846]: I1005 08:37:31.444371 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="12c63dbb-761d-40ba-b5cd-3407b8b3ff10" containerName="extract-content" Oct 05 08:37:31 crc kubenswrapper[4846]: E1005 08:37:31.444392 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="12c63dbb-761d-40ba-b5cd-3407b8b3ff10" containerName="extract-utilities" Oct 05 08:37:31 crc kubenswrapper[4846]: I1005 08:37:31.444401 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="12c63dbb-761d-40ba-b5cd-3407b8b3ff10" containerName="extract-utilities" Oct 05 08:37:31 crc kubenswrapper[4846]: E1005 08:37:31.444633 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="12c63dbb-761d-40ba-b5cd-3407b8b3ff10" containerName="registry-server" Oct 05 08:37:31 crc kubenswrapper[4846]: I1005 08:37:31.444643 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="12c63dbb-761d-40ba-b5cd-3407b8b3ff10" containerName="registry-server" Oct 05 08:37:31 crc kubenswrapper[4846]: I1005 08:37:31.448024 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="12c63dbb-761d-40ba-b5cd-3407b8b3ff10" containerName="registry-server" Oct 05 08:37:31 crc kubenswrapper[4846]: I1005 08:37:31.487767 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-vm5lm"] Oct 05 08:37:31 crc kubenswrapper[4846]: I1005 08:37:31.487907 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-vm5lm" Oct 05 08:37:31 crc kubenswrapper[4846]: I1005 08:37:31.617300 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-72j7q\" (UniqueName: \"kubernetes.io/projected/8a08c380-eac6-445e-9dfc-9410a9db03f9-kube-api-access-72j7q\") pod \"redhat-marketplace-vm5lm\" (UID: \"8a08c380-eac6-445e-9dfc-9410a9db03f9\") " pod="openshift-marketplace/redhat-marketplace-vm5lm" Oct 05 08:37:31 crc kubenswrapper[4846]: I1005 08:37:31.617787 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8a08c380-eac6-445e-9dfc-9410a9db03f9-catalog-content\") pod \"redhat-marketplace-vm5lm\" (UID: \"8a08c380-eac6-445e-9dfc-9410a9db03f9\") " pod="openshift-marketplace/redhat-marketplace-vm5lm" Oct 05 08:37:31 crc kubenswrapper[4846]: I1005 08:37:31.617873 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8a08c380-eac6-445e-9dfc-9410a9db03f9-utilities\") pod \"redhat-marketplace-vm5lm\" (UID: \"8a08c380-eac6-445e-9dfc-9410a9db03f9\") " pod="openshift-marketplace/redhat-marketplace-vm5lm" Oct 05 08:37:31 crc kubenswrapper[4846]: I1005 08:37:31.720020 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8a08c380-eac6-445e-9dfc-9410a9db03f9-catalog-content\") pod \"redhat-marketplace-vm5lm\" (UID: \"8a08c380-eac6-445e-9dfc-9410a9db03f9\") " pod="openshift-marketplace/redhat-marketplace-vm5lm" Oct 05 08:37:31 crc kubenswrapper[4846]: I1005 08:37:31.720133 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8a08c380-eac6-445e-9dfc-9410a9db03f9-utilities\") pod \"redhat-marketplace-vm5lm\" (UID: \"8a08c380-eac6-445e-9dfc-9410a9db03f9\") " pod="openshift-marketplace/redhat-marketplace-vm5lm" Oct 05 08:37:31 crc kubenswrapper[4846]: I1005 08:37:31.720240 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-72j7q\" (UniqueName: \"kubernetes.io/projected/8a08c380-eac6-445e-9dfc-9410a9db03f9-kube-api-access-72j7q\") pod \"redhat-marketplace-vm5lm\" (UID: \"8a08c380-eac6-445e-9dfc-9410a9db03f9\") " pod="openshift-marketplace/redhat-marketplace-vm5lm" Oct 05 08:37:31 crc kubenswrapper[4846]: I1005 08:37:31.720835 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8a08c380-eac6-445e-9dfc-9410a9db03f9-catalog-content\") pod \"redhat-marketplace-vm5lm\" (UID: \"8a08c380-eac6-445e-9dfc-9410a9db03f9\") " pod="openshift-marketplace/redhat-marketplace-vm5lm" Oct 05 08:37:31 crc kubenswrapper[4846]: I1005 08:37:31.720838 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8a08c380-eac6-445e-9dfc-9410a9db03f9-utilities\") pod \"redhat-marketplace-vm5lm\" (UID: \"8a08c380-eac6-445e-9dfc-9410a9db03f9\") " pod="openshift-marketplace/redhat-marketplace-vm5lm" Oct 05 08:37:31 crc kubenswrapper[4846]: I1005 08:37:31.743059 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-72j7q\" (UniqueName: \"kubernetes.io/projected/8a08c380-eac6-445e-9dfc-9410a9db03f9-kube-api-access-72j7q\") pod \"redhat-marketplace-vm5lm\" (UID: \"8a08c380-eac6-445e-9dfc-9410a9db03f9\") " pod="openshift-marketplace/redhat-marketplace-vm5lm" Oct 05 08:37:31 crc kubenswrapper[4846]: I1005 08:37:31.824728 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-vm5lm" Oct 05 08:37:32 crc kubenswrapper[4846]: I1005 08:37:32.301542 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-vm5lm"] Oct 05 08:37:33 crc kubenswrapper[4846]: I1005 08:37:33.258739 4846 generic.go:334] "Generic (PLEG): container finished" podID="8a08c380-eac6-445e-9dfc-9410a9db03f9" containerID="abc852c58a8fc44779944336b66d7afc59e33db0fe1835eb8876a08fd45dc687" exitCode=0 Oct 05 08:37:33 crc kubenswrapper[4846]: I1005 08:37:33.258838 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vm5lm" event={"ID":"8a08c380-eac6-445e-9dfc-9410a9db03f9","Type":"ContainerDied","Data":"abc852c58a8fc44779944336b66d7afc59e33db0fe1835eb8876a08fd45dc687"} Oct 05 08:37:33 crc kubenswrapper[4846]: I1005 08:37:33.259243 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vm5lm" event={"ID":"8a08c380-eac6-445e-9dfc-9410a9db03f9","Type":"ContainerStarted","Data":"06688fe57252d44b2a757eb8ea6dc5bf97c87a4375d2b29d2d79b444a950a490"} Oct 05 08:37:34 crc kubenswrapper[4846]: I1005 08:37:34.277201 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vm5lm" event={"ID":"8a08c380-eac6-445e-9dfc-9410a9db03f9","Type":"ContainerStarted","Data":"bea30c46651c23585e6fbb48397b20c68d33dab85d50c71ac5f643488361de58"} Oct 05 08:37:35 crc kubenswrapper[4846]: I1005 08:37:35.299899 4846 generic.go:334] "Generic (PLEG): container finished" podID="8a08c380-eac6-445e-9dfc-9410a9db03f9" containerID="bea30c46651c23585e6fbb48397b20c68d33dab85d50c71ac5f643488361de58" exitCode=0 Oct 05 08:37:35 crc kubenswrapper[4846]: I1005 08:37:35.300388 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vm5lm" event={"ID":"8a08c380-eac6-445e-9dfc-9410a9db03f9","Type":"ContainerDied","Data":"bea30c46651c23585e6fbb48397b20c68d33dab85d50c71ac5f643488361de58"} Oct 05 08:37:36 crc kubenswrapper[4846]: I1005 08:37:36.313366 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vm5lm" event={"ID":"8a08c380-eac6-445e-9dfc-9410a9db03f9","Type":"ContainerStarted","Data":"0dfb7d34c38bac1a76382e814de2f507c9a6ddbf11a893f7a26fd9cb53f4055b"} Oct 05 08:37:36 crc kubenswrapper[4846]: I1005 08:37:36.357850 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-vm5lm" podStartSLOduration=2.8387299280000002 podStartE2EDuration="5.357825489s" podCreationTimestamp="2025-10-05 08:37:31 +0000 UTC" firstStartedPulling="2025-10-05 08:37:33.261132786 +0000 UTC m=+6575.501985601" lastFinishedPulling="2025-10-05 08:37:35.780228387 +0000 UTC m=+6578.021081162" observedRunningTime="2025-10-05 08:37:36.342929646 +0000 UTC m=+6578.583782441" watchObservedRunningTime="2025-10-05 08:37:36.357825489 +0000 UTC m=+6578.598678294" Oct 05 08:37:37 crc kubenswrapper[4846]: I1005 08:37:37.149402 4846 scope.go:117] "RemoveContainer" containerID="1257dcf06640a3d51ed6a0d23dcee6a8666095312a504ce29cb7966a45415217" Oct 05 08:37:37 crc kubenswrapper[4846]: I1005 08:37:37.178403 4846 scope.go:117] "RemoveContainer" containerID="88776f196cb8122e8a327c2ad8e8e609afdf0184de66732201683b73784627c9" Oct 05 08:37:37 crc kubenswrapper[4846]: I1005 08:37:37.208217 4846 scope.go:117] "RemoveContainer" containerID="8e74fbdd446dcfc85adc4504e1df3064933b2a956c9c05ac76d77b7781fbdf43" Oct 05 08:37:37 crc kubenswrapper[4846]: I1005 08:37:37.240015 4846 scope.go:117] "RemoveContainer" containerID="2c803c74afd6a83d53fac7c8eda7624c53a6da308d32c58796661464b251b0c1" Oct 05 08:37:39 crc kubenswrapper[4846]: I1005 08:37:39.534526 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-79ntl"] Oct 05 08:37:39 crc kubenswrapper[4846]: I1005 08:37:39.537046 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-79ntl" Oct 05 08:37:39 crc kubenswrapper[4846]: I1005 08:37:39.549888 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-79ntl"] Oct 05 08:37:39 crc kubenswrapper[4846]: I1005 08:37:39.727416 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-spjmv\" (UniqueName: \"kubernetes.io/projected/77a4dea6-c1ff-4b0d-9d03-81ec48f79714-kube-api-access-spjmv\") pod \"redhat-operators-79ntl\" (UID: \"77a4dea6-c1ff-4b0d-9d03-81ec48f79714\") " pod="openshift-marketplace/redhat-operators-79ntl" Oct 05 08:37:39 crc kubenswrapper[4846]: I1005 08:37:39.727737 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/77a4dea6-c1ff-4b0d-9d03-81ec48f79714-utilities\") pod \"redhat-operators-79ntl\" (UID: \"77a4dea6-c1ff-4b0d-9d03-81ec48f79714\") " pod="openshift-marketplace/redhat-operators-79ntl" Oct 05 08:37:39 crc kubenswrapper[4846]: I1005 08:37:39.727860 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/77a4dea6-c1ff-4b0d-9d03-81ec48f79714-catalog-content\") pod \"redhat-operators-79ntl\" (UID: \"77a4dea6-c1ff-4b0d-9d03-81ec48f79714\") " pod="openshift-marketplace/redhat-operators-79ntl" Oct 05 08:37:39 crc kubenswrapper[4846]: I1005 08:37:39.830138 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/77a4dea6-c1ff-4b0d-9d03-81ec48f79714-utilities\") pod \"redhat-operators-79ntl\" (UID: \"77a4dea6-c1ff-4b0d-9d03-81ec48f79714\") " pod="openshift-marketplace/redhat-operators-79ntl" Oct 05 08:37:39 crc kubenswrapper[4846]: I1005 08:37:39.830320 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/77a4dea6-c1ff-4b0d-9d03-81ec48f79714-catalog-content\") pod \"redhat-operators-79ntl\" (UID: \"77a4dea6-c1ff-4b0d-9d03-81ec48f79714\") " pod="openshift-marketplace/redhat-operators-79ntl" Oct 05 08:37:39 crc kubenswrapper[4846]: I1005 08:37:39.830451 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-spjmv\" (UniqueName: \"kubernetes.io/projected/77a4dea6-c1ff-4b0d-9d03-81ec48f79714-kube-api-access-spjmv\") pod \"redhat-operators-79ntl\" (UID: \"77a4dea6-c1ff-4b0d-9d03-81ec48f79714\") " pod="openshift-marketplace/redhat-operators-79ntl" Oct 05 08:37:39 crc kubenswrapper[4846]: I1005 08:37:39.830780 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/77a4dea6-c1ff-4b0d-9d03-81ec48f79714-utilities\") pod \"redhat-operators-79ntl\" (UID: \"77a4dea6-c1ff-4b0d-9d03-81ec48f79714\") " pod="openshift-marketplace/redhat-operators-79ntl" Oct 05 08:37:39 crc kubenswrapper[4846]: I1005 08:37:39.830918 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/77a4dea6-c1ff-4b0d-9d03-81ec48f79714-catalog-content\") pod \"redhat-operators-79ntl\" (UID: \"77a4dea6-c1ff-4b0d-9d03-81ec48f79714\") " pod="openshift-marketplace/redhat-operators-79ntl" Oct 05 08:37:39 crc kubenswrapper[4846]: I1005 08:37:39.851107 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-spjmv\" (UniqueName: \"kubernetes.io/projected/77a4dea6-c1ff-4b0d-9d03-81ec48f79714-kube-api-access-spjmv\") pod \"redhat-operators-79ntl\" (UID: \"77a4dea6-c1ff-4b0d-9d03-81ec48f79714\") " pod="openshift-marketplace/redhat-operators-79ntl" Oct 05 08:37:39 crc kubenswrapper[4846]: I1005 08:37:39.879223 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-79ntl" Oct 05 08:37:40 crc kubenswrapper[4846]: I1005 08:37:40.049965 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-db-create-qd22b"] Oct 05 08:37:40 crc kubenswrapper[4846]: I1005 08:37:40.069079 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-db-create-qd22b"] Oct 05 08:37:40 crc kubenswrapper[4846]: I1005 08:37:40.363227 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-79ntl"] Oct 05 08:37:40 crc kubenswrapper[4846]: I1005 08:37:40.509434 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c651e1a1-e229-40f6-8c40-76b40a301d78" path="/var/lib/kubelet/pods/c651e1a1-e229-40f6-8c40-76b40a301d78/volumes" Oct 05 08:37:41 crc kubenswrapper[4846]: I1005 08:37:41.375622 4846 generic.go:334] "Generic (PLEG): container finished" podID="77a4dea6-c1ff-4b0d-9d03-81ec48f79714" containerID="6b77911080b6abd6e019a418e70639857ad413548f0eaa58667d708b5ce7d271" exitCode=0 Oct 05 08:37:41 crc kubenswrapper[4846]: I1005 08:37:41.375992 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-79ntl" event={"ID":"77a4dea6-c1ff-4b0d-9d03-81ec48f79714","Type":"ContainerDied","Data":"6b77911080b6abd6e019a418e70639857ad413548f0eaa58667d708b5ce7d271"} Oct 05 08:37:41 crc kubenswrapper[4846]: I1005 08:37:41.376023 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-79ntl" event={"ID":"77a4dea6-c1ff-4b0d-9d03-81ec48f79714","Type":"ContainerStarted","Data":"1936a6c641a4d9119944a31005a55d345f8d6458428a4b1f83c7be6cbf26fb28"} Oct 05 08:37:41 crc kubenswrapper[4846]: I1005 08:37:41.498227 4846 scope.go:117] "RemoveContainer" containerID="ad5522212e6e288d74747979bc3c3cc6cbe3871d7b66a9bb6a2f8a05f575f23e" Oct 05 08:37:41 crc kubenswrapper[4846]: E1005 08:37:41.498443 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 08:37:41 crc kubenswrapper[4846]: I1005 08:37:41.825992 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-vm5lm" Oct 05 08:37:41 crc kubenswrapper[4846]: I1005 08:37:41.826034 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-vm5lm" Oct 05 08:37:41 crc kubenswrapper[4846]: I1005 08:37:41.908475 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-vm5lm" Oct 05 08:37:42 crc kubenswrapper[4846]: I1005 08:37:42.395512 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-79ntl" event={"ID":"77a4dea6-c1ff-4b0d-9d03-81ec48f79714","Type":"ContainerStarted","Data":"374329af1391983f13e876b316d2911119a1f414cd98836db8d47e583a72024a"} Oct 05 08:37:42 crc kubenswrapper[4846]: I1005 08:37:42.457477 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-vm5lm" Oct 05 08:37:44 crc kubenswrapper[4846]: I1005 08:37:44.210082 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-vm5lm"] Oct 05 08:37:44 crc kubenswrapper[4846]: I1005 08:37:44.425756 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-vm5lm" podUID="8a08c380-eac6-445e-9dfc-9410a9db03f9" containerName="registry-server" containerID="cri-o://0dfb7d34c38bac1a76382e814de2f507c9a6ddbf11a893f7a26fd9cb53f4055b" gracePeriod=2 Oct 05 08:37:46 crc kubenswrapper[4846]: I1005 08:37:46.048651 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-vm5lm" Oct 05 08:37:46 crc kubenswrapper[4846]: I1005 08:37:46.158345 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-72j7q\" (UniqueName: \"kubernetes.io/projected/8a08c380-eac6-445e-9dfc-9410a9db03f9-kube-api-access-72j7q\") pod \"8a08c380-eac6-445e-9dfc-9410a9db03f9\" (UID: \"8a08c380-eac6-445e-9dfc-9410a9db03f9\") " Oct 05 08:37:46 crc kubenswrapper[4846]: I1005 08:37:46.158611 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8a08c380-eac6-445e-9dfc-9410a9db03f9-utilities\") pod \"8a08c380-eac6-445e-9dfc-9410a9db03f9\" (UID: \"8a08c380-eac6-445e-9dfc-9410a9db03f9\") " Oct 05 08:37:46 crc kubenswrapper[4846]: I1005 08:37:46.158677 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8a08c380-eac6-445e-9dfc-9410a9db03f9-catalog-content\") pod \"8a08c380-eac6-445e-9dfc-9410a9db03f9\" (UID: \"8a08c380-eac6-445e-9dfc-9410a9db03f9\") " Oct 05 08:37:46 crc kubenswrapper[4846]: I1005 08:37:46.159532 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8a08c380-eac6-445e-9dfc-9410a9db03f9-utilities" (OuterVolumeSpecName: "utilities") pod "8a08c380-eac6-445e-9dfc-9410a9db03f9" (UID: "8a08c380-eac6-445e-9dfc-9410a9db03f9"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 08:37:46 crc kubenswrapper[4846]: I1005 08:37:46.163169 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8a08c380-eac6-445e-9dfc-9410a9db03f9-kube-api-access-72j7q" (OuterVolumeSpecName: "kube-api-access-72j7q") pod "8a08c380-eac6-445e-9dfc-9410a9db03f9" (UID: "8a08c380-eac6-445e-9dfc-9410a9db03f9"). InnerVolumeSpecName "kube-api-access-72j7q". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:37:46 crc kubenswrapper[4846]: I1005 08:37:46.169290 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8a08c380-eac6-445e-9dfc-9410a9db03f9-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8a08c380-eac6-445e-9dfc-9410a9db03f9" (UID: "8a08c380-eac6-445e-9dfc-9410a9db03f9"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 08:37:46 crc kubenswrapper[4846]: I1005 08:37:46.261594 4846 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8a08c380-eac6-445e-9dfc-9410a9db03f9-utilities\") on node \"crc\" DevicePath \"\"" Oct 05 08:37:46 crc kubenswrapper[4846]: I1005 08:37:46.261641 4846 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8a08c380-eac6-445e-9dfc-9410a9db03f9-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 05 08:37:46 crc kubenswrapper[4846]: I1005 08:37:46.261661 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-72j7q\" (UniqueName: \"kubernetes.io/projected/8a08c380-eac6-445e-9dfc-9410a9db03f9-kube-api-access-72j7q\") on node \"crc\" DevicePath \"\"" Oct 05 08:37:46 crc kubenswrapper[4846]: I1005 08:37:46.454363 4846 generic.go:334] "Generic (PLEG): container finished" podID="8a08c380-eac6-445e-9dfc-9410a9db03f9" containerID="0dfb7d34c38bac1a76382e814de2f507c9a6ddbf11a893f7a26fd9cb53f4055b" exitCode=0 Oct 05 08:37:46 crc kubenswrapper[4846]: I1005 08:37:46.454473 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-vm5lm" Oct 05 08:37:46 crc kubenswrapper[4846]: I1005 08:37:46.454492 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vm5lm" event={"ID":"8a08c380-eac6-445e-9dfc-9410a9db03f9","Type":"ContainerDied","Data":"0dfb7d34c38bac1a76382e814de2f507c9a6ddbf11a893f7a26fd9cb53f4055b"} Oct 05 08:37:46 crc kubenswrapper[4846]: I1005 08:37:46.454626 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vm5lm" event={"ID":"8a08c380-eac6-445e-9dfc-9410a9db03f9","Type":"ContainerDied","Data":"06688fe57252d44b2a757eb8ea6dc5bf97c87a4375d2b29d2d79b444a950a490"} Oct 05 08:37:46 crc kubenswrapper[4846]: I1005 08:37:46.454659 4846 scope.go:117] "RemoveContainer" containerID="0dfb7d34c38bac1a76382e814de2f507c9a6ddbf11a893f7a26fd9cb53f4055b" Oct 05 08:37:46 crc kubenswrapper[4846]: I1005 08:37:46.461728 4846 generic.go:334] "Generic (PLEG): container finished" podID="77a4dea6-c1ff-4b0d-9d03-81ec48f79714" containerID="374329af1391983f13e876b316d2911119a1f414cd98836db8d47e583a72024a" exitCode=0 Oct 05 08:37:46 crc kubenswrapper[4846]: I1005 08:37:46.461836 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-79ntl" event={"ID":"77a4dea6-c1ff-4b0d-9d03-81ec48f79714","Type":"ContainerDied","Data":"374329af1391983f13e876b316d2911119a1f414cd98836db8d47e583a72024a"} Oct 05 08:37:46 crc kubenswrapper[4846]: I1005 08:37:46.498879 4846 scope.go:117] "RemoveContainer" containerID="bea30c46651c23585e6fbb48397b20c68d33dab85d50c71ac5f643488361de58" Oct 05 08:37:46 crc kubenswrapper[4846]: I1005 08:37:46.529857 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-vm5lm"] Oct 05 08:37:46 crc kubenswrapper[4846]: I1005 08:37:46.534742 4846 scope.go:117] "RemoveContainer" containerID="abc852c58a8fc44779944336b66d7afc59e33db0fe1835eb8876a08fd45dc687" Oct 05 08:37:46 crc kubenswrapper[4846]: I1005 08:37:46.540521 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-vm5lm"] Oct 05 08:37:46 crc kubenswrapper[4846]: I1005 08:37:46.597990 4846 scope.go:117] "RemoveContainer" containerID="0dfb7d34c38bac1a76382e814de2f507c9a6ddbf11a893f7a26fd9cb53f4055b" Oct 05 08:37:46 crc kubenswrapper[4846]: E1005 08:37:46.599614 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0dfb7d34c38bac1a76382e814de2f507c9a6ddbf11a893f7a26fd9cb53f4055b\": container with ID starting with 0dfb7d34c38bac1a76382e814de2f507c9a6ddbf11a893f7a26fd9cb53f4055b not found: ID does not exist" containerID="0dfb7d34c38bac1a76382e814de2f507c9a6ddbf11a893f7a26fd9cb53f4055b" Oct 05 08:37:46 crc kubenswrapper[4846]: I1005 08:37:46.599672 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0dfb7d34c38bac1a76382e814de2f507c9a6ddbf11a893f7a26fd9cb53f4055b"} err="failed to get container status \"0dfb7d34c38bac1a76382e814de2f507c9a6ddbf11a893f7a26fd9cb53f4055b\": rpc error: code = NotFound desc = could not find container \"0dfb7d34c38bac1a76382e814de2f507c9a6ddbf11a893f7a26fd9cb53f4055b\": container with ID starting with 0dfb7d34c38bac1a76382e814de2f507c9a6ddbf11a893f7a26fd9cb53f4055b not found: ID does not exist" Oct 05 08:37:46 crc kubenswrapper[4846]: I1005 08:37:46.599705 4846 scope.go:117] "RemoveContainer" containerID="bea30c46651c23585e6fbb48397b20c68d33dab85d50c71ac5f643488361de58" Oct 05 08:37:46 crc kubenswrapper[4846]: E1005 08:37:46.600135 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bea30c46651c23585e6fbb48397b20c68d33dab85d50c71ac5f643488361de58\": container with ID starting with bea30c46651c23585e6fbb48397b20c68d33dab85d50c71ac5f643488361de58 not found: ID does not exist" containerID="bea30c46651c23585e6fbb48397b20c68d33dab85d50c71ac5f643488361de58" Oct 05 08:37:46 crc kubenswrapper[4846]: I1005 08:37:46.600197 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bea30c46651c23585e6fbb48397b20c68d33dab85d50c71ac5f643488361de58"} err="failed to get container status \"bea30c46651c23585e6fbb48397b20c68d33dab85d50c71ac5f643488361de58\": rpc error: code = NotFound desc = could not find container \"bea30c46651c23585e6fbb48397b20c68d33dab85d50c71ac5f643488361de58\": container with ID starting with bea30c46651c23585e6fbb48397b20c68d33dab85d50c71ac5f643488361de58 not found: ID does not exist" Oct 05 08:37:46 crc kubenswrapper[4846]: I1005 08:37:46.600234 4846 scope.go:117] "RemoveContainer" containerID="abc852c58a8fc44779944336b66d7afc59e33db0fe1835eb8876a08fd45dc687" Oct 05 08:37:46 crc kubenswrapper[4846]: E1005 08:37:46.600598 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"abc852c58a8fc44779944336b66d7afc59e33db0fe1835eb8876a08fd45dc687\": container with ID starting with abc852c58a8fc44779944336b66d7afc59e33db0fe1835eb8876a08fd45dc687 not found: ID does not exist" containerID="abc852c58a8fc44779944336b66d7afc59e33db0fe1835eb8876a08fd45dc687" Oct 05 08:37:46 crc kubenswrapper[4846]: I1005 08:37:46.600632 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"abc852c58a8fc44779944336b66d7afc59e33db0fe1835eb8876a08fd45dc687"} err="failed to get container status \"abc852c58a8fc44779944336b66d7afc59e33db0fe1835eb8876a08fd45dc687\": rpc error: code = NotFound desc = could not find container \"abc852c58a8fc44779944336b66d7afc59e33db0fe1835eb8876a08fd45dc687\": container with ID starting with abc852c58a8fc44779944336b66d7afc59e33db0fe1835eb8876a08fd45dc687 not found: ID does not exist" Oct 05 08:37:47 crc kubenswrapper[4846]: I1005 08:37:47.479804 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-79ntl" event={"ID":"77a4dea6-c1ff-4b0d-9d03-81ec48f79714","Type":"ContainerStarted","Data":"cbec6d828ac900aa57379c9b9df0018376cb0cb4384153c8f8e6fda82c934f89"} Oct 05 08:37:47 crc kubenswrapper[4846]: I1005 08:37:47.511504 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-79ntl" podStartSLOduration=2.95033769 podStartE2EDuration="8.511468984s" podCreationTimestamp="2025-10-05 08:37:39 +0000 UTC" firstStartedPulling="2025-10-05 08:37:41.378780127 +0000 UTC m=+6583.619632912" lastFinishedPulling="2025-10-05 08:37:46.939911411 +0000 UTC m=+6589.180764206" observedRunningTime="2025-10-05 08:37:47.503386071 +0000 UTC m=+6589.744238906" watchObservedRunningTime="2025-10-05 08:37:47.511468984 +0000 UTC m=+6589.752321819" Oct 05 08:37:48 crc kubenswrapper[4846]: I1005 08:37:48.513444 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8a08c380-eac6-445e-9dfc-9410a9db03f9" path="/var/lib/kubelet/pods/8a08c380-eac6-445e-9dfc-9410a9db03f9/volumes" Oct 05 08:37:49 crc kubenswrapper[4846]: I1005 08:37:49.880064 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-79ntl" Oct 05 08:37:49 crc kubenswrapper[4846]: I1005 08:37:49.880455 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-79ntl" Oct 05 08:37:50 crc kubenswrapper[4846]: I1005 08:37:50.954205 4846 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-79ntl" podUID="77a4dea6-c1ff-4b0d-9d03-81ec48f79714" containerName="registry-server" probeResult="failure" output=< Oct 05 08:37:50 crc kubenswrapper[4846]: timeout: failed to connect service ":50051" within 1s Oct 05 08:37:50 crc kubenswrapper[4846]: > Oct 05 08:37:51 crc kubenswrapper[4846]: I1005 08:37:51.025725 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-1991-account-create-bp8jt"] Oct 05 08:37:51 crc kubenswrapper[4846]: I1005 08:37:51.032465 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-1991-account-create-bp8jt"] Oct 05 08:37:52 crc kubenswrapper[4846]: I1005 08:37:52.497753 4846 scope.go:117] "RemoveContainer" containerID="ad5522212e6e288d74747979bc3c3cc6cbe3871d7b66a9bb6a2f8a05f575f23e" Oct 05 08:37:52 crc kubenswrapper[4846]: E1005 08:37:52.498389 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 08:37:52 crc kubenswrapper[4846]: I1005 08:37:52.510386 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8307ddbf-1547-41e5-a501-b2d68e00b333" path="/var/lib/kubelet/pods/8307ddbf-1547-41e5-a501-b2d68e00b333/volumes" Oct 05 08:37:59 crc kubenswrapper[4846]: I1005 08:37:59.943514 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-79ntl" Oct 05 08:38:00 crc kubenswrapper[4846]: I1005 08:38:00.001149 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-79ntl" Oct 05 08:38:00 crc kubenswrapper[4846]: I1005 08:38:00.211631 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-79ntl"] Oct 05 08:38:01 crc kubenswrapper[4846]: I1005 08:38:01.636345 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-79ntl" podUID="77a4dea6-c1ff-4b0d-9d03-81ec48f79714" containerName="registry-server" containerID="cri-o://cbec6d828ac900aa57379c9b9df0018376cb0cb4384153c8f8e6fda82c934f89" gracePeriod=2 Oct 05 08:38:02 crc kubenswrapper[4846]: I1005 08:38:02.196279 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-79ntl" Oct 05 08:38:02 crc kubenswrapper[4846]: I1005 08:38:02.341521 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/77a4dea6-c1ff-4b0d-9d03-81ec48f79714-catalog-content\") pod \"77a4dea6-c1ff-4b0d-9d03-81ec48f79714\" (UID: \"77a4dea6-c1ff-4b0d-9d03-81ec48f79714\") " Oct 05 08:38:02 crc kubenswrapper[4846]: I1005 08:38:02.341645 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-spjmv\" (UniqueName: \"kubernetes.io/projected/77a4dea6-c1ff-4b0d-9d03-81ec48f79714-kube-api-access-spjmv\") pod \"77a4dea6-c1ff-4b0d-9d03-81ec48f79714\" (UID: \"77a4dea6-c1ff-4b0d-9d03-81ec48f79714\") " Oct 05 08:38:02 crc kubenswrapper[4846]: I1005 08:38:02.341750 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/77a4dea6-c1ff-4b0d-9d03-81ec48f79714-utilities\") pod \"77a4dea6-c1ff-4b0d-9d03-81ec48f79714\" (UID: \"77a4dea6-c1ff-4b0d-9d03-81ec48f79714\") " Oct 05 08:38:02 crc kubenswrapper[4846]: I1005 08:38:02.342579 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/77a4dea6-c1ff-4b0d-9d03-81ec48f79714-utilities" (OuterVolumeSpecName: "utilities") pod "77a4dea6-c1ff-4b0d-9d03-81ec48f79714" (UID: "77a4dea6-c1ff-4b0d-9d03-81ec48f79714"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 08:38:02 crc kubenswrapper[4846]: I1005 08:38:02.343087 4846 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/77a4dea6-c1ff-4b0d-9d03-81ec48f79714-utilities\") on node \"crc\" DevicePath \"\"" Oct 05 08:38:02 crc kubenswrapper[4846]: I1005 08:38:02.359256 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/77a4dea6-c1ff-4b0d-9d03-81ec48f79714-kube-api-access-spjmv" (OuterVolumeSpecName: "kube-api-access-spjmv") pod "77a4dea6-c1ff-4b0d-9d03-81ec48f79714" (UID: "77a4dea6-c1ff-4b0d-9d03-81ec48f79714"). InnerVolumeSpecName "kube-api-access-spjmv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:38:02 crc kubenswrapper[4846]: I1005 08:38:02.427856 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/77a4dea6-c1ff-4b0d-9d03-81ec48f79714-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "77a4dea6-c1ff-4b0d-9d03-81ec48f79714" (UID: "77a4dea6-c1ff-4b0d-9d03-81ec48f79714"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 08:38:02 crc kubenswrapper[4846]: I1005 08:38:02.445447 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-spjmv\" (UniqueName: \"kubernetes.io/projected/77a4dea6-c1ff-4b0d-9d03-81ec48f79714-kube-api-access-spjmv\") on node \"crc\" DevicePath \"\"" Oct 05 08:38:02 crc kubenswrapper[4846]: I1005 08:38:02.445483 4846 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/77a4dea6-c1ff-4b0d-9d03-81ec48f79714-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 05 08:38:02 crc kubenswrapper[4846]: I1005 08:38:02.654101 4846 generic.go:334] "Generic (PLEG): container finished" podID="77a4dea6-c1ff-4b0d-9d03-81ec48f79714" containerID="cbec6d828ac900aa57379c9b9df0018376cb0cb4384153c8f8e6fda82c934f89" exitCode=0 Oct 05 08:38:02 crc kubenswrapper[4846]: I1005 08:38:02.654153 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-79ntl" event={"ID":"77a4dea6-c1ff-4b0d-9d03-81ec48f79714","Type":"ContainerDied","Data":"cbec6d828ac900aa57379c9b9df0018376cb0cb4384153c8f8e6fda82c934f89"} Oct 05 08:38:02 crc kubenswrapper[4846]: I1005 08:38:02.654258 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-79ntl" event={"ID":"77a4dea6-c1ff-4b0d-9d03-81ec48f79714","Type":"ContainerDied","Data":"1936a6c641a4d9119944a31005a55d345f8d6458428a4b1f83c7be6cbf26fb28"} Oct 05 08:38:02 crc kubenswrapper[4846]: I1005 08:38:02.654285 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-79ntl" Oct 05 08:38:02 crc kubenswrapper[4846]: I1005 08:38:02.654330 4846 scope.go:117] "RemoveContainer" containerID="cbec6d828ac900aa57379c9b9df0018376cb0cb4384153c8f8e6fda82c934f89" Oct 05 08:38:02 crc kubenswrapper[4846]: I1005 08:38:02.686511 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-79ntl"] Oct 05 08:38:02 crc kubenswrapper[4846]: I1005 08:38:02.705465 4846 scope.go:117] "RemoveContainer" containerID="374329af1391983f13e876b316d2911119a1f414cd98836db8d47e583a72024a" Oct 05 08:38:02 crc kubenswrapper[4846]: I1005 08:38:02.708241 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-79ntl"] Oct 05 08:38:02 crc kubenswrapper[4846]: I1005 08:38:02.736859 4846 scope.go:117] "RemoveContainer" containerID="6b77911080b6abd6e019a418e70639857ad413548f0eaa58667d708b5ce7d271" Oct 05 08:38:02 crc kubenswrapper[4846]: I1005 08:38:02.803332 4846 scope.go:117] "RemoveContainer" containerID="cbec6d828ac900aa57379c9b9df0018376cb0cb4384153c8f8e6fda82c934f89" Oct 05 08:38:02 crc kubenswrapper[4846]: E1005 08:38:02.803894 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cbec6d828ac900aa57379c9b9df0018376cb0cb4384153c8f8e6fda82c934f89\": container with ID starting with cbec6d828ac900aa57379c9b9df0018376cb0cb4384153c8f8e6fda82c934f89 not found: ID does not exist" containerID="cbec6d828ac900aa57379c9b9df0018376cb0cb4384153c8f8e6fda82c934f89" Oct 05 08:38:02 crc kubenswrapper[4846]: I1005 08:38:02.803947 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cbec6d828ac900aa57379c9b9df0018376cb0cb4384153c8f8e6fda82c934f89"} err="failed to get container status \"cbec6d828ac900aa57379c9b9df0018376cb0cb4384153c8f8e6fda82c934f89\": rpc error: code = NotFound desc = could not find container \"cbec6d828ac900aa57379c9b9df0018376cb0cb4384153c8f8e6fda82c934f89\": container with ID starting with cbec6d828ac900aa57379c9b9df0018376cb0cb4384153c8f8e6fda82c934f89 not found: ID does not exist" Oct 05 08:38:02 crc kubenswrapper[4846]: I1005 08:38:02.803984 4846 scope.go:117] "RemoveContainer" containerID="374329af1391983f13e876b316d2911119a1f414cd98836db8d47e583a72024a" Oct 05 08:38:02 crc kubenswrapper[4846]: E1005 08:38:02.804493 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"374329af1391983f13e876b316d2911119a1f414cd98836db8d47e583a72024a\": container with ID starting with 374329af1391983f13e876b316d2911119a1f414cd98836db8d47e583a72024a not found: ID does not exist" containerID="374329af1391983f13e876b316d2911119a1f414cd98836db8d47e583a72024a" Oct 05 08:38:02 crc kubenswrapper[4846]: I1005 08:38:02.804560 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"374329af1391983f13e876b316d2911119a1f414cd98836db8d47e583a72024a"} err="failed to get container status \"374329af1391983f13e876b316d2911119a1f414cd98836db8d47e583a72024a\": rpc error: code = NotFound desc = could not find container \"374329af1391983f13e876b316d2911119a1f414cd98836db8d47e583a72024a\": container with ID starting with 374329af1391983f13e876b316d2911119a1f414cd98836db8d47e583a72024a not found: ID does not exist" Oct 05 08:38:02 crc kubenswrapper[4846]: I1005 08:38:02.804600 4846 scope.go:117] "RemoveContainer" containerID="6b77911080b6abd6e019a418e70639857ad413548f0eaa58667d708b5ce7d271" Oct 05 08:38:02 crc kubenswrapper[4846]: E1005 08:38:02.805104 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6b77911080b6abd6e019a418e70639857ad413548f0eaa58667d708b5ce7d271\": container with ID starting with 6b77911080b6abd6e019a418e70639857ad413548f0eaa58667d708b5ce7d271 not found: ID does not exist" containerID="6b77911080b6abd6e019a418e70639857ad413548f0eaa58667d708b5ce7d271" Oct 05 08:38:02 crc kubenswrapper[4846]: I1005 08:38:02.805141 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6b77911080b6abd6e019a418e70639857ad413548f0eaa58667d708b5ce7d271"} err="failed to get container status \"6b77911080b6abd6e019a418e70639857ad413548f0eaa58667d708b5ce7d271\": rpc error: code = NotFound desc = could not find container \"6b77911080b6abd6e019a418e70639857ad413548f0eaa58667d708b5ce7d271\": container with ID starting with 6b77911080b6abd6e019a418e70639857ad413548f0eaa58667d708b5ce7d271 not found: ID does not exist" Oct 05 08:38:04 crc kubenswrapper[4846]: I1005 08:38:04.512618 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="77a4dea6-c1ff-4b0d-9d03-81ec48f79714" path="/var/lib/kubelet/pods/77a4dea6-c1ff-4b0d-9d03-81ec48f79714/volumes" Oct 05 08:38:05 crc kubenswrapper[4846]: I1005 08:38:05.498437 4846 scope.go:117] "RemoveContainer" containerID="ad5522212e6e288d74747979bc3c3cc6cbe3871d7b66a9bb6a2f8a05f575f23e" Oct 05 08:38:05 crc kubenswrapper[4846]: E1005 08:38:05.498957 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 08:38:06 crc kubenswrapper[4846]: I1005 08:38:06.058548 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-db-sync-cxjgw"] Oct 05 08:38:06 crc kubenswrapper[4846]: I1005 08:38:06.073536 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-db-sync-cxjgw"] Oct 05 08:38:06 crc kubenswrapper[4846]: I1005 08:38:06.512569 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d95e7d3b-24b0-4200-af1a-72c6dc45ef56" path="/var/lib/kubelet/pods/d95e7d3b-24b0-4200-af1a-72c6dc45ef56/volumes" Oct 05 08:38:17 crc kubenswrapper[4846]: I1005 08:38:17.498368 4846 scope.go:117] "RemoveContainer" containerID="ad5522212e6e288d74747979bc3c3cc6cbe3871d7b66a9bb6a2f8a05f575f23e" Oct 05 08:38:17 crc kubenswrapper[4846]: E1005 08:38:17.499409 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 08:38:31 crc kubenswrapper[4846]: I1005 08:38:31.497502 4846 scope.go:117] "RemoveContainer" containerID="ad5522212e6e288d74747979bc3c3cc6cbe3871d7b66a9bb6a2f8a05f575f23e" Oct 05 08:38:32 crc kubenswrapper[4846]: I1005 08:38:32.040109 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" event={"ID":"ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7","Type":"ContainerStarted","Data":"07ae868532de3fcc44478d4a563da307f947fa536b8b00fb4c3a114f24e9dad0"} Oct 05 08:38:37 crc kubenswrapper[4846]: I1005 08:38:37.327351 4846 scope.go:117] "RemoveContainer" containerID="4e0247457600bd829ebc59711475f72260f4bcf0734779b1ccf5eeb7694a54df" Oct 05 08:38:37 crc kubenswrapper[4846]: I1005 08:38:37.368150 4846 scope.go:117] "RemoveContainer" containerID="91cd2e2593dec9c125d5555846c0eb484079601ab70ea8aa9be280027c2d6110" Oct 05 08:38:37 crc kubenswrapper[4846]: I1005 08:38:37.447274 4846 scope.go:117] "RemoveContainer" containerID="660a8e9f27bfd3ab4802ae5d8e70d3d15734e986878cce2205c5ef39c55111ca" Oct 05 08:39:49 crc kubenswrapper[4846]: I1005 08:39:49.858347 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-6qm6h"] Oct 05 08:39:49 crc kubenswrapper[4846]: E1005 08:39:49.862862 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="77a4dea6-c1ff-4b0d-9d03-81ec48f79714" containerName="extract-content" Oct 05 08:39:49 crc kubenswrapper[4846]: I1005 08:39:49.862890 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="77a4dea6-c1ff-4b0d-9d03-81ec48f79714" containerName="extract-content" Oct 05 08:39:49 crc kubenswrapper[4846]: E1005 08:39:49.862946 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="77a4dea6-c1ff-4b0d-9d03-81ec48f79714" containerName="extract-utilities" Oct 05 08:39:49 crc kubenswrapper[4846]: I1005 08:39:49.862956 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="77a4dea6-c1ff-4b0d-9d03-81ec48f79714" containerName="extract-utilities" Oct 05 08:39:49 crc kubenswrapper[4846]: E1005 08:39:49.862997 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8a08c380-eac6-445e-9dfc-9410a9db03f9" containerName="extract-content" Oct 05 08:39:49 crc kubenswrapper[4846]: I1005 08:39:49.863009 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="8a08c380-eac6-445e-9dfc-9410a9db03f9" containerName="extract-content" Oct 05 08:39:49 crc kubenswrapper[4846]: E1005 08:39:49.863032 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="77a4dea6-c1ff-4b0d-9d03-81ec48f79714" containerName="registry-server" Oct 05 08:39:49 crc kubenswrapper[4846]: I1005 08:39:49.863039 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="77a4dea6-c1ff-4b0d-9d03-81ec48f79714" containerName="registry-server" Oct 05 08:39:49 crc kubenswrapper[4846]: E1005 08:39:49.863060 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8a08c380-eac6-445e-9dfc-9410a9db03f9" containerName="extract-utilities" Oct 05 08:39:49 crc kubenswrapper[4846]: I1005 08:39:49.863069 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="8a08c380-eac6-445e-9dfc-9410a9db03f9" containerName="extract-utilities" Oct 05 08:39:49 crc kubenswrapper[4846]: E1005 08:39:49.863093 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8a08c380-eac6-445e-9dfc-9410a9db03f9" containerName="registry-server" Oct 05 08:39:49 crc kubenswrapper[4846]: I1005 08:39:49.863106 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="8a08c380-eac6-445e-9dfc-9410a9db03f9" containerName="registry-server" Oct 05 08:39:49 crc kubenswrapper[4846]: I1005 08:39:49.863418 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="8a08c380-eac6-445e-9dfc-9410a9db03f9" containerName="registry-server" Oct 05 08:39:49 crc kubenswrapper[4846]: I1005 08:39:49.863447 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="77a4dea6-c1ff-4b0d-9d03-81ec48f79714" containerName="registry-server" Oct 05 08:39:49 crc kubenswrapper[4846]: I1005 08:39:49.865351 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6qm6h" Oct 05 08:39:49 crc kubenswrapper[4846]: I1005 08:39:49.883033 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-6qm6h"] Oct 05 08:39:50 crc kubenswrapper[4846]: I1005 08:39:50.018461 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6c7627e5-1c55-4af2-a94f-82f1bfd54b02-utilities\") pod \"certified-operators-6qm6h\" (UID: \"6c7627e5-1c55-4af2-a94f-82f1bfd54b02\") " pod="openshift-marketplace/certified-operators-6qm6h" Oct 05 08:39:50 crc kubenswrapper[4846]: I1005 08:39:50.018552 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5kgdl\" (UniqueName: \"kubernetes.io/projected/6c7627e5-1c55-4af2-a94f-82f1bfd54b02-kube-api-access-5kgdl\") pod \"certified-operators-6qm6h\" (UID: \"6c7627e5-1c55-4af2-a94f-82f1bfd54b02\") " pod="openshift-marketplace/certified-operators-6qm6h" Oct 05 08:39:50 crc kubenswrapper[4846]: I1005 08:39:50.018779 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6c7627e5-1c55-4af2-a94f-82f1bfd54b02-catalog-content\") pod \"certified-operators-6qm6h\" (UID: \"6c7627e5-1c55-4af2-a94f-82f1bfd54b02\") " pod="openshift-marketplace/certified-operators-6qm6h" Oct 05 08:39:50 crc kubenswrapper[4846]: I1005 08:39:50.120587 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6c7627e5-1c55-4af2-a94f-82f1bfd54b02-catalog-content\") pod \"certified-operators-6qm6h\" (UID: \"6c7627e5-1c55-4af2-a94f-82f1bfd54b02\") " pod="openshift-marketplace/certified-operators-6qm6h" Oct 05 08:39:50 crc kubenswrapper[4846]: I1005 08:39:50.120748 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6c7627e5-1c55-4af2-a94f-82f1bfd54b02-utilities\") pod \"certified-operators-6qm6h\" (UID: \"6c7627e5-1c55-4af2-a94f-82f1bfd54b02\") " pod="openshift-marketplace/certified-operators-6qm6h" Oct 05 08:39:50 crc kubenswrapper[4846]: I1005 08:39:50.120831 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5kgdl\" (UniqueName: \"kubernetes.io/projected/6c7627e5-1c55-4af2-a94f-82f1bfd54b02-kube-api-access-5kgdl\") pod \"certified-operators-6qm6h\" (UID: \"6c7627e5-1c55-4af2-a94f-82f1bfd54b02\") " pod="openshift-marketplace/certified-operators-6qm6h" Oct 05 08:39:50 crc kubenswrapper[4846]: I1005 08:39:50.121016 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6c7627e5-1c55-4af2-a94f-82f1bfd54b02-catalog-content\") pod \"certified-operators-6qm6h\" (UID: \"6c7627e5-1c55-4af2-a94f-82f1bfd54b02\") " pod="openshift-marketplace/certified-operators-6qm6h" Oct 05 08:39:50 crc kubenswrapper[4846]: I1005 08:39:50.121249 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6c7627e5-1c55-4af2-a94f-82f1bfd54b02-utilities\") pod \"certified-operators-6qm6h\" (UID: \"6c7627e5-1c55-4af2-a94f-82f1bfd54b02\") " pod="openshift-marketplace/certified-operators-6qm6h" Oct 05 08:39:50 crc kubenswrapper[4846]: I1005 08:39:50.149761 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5kgdl\" (UniqueName: \"kubernetes.io/projected/6c7627e5-1c55-4af2-a94f-82f1bfd54b02-kube-api-access-5kgdl\") pod \"certified-operators-6qm6h\" (UID: \"6c7627e5-1c55-4af2-a94f-82f1bfd54b02\") " pod="openshift-marketplace/certified-operators-6qm6h" Oct 05 08:39:50 crc kubenswrapper[4846]: I1005 08:39:50.232332 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6qm6h" Oct 05 08:39:50 crc kubenswrapper[4846]: I1005 08:39:50.743225 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-6qm6h"] Oct 05 08:39:50 crc kubenswrapper[4846]: I1005 08:39:50.980252 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6qm6h" event={"ID":"6c7627e5-1c55-4af2-a94f-82f1bfd54b02","Type":"ContainerStarted","Data":"61be9b2f684d430a2fed4362716b9fb5e5c413d43faa8249a992c1f4cde06f8b"} Oct 05 08:39:50 crc kubenswrapper[4846]: I1005 08:39:50.980541 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6qm6h" event={"ID":"6c7627e5-1c55-4af2-a94f-82f1bfd54b02","Type":"ContainerStarted","Data":"e5160c7a9c902612bb467f77a9a7b31524ccaa82a0cb9b919ab124b41b9d3520"} Oct 05 08:39:52 crc kubenswrapper[4846]: I1005 08:39:52.000322 4846 generic.go:334] "Generic (PLEG): container finished" podID="6c7627e5-1c55-4af2-a94f-82f1bfd54b02" containerID="61be9b2f684d430a2fed4362716b9fb5e5c413d43faa8249a992c1f4cde06f8b" exitCode=0 Oct 05 08:39:52 crc kubenswrapper[4846]: I1005 08:39:52.001269 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6qm6h" event={"ID":"6c7627e5-1c55-4af2-a94f-82f1bfd54b02","Type":"ContainerDied","Data":"61be9b2f684d430a2fed4362716b9fb5e5c413d43faa8249a992c1f4cde06f8b"} Oct 05 08:39:54 crc kubenswrapper[4846]: I1005 08:39:54.026015 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6qm6h" event={"ID":"6c7627e5-1c55-4af2-a94f-82f1bfd54b02","Type":"ContainerStarted","Data":"cfecf8933f2e38db1804440d9c33153d2822456fdaa53573f99e36901fd1358c"} Oct 05 08:39:55 crc kubenswrapper[4846]: I1005 08:39:55.041107 4846 generic.go:334] "Generic (PLEG): container finished" podID="6c7627e5-1c55-4af2-a94f-82f1bfd54b02" containerID="cfecf8933f2e38db1804440d9c33153d2822456fdaa53573f99e36901fd1358c" exitCode=0 Oct 05 08:39:55 crc kubenswrapper[4846]: I1005 08:39:55.041212 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6qm6h" event={"ID":"6c7627e5-1c55-4af2-a94f-82f1bfd54b02","Type":"ContainerDied","Data":"cfecf8933f2e38db1804440d9c33153d2822456fdaa53573f99e36901fd1358c"} Oct 05 08:39:56 crc kubenswrapper[4846]: I1005 08:39:56.054045 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6qm6h" event={"ID":"6c7627e5-1c55-4af2-a94f-82f1bfd54b02","Type":"ContainerStarted","Data":"5dd0349f92aa8df3c3981bc7efd0f96d16ff836466118381b2674a19003bac5e"} Oct 05 08:39:56 crc kubenswrapper[4846]: I1005 08:39:56.089721 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-6qm6h" podStartSLOduration=3.675853047 podStartE2EDuration="7.089701892s" podCreationTimestamp="2025-10-05 08:39:49 +0000 UTC" firstStartedPulling="2025-10-05 08:39:52.002925365 +0000 UTC m=+6714.243778170" lastFinishedPulling="2025-10-05 08:39:55.41677423 +0000 UTC m=+6717.657627015" observedRunningTime="2025-10-05 08:39:56.080652824 +0000 UTC m=+6718.321505639" watchObservedRunningTime="2025-10-05 08:39:56.089701892 +0000 UTC m=+6718.330554677" Oct 05 08:40:00 crc kubenswrapper[4846]: I1005 08:40:00.232578 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-6qm6h" Oct 05 08:40:00 crc kubenswrapper[4846]: I1005 08:40:00.233150 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-6qm6h" Oct 05 08:40:00 crc kubenswrapper[4846]: I1005 08:40:00.304462 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-6qm6h" Oct 05 08:40:01 crc kubenswrapper[4846]: I1005 08:40:01.152868 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-6qm6h" Oct 05 08:40:01 crc kubenswrapper[4846]: I1005 08:40:01.209528 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-6qm6h"] Oct 05 08:40:03 crc kubenswrapper[4846]: I1005 08:40:03.138558 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-6qm6h" podUID="6c7627e5-1c55-4af2-a94f-82f1bfd54b02" containerName="registry-server" containerID="cri-o://5dd0349f92aa8df3c3981bc7efd0f96d16ff836466118381b2674a19003bac5e" gracePeriod=2 Oct 05 08:40:03 crc kubenswrapper[4846]: I1005 08:40:03.653835 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6qm6h" Oct 05 08:40:03 crc kubenswrapper[4846]: I1005 08:40:03.842869 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6c7627e5-1c55-4af2-a94f-82f1bfd54b02-catalog-content\") pod \"6c7627e5-1c55-4af2-a94f-82f1bfd54b02\" (UID: \"6c7627e5-1c55-4af2-a94f-82f1bfd54b02\") " Oct 05 08:40:03 crc kubenswrapper[4846]: I1005 08:40:03.842947 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6c7627e5-1c55-4af2-a94f-82f1bfd54b02-utilities\") pod \"6c7627e5-1c55-4af2-a94f-82f1bfd54b02\" (UID: \"6c7627e5-1c55-4af2-a94f-82f1bfd54b02\") " Oct 05 08:40:03 crc kubenswrapper[4846]: I1005 08:40:03.843394 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5kgdl\" (UniqueName: \"kubernetes.io/projected/6c7627e5-1c55-4af2-a94f-82f1bfd54b02-kube-api-access-5kgdl\") pod \"6c7627e5-1c55-4af2-a94f-82f1bfd54b02\" (UID: \"6c7627e5-1c55-4af2-a94f-82f1bfd54b02\") " Oct 05 08:40:03 crc kubenswrapper[4846]: I1005 08:40:03.844688 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6c7627e5-1c55-4af2-a94f-82f1bfd54b02-utilities" (OuterVolumeSpecName: "utilities") pod "6c7627e5-1c55-4af2-a94f-82f1bfd54b02" (UID: "6c7627e5-1c55-4af2-a94f-82f1bfd54b02"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 08:40:03 crc kubenswrapper[4846]: I1005 08:40:03.845650 4846 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6c7627e5-1c55-4af2-a94f-82f1bfd54b02-utilities\") on node \"crc\" DevicePath \"\"" Oct 05 08:40:03 crc kubenswrapper[4846]: I1005 08:40:03.855008 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6c7627e5-1c55-4af2-a94f-82f1bfd54b02-kube-api-access-5kgdl" (OuterVolumeSpecName: "kube-api-access-5kgdl") pod "6c7627e5-1c55-4af2-a94f-82f1bfd54b02" (UID: "6c7627e5-1c55-4af2-a94f-82f1bfd54b02"). InnerVolumeSpecName "kube-api-access-5kgdl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:40:03 crc kubenswrapper[4846]: I1005 08:40:03.911049 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6c7627e5-1c55-4af2-a94f-82f1bfd54b02-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6c7627e5-1c55-4af2-a94f-82f1bfd54b02" (UID: "6c7627e5-1c55-4af2-a94f-82f1bfd54b02"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 08:40:03 crc kubenswrapper[4846]: I1005 08:40:03.947367 4846 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6c7627e5-1c55-4af2-a94f-82f1bfd54b02-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 05 08:40:03 crc kubenswrapper[4846]: I1005 08:40:03.947399 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5kgdl\" (UniqueName: \"kubernetes.io/projected/6c7627e5-1c55-4af2-a94f-82f1bfd54b02-kube-api-access-5kgdl\") on node \"crc\" DevicePath \"\"" Oct 05 08:40:04 crc kubenswrapper[4846]: I1005 08:40:04.152018 4846 generic.go:334] "Generic (PLEG): container finished" podID="6c7627e5-1c55-4af2-a94f-82f1bfd54b02" containerID="5dd0349f92aa8df3c3981bc7efd0f96d16ff836466118381b2674a19003bac5e" exitCode=0 Oct 05 08:40:04 crc kubenswrapper[4846]: I1005 08:40:04.152095 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6qm6h" Oct 05 08:40:04 crc kubenswrapper[4846]: I1005 08:40:04.152092 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6qm6h" event={"ID":"6c7627e5-1c55-4af2-a94f-82f1bfd54b02","Type":"ContainerDied","Data":"5dd0349f92aa8df3c3981bc7efd0f96d16ff836466118381b2674a19003bac5e"} Oct 05 08:40:04 crc kubenswrapper[4846]: I1005 08:40:04.152336 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6qm6h" event={"ID":"6c7627e5-1c55-4af2-a94f-82f1bfd54b02","Type":"ContainerDied","Data":"e5160c7a9c902612bb467f77a9a7b31524ccaa82a0cb9b919ab124b41b9d3520"} Oct 05 08:40:04 crc kubenswrapper[4846]: I1005 08:40:04.152378 4846 scope.go:117] "RemoveContainer" containerID="5dd0349f92aa8df3c3981bc7efd0f96d16ff836466118381b2674a19003bac5e" Oct 05 08:40:04 crc kubenswrapper[4846]: I1005 08:40:04.176943 4846 scope.go:117] "RemoveContainer" containerID="cfecf8933f2e38db1804440d9c33153d2822456fdaa53573f99e36901fd1358c" Oct 05 08:40:04 crc kubenswrapper[4846]: I1005 08:40:04.192068 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-6qm6h"] Oct 05 08:40:04 crc kubenswrapper[4846]: I1005 08:40:04.200463 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-6qm6h"] Oct 05 08:40:04 crc kubenswrapper[4846]: I1005 08:40:04.220948 4846 scope.go:117] "RemoveContainer" containerID="61be9b2f684d430a2fed4362716b9fb5e5c413d43faa8249a992c1f4cde06f8b" Oct 05 08:40:04 crc kubenswrapper[4846]: I1005 08:40:04.255722 4846 scope.go:117] "RemoveContainer" containerID="5dd0349f92aa8df3c3981bc7efd0f96d16ff836466118381b2674a19003bac5e" Oct 05 08:40:04 crc kubenswrapper[4846]: E1005 08:40:04.256196 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5dd0349f92aa8df3c3981bc7efd0f96d16ff836466118381b2674a19003bac5e\": container with ID starting with 5dd0349f92aa8df3c3981bc7efd0f96d16ff836466118381b2674a19003bac5e not found: ID does not exist" containerID="5dd0349f92aa8df3c3981bc7efd0f96d16ff836466118381b2674a19003bac5e" Oct 05 08:40:04 crc kubenswrapper[4846]: I1005 08:40:04.256236 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5dd0349f92aa8df3c3981bc7efd0f96d16ff836466118381b2674a19003bac5e"} err="failed to get container status \"5dd0349f92aa8df3c3981bc7efd0f96d16ff836466118381b2674a19003bac5e\": rpc error: code = NotFound desc = could not find container \"5dd0349f92aa8df3c3981bc7efd0f96d16ff836466118381b2674a19003bac5e\": container with ID starting with 5dd0349f92aa8df3c3981bc7efd0f96d16ff836466118381b2674a19003bac5e not found: ID does not exist" Oct 05 08:40:04 crc kubenswrapper[4846]: I1005 08:40:04.256261 4846 scope.go:117] "RemoveContainer" containerID="cfecf8933f2e38db1804440d9c33153d2822456fdaa53573f99e36901fd1358c" Oct 05 08:40:04 crc kubenswrapper[4846]: E1005 08:40:04.256611 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cfecf8933f2e38db1804440d9c33153d2822456fdaa53573f99e36901fd1358c\": container with ID starting with cfecf8933f2e38db1804440d9c33153d2822456fdaa53573f99e36901fd1358c not found: ID does not exist" containerID="cfecf8933f2e38db1804440d9c33153d2822456fdaa53573f99e36901fd1358c" Oct 05 08:40:04 crc kubenswrapper[4846]: I1005 08:40:04.256637 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cfecf8933f2e38db1804440d9c33153d2822456fdaa53573f99e36901fd1358c"} err="failed to get container status \"cfecf8933f2e38db1804440d9c33153d2822456fdaa53573f99e36901fd1358c\": rpc error: code = NotFound desc = could not find container \"cfecf8933f2e38db1804440d9c33153d2822456fdaa53573f99e36901fd1358c\": container with ID starting with cfecf8933f2e38db1804440d9c33153d2822456fdaa53573f99e36901fd1358c not found: ID does not exist" Oct 05 08:40:04 crc kubenswrapper[4846]: I1005 08:40:04.256653 4846 scope.go:117] "RemoveContainer" containerID="61be9b2f684d430a2fed4362716b9fb5e5c413d43faa8249a992c1f4cde06f8b" Oct 05 08:40:04 crc kubenswrapper[4846]: E1005 08:40:04.257053 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"61be9b2f684d430a2fed4362716b9fb5e5c413d43faa8249a992c1f4cde06f8b\": container with ID starting with 61be9b2f684d430a2fed4362716b9fb5e5c413d43faa8249a992c1f4cde06f8b not found: ID does not exist" containerID="61be9b2f684d430a2fed4362716b9fb5e5c413d43faa8249a992c1f4cde06f8b" Oct 05 08:40:04 crc kubenswrapper[4846]: I1005 08:40:04.257087 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"61be9b2f684d430a2fed4362716b9fb5e5c413d43faa8249a992c1f4cde06f8b"} err="failed to get container status \"61be9b2f684d430a2fed4362716b9fb5e5c413d43faa8249a992c1f4cde06f8b\": rpc error: code = NotFound desc = could not find container \"61be9b2f684d430a2fed4362716b9fb5e5c413d43faa8249a992c1f4cde06f8b\": container with ID starting with 61be9b2f684d430a2fed4362716b9fb5e5c413d43faa8249a992c1f4cde06f8b not found: ID does not exist" Oct 05 08:40:04 crc kubenswrapper[4846]: I1005 08:40:04.515613 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6c7627e5-1c55-4af2-a94f-82f1bfd54b02" path="/var/lib/kubelet/pods/6c7627e5-1c55-4af2-a94f-82f1bfd54b02/volumes" Oct 05 08:40:17 crc kubenswrapper[4846]: I1005 08:40:17.057169 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-db-create-7sdkh"] Oct 05 08:40:17 crc kubenswrapper[4846]: I1005 08:40:17.073504 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/aodh-db-create-7sdkh"] Oct 05 08:40:18 crc kubenswrapper[4846]: I1005 08:40:18.516435 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e48f4506-d1b2-4dca-b3d8-cd1693af6c01" path="/var/lib/kubelet/pods/e48f4506-d1b2-4dca-b3d8-cd1693af6c01/volumes" Oct 05 08:40:28 crc kubenswrapper[4846]: I1005 08:40:28.058973 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-9865-account-create-fwqsc"] Oct 05 08:40:28 crc kubenswrapper[4846]: I1005 08:40:28.071927 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/aodh-9865-account-create-fwqsc"] Oct 05 08:40:28 crc kubenswrapper[4846]: I1005 08:40:28.527595 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="331c68dd-4707-4e6c-b6cf-9d9686ef9e7e" path="/var/lib/kubelet/pods/331c68dd-4707-4e6c-b6cf-9d9686ef9e7e/volumes" Oct 05 08:40:37 crc kubenswrapper[4846]: I1005 08:40:37.596869 4846 scope.go:117] "RemoveContainer" containerID="9baf95086c181c1170151d82782e7e4354fee9158e2a84f52fa098572c09c14d" Oct 05 08:40:37 crc kubenswrapper[4846]: I1005 08:40:37.659968 4846 scope.go:117] "RemoveContainer" containerID="c579311872b97cc6e9253bc2a70b0c9d987477a88ca327922851f37cd7e6722e" Oct 05 08:40:39 crc kubenswrapper[4846]: I1005 08:40:39.038486 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-db-sync-5tnzg"] Oct 05 08:40:39 crc kubenswrapper[4846]: I1005 08:40:39.050571 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/aodh-db-sync-5tnzg"] Oct 05 08:40:40 crc kubenswrapper[4846]: I1005 08:40:40.522381 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="19856b5e-44a2-4b5e-bee6-2cde2af29996" path="/var/lib/kubelet/pods/19856b5e-44a2-4b5e-bee6-2cde2af29996/volumes" Oct 05 08:40:53 crc kubenswrapper[4846]: I1005 08:40:53.325768 4846 patch_prober.go:28] interesting pod/machine-config-daemon-fscvf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 08:40:53 crc kubenswrapper[4846]: I1005 08:40:53.326566 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 08:41:23 crc kubenswrapper[4846]: I1005 08:41:23.324549 4846 patch_prober.go:28] interesting pod/machine-config-daemon-fscvf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 08:41:23 crc kubenswrapper[4846]: I1005 08:41:23.325347 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 08:41:37 crc kubenswrapper[4846]: I1005 08:41:37.835759 4846 scope.go:117] "RemoveContainer" containerID="2dc7314cc10ac3e793d12d8c243041065e139b7ece7dc8cca1794cec5b2f85f5" Oct 05 08:41:53 crc kubenswrapper[4846]: I1005 08:41:53.325480 4846 patch_prober.go:28] interesting pod/machine-config-daemon-fscvf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 08:41:53 crc kubenswrapper[4846]: I1005 08:41:53.326119 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 08:41:53 crc kubenswrapper[4846]: I1005 08:41:53.326223 4846 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" Oct 05 08:41:53 crc kubenswrapper[4846]: I1005 08:41:53.327102 4846 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"07ae868532de3fcc44478d4a563da307f947fa536b8b00fb4c3a114f24e9dad0"} pod="openshift-machine-config-operator/machine-config-daemon-fscvf" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 05 08:41:53 crc kubenswrapper[4846]: I1005 08:41:53.327196 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" containerName="machine-config-daemon" containerID="cri-o://07ae868532de3fcc44478d4a563da307f947fa536b8b00fb4c3a114f24e9dad0" gracePeriod=600 Oct 05 08:41:54 crc kubenswrapper[4846]: I1005 08:41:54.470482 4846 generic.go:334] "Generic (PLEG): container finished" podID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" containerID="07ae868532de3fcc44478d4a563da307f947fa536b8b00fb4c3a114f24e9dad0" exitCode=0 Oct 05 08:41:54 crc kubenswrapper[4846]: I1005 08:41:54.470556 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" event={"ID":"ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7","Type":"ContainerDied","Data":"07ae868532de3fcc44478d4a563da307f947fa536b8b00fb4c3a114f24e9dad0"} Oct 05 08:41:54 crc kubenswrapper[4846]: I1005 08:41:54.470996 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" event={"ID":"ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7","Type":"ContainerStarted","Data":"0d84b906773347b38ec2ce550f580254430238b4374f3171cb7453fa65259bcc"} Oct 05 08:41:54 crc kubenswrapper[4846]: I1005 08:41:54.471025 4846 scope.go:117] "RemoveContainer" containerID="ad5522212e6e288d74747979bc3c3cc6cbe3871d7b66a9bb6a2f8a05f575f23e" Oct 05 08:42:59 crc kubenswrapper[4846]: I1005 08:42:59.221002 4846 generic.go:334] "Generic (PLEG): container finished" podID="19baf6c0-a301-49fe-be7b-6682904106b9" containerID="5ba7013871b15a45e9c5ea364c64a9710a10033aaa75a7c468cacbde3028d3f9" exitCode=0 Oct 05 08:42:59 crc kubenswrapper[4846]: I1005 08:42:59.221095 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-ptzpk" event={"ID":"19baf6c0-a301-49fe-be7b-6682904106b9","Type":"ContainerDied","Data":"5ba7013871b15a45e9c5ea364c64a9710a10033aaa75a7c468cacbde3028d3f9"} Oct 05 08:43:00 crc kubenswrapper[4846]: I1005 08:43:00.759097 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-ptzpk" Oct 05 08:43:00 crc kubenswrapper[4846]: I1005 08:43:00.895218 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/19baf6c0-a301-49fe-be7b-6682904106b9-ssh-key\") pod \"19baf6c0-a301-49fe-be7b-6682904106b9\" (UID: \"19baf6c0-a301-49fe-be7b-6682904106b9\") " Oct 05 08:43:00 crc kubenswrapper[4846]: I1005 08:43:00.895327 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-brqj8\" (UniqueName: \"kubernetes.io/projected/19baf6c0-a301-49fe-be7b-6682904106b9-kube-api-access-brqj8\") pod \"19baf6c0-a301-49fe-be7b-6682904106b9\" (UID: \"19baf6c0-a301-49fe-be7b-6682904106b9\") " Oct 05 08:43:00 crc kubenswrapper[4846]: I1005 08:43:00.895433 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/19baf6c0-a301-49fe-be7b-6682904106b9-inventory\") pod \"19baf6c0-a301-49fe-be7b-6682904106b9\" (UID: \"19baf6c0-a301-49fe-be7b-6682904106b9\") " Oct 05 08:43:00 crc kubenswrapper[4846]: I1005 08:43:00.895470 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tripleo-cleanup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/19baf6c0-a301-49fe-be7b-6682904106b9-tripleo-cleanup-combined-ca-bundle\") pod \"19baf6c0-a301-49fe-be7b-6682904106b9\" (UID: \"19baf6c0-a301-49fe-be7b-6682904106b9\") " Oct 05 08:43:00 crc kubenswrapper[4846]: I1005 08:43:00.902296 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/19baf6c0-a301-49fe-be7b-6682904106b9-tripleo-cleanup-combined-ca-bundle" (OuterVolumeSpecName: "tripleo-cleanup-combined-ca-bundle") pod "19baf6c0-a301-49fe-be7b-6682904106b9" (UID: "19baf6c0-a301-49fe-be7b-6682904106b9"). InnerVolumeSpecName "tripleo-cleanup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:43:00 crc kubenswrapper[4846]: I1005 08:43:00.902389 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/19baf6c0-a301-49fe-be7b-6682904106b9-kube-api-access-brqj8" (OuterVolumeSpecName: "kube-api-access-brqj8") pod "19baf6c0-a301-49fe-be7b-6682904106b9" (UID: "19baf6c0-a301-49fe-be7b-6682904106b9"). InnerVolumeSpecName "kube-api-access-brqj8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:43:00 crc kubenswrapper[4846]: I1005 08:43:00.936908 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/19baf6c0-a301-49fe-be7b-6682904106b9-inventory" (OuterVolumeSpecName: "inventory") pod "19baf6c0-a301-49fe-be7b-6682904106b9" (UID: "19baf6c0-a301-49fe-be7b-6682904106b9"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:43:00 crc kubenswrapper[4846]: I1005 08:43:00.966636 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/19baf6c0-a301-49fe-be7b-6682904106b9-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "19baf6c0-a301-49fe-be7b-6682904106b9" (UID: "19baf6c0-a301-49fe-be7b-6682904106b9"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:43:00 crc kubenswrapper[4846]: I1005 08:43:00.998563 4846 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/19baf6c0-a301-49fe-be7b-6682904106b9-inventory\") on node \"crc\" DevicePath \"\"" Oct 05 08:43:00 crc kubenswrapper[4846]: I1005 08:43:00.998634 4846 reconciler_common.go:293] "Volume detached for volume \"tripleo-cleanup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/19baf6c0-a301-49fe-be7b-6682904106b9-tripleo-cleanup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 08:43:00 crc kubenswrapper[4846]: I1005 08:43:00.998675 4846 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/19baf6c0-a301-49fe-be7b-6682904106b9-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 05 08:43:00 crc kubenswrapper[4846]: I1005 08:43:00.998694 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-brqj8\" (UniqueName: \"kubernetes.io/projected/19baf6c0-a301-49fe-be7b-6682904106b9-kube-api-access-brqj8\") on node \"crc\" DevicePath \"\"" Oct 05 08:43:01 crc kubenswrapper[4846]: I1005 08:43:01.244557 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-ptzpk" event={"ID":"19baf6c0-a301-49fe-be7b-6682904106b9","Type":"ContainerDied","Data":"6727932025eab23e57fab3c11862dbb665b4f88404d0c9381ae8b575e93100c2"} Oct 05 08:43:01 crc kubenswrapper[4846]: I1005 08:43:01.244626 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6727932025eab23e57fab3c11862dbb665b4f88404d0c9381ae8b575e93100c2" Oct 05 08:43:01 crc kubenswrapper[4846]: I1005 08:43:01.244652 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-ptzpk" Oct 05 08:43:09 crc kubenswrapper[4846]: I1005 08:43:09.874272 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/bootstrap-openstack-openstack-cell1-rxf6k"] Oct 05 08:43:09 crc kubenswrapper[4846]: E1005 08:43:09.877104 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="19baf6c0-a301-49fe-be7b-6682904106b9" containerName="tripleo-cleanup-tripleo-cleanup-openstack-cell1" Oct 05 08:43:09 crc kubenswrapper[4846]: I1005 08:43:09.877142 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="19baf6c0-a301-49fe-be7b-6682904106b9" containerName="tripleo-cleanup-tripleo-cleanup-openstack-cell1" Oct 05 08:43:09 crc kubenswrapper[4846]: E1005 08:43:09.877260 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6c7627e5-1c55-4af2-a94f-82f1bfd54b02" containerName="registry-server" Oct 05 08:43:09 crc kubenswrapper[4846]: I1005 08:43:09.877275 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="6c7627e5-1c55-4af2-a94f-82f1bfd54b02" containerName="registry-server" Oct 05 08:43:09 crc kubenswrapper[4846]: E1005 08:43:09.877322 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6c7627e5-1c55-4af2-a94f-82f1bfd54b02" containerName="extract-utilities" Oct 05 08:43:09 crc kubenswrapper[4846]: I1005 08:43:09.877337 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="6c7627e5-1c55-4af2-a94f-82f1bfd54b02" containerName="extract-utilities" Oct 05 08:43:09 crc kubenswrapper[4846]: E1005 08:43:09.877404 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6c7627e5-1c55-4af2-a94f-82f1bfd54b02" containerName="extract-content" Oct 05 08:43:09 crc kubenswrapper[4846]: I1005 08:43:09.877416 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="6c7627e5-1c55-4af2-a94f-82f1bfd54b02" containerName="extract-content" Oct 05 08:43:09 crc kubenswrapper[4846]: I1005 08:43:09.878816 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="6c7627e5-1c55-4af2-a94f-82f1bfd54b02" containerName="registry-server" Oct 05 08:43:09 crc kubenswrapper[4846]: I1005 08:43:09.878885 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="19baf6c0-a301-49fe-be7b-6682904106b9" containerName="tripleo-cleanup-tripleo-cleanup-openstack-cell1" Oct 05 08:43:09 crc kubenswrapper[4846]: I1005 08:43:09.881581 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-openstack-openstack-cell1-rxf6k" Oct 05 08:43:09 crc kubenswrapper[4846]: I1005 08:43:09.886569 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 05 08:43:09 crc kubenswrapper[4846]: I1005 08:43:09.892154 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Oct 05 08:43:09 crc kubenswrapper[4846]: I1005 08:43:09.897711 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-7687s" Oct 05 08:43:09 crc kubenswrapper[4846]: I1005 08:43:09.899652 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-openstack-openstack-cell1-rxf6k"] Oct 05 08:43:09 crc kubenswrapper[4846]: I1005 08:43:09.902391 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Oct 05 08:43:10 crc kubenswrapper[4846]: I1005 08:43:10.028863 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f55e5fdc-aff8-4ebe-afa7-a570e18bf335-bootstrap-combined-ca-bundle\") pod \"bootstrap-openstack-openstack-cell1-rxf6k\" (UID: \"f55e5fdc-aff8-4ebe-afa7-a570e18bf335\") " pod="openstack/bootstrap-openstack-openstack-cell1-rxf6k" Oct 05 08:43:10 crc kubenswrapper[4846]: I1005 08:43:10.028932 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mpk2z\" (UniqueName: \"kubernetes.io/projected/f55e5fdc-aff8-4ebe-afa7-a570e18bf335-kube-api-access-mpk2z\") pod \"bootstrap-openstack-openstack-cell1-rxf6k\" (UID: \"f55e5fdc-aff8-4ebe-afa7-a570e18bf335\") " pod="openstack/bootstrap-openstack-openstack-cell1-rxf6k" Oct 05 08:43:10 crc kubenswrapper[4846]: I1005 08:43:10.029033 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f55e5fdc-aff8-4ebe-afa7-a570e18bf335-ssh-key\") pod \"bootstrap-openstack-openstack-cell1-rxf6k\" (UID: \"f55e5fdc-aff8-4ebe-afa7-a570e18bf335\") " pod="openstack/bootstrap-openstack-openstack-cell1-rxf6k" Oct 05 08:43:10 crc kubenswrapper[4846]: I1005 08:43:10.029071 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f55e5fdc-aff8-4ebe-afa7-a570e18bf335-inventory\") pod \"bootstrap-openstack-openstack-cell1-rxf6k\" (UID: \"f55e5fdc-aff8-4ebe-afa7-a570e18bf335\") " pod="openstack/bootstrap-openstack-openstack-cell1-rxf6k" Oct 05 08:43:10 crc kubenswrapper[4846]: I1005 08:43:10.131465 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f55e5fdc-aff8-4ebe-afa7-a570e18bf335-ssh-key\") pod \"bootstrap-openstack-openstack-cell1-rxf6k\" (UID: \"f55e5fdc-aff8-4ebe-afa7-a570e18bf335\") " pod="openstack/bootstrap-openstack-openstack-cell1-rxf6k" Oct 05 08:43:10 crc kubenswrapper[4846]: I1005 08:43:10.131549 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f55e5fdc-aff8-4ebe-afa7-a570e18bf335-inventory\") pod \"bootstrap-openstack-openstack-cell1-rxf6k\" (UID: \"f55e5fdc-aff8-4ebe-afa7-a570e18bf335\") " pod="openstack/bootstrap-openstack-openstack-cell1-rxf6k" Oct 05 08:43:10 crc kubenswrapper[4846]: I1005 08:43:10.131681 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f55e5fdc-aff8-4ebe-afa7-a570e18bf335-bootstrap-combined-ca-bundle\") pod \"bootstrap-openstack-openstack-cell1-rxf6k\" (UID: \"f55e5fdc-aff8-4ebe-afa7-a570e18bf335\") " pod="openstack/bootstrap-openstack-openstack-cell1-rxf6k" Oct 05 08:43:10 crc kubenswrapper[4846]: I1005 08:43:10.131725 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mpk2z\" (UniqueName: \"kubernetes.io/projected/f55e5fdc-aff8-4ebe-afa7-a570e18bf335-kube-api-access-mpk2z\") pod \"bootstrap-openstack-openstack-cell1-rxf6k\" (UID: \"f55e5fdc-aff8-4ebe-afa7-a570e18bf335\") " pod="openstack/bootstrap-openstack-openstack-cell1-rxf6k" Oct 05 08:43:10 crc kubenswrapper[4846]: I1005 08:43:10.139515 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f55e5fdc-aff8-4ebe-afa7-a570e18bf335-bootstrap-combined-ca-bundle\") pod \"bootstrap-openstack-openstack-cell1-rxf6k\" (UID: \"f55e5fdc-aff8-4ebe-afa7-a570e18bf335\") " pod="openstack/bootstrap-openstack-openstack-cell1-rxf6k" Oct 05 08:43:10 crc kubenswrapper[4846]: I1005 08:43:10.145303 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f55e5fdc-aff8-4ebe-afa7-a570e18bf335-ssh-key\") pod \"bootstrap-openstack-openstack-cell1-rxf6k\" (UID: \"f55e5fdc-aff8-4ebe-afa7-a570e18bf335\") " pod="openstack/bootstrap-openstack-openstack-cell1-rxf6k" Oct 05 08:43:10 crc kubenswrapper[4846]: I1005 08:43:10.151141 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mpk2z\" (UniqueName: \"kubernetes.io/projected/f55e5fdc-aff8-4ebe-afa7-a570e18bf335-kube-api-access-mpk2z\") pod \"bootstrap-openstack-openstack-cell1-rxf6k\" (UID: \"f55e5fdc-aff8-4ebe-afa7-a570e18bf335\") " pod="openstack/bootstrap-openstack-openstack-cell1-rxf6k" Oct 05 08:43:10 crc kubenswrapper[4846]: I1005 08:43:10.152509 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f55e5fdc-aff8-4ebe-afa7-a570e18bf335-inventory\") pod \"bootstrap-openstack-openstack-cell1-rxf6k\" (UID: \"f55e5fdc-aff8-4ebe-afa7-a570e18bf335\") " pod="openstack/bootstrap-openstack-openstack-cell1-rxf6k" Oct 05 08:43:10 crc kubenswrapper[4846]: I1005 08:43:10.227339 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-openstack-openstack-cell1-rxf6k" Oct 05 08:43:10 crc kubenswrapper[4846]: I1005 08:43:10.785527 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-openstack-openstack-cell1-rxf6k"] Oct 05 08:43:10 crc kubenswrapper[4846]: W1005 08:43:10.788230 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf55e5fdc_aff8_4ebe_afa7_a570e18bf335.slice/crio-a0d4d80548f05771bac2d8b0d6168181e7e5e95074982f9680378d2cdfa01df5 WatchSource:0}: Error finding container a0d4d80548f05771bac2d8b0d6168181e7e5e95074982f9680378d2cdfa01df5: Status 404 returned error can't find the container with id a0d4d80548f05771bac2d8b0d6168181e7e5e95074982f9680378d2cdfa01df5 Oct 05 08:43:10 crc kubenswrapper[4846]: I1005 08:43:10.790966 4846 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 05 08:43:11 crc kubenswrapper[4846]: I1005 08:43:11.353306 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-openstack-openstack-cell1-rxf6k" event={"ID":"f55e5fdc-aff8-4ebe-afa7-a570e18bf335","Type":"ContainerStarted","Data":"a0d4d80548f05771bac2d8b0d6168181e7e5e95074982f9680378d2cdfa01df5"} Oct 05 08:43:12 crc kubenswrapper[4846]: I1005 08:43:12.381317 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-openstack-openstack-cell1-rxf6k" event={"ID":"f55e5fdc-aff8-4ebe-afa7-a570e18bf335","Type":"ContainerStarted","Data":"d56fcf1bb4436455b77d9095e9765666332147eaff35902e39da0154cab4e14d"} Oct 05 08:43:12 crc kubenswrapper[4846]: I1005 08:43:12.412515 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/bootstrap-openstack-openstack-cell1-rxf6k" podStartSLOduration=2.877014601 podStartE2EDuration="3.412489517s" podCreationTimestamp="2025-10-05 08:43:09 +0000 UTC" firstStartedPulling="2025-10-05 08:43:10.790743775 +0000 UTC m=+6913.031596550" lastFinishedPulling="2025-10-05 08:43:11.326218681 +0000 UTC m=+6913.567071466" observedRunningTime="2025-10-05 08:43:12.406138184 +0000 UTC m=+6914.646990979" watchObservedRunningTime="2025-10-05 08:43:12.412489517 +0000 UTC m=+6914.653342302" Oct 05 08:43:53 crc kubenswrapper[4846]: I1005 08:43:53.325074 4846 patch_prober.go:28] interesting pod/machine-config-daemon-fscvf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 08:43:53 crc kubenswrapper[4846]: I1005 08:43:53.325635 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 08:44:23 crc kubenswrapper[4846]: I1005 08:44:23.324451 4846 patch_prober.go:28] interesting pod/machine-config-daemon-fscvf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 08:44:23 crc kubenswrapper[4846]: I1005 08:44:23.326556 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 08:44:53 crc kubenswrapper[4846]: I1005 08:44:53.325119 4846 patch_prober.go:28] interesting pod/machine-config-daemon-fscvf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 08:44:53 crc kubenswrapper[4846]: I1005 08:44:53.325956 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 08:44:53 crc kubenswrapper[4846]: I1005 08:44:53.326030 4846 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" Oct 05 08:44:53 crc kubenswrapper[4846]: I1005 08:44:53.327273 4846 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"0d84b906773347b38ec2ce550f580254430238b4374f3171cb7453fa65259bcc"} pod="openshift-machine-config-operator/machine-config-daemon-fscvf" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 05 08:44:53 crc kubenswrapper[4846]: I1005 08:44:53.327391 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" containerName="machine-config-daemon" containerID="cri-o://0d84b906773347b38ec2ce550f580254430238b4374f3171cb7453fa65259bcc" gracePeriod=600 Oct 05 08:44:53 crc kubenswrapper[4846]: E1005 08:44:53.453477 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 08:44:53 crc kubenswrapper[4846]: I1005 08:44:53.528993 4846 generic.go:334] "Generic (PLEG): container finished" podID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" containerID="0d84b906773347b38ec2ce550f580254430238b4374f3171cb7453fa65259bcc" exitCode=0 Oct 05 08:44:53 crc kubenswrapper[4846]: I1005 08:44:53.529039 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" event={"ID":"ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7","Type":"ContainerDied","Data":"0d84b906773347b38ec2ce550f580254430238b4374f3171cb7453fa65259bcc"} Oct 05 08:44:53 crc kubenswrapper[4846]: I1005 08:44:53.529074 4846 scope.go:117] "RemoveContainer" containerID="07ae868532de3fcc44478d4a563da307f947fa536b8b00fb4c3a114f24e9dad0" Oct 05 08:44:53 crc kubenswrapper[4846]: I1005 08:44:53.529857 4846 scope.go:117] "RemoveContainer" containerID="0d84b906773347b38ec2ce550f580254430238b4374f3171cb7453fa65259bcc" Oct 05 08:44:53 crc kubenswrapper[4846]: E1005 08:44:53.530222 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 08:45:00 crc kubenswrapper[4846]: I1005 08:45:00.168994 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29327565-p7hbh"] Oct 05 08:45:00 crc kubenswrapper[4846]: I1005 08:45:00.171161 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29327565-p7hbh" Oct 05 08:45:00 crc kubenswrapper[4846]: I1005 08:45:00.173598 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 05 08:45:00 crc kubenswrapper[4846]: I1005 08:45:00.175168 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 05 08:45:00 crc kubenswrapper[4846]: I1005 08:45:00.181477 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29327565-p7hbh"] Oct 05 08:45:00 crc kubenswrapper[4846]: I1005 08:45:00.226978 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e9a16960-8bb2-43a3-bbc4-70a135eb2449-secret-volume\") pod \"collect-profiles-29327565-p7hbh\" (UID: \"e9a16960-8bb2-43a3-bbc4-70a135eb2449\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327565-p7hbh" Oct 05 08:45:00 crc kubenswrapper[4846]: I1005 08:45:00.227258 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e9a16960-8bb2-43a3-bbc4-70a135eb2449-config-volume\") pod \"collect-profiles-29327565-p7hbh\" (UID: \"e9a16960-8bb2-43a3-bbc4-70a135eb2449\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327565-p7hbh" Oct 05 08:45:00 crc kubenswrapper[4846]: I1005 08:45:00.227548 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r8g59\" (UniqueName: \"kubernetes.io/projected/e9a16960-8bb2-43a3-bbc4-70a135eb2449-kube-api-access-r8g59\") pod \"collect-profiles-29327565-p7hbh\" (UID: \"e9a16960-8bb2-43a3-bbc4-70a135eb2449\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327565-p7hbh" Oct 05 08:45:00 crc kubenswrapper[4846]: I1005 08:45:00.330030 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r8g59\" (UniqueName: \"kubernetes.io/projected/e9a16960-8bb2-43a3-bbc4-70a135eb2449-kube-api-access-r8g59\") pod \"collect-profiles-29327565-p7hbh\" (UID: \"e9a16960-8bb2-43a3-bbc4-70a135eb2449\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327565-p7hbh" Oct 05 08:45:00 crc kubenswrapper[4846]: I1005 08:45:00.330100 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e9a16960-8bb2-43a3-bbc4-70a135eb2449-secret-volume\") pod \"collect-profiles-29327565-p7hbh\" (UID: \"e9a16960-8bb2-43a3-bbc4-70a135eb2449\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327565-p7hbh" Oct 05 08:45:00 crc kubenswrapper[4846]: I1005 08:45:00.330143 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e9a16960-8bb2-43a3-bbc4-70a135eb2449-config-volume\") pod \"collect-profiles-29327565-p7hbh\" (UID: \"e9a16960-8bb2-43a3-bbc4-70a135eb2449\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327565-p7hbh" Oct 05 08:45:00 crc kubenswrapper[4846]: I1005 08:45:00.331067 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e9a16960-8bb2-43a3-bbc4-70a135eb2449-config-volume\") pod \"collect-profiles-29327565-p7hbh\" (UID: \"e9a16960-8bb2-43a3-bbc4-70a135eb2449\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327565-p7hbh" Oct 05 08:45:00 crc kubenswrapper[4846]: I1005 08:45:00.340022 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e9a16960-8bb2-43a3-bbc4-70a135eb2449-secret-volume\") pod \"collect-profiles-29327565-p7hbh\" (UID: \"e9a16960-8bb2-43a3-bbc4-70a135eb2449\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327565-p7hbh" Oct 05 08:45:00 crc kubenswrapper[4846]: I1005 08:45:00.349816 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r8g59\" (UniqueName: \"kubernetes.io/projected/e9a16960-8bb2-43a3-bbc4-70a135eb2449-kube-api-access-r8g59\") pod \"collect-profiles-29327565-p7hbh\" (UID: \"e9a16960-8bb2-43a3-bbc4-70a135eb2449\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327565-p7hbh" Oct 05 08:45:00 crc kubenswrapper[4846]: I1005 08:45:00.494809 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29327565-p7hbh" Oct 05 08:45:00 crc kubenswrapper[4846]: I1005 08:45:00.948933 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29327565-p7hbh"] Oct 05 08:45:01 crc kubenswrapper[4846]: I1005 08:45:01.642457 4846 generic.go:334] "Generic (PLEG): container finished" podID="e9a16960-8bb2-43a3-bbc4-70a135eb2449" containerID="04048452dbf08a2f8c4be66289e063bae91b3f32184ebb728d5932fbd877ebaf" exitCode=0 Oct 05 08:45:01 crc kubenswrapper[4846]: I1005 08:45:01.642564 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29327565-p7hbh" event={"ID":"e9a16960-8bb2-43a3-bbc4-70a135eb2449","Type":"ContainerDied","Data":"04048452dbf08a2f8c4be66289e063bae91b3f32184ebb728d5932fbd877ebaf"} Oct 05 08:45:01 crc kubenswrapper[4846]: I1005 08:45:01.642847 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29327565-p7hbh" event={"ID":"e9a16960-8bb2-43a3-bbc4-70a135eb2449","Type":"ContainerStarted","Data":"1571fb92575d7a47d29b9292d5a762c7105c54900d3594bfe7cf77449293cff1"} Oct 05 08:45:02 crc kubenswrapper[4846]: I1005 08:45:02.998825 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29327565-p7hbh" Oct 05 08:45:03 crc kubenswrapper[4846]: I1005 08:45:03.094210 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e9a16960-8bb2-43a3-bbc4-70a135eb2449-config-volume\") pod \"e9a16960-8bb2-43a3-bbc4-70a135eb2449\" (UID: \"e9a16960-8bb2-43a3-bbc4-70a135eb2449\") " Oct 05 08:45:03 crc kubenswrapper[4846]: I1005 08:45:03.094386 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r8g59\" (UniqueName: \"kubernetes.io/projected/e9a16960-8bb2-43a3-bbc4-70a135eb2449-kube-api-access-r8g59\") pod \"e9a16960-8bb2-43a3-bbc4-70a135eb2449\" (UID: \"e9a16960-8bb2-43a3-bbc4-70a135eb2449\") " Oct 05 08:45:03 crc kubenswrapper[4846]: I1005 08:45:03.094508 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e9a16960-8bb2-43a3-bbc4-70a135eb2449-secret-volume\") pod \"e9a16960-8bb2-43a3-bbc4-70a135eb2449\" (UID: \"e9a16960-8bb2-43a3-bbc4-70a135eb2449\") " Oct 05 08:45:03 crc kubenswrapper[4846]: I1005 08:45:03.095361 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e9a16960-8bb2-43a3-bbc4-70a135eb2449-config-volume" (OuterVolumeSpecName: "config-volume") pod "e9a16960-8bb2-43a3-bbc4-70a135eb2449" (UID: "e9a16960-8bb2-43a3-bbc4-70a135eb2449"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 08:45:03 crc kubenswrapper[4846]: I1005 08:45:03.101109 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e9a16960-8bb2-43a3-bbc4-70a135eb2449-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "e9a16960-8bb2-43a3-bbc4-70a135eb2449" (UID: "e9a16960-8bb2-43a3-bbc4-70a135eb2449"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:45:03 crc kubenswrapper[4846]: I1005 08:45:03.102755 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e9a16960-8bb2-43a3-bbc4-70a135eb2449-kube-api-access-r8g59" (OuterVolumeSpecName: "kube-api-access-r8g59") pod "e9a16960-8bb2-43a3-bbc4-70a135eb2449" (UID: "e9a16960-8bb2-43a3-bbc4-70a135eb2449"). InnerVolumeSpecName "kube-api-access-r8g59". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:45:03 crc kubenswrapper[4846]: I1005 08:45:03.197574 4846 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e9a16960-8bb2-43a3-bbc4-70a135eb2449-config-volume\") on node \"crc\" DevicePath \"\"" Oct 05 08:45:03 crc kubenswrapper[4846]: I1005 08:45:03.197622 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r8g59\" (UniqueName: \"kubernetes.io/projected/e9a16960-8bb2-43a3-bbc4-70a135eb2449-kube-api-access-r8g59\") on node \"crc\" DevicePath \"\"" Oct 05 08:45:03 crc kubenswrapper[4846]: I1005 08:45:03.197639 4846 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e9a16960-8bb2-43a3-bbc4-70a135eb2449-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 05 08:45:03 crc kubenswrapper[4846]: I1005 08:45:03.667223 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29327565-p7hbh" event={"ID":"e9a16960-8bb2-43a3-bbc4-70a135eb2449","Type":"ContainerDied","Data":"1571fb92575d7a47d29b9292d5a762c7105c54900d3594bfe7cf77449293cff1"} Oct 05 08:45:03 crc kubenswrapper[4846]: I1005 08:45:03.667299 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1571fb92575d7a47d29b9292d5a762c7105c54900d3594bfe7cf77449293cff1" Oct 05 08:45:03 crc kubenswrapper[4846]: I1005 08:45:03.667319 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29327565-p7hbh" Oct 05 08:45:04 crc kubenswrapper[4846]: I1005 08:45:04.079234 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29327520-dcpnc"] Oct 05 08:45:04 crc kubenswrapper[4846]: I1005 08:45:04.087032 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29327520-dcpnc"] Oct 05 08:45:04 crc kubenswrapper[4846]: I1005 08:45:04.517465 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4dc64c01-83c3-4468-9df8-4d0b6562d26a" path="/var/lib/kubelet/pods/4dc64c01-83c3-4468-9df8-4d0b6562d26a/volumes" Oct 05 08:45:06 crc kubenswrapper[4846]: I1005 08:45:06.497824 4846 scope.go:117] "RemoveContainer" containerID="0d84b906773347b38ec2ce550f580254430238b4374f3171cb7453fa65259bcc" Oct 05 08:45:06 crc kubenswrapper[4846]: E1005 08:45:06.498579 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 08:45:19 crc kubenswrapper[4846]: I1005 08:45:19.498358 4846 scope.go:117] "RemoveContainer" containerID="0d84b906773347b38ec2ce550f580254430238b4374f3171cb7453fa65259bcc" Oct 05 08:45:19 crc kubenswrapper[4846]: E1005 08:45:19.499328 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 08:45:33 crc kubenswrapper[4846]: I1005 08:45:33.497931 4846 scope.go:117] "RemoveContainer" containerID="0d84b906773347b38ec2ce550f580254430238b4374f3171cb7453fa65259bcc" Oct 05 08:45:33 crc kubenswrapper[4846]: E1005 08:45:33.500776 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 08:45:38 crc kubenswrapper[4846]: I1005 08:45:38.014828 4846 scope.go:117] "RemoveContainer" containerID="ffde80bfd45c9c3a6ca4b4c833fddc3220ff829cea27ad3ca24a63a713789e68" Oct 05 08:45:44 crc kubenswrapper[4846]: I1005 08:45:44.498683 4846 scope.go:117] "RemoveContainer" containerID="0d84b906773347b38ec2ce550f580254430238b4374f3171cb7453fa65259bcc" Oct 05 08:45:44 crc kubenswrapper[4846]: E1005 08:45:44.499740 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 08:45:56 crc kubenswrapper[4846]: I1005 08:45:56.497933 4846 scope.go:117] "RemoveContainer" containerID="0d84b906773347b38ec2ce550f580254430238b4374f3171cb7453fa65259bcc" Oct 05 08:45:56 crc kubenswrapper[4846]: E1005 08:45:56.498746 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 08:46:08 crc kubenswrapper[4846]: I1005 08:46:08.515545 4846 scope.go:117] "RemoveContainer" containerID="0d84b906773347b38ec2ce550f580254430238b4374f3171cb7453fa65259bcc" Oct 05 08:46:08 crc kubenswrapper[4846]: E1005 08:46:08.516990 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 08:46:11 crc kubenswrapper[4846]: I1005 08:46:11.468236 4846 generic.go:334] "Generic (PLEG): container finished" podID="f55e5fdc-aff8-4ebe-afa7-a570e18bf335" containerID="d56fcf1bb4436455b77d9095e9765666332147eaff35902e39da0154cab4e14d" exitCode=0 Oct 05 08:46:11 crc kubenswrapper[4846]: I1005 08:46:11.468342 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-openstack-openstack-cell1-rxf6k" event={"ID":"f55e5fdc-aff8-4ebe-afa7-a570e18bf335","Type":"ContainerDied","Data":"d56fcf1bb4436455b77d9095e9765666332147eaff35902e39da0154cab4e14d"} Oct 05 08:46:13 crc kubenswrapper[4846]: I1005 08:46:13.016021 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-openstack-openstack-cell1-rxf6k" Oct 05 08:46:13 crc kubenswrapper[4846]: I1005 08:46:13.120134 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f55e5fdc-aff8-4ebe-afa7-a570e18bf335-ssh-key\") pod \"f55e5fdc-aff8-4ebe-afa7-a570e18bf335\" (UID: \"f55e5fdc-aff8-4ebe-afa7-a570e18bf335\") " Oct 05 08:46:13 crc kubenswrapper[4846]: I1005 08:46:13.120311 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f55e5fdc-aff8-4ebe-afa7-a570e18bf335-bootstrap-combined-ca-bundle\") pod \"f55e5fdc-aff8-4ebe-afa7-a570e18bf335\" (UID: \"f55e5fdc-aff8-4ebe-afa7-a570e18bf335\") " Oct 05 08:46:13 crc kubenswrapper[4846]: I1005 08:46:13.120391 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f55e5fdc-aff8-4ebe-afa7-a570e18bf335-inventory\") pod \"f55e5fdc-aff8-4ebe-afa7-a570e18bf335\" (UID: \"f55e5fdc-aff8-4ebe-afa7-a570e18bf335\") " Oct 05 08:46:13 crc kubenswrapper[4846]: I1005 08:46:13.120443 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mpk2z\" (UniqueName: \"kubernetes.io/projected/f55e5fdc-aff8-4ebe-afa7-a570e18bf335-kube-api-access-mpk2z\") pod \"f55e5fdc-aff8-4ebe-afa7-a570e18bf335\" (UID: \"f55e5fdc-aff8-4ebe-afa7-a570e18bf335\") " Oct 05 08:46:13 crc kubenswrapper[4846]: I1005 08:46:13.125639 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f55e5fdc-aff8-4ebe-afa7-a570e18bf335-kube-api-access-mpk2z" (OuterVolumeSpecName: "kube-api-access-mpk2z") pod "f55e5fdc-aff8-4ebe-afa7-a570e18bf335" (UID: "f55e5fdc-aff8-4ebe-afa7-a570e18bf335"). InnerVolumeSpecName "kube-api-access-mpk2z". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:46:13 crc kubenswrapper[4846]: I1005 08:46:13.126158 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f55e5fdc-aff8-4ebe-afa7-a570e18bf335-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "f55e5fdc-aff8-4ebe-afa7-a570e18bf335" (UID: "f55e5fdc-aff8-4ebe-afa7-a570e18bf335"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:46:13 crc kubenswrapper[4846]: I1005 08:46:13.154005 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f55e5fdc-aff8-4ebe-afa7-a570e18bf335-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "f55e5fdc-aff8-4ebe-afa7-a570e18bf335" (UID: "f55e5fdc-aff8-4ebe-afa7-a570e18bf335"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:46:13 crc kubenswrapper[4846]: I1005 08:46:13.156503 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f55e5fdc-aff8-4ebe-afa7-a570e18bf335-inventory" (OuterVolumeSpecName: "inventory") pod "f55e5fdc-aff8-4ebe-afa7-a570e18bf335" (UID: "f55e5fdc-aff8-4ebe-afa7-a570e18bf335"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:46:13 crc kubenswrapper[4846]: I1005 08:46:13.222865 4846 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f55e5fdc-aff8-4ebe-afa7-a570e18bf335-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 05 08:46:13 crc kubenswrapper[4846]: I1005 08:46:13.222900 4846 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f55e5fdc-aff8-4ebe-afa7-a570e18bf335-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 08:46:13 crc kubenswrapper[4846]: I1005 08:46:13.222917 4846 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f55e5fdc-aff8-4ebe-afa7-a570e18bf335-inventory\") on node \"crc\" DevicePath \"\"" Oct 05 08:46:13 crc kubenswrapper[4846]: I1005 08:46:13.222931 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mpk2z\" (UniqueName: \"kubernetes.io/projected/f55e5fdc-aff8-4ebe-afa7-a570e18bf335-kube-api-access-mpk2z\") on node \"crc\" DevicePath \"\"" Oct 05 08:46:13 crc kubenswrapper[4846]: I1005 08:46:13.497638 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-openstack-openstack-cell1-rxf6k" event={"ID":"f55e5fdc-aff8-4ebe-afa7-a570e18bf335","Type":"ContainerDied","Data":"a0d4d80548f05771bac2d8b0d6168181e7e5e95074982f9680378d2cdfa01df5"} Oct 05 08:46:13 crc kubenswrapper[4846]: I1005 08:46:13.497702 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a0d4d80548f05771bac2d8b0d6168181e7e5e95074982f9680378d2cdfa01df5" Oct 05 08:46:13 crc kubenswrapper[4846]: I1005 08:46:13.497719 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-openstack-openstack-cell1-rxf6k" Oct 05 08:46:13 crc kubenswrapper[4846]: I1005 08:46:13.621780 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/download-cache-openstack-openstack-cell1-w242f"] Oct 05 08:46:13 crc kubenswrapper[4846]: E1005 08:46:13.622456 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f55e5fdc-aff8-4ebe-afa7-a570e18bf335" containerName="bootstrap-openstack-openstack-cell1" Oct 05 08:46:13 crc kubenswrapper[4846]: I1005 08:46:13.622488 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="f55e5fdc-aff8-4ebe-afa7-a570e18bf335" containerName="bootstrap-openstack-openstack-cell1" Oct 05 08:46:13 crc kubenswrapper[4846]: E1005 08:46:13.622522 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e9a16960-8bb2-43a3-bbc4-70a135eb2449" containerName="collect-profiles" Oct 05 08:46:13 crc kubenswrapper[4846]: I1005 08:46:13.622532 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="e9a16960-8bb2-43a3-bbc4-70a135eb2449" containerName="collect-profiles" Oct 05 08:46:13 crc kubenswrapper[4846]: I1005 08:46:13.622882 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="e9a16960-8bb2-43a3-bbc4-70a135eb2449" containerName="collect-profiles" Oct 05 08:46:13 crc kubenswrapper[4846]: I1005 08:46:13.622957 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="f55e5fdc-aff8-4ebe-afa7-a570e18bf335" containerName="bootstrap-openstack-openstack-cell1" Oct 05 08:46:13 crc kubenswrapper[4846]: I1005 08:46:13.624204 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-openstack-openstack-cell1-w242f" Oct 05 08:46:13 crc kubenswrapper[4846]: I1005 08:46:13.626682 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-7687s" Oct 05 08:46:13 crc kubenswrapper[4846]: I1005 08:46:13.627935 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 05 08:46:13 crc kubenswrapper[4846]: I1005 08:46:13.628096 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Oct 05 08:46:13 crc kubenswrapper[4846]: I1005 08:46:13.637333 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Oct 05 08:46:13 crc kubenswrapper[4846]: I1005 08:46:13.637906 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-openstack-openstack-cell1-w242f"] Oct 05 08:46:13 crc kubenswrapper[4846]: I1005 08:46:13.744825 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7ecab93f-583b-4a2f-b000-10f96324d002-ssh-key\") pod \"download-cache-openstack-openstack-cell1-w242f\" (UID: \"7ecab93f-583b-4a2f-b000-10f96324d002\") " pod="openstack/download-cache-openstack-openstack-cell1-w242f" Oct 05 08:46:13 crc kubenswrapper[4846]: I1005 08:46:13.744977 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w5czb\" (UniqueName: \"kubernetes.io/projected/7ecab93f-583b-4a2f-b000-10f96324d002-kube-api-access-w5czb\") pod \"download-cache-openstack-openstack-cell1-w242f\" (UID: \"7ecab93f-583b-4a2f-b000-10f96324d002\") " pod="openstack/download-cache-openstack-openstack-cell1-w242f" Oct 05 08:46:13 crc kubenswrapper[4846]: I1005 08:46:13.745041 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7ecab93f-583b-4a2f-b000-10f96324d002-inventory\") pod \"download-cache-openstack-openstack-cell1-w242f\" (UID: \"7ecab93f-583b-4a2f-b000-10f96324d002\") " pod="openstack/download-cache-openstack-openstack-cell1-w242f" Oct 05 08:46:13 crc kubenswrapper[4846]: I1005 08:46:13.846732 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w5czb\" (UniqueName: \"kubernetes.io/projected/7ecab93f-583b-4a2f-b000-10f96324d002-kube-api-access-w5czb\") pod \"download-cache-openstack-openstack-cell1-w242f\" (UID: \"7ecab93f-583b-4a2f-b000-10f96324d002\") " pod="openstack/download-cache-openstack-openstack-cell1-w242f" Oct 05 08:46:13 crc kubenswrapper[4846]: I1005 08:46:13.846898 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7ecab93f-583b-4a2f-b000-10f96324d002-inventory\") pod \"download-cache-openstack-openstack-cell1-w242f\" (UID: \"7ecab93f-583b-4a2f-b000-10f96324d002\") " pod="openstack/download-cache-openstack-openstack-cell1-w242f" Oct 05 08:46:13 crc kubenswrapper[4846]: I1005 08:46:13.847133 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7ecab93f-583b-4a2f-b000-10f96324d002-ssh-key\") pod \"download-cache-openstack-openstack-cell1-w242f\" (UID: \"7ecab93f-583b-4a2f-b000-10f96324d002\") " pod="openstack/download-cache-openstack-openstack-cell1-w242f" Oct 05 08:46:13 crc kubenswrapper[4846]: I1005 08:46:13.856101 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7ecab93f-583b-4a2f-b000-10f96324d002-ssh-key\") pod \"download-cache-openstack-openstack-cell1-w242f\" (UID: \"7ecab93f-583b-4a2f-b000-10f96324d002\") " pod="openstack/download-cache-openstack-openstack-cell1-w242f" Oct 05 08:46:13 crc kubenswrapper[4846]: I1005 08:46:13.856350 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7ecab93f-583b-4a2f-b000-10f96324d002-inventory\") pod \"download-cache-openstack-openstack-cell1-w242f\" (UID: \"7ecab93f-583b-4a2f-b000-10f96324d002\") " pod="openstack/download-cache-openstack-openstack-cell1-w242f" Oct 05 08:46:13 crc kubenswrapper[4846]: I1005 08:46:13.867121 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w5czb\" (UniqueName: \"kubernetes.io/projected/7ecab93f-583b-4a2f-b000-10f96324d002-kube-api-access-w5czb\") pod \"download-cache-openstack-openstack-cell1-w242f\" (UID: \"7ecab93f-583b-4a2f-b000-10f96324d002\") " pod="openstack/download-cache-openstack-openstack-cell1-w242f" Oct 05 08:46:13 crc kubenswrapper[4846]: I1005 08:46:13.951439 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-openstack-openstack-cell1-w242f" Oct 05 08:46:14 crc kubenswrapper[4846]: I1005 08:46:14.515907 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-openstack-openstack-cell1-w242f"] Oct 05 08:46:15 crc kubenswrapper[4846]: I1005 08:46:15.518655 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-openstack-openstack-cell1-w242f" event={"ID":"7ecab93f-583b-4a2f-b000-10f96324d002","Type":"ContainerStarted","Data":"e2956718cac7be5758d55b530476e5da25f4f1cc711675e01697e9d30be2591b"} Oct 05 08:46:15 crc kubenswrapper[4846]: I1005 08:46:15.518940 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-openstack-openstack-cell1-w242f" event={"ID":"7ecab93f-583b-4a2f-b000-10f96324d002","Type":"ContainerStarted","Data":"ca6ab92fcea18e7a730ed792059deec5b592cd722d473cc28f1647f00d71318b"} Oct 05 08:46:15 crc kubenswrapper[4846]: I1005 08:46:15.552102 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/download-cache-openstack-openstack-cell1-w242f" podStartSLOduration=1.853412042 podStartE2EDuration="2.552082312s" podCreationTimestamp="2025-10-05 08:46:13 +0000 UTC" firstStartedPulling="2025-10-05 08:46:14.522583828 +0000 UTC m=+7096.763436603" lastFinishedPulling="2025-10-05 08:46:15.221254068 +0000 UTC m=+7097.462106873" observedRunningTime="2025-10-05 08:46:15.539615361 +0000 UTC m=+7097.780468156" watchObservedRunningTime="2025-10-05 08:46:15.552082312 +0000 UTC m=+7097.792935097" Oct 05 08:46:20 crc kubenswrapper[4846]: I1005 08:46:20.498931 4846 scope.go:117] "RemoveContainer" containerID="0d84b906773347b38ec2ce550f580254430238b4374f3171cb7453fa65259bcc" Oct 05 08:46:20 crc kubenswrapper[4846]: E1005 08:46:20.499992 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 08:46:33 crc kubenswrapper[4846]: I1005 08:46:33.498157 4846 scope.go:117] "RemoveContainer" containerID="0d84b906773347b38ec2ce550f580254430238b4374f3171cb7453fa65259bcc" Oct 05 08:46:33 crc kubenswrapper[4846]: E1005 08:46:33.499440 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 08:46:46 crc kubenswrapper[4846]: I1005 08:46:46.499095 4846 scope.go:117] "RemoveContainer" containerID="0d84b906773347b38ec2ce550f580254430238b4374f3171cb7453fa65259bcc" Oct 05 08:46:46 crc kubenswrapper[4846]: E1005 08:46:46.500342 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 08:47:01 crc kubenswrapper[4846]: I1005 08:47:01.498164 4846 scope.go:117] "RemoveContainer" containerID="0d84b906773347b38ec2ce550f580254430238b4374f3171cb7453fa65259bcc" Oct 05 08:47:01 crc kubenswrapper[4846]: E1005 08:47:01.499540 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 08:47:14 crc kubenswrapper[4846]: I1005 08:47:14.498951 4846 scope.go:117] "RemoveContainer" containerID="0d84b906773347b38ec2ce550f580254430238b4374f3171cb7453fa65259bcc" Oct 05 08:47:14 crc kubenswrapper[4846]: E1005 08:47:14.500054 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 08:47:26 crc kubenswrapper[4846]: I1005 08:47:26.498430 4846 scope.go:117] "RemoveContainer" containerID="0d84b906773347b38ec2ce550f580254430238b4374f3171cb7453fa65259bcc" Oct 05 08:47:26 crc kubenswrapper[4846]: E1005 08:47:26.499433 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 08:47:38 crc kubenswrapper[4846]: I1005 08:47:38.515628 4846 scope.go:117] "RemoveContainer" containerID="0d84b906773347b38ec2ce550f580254430238b4374f3171cb7453fa65259bcc" Oct 05 08:47:38 crc kubenswrapper[4846]: E1005 08:47:38.516598 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 08:47:45 crc kubenswrapper[4846]: I1005 08:47:45.731412 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-bfxmm"] Oct 05 08:47:45 crc kubenswrapper[4846]: I1005 08:47:45.736859 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-bfxmm" Oct 05 08:47:45 crc kubenswrapper[4846]: I1005 08:47:45.750995 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-bfxmm"] Oct 05 08:47:45 crc kubenswrapper[4846]: I1005 08:47:45.812914 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-txg89\" (UniqueName: \"kubernetes.io/projected/0c1a1b66-6c25-4694-afb7-bbeb68089408-kube-api-access-txg89\") pod \"community-operators-bfxmm\" (UID: \"0c1a1b66-6c25-4694-afb7-bbeb68089408\") " pod="openshift-marketplace/community-operators-bfxmm" Oct 05 08:47:45 crc kubenswrapper[4846]: I1005 08:47:45.813025 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0c1a1b66-6c25-4694-afb7-bbeb68089408-catalog-content\") pod \"community-operators-bfxmm\" (UID: \"0c1a1b66-6c25-4694-afb7-bbeb68089408\") " pod="openshift-marketplace/community-operators-bfxmm" Oct 05 08:47:45 crc kubenswrapper[4846]: I1005 08:47:45.813121 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0c1a1b66-6c25-4694-afb7-bbeb68089408-utilities\") pod \"community-operators-bfxmm\" (UID: \"0c1a1b66-6c25-4694-afb7-bbeb68089408\") " pod="openshift-marketplace/community-operators-bfxmm" Oct 05 08:47:45 crc kubenswrapper[4846]: I1005 08:47:45.914951 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-txg89\" (UniqueName: \"kubernetes.io/projected/0c1a1b66-6c25-4694-afb7-bbeb68089408-kube-api-access-txg89\") pod \"community-operators-bfxmm\" (UID: \"0c1a1b66-6c25-4694-afb7-bbeb68089408\") " pod="openshift-marketplace/community-operators-bfxmm" Oct 05 08:47:45 crc kubenswrapper[4846]: I1005 08:47:45.915055 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0c1a1b66-6c25-4694-afb7-bbeb68089408-catalog-content\") pod \"community-operators-bfxmm\" (UID: \"0c1a1b66-6c25-4694-afb7-bbeb68089408\") " pod="openshift-marketplace/community-operators-bfxmm" Oct 05 08:47:45 crc kubenswrapper[4846]: I1005 08:47:45.915133 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0c1a1b66-6c25-4694-afb7-bbeb68089408-utilities\") pod \"community-operators-bfxmm\" (UID: \"0c1a1b66-6c25-4694-afb7-bbeb68089408\") " pod="openshift-marketplace/community-operators-bfxmm" Oct 05 08:47:45 crc kubenswrapper[4846]: I1005 08:47:45.915727 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0c1a1b66-6c25-4694-afb7-bbeb68089408-utilities\") pod \"community-operators-bfxmm\" (UID: \"0c1a1b66-6c25-4694-afb7-bbeb68089408\") " pod="openshift-marketplace/community-operators-bfxmm" Oct 05 08:47:45 crc kubenswrapper[4846]: I1005 08:47:45.916116 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0c1a1b66-6c25-4694-afb7-bbeb68089408-catalog-content\") pod \"community-operators-bfxmm\" (UID: \"0c1a1b66-6c25-4694-afb7-bbeb68089408\") " pod="openshift-marketplace/community-operators-bfxmm" Oct 05 08:47:45 crc kubenswrapper[4846]: I1005 08:47:45.943119 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-txg89\" (UniqueName: \"kubernetes.io/projected/0c1a1b66-6c25-4694-afb7-bbeb68089408-kube-api-access-txg89\") pod \"community-operators-bfxmm\" (UID: \"0c1a1b66-6c25-4694-afb7-bbeb68089408\") " pod="openshift-marketplace/community-operators-bfxmm" Oct 05 08:47:46 crc kubenswrapper[4846]: I1005 08:47:46.084431 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-bfxmm" Oct 05 08:47:46 crc kubenswrapper[4846]: I1005 08:47:46.586982 4846 generic.go:334] "Generic (PLEG): container finished" podID="7ecab93f-583b-4a2f-b000-10f96324d002" containerID="e2956718cac7be5758d55b530476e5da25f4f1cc711675e01697e9d30be2591b" exitCode=0 Oct 05 08:47:46 crc kubenswrapper[4846]: I1005 08:47:46.587088 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-openstack-openstack-cell1-w242f" event={"ID":"7ecab93f-583b-4a2f-b000-10f96324d002","Type":"ContainerDied","Data":"e2956718cac7be5758d55b530476e5da25f4f1cc711675e01697e9d30be2591b"} Oct 05 08:47:46 crc kubenswrapper[4846]: I1005 08:47:46.587955 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-bfxmm"] Oct 05 08:47:47 crc kubenswrapper[4846]: I1005 08:47:47.600531 4846 generic.go:334] "Generic (PLEG): container finished" podID="0c1a1b66-6c25-4694-afb7-bbeb68089408" containerID="7e1f323f362a3c4d472f0929145b0e2e6b4104b2aa60194a06662f6b63d95a69" exitCode=0 Oct 05 08:47:47 crc kubenswrapper[4846]: I1005 08:47:47.600587 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bfxmm" event={"ID":"0c1a1b66-6c25-4694-afb7-bbeb68089408","Type":"ContainerDied","Data":"7e1f323f362a3c4d472f0929145b0e2e6b4104b2aa60194a06662f6b63d95a69"} Oct 05 08:47:47 crc kubenswrapper[4846]: I1005 08:47:47.600976 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bfxmm" event={"ID":"0c1a1b66-6c25-4694-afb7-bbeb68089408","Type":"ContainerStarted","Data":"692cd14fc839604f27b25abb63253a3dda953c7de5a2364087f3cfd30aec7da9"} Oct 05 08:47:48 crc kubenswrapper[4846]: I1005 08:47:48.097703 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-openstack-openstack-cell1-w242f" Oct 05 08:47:48 crc kubenswrapper[4846]: I1005 08:47:48.168053 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7ecab93f-583b-4a2f-b000-10f96324d002-ssh-key\") pod \"7ecab93f-583b-4a2f-b000-10f96324d002\" (UID: \"7ecab93f-583b-4a2f-b000-10f96324d002\") " Oct 05 08:47:48 crc kubenswrapper[4846]: I1005 08:47:48.168143 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w5czb\" (UniqueName: \"kubernetes.io/projected/7ecab93f-583b-4a2f-b000-10f96324d002-kube-api-access-w5czb\") pod \"7ecab93f-583b-4a2f-b000-10f96324d002\" (UID: \"7ecab93f-583b-4a2f-b000-10f96324d002\") " Oct 05 08:47:48 crc kubenswrapper[4846]: I1005 08:47:48.168390 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7ecab93f-583b-4a2f-b000-10f96324d002-inventory\") pod \"7ecab93f-583b-4a2f-b000-10f96324d002\" (UID: \"7ecab93f-583b-4a2f-b000-10f96324d002\") " Oct 05 08:47:48 crc kubenswrapper[4846]: I1005 08:47:48.179196 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7ecab93f-583b-4a2f-b000-10f96324d002-kube-api-access-w5czb" (OuterVolumeSpecName: "kube-api-access-w5czb") pod "7ecab93f-583b-4a2f-b000-10f96324d002" (UID: "7ecab93f-583b-4a2f-b000-10f96324d002"). InnerVolumeSpecName "kube-api-access-w5czb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:47:48 crc kubenswrapper[4846]: I1005 08:47:48.199900 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7ecab93f-583b-4a2f-b000-10f96324d002-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "7ecab93f-583b-4a2f-b000-10f96324d002" (UID: "7ecab93f-583b-4a2f-b000-10f96324d002"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:47:48 crc kubenswrapper[4846]: I1005 08:47:48.209445 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7ecab93f-583b-4a2f-b000-10f96324d002-inventory" (OuterVolumeSpecName: "inventory") pod "7ecab93f-583b-4a2f-b000-10f96324d002" (UID: "7ecab93f-583b-4a2f-b000-10f96324d002"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:47:48 crc kubenswrapper[4846]: I1005 08:47:48.270607 4846 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7ecab93f-583b-4a2f-b000-10f96324d002-inventory\") on node \"crc\" DevicePath \"\"" Oct 05 08:47:48 crc kubenswrapper[4846]: I1005 08:47:48.270650 4846 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7ecab93f-583b-4a2f-b000-10f96324d002-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 05 08:47:48 crc kubenswrapper[4846]: I1005 08:47:48.270664 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w5czb\" (UniqueName: \"kubernetes.io/projected/7ecab93f-583b-4a2f-b000-10f96324d002-kube-api-access-w5czb\") on node \"crc\" DevicePath \"\"" Oct 05 08:47:48 crc kubenswrapper[4846]: I1005 08:47:48.611284 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-openstack-openstack-cell1-w242f" event={"ID":"7ecab93f-583b-4a2f-b000-10f96324d002","Type":"ContainerDied","Data":"ca6ab92fcea18e7a730ed792059deec5b592cd722d473cc28f1647f00d71318b"} Oct 05 08:47:48 crc kubenswrapper[4846]: I1005 08:47:48.611328 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-openstack-openstack-cell1-w242f" Oct 05 08:47:48 crc kubenswrapper[4846]: I1005 08:47:48.611337 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ca6ab92fcea18e7a730ed792059deec5b592cd722d473cc28f1647f00d71318b" Oct 05 08:47:48 crc kubenswrapper[4846]: I1005 08:47:48.614617 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bfxmm" event={"ID":"0c1a1b66-6c25-4694-afb7-bbeb68089408","Type":"ContainerStarted","Data":"1a9e7a9782a692ba92fdd3f59c9689e5c0b605e278b32a9780004974fd4094c1"} Oct 05 08:47:48 crc kubenswrapper[4846]: I1005 08:47:48.706905 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-network-openstack-openstack-cell1-l9jt5"] Oct 05 08:47:48 crc kubenswrapper[4846]: E1005 08:47:48.707504 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7ecab93f-583b-4a2f-b000-10f96324d002" containerName="download-cache-openstack-openstack-cell1" Oct 05 08:47:48 crc kubenswrapper[4846]: I1005 08:47:48.707525 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="7ecab93f-583b-4a2f-b000-10f96324d002" containerName="download-cache-openstack-openstack-cell1" Oct 05 08:47:48 crc kubenswrapper[4846]: I1005 08:47:48.707724 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="7ecab93f-583b-4a2f-b000-10f96324d002" containerName="download-cache-openstack-openstack-cell1" Oct 05 08:47:48 crc kubenswrapper[4846]: I1005 08:47:48.708463 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-openstack-openstack-cell1-l9jt5" Oct 05 08:47:48 crc kubenswrapper[4846]: I1005 08:47:48.710460 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Oct 05 08:47:48 crc kubenswrapper[4846]: I1005 08:47:48.710711 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Oct 05 08:47:48 crc kubenswrapper[4846]: I1005 08:47:48.711284 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 05 08:47:48 crc kubenswrapper[4846]: I1005 08:47:48.715445 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-7687s" Oct 05 08:47:48 crc kubenswrapper[4846]: I1005 08:47:48.716567 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-openstack-openstack-cell1-l9jt5"] Oct 05 08:47:48 crc kubenswrapper[4846]: I1005 08:47:48.781210 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/58a90b64-2441-4ccf-8655-a5a582dee3cc-ssh-key\") pod \"configure-network-openstack-openstack-cell1-l9jt5\" (UID: \"58a90b64-2441-4ccf-8655-a5a582dee3cc\") " pod="openstack/configure-network-openstack-openstack-cell1-l9jt5" Oct 05 08:47:48 crc kubenswrapper[4846]: I1005 08:47:48.781578 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-622w8\" (UniqueName: \"kubernetes.io/projected/58a90b64-2441-4ccf-8655-a5a582dee3cc-kube-api-access-622w8\") pod \"configure-network-openstack-openstack-cell1-l9jt5\" (UID: \"58a90b64-2441-4ccf-8655-a5a582dee3cc\") " pod="openstack/configure-network-openstack-openstack-cell1-l9jt5" Oct 05 08:47:48 crc kubenswrapper[4846]: I1005 08:47:48.781632 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/58a90b64-2441-4ccf-8655-a5a582dee3cc-inventory\") pod \"configure-network-openstack-openstack-cell1-l9jt5\" (UID: \"58a90b64-2441-4ccf-8655-a5a582dee3cc\") " pod="openstack/configure-network-openstack-openstack-cell1-l9jt5" Oct 05 08:47:48 crc kubenswrapper[4846]: I1005 08:47:48.883976 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/58a90b64-2441-4ccf-8655-a5a582dee3cc-ssh-key\") pod \"configure-network-openstack-openstack-cell1-l9jt5\" (UID: \"58a90b64-2441-4ccf-8655-a5a582dee3cc\") " pod="openstack/configure-network-openstack-openstack-cell1-l9jt5" Oct 05 08:47:48 crc kubenswrapper[4846]: I1005 08:47:48.884194 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-622w8\" (UniqueName: \"kubernetes.io/projected/58a90b64-2441-4ccf-8655-a5a582dee3cc-kube-api-access-622w8\") pod \"configure-network-openstack-openstack-cell1-l9jt5\" (UID: \"58a90b64-2441-4ccf-8655-a5a582dee3cc\") " pod="openstack/configure-network-openstack-openstack-cell1-l9jt5" Oct 05 08:47:48 crc kubenswrapper[4846]: I1005 08:47:48.884227 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/58a90b64-2441-4ccf-8655-a5a582dee3cc-inventory\") pod \"configure-network-openstack-openstack-cell1-l9jt5\" (UID: \"58a90b64-2441-4ccf-8655-a5a582dee3cc\") " pod="openstack/configure-network-openstack-openstack-cell1-l9jt5" Oct 05 08:47:48 crc kubenswrapper[4846]: I1005 08:47:48.889906 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/58a90b64-2441-4ccf-8655-a5a582dee3cc-inventory\") pod \"configure-network-openstack-openstack-cell1-l9jt5\" (UID: \"58a90b64-2441-4ccf-8655-a5a582dee3cc\") " pod="openstack/configure-network-openstack-openstack-cell1-l9jt5" Oct 05 08:47:48 crc kubenswrapper[4846]: I1005 08:47:48.896152 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/58a90b64-2441-4ccf-8655-a5a582dee3cc-ssh-key\") pod \"configure-network-openstack-openstack-cell1-l9jt5\" (UID: \"58a90b64-2441-4ccf-8655-a5a582dee3cc\") " pod="openstack/configure-network-openstack-openstack-cell1-l9jt5" Oct 05 08:47:48 crc kubenswrapper[4846]: I1005 08:47:48.909714 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-622w8\" (UniqueName: \"kubernetes.io/projected/58a90b64-2441-4ccf-8655-a5a582dee3cc-kube-api-access-622w8\") pod \"configure-network-openstack-openstack-cell1-l9jt5\" (UID: \"58a90b64-2441-4ccf-8655-a5a582dee3cc\") " pod="openstack/configure-network-openstack-openstack-cell1-l9jt5" Oct 05 08:47:49 crc kubenswrapper[4846]: I1005 08:47:49.024904 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-openstack-openstack-cell1-l9jt5" Oct 05 08:47:49 crc kubenswrapper[4846]: I1005 08:47:49.581989 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-openstack-openstack-cell1-l9jt5"] Oct 05 08:47:49 crc kubenswrapper[4846]: W1005 08:47:49.582046 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod58a90b64_2441_4ccf_8655_a5a582dee3cc.slice/crio-5fba894e1a8a66c6f14fab39fd863bbf1b5173306fd3b63e12c7a2942d262b39 WatchSource:0}: Error finding container 5fba894e1a8a66c6f14fab39fd863bbf1b5173306fd3b63e12c7a2942d262b39: Status 404 returned error can't find the container with id 5fba894e1a8a66c6f14fab39fd863bbf1b5173306fd3b63e12c7a2942d262b39 Oct 05 08:47:49 crc kubenswrapper[4846]: I1005 08:47:49.625467 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-openstack-openstack-cell1-l9jt5" event={"ID":"58a90b64-2441-4ccf-8655-a5a582dee3cc","Type":"ContainerStarted","Data":"5fba894e1a8a66c6f14fab39fd863bbf1b5173306fd3b63e12c7a2942d262b39"} Oct 05 08:47:50 crc kubenswrapper[4846]: I1005 08:47:50.498256 4846 scope.go:117] "RemoveContainer" containerID="0d84b906773347b38ec2ce550f580254430238b4374f3171cb7453fa65259bcc" Oct 05 08:47:50 crc kubenswrapper[4846]: E1005 08:47:50.499275 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 08:47:50 crc kubenswrapper[4846]: I1005 08:47:50.636063 4846 generic.go:334] "Generic (PLEG): container finished" podID="0c1a1b66-6c25-4694-afb7-bbeb68089408" containerID="1a9e7a9782a692ba92fdd3f59c9689e5c0b605e278b32a9780004974fd4094c1" exitCode=0 Oct 05 08:47:50 crc kubenswrapper[4846]: I1005 08:47:50.636158 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bfxmm" event={"ID":"0c1a1b66-6c25-4694-afb7-bbeb68089408","Type":"ContainerDied","Data":"1a9e7a9782a692ba92fdd3f59c9689e5c0b605e278b32a9780004974fd4094c1"} Oct 05 08:47:50 crc kubenswrapper[4846]: I1005 08:47:50.638876 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-openstack-openstack-cell1-l9jt5" event={"ID":"58a90b64-2441-4ccf-8655-a5a582dee3cc","Type":"ContainerStarted","Data":"c3f5ca1c52c00c02111452e4374da5c16c54203abe7e9eaca1618ed2cec9b42c"} Oct 05 08:47:50 crc kubenswrapper[4846]: I1005 08:47:50.697833 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-network-openstack-openstack-cell1-l9jt5" podStartSLOduration=2.159995671 podStartE2EDuration="2.697806037s" podCreationTimestamp="2025-10-05 08:47:48 +0000 UTC" firstStartedPulling="2025-10-05 08:47:49.584042553 +0000 UTC m=+7191.824895338" lastFinishedPulling="2025-10-05 08:47:50.121852919 +0000 UTC m=+7192.362705704" observedRunningTime="2025-10-05 08:47:50.682846402 +0000 UTC m=+7192.923699177" watchObservedRunningTime="2025-10-05 08:47:50.697806037 +0000 UTC m=+7192.938658852" Oct 05 08:47:51 crc kubenswrapper[4846]: I1005 08:47:51.652638 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bfxmm" event={"ID":"0c1a1b66-6c25-4694-afb7-bbeb68089408","Type":"ContainerStarted","Data":"7d4bcade55daf62f7943b2067a35486f5cdc9d2b8c8115e237a74e78a8d81276"} Oct 05 08:47:51 crc kubenswrapper[4846]: I1005 08:47:51.689237 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-bfxmm" podStartSLOduration=3.026152047 podStartE2EDuration="6.68921174s" podCreationTimestamp="2025-10-05 08:47:45 +0000 UTC" firstStartedPulling="2025-10-05 08:47:47.603797885 +0000 UTC m=+7189.844650700" lastFinishedPulling="2025-10-05 08:47:51.266857618 +0000 UTC m=+7193.507710393" observedRunningTime="2025-10-05 08:47:51.679277494 +0000 UTC m=+7193.920130269" watchObservedRunningTime="2025-10-05 08:47:51.68921174 +0000 UTC m=+7193.930064515" Oct 05 08:47:53 crc kubenswrapper[4846]: I1005 08:47:53.900386 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-nb2gm"] Oct 05 08:47:53 crc kubenswrapper[4846]: I1005 08:47:53.903383 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-nb2gm" Oct 05 08:47:53 crc kubenswrapper[4846]: I1005 08:47:53.925424 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-nb2gm"] Oct 05 08:47:53 crc kubenswrapper[4846]: I1005 08:47:53.992345 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/37eb33ef-2623-4413-b67f-2b6b3af85b1e-catalog-content\") pod \"redhat-marketplace-nb2gm\" (UID: \"37eb33ef-2623-4413-b67f-2b6b3af85b1e\") " pod="openshift-marketplace/redhat-marketplace-nb2gm" Oct 05 08:47:53 crc kubenswrapper[4846]: I1005 08:47:53.992450 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-68kt6\" (UniqueName: \"kubernetes.io/projected/37eb33ef-2623-4413-b67f-2b6b3af85b1e-kube-api-access-68kt6\") pod \"redhat-marketplace-nb2gm\" (UID: \"37eb33ef-2623-4413-b67f-2b6b3af85b1e\") " pod="openshift-marketplace/redhat-marketplace-nb2gm" Oct 05 08:47:53 crc kubenswrapper[4846]: I1005 08:47:53.992475 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/37eb33ef-2623-4413-b67f-2b6b3af85b1e-utilities\") pod \"redhat-marketplace-nb2gm\" (UID: \"37eb33ef-2623-4413-b67f-2b6b3af85b1e\") " pod="openshift-marketplace/redhat-marketplace-nb2gm" Oct 05 08:47:54 crc kubenswrapper[4846]: I1005 08:47:54.095808 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/37eb33ef-2623-4413-b67f-2b6b3af85b1e-catalog-content\") pod \"redhat-marketplace-nb2gm\" (UID: \"37eb33ef-2623-4413-b67f-2b6b3af85b1e\") " pod="openshift-marketplace/redhat-marketplace-nb2gm" Oct 05 08:47:54 crc kubenswrapper[4846]: I1005 08:47:54.095969 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-68kt6\" (UniqueName: \"kubernetes.io/projected/37eb33ef-2623-4413-b67f-2b6b3af85b1e-kube-api-access-68kt6\") pod \"redhat-marketplace-nb2gm\" (UID: \"37eb33ef-2623-4413-b67f-2b6b3af85b1e\") " pod="openshift-marketplace/redhat-marketplace-nb2gm" Oct 05 08:47:54 crc kubenswrapper[4846]: I1005 08:47:54.096003 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/37eb33ef-2623-4413-b67f-2b6b3af85b1e-utilities\") pod \"redhat-marketplace-nb2gm\" (UID: \"37eb33ef-2623-4413-b67f-2b6b3af85b1e\") " pod="openshift-marketplace/redhat-marketplace-nb2gm" Oct 05 08:47:54 crc kubenswrapper[4846]: I1005 08:47:54.096441 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/37eb33ef-2623-4413-b67f-2b6b3af85b1e-catalog-content\") pod \"redhat-marketplace-nb2gm\" (UID: \"37eb33ef-2623-4413-b67f-2b6b3af85b1e\") " pod="openshift-marketplace/redhat-marketplace-nb2gm" Oct 05 08:47:54 crc kubenswrapper[4846]: I1005 08:47:54.096659 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/37eb33ef-2623-4413-b67f-2b6b3af85b1e-utilities\") pod \"redhat-marketplace-nb2gm\" (UID: \"37eb33ef-2623-4413-b67f-2b6b3af85b1e\") " pod="openshift-marketplace/redhat-marketplace-nb2gm" Oct 05 08:47:54 crc kubenswrapper[4846]: I1005 08:47:54.125587 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-68kt6\" (UniqueName: \"kubernetes.io/projected/37eb33ef-2623-4413-b67f-2b6b3af85b1e-kube-api-access-68kt6\") pod \"redhat-marketplace-nb2gm\" (UID: \"37eb33ef-2623-4413-b67f-2b6b3af85b1e\") " pod="openshift-marketplace/redhat-marketplace-nb2gm" Oct 05 08:47:54 crc kubenswrapper[4846]: I1005 08:47:54.234865 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-nb2gm" Oct 05 08:47:54 crc kubenswrapper[4846]: I1005 08:47:54.759038 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-nb2gm"] Oct 05 08:47:54 crc kubenswrapper[4846]: W1005 08:47:54.770419 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod37eb33ef_2623_4413_b67f_2b6b3af85b1e.slice/crio-819f61d745e5bc8cc4a52130dd65f449a06f32a8c3c594388d4ab73d0bb70e43 WatchSource:0}: Error finding container 819f61d745e5bc8cc4a52130dd65f449a06f32a8c3c594388d4ab73d0bb70e43: Status 404 returned error can't find the container with id 819f61d745e5bc8cc4a52130dd65f449a06f32a8c3c594388d4ab73d0bb70e43 Oct 05 08:47:55 crc kubenswrapper[4846]: I1005 08:47:55.692931 4846 generic.go:334] "Generic (PLEG): container finished" podID="37eb33ef-2623-4413-b67f-2b6b3af85b1e" containerID="3a81be8ff1a22405f8c1d3101b2879fb1747bfb29bbd3efa21a8b5646d5b576b" exitCode=0 Oct 05 08:47:55 crc kubenswrapper[4846]: I1005 08:47:55.693316 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nb2gm" event={"ID":"37eb33ef-2623-4413-b67f-2b6b3af85b1e","Type":"ContainerDied","Data":"3a81be8ff1a22405f8c1d3101b2879fb1747bfb29bbd3efa21a8b5646d5b576b"} Oct 05 08:47:55 crc kubenswrapper[4846]: I1005 08:47:55.693367 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nb2gm" event={"ID":"37eb33ef-2623-4413-b67f-2b6b3af85b1e","Type":"ContainerStarted","Data":"819f61d745e5bc8cc4a52130dd65f449a06f32a8c3c594388d4ab73d0bb70e43"} Oct 05 08:47:56 crc kubenswrapper[4846]: I1005 08:47:56.085681 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-bfxmm" Oct 05 08:47:56 crc kubenswrapper[4846]: I1005 08:47:56.086373 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-bfxmm" Oct 05 08:47:56 crc kubenswrapper[4846]: I1005 08:47:56.164000 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-bfxmm" Oct 05 08:47:56 crc kubenswrapper[4846]: I1005 08:47:56.707115 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nb2gm" event={"ID":"37eb33ef-2623-4413-b67f-2b6b3af85b1e","Type":"ContainerStarted","Data":"8b022f5d10a38876e0a4c7a179464a64a3463d5c1a5a7c2f1bffc66c2e215dce"} Oct 05 08:47:56 crc kubenswrapper[4846]: I1005 08:47:56.762103 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-bfxmm" Oct 05 08:47:57 crc kubenswrapper[4846]: I1005 08:47:57.719649 4846 generic.go:334] "Generic (PLEG): container finished" podID="37eb33ef-2623-4413-b67f-2b6b3af85b1e" containerID="8b022f5d10a38876e0a4c7a179464a64a3463d5c1a5a7c2f1bffc66c2e215dce" exitCode=0 Oct 05 08:47:57 crc kubenswrapper[4846]: I1005 08:47:57.719707 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nb2gm" event={"ID":"37eb33ef-2623-4413-b67f-2b6b3af85b1e","Type":"ContainerDied","Data":"8b022f5d10a38876e0a4c7a179464a64a3463d5c1a5a7c2f1bffc66c2e215dce"} Oct 05 08:47:58 crc kubenswrapper[4846]: I1005 08:47:58.476073 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-bfxmm"] Oct 05 08:47:58 crc kubenswrapper[4846]: I1005 08:47:58.732913 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nb2gm" event={"ID":"37eb33ef-2623-4413-b67f-2b6b3af85b1e","Type":"ContainerStarted","Data":"e1811ba4672a079a3f5af7253aa3edba89024e39f75b7c80136c935122a9ed38"} Oct 05 08:47:58 crc kubenswrapper[4846]: I1005 08:47:58.762690 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-nb2gm" podStartSLOduration=3.280755023 podStartE2EDuration="5.762665786s" podCreationTimestamp="2025-10-05 08:47:53 +0000 UTC" firstStartedPulling="2025-10-05 08:47:55.696729037 +0000 UTC m=+7197.937581822" lastFinishedPulling="2025-10-05 08:47:58.1786398 +0000 UTC m=+7200.419492585" observedRunningTime="2025-10-05 08:47:58.753596063 +0000 UTC m=+7200.994448858" watchObservedRunningTime="2025-10-05 08:47:58.762665786 +0000 UTC m=+7201.003518601" Oct 05 08:47:59 crc kubenswrapper[4846]: I1005 08:47:59.742639 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-bfxmm" podUID="0c1a1b66-6c25-4694-afb7-bbeb68089408" containerName="registry-server" containerID="cri-o://7d4bcade55daf62f7943b2067a35486f5cdc9d2b8c8115e237a74e78a8d81276" gracePeriod=2 Oct 05 08:48:00 crc kubenswrapper[4846]: I1005 08:48:00.273617 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-bfxmm" Oct 05 08:48:00 crc kubenswrapper[4846]: I1005 08:48:00.336116 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-txg89\" (UniqueName: \"kubernetes.io/projected/0c1a1b66-6c25-4694-afb7-bbeb68089408-kube-api-access-txg89\") pod \"0c1a1b66-6c25-4694-afb7-bbeb68089408\" (UID: \"0c1a1b66-6c25-4694-afb7-bbeb68089408\") " Oct 05 08:48:00 crc kubenswrapper[4846]: I1005 08:48:00.336398 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0c1a1b66-6c25-4694-afb7-bbeb68089408-utilities\") pod \"0c1a1b66-6c25-4694-afb7-bbeb68089408\" (UID: \"0c1a1b66-6c25-4694-afb7-bbeb68089408\") " Oct 05 08:48:00 crc kubenswrapper[4846]: I1005 08:48:00.336508 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0c1a1b66-6c25-4694-afb7-bbeb68089408-catalog-content\") pod \"0c1a1b66-6c25-4694-afb7-bbeb68089408\" (UID: \"0c1a1b66-6c25-4694-afb7-bbeb68089408\") " Oct 05 08:48:00 crc kubenswrapper[4846]: I1005 08:48:00.337312 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0c1a1b66-6c25-4694-afb7-bbeb68089408-utilities" (OuterVolumeSpecName: "utilities") pod "0c1a1b66-6c25-4694-afb7-bbeb68089408" (UID: "0c1a1b66-6c25-4694-afb7-bbeb68089408"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 08:48:00 crc kubenswrapper[4846]: I1005 08:48:00.339413 4846 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0c1a1b66-6c25-4694-afb7-bbeb68089408-utilities\") on node \"crc\" DevicePath \"\"" Oct 05 08:48:00 crc kubenswrapper[4846]: I1005 08:48:00.346484 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0c1a1b66-6c25-4694-afb7-bbeb68089408-kube-api-access-txg89" (OuterVolumeSpecName: "kube-api-access-txg89") pod "0c1a1b66-6c25-4694-afb7-bbeb68089408" (UID: "0c1a1b66-6c25-4694-afb7-bbeb68089408"). InnerVolumeSpecName "kube-api-access-txg89". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:48:00 crc kubenswrapper[4846]: I1005 08:48:00.380887 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0c1a1b66-6c25-4694-afb7-bbeb68089408-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0c1a1b66-6c25-4694-afb7-bbeb68089408" (UID: "0c1a1b66-6c25-4694-afb7-bbeb68089408"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 08:48:00 crc kubenswrapper[4846]: I1005 08:48:00.441358 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-txg89\" (UniqueName: \"kubernetes.io/projected/0c1a1b66-6c25-4694-afb7-bbeb68089408-kube-api-access-txg89\") on node \"crc\" DevicePath \"\"" Oct 05 08:48:00 crc kubenswrapper[4846]: I1005 08:48:00.441410 4846 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0c1a1b66-6c25-4694-afb7-bbeb68089408-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 05 08:48:00 crc kubenswrapper[4846]: I1005 08:48:00.761437 4846 generic.go:334] "Generic (PLEG): container finished" podID="0c1a1b66-6c25-4694-afb7-bbeb68089408" containerID="7d4bcade55daf62f7943b2067a35486f5cdc9d2b8c8115e237a74e78a8d81276" exitCode=0 Oct 05 08:48:00 crc kubenswrapper[4846]: I1005 08:48:00.761505 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bfxmm" event={"ID":"0c1a1b66-6c25-4694-afb7-bbeb68089408","Type":"ContainerDied","Data":"7d4bcade55daf62f7943b2067a35486f5cdc9d2b8c8115e237a74e78a8d81276"} Oct 05 08:48:00 crc kubenswrapper[4846]: I1005 08:48:00.761543 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bfxmm" event={"ID":"0c1a1b66-6c25-4694-afb7-bbeb68089408","Type":"ContainerDied","Data":"692cd14fc839604f27b25abb63253a3dda953c7de5a2364087f3cfd30aec7da9"} Oct 05 08:48:00 crc kubenswrapper[4846]: I1005 08:48:00.761566 4846 scope.go:117] "RemoveContainer" containerID="7d4bcade55daf62f7943b2067a35486f5cdc9d2b8c8115e237a74e78a8d81276" Oct 05 08:48:00 crc kubenswrapper[4846]: I1005 08:48:00.761560 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-bfxmm" Oct 05 08:48:00 crc kubenswrapper[4846]: I1005 08:48:00.810058 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-bfxmm"] Oct 05 08:48:00 crc kubenswrapper[4846]: I1005 08:48:00.814027 4846 scope.go:117] "RemoveContainer" containerID="1a9e7a9782a692ba92fdd3f59c9689e5c0b605e278b32a9780004974fd4094c1" Oct 05 08:48:00 crc kubenswrapper[4846]: I1005 08:48:00.823304 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-bfxmm"] Oct 05 08:48:00 crc kubenswrapper[4846]: I1005 08:48:00.841793 4846 scope.go:117] "RemoveContainer" containerID="7e1f323f362a3c4d472f0929145b0e2e6b4104b2aa60194a06662f6b63d95a69" Oct 05 08:48:00 crc kubenswrapper[4846]: I1005 08:48:00.913170 4846 scope.go:117] "RemoveContainer" containerID="7d4bcade55daf62f7943b2067a35486f5cdc9d2b8c8115e237a74e78a8d81276" Oct 05 08:48:00 crc kubenswrapper[4846]: E1005 08:48:00.913936 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7d4bcade55daf62f7943b2067a35486f5cdc9d2b8c8115e237a74e78a8d81276\": container with ID starting with 7d4bcade55daf62f7943b2067a35486f5cdc9d2b8c8115e237a74e78a8d81276 not found: ID does not exist" containerID="7d4bcade55daf62f7943b2067a35486f5cdc9d2b8c8115e237a74e78a8d81276" Oct 05 08:48:00 crc kubenswrapper[4846]: I1005 08:48:00.913994 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7d4bcade55daf62f7943b2067a35486f5cdc9d2b8c8115e237a74e78a8d81276"} err="failed to get container status \"7d4bcade55daf62f7943b2067a35486f5cdc9d2b8c8115e237a74e78a8d81276\": rpc error: code = NotFound desc = could not find container \"7d4bcade55daf62f7943b2067a35486f5cdc9d2b8c8115e237a74e78a8d81276\": container with ID starting with 7d4bcade55daf62f7943b2067a35486f5cdc9d2b8c8115e237a74e78a8d81276 not found: ID does not exist" Oct 05 08:48:00 crc kubenswrapper[4846]: I1005 08:48:00.914055 4846 scope.go:117] "RemoveContainer" containerID="1a9e7a9782a692ba92fdd3f59c9689e5c0b605e278b32a9780004974fd4094c1" Oct 05 08:48:00 crc kubenswrapper[4846]: E1005 08:48:00.914625 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1a9e7a9782a692ba92fdd3f59c9689e5c0b605e278b32a9780004974fd4094c1\": container with ID starting with 1a9e7a9782a692ba92fdd3f59c9689e5c0b605e278b32a9780004974fd4094c1 not found: ID does not exist" containerID="1a9e7a9782a692ba92fdd3f59c9689e5c0b605e278b32a9780004974fd4094c1" Oct 05 08:48:00 crc kubenswrapper[4846]: I1005 08:48:00.914670 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1a9e7a9782a692ba92fdd3f59c9689e5c0b605e278b32a9780004974fd4094c1"} err="failed to get container status \"1a9e7a9782a692ba92fdd3f59c9689e5c0b605e278b32a9780004974fd4094c1\": rpc error: code = NotFound desc = could not find container \"1a9e7a9782a692ba92fdd3f59c9689e5c0b605e278b32a9780004974fd4094c1\": container with ID starting with 1a9e7a9782a692ba92fdd3f59c9689e5c0b605e278b32a9780004974fd4094c1 not found: ID does not exist" Oct 05 08:48:00 crc kubenswrapper[4846]: I1005 08:48:00.914696 4846 scope.go:117] "RemoveContainer" containerID="7e1f323f362a3c4d472f0929145b0e2e6b4104b2aa60194a06662f6b63d95a69" Oct 05 08:48:00 crc kubenswrapper[4846]: E1005 08:48:00.915009 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7e1f323f362a3c4d472f0929145b0e2e6b4104b2aa60194a06662f6b63d95a69\": container with ID starting with 7e1f323f362a3c4d472f0929145b0e2e6b4104b2aa60194a06662f6b63d95a69 not found: ID does not exist" containerID="7e1f323f362a3c4d472f0929145b0e2e6b4104b2aa60194a06662f6b63d95a69" Oct 05 08:48:00 crc kubenswrapper[4846]: I1005 08:48:00.915049 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7e1f323f362a3c4d472f0929145b0e2e6b4104b2aa60194a06662f6b63d95a69"} err="failed to get container status \"7e1f323f362a3c4d472f0929145b0e2e6b4104b2aa60194a06662f6b63d95a69\": rpc error: code = NotFound desc = could not find container \"7e1f323f362a3c4d472f0929145b0e2e6b4104b2aa60194a06662f6b63d95a69\": container with ID starting with 7e1f323f362a3c4d472f0929145b0e2e6b4104b2aa60194a06662f6b63d95a69 not found: ID does not exist" Oct 05 08:48:02 crc kubenswrapper[4846]: I1005 08:48:02.498927 4846 scope.go:117] "RemoveContainer" containerID="0d84b906773347b38ec2ce550f580254430238b4374f3171cb7453fa65259bcc" Oct 05 08:48:02 crc kubenswrapper[4846]: E1005 08:48:02.499599 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 08:48:02 crc kubenswrapper[4846]: I1005 08:48:02.517796 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0c1a1b66-6c25-4694-afb7-bbeb68089408" path="/var/lib/kubelet/pods/0c1a1b66-6c25-4694-afb7-bbeb68089408/volumes" Oct 05 08:48:04 crc kubenswrapper[4846]: I1005 08:48:04.236124 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-nb2gm" Oct 05 08:48:04 crc kubenswrapper[4846]: I1005 08:48:04.236509 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-nb2gm" Oct 05 08:48:04 crc kubenswrapper[4846]: I1005 08:48:04.289556 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-nb2gm" Oct 05 08:48:04 crc kubenswrapper[4846]: I1005 08:48:04.492103 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-mpm8d"] Oct 05 08:48:04 crc kubenswrapper[4846]: E1005 08:48:04.492608 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0c1a1b66-6c25-4694-afb7-bbeb68089408" containerName="extract-utilities" Oct 05 08:48:04 crc kubenswrapper[4846]: I1005 08:48:04.492628 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="0c1a1b66-6c25-4694-afb7-bbeb68089408" containerName="extract-utilities" Oct 05 08:48:04 crc kubenswrapper[4846]: E1005 08:48:04.492653 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0c1a1b66-6c25-4694-afb7-bbeb68089408" containerName="extract-content" Oct 05 08:48:04 crc kubenswrapper[4846]: I1005 08:48:04.492663 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="0c1a1b66-6c25-4694-afb7-bbeb68089408" containerName="extract-content" Oct 05 08:48:04 crc kubenswrapper[4846]: E1005 08:48:04.492679 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0c1a1b66-6c25-4694-afb7-bbeb68089408" containerName="registry-server" Oct 05 08:48:04 crc kubenswrapper[4846]: I1005 08:48:04.492688 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="0c1a1b66-6c25-4694-afb7-bbeb68089408" containerName="registry-server" Oct 05 08:48:04 crc kubenswrapper[4846]: I1005 08:48:04.492933 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="0c1a1b66-6c25-4694-afb7-bbeb68089408" containerName="registry-server" Oct 05 08:48:04 crc kubenswrapper[4846]: I1005 08:48:04.494851 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-mpm8d" Oct 05 08:48:04 crc kubenswrapper[4846]: I1005 08:48:04.540492 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-mpm8d"] Oct 05 08:48:04 crc kubenswrapper[4846]: I1005 08:48:04.541416 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-frn2m\" (UniqueName: \"kubernetes.io/projected/0c0156d2-8b61-4267-91a6-99d6e4197459-kube-api-access-frn2m\") pod \"redhat-operators-mpm8d\" (UID: \"0c0156d2-8b61-4267-91a6-99d6e4197459\") " pod="openshift-marketplace/redhat-operators-mpm8d" Oct 05 08:48:04 crc kubenswrapper[4846]: I1005 08:48:04.541522 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0c0156d2-8b61-4267-91a6-99d6e4197459-utilities\") pod \"redhat-operators-mpm8d\" (UID: \"0c0156d2-8b61-4267-91a6-99d6e4197459\") " pod="openshift-marketplace/redhat-operators-mpm8d" Oct 05 08:48:04 crc kubenswrapper[4846]: I1005 08:48:04.541704 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0c0156d2-8b61-4267-91a6-99d6e4197459-catalog-content\") pod \"redhat-operators-mpm8d\" (UID: \"0c0156d2-8b61-4267-91a6-99d6e4197459\") " pod="openshift-marketplace/redhat-operators-mpm8d" Oct 05 08:48:04 crc kubenswrapper[4846]: I1005 08:48:04.644063 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0c0156d2-8b61-4267-91a6-99d6e4197459-utilities\") pod \"redhat-operators-mpm8d\" (UID: \"0c0156d2-8b61-4267-91a6-99d6e4197459\") " pod="openshift-marketplace/redhat-operators-mpm8d" Oct 05 08:48:04 crc kubenswrapper[4846]: I1005 08:48:04.644815 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0c0156d2-8b61-4267-91a6-99d6e4197459-utilities\") pod \"redhat-operators-mpm8d\" (UID: \"0c0156d2-8b61-4267-91a6-99d6e4197459\") " pod="openshift-marketplace/redhat-operators-mpm8d" Oct 05 08:48:04 crc kubenswrapper[4846]: I1005 08:48:04.645360 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0c0156d2-8b61-4267-91a6-99d6e4197459-catalog-content\") pod \"redhat-operators-mpm8d\" (UID: \"0c0156d2-8b61-4267-91a6-99d6e4197459\") " pod="openshift-marketplace/redhat-operators-mpm8d" Oct 05 08:48:04 crc kubenswrapper[4846]: I1005 08:48:04.645520 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-frn2m\" (UniqueName: \"kubernetes.io/projected/0c0156d2-8b61-4267-91a6-99d6e4197459-kube-api-access-frn2m\") pod \"redhat-operators-mpm8d\" (UID: \"0c0156d2-8b61-4267-91a6-99d6e4197459\") " pod="openshift-marketplace/redhat-operators-mpm8d" Oct 05 08:48:04 crc kubenswrapper[4846]: I1005 08:48:04.646032 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0c0156d2-8b61-4267-91a6-99d6e4197459-catalog-content\") pod \"redhat-operators-mpm8d\" (UID: \"0c0156d2-8b61-4267-91a6-99d6e4197459\") " pod="openshift-marketplace/redhat-operators-mpm8d" Oct 05 08:48:04 crc kubenswrapper[4846]: I1005 08:48:04.684316 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-frn2m\" (UniqueName: \"kubernetes.io/projected/0c0156d2-8b61-4267-91a6-99d6e4197459-kube-api-access-frn2m\") pod \"redhat-operators-mpm8d\" (UID: \"0c0156d2-8b61-4267-91a6-99d6e4197459\") " pod="openshift-marketplace/redhat-operators-mpm8d" Oct 05 08:48:04 crc kubenswrapper[4846]: I1005 08:48:04.816422 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-mpm8d" Oct 05 08:48:04 crc kubenswrapper[4846]: I1005 08:48:04.882662 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-nb2gm" Oct 05 08:48:05 crc kubenswrapper[4846]: I1005 08:48:05.291051 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-mpm8d"] Oct 05 08:48:05 crc kubenswrapper[4846]: W1005 08:48:05.292072 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0c0156d2_8b61_4267_91a6_99d6e4197459.slice/crio-652f2ac8bc4231868a7174918fac4a20859732e18c57a2eed71a212ba3225d40 WatchSource:0}: Error finding container 652f2ac8bc4231868a7174918fac4a20859732e18c57a2eed71a212ba3225d40: Status 404 returned error can't find the container with id 652f2ac8bc4231868a7174918fac4a20859732e18c57a2eed71a212ba3225d40 Oct 05 08:48:05 crc kubenswrapper[4846]: I1005 08:48:05.813882 4846 generic.go:334] "Generic (PLEG): container finished" podID="0c0156d2-8b61-4267-91a6-99d6e4197459" containerID="787d1f6b62a4aea6ec4b1f13d3f89f5917cfd84353636adbaef74d0672bc19f9" exitCode=0 Oct 05 08:48:05 crc kubenswrapper[4846]: I1005 08:48:05.813985 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mpm8d" event={"ID":"0c0156d2-8b61-4267-91a6-99d6e4197459","Type":"ContainerDied","Data":"787d1f6b62a4aea6ec4b1f13d3f89f5917cfd84353636adbaef74d0672bc19f9"} Oct 05 08:48:05 crc kubenswrapper[4846]: I1005 08:48:05.814041 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mpm8d" event={"ID":"0c0156d2-8b61-4267-91a6-99d6e4197459","Type":"ContainerStarted","Data":"652f2ac8bc4231868a7174918fac4a20859732e18c57a2eed71a212ba3225d40"} Oct 05 08:48:06 crc kubenswrapper[4846]: I1005 08:48:06.668261 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-nb2gm"] Oct 05 08:48:06 crc kubenswrapper[4846]: I1005 08:48:06.829638 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mpm8d" event={"ID":"0c0156d2-8b61-4267-91a6-99d6e4197459","Type":"ContainerStarted","Data":"f68445c8d288d6c301ed5e3c0ec1734846e10d5a378566118fb9c43f313c329c"} Oct 05 08:48:06 crc kubenswrapper[4846]: I1005 08:48:06.829781 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-nb2gm" podUID="37eb33ef-2623-4413-b67f-2b6b3af85b1e" containerName="registry-server" containerID="cri-o://e1811ba4672a079a3f5af7253aa3edba89024e39f75b7c80136c935122a9ed38" gracePeriod=2 Oct 05 08:48:07 crc kubenswrapper[4846]: I1005 08:48:07.316327 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-nb2gm" Oct 05 08:48:07 crc kubenswrapper[4846]: I1005 08:48:07.512511 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/37eb33ef-2623-4413-b67f-2b6b3af85b1e-utilities\") pod \"37eb33ef-2623-4413-b67f-2b6b3af85b1e\" (UID: \"37eb33ef-2623-4413-b67f-2b6b3af85b1e\") " Oct 05 08:48:07 crc kubenswrapper[4846]: I1005 08:48:07.512613 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/37eb33ef-2623-4413-b67f-2b6b3af85b1e-catalog-content\") pod \"37eb33ef-2623-4413-b67f-2b6b3af85b1e\" (UID: \"37eb33ef-2623-4413-b67f-2b6b3af85b1e\") " Oct 05 08:48:07 crc kubenswrapper[4846]: I1005 08:48:07.512820 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-68kt6\" (UniqueName: \"kubernetes.io/projected/37eb33ef-2623-4413-b67f-2b6b3af85b1e-kube-api-access-68kt6\") pod \"37eb33ef-2623-4413-b67f-2b6b3af85b1e\" (UID: \"37eb33ef-2623-4413-b67f-2b6b3af85b1e\") " Oct 05 08:48:07 crc kubenswrapper[4846]: I1005 08:48:07.513641 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/37eb33ef-2623-4413-b67f-2b6b3af85b1e-utilities" (OuterVolumeSpecName: "utilities") pod "37eb33ef-2623-4413-b67f-2b6b3af85b1e" (UID: "37eb33ef-2623-4413-b67f-2b6b3af85b1e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 08:48:07 crc kubenswrapper[4846]: I1005 08:48:07.514682 4846 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/37eb33ef-2623-4413-b67f-2b6b3af85b1e-utilities\") on node \"crc\" DevicePath \"\"" Oct 05 08:48:07 crc kubenswrapper[4846]: I1005 08:48:07.522577 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/37eb33ef-2623-4413-b67f-2b6b3af85b1e-kube-api-access-68kt6" (OuterVolumeSpecName: "kube-api-access-68kt6") pod "37eb33ef-2623-4413-b67f-2b6b3af85b1e" (UID: "37eb33ef-2623-4413-b67f-2b6b3af85b1e"). InnerVolumeSpecName "kube-api-access-68kt6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:48:07 crc kubenswrapper[4846]: I1005 08:48:07.537890 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/37eb33ef-2623-4413-b67f-2b6b3af85b1e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "37eb33ef-2623-4413-b67f-2b6b3af85b1e" (UID: "37eb33ef-2623-4413-b67f-2b6b3af85b1e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 08:48:07 crc kubenswrapper[4846]: I1005 08:48:07.617694 4846 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/37eb33ef-2623-4413-b67f-2b6b3af85b1e-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 05 08:48:07 crc kubenswrapper[4846]: I1005 08:48:07.617724 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-68kt6\" (UniqueName: \"kubernetes.io/projected/37eb33ef-2623-4413-b67f-2b6b3af85b1e-kube-api-access-68kt6\") on node \"crc\" DevicePath \"\"" Oct 05 08:48:07 crc kubenswrapper[4846]: I1005 08:48:07.846547 4846 generic.go:334] "Generic (PLEG): container finished" podID="37eb33ef-2623-4413-b67f-2b6b3af85b1e" containerID="e1811ba4672a079a3f5af7253aa3edba89024e39f75b7c80136c935122a9ed38" exitCode=0 Oct 05 08:48:07 crc kubenswrapper[4846]: I1005 08:48:07.846594 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-nb2gm" Oct 05 08:48:07 crc kubenswrapper[4846]: I1005 08:48:07.846601 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nb2gm" event={"ID":"37eb33ef-2623-4413-b67f-2b6b3af85b1e","Type":"ContainerDied","Data":"e1811ba4672a079a3f5af7253aa3edba89024e39f75b7c80136c935122a9ed38"} Oct 05 08:48:07 crc kubenswrapper[4846]: I1005 08:48:07.846830 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nb2gm" event={"ID":"37eb33ef-2623-4413-b67f-2b6b3af85b1e","Type":"ContainerDied","Data":"819f61d745e5bc8cc4a52130dd65f449a06f32a8c3c594388d4ab73d0bb70e43"} Oct 05 08:48:07 crc kubenswrapper[4846]: I1005 08:48:07.846864 4846 scope.go:117] "RemoveContainer" containerID="e1811ba4672a079a3f5af7253aa3edba89024e39f75b7c80136c935122a9ed38" Oct 05 08:48:07 crc kubenswrapper[4846]: I1005 08:48:07.904338 4846 scope.go:117] "RemoveContainer" containerID="8b022f5d10a38876e0a4c7a179464a64a3463d5c1a5a7c2f1bffc66c2e215dce" Oct 05 08:48:07 crc kubenswrapper[4846]: I1005 08:48:07.908126 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-nb2gm"] Oct 05 08:48:07 crc kubenswrapper[4846]: I1005 08:48:07.918517 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-nb2gm"] Oct 05 08:48:07 crc kubenswrapper[4846]: I1005 08:48:07.928106 4846 scope.go:117] "RemoveContainer" containerID="3a81be8ff1a22405f8c1d3101b2879fb1747bfb29bbd3efa21a8b5646d5b576b" Oct 05 08:48:07 crc kubenswrapper[4846]: I1005 08:48:07.984603 4846 scope.go:117] "RemoveContainer" containerID="e1811ba4672a079a3f5af7253aa3edba89024e39f75b7c80136c935122a9ed38" Oct 05 08:48:07 crc kubenswrapper[4846]: E1005 08:48:07.985805 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e1811ba4672a079a3f5af7253aa3edba89024e39f75b7c80136c935122a9ed38\": container with ID starting with e1811ba4672a079a3f5af7253aa3edba89024e39f75b7c80136c935122a9ed38 not found: ID does not exist" containerID="e1811ba4672a079a3f5af7253aa3edba89024e39f75b7c80136c935122a9ed38" Oct 05 08:48:07 crc kubenswrapper[4846]: I1005 08:48:07.986144 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e1811ba4672a079a3f5af7253aa3edba89024e39f75b7c80136c935122a9ed38"} err="failed to get container status \"e1811ba4672a079a3f5af7253aa3edba89024e39f75b7c80136c935122a9ed38\": rpc error: code = NotFound desc = could not find container \"e1811ba4672a079a3f5af7253aa3edba89024e39f75b7c80136c935122a9ed38\": container with ID starting with e1811ba4672a079a3f5af7253aa3edba89024e39f75b7c80136c935122a9ed38 not found: ID does not exist" Oct 05 08:48:07 crc kubenswrapper[4846]: I1005 08:48:07.986397 4846 scope.go:117] "RemoveContainer" containerID="8b022f5d10a38876e0a4c7a179464a64a3463d5c1a5a7c2f1bffc66c2e215dce" Oct 05 08:48:07 crc kubenswrapper[4846]: E1005 08:48:07.987554 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8b022f5d10a38876e0a4c7a179464a64a3463d5c1a5a7c2f1bffc66c2e215dce\": container with ID starting with 8b022f5d10a38876e0a4c7a179464a64a3463d5c1a5a7c2f1bffc66c2e215dce not found: ID does not exist" containerID="8b022f5d10a38876e0a4c7a179464a64a3463d5c1a5a7c2f1bffc66c2e215dce" Oct 05 08:48:07 crc kubenswrapper[4846]: I1005 08:48:07.987756 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8b022f5d10a38876e0a4c7a179464a64a3463d5c1a5a7c2f1bffc66c2e215dce"} err="failed to get container status \"8b022f5d10a38876e0a4c7a179464a64a3463d5c1a5a7c2f1bffc66c2e215dce\": rpc error: code = NotFound desc = could not find container \"8b022f5d10a38876e0a4c7a179464a64a3463d5c1a5a7c2f1bffc66c2e215dce\": container with ID starting with 8b022f5d10a38876e0a4c7a179464a64a3463d5c1a5a7c2f1bffc66c2e215dce not found: ID does not exist" Oct 05 08:48:07 crc kubenswrapper[4846]: I1005 08:48:07.987945 4846 scope.go:117] "RemoveContainer" containerID="3a81be8ff1a22405f8c1d3101b2879fb1747bfb29bbd3efa21a8b5646d5b576b" Oct 05 08:48:07 crc kubenswrapper[4846]: E1005 08:48:07.988748 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3a81be8ff1a22405f8c1d3101b2879fb1747bfb29bbd3efa21a8b5646d5b576b\": container with ID starting with 3a81be8ff1a22405f8c1d3101b2879fb1747bfb29bbd3efa21a8b5646d5b576b not found: ID does not exist" containerID="3a81be8ff1a22405f8c1d3101b2879fb1747bfb29bbd3efa21a8b5646d5b576b" Oct 05 08:48:07 crc kubenswrapper[4846]: I1005 08:48:07.988797 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3a81be8ff1a22405f8c1d3101b2879fb1747bfb29bbd3efa21a8b5646d5b576b"} err="failed to get container status \"3a81be8ff1a22405f8c1d3101b2879fb1747bfb29bbd3efa21a8b5646d5b576b\": rpc error: code = NotFound desc = could not find container \"3a81be8ff1a22405f8c1d3101b2879fb1747bfb29bbd3efa21a8b5646d5b576b\": container with ID starting with 3a81be8ff1a22405f8c1d3101b2879fb1747bfb29bbd3efa21a8b5646d5b576b not found: ID does not exist" Oct 05 08:48:08 crc kubenswrapper[4846]: I1005 08:48:08.514834 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="37eb33ef-2623-4413-b67f-2b6b3af85b1e" path="/var/lib/kubelet/pods/37eb33ef-2623-4413-b67f-2b6b3af85b1e/volumes" Oct 05 08:48:10 crc kubenswrapper[4846]: I1005 08:48:10.915289 4846 generic.go:334] "Generic (PLEG): container finished" podID="0c0156d2-8b61-4267-91a6-99d6e4197459" containerID="f68445c8d288d6c301ed5e3c0ec1734846e10d5a378566118fb9c43f313c329c" exitCode=0 Oct 05 08:48:10 crc kubenswrapper[4846]: I1005 08:48:10.915645 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mpm8d" event={"ID":"0c0156d2-8b61-4267-91a6-99d6e4197459","Type":"ContainerDied","Data":"f68445c8d288d6c301ed5e3c0ec1734846e10d5a378566118fb9c43f313c329c"} Oct 05 08:48:10 crc kubenswrapper[4846]: I1005 08:48:10.931053 4846 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 05 08:48:11 crc kubenswrapper[4846]: I1005 08:48:11.929656 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mpm8d" event={"ID":"0c0156d2-8b61-4267-91a6-99d6e4197459","Type":"ContainerStarted","Data":"ccf2d9cd245abc05fb38ededfbfbece8d493be5431c1a4a609771ee3ff5d209f"} Oct 05 08:48:11 crc kubenswrapper[4846]: I1005 08:48:11.958145 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-mpm8d" podStartSLOduration=2.38281929 podStartE2EDuration="7.958127849s" podCreationTimestamp="2025-10-05 08:48:04 +0000 UTC" firstStartedPulling="2025-10-05 08:48:05.815503354 +0000 UTC m=+7208.056356129" lastFinishedPulling="2025-10-05 08:48:11.390811913 +0000 UTC m=+7213.631664688" observedRunningTime="2025-10-05 08:48:11.954620019 +0000 UTC m=+7214.195472784" watchObservedRunningTime="2025-10-05 08:48:11.958127849 +0000 UTC m=+7214.198980624" Oct 05 08:48:14 crc kubenswrapper[4846]: I1005 08:48:14.816717 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-mpm8d" Oct 05 08:48:14 crc kubenswrapper[4846]: I1005 08:48:14.817380 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-mpm8d" Oct 05 08:48:15 crc kubenswrapper[4846]: I1005 08:48:15.884666 4846 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-mpm8d" podUID="0c0156d2-8b61-4267-91a6-99d6e4197459" containerName="registry-server" probeResult="failure" output=< Oct 05 08:48:15 crc kubenswrapper[4846]: timeout: failed to connect service ":50051" within 1s Oct 05 08:48:15 crc kubenswrapper[4846]: > Oct 05 08:48:16 crc kubenswrapper[4846]: I1005 08:48:16.497840 4846 scope.go:117] "RemoveContainer" containerID="0d84b906773347b38ec2ce550f580254430238b4374f3171cb7453fa65259bcc" Oct 05 08:48:16 crc kubenswrapper[4846]: E1005 08:48:16.498549 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 08:48:24 crc kubenswrapper[4846]: I1005 08:48:24.884943 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-mpm8d" Oct 05 08:48:24 crc kubenswrapper[4846]: I1005 08:48:24.966431 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-mpm8d" Oct 05 08:48:25 crc kubenswrapper[4846]: I1005 08:48:25.132827 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-mpm8d"] Oct 05 08:48:26 crc kubenswrapper[4846]: I1005 08:48:26.070438 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-mpm8d" podUID="0c0156d2-8b61-4267-91a6-99d6e4197459" containerName="registry-server" containerID="cri-o://ccf2d9cd245abc05fb38ededfbfbece8d493be5431c1a4a609771ee3ff5d209f" gracePeriod=2 Oct 05 08:48:26 crc kubenswrapper[4846]: I1005 08:48:26.601697 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-mpm8d" Oct 05 08:48:26 crc kubenswrapper[4846]: I1005 08:48:26.656873 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0c0156d2-8b61-4267-91a6-99d6e4197459-utilities\") pod \"0c0156d2-8b61-4267-91a6-99d6e4197459\" (UID: \"0c0156d2-8b61-4267-91a6-99d6e4197459\") " Oct 05 08:48:26 crc kubenswrapper[4846]: I1005 08:48:26.657102 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0c0156d2-8b61-4267-91a6-99d6e4197459-catalog-content\") pod \"0c0156d2-8b61-4267-91a6-99d6e4197459\" (UID: \"0c0156d2-8b61-4267-91a6-99d6e4197459\") " Oct 05 08:48:26 crc kubenswrapper[4846]: I1005 08:48:26.657234 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-frn2m\" (UniqueName: \"kubernetes.io/projected/0c0156d2-8b61-4267-91a6-99d6e4197459-kube-api-access-frn2m\") pod \"0c0156d2-8b61-4267-91a6-99d6e4197459\" (UID: \"0c0156d2-8b61-4267-91a6-99d6e4197459\") " Oct 05 08:48:26 crc kubenswrapper[4846]: I1005 08:48:26.657878 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0c0156d2-8b61-4267-91a6-99d6e4197459-utilities" (OuterVolumeSpecName: "utilities") pod "0c0156d2-8b61-4267-91a6-99d6e4197459" (UID: "0c0156d2-8b61-4267-91a6-99d6e4197459"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 08:48:26 crc kubenswrapper[4846]: I1005 08:48:26.673143 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0c0156d2-8b61-4267-91a6-99d6e4197459-kube-api-access-frn2m" (OuterVolumeSpecName: "kube-api-access-frn2m") pod "0c0156d2-8b61-4267-91a6-99d6e4197459" (UID: "0c0156d2-8b61-4267-91a6-99d6e4197459"). InnerVolumeSpecName "kube-api-access-frn2m". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:48:26 crc kubenswrapper[4846]: I1005 08:48:26.749985 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0c0156d2-8b61-4267-91a6-99d6e4197459-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0c0156d2-8b61-4267-91a6-99d6e4197459" (UID: "0c0156d2-8b61-4267-91a6-99d6e4197459"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 08:48:26 crc kubenswrapper[4846]: I1005 08:48:26.759530 4846 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0c0156d2-8b61-4267-91a6-99d6e4197459-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 05 08:48:26 crc kubenswrapper[4846]: I1005 08:48:26.759560 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-frn2m\" (UniqueName: \"kubernetes.io/projected/0c0156d2-8b61-4267-91a6-99d6e4197459-kube-api-access-frn2m\") on node \"crc\" DevicePath \"\"" Oct 05 08:48:26 crc kubenswrapper[4846]: I1005 08:48:26.759574 4846 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0c0156d2-8b61-4267-91a6-99d6e4197459-utilities\") on node \"crc\" DevicePath \"\"" Oct 05 08:48:27 crc kubenswrapper[4846]: I1005 08:48:27.094903 4846 generic.go:334] "Generic (PLEG): container finished" podID="0c0156d2-8b61-4267-91a6-99d6e4197459" containerID="ccf2d9cd245abc05fb38ededfbfbece8d493be5431c1a4a609771ee3ff5d209f" exitCode=0 Oct 05 08:48:27 crc kubenswrapper[4846]: I1005 08:48:27.094974 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mpm8d" event={"ID":"0c0156d2-8b61-4267-91a6-99d6e4197459","Type":"ContainerDied","Data":"ccf2d9cd245abc05fb38ededfbfbece8d493be5431c1a4a609771ee3ff5d209f"} Oct 05 08:48:27 crc kubenswrapper[4846]: I1005 08:48:27.095411 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mpm8d" event={"ID":"0c0156d2-8b61-4267-91a6-99d6e4197459","Type":"ContainerDied","Data":"652f2ac8bc4231868a7174918fac4a20859732e18c57a2eed71a212ba3225d40"} Oct 05 08:48:27 crc kubenswrapper[4846]: I1005 08:48:27.095449 4846 scope.go:117] "RemoveContainer" containerID="ccf2d9cd245abc05fb38ededfbfbece8d493be5431c1a4a609771ee3ff5d209f" Oct 05 08:48:27 crc kubenswrapper[4846]: I1005 08:48:27.095013 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-mpm8d" Oct 05 08:48:27 crc kubenswrapper[4846]: I1005 08:48:27.144099 4846 scope.go:117] "RemoveContainer" containerID="f68445c8d288d6c301ed5e3c0ec1734846e10d5a378566118fb9c43f313c329c" Oct 05 08:48:27 crc kubenswrapper[4846]: I1005 08:48:27.167043 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-mpm8d"] Oct 05 08:48:27 crc kubenswrapper[4846]: I1005 08:48:27.185794 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-mpm8d"] Oct 05 08:48:27 crc kubenswrapper[4846]: I1005 08:48:27.199718 4846 scope.go:117] "RemoveContainer" containerID="787d1f6b62a4aea6ec4b1f13d3f89f5917cfd84353636adbaef74d0672bc19f9" Oct 05 08:48:27 crc kubenswrapper[4846]: I1005 08:48:27.261942 4846 scope.go:117] "RemoveContainer" containerID="ccf2d9cd245abc05fb38ededfbfbece8d493be5431c1a4a609771ee3ff5d209f" Oct 05 08:48:27 crc kubenswrapper[4846]: E1005 08:48:27.262494 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ccf2d9cd245abc05fb38ededfbfbece8d493be5431c1a4a609771ee3ff5d209f\": container with ID starting with ccf2d9cd245abc05fb38ededfbfbece8d493be5431c1a4a609771ee3ff5d209f not found: ID does not exist" containerID="ccf2d9cd245abc05fb38ededfbfbece8d493be5431c1a4a609771ee3ff5d209f" Oct 05 08:48:27 crc kubenswrapper[4846]: I1005 08:48:27.262533 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ccf2d9cd245abc05fb38ededfbfbece8d493be5431c1a4a609771ee3ff5d209f"} err="failed to get container status \"ccf2d9cd245abc05fb38ededfbfbece8d493be5431c1a4a609771ee3ff5d209f\": rpc error: code = NotFound desc = could not find container \"ccf2d9cd245abc05fb38ededfbfbece8d493be5431c1a4a609771ee3ff5d209f\": container with ID starting with ccf2d9cd245abc05fb38ededfbfbece8d493be5431c1a4a609771ee3ff5d209f not found: ID does not exist" Oct 05 08:48:27 crc kubenswrapper[4846]: I1005 08:48:27.262557 4846 scope.go:117] "RemoveContainer" containerID="f68445c8d288d6c301ed5e3c0ec1734846e10d5a378566118fb9c43f313c329c" Oct 05 08:48:27 crc kubenswrapper[4846]: E1005 08:48:27.263033 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f68445c8d288d6c301ed5e3c0ec1734846e10d5a378566118fb9c43f313c329c\": container with ID starting with f68445c8d288d6c301ed5e3c0ec1734846e10d5a378566118fb9c43f313c329c not found: ID does not exist" containerID="f68445c8d288d6c301ed5e3c0ec1734846e10d5a378566118fb9c43f313c329c" Oct 05 08:48:27 crc kubenswrapper[4846]: I1005 08:48:27.263057 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f68445c8d288d6c301ed5e3c0ec1734846e10d5a378566118fb9c43f313c329c"} err="failed to get container status \"f68445c8d288d6c301ed5e3c0ec1734846e10d5a378566118fb9c43f313c329c\": rpc error: code = NotFound desc = could not find container \"f68445c8d288d6c301ed5e3c0ec1734846e10d5a378566118fb9c43f313c329c\": container with ID starting with f68445c8d288d6c301ed5e3c0ec1734846e10d5a378566118fb9c43f313c329c not found: ID does not exist" Oct 05 08:48:27 crc kubenswrapper[4846]: I1005 08:48:27.263076 4846 scope.go:117] "RemoveContainer" containerID="787d1f6b62a4aea6ec4b1f13d3f89f5917cfd84353636adbaef74d0672bc19f9" Oct 05 08:48:27 crc kubenswrapper[4846]: E1005 08:48:27.263354 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"787d1f6b62a4aea6ec4b1f13d3f89f5917cfd84353636adbaef74d0672bc19f9\": container with ID starting with 787d1f6b62a4aea6ec4b1f13d3f89f5917cfd84353636adbaef74d0672bc19f9 not found: ID does not exist" containerID="787d1f6b62a4aea6ec4b1f13d3f89f5917cfd84353636adbaef74d0672bc19f9" Oct 05 08:48:27 crc kubenswrapper[4846]: I1005 08:48:27.263386 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"787d1f6b62a4aea6ec4b1f13d3f89f5917cfd84353636adbaef74d0672bc19f9"} err="failed to get container status \"787d1f6b62a4aea6ec4b1f13d3f89f5917cfd84353636adbaef74d0672bc19f9\": rpc error: code = NotFound desc = could not find container \"787d1f6b62a4aea6ec4b1f13d3f89f5917cfd84353636adbaef74d0672bc19f9\": container with ID starting with 787d1f6b62a4aea6ec4b1f13d3f89f5917cfd84353636adbaef74d0672bc19f9 not found: ID does not exist" Oct 05 08:48:28 crc kubenswrapper[4846]: I1005 08:48:28.513317 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0c0156d2-8b61-4267-91a6-99d6e4197459" path="/var/lib/kubelet/pods/0c0156d2-8b61-4267-91a6-99d6e4197459/volumes" Oct 05 08:48:31 crc kubenswrapper[4846]: I1005 08:48:31.498871 4846 scope.go:117] "RemoveContainer" containerID="0d84b906773347b38ec2ce550f580254430238b4374f3171cb7453fa65259bcc" Oct 05 08:48:31 crc kubenswrapper[4846]: E1005 08:48:31.499797 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 08:48:45 crc kubenswrapper[4846]: I1005 08:48:45.498497 4846 scope.go:117] "RemoveContainer" containerID="0d84b906773347b38ec2ce550f580254430238b4374f3171cb7453fa65259bcc" Oct 05 08:48:45 crc kubenswrapper[4846]: E1005 08:48:45.499562 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 08:48:56 crc kubenswrapper[4846]: I1005 08:48:56.498467 4846 scope.go:117] "RemoveContainer" containerID="0d84b906773347b38ec2ce550f580254430238b4374f3171cb7453fa65259bcc" Oct 05 08:48:56 crc kubenswrapper[4846]: E1005 08:48:56.499850 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 08:49:06 crc kubenswrapper[4846]: I1005 08:49:06.539635 4846 generic.go:334] "Generic (PLEG): container finished" podID="58a90b64-2441-4ccf-8655-a5a582dee3cc" containerID="c3f5ca1c52c00c02111452e4374da5c16c54203abe7e9eaca1618ed2cec9b42c" exitCode=0 Oct 05 08:49:06 crc kubenswrapper[4846]: I1005 08:49:06.539692 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-openstack-openstack-cell1-l9jt5" event={"ID":"58a90b64-2441-4ccf-8655-a5a582dee3cc","Type":"ContainerDied","Data":"c3f5ca1c52c00c02111452e4374da5c16c54203abe7e9eaca1618ed2cec9b42c"} Oct 05 08:49:08 crc kubenswrapper[4846]: I1005 08:49:08.050570 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-openstack-openstack-cell1-l9jt5" Oct 05 08:49:08 crc kubenswrapper[4846]: I1005 08:49:08.181838 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/58a90b64-2441-4ccf-8655-a5a582dee3cc-inventory\") pod \"58a90b64-2441-4ccf-8655-a5a582dee3cc\" (UID: \"58a90b64-2441-4ccf-8655-a5a582dee3cc\") " Oct 05 08:49:08 crc kubenswrapper[4846]: I1005 08:49:08.181916 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/58a90b64-2441-4ccf-8655-a5a582dee3cc-ssh-key\") pod \"58a90b64-2441-4ccf-8655-a5a582dee3cc\" (UID: \"58a90b64-2441-4ccf-8655-a5a582dee3cc\") " Oct 05 08:49:08 crc kubenswrapper[4846]: I1005 08:49:08.182163 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-622w8\" (UniqueName: \"kubernetes.io/projected/58a90b64-2441-4ccf-8655-a5a582dee3cc-kube-api-access-622w8\") pod \"58a90b64-2441-4ccf-8655-a5a582dee3cc\" (UID: \"58a90b64-2441-4ccf-8655-a5a582dee3cc\") " Oct 05 08:49:08 crc kubenswrapper[4846]: I1005 08:49:08.188387 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/58a90b64-2441-4ccf-8655-a5a582dee3cc-kube-api-access-622w8" (OuterVolumeSpecName: "kube-api-access-622w8") pod "58a90b64-2441-4ccf-8655-a5a582dee3cc" (UID: "58a90b64-2441-4ccf-8655-a5a582dee3cc"). InnerVolumeSpecName "kube-api-access-622w8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:49:08 crc kubenswrapper[4846]: I1005 08:49:08.228462 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/58a90b64-2441-4ccf-8655-a5a582dee3cc-inventory" (OuterVolumeSpecName: "inventory") pod "58a90b64-2441-4ccf-8655-a5a582dee3cc" (UID: "58a90b64-2441-4ccf-8655-a5a582dee3cc"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:49:08 crc kubenswrapper[4846]: I1005 08:49:08.247308 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/58a90b64-2441-4ccf-8655-a5a582dee3cc-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "58a90b64-2441-4ccf-8655-a5a582dee3cc" (UID: "58a90b64-2441-4ccf-8655-a5a582dee3cc"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:49:08 crc kubenswrapper[4846]: I1005 08:49:08.284548 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-622w8\" (UniqueName: \"kubernetes.io/projected/58a90b64-2441-4ccf-8655-a5a582dee3cc-kube-api-access-622w8\") on node \"crc\" DevicePath \"\"" Oct 05 08:49:08 crc kubenswrapper[4846]: I1005 08:49:08.284581 4846 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/58a90b64-2441-4ccf-8655-a5a582dee3cc-inventory\") on node \"crc\" DevicePath \"\"" Oct 05 08:49:08 crc kubenswrapper[4846]: I1005 08:49:08.284594 4846 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/58a90b64-2441-4ccf-8655-a5a582dee3cc-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 05 08:49:08 crc kubenswrapper[4846]: I1005 08:49:08.568594 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-openstack-openstack-cell1-l9jt5" event={"ID":"58a90b64-2441-4ccf-8655-a5a582dee3cc","Type":"ContainerDied","Data":"5fba894e1a8a66c6f14fab39fd863bbf1b5173306fd3b63e12c7a2942d262b39"} Oct 05 08:49:08 crc kubenswrapper[4846]: I1005 08:49:08.568657 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5fba894e1a8a66c6f14fab39fd863bbf1b5173306fd3b63e12c7a2942d262b39" Oct 05 08:49:08 crc kubenswrapper[4846]: I1005 08:49:08.568738 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-openstack-openstack-cell1-l9jt5" Oct 05 08:49:08 crc kubenswrapper[4846]: I1005 08:49:08.668363 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/validate-network-openstack-openstack-cell1-fkcjx"] Oct 05 08:49:08 crc kubenswrapper[4846]: E1005 08:49:08.669081 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="58a90b64-2441-4ccf-8655-a5a582dee3cc" containerName="configure-network-openstack-openstack-cell1" Oct 05 08:49:08 crc kubenswrapper[4846]: I1005 08:49:08.669196 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="58a90b64-2441-4ccf-8655-a5a582dee3cc" containerName="configure-network-openstack-openstack-cell1" Oct 05 08:49:08 crc kubenswrapper[4846]: E1005 08:49:08.669293 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0c0156d2-8b61-4267-91a6-99d6e4197459" containerName="registry-server" Oct 05 08:49:08 crc kubenswrapper[4846]: I1005 08:49:08.669373 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="0c0156d2-8b61-4267-91a6-99d6e4197459" containerName="registry-server" Oct 05 08:49:08 crc kubenswrapper[4846]: E1005 08:49:08.669475 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="37eb33ef-2623-4413-b67f-2b6b3af85b1e" containerName="extract-utilities" Oct 05 08:49:08 crc kubenswrapper[4846]: I1005 08:49:08.669557 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="37eb33ef-2623-4413-b67f-2b6b3af85b1e" containerName="extract-utilities" Oct 05 08:49:08 crc kubenswrapper[4846]: E1005 08:49:08.669654 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="37eb33ef-2623-4413-b67f-2b6b3af85b1e" containerName="registry-server" Oct 05 08:49:08 crc kubenswrapper[4846]: I1005 08:49:08.669731 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="37eb33ef-2623-4413-b67f-2b6b3af85b1e" containerName="registry-server" Oct 05 08:49:08 crc kubenswrapper[4846]: E1005 08:49:08.669819 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0c0156d2-8b61-4267-91a6-99d6e4197459" containerName="extract-utilities" Oct 05 08:49:08 crc kubenswrapper[4846]: I1005 08:49:08.669897 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="0c0156d2-8b61-4267-91a6-99d6e4197459" containerName="extract-utilities" Oct 05 08:49:08 crc kubenswrapper[4846]: E1005 08:49:08.669983 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="37eb33ef-2623-4413-b67f-2b6b3af85b1e" containerName="extract-content" Oct 05 08:49:08 crc kubenswrapper[4846]: I1005 08:49:08.670048 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="37eb33ef-2623-4413-b67f-2b6b3af85b1e" containerName="extract-content" Oct 05 08:49:08 crc kubenswrapper[4846]: E1005 08:49:08.670127 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0c0156d2-8b61-4267-91a6-99d6e4197459" containerName="extract-content" Oct 05 08:49:08 crc kubenswrapper[4846]: I1005 08:49:08.670220 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="0c0156d2-8b61-4267-91a6-99d6e4197459" containerName="extract-content" Oct 05 08:49:08 crc kubenswrapper[4846]: I1005 08:49:08.670562 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="0c0156d2-8b61-4267-91a6-99d6e4197459" containerName="registry-server" Oct 05 08:49:08 crc kubenswrapper[4846]: I1005 08:49:08.670667 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="37eb33ef-2623-4413-b67f-2b6b3af85b1e" containerName="registry-server" Oct 05 08:49:08 crc kubenswrapper[4846]: I1005 08:49:08.670768 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="58a90b64-2441-4ccf-8655-a5a582dee3cc" containerName="configure-network-openstack-openstack-cell1" Oct 05 08:49:08 crc kubenswrapper[4846]: I1005 08:49:08.672216 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-openstack-openstack-cell1-fkcjx" Oct 05 08:49:08 crc kubenswrapper[4846]: I1005 08:49:08.678667 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 05 08:49:08 crc kubenswrapper[4846]: I1005 08:49:08.679351 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Oct 05 08:49:08 crc kubenswrapper[4846]: I1005 08:49:08.679700 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-openstack-openstack-cell1-fkcjx"] Oct 05 08:49:08 crc kubenswrapper[4846]: I1005 08:49:08.682383 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-7687s" Oct 05 08:49:08 crc kubenswrapper[4846]: I1005 08:49:08.688601 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Oct 05 08:49:08 crc kubenswrapper[4846]: I1005 08:49:08.795724 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-68z2c\" (UniqueName: \"kubernetes.io/projected/c7fe64f3-1811-49fb-bc47-c9a5a1060954-kube-api-access-68z2c\") pod \"validate-network-openstack-openstack-cell1-fkcjx\" (UID: \"c7fe64f3-1811-49fb-bc47-c9a5a1060954\") " pod="openstack/validate-network-openstack-openstack-cell1-fkcjx" Oct 05 08:49:08 crc kubenswrapper[4846]: I1005 08:49:08.796153 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c7fe64f3-1811-49fb-bc47-c9a5a1060954-inventory\") pod \"validate-network-openstack-openstack-cell1-fkcjx\" (UID: \"c7fe64f3-1811-49fb-bc47-c9a5a1060954\") " pod="openstack/validate-network-openstack-openstack-cell1-fkcjx" Oct 05 08:49:08 crc kubenswrapper[4846]: I1005 08:49:08.796242 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c7fe64f3-1811-49fb-bc47-c9a5a1060954-ssh-key\") pod \"validate-network-openstack-openstack-cell1-fkcjx\" (UID: \"c7fe64f3-1811-49fb-bc47-c9a5a1060954\") " pod="openstack/validate-network-openstack-openstack-cell1-fkcjx" Oct 05 08:49:08 crc kubenswrapper[4846]: I1005 08:49:08.899593 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c7fe64f3-1811-49fb-bc47-c9a5a1060954-inventory\") pod \"validate-network-openstack-openstack-cell1-fkcjx\" (UID: \"c7fe64f3-1811-49fb-bc47-c9a5a1060954\") " pod="openstack/validate-network-openstack-openstack-cell1-fkcjx" Oct 05 08:49:08 crc kubenswrapper[4846]: I1005 08:49:08.899709 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c7fe64f3-1811-49fb-bc47-c9a5a1060954-ssh-key\") pod \"validate-network-openstack-openstack-cell1-fkcjx\" (UID: \"c7fe64f3-1811-49fb-bc47-c9a5a1060954\") " pod="openstack/validate-network-openstack-openstack-cell1-fkcjx" Oct 05 08:49:08 crc kubenswrapper[4846]: I1005 08:49:08.899965 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-68z2c\" (UniqueName: \"kubernetes.io/projected/c7fe64f3-1811-49fb-bc47-c9a5a1060954-kube-api-access-68z2c\") pod \"validate-network-openstack-openstack-cell1-fkcjx\" (UID: \"c7fe64f3-1811-49fb-bc47-c9a5a1060954\") " pod="openstack/validate-network-openstack-openstack-cell1-fkcjx" Oct 05 08:49:08 crc kubenswrapper[4846]: I1005 08:49:08.911770 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c7fe64f3-1811-49fb-bc47-c9a5a1060954-inventory\") pod \"validate-network-openstack-openstack-cell1-fkcjx\" (UID: \"c7fe64f3-1811-49fb-bc47-c9a5a1060954\") " pod="openstack/validate-network-openstack-openstack-cell1-fkcjx" Oct 05 08:49:08 crc kubenswrapper[4846]: I1005 08:49:08.912680 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c7fe64f3-1811-49fb-bc47-c9a5a1060954-ssh-key\") pod \"validate-network-openstack-openstack-cell1-fkcjx\" (UID: \"c7fe64f3-1811-49fb-bc47-c9a5a1060954\") " pod="openstack/validate-network-openstack-openstack-cell1-fkcjx" Oct 05 08:49:08 crc kubenswrapper[4846]: I1005 08:49:08.934348 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-68z2c\" (UniqueName: \"kubernetes.io/projected/c7fe64f3-1811-49fb-bc47-c9a5a1060954-kube-api-access-68z2c\") pod \"validate-network-openstack-openstack-cell1-fkcjx\" (UID: \"c7fe64f3-1811-49fb-bc47-c9a5a1060954\") " pod="openstack/validate-network-openstack-openstack-cell1-fkcjx" Oct 05 08:49:08 crc kubenswrapper[4846]: I1005 08:49:08.992476 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-openstack-openstack-cell1-fkcjx" Oct 05 08:49:09 crc kubenswrapper[4846]: I1005 08:49:09.674517 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-openstack-openstack-cell1-fkcjx"] Oct 05 08:49:10 crc kubenswrapper[4846]: I1005 08:49:10.600192 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-openstack-openstack-cell1-fkcjx" event={"ID":"c7fe64f3-1811-49fb-bc47-c9a5a1060954","Type":"ContainerStarted","Data":"53fdb1368ea62a890f914d5c00d22f89db98b00def7ad53e22d229c20018333c"} Oct 05 08:49:10 crc kubenswrapper[4846]: I1005 08:49:10.600652 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-openstack-openstack-cell1-fkcjx" event={"ID":"c7fe64f3-1811-49fb-bc47-c9a5a1060954","Type":"ContainerStarted","Data":"3c3a431b56faa87e8cc8b1ffea7a2edd7e5de6ae2cc0b81812cf2ffe84ae1d8d"} Oct 05 08:49:10 crc kubenswrapper[4846]: I1005 08:49:10.624989 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/validate-network-openstack-openstack-cell1-fkcjx" podStartSLOduration=2.125712538 podStartE2EDuration="2.624952773s" podCreationTimestamp="2025-10-05 08:49:08 +0000 UTC" firstStartedPulling="2025-10-05 08:49:09.682875816 +0000 UTC m=+7271.923728591" lastFinishedPulling="2025-10-05 08:49:10.182116041 +0000 UTC m=+7272.422968826" observedRunningTime="2025-10-05 08:49:10.614457604 +0000 UTC m=+7272.855310369" watchObservedRunningTime="2025-10-05 08:49:10.624952773 +0000 UTC m=+7272.865805588" Oct 05 08:49:11 crc kubenswrapper[4846]: I1005 08:49:11.497596 4846 scope.go:117] "RemoveContainer" containerID="0d84b906773347b38ec2ce550f580254430238b4374f3171cb7453fa65259bcc" Oct 05 08:49:11 crc kubenswrapper[4846]: E1005 08:49:11.498134 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 08:49:15 crc kubenswrapper[4846]: I1005 08:49:15.651597 4846 generic.go:334] "Generic (PLEG): container finished" podID="c7fe64f3-1811-49fb-bc47-c9a5a1060954" containerID="53fdb1368ea62a890f914d5c00d22f89db98b00def7ad53e22d229c20018333c" exitCode=0 Oct 05 08:49:15 crc kubenswrapper[4846]: I1005 08:49:15.651705 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-openstack-openstack-cell1-fkcjx" event={"ID":"c7fe64f3-1811-49fb-bc47-c9a5a1060954","Type":"ContainerDied","Data":"53fdb1368ea62a890f914d5c00d22f89db98b00def7ad53e22d229c20018333c"} Oct 05 08:49:17 crc kubenswrapper[4846]: I1005 08:49:17.247441 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-openstack-openstack-cell1-fkcjx" Oct 05 08:49:17 crc kubenswrapper[4846]: I1005 08:49:17.388436 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-68z2c\" (UniqueName: \"kubernetes.io/projected/c7fe64f3-1811-49fb-bc47-c9a5a1060954-kube-api-access-68z2c\") pod \"c7fe64f3-1811-49fb-bc47-c9a5a1060954\" (UID: \"c7fe64f3-1811-49fb-bc47-c9a5a1060954\") " Oct 05 08:49:17 crc kubenswrapper[4846]: I1005 08:49:17.388670 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c7fe64f3-1811-49fb-bc47-c9a5a1060954-ssh-key\") pod \"c7fe64f3-1811-49fb-bc47-c9a5a1060954\" (UID: \"c7fe64f3-1811-49fb-bc47-c9a5a1060954\") " Oct 05 08:49:17 crc kubenswrapper[4846]: I1005 08:49:17.388764 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c7fe64f3-1811-49fb-bc47-c9a5a1060954-inventory\") pod \"c7fe64f3-1811-49fb-bc47-c9a5a1060954\" (UID: \"c7fe64f3-1811-49fb-bc47-c9a5a1060954\") " Oct 05 08:49:17 crc kubenswrapper[4846]: I1005 08:49:17.394432 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c7fe64f3-1811-49fb-bc47-c9a5a1060954-kube-api-access-68z2c" (OuterVolumeSpecName: "kube-api-access-68z2c") pod "c7fe64f3-1811-49fb-bc47-c9a5a1060954" (UID: "c7fe64f3-1811-49fb-bc47-c9a5a1060954"). InnerVolumeSpecName "kube-api-access-68z2c". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:49:17 crc kubenswrapper[4846]: I1005 08:49:17.414335 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c7fe64f3-1811-49fb-bc47-c9a5a1060954-inventory" (OuterVolumeSpecName: "inventory") pod "c7fe64f3-1811-49fb-bc47-c9a5a1060954" (UID: "c7fe64f3-1811-49fb-bc47-c9a5a1060954"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:49:17 crc kubenswrapper[4846]: I1005 08:49:17.428144 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c7fe64f3-1811-49fb-bc47-c9a5a1060954-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "c7fe64f3-1811-49fb-bc47-c9a5a1060954" (UID: "c7fe64f3-1811-49fb-bc47-c9a5a1060954"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:49:17 crc kubenswrapper[4846]: I1005 08:49:17.491538 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-68z2c\" (UniqueName: \"kubernetes.io/projected/c7fe64f3-1811-49fb-bc47-c9a5a1060954-kube-api-access-68z2c\") on node \"crc\" DevicePath \"\"" Oct 05 08:49:17 crc kubenswrapper[4846]: I1005 08:49:17.491581 4846 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c7fe64f3-1811-49fb-bc47-c9a5a1060954-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 05 08:49:17 crc kubenswrapper[4846]: I1005 08:49:17.491594 4846 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c7fe64f3-1811-49fb-bc47-c9a5a1060954-inventory\") on node \"crc\" DevicePath \"\"" Oct 05 08:49:17 crc kubenswrapper[4846]: I1005 08:49:17.678006 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-openstack-openstack-cell1-fkcjx" event={"ID":"c7fe64f3-1811-49fb-bc47-c9a5a1060954","Type":"ContainerDied","Data":"3c3a431b56faa87e8cc8b1ffea7a2edd7e5de6ae2cc0b81812cf2ffe84ae1d8d"} Oct 05 08:49:17 crc kubenswrapper[4846]: I1005 08:49:17.678045 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3c3a431b56faa87e8cc8b1ffea7a2edd7e5de6ae2cc0b81812cf2ffe84ae1d8d" Oct 05 08:49:17 crc kubenswrapper[4846]: I1005 08:49:17.678078 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-openstack-openstack-cell1-fkcjx" Oct 05 08:49:17 crc kubenswrapper[4846]: I1005 08:49:17.774383 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-os-openstack-openstack-cell1-l4pz6"] Oct 05 08:49:17 crc kubenswrapper[4846]: E1005 08:49:17.774841 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c7fe64f3-1811-49fb-bc47-c9a5a1060954" containerName="validate-network-openstack-openstack-cell1" Oct 05 08:49:17 crc kubenswrapper[4846]: I1005 08:49:17.774867 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="c7fe64f3-1811-49fb-bc47-c9a5a1060954" containerName="validate-network-openstack-openstack-cell1" Oct 05 08:49:17 crc kubenswrapper[4846]: I1005 08:49:17.775088 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="c7fe64f3-1811-49fb-bc47-c9a5a1060954" containerName="validate-network-openstack-openstack-cell1" Oct 05 08:49:17 crc kubenswrapper[4846]: I1005 08:49:17.775783 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-openstack-openstack-cell1-l4pz6" Oct 05 08:49:17 crc kubenswrapper[4846]: I1005 08:49:17.777894 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Oct 05 08:49:17 crc kubenswrapper[4846]: I1005 08:49:17.778338 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 05 08:49:17 crc kubenswrapper[4846]: I1005 08:49:17.778409 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-7687s" Oct 05 08:49:17 crc kubenswrapper[4846]: I1005 08:49:17.783389 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Oct 05 08:49:17 crc kubenswrapper[4846]: I1005 08:49:17.802029 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-openstack-openstack-cell1-l4pz6"] Oct 05 08:49:17 crc kubenswrapper[4846]: I1005 08:49:17.903945 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bpjs5\" (UniqueName: \"kubernetes.io/projected/b4fc0719-2125-4505-bd67-3052dcfd3f1a-kube-api-access-bpjs5\") pod \"install-os-openstack-openstack-cell1-l4pz6\" (UID: \"b4fc0719-2125-4505-bd67-3052dcfd3f1a\") " pod="openstack/install-os-openstack-openstack-cell1-l4pz6" Oct 05 08:49:17 crc kubenswrapper[4846]: I1005 08:49:17.904040 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b4fc0719-2125-4505-bd67-3052dcfd3f1a-inventory\") pod \"install-os-openstack-openstack-cell1-l4pz6\" (UID: \"b4fc0719-2125-4505-bd67-3052dcfd3f1a\") " pod="openstack/install-os-openstack-openstack-cell1-l4pz6" Oct 05 08:49:17 crc kubenswrapper[4846]: I1005 08:49:17.904292 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b4fc0719-2125-4505-bd67-3052dcfd3f1a-ssh-key\") pod \"install-os-openstack-openstack-cell1-l4pz6\" (UID: \"b4fc0719-2125-4505-bd67-3052dcfd3f1a\") " pod="openstack/install-os-openstack-openstack-cell1-l4pz6" Oct 05 08:49:18 crc kubenswrapper[4846]: I1005 08:49:18.006865 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b4fc0719-2125-4505-bd67-3052dcfd3f1a-inventory\") pod \"install-os-openstack-openstack-cell1-l4pz6\" (UID: \"b4fc0719-2125-4505-bd67-3052dcfd3f1a\") " pod="openstack/install-os-openstack-openstack-cell1-l4pz6" Oct 05 08:49:18 crc kubenswrapper[4846]: I1005 08:49:18.006999 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b4fc0719-2125-4505-bd67-3052dcfd3f1a-ssh-key\") pod \"install-os-openstack-openstack-cell1-l4pz6\" (UID: \"b4fc0719-2125-4505-bd67-3052dcfd3f1a\") " pod="openstack/install-os-openstack-openstack-cell1-l4pz6" Oct 05 08:49:18 crc kubenswrapper[4846]: I1005 08:49:18.007132 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bpjs5\" (UniqueName: \"kubernetes.io/projected/b4fc0719-2125-4505-bd67-3052dcfd3f1a-kube-api-access-bpjs5\") pod \"install-os-openstack-openstack-cell1-l4pz6\" (UID: \"b4fc0719-2125-4505-bd67-3052dcfd3f1a\") " pod="openstack/install-os-openstack-openstack-cell1-l4pz6" Oct 05 08:49:18 crc kubenswrapper[4846]: I1005 08:49:18.011128 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b4fc0719-2125-4505-bd67-3052dcfd3f1a-ssh-key\") pod \"install-os-openstack-openstack-cell1-l4pz6\" (UID: \"b4fc0719-2125-4505-bd67-3052dcfd3f1a\") " pod="openstack/install-os-openstack-openstack-cell1-l4pz6" Oct 05 08:49:18 crc kubenswrapper[4846]: I1005 08:49:18.011381 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b4fc0719-2125-4505-bd67-3052dcfd3f1a-inventory\") pod \"install-os-openstack-openstack-cell1-l4pz6\" (UID: \"b4fc0719-2125-4505-bd67-3052dcfd3f1a\") " pod="openstack/install-os-openstack-openstack-cell1-l4pz6" Oct 05 08:49:18 crc kubenswrapper[4846]: I1005 08:49:18.024929 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bpjs5\" (UniqueName: \"kubernetes.io/projected/b4fc0719-2125-4505-bd67-3052dcfd3f1a-kube-api-access-bpjs5\") pod \"install-os-openstack-openstack-cell1-l4pz6\" (UID: \"b4fc0719-2125-4505-bd67-3052dcfd3f1a\") " pod="openstack/install-os-openstack-openstack-cell1-l4pz6" Oct 05 08:49:18 crc kubenswrapper[4846]: I1005 08:49:18.101862 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-openstack-openstack-cell1-l4pz6" Oct 05 08:49:18 crc kubenswrapper[4846]: I1005 08:49:18.696765 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-openstack-openstack-cell1-l4pz6"] Oct 05 08:49:19 crc kubenswrapper[4846]: I1005 08:49:19.727775 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-openstack-openstack-cell1-l4pz6" event={"ID":"b4fc0719-2125-4505-bd67-3052dcfd3f1a","Type":"ContainerStarted","Data":"e29e4aef6d5cfacc42d2cb47849c014096d6b7daf6a5d999c3f0a659f85b1676"} Oct 05 08:49:19 crc kubenswrapper[4846]: I1005 08:49:19.728300 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-openstack-openstack-cell1-l4pz6" event={"ID":"b4fc0719-2125-4505-bd67-3052dcfd3f1a","Type":"ContainerStarted","Data":"70886c25b857395b6d09e0cad38a622f4fd2374f73bd928090f5d1e9d0578c9c"} Oct 05 08:49:19 crc kubenswrapper[4846]: I1005 08:49:19.763046 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-os-openstack-openstack-cell1-l4pz6" podStartSLOduration=2.373282172 podStartE2EDuration="2.76302004s" podCreationTimestamp="2025-10-05 08:49:17 +0000 UTC" firstStartedPulling="2025-10-05 08:49:18.701950775 +0000 UTC m=+7280.942803550" lastFinishedPulling="2025-10-05 08:49:19.091688643 +0000 UTC m=+7281.332541418" observedRunningTime="2025-10-05 08:49:19.749791317 +0000 UTC m=+7281.990644112" watchObservedRunningTime="2025-10-05 08:49:19.76302004 +0000 UTC m=+7282.003872855" Oct 05 08:49:26 crc kubenswrapper[4846]: I1005 08:49:26.499015 4846 scope.go:117] "RemoveContainer" containerID="0d84b906773347b38ec2ce550f580254430238b4374f3171cb7453fa65259bcc" Oct 05 08:49:26 crc kubenswrapper[4846]: E1005 08:49:26.499660 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 08:49:39 crc kubenswrapper[4846]: I1005 08:49:39.498276 4846 scope.go:117] "RemoveContainer" containerID="0d84b906773347b38ec2ce550f580254430238b4374f3171cb7453fa65259bcc" Oct 05 08:49:39 crc kubenswrapper[4846]: E1005 08:49:39.499049 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 08:49:54 crc kubenswrapper[4846]: I1005 08:49:54.498347 4846 scope.go:117] "RemoveContainer" containerID="0d84b906773347b38ec2ce550f580254430238b4374f3171cb7453fa65259bcc" Oct 05 08:49:55 crc kubenswrapper[4846]: I1005 08:49:55.118534 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" event={"ID":"ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7","Type":"ContainerStarted","Data":"eb023100181ea6025e1663017d08951f92365161214cdfbcbfc8a62d71858fc9"} Oct 05 08:50:03 crc kubenswrapper[4846]: I1005 08:50:03.249982 4846 generic.go:334] "Generic (PLEG): container finished" podID="b4fc0719-2125-4505-bd67-3052dcfd3f1a" containerID="e29e4aef6d5cfacc42d2cb47849c014096d6b7daf6a5d999c3f0a659f85b1676" exitCode=0 Oct 05 08:50:03 crc kubenswrapper[4846]: I1005 08:50:03.250574 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-openstack-openstack-cell1-l4pz6" event={"ID":"b4fc0719-2125-4505-bd67-3052dcfd3f1a","Type":"ContainerDied","Data":"e29e4aef6d5cfacc42d2cb47849c014096d6b7daf6a5d999c3f0a659f85b1676"} Oct 05 08:50:04 crc kubenswrapper[4846]: I1005 08:50:04.813780 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-openstack-openstack-cell1-l4pz6" Oct 05 08:50:04 crc kubenswrapper[4846]: I1005 08:50:04.841081 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bpjs5\" (UniqueName: \"kubernetes.io/projected/b4fc0719-2125-4505-bd67-3052dcfd3f1a-kube-api-access-bpjs5\") pod \"b4fc0719-2125-4505-bd67-3052dcfd3f1a\" (UID: \"b4fc0719-2125-4505-bd67-3052dcfd3f1a\") " Oct 05 08:50:04 crc kubenswrapper[4846]: I1005 08:50:04.841278 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b4fc0719-2125-4505-bd67-3052dcfd3f1a-inventory\") pod \"b4fc0719-2125-4505-bd67-3052dcfd3f1a\" (UID: \"b4fc0719-2125-4505-bd67-3052dcfd3f1a\") " Oct 05 08:50:04 crc kubenswrapper[4846]: I1005 08:50:04.841725 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b4fc0719-2125-4505-bd67-3052dcfd3f1a-ssh-key\") pod \"b4fc0719-2125-4505-bd67-3052dcfd3f1a\" (UID: \"b4fc0719-2125-4505-bd67-3052dcfd3f1a\") " Oct 05 08:50:04 crc kubenswrapper[4846]: I1005 08:50:04.860896 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b4fc0719-2125-4505-bd67-3052dcfd3f1a-kube-api-access-bpjs5" (OuterVolumeSpecName: "kube-api-access-bpjs5") pod "b4fc0719-2125-4505-bd67-3052dcfd3f1a" (UID: "b4fc0719-2125-4505-bd67-3052dcfd3f1a"). InnerVolumeSpecName "kube-api-access-bpjs5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:50:04 crc kubenswrapper[4846]: I1005 08:50:04.894876 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b4fc0719-2125-4505-bd67-3052dcfd3f1a-inventory" (OuterVolumeSpecName: "inventory") pod "b4fc0719-2125-4505-bd67-3052dcfd3f1a" (UID: "b4fc0719-2125-4505-bd67-3052dcfd3f1a"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:50:04 crc kubenswrapper[4846]: I1005 08:50:04.909278 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b4fc0719-2125-4505-bd67-3052dcfd3f1a-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "b4fc0719-2125-4505-bd67-3052dcfd3f1a" (UID: "b4fc0719-2125-4505-bd67-3052dcfd3f1a"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:50:04 crc kubenswrapper[4846]: I1005 08:50:04.945972 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bpjs5\" (UniqueName: \"kubernetes.io/projected/b4fc0719-2125-4505-bd67-3052dcfd3f1a-kube-api-access-bpjs5\") on node \"crc\" DevicePath \"\"" Oct 05 08:50:04 crc kubenswrapper[4846]: I1005 08:50:04.946012 4846 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b4fc0719-2125-4505-bd67-3052dcfd3f1a-inventory\") on node \"crc\" DevicePath \"\"" Oct 05 08:50:04 crc kubenswrapper[4846]: I1005 08:50:04.946024 4846 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b4fc0719-2125-4505-bd67-3052dcfd3f1a-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 05 08:50:05 crc kubenswrapper[4846]: I1005 08:50:05.274511 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-openstack-openstack-cell1-l4pz6" event={"ID":"b4fc0719-2125-4505-bd67-3052dcfd3f1a","Type":"ContainerDied","Data":"70886c25b857395b6d09e0cad38a622f4fd2374f73bd928090f5d1e9d0578c9c"} Oct 05 08:50:05 crc kubenswrapper[4846]: I1005 08:50:05.274894 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="70886c25b857395b6d09e0cad38a622f4fd2374f73bd928090f5d1e9d0578c9c" Oct 05 08:50:05 crc kubenswrapper[4846]: I1005 08:50:05.274577 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-openstack-openstack-cell1-l4pz6" Oct 05 08:50:05 crc kubenswrapper[4846]: I1005 08:50:05.364832 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-openstack-openstack-cell1-wzg2n"] Oct 05 08:50:05 crc kubenswrapper[4846]: E1005 08:50:05.365277 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b4fc0719-2125-4505-bd67-3052dcfd3f1a" containerName="install-os-openstack-openstack-cell1" Oct 05 08:50:05 crc kubenswrapper[4846]: I1005 08:50:05.365295 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="b4fc0719-2125-4505-bd67-3052dcfd3f1a" containerName="install-os-openstack-openstack-cell1" Oct 05 08:50:05 crc kubenswrapper[4846]: I1005 08:50:05.365469 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="b4fc0719-2125-4505-bd67-3052dcfd3f1a" containerName="install-os-openstack-openstack-cell1" Oct 05 08:50:05 crc kubenswrapper[4846]: I1005 08:50:05.366167 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-openstack-openstack-cell1-wzg2n" Oct 05 08:50:05 crc kubenswrapper[4846]: I1005 08:50:05.375008 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Oct 05 08:50:05 crc kubenswrapper[4846]: I1005 08:50:05.379547 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Oct 05 08:50:05 crc kubenswrapper[4846]: I1005 08:50:05.379546 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 05 08:50:05 crc kubenswrapper[4846]: I1005 08:50:05.379707 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-7687s" Oct 05 08:50:05 crc kubenswrapper[4846]: I1005 08:50:05.414575 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-openstack-openstack-cell1-wzg2n"] Oct 05 08:50:05 crc kubenswrapper[4846]: I1005 08:50:05.462741 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c9d6416a-ce5b-41a2-b49b-a67573544f06-inventory\") pod \"configure-os-openstack-openstack-cell1-wzg2n\" (UID: \"c9d6416a-ce5b-41a2-b49b-a67573544f06\") " pod="openstack/configure-os-openstack-openstack-cell1-wzg2n" Oct 05 08:50:05 crc kubenswrapper[4846]: I1005 08:50:05.462858 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c9d6416a-ce5b-41a2-b49b-a67573544f06-ssh-key\") pod \"configure-os-openstack-openstack-cell1-wzg2n\" (UID: \"c9d6416a-ce5b-41a2-b49b-a67573544f06\") " pod="openstack/configure-os-openstack-openstack-cell1-wzg2n" Oct 05 08:50:05 crc kubenswrapper[4846]: I1005 08:50:05.462927 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lsfjx\" (UniqueName: \"kubernetes.io/projected/c9d6416a-ce5b-41a2-b49b-a67573544f06-kube-api-access-lsfjx\") pod \"configure-os-openstack-openstack-cell1-wzg2n\" (UID: \"c9d6416a-ce5b-41a2-b49b-a67573544f06\") " pod="openstack/configure-os-openstack-openstack-cell1-wzg2n" Oct 05 08:50:05 crc kubenswrapper[4846]: I1005 08:50:05.564629 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c9d6416a-ce5b-41a2-b49b-a67573544f06-inventory\") pod \"configure-os-openstack-openstack-cell1-wzg2n\" (UID: \"c9d6416a-ce5b-41a2-b49b-a67573544f06\") " pod="openstack/configure-os-openstack-openstack-cell1-wzg2n" Oct 05 08:50:05 crc kubenswrapper[4846]: I1005 08:50:05.564800 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c9d6416a-ce5b-41a2-b49b-a67573544f06-ssh-key\") pod \"configure-os-openstack-openstack-cell1-wzg2n\" (UID: \"c9d6416a-ce5b-41a2-b49b-a67573544f06\") " pod="openstack/configure-os-openstack-openstack-cell1-wzg2n" Oct 05 08:50:05 crc kubenswrapper[4846]: I1005 08:50:05.564893 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lsfjx\" (UniqueName: \"kubernetes.io/projected/c9d6416a-ce5b-41a2-b49b-a67573544f06-kube-api-access-lsfjx\") pod \"configure-os-openstack-openstack-cell1-wzg2n\" (UID: \"c9d6416a-ce5b-41a2-b49b-a67573544f06\") " pod="openstack/configure-os-openstack-openstack-cell1-wzg2n" Oct 05 08:50:05 crc kubenswrapper[4846]: I1005 08:50:05.568547 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c9d6416a-ce5b-41a2-b49b-a67573544f06-inventory\") pod \"configure-os-openstack-openstack-cell1-wzg2n\" (UID: \"c9d6416a-ce5b-41a2-b49b-a67573544f06\") " pod="openstack/configure-os-openstack-openstack-cell1-wzg2n" Oct 05 08:50:05 crc kubenswrapper[4846]: I1005 08:50:05.569964 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c9d6416a-ce5b-41a2-b49b-a67573544f06-ssh-key\") pod \"configure-os-openstack-openstack-cell1-wzg2n\" (UID: \"c9d6416a-ce5b-41a2-b49b-a67573544f06\") " pod="openstack/configure-os-openstack-openstack-cell1-wzg2n" Oct 05 08:50:05 crc kubenswrapper[4846]: I1005 08:50:05.581433 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lsfjx\" (UniqueName: \"kubernetes.io/projected/c9d6416a-ce5b-41a2-b49b-a67573544f06-kube-api-access-lsfjx\") pod \"configure-os-openstack-openstack-cell1-wzg2n\" (UID: \"c9d6416a-ce5b-41a2-b49b-a67573544f06\") " pod="openstack/configure-os-openstack-openstack-cell1-wzg2n" Oct 05 08:50:05 crc kubenswrapper[4846]: I1005 08:50:05.705848 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-openstack-openstack-cell1-wzg2n" Oct 05 08:50:06 crc kubenswrapper[4846]: I1005 08:50:06.258131 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-openstack-openstack-cell1-wzg2n"] Oct 05 08:50:06 crc kubenswrapper[4846]: I1005 08:50:06.287837 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-openstack-openstack-cell1-wzg2n" event={"ID":"c9d6416a-ce5b-41a2-b49b-a67573544f06","Type":"ContainerStarted","Data":"17b103b093b5999621abd23d79a376a9fdf866e78515c9c16b33a0088cf91230"} Oct 05 08:50:07 crc kubenswrapper[4846]: I1005 08:50:07.310287 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-openstack-openstack-cell1-wzg2n" event={"ID":"c9d6416a-ce5b-41a2-b49b-a67573544f06","Type":"ContainerStarted","Data":"93ff7c5f5d35043aa0f0ba1ab2896064e8ab809c527312e0df9ae6fe3e4b3876"} Oct 05 08:50:07 crc kubenswrapper[4846]: I1005 08:50:07.337713 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-openstack-openstack-cell1-wzg2n" podStartSLOduration=1.925204496 podStartE2EDuration="2.33768316s" podCreationTimestamp="2025-10-05 08:50:05 +0000 UTC" firstStartedPulling="2025-10-05 08:50:06.265041776 +0000 UTC m=+7328.505894551" lastFinishedPulling="2025-10-05 08:50:06.67752043 +0000 UTC m=+7328.918373215" observedRunningTime="2025-10-05 08:50:07.33203807 +0000 UTC m=+7329.572890885" watchObservedRunningTime="2025-10-05 08:50:07.33768316 +0000 UTC m=+7329.578535975" Oct 05 08:50:53 crc kubenswrapper[4846]: I1005 08:50:53.854963 4846 generic.go:334] "Generic (PLEG): container finished" podID="c9d6416a-ce5b-41a2-b49b-a67573544f06" containerID="93ff7c5f5d35043aa0f0ba1ab2896064e8ab809c527312e0df9ae6fe3e4b3876" exitCode=0 Oct 05 08:50:53 crc kubenswrapper[4846]: I1005 08:50:53.855104 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-openstack-openstack-cell1-wzg2n" event={"ID":"c9d6416a-ce5b-41a2-b49b-a67573544f06","Type":"ContainerDied","Data":"93ff7c5f5d35043aa0f0ba1ab2896064e8ab809c527312e0df9ae6fe3e4b3876"} Oct 05 08:50:55 crc kubenswrapper[4846]: I1005 08:50:55.419611 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-openstack-openstack-cell1-wzg2n" Oct 05 08:50:55 crc kubenswrapper[4846]: I1005 08:50:55.482722 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c9d6416a-ce5b-41a2-b49b-a67573544f06-ssh-key\") pod \"c9d6416a-ce5b-41a2-b49b-a67573544f06\" (UID: \"c9d6416a-ce5b-41a2-b49b-a67573544f06\") " Oct 05 08:50:55 crc kubenswrapper[4846]: I1005 08:50:55.482896 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lsfjx\" (UniqueName: \"kubernetes.io/projected/c9d6416a-ce5b-41a2-b49b-a67573544f06-kube-api-access-lsfjx\") pod \"c9d6416a-ce5b-41a2-b49b-a67573544f06\" (UID: \"c9d6416a-ce5b-41a2-b49b-a67573544f06\") " Oct 05 08:50:55 crc kubenswrapper[4846]: I1005 08:50:55.482968 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c9d6416a-ce5b-41a2-b49b-a67573544f06-inventory\") pod \"c9d6416a-ce5b-41a2-b49b-a67573544f06\" (UID: \"c9d6416a-ce5b-41a2-b49b-a67573544f06\") " Oct 05 08:50:55 crc kubenswrapper[4846]: I1005 08:50:55.496713 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c9d6416a-ce5b-41a2-b49b-a67573544f06-kube-api-access-lsfjx" (OuterVolumeSpecName: "kube-api-access-lsfjx") pod "c9d6416a-ce5b-41a2-b49b-a67573544f06" (UID: "c9d6416a-ce5b-41a2-b49b-a67573544f06"). InnerVolumeSpecName "kube-api-access-lsfjx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:50:55 crc kubenswrapper[4846]: I1005 08:50:55.514053 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c9d6416a-ce5b-41a2-b49b-a67573544f06-inventory" (OuterVolumeSpecName: "inventory") pod "c9d6416a-ce5b-41a2-b49b-a67573544f06" (UID: "c9d6416a-ce5b-41a2-b49b-a67573544f06"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:50:55 crc kubenswrapper[4846]: I1005 08:50:55.514632 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c9d6416a-ce5b-41a2-b49b-a67573544f06-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "c9d6416a-ce5b-41a2-b49b-a67573544f06" (UID: "c9d6416a-ce5b-41a2-b49b-a67573544f06"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:50:55 crc kubenswrapper[4846]: I1005 08:50:55.585680 4846 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c9d6416a-ce5b-41a2-b49b-a67573544f06-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 05 08:50:55 crc kubenswrapper[4846]: I1005 08:50:55.585744 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lsfjx\" (UniqueName: \"kubernetes.io/projected/c9d6416a-ce5b-41a2-b49b-a67573544f06-kube-api-access-lsfjx\") on node \"crc\" DevicePath \"\"" Oct 05 08:50:55 crc kubenswrapper[4846]: I1005 08:50:55.585760 4846 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c9d6416a-ce5b-41a2-b49b-a67573544f06-inventory\") on node \"crc\" DevicePath \"\"" Oct 05 08:50:55 crc kubenswrapper[4846]: I1005 08:50:55.879097 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-openstack-openstack-cell1-wzg2n" event={"ID":"c9d6416a-ce5b-41a2-b49b-a67573544f06","Type":"ContainerDied","Data":"17b103b093b5999621abd23d79a376a9fdf866e78515c9c16b33a0088cf91230"} Oct 05 08:50:55 crc kubenswrapper[4846]: I1005 08:50:55.879137 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="17b103b093b5999621abd23d79a376a9fdf866e78515c9c16b33a0088cf91230" Oct 05 08:50:55 crc kubenswrapper[4846]: I1005 08:50:55.879175 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-openstack-openstack-cell1-wzg2n" Oct 05 08:50:55 crc kubenswrapper[4846]: I1005 08:50:55.974400 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ssh-known-hosts-openstack-x82zm"] Oct 05 08:50:55 crc kubenswrapper[4846]: E1005 08:50:55.994138 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c9d6416a-ce5b-41a2-b49b-a67573544f06" containerName="configure-os-openstack-openstack-cell1" Oct 05 08:50:55 crc kubenswrapper[4846]: I1005 08:50:55.994405 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="c9d6416a-ce5b-41a2-b49b-a67573544f06" containerName="configure-os-openstack-openstack-cell1" Oct 05 08:50:55 crc kubenswrapper[4846]: I1005 08:50:55.994826 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="c9d6416a-ce5b-41a2-b49b-a67573544f06" containerName="configure-os-openstack-openstack-cell1" Oct 05 08:50:55 crc kubenswrapper[4846]: I1005 08:50:55.997030 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-openstack-x82zm"] Oct 05 08:50:55 crc kubenswrapper[4846]: I1005 08:50:55.997126 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-openstack-x82zm" Oct 05 08:50:56 crc kubenswrapper[4846]: I1005 08:50:56.001619 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 05 08:50:56 crc kubenswrapper[4846]: I1005 08:50:56.003124 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-7687s" Oct 05 08:50:56 crc kubenswrapper[4846]: I1005 08:50:56.003398 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Oct 05 08:50:56 crc kubenswrapper[4846]: I1005 08:50:56.003661 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Oct 05 08:50:56 crc kubenswrapper[4846]: I1005 08:50:56.097983 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/a87e92c9-28ef-42d1-a662-dea1a080b896-inventory-0\") pod \"ssh-known-hosts-openstack-x82zm\" (UID: \"a87e92c9-28ef-42d1-a662-dea1a080b896\") " pod="openstack/ssh-known-hosts-openstack-x82zm" Oct 05 08:50:56 crc kubenswrapper[4846]: I1005 08:50:56.098408 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/a87e92c9-28ef-42d1-a662-dea1a080b896-ssh-key-openstack-cell1\") pod \"ssh-known-hosts-openstack-x82zm\" (UID: \"a87e92c9-28ef-42d1-a662-dea1a080b896\") " pod="openstack/ssh-known-hosts-openstack-x82zm" Oct 05 08:50:56 crc kubenswrapper[4846]: I1005 08:50:56.098438 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8ms4j\" (UniqueName: \"kubernetes.io/projected/a87e92c9-28ef-42d1-a662-dea1a080b896-kube-api-access-8ms4j\") pod \"ssh-known-hosts-openstack-x82zm\" (UID: \"a87e92c9-28ef-42d1-a662-dea1a080b896\") " pod="openstack/ssh-known-hosts-openstack-x82zm" Oct 05 08:50:56 crc kubenswrapper[4846]: I1005 08:50:56.201212 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/a87e92c9-28ef-42d1-a662-dea1a080b896-ssh-key-openstack-cell1\") pod \"ssh-known-hosts-openstack-x82zm\" (UID: \"a87e92c9-28ef-42d1-a662-dea1a080b896\") " pod="openstack/ssh-known-hosts-openstack-x82zm" Oct 05 08:50:56 crc kubenswrapper[4846]: I1005 08:50:56.201279 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8ms4j\" (UniqueName: \"kubernetes.io/projected/a87e92c9-28ef-42d1-a662-dea1a080b896-kube-api-access-8ms4j\") pod \"ssh-known-hosts-openstack-x82zm\" (UID: \"a87e92c9-28ef-42d1-a662-dea1a080b896\") " pod="openstack/ssh-known-hosts-openstack-x82zm" Oct 05 08:50:56 crc kubenswrapper[4846]: I1005 08:50:56.201508 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/a87e92c9-28ef-42d1-a662-dea1a080b896-inventory-0\") pod \"ssh-known-hosts-openstack-x82zm\" (UID: \"a87e92c9-28ef-42d1-a662-dea1a080b896\") " pod="openstack/ssh-known-hosts-openstack-x82zm" Oct 05 08:50:56 crc kubenswrapper[4846]: I1005 08:50:56.218257 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/a87e92c9-28ef-42d1-a662-dea1a080b896-ssh-key-openstack-cell1\") pod \"ssh-known-hosts-openstack-x82zm\" (UID: \"a87e92c9-28ef-42d1-a662-dea1a080b896\") " pod="openstack/ssh-known-hosts-openstack-x82zm" Oct 05 08:50:56 crc kubenswrapper[4846]: I1005 08:50:56.222554 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/a87e92c9-28ef-42d1-a662-dea1a080b896-inventory-0\") pod \"ssh-known-hosts-openstack-x82zm\" (UID: \"a87e92c9-28ef-42d1-a662-dea1a080b896\") " pod="openstack/ssh-known-hosts-openstack-x82zm" Oct 05 08:50:56 crc kubenswrapper[4846]: I1005 08:50:56.233002 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8ms4j\" (UniqueName: \"kubernetes.io/projected/a87e92c9-28ef-42d1-a662-dea1a080b896-kube-api-access-8ms4j\") pod \"ssh-known-hosts-openstack-x82zm\" (UID: \"a87e92c9-28ef-42d1-a662-dea1a080b896\") " pod="openstack/ssh-known-hosts-openstack-x82zm" Oct 05 08:50:56 crc kubenswrapper[4846]: I1005 08:50:56.281406 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-wq9mr"] Oct 05 08:50:56 crc kubenswrapper[4846]: I1005 08:50:56.284368 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wq9mr" Oct 05 08:50:56 crc kubenswrapper[4846]: I1005 08:50:56.297546 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-wq9mr"] Oct 05 08:50:56 crc kubenswrapper[4846]: I1005 08:50:56.322563 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-openstack-x82zm" Oct 05 08:50:56 crc kubenswrapper[4846]: I1005 08:50:56.425298 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fjdfg\" (UniqueName: \"kubernetes.io/projected/b0582274-1b4f-4050-979d-053265764498-kube-api-access-fjdfg\") pod \"certified-operators-wq9mr\" (UID: \"b0582274-1b4f-4050-979d-053265764498\") " pod="openshift-marketplace/certified-operators-wq9mr" Oct 05 08:50:56 crc kubenswrapper[4846]: I1005 08:50:56.425392 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b0582274-1b4f-4050-979d-053265764498-utilities\") pod \"certified-operators-wq9mr\" (UID: \"b0582274-1b4f-4050-979d-053265764498\") " pod="openshift-marketplace/certified-operators-wq9mr" Oct 05 08:50:56 crc kubenswrapper[4846]: I1005 08:50:56.425449 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b0582274-1b4f-4050-979d-053265764498-catalog-content\") pod \"certified-operators-wq9mr\" (UID: \"b0582274-1b4f-4050-979d-053265764498\") " pod="openshift-marketplace/certified-operators-wq9mr" Oct 05 08:50:56 crc kubenswrapper[4846]: I1005 08:50:56.530202 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fjdfg\" (UniqueName: \"kubernetes.io/projected/b0582274-1b4f-4050-979d-053265764498-kube-api-access-fjdfg\") pod \"certified-operators-wq9mr\" (UID: \"b0582274-1b4f-4050-979d-053265764498\") " pod="openshift-marketplace/certified-operators-wq9mr" Oct 05 08:50:56 crc kubenswrapper[4846]: I1005 08:50:56.530781 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b0582274-1b4f-4050-979d-053265764498-utilities\") pod \"certified-operators-wq9mr\" (UID: \"b0582274-1b4f-4050-979d-053265764498\") " pod="openshift-marketplace/certified-operators-wq9mr" Oct 05 08:50:56 crc kubenswrapper[4846]: I1005 08:50:56.531054 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b0582274-1b4f-4050-979d-053265764498-catalog-content\") pod \"certified-operators-wq9mr\" (UID: \"b0582274-1b4f-4050-979d-053265764498\") " pod="openshift-marketplace/certified-operators-wq9mr" Oct 05 08:50:56 crc kubenswrapper[4846]: I1005 08:50:56.531506 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b0582274-1b4f-4050-979d-053265764498-catalog-content\") pod \"certified-operators-wq9mr\" (UID: \"b0582274-1b4f-4050-979d-053265764498\") " pod="openshift-marketplace/certified-operators-wq9mr" Oct 05 08:50:56 crc kubenswrapper[4846]: I1005 08:50:56.531592 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b0582274-1b4f-4050-979d-053265764498-utilities\") pod \"certified-operators-wq9mr\" (UID: \"b0582274-1b4f-4050-979d-053265764498\") " pod="openshift-marketplace/certified-operators-wq9mr" Oct 05 08:50:56 crc kubenswrapper[4846]: I1005 08:50:56.547087 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fjdfg\" (UniqueName: \"kubernetes.io/projected/b0582274-1b4f-4050-979d-053265764498-kube-api-access-fjdfg\") pod \"certified-operators-wq9mr\" (UID: \"b0582274-1b4f-4050-979d-053265764498\") " pod="openshift-marketplace/certified-operators-wq9mr" Oct 05 08:50:56 crc kubenswrapper[4846]: I1005 08:50:56.756667 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wq9mr" Oct 05 08:50:56 crc kubenswrapper[4846]: I1005 08:50:56.940819 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-openstack-x82zm"] Oct 05 08:50:57 crc kubenswrapper[4846]: W1005 08:50:57.270435 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb0582274_1b4f_4050_979d_053265764498.slice/crio-3eba90ce14910da9eeb26b729eb5d0000bce14569b08c9551dd5d03977df9c92 WatchSource:0}: Error finding container 3eba90ce14910da9eeb26b729eb5d0000bce14569b08c9551dd5d03977df9c92: Status 404 returned error can't find the container with id 3eba90ce14910da9eeb26b729eb5d0000bce14569b08c9551dd5d03977df9c92 Oct 05 08:50:57 crc kubenswrapper[4846]: I1005 08:50:57.278029 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-wq9mr"] Oct 05 08:50:57 crc kubenswrapper[4846]: I1005 08:50:57.902019 4846 generic.go:334] "Generic (PLEG): container finished" podID="b0582274-1b4f-4050-979d-053265764498" containerID="9b2e1f36d4ea69556237161b28f2ccec8ffa73e742c764717174625d2e8a5807" exitCode=0 Oct 05 08:50:57 crc kubenswrapper[4846]: I1005 08:50:57.902129 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wq9mr" event={"ID":"b0582274-1b4f-4050-979d-053265764498","Type":"ContainerDied","Data":"9b2e1f36d4ea69556237161b28f2ccec8ffa73e742c764717174625d2e8a5807"} Oct 05 08:50:57 crc kubenswrapper[4846]: I1005 08:50:57.903000 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wq9mr" event={"ID":"b0582274-1b4f-4050-979d-053265764498","Type":"ContainerStarted","Data":"3eba90ce14910da9eeb26b729eb5d0000bce14569b08c9551dd5d03977df9c92"} Oct 05 08:50:57 crc kubenswrapper[4846]: I1005 08:50:57.908726 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-openstack-x82zm" event={"ID":"a87e92c9-28ef-42d1-a662-dea1a080b896","Type":"ContainerStarted","Data":"9a04700732d20fcb9e05e021beae61e791455d1594b2a0aefad7a55979f2b29f"} Oct 05 08:50:57 crc kubenswrapper[4846]: I1005 08:50:57.908773 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-openstack-x82zm" event={"ID":"a87e92c9-28ef-42d1-a662-dea1a080b896","Type":"ContainerStarted","Data":"11c34cf1c25dae76b115c9c7fd016cbd7f410ad55a502910ea60acbf07d65e6f"} Oct 05 08:50:57 crc kubenswrapper[4846]: I1005 08:50:57.949394 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ssh-known-hosts-openstack-x82zm" podStartSLOduration=2.292071008 podStartE2EDuration="2.949376692s" podCreationTimestamp="2025-10-05 08:50:55 +0000 UTC" firstStartedPulling="2025-10-05 08:50:56.954417037 +0000 UTC m=+7379.195269812" lastFinishedPulling="2025-10-05 08:50:57.611722721 +0000 UTC m=+7379.852575496" observedRunningTime="2025-10-05 08:50:57.946026183 +0000 UTC m=+7380.186878968" watchObservedRunningTime="2025-10-05 08:50:57.949376692 +0000 UTC m=+7380.190229467" Oct 05 08:50:58 crc kubenswrapper[4846]: I1005 08:50:58.924644 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wq9mr" event={"ID":"b0582274-1b4f-4050-979d-053265764498","Type":"ContainerStarted","Data":"5dad89d929aa074de1e9d3e822dfd8e8cbecbc6f1cb7c126bdc0d23d2a86f56f"} Oct 05 08:51:00 crc kubenswrapper[4846]: I1005 08:51:00.955833 4846 generic.go:334] "Generic (PLEG): container finished" podID="b0582274-1b4f-4050-979d-053265764498" containerID="5dad89d929aa074de1e9d3e822dfd8e8cbecbc6f1cb7c126bdc0d23d2a86f56f" exitCode=0 Oct 05 08:51:00 crc kubenswrapper[4846]: I1005 08:51:00.956287 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wq9mr" event={"ID":"b0582274-1b4f-4050-979d-053265764498","Type":"ContainerDied","Data":"5dad89d929aa074de1e9d3e822dfd8e8cbecbc6f1cb7c126bdc0d23d2a86f56f"} Oct 05 08:51:01 crc kubenswrapper[4846]: I1005 08:51:01.971328 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wq9mr" event={"ID":"b0582274-1b4f-4050-979d-053265764498","Type":"ContainerStarted","Data":"7d1585e8f37ce1c2d29f7e87768d757bcf1afbd7d805ffb09671c5949d319a06"} Oct 05 08:51:02 crc kubenswrapper[4846]: I1005 08:51:02.005324 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-wq9mr" podStartSLOduration=2.545666281 podStartE2EDuration="6.005293387s" podCreationTimestamp="2025-10-05 08:50:56 +0000 UTC" firstStartedPulling="2025-10-05 08:50:57.90537078 +0000 UTC m=+7380.146223605" lastFinishedPulling="2025-10-05 08:51:01.364997896 +0000 UTC m=+7383.605850711" observedRunningTime="2025-10-05 08:51:01.994339175 +0000 UTC m=+7384.235191990" watchObservedRunningTime="2025-10-05 08:51:02.005293387 +0000 UTC m=+7384.246146202" Oct 05 08:51:06 crc kubenswrapper[4846]: I1005 08:51:06.757158 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-wq9mr" Oct 05 08:51:06 crc kubenswrapper[4846]: I1005 08:51:06.757590 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-wq9mr" Oct 05 08:51:06 crc kubenswrapper[4846]: I1005 08:51:06.817087 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-wq9mr" Oct 05 08:51:07 crc kubenswrapper[4846]: I1005 08:51:07.033552 4846 generic.go:334] "Generic (PLEG): container finished" podID="a87e92c9-28ef-42d1-a662-dea1a080b896" containerID="9a04700732d20fcb9e05e021beae61e791455d1594b2a0aefad7a55979f2b29f" exitCode=0 Oct 05 08:51:07 crc kubenswrapper[4846]: I1005 08:51:07.033698 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-openstack-x82zm" event={"ID":"a87e92c9-28ef-42d1-a662-dea1a080b896","Type":"ContainerDied","Data":"9a04700732d20fcb9e05e021beae61e791455d1594b2a0aefad7a55979f2b29f"} Oct 05 08:51:07 crc kubenswrapper[4846]: I1005 08:51:07.124953 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-wq9mr" Oct 05 08:51:07 crc kubenswrapper[4846]: I1005 08:51:07.176964 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-wq9mr"] Oct 05 08:51:08 crc kubenswrapper[4846]: I1005 08:51:08.632271 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-openstack-x82zm" Oct 05 08:51:08 crc kubenswrapper[4846]: I1005 08:51:08.803809 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/a87e92c9-28ef-42d1-a662-dea1a080b896-ssh-key-openstack-cell1\") pod \"a87e92c9-28ef-42d1-a662-dea1a080b896\" (UID: \"a87e92c9-28ef-42d1-a662-dea1a080b896\") " Oct 05 08:51:08 crc kubenswrapper[4846]: I1005 08:51:08.803913 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/a87e92c9-28ef-42d1-a662-dea1a080b896-inventory-0\") pod \"a87e92c9-28ef-42d1-a662-dea1a080b896\" (UID: \"a87e92c9-28ef-42d1-a662-dea1a080b896\") " Oct 05 08:51:08 crc kubenswrapper[4846]: I1005 08:51:08.804118 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8ms4j\" (UniqueName: \"kubernetes.io/projected/a87e92c9-28ef-42d1-a662-dea1a080b896-kube-api-access-8ms4j\") pod \"a87e92c9-28ef-42d1-a662-dea1a080b896\" (UID: \"a87e92c9-28ef-42d1-a662-dea1a080b896\") " Oct 05 08:51:08 crc kubenswrapper[4846]: I1005 08:51:08.811898 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a87e92c9-28ef-42d1-a662-dea1a080b896-kube-api-access-8ms4j" (OuterVolumeSpecName: "kube-api-access-8ms4j") pod "a87e92c9-28ef-42d1-a662-dea1a080b896" (UID: "a87e92c9-28ef-42d1-a662-dea1a080b896"). InnerVolumeSpecName "kube-api-access-8ms4j". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:51:08 crc kubenswrapper[4846]: I1005 08:51:08.841262 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a87e92c9-28ef-42d1-a662-dea1a080b896-ssh-key-openstack-cell1" (OuterVolumeSpecName: "ssh-key-openstack-cell1") pod "a87e92c9-28ef-42d1-a662-dea1a080b896" (UID: "a87e92c9-28ef-42d1-a662-dea1a080b896"). InnerVolumeSpecName "ssh-key-openstack-cell1". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:51:08 crc kubenswrapper[4846]: I1005 08:51:08.853001 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a87e92c9-28ef-42d1-a662-dea1a080b896-inventory-0" (OuterVolumeSpecName: "inventory-0") pod "a87e92c9-28ef-42d1-a662-dea1a080b896" (UID: "a87e92c9-28ef-42d1-a662-dea1a080b896"). InnerVolumeSpecName "inventory-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:51:08 crc kubenswrapper[4846]: I1005 08:51:08.907862 4846 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/a87e92c9-28ef-42d1-a662-dea1a080b896-ssh-key-openstack-cell1\") on node \"crc\" DevicePath \"\"" Oct 05 08:51:08 crc kubenswrapper[4846]: I1005 08:51:08.907942 4846 reconciler_common.go:293] "Volume detached for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/a87e92c9-28ef-42d1-a662-dea1a080b896-inventory-0\") on node \"crc\" DevicePath \"\"" Oct 05 08:51:08 crc kubenswrapper[4846]: I1005 08:51:08.907973 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8ms4j\" (UniqueName: \"kubernetes.io/projected/a87e92c9-28ef-42d1-a662-dea1a080b896-kube-api-access-8ms4j\") on node \"crc\" DevicePath \"\"" Oct 05 08:51:09 crc kubenswrapper[4846]: I1005 08:51:09.059126 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-openstack-x82zm" Oct 05 08:51:09 crc kubenswrapper[4846]: I1005 08:51:09.059115 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-openstack-x82zm" event={"ID":"a87e92c9-28ef-42d1-a662-dea1a080b896","Type":"ContainerDied","Data":"11c34cf1c25dae76b115c9c7fd016cbd7f410ad55a502910ea60acbf07d65e6f"} Oct 05 08:51:09 crc kubenswrapper[4846]: I1005 08:51:09.059209 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="11c34cf1c25dae76b115c9c7fd016cbd7f410ad55a502910ea60acbf07d65e6f" Oct 05 08:51:09 crc kubenswrapper[4846]: I1005 08:51:09.059386 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-wq9mr" podUID="b0582274-1b4f-4050-979d-053265764498" containerName="registry-server" containerID="cri-o://7d1585e8f37ce1c2d29f7e87768d757bcf1afbd7d805ffb09671c5949d319a06" gracePeriod=2 Oct 05 08:51:09 crc kubenswrapper[4846]: I1005 08:51:09.169300 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/run-os-openstack-openstack-cell1-m2snw"] Oct 05 08:51:09 crc kubenswrapper[4846]: E1005 08:51:09.169860 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a87e92c9-28ef-42d1-a662-dea1a080b896" containerName="ssh-known-hosts-openstack" Oct 05 08:51:09 crc kubenswrapper[4846]: I1005 08:51:09.169884 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="a87e92c9-28ef-42d1-a662-dea1a080b896" containerName="ssh-known-hosts-openstack" Oct 05 08:51:09 crc kubenswrapper[4846]: I1005 08:51:09.170147 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="a87e92c9-28ef-42d1-a662-dea1a080b896" containerName="ssh-known-hosts-openstack" Oct 05 08:51:09 crc kubenswrapper[4846]: I1005 08:51:09.171619 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-openstack-openstack-cell1-m2snw" Oct 05 08:51:09 crc kubenswrapper[4846]: I1005 08:51:09.173343 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-7687s" Oct 05 08:51:09 crc kubenswrapper[4846]: I1005 08:51:09.174360 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 05 08:51:09 crc kubenswrapper[4846]: I1005 08:51:09.174717 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Oct 05 08:51:09 crc kubenswrapper[4846]: I1005 08:51:09.177805 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Oct 05 08:51:09 crc kubenswrapper[4846]: I1005 08:51:09.200367 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-openstack-openstack-cell1-m2snw"] Oct 05 08:51:09 crc kubenswrapper[4846]: I1005 08:51:09.315000 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fe44fec2-66d0-4b82-909b-3a27f42340db-inventory\") pod \"run-os-openstack-openstack-cell1-m2snw\" (UID: \"fe44fec2-66d0-4b82-909b-3a27f42340db\") " pod="openstack/run-os-openstack-openstack-cell1-m2snw" Oct 05 08:51:09 crc kubenswrapper[4846]: I1005 08:51:09.315439 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fe44fec2-66d0-4b82-909b-3a27f42340db-ssh-key\") pod \"run-os-openstack-openstack-cell1-m2snw\" (UID: \"fe44fec2-66d0-4b82-909b-3a27f42340db\") " pod="openstack/run-os-openstack-openstack-cell1-m2snw" Oct 05 08:51:09 crc kubenswrapper[4846]: I1005 08:51:09.315491 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6ncpj\" (UniqueName: \"kubernetes.io/projected/fe44fec2-66d0-4b82-909b-3a27f42340db-kube-api-access-6ncpj\") pod \"run-os-openstack-openstack-cell1-m2snw\" (UID: \"fe44fec2-66d0-4b82-909b-3a27f42340db\") " pod="openstack/run-os-openstack-openstack-cell1-m2snw" Oct 05 08:51:09 crc kubenswrapper[4846]: I1005 08:51:09.417641 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fe44fec2-66d0-4b82-909b-3a27f42340db-inventory\") pod \"run-os-openstack-openstack-cell1-m2snw\" (UID: \"fe44fec2-66d0-4b82-909b-3a27f42340db\") " pod="openstack/run-os-openstack-openstack-cell1-m2snw" Oct 05 08:51:09 crc kubenswrapper[4846]: I1005 08:51:09.417697 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fe44fec2-66d0-4b82-909b-3a27f42340db-ssh-key\") pod \"run-os-openstack-openstack-cell1-m2snw\" (UID: \"fe44fec2-66d0-4b82-909b-3a27f42340db\") " pod="openstack/run-os-openstack-openstack-cell1-m2snw" Oct 05 08:51:09 crc kubenswrapper[4846]: I1005 08:51:09.417752 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6ncpj\" (UniqueName: \"kubernetes.io/projected/fe44fec2-66d0-4b82-909b-3a27f42340db-kube-api-access-6ncpj\") pod \"run-os-openstack-openstack-cell1-m2snw\" (UID: \"fe44fec2-66d0-4b82-909b-3a27f42340db\") " pod="openstack/run-os-openstack-openstack-cell1-m2snw" Oct 05 08:51:09 crc kubenswrapper[4846]: I1005 08:51:09.425001 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fe44fec2-66d0-4b82-909b-3a27f42340db-inventory\") pod \"run-os-openstack-openstack-cell1-m2snw\" (UID: \"fe44fec2-66d0-4b82-909b-3a27f42340db\") " pod="openstack/run-os-openstack-openstack-cell1-m2snw" Oct 05 08:51:09 crc kubenswrapper[4846]: I1005 08:51:09.427648 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fe44fec2-66d0-4b82-909b-3a27f42340db-ssh-key\") pod \"run-os-openstack-openstack-cell1-m2snw\" (UID: \"fe44fec2-66d0-4b82-909b-3a27f42340db\") " pod="openstack/run-os-openstack-openstack-cell1-m2snw" Oct 05 08:51:09 crc kubenswrapper[4846]: I1005 08:51:09.433272 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6ncpj\" (UniqueName: \"kubernetes.io/projected/fe44fec2-66d0-4b82-909b-3a27f42340db-kube-api-access-6ncpj\") pod \"run-os-openstack-openstack-cell1-m2snw\" (UID: \"fe44fec2-66d0-4b82-909b-3a27f42340db\") " pod="openstack/run-os-openstack-openstack-cell1-m2snw" Oct 05 08:51:09 crc kubenswrapper[4846]: I1005 08:51:09.437171 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wq9mr" Oct 05 08:51:09 crc kubenswrapper[4846]: I1005 08:51:09.519424 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b0582274-1b4f-4050-979d-053265764498-utilities\") pod \"b0582274-1b4f-4050-979d-053265764498\" (UID: \"b0582274-1b4f-4050-979d-053265764498\") " Oct 05 08:51:09 crc kubenswrapper[4846]: I1005 08:51:09.519665 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b0582274-1b4f-4050-979d-053265764498-catalog-content\") pod \"b0582274-1b4f-4050-979d-053265764498\" (UID: \"b0582274-1b4f-4050-979d-053265764498\") " Oct 05 08:51:09 crc kubenswrapper[4846]: I1005 08:51:09.519738 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fjdfg\" (UniqueName: \"kubernetes.io/projected/b0582274-1b4f-4050-979d-053265764498-kube-api-access-fjdfg\") pod \"b0582274-1b4f-4050-979d-053265764498\" (UID: \"b0582274-1b4f-4050-979d-053265764498\") " Oct 05 08:51:09 crc kubenswrapper[4846]: I1005 08:51:09.521505 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b0582274-1b4f-4050-979d-053265764498-utilities" (OuterVolumeSpecName: "utilities") pod "b0582274-1b4f-4050-979d-053265764498" (UID: "b0582274-1b4f-4050-979d-053265764498"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 08:51:09 crc kubenswrapper[4846]: I1005 08:51:09.525052 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b0582274-1b4f-4050-979d-053265764498-kube-api-access-fjdfg" (OuterVolumeSpecName: "kube-api-access-fjdfg") pod "b0582274-1b4f-4050-979d-053265764498" (UID: "b0582274-1b4f-4050-979d-053265764498"). InnerVolumeSpecName "kube-api-access-fjdfg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:51:09 crc kubenswrapper[4846]: I1005 08:51:09.533228 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-openstack-openstack-cell1-m2snw" Oct 05 08:51:09 crc kubenswrapper[4846]: I1005 08:51:09.576297 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b0582274-1b4f-4050-979d-053265764498-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b0582274-1b4f-4050-979d-053265764498" (UID: "b0582274-1b4f-4050-979d-053265764498"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 08:51:09 crc kubenswrapper[4846]: I1005 08:51:09.622576 4846 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b0582274-1b4f-4050-979d-053265764498-utilities\") on node \"crc\" DevicePath \"\"" Oct 05 08:51:09 crc kubenswrapper[4846]: I1005 08:51:09.622616 4846 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b0582274-1b4f-4050-979d-053265764498-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 05 08:51:09 crc kubenswrapper[4846]: I1005 08:51:09.622630 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fjdfg\" (UniqueName: \"kubernetes.io/projected/b0582274-1b4f-4050-979d-053265764498-kube-api-access-fjdfg\") on node \"crc\" DevicePath \"\"" Oct 05 08:51:09 crc kubenswrapper[4846]: I1005 08:51:09.878511 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-openstack-openstack-cell1-m2snw"] Oct 05 08:51:10 crc kubenswrapper[4846]: I1005 08:51:10.079988 4846 generic.go:334] "Generic (PLEG): container finished" podID="b0582274-1b4f-4050-979d-053265764498" containerID="7d1585e8f37ce1c2d29f7e87768d757bcf1afbd7d805ffb09671c5949d319a06" exitCode=0 Oct 05 08:51:10 crc kubenswrapper[4846]: I1005 08:51:10.080090 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wq9mr" Oct 05 08:51:10 crc kubenswrapper[4846]: I1005 08:51:10.080109 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wq9mr" event={"ID":"b0582274-1b4f-4050-979d-053265764498","Type":"ContainerDied","Data":"7d1585e8f37ce1c2d29f7e87768d757bcf1afbd7d805ffb09671c5949d319a06"} Oct 05 08:51:10 crc kubenswrapper[4846]: I1005 08:51:10.080204 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wq9mr" event={"ID":"b0582274-1b4f-4050-979d-053265764498","Type":"ContainerDied","Data":"3eba90ce14910da9eeb26b729eb5d0000bce14569b08c9551dd5d03977df9c92"} Oct 05 08:51:10 crc kubenswrapper[4846]: I1005 08:51:10.080240 4846 scope.go:117] "RemoveContainer" containerID="7d1585e8f37ce1c2d29f7e87768d757bcf1afbd7d805ffb09671c5949d319a06" Oct 05 08:51:10 crc kubenswrapper[4846]: I1005 08:51:10.082577 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-openstack-openstack-cell1-m2snw" event={"ID":"fe44fec2-66d0-4b82-909b-3a27f42340db","Type":"ContainerStarted","Data":"8c9d4b403dbc06ffaf68621f6f1f583ebb9603f1950552e8e7bfd6ec55048120"} Oct 05 08:51:10 crc kubenswrapper[4846]: I1005 08:51:10.108894 4846 scope.go:117] "RemoveContainer" containerID="5dad89d929aa074de1e9d3e822dfd8e8cbecbc6f1cb7c126bdc0d23d2a86f56f" Oct 05 08:51:10 crc kubenswrapper[4846]: I1005 08:51:10.143692 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-wq9mr"] Oct 05 08:51:10 crc kubenswrapper[4846]: I1005 08:51:10.154449 4846 scope.go:117] "RemoveContainer" containerID="9b2e1f36d4ea69556237161b28f2ccec8ffa73e742c764717174625d2e8a5807" Oct 05 08:51:10 crc kubenswrapper[4846]: I1005 08:51:10.160737 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-wq9mr"] Oct 05 08:51:10 crc kubenswrapper[4846]: I1005 08:51:10.216846 4846 scope.go:117] "RemoveContainer" containerID="7d1585e8f37ce1c2d29f7e87768d757bcf1afbd7d805ffb09671c5949d319a06" Oct 05 08:51:10 crc kubenswrapper[4846]: E1005 08:51:10.217505 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7d1585e8f37ce1c2d29f7e87768d757bcf1afbd7d805ffb09671c5949d319a06\": container with ID starting with 7d1585e8f37ce1c2d29f7e87768d757bcf1afbd7d805ffb09671c5949d319a06 not found: ID does not exist" containerID="7d1585e8f37ce1c2d29f7e87768d757bcf1afbd7d805ffb09671c5949d319a06" Oct 05 08:51:10 crc kubenswrapper[4846]: I1005 08:51:10.217551 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7d1585e8f37ce1c2d29f7e87768d757bcf1afbd7d805ffb09671c5949d319a06"} err="failed to get container status \"7d1585e8f37ce1c2d29f7e87768d757bcf1afbd7d805ffb09671c5949d319a06\": rpc error: code = NotFound desc = could not find container \"7d1585e8f37ce1c2d29f7e87768d757bcf1afbd7d805ffb09671c5949d319a06\": container with ID starting with 7d1585e8f37ce1c2d29f7e87768d757bcf1afbd7d805ffb09671c5949d319a06 not found: ID does not exist" Oct 05 08:51:10 crc kubenswrapper[4846]: I1005 08:51:10.217584 4846 scope.go:117] "RemoveContainer" containerID="5dad89d929aa074de1e9d3e822dfd8e8cbecbc6f1cb7c126bdc0d23d2a86f56f" Oct 05 08:51:10 crc kubenswrapper[4846]: E1005 08:51:10.218073 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5dad89d929aa074de1e9d3e822dfd8e8cbecbc6f1cb7c126bdc0d23d2a86f56f\": container with ID starting with 5dad89d929aa074de1e9d3e822dfd8e8cbecbc6f1cb7c126bdc0d23d2a86f56f not found: ID does not exist" containerID="5dad89d929aa074de1e9d3e822dfd8e8cbecbc6f1cb7c126bdc0d23d2a86f56f" Oct 05 08:51:10 crc kubenswrapper[4846]: I1005 08:51:10.218265 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5dad89d929aa074de1e9d3e822dfd8e8cbecbc6f1cb7c126bdc0d23d2a86f56f"} err="failed to get container status \"5dad89d929aa074de1e9d3e822dfd8e8cbecbc6f1cb7c126bdc0d23d2a86f56f\": rpc error: code = NotFound desc = could not find container \"5dad89d929aa074de1e9d3e822dfd8e8cbecbc6f1cb7c126bdc0d23d2a86f56f\": container with ID starting with 5dad89d929aa074de1e9d3e822dfd8e8cbecbc6f1cb7c126bdc0d23d2a86f56f not found: ID does not exist" Oct 05 08:51:10 crc kubenswrapper[4846]: I1005 08:51:10.218418 4846 scope.go:117] "RemoveContainer" containerID="9b2e1f36d4ea69556237161b28f2ccec8ffa73e742c764717174625d2e8a5807" Oct 05 08:51:10 crc kubenswrapper[4846]: E1005 08:51:10.219066 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9b2e1f36d4ea69556237161b28f2ccec8ffa73e742c764717174625d2e8a5807\": container with ID starting with 9b2e1f36d4ea69556237161b28f2ccec8ffa73e742c764717174625d2e8a5807 not found: ID does not exist" containerID="9b2e1f36d4ea69556237161b28f2ccec8ffa73e742c764717174625d2e8a5807" Oct 05 08:51:10 crc kubenswrapper[4846]: I1005 08:51:10.219102 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9b2e1f36d4ea69556237161b28f2ccec8ffa73e742c764717174625d2e8a5807"} err="failed to get container status \"9b2e1f36d4ea69556237161b28f2ccec8ffa73e742c764717174625d2e8a5807\": rpc error: code = NotFound desc = could not find container \"9b2e1f36d4ea69556237161b28f2ccec8ffa73e742c764717174625d2e8a5807\": container with ID starting with 9b2e1f36d4ea69556237161b28f2ccec8ffa73e742c764717174625d2e8a5807 not found: ID does not exist" Oct 05 08:51:10 crc kubenswrapper[4846]: I1005 08:51:10.515071 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b0582274-1b4f-4050-979d-053265764498" path="/var/lib/kubelet/pods/b0582274-1b4f-4050-979d-053265764498/volumes" Oct 05 08:51:11 crc kubenswrapper[4846]: I1005 08:51:11.105364 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-openstack-openstack-cell1-m2snw" event={"ID":"fe44fec2-66d0-4b82-909b-3a27f42340db","Type":"ContainerStarted","Data":"355eeb4f237cedf8f917d303f2f665a96d68446e90c4541d1951c2c9932295cf"} Oct 05 08:51:11 crc kubenswrapper[4846]: I1005 08:51:11.130509 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/run-os-openstack-openstack-cell1-m2snw" podStartSLOduration=1.747337029 podStartE2EDuration="2.130477441s" podCreationTimestamp="2025-10-05 08:51:09 +0000 UTC" firstStartedPulling="2025-10-05 08:51:09.881892002 +0000 UTC m=+7392.122744777" lastFinishedPulling="2025-10-05 08:51:10.265032374 +0000 UTC m=+7392.505885189" observedRunningTime="2025-10-05 08:51:11.127126442 +0000 UTC m=+7393.367979227" watchObservedRunningTime="2025-10-05 08:51:11.130477441 +0000 UTC m=+7393.371330256" Oct 05 08:51:19 crc kubenswrapper[4846]: I1005 08:51:19.201498 4846 generic.go:334] "Generic (PLEG): container finished" podID="fe44fec2-66d0-4b82-909b-3a27f42340db" containerID="355eeb4f237cedf8f917d303f2f665a96d68446e90c4541d1951c2c9932295cf" exitCode=0 Oct 05 08:51:19 crc kubenswrapper[4846]: I1005 08:51:19.201581 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-openstack-openstack-cell1-m2snw" event={"ID":"fe44fec2-66d0-4b82-909b-3a27f42340db","Type":"ContainerDied","Data":"355eeb4f237cedf8f917d303f2f665a96d68446e90c4541d1951c2c9932295cf"} Oct 05 08:51:20 crc kubenswrapper[4846]: I1005 08:51:20.713157 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-openstack-openstack-cell1-m2snw" Oct 05 08:51:20 crc kubenswrapper[4846]: I1005 08:51:20.809956 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ncpj\" (UniqueName: \"kubernetes.io/projected/fe44fec2-66d0-4b82-909b-3a27f42340db-kube-api-access-6ncpj\") pod \"fe44fec2-66d0-4b82-909b-3a27f42340db\" (UID: \"fe44fec2-66d0-4b82-909b-3a27f42340db\") " Oct 05 08:51:20 crc kubenswrapper[4846]: I1005 08:51:20.810137 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fe44fec2-66d0-4b82-909b-3a27f42340db-ssh-key\") pod \"fe44fec2-66d0-4b82-909b-3a27f42340db\" (UID: \"fe44fec2-66d0-4b82-909b-3a27f42340db\") " Oct 05 08:51:20 crc kubenswrapper[4846]: I1005 08:51:20.810515 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fe44fec2-66d0-4b82-909b-3a27f42340db-inventory\") pod \"fe44fec2-66d0-4b82-909b-3a27f42340db\" (UID: \"fe44fec2-66d0-4b82-909b-3a27f42340db\") " Oct 05 08:51:20 crc kubenswrapper[4846]: I1005 08:51:20.818760 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fe44fec2-66d0-4b82-909b-3a27f42340db-kube-api-access-6ncpj" (OuterVolumeSpecName: "kube-api-access-6ncpj") pod "fe44fec2-66d0-4b82-909b-3a27f42340db" (UID: "fe44fec2-66d0-4b82-909b-3a27f42340db"). InnerVolumeSpecName "kube-api-access-6ncpj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:51:20 crc kubenswrapper[4846]: I1005 08:51:20.861617 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fe44fec2-66d0-4b82-909b-3a27f42340db-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "fe44fec2-66d0-4b82-909b-3a27f42340db" (UID: "fe44fec2-66d0-4b82-909b-3a27f42340db"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:51:20 crc kubenswrapper[4846]: I1005 08:51:20.861995 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fe44fec2-66d0-4b82-909b-3a27f42340db-inventory" (OuterVolumeSpecName: "inventory") pod "fe44fec2-66d0-4b82-909b-3a27f42340db" (UID: "fe44fec2-66d0-4b82-909b-3a27f42340db"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:51:20 crc kubenswrapper[4846]: I1005 08:51:20.914242 4846 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fe44fec2-66d0-4b82-909b-3a27f42340db-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 05 08:51:20 crc kubenswrapper[4846]: I1005 08:51:20.914284 4846 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fe44fec2-66d0-4b82-909b-3a27f42340db-inventory\") on node \"crc\" DevicePath \"\"" Oct 05 08:51:20 crc kubenswrapper[4846]: I1005 08:51:20.914299 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ncpj\" (UniqueName: \"kubernetes.io/projected/fe44fec2-66d0-4b82-909b-3a27f42340db-kube-api-access-6ncpj\") on node \"crc\" DevicePath \"\"" Oct 05 08:51:21 crc kubenswrapper[4846]: I1005 08:51:21.234241 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-openstack-openstack-cell1-m2snw" event={"ID":"fe44fec2-66d0-4b82-909b-3a27f42340db","Type":"ContainerDied","Data":"8c9d4b403dbc06ffaf68621f6f1f583ebb9603f1950552e8e7bfd6ec55048120"} Oct 05 08:51:21 crc kubenswrapper[4846]: I1005 08:51:21.234308 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8c9d4b403dbc06ffaf68621f6f1f583ebb9603f1950552e8e7bfd6ec55048120" Oct 05 08:51:21 crc kubenswrapper[4846]: I1005 08:51:21.234390 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-openstack-openstack-cell1-m2snw" Oct 05 08:51:21 crc kubenswrapper[4846]: I1005 08:51:21.343648 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/reboot-os-openstack-openstack-cell1-fdct8"] Oct 05 08:51:21 crc kubenswrapper[4846]: E1005 08:51:21.344149 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b0582274-1b4f-4050-979d-053265764498" containerName="extract-content" Oct 05 08:51:21 crc kubenswrapper[4846]: I1005 08:51:21.344197 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="b0582274-1b4f-4050-979d-053265764498" containerName="extract-content" Oct 05 08:51:21 crc kubenswrapper[4846]: E1005 08:51:21.344215 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fe44fec2-66d0-4b82-909b-3a27f42340db" containerName="run-os-openstack-openstack-cell1" Oct 05 08:51:21 crc kubenswrapper[4846]: I1005 08:51:21.344252 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="fe44fec2-66d0-4b82-909b-3a27f42340db" containerName="run-os-openstack-openstack-cell1" Oct 05 08:51:21 crc kubenswrapper[4846]: E1005 08:51:21.344278 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b0582274-1b4f-4050-979d-053265764498" containerName="extract-utilities" Oct 05 08:51:21 crc kubenswrapper[4846]: I1005 08:51:21.344308 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="b0582274-1b4f-4050-979d-053265764498" containerName="extract-utilities" Oct 05 08:51:21 crc kubenswrapper[4846]: E1005 08:51:21.344348 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b0582274-1b4f-4050-979d-053265764498" containerName="registry-server" Oct 05 08:51:21 crc kubenswrapper[4846]: I1005 08:51:21.344404 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="b0582274-1b4f-4050-979d-053265764498" containerName="registry-server" Oct 05 08:51:21 crc kubenswrapper[4846]: I1005 08:51:21.344641 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="fe44fec2-66d0-4b82-909b-3a27f42340db" containerName="run-os-openstack-openstack-cell1" Oct 05 08:51:21 crc kubenswrapper[4846]: I1005 08:51:21.344668 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="b0582274-1b4f-4050-979d-053265764498" containerName="registry-server" Oct 05 08:51:21 crc kubenswrapper[4846]: I1005 08:51:21.345481 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-openstack-openstack-cell1-fdct8" Oct 05 08:51:21 crc kubenswrapper[4846]: I1005 08:51:21.351867 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-7687s" Oct 05 08:51:21 crc kubenswrapper[4846]: I1005 08:51:21.352598 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Oct 05 08:51:21 crc kubenswrapper[4846]: I1005 08:51:21.352788 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 05 08:51:21 crc kubenswrapper[4846]: I1005 08:51:21.352733 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Oct 05 08:51:21 crc kubenswrapper[4846]: I1005 08:51:21.365871 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-openstack-openstack-cell1-fdct8"] Oct 05 08:51:21 crc kubenswrapper[4846]: I1005 08:51:21.425992 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1a2802cf-4055-4f27-af97-e9dbd5266a3d-ssh-key\") pod \"reboot-os-openstack-openstack-cell1-fdct8\" (UID: \"1a2802cf-4055-4f27-af97-e9dbd5266a3d\") " pod="openstack/reboot-os-openstack-openstack-cell1-fdct8" Oct 05 08:51:21 crc kubenswrapper[4846]: I1005 08:51:21.426510 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1a2802cf-4055-4f27-af97-e9dbd5266a3d-inventory\") pod \"reboot-os-openstack-openstack-cell1-fdct8\" (UID: \"1a2802cf-4055-4f27-af97-e9dbd5266a3d\") " pod="openstack/reboot-os-openstack-openstack-cell1-fdct8" Oct 05 08:51:21 crc kubenswrapper[4846]: I1005 08:51:21.426723 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7wk6d\" (UniqueName: \"kubernetes.io/projected/1a2802cf-4055-4f27-af97-e9dbd5266a3d-kube-api-access-7wk6d\") pod \"reboot-os-openstack-openstack-cell1-fdct8\" (UID: \"1a2802cf-4055-4f27-af97-e9dbd5266a3d\") " pod="openstack/reboot-os-openstack-openstack-cell1-fdct8" Oct 05 08:51:21 crc kubenswrapper[4846]: I1005 08:51:21.528948 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7wk6d\" (UniqueName: \"kubernetes.io/projected/1a2802cf-4055-4f27-af97-e9dbd5266a3d-kube-api-access-7wk6d\") pod \"reboot-os-openstack-openstack-cell1-fdct8\" (UID: \"1a2802cf-4055-4f27-af97-e9dbd5266a3d\") " pod="openstack/reboot-os-openstack-openstack-cell1-fdct8" Oct 05 08:51:21 crc kubenswrapper[4846]: I1005 08:51:21.529041 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1a2802cf-4055-4f27-af97-e9dbd5266a3d-ssh-key\") pod \"reboot-os-openstack-openstack-cell1-fdct8\" (UID: \"1a2802cf-4055-4f27-af97-e9dbd5266a3d\") " pod="openstack/reboot-os-openstack-openstack-cell1-fdct8" Oct 05 08:51:21 crc kubenswrapper[4846]: I1005 08:51:21.529076 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1a2802cf-4055-4f27-af97-e9dbd5266a3d-inventory\") pod \"reboot-os-openstack-openstack-cell1-fdct8\" (UID: \"1a2802cf-4055-4f27-af97-e9dbd5266a3d\") " pod="openstack/reboot-os-openstack-openstack-cell1-fdct8" Oct 05 08:51:21 crc kubenswrapper[4846]: I1005 08:51:21.534081 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1a2802cf-4055-4f27-af97-e9dbd5266a3d-inventory\") pod \"reboot-os-openstack-openstack-cell1-fdct8\" (UID: \"1a2802cf-4055-4f27-af97-e9dbd5266a3d\") " pod="openstack/reboot-os-openstack-openstack-cell1-fdct8" Oct 05 08:51:21 crc kubenswrapper[4846]: I1005 08:51:21.534860 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1a2802cf-4055-4f27-af97-e9dbd5266a3d-ssh-key\") pod \"reboot-os-openstack-openstack-cell1-fdct8\" (UID: \"1a2802cf-4055-4f27-af97-e9dbd5266a3d\") " pod="openstack/reboot-os-openstack-openstack-cell1-fdct8" Oct 05 08:51:21 crc kubenswrapper[4846]: I1005 08:51:21.550086 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7wk6d\" (UniqueName: \"kubernetes.io/projected/1a2802cf-4055-4f27-af97-e9dbd5266a3d-kube-api-access-7wk6d\") pod \"reboot-os-openstack-openstack-cell1-fdct8\" (UID: \"1a2802cf-4055-4f27-af97-e9dbd5266a3d\") " pod="openstack/reboot-os-openstack-openstack-cell1-fdct8" Oct 05 08:51:21 crc kubenswrapper[4846]: I1005 08:51:21.721627 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-openstack-openstack-cell1-fdct8" Oct 05 08:51:22 crc kubenswrapper[4846]: I1005 08:51:22.418880 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-openstack-openstack-cell1-fdct8"] Oct 05 08:51:22 crc kubenswrapper[4846]: W1005 08:51:22.429531 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1a2802cf_4055_4f27_af97_e9dbd5266a3d.slice/crio-e62b1bbb689524d6582f628b1c0e07bb477157e1844e2a5154ab0b47c202982b WatchSource:0}: Error finding container e62b1bbb689524d6582f628b1c0e07bb477157e1844e2a5154ab0b47c202982b: Status 404 returned error can't find the container with id e62b1bbb689524d6582f628b1c0e07bb477157e1844e2a5154ab0b47c202982b Oct 05 08:51:23 crc kubenswrapper[4846]: I1005 08:51:23.256152 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-openstack-openstack-cell1-fdct8" event={"ID":"1a2802cf-4055-4f27-af97-e9dbd5266a3d","Type":"ContainerStarted","Data":"2d2ac1f21ac2934e87c95444ef567461ddd30a52b1fcca836aed5b63d3f55964"} Oct 05 08:51:23 crc kubenswrapper[4846]: I1005 08:51:23.256540 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-openstack-openstack-cell1-fdct8" event={"ID":"1a2802cf-4055-4f27-af97-e9dbd5266a3d","Type":"ContainerStarted","Data":"e62b1bbb689524d6582f628b1c0e07bb477157e1844e2a5154ab0b47c202982b"} Oct 05 08:51:23 crc kubenswrapper[4846]: I1005 08:51:23.275917 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/reboot-os-openstack-openstack-cell1-fdct8" podStartSLOduration=1.818938012 podStartE2EDuration="2.27589679s" podCreationTimestamp="2025-10-05 08:51:21 +0000 UTC" firstStartedPulling="2025-10-05 08:51:22.43357743 +0000 UTC m=+7404.674430215" lastFinishedPulling="2025-10-05 08:51:22.890536208 +0000 UTC m=+7405.131388993" observedRunningTime="2025-10-05 08:51:23.274200465 +0000 UTC m=+7405.515053240" watchObservedRunningTime="2025-10-05 08:51:23.27589679 +0000 UTC m=+7405.516749585" Oct 05 08:51:39 crc kubenswrapper[4846]: I1005 08:51:39.479784 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-openstack-openstack-cell1-fdct8" event={"ID":"1a2802cf-4055-4f27-af97-e9dbd5266a3d","Type":"ContainerDied","Data":"2d2ac1f21ac2934e87c95444ef567461ddd30a52b1fcca836aed5b63d3f55964"} Oct 05 08:51:39 crc kubenswrapper[4846]: I1005 08:51:39.479759 4846 generic.go:334] "Generic (PLEG): container finished" podID="1a2802cf-4055-4f27-af97-e9dbd5266a3d" containerID="2d2ac1f21ac2934e87c95444ef567461ddd30a52b1fcca836aed5b63d3f55964" exitCode=0 Oct 05 08:51:41 crc kubenswrapper[4846]: I1005 08:51:41.094239 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-openstack-openstack-cell1-fdct8" Oct 05 08:51:41 crc kubenswrapper[4846]: I1005 08:51:41.137720 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1a2802cf-4055-4f27-af97-e9dbd5266a3d-ssh-key\") pod \"1a2802cf-4055-4f27-af97-e9dbd5266a3d\" (UID: \"1a2802cf-4055-4f27-af97-e9dbd5266a3d\") " Oct 05 08:51:41 crc kubenswrapper[4846]: I1005 08:51:41.137885 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1a2802cf-4055-4f27-af97-e9dbd5266a3d-inventory\") pod \"1a2802cf-4055-4f27-af97-e9dbd5266a3d\" (UID: \"1a2802cf-4055-4f27-af97-e9dbd5266a3d\") " Oct 05 08:51:41 crc kubenswrapper[4846]: I1005 08:51:41.137980 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7wk6d\" (UniqueName: \"kubernetes.io/projected/1a2802cf-4055-4f27-af97-e9dbd5266a3d-kube-api-access-7wk6d\") pod \"1a2802cf-4055-4f27-af97-e9dbd5266a3d\" (UID: \"1a2802cf-4055-4f27-af97-e9dbd5266a3d\") " Oct 05 08:51:41 crc kubenswrapper[4846]: I1005 08:51:41.145449 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1a2802cf-4055-4f27-af97-e9dbd5266a3d-kube-api-access-7wk6d" (OuterVolumeSpecName: "kube-api-access-7wk6d") pod "1a2802cf-4055-4f27-af97-e9dbd5266a3d" (UID: "1a2802cf-4055-4f27-af97-e9dbd5266a3d"). InnerVolumeSpecName "kube-api-access-7wk6d". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:51:41 crc kubenswrapper[4846]: I1005 08:51:41.170068 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1a2802cf-4055-4f27-af97-e9dbd5266a3d-inventory" (OuterVolumeSpecName: "inventory") pod "1a2802cf-4055-4f27-af97-e9dbd5266a3d" (UID: "1a2802cf-4055-4f27-af97-e9dbd5266a3d"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:51:41 crc kubenswrapper[4846]: I1005 08:51:41.178928 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1a2802cf-4055-4f27-af97-e9dbd5266a3d-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "1a2802cf-4055-4f27-af97-e9dbd5266a3d" (UID: "1a2802cf-4055-4f27-af97-e9dbd5266a3d"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:51:41 crc kubenswrapper[4846]: I1005 08:51:41.239890 4846 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1a2802cf-4055-4f27-af97-e9dbd5266a3d-inventory\") on node \"crc\" DevicePath \"\"" Oct 05 08:51:41 crc kubenswrapper[4846]: I1005 08:51:41.239937 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7wk6d\" (UniqueName: \"kubernetes.io/projected/1a2802cf-4055-4f27-af97-e9dbd5266a3d-kube-api-access-7wk6d\") on node \"crc\" DevicePath \"\"" Oct 05 08:51:41 crc kubenswrapper[4846]: I1005 08:51:41.239951 4846 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1a2802cf-4055-4f27-af97-e9dbd5266a3d-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 05 08:51:41 crc kubenswrapper[4846]: I1005 08:51:41.513561 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-openstack-openstack-cell1-fdct8" event={"ID":"1a2802cf-4055-4f27-af97-e9dbd5266a3d","Type":"ContainerDied","Data":"e62b1bbb689524d6582f628b1c0e07bb477157e1844e2a5154ab0b47c202982b"} Oct 05 08:51:41 crc kubenswrapper[4846]: I1005 08:51:41.513642 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e62b1bbb689524d6582f628b1c0e07bb477157e1844e2a5154ab0b47c202982b" Oct 05 08:51:41 crc kubenswrapper[4846]: I1005 08:51:41.513660 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-openstack-openstack-cell1-fdct8" Oct 05 08:51:41 crc kubenswrapper[4846]: I1005 08:51:41.639169 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-certs-openstack-openstack-cell1-45cc8"] Oct 05 08:51:41 crc kubenswrapper[4846]: E1005 08:51:41.639853 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1a2802cf-4055-4f27-af97-e9dbd5266a3d" containerName="reboot-os-openstack-openstack-cell1" Oct 05 08:51:41 crc kubenswrapper[4846]: I1005 08:51:41.639884 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="1a2802cf-4055-4f27-af97-e9dbd5266a3d" containerName="reboot-os-openstack-openstack-cell1" Oct 05 08:51:41 crc kubenswrapper[4846]: I1005 08:51:41.640350 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="1a2802cf-4055-4f27-af97-e9dbd5266a3d" containerName="reboot-os-openstack-openstack-cell1" Oct 05 08:51:41 crc kubenswrapper[4846]: I1005 08:51:41.641607 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-openstack-openstack-cell1-45cc8" Oct 05 08:51:41 crc kubenswrapper[4846]: I1005 08:51:41.643883 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-telemetry-default-certs-0" Oct 05 08:51:41 crc kubenswrapper[4846]: I1005 08:51:41.645328 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 05 08:51:41 crc kubenswrapper[4846]: I1005 08:51:41.645347 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-7687s" Oct 05 08:51:41 crc kubenswrapper[4846]: I1005 08:51:41.645588 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-neutron-metadata-default-certs-0" Oct 05 08:51:41 crc kubenswrapper[4846]: I1005 08:51:41.645805 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Oct 05 08:51:41 crc kubenswrapper[4846]: I1005 08:51:41.646040 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Oct 05 08:51:41 crc kubenswrapper[4846]: I1005 08:51:41.646101 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-ovn-default-certs-0" Oct 05 08:51:41 crc kubenswrapper[4846]: I1005 08:51:41.646493 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-libvirt-default-certs-0" Oct 05 08:51:41 crc kubenswrapper[4846]: I1005 08:51:41.652317 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-openstack-openstack-cell1-45cc8"] Oct 05 08:51:41 crc kubenswrapper[4846]: I1005 08:51:41.752205 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dda9bfde-287c-49de-aacc-57e6b1ef2950-bootstrap-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-45cc8\" (UID: \"dda9bfde-287c-49de-aacc-57e6b1ef2950\") " pod="openstack/install-certs-openstack-openstack-cell1-45cc8" Oct 05 08:51:41 crc kubenswrapper[4846]: I1005 08:51:41.752279 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-cell1-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/dda9bfde-287c-49de-aacc-57e6b1ef2950-openstack-cell1-libvirt-default-certs-0\") pod \"install-certs-openstack-openstack-cell1-45cc8\" (UID: \"dda9bfde-287c-49de-aacc-57e6b1ef2950\") " pod="openstack/install-certs-openstack-openstack-cell1-45cc8" Oct 05 08:51:41 crc kubenswrapper[4846]: I1005 08:51:41.752321 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dda9bfde-287c-49de-aacc-57e6b1ef2950-neutron-sriov-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-45cc8\" (UID: \"dda9bfde-287c-49de-aacc-57e6b1ef2950\") " pod="openstack/install-certs-openstack-openstack-cell1-45cc8" Oct 05 08:51:41 crc kubenswrapper[4846]: I1005 08:51:41.752359 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-cell1-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/dda9bfde-287c-49de-aacc-57e6b1ef2950-openstack-cell1-telemetry-default-certs-0\") pod \"install-certs-openstack-openstack-cell1-45cc8\" (UID: \"dda9bfde-287c-49de-aacc-57e6b1ef2950\") " pod="openstack/install-certs-openstack-openstack-cell1-45cc8" Oct 05 08:51:41 crc kubenswrapper[4846]: I1005 08:51:41.752434 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-cell1-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/dda9bfde-287c-49de-aacc-57e6b1ef2950-openstack-cell1-neutron-metadata-default-certs-0\") pod \"install-certs-openstack-openstack-cell1-45cc8\" (UID: \"dda9bfde-287c-49de-aacc-57e6b1ef2950\") " pod="openstack/install-certs-openstack-openstack-cell1-45cc8" Oct 05 08:51:41 crc kubenswrapper[4846]: I1005 08:51:41.752463 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dda9bfde-287c-49de-aacc-57e6b1ef2950-ovn-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-45cc8\" (UID: \"dda9bfde-287c-49de-aacc-57e6b1ef2950\") " pod="openstack/install-certs-openstack-openstack-cell1-45cc8" Oct 05 08:51:41 crc kubenswrapper[4846]: I1005 08:51:41.752578 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-cell1-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/dda9bfde-287c-49de-aacc-57e6b1ef2950-openstack-cell1-ovn-default-certs-0\") pod \"install-certs-openstack-openstack-cell1-45cc8\" (UID: \"dda9bfde-287c-49de-aacc-57e6b1ef2950\") " pod="openstack/install-certs-openstack-openstack-cell1-45cc8" Oct 05 08:51:41 crc kubenswrapper[4846]: I1005 08:51:41.752702 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dda9bfde-287c-49de-aacc-57e6b1ef2950-nova-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-45cc8\" (UID: \"dda9bfde-287c-49de-aacc-57e6b1ef2950\") " pod="openstack/install-certs-openstack-openstack-cell1-45cc8" Oct 05 08:51:41 crc kubenswrapper[4846]: I1005 08:51:41.752801 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dda9bfde-287c-49de-aacc-57e6b1ef2950-libvirt-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-45cc8\" (UID: \"dda9bfde-287c-49de-aacc-57e6b1ef2950\") " pod="openstack/install-certs-openstack-openstack-cell1-45cc8" Oct 05 08:51:41 crc kubenswrapper[4846]: I1005 08:51:41.752842 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/dda9bfde-287c-49de-aacc-57e6b1ef2950-ssh-key\") pod \"install-certs-openstack-openstack-cell1-45cc8\" (UID: \"dda9bfde-287c-49de-aacc-57e6b1ef2950\") " pod="openstack/install-certs-openstack-openstack-cell1-45cc8" Oct 05 08:51:41 crc kubenswrapper[4846]: I1005 08:51:41.752980 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-td4q4\" (UniqueName: \"kubernetes.io/projected/dda9bfde-287c-49de-aacc-57e6b1ef2950-kube-api-access-td4q4\") pod \"install-certs-openstack-openstack-cell1-45cc8\" (UID: \"dda9bfde-287c-49de-aacc-57e6b1ef2950\") " pod="openstack/install-certs-openstack-openstack-cell1-45cc8" Oct 05 08:51:41 crc kubenswrapper[4846]: I1005 08:51:41.753041 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dda9bfde-287c-49de-aacc-57e6b1ef2950-neutron-dhcp-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-45cc8\" (UID: \"dda9bfde-287c-49de-aacc-57e6b1ef2950\") " pod="openstack/install-certs-openstack-openstack-cell1-45cc8" Oct 05 08:51:41 crc kubenswrapper[4846]: I1005 08:51:41.753073 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dda9bfde-287c-49de-aacc-57e6b1ef2950-telemetry-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-45cc8\" (UID: \"dda9bfde-287c-49de-aacc-57e6b1ef2950\") " pod="openstack/install-certs-openstack-openstack-cell1-45cc8" Oct 05 08:51:41 crc kubenswrapper[4846]: I1005 08:51:41.753124 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dda9bfde-287c-49de-aacc-57e6b1ef2950-neutron-metadata-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-45cc8\" (UID: \"dda9bfde-287c-49de-aacc-57e6b1ef2950\") " pod="openstack/install-certs-openstack-openstack-cell1-45cc8" Oct 05 08:51:41 crc kubenswrapper[4846]: I1005 08:51:41.753158 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/dda9bfde-287c-49de-aacc-57e6b1ef2950-inventory\") pod \"install-certs-openstack-openstack-cell1-45cc8\" (UID: \"dda9bfde-287c-49de-aacc-57e6b1ef2950\") " pod="openstack/install-certs-openstack-openstack-cell1-45cc8" Oct 05 08:51:41 crc kubenswrapper[4846]: I1005 08:51:41.855881 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dda9bfde-287c-49de-aacc-57e6b1ef2950-libvirt-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-45cc8\" (UID: \"dda9bfde-287c-49de-aacc-57e6b1ef2950\") " pod="openstack/install-certs-openstack-openstack-cell1-45cc8" Oct 05 08:51:41 crc kubenswrapper[4846]: I1005 08:51:41.855973 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/dda9bfde-287c-49de-aacc-57e6b1ef2950-ssh-key\") pod \"install-certs-openstack-openstack-cell1-45cc8\" (UID: \"dda9bfde-287c-49de-aacc-57e6b1ef2950\") " pod="openstack/install-certs-openstack-openstack-cell1-45cc8" Oct 05 08:51:41 crc kubenswrapper[4846]: I1005 08:51:41.856080 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-td4q4\" (UniqueName: \"kubernetes.io/projected/dda9bfde-287c-49de-aacc-57e6b1ef2950-kube-api-access-td4q4\") pod \"install-certs-openstack-openstack-cell1-45cc8\" (UID: \"dda9bfde-287c-49de-aacc-57e6b1ef2950\") " pod="openstack/install-certs-openstack-openstack-cell1-45cc8" Oct 05 08:51:41 crc kubenswrapper[4846]: I1005 08:51:41.856154 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dda9bfde-287c-49de-aacc-57e6b1ef2950-neutron-dhcp-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-45cc8\" (UID: \"dda9bfde-287c-49de-aacc-57e6b1ef2950\") " pod="openstack/install-certs-openstack-openstack-cell1-45cc8" Oct 05 08:51:41 crc kubenswrapper[4846]: I1005 08:51:41.856222 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dda9bfde-287c-49de-aacc-57e6b1ef2950-telemetry-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-45cc8\" (UID: \"dda9bfde-287c-49de-aacc-57e6b1ef2950\") " pod="openstack/install-certs-openstack-openstack-cell1-45cc8" Oct 05 08:51:41 crc kubenswrapper[4846]: I1005 08:51:41.856292 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dda9bfde-287c-49de-aacc-57e6b1ef2950-neutron-metadata-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-45cc8\" (UID: \"dda9bfde-287c-49de-aacc-57e6b1ef2950\") " pod="openstack/install-certs-openstack-openstack-cell1-45cc8" Oct 05 08:51:41 crc kubenswrapper[4846]: I1005 08:51:41.856339 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/dda9bfde-287c-49de-aacc-57e6b1ef2950-inventory\") pod \"install-certs-openstack-openstack-cell1-45cc8\" (UID: \"dda9bfde-287c-49de-aacc-57e6b1ef2950\") " pod="openstack/install-certs-openstack-openstack-cell1-45cc8" Oct 05 08:51:41 crc kubenswrapper[4846]: I1005 08:51:41.856389 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dda9bfde-287c-49de-aacc-57e6b1ef2950-bootstrap-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-45cc8\" (UID: \"dda9bfde-287c-49de-aacc-57e6b1ef2950\") " pod="openstack/install-certs-openstack-openstack-cell1-45cc8" Oct 05 08:51:41 crc kubenswrapper[4846]: I1005 08:51:41.856445 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-cell1-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/dda9bfde-287c-49de-aacc-57e6b1ef2950-openstack-cell1-libvirt-default-certs-0\") pod \"install-certs-openstack-openstack-cell1-45cc8\" (UID: \"dda9bfde-287c-49de-aacc-57e6b1ef2950\") " pod="openstack/install-certs-openstack-openstack-cell1-45cc8" Oct 05 08:51:41 crc kubenswrapper[4846]: I1005 08:51:41.856491 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dda9bfde-287c-49de-aacc-57e6b1ef2950-neutron-sriov-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-45cc8\" (UID: \"dda9bfde-287c-49de-aacc-57e6b1ef2950\") " pod="openstack/install-certs-openstack-openstack-cell1-45cc8" Oct 05 08:51:41 crc kubenswrapper[4846]: I1005 08:51:41.856542 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-cell1-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/dda9bfde-287c-49de-aacc-57e6b1ef2950-openstack-cell1-telemetry-default-certs-0\") pod \"install-certs-openstack-openstack-cell1-45cc8\" (UID: \"dda9bfde-287c-49de-aacc-57e6b1ef2950\") " pod="openstack/install-certs-openstack-openstack-cell1-45cc8" Oct 05 08:51:41 crc kubenswrapper[4846]: I1005 08:51:41.856587 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-cell1-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/dda9bfde-287c-49de-aacc-57e6b1ef2950-openstack-cell1-neutron-metadata-default-certs-0\") pod \"install-certs-openstack-openstack-cell1-45cc8\" (UID: \"dda9bfde-287c-49de-aacc-57e6b1ef2950\") " pod="openstack/install-certs-openstack-openstack-cell1-45cc8" Oct 05 08:51:41 crc kubenswrapper[4846]: I1005 08:51:41.856632 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dda9bfde-287c-49de-aacc-57e6b1ef2950-ovn-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-45cc8\" (UID: \"dda9bfde-287c-49de-aacc-57e6b1ef2950\") " pod="openstack/install-certs-openstack-openstack-cell1-45cc8" Oct 05 08:51:41 crc kubenswrapper[4846]: I1005 08:51:41.856699 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-cell1-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/dda9bfde-287c-49de-aacc-57e6b1ef2950-openstack-cell1-ovn-default-certs-0\") pod \"install-certs-openstack-openstack-cell1-45cc8\" (UID: \"dda9bfde-287c-49de-aacc-57e6b1ef2950\") " pod="openstack/install-certs-openstack-openstack-cell1-45cc8" Oct 05 08:51:41 crc kubenswrapper[4846]: I1005 08:51:41.856757 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dda9bfde-287c-49de-aacc-57e6b1ef2950-nova-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-45cc8\" (UID: \"dda9bfde-287c-49de-aacc-57e6b1ef2950\") " pod="openstack/install-certs-openstack-openstack-cell1-45cc8" Oct 05 08:51:41 crc kubenswrapper[4846]: I1005 08:51:41.863034 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dda9bfde-287c-49de-aacc-57e6b1ef2950-neutron-dhcp-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-45cc8\" (UID: \"dda9bfde-287c-49de-aacc-57e6b1ef2950\") " pod="openstack/install-certs-openstack-openstack-cell1-45cc8" Oct 05 08:51:41 crc kubenswrapper[4846]: I1005 08:51:41.866400 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dda9bfde-287c-49de-aacc-57e6b1ef2950-telemetry-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-45cc8\" (UID: \"dda9bfde-287c-49de-aacc-57e6b1ef2950\") " pod="openstack/install-certs-openstack-openstack-cell1-45cc8" Oct 05 08:51:41 crc kubenswrapper[4846]: I1005 08:51:41.868094 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-cell1-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/dda9bfde-287c-49de-aacc-57e6b1ef2950-openstack-cell1-neutron-metadata-default-certs-0\") pod \"install-certs-openstack-openstack-cell1-45cc8\" (UID: \"dda9bfde-287c-49de-aacc-57e6b1ef2950\") " pod="openstack/install-certs-openstack-openstack-cell1-45cc8" Oct 05 08:51:41 crc kubenswrapper[4846]: I1005 08:51:41.869611 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-cell1-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/dda9bfde-287c-49de-aacc-57e6b1ef2950-openstack-cell1-libvirt-default-certs-0\") pod \"install-certs-openstack-openstack-cell1-45cc8\" (UID: \"dda9bfde-287c-49de-aacc-57e6b1ef2950\") " pod="openstack/install-certs-openstack-openstack-cell1-45cc8" Oct 05 08:51:41 crc kubenswrapper[4846]: I1005 08:51:41.870047 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/dda9bfde-287c-49de-aacc-57e6b1ef2950-ssh-key\") pod \"install-certs-openstack-openstack-cell1-45cc8\" (UID: \"dda9bfde-287c-49de-aacc-57e6b1ef2950\") " pod="openstack/install-certs-openstack-openstack-cell1-45cc8" Oct 05 08:51:41 crc kubenswrapper[4846]: I1005 08:51:41.870973 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dda9bfde-287c-49de-aacc-57e6b1ef2950-neutron-metadata-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-45cc8\" (UID: \"dda9bfde-287c-49de-aacc-57e6b1ef2950\") " pod="openstack/install-certs-openstack-openstack-cell1-45cc8" Oct 05 08:51:41 crc kubenswrapper[4846]: I1005 08:51:41.871832 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-cell1-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/dda9bfde-287c-49de-aacc-57e6b1ef2950-openstack-cell1-telemetry-default-certs-0\") pod \"install-certs-openstack-openstack-cell1-45cc8\" (UID: \"dda9bfde-287c-49de-aacc-57e6b1ef2950\") " pod="openstack/install-certs-openstack-openstack-cell1-45cc8" Oct 05 08:51:41 crc kubenswrapper[4846]: I1005 08:51:41.872302 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dda9bfde-287c-49de-aacc-57e6b1ef2950-ovn-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-45cc8\" (UID: \"dda9bfde-287c-49de-aacc-57e6b1ef2950\") " pod="openstack/install-certs-openstack-openstack-cell1-45cc8" Oct 05 08:51:41 crc kubenswrapper[4846]: I1005 08:51:41.872579 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dda9bfde-287c-49de-aacc-57e6b1ef2950-libvirt-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-45cc8\" (UID: \"dda9bfde-287c-49de-aacc-57e6b1ef2950\") " pod="openstack/install-certs-openstack-openstack-cell1-45cc8" Oct 05 08:51:41 crc kubenswrapper[4846]: I1005 08:51:41.873989 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dda9bfde-287c-49de-aacc-57e6b1ef2950-bootstrap-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-45cc8\" (UID: \"dda9bfde-287c-49de-aacc-57e6b1ef2950\") " pod="openstack/install-certs-openstack-openstack-cell1-45cc8" Oct 05 08:51:41 crc kubenswrapper[4846]: I1005 08:51:41.874921 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dda9bfde-287c-49de-aacc-57e6b1ef2950-nova-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-45cc8\" (UID: \"dda9bfde-287c-49de-aacc-57e6b1ef2950\") " pod="openstack/install-certs-openstack-openstack-cell1-45cc8" Oct 05 08:51:41 crc kubenswrapper[4846]: I1005 08:51:41.876159 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dda9bfde-287c-49de-aacc-57e6b1ef2950-neutron-sriov-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-45cc8\" (UID: \"dda9bfde-287c-49de-aacc-57e6b1ef2950\") " pod="openstack/install-certs-openstack-openstack-cell1-45cc8" Oct 05 08:51:41 crc kubenswrapper[4846]: I1005 08:51:41.877619 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-cell1-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/dda9bfde-287c-49de-aacc-57e6b1ef2950-openstack-cell1-ovn-default-certs-0\") pod \"install-certs-openstack-openstack-cell1-45cc8\" (UID: \"dda9bfde-287c-49de-aacc-57e6b1ef2950\") " pod="openstack/install-certs-openstack-openstack-cell1-45cc8" Oct 05 08:51:41 crc kubenswrapper[4846]: I1005 08:51:41.880277 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/dda9bfde-287c-49de-aacc-57e6b1ef2950-inventory\") pod \"install-certs-openstack-openstack-cell1-45cc8\" (UID: \"dda9bfde-287c-49de-aacc-57e6b1ef2950\") " pod="openstack/install-certs-openstack-openstack-cell1-45cc8" Oct 05 08:51:41 crc kubenswrapper[4846]: I1005 08:51:41.891329 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-td4q4\" (UniqueName: \"kubernetes.io/projected/dda9bfde-287c-49de-aacc-57e6b1ef2950-kube-api-access-td4q4\") pod \"install-certs-openstack-openstack-cell1-45cc8\" (UID: \"dda9bfde-287c-49de-aacc-57e6b1ef2950\") " pod="openstack/install-certs-openstack-openstack-cell1-45cc8" Oct 05 08:51:41 crc kubenswrapper[4846]: I1005 08:51:41.978831 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-openstack-openstack-cell1-45cc8" Oct 05 08:51:42 crc kubenswrapper[4846]: I1005 08:51:42.626759 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-openstack-openstack-cell1-45cc8"] Oct 05 08:51:43 crc kubenswrapper[4846]: I1005 08:51:43.540911 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-openstack-openstack-cell1-45cc8" event={"ID":"dda9bfde-287c-49de-aacc-57e6b1ef2950","Type":"ContainerStarted","Data":"b42a62b6437704f22e26874154c9045c566a89afb9c54adb1858850221c2014d"} Oct 05 08:51:43 crc kubenswrapper[4846]: I1005 08:51:43.541242 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-openstack-openstack-cell1-45cc8" event={"ID":"dda9bfde-287c-49de-aacc-57e6b1ef2950","Type":"ContainerStarted","Data":"10bb94e4dee77f6cc45c020ed31ab089954160769fb4e8d34ee8f099f537da1a"} Oct 05 08:51:43 crc kubenswrapper[4846]: I1005 08:51:43.586519 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-certs-openstack-openstack-cell1-45cc8" podStartSLOduration=2.086531294 podStartE2EDuration="2.586485187s" podCreationTimestamp="2025-10-05 08:51:41 +0000 UTC" firstStartedPulling="2025-10-05 08:51:42.643817325 +0000 UTC m=+7424.884670100" lastFinishedPulling="2025-10-05 08:51:43.143771208 +0000 UTC m=+7425.384623993" observedRunningTime="2025-10-05 08:51:43.572650109 +0000 UTC m=+7425.813502924" watchObservedRunningTime="2025-10-05 08:51:43.586485187 +0000 UTC m=+7425.827338002" Oct 05 08:52:22 crc kubenswrapper[4846]: I1005 08:52:22.996098 4846 generic.go:334] "Generic (PLEG): container finished" podID="dda9bfde-287c-49de-aacc-57e6b1ef2950" containerID="b42a62b6437704f22e26874154c9045c566a89afb9c54adb1858850221c2014d" exitCode=0 Oct 05 08:52:22 crc kubenswrapper[4846]: I1005 08:52:22.996195 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-openstack-openstack-cell1-45cc8" event={"ID":"dda9bfde-287c-49de-aacc-57e6b1ef2950","Type":"ContainerDied","Data":"b42a62b6437704f22e26874154c9045c566a89afb9c54adb1858850221c2014d"} Oct 05 08:52:23 crc kubenswrapper[4846]: I1005 08:52:23.325693 4846 patch_prober.go:28] interesting pod/machine-config-daemon-fscvf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 08:52:23 crc kubenswrapper[4846]: I1005 08:52:23.325757 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 08:52:24 crc kubenswrapper[4846]: I1005 08:52:24.451568 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-openstack-openstack-cell1-45cc8" Oct 05 08:52:24 crc kubenswrapper[4846]: I1005 08:52:24.549519 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/dda9bfde-287c-49de-aacc-57e6b1ef2950-inventory\") pod \"dda9bfde-287c-49de-aacc-57e6b1ef2950\" (UID: \"dda9bfde-287c-49de-aacc-57e6b1ef2950\") " Oct 05 08:52:24 crc kubenswrapper[4846]: I1005 08:52:24.549576 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dda9bfde-287c-49de-aacc-57e6b1ef2950-bootstrap-combined-ca-bundle\") pod \"dda9bfde-287c-49de-aacc-57e6b1ef2950\" (UID: \"dda9bfde-287c-49de-aacc-57e6b1ef2950\") " Oct 05 08:52:24 crc kubenswrapper[4846]: I1005 08:52:24.549595 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-cell1-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/dda9bfde-287c-49de-aacc-57e6b1ef2950-openstack-cell1-ovn-default-certs-0\") pod \"dda9bfde-287c-49de-aacc-57e6b1ef2950\" (UID: \"dda9bfde-287c-49de-aacc-57e6b1ef2950\") " Oct 05 08:52:24 crc kubenswrapper[4846]: I1005 08:52:24.549685 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/dda9bfde-287c-49de-aacc-57e6b1ef2950-ssh-key\") pod \"dda9bfde-287c-49de-aacc-57e6b1ef2950\" (UID: \"dda9bfde-287c-49de-aacc-57e6b1ef2950\") " Oct 05 08:52:24 crc kubenswrapper[4846]: I1005 08:52:24.549709 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dda9bfde-287c-49de-aacc-57e6b1ef2950-nova-combined-ca-bundle\") pod \"dda9bfde-287c-49de-aacc-57e6b1ef2950\" (UID: \"dda9bfde-287c-49de-aacc-57e6b1ef2950\") " Oct 05 08:52:24 crc kubenswrapper[4846]: I1005 08:52:24.549734 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-cell1-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/dda9bfde-287c-49de-aacc-57e6b1ef2950-openstack-cell1-neutron-metadata-default-certs-0\") pod \"dda9bfde-287c-49de-aacc-57e6b1ef2950\" (UID: \"dda9bfde-287c-49de-aacc-57e6b1ef2950\") " Oct 05 08:52:24 crc kubenswrapper[4846]: I1005 08:52:24.549773 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-td4q4\" (UniqueName: \"kubernetes.io/projected/dda9bfde-287c-49de-aacc-57e6b1ef2950-kube-api-access-td4q4\") pod \"dda9bfde-287c-49de-aacc-57e6b1ef2950\" (UID: \"dda9bfde-287c-49de-aacc-57e6b1ef2950\") " Oct 05 08:52:24 crc kubenswrapper[4846]: I1005 08:52:24.549802 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-cell1-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/dda9bfde-287c-49de-aacc-57e6b1ef2950-openstack-cell1-libvirt-default-certs-0\") pod \"dda9bfde-287c-49de-aacc-57e6b1ef2950\" (UID: \"dda9bfde-287c-49de-aacc-57e6b1ef2950\") " Oct 05 08:52:24 crc kubenswrapper[4846]: I1005 08:52:24.549821 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dda9bfde-287c-49de-aacc-57e6b1ef2950-neutron-sriov-combined-ca-bundle\") pod \"dda9bfde-287c-49de-aacc-57e6b1ef2950\" (UID: \"dda9bfde-287c-49de-aacc-57e6b1ef2950\") " Oct 05 08:52:24 crc kubenswrapper[4846]: I1005 08:52:24.549846 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dda9bfde-287c-49de-aacc-57e6b1ef2950-ovn-combined-ca-bundle\") pod \"dda9bfde-287c-49de-aacc-57e6b1ef2950\" (UID: \"dda9bfde-287c-49de-aacc-57e6b1ef2950\") " Oct 05 08:52:24 crc kubenswrapper[4846]: I1005 08:52:24.549862 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dda9bfde-287c-49de-aacc-57e6b1ef2950-libvirt-combined-ca-bundle\") pod \"dda9bfde-287c-49de-aacc-57e6b1ef2950\" (UID: \"dda9bfde-287c-49de-aacc-57e6b1ef2950\") " Oct 05 08:52:24 crc kubenswrapper[4846]: I1005 08:52:24.549937 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dda9bfde-287c-49de-aacc-57e6b1ef2950-telemetry-combined-ca-bundle\") pod \"dda9bfde-287c-49de-aacc-57e6b1ef2950\" (UID: \"dda9bfde-287c-49de-aacc-57e6b1ef2950\") " Oct 05 08:52:24 crc kubenswrapper[4846]: I1005 08:52:24.549964 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dda9bfde-287c-49de-aacc-57e6b1ef2950-neutron-dhcp-combined-ca-bundle\") pod \"dda9bfde-287c-49de-aacc-57e6b1ef2950\" (UID: \"dda9bfde-287c-49de-aacc-57e6b1ef2950\") " Oct 05 08:52:24 crc kubenswrapper[4846]: I1005 08:52:24.550044 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-cell1-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/dda9bfde-287c-49de-aacc-57e6b1ef2950-openstack-cell1-telemetry-default-certs-0\") pod \"dda9bfde-287c-49de-aacc-57e6b1ef2950\" (UID: \"dda9bfde-287c-49de-aacc-57e6b1ef2950\") " Oct 05 08:52:24 crc kubenswrapper[4846]: I1005 08:52:24.550063 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dda9bfde-287c-49de-aacc-57e6b1ef2950-neutron-metadata-combined-ca-bundle\") pod \"dda9bfde-287c-49de-aacc-57e6b1ef2950\" (UID: \"dda9bfde-287c-49de-aacc-57e6b1ef2950\") " Oct 05 08:52:24 crc kubenswrapper[4846]: I1005 08:52:24.556166 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dda9bfde-287c-49de-aacc-57e6b1ef2950-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "dda9bfde-287c-49de-aacc-57e6b1ef2950" (UID: "dda9bfde-287c-49de-aacc-57e6b1ef2950"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:52:24 crc kubenswrapper[4846]: I1005 08:52:24.556160 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dda9bfde-287c-49de-aacc-57e6b1ef2950-neutron-sriov-combined-ca-bundle" (OuterVolumeSpecName: "neutron-sriov-combined-ca-bundle") pod "dda9bfde-287c-49de-aacc-57e6b1ef2950" (UID: "dda9bfde-287c-49de-aacc-57e6b1ef2950"). InnerVolumeSpecName "neutron-sriov-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:52:24 crc kubenswrapper[4846]: I1005 08:52:24.557126 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dda9bfde-287c-49de-aacc-57e6b1ef2950-openstack-cell1-telemetry-default-certs-0" (OuterVolumeSpecName: "openstack-cell1-telemetry-default-certs-0") pod "dda9bfde-287c-49de-aacc-57e6b1ef2950" (UID: "dda9bfde-287c-49de-aacc-57e6b1ef2950"). InnerVolumeSpecName "openstack-cell1-telemetry-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:52:24 crc kubenswrapper[4846]: I1005 08:52:24.559267 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dda9bfde-287c-49de-aacc-57e6b1ef2950-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "dda9bfde-287c-49de-aacc-57e6b1ef2950" (UID: "dda9bfde-287c-49de-aacc-57e6b1ef2950"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:52:24 crc kubenswrapper[4846]: I1005 08:52:24.559540 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dda9bfde-287c-49de-aacc-57e6b1ef2950-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "dda9bfde-287c-49de-aacc-57e6b1ef2950" (UID: "dda9bfde-287c-49de-aacc-57e6b1ef2950"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:52:24 crc kubenswrapper[4846]: I1005 08:52:24.560579 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dda9bfde-287c-49de-aacc-57e6b1ef2950-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "dda9bfde-287c-49de-aacc-57e6b1ef2950" (UID: "dda9bfde-287c-49de-aacc-57e6b1ef2950"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:52:24 crc kubenswrapper[4846]: I1005 08:52:24.561431 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dda9bfde-287c-49de-aacc-57e6b1ef2950-kube-api-access-td4q4" (OuterVolumeSpecName: "kube-api-access-td4q4") pod "dda9bfde-287c-49de-aacc-57e6b1ef2950" (UID: "dda9bfde-287c-49de-aacc-57e6b1ef2950"). InnerVolumeSpecName "kube-api-access-td4q4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:52:24 crc kubenswrapper[4846]: I1005 08:52:24.562128 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dda9bfde-287c-49de-aacc-57e6b1ef2950-neutron-dhcp-combined-ca-bundle" (OuterVolumeSpecName: "neutron-dhcp-combined-ca-bundle") pod "dda9bfde-287c-49de-aacc-57e6b1ef2950" (UID: "dda9bfde-287c-49de-aacc-57e6b1ef2950"). InnerVolumeSpecName "neutron-dhcp-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:52:24 crc kubenswrapper[4846]: I1005 08:52:24.562513 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dda9bfde-287c-49de-aacc-57e6b1ef2950-openstack-cell1-libvirt-default-certs-0" (OuterVolumeSpecName: "openstack-cell1-libvirt-default-certs-0") pod "dda9bfde-287c-49de-aacc-57e6b1ef2950" (UID: "dda9bfde-287c-49de-aacc-57e6b1ef2950"). InnerVolumeSpecName "openstack-cell1-libvirt-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:52:24 crc kubenswrapper[4846]: I1005 08:52:24.563947 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dda9bfde-287c-49de-aacc-57e6b1ef2950-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "dda9bfde-287c-49de-aacc-57e6b1ef2950" (UID: "dda9bfde-287c-49de-aacc-57e6b1ef2950"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:52:24 crc kubenswrapper[4846]: I1005 08:52:24.566700 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dda9bfde-287c-49de-aacc-57e6b1ef2950-openstack-cell1-ovn-default-certs-0" (OuterVolumeSpecName: "openstack-cell1-ovn-default-certs-0") pod "dda9bfde-287c-49de-aacc-57e6b1ef2950" (UID: "dda9bfde-287c-49de-aacc-57e6b1ef2950"). InnerVolumeSpecName "openstack-cell1-ovn-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:52:24 crc kubenswrapper[4846]: I1005 08:52:24.568814 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dda9bfde-287c-49de-aacc-57e6b1ef2950-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "dda9bfde-287c-49de-aacc-57e6b1ef2950" (UID: "dda9bfde-287c-49de-aacc-57e6b1ef2950"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:52:24 crc kubenswrapper[4846]: I1005 08:52:24.569413 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dda9bfde-287c-49de-aacc-57e6b1ef2950-openstack-cell1-neutron-metadata-default-certs-0" (OuterVolumeSpecName: "openstack-cell1-neutron-metadata-default-certs-0") pod "dda9bfde-287c-49de-aacc-57e6b1ef2950" (UID: "dda9bfde-287c-49de-aacc-57e6b1ef2950"). InnerVolumeSpecName "openstack-cell1-neutron-metadata-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:52:24 crc kubenswrapper[4846]: I1005 08:52:24.592905 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dda9bfde-287c-49de-aacc-57e6b1ef2950-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "dda9bfde-287c-49de-aacc-57e6b1ef2950" (UID: "dda9bfde-287c-49de-aacc-57e6b1ef2950"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:52:24 crc kubenswrapper[4846]: I1005 08:52:24.594405 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dda9bfde-287c-49de-aacc-57e6b1ef2950-inventory" (OuterVolumeSpecName: "inventory") pod "dda9bfde-287c-49de-aacc-57e6b1ef2950" (UID: "dda9bfde-287c-49de-aacc-57e6b1ef2950"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:52:24 crc kubenswrapper[4846]: I1005 08:52:24.654106 4846 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/dda9bfde-287c-49de-aacc-57e6b1ef2950-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 05 08:52:24 crc kubenswrapper[4846]: I1005 08:52:24.654141 4846 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dda9bfde-287c-49de-aacc-57e6b1ef2950-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 08:52:24 crc kubenswrapper[4846]: I1005 08:52:24.654154 4846 reconciler_common.go:293] "Volume detached for volume \"openstack-cell1-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/dda9bfde-287c-49de-aacc-57e6b1ef2950-openstack-cell1-neutron-metadata-default-certs-0\") on node \"crc\" DevicePath \"\"" Oct 05 08:52:24 crc kubenswrapper[4846]: I1005 08:52:24.654192 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-td4q4\" (UniqueName: \"kubernetes.io/projected/dda9bfde-287c-49de-aacc-57e6b1ef2950-kube-api-access-td4q4\") on node \"crc\" DevicePath \"\"" Oct 05 08:52:24 crc kubenswrapper[4846]: I1005 08:52:24.654208 4846 reconciler_common.go:293] "Volume detached for volume \"openstack-cell1-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/dda9bfde-287c-49de-aacc-57e6b1ef2950-openstack-cell1-libvirt-default-certs-0\") on node \"crc\" DevicePath \"\"" Oct 05 08:52:24 crc kubenswrapper[4846]: I1005 08:52:24.654224 4846 reconciler_common.go:293] "Volume detached for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dda9bfde-287c-49de-aacc-57e6b1ef2950-neutron-sriov-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 08:52:24 crc kubenswrapper[4846]: I1005 08:52:24.654235 4846 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dda9bfde-287c-49de-aacc-57e6b1ef2950-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 08:52:24 crc kubenswrapper[4846]: I1005 08:52:24.654265 4846 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dda9bfde-287c-49de-aacc-57e6b1ef2950-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 08:52:24 crc kubenswrapper[4846]: I1005 08:52:24.654276 4846 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dda9bfde-287c-49de-aacc-57e6b1ef2950-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 08:52:24 crc kubenswrapper[4846]: I1005 08:52:24.654285 4846 reconciler_common.go:293] "Volume detached for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dda9bfde-287c-49de-aacc-57e6b1ef2950-neutron-dhcp-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 08:52:24 crc kubenswrapper[4846]: I1005 08:52:24.654294 4846 reconciler_common.go:293] "Volume detached for volume \"openstack-cell1-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/dda9bfde-287c-49de-aacc-57e6b1ef2950-openstack-cell1-telemetry-default-certs-0\") on node \"crc\" DevicePath \"\"" Oct 05 08:52:24 crc kubenswrapper[4846]: I1005 08:52:24.654306 4846 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dda9bfde-287c-49de-aacc-57e6b1ef2950-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 08:52:24 crc kubenswrapper[4846]: I1005 08:52:24.654318 4846 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/dda9bfde-287c-49de-aacc-57e6b1ef2950-inventory\") on node \"crc\" DevicePath \"\"" Oct 05 08:52:24 crc kubenswrapper[4846]: I1005 08:52:24.654332 4846 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dda9bfde-287c-49de-aacc-57e6b1ef2950-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 08:52:24 crc kubenswrapper[4846]: I1005 08:52:24.654361 4846 reconciler_common.go:293] "Volume detached for volume \"openstack-cell1-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/dda9bfde-287c-49de-aacc-57e6b1ef2950-openstack-cell1-ovn-default-certs-0\") on node \"crc\" DevicePath \"\"" Oct 05 08:52:25 crc kubenswrapper[4846]: I1005 08:52:25.021746 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-openstack-openstack-cell1-45cc8" event={"ID":"dda9bfde-287c-49de-aacc-57e6b1ef2950","Type":"ContainerDied","Data":"10bb94e4dee77f6cc45c020ed31ab089954160769fb4e8d34ee8f099f537da1a"} Oct 05 08:52:25 crc kubenswrapper[4846]: I1005 08:52:25.022332 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="10bb94e4dee77f6cc45c020ed31ab089954160769fb4e8d34ee8f099f537da1a" Oct 05 08:52:25 crc kubenswrapper[4846]: I1005 08:52:25.022005 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-openstack-openstack-cell1-45cc8" Oct 05 08:52:25 crc kubenswrapper[4846]: I1005 08:52:25.150619 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-openstack-openstack-cell1-5bb6g"] Oct 05 08:52:25 crc kubenswrapper[4846]: E1005 08:52:25.151359 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dda9bfde-287c-49de-aacc-57e6b1ef2950" containerName="install-certs-openstack-openstack-cell1" Oct 05 08:52:25 crc kubenswrapper[4846]: I1005 08:52:25.151432 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="dda9bfde-287c-49de-aacc-57e6b1ef2950" containerName="install-certs-openstack-openstack-cell1" Oct 05 08:52:25 crc kubenswrapper[4846]: I1005 08:52:25.151705 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="dda9bfde-287c-49de-aacc-57e6b1ef2950" containerName="install-certs-openstack-openstack-cell1" Oct 05 08:52:25 crc kubenswrapper[4846]: I1005 08:52:25.152523 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-openstack-openstack-cell1-5bb6g" Oct 05 08:52:25 crc kubenswrapper[4846]: I1005 08:52:25.156316 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-config" Oct 05 08:52:25 crc kubenswrapper[4846]: I1005 08:52:25.156413 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Oct 05 08:52:25 crc kubenswrapper[4846]: I1005 08:52:25.156422 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-7687s" Oct 05 08:52:25 crc kubenswrapper[4846]: I1005 08:52:25.156559 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Oct 05 08:52:25 crc kubenswrapper[4846]: I1005 08:52:25.156642 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 05 08:52:25 crc kubenswrapper[4846]: I1005 08:52:25.176615 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-openstack-openstack-cell1-5bb6g"] Oct 05 08:52:25 crc kubenswrapper[4846]: I1005 08:52:25.269314 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b6c4e5a0-3410-406d-bb32-6046cf340cf1-inventory\") pod \"ovn-openstack-openstack-cell1-5bb6g\" (UID: \"b6c4e5a0-3410-406d-bb32-6046cf340cf1\") " pod="openstack/ovn-openstack-openstack-cell1-5bb6g" Oct 05 08:52:25 crc kubenswrapper[4846]: I1005 08:52:25.269400 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b6c4e5a0-3410-406d-bb32-6046cf340cf1-ssh-key\") pod \"ovn-openstack-openstack-cell1-5bb6g\" (UID: \"b6c4e5a0-3410-406d-bb32-6046cf340cf1\") " pod="openstack/ovn-openstack-openstack-cell1-5bb6g" Oct 05 08:52:25 crc kubenswrapper[4846]: I1005 08:52:25.269479 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h6qjw\" (UniqueName: \"kubernetes.io/projected/b6c4e5a0-3410-406d-bb32-6046cf340cf1-kube-api-access-h6qjw\") pod \"ovn-openstack-openstack-cell1-5bb6g\" (UID: \"b6c4e5a0-3410-406d-bb32-6046cf340cf1\") " pod="openstack/ovn-openstack-openstack-cell1-5bb6g" Oct 05 08:52:25 crc kubenswrapper[4846]: I1005 08:52:25.269516 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b6c4e5a0-3410-406d-bb32-6046cf340cf1-ovn-combined-ca-bundle\") pod \"ovn-openstack-openstack-cell1-5bb6g\" (UID: \"b6c4e5a0-3410-406d-bb32-6046cf340cf1\") " pod="openstack/ovn-openstack-openstack-cell1-5bb6g" Oct 05 08:52:25 crc kubenswrapper[4846]: I1005 08:52:25.269580 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/b6c4e5a0-3410-406d-bb32-6046cf340cf1-ovncontroller-config-0\") pod \"ovn-openstack-openstack-cell1-5bb6g\" (UID: \"b6c4e5a0-3410-406d-bb32-6046cf340cf1\") " pod="openstack/ovn-openstack-openstack-cell1-5bb6g" Oct 05 08:52:25 crc kubenswrapper[4846]: I1005 08:52:25.371395 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/b6c4e5a0-3410-406d-bb32-6046cf340cf1-ovncontroller-config-0\") pod \"ovn-openstack-openstack-cell1-5bb6g\" (UID: \"b6c4e5a0-3410-406d-bb32-6046cf340cf1\") " pod="openstack/ovn-openstack-openstack-cell1-5bb6g" Oct 05 08:52:25 crc kubenswrapper[4846]: I1005 08:52:25.371550 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b6c4e5a0-3410-406d-bb32-6046cf340cf1-inventory\") pod \"ovn-openstack-openstack-cell1-5bb6g\" (UID: \"b6c4e5a0-3410-406d-bb32-6046cf340cf1\") " pod="openstack/ovn-openstack-openstack-cell1-5bb6g" Oct 05 08:52:25 crc kubenswrapper[4846]: I1005 08:52:25.371606 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b6c4e5a0-3410-406d-bb32-6046cf340cf1-ssh-key\") pod \"ovn-openstack-openstack-cell1-5bb6g\" (UID: \"b6c4e5a0-3410-406d-bb32-6046cf340cf1\") " pod="openstack/ovn-openstack-openstack-cell1-5bb6g" Oct 05 08:52:25 crc kubenswrapper[4846]: I1005 08:52:25.371701 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h6qjw\" (UniqueName: \"kubernetes.io/projected/b6c4e5a0-3410-406d-bb32-6046cf340cf1-kube-api-access-h6qjw\") pod \"ovn-openstack-openstack-cell1-5bb6g\" (UID: \"b6c4e5a0-3410-406d-bb32-6046cf340cf1\") " pod="openstack/ovn-openstack-openstack-cell1-5bb6g" Oct 05 08:52:25 crc kubenswrapper[4846]: I1005 08:52:25.372398 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b6c4e5a0-3410-406d-bb32-6046cf340cf1-ovn-combined-ca-bundle\") pod \"ovn-openstack-openstack-cell1-5bb6g\" (UID: \"b6c4e5a0-3410-406d-bb32-6046cf340cf1\") " pod="openstack/ovn-openstack-openstack-cell1-5bb6g" Oct 05 08:52:25 crc kubenswrapper[4846]: I1005 08:52:25.374406 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/b6c4e5a0-3410-406d-bb32-6046cf340cf1-ovncontroller-config-0\") pod \"ovn-openstack-openstack-cell1-5bb6g\" (UID: \"b6c4e5a0-3410-406d-bb32-6046cf340cf1\") " pod="openstack/ovn-openstack-openstack-cell1-5bb6g" Oct 05 08:52:25 crc kubenswrapper[4846]: I1005 08:52:25.376823 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b6c4e5a0-3410-406d-bb32-6046cf340cf1-ssh-key\") pod \"ovn-openstack-openstack-cell1-5bb6g\" (UID: \"b6c4e5a0-3410-406d-bb32-6046cf340cf1\") " pod="openstack/ovn-openstack-openstack-cell1-5bb6g" Oct 05 08:52:25 crc kubenswrapper[4846]: I1005 08:52:25.377209 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b6c4e5a0-3410-406d-bb32-6046cf340cf1-inventory\") pod \"ovn-openstack-openstack-cell1-5bb6g\" (UID: \"b6c4e5a0-3410-406d-bb32-6046cf340cf1\") " pod="openstack/ovn-openstack-openstack-cell1-5bb6g" Oct 05 08:52:25 crc kubenswrapper[4846]: I1005 08:52:25.377952 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b6c4e5a0-3410-406d-bb32-6046cf340cf1-ovn-combined-ca-bundle\") pod \"ovn-openstack-openstack-cell1-5bb6g\" (UID: \"b6c4e5a0-3410-406d-bb32-6046cf340cf1\") " pod="openstack/ovn-openstack-openstack-cell1-5bb6g" Oct 05 08:52:25 crc kubenswrapper[4846]: I1005 08:52:25.395262 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h6qjw\" (UniqueName: \"kubernetes.io/projected/b6c4e5a0-3410-406d-bb32-6046cf340cf1-kube-api-access-h6qjw\") pod \"ovn-openstack-openstack-cell1-5bb6g\" (UID: \"b6c4e5a0-3410-406d-bb32-6046cf340cf1\") " pod="openstack/ovn-openstack-openstack-cell1-5bb6g" Oct 05 08:52:25 crc kubenswrapper[4846]: I1005 08:52:25.483643 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-openstack-openstack-cell1-5bb6g" Oct 05 08:52:26 crc kubenswrapper[4846]: I1005 08:52:26.130360 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-openstack-openstack-cell1-5bb6g"] Oct 05 08:52:27 crc kubenswrapper[4846]: I1005 08:52:27.042124 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-openstack-openstack-cell1-5bb6g" event={"ID":"b6c4e5a0-3410-406d-bb32-6046cf340cf1","Type":"ContainerStarted","Data":"5984f33935fe0ca704de0bb556ac93a0ed3e54871c840702e0524caf4127a039"} Oct 05 08:52:27 crc kubenswrapper[4846]: I1005 08:52:27.042488 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-openstack-openstack-cell1-5bb6g" event={"ID":"b6c4e5a0-3410-406d-bb32-6046cf340cf1","Type":"ContainerStarted","Data":"0656f519ea78d2466e5b195e4beb940d1b6daa5847cb3e2b27ab21efaad20a81"} Oct 05 08:52:27 crc kubenswrapper[4846]: I1005 08:52:27.068593 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-openstack-openstack-cell1-5bb6g" podStartSLOduration=1.5728804589999998 podStartE2EDuration="2.068574859s" podCreationTimestamp="2025-10-05 08:52:25 +0000 UTC" firstStartedPulling="2025-10-05 08:52:26.135655136 +0000 UTC m=+7468.376507911" lastFinishedPulling="2025-10-05 08:52:26.631349536 +0000 UTC m=+7468.872202311" observedRunningTime="2025-10-05 08:52:27.060801952 +0000 UTC m=+7469.301654727" watchObservedRunningTime="2025-10-05 08:52:27.068574859 +0000 UTC m=+7469.309427634" Oct 05 08:52:53 crc kubenswrapper[4846]: I1005 08:52:53.325689 4846 patch_prober.go:28] interesting pod/machine-config-daemon-fscvf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 08:52:53 crc kubenswrapper[4846]: I1005 08:52:53.327420 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 08:53:23 crc kubenswrapper[4846]: I1005 08:53:23.325209 4846 patch_prober.go:28] interesting pod/machine-config-daemon-fscvf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 08:53:23 crc kubenswrapper[4846]: I1005 08:53:23.325768 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 08:53:23 crc kubenswrapper[4846]: I1005 08:53:23.325820 4846 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" Oct 05 08:53:23 crc kubenswrapper[4846]: I1005 08:53:23.326712 4846 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"eb023100181ea6025e1663017d08951f92365161214cdfbcbfc8a62d71858fc9"} pod="openshift-machine-config-operator/machine-config-daemon-fscvf" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 05 08:53:23 crc kubenswrapper[4846]: I1005 08:53:23.326775 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" containerName="machine-config-daemon" containerID="cri-o://eb023100181ea6025e1663017d08951f92365161214cdfbcbfc8a62d71858fc9" gracePeriod=600 Oct 05 08:53:23 crc kubenswrapper[4846]: I1005 08:53:23.713752 4846 generic.go:334] "Generic (PLEG): container finished" podID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" containerID="eb023100181ea6025e1663017d08951f92365161214cdfbcbfc8a62d71858fc9" exitCode=0 Oct 05 08:53:23 crc kubenswrapper[4846]: I1005 08:53:23.713815 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" event={"ID":"ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7","Type":"ContainerDied","Data":"eb023100181ea6025e1663017d08951f92365161214cdfbcbfc8a62d71858fc9"} Oct 05 08:53:23 crc kubenswrapper[4846]: I1005 08:53:23.714895 4846 scope.go:117] "RemoveContainer" containerID="0d84b906773347b38ec2ce550f580254430238b4374f3171cb7453fa65259bcc" Oct 05 08:53:24 crc kubenswrapper[4846]: I1005 08:53:24.729290 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" event={"ID":"ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7","Type":"ContainerStarted","Data":"e6df7a4d2c7027b882a7e838679edd7570422c9a8553ab74b0bddbf3e4bf4b45"} Oct 05 08:53:34 crc kubenswrapper[4846]: I1005 08:53:34.856835 4846 generic.go:334] "Generic (PLEG): container finished" podID="b6c4e5a0-3410-406d-bb32-6046cf340cf1" containerID="5984f33935fe0ca704de0bb556ac93a0ed3e54871c840702e0524caf4127a039" exitCode=0 Oct 05 08:53:34 crc kubenswrapper[4846]: I1005 08:53:34.856965 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-openstack-openstack-cell1-5bb6g" event={"ID":"b6c4e5a0-3410-406d-bb32-6046cf340cf1","Type":"ContainerDied","Data":"5984f33935fe0ca704de0bb556ac93a0ed3e54871c840702e0524caf4127a039"} Oct 05 08:53:36 crc kubenswrapper[4846]: I1005 08:53:36.366439 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-openstack-openstack-cell1-5bb6g" Oct 05 08:53:36 crc kubenswrapper[4846]: I1005 08:53:36.461363 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h6qjw\" (UniqueName: \"kubernetes.io/projected/b6c4e5a0-3410-406d-bb32-6046cf340cf1-kube-api-access-h6qjw\") pod \"b6c4e5a0-3410-406d-bb32-6046cf340cf1\" (UID: \"b6c4e5a0-3410-406d-bb32-6046cf340cf1\") " Oct 05 08:53:36 crc kubenswrapper[4846]: I1005 08:53:36.461495 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b6c4e5a0-3410-406d-bb32-6046cf340cf1-inventory\") pod \"b6c4e5a0-3410-406d-bb32-6046cf340cf1\" (UID: \"b6c4e5a0-3410-406d-bb32-6046cf340cf1\") " Oct 05 08:53:36 crc kubenswrapper[4846]: I1005 08:53:36.461556 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b6c4e5a0-3410-406d-bb32-6046cf340cf1-ssh-key\") pod \"b6c4e5a0-3410-406d-bb32-6046cf340cf1\" (UID: \"b6c4e5a0-3410-406d-bb32-6046cf340cf1\") " Oct 05 08:53:36 crc kubenswrapper[4846]: I1005 08:53:36.461585 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/b6c4e5a0-3410-406d-bb32-6046cf340cf1-ovncontroller-config-0\") pod \"b6c4e5a0-3410-406d-bb32-6046cf340cf1\" (UID: \"b6c4e5a0-3410-406d-bb32-6046cf340cf1\") " Oct 05 08:53:36 crc kubenswrapper[4846]: I1005 08:53:36.461612 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b6c4e5a0-3410-406d-bb32-6046cf340cf1-ovn-combined-ca-bundle\") pod \"b6c4e5a0-3410-406d-bb32-6046cf340cf1\" (UID: \"b6c4e5a0-3410-406d-bb32-6046cf340cf1\") " Oct 05 08:53:36 crc kubenswrapper[4846]: I1005 08:53:36.467383 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6c4e5a0-3410-406d-bb32-6046cf340cf1-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "b6c4e5a0-3410-406d-bb32-6046cf340cf1" (UID: "b6c4e5a0-3410-406d-bb32-6046cf340cf1"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:53:36 crc kubenswrapper[4846]: I1005 08:53:36.475567 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6c4e5a0-3410-406d-bb32-6046cf340cf1-kube-api-access-h6qjw" (OuterVolumeSpecName: "kube-api-access-h6qjw") pod "b6c4e5a0-3410-406d-bb32-6046cf340cf1" (UID: "b6c4e5a0-3410-406d-bb32-6046cf340cf1"). InnerVolumeSpecName "kube-api-access-h6qjw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:53:36 crc kubenswrapper[4846]: I1005 08:53:36.490884 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6c4e5a0-3410-406d-bb32-6046cf340cf1-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "b6c4e5a0-3410-406d-bb32-6046cf340cf1" (UID: "b6c4e5a0-3410-406d-bb32-6046cf340cf1"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:53:36 crc kubenswrapper[4846]: I1005 08:53:36.491245 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6c4e5a0-3410-406d-bb32-6046cf340cf1-ovncontroller-config-0" (OuterVolumeSpecName: "ovncontroller-config-0") pod "b6c4e5a0-3410-406d-bb32-6046cf340cf1" (UID: "b6c4e5a0-3410-406d-bb32-6046cf340cf1"). InnerVolumeSpecName "ovncontroller-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 08:53:36 crc kubenswrapper[4846]: I1005 08:53:36.494142 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6c4e5a0-3410-406d-bb32-6046cf340cf1-inventory" (OuterVolumeSpecName: "inventory") pod "b6c4e5a0-3410-406d-bb32-6046cf340cf1" (UID: "b6c4e5a0-3410-406d-bb32-6046cf340cf1"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:53:36 crc kubenswrapper[4846]: I1005 08:53:36.563840 4846 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b6c4e5a0-3410-406d-bb32-6046cf340cf1-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 05 08:53:36 crc kubenswrapper[4846]: I1005 08:53:36.563879 4846 reconciler_common.go:293] "Volume detached for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/b6c4e5a0-3410-406d-bb32-6046cf340cf1-ovncontroller-config-0\") on node \"crc\" DevicePath \"\"" Oct 05 08:53:36 crc kubenswrapper[4846]: I1005 08:53:36.563893 4846 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b6c4e5a0-3410-406d-bb32-6046cf340cf1-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 08:53:36 crc kubenswrapper[4846]: I1005 08:53:36.563905 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h6qjw\" (UniqueName: \"kubernetes.io/projected/b6c4e5a0-3410-406d-bb32-6046cf340cf1-kube-api-access-h6qjw\") on node \"crc\" DevicePath \"\"" Oct 05 08:53:36 crc kubenswrapper[4846]: I1005 08:53:36.563916 4846 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b6c4e5a0-3410-406d-bb32-6046cf340cf1-inventory\") on node \"crc\" DevicePath \"\"" Oct 05 08:53:36 crc kubenswrapper[4846]: I1005 08:53:36.890573 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-openstack-openstack-cell1-5bb6g" event={"ID":"b6c4e5a0-3410-406d-bb32-6046cf340cf1","Type":"ContainerDied","Data":"0656f519ea78d2466e5b195e4beb940d1b6daa5847cb3e2b27ab21efaad20a81"} Oct 05 08:53:36 crc kubenswrapper[4846]: I1005 08:53:36.891026 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0656f519ea78d2466e5b195e4beb940d1b6daa5847cb3e2b27ab21efaad20a81" Oct 05 08:53:36 crc kubenswrapper[4846]: I1005 08:53:36.890682 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-openstack-openstack-cell1-5bb6g" Oct 05 08:53:36 crc kubenswrapper[4846]: I1005 08:53:36.994677 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-metadata-openstack-openstack-cell1-xnhm4"] Oct 05 08:53:36 crc kubenswrapper[4846]: E1005 08:53:36.995605 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b6c4e5a0-3410-406d-bb32-6046cf340cf1" containerName="ovn-openstack-openstack-cell1" Oct 05 08:53:36 crc kubenswrapper[4846]: I1005 08:53:36.995763 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="b6c4e5a0-3410-406d-bb32-6046cf340cf1" containerName="ovn-openstack-openstack-cell1" Oct 05 08:53:36 crc kubenswrapper[4846]: I1005 08:53:36.996301 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="b6c4e5a0-3410-406d-bb32-6046cf340cf1" containerName="ovn-openstack-openstack-cell1" Oct 05 08:53:36 crc kubenswrapper[4846]: I1005 08:53:36.997547 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-openstack-openstack-cell1-xnhm4" Oct 05 08:53:37 crc kubenswrapper[4846]: I1005 08:53:37.001118 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 05 08:53:37 crc kubenswrapper[4846]: I1005 08:53:37.001127 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Oct 05 08:53:37 crc kubenswrapper[4846]: I1005 08:53:37.001257 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-neutron-config" Oct 05 08:53:37 crc kubenswrapper[4846]: I1005 08:53:37.001700 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Oct 05 08:53:37 crc kubenswrapper[4846]: I1005 08:53:37.001988 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-7687s" Oct 05 08:53:37 crc kubenswrapper[4846]: I1005 08:53:37.002084 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-ovn-metadata-agent-neutron-config" Oct 05 08:53:37 crc kubenswrapper[4846]: I1005 08:53:37.003442 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-openstack-openstack-cell1-xnhm4"] Oct 05 08:53:37 crc kubenswrapper[4846]: I1005 08:53:37.073794 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5gz2d\" (UniqueName: \"kubernetes.io/projected/22023b6c-2796-4c84-96ad-105022a558fb-kube-api-access-5gz2d\") pod \"neutron-metadata-openstack-openstack-cell1-xnhm4\" (UID: \"22023b6c-2796-4c84-96ad-105022a558fb\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-xnhm4" Oct 05 08:53:37 crc kubenswrapper[4846]: I1005 08:53:37.074040 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/22023b6c-2796-4c84-96ad-105022a558fb-ssh-key\") pod \"neutron-metadata-openstack-openstack-cell1-xnhm4\" (UID: \"22023b6c-2796-4c84-96ad-105022a558fb\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-xnhm4" Oct 05 08:53:37 crc kubenswrapper[4846]: I1005 08:53:37.074097 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/22023b6c-2796-4c84-96ad-105022a558fb-nova-metadata-neutron-config-0\") pod \"neutron-metadata-openstack-openstack-cell1-xnhm4\" (UID: \"22023b6c-2796-4c84-96ad-105022a558fb\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-xnhm4" Oct 05 08:53:37 crc kubenswrapper[4846]: I1005 08:53:37.074161 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/22023b6c-2796-4c84-96ad-105022a558fb-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-openstack-openstack-cell1-xnhm4\" (UID: \"22023b6c-2796-4c84-96ad-105022a558fb\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-xnhm4" Oct 05 08:53:37 crc kubenswrapper[4846]: I1005 08:53:37.074246 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/22023b6c-2796-4c84-96ad-105022a558fb-inventory\") pod \"neutron-metadata-openstack-openstack-cell1-xnhm4\" (UID: \"22023b6c-2796-4c84-96ad-105022a558fb\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-xnhm4" Oct 05 08:53:37 crc kubenswrapper[4846]: I1005 08:53:37.074290 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/22023b6c-2796-4c84-96ad-105022a558fb-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-openstack-openstack-cell1-xnhm4\" (UID: \"22023b6c-2796-4c84-96ad-105022a558fb\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-xnhm4" Oct 05 08:53:37 crc kubenswrapper[4846]: I1005 08:53:37.176282 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/22023b6c-2796-4c84-96ad-105022a558fb-ssh-key\") pod \"neutron-metadata-openstack-openstack-cell1-xnhm4\" (UID: \"22023b6c-2796-4c84-96ad-105022a558fb\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-xnhm4" Oct 05 08:53:37 crc kubenswrapper[4846]: I1005 08:53:37.176347 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/22023b6c-2796-4c84-96ad-105022a558fb-nova-metadata-neutron-config-0\") pod \"neutron-metadata-openstack-openstack-cell1-xnhm4\" (UID: \"22023b6c-2796-4c84-96ad-105022a558fb\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-xnhm4" Oct 05 08:53:37 crc kubenswrapper[4846]: I1005 08:53:37.176392 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/22023b6c-2796-4c84-96ad-105022a558fb-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-openstack-openstack-cell1-xnhm4\" (UID: \"22023b6c-2796-4c84-96ad-105022a558fb\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-xnhm4" Oct 05 08:53:37 crc kubenswrapper[4846]: I1005 08:53:37.176429 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/22023b6c-2796-4c84-96ad-105022a558fb-inventory\") pod \"neutron-metadata-openstack-openstack-cell1-xnhm4\" (UID: \"22023b6c-2796-4c84-96ad-105022a558fb\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-xnhm4" Oct 05 08:53:37 crc kubenswrapper[4846]: I1005 08:53:37.176466 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/22023b6c-2796-4c84-96ad-105022a558fb-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-openstack-openstack-cell1-xnhm4\" (UID: \"22023b6c-2796-4c84-96ad-105022a558fb\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-xnhm4" Oct 05 08:53:37 crc kubenswrapper[4846]: I1005 08:53:37.176547 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5gz2d\" (UniqueName: \"kubernetes.io/projected/22023b6c-2796-4c84-96ad-105022a558fb-kube-api-access-5gz2d\") pod \"neutron-metadata-openstack-openstack-cell1-xnhm4\" (UID: \"22023b6c-2796-4c84-96ad-105022a558fb\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-xnhm4" Oct 05 08:53:37 crc kubenswrapper[4846]: I1005 08:53:37.181637 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/22023b6c-2796-4c84-96ad-105022a558fb-inventory\") pod \"neutron-metadata-openstack-openstack-cell1-xnhm4\" (UID: \"22023b6c-2796-4c84-96ad-105022a558fb\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-xnhm4" Oct 05 08:53:37 crc kubenswrapper[4846]: I1005 08:53:37.181673 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/22023b6c-2796-4c84-96ad-105022a558fb-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-openstack-openstack-cell1-xnhm4\" (UID: \"22023b6c-2796-4c84-96ad-105022a558fb\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-xnhm4" Oct 05 08:53:37 crc kubenswrapper[4846]: I1005 08:53:37.181678 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/22023b6c-2796-4c84-96ad-105022a558fb-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-openstack-openstack-cell1-xnhm4\" (UID: \"22023b6c-2796-4c84-96ad-105022a558fb\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-xnhm4" Oct 05 08:53:37 crc kubenswrapper[4846]: I1005 08:53:37.181737 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/22023b6c-2796-4c84-96ad-105022a558fb-ssh-key\") pod \"neutron-metadata-openstack-openstack-cell1-xnhm4\" (UID: \"22023b6c-2796-4c84-96ad-105022a558fb\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-xnhm4" Oct 05 08:53:37 crc kubenswrapper[4846]: I1005 08:53:37.182684 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/22023b6c-2796-4c84-96ad-105022a558fb-nova-metadata-neutron-config-0\") pod \"neutron-metadata-openstack-openstack-cell1-xnhm4\" (UID: \"22023b6c-2796-4c84-96ad-105022a558fb\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-xnhm4" Oct 05 08:53:37 crc kubenswrapper[4846]: I1005 08:53:37.195852 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5gz2d\" (UniqueName: \"kubernetes.io/projected/22023b6c-2796-4c84-96ad-105022a558fb-kube-api-access-5gz2d\") pod \"neutron-metadata-openstack-openstack-cell1-xnhm4\" (UID: \"22023b6c-2796-4c84-96ad-105022a558fb\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-xnhm4" Oct 05 08:53:37 crc kubenswrapper[4846]: I1005 08:53:37.325773 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-openstack-openstack-cell1-xnhm4" Oct 05 08:53:37 crc kubenswrapper[4846]: W1005 08:53:37.926215 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod22023b6c_2796_4c84_96ad_105022a558fb.slice/crio-1ac769455a812ce479cf8a88c944d82894142cd8145c1d323841f3b81a50c753 WatchSource:0}: Error finding container 1ac769455a812ce479cf8a88c944d82894142cd8145c1d323841f3b81a50c753: Status 404 returned error can't find the container with id 1ac769455a812ce479cf8a88c944d82894142cd8145c1d323841f3b81a50c753 Oct 05 08:53:37 crc kubenswrapper[4846]: I1005 08:53:37.927277 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-openstack-openstack-cell1-xnhm4"] Oct 05 08:53:37 crc kubenswrapper[4846]: I1005 08:53:37.930547 4846 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 05 08:53:38 crc kubenswrapper[4846]: I1005 08:53:38.920738 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-openstack-openstack-cell1-xnhm4" event={"ID":"22023b6c-2796-4c84-96ad-105022a558fb","Type":"ContainerStarted","Data":"89c2d611d9288cf38282c5eb40bb39a8c9b57dd21cf5db7f73b9e0293df7857f"} Oct 05 08:53:38 crc kubenswrapper[4846]: I1005 08:53:38.921421 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-openstack-openstack-cell1-xnhm4" event={"ID":"22023b6c-2796-4c84-96ad-105022a558fb","Type":"ContainerStarted","Data":"1ac769455a812ce479cf8a88c944d82894142cd8145c1d323841f3b81a50c753"} Oct 05 08:53:38 crc kubenswrapper[4846]: I1005 08:53:38.963494 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-metadata-openstack-openstack-cell1-xnhm4" podStartSLOduration=2.472116547 podStartE2EDuration="2.963465142s" podCreationTimestamp="2025-10-05 08:53:36 +0000 UTC" firstStartedPulling="2025-10-05 08:53:37.930263398 +0000 UTC m=+7540.171116183" lastFinishedPulling="2025-10-05 08:53:38.421611993 +0000 UTC m=+7540.662464778" observedRunningTime="2025-10-05 08:53:38.949982353 +0000 UTC m=+7541.190835158" watchObservedRunningTime="2025-10-05 08:53:38.963465142 +0000 UTC m=+7541.204317957" Oct 05 08:54:34 crc kubenswrapper[4846]: I1005 08:54:34.515809 4846 generic.go:334] "Generic (PLEG): container finished" podID="22023b6c-2796-4c84-96ad-105022a558fb" containerID="89c2d611d9288cf38282c5eb40bb39a8c9b57dd21cf5db7f73b9e0293df7857f" exitCode=0 Oct 05 08:54:34 crc kubenswrapper[4846]: I1005 08:54:34.515923 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-openstack-openstack-cell1-xnhm4" event={"ID":"22023b6c-2796-4c84-96ad-105022a558fb","Type":"ContainerDied","Data":"89c2d611d9288cf38282c5eb40bb39a8c9b57dd21cf5db7f73b9e0293df7857f"} Oct 05 08:54:35 crc kubenswrapper[4846]: I1005 08:54:35.974760 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-openstack-openstack-cell1-xnhm4" Oct 05 08:54:36 crc kubenswrapper[4846]: I1005 08:54:36.045434 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/22023b6c-2796-4c84-96ad-105022a558fb-neutron-metadata-combined-ca-bundle\") pod \"22023b6c-2796-4c84-96ad-105022a558fb\" (UID: \"22023b6c-2796-4c84-96ad-105022a558fb\") " Oct 05 08:54:36 crc kubenswrapper[4846]: I1005 08:54:36.045648 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5gz2d\" (UniqueName: \"kubernetes.io/projected/22023b6c-2796-4c84-96ad-105022a558fb-kube-api-access-5gz2d\") pod \"22023b6c-2796-4c84-96ad-105022a558fb\" (UID: \"22023b6c-2796-4c84-96ad-105022a558fb\") " Oct 05 08:54:36 crc kubenswrapper[4846]: I1005 08:54:36.045695 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/22023b6c-2796-4c84-96ad-105022a558fb-inventory\") pod \"22023b6c-2796-4c84-96ad-105022a558fb\" (UID: \"22023b6c-2796-4c84-96ad-105022a558fb\") " Oct 05 08:54:36 crc kubenswrapper[4846]: I1005 08:54:36.045721 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/22023b6c-2796-4c84-96ad-105022a558fb-neutron-ovn-metadata-agent-neutron-config-0\") pod \"22023b6c-2796-4c84-96ad-105022a558fb\" (UID: \"22023b6c-2796-4c84-96ad-105022a558fb\") " Oct 05 08:54:36 crc kubenswrapper[4846]: I1005 08:54:36.045768 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/22023b6c-2796-4c84-96ad-105022a558fb-ssh-key\") pod \"22023b6c-2796-4c84-96ad-105022a558fb\" (UID: \"22023b6c-2796-4c84-96ad-105022a558fb\") " Oct 05 08:54:36 crc kubenswrapper[4846]: I1005 08:54:36.045793 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/22023b6c-2796-4c84-96ad-105022a558fb-nova-metadata-neutron-config-0\") pod \"22023b6c-2796-4c84-96ad-105022a558fb\" (UID: \"22023b6c-2796-4c84-96ad-105022a558fb\") " Oct 05 08:54:36 crc kubenswrapper[4846]: I1005 08:54:36.053403 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22023b6c-2796-4c84-96ad-105022a558fb-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "22023b6c-2796-4c84-96ad-105022a558fb" (UID: "22023b6c-2796-4c84-96ad-105022a558fb"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:54:36 crc kubenswrapper[4846]: I1005 08:54:36.059015 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22023b6c-2796-4c84-96ad-105022a558fb-kube-api-access-5gz2d" (OuterVolumeSpecName: "kube-api-access-5gz2d") pod "22023b6c-2796-4c84-96ad-105022a558fb" (UID: "22023b6c-2796-4c84-96ad-105022a558fb"). InnerVolumeSpecName "kube-api-access-5gz2d". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:54:36 crc kubenswrapper[4846]: I1005 08:54:36.077937 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22023b6c-2796-4c84-96ad-105022a558fb-nova-metadata-neutron-config-0" (OuterVolumeSpecName: "nova-metadata-neutron-config-0") pod "22023b6c-2796-4c84-96ad-105022a558fb" (UID: "22023b6c-2796-4c84-96ad-105022a558fb"). InnerVolumeSpecName "nova-metadata-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:54:36 crc kubenswrapper[4846]: I1005 08:54:36.079019 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22023b6c-2796-4c84-96ad-105022a558fb-inventory" (OuterVolumeSpecName: "inventory") pod "22023b6c-2796-4c84-96ad-105022a558fb" (UID: "22023b6c-2796-4c84-96ad-105022a558fb"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:54:36 crc kubenswrapper[4846]: I1005 08:54:36.079982 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22023b6c-2796-4c84-96ad-105022a558fb-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "22023b6c-2796-4c84-96ad-105022a558fb" (UID: "22023b6c-2796-4c84-96ad-105022a558fb"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:54:36 crc kubenswrapper[4846]: I1005 08:54:36.099835 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22023b6c-2796-4c84-96ad-105022a558fb-neutron-ovn-metadata-agent-neutron-config-0" (OuterVolumeSpecName: "neutron-ovn-metadata-agent-neutron-config-0") pod "22023b6c-2796-4c84-96ad-105022a558fb" (UID: "22023b6c-2796-4c84-96ad-105022a558fb"). InnerVolumeSpecName "neutron-ovn-metadata-agent-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:54:36 crc kubenswrapper[4846]: I1005 08:54:36.147963 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5gz2d\" (UniqueName: \"kubernetes.io/projected/22023b6c-2796-4c84-96ad-105022a558fb-kube-api-access-5gz2d\") on node \"crc\" DevicePath \"\"" Oct 05 08:54:36 crc kubenswrapper[4846]: I1005 08:54:36.147999 4846 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/22023b6c-2796-4c84-96ad-105022a558fb-inventory\") on node \"crc\" DevicePath \"\"" Oct 05 08:54:36 crc kubenswrapper[4846]: I1005 08:54:36.148011 4846 reconciler_common.go:293] "Volume detached for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/22023b6c-2796-4c84-96ad-105022a558fb-neutron-ovn-metadata-agent-neutron-config-0\") on node \"crc\" DevicePath \"\"" Oct 05 08:54:36 crc kubenswrapper[4846]: I1005 08:54:36.148021 4846 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/22023b6c-2796-4c84-96ad-105022a558fb-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 05 08:54:36 crc kubenswrapper[4846]: I1005 08:54:36.148031 4846 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/22023b6c-2796-4c84-96ad-105022a558fb-nova-metadata-neutron-config-0\") on node \"crc\" DevicePath \"\"" Oct 05 08:54:36 crc kubenswrapper[4846]: I1005 08:54:36.148044 4846 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/22023b6c-2796-4c84-96ad-105022a558fb-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 08:54:36 crc kubenswrapper[4846]: I1005 08:54:36.539320 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-openstack-openstack-cell1-xnhm4" event={"ID":"22023b6c-2796-4c84-96ad-105022a558fb","Type":"ContainerDied","Data":"1ac769455a812ce479cf8a88c944d82894142cd8145c1d323841f3b81a50c753"} Oct 05 08:54:36 crc kubenswrapper[4846]: I1005 08:54:36.539359 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1ac769455a812ce479cf8a88c944d82894142cd8145c1d323841f3b81a50c753" Oct 05 08:54:36 crc kubenswrapper[4846]: I1005 08:54:36.539366 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-openstack-openstack-cell1-xnhm4" Oct 05 08:54:36 crc kubenswrapper[4846]: I1005 08:54:36.679919 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/libvirt-openstack-openstack-cell1-2f7vt"] Oct 05 08:54:36 crc kubenswrapper[4846]: E1005 08:54:36.680561 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="22023b6c-2796-4c84-96ad-105022a558fb" containerName="neutron-metadata-openstack-openstack-cell1" Oct 05 08:54:36 crc kubenswrapper[4846]: I1005 08:54:36.680594 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="22023b6c-2796-4c84-96ad-105022a558fb" containerName="neutron-metadata-openstack-openstack-cell1" Oct 05 08:54:36 crc kubenswrapper[4846]: I1005 08:54:36.680923 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="22023b6c-2796-4c84-96ad-105022a558fb" containerName="neutron-metadata-openstack-openstack-cell1" Oct 05 08:54:36 crc kubenswrapper[4846]: I1005 08:54:36.682225 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-openstack-openstack-cell1-2f7vt" Oct 05 08:54:36 crc kubenswrapper[4846]: I1005 08:54:36.685417 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 05 08:54:36 crc kubenswrapper[4846]: I1005 08:54:36.685489 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Oct 05 08:54:36 crc kubenswrapper[4846]: I1005 08:54:36.685706 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"libvirt-secret" Oct 05 08:54:36 crc kubenswrapper[4846]: I1005 08:54:36.685819 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-7687s" Oct 05 08:54:36 crc kubenswrapper[4846]: I1005 08:54:36.685844 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Oct 05 08:54:36 crc kubenswrapper[4846]: I1005 08:54:36.723979 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-openstack-openstack-cell1-2f7vt"] Oct 05 08:54:36 crc kubenswrapper[4846]: I1005 08:54:36.758788 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2f3f1db8-f42b-4f2b-aeb3-05b9523aa4ea-inventory\") pod \"libvirt-openstack-openstack-cell1-2f7vt\" (UID: \"2f3f1db8-f42b-4f2b-aeb3-05b9523aa4ea\") " pod="openstack/libvirt-openstack-openstack-cell1-2f7vt" Oct 05 08:54:36 crc kubenswrapper[4846]: I1005 08:54:36.758883 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/2f3f1db8-f42b-4f2b-aeb3-05b9523aa4ea-libvirt-secret-0\") pod \"libvirt-openstack-openstack-cell1-2f7vt\" (UID: \"2f3f1db8-f42b-4f2b-aeb3-05b9523aa4ea\") " pod="openstack/libvirt-openstack-openstack-cell1-2f7vt" Oct 05 08:54:36 crc kubenswrapper[4846]: I1005 08:54:36.758954 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2f3f1db8-f42b-4f2b-aeb3-05b9523aa4ea-ssh-key\") pod \"libvirt-openstack-openstack-cell1-2f7vt\" (UID: \"2f3f1db8-f42b-4f2b-aeb3-05b9523aa4ea\") " pod="openstack/libvirt-openstack-openstack-cell1-2f7vt" Oct 05 08:54:36 crc kubenswrapper[4846]: I1005 08:54:36.758982 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2f3f1db8-f42b-4f2b-aeb3-05b9523aa4ea-libvirt-combined-ca-bundle\") pod \"libvirt-openstack-openstack-cell1-2f7vt\" (UID: \"2f3f1db8-f42b-4f2b-aeb3-05b9523aa4ea\") " pod="openstack/libvirt-openstack-openstack-cell1-2f7vt" Oct 05 08:54:36 crc kubenswrapper[4846]: I1005 08:54:36.759021 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mvcsx\" (UniqueName: \"kubernetes.io/projected/2f3f1db8-f42b-4f2b-aeb3-05b9523aa4ea-kube-api-access-mvcsx\") pod \"libvirt-openstack-openstack-cell1-2f7vt\" (UID: \"2f3f1db8-f42b-4f2b-aeb3-05b9523aa4ea\") " pod="openstack/libvirt-openstack-openstack-cell1-2f7vt" Oct 05 08:54:36 crc kubenswrapper[4846]: I1005 08:54:36.861221 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2f3f1db8-f42b-4f2b-aeb3-05b9523aa4ea-inventory\") pod \"libvirt-openstack-openstack-cell1-2f7vt\" (UID: \"2f3f1db8-f42b-4f2b-aeb3-05b9523aa4ea\") " pod="openstack/libvirt-openstack-openstack-cell1-2f7vt" Oct 05 08:54:36 crc kubenswrapper[4846]: I1005 08:54:36.861632 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/2f3f1db8-f42b-4f2b-aeb3-05b9523aa4ea-libvirt-secret-0\") pod \"libvirt-openstack-openstack-cell1-2f7vt\" (UID: \"2f3f1db8-f42b-4f2b-aeb3-05b9523aa4ea\") " pod="openstack/libvirt-openstack-openstack-cell1-2f7vt" Oct 05 08:54:36 crc kubenswrapper[4846]: I1005 08:54:36.861734 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2f3f1db8-f42b-4f2b-aeb3-05b9523aa4ea-ssh-key\") pod \"libvirt-openstack-openstack-cell1-2f7vt\" (UID: \"2f3f1db8-f42b-4f2b-aeb3-05b9523aa4ea\") " pod="openstack/libvirt-openstack-openstack-cell1-2f7vt" Oct 05 08:54:36 crc kubenswrapper[4846]: I1005 08:54:36.861773 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2f3f1db8-f42b-4f2b-aeb3-05b9523aa4ea-libvirt-combined-ca-bundle\") pod \"libvirt-openstack-openstack-cell1-2f7vt\" (UID: \"2f3f1db8-f42b-4f2b-aeb3-05b9523aa4ea\") " pod="openstack/libvirt-openstack-openstack-cell1-2f7vt" Oct 05 08:54:36 crc kubenswrapper[4846]: I1005 08:54:36.861823 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mvcsx\" (UniqueName: \"kubernetes.io/projected/2f3f1db8-f42b-4f2b-aeb3-05b9523aa4ea-kube-api-access-mvcsx\") pod \"libvirt-openstack-openstack-cell1-2f7vt\" (UID: \"2f3f1db8-f42b-4f2b-aeb3-05b9523aa4ea\") " pod="openstack/libvirt-openstack-openstack-cell1-2f7vt" Oct 05 08:54:36 crc kubenswrapper[4846]: I1005 08:54:36.866138 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2f3f1db8-f42b-4f2b-aeb3-05b9523aa4ea-inventory\") pod \"libvirt-openstack-openstack-cell1-2f7vt\" (UID: \"2f3f1db8-f42b-4f2b-aeb3-05b9523aa4ea\") " pod="openstack/libvirt-openstack-openstack-cell1-2f7vt" Oct 05 08:54:36 crc kubenswrapper[4846]: I1005 08:54:36.866138 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/2f3f1db8-f42b-4f2b-aeb3-05b9523aa4ea-libvirt-secret-0\") pod \"libvirt-openstack-openstack-cell1-2f7vt\" (UID: \"2f3f1db8-f42b-4f2b-aeb3-05b9523aa4ea\") " pod="openstack/libvirt-openstack-openstack-cell1-2f7vt" Oct 05 08:54:36 crc kubenswrapper[4846]: I1005 08:54:36.868458 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2f3f1db8-f42b-4f2b-aeb3-05b9523aa4ea-libvirt-combined-ca-bundle\") pod \"libvirt-openstack-openstack-cell1-2f7vt\" (UID: \"2f3f1db8-f42b-4f2b-aeb3-05b9523aa4ea\") " pod="openstack/libvirt-openstack-openstack-cell1-2f7vt" Oct 05 08:54:36 crc kubenswrapper[4846]: I1005 08:54:36.868597 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2f3f1db8-f42b-4f2b-aeb3-05b9523aa4ea-ssh-key\") pod \"libvirt-openstack-openstack-cell1-2f7vt\" (UID: \"2f3f1db8-f42b-4f2b-aeb3-05b9523aa4ea\") " pod="openstack/libvirt-openstack-openstack-cell1-2f7vt" Oct 05 08:54:36 crc kubenswrapper[4846]: I1005 08:54:36.886814 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mvcsx\" (UniqueName: \"kubernetes.io/projected/2f3f1db8-f42b-4f2b-aeb3-05b9523aa4ea-kube-api-access-mvcsx\") pod \"libvirt-openstack-openstack-cell1-2f7vt\" (UID: \"2f3f1db8-f42b-4f2b-aeb3-05b9523aa4ea\") " pod="openstack/libvirt-openstack-openstack-cell1-2f7vt" Oct 05 08:54:37 crc kubenswrapper[4846]: I1005 08:54:37.008540 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-openstack-openstack-cell1-2f7vt" Oct 05 08:54:37 crc kubenswrapper[4846]: I1005 08:54:37.984377 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-openstack-openstack-cell1-2f7vt"] Oct 05 08:54:38 crc kubenswrapper[4846]: I1005 08:54:38.582096 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-openstack-openstack-cell1-2f7vt" event={"ID":"2f3f1db8-f42b-4f2b-aeb3-05b9523aa4ea","Type":"ContainerStarted","Data":"745cab4fa146d9e3ba50726e19829201855cf4b53c7a4420a77c71d66492c3df"} Oct 05 08:54:39 crc kubenswrapper[4846]: I1005 08:54:39.596058 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-openstack-openstack-cell1-2f7vt" event={"ID":"2f3f1db8-f42b-4f2b-aeb3-05b9523aa4ea","Type":"ContainerStarted","Data":"c31ab2db9ede40146fb3c40c7065bae2a65c725c17b44e74a9a3061bb75267b3"} Oct 05 08:54:39 crc kubenswrapper[4846]: I1005 08:54:39.625631 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/libvirt-openstack-openstack-cell1-2f7vt" podStartSLOduration=3.218211447 podStartE2EDuration="3.625611426s" podCreationTimestamp="2025-10-05 08:54:36 +0000 UTC" firstStartedPulling="2025-10-05 08:54:37.965859808 +0000 UTC m=+7600.206712593" lastFinishedPulling="2025-10-05 08:54:38.373259797 +0000 UTC m=+7600.614112572" observedRunningTime="2025-10-05 08:54:39.61523065 +0000 UTC m=+7601.856083455" watchObservedRunningTime="2025-10-05 08:54:39.625611426 +0000 UTC m=+7601.866464201" Oct 05 08:55:23 crc kubenswrapper[4846]: I1005 08:55:23.324804 4846 patch_prober.go:28] interesting pod/machine-config-daemon-fscvf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 08:55:23 crc kubenswrapper[4846]: I1005 08:55:23.325460 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 08:55:53 crc kubenswrapper[4846]: I1005 08:55:53.325528 4846 patch_prober.go:28] interesting pod/machine-config-daemon-fscvf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 08:55:53 crc kubenswrapper[4846]: I1005 08:55:53.326107 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 08:56:23 crc kubenswrapper[4846]: I1005 08:56:23.324723 4846 patch_prober.go:28] interesting pod/machine-config-daemon-fscvf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 08:56:23 crc kubenswrapper[4846]: I1005 08:56:23.325268 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 08:56:23 crc kubenswrapper[4846]: I1005 08:56:23.325314 4846 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" Oct 05 08:56:23 crc kubenswrapper[4846]: I1005 08:56:23.326413 4846 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"e6df7a4d2c7027b882a7e838679edd7570422c9a8553ab74b0bddbf3e4bf4b45"} pod="openshift-machine-config-operator/machine-config-daemon-fscvf" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 05 08:56:23 crc kubenswrapper[4846]: I1005 08:56:23.326472 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" containerName="machine-config-daemon" containerID="cri-o://e6df7a4d2c7027b882a7e838679edd7570422c9a8553ab74b0bddbf3e4bf4b45" gracePeriod=600 Oct 05 08:56:23 crc kubenswrapper[4846]: E1005 08:56:23.443861 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 08:56:23 crc kubenswrapper[4846]: I1005 08:56:23.929266 4846 generic.go:334] "Generic (PLEG): container finished" podID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" containerID="e6df7a4d2c7027b882a7e838679edd7570422c9a8553ab74b0bddbf3e4bf4b45" exitCode=0 Oct 05 08:56:23 crc kubenswrapper[4846]: I1005 08:56:23.929308 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" event={"ID":"ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7","Type":"ContainerDied","Data":"e6df7a4d2c7027b882a7e838679edd7570422c9a8553ab74b0bddbf3e4bf4b45"} Oct 05 08:56:23 crc kubenswrapper[4846]: I1005 08:56:23.929340 4846 scope.go:117] "RemoveContainer" containerID="eb023100181ea6025e1663017d08951f92365161214cdfbcbfc8a62d71858fc9" Oct 05 08:56:23 crc kubenswrapper[4846]: I1005 08:56:23.929941 4846 scope.go:117] "RemoveContainer" containerID="e6df7a4d2c7027b882a7e838679edd7570422c9a8553ab74b0bddbf3e4bf4b45" Oct 05 08:56:23 crc kubenswrapper[4846]: E1005 08:56:23.930537 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 08:56:35 crc kubenswrapper[4846]: I1005 08:56:35.498311 4846 scope.go:117] "RemoveContainer" containerID="e6df7a4d2c7027b882a7e838679edd7570422c9a8553ab74b0bddbf3e4bf4b45" Oct 05 08:56:35 crc kubenswrapper[4846]: E1005 08:56:35.499099 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 08:56:49 crc kubenswrapper[4846]: I1005 08:56:49.498173 4846 scope.go:117] "RemoveContainer" containerID="e6df7a4d2c7027b882a7e838679edd7570422c9a8553ab74b0bddbf3e4bf4b45" Oct 05 08:56:49 crc kubenswrapper[4846]: E1005 08:56:49.499398 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 08:57:02 crc kubenswrapper[4846]: I1005 08:57:02.498599 4846 scope.go:117] "RemoveContainer" containerID="e6df7a4d2c7027b882a7e838679edd7570422c9a8553ab74b0bddbf3e4bf4b45" Oct 05 08:57:02 crc kubenswrapper[4846]: E1005 08:57:02.501304 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 08:57:13 crc kubenswrapper[4846]: I1005 08:57:13.498160 4846 scope.go:117] "RemoveContainer" containerID="e6df7a4d2c7027b882a7e838679edd7570422c9a8553ab74b0bddbf3e4bf4b45" Oct 05 08:57:13 crc kubenswrapper[4846]: E1005 08:57:13.499138 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 08:57:28 crc kubenswrapper[4846]: I1005 08:57:28.505076 4846 scope.go:117] "RemoveContainer" containerID="e6df7a4d2c7027b882a7e838679edd7570422c9a8553ab74b0bddbf3e4bf4b45" Oct 05 08:57:28 crc kubenswrapper[4846]: E1005 08:57:28.507258 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 08:57:39 crc kubenswrapper[4846]: I1005 08:57:39.498876 4846 scope.go:117] "RemoveContainer" containerID="e6df7a4d2c7027b882a7e838679edd7570422c9a8553ab74b0bddbf3e4bf4b45" Oct 05 08:57:39 crc kubenswrapper[4846]: E1005 08:57:39.499572 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 08:57:52 crc kubenswrapper[4846]: I1005 08:57:52.497867 4846 scope.go:117] "RemoveContainer" containerID="e6df7a4d2c7027b882a7e838679edd7570422c9a8553ab74b0bddbf3e4bf4b45" Oct 05 08:57:52 crc kubenswrapper[4846]: E1005 08:57:52.499224 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 08:58:05 crc kubenswrapper[4846]: I1005 08:58:05.498397 4846 scope.go:117] "RemoveContainer" containerID="e6df7a4d2c7027b882a7e838679edd7570422c9a8553ab74b0bddbf3e4bf4b45" Oct 05 08:58:05 crc kubenswrapper[4846]: E1005 08:58:05.499708 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 08:58:16 crc kubenswrapper[4846]: I1005 08:58:16.500018 4846 scope.go:117] "RemoveContainer" containerID="e6df7a4d2c7027b882a7e838679edd7570422c9a8553ab74b0bddbf3e4bf4b45" Oct 05 08:58:16 crc kubenswrapper[4846]: E1005 08:58:16.501326 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 08:58:31 crc kubenswrapper[4846]: I1005 08:58:31.498579 4846 scope.go:117] "RemoveContainer" containerID="e6df7a4d2c7027b882a7e838679edd7570422c9a8553ab74b0bddbf3e4bf4b45" Oct 05 08:58:31 crc kubenswrapper[4846]: E1005 08:58:31.500180 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 08:58:35 crc kubenswrapper[4846]: I1005 08:58:35.473542 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-2cf94"] Oct 05 08:58:35 crc kubenswrapper[4846]: I1005 08:58:35.476636 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-2cf94" Oct 05 08:58:35 crc kubenswrapper[4846]: I1005 08:58:35.523929 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-2cf94"] Oct 05 08:58:35 crc kubenswrapper[4846]: I1005 08:58:35.562808 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/45f5d41a-602e-4333-9d6c-6ecc84c011fd-utilities\") pod \"community-operators-2cf94\" (UID: \"45f5d41a-602e-4333-9d6c-6ecc84c011fd\") " pod="openshift-marketplace/community-operators-2cf94" Oct 05 08:58:35 crc kubenswrapper[4846]: I1005 08:58:35.562891 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-45c77\" (UniqueName: \"kubernetes.io/projected/45f5d41a-602e-4333-9d6c-6ecc84c011fd-kube-api-access-45c77\") pod \"community-operators-2cf94\" (UID: \"45f5d41a-602e-4333-9d6c-6ecc84c011fd\") " pod="openshift-marketplace/community-operators-2cf94" Oct 05 08:58:35 crc kubenswrapper[4846]: I1005 08:58:35.562930 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/45f5d41a-602e-4333-9d6c-6ecc84c011fd-catalog-content\") pod \"community-operators-2cf94\" (UID: \"45f5d41a-602e-4333-9d6c-6ecc84c011fd\") " pod="openshift-marketplace/community-operators-2cf94" Oct 05 08:58:35 crc kubenswrapper[4846]: I1005 08:58:35.665390 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-45c77\" (UniqueName: \"kubernetes.io/projected/45f5d41a-602e-4333-9d6c-6ecc84c011fd-kube-api-access-45c77\") pod \"community-operators-2cf94\" (UID: \"45f5d41a-602e-4333-9d6c-6ecc84c011fd\") " pod="openshift-marketplace/community-operators-2cf94" Oct 05 08:58:35 crc kubenswrapper[4846]: I1005 08:58:35.665452 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/45f5d41a-602e-4333-9d6c-6ecc84c011fd-catalog-content\") pod \"community-operators-2cf94\" (UID: \"45f5d41a-602e-4333-9d6c-6ecc84c011fd\") " pod="openshift-marketplace/community-operators-2cf94" Oct 05 08:58:35 crc kubenswrapper[4846]: I1005 08:58:35.665616 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/45f5d41a-602e-4333-9d6c-6ecc84c011fd-utilities\") pod \"community-operators-2cf94\" (UID: \"45f5d41a-602e-4333-9d6c-6ecc84c011fd\") " pod="openshift-marketplace/community-operators-2cf94" Oct 05 08:58:35 crc kubenswrapper[4846]: I1005 08:58:35.666093 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/45f5d41a-602e-4333-9d6c-6ecc84c011fd-utilities\") pod \"community-operators-2cf94\" (UID: \"45f5d41a-602e-4333-9d6c-6ecc84c011fd\") " pod="openshift-marketplace/community-operators-2cf94" Oct 05 08:58:35 crc kubenswrapper[4846]: I1005 08:58:35.666209 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/45f5d41a-602e-4333-9d6c-6ecc84c011fd-catalog-content\") pod \"community-operators-2cf94\" (UID: \"45f5d41a-602e-4333-9d6c-6ecc84c011fd\") " pod="openshift-marketplace/community-operators-2cf94" Oct 05 08:58:35 crc kubenswrapper[4846]: I1005 08:58:35.693669 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-45c77\" (UniqueName: \"kubernetes.io/projected/45f5d41a-602e-4333-9d6c-6ecc84c011fd-kube-api-access-45c77\") pod \"community-operators-2cf94\" (UID: \"45f5d41a-602e-4333-9d6c-6ecc84c011fd\") " pod="openshift-marketplace/community-operators-2cf94" Oct 05 08:58:35 crc kubenswrapper[4846]: I1005 08:58:35.809187 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-2cf94" Oct 05 08:58:36 crc kubenswrapper[4846]: W1005 08:58:36.332343 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod45f5d41a_602e_4333_9d6c_6ecc84c011fd.slice/crio-19d6bd75dd010b0430e156b7183b36deb1fc91e41b2c9b16021f0271f9f647a8 WatchSource:0}: Error finding container 19d6bd75dd010b0430e156b7183b36deb1fc91e41b2c9b16021f0271f9f647a8: Status 404 returned error can't find the container with id 19d6bd75dd010b0430e156b7183b36deb1fc91e41b2c9b16021f0271f9f647a8 Oct 05 08:58:36 crc kubenswrapper[4846]: I1005 08:58:36.336136 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-2cf94"] Oct 05 08:58:36 crc kubenswrapper[4846]: I1005 08:58:36.593125 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2cf94" event={"ID":"45f5d41a-602e-4333-9d6c-6ecc84c011fd","Type":"ContainerStarted","Data":"d0ed263d862e84374edd9bdf562e19e2fb4571e5cd22528d636c79ce3036738a"} Oct 05 08:58:36 crc kubenswrapper[4846]: I1005 08:58:36.594159 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2cf94" event={"ID":"45f5d41a-602e-4333-9d6c-6ecc84c011fd","Type":"ContainerStarted","Data":"19d6bd75dd010b0430e156b7183b36deb1fc91e41b2c9b16021f0271f9f647a8"} Oct 05 08:58:37 crc kubenswrapper[4846]: I1005 08:58:37.610535 4846 generic.go:334] "Generic (PLEG): container finished" podID="45f5d41a-602e-4333-9d6c-6ecc84c011fd" containerID="d0ed263d862e84374edd9bdf562e19e2fb4571e5cd22528d636c79ce3036738a" exitCode=0 Oct 05 08:58:37 crc kubenswrapper[4846]: I1005 08:58:37.610630 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2cf94" event={"ID":"45f5d41a-602e-4333-9d6c-6ecc84c011fd","Type":"ContainerDied","Data":"d0ed263d862e84374edd9bdf562e19e2fb4571e5cd22528d636c79ce3036738a"} Oct 05 08:58:39 crc kubenswrapper[4846]: I1005 08:58:39.654699 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2cf94" event={"ID":"45f5d41a-602e-4333-9d6c-6ecc84c011fd","Type":"ContainerStarted","Data":"45a1c03a1c673d45f790a27aabf85b49279f023057b1629e9694dffb1380621b"} Oct 05 08:58:40 crc kubenswrapper[4846]: I1005 08:58:40.668810 4846 generic.go:334] "Generic (PLEG): container finished" podID="45f5d41a-602e-4333-9d6c-6ecc84c011fd" containerID="45a1c03a1c673d45f790a27aabf85b49279f023057b1629e9694dffb1380621b" exitCode=0 Oct 05 08:58:40 crc kubenswrapper[4846]: I1005 08:58:40.668892 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2cf94" event={"ID":"45f5d41a-602e-4333-9d6c-6ecc84c011fd","Type":"ContainerDied","Data":"45a1c03a1c673d45f790a27aabf85b49279f023057b1629e9694dffb1380621b"} Oct 05 08:58:40 crc kubenswrapper[4846]: I1005 08:58:40.671261 4846 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 05 08:58:41 crc kubenswrapper[4846]: I1005 08:58:41.681575 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2cf94" event={"ID":"45f5d41a-602e-4333-9d6c-6ecc84c011fd","Type":"ContainerStarted","Data":"bdef2432020a0b67fa4fe00c95435edf638b4d58ceb54aee215c99d4cbc94e47"} Oct 05 08:58:41 crc kubenswrapper[4846]: I1005 08:58:41.708933 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-2cf94" podStartSLOduration=3.199347739 podStartE2EDuration="6.708900853s" podCreationTimestamp="2025-10-05 08:58:35 +0000 UTC" firstStartedPulling="2025-10-05 08:58:37.613404369 +0000 UTC m=+7839.854257174" lastFinishedPulling="2025-10-05 08:58:41.122957523 +0000 UTC m=+7843.363810288" observedRunningTime="2025-10-05 08:58:41.701636732 +0000 UTC m=+7843.942489517" watchObservedRunningTime="2025-10-05 08:58:41.708900853 +0000 UTC m=+7843.949753668" Oct 05 08:58:42 crc kubenswrapper[4846]: I1005 08:58:42.497831 4846 scope.go:117] "RemoveContainer" containerID="e6df7a4d2c7027b882a7e838679edd7570422c9a8553ab74b0bddbf3e4bf4b45" Oct 05 08:58:42 crc kubenswrapper[4846]: E1005 08:58:42.498688 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 08:58:45 crc kubenswrapper[4846]: I1005 08:58:45.809841 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-2cf94" Oct 05 08:58:45 crc kubenswrapper[4846]: I1005 08:58:45.810398 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-2cf94" Oct 05 08:58:45 crc kubenswrapper[4846]: I1005 08:58:45.896731 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-2cf94" Oct 05 08:58:46 crc kubenswrapper[4846]: I1005 08:58:46.817379 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-2cf94" Oct 05 08:58:46 crc kubenswrapper[4846]: I1005 08:58:46.880965 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-2cf94"] Oct 05 08:58:48 crc kubenswrapper[4846]: I1005 08:58:48.763718 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-2cf94" podUID="45f5d41a-602e-4333-9d6c-6ecc84c011fd" containerName="registry-server" containerID="cri-o://bdef2432020a0b67fa4fe00c95435edf638b4d58ceb54aee215c99d4cbc94e47" gracePeriod=2 Oct 05 08:58:49 crc kubenswrapper[4846]: I1005 08:58:49.266718 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-2cf94" Oct 05 08:58:49 crc kubenswrapper[4846]: I1005 08:58:49.398452 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/45f5d41a-602e-4333-9d6c-6ecc84c011fd-utilities\") pod \"45f5d41a-602e-4333-9d6c-6ecc84c011fd\" (UID: \"45f5d41a-602e-4333-9d6c-6ecc84c011fd\") " Oct 05 08:58:49 crc kubenswrapper[4846]: I1005 08:58:49.398542 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/45f5d41a-602e-4333-9d6c-6ecc84c011fd-catalog-content\") pod \"45f5d41a-602e-4333-9d6c-6ecc84c011fd\" (UID: \"45f5d41a-602e-4333-9d6c-6ecc84c011fd\") " Oct 05 08:58:49 crc kubenswrapper[4846]: I1005 08:58:49.398688 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-45c77\" (UniqueName: \"kubernetes.io/projected/45f5d41a-602e-4333-9d6c-6ecc84c011fd-kube-api-access-45c77\") pod \"45f5d41a-602e-4333-9d6c-6ecc84c011fd\" (UID: \"45f5d41a-602e-4333-9d6c-6ecc84c011fd\") " Oct 05 08:58:49 crc kubenswrapper[4846]: I1005 08:58:49.401203 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/45f5d41a-602e-4333-9d6c-6ecc84c011fd-utilities" (OuterVolumeSpecName: "utilities") pod "45f5d41a-602e-4333-9d6c-6ecc84c011fd" (UID: "45f5d41a-602e-4333-9d6c-6ecc84c011fd"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 08:58:49 crc kubenswrapper[4846]: I1005 08:58:49.405392 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/45f5d41a-602e-4333-9d6c-6ecc84c011fd-kube-api-access-45c77" (OuterVolumeSpecName: "kube-api-access-45c77") pod "45f5d41a-602e-4333-9d6c-6ecc84c011fd" (UID: "45f5d41a-602e-4333-9d6c-6ecc84c011fd"). InnerVolumeSpecName "kube-api-access-45c77". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:58:49 crc kubenswrapper[4846]: I1005 08:58:49.501476 4846 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/45f5d41a-602e-4333-9d6c-6ecc84c011fd-utilities\") on node \"crc\" DevicePath \"\"" Oct 05 08:58:49 crc kubenswrapper[4846]: I1005 08:58:49.501522 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-45c77\" (UniqueName: \"kubernetes.io/projected/45f5d41a-602e-4333-9d6c-6ecc84c011fd-kube-api-access-45c77\") on node \"crc\" DevicePath \"\"" Oct 05 08:58:49 crc kubenswrapper[4846]: I1005 08:58:49.779069 4846 generic.go:334] "Generic (PLEG): container finished" podID="45f5d41a-602e-4333-9d6c-6ecc84c011fd" containerID="bdef2432020a0b67fa4fe00c95435edf638b4d58ceb54aee215c99d4cbc94e47" exitCode=0 Oct 05 08:58:49 crc kubenswrapper[4846]: I1005 08:58:49.779131 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2cf94" event={"ID":"45f5d41a-602e-4333-9d6c-6ecc84c011fd","Type":"ContainerDied","Data":"bdef2432020a0b67fa4fe00c95435edf638b4d58ceb54aee215c99d4cbc94e47"} Oct 05 08:58:49 crc kubenswrapper[4846]: I1005 08:58:49.779209 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2cf94" event={"ID":"45f5d41a-602e-4333-9d6c-6ecc84c011fd","Type":"ContainerDied","Data":"19d6bd75dd010b0430e156b7183b36deb1fc91e41b2c9b16021f0271f9f647a8"} Oct 05 08:58:49 crc kubenswrapper[4846]: I1005 08:58:49.779236 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-2cf94" Oct 05 08:58:49 crc kubenswrapper[4846]: I1005 08:58:49.779264 4846 scope.go:117] "RemoveContainer" containerID="bdef2432020a0b67fa4fe00c95435edf638b4d58ceb54aee215c99d4cbc94e47" Oct 05 08:58:49 crc kubenswrapper[4846]: I1005 08:58:49.817554 4846 scope.go:117] "RemoveContainer" containerID="45a1c03a1c673d45f790a27aabf85b49279f023057b1629e9694dffb1380621b" Oct 05 08:58:49 crc kubenswrapper[4846]: I1005 08:58:49.868238 4846 scope.go:117] "RemoveContainer" containerID="d0ed263d862e84374edd9bdf562e19e2fb4571e5cd22528d636c79ce3036738a" Oct 05 08:58:49 crc kubenswrapper[4846]: I1005 08:58:49.958285 4846 scope.go:117] "RemoveContainer" containerID="bdef2432020a0b67fa4fe00c95435edf638b4d58ceb54aee215c99d4cbc94e47" Oct 05 08:58:49 crc kubenswrapper[4846]: E1005 08:58:49.958654 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bdef2432020a0b67fa4fe00c95435edf638b4d58ceb54aee215c99d4cbc94e47\": container with ID starting with bdef2432020a0b67fa4fe00c95435edf638b4d58ceb54aee215c99d4cbc94e47 not found: ID does not exist" containerID="bdef2432020a0b67fa4fe00c95435edf638b4d58ceb54aee215c99d4cbc94e47" Oct 05 08:58:49 crc kubenswrapper[4846]: I1005 08:58:49.958708 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bdef2432020a0b67fa4fe00c95435edf638b4d58ceb54aee215c99d4cbc94e47"} err="failed to get container status \"bdef2432020a0b67fa4fe00c95435edf638b4d58ceb54aee215c99d4cbc94e47\": rpc error: code = NotFound desc = could not find container \"bdef2432020a0b67fa4fe00c95435edf638b4d58ceb54aee215c99d4cbc94e47\": container with ID starting with bdef2432020a0b67fa4fe00c95435edf638b4d58ceb54aee215c99d4cbc94e47 not found: ID does not exist" Oct 05 08:58:49 crc kubenswrapper[4846]: I1005 08:58:49.958745 4846 scope.go:117] "RemoveContainer" containerID="45a1c03a1c673d45f790a27aabf85b49279f023057b1629e9694dffb1380621b" Oct 05 08:58:49 crc kubenswrapper[4846]: E1005 08:58:49.959241 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"45a1c03a1c673d45f790a27aabf85b49279f023057b1629e9694dffb1380621b\": container with ID starting with 45a1c03a1c673d45f790a27aabf85b49279f023057b1629e9694dffb1380621b not found: ID does not exist" containerID="45a1c03a1c673d45f790a27aabf85b49279f023057b1629e9694dffb1380621b" Oct 05 08:58:49 crc kubenswrapper[4846]: I1005 08:58:49.959280 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"45a1c03a1c673d45f790a27aabf85b49279f023057b1629e9694dffb1380621b"} err="failed to get container status \"45a1c03a1c673d45f790a27aabf85b49279f023057b1629e9694dffb1380621b\": rpc error: code = NotFound desc = could not find container \"45a1c03a1c673d45f790a27aabf85b49279f023057b1629e9694dffb1380621b\": container with ID starting with 45a1c03a1c673d45f790a27aabf85b49279f023057b1629e9694dffb1380621b not found: ID does not exist" Oct 05 08:58:49 crc kubenswrapper[4846]: I1005 08:58:49.959307 4846 scope.go:117] "RemoveContainer" containerID="d0ed263d862e84374edd9bdf562e19e2fb4571e5cd22528d636c79ce3036738a" Oct 05 08:58:49 crc kubenswrapper[4846]: E1005 08:58:49.959801 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d0ed263d862e84374edd9bdf562e19e2fb4571e5cd22528d636c79ce3036738a\": container with ID starting with d0ed263d862e84374edd9bdf562e19e2fb4571e5cd22528d636c79ce3036738a not found: ID does not exist" containerID="d0ed263d862e84374edd9bdf562e19e2fb4571e5cd22528d636c79ce3036738a" Oct 05 08:58:49 crc kubenswrapper[4846]: I1005 08:58:49.959847 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d0ed263d862e84374edd9bdf562e19e2fb4571e5cd22528d636c79ce3036738a"} err="failed to get container status \"d0ed263d862e84374edd9bdf562e19e2fb4571e5cd22528d636c79ce3036738a\": rpc error: code = NotFound desc = could not find container \"d0ed263d862e84374edd9bdf562e19e2fb4571e5cd22528d636c79ce3036738a\": container with ID starting with d0ed263d862e84374edd9bdf562e19e2fb4571e5cd22528d636c79ce3036738a not found: ID does not exist" Oct 05 08:58:50 crc kubenswrapper[4846]: I1005 08:58:50.083849 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/45f5d41a-602e-4333-9d6c-6ecc84c011fd-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "45f5d41a-602e-4333-9d6c-6ecc84c011fd" (UID: "45f5d41a-602e-4333-9d6c-6ecc84c011fd"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 08:58:50 crc kubenswrapper[4846]: I1005 08:58:50.114966 4846 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/45f5d41a-602e-4333-9d6c-6ecc84c011fd-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 05 08:58:50 crc kubenswrapper[4846]: I1005 08:58:50.422150 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-2cf94"] Oct 05 08:58:50 crc kubenswrapper[4846]: I1005 08:58:50.434852 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-2cf94"] Oct 05 08:58:50 crc kubenswrapper[4846]: I1005 08:58:50.515244 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="45f5d41a-602e-4333-9d6c-6ecc84c011fd" path="/var/lib/kubelet/pods/45f5d41a-602e-4333-9d6c-6ecc84c011fd/volumes" Oct 05 08:58:55 crc kubenswrapper[4846]: I1005 08:58:55.499886 4846 scope.go:117] "RemoveContainer" containerID="e6df7a4d2c7027b882a7e838679edd7570422c9a8553ab74b0bddbf3e4bf4b45" Oct 05 08:58:55 crc kubenswrapper[4846]: E1005 08:58:55.500882 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 08:59:08 crc kubenswrapper[4846]: I1005 08:59:08.504872 4846 scope.go:117] "RemoveContainer" containerID="e6df7a4d2c7027b882a7e838679edd7570422c9a8553ab74b0bddbf3e4bf4b45" Oct 05 08:59:08 crc kubenswrapper[4846]: E1005 08:59:08.505839 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 08:59:11 crc kubenswrapper[4846]: I1005 08:59:11.746152 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-9bxg2"] Oct 05 08:59:11 crc kubenswrapper[4846]: E1005 08:59:11.747795 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="45f5d41a-602e-4333-9d6c-6ecc84c011fd" containerName="extract-content" Oct 05 08:59:11 crc kubenswrapper[4846]: I1005 08:59:11.747824 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="45f5d41a-602e-4333-9d6c-6ecc84c011fd" containerName="extract-content" Oct 05 08:59:11 crc kubenswrapper[4846]: E1005 08:59:11.747865 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="45f5d41a-602e-4333-9d6c-6ecc84c011fd" containerName="registry-server" Oct 05 08:59:11 crc kubenswrapper[4846]: I1005 08:59:11.747876 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="45f5d41a-602e-4333-9d6c-6ecc84c011fd" containerName="registry-server" Oct 05 08:59:11 crc kubenswrapper[4846]: E1005 08:59:11.747891 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="45f5d41a-602e-4333-9d6c-6ecc84c011fd" containerName="extract-utilities" Oct 05 08:59:11 crc kubenswrapper[4846]: I1005 08:59:11.747899 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="45f5d41a-602e-4333-9d6c-6ecc84c011fd" containerName="extract-utilities" Oct 05 08:59:11 crc kubenswrapper[4846]: I1005 08:59:11.748253 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="45f5d41a-602e-4333-9d6c-6ecc84c011fd" containerName="registry-server" Oct 05 08:59:11 crc kubenswrapper[4846]: I1005 08:59:11.750111 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-9bxg2" Oct 05 08:59:11 crc kubenswrapper[4846]: I1005 08:59:11.760001 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-9bxg2"] Oct 05 08:59:11 crc kubenswrapper[4846]: I1005 08:59:11.884738 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/beef8490-c87d-4459-ab70-ad20d33bfb63-catalog-content\") pod \"redhat-marketplace-9bxg2\" (UID: \"beef8490-c87d-4459-ab70-ad20d33bfb63\") " pod="openshift-marketplace/redhat-marketplace-9bxg2" Oct 05 08:59:11 crc kubenswrapper[4846]: I1005 08:59:11.884898 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/beef8490-c87d-4459-ab70-ad20d33bfb63-utilities\") pod \"redhat-marketplace-9bxg2\" (UID: \"beef8490-c87d-4459-ab70-ad20d33bfb63\") " pod="openshift-marketplace/redhat-marketplace-9bxg2" Oct 05 08:59:11 crc kubenswrapper[4846]: I1005 08:59:11.884952 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wh28b\" (UniqueName: \"kubernetes.io/projected/beef8490-c87d-4459-ab70-ad20d33bfb63-kube-api-access-wh28b\") pod \"redhat-marketplace-9bxg2\" (UID: \"beef8490-c87d-4459-ab70-ad20d33bfb63\") " pod="openshift-marketplace/redhat-marketplace-9bxg2" Oct 05 08:59:11 crc kubenswrapper[4846]: I1005 08:59:11.986956 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/beef8490-c87d-4459-ab70-ad20d33bfb63-utilities\") pod \"redhat-marketplace-9bxg2\" (UID: \"beef8490-c87d-4459-ab70-ad20d33bfb63\") " pod="openshift-marketplace/redhat-marketplace-9bxg2" Oct 05 08:59:11 crc kubenswrapper[4846]: I1005 08:59:11.987055 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wh28b\" (UniqueName: \"kubernetes.io/projected/beef8490-c87d-4459-ab70-ad20d33bfb63-kube-api-access-wh28b\") pod \"redhat-marketplace-9bxg2\" (UID: \"beef8490-c87d-4459-ab70-ad20d33bfb63\") " pod="openshift-marketplace/redhat-marketplace-9bxg2" Oct 05 08:59:11 crc kubenswrapper[4846]: I1005 08:59:11.987274 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/beef8490-c87d-4459-ab70-ad20d33bfb63-catalog-content\") pod \"redhat-marketplace-9bxg2\" (UID: \"beef8490-c87d-4459-ab70-ad20d33bfb63\") " pod="openshift-marketplace/redhat-marketplace-9bxg2" Oct 05 08:59:11 crc kubenswrapper[4846]: I1005 08:59:11.987606 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/beef8490-c87d-4459-ab70-ad20d33bfb63-utilities\") pod \"redhat-marketplace-9bxg2\" (UID: \"beef8490-c87d-4459-ab70-ad20d33bfb63\") " pod="openshift-marketplace/redhat-marketplace-9bxg2" Oct 05 08:59:11 crc kubenswrapper[4846]: I1005 08:59:11.987770 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/beef8490-c87d-4459-ab70-ad20d33bfb63-catalog-content\") pod \"redhat-marketplace-9bxg2\" (UID: \"beef8490-c87d-4459-ab70-ad20d33bfb63\") " pod="openshift-marketplace/redhat-marketplace-9bxg2" Oct 05 08:59:12 crc kubenswrapper[4846]: I1005 08:59:12.015327 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wh28b\" (UniqueName: \"kubernetes.io/projected/beef8490-c87d-4459-ab70-ad20d33bfb63-kube-api-access-wh28b\") pod \"redhat-marketplace-9bxg2\" (UID: \"beef8490-c87d-4459-ab70-ad20d33bfb63\") " pod="openshift-marketplace/redhat-marketplace-9bxg2" Oct 05 08:59:12 crc kubenswrapper[4846]: I1005 08:59:12.078034 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-9bxg2" Oct 05 08:59:12 crc kubenswrapper[4846]: I1005 08:59:12.790604 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-9bxg2"] Oct 05 08:59:13 crc kubenswrapper[4846]: I1005 08:59:13.054516 4846 generic.go:334] "Generic (PLEG): container finished" podID="beef8490-c87d-4459-ab70-ad20d33bfb63" containerID="1c4bd9946bab7202646303d842707fcd8738710eefbca17ec827e1cd197b383a" exitCode=0 Oct 05 08:59:13 crc kubenswrapper[4846]: I1005 08:59:13.054577 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9bxg2" event={"ID":"beef8490-c87d-4459-ab70-ad20d33bfb63","Type":"ContainerDied","Data":"1c4bd9946bab7202646303d842707fcd8738710eefbca17ec827e1cd197b383a"} Oct 05 08:59:13 crc kubenswrapper[4846]: I1005 08:59:13.054929 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9bxg2" event={"ID":"beef8490-c87d-4459-ab70-ad20d33bfb63","Type":"ContainerStarted","Data":"0ee0ef9a25968076eb6cfbc7e25b653e0195b7d66cf25bccaca2a71106587d68"} Oct 05 08:59:14 crc kubenswrapper[4846]: I1005 08:59:14.070216 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9bxg2" event={"ID":"beef8490-c87d-4459-ab70-ad20d33bfb63","Type":"ContainerStarted","Data":"667c974facb5f2a7db8d43f1c0e4da4af4a8b329ac9a59ee019cdd23fde47aec"} Oct 05 08:59:15 crc kubenswrapper[4846]: I1005 08:59:15.083850 4846 generic.go:334] "Generic (PLEG): container finished" podID="beef8490-c87d-4459-ab70-ad20d33bfb63" containerID="667c974facb5f2a7db8d43f1c0e4da4af4a8b329ac9a59ee019cdd23fde47aec" exitCode=0 Oct 05 08:59:15 crc kubenswrapper[4846]: I1005 08:59:15.083997 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9bxg2" event={"ID":"beef8490-c87d-4459-ab70-ad20d33bfb63","Type":"ContainerDied","Data":"667c974facb5f2a7db8d43f1c0e4da4af4a8b329ac9a59ee019cdd23fde47aec"} Oct 05 08:59:16 crc kubenswrapper[4846]: I1005 08:59:16.096265 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9bxg2" event={"ID":"beef8490-c87d-4459-ab70-ad20d33bfb63","Type":"ContainerStarted","Data":"3908c41cae61a0654fff2d77459ff2858aeac597a12dfe7e6b75bc970c5d57ce"} Oct 05 08:59:16 crc kubenswrapper[4846]: I1005 08:59:16.125497 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-9bxg2" podStartSLOduration=2.416968052 podStartE2EDuration="5.125472388s" podCreationTimestamp="2025-10-05 08:59:11 +0000 UTC" firstStartedPulling="2025-10-05 08:59:13.056562736 +0000 UTC m=+7875.297415511" lastFinishedPulling="2025-10-05 08:59:15.765067052 +0000 UTC m=+7878.005919847" observedRunningTime="2025-10-05 08:59:16.117765985 +0000 UTC m=+7878.358618760" watchObservedRunningTime="2025-10-05 08:59:16.125472388 +0000 UTC m=+7878.366325173" Oct 05 08:59:22 crc kubenswrapper[4846]: I1005 08:59:22.078257 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-9bxg2" Oct 05 08:59:22 crc kubenswrapper[4846]: I1005 08:59:22.078971 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-9bxg2" Oct 05 08:59:22 crc kubenswrapper[4846]: I1005 08:59:22.162418 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-9bxg2" Oct 05 08:59:22 crc kubenswrapper[4846]: I1005 08:59:22.249833 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-9bxg2" Oct 05 08:59:22 crc kubenswrapper[4846]: I1005 08:59:22.407798 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-9bxg2"] Oct 05 08:59:22 crc kubenswrapper[4846]: I1005 08:59:22.497968 4846 scope.go:117] "RemoveContainer" containerID="e6df7a4d2c7027b882a7e838679edd7570422c9a8553ab74b0bddbf3e4bf4b45" Oct 05 08:59:22 crc kubenswrapper[4846]: E1005 08:59:22.498536 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 08:59:24 crc kubenswrapper[4846]: I1005 08:59:24.193606 4846 generic.go:334] "Generic (PLEG): container finished" podID="2f3f1db8-f42b-4f2b-aeb3-05b9523aa4ea" containerID="c31ab2db9ede40146fb3c40c7065bae2a65c725c17b44e74a9a3061bb75267b3" exitCode=0 Oct 05 08:59:24 crc kubenswrapper[4846]: I1005 08:59:24.193698 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-openstack-openstack-cell1-2f7vt" event={"ID":"2f3f1db8-f42b-4f2b-aeb3-05b9523aa4ea","Type":"ContainerDied","Data":"c31ab2db9ede40146fb3c40c7065bae2a65c725c17b44e74a9a3061bb75267b3"} Oct 05 08:59:24 crc kubenswrapper[4846]: I1005 08:59:24.194171 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-9bxg2" podUID="beef8490-c87d-4459-ab70-ad20d33bfb63" containerName="registry-server" containerID="cri-o://3908c41cae61a0654fff2d77459ff2858aeac597a12dfe7e6b75bc970c5d57ce" gracePeriod=2 Oct 05 08:59:24 crc kubenswrapper[4846]: I1005 08:59:24.668337 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-9bxg2" Oct 05 08:59:24 crc kubenswrapper[4846]: I1005 08:59:24.769345 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/beef8490-c87d-4459-ab70-ad20d33bfb63-catalog-content\") pod \"beef8490-c87d-4459-ab70-ad20d33bfb63\" (UID: \"beef8490-c87d-4459-ab70-ad20d33bfb63\") " Oct 05 08:59:24 crc kubenswrapper[4846]: I1005 08:59:24.769417 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/beef8490-c87d-4459-ab70-ad20d33bfb63-utilities\") pod \"beef8490-c87d-4459-ab70-ad20d33bfb63\" (UID: \"beef8490-c87d-4459-ab70-ad20d33bfb63\") " Oct 05 08:59:24 crc kubenswrapper[4846]: I1005 08:59:24.769479 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wh28b\" (UniqueName: \"kubernetes.io/projected/beef8490-c87d-4459-ab70-ad20d33bfb63-kube-api-access-wh28b\") pod \"beef8490-c87d-4459-ab70-ad20d33bfb63\" (UID: \"beef8490-c87d-4459-ab70-ad20d33bfb63\") " Oct 05 08:59:24 crc kubenswrapper[4846]: I1005 08:59:24.772132 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/beef8490-c87d-4459-ab70-ad20d33bfb63-utilities" (OuterVolumeSpecName: "utilities") pod "beef8490-c87d-4459-ab70-ad20d33bfb63" (UID: "beef8490-c87d-4459-ab70-ad20d33bfb63"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 08:59:24 crc kubenswrapper[4846]: I1005 08:59:24.776171 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/beef8490-c87d-4459-ab70-ad20d33bfb63-kube-api-access-wh28b" (OuterVolumeSpecName: "kube-api-access-wh28b") pod "beef8490-c87d-4459-ab70-ad20d33bfb63" (UID: "beef8490-c87d-4459-ab70-ad20d33bfb63"). InnerVolumeSpecName "kube-api-access-wh28b". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:59:24 crc kubenswrapper[4846]: I1005 08:59:24.784503 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/beef8490-c87d-4459-ab70-ad20d33bfb63-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "beef8490-c87d-4459-ab70-ad20d33bfb63" (UID: "beef8490-c87d-4459-ab70-ad20d33bfb63"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 08:59:24 crc kubenswrapper[4846]: I1005 08:59:24.872726 4846 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/beef8490-c87d-4459-ab70-ad20d33bfb63-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 05 08:59:24 crc kubenswrapper[4846]: I1005 08:59:24.872768 4846 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/beef8490-c87d-4459-ab70-ad20d33bfb63-utilities\") on node \"crc\" DevicePath \"\"" Oct 05 08:59:24 crc kubenswrapper[4846]: I1005 08:59:24.872783 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wh28b\" (UniqueName: \"kubernetes.io/projected/beef8490-c87d-4459-ab70-ad20d33bfb63-kube-api-access-wh28b\") on node \"crc\" DevicePath \"\"" Oct 05 08:59:25 crc kubenswrapper[4846]: I1005 08:59:25.206303 4846 generic.go:334] "Generic (PLEG): container finished" podID="beef8490-c87d-4459-ab70-ad20d33bfb63" containerID="3908c41cae61a0654fff2d77459ff2858aeac597a12dfe7e6b75bc970c5d57ce" exitCode=0 Oct 05 08:59:25 crc kubenswrapper[4846]: I1005 08:59:25.206515 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9bxg2" event={"ID":"beef8490-c87d-4459-ab70-ad20d33bfb63","Type":"ContainerDied","Data":"3908c41cae61a0654fff2d77459ff2858aeac597a12dfe7e6b75bc970c5d57ce"} Oct 05 08:59:25 crc kubenswrapper[4846]: I1005 08:59:25.206803 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9bxg2" event={"ID":"beef8490-c87d-4459-ab70-ad20d33bfb63","Type":"ContainerDied","Data":"0ee0ef9a25968076eb6cfbc7e25b653e0195b7d66cf25bccaca2a71106587d68"} Oct 05 08:59:25 crc kubenswrapper[4846]: I1005 08:59:25.206862 4846 scope.go:117] "RemoveContainer" containerID="3908c41cae61a0654fff2d77459ff2858aeac597a12dfe7e6b75bc970c5d57ce" Oct 05 08:59:25 crc kubenswrapper[4846]: I1005 08:59:25.207393 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-9bxg2" Oct 05 08:59:25 crc kubenswrapper[4846]: I1005 08:59:25.265509 4846 scope.go:117] "RemoveContainer" containerID="667c974facb5f2a7db8d43f1c0e4da4af4a8b329ac9a59ee019cdd23fde47aec" Oct 05 08:59:25 crc kubenswrapper[4846]: I1005 08:59:25.269432 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-9bxg2"] Oct 05 08:59:25 crc kubenswrapper[4846]: I1005 08:59:25.279415 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-9bxg2"] Oct 05 08:59:25 crc kubenswrapper[4846]: I1005 08:59:25.320996 4846 scope.go:117] "RemoveContainer" containerID="1c4bd9946bab7202646303d842707fcd8738710eefbca17ec827e1cd197b383a" Oct 05 08:59:25 crc kubenswrapper[4846]: I1005 08:59:25.370763 4846 scope.go:117] "RemoveContainer" containerID="3908c41cae61a0654fff2d77459ff2858aeac597a12dfe7e6b75bc970c5d57ce" Oct 05 08:59:25 crc kubenswrapper[4846]: E1005 08:59:25.371425 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3908c41cae61a0654fff2d77459ff2858aeac597a12dfe7e6b75bc970c5d57ce\": container with ID starting with 3908c41cae61a0654fff2d77459ff2858aeac597a12dfe7e6b75bc970c5d57ce not found: ID does not exist" containerID="3908c41cae61a0654fff2d77459ff2858aeac597a12dfe7e6b75bc970c5d57ce" Oct 05 08:59:25 crc kubenswrapper[4846]: I1005 08:59:25.371464 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3908c41cae61a0654fff2d77459ff2858aeac597a12dfe7e6b75bc970c5d57ce"} err="failed to get container status \"3908c41cae61a0654fff2d77459ff2858aeac597a12dfe7e6b75bc970c5d57ce\": rpc error: code = NotFound desc = could not find container \"3908c41cae61a0654fff2d77459ff2858aeac597a12dfe7e6b75bc970c5d57ce\": container with ID starting with 3908c41cae61a0654fff2d77459ff2858aeac597a12dfe7e6b75bc970c5d57ce not found: ID does not exist" Oct 05 08:59:25 crc kubenswrapper[4846]: I1005 08:59:25.371490 4846 scope.go:117] "RemoveContainer" containerID="667c974facb5f2a7db8d43f1c0e4da4af4a8b329ac9a59ee019cdd23fde47aec" Oct 05 08:59:25 crc kubenswrapper[4846]: E1005 08:59:25.371828 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"667c974facb5f2a7db8d43f1c0e4da4af4a8b329ac9a59ee019cdd23fde47aec\": container with ID starting with 667c974facb5f2a7db8d43f1c0e4da4af4a8b329ac9a59ee019cdd23fde47aec not found: ID does not exist" containerID="667c974facb5f2a7db8d43f1c0e4da4af4a8b329ac9a59ee019cdd23fde47aec" Oct 05 08:59:25 crc kubenswrapper[4846]: I1005 08:59:25.371858 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"667c974facb5f2a7db8d43f1c0e4da4af4a8b329ac9a59ee019cdd23fde47aec"} err="failed to get container status \"667c974facb5f2a7db8d43f1c0e4da4af4a8b329ac9a59ee019cdd23fde47aec\": rpc error: code = NotFound desc = could not find container \"667c974facb5f2a7db8d43f1c0e4da4af4a8b329ac9a59ee019cdd23fde47aec\": container with ID starting with 667c974facb5f2a7db8d43f1c0e4da4af4a8b329ac9a59ee019cdd23fde47aec not found: ID does not exist" Oct 05 08:59:25 crc kubenswrapper[4846]: I1005 08:59:25.371875 4846 scope.go:117] "RemoveContainer" containerID="1c4bd9946bab7202646303d842707fcd8738710eefbca17ec827e1cd197b383a" Oct 05 08:59:25 crc kubenswrapper[4846]: E1005 08:59:25.372257 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1c4bd9946bab7202646303d842707fcd8738710eefbca17ec827e1cd197b383a\": container with ID starting with 1c4bd9946bab7202646303d842707fcd8738710eefbca17ec827e1cd197b383a not found: ID does not exist" containerID="1c4bd9946bab7202646303d842707fcd8738710eefbca17ec827e1cd197b383a" Oct 05 08:59:25 crc kubenswrapper[4846]: I1005 08:59:25.372284 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1c4bd9946bab7202646303d842707fcd8738710eefbca17ec827e1cd197b383a"} err="failed to get container status \"1c4bd9946bab7202646303d842707fcd8738710eefbca17ec827e1cd197b383a\": rpc error: code = NotFound desc = could not find container \"1c4bd9946bab7202646303d842707fcd8738710eefbca17ec827e1cd197b383a\": container with ID starting with 1c4bd9946bab7202646303d842707fcd8738710eefbca17ec827e1cd197b383a not found: ID does not exist" Oct 05 08:59:25 crc kubenswrapper[4846]: I1005 08:59:25.660465 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-openstack-openstack-cell1-2f7vt" Oct 05 08:59:25 crc kubenswrapper[4846]: I1005 08:59:25.800944 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/2f3f1db8-f42b-4f2b-aeb3-05b9523aa4ea-libvirt-secret-0\") pod \"2f3f1db8-f42b-4f2b-aeb3-05b9523aa4ea\" (UID: \"2f3f1db8-f42b-4f2b-aeb3-05b9523aa4ea\") " Oct 05 08:59:25 crc kubenswrapper[4846]: I1005 08:59:25.801104 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2f3f1db8-f42b-4f2b-aeb3-05b9523aa4ea-libvirt-combined-ca-bundle\") pod \"2f3f1db8-f42b-4f2b-aeb3-05b9523aa4ea\" (UID: \"2f3f1db8-f42b-4f2b-aeb3-05b9523aa4ea\") " Oct 05 08:59:25 crc kubenswrapper[4846]: I1005 08:59:25.801184 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2f3f1db8-f42b-4f2b-aeb3-05b9523aa4ea-inventory\") pod \"2f3f1db8-f42b-4f2b-aeb3-05b9523aa4ea\" (UID: \"2f3f1db8-f42b-4f2b-aeb3-05b9523aa4ea\") " Oct 05 08:59:25 crc kubenswrapper[4846]: I1005 08:59:25.801796 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mvcsx\" (UniqueName: \"kubernetes.io/projected/2f3f1db8-f42b-4f2b-aeb3-05b9523aa4ea-kube-api-access-mvcsx\") pod \"2f3f1db8-f42b-4f2b-aeb3-05b9523aa4ea\" (UID: \"2f3f1db8-f42b-4f2b-aeb3-05b9523aa4ea\") " Oct 05 08:59:25 crc kubenswrapper[4846]: I1005 08:59:25.801848 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2f3f1db8-f42b-4f2b-aeb3-05b9523aa4ea-ssh-key\") pod \"2f3f1db8-f42b-4f2b-aeb3-05b9523aa4ea\" (UID: \"2f3f1db8-f42b-4f2b-aeb3-05b9523aa4ea\") " Oct 05 08:59:25 crc kubenswrapper[4846]: I1005 08:59:25.805329 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2f3f1db8-f42b-4f2b-aeb3-05b9523aa4ea-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "2f3f1db8-f42b-4f2b-aeb3-05b9523aa4ea" (UID: "2f3f1db8-f42b-4f2b-aeb3-05b9523aa4ea"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:59:25 crc kubenswrapper[4846]: I1005 08:59:25.807456 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2f3f1db8-f42b-4f2b-aeb3-05b9523aa4ea-kube-api-access-mvcsx" (OuterVolumeSpecName: "kube-api-access-mvcsx") pod "2f3f1db8-f42b-4f2b-aeb3-05b9523aa4ea" (UID: "2f3f1db8-f42b-4f2b-aeb3-05b9523aa4ea"). InnerVolumeSpecName "kube-api-access-mvcsx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 08:59:25 crc kubenswrapper[4846]: I1005 08:59:25.827731 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2f3f1db8-f42b-4f2b-aeb3-05b9523aa4ea-libvirt-secret-0" (OuterVolumeSpecName: "libvirt-secret-0") pod "2f3f1db8-f42b-4f2b-aeb3-05b9523aa4ea" (UID: "2f3f1db8-f42b-4f2b-aeb3-05b9523aa4ea"). InnerVolumeSpecName "libvirt-secret-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:59:25 crc kubenswrapper[4846]: I1005 08:59:25.829322 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2f3f1db8-f42b-4f2b-aeb3-05b9523aa4ea-inventory" (OuterVolumeSpecName: "inventory") pod "2f3f1db8-f42b-4f2b-aeb3-05b9523aa4ea" (UID: "2f3f1db8-f42b-4f2b-aeb3-05b9523aa4ea"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:59:25 crc kubenswrapper[4846]: I1005 08:59:25.837620 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2f3f1db8-f42b-4f2b-aeb3-05b9523aa4ea-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "2f3f1db8-f42b-4f2b-aeb3-05b9523aa4ea" (UID: "2f3f1db8-f42b-4f2b-aeb3-05b9523aa4ea"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 08:59:25 crc kubenswrapper[4846]: I1005 08:59:25.903937 4846 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2f3f1db8-f42b-4f2b-aeb3-05b9523aa4ea-inventory\") on node \"crc\" DevicePath \"\"" Oct 05 08:59:25 crc kubenswrapper[4846]: I1005 08:59:25.903975 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mvcsx\" (UniqueName: \"kubernetes.io/projected/2f3f1db8-f42b-4f2b-aeb3-05b9523aa4ea-kube-api-access-mvcsx\") on node \"crc\" DevicePath \"\"" Oct 05 08:59:25 crc kubenswrapper[4846]: I1005 08:59:25.903986 4846 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2f3f1db8-f42b-4f2b-aeb3-05b9523aa4ea-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 05 08:59:25 crc kubenswrapper[4846]: I1005 08:59:25.903996 4846 reconciler_common.go:293] "Volume detached for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/2f3f1db8-f42b-4f2b-aeb3-05b9523aa4ea-libvirt-secret-0\") on node \"crc\" DevicePath \"\"" Oct 05 08:59:25 crc kubenswrapper[4846]: I1005 08:59:25.904005 4846 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2f3f1db8-f42b-4f2b-aeb3-05b9523aa4ea-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 08:59:26 crc kubenswrapper[4846]: I1005 08:59:26.224907 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-openstack-openstack-cell1-2f7vt" event={"ID":"2f3f1db8-f42b-4f2b-aeb3-05b9523aa4ea","Type":"ContainerDied","Data":"745cab4fa146d9e3ba50726e19829201855cf4b53c7a4420a77c71d66492c3df"} Oct 05 08:59:26 crc kubenswrapper[4846]: I1005 08:59:26.224994 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="745cab4fa146d9e3ba50726e19829201855cf4b53c7a4420a77c71d66492c3df" Oct 05 08:59:26 crc kubenswrapper[4846]: I1005 08:59:26.225101 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-openstack-openstack-cell1-2f7vt" Oct 05 08:59:26 crc kubenswrapper[4846]: I1005 08:59:26.304596 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-openstack-openstack-cell1-kjczt"] Oct 05 08:59:26 crc kubenswrapper[4846]: E1005 08:59:26.305009 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2f3f1db8-f42b-4f2b-aeb3-05b9523aa4ea" containerName="libvirt-openstack-openstack-cell1" Oct 05 08:59:26 crc kubenswrapper[4846]: I1005 08:59:26.305024 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="2f3f1db8-f42b-4f2b-aeb3-05b9523aa4ea" containerName="libvirt-openstack-openstack-cell1" Oct 05 08:59:26 crc kubenswrapper[4846]: E1005 08:59:26.305037 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="beef8490-c87d-4459-ab70-ad20d33bfb63" containerName="registry-server" Oct 05 08:59:26 crc kubenswrapper[4846]: I1005 08:59:26.305045 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="beef8490-c87d-4459-ab70-ad20d33bfb63" containerName="registry-server" Oct 05 08:59:26 crc kubenswrapper[4846]: E1005 08:59:26.305062 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="beef8490-c87d-4459-ab70-ad20d33bfb63" containerName="extract-utilities" Oct 05 08:59:26 crc kubenswrapper[4846]: I1005 08:59:26.305068 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="beef8490-c87d-4459-ab70-ad20d33bfb63" containerName="extract-utilities" Oct 05 08:59:26 crc kubenswrapper[4846]: E1005 08:59:26.305100 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="beef8490-c87d-4459-ab70-ad20d33bfb63" containerName="extract-content" Oct 05 08:59:26 crc kubenswrapper[4846]: I1005 08:59:26.305106 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="beef8490-c87d-4459-ab70-ad20d33bfb63" containerName="extract-content" Oct 05 08:59:26 crc kubenswrapper[4846]: I1005 08:59:26.305315 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="2f3f1db8-f42b-4f2b-aeb3-05b9523aa4ea" containerName="libvirt-openstack-openstack-cell1" Oct 05 08:59:26 crc kubenswrapper[4846]: I1005 08:59:26.305335 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="beef8490-c87d-4459-ab70-ad20d33bfb63" containerName="registry-server" Oct 05 08:59:26 crc kubenswrapper[4846]: I1005 08:59:26.332691 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-openstack-openstack-cell1-kjczt"] Oct 05 08:59:26 crc kubenswrapper[4846]: I1005 08:59:26.332790 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-openstack-openstack-cell1-kjczt" Oct 05 08:59:26 crc kubenswrapper[4846]: I1005 08:59:26.336928 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"nova-cells-global-config" Oct 05 08:59:26 crc kubenswrapper[4846]: I1005 08:59:26.337169 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 05 08:59:26 crc kubenswrapper[4846]: I1005 08:59:26.337376 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Oct 05 08:59:26 crc kubenswrapper[4846]: I1005 08:59:26.337495 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-7687s" Oct 05 08:59:26 crc kubenswrapper[4846]: I1005 08:59:26.337552 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-compute-config" Oct 05 08:59:26 crc kubenswrapper[4846]: I1005 08:59:26.337737 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-migration-ssh-key" Oct 05 08:59:26 crc kubenswrapper[4846]: I1005 08:59:26.337890 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Oct 05 08:59:26 crc kubenswrapper[4846]: I1005 08:59:26.413808 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6b81e192-94b8-4a4b-918b-feab5d5a7860-ssh-key\") pod \"nova-cell1-openstack-openstack-cell1-kjczt\" (UID: \"6b81e192-94b8-4a4b-918b-feab5d5a7860\") " pod="openstack/nova-cell1-openstack-openstack-cell1-kjczt" Oct 05 08:59:26 crc kubenswrapper[4846]: I1005 08:59:26.413867 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/6b81e192-94b8-4a4b-918b-feab5d5a7860-nova-migration-ssh-key-1\") pod \"nova-cell1-openstack-openstack-cell1-kjczt\" (UID: \"6b81e192-94b8-4a4b-918b-feab5d5a7860\") " pod="openstack/nova-cell1-openstack-openstack-cell1-kjczt" Oct 05 08:59:26 crc kubenswrapper[4846]: I1005 08:59:26.413912 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b81e192-94b8-4a4b-918b-feab5d5a7860-nova-cell1-combined-ca-bundle\") pod \"nova-cell1-openstack-openstack-cell1-kjczt\" (UID: \"6b81e192-94b8-4a4b-918b-feab5d5a7860\") " pod="openstack/nova-cell1-openstack-openstack-cell1-kjczt" Oct 05 08:59:26 crc kubenswrapper[4846]: I1005 08:59:26.414009 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/6b81e192-94b8-4a4b-918b-feab5d5a7860-nova-cells-global-config-0\") pod \"nova-cell1-openstack-openstack-cell1-kjczt\" (UID: \"6b81e192-94b8-4a4b-918b-feab5d5a7860\") " pod="openstack/nova-cell1-openstack-openstack-cell1-kjczt" Oct 05 08:59:26 crc kubenswrapper[4846]: I1005 08:59:26.414066 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6b81e192-94b8-4a4b-918b-feab5d5a7860-inventory\") pod \"nova-cell1-openstack-openstack-cell1-kjczt\" (UID: \"6b81e192-94b8-4a4b-918b-feab5d5a7860\") " pod="openstack/nova-cell1-openstack-openstack-cell1-kjczt" Oct 05 08:59:26 crc kubenswrapper[4846]: I1005 08:59:26.414198 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4z4nq\" (UniqueName: \"kubernetes.io/projected/6b81e192-94b8-4a4b-918b-feab5d5a7860-kube-api-access-4z4nq\") pod \"nova-cell1-openstack-openstack-cell1-kjczt\" (UID: \"6b81e192-94b8-4a4b-918b-feab5d5a7860\") " pod="openstack/nova-cell1-openstack-openstack-cell1-kjczt" Oct 05 08:59:26 crc kubenswrapper[4846]: I1005 08:59:26.414222 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/6b81e192-94b8-4a4b-918b-feab5d5a7860-nova-cell1-compute-config-1\") pod \"nova-cell1-openstack-openstack-cell1-kjczt\" (UID: \"6b81e192-94b8-4a4b-918b-feab5d5a7860\") " pod="openstack/nova-cell1-openstack-openstack-cell1-kjczt" Oct 05 08:59:26 crc kubenswrapper[4846]: I1005 08:59:26.414380 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/6b81e192-94b8-4a4b-918b-feab5d5a7860-nova-cell1-compute-config-0\") pod \"nova-cell1-openstack-openstack-cell1-kjczt\" (UID: \"6b81e192-94b8-4a4b-918b-feab5d5a7860\") " pod="openstack/nova-cell1-openstack-openstack-cell1-kjczt" Oct 05 08:59:26 crc kubenswrapper[4846]: I1005 08:59:26.414447 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/6b81e192-94b8-4a4b-918b-feab5d5a7860-nova-migration-ssh-key-0\") pod \"nova-cell1-openstack-openstack-cell1-kjczt\" (UID: \"6b81e192-94b8-4a4b-918b-feab5d5a7860\") " pod="openstack/nova-cell1-openstack-openstack-cell1-kjczt" Oct 05 08:59:26 crc kubenswrapper[4846]: I1005 08:59:26.510549 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="beef8490-c87d-4459-ab70-ad20d33bfb63" path="/var/lib/kubelet/pods/beef8490-c87d-4459-ab70-ad20d33bfb63/volumes" Oct 05 08:59:26 crc kubenswrapper[4846]: I1005 08:59:26.516617 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4z4nq\" (UniqueName: \"kubernetes.io/projected/6b81e192-94b8-4a4b-918b-feab5d5a7860-kube-api-access-4z4nq\") pod \"nova-cell1-openstack-openstack-cell1-kjczt\" (UID: \"6b81e192-94b8-4a4b-918b-feab5d5a7860\") " pod="openstack/nova-cell1-openstack-openstack-cell1-kjczt" Oct 05 08:59:26 crc kubenswrapper[4846]: I1005 08:59:26.517107 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/6b81e192-94b8-4a4b-918b-feab5d5a7860-nova-cell1-compute-config-1\") pod \"nova-cell1-openstack-openstack-cell1-kjczt\" (UID: \"6b81e192-94b8-4a4b-918b-feab5d5a7860\") " pod="openstack/nova-cell1-openstack-openstack-cell1-kjczt" Oct 05 08:59:26 crc kubenswrapper[4846]: I1005 08:59:26.517960 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/6b81e192-94b8-4a4b-918b-feab5d5a7860-nova-cell1-compute-config-0\") pod \"nova-cell1-openstack-openstack-cell1-kjczt\" (UID: \"6b81e192-94b8-4a4b-918b-feab5d5a7860\") " pod="openstack/nova-cell1-openstack-openstack-cell1-kjczt" Oct 05 08:59:26 crc kubenswrapper[4846]: I1005 08:59:26.518017 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/6b81e192-94b8-4a4b-918b-feab5d5a7860-nova-migration-ssh-key-0\") pod \"nova-cell1-openstack-openstack-cell1-kjczt\" (UID: \"6b81e192-94b8-4a4b-918b-feab5d5a7860\") " pod="openstack/nova-cell1-openstack-openstack-cell1-kjczt" Oct 05 08:59:26 crc kubenswrapper[4846]: I1005 08:59:26.518128 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6b81e192-94b8-4a4b-918b-feab5d5a7860-ssh-key\") pod \"nova-cell1-openstack-openstack-cell1-kjczt\" (UID: \"6b81e192-94b8-4a4b-918b-feab5d5a7860\") " pod="openstack/nova-cell1-openstack-openstack-cell1-kjczt" Oct 05 08:59:26 crc kubenswrapper[4846]: I1005 08:59:26.518536 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/6b81e192-94b8-4a4b-918b-feab5d5a7860-nova-migration-ssh-key-1\") pod \"nova-cell1-openstack-openstack-cell1-kjczt\" (UID: \"6b81e192-94b8-4a4b-918b-feab5d5a7860\") " pod="openstack/nova-cell1-openstack-openstack-cell1-kjczt" Oct 05 08:59:26 crc kubenswrapper[4846]: I1005 08:59:26.518594 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b81e192-94b8-4a4b-918b-feab5d5a7860-nova-cell1-combined-ca-bundle\") pod \"nova-cell1-openstack-openstack-cell1-kjczt\" (UID: \"6b81e192-94b8-4a4b-918b-feab5d5a7860\") " pod="openstack/nova-cell1-openstack-openstack-cell1-kjczt" Oct 05 08:59:26 crc kubenswrapper[4846]: I1005 08:59:26.518637 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/6b81e192-94b8-4a4b-918b-feab5d5a7860-nova-cells-global-config-0\") pod \"nova-cell1-openstack-openstack-cell1-kjczt\" (UID: \"6b81e192-94b8-4a4b-918b-feab5d5a7860\") " pod="openstack/nova-cell1-openstack-openstack-cell1-kjczt" Oct 05 08:59:26 crc kubenswrapper[4846]: I1005 08:59:26.518664 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6b81e192-94b8-4a4b-918b-feab5d5a7860-inventory\") pod \"nova-cell1-openstack-openstack-cell1-kjczt\" (UID: \"6b81e192-94b8-4a4b-918b-feab5d5a7860\") " pod="openstack/nova-cell1-openstack-openstack-cell1-kjczt" Oct 05 08:59:26 crc kubenswrapper[4846]: I1005 08:59:26.525462 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/6b81e192-94b8-4a4b-918b-feab5d5a7860-nova-cell1-compute-config-0\") pod \"nova-cell1-openstack-openstack-cell1-kjczt\" (UID: \"6b81e192-94b8-4a4b-918b-feab5d5a7860\") " pod="openstack/nova-cell1-openstack-openstack-cell1-kjczt" Oct 05 08:59:26 crc kubenswrapper[4846]: I1005 08:59:26.529905 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/6b81e192-94b8-4a4b-918b-feab5d5a7860-nova-cell1-compute-config-1\") pod \"nova-cell1-openstack-openstack-cell1-kjczt\" (UID: \"6b81e192-94b8-4a4b-918b-feab5d5a7860\") " pod="openstack/nova-cell1-openstack-openstack-cell1-kjczt" Oct 05 08:59:26 crc kubenswrapper[4846]: I1005 08:59:26.529927 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/6b81e192-94b8-4a4b-918b-feab5d5a7860-nova-migration-ssh-key-0\") pod \"nova-cell1-openstack-openstack-cell1-kjczt\" (UID: \"6b81e192-94b8-4a4b-918b-feab5d5a7860\") " pod="openstack/nova-cell1-openstack-openstack-cell1-kjczt" Oct 05 08:59:26 crc kubenswrapper[4846]: I1005 08:59:26.530122 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6b81e192-94b8-4a4b-918b-feab5d5a7860-inventory\") pod \"nova-cell1-openstack-openstack-cell1-kjczt\" (UID: \"6b81e192-94b8-4a4b-918b-feab5d5a7860\") " pod="openstack/nova-cell1-openstack-openstack-cell1-kjczt" Oct 05 08:59:26 crc kubenswrapper[4846]: I1005 08:59:26.530789 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/6b81e192-94b8-4a4b-918b-feab5d5a7860-nova-cells-global-config-0\") pod \"nova-cell1-openstack-openstack-cell1-kjczt\" (UID: \"6b81e192-94b8-4a4b-918b-feab5d5a7860\") " pod="openstack/nova-cell1-openstack-openstack-cell1-kjczt" Oct 05 08:59:26 crc kubenswrapper[4846]: I1005 08:59:26.531465 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6b81e192-94b8-4a4b-918b-feab5d5a7860-ssh-key\") pod \"nova-cell1-openstack-openstack-cell1-kjczt\" (UID: \"6b81e192-94b8-4a4b-918b-feab5d5a7860\") " pod="openstack/nova-cell1-openstack-openstack-cell1-kjczt" Oct 05 08:59:26 crc kubenswrapper[4846]: I1005 08:59:26.534959 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/6b81e192-94b8-4a4b-918b-feab5d5a7860-nova-migration-ssh-key-1\") pod \"nova-cell1-openstack-openstack-cell1-kjczt\" (UID: \"6b81e192-94b8-4a4b-918b-feab5d5a7860\") " pod="openstack/nova-cell1-openstack-openstack-cell1-kjczt" Oct 05 08:59:26 crc kubenswrapper[4846]: I1005 08:59:26.535723 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b81e192-94b8-4a4b-918b-feab5d5a7860-nova-cell1-combined-ca-bundle\") pod \"nova-cell1-openstack-openstack-cell1-kjczt\" (UID: \"6b81e192-94b8-4a4b-918b-feab5d5a7860\") " pod="openstack/nova-cell1-openstack-openstack-cell1-kjczt" Oct 05 08:59:26 crc kubenswrapper[4846]: I1005 08:59:26.556625 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4z4nq\" (UniqueName: \"kubernetes.io/projected/6b81e192-94b8-4a4b-918b-feab5d5a7860-kube-api-access-4z4nq\") pod \"nova-cell1-openstack-openstack-cell1-kjczt\" (UID: \"6b81e192-94b8-4a4b-918b-feab5d5a7860\") " pod="openstack/nova-cell1-openstack-openstack-cell1-kjczt" Oct 05 08:59:26 crc kubenswrapper[4846]: I1005 08:59:26.682470 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-openstack-openstack-cell1-kjczt" Oct 05 08:59:27 crc kubenswrapper[4846]: I1005 08:59:27.233005 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-openstack-openstack-cell1-kjczt"] Oct 05 08:59:28 crc kubenswrapper[4846]: I1005 08:59:28.248385 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-openstack-openstack-cell1-kjczt" event={"ID":"6b81e192-94b8-4a4b-918b-feab5d5a7860","Type":"ContainerStarted","Data":"713c39f65f205843e64c0b8e05f448abc94fe0e34b8bbca868742e22d623cd67"} Oct 05 08:59:28 crc kubenswrapper[4846]: I1005 08:59:28.248685 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-openstack-openstack-cell1-kjczt" event={"ID":"6b81e192-94b8-4a4b-918b-feab5d5a7860","Type":"ContainerStarted","Data":"c378c0a0013f39a4dfe65c7fc3e538963b2c1a5d22b38e73b759914d9e516fe0"} Oct 05 08:59:28 crc kubenswrapper[4846]: I1005 08:59:28.264589 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-openstack-openstack-cell1-kjczt" podStartSLOduration=1.818317069 podStartE2EDuration="2.264564989s" podCreationTimestamp="2025-10-05 08:59:26 +0000 UTC" firstStartedPulling="2025-10-05 08:59:27.241714615 +0000 UTC m=+7889.482567390" lastFinishedPulling="2025-10-05 08:59:27.687962525 +0000 UTC m=+7889.928815310" observedRunningTime="2025-10-05 08:59:28.262477774 +0000 UTC m=+7890.503330569" watchObservedRunningTime="2025-10-05 08:59:28.264564989 +0000 UTC m=+7890.505417784" Oct 05 08:59:36 crc kubenswrapper[4846]: I1005 08:59:36.498148 4846 scope.go:117] "RemoveContainer" containerID="e6df7a4d2c7027b882a7e838679edd7570422c9a8553ab74b0bddbf3e4bf4b45" Oct 05 08:59:36 crc kubenswrapper[4846]: E1005 08:59:36.499120 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 08:59:49 crc kubenswrapper[4846]: I1005 08:59:49.498162 4846 scope.go:117] "RemoveContainer" containerID="e6df7a4d2c7027b882a7e838679edd7570422c9a8553ab74b0bddbf3e4bf4b45" Oct 05 08:59:49 crc kubenswrapper[4846]: E1005 08:59:49.499288 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 09:00:00 crc kubenswrapper[4846]: I1005 09:00:00.163108 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29327580-shk64"] Oct 05 09:00:00 crc kubenswrapper[4846]: I1005 09:00:00.167054 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29327580-shk64" Oct 05 09:00:00 crc kubenswrapper[4846]: I1005 09:00:00.170435 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 05 09:00:00 crc kubenswrapper[4846]: I1005 09:00:00.170879 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 05 09:00:00 crc kubenswrapper[4846]: I1005 09:00:00.188272 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29327580-shk64"] Oct 05 09:00:00 crc kubenswrapper[4846]: I1005 09:00:00.283078 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b885295e-73f2-4fc4-8c83-149b88556acd-config-volume\") pod \"collect-profiles-29327580-shk64\" (UID: \"b885295e-73f2-4fc4-8c83-149b88556acd\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327580-shk64" Oct 05 09:00:00 crc kubenswrapper[4846]: I1005 09:00:00.283203 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b885295e-73f2-4fc4-8c83-149b88556acd-secret-volume\") pod \"collect-profiles-29327580-shk64\" (UID: \"b885295e-73f2-4fc4-8c83-149b88556acd\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327580-shk64" Oct 05 09:00:00 crc kubenswrapper[4846]: I1005 09:00:00.283495 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2lwmd\" (UniqueName: \"kubernetes.io/projected/b885295e-73f2-4fc4-8c83-149b88556acd-kube-api-access-2lwmd\") pod \"collect-profiles-29327580-shk64\" (UID: \"b885295e-73f2-4fc4-8c83-149b88556acd\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327580-shk64" Oct 05 09:00:00 crc kubenswrapper[4846]: I1005 09:00:00.385748 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b885295e-73f2-4fc4-8c83-149b88556acd-config-volume\") pod \"collect-profiles-29327580-shk64\" (UID: \"b885295e-73f2-4fc4-8c83-149b88556acd\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327580-shk64" Oct 05 09:00:00 crc kubenswrapper[4846]: I1005 09:00:00.385992 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b885295e-73f2-4fc4-8c83-149b88556acd-secret-volume\") pod \"collect-profiles-29327580-shk64\" (UID: \"b885295e-73f2-4fc4-8c83-149b88556acd\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327580-shk64" Oct 05 09:00:00 crc kubenswrapper[4846]: I1005 09:00:00.386073 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2lwmd\" (UniqueName: \"kubernetes.io/projected/b885295e-73f2-4fc4-8c83-149b88556acd-kube-api-access-2lwmd\") pod \"collect-profiles-29327580-shk64\" (UID: \"b885295e-73f2-4fc4-8c83-149b88556acd\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327580-shk64" Oct 05 09:00:00 crc kubenswrapper[4846]: I1005 09:00:00.387108 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b885295e-73f2-4fc4-8c83-149b88556acd-config-volume\") pod \"collect-profiles-29327580-shk64\" (UID: \"b885295e-73f2-4fc4-8c83-149b88556acd\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327580-shk64" Oct 05 09:00:00 crc kubenswrapper[4846]: I1005 09:00:00.393528 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b885295e-73f2-4fc4-8c83-149b88556acd-secret-volume\") pod \"collect-profiles-29327580-shk64\" (UID: \"b885295e-73f2-4fc4-8c83-149b88556acd\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327580-shk64" Oct 05 09:00:00 crc kubenswrapper[4846]: I1005 09:00:00.405935 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2lwmd\" (UniqueName: \"kubernetes.io/projected/b885295e-73f2-4fc4-8c83-149b88556acd-kube-api-access-2lwmd\") pod \"collect-profiles-29327580-shk64\" (UID: \"b885295e-73f2-4fc4-8c83-149b88556acd\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327580-shk64" Oct 05 09:00:00 crc kubenswrapper[4846]: I1005 09:00:00.497845 4846 scope.go:117] "RemoveContainer" containerID="e6df7a4d2c7027b882a7e838679edd7570422c9a8553ab74b0bddbf3e4bf4b45" Oct 05 09:00:00 crc kubenswrapper[4846]: E1005 09:00:00.498165 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 09:00:00 crc kubenswrapper[4846]: I1005 09:00:00.503715 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29327580-shk64" Oct 05 09:00:01 crc kubenswrapper[4846]: I1005 09:00:01.003010 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29327580-shk64"] Oct 05 09:00:01 crc kubenswrapper[4846]: W1005 09:00:01.012066 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb885295e_73f2_4fc4_8c83_149b88556acd.slice/crio-7b3f067e3832188717990859f7be88d92c54f5a12154847048a9a8c764d90ce1 WatchSource:0}: Error finding container 7b3f067e3832188717990859f7be88d92c54f5a12154847048a9a8c764d90ce1: Status 404 returned error can't find the container with id 7b3f067e3832188717990859f7be88d92c54f5a12154847048a9a8c764d90ce1 Oct 05 09:00:01 crc kubenswrapper[4846]: I1005 09:00:01.636220 4846 generic.go:334] "Generic (PLEG): container finished" podID="b885295e-73f2-4fc4-8c83-149b88556acd" containerID="24d95c8a26fdaa4b41a0694536bd8091b81e5b92cd86f9eaa23e21b644f20d45" exitCode=0 Oct 05 09:00:01 crc kubenswrapper[4846]: I1005 09:00:01.636611 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29327580-shk64" event={"ID":"b885295e-73f2-4fc4-8c83-149b88556acd","Type":"ContainerDied","Data":"24d95c8a26fdaa4b41a0694536bd8091b81e5b92cd86f9eaa23e21b644f20d45"} Oct 05 09:00:01 crc kubenswrapper[4846]: I1005 09:00:01.636652 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29327580-shk64" event={"ID":"b885295e-73f2-4fc4-8c83-149b88556acd","Type":"ContainerStarted","Data":"7b3f067e3832188717990859f7be88d92c54f5a12154847048a9a8c764d90ce1"} Oct 05 09:00:03 crc kubenswrapper[4846]: I1005 09:00:03.005029 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29327580-shk64" Oct 05 09:00:03 crc kubenswrapper[4846]: I1005 09:00:03.040980 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b885295e-73f2-4fc4-8c83-149b88556acd-config-volume\") pod \"b885295e-73f2-4fc4-8c83-149b88556acd\" (UID: \"b885295e-73f2-4fc4-8c83-149b88556acd\") " Oct 05 09:00:03 crc kubenswrapper[4846]: I1005 09:00:03.041232 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b885295e-73f2-4fc4-8c83-149b88556acd-secret-volume\") pod \"b885295e-73f2-4fc4-8c83-149b88556acd\" (UID: \"b885295e-73f2-4fc4-8c83-149b88556acd\") " Oct 05 09:00:03 crc kubenswrapper[4846]: I1005 09:00:03.041532 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2lwmd\" (UniqueName: \"kubernetes.io/projected/b885295e-73f2-4fc4-8c83-149b88556acd-kube-api-access-2lwmd\") pod \"b885295e-73f2-4fc4-8c83-149b88556acd\" (UID: \"b885295e-73f2-4fc4-8c83-149b88556acd\") " Oct 05 09:00:03 crc kubenswrapper[4846]: I1005 09:00:03.042504 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b885295e-73f2-4fc4-8c83-149b88556acd-config-volume" (OuterVolumeSpecName: "config-volume") pod "b885295e-73f2-4fc4-8c83-149b88556acd" (UID: "b885295e-73f2-4fc4-8c83-149b88556acd"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 09:00:03 crc kubenswrapper[4846]: I1005 09:00:03.051690 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b885295e-73f2-4fc4-8c83-149b88556acd-kube-api-access-2lwmd" (OuterVolumeSpecName: "kube-api-access-2lwmd") pod "b885295e-73f2-4fc4-8c83-149b88556acd" (UID: "b885295e-73f2-4fc4-8c83-149b88556acd"). InnerVolumeSpecName "kube-api-access-2lwmd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 09:00:03 crc kubenswrapper[4846]: I1005 09:00:03.055246 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b885295e-73f2-4fc4-8c83-149b88556acd-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "b885295e-73f2-4fc4-8c83-149b88556acd" (UID: "b885295e-73f2-4fc4-8c83-149b88556acd"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 09:00:03 crc kubenswrapper[4846]: I1005 09:00:03.145797 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2lwmd\" (UniqueName: \"kubernetes.io/projected/b885295e-73f2-4fc4-8c83-149b88556acd-kube-api-access-2lwmd\") on node \"crc\" DevicePath \"\"" Oct 05 09:00:03 crc kubenswrapper[4846]: I1005 09:00:03.145998 4846 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b885295e-73f2-4fc4-8c83-149b88556acd-config-volume\") on node \"crc\" DevicePath \"\"" Oct 05 09:00:03 crc kubenswrapper[4846]: I1005 09:00:03.146100 4846 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b885295e-73f2-4fc4-8c83-149b88556acd-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 05 09:00:03 crc kubenswrapper[4846]: I1005 09:00:03.663270 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29327580-shk64" event={"ID":"b885295e-73f2-4fc4-8c83-149b88556acd","Type":"ContainerDied","Data":"7b3f067e3832188717990859f7be88d92c54f5a12154847048a9a8c764d90ce1"} Oct 05 09:00:03 crc kubenswrapper[4846]: I1005 09:00:03.663333 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7b3f067e3832188717990859f7be88d92c54f5a12154847048a9a8c764d90ce1" Oct 05 09:00:03 crc kubenswrapper[4846]: I1005 09:00:03.663364 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29327580-shk64" Oct 05 09:00:04 crc kubenswrapper[4846]: I1005 09:00:04.095415 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29327535-9qgh9"] Oct 05 09:00:04 crc kubenswrapper[4846]: I1005 09:00:04.105806 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29327535-9qgh9"] Oct 05 09:00:04 crc kubenswrapper[4846]: I1005 09:00:04.520140 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5612d2ab-7b8a-452c-8b97-8a968620b14f" path="/var/lib/kubelet/pods/5612d2ab-7b8a-452c-8b97-8a968620b14f/volumes" Oct 05 09:00:11 crc kubenswrapper[4846]: I1005 09:00:11.498307 4846 scope.go:117] "RemoveContainer" containerID="e6df7a4d2c7027b882a7e838679edd7570422c9a8553ab74b0bddbf3e4bf4b45" Oct 05 09:00:11 crc kubenswrapper[4846]: E1005 09:00:11.499633 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 09:00:25 crc kubenswrapper[4846]: I1005 09:00:25.498658 4846 scope.go:117] "RemoveContainer" containerID="e6df7a4d2c7027b882a7e838679edd7570422c9a8553ab74b0bddbf3e4bf4b45" Oct 05 09:00:25 crc kubenswrapper[4846]: E1005 09:00:25.499760 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 09:00:36 crc kubenswrapper[4846]: I1005 09:00:36.498098 4846 scope.go:117] "RemoveContainer" containerID="e6df7a4d2c7027b882a7e838679edd7570422c9a8553ab74b0bddbf3e4bf4b45" Oct 05 09:00:36 crc kubenswrapper[4846]: E1005 09:00:36.498879 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 09:00:38 crc kubenswrapper[4846]: I1005 09:00:38.577496 4846 scope.go:117] "RemoveContainer" containerID="38fc1323ac49b5bbf03cda77705f78df164686147e3bd467d7059633eddc03e6" Oct 05 09:00:50 crc kubenswrapper[4846]: I1005 09:00:50.499034 4846 scope.go:117] "RemoveContainer" containerID="e6df7a4d2c7027b882a7e838679edd7570422c9a8553ab74b0bddbf3e4bf4b45" Oct 05 09:00:50 crc kubenswrapper[4846]: E1005 09:00:50.500595 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 09:01:00 crc kubenswrapper[4846]: I1005 09:01:00.176224 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-cron-29327581-f9wqf"] Oct 05 09:01:00 crc kubenswrapper[4846]: E1005 09:01:00.177536 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b885295e-73f2-4fc4-8c83-149b88556acd" containerName="collect-profiles" Oct 05 09:01:00 crc kubenswrapper[4846]: I1005 09:01:00.177560 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="b885295e-73f2-4fc4-8c83-149b88556acd" containerName="collect-profiles" Oct 05 09:01:00 crc kubenswrapper[4846]: I1005 09:01:00.177928 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="b885295e-73f2-4fc4-8c83-149b88556acd" containerName="collect-profiles" Oct 05 09:01:00 crc kubenswrapper[4846]: I1005 09:01:00.179218 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29327581-f9wqf" Oct 05 09:01:00 crc kubenswrapper[4846]: I1005 09:01:00.190779 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29327581-f9wqf"] Oct 05 09:01:00 crc kubenswrapper[4846]: I1005 09:01:00.243963 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/9cb75497-a34f-4ff3-8140-4be43217b17f-fernet-keys\") pod \"keystone-cron-29327581-f9wqf\" (UID: \"9cb75497-a34f-4ff3-8140-4be43217b17f\") " pod="openstack/keystone-cron-29327581-f9wqf" Oct 05 09:01:00 crc kubenswrapper[4846]: I1005 09:01:00.244139 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9cb75497-a34f-4ff3-8140-4be43217b17f-combined-ca-bundle\") pod \"keystone-cron-29327581-f9wqf\" (UID: \"9cb75497-a34f-4ff3-8140-4be43217b17f\") " pod="openstack/keystone-cron-29327581-f9wqf" Oct 05 09:01:00 crc kubenswrapper[4846]: I1005 09:01:00.244248 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bwr9h\" (UniqueName: \"kubernetes.io/projected/9cb75497-a34f-4ff3-8140-4be43217b17f-kube-api-access-bwr9h\") pod \"keystone-cron-29327581-f9wqf\" (UID: \"9cb75497-a34f-4ff3-8140-4be43217b17f\") " pod="openstack/keystone-cron-29327581-f9wqf" Oct 05 09:01:00 crc kubenswrapper[4846]: I1005 09:01:00.244323 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9cb75497-a34f-4ff3-8140-4be43217b17f-config-data\") pod \"keystone-cron-29327581-f9wqf\" (UID: \"9cb75497-a34f-4ff3-8140-4be43217b17f\") " pod="openstack/keystone-cron-29327581-f9wqf" Oct 05 09:01:00 crc kubenswrapper[4846]: I1005 09:01:00.345791 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/9cb75497-a34f-4ff3-8140-4be43217b17f-fernet-keys\") pod \"keystone-cron-29327581-f9wqf\" (UID: \"9cb75497-a34f-4ff3-8140-4be43217b17f\") " pod="openstack/keystone-cron-29327581-f9wqf" Oct 05 09:01:00 crc kubenswrapper[4846]: I1005 09:01:00.345851 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9cb75497-a34f-4ff3-8140-4be43217b17f-combined-ca-bundle\") pod \"keystone-cron-29327581-f9wqf\" (UID: \"9cb75497-a34f-4ff3-8140-4be43217b17f\") " pod="openstack/keystone-cron-29327581-f9wqf" Oct 05 09:01:00 crc kubenswrapper[4846]: I1005 09:01:00.345878 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bwr9h\" (UniqueName: \"kubernetes.io/projected/9cb75497-a34f-4ff3-8140-4be43217b17f-kube-api-access-bwr9h\") pod \"keystone-cron-29327581-f9wqf\" (UID: \"9cb75497-a34f-4ff3-8140-4be43217b17f\") " pod="openstack/keystone-cron-29327581-f9wqf" Oct 05 09:01:00 crc kubenswrapper[4846]: I1005 09:01:00.345897 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9cb75497-a34f-4ff3-8140-4be43217b17f-config-data\") pod \"keystone-cron-29327581-f9wqf\" (UID: \"9cb75497-a34f-4ff3-8140-4be43217b17f\") " pod="openstack/keystone-cron-29327581-f9wqf" Oct 05 09:01:00 crc kubenswrapper[4846]: I1005 09:01:00.354136 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9cb75497-a34f-4ff3-8140-4be43217b17f-combined-ca-bundle\") pod \"keystone-cron-29327581-f9wqf\" (UID: \"9cb75497-a34f-4ff3-8140-4be43217b17f\") " pod="openstack/keystone-cron-29327581-f9wqf" Oct 05 09:01:00 crc kubenswrapper[4846]: I1005 09:01:00.357755 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/9cb75497-a34f-4ff3-8140-4be43217b17f-fernet-keys\") pod \"keystone-cron-29327581-f9wqf\" (UID: \"9cb75497-a34f-4ff3-8140-4be43217b17f\") " pod="openstack/keystone-cron-29327581-f9wqf" Oct 05 09:01:00 crc kubenswrapper[4846]: I1005 09:01:00.369872 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9cb75497-a34f-4ff3-8140-4be43217b17f-config-data\") pod \"keystone-cron-29327581-f9wqf\" (UID: \"9cb75497-a34f-4ff3-8140-4be43217b17f\") " pod="openstack/keystone-cron-29327581-f9wqf" Oct 05 09:01:00 crc kubenswrapper[4846]: I1005 09:01:00.372212 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bwr9h\" (UniqueName: \"kubernetes.io/projected/9cb75497-a34f-4ff3-8140-4be43217b17f-kube-api-access-bwr9h\") pod \"keystone-cron-29327581-f9wqf\" (UID: \"9cb75497-a34f-4ff3-8140-4be43217b17f\") " pod="openstack/keystone-cron-29327581-f9wqf" Oct 05 09:01:00 crc kubenswrapper[4846]: I1005 09:01:00.525498 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29327581-f9wqf" Oct 05 09:01:01 crc kubenswrapper[4846]: I1005 09:01:01.054418 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29327581-f9wqf"] Oct 05 09:01:01 crc kubenswrapper[4846]: I1005 09:01:01.327166 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29327581-f9wqf" event={"ID":"9cb75497-a34f-4ff3-8140-4be43217b17f","Type":"ContainerStarted","Data":"3c1b79dcd489134ddc5f260d70724cfe8298f25f4815eed657931721c26fd4a3"} Oct 05 09:01:01 crc kubenswrapper[4846]: I1005 09:01:01.327641 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29327581-f9wqf" event={"ID":"9cb75497-a34f-4ff3-8140-4be43217b17f","Type":"ContainerStarted","Data":"c745c305fb4ab9a2340aa98c72b39522ab5dd37b1e85cd3db10b19ae197edcf8"} Oct 05 09:01:01 crc kubenswrapper[4846]: I1005 09:01:01.345663 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-cron-29327581-f9wqf" podStartSLOduration=1.345634233 podStartE2EDuration="1.345634233s" podCreationTimestamp="2025-10-05 09:01:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 09:01:01.345393126 +0000 UTC m=+7983.586245921" watchObservedRunningTime="2025-10-05 09:01:01.345634233 +0000 UTC m=+7983.586487018" Oct 05 09:01:04 crc kubenswrapper[4846]: I1005 09:01:04.363819 4846 generic.go:334] "Generic (PLEG): container finished" podID="9cb75497-a34f-4ff3-8140-4be43217b17f" containerID="3c1b79dcd489134ddc5f260d70724cfe8298f25f4815eed657931721c26fd4a3" exitCode=0 Oct 05 09:01:04 crc kubenswrapper[4846]: I1005 09:01:04.363884 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29327581-f9wqf" event={"ID":"9cb75497-a34f-4ff3-8140-4be43217b17f","Type":"ContainerDied","Data":"3c1b79dcd489134ddc5f260d70724cfe8298f25f4815eed657931721c26fd4a3"} Oct 05 09:01:05 crc kubenswrapper[4846]: I1005 09:01:05.497695 4846 scope.go:117] "RemoveContainer" containerID="e6df7a4d2c7027b882a7e838679edd7570422c9a8553ab74b0bddbf3e4bf4b45" Oct 05 09:01:05 crc kubenswrapper[4846]: E1005 09:01:05.498213 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 09:01:05 crc kubenswrapper[4846]: I1005 09:01:05.781425 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29327581-f9wqf" Oct 05 09:01:05 crc kubenswrapper[4846]: I1005 09:01:05.886654 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9cb75497-a34f-4ff3-8140-4be43217b17f-config-data\") pod \"9cb75497-a34f-4ff3-8140-4be43217b17f\" (UID: \"9cb75497-a34f-4ff3-8140-4be43217b17f\") " Oct 05 09:01:05 crc kubenswrapper[4846]: I1005 09:01:05.886777 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9cb75497-a34f-4ff3-8140-4be43217b17f-combined-ca-bundle\") pod \"9cb75497-a34f-4ff3-8140-4be43217b17f\" (UID: \"9cb75497-a34f-4ff3-8140-4be43217b17f\") " Oct 05 09:01:05 crc kubenswrapper[4846]: I1005 09:01:05.886918 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bwr9h\" (UniqueName: \"kubernetes.io/projected/9cb75497-a34f-4ff3-8140-4be43217b17f-kube-api-access-bwr9h\") pod \"9cb75497-a34f-4ff3-8140-4be43217b17f\" (UID: \"9cb75497-a34f-4ff3-8140-4be43217b17f\") " Oct 05 09:01:05 crc kubenswrapper[4846]: I1005 09:01:05.886938 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/9cb75497-a34f-4ff3-8140-4be43217b17f-fernet-keys\") pod \"9cb75497-a34f-4ff3-8140-4be43217b17f\" (UID: \"9cb75497-a34f-4ff3-8140-4be43217b17f\") " Oct 05 09:01:05 crc kubenswrapper[4846]: I1005 09:01:05.894140 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9cb75497-a34f-4ff3-8140-4be43217b17f-kube-api-access-bwr9h" (OuterVolumeSpecName: "kube-api-access-bwr9h") pod "9cb75497-a34f-4ff3-8140-4be43217b17f" (UID: "9cb75497-a34f-4ff3-8140-4be43217b17f"). InnerVolumeSpecName "kube-api-access-bwr9h". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 09:01:05 crc kubenswrapper[4846]: I1005 09:01:05.894509 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9cb75497-a34f-4ff3-8140-4be43217b17f-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "9cb75497-a34f-4ff3-8140-4be43217b17f" (UID: "9cb75497-a34f-4ff3-8140-4be43217b17f"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 09:01:05 crc kubenswrapper[4846]: I1005 09:01:05.939069 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9cb75497-a34f-4ff3-8140-4be43217b17f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9cb75497-a34f-4ff3-8140-4be43217b17f" (UID: "9cb75497-a34f-4ff3-8140-4be43217b17f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 09:01:05 crc kubenswrapper[4846]: I1005 09:01:05.948276 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9cb75497-a34f-4ff3-8140-4be43217b17f-config-data" (OuterVolumeSpecName: "config-data") pod "9cb75497-a34f-4ff3-8140-4be43217b17f" (UID: "9cb75497-a34f-4ff3-8140-4be43217b17f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 09:01:05 crc kubenswrapper[4846]: I1005 09:01:05.989697 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bwr9h\" (UniqueName: \"kubernetes.io/projected/9cb75497-a34f-4ff3-8140-4be43217b17f-kube-api-access-bwr9h\") on node \"crc\" DevicePath \"\"" Oct 05 09:01:05 crc kubenswrapper[4846]: I1005 09:01:05.989741 4846 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/9cb75497-a34f-4ff3-8140-4be43217b17f-fernet-keys\") on node \"crc\" DevicePath \"\"" Oct 05 09:01:05 crc kubenswrapper[4846]: I1005 09:01:05.989755 4846 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9cb75497-a34f-4ff3-8140-4be43217b17f-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 09:01:05 crc kubenswrapper[4846]: I1005 09:01:05.989766 4846 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9cb75497-a34f-4ff3-8140-4be43217b17f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 09:01:06 crc kubenswrapper[4846]: I1005 09:01:06.393032 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29327581-f9wqf" event={"ID":"9cb75497-a34f-4ff3-8140-4be43217b17f","Type":"ContainerDied","Data":"c745c305fb4ab9a2340aa98c72b39522ab5dd37b1e85cd3db10b19ae197edcf8"} Oct 05 09:01:06 crc kubenswrapper[4846]: I1005 09:01:06.393095 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c745c305fb4ab9a2340aa98c72b39522ab5dd37b1e85cd3db10b19ae197edcf8" Oct 05 09:01:06 crc kubenswrapper[4846]: I1005 09:01:06.393204 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29327581-f9wqf" Oct 05 09:01:16 crc kubenswrapper[4846]: I1005 09:01:16.498768 4846 scope.go:117] "RemoveContainer" containerID="e6df7a4d2c7027b882a7e838679edd7570422c9a8553ab74b0bddbf3e4bf4b45" Oct 05 09:01:16 crc kubenswrapper[4846]: E1005 09:01:16.500002 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 09:01:17 crc kubenswrapper[4846]: I1005 09:01:17.211549 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-rfkzt"] Oct 05 09:01:17 crc kubenswrapper[4846]: E1005 09:01:17.212862 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9cb75497-a34f-4ff3-8140-4be43217b17f" containerName="keystone-cron" Oct 05 09:01:17 crc kubenswrapper[4846]: I1005 09:01:17.213112 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="9cb75497-a34f-4ff3-8140-4be43217b17f" containerName="keystone-cron" Oct 05 09:01:17 crc kubenswrapper[4846]: I1005 09:01:17.213765 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="9cb75497-a34f-4ff3-8140-4be43217b17f" containerName="keystone-cron" Oct 05 09:01:17 crc kubenswrapper[4846]: I1005 09:01:17.216797 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rfkzt" Oct 05 09:01:17 crc kubenswrapper[4846]: I1005 09:01:17.232960 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-rfkzt"] Oct 05 09:01:17 crc kubenswrapper[4846]: I1005 09:01:17.291958 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aed1876b-77bb-4605-920d-5a64dc4d0130-utilities\") pod \"certified-operators-rfkzt\" (UID: \"aed1876b-77bb-4605-920d-5a64dc4d0130\") " pod="openshift-marketplace/certified-operators-rfkzt" Oct 05 09:01:17 crc kubenswrapper[4846]: I1005 09:01:17.292289 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4p7s4\" (UniqueName: \"kubernetes.io/projected/aed1876b-77bb-4605-920d-5a64dc4d0130-kube-api-access-4p7s4\") pod \"certified-operators-rfkzt\" (UID: \"aed1876b-77bb-4605-920d-5a64dc4d0130\") " pod="openshift-marketplace/certified-operators-rfkzt" Oct 05 09:01:17 crc kubenswrapper[4846]: I1005 09:01:17.292467 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aed1876b-77bb-4605-920d-5a64dc4d0130-catalog-content\") pod \"certified-operators-rfkzt\" (UID: \"aed1876b-77bb-4605-920d-5a64dc4d0130\") " pod="openshift-marketplace/certified-operators-rfkzt" Oct 05 09:01:17 crc kubenswrapper[4846]: I1005 09:01:17.394214 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aed1876b-77bb-4605-920d-5a64dc4d0130-catalog-content\") pod \"certified-operators-rfkzt\" (UID: \"aed1876b-77bb-4605-920d-5a64dc4d0130\") " pod="openshift-marketplace/certified-operators-rfkzt" Oct 05 09:01:17 crc kubenswrapper[4846]: I1005 09:01:17.394312 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aed1876b-77bb-4605-920d-5a64dc4d0130-utilities\") pod \"certified-operators-rfkzt\" (UID: \"aed1876b-77bb-4605-920d-5a64dc4d0130\") " pod="openshift-marketplace/certified-operators-rfkzt" Oct 05 09:01:17 crc kubenswrapper[4846]: I1005 09:01:17.394370 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4p7s4\" (UniqueName: \"kubernetes.io/projected/aed1876b-77bb-4605-920d-5a64dc4d0130-kube-api-access-4p7s4\") pod \"certified-operators-rfkzt\" (UID: \"aed1876b-77bb-4605-920d-5a64dc4d0130\") " pod="openshift-marketplace/certified-operators-rfkzt" Oct 05 09:01:17 crc kubenswrapper[4846]: I1005 09:01:17.395361 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aed1876b-77bb-4605-920d-5a64dc4d0130-utilities\") pod \"certified-operators-rfkzt\" (UID: \"aed1876b-77bb-4605-920d-5a64dc4d0130\") " pod="openshift-marketplace/certified-operators-rfkzt" Oct 05 09:01:17 crc kubenswrapper[4846]: I1005 09:01:17.395361 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aed1876b-77bb-4605-920d-5a64dc4d0130-catalog-content\") pod \"certified-operators-rfkzt\" (UID: \"aed1876b-77bb-4605-920d-5a64dc4d0130\") " pod="openshift-marketplace/certified-operators-rfkzt" Oct 05 09:01:17 crc kubenswrapper[4846]: I1005 09:01:17.418557 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4p7s4\" (UniqueName: \"kubernetes.io/projected/aed1876b-77bb-4605-920d-5a64dc4d0130-kube-api-access-4p7s4\") pod \"certified-operators-rfkzt\" (UID: \"aed1876b-77bb-4605-920d-5a64dc4d0130\") " pod="openshift-marketplace/certified-operators-rfkzt" Oct 05 09:01:17 crc kubenswrapper[4846]: I1005 09:01:17.561131 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rfkzt" Oct 05 09:01:18 crc kubenswrapper[4846]: I1005 09:01:18.169272 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-rfkzt"] Oct 05 09:01:18 crc kubenswrapper[4846]: I1005 09:01:18.547876 4846 generic.go:334] "Generic (PLEG): container finished" podID="aed1876b-77bb-4605-920d-5a64dc4d0130" containerID="42bb399da018b1e3c5a40e0692e38cf759d61e3c2bc8e4176f5fb12ec0c6c6e0" exitCode=0 Oct 05 09:01:18 crc kubenswrapper[4846]: I1005 09:01:18.548377 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rfkzt" event={"ID":"aed1876b-77bb-4605-920d-5a64dc4d0130","Type":"ContainerDied","Data":"42bb399da018b1e3c5a40e0692e38cf759d61e3c2bc8e4176f5fb12ec0c6c6e0"} Oct 05 09:01:18 crc kubenswrapper[4846]: I1005 09:01:18.548442 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rfkzt" event={"ID":"aed1876b-77bb-4605-920d-5a64dc4d0130","Type":"ContainerStarted","Data":"e579f7fe3b4d144bb6dbb76c9e422f5cdf313a3d7a653d832b85cc64395918c8"} Oct 05 09:01:19 crc kubenswrapper[4846]: I1005 09:01:19.387056 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-krgq6"] Oct 05 09:01:19 crc kubenswrapper[4846]: I1005 09:01:19.392642 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-krgq6" Oct 05 09:01:19 crc kubenswrapper[4846]: I1005 09:01:19.400771 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-krgq6"] Oct 05 09:01:19 crc kubenswrapper[4846]: I1005 09:01:19.443921 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/62003eb9-c4da-4b5d-aac5-5ee422c2f46f-catalog-content\") pod \"redhat-operators-krgq6\" (UID: \"62003eb9-c4da-4b5d-aac5-5ee422c2f46f\") " pod="openshift-marketplace/redhat-operators-krgq6" Oct 05 09:01:19 crc kubenswrapper[4846]: I1005 09:01:19.444090 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9wjjk\" (UniqueName: \"kubernetes.io/projected/62003eb9-c4da-4b5d-aac5-5ee422c2f46f-kube-api-access-9wjjk\") pod \"redhat-operators-krgq6\" (UID: \"62003eb9-c4da-4b5d-aac5-5ee422c2f46f\") " pod="openshift-marketplace/redhat-operators-krgq6" Oct 05 09:01:19 crc kubenswrapper[4846]: I1005 09:01:19.444246 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/62003eb9-c4da-4b5d-aac5-5ee422c2f46f-utilities\") pod \"redhat-operators-krgq6\" (UID: \"62003eb9-c4da-4b5d-aac5-5ee422c2f46f\") " pod="openshift-marketplace/redhat-operators-krgq6" Oct 05 09:01:19 crc kubenswrapper[4846]: I1005 09:01:19.548091 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/62003eb9-c4da-4b5d-aac5-5ee422c2f46f-catalog-content\") pod \"redhat-operators-krgq6\" (UID: \"62003eb9-c4da-4b5d-aac5-5ee422c2f46f\") " pod="openshift-marketplace/redhat-operators-krgq6" Oct 05 09:01:19 crc kubenswrapper[4846]: I1005 09:01:19.549030 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/62003eb9-c4da-4b5d-aac5-5ee422c2f46f-catalog-content\") pod \"redhat-operators-krgq6\" (UID: \"62003eb9-c4da-4b5d-aac5-5ee422c2f46f\") " pod="openshift-marketplace/redhat-operators-krgq6" Oct 05 09:01:19 crc kubenswrapper[4846]: I1005 09:01:19.549117 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9wjjk\" (UniqueName: \"kubernetes.io/projected/62003eb9-c4da-4b5d-aac5-5ee422c2f46f-kube-api-access-9wjjk\") pod \"redhat-operators-krgq6\" (UID: \"62003eb9-c4da-4b5d-aac5-5ee422c2f46f\") " pod="openshift-marketplace/redhat-operators-krgq6" Oct 05 09:01:19 crc kubenswrapper[4846]: I1005 09:01:19.549249 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/62003eb9-c4da-4b5d-aac5-5ee422c2f46f-utilities\") pod \"redhat-operators-krgq6\" (UID: \"62003eb9-c4da-4b5d-aac5-5ee422c2f46f\") " pod="openshift-marketplace/redhat-operators-krgq6" Oct 05 09:01:19 crc kubenswrapper[4846]: I1005 09:01:19.549732 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/62003eb9-c4da-4b5d-aac5-5ee422c2f46f-utilities\") pod \"redhat-operators-krgq6\" (UID: \"62003eb9-c4da-4b5d-aac5-5ee422c2f46f\") " pod="openshift-marketplace/redhat-operators-krgq6" Oct 05 09:01:19 crc kubenswrapper[4846]: I1005 09:01:19.576103 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9wjjk\" (UniqueName: \"kubernetes.io/projected/62003eb9-c4da-4b5d-aac5-5ee422c2f46f-kube-api-access-9wjjk\") pod \"redhat-operators-krgq6\" (UID: \"62003eb9-c4da-4b5d-aac5-5ee422c2f46f\") " pod="openshift-marketplace/redhat-operators-krgq6" Oct 05 09:01:19 crc kubenswrapper[4846]: I1005 09:01:19.719736 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-krgq6" Oct 05 09:01:20 crc kubenswrapper[4846]: W1005 09:01:20.211362 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod62003eb9_c4da_4b5d_aac5_5ee422c2f46f.slice/crio-00f9e9874a02a20e7141ffab1ac1b5a14a3db1fa705c05a42798547d858d0466 WatchSource:0}: Error finding container 00f9e9874a02a20e7141ffab1ac1b5a14a3db1fa705c05a42798547d858d0466: Status 404 returned error can't find the container with id 00f9e9874a02a20e7141ffab1ac1b5a14a3db1fa705c05a42798547d858d0466 Oct 05 09:01:20 crc kubenswrapper[4846]: I1005 09:01:20.213669 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-krgq6"] Oct 05 09:01:20 crc kubenswrapper[4846]: I1005 09:01:20.570101 4846 generic.go:334] "Generic (PLEG): container finished" podID="62003eb9-c4da-4b5d-aac5-5ee422c2f46f" containerID="4e8e81e91c67642c54b3b57426b4268a70dd14ca2ce37213d0727c7330a019e8" exitCode=0 Oct 05 09:01:20 crc kubenswrapper[4846]: I1005 09:01:20.570155 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-krgq6" event={"ID":"62003eb9-c4da-4b5d-aac5-5ee422c2f46f","Type":"ContainerDied","Data":"4e8e81e91c67642c54b3b57426b4268a70dd14ca2ce37213d0727c7330a019e8"} Oct 05 09:01:20 crc kubenswrapper[4846]: I1005 09:01:20.570488 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-krgq6" event={"ID":"62003eb9-c4da-4b5d-aac5-5ee422c2f46f","Type":"ContainerStarted","Data":"00f9e9874a02a20e7141ffab1ac1b5a14a3db1fa705c05a42798547d858d0466"} Oct 05 09:01:20 crc kubenswrapper[4846]: I1005 09:01:20.576577 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rfkzt" event={"ID":"aed1876b-77bb-4605-920d-5a64dc4d0130","Type":"ContainerStarted","Data":"328c68bcc4e01653178817fd15ae31ce2e495e3dde729cb29d88d166891faf44"} Oct 05 09:01:21 crc kubenswrapper[4846]: I1005 09:01:21.605946 4846 generic.go:334] "Generic (PLEG): container finished" podID="aed1876b-77bb-4605-920d-5a64dc4d0130" containerID="328c68bcc4e01653178817fd15ae31ce2e495e3dde729cb29d88d166891faf44" exitCode=0 Oct 05 09:01:21 crc kubenswrapper[4846]: I1005 09:01:21.606251 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rfkzt" event={"ID":"aed1876b-77bb-4605-920d-5a64dc4d0130","Type":"ContainerDied","Data":"328c68bcc4e01653178817fd15ae31ce2e495e3dde729cb29d88d166891faf44"} Oct 05 09:01:22 crc kubenswrapper[4846]: I1005 09:01:22.623164 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-krgq6" event={"ID":"62003eb9-c4da-4b5d-aac5-5ee422c2f46f","Type":"ContainerStarted","Data":"b0bea931aefc296d891b029339d471f46403901204a870fc0b8c73e77c5b6963"} Oct 05 09:01:23 crc kubenswrapper[4846]: I1005 09:01:23.635404 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rfkzt" event={"ID":"aed1876b-77bb-4605-920d-5a64dc4d0130","Type":"ContainerStarted","Data":"656d84447b69c27a29ffee681dabd9ef763036e8d46eea55c320fc9100bc334a"} Oct 05 09:01:23 crc kubenswrapper[4846]: I1005 09:01:23.667340 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-rfkzt" podStartSLOduration=2.660196285 podStartE2EDuration="6.667320277s" podCreationTimestamp="2025-10-05 09:01:17 +0000 UTC" firstStartedPulling="2025-10-05 09:01:18.550353365 +0000 UTC m=+8000.791206130" lastFinishedPulling="2025-10-05 09:01:22.557477307 +0000 UTC m=+8004.798330122" observedRunningTime="2025-10-05 09:01:23.661170385 +0000 UTC m=+8005.902023170" watchObservedRunningTime="2025-10-05 09:01:23.667320277 +0000 UTC m=+8005.908173062" Oct 05 09:01:27 crc kubenswrapper[4846]: I1005 09:01:27.561340 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-rfkzt" Oct 05 09:01:27 crc kubenswrapper[4846]: I1005 09:01:27.562659 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-rfkzt" Oct 05 09:01:27 crc kubenswrapper[4846]: I1005 09:01:27.682498 4846 generic.go:334] "Generic (PLEG): container finished" podID="62003eb9-c4da-4b5d-aac5-5ee422c2f46f" containerID="b0bea931aefc296d891b029339d471f46403901204a870fc0b8c73e77c5b6963" exitCode=0 Oct 05 09:01:27 crc kubenswrapper[4846]: I1005 09:01:27.682788 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-krgq6" event={"ID":"62003eb9-c4da-4b5d-aac5-5ee422c2f46f","Type":"ContainerDied","Data":"b0bea931aefc296d891b029339d471f46403901204a870fc0b8c73e77c5b6963"} Oct 05 09:01:28 crc kubenswrapper[4846]: I1005 09:01:28.512404 4846 scope.go:117] "RemoveContainer" containerID="e6df7a4d2c7027b882a7e838679edd7570422c9a8553ab74b0bddbf3e4bf4b45" Oct 05 09:01:28 crc kubenswrapper[4846]: I1005 09:01:28.642131 4846 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-rfkzt" podUID="aed1876b-77bb-4605-920d-5a64dc4d0130" containerName="registry-server" probeResult="failure" output=< Oct 05 09:01:28 crc kubenswrapper[4846]: timeout: failed to connect service ":50051" within 1s Oct 05 09:01:28 crc kubenswrapper[4846]: > Oct 05 09:01:28 crc kubenswrapper[4846]: I1005 09:01:28.695695 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-krgq6" event={"ID":"62003eb9-c4da-4b5d-aac5-5ee422c2f46f","Type":"ContainerStarted","Data":"9d722d55a59e46cb1e97594f5dccc17cca6660c72adb45bf49cf82d0da3983f9"} Oct 05 09:01:28 crc kubenswrapper[4846]: I1005 09:01:28.716917 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-krgq6" podStartSLOduration=1.8998463330000002 podStartE2EDuration="9.716898909s" podCreationTimestamp="2025-10-05 09:01:19 +0000 UTC" firstStartedPulling="2025-10-05 09:01:20.571694934 +0000 UTC m=+8002.812547729" lastFinishedPulling="2025-10-05 09:01:28.38874752 +0000 UTC m=+8010.629600305" observedRunningTime="2025-10-05 09:01:28.709394641 +0000 UTC m=+8010.950247456" watchObservedRunningTime="2025-10-05 09:01:28.716898909 +0000 UTC m=+8010.957751694" Oct 05 09:01:29 crc kubenswrapper[4846]: I1005 09:01:29.708339 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" event={"ID":"ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7","Type":"ContainerStarted","Data":"cf2bd54f0b253e45b6206216040988aa7cfaf6fc9fa46534f25c820448050b4f"} Oct 05 09:01:29 crc kubenswrapper[4846]: I1005 09:01:29.720269 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-krgq6" Oct 05 09:01:29 crc kubenswrapper[4846]: I1005 09:01:29.720325 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-krgq6" Oct 05 09:01:30 crc kubenswrapper[4846]: I1005 09:01:30.780103 4846 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-krgq6" podUID="62003eb9-c4da-4b5d-aac5-5ee422c2f46f" containerName="registry-server" probeResult="failure" output=< Oct 05 09:01:30 crc kubenswrapper[4846]: timeout: failed to connect service ":50051" within 1s Oct 05 09:01:30 crc kubenswrapper[4846]: > Oct 05 09:01:37 crc kubenswrapper[4846]: I1005 09:01:37.612859 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-rfkzt" Oct 05 09:01:37 crc kubenswrapper[4846]: I1005 09:01:37.688692 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-rfkzt" Oct 05 09:01:37 crc kubenswrapper[4846]: I1005 09:01:37.864231 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-rfkzt"] Oct 05 09:01:38 crc kubenswrapper[4846]: I1005 09:01:38.803351 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-rfkzt" podUID="aed1876b-77bb-4605-920d-5a64dc4d0130" containerName="registry-server" containerID="cri-o://656d84447b69c27a29ffee681dabd9ef763036e8d46eea55c320fc9100bc334a" gracePeriod=2 Oct 05 09:01:39 crc kubenswrapper[4846]: I1005 09:01:39.333433 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rfkzt" Oct 05 09:01:39 crc kubenswrapper[4846]: I1005 09:01:39.405283 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aed1876b-77bb-4605-920d-5a64dc4d0130-utilities\") pod \"aed1876b-77bb-4605-920d-5a64dc4d0130\" (UID: \"aed1876b-77bb-4605-920d-5a64dc4d0130\") " Oct 05 09:01:39 crc kubenswrapper[4846]: I1005 09:01:39.405429 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aed1876b-77bb-4605-920d-5a64dc4d0130-catalog-content\") pod \"aed1876b-77bb-4605-920d-5a64dc4d0130\" (UID: \"aed1876b-77bb-4605-920d-5a64dc4d0130\") " Oct 05 09:01:39 crc kubenswrapper[4846]: I1005 09:01:39.405677 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4p7s4\" (UniqueName: \"kubernetes.io/projected/aed1876b-77bb-4605-920d-5a64dc4d0130-kube-api-access-4p7s4\") pod \"aed1876b-77bb-4605-920d-5a64dc4d0130\" (UID: \"aed1876b-77bb-4605-920d-5a64dc4d0130\") " Oct 05 09:01:39 crc kubenswrapper[4846]: I1005 09:01:39.406056 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/aed1876b-77bb-4605-920d-5a64dc4d0130-utilities" (OuterVolumeSpecName: "utilities") pod "aed1876b-77bb-4605-920d-5a64dc4d0130" (UID: "aed1876b-77bb-4605-920d-5a64dc4d0130"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 09:01:39 crc kubenswrapper[4846]: I1005 09:01:39.406771 4846 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aed1876b-77bb-4605-920d-5a64dc4d0130-utilities\") on node \"crc\" DevicePath \"\"" Oct 05 09:01:39 crc kubenswrapper[4846]: I1005 09:01:39.413086 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aed1876b-77bb-4605-920d-5a64dc4d0130-kube-api-access-4p7s4" (OuterVolumeSpecName: "kube-api-access-4p7s4") pod "aed1876b-77bb-4605-920d-5a64dc4d0130" (UID: "aed1876b-77bb-4605-920d-5a64dc4d0130"). InnerVolumeSpecName "kube-api-access-4p7s4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 09:01:39 crc kubenswrapper[4846]: I1005 09:01:39.449403 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/aed1876b-77bb-4605-920d-5a64dc4d0130-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "aed1876b-77bb-4605-920d-5a64dc4d0130" (UID: "aed1876b-77bb-4605-920d-5a64dc4d0130"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 09:01:39 crc kubenswrapper[4846]: I1005 09:01:39.508617 4846 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aed1876b-77bb-4605-920d-5a64dc4d0130-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 05 09:01:39 crc kubenswrapper[4846]: I1005 09:01:39.508926 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4p7s4\" (UniqueName: \"kubernetes.io/projected/aed1876b-77bb-4605-920d-5a64dc4d0130-kube-api-access-4p7s4\") on node \"crc\" DevicePath \"\"" Oct 05 09:01:39 crc kubenswrapper[4846]: I1005 09:01:39.819504 4846 generic.go:334] "Generic (PLEG): container finished" podID="aed1876b-77bb-4605-920d-5a64dc4d0130" containerID="656d84447b69c27a29ffee681dabd9ef763036e8d46eea55c320fc9100bc334a" exitCode=0 Oct 05 09:01:39 crc kubenswrapper[4846]: I1005 09:01:39.819577 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rfkzt" event={"ID":"aed1876b-77bb-4605-920d-5a64dc4d0130","Type":"ContainerDied","Data":"656d84447b69c27a29ffee681dabd9ef763036e8d46eea55c320fc9100bc334a"} Oct 05 09:01:39 crc kubenswrapper[4846]: I1005 09:01:39.819601 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rfkzt" Oct 05 09:01:39 crc kubenswrapper[4846]: I1005 09:01:39.819626 4846 scope.go:117] "RemoveContainer" containerID="656d84447b69c27a29ffee681dabd9ef763036e8d46eea55c320fc9100bc334a" Oct 05 09:01:39 crc kubenswrapper[4846]: I1005 09:01:39.819615 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rfkzt" event={"ID":"aed1876b-77bb-4605-920d-5a64dc4d0130","Type":"ContainerDied","Data":"e579f7fe3b4d144bb6dbb76c9e422f5cdf313a3d7a653d832b85cc64395918c8"} Oct 05 09:01:39 crc kubenswrapper[4846]: I1005 09:01:39.871965 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-rfkzt"] Oct 05 09:01:39 crc kubenswrapper[4846]: I1005 09:01:39.879474 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-rfkzt"] Oct 05 09:01:39 crc kubenswrapper[4846]: I1005 09:01:39.882131 4846 scope.go:117] "RemoveContainer" containerID="328c68bcc4e01653178817fd15ae31ce2e495e3dde729cb29d88d166891faf44" Oct 05 09:01:39 crc kubenswrapper[4846]: I1005 09:01:39.915234 4846 scope.go:117] "RemoveContainer" containerID="42bb399da018b1e3c5a40e0692e38cf759d61e3c2bc8e4176f5fb12ec0c6c6e0" Oct 05 09:01:39 crc kubenswrapper[4846]: I1005 09:01:39.969282 4846 scope.go:117] "RemoveContainer" containerID="656d84447b69c27a29ffee681dabd9ef763036e8d46eea55c320fc9100bc334a" Oct 05 09:01:39 crc kubenswrapper[4846]: E1005 09:01:39.969729 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"656d84447b69c27a29ffee681dabd9ef763036e8d46eea55c320fc9100bc334a\": container with ID starting with 656d84447b69c27a29ffee681dabd9ef763036e8d46eea55c320fc9100bc334a not found: ID does not exist" containerID="656d84447b69c27a29ffee681dabd9ef763036e8d46eea55c320fc9100bc334a" Oct 05 09:01:39 crc kubenswrapper[4846]: I1005 09:01:39.969828 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"656d84447b69c27a29ffee681dabd9ef763036e8d46eea55c320fc9100bc334a"} err="failed to get container status \"656d84447b69c27a29ffee681dabd9ef763036e8d46eea55c320fc9100bc334a\": rpc error: code = NotFound desc = could not find container \"656d84447b69c27a29ffee681dabd9ef763036e8d46eea55c320fc9100bc334a\": container with ID starting with 656d84447b69c27a29ffee681dabd9ef763036e8d46eea55c320fc9100bc334a not found: ID does not exist" Oct 05 09:01:39 crc kubenswrapper[4846]: I1005 09:01:39.969862 4846 scope.go:117] "RemoveContainer" containerID="328c68bcc4e01653178817fd15ae31ce2e495e3dde729cb29d88d166891faf44" Oct 05 09:01:39 crc kubenswrapper[4846]: E1005 09:01:39.970306 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"328c68bcc4e01653178817fd15ae31ce2e495e3dde729cb29d88d166891faf44\": container with ID starting with 328c68bcc4e01653178817fd15ae31ce2e495e3dde729cb29d88d166891faf44 not found: ID does not exist" containerID="328c68bcc4e01653178817fd15ae31ce2e495e3dde729cb29d88d166891faf44" Oct 05 09:01:39 crc kubenswrapper[4846]: I1005 09:01:39.970359 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"328c68bcc4e01653178817fd15ae31ce2e495e3dde729cb29d88d166891faf44"} err="failed to get container status \"328c68bcc4e01653178817fd15ae31ce2e495e3dde729cb29d88d166891faf44\": rpc error: code = NotFound desc = could not find container \"328c68bcc4e01653178817fd15ae31ce2e495e3dde729cb29d88d166891faf44\": container with ID starting with 328c68bcc4e01653178817fd15ae31ce2e495e3dde729cb29d88d166891faf44 not found: ID does not exist" Oct 05 09:01:39 crc kubenswrapper[4846]: I1005 09:01:39.970392 4846 scope.go:117] "RemoveContainer" containerID="42bb399da018b1e3c5a40e0692e38cf759d61e3c2bc8e4176f5fb12ec0c6c6e0" Oct 05 09:01:39 crc kubenswrapper[4846]: E1005 09:01:39.970804 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"42bb399da018b1e3c5a40e0692e38cf759d61e3c2bc8e4176f5fb12ec0c6c6e0\": container with ID starting with 42bb399da018b1e3c5a40e0692e38cf759d61e3c2bc8e4176f5fb12ec0c6c6e0 not found: ID does not exist" containerID="42bb399da018b1e3c5a40e0692e38cf759d61e3c2bc8e4176f5fb12ec0c6c6e0" Oct 05 09:01:39 crc kubenswrapper[4846]: I1005 09:01:39.970836 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"42bb399da018b1e3c5a40e0692e38cf759d61e3c2bc8e4176f5fb12ec0c6c6e0"} err="failed to get container status \"42bb399da018b1e3c5a40e0692e38cf759d61e3c2bc8e4176f5fb12ec0c6c6e0\": rpc error: code = NotFound desc = could not find container \"42bb399da018b1e3c5a40e0692e38cf759d61e3c2bc8e4176f5fb12ec0c6c6e0\": container with ID starting with 42bb399da018b1e3c5a40e0692e38cf759d61e3c2bc8e4176f5fb12ec0c6c6e0 not found: ID does not exist" Oct 05 09:01:40 crc kubenswrapper[4846]: I1005 09:01:40.514674 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aed1876b-77bb-4605-920d-5a64dc4d0130" path="/var/lib/kubelet/pods/aed1876b-77bb-4605-920d-5a64dc4d0130/volumes" Oct 05 09:01:40 crc kubenswrapper[4846]: I1005 09:01:40.793222 4846 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-krgq6" podUID="62003eb9-c4da-4b5d-aac5-5ee422c2f46f" containerName="registry-server" probeResult="failure" output=< Oct 05 09:01:40 crc kubenswrapper[4846]: timeout: failed to connect service ":50051" within 1s Oct 05 09:01:40 crc kubenswrapper[4846]: > Oct 05 09:01:50 crc kubenswrapper[4846]: I1005 09:01:50.766660 4846 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-krgq6" podUID="62003eb9-c4da-4b5d-aac5-5ee422c2f46f" containerName="registry-server" probeResult="failure" output=< Oct 05 09:01:50 crc kubenswrapper[4846]: timeout: failed to connect service ":50051" within 1s Oct 05 09:01:50 crc kubenswrapper[4846]: > Oct 05 09:01:59 crc kubenswrapper[4846]: I1005 09:01:59.784113 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-krgq6" Oct 05 09:01:59 crc kubenswrapper[4846]: I1005 09:01:59.881077 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-krgq6" Oct 05 09:02:00 crc kubenswrapper[4846]: I1005 09:02:00.026290 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-krgq6"] Oct 05 09:02:01 crc kubenswrapper[4846]: I1005 09:02:01.120047 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-krgq6" podUID="62003eb9-c4da-4b5d-aac5-5ee422c2f46f" containerName="registry-server" containerID="cri-o://9d722d55a59e46cb1e97594f5dccc17cca6660c72adb45bf49cf82d0da3983f9" gracePeriod=2 Oct 05 09:02:01 crc kubenswrapper[4846]: I1005 09:02:01.650786 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-krgq6" Oct 05 09:02:01 crc kubenswrapper[4846]: I1005 09:02:01.740281 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9wjjk\" (UniqueName: \"kubernetes.io/projected/62003eb9-c4da-4b5d-aac5-5ee422c2f46f-kube-api-access-9wjjk\") pod \"62003eb9-c4da-4b5d-aac5-5ee422c2f46f\" (UID: \"62003eb9-c4da-4b5d-aac5-5ee422c2f46f\") " Oct 05 09:02:01 crc kubenswrapper[4846]: I1005 09:02:01.740519 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/62003eb9-c4da-4b5d-aac5-5ee422c2f46f-catalog-content\") pod \"62003eb9-c4da-4b5d-aac5-5ee422c2f46f\" (UID: \"62003eb9-c4da-4b5d-aac5-5ee422c2f46f\") " Oct 05 09:02:01 crc kubenswrapper[4846]: I1005 09:02:01.740584 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/62003eb9-c4da-4b5d-aac5-5ee422c2f46f-utilities\") pod \"62003eb9-c4da-4b5d-aac5-5ee422c2f46f\" (UID: \"62003eb9-c4da-4b5d-aac5-5ee422c2f46f\") " Oct 05 09:02:01 crc kubenswrapper[4846]: I1005 09:02:01.741388 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/62003eb9-c4da-4b5d-aac5-5ee422c2f46f-utilities" (OuterVolumeSpecName: "utilities") pod "62003eb9-c4da-4b5d-aac5-5ee422c2f46f" (UID: "62003eb9-c4da-4b5d-aac5-5ee422c2f46f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 09:02:01 crc kubenswrapper[4846]: I1005 09:02:01.748003 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/62003eb9-c4da-4b5d-aac5-5ee422c2f46f-kube-api-access-9wjjk" (OuterVolumeSpecName: "kube-api-access-9wjjk") pod "62003eb9-c4da-4b5d-aac5-5ee422c2f46f" (UID: "62003eb9-c4da-4b5d-aac5-5ee422c2f46f"). InnerVolumeSpecName "kube-api-access-9wjjk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 09:02:01 crc kubenswrapper[4846]: I1005 09:02:01.831639 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/62003eb9-c4da-4b5d-aac5-5ee422c2f46f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "62003eb9-c4da-4b5d-aac5-5ee422c2f46f" (UID: "62003eb9-c4da-4b5d-aac5-5ee422c2f46f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 09:02:01 crc kubenswrapper[4846]: I1005 09:02:01.843375 4846 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/62003eb9-c4da-4b5d-aac5-5ee422c2f46f-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 05 09:02:01 crc kubenswrapper[4846]: I1005 09:02:01.843404 4846 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/62003eb9-c4da-4b5d-aac5-5ee422c2f46f-utilities\") on node \"crc\" DevicePath \"\"" Oct 05 09:02:01 crc kubenswrapper[4846]: I1005 09:02:01.843414 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9wjjk\" (UniqueName: \"kubernetes.io/projected/62003eb9-c4da-4b5d-aac5-5ee422c2f46f-kube-api-access-9wjjk\") on node \"crc\" DevicePath \"\"" Oct 05 09:02:02 crc kubenswrapper[4846]: I1005 09:02:02.131098 4846 generic.go:334] "Generic (PLEG): container finished" podID="62003eb9-c4da-4b5d-aac5-5ee422c2f46f" containerID="9d722d55a59e46cb1e97594f5dccc17cca6660c72adb45bf49cf82d0da3983f9" exitCode=0 Oct 05 09:02:02 crc kubenswrapper[4846]: I1005 09:02:02.131140 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-krgq6" event={"ID":"62003eb9-c4da-4b5d-aac5-5ee422c2f46f","Type":"ContainerDied","Data":"9d722d55a59e46cb1e97594f5dccc17cca6660c72adb45bf49cf82d0da3983f9"} Oct 05 09:02:02 crc kubenswrapper[4846]: I1005 09:02:02.131164 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-krgq6" event={"ID":"62003eb9-c4da-4b5d-aac5-5ee422c2f46f","Type":"ContainerDied","Data":"00f9e9874a02a20e7141ffab1ac1b5a14a3db1fa705c05a42798547d858d0466"} Oct 05 09:02:02 crc kubenswrapper[4846]: I1005 09:02:02.131194 4846 scope.go:117] "RemoveContainer" containerID="9d722d55a59e46cb1e97594f5dccc17cca6660c72adb45bf49cf82d0da3983f9" Oct 05 09:02:02 crc kubenswrapper[4846]: I1005 09:02:02.131330 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-krgq6" Oct 05 09:02:02 crc kubenswrapper[4846]: I1005 09:02:02.162075 4846 scope.go:117] "RemoveContainer" containerID="b0bea931aefc296d891b029339d471f46403901204a870fc0b8c73e77c5b6963" Oct 05 09:02:02 crc kubenswrapper[4846]: I1005 09:02:02.167278 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-krgq6"] Oct 05 09:02:02 crc kubenswrapper[4846]: I1005 09:02:02.183439 4846 scope.go:117] "RemoveContainer" containerID="4e8e81e91c67642c54b3b57426b4268a70dd14ca2ce37213d0727c7330a019e8" Oct 05 09:02:02 crc kubenswrapper[4846]: I1005 09:02:02.194058 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-krgq6"] Oct 05 09:02:02 crc kubenswrapper[4846]: I1005 09:02:02.246215 4846 scope.go:117] "RemoveContainer" containerID="9d722d55a59e46cb1e97594f5dccc17cca6660c72adb45bf49cf82d0da3983f9" Oct 05 09:02:02 crc kubenswrapper[4846]: E1005 09:02:02.246703 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9d722d55a59e46cb1e97594f5dccc17cca6660c72adb45bf49cf82d0da3983f9\": container with ID starting with 9d722d55a59e46cb1e97594f5dccc17cca6660c72adb45bf49cf82d0da3983f9 not found: ID does not exist" containerID="9d722d55a59e46cb1e97594f5dccc17cca6660c72adb45bf49cf82d0da3983f9" Oct 05 09:02:02 crc kubenswrapper[4846]: I1005 09:02:02.246749 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9d722d55a59e46cb1e97594f5dccc17cca6660c72adb45bf49cf82d0da3983f9"} err="failed to get container status \"9d722d55a59e46cb1e97594f5dccc17cca6660c72adb45bf49cf82d0da3983f9\": rpc error: code = NotFound desc = could not find container \"9d722d55a59e46cb1e97594f5dccc17cca6660c72adb45bf49cf82d0da3983f9\": container with ID starting with 9d722d55a59e46cb1e97594f5dccc17cca6660c72adb45bf49cf82d0da3983f9 not found: ID does not exist" Oct 05 09:02:02 crc kubenswrapper[4846]: I1005 09:02:02.246770 4846 scope.go:117] "RemoveContainer" containerID="b0bea931aefc296d891b029339d471f46403901204a870fc0b8c73e77c5b6963" Oct 05 09:02:02 crc kubenswrapper[4846]: E1005 09:02:02.247310 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b0bea931aefc296d891b029339d471f46403901204a870fc0b8c73e77c5b6963\": container with ID starting with b0bea931aefc296d891b029339d471f46403901204a870fc0b8c73e77c5b6963 not found: ID does not exist" containerID="b0bea931aefc296d891b029339d471f46403901204a870fc0b8c73e77c5b6963" Oct 05 09:02:02 crc kubenswrapper[4846]: I1005 09:02:02.247331 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b0bea931aefc296d891b029339d471f46403901204a870fc0b8c73e77c5b6963"} err="failed to get container status \"b0bea931aefc296d891b029339d471f46403901204a870fc0b8c73e77c5b6963\": rpc error: code = NotFound desc = could not find container \"b0bea931aefc296d891b029339d471f46403901204a870fc0b8c73e77c5b6963\": container with ID starting with b0bea931aefc296d891b029339d471f46403901204a870fc0b8c73e77c5b6963 not found: ID does not exist" Oct 05 09:02:02 crc kubenswrapper[4846]: I1005 09:02:02.247346 4846 scope.go:117] "RemoveContainer" containerID="4e8e81e91c67642c54b3b57426b4268a70dd14ca2ce37213d0727c7330a019e8" Oct 05 09:02:02 crc kubenswrapper[4846]: E1005 09:02:02.247905 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4e8e81e91c67642c54b3b57426b4268a70dd14ca2ce37213d0727c7330a019e8\": container with ID starting with 4e8e81e91c67642c54b3b57426b4268a70dd14ca2ce37213d0727c7330a019e8 not found: ID does not exist" containerID="4e8e81e91c67642c54b3b57426b4268a70dd14ca2ce37213d0727c7330a019e8" Oct 05 09:02:02 crc kubenswrapper[4846]: I1005 09:02:02.247924 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4e8e81e91c67642c54b3b57426b4268a70dd14ca2ce37213d0727c7330a019e8"} err="failed to get container status \"4e8e81e91c67642c54b3b57426b4268a70dd14ca2ce37213d0727c7330a019e8\": rpc error: code = NotFound desc = could not find container \"4e8e81e91c67642c54b3b57426b4268a70dd14ca2ce37213d0727c7330a019e8\": container with ID starting with 4e8e81e91c67642c54b3b57426b4268a70dd14ca2ce37213d0727c7330a019e8 not found: ID does not exist" Oct 05 09:02:02 crc kubenswrapper[4846]: I1005 09:02:02.516263 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="62003eb9-c4da-4b5d-aac5-5ee422c2f46f" path="/var/lib/kubelet/pods/62003eb9-c4da-4b5d-aac5-5ee422c2f46f/volumes" Oct 05 09:03:17 crc kubenswrapper[4846]: I1005 09:03:17.018672 4846 generic.go:334] "Generic (PLEG): container finished" podID="6b81e192-94b8-4a4b-918b-feab5d5a7860" containerID="713c39f65f205843e64c0b8e05f448abc94fe0e34b8bbca868742e22d623cd67" exitCode=0 Oct 05 09:03:17 crc kubenswrapper[4846]: I1005 09:03:17.018822 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-openstack-openstack-cell1-kjczt" event={"ID":"6b81e192-94b8-4a4b-918b-feab5d5a7860","Type":"ContainerDied","Data":"713c39f65f205843e64c0b8e05f448abc94fe0e34b8bbca868742e22d623cd67"} Oct 05 09:03:18 crc kubenswrapper[4846]: I1005 09:03:18.503466 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-openstack-openstack-cell1-kjczt" Oct 05 09:03:18 crc kubenswrapper[4846]: I1005 09:03:18.666948 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/6b81e192-94b8-4a4b-918b-feab5d5a7860-nova-cells-global-config-0\") pod \"6b81e192-94b8-4a4b-918b-feab5d5a7860\" (UID: \"6b81e192-94b8-4a4b-918b-feab5d5a7860\") " Oct 05 09:03:18 crc kubenswrapper[4846]: I1005 09:03:18.667375 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4z4nq\" (UniqueName: \"kubernetes.io/projected/6b81e192-94b8-4a4b-918b-feab5d5a7860-kube-api-access-4z4nq\") pod \"6b81e192-94b8-4a4b-918b-feab5d5a7860\" (UID: \"6b81e192-94b8-4a4b-918b-feab5d5a7860\") " Oct 05 09:03:18 crc kubenswrapper[4846]: I1005 09:03:18.667659 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6b81e192-94b8-4a4b-918b-feab5d5a7860-ssh-key\") pod \"6b81e192-94b8-4a4b-918b-feab5d5a7860\" (UID: \"6b81e192-94b8-4a4b-918b-feab5d5a7860\") " Oct 05 09:03:18 crc kubenswrapper[4846]: I1005 09:03:18.667745 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6b81e192-94b8-4a4b-918b-feab5d5a7860-inventory\") pod \"6b81e192-94b8-4a4b-918b-feab5d5a7860\" (UID: \"6b81e192-94b8-4a4b-918b-feab5d5a7860\") " Oct 05 09:03:18 crc kubenswrapper[4846]: I1005 09:03:18.667880 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/6b81e192-94b8-4a4b-918b-feab5d5a7860-nova-migration-ssh-key-0\") pod \"6b81e192-94b8-4a4b-918b-feab5d5a7860\" (UID: \"6b81e192-94b8-4a4b-918b-feab5d5a7860\") " Oct 05 09:03:18 crc kubenswrapper[4846]: I1005 09:03:18.667933 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b81e192-94b8-4a4b-918b-feab5d5a7860-nova-cell1-combined-ca-bundle\") pod \"6b81e192-94b8-4a4b-918b-feab5d5a7860\" (UID: \"6b81e192-94b8-4a4b-918b-feab5d5a7860\") " Oct 05 09:03:18 crc kubenswrapper[4846]: I1005 09:03:18.667968 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/6b81e192-94b8-4a4b-918b-feab5d5a7860-nova-cell1-compute-config-1\") pod \"6b81e192-94b8-4a4b-918b-feab5d5a7860\" (UID: \"6b81e192-94b8-4a4b-918b-feab5d5a7860\") " Oct 05 09:03:18 crc kubenswrapper[4846]: I1005 09:03:18.668043 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/6b81e192-94b8-4a4b-918b-feab5d5a7860-nova-cell1-compute-config-0\") pod \"6b81e192-94b8-4a4b-918b-feab5d5a7860\" (UID: \"6b81e192-94b8-4a4b-918b-feab5d5a7860\") " Oct 05 09:03:18 crc kubenswrapper[4846]: I1005 09:03:18.668093 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/6b81e192-94b8-4a4b-918b-feab5d5a7860-nova-migration-ssh-key-1\") pod \"6b81e192-94b8-4a4b-918b-feab5d5a7860\" (UID: \"6b81e192-94b8-4a4b-918b-feab5d5a7860\") " Oct 05 09:03:18 crc kubenswrapper[4846]: I1005 09:03:18.677345 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6b81e192-94b8-4a4b-918b-feab5d5a7860-nova-cell1-combined-ca-bundle" (OuterVolumeSpecName: "nova-cell1-combined-ca-bundle") pod "6b81e192-94b8-4a4b-918b-feab5d5a7860" (UID: "6b81e192-94b8-4a4b-918b-feab5d5a7860"). InnerVolumeSpecName "nova-cell1-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 09:03:18 crc kubenswrapper[4846]: I1005 09:03:18.680108 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6b81e192-94b8-4a4b-918b-feab5d5a7860-kube-api-access-4z4nq" (OuterVolumeSpecName: "kube-api-access-4z4nq") pod "6b81e192-94b8-4a4b-918b-feab5d5a7860" (UID: "6b81e192-94b8-4a4b-918b-feab5d5a7860"). InnerVolumeSpecName "kube-api-access-4z4nq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 09:03:18 crc kubenswrapper[4846]: I1005 09:03:18.720503 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6b81e192-94b8-4a4b-918b-feab5d5a7860-nova-cells-global-config-0" (OuterVolumeSpecName: "nova-cells-global-config-0") pod "6b81e192-94b8-4a4b-918b-feab5d5a7860" (UID: "6b81e192-94b8-4a4b-918b-feab5d5a7860"). InnerVolumeSpecName "nova-cells-global-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 09:03:18 crc kubenswrapper[4846]: I1005 09:03:18.720509 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6b81e192-94b8-4a4b-918b-feab5d5a7860-nova-migration-ssh-key-1" (OuterVolumeSpecName: "nova-migration-ssh-key-1") pod "6b81e192-94b8-4a4b-918b-feab5d5a7860" (UID: "6b81e192-94b8-4a4b-918b-feab5d5a7860"). InnerVolumeSpecName "nova-migration-ssh-key-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 09:03:18 crc kubenswrapper[4846]: I1005 09:03:18.720998 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6b81e192-94b8-4a4b-918b-feab5d5a7860-nova-migration-ssh-key-0" (OuterVolumeSpecName: "nova-migration-ssh-key-0") pod "6b81e192-94b8-4a4b-918b-feab5d5a7860" (UID: "6b81e192-94b8-4a4b-918b-feab5d5a7860"). InnerVolumeSpecName "nova-migration-ssh-key-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 09:03:18 crc kubenswrapper[4846]: I1005 09:03:18.738666 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6b81e192-94b8-4a4b-918b-feab5d5a7860-nova-cell1-compute-config-0" (OuterVolumeSpecName: "nova-cell1-compute-config-0") pod "6b81e192-94b8-4a4b-918b-feab5d5a7860" (UID: "6b81e192-94b8-4a4b-918b-feab5d5a7860"). InnerVolumeSpecName "nova-cell1-compute-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 09:03:18 crc kubenswrapper[4846]: I1005 09:03:18.748243 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6b81e192-94b8-4a4b-918b-feab5d5a7860-inventory" (OuterVolumeSpecName: "inventory") pod "6b81e192-94b8-4a4b-918b-feab5d5a7860" (UID: "6b81e192-94b8-4a4b-918b-feab5d5a7860"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 09:03:18 crc kubenswrapper[4846]: I1005 09:03:18.751215 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6b81e192-94b8-4a4b-918b-feab5d5a7860-nova-cell1-compute-config-1" (OuterVolumeSpecName: "nova-cell1-compute-config-1") pod "6b81e192-94b8-4a4b-918b-feab5d5a7860" (UID: "6b81e192-94b8-4a4b-918b-feab5d5a7860"). InnerVolumeSpecName "nova-cell1-compute-config-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 09:03:18 crc kubenswrapper[4846]: I1005 09:03:18.754629 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6b81e192-94b8-4a4b-918b-feab5d5a7860-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "6b81e192-94b8-4a4b-918b-feab5d5a7860" (UID: "6b81e192-94b8-4a4b-918b-feab5d5a7860"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 09:03:18 crc kubenswrapper[4846]: I1005 09:03:18.770855 4846 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/6b81e192-94b8-4a4b-918b-feab5d5a7860-nova-migration-ssh-key-0\") on node \"crc\" DevicePath \"\"" Oct 05 09:03:18 crc kubenswrapper[4846]: I1005 09:03:18.770887 4846 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b81e192-94b8-4a4b-918b-feab5d5a7860-nova-cell1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 09:03:18 crc kubenswrapper[4846]: I1005 09:03:18.770900 4846 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/6b81e192-94b8-4a4b-918b-feab5d5a7860-nova-cell1-compute-config-1\") on node \"crc\" DevicePath \"\"" Oct 05 09:03:18 crc kubenswrapper[4846]: I1005 09:03:18.770912 4846 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/6b81e192-94b8-4a4b-918b-feab5d5a7860-nova-cell1-compute-config-0\") on node \"crc\" DevicePath \"\"" Oct 05 09:03:18 crc kubenswrapper[4846]: I1005 09:03:18.770925 4846 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/6b81e192-94b8-4a4b-918b-feab5d5a7860-nova-migration-ssh-key-1\") on node \"crc\" DevicePath \"\"" Oct 05 09:03:18 crc kubenswrapper[4846]: I1005 09:03:18.770938 4846 reconciler_common.go:293] "Volume detached for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/6b81e192-94b8-4a4b-918b-feab5d5a7860-nova-cells-global-config-0\") on node \"crc\" DevicePath \"\"" Oct 05 09:03:18 crc kubenswrapper[4846]: I1005 09:03:18.770949 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4z4nq\" (UniqueName: \"kubernetes.io/projected/6b81e192-94b8-4a4b-918b-feab5d5a7860-kube-api-access-4z4nq\") on node \"crc\" DevicePath \"\"" Oct 05 09:03:18 crc kubenswrapper[4846]: I1005 09:03:18.770961 4846 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6b81e192-94b8-4a4b-918b-feab5d5a7860-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 05 09:03:18 crc kubenswrapper[4846]: I1005 09:03:18.770973 4846 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6b81e192-94b8-4a4b-918b-feab5d5a7860-inventory\") on node \"crc\" DevicePath \"\"" Oct 05 09:03:19 crc kubenswrapper[4846]: I1005 09:03:19.048430 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-openstack-openstack-cell1-kjczt" event={"ID":"6b81e192-94b8-4a4b-918b-feab5d5a7860","Type":"ContainerDied","Data":"c378c0a0013f39a4dfe65c7fc3e538963b2c1a5d22b38e73b759914d9e516fe0"} Oct 05 09:03:19 crc kubenswrapper[4846]: I1005 09:03:19.048503 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c378c0a0013f39a4dfe65c7fc3e538963b2c1a5d22b38e73b759914d9e516fe0" Oct 05 09:03:19 crc kubenswrapper[4846]: I1005 09:03:19.048580 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-openstack-openstack-cell1-kjczt" Oct 05 09:03:19 crc kubenswrapper[4846]: I1005 09:03:19.224841 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/telemetry-openstack-openstack-cell1-fl275"] Oct 05 09:03:19 crc kubenswrapper[4846]: E1005 09:03:19.225563 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="62003eb9-c4da-4b5d-aac5-5ee422c2f46f" containerName="extract-utilities" Oct 05 09:03:19 crc kubenswrapper[4846]: I1005 09:03:19.225593 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="62003eb9-c4da-4b5d-aac5-5ee422c2f46f" containerName="extract-utilities" Oct 05 09:03:19 crc kubenswrapper[4846]: E1005 09:03:19.225627 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aed1876b-77bb-4605-920d-5a64dc4d0130" containerName="extract-content" Oct 05 09:03:19 crc kubenswrapper[4846]: I1005 09:03:19.225642 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="aed1876b-77bb-4605-920d-5a64dc4d0130" containerName="extract-content" Oct 05 09:03:19 crc kubenswrapper[4846]: E1005 09:03:19.225672 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="62003eb9-c4da-4b5d-aac5-5ee422c2f46f" containerName="extract-content" Oct 05 09:03:19 crc kubenswrapper[4846]: I1005 09:03:19.225686 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="62003eb9-c4da-4b5d-aac5-5ee422c2f46f" containerName="extract-content" Oct 05 09:03:19 crc kubenswrapper[4846]: E1005 09:03:19.225729 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6b81e192-94b8-4a4b-918b-feab5d5a7860" containerName="nova-cell1-openstack-openstack-cell1" Oct 05 09:03:19 crc kubenswrapper[4846]: I1005 09:03:19.225743 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="6b81e192-94b8-4a4b-918b-feab5d5a7860" containerName="nova-cell1-openstack-openstack-cell1" Oct 05 09:03:19 crc kubenswrapper[4846]: E1005 09:03:19.225779 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="62003eb9-c4da-4b5d-aac5-5ee422c2f46f" containerName="registry-server" Oct 05 09:03:19 crc kubenswrapper[4846]: I1005 09:03:19.225794 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="62003eb9-c4da-4b5d-aac5-5ee422c2f46f" containerName="registry-server" Oct 05 09:03:19 crc kubenswrapper[4846]: E1005 09:03:19.225812 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aed1876b-77bb-4605-920d-5a64dc4d0130" containerName="registry-server" Oct 05 09:03:19 crc kubenswrapper[4846]: I1005 09:03:19.225826 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="aed1876b-77bb-4605-920d-5a64dc4d0130" containerName="registry-server" Oct 05 09:03:19 crc kubenswrapper[4846]: E1005 09:03:19.225848 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aed1876b-77bb-4605-920d-5a64dc4d0130" containerName="extract-utilities" Oct 05 09:03:19 crc kubenswrapper[4846]: I1005 09:03:19.225862 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="aed1876b-77bb-4605-920d-5a64dc4d0130" containerName="extract-utilities" Oct 05 09:03:19 crc kubenswrapper[4846]: I1005 09:03:19.226278 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="aed1876b-77bb-4605-920d-5a64dc4d0130" containerName="registry-server" Oct 05 09:03:19 crc kubenswrapper[4846]: I1005 09:03:19.226319 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="62003eb9-c4da-4b5d-aac5-5ee422c2f46f" containerName="registry-server" Oct 05 09:03:19 crc kubenswrapper[4846]: I1005 09:03:19.226344 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="6b81e192-94b8-4a4b-918b-feab5d5a7860" containerName="nova-cell1-openstack-openstack-cell1" Oct 05 09:03:19 crc kubenswrapper[4846]: I1005 09:03:19.227652 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-openstack-openstack-cell1-fl275" Oct 05 09:03:19 crc kubenswrapper[4846]: I1005 09:03:19.231091 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-compute-config-data" Oct 05 09:03:19 crc kubenswrapper[4846]: I1005 09:03:19.231320 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-7687s" Oct 05 09:03:19 crc kubenswrapper[4846]: I1005 09:03:19.231481 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Oct 05 09:03:19 crc kubenswrapper[4846]: I1005 09:03:19.231765 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Oct 05 09:03:19 crc kubenswrapper[4846]: I1005 09:03:19.233600 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 05 09:03:19 crc kubenswrapper[4846]: I1005 09:03:19.234335 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-openstack-openstack-cell1-fl275"] Oct 05 09:03:19 crc kubenswrapper[4846]: I1005 09:03:19.383160 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/8b360825-bbe2-4522-b0b0-1f0bf6c4bc9f-ceilometer-compute-config-data-2\") pod \"telemetry-openstack-openstack-cell1-fl275\" (UID: \"8b360825-bbe2-4522-b0b0-1f0bf6c4bc9f\") " pod="openstack/telemetry-openstack-openstack-cell1-fl275" Oct 05 09:03:19 crc kubenswrapper[4846]: I1005 09:03:19.383231 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/8b360825-bbe2-4522-b0b0-1f0bf6c4bc9f-ceilometer-compute-config-data-1\") pod \"telemetry-openstack-openstack-cell1-fl275\" (UID: \"8b360825-bbe2-4522-b0b0-1f0bf6c4bc9f\") " pod="openstack/telemetry-openstack-openstack-cell1-fl275" Oct 05 09:03:19 crc kubenswrapper[4846]: I1005 09:03:19.383571 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8b360825-bbe2-4522-b0b0-1f0bf6c4bc9f-inventory\") pod \"telemetry-openstack-openstack-cell1-fl275\" (UID: \"8b360825-bbe2-4522-b0b0-1f0bf6c4bc9f\") " pod="openstack/telemetry-openstack-openstack-cell1-fl275" Oct 05 09:03:19 crc kubenswrapper[4846]: I1005 09:03:19.383731 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8b360825-bbe2-4522-b0b0-1f0bf6c4bc9f-ssh-key\") pod \"telemetry-openstack-openstack-cell1-fl275\" (UID: \"8b360825-bbe2-4522-b0b0-1f0bf6c4bc9f\") " pod="openstack/telemetry-openstack-openstack-cell1-fl275" Oct 05 09:03:19 crc kubenswrapper[4846]: I1005 09:03:19.383808 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8b360825-bbe2-4522-b0b0-1f0bf6c4bc9f-telemetry-combined-ca-bundle\") pod \"telemetry-openstack-openstack-cell1-fl275\" (UID: \"8b360825-bbe2-4522-b0b0-1f0bf6c4bc9f\") " pod="openstack/telemetry-openstack-openstack-cell1-fl275" Oct 05 09:03:19 crc kubenswrapper[4846]: I1005 09:03:19.383996 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p8b6v\" (UniqueName: \"kubernetes.io/projected/8b360825-bbe2-4522-b0b0-1f0bf6c4bc9f-kube-api-access-p8b6v\") pod \"telemetry-openstack-openstack-cell1-fl275\" (UID: \"8b360825-bbe2-4522-b0b0-1f0bf6c4bc9f\") " pod="openstack/telemetry-openstack-openstack-cell1-fl275" Oct 05 09:03:19 crc kubenswrapper[4846]: I1005 09:03:19.384331 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/8b360825-bbe2-4522-b0b0-1f0bf6c4bc9f-ceilometer-compute-config-data-0\") pod \"telemetry-openstack-openstack-cell1-fl275\" (UID: \"8b360825-bbe2-4522-b0b0-1f0bf6c4bc9f\") " pod="openstack/telemetry-openstack-openstack-cell1-fl275" Oct 05 09:03:19 crc kubenswrapper[4846]: I1005 09:03:19.486570 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p8b6v\" (UniqueName: \"kubernetes.io/projected/8b360825-bbe2-4522-b0b0-1f0bf6c4bc9f-kube-api-access-p8b6v\") pod \"telemetry-openstack-openstack-cell1-fl275\" (UID: \"8b360825-bbe2-4522-b0b0-1f0bf6c4bc9f\") " pod="openstack/telemetry-openstack-openstack-cell1-fl275" Oct 05 09:03:19 crc kubenswrapper[4846]: I1005 09:03:19.486794 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/8b360825-bbe2-4522-b0b0-1f0bf6c4bc9f-ceilometer-compute-config-data-0\") pod \"telemetry-openstack-openstack-cell1-fl275\" (UID: \"8b360825-bbe2-4522-b0b0-1f0bf6c4bc9f\") " pod="openstack/telemetry-openstack-openstack-cell1-fl275" Oct 05 09:03:19 crc kubenswrapper[4846]: I1005 09:03:19.486851 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/8b360825-bbe2-4522-b0b0-1f0bf6c4bc9f-ceilometer-compute-config-data-2\") pod \"telemetry-openstack-openstack-cell1-fl275\" (UID: \"8b360825-bbe2-4522-b0b0-1f0bf6c4bc9f\") " pod="openstack/telemetry-openstack-openstack-cell1-fl275" Oct 05 09:03:19 crc kubenswrapper[4846]: I1005 09:03:19.486893 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/8b360825-bbe2-4522-b0b0-1f0bf6c4bc9f-ceilometer-compute-config-data-1\") pod \"telemetry-openstack-openstack-cell1-fl275\" (UID: \"8b360825-bbe2-4522-b0b0-1f0bf6c4bc9f\") " pod="openstack/telemetry-openstack-openstack-cell1-fl275" Oct 05 09:03:19 crc kubenswrapper[4846]: I1005 09:03:19.487024 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8b360825-bbe2-4522-b0b0-1f0bf6c4bc9f-inventory\") pod \"telemetry-openstack-openstack-cell1-fl275\" (UID: \"8b360825-bbe2-4522-b0b0-1f0bf6c4bc9f\") " pod="openstack/telemetry-openstack-openstack-cell1-fl275" Oct 05 09:03:19 crc kubenswrapper[4846]: I1005 09:03:19.487089 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8b360825-bbe2-4522-b0b0-1f0bf6c4bc9f-ssh-key\") pod \"telemetry-openstack-openstack-cell1-fl275\" (UID: \"8b360825-bbe2-4522-b0b0-1f0bf6c4bc9f\") " pod="openstack/telemetry-openstack-openstack-cell1-fl275" Oct 05 09:03:19 crc kubenswrapper[4846]: I1005 09:03:19.487138 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8b360825-bbe2-4522-b0b0-1f0bf6c4bc9f-telemetry-combined-ca-bundle\") pod \"telemetry-openstack-openstack-cell1-fl275\" (UID: \"8b360825-bbe2-4522-b0b0-1f0bf6c4bc9f\") " pod="openstack/telemetry-openstack-openstack-cell1-fl275" Oct 05 09:03:19 crc kubenswrapper[4846]: I1005 09:03:19.493494 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/8b360825-bbe2-4522-b0b0-1f0bf6c4bc9f-ceilometer-compute-config-data-0\") pod \"telemetry-openstack-openstack-cell1-fl275\" (UID: \"8b360825-bbe2-4522-b0b0-1f0bf6c4bc9f\") " pod="openstack/telemetry-openstack-openstack-cell1-fl275" Oct 05 09:03:19 crc kubenswrapper[4846]: I1005 09:03:19.493551 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8b360825-bbe2-4522-b0b0-1f0bf6c4bc9f-inventory\") pod \"telemetry-openstack-openstack-cell1-fl275\" (UID: \"8b360825-bbe2-4522-b0b0-1f0bf6c4bc9f\") " pod="openstack/telemetry-openstack-openstack-cell1-fl275" Oct 05 09:03:19 crc kubenswrapper[4846]: I1005 09:03:19.493494 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/8b360825-bbe2-4522-b0b0-1f0bf6c4bc9f-ceilometer-compute-config-data-1\") pod \"telemetry-openstack-openstack-cell1-fl275\" (UID: \"8b360825-bbe2-4522-b0b0-1f0bf6c4bc9f\") " pod="openstack/telemetry-openstack-openstack-cell1-fl275" Oct 05 09:03:19 crc kubenswrapper[4846]: I1005 09:03:19.494955 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8b360825-bbe2-4522-b0b0-1f0bf6c4bc9f-ssh-key\") pod \"telemetry-openstack-openstack-cell1-fl275\" (UID: \"8b360825-bbe2-4522-b0b0-1f0bf6c4bc9f\") " pod="openstack/telemetry-openstack-openstack-cell1-fl275" Oct 05 09:03:19 crc kubenswrapper[4846]: I1005 09:03:19.496793 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/8b360825-bbe2-4522-b0b0-1f0bf6c4bc9f-ceilometer-compute-config-data-2\") pod \"telemetry-openstack-openstack-cell1-fl275\" (UID: \"8b360825-bbe2-4522-b0b0-1f0bf6c4bc9f\") " pod="openstack/telemetry-openstack-openstack-cell1-fl275" Oct 05 09:03:19 crc kubenswrapper[4846]: I1005 09:03:19.502518 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8b360825-bbe2-4522-b0b0-1f0bf6c4bc9f-telemetry-combined-ca-bundle\") pod \"telemetry-openstack-openstack-cell1-fl275\" (UID: \"8b360825-bbe2-4522-b0b0-1f0bf6c4bc9f\") " pod="openstack/telemetry-openstack-openstack-cell1-fl275" Oct 05 09:03:19 crc kubenswrapper[4846]: I1005 09:03:19.517252 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p8b6v\" (UniqueName: \"kubernetes.io/projected/8b360825-bbe2-4522-b0b0-1f0bf6c4bc9f-kube-api-access-p8b6v\") pod \"telemetry-openstack-openstack-cell1-fl275\" (UID: \"8b360825-bbe2-4522-b0b0-1f0bf6c4bc9f\") " pod="openstack/telemetry-openstack-openstack-cell1-fl275" Oct 05 09:03:19 crc kubenswrapper[4846]: I1005 09:03:19.550162 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-openstack-openstack-cell1-fl275" Oct 05 09:03:19 crc kubenswrapper[4846]: I1005 09:03:19.985642 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-openstack-openstack-cell1-fl275"] Oct 05 09:03:20 crc kubenswrapper[4846]: I1005 09:03:20.059633 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-openstack-openstack-cell1-fl275" event={"ID":"8b360825-bbe2-4522-b0b0-1f0bf6c4bc9f","Type":"ContainerStarted","Data":"88f78d7e318bf085552a2a3590d51e9ce467abd32e041168b3f67c27a97c7101"} Oct 05 09:03:21 crc kubenswrapper[4846]: I1005 09:03:21.074593 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-openstack-openstack-cell1-fl275" event={"ID":"8b360825-bbe2-4522-b0b0-1f0bf6c4bc9f","Type":"ContainerStarted","Data":"6a06786ddcedfc15ef38d9b2e290e9c711dafa86da3ab89eef356ef1ef248842"} Oct 05 09:03:21 crc kubenswrapper[4846]: I1005 09:03:21.100915 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/telemetry-openstack-openstack-cell1-fl275" podStartSLOduration=1.555395624 podStartE2EDuration="2.10089825s" podCreationTimestamp="2025-10-05 09:03:19 +0000 UTC" firstStartedPulling="2025-10-05 09:03:19.991302589 +0000 UTC m=+8122.232155384" lastFinishedPulling="2025-10-05 09:03:20.536805215 +0000 UTC m=+8122.777658010" observedRunningTime="2025-10-05 09:03:21.09175705 +0000 UTC m=+8123.332609835" watchObservedRunningTime="2025-10-05 09:03:21.10089825 +0000 UTC m=+8123.341751025" Oct 05 09:03:53 crc kubenswrapper[4846]: I1005 09:03:53.325042 4846 patch_prober.go:28] interesting pod/machine-config-daemon-fscvf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 09:03:53 crc kubenswrapper[4846]: I1005 09:03:53.326200 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 09:04:23 crc kubenswrapper[4846]: I1005 09:04:23.325744 4846 patch_prober.go:28] interesting pod/machine-config-daemon-fscvf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 09:04:23 crc kubenswrapper[4846]: I1005 09:04:23.326406 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 09:04:53 crc kubenswrapper[4846]: I1005 09:04:53.324630 4846 patch_prober.go:28] interesting pod/machine-config-daemon-fscvf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 09:04:53 crc kubenswrapper[4846]: I1005 09:04:53.325182 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 09:04:53 crc kubenswrapper[4846]: I1005 09:04:53.325264 4846 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" Oct 05 09:04:53 crc kubenswrapper[4846]: I1005 09:04:53.326058 4846 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"cf2bd54f0b253e45b6206216040988aa7cfaf6fc9fa46534f25c820448050b4f"} pod="openshift-machine-config-operator/machine-config-daemon-fscvf" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 05 09:04:53 crc kubenswrapper[4846]: I1005 09:04:53.326113 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" containerName="machine-config-daemon" containerID="cri-o://cf2bd54f0b253e45b6206216040988aa7cfaf6fc9fa46534f25c820448050b4f" gracePeriod=600 Oct 05 09:04:54 crc kubenswrapper[4846]: I1005 09:04:54.207146 4846 generic.go:334] "Generic (PLEG): container finished" podID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" containerID="cf2bd54f0b253e45b6206216040988aa7cfaf6fc9fa46534f25c820448050b4f" exitCode=0 Oct 05 09:04:54 crc kubenswrapper[4846]: I1005 09:04:54.207185 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" event={"ID":"ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7","Type":"ContainerDied","Data":"cf2bd54f0b253e45b6206216040988aa7cfaf6fc9fa46534f25c820448050b4f"} Oct 05 09:04:54 crc kubenswrapper[4846]: I1005 09:04:54.207731 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" event={"ID":"ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7","Type":"ContainerStarted","Data":"e39593bcdffa488d1cfc3d5c184d6e46655bb5c018d40df0f20d590de5498b3d"} Oct 05 09:04:54 crc kubenswrapper[4846]: I1005 09:04:54.207773 4846 scope.go:117] "RemoveContainer" containerID="e6df7a4d2c7027b882a7e838679edd7570422c9a8553ab74b0bddbf3e4bf4b45" Oct 05 09:06:53 crc kubenswrapper[4846]: I1005 09:06:53.325428 4846 patch_prober.go:28] interesting pod/machine-config-daemon-fscvf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 09:06:53 crc kubenswrapper[4846]: I1005 09:06:53.325912 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 09:07:23 crc kubenswrapper[4846]: I1005 09:07:23.325438 4846 patch_prober.go:28] interesting pod/machine-config-daemon-fscvf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 09:07:23 crc kubenswrapper[4846]: I1005 09:07:23.326035 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 09:07:53 crc kubenswrapper[4846]: I1005 09:07:53.325070 4846 patch_prober.go:28] interesting pod/machine-config-daemon-fscvf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 09:07:53 crc kubenswrapper[4846]: I1005 09:07:53.325596 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 09:07:53 crc kubenswrapper[4846]: I1005 09:07:53.325649 4846 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" Oct 05 09:07:53 crc kubenswrapper[4846]: I1005 09:07:53.326174 4846 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"e39593bcdffa488d1cfc3d5c184d6e46655bb5c018d40df0f20d590de5498b3d"} pod="openshift-machine-config-operator/machine-config-daemon-fscvf" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 05 09:07:53 crc kubenswrapper[4846]: I1005 09:07:53.326295 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" containerName="machine-config-daemon" containerID="cri-o://e39593bcdffa488d1cfc3d5c184d6e46655bb5c018d40df0f20d590de5498b3d" gracePeriod=600 Oct 05 09:07:53 crc kubenswrapper[4846]: E1005 09:07:53.457859 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 09:07:53 crc kubenswrapper[4846]: I1005 09:07:53.579115 4846 generic.go:334] "Generic (PLEG): container finished" podID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" containerID="e39593bcdffa488d1cfc3d5c184d6e46655bb5c018d40df0f20d590de5498b3d" exitCode=0 Oct 05 09:07:53 crc kubenswrapper[4846]: I1005 09:07:53.579345 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" event={"ID":"ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7","Type":"ContainerDied","Data":"e39593bcdffa488d1cfc3d5c184d6e46655bb5c018d40df0f20d590de5498b3d"} Oct 05 09:07:53 crc kubenswrapper[4846]: I1005 09:07:53.579781 4846 scope.go:117] "RemoveContainer" containerID="cf2bd54f0b253e45b6206216040988aa7cfaf6fc9fa46534f25c820448050b4f" Oct 05 09:07:53 crc kubenswrapper[4846]: I1005 09:07:53.580638 4846 scope.go:117] "RemoveContainer" containerID="e39593bcdffa488d1cfc3d5c184d6e46655bb5c018d40df0f20d590de5498b3d" Oct 05 09:07:53 crc kubenswrapper[4846]: E1005 09:07:53.581040 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 09:08:08 crc kubenswrapper[4846]: I1005 09:08:08.510959 4846 scope.go:117] "RemoveContainer" containerID="e39593bcdffa488d1cfc3d5c184d6e46655bb5c018d40df0f20d590de5498b3d" Oct 05 09:08:08 crc kubenswrapper[4846]: E1005 09:08:08.511826 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 09:08:22 crc kubenswrapper[4846]: I1005 09:08:22.498272 4846 scope.go:117] "RemoveContainer" containerID="e39593bcdffa488d1cfc3d5c184d6e46655bb5c018d40df0f20d590de5498b3d" Oct 05 09:08:22 crc kubenswrapper[4846]: E1005 09:08:22.499464 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 09:08:24 crc kubenswrapper[4846]: I1005 09:08:24.949753 4846 generic.go:334] "Generic (PLEG): container finished" podID="8b360825-bbe2-4522-b0b0-1f0bf6c4bc9f" containerID="6a06786ddcedfc15ef38d9b2e290e9c711dafa86da3ab89eef356ef1ef248842" exitCode=0 Oct 05 09:08:24 crc kubenswrapper[4846]: I1005 09:08:24.949855 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-openstack-openstack-cell1-fl275" event={"ID":"8b360825-bbe2-4522-b0b0-1f0bf6c4bc9f","Type":"ContainerDied","Data":"6a06786ddcedfc15ef38d9b2e290e9c711dafa86da3ab89eef356ef1ef248842"} Oct 05 09:08:26 crc kubenswrapper[4846]: I1005 09:08:26.445387 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-openstack-openstack-cell1-fl275" Oct 05 09:08:26 crc kubenswrapper[4846]: I1005 09:08:26.571818 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/8b360825-bbe2-4522-b0b0-1f0bf6c4bc9f-ceilometer-compute-config-data-1\") pod \"8b360825-bbe2-4522-b0b0-1f0bf6c4bc9f\" (UID: \"8b360825-bbe2-4522-b0b0-1f0bf6c4bc9f\") " Oct 05 09:08:26 crc kubenswrapper[4846]: I1005 09:08:26.571892 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8b360825-bbe2-4522-b0b0-1f0bf6c4bc9f-inventory\") pod \"8b360825-bbe2-4522-b0b0-1f0bf6c4bc9f\" (UID: \"8b360825-bbe2-4522-b0b0-1f0bf6c4bc9f\") " Oct 05 09:08:26 crc kubenswrapper[4846]: I1005 09:08:26.571943 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8b360825-bbe2-4522-b0b0-1f0bf6c4bc9f-ssh-key\") pod \"8b360825-bbe2-4522-b0b0-1f0bf6c4bc9f\" (UID: \"8b360825-bbe2-4522-b0b0-1f0bf6c4bc9f\") " Oct 05 09:08:26 crc kubenswrapper[4846]: I1005 09:08:26.572123 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/8b360825-bbe2-4522-b0b0-1f0bf6c4bc9f-ceilometer-compute-config-data-2\") pod \"8b360825-bbe2-4522-b0b0-1f0bf6c4bc9f\" (UID: \"8b360825-bbe2-4522-b0b0-1f0bf6c4bc9f\") " Oct 05 09:08:26 crc kubenswrapper[4846]: I1005 09:08:26.572149 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/8b360825-bbe2-4522-b0b0-1f0bf6c4bc9f-ceilometer-compute-config-data-0\") pod \"8b360825-bbe2-4522-b0b0-1f0bf6c4bc9f\" (UID: \"8b360825-bbe2-4522-b0b0-1f0bf6c4bc9f\") " Oct 05 09:08:26 crc kubenswrapper[4846]: I1005 09:08:26.572269 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p8b6v\" (UniqueName: \"kubernetes.io/projected/8b360825-bbe2-4522-b0b0-1f0bf6c4bc9f-kube-api-access-p8b6v\") pod \"8b360825-bbe2-4522-b0b0-1f0bf6c4bc9f\" (UID: \"8b360825-bbe2-4522-b0b0-1f0bf6c4bc9f\") " Oct 05 09:08:26 crc kubenswrapper[4846]: I1005 09:08:26.572302 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8b360825-bbe2-4522-b0b0-1f0bf6c4bc9f-telemetry-combined-ca-bundle\") pod \"8b360825-bbe2-4522-b0b0-1f0bf6c4bc9f\" (UID: \"8b360825-bbe2-4522-b0b0-1f0bf6c4bc9f\") " Oct 05 09:08:26 crc kubenswrapper[4846]: I1005 09:08:26.588135 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8b360825-bbe2-4522-b0b0-1f0bf6c4bc9f-kube-api-access-p8b6v" (OuterVolumeSpecName: "kube-api-access-p8b6v") pod "8b360825-bbe2-4522-b0b0-1f0bf6c4bc9f" (UID: "8b360825-bbe2-4522-b0b0-1f0bf6c4bc9f"). InnerVolumeSpecName "kube-api-access-p8b6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 09:08:26 crc kubenswrapper[4846]: I1005 09:08:26.588471 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8b360825-bbe2-4522-b0b0-1f0bf6c4bc9f-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "8b360825-bbe2-4522-b0b0-1f0bf6c4bc9f" (UID: "8b360825-bbe2-4522-b0b0-1f0bf6c4bc9f"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 09:08:26 crc kubenswrapper[4846]: I1005 09:08:26.601816 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8b360825-bbe2-4522-b0b0-1f0bf6c4bc9f-ceilometer-compute-config-data-2" (OuterVolumeSpecName: "ceilometer-compute-config-data-2") pod "8b360825-bbe2-4522-b0b0-1f0bf6c4bc9f" (UID: "8b360825-bbe2-4522-b0b0-1f0bf6c4bc9f"). InnerVolumeSpecName "ceilometer-compute-config-data-2". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 09:08:26 crc kubenswrapper[4846]: I1005 09:08:26.610139 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8b360825-bbe2-4522-b0b0-1f0bf6c4bc9f-ceilometer-compute-config-data-0" (OuterVolumeSpecName: "ceilometer-compute-config-data-0") pod "8b360825-bbe2-4522-b0b0-1f0bf6c4bc9f" (UID: "8b360825-bbe2-4522-b0b0-1f0bf6c4bc9f"). InnerVolumeSpecName "ceilometer-compute-config-data-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 09:08:26 crc kubenswrapper[4846]: I1005 09:08:26.614238 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8b360825-bbe2-4522-b0b0-1f0bf6c4bc9f-inventory" (OuterVolumeSpecName: "inventory") pod "8b360825-bbe2-4522-b0b0-1f0bf6c4bc9f" (UID: "8b360825-bbe2-4522-b0b0-1f0bf6c4bc9f"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 09:08:26 crc kubenswrapper[4846]: I1005 09:08:26.622717 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8b360825-bbe2-4522-b0b0-1f0bf6c4bc9f-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "8b360825-bbe2-4522-b0b0-1f0bf6c4bc9f" (UID: "8b360825-bbe2-4522-b0b0-1f0bf6c4bc9f"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 09:08:26 crc kubenswrapper[4846]: I1005 09:08:26.627459 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8b360825-bbe2-4522-b0b0-1f0bf6c4bc9f-ceilometer-compute-config-data-1" (OuterVolumeSpecName: "ceilometer-compute-config-data-1") pod "8b360825-bbe2-4522-b0b0-1f0bf6c4bc9f" (UID: "8b360825-bbe2-4522-b0b0-1f0bf6c4bc9f"). InnerVolumeSpecName "ceilometer-compute-config-data-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 09:08:26 crc kubenswrapper[4846]: I1005 09:08:26.675869 4846 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8b360825-bbe2-4522-b0b0-1f0bf6c4bc9f-inventory\") on node \"crc\" DevicePath \"\"" Oct 05 09:08:26 crc kubenswrapper[4846]: I1005 09:08:26.676359 4846 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8b360825-bbe2-4522-b0b0-1f0bf6c4bc9f-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 05 09:08:26 crc kubenswrapper[4846]: I1005 09:08:26.676377 4846 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/8b360825-bbe2-4522-b0b0-1f0bf6c4bc9f-ceilometer-compute-config-data-2\") on node \"crc\" DevicePath \"\"" Oct 05 09:08:26 crc kubenswrapper[4846]: I1005 09:08:26.676390 4846 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/8b360825-bbe2-4522-b0b0-1f0bf6c4bc9f-ceilometer-compute-config-data-0\") on node \"crc\" DevicePath \"\"" Oct 05 09:08:26 crc kubenswrapper[4846]: I1005 09:08:26.676406 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p8b6v\" (UniqueName: \"kubernetes.io/projected/8b360825-bbe2-4522-b0b0-1f0bf6c4bc9f-kube-api-access-p8b6v\") on node \"crc\" DevicePath \"\"" Oct 05 09:08:26 crc kubenswrapper[4846]: I1005 09:08:26.676420 4846 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8b360825-bbe2-4522-b0b0-1f0bf6c4bc9f-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 09:08:26 crc kubenswrapper[4846]: I1005 09:08:26.676431 4846 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/8b360825-bbe2-4522-b0b0-1f0bf6c4bc9f-ceilometer-compute-config-data-1\") on node \"crc\" DevicePath \"\"" Oct 05 09:08:26 crc kubenswrapper[4846]: I1005 09:08:26.975282 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-openstack-openstack-cell1-fl275" event={"ID":"8b360825-bbe2-4522-b0b0-1f0bf6c4bc9f","Type":"ContainerDied","Data":"88f78d7e318bf085552a2a3590d51e9ce467abd32e041168b3f67c27a97c7101"} Oct 05 09:08:26 crc kubenswrapper[4846]: I1005 09:08:26.975327 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="88f78d7e318bf085552a2a3590d51e9ce467abd32e041168b3f67c27a97c7101" Oct 05 09:08:26 crc kubenswrapper[4846]: I1005 09:08:26.975892 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-openstack-openstack-cell1-fl275" Oct 05 09:08:27 crc kubenswrapper[4846]: I1005 09:08:27.083469 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-sriov-openstack-openstack-cell1-ds9h4"] Oct 05 09:08:27 crc kubenswrapper[4846]: E1005 09:08:27.084208 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8b360825-bbe2-4522-b0b0-1f0bf6c4bc9f" containerName="telemetry-openstack-openstack-cell1" Oct 05 09:08:27 crc kubenswrapper[4846]: I1005 09:08:27.084292 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="8b360825-bbe2-4522-b0b0-1f0bf6c4bc9f" containerName="telemetry-openstack-openstack-cell1" Oct 05 09:08:27 crc kubenswrapper[4846]: I1005 09:08:27.084574 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="8b360825-bbe2-4522-b0b0-1f0bf6c4bc9f" containerName="telemetry-openstack-openstack-cell1" Oct 05 09:08:27 crc kubenswrapper[4846]: I1005 09:08:27.085720 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-sriov-openstack-openstack-cell1-ds9h4" Oct 05 09:08:27 crc kubenswrapper[4846]: I1005 09:08:27.089686 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-sriov-agent-neutron-config" Oct 05 09:08:27 crc kubenswrapper[4846]: I1005 09:08:27.090016 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 05 09:08:27 crc kubenswrapper[4846]: I1005 09:08:27.090198 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Oct 05 09:08:27 crc kubenswrapper[4846]: I1005 09:08:27.090696 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-7687s" Oct 05 09:08:27 crc kubenswrapper[4846]: I1005 09:08:27.090727 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Oct 05 09:08:27 crc kubenswrapper[4846]: I1005 09:08:27.106100 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-sriov-openstack-openstack-cell1-ds9h4"] Oct 05 09:08:27 crc kubenswrapper[4846]: I1005 09:08:27.187384 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e844e15c-7dd3-4056-8727-a24c9b65dbd8-ssh-key\") pod \"neutron-sriov-openstack-openstack-cell1-ds9h4\" (UID: \"e844e15c-7dd3-4056-8727-a24c9b65dbd8\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-ds9h4" Oct 05 09:08:27 crc kubenswrapper[4846]: I1005 09:08:27.187435 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2hhbj\" (UniqueName: \"kubernetes.io/projected/e844e15c-7dd3-4056-8727-a24c9b65dbd8-kube-api-access-2hhbj\") pod \"neutron-sriov-openstack-openstack-cell1-ds9h4\" (UID: \"e844e15c-7dd3-4056-8727-a24c9b65dbd8\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-ds9h4" Oct 05 09:08:27 crc kubenswrapper[4846]: I1005 09:08:27.187491 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e844e15c-7dd3-4056-8727-a24c9b65dbd8-inventory\") pod \"neutron-sriov-openstack-openstack-cell1-ds9h4\" (UID: \"e844e15c-7dd3-4056-8727-a24c9b65dbd8\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-ds9h4" Oct 05 09:08:27 crc kubenswrapper[4846]: I1005 09:08:27.187590 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e844e15c-7dd3-4056-8727-a24c9b65dbd8-neutron-sriov-combined-ca-bundle\") pod \"neutron-sriov-openstack-openstack-cell1-ds9h4\" (UID: \"e844e15c-7dd3-4056-8727-a24c9b65dbd8\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-ds9h4" Oct 05 09:08:27 crc kubenswrapper[4846]: I1005 09:08:27.187662 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-sriov-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/e844e15c-7dd3-4056-8727-a24c9b65dbd8-neutron-sriov-agent-neutron-config-0\") pod \"neutron-sriov-openstack-openstack-cell1-ds9h4\" (UID: \"e844e15c-7dd3-4056-8727-a24c9b65dbd8\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-ds9h4" Oct 05 09:08:27 crc kubenswrapper[4846]: I1005 09:08:27.290289 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e844e15c-7dd3-4056-8727-a24c9b65dbd8-ssh-key\") pod \"neutron-sriov-openstack-openstack-cell1-ds9h4\" (UID: \"e844e15c-7dd3-4056-8727-a24c9b65dbd8\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-ds9h4" Oct 05 09:08:27 crc kubenswrapper[4846]: I1005 09:08:27.290373 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2hhbj\" (UniqueName: \"kubernetes.io/projected/e844e15c-7dd3-4056-8727-a24c9b65dbd8-kube-api-access-2hhbj\") pod \"neutron-sriov-openstack-openstack-cell1-ds9h4\" (UID: \"e844e15c-7dd3-4056-8727-a24c9b65dbd8\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-ds9h4" Oct 05 09:08:27 crc kubenswrapper[4846]: I1005 09:08:27.290442 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e844e15c-7dd3-4056-8727-a24c9b65dbd8-inventory\") pod \"neutron-sriov-openstack-openstack-cell1-ds9h4\" (UID: \"e844e15c-7dd3-4056-8727-a24c9b65dbd8\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-ds9h4" Oct 05 09:08:27 crc kubenswrapper[4846]: I1005 09:08:27.290547 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e844e15c-7dd3-4056-8727-a24c9b65dbd8-neutron-sriov-combined-ca-bundle\") pod \"neutron-sriov-openstack-openstack-cell1-ds9h4\" (UID: \"e844e15c-7dd3-4056-8727-a24c9b65dbd8\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-ds9h4" Oct 05 09:08:27 crc kubenswrapper[4846]: I1005 09:08:27.290639 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-sriov-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/e844e15c-7dd3-4056-8727-a24c9b65dbd8-neutron-sriov-agent-neutron-config-0\") pod \"neutron-sriov-openstack-openstack-cell1-ds9h4\" (UID: \"e844e15c-7dd3-4056-8727-a24c9b65dbd8\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-ds9h4" Oct 05 09:08:27 crc kubenswrapper[4846]: I1005 09:08:27.295501 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e844e15c-7dd3-4056-8727-a24c9b65dbd8-ssh-key\") pod \"neutron-sriov-openstack-openstack-cell1-ds9h4\" (UID: \"e844e15c-7dd3-4056-8727-a24c9b65dbd8\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-ds9h4" Oct 05 09:08:27 crc kubenswrapper[4846]: I1005 09:08:27.295943 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-sriov-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/e844e15c-7dd3-4056-8727-a24c9b65dbd8-neutron-sriov-agent-neutron-config-0\") pod \"neutron-sriov-openstack-openstack-cell1-ds9h4\" (UID: \"e844e15c-7dd3-4056-8727-a24c9b65dbd8\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-ds9h4" Oct 05 09:08:27 crc kubenswrapper[4846]: I1005 09:08:27.297394 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e844e15c-7dd3-4056-8727-a24c9b65dbd8-neutron-sriov-combined-ca-bundle\") pod \"neutron-sriov-openstack-openstack-cell1-ds9h4\" (UID: \"e844e15c-7dd3-4056-8727-a24c9b65dbd8\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-ds9h4" Oct 05 09:08:27 crc kubenswrapper[4846]: I1005 09:08:27.300234 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e844e15c-7dd3-4056-8727-a24c9b65dbd8-inventory\") pod \"neutron-sriov-openstack-openstack-cell1-ds9h4\" (UID: \"e844e15c-7dd3-4056-8727-a24c9b65dbd8\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-ds9h4" Oct 05 09:08:27 crc kubenswrapper[4846]: I1005 09:08:27.318863 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2hhbj\" (UniqueName: \"kubernetes.io/projected/e844e15c-7dd3-4056-8727-a24c9b65dbd8-kube-api-access-2hhbj\") pod \"neutron-sriov-openstack-openstack-cell1-ds9h4\" (UID: \"e844e15c-7dd3-4056-8727-a24c9b65dbd8\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-ds9h4" Oct 05 09:08:27 crc kubenswrapper[4846]: I1005 09:08:27.404066 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-sriov-openstack-openstack-cell1-ds9h4" Oct 05 09:08:28 crc kubenswrapper[4846]: I1005 09:08:28.025100 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-sriov-openstack-openstack-cell1-ds9h4"] Oct 05 09:08:28 crc kubenswrapper[4846]: I1005 09:08:28.035067 4846 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 05 09:08:29 crc kubenswrapper[4846]: I1005 09:08:29.007942 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-sriov-openstack-openstack-cell1-ds9h4" event={"ID":"e844e15c-7dd3-4056-8727-a24c9b65dbd8","Type":"ContainerStarted","Data":"62450a6f22dcdd130ddefa63260041965f680f613c71ffe53653bc4acfffd33e"} Oct 05 09:08:29 crc kubenswrapper[4846]: I1005 09:08:29.008415 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-sriov-openstack-openstack-cell1-ds9h4" event={"ID":"e844e15c-7dd3-4056-8727-a24c9b65dbd8","Type":"ContainerStarted","Data":"241c0717a47a60438d90773eac612192320e1ba8dce1f1e4b93dbfe501c6d0e1"} Oct 05 09:08:29 crc kubenswrapper[4846]: I1005 09:08:29.034729 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-sriov-openstack-openstack-cell1-ds9h4" podStartSLOduration=1.4776374780000001 podStartE2EDuration="2.034380818s" podCreationTimestamp="2025-10-05 09:08:27 +0000 UTC" firstStartedPulling="2025-10-05 09:08:28.034668091 +0000 UTC m=+8430.275520906" lastFinishedPulling="2025-10-05 09:08:28.591411471 +0000 UTC m=+8430.832264246" observedRunningTime="2025-10-05 09:08:29.025456931 +0000 UTC m=+8431.266309706" watchObservedRunningTime="2025-10-05 09:08:29.034380818 +0000 UTC m=+8431.275233593" Oct 05 09:08:34 crc kubenswrapper[4846]: I1005 09:08:34.499616 4846 scope.go:117] "RemoveContainer" containerID="e39593bcdffa488d1cfc3d5c184d6e46655bb5c018d40df0f20d590de5498b3d" Oct 05 09:08:34 crc kubenswrapper[4846]: E1005 09:08:34.500008 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 09:08:46 crc kubenswrapper[4846]: I1005 09:08:46.499611 4846 scope.go:117] "RemoveContainer" containerID="e39593bcdffa488d1cfc3d5c184d6e46655bb5c018d40df0f20d590de5498b3d" Oct 05 09:08:46 crc kubenswrapper[4846]: E1005 09:08:46.501000 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 09:08:53 crc kubenswrapper[4846]: I1005 09:08:53.534230 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-9b72n"] Oct 05 09:08:53 crc kubenswrapper[4846]: I1005 09:08:53.536875 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9b72n" Oct 05 09:08:53 crc kubenswrapper[4846]: I1005 09:08:53.546387 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-9b72n"] Oct 05 09:08:53 crc kubenswrapper[4846]: I1005 09:08:53.632850 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-btkfm\" (UniqueName: \"kubernetes.io/projected/d38574e2-d394-42c0-ab3f-17e9c41b151e-kube-api-access-btkfm\") pod \"community-operators-9b72n\" (UID: \"d38574e2-d394-42c0-ab3f-17e9c41b151e\") " pod="openshift-marketplace/community-operators-9b72n" Oct 05 09:08:53 crc kubenswrapper[4846]: I1005 09:08:53.633505 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d38574e2-d394-42c0-ab3f-17e9c41b151e-utilities\") pod \"community-operators-9b72n\" (UID: \"d38574e2-d394-42c0-ab3f-17e9c41b151e\") " pod="openshift-marketplace/community-operators-9b72n" Oct 05 09:08:53 crc kubenswrapper[4846]: I1005 09:08:53.633637 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d38574e2-d394-42c0-ab3f-17e9c41b151e-catalog-content\") pod \"community-operators-9b72n\" (UID: \"d38574e2-d394-42c0-ab3f-17e9c41b151e\") " pod="openshift-marketplace/community-operators-9b72n" Oct 05 09:08:53 crc kubenswrapper[4846]: I1005 09:08:53.736955 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d38574e2-d394-42c0-ab3f-17e9c41b151e-catalog-content\") pod \"community-operators-9b72n\" (UID: \"d38574e2-d394-42c0-ab3f-17e9c41b151e\") " pod="openshift-marketplace/community-operators-9b72n" Oct 05 09:08:53 crc kubenswrapper[4846]: I1005 09:08:53.737045 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-btkfm\" (UniqueName: \"kubernetes.io/projected/d38574e2-d394-42c0-ab3f-17e9c41b151e-kube-api-access-btkfm\") pod \"community-operators-9b72n\" (UID: \"d38574e2-d394-42c0-ab3f-17e9c41b151e\") " pod="openshift-marketplace/community-operators-9b72n" Oct 05 09:08:53 crc kubenswrapper[4846]: I1005 09:08:53.737171 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d38574e2-d394-42c0-ab3f-17e9c41b151e-utilities\") pod \"community-operators-9b72n\" (UID: \"d38574e2-d394-42c0-ab3f-17e9c41b151e\") " pod="openshift-marketplace/community-operators-9b72n" Oct 05 09:08:53 crc kubenswrapper[4846]: I1005 09:08:53.737628 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d38574e2-d394-42c0-ab3f-17e9c41b151e-utilities\") pod \"community-operators-9b72n\" (UID: \"d38574e2-d394-42c0-ab3f-17e9c41b151e\") " pod="openshift-marketplace/community-operators-9b72n" Oct 05 09:08:53 crc kubenswrapper[4846]: I1005 09:08:53.737894 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d38574e2-d394-42c0-ab3f-17e9c41b151e-catalog-content\") pod \"community-operators-9b72n\" (UID: \"d38574e2-d394-42c0-ab3f-17e9c41b151e\") " pod="openshift-marketplace/community-operators-9b72n" Oct 05 09:08:53 crc kubenswrapper[4846]: I1005 09:08:53.763337 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-btkfm\" (UniqueName: \"kubernetes.io/projected/d38574e2-d394-42c0-ab3f-17e9c41b151e-kube-api-access-btkfm\") pod \"community-operators-9b72n\" (UID: \"d38574e2-d394-42c0-ab3f-17e9c41b151e\") " pod="openshift-marketplace/community-operators-9b72n" Oct 05 09:08:53 crc kubenswrapper[4846]: I1005 09:08:53.892570 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9b72n" Oct 05 09:08:54 crc kubenswrapper[4846]: I1005 09:08:54.384273 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-9b72n"] Oct 05 09:08:55 crc kubenswrapper[4846]: I1005 09:08:55.349495 4846 generic.go:334] "Generic (PLEG): container finished" podID="d38574e2-d394-42c0-ab3f-17e9c41b151e" containerID="b3f10e2dc372915041fcb8b0a951e6c96a916b60b6d5871a8f0e013c1d3ce7a7" exitCode=0 Oct 05 09:08:55 crc kubenswrapper[4846]: I1005 09:08:55.349568 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9b72n" event={"ID":"d38574e2-d394-42c0-ab3f-17e9c41b151e","Type":"ContainerDied","Data":"b3f10e2dc372915041fcb8b0a951e6c96a916b60b6d5871a8f0e013c1d3ce7a7"} Oct 05 09:08:55 crc kubenswrapper[4846]: I1005 09:08:55.350118 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9b72n" event={"ID":"d38574e2-d394-42c0-ab3f-17e9c41b151e","Type":"ContainerStarted","Data":"b1370742019f6158d7e8ce989c25e69e546769a4ebb0eaa85963f4dbb44ef948"} Oct 05 09:08:56 crc kubenswrapper[4846]: I1005 09:08:56.363351 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9b72n" event={"ID":"d38574e2-d394-42c0-ab3f-17e9c41b151e","Type":"ContainerStarted","Data":"5154b7896562620a5a19a07c25773292f1d5387049a356bf19c17840b0f8f2a5"} Oct 05 09:08:57 crc kubenswrapper[4846]: I1005 09:08:57.497646 4846 scope.go:117] "RemoveContainer" containerID="e39593bcdffa488d1cfc3d5c184d6e46655bb5c018d40df0f20d590de5498b3d" Oct 05 09:08:57 crc kubenswrapper[4846]: E1005 09:08:57.498344 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 09:08:58 crc kubenswrapper[4846]: I1005 09:08:58.393397 4846 generic.go:334] "Generic (PLEG): container finished" podID="d38574e2-d394-42c0-ab3f-17e9c41b151e" containerID="5154b7896562620a5a19a07c25773292f1d5387049a356bf19c17840b0f8f2a5" exitCode=0 Oct 05 09:08:58 crc kubenswrapper[4846]: I1005 09:08:58.393461 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9b72n" event={"ID":"d38574e2-d394-42c0-ab3f-17e9c41b151e","Type":"ContainerDied","Data":"5154b7896562620a5a19a07c25773292f1d5387049a356bf19c17840b0f8f2a5"} Oct 05 09:08:59 crc kubenswrapper[4846]: I1005 09:08:59.406767 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9b72n" event={"ID":"d38574e2-d394-42c0-ab3f-17e9c41b151e","Type":"ContainerStarted","Data":"8824a0264cb0f6120e3f8835426ea8279a34e6d90984890558e01b888d1d59d8"} Oct 05 09:08:59 crc kubenswrapper[4846]: I1005 09:08:59.437066 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-9b72n" podStartSLOduration=2.9575231 podStartE2EDuration="6.437040141s" podCreationTimestamp="2025-10-05 09:08:53 +0000 UTC" firstStartedPulling="2025-10-05 09:08:55.352787255 +0000 UTC m=+8457.593640040" lastFinishedPulling="2025-10-05 09:08:58.832304266 +0000 UTC m=+8461.073157081" observedRunningTime="2025-10-05 09:08:59.429588963 +0000 UTC m=+8461.670441758" watchObservedRunningTime="2025-10-05 09:08:59.437040141 +0000 UTC m=+8461.677892926" Oct 05 09:09:03 crc kubenswrapper[4846]: I1005 09:09:03.894525 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-9b72n" Oct 05 09:09:03 crc kubenswrapper[4846]: I1005 09:09:03.894969 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-9b72n" Oct 05 09:09:03 crc kubenswrapper[4846]: I1005 09:09:03.987438 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-9b72n" Oct 05 09:09:04 crc kubenswrapper[4846]: I1005 09:09:04.554279 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-9b72n" Oct 05 09:09:04 crc kubenswrapper[4846]: I1005 09:09:04.649533 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-9b72n"] Oct 05 09:09:06 crc kubenswrapper[4846]: I1005 09:09:06.500367 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-9b72n" podUID="d38574e2-d394-42c0-ab3f-17e9c41b151e" containerName="registry-server" containerID="cri-o://8824a0264cb0f6120e3f8835426ea8279a34e6d90984890558e01b888d1d59d8" gracePeriod=2 Oct 05 09:09:06 crc kubenswrapper[4846]: I1005 09:09:06.956648 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9b72n" Oct 05 09:09:07 crc kubenswrapper[4846]: I1005 09:09:07.039341 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d38574e2-d394-42c0-ab3f-17e9c41b151e-utilities\") pod \"d38574e2-d394-42c0-ab3f-17e9c41b151e\" (UID: \"d38574e2-d394-42c0-ab3f-17e9c41b151e\") " Oct 05 09:09:07 crc kubenswrapper[4846]: I1005 09:09:07.039595 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-btkfm\" (UniqueName: \"kubernetes.io/projected/d38574e2-d394-42c0-ab3f-17e9c41b151e-kube-api-access-btkfm\") pod \"d38574e2-d394-42c0-ab3f-17e9c41b151e\" (UID: \"d38574e2-d394-42c0-ab3f-17e9c41b151e\") " Oct 05 09:09:07 crc kubenswrapper[4846]: I1005 09:09:07.039826 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d38574e2-d394-42c0-ab3f-17e9c41b151e-utilities" (OuterVolumeSpecName: "utilities") pod "d38574e2-d394-42c0-ab3f-17e9c41b151e" (UID: "d38574e2-d394-42c0-ab3f-17e9c41b151e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 09:09:07 crc kubenswrapper[4846]: I1005 09:09:07.040558 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d38574e2-d394-42c0-ab3f-17e9c41b151e-catalog-content\") pod \"d38574e2-d394-42c0-ab3f-17e9c41b151e\" (UID: \"d38574e2-d394-42c0-ab3f-17e9c41b151e\") " Oct 05 09:09:07 crc kubenswrapper[4846]: I1005 09:09:07.041451 4846 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d38574e2-d394-42c0-ab3f-17e9c41b151e-utilities\") on node \"crc\" DevicePath \"\"" Oct 05 09:09:07 crc kubenswrapper[4846]: I1005 09:09:07.045047 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d38574e2-d394-42c0-ab3f-17e9c41b151e-kube-api-access-btkfm" (OuterVolumeSpecName: "kube-api-access-btkfm") pod "d38574e2-d394-42c0-ab3f-17e9c41b151e" (UID: "d38574e2-d394-42c0-ab3f-17e9c41b151e"). InnerVolumeSpecName "kube-api-access-btkfm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 09:09:07 crc kubenswrapper[4846]: I1005 09:09:07.143545 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d38574e2-d394-42c0-ab3f-17e9c41b151e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d38574e2-d394-42c0-ab3f-17e9c41b151e" (UID: "d38574e2-d394-42c0-ab3f-17e9c41b151e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 09:09:07 crc kubenswrapper[4846]: I1005 09:09:07.143667 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d38574e2-d394-42c0-ab3f-17e9c41b151e-catalog-content\") pod \"d38574e2-d394-42c0-ab3f-17e9c41b151e\" (UID: \"d38574e2-d394-42c0-ab3f-17e9c41b151e\") " Oct 05 09:09:07 crc kubenswrapper[4846]: W1005 09:09:07.143780 4846 empty_dir.go:500] Warning: Unmount skipped because path does not exist: /var/lib/kubelet/pods/d38574e2-d394-42c0-ab3f-17e9c41b151e/volumes/kubernetes.io~empty-dir/catalog-content Oct 05 09:09:07 crc kubenswrapper[4846]: I1005 09:09:07.143792 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d38574e2-d394-42c0-ab3f-17e9c41b151e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d38574e2-d394-42c0-ab3f-17e9c41b151e" (UID: "d38574e2-d394-42c0-ab3f-17e9c41b151e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 09:09:07 crc kubenswrapper[4846]: I1005 09:09:07.144230 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-btkfm\" (UniqueName: \"kubernetes.io/projected/d38574e2-d394-42c0-ab3f-17e9c41b151e-kube-api-access-btkfm\") on node \"crc\" DevicePath \"\"" Oct 05 09:09:07 crc kubenswrapper[4846]: I1005 09:09:07.144250 4846 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d38574e2-d394-42c0-ab3f-17e9c41b151e-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 05 09:09:07 crc kubenswrapper[4846]: I1005 09:09:07.516446 4846 generic.go:334] "Generic (PLEG): container finished" podID="d38574e2-d394-42c0-ab3f-17e9c41b151e" containerID="8824a0264cb0f6120e3f8835426ea8279a34e6d90984890558e01b888d1d59d8" exitCode=0 Oct 05 09:09:07 crc kubenswrapper[4846]: I1005 09:09:07.516508 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9b72n" event={"ID":"d38574e2-d394-42c0-ab3f-17e9c41b151e","Type":"ContainerDied","Data":"8824a0264cb0f6120e3f8835426ea8279a34e6d90984890558e01b888d1d59d8"} Oct 05 09:09:07 crc kubenswrapper[4846]: I1005 09:09:07.516547 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9b72n" event={"ID":"d38574e2-d394-42c0-ab3f-17e9c41b151e","Type":"ContainerDied","Data":"b1370742019f6158d7e8ce989c25e69e546769a4ebb0eaa85963f4dbb44ef948"} Oct 05 09:09:07 crc kubenswrapper[4846]: I1005 09:09:07.516576 4846 scope.go:117] "RemoveContainer" containerID="8824a0264cb0f6120e3f8835426ea8279a34e6d90984890558e01b888d1d59d8" Oct 05 09:09:07 crc kubenswrapper[4846]: I1005 09:09:07.517319 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9b72n" Oct 05 09:09:07 crc kubenswrapper[4846]: I1005 09:09:07.549367 4846 scope.go:117] "RemoveContainer" containerID="5154b7896562620a5a19a07c25773292f1d5387049a356bf19c17840b0f8f2a5" Oct 05 09:09:07 crc kubenswrapper[4846]: I1005 09:09:07.584735 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-9b72n"] Oct 05 09:09:07 crc kubenswrapper[4846]: I1005 09:09:07.594336 4846 scope.go:117] "RemoveContainer" containerID="b3f10e2dc372915041fcb8b0a951e6c96a916b60b6d5871a8f0e013c1d3ce7a7" Oct 05 09:09:07 crc kubenswrapper[4846]: I1005 09:09:07.601321 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-9b72n"] Oct 05 09:09:07 crc kubenswrapper[4846]: I1005 09:09:07.638169 4846 scope.go:117] "RemoveContainer" containerID="8824a0264cb0f6120e3f8835426ea8279a34e6d90984890558e01b888d1d59d8" Oct 05 09:09:07 crc kubenswrapper[4846]: E1005 09:09:07.638920 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8824a0264cb0f6120e3f8835426ea8279a34e6d90984890558e01b888d1d59d8\": container with ID starting with 8824a0264cb0f6120e3f8835426ea8279a34e6d90984890558e01b888d1d59d8 not found: ID does not exist" containerID="8824a0264cb0f6120e3f8835426ea8279a34e6d90984890558e01b888d1d59d8" Oct 05 09:09:07 crc kubenswrapper[4846]: I1005 09:09:07.639012 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8824a0264cb0f6120e3f8835426ea8279a34e6d90984890558e01b888d1d59d8"} err="failed to get container status \"8824a0264cb0f6120e3f8835426ea8279a34e6d90984890558e01b888d1d59d8\": rpc error: code = NotFound desc = could not find container \"8824a0264cb0f6120e3f8835426ea8279a34e6d90984890558e01b888d1d59d8\": container with ID starting with 8824a0264cb0f6120e3f8835426ea8279a34e6d90984890558e01b888d1d59d8 not found: ID does not exist" Oct 05 09:09:07 crc kubenswrapper[4846]: I1005 09:09:07.639069 4846 scope.go:117] "RemoveContainer" containerID="5154b7896562620a5a19a07c25773292f1d5387049a356bf19c17840b0f8f2a5" Oct 05 09:09:07 crc kubenswrapper[4846]: E1005 09:09:07.639480 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5154b7896562620a5a19a07c25773292f1d5387049a356bf19c17840b0f8f2a5\": container with ID starting with 5154b7896562620a5a19a07c25773292f1d5387049a356bf19c17840b0f8f2a5 not found: ID does not exist" containerID="5154b7896562620a5a19a07c25773292f1d5387049a356bf19c17840b0f8f2a5" Oct 05 09:09:07 crc kubenswrapper[4846]: I1005 09:09:07.639568 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5154b7896562620a5a19a07c25773292f1d5387049a356bf19c17840b0f8f2a5"} err="failed to get container status \"5154b7896562620a5a19a07c25773292f1d5387049a356bf19c17840b0f8f2a5\": rpc error: code = NotFound desc = could not find container \"5154b7896562620a5a19a07c25773292f1d5387049a356bf19c17840b0f8f2a5\": container with ID starting with 5154b7896562620a5a19a07c25773292f1d5387049a356bf19c17840b0f8f2a5 not found: ID does not exist" Oct 05 09:09:07 crc kubenswrapper[4846]: I1005 09:09:07.639599 4846 scope.go:117] "RemoveContainer" containerID="b3f10e2dc372915041fcb8b0a951e6c96a916b60b6d5871a8f0e013c1d3ce7a7" Oct 05 09:09:07 crc kubenswrapper[4846]: E1005 09:09:07.640024 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b3f10e2dc372915041fcb8b0a951e6c96a916b60b6d5871a8f0e013c1d3ce7a7\": container with ID starting with b3f10e2dc372915041fcb8b0a951e6c96a916b60b6d5871a8f0e013c1d3ce7a7 not found: ID does not exist" containerID="b3f10e2dc372915041fcb8b0a951e6c96a916b60b6d5871a8f0e013c1d3ce7a7" Oct 05 09:09:07 crc kubenswrapper[4846]: I1005 09:09:07.640096 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b3f10e2dc372915041fcb8b0a951e6c96a916b60b6d5871a8f0e013c1d3ce7a7"} err="failed to get container status \"b3f10e2dc372915041fcb8b0a951e6c96a916b60b6d5871a8f0e013c1d3ce7a7\": rpc error: code = NotFound desc = could not find container \"b3f10e2dc372915041fcb8b0a951e6c96a916b60b6d5871a8f0e013c1d3ce7a7\": container with ID starting with b3f10e2dc372915041fcb8b0a951e6c96a916b60b6d5871a8f0e013c1d3ce7a7 not found: ID does not exist" Oct 05 09:09:08 crc kubenswrapper[4846]: I1005 09:09:08.517292 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d38574e2-d394-42c0-ab3f-17e9c41b151e" path="/var/lib/kubelet/pods/d38574e2-d394-42c0-ab3f-17e9c41b151e/volumes" Oct 05 09:09:11 crc kubenswrapper[4846]: I1005 09:09:11.498675 4846 scope.go:117] "RemoveContainer" containerID="e39593bcdffa488d1cfc3d5c184d6e46655bb5c018d40df0f20d590de5498b3d" Oct 05 09:09:11 crc kubenswrapper[4846]: E1005 09:09:11.499134 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 09:09:23 crc kubenswrapper[4846]: I1005 09:09:23.498078 4846 scope.go:117] "RemoveContainer" containerID="e39593bcdffa488d1cfc3d5c184d6e46655bb5c018d40df0f20d590de5498b3d" Oct 05 09:09:23 crc kubenswrapper[4846]: E1005 09:09:23.498944 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 09:09:33 crc kubenswrapper[4846]: I1005 09:09:33.723389 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-4dmmj"] Oct 05 09:09:33 crc kubenswrapper[4846]: E1005 09:09:33.724674 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d38574e2-d394-42c0-ab3f-17e9c41b151e" containerName="extract-utilities" Oct 05 09:09:33 crc kubenswrapper[4846]: I1005 09:09:33.724696 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="d38574e2-d394-42c0-ab3f-17e9c41b151e" containerName="extract-utilities" Oct 05 09:09:33 crc kubenswrapper[4846]: E1005 09:09:33.724740 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d38574e2-d394-42c0-ab3f-17e9c41b151e" containerName="extract-content" Oct 05 09:09:33 crc kubenswrapper[4846]: I1005 09:09:33.724752 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="d38574e2-d394-42c0-ab3f-17e9c41b151e" containerName="extract-content" Oct 05 09:09:33 crc kubenswrapper[4846]: E1005 09:09:33.724777 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d38574e2-d394-42c0-ab3f-17e9c41b151e" containerName="registry-server" Oct 05 09:09:33 crc kubenswrapper[4846]: I1005 09:09:33.724789 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="d38574e2-d394-42c0-ab3f-17e9c41b151e" containerName="registry-server" Oct 05 09:09:33 crc kubenswrapper[4846]: I1005 09:09:33.725206 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="d38574e2-d394-42c0-ab3f-17e9c41b151e" containerName="registry-server" Oct 05 09:09:33 crc kubenswrapper[4846]: I1005 09:09:33.728346 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4dmmj" Oct 05 09:09:33 crc kubenswrapper[4846]: I1005 09:09:33.732615 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-4dmmj"] Oct 05 09:09:33 crc kubenswrapper[4846]: I1005 09:09:33.842433 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5e60b3c1-9577-40d9-b81c-5a44dd7e0df1-utilities\") pod \"redhat-marketplace-4dmmj\" (UID: \"5e60b3c1-9577-40d9-b81c-5a44dd7e0df1\") " pod="openshift-marketplace/redhat-marketplace-4dmmj" Oct 05 09:09:33 crc kubenswrapper[4846]: I1005 09:09:33.842663 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5e60b3c1-9577-40d9-b81c-5a44dd7e0df1-catalog-content\") pod \"redhat-marketplace-4dmmj\" (UID: \"5e60b3c1-9577-40d9-b81c-5a44dd7e0df1\") " pod="openshift-marketplace/redhat-marketplace-4dmmj" Oct 05 09:09:33 crc kubenswrapper[4846]: I1005 09:09:33.842722 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xgvhp\" (UniqueName: \"kubernetes.io/projected/5e60b3c1-9577-40d9-b81c-5a44dd7e0df1-kube-api-access-xgvhp\") pod \"redhat-marketplace-4dmmj\" (UID: \"5e60b3c1-9577-40d9-b81c-5a44dd7e0df1\") " pod="openshift-marketplace/redhat-marketplace-4dmmj" Oct 05 09:09:33 crc kubenswrapper[4846]: I1005 09:09:33.944518 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5e60b3c1-9577-40d9-b81c-5a44dd7e0df1-catalog-content\") pod \"redhat-marketplace-4dmmj\" (UID: \"5e60b3c1-9577-40d9-b81c-5a44dd7e0df1\") " pod="openshift-marketplace/redhat-marketplace-4dmmj" Oct 05 09:09:33 crc kubenswrapper[4846]: I1005 09:09:33.944895 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xgvhp\" (UniqueName: \"kubernetes.io/projected/5e60b3c1-9577-40d9-b81c-5a44dd7e0df1-kube-api-access-xgvhp\") pod \"redhat-marketplace-4dmmj\" (UID: \"5e60b3c1-9577-40d9-b81c-5a44dd7e0df1\") " pod="openshift-marketplace/redhat-marketplace-4dmmj" Oct 05 09:09:33 crc kubenswrapper[4846]: I1005 09:09:33.944972 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5e60b3c1-9577-40d9-b81c-5a44dd7e0df1-catalog-content\") pod \"redhat-marketplace-4dmmj\" (UID: \"5e60b3c1-9577-40d9-b81c-5a44dd7e0df1\") " pod="openshift-marketplace/redhat-marketplace-4dmmj" Oct 05 09:09:33 crc kubenswrapper[4846]: I1005 09:09:33.945070 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5e60b3c1-9577-40d9-b81c-5a44dd7e0df1-utilities\") pod \"redhat-marketplace-4dmmj\" (UID: \"5e60b3c1-9577-40d9-b81c-5a44dd7e0df1\") " pod="openshift-marketplace/redhat-marketplace-4dmmj" Oct 05 09:09:33 crc kubenswrapper[4846]: I1005 09:09:33.945324 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5e60b3c1-9577-40d9-b81c-5a44dd7e0df1-utilities\") pod \"redhat-marketplace-4dmmj\" (UID: \"5e60b3c1-9577-40d9-b81c-5a44dd7e0df1\") " pod="openshift-marketplace/redhat-marketplace-4dmmj" Oct 05 09:09:33 crc kubenswrapper[4846]: I1005 09:09:33.976756 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xgvhp\" (UniqueName: \"kubernetes.io/projected/5e60b3c1-9577-40d9-b81c-5a44dd7e0df1-kube-api-access-xgvhp\") pod \"redhat-marketplace-4dmmj\" (UID: \"5e60b3c1-9577-40d9-b81c-5a44dd7e0df1\") " pod="openshift-marketplace/redhat-marketplace-4dmmj" Oct 05 09:09:34 crc kubenswrapper[4846]: I1005 09:09:34.061010 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4dmmj" Oct 05 09:09:34 crc kubenswrapper[4846]: I1005 09:09:34.498586 4846 scope.go:117] "RemoveContainer" containerID="e39593bcdffa488d1cfc3d5c184d6e46655bb5c018d40df0f20d590de5498b3d" Oct 05 09:09:34 crc kubenswrapper[4846]: E1005 09:09:34.499830 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 09:09:34 crc kubenswrapper[4846]: I1005 09:09:34.577601 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-4dmmj"] Oct 05 09:09:34 crc kubenswrapper[4846]: W1005 09:09:34.586738 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5e60b3c1_9577_40d9_b81c_5a44dd7e0df1.slice/crio-78300a70d82d504878af62b697e91d45c5a311784a88772a8c5277b69c12e3ac WatchSource:0}: Error finding container 78300a70d82d504878af62b697e91d45c5a311784a88772a8c5277b69c12e3ac: Status 404 returned error can't find the container with id 78300a70d82d504878af62b697e91d45c5a311784a88772a8c5277b69c12e3ac Oct 05 09:09:34 crc kubenswrapper[4846]: I1005 09:09:34.853293 4846 generic.go:334] "Generic (PLEG): container finished" podID="5e60b3c1-9577-40d9-b81c-5a44dd7e0df1" containerID="e1ca40fb9e30d51ff5f94207cbfbb5dd3bb8e31bcd467cfae71def0b5a8a2463" exitCode=0 Oct 05 09:09:34 crc kubenswrapper[4846]: I1005 09:09:34.853368 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4dmmj" event={"ID":"5e60b3c1-9577-40d9-b81c-5a44dd7e0df1","Type":"ContainerDied","Data":"e1ca40fb9e30d51ff5f94207cbfbb5dd3bb8e31bcd467cfae71def0b5a8a2463"} Oct 05 09:09:34 crc kubenswrapper[4846]: I1005 09:09:34.853628 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4dmmj" event={"ID":"5e60b3c1-9577-40d9-b81c-5a44dd7e0df1","Type":"ContainerStarted","Data":"78300a70d82d504878af62b697e91d45c5a311784a88772a8c5277b69c12e3ac"} Oct 05 09:09:36 crc kubenswrapper[4846]: I1005 09:09:36.882546 4846 generic.go:334] "Generic (PLEG): container finished" podID="5e60b3c1-9577-40d9-b81c-5a44dd7e0df1" containerID="1ff01c4d6ea43def20eb5f1b4899cc185db32e187e4cd4b68e1c9a1bfa0400e2" exitCode=0 Oct 05 09:09:36 crc kubenswrapper[4846]: I1005 09:09:36.882615 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4dmmj" event={"ID":"5e60b3c1-9577-40d9-b81c-5a44dd7e0df1","Type":"ContainerDied","Data":"1ff01c4d6ea43def20eb5f1b4899cc185db32e187e4cd4b68e1c9a1bfa0400e2"} Oct 05 09:09:37 crc kubenswrapper[4846]: I1005 09:09:37.899760 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4dmmj" event={"ID":"5e60b3c1-9577-40d9-b81c-5a44dd7e0df1","Type":"ContainerStarted","Data":"0e3e09da8dc0c586451462b43675f911b1c42d9fd48a8ce3b4c96a19ab98bc9e"} Oct 05 09:09:37 crc kubenswrapper[4846]: I1005 09:09:37.928501 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-4dmmj" podStartSLOduration=2.507838666 podStartE2EDuration="4.92848357s" podCreationTimestamp="2025-10-05 09:09:33 +0000 UTC" firstStartedPulling="2025-10-05 09:09:34.856087333 +0000 UTC m=+8497.096940138" lastFinishedPulling="2025-10-05 09:09:37.276732257 +0000 UTC m=+8499.517585042" observedRunningTime="2025-10-05 09:09:37.923125058 +0000 UTC m=+8500.163977843" watchObservedRunningTime="2025-10-05 09:09:37.92848357 +0000 UTC m=+8500.169336355" Oct 05 09:09:44 crc kubenswrapper[4846]: I1005 09:09:44.062068 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-4dmmj" Oct 05 09:09:44 crc kubenswrapper[4846]: I1005 09:09:44.062831 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-4dmmj" Oct 05 09:09:44 crc kubenswrapper[4846]: I1005 09:09:44.132594 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-4dmmj" Oct 05 09:09:45 crc kubenswrapper[4846]: I1005 09:09:45.072595 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-4dmmj" Oct 05 09:09:45 crc kubenswrapper[4846]: I1005 09:09:45.142628 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-4dmmj"] Oct 05 09:09:47 crc kubenswrapper[4846]: I1005 09:09:47.010076 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-4dmmj" podUID="5e60b3c1-9577-40d9-b81c-5a44dd7e0df1" containerName="registry-server" containerID="cri-o://0e3e09da8dc0c586451462b43675f911b1c42d9fd48a8ce3b4c96a19ab98bc9e" gracePeriod=2 Oct 05 09:09:47 crc kubenswrapper[4846]: I1005 09:09:47.551160 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4dmmj" Oct 05 09:09:47 crc kubenswrapper[4846]: I1005 09:09:47.673885 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5e60b3c1-9577-40d9-b81c-5a44dd7e0df1-catalog-content\") pod \"5e60b3c1-9577-40d9-b81c-5a44dd7e0df1\" (UID: \"5e60b3c1-9577-40d9-b81c-5a44dd7e0df1\") " Oct 05 09:09:47 crc kubenswrapper[4846]: I1005 09:09:47.674062 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5e60b3c1-9577-40d9-b81c-5a44dd7e0df1-utilities\") pod \"5e60b3c1-9577-40d9-b81c-5a44dd7e0df1\" (UID: \"5e60b3c1-9577-40d9-b81c-5a44dd7e0df1\") " Oct 05 09:09:47 crc kubenswrapper[4846]: I1005 09:09:47.674098 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xgvhp\" (UniqueName: \"kubernetes.io/projected/5e60b3c1-9577-40d9-b81c-5a44dd7e0df1-kube-api-access-xgvhp\") pod \"5e60b3c1-9577-40d9-b81c-5a44dd7e0df1\" (UID: \"5e60b3c1-9577-40d9-b81c-5a44dd7e0df1\") " Oct 05 09:09:47 crc kubenswrapper[4846]: I1005 09:09:47.674882 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5e60b3c1-9577-40d9-b81c-5a44dd7e0df1-utilities" (OuterVolumeSpecName: "utilities") pod "5e60b3c1-9577-40d9-b81c-5a44dd7e0df1" (UID: "5e60b3c1-9577-40d9-b81c-5a44dd7e0df1"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 09:09:47 crc kubenswrapper[4846]: I1005 09:09:47.680726 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5e60b3c1-9577-40d9-b81c-5a44dd7e0df1-kube-api-access-xgvhp" (OuterVolumeSpecName: "kube-api-access-xgvhp") pod "5e60b3c1-9577-40d9-b81c-5a44dd7e0df1" (UID: "5e60b3c1-9577-40d9-b81c-5a44dd7e0df1"). InnerVolumeSpecName "kube-api-access-xgvhp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 09:09:47 crc kubenswrapper[4846]: I1005 09:09:47.690240 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5e60b3c1-9577-40d9-b81c-5a44dd7e0df1-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5e60b3c1-9577-40d9-b81c-5a44dd7e0df1" (UID: "5e60b3c1-9577-40d9-b81c-5a44dd7e0df1"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 09:09:47 crc kubenswrapper[4846]: I1005 09:09:47.776490 4846 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5e60b3c1-9577-40d9-b81c-5a44dd7e0df1-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 05 09:09:47 crc kubenswrapper[4846]: I1005 09:09:47.776531 4846 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5e60b3c1-9577-40d9-b81c-5a44dd7e0df1-utilities\") on node \"crc\" DevicePath \"\"" Oct 05 09:09:47 crc kubenswrapper[4846]: I1005 09:09:47.776547 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xgvhp\" (UniqueName: \"kubernetes.io/projected/5e60b3c1-9577-40d9-b81c-5a44dd7e0df1-kube-api-access-xgvhp\") on node \"crc\" DevicePath \"\"" Oct 05 09:09:48 crc kubenswrapper[4846]: I1005 09:09:48.025892 4846 generic.go:334] "Generic (PLEG): container finished" podID="5e60b3c1-9577-40d9-b81c-5a44dd7e0df1" containerID="0e3e09da8dc0c586451462b43675f911b1c42d9fd48a8ce3b4c96a19ab98bc9e" exitCode=0 Oct 05 09:09:48 crc kubenswrapper[4846]: I1005 09:09:48.025959 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4dmmj" event={"ID":"5e60b3c1-9577-40d9-b81c-5a44dd7e0df1","Type":"ContainerDied","Data":"0e3e09da8dc0c586451462b43675f911b1c42d9fd48a8ce3b4c96a19ab98bc9e"} Oct 05 09:09:48 crc kubenswrapper[4846]: I1005 09:09:48.026001 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4dmmj" event={"ID":"5e60b3c1-9577-40d9-b81c-5a44dd7e0df1","Type":"ContainerDied","Data":"78300a70d82d504878af62b697e91d45c5a311784a88772a8c5277b69c12e3ac"} Oct 05 09:09:48 crc kubenswrapper[4846]: I1005 09:09:48.026029 4846 scope.go:117] "RemoveContainer" containerID="0e3e09da8dc0c586451462b43675f911b1c42d9fd48a8ce3b4c96a19ab98bc9e" Oct 05 09:09:48 crc kubenswrapper[4846]: I1005 09:09:48.027459 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4dmmj" Oct 05 09:09:48 crc kubenswrapper[4846]: I1005 09:09:48.061038 4846 scope.go:117] "RemoveContainer" containerID="1ff01c4d6ea43def20eb5f1b4899cc185db32e187e4cd4b68e1c9a1bfa0400e2" Oct 05 09:09:48 crc kubenswrapper[4846]: I1005 09:09:48.095767 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-4dmmj"] Oct 05 09:09:48 crc kubenswrapper[4846]: I1005 09:09:48.104332 4846 scope.go:117] "RemoveContainer" containerID="e1ca40fb9e30d51ff5f94207cbfbb5dd3bb8e31bcd467cfae71def0b5a8a2463" Oct 05 09:09:48 crc kubenswrapper[4846]: I1005 09:09:48.122412 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-4dmmj"] Oct 05 09:09:48 crc kubenswrapper[4846]: I1005 09:09:48.179537 4846 scope.go:117] "RemoveContainer" containerID="0e3e09da8dc0c586451462b43675f911b1c42d9fd48a8ce3b4c96a19ab98bc9e" Oct 05 09:09:48 crc kubenswrapper[4846]: E1005 09:09:48.180590 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0e3e09da8dc0c586451462b43675f911b1c42d9fd48a8ce3b4c96a19ab98bc9e\": container with ID starting with 0e3e09da8dc0c586451462b43675f911b1c42d9fd48a8ce3b4c96a19ab98bc9e not found: ID does not exist" containerID="0e3e09da8dc0c586451462b43675f911b1c42d9fd48a8ce3b4c96a19ab98bc9e" Oct 05 09:09:48 crc kubenswrapper[4846]: I1005 09:09:48.180635 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0e3e09da8dc0c586451462b43675f911b1c42d9fd48a8ce3b4c96a19ab98bc9e"} err="failed to get container status \"0e3e09da8dc0c586451462b43675f911b1c42d9fd48a8ce3b4c96a19ab98bc9e\": rpc error: code = NotFound desc = could not find container \"0e3e09da8dc0c586451462b43675f911b1c42d9fd48a8ce3b4c96a19ab98bc9e\": container with ID starting with 0e3e09da8dc0c586451462b43675f911b1c42d9fd48a8ce3b4c96a19ab98bc9e not found: ID does not exist" Oct 05 09:09:48 crc kubenswrapper[4846]: I1005 09:09:48.180663 4846 scope.go:117] "RemoveContainer" containerID="1ff01c4d6ea43def20eb5f1b4899cc185db32e187e4cd4b68e1c9a1bfa0400e2" Oct 05 09:09:48 crc kubenswrapper[4846]: E1005 09:09:48.181083 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1ff01c4d6ea43def20eb5f1b4899cc185db32e187e4cd4b68e1c9a1bfa0400e2\": container with ID starting with 1ff01c4d6ea43def20eb5f1b4899cc185db32e187e4cd4b68e1c9a1bfa0400e2 not found: ID does not exist" containerID="1ff01c4d6ea43def20eb5f1b4899cc185db32e187e4cd4b68e1c9a1bfa0400e2" Oct 05 09:09:48 crc kubenswrapper[4846]: I1005 09:09:48.181116 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1ff01c4d6ea43def20eb5f1b4899cc185db32e187e4cd4b68e1c9a1bfa0400e2"} err="failed to get container status \"1ff01c4d6ea43def20eb5f1b4899cc185db32e187e4cd4b68e1c9a1bfa0400e2\": rpc error: code = NotFound desc = could not find container \"1ff01c4d6ea43def20eb5f1b4899cc185db32e187e4cd4b68e1c9a1bfa0400e2\": container with ID starting with 1ff01c4d6ea43def20eb5f1b4899cc185db32e187e4cd4b68e1c9a1bfa0400e2 not found: ID does not exist" Oct 05 09:09:48 crc kubenswrapper[4846]: I1005 09:09:48.181139 4846 scope.go:117] "RemoveContainer" containerID="e1ca40fb9e30d51ff5f94207cbfbb5dd3bb8e31bcd467cfae71def0b5a8a2463" Oct 05 09:09:48 crc kubenswrapper[4846]: E1005 09:09:48.181536 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e1ca40fb9e30d51ff5f94207cbfbb5dd3bb8e31bcd467cfae71def0b5a8a2463\": container with ID starting with e1ca40fb9e30d51ff5f94207cbfbb5dd3bb8e31bcd467cfae71def0b5a8a2463 not found: ID does not exist" containerID="e1ca40fb9e30d51ff5f94207cbfbb5dd3bb8e31bcd467cfae71def0b5a8a2463" Oct 05 09:09:48 crc kubenswrapper[4846]: I1005 09:09:48.181592 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e1ca40fb9e30d51ff5f94207cbfbb5dd3bb8e31bcd467cfae71def0b5a8a2463"} err="failed to get container status \"e1ca40fb9e30d51ff5f94207cbfbb5dd3bb8e31bcd467cfae71def0b5a8a2463\": rpc error: code = NotFound desc = could not find container \"e1ca40fb9e30d51ff5f94207cbfbb5dd3bb8e31bcd467cfae71def0b5a8a2463\": container with ID starting with e1ca40fb9e30d51ff5f94207cbfbb5dd3bb8e31bcd467cfae71def0b5a8a2463 not found: ID does not exist" Oct 05 09:09:48 crc kubenswrapper[4846]: I1005 09:09:48.515527 4846 scope.go:117] "RemoveContainer" containerID="e39593bcdffa488d1cfc3d5c184d6e46655bb5c018d40df0f20d590de5498b3d" Oct 05 09:09:48 crc kubenswrapper[4846]: E1005 09:09:48.516165 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 09:09:48 crc kubenswrapper[4846]: I1005 09:09:48.520722 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5e60b3c1-9577-40d9-b81c-5a44dd7e0df1" path="/var/lib/kubelet/pods/5e60b3c1-9577-40d9-b81c-5a44dd7e0df1/volumes" Oct 05 09:10:04 crc kubenswrapper[4846]: I1005 09:10:04.512037 4846 scope.go:117] "RemoveContainer" containerID="e39593bcdffa488d1cfc3d5c184d6e46655bb5c018d40df0f20d590de5498b3d" Oct 05 09:10:04 crc kubenswrapper[4846]: E1005 09:10:04.513012 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 09:10:15 crc kubenswrapper[4846]: I1005 09:10:15.497940 4846 scope.go:117] "RemoveContainer" containerID="e39593bcdffa488d1cfc3d5c184d6e46655bb5c018d40df0f20d590de5498b3d" Oct 05 09:10:15 crc kubenswrapper[4846]: E1005 09:10:15.498633 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 09:10:28 crc kubenswrapper[4846]: I1005 09:10:28.503755 4846 scope.go:117] "RemoveContainer" containerID="e39593bcdffa488d1cfc3d5c184d6e46655bb5c018d40df0f20d590de5498b3d" Oct 05 09:10:28 crc kubenswrapper[4846]: E1005 09:10:28.504764 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 09:10:43 crc kubenswrapper[4846]: I1005 09:10:43.497958 4846 scope.go:117] "RemoveContainer" containerID="e39593bcdffa488d1cfc3d5c184d6e46655bb5c018d40df0f20d590de5498b3d" Oct 05 09:10:43 crc kubenswrapper[4846]: E1005 09:10:43.499644 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 09:10:57 crc kubenswrapper[4846]: I1005 09:10:57.498316 4846 scope.go:117] "RemoveContainer" containerID="e39593bcdffa488d1cfc3d5c184d6e46655bb5c018d40df0f20d590de5498b3d" Oct 05 09:10:57 crc kubenswrapper[4846]: E1005 09:10:57.499558 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 09:11:11 crc kubenswrapper[4846]: I1005 09:11:11.498721 4846 scope.go:117] "RemoveContainer" containerID="e39593bcdffa488d1cfc3d5c184d6e46655bb5c018d40df0f20d590de5498b3d" Oct 05 09:11:11 crc kubenswrapper[4846]: E1005 09:11:11.500768 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 09:11:24 crc kubenswrapper[4846]: I1005 09:11:24.499079 4846 scope.go:117] "RemoveContainer" containerID="e39593bcdffa488d1cfc3d5c184d6e46655bb5c018d40df0f20d590de5498b3d" Oct 05 09:11:24 crc kubenswrapper[4846]: E1005 09:11:24.500482 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 09:11:36 crc kubenswrapper[4846]: I1005 09:11:36.498038 4846 scope.go:117] "RemoveContainer" containerID="e39593bcdffa488d1cfc3d5c184d6e46655bb5c018d40df0f20d590de5498b3d" Oct 05 09:11:36 crc kubenswrapper[4846]: E1005 09:11:36.498800 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 09:11:42 crc kubenswrapper[4846]: I1005 09:11:42.471340 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-zfjb5"] Oct 05 09:11:42 crc kubenswrapper[4846]: E1005 09:11:42.472698 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5e60b3c1-9577-40d9-b81c-5a44dd7e0df1" containerName="extract-utilities" Oct 05 09:11:42 crc kubenswrapper[4846]: I1005 09:11:42.472728 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="5e60b3c1-9577-40d9-b81c-5a44dd7e0df1" containerName="extract-utilities" Oct 05 09:11:42 crc kubenswrapper[4846]: E1005 09:11:42.472757 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5e60b3c1-9577-40d9-b81c-5a44dd7e0df1" containerName="registry-server" Oct 05 09:11:42 crc kubenswrapper[4846]: I1005 09:11:42.472769 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="5e60b3c1-9577-40d9-b81c-5a44dd7e0df1" containerName="registry-server" Oct 05 09:11:42 crc kubenswrapper[4846]: E1005 09:11:42.472847 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5e60b3c1-9577-40d9-b81c-5a44dd7e0df1" containerName="extract-content" Oct 05 09:11:42 crc kubenswrapper[4846]: I1005 09:11:42.472861 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="5e60b3c1-9577-40d9-b81c-5a44dd7e0df1" containerName="extract-content" Oct 05 09:11:42 crc kubenswrapper[4846]: I1005 09:11:42.473257 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="5e60b3c1-9577-40d9-b81c-5a44dd7e0df1" containerName="registry-server" Oct 05 09:11:42 crc kubenswrapper[4846]: I1005 09:11:42.476091 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-zfjb5" Oct 05 09:11:42 crc kubenswrapper[4846]: I1005 09:11:42.496276 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-zfjb5"] Oct 05 09:11:42 crc kubenswrapper[4846]: I1005 09:11:42.648210 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a1dbb719-cf21-487d-9605-0dc1b615046d-utilities\") pod \"redhat-operators-zfjb5\" (UID: \"a1dbb719-cf21-487d-9605-0dc1b615046d\") " pod="openshift-marketplace/redhat-operators-zfjb5" Oct 05 09:11:42 crc kubenswrapper[4846]: I1005 09:11:42.648383 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a1dbb719-cf21-487d-9605-0dc1b615046d-catalog-content\") pod \"redhat-operators-zfjb5\" (UID: \"a1dbb719-cf21-487d-9605-0dc1b615046d\") " pod="openshift-marketplace/redhat-operators-zfjb5" Oct 05 09:11:42 crc kubenswrapper[4846]: I1005 09:11:42.648420 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rfljh\" (UniqueName: \"kubernetes.io/projected/a1dbb719-cf21-487d-9605-0dc1b615046d-kube-api-access-rfljh\") pod \"redhat-operators-zfjb5\" (UID: \"a1dbb719-cf21-487d-9605-0dc1b615046d\") " pod="openshift-marketplace/redhat-operators-zfjb5" Oct 05 09:11:42 crc kubenswrapper[4846]: I1005 09:11:42.750284 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a1dbb719-cf21-487d-9605-0dc1b615046d-catalog-content\") pod \"redhat-operators-zfjb5\" (UID: \"a1dbb719-cf21-487d-9605-0dc1b615046d\") " pod="openshift-marketplace/redhat-operators-zfjb5" Oct 05 09:11:42 crc kubenswrapper[4846]: I1005 09:11:42.750366 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rfljh\" (UniqueName: \"kubernetes.io/projected/a1dbb719-cf21-487d-9605-0dc1b615046d-kube-api-access-rfljh\") pod \"redhat-operators-zfjb5\" (UID: \"a1dbb719-cf21-487d-9605-0dc1b615046d\") " pod="openshift-marketplace/redhat-operators-zfjb5" Oct 05 09:11:42 crc kubenswrapper[4846]: I1005 09:11:42.750451 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a1dbb719-cf21-487d-9605-0dc1b615046d-utilities\") pod \"redhat-operators-zfjb5\" (UID: \"a1dbb719-cf21-487d-9605-0dc1b615046d\") " pod="openshift-marketplace/redhat-operators-zfjb5" Oct 05 09:11:42 crc kubenswrapper[4846]: I1005 09:11:42.750934 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a1dbb719-cf21-487d-9605-0dc1b615046d-catalog-content\") pod \"redhat-operators-zfjb5\" (UID: \"a1dbb719-cf21-487d-9605-0dc1b615046d\") " pod="openshift-marketplace/redhat-operators-zfjb5" Oct 05 09:11:42 crc kubenswrapper[4846]: I1005 09:11:42.751000 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a1dbb719-cf21-487d-9605-0dc1b615046d-utilities\") pod \"redhat-operators-zfjb5\" (UID: \"a1dbb719-cf21-487d-9605-0dc1b615046d\") " pod="openshift-marketplace/redhat-operators-zfjb5" Oct 05 09:11:42 crc kubenswrapper[4846]: I1005 09:11:42.769980 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rfljh\" (UniqueName: \"kubernetes.io/projected/a1dbb719-cf21-487d-9605-0dc1b615046d-kube-api-access-rfljh\") pod \"redhat-operators-zfjb5\" (UID: \"a1dbb719-cf21-487d-9605-0dc1b615046d\") " pod="openshift-marketplace/redhat-operators-zfjb5" Oct 05 09:11:42 crc kubenswrapper[4846]: I1005 09:11:42.815831 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-zfjb5" Oct 05 09:11:43 crc kubenswrapper[4846]: I1005 09:11:43.287815 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-zfjb5"] Oct 05 09:11:43 crc kubenswrapper[4846]: I1005 09:11:43.474338 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zfjb5" event={"ID":"a1dbb719-cf21-487d-9605-0dc1b615046d","Type":"ContainerStarted","Data":"8b49b1acd5fb4ee2ab3ea0156408a3bcf054bf13b219afa66e84e720273919fe"} Oct 05 09:11:44 crc kubenswrapper[4846]: I1005 09:11:44.492511 4846 generic.go:334] "Generic (PLEG): container finished" podID="a1dbb719-cf21-487d-9605-0dc1b615046d" containerID="66edc6152d673d134a40607c9a04a57c45b8dbb3c675e0ed3a70082ca1746cf0" exitCode=0 Oct 05 09:11:44 crc kubenswrapper[4846]: I1005 09:11:44.492630 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zfjb5" event={"ID":"a1dbb719-cf21-487d-9605-0dc1b615046d","Type":"ContainerDied","Data":"66edc6152d673d134a40607c9a04a57c45b8dbb3c675e0ed3a70082ca1746cf0"} Oct 05 09:11:45 crc kubenswrapper[4846]: I1005 09:11:45.508644 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zfjb5" event={"ID":"a1dbb719-cf21-487d-9605-0dc1b615046d","Type":"ContainerStarted","Data":"dda56d45828d6dce9e0f4fd891c56df32c214397f40f33d0483a1a12f46da4c0"} Oct 05 09:11:49 crc kubenswrapper[4846]: I1005 09:11:49.559776 4846 generic.go:334] "Generic (PLEG): container finished" podID="a1dbb719-cf21-487d-9605-0dc1b615046d" containerID="dda56d45828d6dce9e0f4fd891c56df32c214397f40f33d0483a1a12f46da4c0" exitCode=0 Oct 05 09:11:49 crc kubenswrapper[4846]: I1005 09:11:49.559846 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zfjb5" event={"ID":"a1dbb719-cf21-487d-9605-0dc1b615046d","Type":"ContainerDied","Data":"dda56d45828d6dce9e0f4fd891c56df32c214397f40f33d0483a1a12f46da4c0"} Oct 05 09:11:50 crc kubenswrapper[4846]: I1005 09:11:50.497538 4846 scope.go:117] "RemoveContainer" containerID="e39593bcdffa488d1cfc3d5c184d6e46655bb5c018d40df0f20d590de5498b3d" Oct 05 09:11:50 crc kubenswrapper[4846]: E1005 09:11:50.498344 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 09:11:50 crc kubenswrapper[4846]: I1005 09:11:50.575092 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zfjb5" event={"ID":"a1dbb719-cf21-487d-9605-0dc1b615046d","Type":"ContainerStarted","Data":"2a3f4982e534a1a834edbce86aa9bf2ec7b38d69f037f1ecf5f9a1c4e37ff9d8"} Oct 05 09:11:52 crc kubenswrapper[4846]: I1005 09:11:52.816049 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-zfjb5" Oct 05 09:11:52 crc kubenswrapper[4846]: I1005 09:11:52.816435 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-zfjb5" Oct 05 09:11:53 crc kubenswrapper[4846]: I1005 09:11:53.883381 4846 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-zfjb5" podUID="a1dbb719-cf21-487d-9605-0dc1b615046d" containerName="registry-server" probeResult="failure" output=< Oct 05 09:11:53 crc kubenswrapper[4846]: timeout: failed to connect service ":50051" within 1s Oct 05 09:11:53 crc kubenswrapper[4846]: > Oct 05 09:12:01 crc kubenswrapper[4846]: I1005 09:12:01.498041 4846 scope.go:117] "RemoveContainer" containerID="e39593bcdffa488d1cfc3d5c184d6e46655bb5c018d40df0f20d590de5498b3d" Oct 05 09:12:01 crc kubenswrapper[4846]: E1005 09:12:01.498892 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 09:12:03 crc kubenswrapper[4846]: I1005 09:12:03.868028 4846 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-zfjb5" podUID="a1dbb719-cf21-487d-9605-0dc1b615046d" containerName="registry-server" probeResult="failure" output=< Oct 05 09:12:03 crc kubenswrapper[4846]: timeout: failed to connect service ":50051" within 1s Oct 05 09:12:03 crc kubenswrapper[4846]: > Oct 05 09:12:06 crc kubenswrapper[4846]: I1005 09:12:06.764124 4846 generic.go:334] "Generic (PLEG): container finished" podID="e844e15c-7dd3-4056-8727-a24c9b65dbd8" containerID="62450a6f22dcdd130ddefa63260041965f680f613c71ffe53653bc4acfffd33e" exitCode=0 Oct 05 09:12:06 crc kubenswrapper[4846]: I1005 09:12:06.764215 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-sriov-openstack-openstack-cell1-ds9h4" event={"ID":"e844e15c-7dd3-4056-8727-a24c9b65dbd8","Type":"ContainerDied","Data":"62450a6f22dcdd130ddefa63260041965f680f613c71ffe53653bc4acfffd33e"} Oct 05 09:12:06 crc kubenswrapper[4846]: I1005 09:12:06.789352 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-zfjb5" podStartSLOduration=19.28638172 podStartE2EDuration="24.789333743s" podCreationTimestamp="2025-10-05 09:11:42 +0000 UTC" firstStartedPulling="2025-10-05 09:11:44.495692952 +0000 UTC m=+8626.736545737" lastFinishedPulling="2025-10-05 09:11:49.998644945 +0000 UTC m=+8632.239497760" observedRunningTime="2025-10-05 09:11:50.594779092 +0000 UTC m=+8632.835631877" watchObservedRunningTime="2025-10-05 09:12:06.789333743 +0000 UTC m=+8649.030186528" Oct 05 09:12:08 crc kubenswrapper[4846]: I1005 09:12:08.185241 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-sriov-openstack-openstack-cell1-ds9h4" Oct 05 09:12:08 crc kubenswrapper[4846]: I1005 09:12:08.214301 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-sriov-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/e844e15c-7dd3-4056-8727-a24c9b65dbd8-neutron-sriov-agent-neutron-config-0\") pod \"e844e15c-7dd3-4056-8727-a24c9b65dbd8\" (UID: \"e844e15c-7dd3-4056-8727-a24c9b65dbd8\") " Oct 05 09:12:08 crc kubenswrapper[4846]: I1005 09:12:08.214352 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e844e15c-7dd3-4056-8727-a24c9b65dbd8-neutron-sriov-combined-ca-bundle\") pod \"e844e15c-7dd3-4056-8727-a24c9b65dbd8\" (UID: \"e844e15c-7dd3-4056-8727-a24c9b65dbd8\") " Oct 05 09:12:08 crc kubenswrapper[4846]: I1005 09:12:08.214447 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2hhbj\" (UniqueName: \"kubernetes.io/projected/e844e15c-7dd3-4056-8727-a24c9b65dbd8-kube-api-access-2hhbj\") pod \"e844e15c-7dd3-4056-8727-a24c9b65dbd8\" (UID: \"e844e15c-7dd3-4056-8727-a24c9b65dbd8\") " Oct 05 09:12:08 crc kubenswrapper[4846]: I1005 09:12:08.214514 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e844e15c-7dd3-4056-8727-a24c9b65dbd8-ssh-key\") pod \"e844e15c-7dd3-4056-8727-a24c9b65dbd8\" (UID: \"e844e15c-7dd3-4056-8727-a24c9b65dbd8\") " Oct 05 09:12:08 crc kubenswrapper[4846]: I1005 09:12:08.214708 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e844e15c-7dd3-4056-8727-a24c9b65dbd8-inventory\") pod \"e844e15c-7dd3-4056-8727-a24c9b65dbd8\" (UID: \"e844e15c-7dd3-4056-8727-a24c9b65dbd8\") " Oct 05 09:12:08 crc kubenswrapper[4846]: I1005 09:12:08.222534 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e844e15c-7dd3-4056-8727-a24c9b65dbd8-kube-api-access-2hhbj" (OuterVolumeSpecName: "kube-api-access-2hhbj") pod "e844e15c-7dd3-4056-8727-a24c9b65dbd8" (UID: "e844e15c-7dd3-4056-8727-a24c9b65dbd8"). InnerVolumeSpecName "kube-api-access-2hhbj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 09:12:08 crc kubenswrapper[4846]: I1005 09:12:08.224489 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e844e15c-7dd3-4056-8727-a24c9b65dbd8-neutron-sriov-combined-ca-bundle" (OuterVolumeSpecName: "neutron-sriov-combined-ca-bundle") pod "e844e15c-7dd3-4056-8727-a24c9b65dbd8" (UID: "e844e15c-7dd3-4056-8727-a24c9b65dbd8"). InnerVolumeSpecName "neutron-sriov-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 09:12:08 crc kubenswrapper[4846]: I1005 09:12:08.249560 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e844e15c-7dd3-4056-8727-a24c9b65dbd8-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "e844e15c-7dd3-4056-8727-a24c9b65dbd8" (UID: "e844e15c-7dd3-4056-8727-a24c9b65dbd8"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 09:12:08 crc kubenswrapper[4846]: I1005 09:12:08.284071 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e844e15c-7dd3-4056-8727-a24c9b65dbd8-neutron-sriov-agent-neutron-config-0" (OuterVolumeSpecName: "neutron-sriov-agent-neutron-config-0") pod "e844e15c-7dd3-4056-8727-a24c9b65dbd8" (UID: "e844e15c-7dd3-4056-8727-a24c9b65dbd8"). InnerVolumeSpecName "neutron-sriov-agent-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 09:12:08 crc kubenswrapper[4846]: I1005 09:12:08.284674 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e844e15c-7dd3-4056-8727-a24c9b65dbd8-inventory" (OuterVolumeSpecName: "inventory") pod "e844e15c-7dd3-4056-8727-a24c9b65dbd8" (UID: "e844e15c-7dd3-4056-8727-a24c9b65dbd8"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 09:12:08 crc kubenswrapper[4846]: I1005 09:12:08.317057 4846 reconciler_common.go:293] "Volume detached for volume \"neutron-sriov-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/e844e15c-7dd3-4056-8727-a24c9b65dbd8-neutron-sriov-agent-neutron-config-0\") on node \"crc\" DevicePath \"\"" Oct 05 09:12:08 crc kubenswrapper[4846]: I1005 09:12:08.317097 4846 reconciler_common.go:293] "Volume detached for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e844e15c-7dd3-4056-8727-a24c9b65dbd8-neutron-sriov-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 09:12:08 crc kubenswrapper[4846]: I1005 09:12:08.317114 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2hhbj\" (UniqueName: \"kubernetes.io/projected/e844e15c-7dd3-4056-8727-a24c9b65dbd8-kube-api-access-2hhbj\") on node \"crc\" DevicePath \"\"" Oct 05 09:12:08 crc kubenswrapper[4846]: I1005 09:12:08.317127 4846 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e844e15c-7dd3-4056-8727-a24c9b65dbd8-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 05 09:12:08 crc kubenswrapper[4846]: I1005 09:12:08.317138 4846 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e844e15c-7dd3-4056-8727-a24c9b65dbd8-inventory\") on node \"crc\" DevicePath \"\"" Oct 05 09:12:08 crc kubenswrapper[4846]: I1005 09:12:08.790814 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-sriov-openstack-openstack-cell1-ds9h4" event={"ID":"e844e15c-7dd3-4056-8727-a24c9b65dbd8","Type":"ContainerDied","Data":"241c0717a47a60438d90773eac612192320e1ba8dce1f1e4b93dbfe501c6d0e1"} Oct 05 09:12:08 crc kubenswrapper[4846]: I1005 09:12:08.790890 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-sriov-openstack-openstack-cell1-ds9h4" Oct 05 09:12:08 crc kubenswrapper[4846]: I1005 09:12:08.790895 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="241c0717a47a60438d90773eac612192320e1ba8dce1f1e4b93dbfe501c6d0e1" Oct 05 09:12:08 crc kubenswrapper[4846]: I1005 09:12:08.907559 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-dhcp-openstack-openstack-cell1-hl2xj"] Oct 05 09:12:08 crc kubenswrapper[4846]: E1005 09:12:08.907997 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e844e15c-7dd3-4056-8727-a24c9b65dbd8" containerName="neutron-sriov-openstack-openstack-cell1" Oct 05 09:12:08 crc kubenswrapper[4846]: I1005 09:12:08.908015 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="e844e15c-7dd3-4056-8727-a24c9b65dbd8" containerName="neutron-sriov-openstack-openstack-cell1" Oct 05 09:12:08 crc kubenswrapper[4846]: I1005 09:12:08.908223 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="e844e15c-7dd3-4056-8727-a24c9b65dbd8" containerName="neutron-sriov-openstack-openstack-cell1" Oct 05 09:12:08 crc kubenswrapper[4846]: I1005 09:12:08.908952 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-dhcp-openstack-openstack-cell1-hl2xj" Oct 05 09:12:08 crc kubenswrapper[4846]: I1005 09:12:08.912706 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-dhcp-agent-neutron-config" Oct 05 09:12:08 crc kubenswrapper[4846]: I1005 09:12:08.913256 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Oct 05 09:12:08 crc kubenswrapper[4846]: I1005 09:12:08.913448 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-7687s" Oct 05 09:12:08 crc kubenswrapper[4846]: I1005 09:12:08.913824 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 05 09:12:08 crc kubenswrapper[4846]: I1005 09:12:08.914092 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Oct 05 09:12:08 crc kubenswrapper[4846]: I1005 09:12:08.934606 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-dhcp-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/3d69d7c2-0c3a-4850-9ad0-e9654b3e8166-neutron-dhcp-agent-neutron-config-0\") pod \"neutron-dhcp-openstack-openstack-cell1-hl2xj\" (UID: \"3d69d7c2-0c3a-4850-9ad0-e9654b3e8166\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-hl2xj" Oct 05 09:12:08 crc kubenswrapper[4846]: I1005 09:12:08.934784 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3d69d7c2-0c3a-4850-9ad0-e9654b3e8166-inventory\") pod \"neutron-dhcp-openstack-openstack-cell1-hl2xj\" (UID: \"3d69d7c2-0c3a-4850-9ad0-e9654b3e8166\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-hl2xj" Oct 05 09:12:08 crc kubenswrapper[4846]: I1005 09:12:08.934971 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3d69d7c2-0c3a-4850-9ad0-e9654b3e8166-ssh-key\") pod \"neutron-dhcp-openstack-openstack-cell1-hl2xj\" (UID: \"3d69d7c2-0c3a-4850-9ad0-e9654b3e8166\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-hl2xj" Oct 05 09:12:08 crc kubenswrapper[4846]: I1005 09:12:08.935218 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3d69d7c2-0c3a-4850-9ad0-e9654b3e8166-neutron-dhcp-combined-ca-bundle\") pod \"neutron-dhcp-openstack-openstack-cell1-hl2xj\" (UID: \"3d69d7c2-0c3a-4850-9ad0-e9654b3e8166\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-hl2xj" Oct 05 09:12:08 crc kubenswrapper[4846]: I1005 09:12:08.935356 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f8gxp\" (UniqueName: \"kubernetes.io/projected/3d69d7c2-0c3a-4850-9ad0-e9654b3e8166-kube-api-access-f8gxp\") pod \"neutron-dhcp-openstack-openstack-cell1-hl2xj\" (UID: \"3d69d7c2-0c3a-4850-9ad0-e9654b3e8166\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-hl2xj" Oct 05 09:12:08 crc kubenswrapper[4846]: I1005 09:12:08.936499 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-dhcp-openstack-openstack-cell1-hl2xj"] Oct 05 09:12:09 crc kubenswrapper[4846]: I1005 09:12:09.037487 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-dhcp-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/3d69d7c2-0c3a-4850-9ad0-e9654b3e8166-neutron-dhcp-agent-neutron-config-0\") pod \"neutron-dhcp-openstack-openstack-cell1-hl2xj\" (UID: \"3d69d7c2-0c3a-4850-9ad0-e9654b3e8166\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-hl2xj" Oct 05 09:12:09 crc kubenswrapper[4846]: I1005 09:12:09.037634 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3d69d7c2-0c3a-4850-9ad0-e9654b3e8166-inventory\") pod \"neutron-dhcp-openstack-openstack-cell1-hl2xj\" (UID: \"3d69d7c2-0c3a-4850-9ad0-e9654b3e8166\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-hl2xj" Oct 05 09:12:09 crc kubenswrapper[4846]: I1005 09:12:09.037803 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3d69d7c2-0c3a-4850-9ad0-e9654b3e8166-ssh-key\") pod \"neutron-dhcp-openstack-openstack-cell1-hl2xj\" (UID: \"3d69d7c2-0c3a-4850-9ad0-e9654b3e8166\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-hl2xj" Oct 05 09:12:09 crc kubenswrapper[4846]: I1005 09:12:09.037845 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3d69d7c2-0c3a-4850-9ad0-e9654b3e8166-neutron-dhcp-combined-ca-bundle\") pod \"neutron-dhcp-openstack-openstack-cell1-hl2xj\" (UID: \"3d69d7c2-0c3a-4850-9ad0-e9654b3e8166\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-hl2xj" Oct 05 09:12:09 crc kubenswrapper[4846]: I1005 09:12:09.037885 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f8gxp\" (UniqueName: \"kubernetes.io/projected/3d69d7c2-0c3a-4850-9ad0-e9654b3e8166-kube-api-access-f8gxp\") pod \"neutron-dhcp-openstack-openstack-cell1-hl2xj\" (UID: \"3d69d7c2-0c3a-4850-9ad0-e9654b3e8166\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-hl2xj" Oct 05 09:12:09 crc kubenswrapper[4846]: I1005 09:12:09.044386 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3d69d7c2-0c3a-4850-9ad0-e9654b3e8166-ssh-key\") pod \"neutron-dhcp-openstack-openstack-cell1-hl2xj\" (UID: \"3d69d7c2-0c3a-4850-9ad0-e9654b3e8166\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-hl2xj" Oct 05 09:12:09 crc kubenswrapper[4846]: I1005 09:12:09.044615 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3d69d7c2-0c3a-4850-9ad0-e9654b3e8166-inventory\") pod \"neutron-dhcp-openstack-openstack-cell1-hl2xj\" (UID: \"3d69d7c2-0c3a-4850-9ad0-e9654b3e8166\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-hl2xj" Oct 05 09:12:09 crc kubenswrapper[4846]: I1005 09:12:09.044622 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3d69d7c2-0c3a-4850-9ad0-e9654b3e8166-neutron-dhcp-combined-ca-bundle\") pod \"neutron-dhcp-openstack-openstack-cell1-hl2xj\" (UID: \"3d69d7c2-0c3a-4850-9ad0-e9654b3e8166\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-hl2xj" Oct 05 09:12:09 crc kubenswrapper[4846]: I1005 09:12:09.045045 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-dhcp-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/3d69d7c2-0c3a-4850-9ad0-e9654b3e8166-neutron-dhcp-agent-neutron-config-0\") pod \"neutron-dhcp-openstack-openstack-cell1-hl2xj\" (UID: \"3d69d7c2-0c3a-4850-9ad0-e9654b3e8166\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-hl2xj" Oct 05 09:12:09 crc kubenswrapper[4846]: I1005 09:12:09.056724 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f8gxp\" (UniqueName: \"kubernetes.io/projected/3d69d7c2-0c3a-4850-9ad0-e9654b3e8166-kube-api-access-f8gxp\") pod \"neutron-dhcp-openstack-openstack-cell1-hl2xj\" (UID: \"3d69d7c2-0c3a-4850-9ad0-e9654b3e8166\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-hl2xj" Oct 05 09:12:09 crc kubenswrapper[4846]: I1005 09:12:09.228113 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-dhcp-openstack-openstack-cell1-hl2xj" Oct 05 09:12:09 crc kubenswrapper[4846]: I1005 09:12:09.900906 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-dhcp-openstack-openstack-cell1-hl2xj"] Oct 05 09:12:10 crc kubenswrapper[4846]: I1005 09:12:10.834720 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-dhcp-openstack-openstack-cell1-hl2xj" event={"ID":"3d69d7c2-0c3a-4850-9ad0-e9654b3e8166","Type":"ContainerStarted","Data":"17041d0d7b8e960e1186bb520c1c4b6d2c7c2fbadf9bb15ab8b35acf03ea5352"} Oct 05 09:12:10 crc kubenswrapper[4846]: I1005 09:12:10.835254 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-dhcp-openstack-openstack-cell1-hl2xj" event={"ID":"3d69d7c2-0c3a-4850-9ad0-e9654b3e8166","Type":"ContainerStarted","Data":"29362ad5647e0db3feeb4f97c77cd492f07928f66b498a8a702c40a47ec97907"} Oct 05 09:12:10 crc kubenswrapper[4846]: I1005 09:12:10.884421 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-dhcp-openstack-openstack-cell1-hl2xj" podStartSLOduration=2.469346631 podStartE2EDuration="2.884390916s" podCreationTimestamp="2025-10-05 09:12:08 +0000 UTC" firstStartedPulling="2025-10-05 09:12:09.899298218 +0000 UTC m=+8652.140151033" lastFinishedPulling="2025-10-05 09:12:10.314342513 +0000 UTC m=+8652.555195318" observedRunningTime="2025-10-05 09:12:10.864152659 +0000 UTC m=+8653.105005474" watchObservedRunningTime="2025-10-05 09:12:10.884390916 +0000 UTC m=+8653.125243711" Oct 05 09:12:12 crc kubenswrapper[4846]: I1005 09:12:12.913369 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-zfjb5" Oct 05 09:12:12 crc kubenswrapper[4846]: I1005 09:12:12.999388 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-zfjb5" Oct 05 09:12:13 crc kubenswrapper[4846]: I1005 09:12:13.667245 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-zfjb5"] Oct 05 09:12:14 crc kubenswrapper[4846]: I1005 09:12:14.882620 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-zfjb5" podUID="a1dbb719-cf21-487d-9605-0dc1b615046d" containerName="registry-server" containerID="cri-o://2a3f4982e534a1a834edbce86aa9bf2ec7b38d69f037f1ecf5f9a1c4e37ff9d8" gracePeriod=2 Oct 05 09:12:15 crc kubenswrapper[4846]: I1005 09:12:15.442076 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-zfjb5" Oct 05 09:12:15 crc kubenswrapper[4846]: I1005 09:12:15.497337 4846 scope.go:117] "RemoveContainer" containerID="e39593bcdffa488d1cfc3d5c184d6e46655bb5c018d40df0f20d590de5498b3d" Oct 05 09:12:15 crc kubenswrapper[4846]: I1005 09:12:15.497415 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a1dbb719-cf21-487d-9605-0dc1b615046d-catalog-content\") pod \"a1dbb719-cf21-487d-9605-0dc1b615046d\" (UID: \"a1dbb719-cf21-487d-9605-0dc1b615046d\") " Oct 05 09:12:15 crc kubenswrapper[4846]: I1005 09:12:15.497475 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rfljh\" (UniqueName: \"kubernetes.io/projected/a1dbb719-cf21-487d-9605-0dc1b615046d-kube-api-access-rfljh\") pod \"a1dbb719-cf21-487d-9605-0dc1b615046d\" (UID: \"a1dbb719-cf21-487d-9605-0dc1b615046d\") " Oct 05 09:12:15 crc kubenswrapper[4846]: I1005 09:12:15.497640 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a1dbb719-cf21-487d-9605-0dc1b615046d-utilities\") pod \"a1dbb719-cf21-487d-9605-0dc1b615046d\" (UID: \"a1dbb719-cf21-487d-9605-0dc1b615046d\") " Oct 05 09:12:15 crc kubenswrapper[4846]: E1005 09:12:15.497661 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 09:12:15 crc kubenswrapper[4846]: I1005 09:12:15.498271 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a1dbb719-cf21-487d-9605-0dc1b615046d-utilities" (OuterVolumeSpecName: "utilities") pod "a1dbb719-cf21-487d-9605-0dc1b615046d" (UID: "a1dbb719-cf21-487d-9605-0dc1b615046d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 09:12:15 crc kubenswrapper[4846]: I1005 09:12:15.509484 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a1dbb719-cf21-487d-9605-0dc1b615046d-kube-api-access-rfljh" (OuterVolumeSpecName: "kube-api-access-rfljh") pod "a1dbb719-cf21-487d-9605-0dc1b615046d" (UID: "a1dbb719-cf21-487d-9605-0dc1b615046d"). InnerVolumeSpecName "kube-api-access-rfljh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 09:12:15 crc kubenswrapper[4846]: I1005 09:12:15.566364 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a1dbb719-cf21-487d-9605-0dc1b615046d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a1dbb719-cf21-487d-9605-0dc1b615046d" (UID: "a1dbb719-cf21-487d-9605-0dc1b615046d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 09:12:15 crc kubenswrapper[4846]: I1005 09:12:15.600667 4846 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a1dbb719-cf21-487d-9605-0dc1b615046d-utilities\") on node \"crc\" DevicePath \"\"" Oct 05 09:12:15 crc kubenswrapper[4846]: I1005 09:12:15.600703 4846 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a1dbb719-cf21-487d-9605-0dc1b615046d-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 05 09:12:15 crc kubenswrapper[4846]: I1005 09:12:15.600718 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rfljh\" (UniqueName: \"kubernetes.io/projected/a1dbb719-cf21-487d-9605-0dc1b615046d-kube-api-access-rfljh\") on node \"crc\" DevicePath \"\"" Oct 05 09:12:15 crc kubenswrapper[4846]: I1005 09:12:15.895073 4846 generic.go:334] "Generic (PLEG): container finished" podID="a1dbb719-cf21-487d-9605-0dc1b615046d" containerID="2a3f4982e534a1a834edbce86aa9bf2ec7b38d69f037f1ecf5f9a1c4e37ff9d8" exitCode=0 Oct 05 09:12:15 crc kubenswrapper[4846]: I1005 09:12:15.895166 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zfjb5" event={"ID":"a1dbb719-cf21-487d-9605-0dc1b615046d","Type":"ContainerDied","Data":"2a3f4982e534a1a834edbce86aa9bf2ec7b38d69f037f1ecf5f9a1c4e37ff9d8"} Oct 05 09:12:15 crc kubenswrapper[4846]: I1005 09:12:15.895235 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zfjb5" event={"ID":"a1dbb719-cf21-487d-9605-0dc1b615046d","Type":"ContainerDied","Data":"8b49b1acd5fb4ee2ab3ea0156408a3bcf054bf13b219afa66e84e720273919fe"} Oct 05 09:12:15 crc kubenswrapper[4846]: I1005 09:12:15.895265 4846 scope.go:117] "RemoveContainer" containerID="2a3f4982e534a1a834edbce86aa9bf2ec7b38d69f037f1ecf5f9a1c4e37ff9d8" Oct 05 09:12:15 crc kubenswrapper[4846]: I1005 09:12:15.895173 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-zfjb5" Oct 05 09:12:15 crc kubenswrapper[4846]: I1005 09:12:15.921264 4846 scope.go:117] "RemoveContainer" containerID="dda56d45828d6dce9e0f4fd891c56df32c214397f40f33d0483a1a12f46da4c0" Oct 05 09:12:15 crc kubenswrapper[4846]: I1005 09:12:15.944145 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-zfjb5"] Oct 05 09:12:15 crc kubenswrapper[4846]: I1005 09:12:15.955546 4846 scope.go:117] "RemoveContainer" containerID="66edc6152d673d134a40607c9a04a57c45b8dbb3c675e0ed3a70082ca1746cf0" Oct 05 09:12:15 crc kubenswrapper[4846]: I1005 09:12:15.960521 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-zfjb5"] Oct 05 09:12:16 crc kubenswrapper[4846]: I1005 09:12:16.026080 4846 scope.go:117] "RemoveContainer" containerID="2a3f4982e534a1a834edbce86aa9bf2ec7b38d69f037f1ecf5f9a1c4e37ff9d8" Oct 05 09:12:16 crc kubenswrapper[4846]: E1005 09:12:16.026729 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2a3f4982e534a1a834edbce86aa9bf2ec7b38d69f037f1ecf5f9a1c4e37ff9d8\": container with ID starting with 2a3f4982e534a1a834edbce86aa9bf2ec7b38d69f037f1ecf5f9a1c4e37ff9d8 not found: ID does not exist" containerID="2a3f4982e534a1a834edbce86aa9bf2ec7b38d69f037f1ecf5f9a1c4e37ff9d8" Oct 05 09:12:16 crc kubenswrapper[4846]: I1005 09:12:16.026798 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2a3f4982e534a1a834edbce86aa9bf2ec7b38d69f037f1ecf5f9a1c4e37ff9d8"} err="failed to get container status \"2a3f4982e534a1a834edbce86aa9bf2ec7b38d69f037f1ecf5f9a1c4e37ff9d8\": rpc error: code = NotFound desc = could not find container \"2a3f4982e534a1a834edbce86aa9bf2ec7b38d69f037f1ecf5f9a1c4e37ff9d8\": container with ID starting with 2a3f4982e534a1a834edbce86aa9bf2ec7b38d69f037f1ecf5f9a1c4e37ff9d8 not found: ID does not exist" Oct 05 09:12:16 crc kubenswrapper[4846]: I1005 09:12:16.026842 4846 scope.go:117] "RemoveContainer" containerID="dda56d45828d6dce9e0f4fd891c56df32c214397f40f33d0483a1a12f46da4c0" Oct 05 09:12:16 crc kubenswrapper[4846]: E1005 09:12:16.027375 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dda56d45828d6dce9e0f4fd891c56df32c214397f40f33d0483a1a12f46da4c0\": container with ID starting with dda56d45828d6dce9e0f4fd891c56df32c214397f40f33d0483a1a12f46da4c0 not found: ID does not exist" containerID="dda56d45828d6dce9e0f4fd891c56df32c214397f40f33d0483a1a12f46da4c0" Oct 05 09:12:16 crc kubenswrapper[4846]: I1005 09:12:16.027425 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dda56d45828d6dce9e0f4fd891c56df32c214397f40f33d0483a1a12f46da4c0"} err="failed to get container status \"dda56d45828d6dce9e0f4fd891c56df32c214397f40f33d0483a1a12f46da4c0\": rpc error: code = NotFound desc = could not find container \"dda56d45828d6dce9e0f4fd891c56df32c214397f40f33d0483a1a12f46da4c0\": container with ID starting with dda56d45828d6dce9e0f4fd891c56df32c214397f40f33d0483a1a12f46da4c0 not found: ID does not exist" Oct 05 09:12:16 crc kubenswrapper[4846]: I1005 09:12:16.027456 4846 scope.go:117] "RemoveContainer" containerID="66edc6152d673d134a40607c9a04a57c45b8dbb3c675e0ed3a70082ca1746cf0" Oct 05 09:12:16 crc kubenswrapper[4846]: E1005 09:12:16.027811 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"66edc6152d673d134a40607c9a04a57c45b8dbb3c675e0ed3a70082ca1746cf0\": container with ID starting with 66edc6152d673d134a40607c9a04a57c45b8dbb3c675e0ed3a70082ca1746cf0 not found: ID does not exist" containerID="66edc6152d673d134a40607c9a04a57c45b8dbb3c675e0ed3a70082ca1746cf0" Oct 05 09:12:16 crc kubenswrapper[4846]: I1005 09:12:16.027854 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"66edc6152d673d134a40607c9a04a57c45b8dbb3c675e0ed3a70082ca1746cf0"} err="failed to get container status \"66edc6152d673d134a40607c9a04a57c45b8dbb3c675e0ed3a70082ca1746cf0\": rpc error: code = NotFound desc = could not find container \"66edc6152d673d134a40607c9a04a57c45b8dbb3c675e0ed3a70082ca1746cf0\": container with ID starting with 66edc6152d673d134a40607c9a04a57c45b8dbb3c675e0ed3a70082ca1746cf0 not found: ID does not exist" Oct 05 09:12:16 crc kubenswrapper[4846]: I1005 09:12:16.271081 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-vprnc"] Oct 05 09:12:16 crc kubenswrapper[4846]: E1005 09:12:16.271659 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a1dbb719-cf21-487d-9605-0dc1b615046d" containerName="extract-content" Oct 05 09:12:16 crc kubenswrapper[4846]: I1005 09:12:16.271684 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="a1dbb719-cf21-487d-9605-0dc1b615046d" containerName="extract-content" Oct 05 09:12:16 crc kubenswrapper[4846]: E1005 09:12:16.271705 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a1dbb719-cf21-487d-9605-0dc1b615046d" containerName="registry-server" Oct 05 09:12:16 crc kubenswrapper[4846]: I1005 09:12:16.271714 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="a1dbb719-cf21-487d-9605-0dc1b615046d" containerName="registry-server" Oct 05 09:12:16 crc kubenswrapper[4846]: E1005 09:12:16.271734 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a1dbb719-cf21-487d-9605-0dc1b615046d" containerName="extract-utilities" Oct 05 09:12:16 crc kubenswrapper[4846]: I1005 09:12:16.271742 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="a1dbb719-cf21-487d-9605-0dc1b615046d" containerName="extract-utilities" Oct 05 09:12:16 crc kubenswrapper[4846]: I1005 09:12:16.272001 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="a1dbb719-cf21-487d-9605-0dc1b615046d" containerName="registry-server" Oct 05 09:12:16 crc kubenswrapper[4846]: I1005 09:12:16.273911 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-vprnc" Oct 05 09:12:16 crc kubenswrapper[4846]: I1005 09:12:16.317854 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-vprnc"] Oct 05 09:12:16 crc kubenswrapper[4846]: I1005 09:12:16.318040 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rbgvv\" (UniqueName: \"kubernetes.io/projected/51ec6e14-e3cc-40dc-9033-ac38bf2f74b3-kube-api-access-rbgvv\") pod \"certified-operators-vprnc\" (UID: \"51ec6e14-e3cc-40dc-9033-ac38bf2f74b3\") " pod="openshift-marketplace/certified-operators-vprnc" Oct 05 09:12:16 crc kubenswrapper[4846]: I1005 09:12:16.318282 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/51ec6e14-e3cc-40dc-9033-ac38bf2f74b3-utilities\") pod \"certified-operators-vprnc\" (UID: \"51ec6e14-e3cc-40dc-9033-ac38bf2f74b3\") " pod="openshift-marketplace/certified-operators-vprnc" Oct 05 09:12:16 crc kubenswrapper[4846]: I1005 09:12:16.318406 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/51ec6e14-e3cc-40dc-9033-ac38bf2f74b3-catalog-content\") pod \"certified-operators-vprnc\" (UID: \"51ec6e14-e3cc-40dc-9033-ac38bf2f74b3\") " pod="openshift-marketplace/certified-operators-vprnc" Oct 05 09:12:16 crc kubenswrapper[4846]: I1005 09:12:16.422302 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rbgvv\" (UniqueName: \"kubernetes.io/projected/51ec6e14-e3cc-40dc-9033-ac38bf2f74b3-kube-api-access-rbgvv\") pod \"certified-operators-vprnc\" (UID: \"51ec6e14-e3cc-40dc-9033-ac38bf2f74b3\") " pod="openshift-marketplace/certified-operators-vprnc" Oct 05 09:12:16 crc kubenswrapper[4846]: I1005 09:12:16.422432 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/51ec6e14-e3cc-40dc-9033-ac38bf2f74b3-utilities\") pod \"certified-operators-vprnc\" (UID: \"51ec6e14-e3cc-40dc-9033-ac38bf2f74b3\") " pod="openshift-marketplace/certified-operators-vprnc" Oct 05 09:12:16 crc kubenswrapper[4846]: I1005 09:12:16.422488 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/51ec6e14-e3cc-40dc-9033-ac38bf2f74b3-catalog-content\") pod \"certified-operators-vprnc\" (UID: \"51ec6e14-e3cc-40dc-9033-ac38bf2f74b3\") " pod="openshift-marketplace/certified-operators-vprnc" Oct 05 09:12:16 crc kubenswrapper[4846]: I1005 09:12:16.423149 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/51ec6e14-e3cc-40dc-9033-ac38bf2f74b3-utilities\") pod \"certified-operators-vprnc\" (UID: \"51ec6e14-e3cc-40dc-9033-ac38bf2f74b3\") " pod="openshift-marketplace/certified-operators-vprnc" Oct 05 09:12:16 crc kubenswrapper[4846]: I1005 09:12:16.423997 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/51ec6e14-e3cc-40dc-9033-ac38bf2f74b3-catalog-content\") pod \"certified-operators-vprnc\" (UID: \"51ec6e14-e3cc-40dc-9033-ac38bf2f74b3\") " pod="openshift-marketplace/certified-operators-vprnc" Oct 05 09:12:16 crc kubenswrapper[4846]: I1005 09:12:16.445559 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rbgvv\" (UniqueName: \"kubernetes.io/projected/51ec6e14-e3cc-40dc-9033-ac38bf2f74b3-kube-api-access-rbgvv\") pod \"certified-operators-vprnc\" (UID: \"51ec6e14-e3cc-40dc-9033-ac38bf2f74b3\") " pod="openshift-marketplace/certified-operators-vprnc" Oct 05 09:12:16 crc kubenswrapper[4846]: I1005 09:12:16.512382 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a1dbb719-cf21-487d-9605-0dc1b615046d" path="/var/lib/kubelet/pods/a1dbb719-cf21-487d-9605-0dc1b615046d/volumes" Oct 05 09:12:16 crc kubenswrapper[4846]: I1005 09:12:16.617154 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-vprnc" Oct 05 09:12:17 crc kubenswrapper[4846]: I1005 09:12:17.204363 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-vprnc"] Oct 05 09:12:17 crc kubenswrapper[4846]: I1005 09:12:17.938083 4846 generic.go:334] "Generic (PLEG): container finished" podID="51ec6e14-e3cc-40dc-9033-ac38bf2f74b3" containerID="610bb3bc9297a6e2905343a1d286aaf29bf49d23ff8206062e2b4b0ae4092f27" exitCode=0 Oct 05 09:12:17 crc kubenswrapper[4846]: I1005 09:12:17.938213 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vprnc" event={"ID":"51ec6e14-e3cc-40dc-9033-ac38bf2f74b3","Type":"ContainerDied","Data":"610bb3bc9297a6e2905343a1d286aaf29bf49d23ff8206062e2b4b0ae4092f27"} Oct 05 09:12:17 crc kubenswrapper[4846]: I1005 09:12:17.938471 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vprnc" event={"ID":"51ec6e14-e3cc-40dc-9033-ac38bf2f74b3","Type":"ContainerStarted","Data":"c392d5c7052563ff37a115dce17f7959b944704c12ae97671be40c575d4f7f2f"} Oct 05 09:12:18 crc kubenswrapper[4846]: I1005 09:12:18.960742 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vprnc" event={"ID":"51ec6e14-e3cc-40dc-9033-ac38bf2f74b3","Type":"ContainerStarted","Data":"b8cea438c0f83ee3602db6928dc5e3b1535ad134c0c0a7e58a1d5330efc0ca1a"} Oct 05 09:12:20 crc kubenswrapper[4846]: I1005 09:12:20.984580 4846 generic.go:334] "Generic (PLEG): container finished" podID="51ec6e14-e3cc-40dc-9033-ac38bf2f74b3" containerID="b8cea438c0f83ee3602db6928dc5e3b1535ad134c0c0a7e58a1d5330efc0ca1a" exitCode=0 Oct 05 09:12:20 crc kubenswrapper[4846]: I1005 09:12:20.984641 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vprnc" event={"ID":"51ec6e14-e3cc-40dc-9033-ac38bf2f74b3","Type":"ContainerDied","Data":"b8cea438c0f83ee3602db6928dc5e3b1535ad134c0c0a7e58a1d5330efc0ca1a"} Oct 05 09:12:21 crc kubenswrapper[4846]: I1005 09:12:21.996991 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vprnc" event={"ID":"51ec6e14-e3cc-40dc-9033-ac38bf2f74b3","Type":"ContainerStarted","Data":"8aa526147ea1e8dfe1ec302b08dd8253fc6c9827ddc39aec47f5b806b11e95c8"} Oct 05 09:12:22 crc kubenswrapper[4846]: I1005 09:12:22.019734 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-vprnc" podStartSLOduration=2.498927824 podStartE2EDuration="6.019716622s" podCreationTimestamp="2025-10-05 09:12:16 +0000 UTC" firstStartedPulling="2025-10-05 09:12:17.94203875 +0000 UTC m=+8660.182891535" lastFinishedPulling="2025-10-05 09:12:21.462827548 +0000 UTC m=+8663.703680333" observedRunningTime="2025-10-05 09:12:22.014087132 +0000 UTC m=+8664.254939907" watchObservedRunningTime="2025-10-05 09:12:22.019716622 +0000 UTC m=+8664.260569397" Oct 05 09:12:26 crc kubenswrapper[4846]: I1005 09:12:26.617777 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-vprnc" Oct 05 09:12:26 crc kubenswrapper[4846]: I1005 09:12:26.618395 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-vprnc" Oct 05 09:12:26 crc kubenswrapper[4846]: I1005 09:12:26.689539 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-vprnc" Oct 05 09:12:27 crc kubenswrapper[4846]: I1005 09:12:27.130506 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-vprnc" Oct 05 09:12:27 crc kubenswrapper[4846]: I1005 09:12:27.183629 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-vprnc"] Oct 05 09:12:27 crc kubenswrapper[4846]: I1005 09:12:27.498075 4846 scope.go:117] "RemoveContainer" containerID="e39593bcdffa488d1cfc3d5c184d6e46655bb5c018d40df0f20d590de5498b3d" Oct 05 09:12:27 crc kubenswrapper[4846]: E1005 09:12:27.498329 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 09:12:29 crc kubenswrapper[4846]: I1005 09:12:29.083907 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-vprnc" podUID="51ec6e14-e3cc-40dc-9033-ac38bf2f74b3" containerName="registry-server" containerID="cri-o://8aa526147ea1e8dfe1ec302b08dd8253fc6c9827ddc39aec47f5b806b11e95c8" gracePeriod=2 Oct 05 09:12:30 crc kubenswrapper[4846]: I1005 09:12:30.096101 4846 generic.go:334] "Generic (PLEG): container finished" podID="51ec6e14-e3cc-40dc-9033-ac38bf2f74b3" containerID="8aa526147ea1e8dfe1ec302b08dd8253fc6c9827ddc39aec47f5b806b11e95c8" exitCode=0 Oct 05 09:12:30 crc kubenswrapper[4846]: I1005 09:12:30.096208 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vprnc" event={"ID":"51ec6e14-e3cc-40dc-9033-ac38bf2f74b3","Type":"ContainerDied","Data":"8aa526147ea1e8dfe1ec302b08dd8253fc6c9827ddc39aec47f5b806b11e95c8"} Oct 05 09:12:30 crc kubenswrapper[4846]: I1005 09:12:30.096525 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vprnc" event={"ID":"51ec6e14-e3cc-40dc-9033-ac38bf2f74b3","Type":"ContainerDied","Data":"c392d5c7052563ff37a115dce17f7959b944704c12ae97671be40c575d4f7f2f"} Oct 05 09:12:30 crc kubenswrapper[4846]: I1005 09:12:30.096554 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c392d5c7052563ff37a115dce17f7959b944704c12ae97671be40c575d4f7f2f" Oct 05 09:12:30 crc kubenswrapper[4846]: I1005 09:12:30.169792 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-vprnc" Oct 05 09:12:30 crc kubenswrapper[4846]: I1005 09:12:30.255908 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rbgvv\" (UniqueName: \"kubernetes.io/projected/51ec6e14-e3cc-40dc-9033-ac38bf2f74b3-kube-api-access-rbgvv\") pod \"51ec6e14-e3cc-40dc-9033-ac38bf2f74b3\" (UID: \"51ec6e14-e3cc-40dc-9033-ac38bf2f74b3\") " Oct 05 09:12:30 crc kubenswrapper[4846]: I1005 09:12:30.256031 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/51ec6e14-e3cc-40dc-9033-ac38bf2f74b3-utilities\") pod \"51ec6e14-e3cc-40dc-9033-ac38bf2f74b3\" (UID: \"51ec6e14-e3cc-40dc-9033-ac38bf2f74b3\") " Oct 05 09:12:30 crc kubenswrapper[4846]: I1005 09:12:30.256097 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/51ec6e14-e3cc-40dc-9033-ac38bf2f74b3-catalog-content\") pod \"51ec6e14-e3cc-40dc-9033-ac38bf2f74b3\" (UID: \"51ec6e14-e3cc-40dc-9033-ac38bf2f74b3\") " Oct 05 09:12:30 crc kubenswrapper[4846]: I1005 09:12:30.257096 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/51ec6e14-e3cc-40dc-9033-ac38bf2f74b3-utilities" (OuterVolumeSpecName: "utilities") pod "51ec6e14-e3cc-40dc-9033-ac38bf2f74b3" (UID: "51ec6e14-e3cc-40dc-9033-ac38bf2f74b3"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 09:12:30 crc kubenswrapper[4846]: I1005 09:12:30.265677 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/51ec6e14-e3cc-40dc-9033-ac38bf2f74b3-kube-api-access-rbgvv" (OuterVolumeSpecName: "kube-api-access-rbgvv") pod "51ec6e14-e3cc-40dc-9033-ac38bf2f74b3" (UID: "51ec6e14-e3cc-40dc-9033-ac38bf2f74b3"). InnerVolumeSpecName "kube-api-access-rbgvv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 09:12:30 crc kubenswrapper[4846]: I1005 09:12:30.300067 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/51ec6e14-e3cc-40dc-9033-ac38bf2f74b3-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "51ec6e14-e3cc-40dc-9033-ac38bf2f74b3" (UID: "51ec6e14-e3cc-40dc-9033-ac38bf2f74b3"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 09:12:30 crc kubenswrapper[4846]: I1005 09:12:30.358279 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rbgvv\" (UniqueName: \"kubernetes.io/projected/51ec6e14-e3cc-40dc-9033-ac38bf2f74b3-kube-api-access-rbgvv\") on node \"crc\" DevicePath \"\"" Oct 05 09:12:30 crc kubenswrapper[4846]: I1005 09:12:30.358317 4846 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/51ec6e14-e3cc-40dc-9033-ac38bf2f74b3-utilities\") on node \"crc\" DevicePath \"\"" Oct 05 09:12:30 crc kubenswrapper[4846]: I1005 09:12:30.358332 4846 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/51ec6e14-e3cc-40dc-9033-ac38bf2f74b3-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 05 09:12:31 crc kubenswrapper[4846]: I1005 09:12:31.106658 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-vprnc" Oct 05 09:12:31 crc kubenswrapper[4846]: I1005 09:12:31.136416 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-vprnc"] Oct 05 09:12:31 crc kubenswrapper[4846]: I1005 09:12:31.147020 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-vprnc"] Oct 05 09:12:32 crc kubenswrapper[4846]: I1005 09:12:32.512436 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="51ec6e14-e3cc-40dc-9033-ac38bf2f74b3" path="/var/lib/kubelet/pods/51ec6e14-e3cc-40dc-9033-ac38bf2f74b3/volumes" Oct 05 09:12:40 crc kubenswrapper[4846]: I1005 09:12:40.498788 4846 scope.go:117] "RemoveContainer" containerID="e39593bcdffa488d1cfc3d5c184d6e46655bb5c018d40df0f20d590de5498b3d" Oct 05 09:12:40 crc kubenswrapper[4846]: E1005 09:12:40.499759 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 09:12:55 crc kubenswrapper[4846]: I1005 09:12:55.498602 4846 scope.go:117] "RemoveContainer" containerID="e39593bcdffa488d1cfc3d5c184d6e46655bb5c018d40df0f20d590de5498b3d" Oct 05 09:12:56 crc kubenswrapper[4846]: I1005 09:12:56.488986 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" event={"ID":"ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7","Type":"ContainerStarted","Data":"206a9f8fee4cb0430eaa85104168af858cba45784f821a5ea04393d05e733060"} Oct 05 09:15:00 crc kubenswrapper[4846]: I1005 09:15:00.173136 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29327595-rcghk"] Oct 05 09:15:00 crc kubenswrapper[4846]: E1005 09:15:00.174019 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="51ec6e14-e3cc-40dc-9033-ac38bf2f74b3" containerName="extract-content" Oct 05 09:15:00 crc kubenswrapper[4846]: I1005 09:15:00.174032 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="51ec6e14-e3cc-40dc-9033-ac38bf2f74b3" containerName="extract-content" Oct 05 09:15:00 crc kubenswrapper[4846]: E1005 09:15:00.174087 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="51ec6e14-e3cc-40dc-9033-ac38bf2f74b3" containerName="registry-server" Oct 05 09:15:00 crc kubenswrapper[4846]: I1005 09:15:00.174094 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="51ec6e14-e3cc-40dc-9033-ac38bf2f74b3" containerName="registry-server" Oct 05 09:15:00 crc kubenswrapper[4846]: E1005 09:15:00.174110 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="51ec6e14-e3cc-40dc-9033-ac38bf2f74b3" containerName="extract-utilities" Oct 05 09:15:00 crc kubenswrapper[4846]: I1005 09:15:00.174115 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="51ec6e14-e3cc-40dc-9033-ac38bf2f74b3" containerName="extract-utilities" Oct 05 09:15:00 crc kubenswrapper[4846]: I1005 09:15:00.174371 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="51ec6e14-e3cc-40dc-9033-ac38bf2f74b3" containerName="registry-server" Oct 05 09:15:00 crc kubenswrapper[4846]: I1005 09:15:00.175127 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29327595-rcghk" Oct 05 09:15:00 crc kubenswrapper[4846]: I1005 09:15:00.182327 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 05 09:15:00 crc kubenswrapper[4846]: I1005 09:15:00.183009 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 05 09:15:00 crc kubenswrapper[4846]: I1005 09:15:00.192014 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29327595-rcghk"] Oct 05 09:15:00 crc kubenswrapper[4846]: I1005 09:15:00.283110 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/155d6821-91ed-4974-aab5-1ba06684c996-config-volume\") pod \"collect-profiles-29327595-rcghk\" (UID: \"155d6821-91ed-4974-aab5-1ba06684c996\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327595-rcghk" Oct 05 09:15:00 crc kubenswrapper[4846]: I1005 09:15:00.283256 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/155d6821-91ed-4974-aab5-1ba06684c996-secret-volume\") pod \"collect-profiles-29327595-rcghk\" (UID: \"155d6821-91ed-4974-aab5-1ba06684c996\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327595-rcghk" Oct 05 09:15:00 crc kubenswrapper[4846]: I1005 09:15:00.283384 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f4wqn\" (UniqueName: \"kubernetes.io/projected/155d6821-91ed-4974-aab5-1ba06684c996-kube-api-access-f4wqn\") pod \"collect-profiles-29327595-rcghk\" (UID: \"155d6821-91ed-4974-aab5-1ba06684c996\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327595-rcghk" Oct 05 09:15:00 crc kubenswrapper[4846]: I1005 09:15:00.385465 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f4wqn\" (UniqueName: \"kubernetes.io/projected/155d6821-91ed-4974-aab5-1ba06684c996-kube-api-access-f4wqn\") pod \"collect-profiles-29327595-rcghk\" (UID: \"155d6821-91ed-4974-aab5-1ba06684c996\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327595-rcghk" Oct 05 09:15:00 crc kubenswrapper[4846]: I1005 09:15:00.385639 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/155d6821-91ed-4974-aab5-1ba06684c996-config-volume\") pod \"collect-profiles-29327595-rcghk\" (UID: \"155d6821-91ed-4974-aab5-1ba06684c996\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327595-rcghk" Oct 05 09:15:00 crc kubenswrapper[4846]: I1005 09:15:00.385700 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/155d6821-91ed-4974-aab5-1ba06684c996-secret-volume\") pod \"collect-profiles-29327595-rcghk\" (UID: \"155d6821-91ed-4974-aab5-1ba06684c996\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327595-rcghk" Oct 05 09:15:00 crc kubenswrapper[4846]: I1005 09:15:00.387934 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/155d6821-91ed-4974-aab5-1ba06684c996-config-volume\") pod \"collect-profiles-29327595-rcghk\" (UID: \"155d6821-91ed-4974-aab5-1ba06684c996\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327595-rcghk" Oct 05 09:15:00 crc kubenswrapper[4846]: I1005 09:15:00.401559 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f4wqn\" (UniqueName: \"kubernetes.io/projected/155d6821-91ed-4974-aab5-1ba06684c996-kube-api-access-f4wqn\") pod \"collect-profiles-29327595-rcghk\" (UID: \"155d6821-91ed-4974-aab5-1ba06684c996\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327595-rcghk" Oct 05 09:15:00 crc kubenswrapper[4846]: I1005 09:15:00.402503 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/155d6821-91ed-4974-aab5-1ba06684c996-secret-volume\") pod \"collect-profiles-29327595-rcghk\" (UID: \"155d6821-91ed-4974-aab5-1ba06684c996\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327595-rcghk" Oct 05 09:15:00 crc kubenswrapper[4846]: I1005 09:15:00.499069 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29327595-rcghk" Oct 05 09:15:00 crc kubenswrapper[4846]: I1005 09:15:00.973809 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29327595-rcghk"] Oct 05 09:15:01 crc kubenswrapper[4846]: I1005 09:15:01.071255 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29327595-rcghk" event={"ID":"155d6821-91ed-4974-aab5-1ba06684c996","Type":"ContainerStarted","Data":"3d883900e5390251e5113ddccbf1cc0b508aaf481396f81f772eeb0cb4b6158e"} Oct 05 09:15:02 crc kubenswrapper[4846]: I1005 09:15:02.087659 4846 generic.go:334] "Generic (PLEG): container finished" podID="155d6821-91ed-4974-aab5-1ba06684c996" containerID="a35dc12a6f05f1dab426fc0134036a32267aac38347eb8f20f44fc323633fb63" exitCode=0 Oct 05 09:15:02 crc kubenswrapper[4846]: I1005 09:15:02.087726 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29327595-rcghk" event={"ID":"155d6821-91ed-4974-aab5-1ba06684c996","Type":"ContainerDied","Data":"a35dc12a6f05f1dab426fc0134036a32267aac38347eb8f20f44fc323633fb63"} Oct 05 09:15:03 crc kubenswrapper[4846]: I1005 09:15:03.515288 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29327595-rcghk" Oct 05 09:15:03 crc kubenswrapper[4846]: I1005 09:15:03.672721 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/155d6821-91ed-4974-aab5-1ba06684c996-secret-volume\") pod \"155d6821-91ed-4974-aab5-1ba06684c996\" (UID: \"155d6821-91ed-4974-aab5-1ba06684c996\") " Oct 05 09:15:03 crc kubenswrapper[4846]: I1005 09:15:03.673010 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f4wqn\" (UniqueName: \"kubernetes.io/projected/155d6821-91ed-4974-aab5-1ba06684c996-kube-api-access-f4wqn\") pod \"155d6821-91ed-4974-aab5-1ba06684c996\" (UID: \"155d6821-91ed-4974-aab5-1ba06684c996\") " Oct 05 09:15:03 crc kubenswrapper[4846]: I1005 09:15:03.673273 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/155d6821-91ed-4974-aab5-1ba06684c996-config-volume\") pod \"155d6821-91ed-4974-aab5-1ba06684c996\" (UID: \"155d6821-91ed-4974-aab5-1ba06684c996\") " Oct 05 09:15:03 crc kubenswrapper[4846]: I1005 09:15:03.673872 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/155d6821-91ed-4974-aab5-1ba06684c996-config-volume" (OuterVolumeSpecName: "config-volume") pod "155d6821-91ed-4974-aab5-1ba06684c996" (UID: "155d6821-91ed-4974-aab5-1ba06684c996"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 09:15:03 crc kubenswrapper[4846]: I1005 09:15:03.674355 4846 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/155d6821-91ed-4974-aab5-1ba06684c996-config-volume\") on node \"crc\" DevicePath \"\"" Oct 05 09:15:03 crc kubenswrapper[4846]: I1005 09:15:03.680316 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/155d6821-91ed-4974-aab5-1ba06684c996-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "155d6821-91ed-4974-aab5-1ba06684c996" (UID: "155d6821-91ed-4974-aab5-1ba06684c996"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 09:15:03 crc kubenswrapper[4846]: I1005 09:15:03.689693 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/155d6821-91ed-4974-aab5-1ba06684c996-kube-api-access-f4wqn" (OuterVolumeSpecName: "kube-api-access-f4wqn") pod "155d6821-91ed-4974-aab5-1ba06684c996" (UID: "155d6821-91ed-4974-aab5-1ba06684c996"). InnerVolumeSpecName "kube-api-access-f4wqn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 09:15:03 crc kubenswrapper[4846]: I1005 09:15:03.775539 4846 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/155d6821-91ed-4974-aab5-1ba06684c996-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 05 09:15:03 crc kubenswrapper[4846]: I1005 09:15:03.776032 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f4wqn\" (UniqueName: \"kubernetes.io/projected/155d6821-91ed-4974-aab5-1ba06684c996-kube-api-access-f4wqn\") on node \"crc\" DevicePath \"\"" Oct 05 09:15:04 crc kubenswrapper[4846]: I1005 09:15:04.115707 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29327595-rcghk" event={"ID":"155d6821-91ed-4974-aab5-1ba06684c996","Type":"ContainerDied","Data":"3d883900e5390251e5113ddccbf1cc0b508aaf481396f81f772eeb0cb4b6158e"} Oct 05 09:15:04 crc kubenswrapper[4846]: I1005 09:15:04.115764 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3d883900e5390251e5113ddccbf1cc0b508aaf481396f81f772eeb0cb4b6158e" Oct 05 09:15:04 crc kubenswrapper[4846]: I1005 09:15:04.116153 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29327595-rcghk" Oct 05 09:15:04 crc kubenswrapper[4846]: I1005 09:15:04.657538 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29327550-2445r"] Oct 05 09:15:04 crc kubenswrapper[4846]: I1005 09:15:04.694989 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29327550-2445r"] Oct 05 09:15:06 crc kubenswrapper[4846]: I1005 09:15:06.511641 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c4867d81-361b-4e63-9860-2099edac282c" path="/var/lib/kubelet/pods/c4867d81-361b-4e63-9860-2099edac282c/volumes" Oct 05 09:15:23 crc kubenswrapper[4846]: I1005 09:15:23.325648 4846 patch_prober.go:28] interesting pod/machine-config-daemon-fscvf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 09:15:23 crc kubenswrapper[4846]: I1005 09:15:23.326302 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 09:15:39 crc kubenswrapper[4846]: I1005 09:15:39.176422 4846 scope.go:117] "RemoveContainer" containerID="c4ac4f425d1a512844e1cd0c566e532368ed403a9743d1c0c2073c2db29e0862" Oct 05 09:15:53 crc kubenswrapper[4846]: I1005 09:15:53.325347 4846 patch_prober.go:28] interesting pod/machine-config-daemon-fscvf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 09:15:53 crc kubenswrapper[4846]: I1005 09:15:53.325914 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 09:16:08 crc kubenswrapper[4846]: I1005 09:16:08.870383 4846 generic.go:334] "Generic (PLEG): container finished" podID="3d69d7c2-0c3a-4850-9ad0-e9654b3e8166" containerID="17041d0d7b8e960e1186bb520c1c4b6d2c7c2fbadf9bb15ab8b35acf03ea5352" exitCode=0 Oct 05 09:16:08 crc kubenswrapper[4846]: I1005 09:16:08.870495 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-dhcp-openstack-openstack-cell1-hl2xj" event={"ID":"3d69d7c2-0c3a-4850-9ad0-e9654b3e8166","Type":"ContainerDied","Data":"17041d0d7b8e960e1186bb520c1c4b6d2c7c2fbadf9bb15ab8b35acf03ea5352"} Oct 05 09:16:10 crc kubenswrapper[4846]: I1005 09:16:10.464952 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-dhcp-openstack-openstack-cell1-hl2xj" Oct 05 09:16:10 crc kubenswrapper[4846]: I1005 09:16:10.566765 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f8gxp\" (UniqueName: \"kubernetes.io/projected/3d69d7c2-0c3a-4850-9ad0-e9654b3e8166-kube-api-access-f8gxp\") pod \"3d69d7c2-0c3a-4850-9ad0-e9654b3e8166\" (UID: \"3d69d7c2-0c3a-4850-9ad0-e9654b3e8166\") " Oct 05 09:16:10 crc kubenswrapper[4846]: I1005 09:16:10.566845 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3d69d7c2-0c3a-4850-9ad0-e9654b3e8166-inventory\") pod \"3d69d7c2-0c3a-4850-9ad0-e9654b3e8166\" (UID: \"3d69d7c2-0c3a-4850-9ad0-e9654b3e8166\") " Oct 05 09:16:10 crc kubenswrapper[4846]: I1005 09:16:10.566947 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3d69d7c2-0c3a-4850-9ad0-e9654b3e8166-neutron-dhcp-combined-ca-bundle\") pod \"3d69d7c2-0c3a-4850-9ad0-e9654b3e8166\" (UID: \"3d69d7c2-0c3a-4850-9ad0-e9654b3e8166\") " Oct 05 09:16:10 crc kubenswrapper[4846]: I1005 09:16:10.567268 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-dhcp-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/3d69d7c2-0c3a-4850-9ad0-e9654b3e8166-neutron-dhcp-agent-neutron-config-0\") pod \"3d69d7c2-0c3a-4850-9ad0-e9654b3e8166\" (UID: \"3d69d7c2-0c3a-4850-9ad0-e9654b3e8166\") " Oct 05 09:16:10 crc kubenswrapper[4846]: I1005 09:16:10.567305 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3d69d7c2-0c3a-4850-9ad0-e9654b3e8166-ssh-key\") pod \"3d69d7c2-0c3a-4850-9ad0-e9654b3e8166\" (UID: \"3d69d7c2-0c3a-4850-9ad0-e9654b3e8166\") " Oct 05 09:16:10 crc kubenswrapper[4846]: I1005 09:16:10.579784 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3d69d7c2-0c3a-4850-9ad0-e9654b3e8166-neutron-dhcp-combined-ca-bundle" (OuterVolumeSpecName: "neutron-dhcp-combined-ca-bundle") pod "3d69d7c2-0c3a-4850-9ad0-e9654b3e8166" (UID: "3d69d7c2-0c3a-4850-9ad0-e9654b3e8166"). InnerVolumeSpecName "neutron-dhcp-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 09:16:10 crc kubenswrapper[4846]: I1005 09:16:10.579867 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3d69d7c2-0c3a-4850-9ad0-e9654b3e8166-kube-api-access-f8gxp" (OuterVolumeSpecName: "kube-api-access-f8gxp") pod "3d69d7c2-0c3a-4850-9ad0-e9654b3e8166" (UID: "3d69d7c2-0c3a-4850-9ad0-e9654b3e8166"). InnerVolumeSpecName "kube-api-access-f8gxp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 09:16:10 crc kubenswrapper[4846]: I1005 09:16:10.606480 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3d69d7c2-0c3a-4850-9ad0-e9654b3e8166-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "3d69d7c2-0c3a-4850-9ad0-e9654b3e8166" (UID: "3d69d7c2-0c3a-4850-9ad0-e9654b3e8166"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 09:16:10 crc kubenswrapper[4846]: I1005 09:16:10.610839 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3d69d7c2-0c3a-4850-9ad0-e9654b3e8166-neutron-dhcp-agent-neutron-config-0" (OuterVolumeSpecName: "neutron-dhcp-agent-neutron-config-0") pod "3d69d7c2-0c3a-4850-9ad0-e9654b3e8166" (UID: "3d69d7c2-0c3a-4850-9ad0-e9654b3e8166"). InnerVolumeSpecName "neutron-dhcp-agent-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 09:16:10 crc kubenswrapper[4846]: I1005 09:16:10.616036 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3d69d7c2-0c3a-4850-9ad0-e9654b3e8166-inventory" (OuterVolumeSpecName: "inventory") pod "3d69d7c2-0c3a-4850-9ad0-e9654b3e8166" (UID: "3d69d7c2-0c3a-4850-9ad0-e9654b3e8166"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 09:16:10 crc kubenswrapper[4846]: I1005 09:16:10.670376 4846 reconciler_common.go:293] "Volume detached for volume \"neutron-dhcp-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/3d69d7c2-0c3a-4850-9ad0-e9654b3e8166-neutron-dhcp-agent-neutron-config-0\") on node \"crc\" DevicePath \"\"" Oct 05 09:16:10 crc kubenswrapper[4846]: I1005 09:16:10.670420 4846 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3d69d7c2-0c3a-4850-9ad0-e9654b3e8166-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 05 09:16:10 crc kubenswrapper[4846]: I1005 09:16:10.670437 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f8gxp\" (UniqueName: \"kubernetes.io/projected/3d69d7c2-0c3a-4850-9ad0-e9654b3e8166-kube-api-access-f8gxp\") on node \"crc\" DevicePath \"\"" Oct 05 09:16:10 crc kubenswrapper[4846]: I1005 09:16:10.670451 4846 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3d69d7c2-0c3a-4850-9ad0-e9654b3e8166-inventory\") on node \"crc\" DevicePath \"\"" Oct 05 09:16:10 crc kubenswrapper[4846]: I1005 09:16:10.670465 4846 reconciler_common.go:293] "Volume detached for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3d69d7c2-0c3a-4850-9ad0-e9654b3e8166-neutron-dhcp-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 09:16:10 crc kubenswrapper[4846]: I1005 09:16:10.898811 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-dhcp-openstack-openstack-cell1-hl2xj" event={"ID":"3d69d7c2-0c3a-4850-9ad0-e9654b3e8166","Type":"ContainerDied","Data":"29362ad5647e0db3feeb4f97c77cd492f07928f66b498a8a702c40a47ec97907"} Oct 05 09:16:10 crc kubenswrapper[4846]: I1005 09:16:10.898859 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="29362ad5647e0db3feeb4f97c77cd492f07928f66b498a8a702c40a47ec97907" Oct 05 09:16:10 crc kubenswrapper[4846]: I1005 09:16:10.898966 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-dhcp-openstack-openstack-cell1-hl2xj" Oct 05 09:16:23 crc kubenswrapper[4846]: I1005 09:16:23.324814 4846 patch_prober.go:28] interesting pod/machine-config-daemon-fscvf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 09:16:23 crc kubenswrapper[4846]: I1005 09:16:23.325313 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 09:16:23 crc kubenswrapper[4846]: I1005 09:16:23.325365 4846 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" Oct 05 09:16:23 crc kubenswrapper[4846]: I1005 09:16:23.326201 4846 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"206a9f8fee4cb0430eaa85104168af858cba45784f821a5ea04393d05e733060"} pod="openshift-machine-config-operator/machine-config-daemon-fscvf" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 05 09:16:23 crc kubenswrapper[4846]: I1005 09:16:23.326266 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" containerName="machine-config-daemon" containerID="cri-o://206a9f8fee4cb0430eaa85104168af858cba45784f821a5ea04393d05e733060" gracePeriod=600 Oct 05 09:16:24 crc kubenswrapper[4846]: I1005 09:16:24.055072 4846 generic.go:334] "Generic (PLEG): container finished" podID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" containerID="206a9f8fee4cb0430eaa85104168af858cba45784f821a5ea04393d05e733060" exitCode=0 Oct 05 09:16:24 crc kubenswrapper[4846]: I1005 09:16:24.055144 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" event={"ID":"ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7","Type":"ContainerDied","Data":"206a9f8fee4cb0430eaa85104168af858cba45784f821a5ea04393d05e733060"} Oct 05 09:16:24 crc kubenswrapper[4846]: I1005 09:16:24.055795 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" event={"ID":"ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7","Type":"ContainerStarted","Data":"2ddf4e92e4eafd2c8e76240cc0766ba587ed44b22d7136c95e5159c1c5e8c32e"} Oct 05 09:16:24 crc kubenswrapper[4846]: I1005 09:16:24.055820 4846 scope.go:117] "RemoveContainer" containerID="e39593bcdffa488d1cfc3d5c184d6e46655bb5c018d40df0f20d590de5498b3d" Oct 05 09:16:33 crc kubenswrapper[4846]: I1005 09:16:33.798440 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 05 09:16:33 crc kubenswrapper[4846]: I1005 09:16:33.799297 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell0-conductor-0" podUID="672ba7aa-607e-4872-9083-cb2847606b7c" containerName="nova-cell0-conductor-conductor" containerID="cri-o://7411f369dc75ad75e553fd68df2151115121338637f010473f2d659c8276b45e" gracePeriod=30 Oct 05 09:16:34 crc kubenswrapper[4846]: I1005 09:16:34.398489 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 05 09:16:34 crc kubenswrapper[4846]: I1005 09:16:34.398872 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-conductor-0" podUID="4950f233-520f-43d7-bba1-6695f817b029" containerName="nova-cell1-conductor-conductor" containerID="cri-o://55353d85045e1540708f470739a6fef7985bc1caeacccdc149c6f115c318e396" gracePeriod=30 Oct 05 09:16:34 crc kubenswrapper[4846]: I1005 09:16:34.658925 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 05 09:16:34 crc kubenswrapper[4846]: I1005 09:16:34.659133 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="3309d2cf-a492-4302-9aaf-30c3c1cd3f57" containerName="nova-scheduler-scheduler" containerID="cri-o://be07c0f8ff32f38239cb958ae582e85c424e467a057d0930617ab06cf7454997" gracePeriod=30 Oct 05 09:16:34 crc kubenswrapper[4846]: I1005 09:16:34.669479 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 05 09:16:34 crc kubenswrapper[4846]: I1005 09:16:34.669765 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="29bec123-d7d4-47d6-ae64-68b2aab524e4" containerName="nova-api-log" containerID="cri-o://a472b6104863e77453086f959d88f81c611f09276e87a825918364149a86c5b2" gracePeriod=30 Oct 05 09:16:34 crc kubenswrapper[4846]: I1005 09:16:34.669943 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="29bec123-d7d4-47d6-ae64-68b2aab524e4" containerName="nova-api-api" containerID="cri-o://0b1b66903a1ae5452af5f6f5c676e5e72227f1509e9438272a22b31171c8814c" gracePeriod=30 Oct 05 09:16:34 crc kubenswrapper[4846]: I1005 09:16:34.742534 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 05 09:16:34 crc kubenswrapper[4846]: I1005 09:16:34.743091 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="a0ed1ebb-faf2-48bb-8fe4-efcecaa2bb05" containerName="nova-metadata-log" containerID="cri-o://61f00a73cfd19cacff2ab04863a43cce8b74f56a2ee186d97ecc5edaa8983aec" gracePeriod=30 Oct 05 09:16:34 crc kubenswrapper[4846]: I1005 09:16:34.743283 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="a0ed1ebb-faf2-48bb-8fe4-efcecaa2bb05" containerName="nova-metadata-metadata" containerID="cri-o://426817ac622bb93bbce4d37648f9e92d5a85879c645c126fde022f827fbd461a" gracePeriod=30 Oct 05 09:16:35 crc kubenswrapper[4846]: I1005 09:16:35.222127 4846 generic.go:334] "Generic (PLEG): container finished" podID="a0ed1ebb-faf2-48bb-8fe4-efcecaa2bb05" containerID="61f00a73cfd19cacff2ab04863a43cce8b74f56a2ee186d97ecc5edaa8983aec" exitCode=143 Oct 05 09:16:35 crc kubenswrapper[4846]: I1005 09:16:35.222306 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"a0ed1ebb-faf2-48bb-8fe4-efcecaa2bb05","Type":"ContainerDied","Data":"61f00a73cfd19cacff2ab04863a43cce8b74f56a2ee186d97ecc5edaa8983aec"} Oct 05 09:16:35 crc kubenswrapper[4846]: I1005 09:16:35.228833 4846 generic.go:334] "Generic (PLEG): container finished" podID="29bec123-d7d4-47d6-ae64-68b2aab524e4" containerID="a472b6104863e77453086f959d88f81c611f09276e87a825918364149a86c5b2" exitCode=143 Oct 05 09:16:35 crc kubenswrapper[4846]: I1005 09:16:35.228873 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"29bec123-d7d4-47d6-ae64-68b2aab524e4","Type":"ContainerDied","Data":"a472b6104863e77453086f959d88f81c611f09276e87a825918364149a86c5b2"} Oct 05 09:16:35 crc kubenswrapper[4846]: E1005 09:16:35.930864 4846 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="7411f369dc75ad75e553fd68df2151115121338637f010473f2d659c8276b45e" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Oct 05 09:16:35 crc kubenswrapper[4846]: E1005 09:16:35.941605 4846 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="7411f369dc75ad75e553fd68df2151115121338637f010473f2d659c8276b45e" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Oct 05 09:16:35 crc kubenswrapper[4846]: E1005 09:16:35.947567 4846 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="7411f369dc75ad75e553fd68df2151115121338637f010473f2d659c8276b45e" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Oct 05 09:16:35 crc kubenswrapper[4846]: E1005 09:16:35.947683 4846 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-cell0-conductor-0" podUID="672ba7aa-607e-4872-9083-cb2847606b7c" containerName="nova-cell0-conductor-conductor" Oct 05 09:16:36 crc kubenswrapper[4846]: I1005 09:16:36.141400 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Oct 05 09:16:36 crc kubenswrapper[4846]: I1005 09:16:36.242499 4846 generic.go:334] "Generic (PLEG): container finished" podID="4950f233-520f-43d7-bba1-6695f817b029" containerID="55353d85045e1540708f470739a6fef7985bc1caeacccdc149c6f115c318e396" exitCode=0 Oct 05 09:16:36 crc kubenswrapper[4846]: I1005 09:16:36.243457 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"4950f233-520f-43d7-bba1-6695f817b029","Type":"ContainerDied","Data":"55353d85045e1540708f470739a6fef7985bc1caeacccdc149c6f115c318e396"} Oct 05 09:16:36 crc kubenswrapper[4846]: I1005 09:16:36.243572 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"4950f233-520f-43d7-bba1-6695f817b029","Type":"ContainerDied","Data":"70fc77f261d8108bb745ceb205f4867a778c0f57ffaea865933e28a3ed7a491f"} Oct 05 09:16:36 crc kubenswrapper[4846]: I1005 09:16:36.243663 4846 scope.go:117] "RemoveContainer" containerID="55353d85045e1540708f470739a6fef7985bc1caeacccdc149c6f115c318e396" Oct 05 09:16:36 crc kubenswrapper[4846]: I1005 09:16:36.243883 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Oct 05 09:16:36 crc kubenswrapper[4846]: I1005 09:16:36.286326 4846 scope.go:117] "RemoveContainer" containerID="55353d85045e1540708f470739a6fef7985bc1caeacccdc149c6f115c318e396" Oct 05 09:16:36 crc kubenswrapper[4846]: I1005 09:16:36.289990 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4950f233-520f-43d7-bba1-6695f817b029-config-data\") pod \"4950f233-520f-43d7-bba1-6695f817b029\" (UID: \"4950f233-520f-43d7-bba1-6695f817b029\") " Oct 05 09:16:36 crc kubenswrapper[4846]: I1005 09:16:36.290238 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7ng45\" (UniqueName: \"kubernetes.io/projected/4950f233-520f-43d7-bba1-6695f817b029-kube-api-access-7ng45\") pod \"4950f233-520f-43d7-bba1-6695f817b029\" (UID: \"4950f233-520f-43d7-bba1-6695f817b029\") " Oct 05 09:16:36 crc kubenswrapper[4846]: I1005 09:16:36.290291 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4950f233-520f-43d7-bba1-6695f817b029-combined-ca-bundle\") pod \"4950f233-520f-43d7-bba1-6695f817b029\" (UID: \"4950f233-520f-43d7-bba1-6695f817b029\") " Oct 05 09:16:36 crc kubenswrapper[4846]: E1005 09:16:36.291317 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"55353d85045e1540708f470739a6fef7985bc1caeacccdc149c6f115c318e396\": container with ID starting with 55353d85045e1540708f470739a6fef7985bc1caeacccdc149c6f115c318e396 not found: ID does not exist" containerID="55353d85045e1540708f470739a6fef7985bc1caeacccdc149c6f115c318e396" Oct 05 09:16:36 crc kubenswrapper[4846]: I1005 09:16:36.291362 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"55353d85045e1540708f470739a6fef7985bc1caeacccdc149c6f115c318e396"} err="failed to get container status \"55353d85045e1540708f470739a6fef7985bc1caeacccdc149c6f115c318e396\": rpc error: code = NotFound desc = could not find container \"55353d85045e1540708f470739a6fef7985bc1caeacccdc149c6f115c318e396\": container with ID starting with 55353d85045e1540708f470739a6fef7985bc1caeacccdc149c6f115c318e396 not found: ID does not exist" Oct 05 09:16:36 crc kubenswrapper[4846]: I1005 09:16:36.308799 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4950f233-520f-43d7-bba1-6695f817b029-kube-api-access-7ng45" (OuterVolumeSpecName: "kube-api-access-7ng45") pod "4950f233-520f-43d7-bba1-6695f817b029" (UID: "4950f233-520f-43d7-bba1-6695f817b029"). InnerVolumeSpecName "kube-api-access-7ng45". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 09:16:36 crc kubenswrapper[4846]: I1005 09:16:36.390636 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4950f233-520f-43d7-bba1-6695f817b029-config-data" (OuterVolumeSpecName: "config-data") pod "4950f233-520f-43d7-bba1-6695f817b029" (UID: "4950f233-520f-43d7-bba1-6695f817b029"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 09:16:36 crc kubenswrapper[4846]: I1005 09:16:36.391446 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4950f233-520f-43d7-bba1-6695f817b029-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4950f233-520f-43d7-bba1-6695f817b029" (UID: "4950f233-520f-43d7-bba1-6695f817b029"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 09:16:36 crc kubenswrapper[4846]: I1005 09:16:36.392538 4846 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4950f233-520f-43d7-bba1-6695f817b029-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 09:16:36 crc kubenswrapper[4846]: I1005 09:16:36.392559 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7ng45\" (UniqueName: \"kubernetes.io/projected/4950f233-520f-43d7-bba1-6695f817b029-kube-api-access-7ng45\") on node \"crc\" DevicePath \"\"" Oct 05 09:16:36 crc kubenswrapper[4846]: I1005 09:16:36.392570 4846 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4950f233-520f-43d7-bba1-6695f817b029-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 09:16:36 crc kubenswrapper[4846]: I1005 09:16:36.568926 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 05 09:16:36 crc kubenswrapper[4846]: I1005 09:16:36.581339 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 05 09:16:36 crc kubenswrapper[4846]: I1005 09:16:36.593584 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 05 09:16:36 crc kubenswrapper[4846]: E1005 09:16:36.594103 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4950f233-520f-43d7-bba1-6695f817b029" containerName="nova-cell1-conductor-conductor" Oct 05 09:16:36 crc kubenswrapper[4846]: I1005 09:16:36.594126 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="4950f233-520f-43d7-bba1-6695f817b029" containerName="nova-cell1-conductor-conductor" Oct 05 09:16:36 crc kubenswrapper[4846]: E1005 09:16:36.594140 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3d69d7c2-0c3a-4850-9ad0-e9654b3e8166" containerName="neutron-dhcp-openstack-openstack-cell1" Oct 05 09:16:36 crc kubenswrapper[4846]: I1005 09:16:36.594150 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="3d69d7c2-0c3a-4850-9ad0-e9654b3e8166" containerName="neutron-dhcp-openstack-openstack-cell1" Oct 05 09:16:36 crc kubenswrapper[4846]: E1005 09:16:36.594210 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="155d6821-91ed-4974-aab5-1ba06684c996" containerName="collect-profiles" Oct 05 09:16:36 crc kubenswrapper[4846]: I1005 09:16:36.594220 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="155d6821-91ed-4974-aab5-1ba06684c996" containerName="collect-profiles" Oct 05 09:16:36 crc kubenswrapper[4846]: I1005 09:16:36.594477 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="3d69d7c2-0c3a-4850-9ad0-e9654b3e8166" containerName="neutron-dhcp-openstack-openstack-cell1" Oct 05 09:16:36 crc kubenswrapper[4846]: I1005 09:16:36.594501 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="155d6821-91ed-4974-aab5-1ba06684c996" containerName="collect-profiles" Oct 05 09:16:36 crc kubenswrapper[4846]: I1005 09:16:36.594520 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="4950f233-520f-43d7-bba1-6695f817b029" containerName="nova-cell1-conductor-conductor" Oct 05 09:16:36 crc kubenswrapper[4846]: I1005 09:16:36.595363 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Oct 05 09:16:36 crc kubenswrapper[4846]: I1005 09:16:36.598800 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Oct 05 09:16:36 crc kubenswrapper[4846]: I1005 09:16:36.616991 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 05 09:16:36 crc kubenswrapper[4846]: I1005 09:16:36.697952 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2ggj6\" (UniqueName: \"kubernetes.io/projected/0fccb0ce-ab1d-495e-b080-8c1310a2297a-kube-api-access-2ggj6\") pod \"nova-cell1-conductor-0\" (UID: \"0fccb0ce-ab1d-495e-b080-8c1310a2297a\") " pod="openstack/nova-cell1-conductor-0" Oct 05 09:16:36 crc kubenswrapper[4846]: I1005 09:16:36.698104 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0fccb0ce-ab1d-495e-b080-8c1310a2297a-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"0fccb0ce-ab1d-495e-b080-8c1310a2297a\") " pod="openstack/nova-cell1-conductor-0" Oct 05 09:16:36 crc kubenswrapper[4846]: I1005 09:16:36.698235 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0fccb0ce-ab1d-495e-b080-8c1310a2297a-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"0fccb0ce-ab1d-495e-b080-8c1310a2297a\") " pod="openstack/nova-cell1-conductor-0" Oct 05 09:16:36 crc kubenswrapper[4846]: I1005 09:16:36.800028 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0fccb0ce-ab1d-495e-b080-8c1310a2297a-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"0fccb0ce-ab1d-495e-b080-8c1310a2297a\") " pod="openstack/nova-cell1-conductor-0" Oct 05 09:16:36 crc kubenswrapper[4846]: I1005 09:16:36.800142 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0fccb0ce-ab1d-495e-b080-8c1310a2297a-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"0fccb0ce-ab1d-495e-b080-8c1310a2297a\") " pod="openstack/nova-cell1-conductor-0" Oct 05 09:16:36 crc kubenswrapper[4846]: I1005 09:16:36.800192 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2ggj6\" (UniqueName: \"kubernetes.io/projected/0fccb0ce-ab1d-495e-b080-8c1310a2297a-kube-api-access-2ggj6\") pod \"nova-cell1-conductor-0\" (UID: \"0fccb0ce-ab1d-495e-b080-8c1310a2297a\") " pod="openstack/nova-cell1-conductor-0" Oct 05 09:16:36 crc kubenswrapper[4846]: I1005 09:16:36.804290 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0fccb0ce-ab1d-495e-b080-8c1310a2297a-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"0fccb0ce-ab1d-495e-b080-8c1310a2297a\") " pod="openstack/nova-cell1-conductor-0" Oct 05 09:16:36 crc kubenswrapper[4846]: I1005 09:16:36.806047 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0fccb0ce-ab1d-495e-b080-8c1310a2297a-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"0fccb0ce-ab1d-495e-b080-8c1310a2297a\") " pod="openstack/nova-cell1-conductor-0" Oct 05 09:16:36 crc kubenswrapper[4846]: I1005 09:16:36.819612 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2ggj6\" (UniqueName: \"kubernetes.io/projected/0fccb0ce-ab1d-495e-b080-8c1310a2297a-kube-api-access-2ggj6\") pod \"nova-cell1-conductor-0\" (UID: \"0fccb0ce-ab1d-495e-b080-8c1310a2297a\") " pod="openstack/nova-cell1-conductor-0" Oct 05 09:16:36 crc kubenswrapper[4846]: I1005 09:16:36.916355 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Oct 05 09:16:37 crc kubenswrapper[4846]: I1005 09:16:37.380764 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 05 09:16:38 crc kubenswrapper[4846]: I1005 09:16:38.183978 4846 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="a0ed1ebb-faf2-48bb-8fe4-efcecaa2bb05" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.1.96:8775/\": dial tcp 10.217.1.96:8775: connect: connection refused" Oct 05 09:16:38 crc kubenswrapper[4846]: I1005 09:16:38.184648 4846 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="a0ed1ebb-faf2-48bb-8fe4-efcecaa2bb05" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.1.96:8775/\": dial tcp 10.217.1.96:8775: connect: connection refused" Oct 05 09:16:38 crc kubenswrapper[4846]: I1005 09:16:38.270821 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"0fccb0ce-ab1d-495e-b080-8c1310a2297a","Type":"ContainerStarted","Data":"56c7705e376f9e81a6870d732108904ba893356494387512dba4d365bc69da22"} Oct 05 09:16:38 crc kubenswrapper[4846]: I1005 09:16:38.271045 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"0fccb0ce-ab1d-495e-b080-8c1310a2297a","Type":"ContainerStarted","Data":"772ab525bed6a9a04a9d9a80c69b762fec50a4c74c4b49d70f1a8322983427b1"} Oct 05 09:16:38 crc kubenswrapper[4846]: I1005 09:16:38.272215 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Oct 05 09:16:38 crc kubenswrapper[4846]: I1005 09:16:38.274606 4846 generic.go:334] "Generic (PLEG): container finished" podID="a0ed1ebb-faf2-48bb-8fe4-efcecaa2bb05" containerID="426817ac622bb93bbce4d37648f9e92d5a85879c645c126fde022f827fbd461a" exitCode=0 Oct 05 09:16:38 crc kubenswrapper[4846]: I1005 09:16:38.274646 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"a0ed1ebb-faf2-48bb-8fe4-efcecaa2bb05","Type":"ContainerDied","Data":"426817ac622bb93bbce4d37648f9e92d5a85879c645c126fde022f827fbd461a"} Oct 05 09:16:38 crc kubenswrapper[4846]: I1005 09:16:38.278313 4846 generic.go:334] "Generic (PLEG): container finished" podID="29bec123-d7d4-47d6-ae64-68b2aab524e4" containerID="0b1b66903a1ae5452af5f6f5c676e5e72227f1509e9438272a22b31171c8814c" exitCode=0 Oct 05 09:16:38 crc kubenswrapper[4846]: I1005 09:16:38.278361 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"29bec123-d7d4-47d6-ae64-68b2aab524e4","Type":"ContainerDied","Data":"0b1b66903a1ae5452af5f6f5c676e5e72227f1509e9438272a22b31171c8814c"} Oct 05 09:16:38 crc kubenswrapper[4846]: I1005 09:16:38.278387 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"29bec123-d7d4-47d6-ae64-68b2aab524e4","Type":"ContainerDied","Data":"969d76578245005e846856a5ed7bfb101ce7d29d780cc01d9b4d53ae750adb14"} Oct 05 09:16:38 crc kubenswrapper[4846]: I1005 09:16:38.278399 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="969d76578245005e846856a5ed7bfb101ce7d29d780cc01d9b4d53ae750adb14" Oct 05 09:16:38 crc kubenswrapper[4846]: I1005 09:16:38.302620 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=2.302602584 podStartE2EDuration="2.302602584s" podCreationTimestamp="2025-10-05 09:16:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 09:16:38.29250315 +0000 UTC m=+8920.533355935" watchObservedRunningTime="2025-10-05 09:16:38.302602584 +0000 UTC m=+8920.543455359" Oct 05 09:16:38 crc kubenswrapper[4846]: I1005 09:16:38.359448 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 05 09:16:38 crc kubenswrapper[4846]: I1005 09:16:38.514335 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4950f233-520f-43d7-bba1-6695f817b029" path="/var/lib/kubelet/pods/4950f233-520f-43d7-bba1-6695f817b029/volumes" Oct 05 09:16:38 crc kubenswrapper[4846]: I1005 09:16:38.548644 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/29bec123-d7d4-47d6-ae64-68b2aab524e4-public-tls-certs\") pod \"29bec123-d7d4-47d6-ae64-68b2aab524e4\" (UID: \"29bec123-d7d4-47d6-ae64-68b2aab524e4\") " Oct 05 09:16:38 crc kubenswrapper[4846]: I1005 09:16:38.548713 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnbd4\" (UniqueName: \"kubernetes.io/projected/29bec123-d7d4-47d6-ae64-68b2aab524e4-kube-api-access-rnbd4\") pod \"29bec123-d7d4-47d6-ae64-68b2aab524e4\" (UID: \"29bec123-d7d4-47d6-ae64-68b2aab524e4\") " Oct 05 09:16:38 crc kubenswrapper[4846]: I1005 09:16:38.548745 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/29bec123-d7d4-47d6-ae64-68b2aab524e4-logs\") pod \"29bec123-d7d4-47d6-ae64-68b2aab524e4\" (UID: \"29bec123-d7d4-47d6-ae64-68b2aab524e4\") " Oct 05 09:16:38 crc kubenswrapper[4846]: I1005 09:16:38.548761 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/29bec123-d7d4-47d6-ae64-68b2aab524e4-config-data\") pod \"29bec123-d7d4-47d6-ae64-68b2aab524e4\" (UID: \"29bec123-d7d4-47d6-ae64-68b2aab524e4\") " Oct 05 09:16:38 crc kubenswrapper[4846]: I1005 09:16:38.548783 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/29bec123-d7d4-47d6-ae64-68b2aab524e4-internal-tls-certs\") pod \"29bec123-d7d4-47d6-ae64-68b2aab524e4\" (UID: \"29bec123-d7d4-47d6-ae64-68b2aab524e4\") " Oct 05 09:16:38 crc kubenswrapper[4846]: I1005 09:16:38.548847 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/29bec123-d7d4-47d6-ae64-68b2aab524e4-combined-ca-bundle\") pod \"29bec123-d7d4-47d6-ae64-68b2aab524e4\" (UID: \"29bec123-d7d4-47d6-ae64-68b2aab524e4\") " Oct 05 09:16:38 crc kubenswrapper[4846]: I1005 09:16:38.550720 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/29bec123-d7d4-47d6-ae64-68b2aab524e4-logs" (OuterVolumeSpecName: "logs") pod "29bec123-d7d4-47d6-ae64-68b2aab524e4" (UID: "29bec123-d7d4-47d6-ae64-68b2aab524e4"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 09:16:38 crc kubenswrapper[4846]: I1005 09:16:38.555524 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 05 09:16:38 crc kubenswrapper[4846]: I1005 09:16:38.560442 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/29bec123-d7d4-47d6-ae64-68b2aab524e4-kube-api-access-rnbd4" (OuterVolumeSpecName: "kube-api-access-rnbd4") pod "29bec123-d7d4-47d6-ae64-68b2aab524e4" (UID: "29bec123-d7d4-47d6-ae64-68b2aab524e4"). InnerVolumeSpecName "kube-api-access-rnbd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 09:16:38 crc kubenswrapper[4846]: I1005 09:16:38.600560 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/29bec123-d7d4-47d6-ae64-68b2aab524e4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "29bec123-d7d4-47d6-ae64-68b2aab524e4" (UID: "29bec123-d7d4-47d6-ae64-68b2aab524e4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 09:16:38 crc kubenswrapper[4846]: I1005 09:16:38.609775 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/29bec123-d7d4-47d6-ae64-68b2aab524e4-config-data" (OuterVolumeSpecName: "config-data") pod "29bec123-d7d4-47d6-ae64-68b2aab524e4" (UID: "29bec123-d7d4-47d6-ae64-68b2aab524e4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 09:16:38 crc kubenswrapper[4846]: I1005 09:16:38.614650 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/29bec123-d7d4-47d6-ae64-68b2aab524e4-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "29bec123-d7d4-47d6-ae64-68b2aab524e4" (UID: "29bec123-d7d4-47d6-ae64-68b2aab524e4"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 09:16:38 crc kubenswrapper[4846]: I1005 09:16:38.614795 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/29bec123-d7d4-47d6-ae64-68b2aab524e4-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "29bec123-d7d4-47d6-ae64-68b2aab524e4" (UID: "29bec123-d7d4-47d6-ae64-68b2aab524e4"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 09:16:38 crc kubenswrapper[4846]: I1005 09:16:38.650390 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a0ed1ebb-faf2-48bb-8fe4-efcecaa2bb05-config-data\") pod \"a0ed1ebb-faf2-48bb-8fe4-efcecaa2bb05\" (UID: \"a0ed1ebb-faf2-48bb-8fe4-efcecaa2bb05\") " Oct 05 09:16:38 crc kubenswrapper[4846]: I1005 09:16:38.650439 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nf499\" (UniqueName: \"kubernetes.io/projected/a0ed1ebb-faf2-48bb-8fe4-efcecaa2bb05-kube-api-access-nf499\") pod \"a0ed1ebb-faf2-48bb-8fe4-efcecaa2bb05\" (UID: \"a0ed1ebb-faf2-48bb-8fe4-efcecaa2bb05\") " Oct 05 09:16:38 crc kubenswrapper[4846]: I1005 09:16:38.650515 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/a0ed1ebb-faf2-48bb-8fe4-efcecaa2bb05-nova-metadata-tls-certs\") pod \"a0ed1ebb-faf2-48bb-8fe4-efcecaa2bb05\" (UID: \"a0ed1ebb-faf2-48bb-8fe4-efcecaa2bb05\") " Oct 05 09:16:38 crc kubenswrapper[4846]: I1005 09:16:38.650532 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a0ed1ebb-faf2-48bb-8fe4-efcecaa2bb05-combined-ca-bundle\") pod \"a0ed1ebb-faf2-48bb-8fe4-efcecaa2bb05\" (UID: \"a0ed1ebb-faf2-48bb-8fe4-efcecaa2bb05\") " Oct 05 09:16:38 crc kubenswrapper[4846]: I1005 09:16:38.650569 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a0ed1ebb-faf2-48bb-8fe4-efcecaa2bb05-logs\") pod \"a0ed1ebb-faf2-48bb-8fe4-efcecaa2bb05\" (UID: \"a0ed1ebb-faf2-48bb-8fe4-efcecaa2bb05\") " Oct 05 09:16:38 crc kubenswrapper[4846]: I1005 09:16:38.651071 4846 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/29bec123-d7d4-47d6-ae64-68b2aab524e4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 09:16:38 crc kubenswrapper[4846]: I1005 09:16:38.651082 4846 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/29bec123-d7d4-47d6-ae64-68b2aab524e4-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 05 09:16:38 crc kubenswrapper[4846]: I1005 09:16:38.651091 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnbd4\" (UniqueName: \"kubernetes.io/projected/29bec123-d7d4-47d6-ae64-68b2aab524e4-kube-api-access-rnbd4\") on node \"crc\" DevicePath \"\"" Oct 05 09:16:38 crc kubenswrapper[4846]: I1005 09:16:38.651101 4846 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/29bec123-d7d4-47d6-ae64-68b2aab524e4-logs\") on node \"crc\" DevicePath \"\"" Oct 05 09:16:38 crc kubenswrapper[4846]: I1005 09:16:38.651111 4846 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/29bec123-d7d4-47d6-ae64-68b2aab524e4-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 09:16:38 crc kubenswrapper[4846]: I1005 09:16:38.651119 4846 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/29bec123-d7d4-47d6-ae64-68b2aab524e4-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 05 09:16:38 crc kubenswrapper[4846]: I1005 09:16:38.651425 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a0ed1ebb-faf2-48bb-8fe4-efcecaa2bb05-logs" (OuterVolumeSpecName: "logs") pod "a0ed1ebb-faf2-48bb-8fe4-efcecaa2bb05" (UID: "a0ed1ebb-faf2-48bb-8fe4-efcecaa2bb05"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 09:16:38 crc kubenswrapper[4846]: I1005 09:16:38.655878 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0ed1ebb-faf2-48bb-8fe4-efcecaa2bb05-kube-api-access-nf499" (OuterVolumeSpecName: "kube-api-access-nf499") pod "a0ed1ebb-faf2-48bb-8fe4-efcecaa2bb05" (UID: "a0ed1ebb-faf2-48bb-8fe4-efcecaa2bb05"). InnerVolumeSpecName "kube-api-access-nf499". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 09:16:38 crc kubenswrapper[4846]: I1005 09:16:38.691293 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0ed1ebb-faf2-48bb-8fe4-efcecaa2bb05-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a0ed1ebb-faf2-48bb-8fe4-efcecaa2bb05" (UID: "a0ed1ebb-faf2-48bb-8fe4-efcecaa2bb05"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 09:16:38 crc kubenswrapper[4846]: I1005 09:16:38.722755 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0ed1ebb-faf2-48bb-8fe4-efcecaa2bb05-config-data" (OuterVolumeSpecName: "config-data") pod "a0ed1ebb-faf2-48bb-8fe4-efcecaa2bb05" (UID: "a0ed1ebb-faf2-48bb-8fe4-efcecaa2bb05"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 09:16:38 crc kubenswrapper[4846]: I1005 09:16:38.736328 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0ed1ebb-faf2-48bb-8fe4-efcecaa2bb05-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "a0ed1ebb-faf2-48bb-8fe4-efcecaa2bb05" (UID: "a0ed1ebb-faf2-48bb-8fe4-efcecaa2bb05"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 09:16:38 crc kubenswrapper[4846]: I1005 09:16:38.752302 4846 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a0ed1ebb-faf2-48bb-8fe4-efcecaa2bb05-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 09:16:38 crc kubenswrapper[4846]: I1005 09:16:38.752335 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nf499\" (UniqueName: \"kubernetes.io/projected/a0ed1ebb-faf2-48bb-8fe4-efcecaa2bb05-kube-api-access-nf499\") on node \"crc\" DevicePath \"\"" Oct 05 09:16:38 crc kubenswrapper[4846]: I1005 09:16:38.752347 4846 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/a0ed1ebb-faf2-48bb-8fe4-efcecaa2bb05-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 05 09:16:38 crc kubenswrapper[4846]: I1005 09:16:38.752357 4846 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a0ed1ebb-faf2-48bb-8fe4-efcecaa2bb05-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 09:16:38 crc kubenswrapper[4846]: I1005 09:16:38.752367 4846 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a0ed1ebb-faf2-48bb-8fe4-efcecaa2bb05-logs\") on node \"crc\" DevicePath \"\"" Oct 05 09:16:38 crc kubenswrapper[4846]: I1005 09:16:38.811835 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 05 09:16:38 crc kubenswrapper[4846]: I1005 09:16:38.853821 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3309d2cf-a492-4302-9aaf-30c3c1cd3f57-combined-ca-bundle\") pod \"3309d2cf-a492-4302-9aaf-30c3c1cd3f57\" (UID: \"3309d2cf-a492-4302-9aaf-30c3c1cd3f57\") " Oct 05 09:16:38 crc kubenswrapper[4846]: I1005 09:16:38.854089 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3309d2cf-a492-4302-9aaf-30c3c1cd3f57-config-data\") pod \"3309d2cf-a492-4302-9aaf-30c3c1cd3f57\" (UID: \"3309d2cf-a492-4302-9aaf-30c3c1cd3f57\") " Oct 05 09:16:38 crc kubenswrapper[4846]: I1005 09:16:38.854135 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7qkkk\" (UniqueName: \"kubernetes.io/projected/3309d2cf-a492-4302-9aaf-30c3c1cd3f57-kube-api-access-7qkkk\") pod \"3309d2cf-a492-4302-9aaf-30c3c1cd3f57\" (UID: \"3309d2cf-a492-4302-9aaf-30c3c1cd3f57\") " Oct 05 09:16:38 crc kubenswrapper[4846]: I1005 09:16:38.857653 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3309d2cf-a492-4302-9aaf-30c3c1cd3f57-kube-api-access-7qkkk" (OuterVolumeSpecName: "kube-api-access-7qkkk") pod "3309d2cf-a492-4302-9aaf-30c3c1cd3f57" (UID: "3309d2cf-a492-4302-9aaf-30c3c1cd3f57"). InnerVolumeSpecName "kube-api-access-7qkkk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 09:16:38 crc kubenswrapper[4846]: I1005 09:16:38.884082 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3309d2cf-a492-4302-9aaf-30c3c1cd3f57-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3309d2cf-a492-4302-9aaf-30c3c1cd3f57" (UID: "3309d2cf-a492-4302-9aaf-30c3c1cd3f57"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 09:16:38 crc kubenswrapper[4846]: I1005 09:16:38.889487 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3309d2cf-a492-4302-9aaf-30c3c1cd3f57-config-data" (OuterVolumeSpecName: "config-data") pod "3309d2cf-a492-4302-9aaf-30c3c1cd3f57" (UID: "3309d2cf-a492-4302-9aaf-30c3c1cd3f57"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 09:16:38 crc kubenswrapper[4846]: I1005 09:16:38.956572 4846 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3309d2cf-a492-4302-9aaf-30c3c1cd3f57-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 09:16:38 crc kubenswrapper[4846]: I1005 09:16:38.956616 4846 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3309d2cf-a492-4302-9aaf-30c3c1cd3f57-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 09:16:38 crc kubenswrapper[4846]: I1005 09:16:38.956629 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7qkkk\" (UniqueName: \"kubernetes.io/projected/3309d2cf-a492-4302-9aaf-30c3c1cd3f57-kube-api-access-7qkkk\") on node \"crc\" DevicePath \"\"" Oct 05 09:16:39 crc kubenswrapper[4846]: I1005 09:16:39.289943 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"a0ed1ebb-faf2-48bb-8fe4-efcecaa2bb05","Type":"ContainerDied","Data":"c49958b7cdfe3ce6aceb97827968e9db73b3479550a08d910e4c331710a10cce"} Oct 05 09:16:39 crc kubenswrapper[4846]: I1005 09:16:39.290093 4846 scope.go:117] "RemoveContainer" containerID="426817ac622bb93bbce4d37648f9e92d5a85879c645c126fde022f827fbd461a" Oct 05 09:16:39 crc kubenswrapper[4846]: I1005 09:16:39.290049 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 05 09:16:39 crc kubenswrapper[4846]: I1005 09:16:39.292133 4846 generic.go:334] "Generic (PLEG): container finished" podID="3309d2cf-a492-4302-9aaf-30c3c1cd3f57" containerID="be07c0f8ff32f38239cb958ae582e85c424e467a057d0930617ab06cf7454997" exitCode=0 Oct 05 09:16:39 crc kubenswrapper[4846]: I1005 09:16:39.292238 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 05 09:16:39 crc kubenswrapper[4846]: I1005 09:16:39.292267 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 05 09:16:39 crc kubenswrapper[4846]: I1005 09:16:39.292236 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"3309d2cf-a492-4302-9aaf-30c3c1cd3f57","Type":"ContainerDied","Data":"be07c0f8ff32f38239cb958ae582e85c424e467a057d0930617ab06cf7454997"} Oct 05 09:16:39 crc kubenswrapper[4846]: I1005 09:16:39.292386 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"3309d2cf-a492-4302-9aaf-30c3c1cd3f57","Type":"ContainerDied","Data":"95bee5b37d68db33c99d15b6d7af39ba404887ee4e81b2bf1e2cafb643e3ae38"} Oct 05 09:16:39 crc kubenswrapper[4846]: I1005 09:16:39.296525 4846 scope.go:117] "RemoveContainer" containerID="be07c0f8ff32f38239cb958ae582e85c424e467a057d0930617ab06cf7454997" Oct 05 09:16:39 crc kubenswrapper[4846]: I1005 09:16:39.321617 4846 scope.go:117] "RemoveContainer" containerID="0b1b66903a1ae5452af5f6f5c676e5e72227f1509e9438272a22b31171c8814c" Oct 05 09:16:39 crc kubenswrapper[4846]: I1005 09:16:39.351400 4846 scope.go:117] "RemoveContainer" containerID="61f00a73cfd19cacff2ab04863a43cce8b74f56a2ee186d97ecc5edaa8983aec" Oct 05 09:16:39 crc kubenswrapper[4846]: I1005 09:16:39.354240 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 05 09:16:39 crc kubenswrapper[4846]: I1005 09:16:39.374964 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Oct 05 09:16:39 crc kubenswrapper[4846]: I1005 09:16:39.384706 4846 scope.go:117] "RemoveContainer" containerID="426817ac622bb93bbce4d37648f9e92d5a85879c645c126fde022f827fbd461a" Oct 05 09:16:39 crc kubenswrapper[4846]: I1005 09:16:39.388927 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 05 09:16:39 crc kubenswrapper[4846]: E1005 09:16:39.394142 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"426817ac622bb93bbce4d37648f9e92d5a85879c645c126fde022f827fbd461a\": container with ID starting with 426817ac622bb93bbce4d37648f9e92d5a85879c645c126fde022f827fbd461a not found: ID does not exist" containerID="426817ac622bb93bbce4d37648f9e92d5a85879c645c126fde022f827fbd461a" Oct 05 09:16:39 crc kubenswrapper[4846]: E1005 09:16:39.394197 4846 kuberuntime_gc.go:150] "Failed to remove container" err="failed to get container status \"426817ac622bb93bbce4d37648f9e92d5a85879c645c126fde022f827fbd461a\": rpc error: code = NotFound desc = could not find container \"426817ac622bb93bbce4d37648f9e92d5a85879c645c126fde022f827fbd461a\": container with ID starting with 426817ac622bb93bbce4d37648f9e92d5a85879c645c126fde022f827fbd461a not found: ID does not exist" containerID="426817ac622bb93bbce4d37648f9e92d5a85879c645c126fde022f827fbd461a" Oct 05 09:16:39 crc kubenswrapper[4846]: I1005 09:16:39.394239 4846 scope.go:117] "RemoveContainer" containerID="61f00a73cfd19cacff2ab04863a43cce8b74f56a2ee186d97ecc5edaa8983aec" Oct 05 09:16:39 crc kubenswrapper[4846]: I1005 09:16:39.397488 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Oct 05 09:16:39 crc kubenswrapper[4846]: I1005 09:16:39.411970 4846 scope.go:117] "RemoveContainer" containerID="be07c0f8ff32f38239cb958ae582e85c424e467a057d0930617ab06cf7454997" Oct 05 09:16:39 crc kubenswrapper[4846]: I1005 09:16:39.413164 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 05 09:16:39 crc kubenswrapper[4846]: E1005 09:16:39.413629 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="29bec123-d7d4-47d6-ae64-68b2aab524e4" containerName="nova-api-log" Oct 05 09:16:39 crc kubenswrapper[4846]: I1005 09:16:39.413640 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="29bec123-d7d4-47d6-ae64-68b2aab524e4" containerName="nova-api-log" Oct 05 09:16:39 crc kubenswrapper[4846]: E1005 09:16:39.413652 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a0ed1ebb-faf2-48bb-8fe4-efcecaa2bb05" containerName="nova-metadata-log" Oct 05 09:16:39 crc kubenswrapper[4846]: I1005 09:16:39.413658 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="a0ed1ebb-faf2-48bb-8fe4-efcecaa2bb05" containerName="nova-metadata-log" Oct 05 09:16:39 crc kubenswrapper[4846]: E1005 09:16:39.413667 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a0ed1ebb-faf2-48bb-8fe4-efcecaa2bb05" containerName="nova-metadata-metadata" Oct 05 09:16:39 crc kubenswrapper[4846]: I1005 09:16:39.413674 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="a0ed1ebb-faf2-48bb-8fe4-efcecaa2bb05" containerName="nova-metadata-metadata" Oct 05 09:16:39 crc kubenswrapper[4846]: E1005 09:16:39.413683 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3309d2cf-a492-4302-9aaf-30c3c1cd3f57" containerName="nova-scheduler-scheduler" Oct 05 09:16:39 crc kubenswrapper[4846]: I1005 09:16:39.413689 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="3309d2cf-a492-4302-9aaf-30c3c1cd3f57" containerName="nova-scheduler-scheduler" Oct 05 09:16:39 crc kubenswrapper[4846]: E1005 09:16:39.413706 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="29bec123-d7d4-47d6-ae64-68b2aab524e4" containerName="nova-api-api" Oct 05 09:16:39 crc kubenswrapper[4846]: I1005 09:16:39.413711 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="29bec123-d7d4-47d6-ae64-68b2aab524e4" containerName="nova-api-api" Oct 05 09:16:39 crc kubenswrapper[4846]: I1005 09:16:39.413888 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="3309d2cf-a492-4302-9aaf-30c3c1cd3f57" containerName="nova-scheduler-scheduler" Oct 05 09:16:39 crc kubenswrapper[4846]: I1005 09:16:39.413906 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="29bec123-d7d4-47d6-ae64-68b2aab524e4" containerName="nova-api-api" Oct 05 09:16:39 crc kubenswrapper[4846]: I1005 09:16:39.413920 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="a0ed1ebb-faf2-48bb-8fe4-efcecaa2bb05" containerName="nova-metadata-log" Oct 05 09:16:39 crc kubenswrapper[4846]: I1005 09:16:39.413938 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="a0ed1ebb-faf2-48bb-8fe4-efcecaa2bb05" containerName="nova-metadata-metadata" Oct 05 09:16:39 crc kubenswrapper[4846]: I1005 09:16:39.413945 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="29bec123-d7d4-47d6-ae64-68b2aab524e4" containerName="nova-api-log" Oct 05 09:16:39 crc kubenswrapper[4846]: I1005 09:16:39.414944 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 05 09:16:39 crc kubenswrapper[4846]: E1005 09:16:39.415932 4846 log.go:32] "RemoveContainer from runtime service failed" err="rpc error: code = Unknown desc = failed to delete container k8s_nova-metadata-log_nova-metadata-0_openstack_a0ed1ebb-faf2-48bb-8fe4-efcecaa2bb05_0 in pod sandbox c49958b7cdfe3ce6aceb97827968e9db73b3479550a08d910e4c331710a10cce from index: no such id: '61f00a73cfd19cacff2ab04863a43cce8b74f56a2ee186d97ecc5edaa8983aec'" containerID="61f00a73cfd19cacff2ab04863a43cce8b74f56a2ee186d97ecc5edaa8983aec" Oct 05 09:16:39 crc kubenswrapper[4846]: E1005 09:16:39.415958 4846 kuberuntime_gc.go:150] "Failed to remove container" err="rpc error: code = Unknown desc = failed to delete container k8s_nova-metadata-log_nova-metadata-0_openstack_a0ed1ebb-faf2-48bb-8fe4-efcecaa2bb05_0 in pod sandbox c49958b7cdfe3ce6aceb97827968e9db73b3479550a08d910e4c331710a10cce from index: no such id: '61f00a73cfd19cacff2ab04863a43cce8b74f56a2ee186d97ecc5edaa8983aec'" containerID="61f00a73cfd19cacff2ab04863a43cce8b74f56a2ee186d97ecc5edaa8983aec" Oct 05 09:16:39 crc kubenswrapper[4846]: I1005 09:16:39.415984 4846 scope.go:117] "RemoveContainer" containerID="a472b6104863e77453086f959d88f81c611f09276e87a825918364149a86c5b2" Oct 05 09:16:39 crc kubenswrapper[4846]: E1005 09:16:39.416018 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"be07c0f8ff32f38239cb958ae582e85c424e467a057d0930617ab06cf7454997\": container with ID starting with be07c0f8ff32f38239cb958ae582e85c424e467a057d0930617ab06cf7454997 not found: ID does not exist" containerID="be07c0f8ff32f38239cb958ae582e85c424e467a057d0930617ab06cf7454997" Oct 05 09:16:39 crc kubenswrapper[4846]: I1005 09:16:39.416036 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"be07c0f8ff32f38239cb958ae582e85c424e467a057d0930617ab06cf7454997"} err="failed to get container status \"be07c0f8ff32f38239cb958ae582e85c424e467a057d0930617ab06cf7454997\": rpc error: code = NotFound desc = could not find container \"be07c0f8ff32f38239cb958ae582e85c424e467a057d0930617ab06cf7454997\": container with ID starting with be07c0f8ff32f38239cb958ae582e85c424e467a057d0930617ab06cf7454997 not found: ID does not exist" Oct 05 09:16:39 crc kubenswrapper[4846]: I1005 09:16:39.416052 4846 scope.go:117] "RemoveContainer" containerID="be07c0f8ff32f38239cb958ae582e85c424e467a057d0930617ab06cf7454997" Oct 05 09:16:39 crc kubenswrapper[4846]: I1005 09:16:39.416592 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"be07c0f8ff32f38239cb958ae582e85c424e467a057d0930617ab06cf7454997"} err="failed to get container status \"be07c0f8ff32f38239cb958ae582e85c424e467a057d0930617ab06cf7454997\": rpc error: code = NotFound desc = could not find container \"be07c0f8ff32f38239cb958ae582e85c424e467a057d0930617ab06cf7454997\": container with ID starting with be07c0f8ff32f38239cb958ae582e85c424e467a057d0930617ab06cf7454997 not found: ID does not exist" Oct 05 09:16:39 crc kubenswrapper[4846]: I1005 09:16:39.418314 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Oct 05 09:16:39 crc kubenswrapper[4846]: I1005 09:16:39.418463 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 05 09:16:39 crc kubenswrapper[4846]: I1005 09:16:39.418563 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Oct 05 09:16:39 crc kubenswrapper[4846]: I1005 09:16:39.426876 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 05 09:16:39 crc kubenswrapper[4846]: I1005 09:16:39.435603 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 05 09:16:39 crc kubenswrapper[4846]: I1005 09:16:39.437845 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 05 09:16:39 crc kubenswrapper[4846]: I1005 09:16:39.445511 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 05 09:16:39 crc kubenswrapper[4846]: I1005 09:16:39.446271 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 05 09:16:39 crc kubenswrapper[4846]: I1005 09:16:39.446320 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Oct 05 09:16:39 crc kubenswrapper[4846]: I1005 09:16:39.458238 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Oct 05 09:16:39 crc kubenswrapper[4846]: I1005 09:16:39.466201 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/563f91c3-63c3-4dc0-886a-e6a1a8a5672e-config-data\") pod \"nova-api-0\" (UID: \"563f91c3-63c3-4dc0-886a-e6a1a8a5672e\") " pod="openstack/nova-api-0" Oct 05 09:16:39 crc kubenswrapper[4846]: I1005 09:16:39.466234 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/95c1c757-bd73-4e9a-9d5e-59c5970491cf-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"95c1c757-bd73-4e9a-9d5e-59c5970491cf\") " pod="openstack/nova-metadata-0" Oct 05 09:16:39 crc kubenswrapper[4846]: I1005 09:16:39.466267 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/95c1c757-bd73-4e9a-9d5e-59c5970491cf-logs\") pod \"nova-metadata-0\" (UID: \"95c1c757-bd73-4e9a-9d5e-59c5970491cf\") " pod="openstack/nova-metadata-0" Oct 05 09:16:39 crc kubenswrapper[4846]: I1005 09:16:39.466287 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/563f91c3-63c3-4dc0-886a-e6a1a8a5672e-logs\") pod \"nova-api-0\" (UID: \"563f91c3-63c3-4dc0-886a-e6a1a8a5672e\") " pod="openstack/nova-api-0" Oct 05 09:16:39 crc kubenswrapper[4846]: I1005 09:16:39.466303 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/563f91c3-63c3-4dc0-886a-e6a1a8a5672e-internal-tls-certs\") pod \"nova-api-0\" (UID: \"563f91c3-63c3-4dc0-886a-e6a1a8a5672e\") " pod="openstack/nova-api-0" Oct 05 09:16:39 crc kubenswrapper[4846]: I1005 09:16:39.466320 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/563f91c3-63c3-4dc0-886a-e6a1a8a5672e-public-tls-certs\") pod \"nova-api-0\" (UID: \"563f91c3-63c3-4dc0-886a-e6a1a8a5672e\") " pod="openstack/nova-api-0" Oct 05 09:16:39 crc kubenswrapper[4846]: I1005 09:16:39.466354 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/95c1c757-bd73-4e9a-9d5e-59c5970491cf-config-data\") pod \"nova-metadata-0\" (UID: \"95c1c757-bd73-4e9a-9d5e-59c5970491cf\") " pod="openstack/nova-metadata-0" Oct 05 09:16:39 crc kubenswrapper[4846]: I1005 09:16:39.466390 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zhjst\" (UniqueName: \"kubernetes.io/projected/563f91c3-63c3-4dc0-886a-e6a1a8a5672e-kube-api-access-zhjst\") pod \"nova-api-0\" (UID: \"563f91c3-63c3-4dc0-886a-e6a1a8a5672e\") " pod="openstack/nova-api-0" Oct 05 09:16:39 crc kubenswrapper[4846]: I1005 09:16:39.466420 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/95c1c757-bd73-4e9a-9d5e-59c5970491cf-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"95c1c757-bd73-4e9a-9d5e-59c5970491cf\") " pod="openstack/nova-metadata-0" Oct 05 09:16:39 crc kubenswrapper[4846]: I1005 09:16:39.466438 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6crb2\" (UniqueName: \"kubernetes.io/projected/95c1c757-bd73-4e9a-9d5e-59c5970491cf-kube-api-access-6crb2\") pod \"nova-metadata-0\" (UID: \"95c1c757-bd73-4e9a-9d5e-59c5970491cf\") " pod="openstack/nova-metadata-0" Oct 05 09:16:39 crc kubenswrapper[4846]: I1005 09:16:39.466463 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/563f91c3-63c3-4dc0-886a-e6a1a8a5672e-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"563f91c3-63c3-4dc0-886a-e6a1a8a5672e\") " pod="openstack/nova-api-0" Oct 05 09:16:39 crc kubenswrapper[4846]: I1005 09:16:39.485682 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Oct 05 09:16:39 crc kubenswrapper[4846]: I1005 09:16:39.487090 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 05 09:16:39 crc kubenswrapper[4846]: I1005 09:16:39.488955 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Oct 05 09:16:39 crc kubenswrapper[4846]: I1005 09:16:39.494714 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 05 09:16:39 crc kubenswrapper[4846]: I1005 09:16:39.505410 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 05 09:16:39 crc kubenswrapper[4846]: I1005 09:16:39.568833 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/563f91c3-63c3-4dc0-886a-e6a1a8a5672e-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"563f91c3-63c3-4dc0-886a-e6a1a8a5672e\") " pod="openstack/nova-api-0" Oct 05 09:16:39 crc kubenswrapper[4846]: I1005 09:16:39.568942 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9thfk\" (UniqueName: \"kubernetes.io/projected/81e95869-36b9-41c5-a1cc-e3cea1471c26-kube-api-access-9thfk\") pod \"nova-scheduler-0\" (UID: \"81e95869-36b9-41c5-a1cc-e3cea1471c26\") " pod="openstack/nova-scheduler-0" Oct 05 09:16:39 crc kubenswrapper[4846]: I1005 09:16:39.568997 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/563f91c3-63c3-4dc0-886a-e6a1a8a5672e-config-data\") pod \"nova-api-0\" (UID: \"563f91c3-63c3-4dc0-886a-e6a1a8a5672e\") " pod="openstack/nova-api-0" Oct 05 09:16:39 crc kubenswrapper[4846]: I1005 09:16:39.569016 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/95c1c757-bd73-4e9a-9d5e-59c5970491cf-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"95c1c757-bd73-4e9a-9d5e-59c5970491cf\") " pod="openstack/nova-metadata-0" Oct 05 09:16:39 crc kubenswrapper[4846]: I1005 09:16:39.569033 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/81e95869-36b9-41c5-a1cc-e3cea1471c26-config-data\") pod \"nova-scheduler-0\" (UID: \"81e95869-36b9-41c5-a1cc-e3cea1471c26\") " pod="openstack/nova-scheduler-0" Oct 05 09:16:39 crc kubenswrapper[4846]: I1005 09:16:39.569064 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/95c1c757-bd73-4e9a-9d5e-59c5970491cf-logs\") pod \"nova-metadata-0\" (UID: \"95c1c757-bd73-4e9a-9d5e-59c5970491cf\") " pod="openstack/nova-metadata-0" Oct 05 09:16:39 crc kubenswrapper[4846]: I1005 09:16:39.569086 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/563f91c3-63c3-4dc0-886a-e6a1a8a5672e-logs\") pod \"nova-api-0\" (UID: \"563f91c3-63c3-4dc0-886a-e6a1a8a5672e\") " pod="openstack/nova-api-0" Oct 05 09:16:39 crc kubenswrapper[4846]: I1005 09:16:39.569101 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/563f91c3-63c3-4dc0-886a-e6a1a8a5672e-internal-tls-certs\") pod \"nova-api-0\" (UID: \"563f91c3-63c3-4dc0-886a-e6a1a8a5672e\") " pod="openstack/nova-api-0" Oct 05 09:16:39 crc kubenswrapper[4846]: I1005 09:16:39.569116 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/563f91c3-63c3-4dc0-886a-e6a1a8a5672e-public-tls-certs\") pod \"nova-api-0\" (UID: \"563f91c3-63c3-4dc0-886a-e6a1a8a5672e\") " pod="openstack/nova-api-0" Oct 05 09:16:39 crc kubenswrapper[4846]: I1005 09:16:39.569154 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/81e95869-36b9-41c5-a1cc-e3cea1471c26-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"81e95869-36b9-41c5-a1cc-e3cea1471c26\") " pod="openstack/nova-scheduler-0" Oct 05 09:16:39 crc kubenswrapper[4846]: I1005 09:16:39.569201 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/95c1c757-bd73-4e9a-9d5e-59c5970491cf-config-data\") pod \"nova-metadata-0\" (UID: \"95c1c757-bd73-4e9a-9d5e-59c5970491cf\") " pod="openstack/nova-metadata-0" Oct 05 09:16:39 crc kubenswrapper[4846]: I1005 09:16:39.569256 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zhjst\" (UniqueName: \"kubernetes.io/projected/563f91c3-63c3-4dc0-886a-e6a1a8a5672e-kube-api-access-zhjst\") pod \"nova-api-0\" (UID: \"563f91c3-63c3-4dc0-886a-e6a1a8a5672e\") " pod="openstack/nova-api-0" Oct 05 09:16:39 crc kubenswrapper[4846]: I1005 09:16:39.569291 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/95c1c757-bd73-4e9a-9d5e-59c5970491cf-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"95c1c757-bd73-4e9a-9d5e-59c5970491cf\") " pod="openstack/nova-metadata-0" Oct 05 09:16:39 crc kubenswrapper[4846]: I1005 09:16:39.569311 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6crb2\" (UniqueName: \"kubernetes.io/projected/95c1c757-bd73-4e9a-9d5e-59c5970491cf-kube-api-access-6crb2\") pod \"nova-metadata-0\" (UID: \"95c1c757-bd73-4e9a-9d5e-59c5970491cf\") " pod="openstack/nova-metadata-0" Oct 05 09:16:39 crc kubenswrapper[4846]: I1005 09:16:39.570048 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/563f91c3-63c3-4dc0-886a-e6a1a8a5672e-logs\") pod \"nova-api-0\" (UID: \"563f91c3-63c3-4dc0-886a-e6a1a8a5672e\") " pod="openstack/nova-api-0" Oct 05 09:16:39 crc kubenswrapper[4846]: I1005 09:16:39.570323 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/95c1c757-bd73-4e9a-9d5e-59c5970491cf-logs\") pod \"nova-metadata-0\" (UID: \"95c1c757-bd73-4e9a-9d5e-59c5970491cf\") " pod="openstack/nova-metadata-0" Oct 05 09:16:39 crc kubenswrapper[4846]: I1005 09:16:39.575632 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/563f91c3-63c3-4dc0-886a-e6a1a8a5672e-internal-tls-certs\") pod \"nova-api-0\" (UID: \"563f91c3-63c3-4dc0-886a-e6a1a8a5672e\") " pod="openstack/nova-api-0" Oct 05 09:16:39 crc kubenswrapper[4846]: I1005 09:16:39.575691 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/95c1c757-bd73-4e9a-9d5e-59c5970491cf-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"95c1c757-bd73-4e9a-9d5e-59c5970491cf\") " pod="openstack/nova-metadata-0" Oct 05 09:16:39 crc kubenswrapper[4846]: I1005 09:16:39.576138 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/95c1c757-bd73-4e9a-9d5e-59c5970491cf-config-data\") pod \"nova-metadata-0\" (UID: \"95c1c757-bd73-4e9a-9d5e-59c5970491cf\") " pod="openstack/nova-metadata-0" Oct 05 09:16:39 crc kubenswrapper[4846]: I1005 09:16:39.577653 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/563f91c3-63c3-4dc0-886a-e6a1a8a5672e-public-tls-certs\") pod \"nova-api-0\" (UID: \"563f91c3-63c3-4dc0-886a-e6a1a8a5672e\") " pod="openstack/nova-api-0" Oct 05 09:16:39 crc kubenswrapper[4846]: I1005 09:16:39.578064 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/563f91c3-63c3-4dc0-886a-e6a1a8a5672e-config-data\") pod \"nova-api-0\" (UID: \"563f91c3-63c3-4dc0-886a-e6a1a8a5672e\") " pod="openstack/nova-api-0" Oct 05 09:16:39 crc kubenswrapper[4846]: I1005 09:16:39.582596 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/95c1c757-bd73-4e9a-9d5e-59c5970491cf-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"95c1c757-bd73-4e9a-9d5e-59c5970491cf\") " pod="openstack/nova-metadata-0" Oct 05 09:16:39 crc kubenswrapper[4846]: I1005 09:16:39.584981 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6crb2\" (UniqueName: \"kubernetes.io/projected/95c1c757-bd73-4e9a-9d5e-59c5970491cf-kube-api-access-6crb2\") pod \"nova-metadata-0\" (UID: \"95c1c757-bd73-4e9a-9d5e-59c5970491cf\") " pod="openstack/nova-metadata-0" Oct 05 09:16:39 crc kubenswrapper[4846]: I1005 09:16:39.585595 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/563f91c3-63c3-4dc0-886a-e6a1a8a5672e-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"563f91c3-63c3-4dc0-886a-e6a1a8a5672e\") " pod="openstack/nova-api-0" Oct 05 09:16:39 crc kubenswrapper[4846]: I1005 09:16:39.586942 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zhjst\" (UniqueName: \"kubernetes.io/projected/563f91c3-63c3-4dc0-886a-e6a1a8a5672e-kube-api-access-zhjst\") pod \"nova-api-0\" (UID: \"563f91c3-63c3-4dc0-886a-e6a1a8a5672e\") " pod="openstack/nova-api-0" Oct 05 09:16:39 crc kubenswrapper[4846]: I1005 09:16:39.670094 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9thfk\" (UniqueName: \"kubernetes.io/projected/81e95869-36b9-41c5-a1cc-e3cea1471c26-kube-api-access-9thfk\") pod \"nova-scheduler-0\" (UID: \"81e95869-36b9-41c5-a1cc-e3cea1471c26\") " pod="openstack/nova-scheduler-0" Oct 05 09:16:39 crc kubenswrapper[4846]: I1005 09:16:39.670159 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/81e95869-36b9-41c5-a1cc-e3cea1471c26-config-data\") pod \"nova-scheduler-0\" (UID: \"81e95869-36b9-41c5-a1cc-e3cea1471c26\") " pod="openstack/nova-scheduler-0" Oct 05 09:16:39 crc kubenswrapper[4846]: I1005 09:16:39.670229 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/81e95869-36b9-41c5-a1cc-e3cea1471c26-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"81e95869-36b9-41c5-a1cc-e3cea1471c26\") " pod="openstack/nova-scheduler-0" Oct 05 09:16:39 crc kubenswrapper[4846]: I1005 09:16:39.675649 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/81e95869-36b9-41c5-a1cc-e3cea1471c26-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"81e95869-36b9-41c5-a1cc-e3cea1471c26\") " pod="openstack/nova-scheduler-0" Oct 05 09:16:39 crc kubenswrapper[4846]: I1005 09:16:39.684604 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/81e95869-36b9-41c5-a1cc-e3cea1471c26-config-data\") pod \"nova-scheduler-0\" (UID: \"81e95869-36b9-41c5-a1cc-e3cea1471c26\") " pod="openstack/nova-scheduler-0" Oct 05 09:16:39 crc kubenswrapper[4846]: I1005 09:16:39.699571 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9thfk\" (UniqueName: \"kubernetes.io/projected/81e95869-36b9-41c5-a1cc-e3cea1471c26-kube-api-access-9thfk\") pod \"nova-scheduler-0\" (UID: \"81e95869-36b9-41c5-a1cc-e3cea1471c26\") " pod="openstack/nova-scheduler-0" Oct 05 09:16:39 crc kubenswrapper[4846]: I1005 09:16:39.747018 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 05 09:16:39 crc kubenswrapper[4846]: I1005 09:16:39.766030 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 05 09:16:39 crc kubenswrapper[4846]: I1005 09:16:39.820393 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 05 09:16:40 crc kubenswrapper[4846]: I1005 09:16:40.336788 4846 generic.go:334] "Generic (PLEG): container finished" podID="672ba7aa-607e-4872-9083-cb2847606b7c" containerID="7411f369dc75ad75e553fd68df2151115121338637f010473f2d659c8276b45e" exitCode=0 Oct 05 09:16:40 crc kubenswrapper[4846]: I1005 09:16:40.338354 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"672ba7aa-607e-4872-9083-cb2847606b7c","Type":"ContainerDied","Data":"7411f369dc75ad75e553fd68df2151115121338637f010473f2d659c8276b45e"} Oct 05 09:16:40 crc kubenswrapper[4846]: I1005 09:16:40.338381 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"672ba7aa-607e-4872-9083-cb2847606b7c","Type":"ContainerDied","Data":"16e9d57bae311cbe11b32d992f81e5b45f1c88394da60fc86883e372c4962abc"} Oct 05 09:16:40 crc kubenswrapper[4846]: I1005 09:16:40.338393 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="16e9d57bae311cbe11b32d992f81e5b45f1c88394da60fc86883e372c4962abc" Oct 05 09:16:40 crc kubenswrapper[4846]: I1005 09:16:40.387963 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 05 09:16:40 crc kubenswrapper[4846]: W1005 09:16:40.414750 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod563f91c3_63c3_4dc0_886a_e6a1a8a5672e.slice/crio-41d2b38f3ae9a2f6fcf62be44e913d330da99850dfe011d5d464f08667e855a7 WatchSource:0}: Error finding container 41d2b38f3ae9a2f6fcf62be44e913d330da99850dfe011d5d464f08667e855a7: Status 404 returned error can't find the container with id 41d2b38f3ae9a2f6fcf62be44e913d330da99850dfe011d5d464f08667e855a7 Oct 05 09:16:40 crc kubenswrapper[4846]: I1005 09:16:40.481436 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 05 09:16:40 crc kubenswrapper[4846]: I1005 09:16:40.485615 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Oct 05 09:16:40 crc kubenswrapper[4846]: I1005 09:16:40.496739 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 05 09:16:40 crc kubenswrapper[4846]: I1005 09:16:40.538976 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="29bec123-d7d4-47d6-ae64-68b2aab524e4" path="/var/lib/kubelet/pods/29bec123-d7d4-47d6-ae64-68b2aab524e4/volumes" Oct 05 09:16:40 crc kubenswrapper[4846]: I1005 09:16:40.550708 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3309d2cf-a492-4302-9aaf-30c3c1cd3f57" path="/var/lib/kubelet/pods/3309d2cf-a492-4302-9aaf-30c3c1cd3f57/volumes" Oct 05 09:16:40 crc kubenswrapper[4846]: I1005 09:16:40.551719 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0ed1ebb-faf2-48bb-8fe4-efcecaa2bb05" path="/var/lib/kubelet/pods/a0ed1ebb-faf2-48bb-8fe4-efcecaa2bb05/volumes" Oct 05 09:16:40 crc kubenswrapper[4846]: I1005 09:16:40.594000 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-746n7\" (UniqueName: \"kubernetes.io/projected/672ba7aa-607e-4872-9083-cb2847606b7c-kube-api-access-746n7\") pod \"672ba7aa-607e-4872-9083-cb2847606b7c\" (UID: \"672ba7aa-607e-4872-9083-cb2847606b7c\") " Oct 05 09:16:40 crc kubenswrapper[4846]: I1005 09:16:40.594058 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/672ba7aa-607e-4872-9083-cb2847606b7c-config-data\") pod \"672ba7aa-607e-4872-9083-cb2847606b7c\" (UID: \"672ba7aa-607e-4872-9083-cb2847606b7c\") " Oct 05 09:16:40 crc kubenswrapper[4846]: I1005 09:16:40.594129 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/672ba7aa-607e-4872-9083-cb2847606b7c-combined-ca-bundle\") pod \"672ba7aa-607e-4872-9083-cb2847606b7c\" (UID: \"672ba7aa-607e-4872-9083-cb2847606b7c\") " Oct 05 09:16:40 crc kubenswrapper[4846]: I1005 09:16:40.600393 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/672ba7aa-607e-4872-9083-cb2847606b7c-kube-api-access-746n7" (OuterVolumeSpecName: "kube-api-access-746n7") pod "672ba7aa-607e-4872-9083-cb2847606b7c" (UID: "672ba7aa-607e-4872-9083-cb2847606b7c"). InnerVolumeSpecName "kube-api-access-746n7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 09:16:40 crc kubenswrapper[4846]: I1005 09:16:40.680190 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/672ba7aa-607e-4872-9083-cb2847606b7c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "672ba7aa-607e-4872-9083-cb2847606b7c" (UID: "672ba7aa-607e-4872-9083-cb2847606b7c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 09:16:40 crc kubenswrapper[4846]: I1005 09:16:40.685083 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/672ba7aa-607e-4872-9083-cb2847606b7c-config-data" (OuterVolumeSpecName: "config-data") pod "672ba7aa-607e-4872-9083-cb2847606b7c" (UID: "672ba7aa-607e-4872-9083-cb2847606b7c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 09:16:40 crc kubenswrapper[4846]: I1005 09:16:40.696653 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-746n7\" (UniqueName: \"kubernetes.io/projected/672ba7aa-607e-4872-9083-cb2847606b7c-kube-api-access-746n7\") on node \"crc\" DevicePath \"\"" Oct 05 09:16:40 crc kubenswrapper[4846]: I1005 09:16:40.696685 4846 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/672ba7aa-607e-4872-9083-cb2847606b7c-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 09:16:40 crc kubenswrapper[4846]: I1005 09:16:40.696695 4846 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/672ba7aa-607e-4872-9083-cb2847606b7c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 09:16:41 crc kubenswrapper[4846]: I1005 09:16:41.354916 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"563f91c3-63c3-4dc0-886a-e6a1a8a5672e","Type":"ContainerStarted","Data":"9f239fb33014b6bbb7b12338419696e7e7d8cfdf8112dc9636e208a6d9db863c"} Oct 05 09:16:41 crc kubenswrapper[4846]: I1005 09:16:41.356848 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"563f91c3-63c3-4dc0-886a-e6a1a8a5672e","Type":"ContainerStarted","Data":"fca9afea702b6371ecddeb920f15e589a907d799c7f26b0f8927f5df2c79eccf"} Oct 05 09:16:41 crc kubenswrapper[4846]: I1005 09:16:41.356871 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"563f91c3-63c3-4dc0-886a-e6a1a8a5672e","Type":"ContainerStarted","Data":"41d2b38f3ae9a2f6fcf62be44e913d330da99850dfe011d5d464f08667e855a7"} Oct 05 09:16:41 crc kubenswrapper[4846]: I1005 09:16:41.373359 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"95c1c757-bd73-4e9a-9d5e-59c5970491cf","Type":"ContainerStarted","Data":"ae4dbfca2db023bcd5ccf58492d3788ed39e149243de8f63234f8a22f6eb2584"} Oct 05 09:16:41 crc kubenswrapper[4846]: I1005 09:16:41.373456 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"95c1c757-bd73-4e9a-9d5e-59c5970491cf","Type":"ContainerStarted","Data":"1add530d50852ea7e410259bf82632df00b1cc8be21bb6f8d0d87536b7814f92"} Oct 05 09:16:41 crc kubenswrapper[4846]: I1005 09:16:41.373479 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"95c1c757-bd73-4e9a-9d5e-59c5970491cf","Type":"ContainerStarted","Data":"950d553c3635de32e3f2e1f8465f6117b6de103e51a3656daa95ed20a255e11c"} Oct 05 09:16:41 crc kubenswrapper[4846]: I1005 09:16:41.396962 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Oct 05 09:16:41 crc kubenswrapper[4846]: I1005 09:16:41.400272 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"81e95869-36b9-41c5-a1cc-e3cea1471c26","Type":"ContainerStarted","Data":"75366df709f778f06bae94af592636b0f5beacb1c8c86c2fe6811a01924ea268"} Oct 05 09:16:41 crc kubenswrapper[4846]: I1005 09:16:41.400327 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"81e95869-36b9-41c5-a1cc-e3cea1471c26","Type":"ContainerStarted","Data":"94dd2997861e532c17f13fa01bcaddfc3315f24d685ffb4a30e8150547507e52"} Oct 05 09:16:41 crc kubenswrapper[4846]: I1005 09:16:41.404988 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.404962192 podStartE2EDuration="2.404962192s" podCreationTimestamp="2025-10-05 09:16:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 09:16:41.378428118 +0000 UTC m=+8923.619280893" watchObservedRunningTime="2025-10-05 09:16:41.404962192 +0000 UTC m=+8923.645814967" Oct 05 09:16:41 crc kubenswrapper[4846]: I1005 09:16:41.413469 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.413424903 podStartE2EDuration="2.413424903s" podCreationTimestamp="2025-10-05 09:16:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 09:16:41.40146895 +0000 UTC m=+8923.642321725" watchObservedRunningTime="2025-10-05 09:16:41.413424903 +0000 UTC m=+8923.654277678" Oct 05 09:16:41 crc kubenswrapper[4846]: I1005 09:16:41.431819 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.431798764 podStartE2EDuration="2.431798764s" podCreationTimestamp="2025-10-05 09:16:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 09:16:41.427681616 +0000 UTC m=+8923.668534391" watchObservedRunningTime="2025-10-05 09:16:41.431798764 +0000 UTC m=+8923.672651529" Oct 05 09:16:41 crc kubenswrapper[4846]: I1005 09:16:41.451614 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 05 09:16:41 crc kubenswrapper[4846]: I1005 09:16:41.470503 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 05 09:16:41 crc kubenswrapper[4846]: I1005 09:16:41.488414 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 05 09:16:41 crc kubenswrapper[4846]: E1005 09:16:41.489068 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="672ba7aa-607e-4872-9083-cb2847606b7c" containerName="nova-cell0-conductor-conductor" Oct 05 09:16:41 crc kubenswrapper[4846]: I1005 09:16:41.489087 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="672ba7aa-607e-4872-9083-cb2847606b7c" containerName="nova-cell0-conductor-conductor" Oct 05 09:16:41 crc kubenswrapper[4846]: I1005 09:16:41.489315 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="672ba7aa-607e-4872-9083-cb2847606b7c" containerName="nova-cell0-conductor-conductor" Oct 05 09:16:41 crc kubenswrapper[4846]: I1005 09:16:41.490047 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Oct 05 09:16:41 crc kubenswrapper[4846]: I1005 09:16:41.493760 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Oct 05 09:16:41 crc kubenswrapper[4846]: I1005 09:16:41.501449 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 05 09:16:41 crc kubenswrapper[4846]: I1005 09:16:41.615060 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vphng\" (UniqueName: \"kubernetes.io/projected/74d21914-81a3-47ce-8a85-3d52430d9422-kube-api-access-vphng\") pod \"nova-cell0-conductor-0\" (UID: \"74d21914-81a3-47ce-8a85-3d52430d9422\") " pod="openstack/nova-cell0-conductor-0" Oct 05 09:16:41 crc kubenswrapper[4846]: I1005 09:16:41.615259 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/74d21914-81a3-47ce-8a85-3d52430d9422-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"74d21914-81a3-47ce-8a85-3d52430d9422\") " pod="openstack/nova-cell0-conductor-0" Oct 05 09:16:41 crc kubenswrapper[4846]: I1005 09:16:41.615387 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/74d21914-81a3-47ce-8a85-3d52430d9422-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"74d21914-81a3-47ce-8a85-3d52430d9422\") " pod="openstack/nova-cell0-conductor-0" Oct 05 09:16:41 crc kubenswrapper[4846]: I1005 09:16:41.717513 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vphng\" (UniqueName: \"kubernetes.io/projected/74d21914-81a3-47ce-8a85-3d52430d9422-kube-api-access-vphng\") pod \"nova-cell0-conductor-0\" (UID: \"74d21914-81a3-47ce-8a85-3d52430d9422\") " pod="openstack/nova-cell0-conductor-0" Oct 05 09:16:41 crc kubenswrapper[4846]: I1005 09:16:41.717589 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/74d21914-81a3-47ce-8a85-3d52430d9422-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"74d21914-81a3-47ce-8a85-3d52430d9422\") " pod="openstack/nova-cell0-conductor-0" Oct 05 09:16:41 crc kubenswrapper[4846]: I1005 09:16:41.717654 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/74d21914-81a3-47ce-8a85-3d52430d9422-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"74d21914-81a3-47ce-8a85-3d52430d9422\") " pod="openstack/nova-cell0-conductor-0" Oct 05 09:16:41 crc kubenswrapper[4846]: I1005 09:16:41.723722 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/74d21914-81a3-47ce-8a85-3d52430d9422-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"74d21914-81a3-47ce-8a85-3d52430d9422\") " pod="openstack/nova-cell0-conductor-0" Oct 05 09:16:41 crc kubenswrapper[4846]: I1005 09:16:41.724078 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/74d21914-81a3-47ce-8a85-3d52430d9422-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"74d21914-81a3-47ce-8a85-3d52430d9422\") " pod="openstack/nova-cell0-conductor-0" Oct 05 09:16:41 crc kubenswrapper[4846]: I1005 09:16:41.736305 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vphng\" (UniqueName: \"kubernetes.io/projected/74d21914-81a3-47ce-8a85-3d52430d9422-kube-api-access-vphng\") pod \"nova-cell0-conductor-0\" (UID: \"74d21914-81a3-47ce-8a85-3d52430d9422\") " pod="openstack/nova-cell0-conductor-0" Oct 05 09:16:41 crc kubenswrapper[4846]: I1005 09:16:41.808895 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Oct 05 09:16:42 crc kubenswrapper[4846]: I1005 09:16:42.331143 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 05 09:16:42 crc kubenswrapper[4846]: W1005 09:16:42.337255 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod74d21914_81a3_47ce_8a85_3d52430d9422.slice/crio-a3b05a0dbae81840c35c6b857781d77efa9b608dba2699e04acfad3a7b5e674b WatchSource:0}: Error finding container a3b05a0dbae81840c35c6b857781d77efa9b608dba2699e04acfad3a7b5e674b: Status 404 returned error can't find the container with id a3b05a0dbae81840c35c6b857781d77efa9b608dba2699e04acfad3a7b5e674b Oct 05 09:16:42 crc kubenswrapper[4846]: I1005 09:16:42.414920 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"74d21914-81a3-47ce-8a85-3d52430d9422","Type":"ContainerStarted","Data":"a3b05a0dbae81840c35c6b857781d77efa9b608dba2699e04acfad3a7b5e674b"} Oct 05 09:16:42 crc kubenswrapper[4846]: I1005 09:16:42.516602 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="672ba7aa-607e-4872-9083-cb2847606b7c" path="/var/lib/kubelet/pods/672ba7aa-607e-4872-9083-cb2847606b7c/volumes" Oct 05 09:16:43 crc kubenswrapper[4846]: I1005 09:16:43.425956 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"74d21914-81a3-47ce-8a85-3d52430d9422","Type":"ContainerStarted","Data":"5fa10366d9298ee3bbdef230aa5dcb99f6312f093bc421c64dc8764e2f31dd7d"} Oct 05 09:16:43 crc kubenswrapper[4846]: I1005 09:16:43.426279 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Oct 05 09:16:44 crc kubenswrapper[4846]: I1005 09:16:44.766915 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 05 09:16:44 crc kubenswrapper[4846]: I1005 09:16:44.767257 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 05 09:16:44 crc kubenswrapper[4846]: I1005 09:16:44.820473 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Oct 05 09:16:46 crc kubenswrapper[4846]: I1005 09:16:46.949892 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Oct 05 09:16:46 crc kubenswrapper[4846]: I1005 09:16:46.976461 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=5.97644097 podStartE2EDuration="5.97644097s" podCreationTimestamp="2025-10-05 09:16:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 09:16:43.445123381 +0000 UTC m=+8925.685976156" watchObservedRunningTime="2025-10-05 09:16:46.97644097 +0000 UTC m=+8929.217293755" Oct 05 09:16:49 crc kubenswrapper[4846]: I1005 09:16:49.747435 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 05 09:16:49 crc kubenswrapper[4846]: I1005 09:16:49.747795 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 05 09:16:49 crc kubenswrapper[4846]: I1005 09:16:49.766784 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 05 09:16:49 crc kubenswrapper[4846]: I1005 09:16:49.766849 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 05 09:16:49 crc kubenswrapper[4846]: I1005 09:16:49.821070 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Oct 05 09:16:49 crc kubenswrapper[4846]: I1005 09:16:49.872839 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Oct 05 09:16:50 crc kubenswrapper[4846]: I1005 09:16:50.542245 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Oct 05 09:16:50 crc kubenswrapper[4846]: I1005 09:16:50.765391 4846 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="563f91c3-63c3-4dc0-886a-e6a1a8a5672e" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.1.187:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 05 09:16:50 crc kubenswrapper[4846]: I1005 09:16:50.765450 4846 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="563f91c3-63c3-4dc0-886a-e6a1a8a5672e" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.1.187:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 05 09:16:50 crc kubenswrapper[4846]: I1005 09:16:50.789350 4846 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="95c1c757-bd73-4e9a-9d5e-59c5970491cf" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.1.188:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 05 09:16:50 crc kubenswrapper[4846]: I1005 09:16:50.789392 4846 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="95c1c757-bd73-4e9a-9d5e-59c5970491cf" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.1.188:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 05 09:16:51 crc kubenswrapper[4846]: I1005 09:16:51.860750 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Oct 05 09:16:59 crc kubenswrapper[4846]: I1005 09:16:59.761316 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 05 09:16:59 crc kubenswrapper[4846]: I1005 09:16:59.762655 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 05 09:16:59 crc kubenswrapper[4846]: I1005 09:16:59.762762 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 05 09:16:59 crc kubenswrapper[4846]: I1005 09:16:59.775210 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 05 09:16:59 crc kubenswrapper[4846]: I1005 09:16:59.780264 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 05 09:16:59 crc kubenswrapper[4846]: I1005 09:16:59.780599 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 05 09:16:59 crc kubenswrapper[4846]: I1005 09:16:59.783077 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 05 09:17:00 crc kubenswrapper[4846]: I1005 09:17:00.627792 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 05 09:17:00 crc kubenswrapper[4846]: I1005 09:17:00.634232 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 05 09:17:00 crc kubenswrapper[4846]: I1005 09:17:00.635311 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 05 09:17:01 crc kubenswrapper[4846]: I1005 09:17:01.942541 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celldnhqr"] Oct 05 09:17:01 crc kubenswrapper[4846]: I1005 09:17:01.945358 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celldnhqr" Oct 05 09:17:01 crc kubenswrapper[4846]: I1005 09:17:01.949633 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Oct 05 09:17:01 crc kubenswrapper[4846]: I1005 09:17:01.950649 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-migration-ssh-key" Oct 05 09:17:01 crc kubenswrapper[4846]: I1005 09:17:01.951008 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 05 09:17:01 crc kubenswrapper[4846]: I1005 09:17:01.951543 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"nova-cells-global-config" Oct 05 09:17:01 crc kubenswrapper[4846]: I1005 09:17:01.953689 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-7687s" Oct 05 09:17:01 crc kubenswrapper[4846]: I1005 09:17:01.954009 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-compute-config" Oct 05 09:17:01 crc kubenswrapper[4846]: I1005 09:17:01.954631 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Oct 05 09:17:01 crc kubenswrapper[4846]: I1005 09:17:01.961125 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celldnhqr"] Oct 05 09:17:02 crc kubenswrapper[4846]: I1005 09:17:02.094105 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k9xc7\" (UniqueName: \"kubernetes.io/projected/4e445bd2-1270-4c05-9e0b-bcdbec35fa74-kube-api-access-k9xc7\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celldnhqr\" (UID: \"4e445bd2-1270-4c05-9e0b-bcdbec35fa74\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celldnhqr" Oct 05 09:17:02 crc kubenswrapper[4846]: I1005 09:17:02.094259 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/4e445bd2-1270-4c05-9e0b-bcdbec35fa74-nova-cell1-compute-config-1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celldnhqr\" (UID: \"4e445bd2-1270-4c05-9e0b-bcdbec35fa74\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celldnhqr" Oct 05 09:17:02 crc kubenswrapper[4846]: I1005 09:17:02.094315 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4e445bd2-1270-4c05-9e0b-bcdbec35fa74-nova-cell1-combined-ca-bundle\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celldnhqr\" (UID: \"4e445bd2-1270-4c05-9e0b-bcdbec35fa74\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celldnhqr" Oct 05 09:17:02 crc kubenswrapper[4846]: I1005 09:17:02.094448 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/4e445bd2-1270-4c05-9e0b-bcdbec35fa74-nova-migration-ssh-key-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celldnhqr\" (UID: \"4e445bd2-1270-4c05-9e0b-bcdbec35fa74\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celldnhqr" Oct 05 09:17:02 crc kubenswrapper[4846]: I1005 09:17:02.094496 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/4e445bd2-1270-4c05-9e0b-bcdbec35fa74-nova-cell1-compute-config-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celldnhqr\" (UID: \"4e445bd2-1270-4c05-9e0b-bcdbec35fa74\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celldnhqr" Oct 05 09:17:02 crc kubenswrapper[4846]: I1005 09:17:02.094546 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/4e445bd2-1270-4c05-9e0b-bcdbec35fa74-nova-cells-global-config-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celldnhqr\" (UID: \"4e445bd2-1270-4c05-9e0b-bcdbec35fa74\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celldnhqr" Oct 05 09:17:02 crc kubenswrapper[4846]: I1005 09:17:02.094746 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4e445bd2-1270-4c05-9e0b-bcdbec35fa74-inventory\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celldnhqr\" (UID: \"4e445bd2-1270-4c05-9e0b-bcdbec35fa74\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celldnhqr" Oct 05 09:17:02 crc kubenswrapper[4846]: I1005 09:17:02.094931 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/4e445bd2-1270-4c05-9e0b-bcdbec35fa74-nova-migration-ssh-key-1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celldnhqr\" (UID: \"4e445bd2-1270-4c05-9e0b-bcdbec35fa74\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celldnhqr" Oct 05 09:17:02 crc kubenswrapper[4846]: I1005 09:17:02.095043 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4e445bd2-1270-4c05-9e0b-bcdbec35fa74-ssh-key\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celldnhqr\" (UID: \"4e445bd2-1270-4c05-9e0b-bcdbec35fa74\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celldnhqr" Oct 05 09:17:02 crc kubenswrapper[4846]: I1005 09:17:02.197081 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4e445bd2-1270-4c05-9e0b-bcdbec35fa74-inventory\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celldnhqr\" (UID: \"4e445bd2-1270-4c05-9e0b-bcdbec35fa74\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celldnhqr" Oct 05 09:17:02 crc kubenswrapper[4846]: I1005 09:17:02.197537 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/4e445bd2-1270-4c05-9e0b-bcdbec35fa74-nova-migration-ssh-key-1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celldnhqr\" (UID: \"4e445bd2-1270-4c05-9e0b-bcdbec35fa74\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celldnhqr" Oct 05 09:17:02 crc kubenswrapper[4846]: I1005 09:17:02.197626 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4e445bd2-1270-4c05-9e0b-bcdbec35fa74-ssh-key\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celldnhqr\" (UID: \"4e445bd2-1270-4c05-9e0b-bcdbec35fa74\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celldnhqr" Oct 05 09:17:02 crc kubenswrapper[4846]: I1005 09:17:02.197753 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k9xc7\" (UniqueName: \"kubernetes.io/projected/4e445bd2-1270-4c05-9e0b-bcdbec35fa74-kube-api-access-k9xc7\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celldnhqr\" (UID: \"4e445bd2-1270-4c05-9e0b-bcdbec35fa74\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celldnhqr" Oct 05 09:17:02 crc kubenswrapper[4846]: I1005 09:17:02.197847 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/4e445bd2-1270-4c05-9e0b-bcdbec35fa74-nova-cell1-compute-config-1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celldnhqr\" (UID: \"4e445bd2-1270-4c05-9e0b-bcdbec35fa74\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celldnhqr" Oct 05 09:17:02 crc kubenswrapper[4846]: I1005 09:17:02.197888 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4e445bd2-1270-4c05-9e0b-bcdbec35fa74-nova-cell1-combined-ca-bundle\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celldnhqr\" (UID: \"4e445bd2-1270-4c05-9e0b-bcdbec35fa74\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celldnhqr" Oct 05 09:17:02 crc kubenswrapper[4846]: I1005 09:17:02.198000 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/4e445bd2-1270-4c05-9e0b-bcdbec35fa74-nova-migration-ssh-key-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celldnhqr\" (UID: \"4e445bd2-1270-4c05-9e0b-bcdbec35fa74\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celldnhqr" Oct 05 09:17:02 crc kubenswrapper[4846]: I1005 09:17:02.198041 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/4e445bd2-1270-4c05-9e0b-bcdbec35fa74-nova-cell1-compute-config-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celldnhqr\" (UID: \"4e445bd2-1270-4c05-9e0b-bcdbec35fa74\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celldnhqr" Oct 05 09:17:02 crc kubenswrapper[4846]: I1005 09:17:02.198075 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/4e445bd2-1270-4c05-9e0b-bcdbec35fa74-nova-cells-global-config-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celldnhqr\" (UID: \"4e445bd2-1270-4c05-9e0b-bcdbec35fa74\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celldnhqr" Oct 05 09:17:02 crc kubenswrapper[4846]: I1005 09:17:02.199442 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/4e445bd2-1270-4c05-9e0b-bcdbec35fa74-nova-cells-global-config-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celldnhqr\" (UID: \"4e445bd2-1270-4c05-9e0b-bcdbec35fa74\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celldnhqr" Oct 05 09:17:02 crc kubenswrapper[4846]: I1005 09:17:02.207746 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4e445bd2-1270-4c05-9e0b-bcdbec35fa74-ssh-key\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celldnhqr\" (UID: \"4e445bd2-1270-4c05-9e0b-bcdbec35fa74\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celldnhqr" Oct 05 09:17:02 crc kubenswrapper[4846]: I1005 09:17:02.208486 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4e445bd2-1270-4c05-9e0b-bcdbec35fa74-inventory\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celldnhqr\" (UID: \"4e445bd2-1270-4c05-9e0b-bcdbec35fa74\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celldnhqr" Oct 05 09:17:02 crc kubenswrapper[4846]: I1005 09:17:02.209547 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/4e445bd2-1270-4c05-9e0b-bcdbec35fa74-nova-cell1-compute-config-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celldnhqr\" (UID: \"4e445bd2-1270-4c05-9e0b-bcdbec35fa74\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celldnhqr" Oct 05 09:17:02 crc kubenswrapper[4846]: I1005 09:17:02.209609 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4e445bd2-1270-4c05-9e0b-bcdbec35fa74-nova-cell1-combined-ca-bundle\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celldnhqr\" (UID: \"4e445bd2-1270-4c05-9e0b-bcdbec35fa74\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celldnhqr" Oct 05 09:17:02 crc kubenswrapper[4846]: I1005 09:17:02.221124 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/4e445bd2-1270-4c05-9e0b-bcdbec35fa74-nova-migration-ssh-key-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celldnhqr\" (UID: \"4e445bd2-1270-4c05-9e0b-bcdbec35fa74\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celldnhqr" Oct 05 09:17:02 crc kubenswrapper[4846]: I1005 09:17:02.221352 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/4e445bd2-1270-4c05-9e0b-bcdbec35fa74-nova-migration-ssh-key-1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celldnhqr\" (UID: \"4e445bd2-1270-4c05-9e0b-bcdbec35fa74\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celldnhqr" Oct 05 09:17:02 crc kubenswrapper[4846]: I1005 09:17:02.226591 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/4e445bd2-1270-4c05-9e0b-bcdbec35fa74-nova-cell1-compute-config-1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celldnhqr\" (UID: \"4e445bd2-1270-4c05-9e0b-bcdbec35fa74\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celldnhqr" Oct 05 09:17:02 crc kubenswrapper[4846]: I1005 09:17:02.240792 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k9xc7\" (UniqueName: \"kubernetes.io/projected/4e445bd2-1270-4c05-9e0b-bcdbec35fa74-kube-api-access-k9xc7\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celldnhqr\" (UID: \"4e445bd2-1270-4c05-9e0b-bcdbec35fa74\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celldnhqr" Oct 05 09:17:02 crc kubenswrapper[4846]: I1005 09:17:02.271924 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celldnhqr" Oct 05 09:17:02 crc kubenswrapper[4846]: I1005 09:17:02.843576 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celldnhqr"] Oct 05 09:17:02 crc kubenswrapper[4846]: W1005 09:17:02.849880 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4e445bd2_1270_4c05_9e0b_bcdbec35fa74.slice/crio-4e094dc1e72e556a087a8e51a20b8654c7e32d2bb4dc6a2b32e16f933efda61e WatchSource:0}: Error finding container 4e094dc1e72e556a087a8e51a20b8654c7e32d2bb4dc6a2b32e16f933efda61e: Status 404 returned error can't find the container with id 4e094dc1e72e556a087a8e51a20b8654c7e32d2bb4dc6a2b32e16f933efda61e Oct 05 09:17:02 crc kubenswrapper[4846]: I1005 09:17:02.857540 4846 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 05 09:17:03 crc kubenswrapper[4846]: I1005 09:17:03.661681 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celldnhqr" event={"ID":"4e445bd2-1270-4c05-9e0b-bcdbec35fa74","Type":"ContainerStarted","Data":"4e094dc1e72e556a087a8e51a20b8654c7e32d2bb4dc6a2b32e16f933efda61e"} Oct 05 09:17:04 crc kubenswrapper[4846]: I1005 09:17:04.705760 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celldnhqr" event={"ID":"4e445bd2-1270-4c05-9e0b-bcdbec35fa74","Type":"ContainerStarted","Data":"dbbcf7bfea17739c7d2f5e98e099b63bbc5451902575a503bbfe34e745ba064b"} Oct 05 09:17:04 crc kubenswrapper[4846]: I1005 09:17:04.745721 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celldnhqr" podStartSLOduration=3.255401834 podStartE2EDuration="3.745685219s" podCreationTimestamp="2025-10-05 09:17:01 +0000 UTC" firstStartedPulling="2025-10-05 09:17:02.857211377 +0000 UTC m=+8945.098064162" lastFinishedPulling="2025-10-05 09:17:03.347494732 +0000 UTC m=+8945.588347547" observedRunningTime="2025-10-05 09:17:04.738757067 +0000 UTC m=+8946.979609882" watchObservedRunningTime="2025-10-05 09:17:04.745685219 +0000 UTC m=+8946.986538004" Oct 05 09:17:39 crc kubenswrapper[4846]: I1005 09:17:39.569974 4846 scope.go:117] "RemoveContainer" containerID="7411f369dc75ad75e553fd68df2151115121338637f010473f2d659c8276b45e" Oct 05 09:18:23 crc kubenswrapper[4846]: I1005 09:18:23.325420 4846 patch_prober.go:28] interesting pod/machine-config-daemon-fscvf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 09:18:23 crc kubenswrapper[4846]: I1005 09:18:23.326158 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 09:18:39 crc kubenswrapper[4846]: I1005 09:18:39.657610 4846 scope.go:117] "RemoveContainer" containerID="b8cea438c0f83ee3602db6928dc5e3b1535ad134c0c0a7e58a1d5330efc0ca1a" Oct 05 09:18:39 crc kubenswrapper[4846]: I1005 09:18:39.697041 4846 scope.go:117] "RemoveContainer" containerID="8aa526147ea1e8dfe1ec302b08dd8253fc6c9827ddc39aec47f5b806b11e95c8" Oct 05 09:18:39 crc kubenswrapper[4846]: I1005 09:18:39.777281 4846 scope.go:117] "RemoveContainer" containerID="610bb3bc9297a6e2905343a1d286aaf29bf49d23ff8206062e2b4b0ae4092f27" Oct 05 09:18:53 crc kubenswrapper[4846]: I1005 09:18:53.324873 4846 patch_prober.go:28] interesting pod/machine-config-daemon-fscvf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 09:18:53 crc kubenswrapper[4846]: I1005 09:18:53.325553 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 09:19:23 crc kubenswrapper[4846]: I1005 09:19:23.325031 4846 patch_prober.go:28] interesting pod/machine-config-daemon-fscvf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 09:19:23 crc kubenswrapper[4846]: I1005 09:19:23.325650 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 09:19:23 crc kubenswrapper[4846]: I1005 09:19:23.325708 4846 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" Oct 05 09:19:23 crc kubenswrapper[4846]: I1005 09:19:23.326489 4846 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"2ddf4e92e4eafd2c8e76240cc0766ba587ed44b22d7136c95e5159c1c5e8c32e"} pod="openshift-machine-config-operator/machine-config-daemon-fscvf" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 05 09:19:23 crc kubenswrapper[4846]: I1005 09:19:23.326575 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" containerName="machine-config-daemon" containerID="cri-o://2ddf4e92e4eafd2c8e76240cc0766ba587ed44b22d7136c95e5159c1c5e8c32e" gracePeriod=600 Oct 05 09:19:23 crc kubenswrapper[4846]: E1005 09:19:23.474118 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 09:19:23 crc kubenswrapper[4846]: I1005 09:19:23.489035 4846 generic.go:334] "Generic (PLEG): container finished" podID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" containerID="2ddf4e92e4eafd2c8e76240cc0766ba587ed44b22d7136c95e5159c1c5e8c32e" exitCode=0 Oct 05 09:19:23 crc kubenswrapper[4846]: I1005 09:19:23.489078 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" event={"ID":"ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7","Type":"ContainerDied","Data":"2ddf4e92e4eafd2c8e76240cc0766ba587ed44b22d7136c95e5159c1c5e8c32e"} Oct 05 09:19:23 crc kubenswrapper[4846]: I1005 09:19:23.489114 4846 scope.go:117] "RemoveContainer" containerID="206a9f8fee4cb0430eaa85104168af858cba45784f821a5ea04393d05e733060" Oct 05 09:19:23 crc kubenswrapper[4846]: I1005 09:19:23.490170 4846 scope.go:117] "RemoveContainer" containerID="2ddf4e92e4eafd2c8e76240cc0766ba587ed44b22d7136c95e5159c1c5e8c32e" Oct 05 09:19:23 crc kubenswrapper[4846]: E1005 09:19:23.490676 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 09:19:25 crc kubenswrapper[4846]: I1005 09:19:25.514418 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-rnh9j"] Oct 05 09:19:25 crc kubenswrapper[4846]: I1005 09:19:25.523678 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rnh9j" Oct 05 09:19:25 crc kubenswrapper[4846]: I1005 09:19:25.572609 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-rnh9j"] Oct 05 09:19:25 crc kubenswrapper[4846]: I1005 09:19:25.677436 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cf4ad9d5-3f48-47d0-867c-a1f0ffbd935f-utilities\") pod \"community-operators-rnh9j\" (UID: \"cf4ad9d5-3f48-47d0-867c-a1f0ffbd935f\") " pod="openshift-marketplace/community-operators-rnh9j" Oct 05 09:19:25 crc kubenswrapper[4846]: I1005 09:19:25.677800 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cf4ad9d5-3f48-47d0-867c-a1f0ffbd935f-catalog-content\") pod \"community-operators-rnh9j\" (UID: \"cf4ad9d5-3f48-47d0-867c-a1f0ffbd935f\") " pod="openshift-marketplace/community-operators-rnh9j" Oct 05 09:19:25 crc kubenswrapper[4846]: I1005 09:19:25.677924 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zvn7w\" (UniqueName: \"kubernetes.io/projected/cf4ad9d5-3f48-47d0-867c-a1f0ffbd935f-kube-api-access-zvn7w\") pod \"community-operators-rnh9j\" (UID: \"cf4ad9d5-3f48-47d0-867c-a1f0ffbd935f\") " pod="openshift-marketplace/community-operators-rnh9j" Oct 05 09:19:25 crc kubenswrapper[4846]: I1005 09:19:25.779977 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cf4ad9d5-3f48-47d0-867c-a1f0ffbd935f-catalog-content\") pod \"community-operators-rnh9j\" (UID: \"cf4ad9d5-3f48-47d0-867c-a1f0ffbd935f\") " pod="openshift-marketplace/community-operators-rnh9j" Oct 05 09:19:25 crc kubenswrapper[4846]: I1005 09:19:25.780063 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zvn7w\" (UniqueName: \"kubernetes.io/projected/cf4ad9d5-3f48-47d0-867c-a1f0ffbd935f-kube-api-access-zvn7w\") pod \"community-operators-rnh9j\" (UID: \"cf4ad9d5-3f48-47d0-867c-a1f0ffbd935f\") " pod="openshift-marketplace/community-operators-rnh9j" Oct 05 09:19:25 crc kubenswrapper[4846]: I1005 09:19:25.780181 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cf4ad9d5-3f48-47d0-867c-a1f0ffbd935f-utilities\") pod \"community-operators-rnh9j\" (UID: \"cf4ad9d5-3f48-47d0-867c-a1f0ffbd935f\") " pod="openshift-marketplace/community-operators-rnh9j" Oct 05 09:19:25 crc kubenswrapper[4846]: I1005 09:19:25.780709 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cf4ad9d5-3f48-47d0-867c-a1f0ffbd935f-catalog-content\") pod \"community-operators-rnh9j\" (UID: \"cf4ad9d5-3f48-47d0-867c-a1f0ffbd935f\") " pod="openshift-marketplace/community-operators-rnh9j" Oct 05 09:19:25 crc kubenswrapper[4846]: I1005 09:19:25.780773 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cf4ad9d5-3f48-47d0-867c-a1f0ffbd935f-utilities\") pod \"community-operators-rnh9j\" (UID: \"cf4ad9d5-3f48-47d0-867c-a1f0ffbd935f\") " pod="openshift-marketplace/community-operators-rnh9j" Oct 05 09:19:25 crc kubenswrapper[4846]: I1005 09:19:25.801757 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zvn7w\" (UniqueName: \"kubernetes.io/projected/cf4ad9d5-3f48-47d0-867c-a1f0ffbd935f-kube-api-access-zvn7w\") pod \"community-operators-rnh9j\" (UID: \"cf4ad9d5-3f48-47d0-867c-a1f0ffbd935f\") " pod="openshift-marketplace/community-operators-rnh9j" Oct 05 09:19:25 crc kubenswrapper[4846]: I1005 09:19:25.861762 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rnh9j" Oct 05 09:19:26 crc kubenswrapper[4846]: I1005 09:19:26.248870 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-rnh9j"] Oct 05 09:19:26 crc kubenswrapper[4846]: I1005 09:19:26.533497 4846 generic.go:334] "Generic (PLEG): container finished" podID="cf4ad9d5-3f48-47d0-867c-a1f0ffbd935f" containerID="0c1048b586ab5970be9cbf848fa829a549a44ae420003566b667d6632e386e43" exitCode=0 Oct 05 09:19:26 crc kubenswrapper[4846]: I1005 09:19:26.533769 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rnh9j" event={"ID":"cf4ad9d5-3f48-47d0-867c-a1f0ffbd935f","Type":"ContainerDied","Data":"0c1048b586ab5970be9cbf848fa829a549a44ae420003566b667d6632e386e43"} Oct 05 09:19:26 crc kubenswrapper[4846]: I1005 09:19:26.533797 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rnh9j" event={"ID":"cf4ad9d5-3f48-47d0-867c-a1f0ffbd935f","Type":"ContainerStarted","Data":"e22eb8d8baa0bfb84c71a6ca0eff94e554b41a7f18c1c0c3342bd10fd36f5246"} Oct 05 09:19:27 crc kubenswrapper[4846]: I1005 09:19:27.544648 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rnh9j" event={"ID":"cf4ad9d5-3f48-47d0-867c-a1f0ffbd935f","Type":"ContainerStarted","Data":"9c511031dddca44362770eea8de17c88c9a478909ef054269760523a824ee83e"} Oct 05 09:19:29 crc kubenswrapper[4846]: I1005 09:19:29.566725 4846 generic.go:334] "Generic (PLEG): container finished" podID="cf4ad9d5-3f48-47d0-867c-a1f0ffbd935f" containerID="9c511031dddca44362770eea8de17c88c9a478909ef054269760523a824ee83e" exitCode=0 Oct 05 09:19:29 crc kubenswrapper[4846]: I1005 09:19:29.566795 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rnh9j" event={"ID":"cf4ad9d5-3f48-47d0-867c-a1f0ffbd935f","Type":"ContainerDied","Data":"9c511031dddca44362770eea8de17c88c9a478909ef054269760523a824ee83e"} Oct 05 09:19:31 crc kubenswrapper[4846]: I1005 09:19:31.601110 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rnh9j" event={"ID":"cf4ad9d5-3f48-47d0-867c-a1f0ffbd935f","Type":"ContainerStarted","Data":"9742bbcbaa01f1c65e9a43dd00808e6e8c9d030e723970e5c56fff4c2f58b05c"} Oct 05 09:19:31 crc kubenswrapper[4846]: I1005 09:19:31.630220 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-rnh9j" podStartSLOduration=2.146518043 podStartE2EDuration="6.630175634s" podCreationTimestamp="2025-10-05 09:19:25 +0000 UTC" firstStartedPulling="2025-10-05 09:19:26.537216783 +0000 UTC m=+9088.778069548" lastFinishedPulling="2025-10-05 09:19:31.020874334 +0000 UTC m=+9093.261727139" observedRunningTime="2025-10-05 09:19:31.627971786 +0000 UTC m=+9093.868824591" watchObservedRunningTime="2025-10-05 09:19:31.630175634 +0000 UTC m=+9093.871028429" Oct 05 09:19:35 crc kubenswrapper[4846]: I1005 09:19:35.863384 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-rnh9j" Oct 05 09:19:35 crc kubenswrapper[4846]: I1005 09:19:35.863790 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-rnh9j" Oct 05 09:19:36 crc kubenswrapper[4846]: I1005 09:19:36.513401 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-b9k7t"] Oct 05 09:19:36 crc kubenswrapper[4846]: I1005 09:19:36.517869 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-b9k7t" Oct 05 09:19:36 crc kubenswrapper[4846]: I1005 09:19:36.521022 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-b9k7t"] Oct 05 09:19:36 crc kubenswrapper[4846]: I1005 09:19:36.610764 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-98nlb\" (UniqueName: \"kubernetes.io/projected/aec84391-61b6-4f99-87cc-723ed83d1626-kube-api-access-98nlb\") pod \"redhat-marketplace-b9k7t\" (UID: \"aec84391-61b6-4f99-87cc-723ed83d1626\") " pod="openshift-marketplace/redhat-marketplace-b9k7t" Oct 05 09:19:36 crc kubenswrapper[4846]: I1005 09:19:36.611309 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aec84391-61b6-4f99-87cc-723ed83d1626-catalog-content\") pod \"redhat-marketplace-b9k7t\" (UID: \"aec84391-61b6-4f99-87cc-723ed83d1626\") " pod="openshift-marketplace/redhat-marketplace-b9k7t" Oct 05 09:19:36 crc kubenswrapper[4846]: I1005 09:19:36.611567 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aec84391-61b6-4f99-87cc-723ed83d1626-utilities\") pod \"redhat-marketplace-b9k7t\" (UID: \"aec84391-61b6-4f99-87cc-723ed83d1626\") " pod="openshift-marketplace/redhat-marketplace-b9k7t" Oct 05 09:19:36 crc kubenswrapper[4846]: I1005 09:19:36.714484 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-98nlb\" (UniqueName: \"kubernetes.io/projected/aec84391-61b6-4f99-87cc-723ed83d1626-kube-api-access-98nlb\") pod \"redhat-marketplace-b9k7t\" (UID: \"aec84391-61b6-4f99-87cc-723ed83d1626\") " pod="openshift-marketplace/redhat-marketplace-b9k7t" Oct 05 09:19:36 crc kubenswrapper[4846]: I1005 09:19:36.714634 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aec84391-61b6-4f99-87cc-723ed83d1626-catalog-content\") pod \"redhat-marketplace-b9k7t\" (UID: \"aec84391-61b6-4f99-87cc-723ed83d1626\") " pod="openshift-marketplace/redhat-marketplace-b9k7t" Oct 05 09:19:36 crc kubenswrapper[4846]: I1005 09:19:36.714692 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aec84391-61b6-4f99-87cc-723ed83d1626-utilities\") pod \"redhat-marketplace-b9k7t\" (UID: \"aec84391-61b6-4f99-87cc-723ed83d1626\") " pod="openshift-marketplace/redhat-marketplace-b9k7t" Oct 05 09:19:36 crc kubenswrapper[4846]: I1005 09:19:36.715251 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aec84391-61b6-4f99-87cc-723ed83d1626-catalog-content\") pod \"redhat-marketplace-b9k7t\" (UID: \"aec84391-61b6-4f99-87cc-723ed83d1626\") " pod="openshift-marketplace/redhat-marketplace-b9k7t" Oct 05 09:19:36 crc kubenswrapper[4846]: I1005 09:19:36.715378 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aec84391-61b6-4f99-87cc-723ed83d1626-utilities\") pod \"redhat-marketplace-b9k7t\" (UID: \"aec84391-61b6-4f99-87cc-723ed83d1626\") " pod="openshift-marketplace/redhat-marketplace-b9k7t" Oct 05 09:19:36 crc kubenswrapper[4846]: I1005 09:19:36.734634 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-98nlb\" (UniqueName: \"kubernetes.io/projected/aec84391-61b6-4f99-87cc-723ed83d1626-kube-api-access-98nlb\") pod \"redhat-marketplace-b9k7t\" (UID: \"aec84391-61b6-4f99-87cc-723ed83d1626\") " pod="openshift-marketplace/redhat-marketplace-b9k7t" Oct 05 09:19:36 crc kubenswrapper[4846]: I1005 09:19:36.856606 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-b9k7t" Oct 05 09:19:36 crc kubenswrapper[4846]: I1005 09:19:36.928761 4846 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-rnh9j" podUID="cf4ad9d5-3f48-47d0-867c-a1f0ffbd935f" containerName="registry-server" probeResult="failure" output=< Oct 05 09:19:36 crc kubenswrapper[4846]: timeout: failed to connect service ":50051" within 1s Oct 05 09:19:36 crc kubenswrapper[4846]: > Oct 05 09:19:37 crc kubenswrapper[4846]: I1005 09:19:37.350604 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-b9k7t"] Oct 05 09:19:37 crc kubenswrapper[4846]: I1005 09:19:37.498639 4846 scope.go:117] "RemoveContainer" containerID="2ddf4e92e4eafd2c8e76240cc0766ba587ed44b22d7136c95e5159c1c5e8c32e" Oct 05 09:19:37 crc kubenswrapper[4846]: E1005 09:19:37.499357 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 09:19:37 crc kubenswrapper[4846]: I1005 09:19:37.658071 4846 generic.go:334] "Generic (PLEG): container finished" podID="aec84391-61b6-4f99-87cc-723ed83d1626" containerID="63be0a54a6a1758e1e5bd71334414105c325f82b3cc45f854e0c3aff65ab31cb" exitCode=0 Oct 05 09:19:37 crc kubenswrapper[4846]: I1005 09:19:37.658118 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-b9k7t" event={"ID":"aec84391-61b6-4f99-87cc-723ed83d1626","Type":"ContainerDied","Data":"63be0a54a6a1758e1e5bd71334414105c325f82b3cc45f854e0c3aff65ab31cb"} Oct 05 09:19:37 crc kubenswrapper[4846]: I1005 09:19:37.658148 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-b9k7t" event={"ID":"aec84391-61b6-4f99-87cc-723ed83d1626","Type":"ContainerStarted","Data":"c4698f87fb583fd22135d67c3d08c42f7790933cce93aa24babd8cb682b3a8ba"} Oct 05 09:19:38 crc kubenswrapper[4846]: I1005 09:19:38.671695 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-b9k7t" event={"ID":"aec84391-61b6-4f99-87cc-723ed83d1626","Type":"ContainerStarted","Data":"a22aa83760413f05b5cc8ff11761535de907053350fdf93d15fe9be94233cc63"} Oct 05 09:19:39 crc kubenswrapper[4846]: I1005 09:19:39.688902 4846 generic.go:334] "Generic (PLEG): container finished" podID="aec84391-61b6-4f99-87cc-723ed83d1626" containerID="a22aa83760413f05b5cc8ff11761535de907053350fdf93d15fe9be94233cc63" exitCode=0 Oct 05 09:19:39 crc kubenswrapper[4846]: I1005 09:19:39.689479 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-b9k7t" event={"ID":"aec84391-61b6-4f99-87cc-723ed83d1626","Type":"ContainerDied","Data":"a22aa83760413f05b5cc8ff11761535de907053350fdf93d15fe9be94233cc63"} Oct 05 09:19:40 crc kubenswrapper[4846]: I1005 09:19:40.705129 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-b9k7t" event={"ID":"aec84391-61b6-4f99-87cc-723ed83d1626","Type":"ContainerStarted","Data":"54792b0b9e2f53a6566189ea3b286a204f4e9fc605021ed11336541a15ac8460"} Oct 05 09:19:40 crc kubenswrapper[4846]: I1005 09:19:40.732366 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-b9k7t" podStartSLOduration=2.268546412 podStartE2EDuration="4.732344686s" podCreationTimestamp="2025-10-05 09:19:36 +0000 UTC" firstStartedPulling="2025-10-05 09:19:37.659746926 +0000 UTC m=+9099.900599701" lastFinishedPulling="2025-10-05 09:19:40.12354517 +0000 UTC m=+9102.364397975" observedRunningTime="2025-10-05 09:19:40.727005107 +0000 UTC m=+9102.967857922" watchObservedRunningTime="2025-10-05 09:19:40.732344686 +0000 UTC m=+9102.973197461" Oct 05 09:19:45 crc kubenswrapper[4846]: I1005 09:19:45.941086 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-rnh9j" Oct 05 09:19:46 crc kubenswrapper[4846]: I1005 09:19:46.010703 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-rnh9j" Oct 05 09:19:46 crc kubenswrapper[4846]: I1005 09:19:46.181806 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-rnh9j"] Oct 05 09:19:46 crc kubenswrapper[4846]: I1005 09:19:46.857083 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-b9k7t" Oct 05 09:19:46 crc kubenswrapper[4846]: I1005 09:19:46.857140 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-b9k7t" Oct 05 09:19:46 crc kubenswrapper[4846]: I1005 09:19:46.925382 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-b9k7t" Oct 05 09:19:47 crc kubenswrapper[4846]: I1005 09:19:47.859228 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-rnh9j" podUID="cf4ad9d5-3f48-47d0-867c-a1f0ffbd935f" containerName="registry-server" containerID="cri-o://9742bbcbaa01f1c65e9a43dd00808e6e8c9d030e723970e5c56fff4c2f58b05c" gracePeriod=2 Oct 05 09:19:47 crc kubenswrapper[4846]: I1005 09:19:47.928869 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-b9k7t" Oct 05 09:19:48 crc kubenswrapper[4846]: I1005 09:19:48.358509 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rnh9j" Oct 05 09:19:48 crc kubenswrapper[4846]: I1005 09:19:48.505653 4846 scope.go:117] "RemoveContainer" containerID="2ddf4e92e4eafd2c8e76240cc0766ba587ed44b22d7136c95e5159c1c5e8c32e" Oct 05 09:19:48 crc kubenswrapper[4846]: E1005 09:19:48.506236 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 09:19:48 crc kubenswrapper[4846]: I1005 09:19:48.527878 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zvn7w\" (UniqueName: \"kubernetes.io/projected/cf4ad9d5-3f48-47d0-867c-a1f0ffbd935f-kube-api-access-zvn7w\") pod \"cf4ad9d5-3f48-47d0-867c-a1f0ffbd935f\" (UID: \"cf4ad9d5-3f48-47d0-867c-a1f0ffbd935f\") " Oct 05 09:19:48 crc kubenswrapper[4846]: I1005 09:19:48.528059 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cf4ad9d5-3f48-47d0-867c-a1f0ffbd935f-utilities\") pod \"cf4ad9d5-3f48-47d0-867c-a1f0ffbd935f\" (UID: \"cf4ad9d5-3f48-47d0-867c-a1f0ffbd935f\") " Oct 05 09:19:48 crc kubenswrapper[4846]: I1005 09:19:48.528221 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cf4ad9d5-3f48-47d0-867c-a1f0ffbd935f-catalog-content\") pod \"cf4ad9d5-3f48-47d0-867c-a1f0ffbd935f\" (UID: \"cf4ad9d5-3f48-47d0-867c-a1f0ffbd935f\") " Oct 05 09:19:48 crc kubenswrapper[4846]: I1005 09:19:48.529855 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cf4ad9d5-3f48-47d0-867c-a1f0ffbd935f-utilities" (OuterVolumeSpecName: "utilities") pod "cf4ad9d5-3f48-47d0-867c-a1f0ffbd935f" (UID: "cf4ad9d5-3f48-47d0-867c-a1f0ffbd935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 09:19:48 crc kubenswrapper[4846]: I1005 09:19:48.537267 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cf4ad9d5-3f48-47d0-867c-a1f0ffbd935f-kube-api-access-zvn7w" (OuterVolumeSpecName: "kube-api-access-zvn7w") pod "cf4ad9d5-3f48-47d0-867c-a1f0ffbd935f" (UID: "cf4ad9d5-3f48-47d0-867c-a1f0ffbd935f"). InnerVolumeSpecName "kube-api-access-zvn7w". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 09:19:48 crc kubenswrapper[4846]: I1005 09:19:48.591629 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-b9k7t"] Oct 05 09:19:48 crc kubenswrapper[4846]: I1005 09:19:48.601967 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cf4ad9d5-3f48-47d0-867c-a1f0ffbd935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "cf4ad9d5-3f48-47d0-867c-a1f0ffbd935f" (UID: "cf4ad9d5-3f48-47d0-867c-a1f0ffbd935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 09:19:48 crc kubenswrapper[4846]: I1005 09:19:48.632339 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zvn7w\" (UniqueName: \"kubernetes.io/projected/cf4ad9d5-3f48-47d0-867c-a1f0ffbd935f-kube-api-access-zvn7w\") on node \"crc\" DevicePath \"\"" Oct 05 09:19:48 crc kubenswrapper[4846]: I1005 09:19:48.632406 4846 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cf4ad9d5-3f48-47d0-867c-a1f0ffbd935f-utilities\") on node \"crc\" DevicePath \"\"" Oct 05 09:19:48 crc kubenswrapper[4846]: I1005 09:19:48.632424 4846 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cf4ad9d5-3f48-47d0-867c-a1f0ffbd935f-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 05 09:19:48 crc kubenswrapper[4846]: I1005 09:19:48.869647 4846 generic.go:334] "Generic (PLEG): container finished" podID="cf4ad9d5-3f48-47d0-867c-a1f0ffbd935f" containerID="9742bbcbaa01f1c65e9a43dd00808e6e8c9d030e723970e5c56fff4c2f58b05c" exitCode=0 Oct 05 09:19:48 crc kubenswrapper[4846]: I1005 09:19:48.869685 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rnh9j" event={"ID":"cf4ad9d5-3f48-47d0-867c-a1f0ffbd935f","Type":"ContainerDied","Data":"9742bbcbaa01f1c65e9a43dd00808e6e8c9d030e723970e5c56fff4c2f58b05c"} Oct 05 09:19:48 crc kubenswrapper[4846]: I1005 09:19:48.869725 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rnh9j" Oct 05 09:19:48 crc kubenswrapper[4846]: I1005 09:19:48.869755 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rnh9j" event={"ID":"cf4ad9d5-3f48-47d0-867c-a1f0ffbd935f","Type":"ContainerDied","Data":"e22eb8d8baa0bfb84c71a6ca0eff94e554b41a7f18c1c0c3342bd10fd36f5246"} Oct 05 09:19:48 crc kubenswrapper[4846]: I1005 09:19:48.869778 4846 scope.go:117] "RemoveContainer" containerID="9742bbcbaa01f1c65e9a43dd00808e6e8c9d030e723970e5c56fff4c2f58b05c" Oct 05 09:19:48 crc kubenswrapper[4846]: I1005 09:19:48.908452 4846 scope.go:117] "RemoveContainer" containerID="9c511031dddca44362770eea8de17c88c9a478909ef054269760523a824ee83e" Oct 05 09:19:48 crc kubenswrapper[4846]: I1005 09:19:48.913865 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-rnh9j"] Oct 05 09:19:48 crc kubenswrapper[4846]: I1005 09:19:48.924357 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-rnh9j"] Oct 05 09:19:48 crc kubenswrapper[4846]: I1005 09:19:48.951127 4846 scope.go:117] "RemoveContainer" containerID="0c1048b586ab5970be9cbf848fa829a549a44ae420003566b667d6632e386e43" Oct 05 09:19:49 crc kubenswrapper[4846]: I1005 09:19:49.002530 4846 scope.go:117] "RemoveContainer" containerID="9742bbcbaa01f1c65e9a43dd00808e6e8c9d030e723970e5c56fff4c2f58b05c" Oct 05 09:19:49 crc kubenswrapper[4846]: E1005 09:19:49.003129 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9742bbcbaa01f1c65e9a43dd00808e6e8c9d030e723970e5c56fff4c2f58b05c\": container with ID starting with 9742bbcbaa01f1c65e9a43dd00808e6e8c9d030e723970e5c56fff4c2f58b05c not found: ID does not exist" containerID="9742bbcbaa01f1c65e9a43dd00808e6e8c9d030e723970e5c56fff4c2f58b05c" Oct 05 09:19:49 crc kubenswrapper[4846]: I1005 09:19:49.003162 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9742bbcbaa01f1c65e9a43dd00808e6e8c9d030e723970e5c56fff4c2f58b05c"} err="failed to get container status \"9742bbcbaa01f1c65e9a43dd00808e6e8c9d030e723970e5c56fff4c2f58b05c\": rpc error: code = NotFound desc = could not find container \"9742bbcbaa01f1c65e9a43dd00808e6e8c9d030e723970e5c56fff4c2f58b05c\": container with ID starting with 9742bbcbaa01f1c65e9a43dd00808e6e8c9d030e723970e5c56fff4c2f58b05c not found: ID does not exist" Oct 05 09:19:49 crc kubenswrapper[4846]: I1005 09:19:49.003192 4846 scope.go:117] "RemoveContainer" containerID="9c511031dddca44362770eea8de17c88c9a478909ef054269760523a824ee83e" Oct 05 09:19:49 crc kubenswrapper[4846]: E1005 09:19:49.003649 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9c511031dddca44362770eea8de17c88c9a478909ef054269760523a824ee83e\": container with ID starting with 9c511031dddca44362770eea8de17c88c9a478909ef054269760523a824ee83e not found: ID does not exist" containerID="9c511031dddca44362770eea8de17c88c9a478909ef054269760523a824ee83e" Oct 05 09:19:49 crc kubenswrapper[4846]: I1005 09:19:49.003676 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9c511031dddca44362770eea8de17c88c9a478909ef054269760523a824ee83e"} err="failed to get container status \"9c511031dddca44362770eea8de17c88c9a478909ef054269760523a824ee83e\": rpc error: code = NotFound desc = could not find container \"9c511031dddca44362770eea8de17c88c9a478909ef054269760523a824ee83e\": container with ID starting with 9c511031dddca44362770eea8de17c88c9a478909ef054269760523a824ee83e not found: ID does not exist" Oct 05 09:19:49 crc kubenswrapper[4846]: I1005 09:19:49.003689 4846 scope.go:117] "RemoveContainer" containerID="0c1048b586ab5970be9cbf848fa829a549a44ae420003566b667d6632e386e43" Oct 05 09:19:49 crc kubenswrapper[4846]: E1005 09:19:49.004090 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0c1048b586ab5970be9cbf848fa829a549a44ae420003566b667d6632e386e43\": container with ID starting with 0c1048b586ab5970be9cbf848fa829a549a44ae420003566b667d6632e386e43 not found: ID does not exist" containerID="0c1048b586ab5970be9cbf848fa829a549a44ae420003566b667d6632e386e43" Oct 05 09:19:49 crc kubenswrapper[4846]: I1005 09:19:49.004114 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0c1048b586ab5970be9cbf848fa829a549a44ae420003566b667d6632e386e43"} err="failed to get container status \"0c1048b586ab5970be9cbf848fa829a549a44ae420003566b667d6632e386e43\": rpc error: code = NotFound desc = could not find container \"0c1048b586ab5970be9cbf848fa829a549a44ae420003566b667d6632e386e43\": container with ID starting with 0c1048b586ab5970be9cbf848fa829a549a44ae420003566b667d6632e386e43 not found: ID does not exist" Oct 05 09:19:49 crc kubenswrapper[4846]: I1005 09:19:49.883263 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-b9k7t" podUID="aec84391-61b6-4f99-87cc-723ed83d1626" containerName="registry-server" containerID="cri-o://54792b0b9e2f53a6566189ea3b286a204f4e9fc605021ed11336541a15ac8460" gracePeriod=2 Oct 05 09:19:50 crc kubenswrapper[4846]: I1005 09:19:50.521427 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cf4ad9d5-3f48-47d0-867c-a1f0ffbd935f" path="/var/lib/kubelet/pods/cf4ad9d5-3f48-47d0-867c-a1f0ffbd935f/volumes" Oct 05 09:19:50 crc kubenswrapper[4846]: I1005 09:19:50.897292 4846 generic.go:334] "Generic (PLEG): container finished" podID="aec84391-61b6-4f99-87cc-723ed83d1626" containerID="54792b0b9e2f53a6566189ea3b286a204f4e9fc605021ed11336541a15ac8460" exitCode=0 Oct 05 09:19:50 crc kubenswrapper[4846]: I1005 09:19:50.897342 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-b9k7t" event={"ID":"aec84391-61b6-4f99-87cc-723ed83d1626","Type":"ContainerDied","Data":"54792b0b9e2f53a6566189ea3b286a204f4e9fc605021ed11336541a15ac8460"} Oct 05 09:19:50 crc kubenswrapper[4846]: I1005 09:19:50.897374 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-b9k7t" event={"ID":"aec84391-61b6-4f99-87cc-723ed83d1626","Type":"ContainerDied","Data":"c4698f87fb583fd22135d67c3d08c42f7790933cce93aa24babd8cb682b3a8ba"} Oct 05 09:19:50 crc kubenswrapper[4846]: I1005 09:19:50.897390 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c4698f87fb583fd22135d67c3d08c42f7790933cce93aa24babd8cb682b3a8ba" Oct 05 09:19:51 crc kubenswrapper[4846]: I1005 09:19:51.497594 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-b9k7t" Oct 05 09:19:51 crc kubenswrapper[4846]: I1005 09:19:51.603391 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aec84391-61b6-4f99-87cc-723ed83d1626-utilities\") pod \"aec84391-61b6-4f99-87cc-723ed83d1626\" (UID: \"aec84391-61b6-4f99-87cc-723ed83d1626\") " Oct 05 09:19:51 crc kubenswrapper[4846]: I1005 09:19:51.603628 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-98nlb\" (UniqueName: \"kubernetes.io/projected/aec84391-61b6-4f99-87cc-723ed83d1626-kube-api-access-98nlb\") pod \"aec84391-61b6-4f99-87cc-723ed83d1626\" (UID: \"aec84391-61b6-4f99-87cc-723ed83d1626\") " Oct 05 09:19:51 crc kubenswrapper[4846]: I1005 09:19:51.603651 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aec84391-61b6-4f99-87cc-723ed83d1626-catalog-content\") pod \"aec84391-61b6-4f99-87cc-723ed83d1626\" (UID: \"aec84391-61b6-4f99-87cc-723ed83d1626\") " Oct 05 09:19:51 crc kubenswrapper[4846]: I1005 09:19:51.604561 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/aec84391-61b6-4f99-87cc-723ed83d1626-utilities" (OuterVolumeSpecName: "utilities") pod "aec84391-61b6-4f99-87cc-723ed83d1626" (UID: "aec84391-61b6-4f99-87cc-723ed83d1626"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 09:19:51 crc kubenswrapper[4846]: I1005 09:19:51.610532 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aec84391-61b6-4f99-87cc-723ed83d1626-kube-api-access-98nlb" (OuterVolumeSpecName: "kube-api-access-98nlb") pod "aec84391-61b6-4f99-87cc-723ed83d1626" (UID: "aec84391-61b6-4f99-87cc-723ed83d1626"). InnerVolumeSpecName "kube-api-access-98nlb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 09:19:51 crc kubenswrapper[4846]: I1005 09:19:51.632904 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/aec84391-61b6-4f99-87cc-723ed83d1626-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "aec84391-61b6-4f99-87cc-723ed83d1626" (UID: "aec84391-61b6-4f99-87cc-723ed83d1626"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 09:19:51 crc kubenswrapper[4846]: I1005 09:19:51.706334 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-98nlb\" (UniqueName: \"kubernetes.io/projected/aec84391-61b6-4f99-87cc-723ed83d1626-kube-api-access-98nlb\") on node \"crc\" DevicePath \"\"" Oct 05 09:19:51 crc kubenswrapper[4846]: I1005 09:19:51.706423 4846 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aec84391-61b6-4f99-87cc-723ed83d1626-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 05 09:19:51 crc kubenswrapper[4846]: I1005 09:19:51.706444 4846 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aec84391-61b6-4f99-87cc-723ed83d1626-utilities\") on node \"crc\" DevicePath \"\"" Oct 05 09:19:51 crc kubenswrapper[4846]: I1005 09:19:51.914589 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-b9k7t" Oct 05 09:19:51 crc kubenswrapper[4846]: I1005 09:19:51.979203 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-b9k7t"] Oct 05 09:19:51 crc kubenswrapper[4846]: I1005 09:19:51.991373 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-b9k7t"] Oct 05 09:19:52 crc kubenswrapper[4846]: I1005 09:19:52.510526 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aec84391-61b6-4f99-87cc-723ed83d1626" path="/var/lib/kubelet/pods/aec84391-61b6-4f99-87cc-723ed83d1626/volumes" Oct 05 09:20:00 crc kubenswrapper[4846]: I1005 09:20:00.498450 4846 scope.go:117] "RemoveContainer" containerID="2ddf4e92e4eafd2c8e76240cc0766ba587ed44b22d7136c95e5159c1c5e8c32e" Oct 05 09:20:00 crc kubenswrapper[4846]: E1005 09:20:00.499466 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 09:20:15 crc kubenswrapper[4846]: I1005 09:20:15.498926 4846 scope.go:117] "RemoveContainer" containerID="2ddf4e92e4eafd2c8e76240cc0766ba587ed44b22d7136c95e5159c1c5e8c32e" Oct 05 09:20:15 crc kubenswrapper[4846]: E1005 09:20:15.499547 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 09:20:26 crc kubenswrapper[4846]: I1005 09:20:26.499535 4846 scope.go:117] "RemoveContainer" containerID="2ddf4e92e4eafd2c8e76240cc0766ba587ed44b22d7136c95e5159c1c5e8c32e" Oct 05 09:20:26 crc kubenswrapper[4846]: E1005 09:20:26.500573 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 09:20:39 crc kubenswrapper[4846]: I1005 09:20:39.499094 4846 scope.go:117] "RemoveContainer" containerID="2ddf4e92e4eafd2c8e76240cc0766ba587ed44b22d7136c95e5159c1c5e8c32e" Oct 05 09:20:39 crc kubenswrapper[4846]: E1005 09:20:39.500056 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 09:20:50 crc kubenswrapper[4846]: I1005 09:20:50.499149 4846 scope.go:117] "RemoveContainer" containerID="2ddf4e92e4eafd2c8e76240cc0766ba587ed44b22d7136c95e5159c1c5e8c32e" Oct 05 09:20:50 crc kubenswrapper[4846]: E1005 09:20:50.500325 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 09:21:04 crc kubenswrapper[4846]: I1005 09:21:04.498277 4846 scope.go:117] "RemoveContainer" containerID="2ddf4e92e4eafd2c8e76240cc0766ba587ed44b22d7136c95e5159c1c5e8c32e" Oct 05 09:21:04 crc kubenswrapper[4846]: E1005 09:21:04.499351 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 09:21:16 crc kubenswrapper[4846]: I1005 09:21:16.498493 4846 scope.go:117] "RemoveContainer" containerID="2ddf4e92e4eafd2c8e76240cc0766ba587ed44b22d7136c95e5159c1c5e8c32e" Oct 05 09:21:16 crc kubenswrapper[4846]: E1005 09:21:16.499764 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 09:21:27 crc kubenswrapper[4846]: I1005 09:21:27.497166 4846 scope.go:117] "RemoveContainer" containerID="2ddf4e92e4eafd2c8e76240cc0766ba587ed44b22d7136c95e5159c1c5e8c32e" Oct 05 09:21:27 crc kubenswrapper[4846]: E1005 09:21:27.497946 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 09:21:40 crc kubenswrapper[4846]: I1005 09:21:40.498436 4846 scope.go:117] "RemoveContainer" containerID="2ddf4e92e4eafd2c8e76240cc0766ba587ed44b22d7136c95e5159c1c5e8c32e" Oct 05 09:21:40 crc kubenswrapper[4846]: E1005 09:21:40.499308 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 09:21:54 crc kubenswrapper[4846]: I1005 09:21:54.498237 4846 scope.go:117] "RemoveContainer" containerID="2ddf4e92e4eafd2c8e76240cc0766ba587ed44b22d7136c95e5159c1c5e8c32e" Oct 05 09:21:54 crc kubenswrapper[4846]: E1005 09:21:54.499398 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 09:22:05 crc kubenswrapper[4846]: I1005 09:22:05.498495 4846 scope.go:117] "RemoveContainer" containerID="2ddf4e92e4eafd2c8e76240cc0766ba587ed44b22d7136c95e5159c1c5e8c32e" Oct 05 09:22:05 crc kubenswrapper[4846]: E1005 09:22:05.499763 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 09:22:20 crc kubenswrapper[4846]: I1005 09:22:20.499682 4846 scope.go:117] "RemoveContainer" containerID="2ddf4e92e4eafd2c8e76240cc0766ba587ed44b22d7136c95e5159c1c5e8c32e" Oct 05 09:22:20 crc kubenswrapper[4846]: E1005 09:22:20.500855 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 09:22:31 crc kubenswrapper[4846]: I1005 09:22:31.498375 4846 scope.go:117] "RemoveContainer" containerID="2ddf4e92e4eafd2c8e76240cc0766ba587ed44b22d7136c95e5159c1c5e8c32e" Oct 05 09:22:31 crc kubenswrapper[4846]: E1005 09:22:31.501592 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 09:22:45 crc kubenswrapper[4846]: I1005 09:22:45.497665 4846 scope.go:117] "RemoveContainer" containerID="2ddf4e92e4eafd2c8e76240cc0766ba587ed44b22d7136c95e5159c1c5e8c32e" Oct 05 09:22:45 crc kubenswrapper[4846]: E1005 09:22:45.498571 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 09:22:59 crc kubenswrapper[4846]: I1005 09:22:59.498453 4846 scope.go:117] "RemoveContainer" containerID="2ddf4e92e4eafd2c8e76240cc0766ba587ed44b22d7136c95e5159c1c5e8c32e" Oct 05 09:22:59 crc kubenswrapper[4846]: E1005 09:22:59.500890 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 09:23:13 crc kubenswrapper[4846]: I1005 09:23:13.498103 4846 scope.go:117] "RemoveContainer" containerID="2ddf4e92e4eafd2c8e76240cc0766ba587ed44b22d7136c95e5159c1c5e8c32e" Oct 05 09:23:13 crc kubenswrapper[4846]: E1005 09:23:13.500390 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 09:23:15 crc kubenswrapper[4846]: I1005 09:23:15.991686 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-pprrt"] Oct 05 09:23:15 crc kubenswrapper[4846]: E1005 09:23:15.992742 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aec84391-61b6-4f99-87cc-723ed83d1626" containerName="extract-content" Oct 05 09:23:15 crc kubenswrapper[4846]: I1005 09:23:15.993063 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="aec84391-61b6-4f99-87cc-723ed83d1626" containerName="extract-content" Oct 05 09:23:15 crc kubenswrapper[4846]: E1005 09:23:15.993082 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cf4ad9d5-3f48-47d0-867c-a1f0ffbd935f" containerName="extract-content" Oct 05 09:23:15 crc kubenswrapper[4846]: I1005 09:23:15.993094 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="cf4ad9d5-3f48-47d0-867c-a1f0ffbd935f" containerName="extract-content" Oct 05 09:23:15 crc kubenswrapper[4846]: E1005 09:23:15.993127 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cf4ad9d5-3f48-47d0-867c-a1f0ffbd935f" containerName="registry-server" Oct 05 09:23:15 crc kubenswrapper[4846]: I1005 09:23:15.993139 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="cf4ad9d5-3f48-47d0-867c-a1f0ffbd935f" containerName="registry-server" Oct 05 09:23:15 crc kubenswrapper[4846]: E1005 09:23:15.993173 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aec84391-61b6-4f99-87cc-723ed83d1626" containerName="registry-server" Oct 05 09:23:15 crc kubenswrapper[4846]: I1005 09:23:15.993218 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="aec84391-61b6-4f99-87cc-723ed83d1626" containerName="registry-server" Oct 05 09:23:15 crc kubenswrapper[4846]: E1005 09:23:15.993247 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cf4ad9d5-3f48-47d0-867c-a1f0ffbd935f" containerName="extract-utilities" Oct 05 09:23:15 crc kubenswrapper[4846]: I1005 09:23:15.993257 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="cf4ad9d5-3f48-47d0-867c-a1f0ffbd935f" containerName="extract-utilities" Oct 05 09:23:15 crc kubenswrapper[4846]: E1005 09:23:15.993278 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aec84391-61b6-4f99-87cc-723ed83d1626" containerName="extract-utilities" Oct 05 09:23:15 crc kubenswrapper[4846]: I1005 09:23:15.993289 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="aec84391-61b6-4f99-87cc-723ed83d1626" containerName="extract-utilities" Oct 05 09:23:15 crc kubenswrapper[4846]: I1005 09:23:15.993654 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="cf4ad9d5-3f48-47d0-867c-a1f0ffbd935f" containerName="registry-server" Oct 05 09:23:15 crc kubenswrapper[4846]: I1005 09:23:15.993685 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="aec84391-61b6-4f99-87cc-723ed83d1626" containerName="registry-server" Oct 05 09:23:15 crc kubenswrapper[4846]: I1005 09:23:15.995626 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-pprrt" Oct 05 09:23:16 crc kubenswrapper[4846]: I1005 09:23:16.007505 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-pprrt"] Oct 05 09:23:16 crc kubenswrapper[4846]: I1005 09:23:16.032936 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f724ea50-bee0-4a90-a0f3-def9dbd56b0a-utilities\") pod \"redhat-operators-pprrt\" (UID: \"f724ea50-bee0-4a90-a0f3-def9dbd56b0a\") " pod="openshift-marketplace/redhat-operators-pprrt" Oct 05 09:23:16 crc kubenswrapper[4846]: I1005 09:23:16.033527 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f724ea50-bee0-4a90-a0f3-def9dbd56b0a-catalog-content\") pod \"redhat-operators-pprrt\" (UID: \"f724ea50-bee0-4a90-a0f3-def9dbd56b0a\") " pod="openshift-marketplace/redhat-operators-pprrt" Oct 05 09:23:16 crc kubenswrapper[4846]: I1005 09:23:16.033595 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xmpjb\" (UniqueName: \"kubernetes.io/projected/f724ea50-bee0-4a90-a0f3-def9dbd56b0a-kube-api-access-xmpjb\") pod \"redhat-operators-pprrt\" (UID: \"f724ea50-bee0-4a90-a0f3-def9dbd56b0a\") " pod="openshift-marketplace/redhat-operators-pprrt" Oct 05 09:23:16 crc kubenswrapper[4846]: I1005 09:23:16.136255 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f724ea50-bee0-4a90-a0f3-def9dbd56b0a-catalog-content\") pod \"redhat-operators-pprrt\" (UID: \"f724ea50-bee0-4a90-a0f3-def9dbd56b0a\") " pod="openshift-marketplace/redhat-operators-pprrt" Oct 05 09:23:16 crc kubenswrapper[4846]: I1005 09:23:16.136319 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xmpjb\" (UniqueName: \"kubernetes.io/projected/f724ea50-bee0-4a90-a0f3-def9dbd56b0a-kube-api-access-xmpjb\") pod \"redhat-operators-pprrt\" (UID: \"f724ea50-bee0-4a90-a0f3-def9dbd56b0a\") " pod="openshift-marketplace/redhat-operators-pprrt" Oct 05 09:23:16 crc kubenswrapper[4846]: I1005 09:23:16.136374 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f724ea50-bee0-4a90-a0f3-def9dbd56b0a-utilities\") pod \"redhat-operators-pprrt\" (UID: \"f724ea50-bee0-4a90-a0f3-def9dbd56b0a\") " pod="openshift-marketplace/redhat-operators-pprrt" Oct 05 09:23:16 crc kubenswrapper[4846]: I1005 09:23:16.136826 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f724ea50-bee0-4a90-a0f3-def9dbd56b0a-catalog-content\") pod \"redhat-operators-pprrt\" (UID: \"f724ea50-bee0-4a90-a0f3-def9dbd56b0a\") " pod="openshift-marketplace/redhat-operators-pprrt" Oct 05 09:23:16 crc kubenswrapper[4846]: I1005 09:23:16.136923 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f724ea50-bee0-4a90-a0f3-def9dbd56b0a-utilities\") pod \"redhat-operators-pprrt\" (UID: \"f724ea50-bee0-4a90-a0f3-def9dbd56b0a\") " pod="openshift-marketplace/redhat-operators-pprrt" Oct 05 09:23:16 crc kubenswrapper[4846]: I1005 09:23:16.160058 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xmpjb\" (UniqueName: \"kubernetes.io/projected/f724ea50-bee0-4a90-a0f3-def9dbd56b0a-kube-api-access-xmpjb\") pod \"redhat-operators-pprrt\" (UID: \"f724ea50-bee0-4a90-a0f3-def9dbd56b0a\") " pod="openshift-marketplace/redhat-operators-pprrt" Oct 05 09:23:16 crc kubenswrapper[4846]: I1005 09:23:16.330615 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-pprrt" Oct 05 09:23:16 crc kubenswrapper[4846]: I1005 09:23:16.852836 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-pprrt"] Oct 05 09:23:17 crc kubenswrapper[4846]: I1005 09:23:17.377955 4846 generic.go:334] "Generic (PLEG): container finished" podID="f724ea50-bee0-4a90-a0f3-def9dbd56b0a" containerID="8ca55153530905c32876af2cab5c26abe18fbe153136ba0470926236f7c3c6fa" exitCode=0 Oct 05 09:23:17 crc kubenswrapper[4846]: I1005 09:23:17.378053 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pprrt" event={"ID":"f724ea50-bee0-4a90-a0f3-def9dbd56b0a","Type":"ContainerDied","Data":"8ca55153530905c32876af2cab5c26abe18fbe153136ba0470926236f7c3c6fa"} Oct 05 09:23:17 crc kubenswrapper[4846]: I1005 09:23:17.378457 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pprrt" event={"ID":"f724ea50-bee0-4a90-a0f3-def9dbd56b0a","Type":"ContainerStarted","Data":"cd0ea672880c96895116ffe009480f1a64ae07266d46d982e446a965f7330db6"} Oct 05 09:23:17 crc kubenswrapper[4846]: I1005 09:23:17.380591 4846 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 05 09:23:18 crc kubenswrapper[4846]: I1005 09:23:18.388593 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pprrt" event={"ID":"f724ea50-bee0-4a90-a0f3-def9dbd56b0a","Type":"ContainerStarted","Data":"9962bddd6e832f0e1055e1e0d7236e065de89bf0216c30e62d7eb52f9849a87a"} Oct 05 09:23:22 crc kubenswrapper[4846]: I1005 09:23:22.442574 4846 generic.go:334] "Generic (PLEG): container finished" podID="f724ea50-bee0-4a90-a0f3-def9dbd56b0a" containerID="9962bddd6e832f0e1055e1e0d7236e065de89bf0216c30e62d7eb52f9849a87a" exitCode=0 Oct 05 09:23:22 crc kubenswrapper[4846]: I1005 09:23:22.442643 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pprrt" event={"ID":"f724ea50-bee0-4a90-a0f3-def9dbd56b0a","Type":"ContainerDied","Data":"9962bddd6e832f0e1055e1e0d7236e065de89bf0216c30e62d7eb52f9849a87a"} Oct 05 09:23:23 crc kubenswrapper[4846]: I1005 09:23:23.462658 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pprrt" event={"ID":"f724ea50-bee0-4a90-a0f3-def9dbd56b0a","Type":"ContainerStarted","Data":"5726ba891cba54ccc0ba32ee24fbd7218f9131ef25b7ffddf0a8d8cf53df5e00"} Oct 05 09:23:23 crc kubenswrapper[4846]: I1005 09:23:23.492718 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-pprrt" podStartSLOduration=2.9713977910000002 podStartE2EDuration="8.492699368s" podCreationTimestamp="2025-10-05 09:23:15 +0000 UTC" firstStartedPulling="2025-10-05 09:23:17.380288728 +0000 UTC m=+9319.621141503" lastFinishedPulling="2025-10-05 09:23:22.901590295 +0000 UTC m=+9325.142443080" observedRunningTime="2025-10-05 09:23:23.487650666 +0000 UTC m=+9325.728503451" watchObservedRunningTime="2025-10-05 09:23:23.492699368 +0000 UTC m=+9325.733552153" Oct 05 09:23:24 crc kubenswrapper[4846]: I1005 09:23:24.362924 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-r8mf2"] Oct 05 09:23:24 crc kubenswrapper[4846]: I1005 09:23:24.366844 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-r8mf2" Oct 05 09:23:24 crc kubenswrapper[4846]: I1005 09:23:24.388244 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-r8mf2"] Oct 05 09:23:24 crc kubenswrapper[4846]: I1005 09:23:24.546035 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mdtjt\" (UniqueName: \"kubernetes.io/projected/dca9ba8b-113d-4f55-9a66-7350376e0918-kube-api-access-mdtjt\") pod \"certified-operators-r8mf2\" (UID: \"dca9ba8b-113d-4f55-9a66-7350376e0918\") " pod="openshift-marketplace/certified-operators-r8mf2" Oct 05 09:23:24 crc kubenswrapper[4846]: I1005 09:23:24.546324 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dca9ba8b-113d-4f55-9a66-7350376e0918-utilities\") pod \"certified-operators-r8mf2\" (UID: \"dca9ba8b-113d-4f55-9a66-7350376e0918\") " pod="openshift-marketplace/certified-operators-r8mf2" Oct 05 09:23:24 crc kubenswrapper[4846]: I1005 09:23:24.546345 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dca9ba8b-113d-4f55-9a66-7350376e0918-catalog-content\") pod \"certified-operators-r8mf2\" (UID: \"dca9ba8b-113d-4f55-9a66-7350376e0918\") " pod="openshift-marketplace/certified-operators-r8mf2" Oct 05 09:23:24 crc kubenswrapper[4846]: I1005 09:23:24.648627 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mdtjt\" (UniqueName: \"kubernetes.io/projected/dca9ba8b-113d-4f55-9a66-7350376e0918-kube-api-access-mdtjt\") pod \"certified-operators-r8mf2\" (UID: \"dca9ba8b-113d-4f55-9a66-7350376e0918\") " pod="openshift-marketplace/certified-operators-r8mf2" Oct 05 09:23:24 crc kubenswrapper[4846]: I1005 09:23:24.648684 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dca9ba8b-113d-4f55-9a66-7350376e0918-utilities\") pod \"certified-operators-r8mf2\" (UID: \"dca9ba8b-113d-4f55-9a66-7350376e0918\") " pod="openshift-marketplace/certified-operators-r8mf2" Oct 05 09:23:24 crc kubenswrapper[4846]: I1005 09:23:24.648712 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dca9ba8b-113d-4f55-9a66-7350376e0918-catalog-content\") pod \"certified-operators-r8mf2\" (UID: \"dca9ba8b-113d-4f55-9a66-7350376e0918\") " pod="openshift-marketplace/certified-operators-r8mf2" Oct 05 09:23:24 crc kubenswrapper[4846]: I1005 09:23:24.651558 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dca9ba8b-113d-4f55-9a66-7350376e0918-utilities\") pod \"certified-operators-r8mf2\" (UID: \"dca9ba8b-113d-4f55-9a66-7350376e0918\") " pod="openshift-marketplace/certified-operators-r8mf2" Oct 05 09:23:24 crc kubenswrapper[4846]: I1005 09:23:24.651837 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dca9ba8b-113d-4f55-9a66-7350376e0918-catalog-content\") pod \"certified-operators-r8mf2\" (UID: \"dca9ba8b-113d-4f55-9a66-7350376e0918\") " pod="openshift-marketplace/certified-operators-r8mf2" Oct 05 09:23:24 crc kubenswrapper[4846]: I1005 09:23:24.684361 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mdtjt\" (UniqueName: \"kubernetes.io/projected/dca9ba8b-113d-4f55-9a66-7350376e0918-kube-api-access-mdtjt\") pod \"certified-operators-r8mf2\" (UID: \"dca9ba8b-113d-4f55-9a66-7350376e0918\") " pod="openshift-marketplace/certified-operators-r8mf2" Oct 05 09:23:24 crc kubenswrapper[4846]: I1005 09:23:24.699661 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-r8mf2" Oct 05 09:23:25 crc kubenswrapper[4846]: I1005 09:23:25.276844 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-r8mf2"] Oct 05 09:23:25 crc kubenswrapper[4846]: W1005 09:23:25.290026 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddca9ba8b_113d_4f55_9a66_7350376e0918.slice/crio-a0aae590fd4c01ad7905cc7dddb60eda34e34133bef9f5f62a82c5969f646f08 WatchSource:0}: Error finding container a0aae590fd4c01ad7905cc7dddb60eda34e34133bef9f5f62a82c5969f646f08: Status 404 returned error can't find the container with id a0aae590fd4c01ad7905cc7dddb60eda34e34133bef9f5f62a82c5969f646f08 Oct 05 09:23:25 crc kubenswrapper[4846]: I1005 09:23:25.492434 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-r8mf2" event={"ID":"dca9ba8b-113d-4f55-9a66-7350376e0918","Type":"ContainerStarted","Data":"a0aae590fd4c01ad7905cc7dddb60eda34e34133bef9f5f62a82c5969f646f08"} Oct 05 09:23:26 crc kubenswrapper[4846]: I1005 09:23:26.332522 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-pprrt" Oct 05 09:23:26 crc kubenswrapper[4846]: I1005 09:23:26.332591 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-pprrt" Oct 05 09:23:26 crc kubenswrapper[4846]: I1005 09:23:26.510733 4846 generic.go:334] "Generic (PLEG): container finished" podID="dca9ba8b-113d-4f55-9a66-7350376e0918" containerID="b2c0ccfaea9a4d5ac3a3ae9be998766b0e377f43405432e4f31557bc2e9ffdb8" exitCode=0 Oct 05 09:23:26 crc kubenswrapper[4846]: I1005 09:23:26.523978 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-r8mf2" event={"ID":"dca9ba8b-113d-4f55-9a66-7350376e0918","Type":"ContainerDied","Data":"b2c0ccfaea9a4d5ac3a3ae9be998766b0e377f43405432e4f31557bc2e9ffdb8"} Oct 05 09:23:27 crc kubenswrapper[4846]: I1005 09:23:27.522972 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-r8mf2" event={"ID":"dca9ba8b-113d-4f55-9a66-7350376e0918","Type":"ContainerStarted","Data":"4044341c9766860086af42ff6587fcb6439924cbb39773f4fa4b02724f19e59a"} Oct 05 09:23:27 crc kubenswrapper[4846]: I1005 09:23:27.551738 4846 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-pprrt" podUID="f724ea50-bee0-4a90-a0f3-def9dbd56b0a" containerName="registry-server" probeResult="failure" output=< Oct 05 09:23:27 crc kubenswrapper[4846]: timeout: failed to connect service ":50051" within 1s Oct 05 09:23:27 crc kubenswrapper[4846]: > Oct 05 09:23:28 crc kubenswrapper[4846]: I1005 09:23:28.506757 4846 scope.go:117] "RemoveContainer" containerID="2ddf4e92e4eafd2c8e76240cc0766ba587ed44b22d7136c95e5159c1c5e8c32e" Oct 05 09:23:28 crc kubenswrapper[4846]: E1005 09:23:28.507724 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 09:23:29 crc kubenswrapper[4846]: I1005 09:23:29.546454 4846 generic.go:334] "Generic (PLEG): container finished" podID="dca9ba8b-113d-4f55-9a66-7350376e0918" containerID="4044341c9766860086af42ff6587fcb6439924cbb39773f4fa4b02724f19e59a" exitCode=0 Oct 05 09:23:29 crc kubenswrapper[4846]: I1005 09:23:29.546501 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-r8mf2" event={"ID":"dca9ba8b-113d-4f55-9a66-7350376e0918","Type":"ContainerDied","Data":"4044341c9766860086af42ff6587fcb6439924cbb39773f4fa4b02724f19e59a"} Oct 05 09:23:30 crc kubenswrapper[4846]: I1005 09:23:30.560052 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-r8mf2" event={"ID":"dca9ba8b-113d-4f55-9a66-7350376e0918","Type":"ContainerStarted","Data":"f46f3f7fe9829bdb7a499a84a9b2fcce72aa736c27ef4ec62db04ceb039a2425"} Oct 05 09:23:30 crc kubenswrapper[4846]: I1005 09:23:30.576380 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-r8mf2" podStartSLOduration=3.112740645 podStartE2EDuration="6.576365224s" podCreationTimestamp="2025-10-05 09:23:24 +0000 UTC" firstStartedPulling="2025-10-05 09:23:26.51919687 +0000 UTC m=+9328.760049645" lastFinishedPulling="2025-10-05 09:23:29.982821449 +0000 UTC m=+9332.223674224" observedRunningTime="2025-10-05 09:23:30.573596642 +0000 UTC m=+9332.814449417" watchObservedRunningTime="2025-10-05 09:23:30.576365224 +0000 UTC m=+9332.817218009" Oct 05 09:23:34 crc kubenswrapper[4846]: I1005 09:23:34.700594 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-r8mf2" Oct 05 09:23:34 crc kubenswrapper[4846]: I1005 09:23:34.701233 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-r8mf2" Oct 05 09:23:35 crc kubenswrapper[4846]: I1005 09:23:35.753195 4846 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-r8mf2" podUID="dca9ba8b-113d-4f55-9a66-7350376e0918" containerName="registry-server" probeResult="failure" output=< Oct 05 09:23:35 crc kubenswrapper[4846]: timeout: failed to connect service ":50051" within 1s Oct 05 09:23:35 crc kubenswrapper[4846]: > Oct 05 09:23:37 crc kubenswrapper[4846]: I1005 09:23:37.396717 4846 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-pprrt" podUID="f724ea50-bee0-4a90-a0f3-def9dbd56b0a" containerName="registry-server" probeResult="failure" output=< Oct 05 09:23:37 crc kubenswrapper[4846]: timeout: failed to connect service ":50051" within 1s Oct 05 09:23:37 crc kubenswrapper[4846]: > Oct 05 09:23:39 crc kubenswrapper[4846]: I1005 09:23:39.498276 4846 scope.go:117] "RemoveContainer" containerID="2ddf4e92e4eafd2c8e76240cc0766ba587ed44b22d7136c95e5159c1c5e8c32e" Oct 05 09:23:39 crc kubenswrapper[4846]: E1005 09:23:39.498763 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 09:23:44 crc kubenswrapper[4846]: I1005 09:23:44.782824 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-r8mf2" Oct 05 09:23:44 crc kubenswrapper[4846]: I1005 09:23:44.847409 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-r8mf2" Oct 05 09:23:45 crc kubenswrapper[4846]: I1005 09:23:45.026072 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-r8mf2"] Oct 05 09:23:46 crc kubenswrapper[4846]: I1005 09:23:46.720719 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-r8mf2" podUID="dca9ba8b-113d-4f55-9a66-7350376e0918" containerName="registry-server" containerID="cri-o://f46f3f7fe9829bdb7a499a84a9b2fcce72aa736c27ef4ec62db04ceb039a2425" gracePeriod=2 Oct 05 09:23:47 crc kubenswrapper[4846]: I1005 09:23:47.241298 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-r8mf2" Oct 05 09:23:47 crc kubenswrapper[4846]: I1005 09:23:47.357845 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mdtjt\" (UniqueName: \"kubernetes.io/projected/dca9ba8b-113d-4f55-9a66-7350376e0918-kube-api-access-mdtjt\") pod \"dca9ba8b-113d-4f55-9a66-7350376e0918\" (UID: \"dca9ba8b-113d-4f55-9a66-7350376e0918\") " Oct 05 09:23:47 crc kubenswrapper[4846]: I1005 09:23:47.357941 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dca9ba8b-113d-4f55-9a66-7350376e0918-catalog-content\") pod \"dca9ba8b-113d-4f55-9a66-7350376e0918\" (UID: \"dca9ba8b-113d-4f55-9a66-7350376e0918\") " Oct 05 09:23:47 crc kubenswrapper[4846]: I1005 09:23:47.358003 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dca9ba8b-113d-4f55-9a66-7350376e0918-utilities\") pod \"dca9ba8b-113d-4f55-9a66-7350376e0918\" (UID: \"dca9ba8b-113d-4f55-9a66-7350376e0918\") " Oct 05 09:23:47 crc kubenswrapper[4846]: I1005 09:23:47.358834 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dca9ba8b-113d-4f55-9a66-7350376e0918-utilities" (OuterVolumeSpecName: "utilities") pod "dca9ba8b-113d-4f55-9a66-7350376e0918" (UID: "dca9ba8b-113d-4f55-9a66-7350376e0918"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 09:23:47 crc kubenswrapper[4846]: I1005 09:23:47.363524 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dca9ba8b-113d-4f55-9a66-7350376e0918-kube-api-access-mdtjt" (OuterVolumeSpecName: "kube-api-access-mdtjt") pod "dca9ba8b-113d-4f55-9a66-7350376e0918" (UID: "dca9ba8b-113d-4f55-9a66-7350376e0918"). InnerVolumeSpecName "kube-api-access-mdtjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 09:23:47 crc kubenswrapper[4846]: I1005 09:23:47.379649 4846 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-pprrt" podUID="f724ea50-bee0-4a90-a0f3-def9dbd56b0a" containerName="registry-server" probeResult="failure" output=< Oct 05 09:23:47 crc kubenswrapper[4846]: timeout: failed to connect service ":50051" within 1s Oct 05 09:23:47 crc kubenswrapper[4846]: > Oct 05 09:23:47 crc kubenswrapper[4846]: I1005 09:23:47.411453 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dca9ba8b-113d-4f55-9a66-7350376e0918-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "dca9ba8b-113d-4f55-9a66-7350376e0918" (UID: "dca9ba8b-113d-4f55-9a66-7350376e0918"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 09:23:47 crc kubenswrapper[4846]: I1005 09:23:47.460795 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mdtjt\" (UniqueName: \"kubernetes.io/projected/dca9ba8b-113d-4f55-9a66-7350376e0918-kube-api-access-mdtjt\") on node \"crc\" DevicePath \"\"" Oct 05 09:23:47 crc kubenswrapper[4846]: I1005 09:23:47.460836 4846 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dca9ba8b-113d-4f55-9a66-7350376e0918-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 05 09:23:47 crc kubenswrapper[4846]: I1005 09:23:47.460849 4846 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dca9ba8b-113d-4f55-9a66-7350376e0918-utilities\") on node \"crc\" DevicePath \"\"" Oct 05 09:23:47 crc kubenswrapper[4846]: I1005 09:23:47.731479 4846 generic.go:334] "Generic (PLEG): container finished" podID="dca9ba8b-113d-4f55-9a66-7350376e0918" containerID="f46f3f7fe9829bdb7a499a84a9b2fcce72aa736c27ef4ec62db04ceb039a2425" exitCode=0 Oct 05 09:23:47 crc kubenswrapper[4846]: I1005 09:23:47.731564 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-r8mf2" Oct 05 09:23:47 crc kubenswrapper[4846]: I1005 09:23:47.731564 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-r8mf2" event={"ID":"dca9ba8b-113d-4f55-9a66-7350376e0918","Type":"ContainerDied","Data":"f46f3f7fe9829bdb7a499a84a9b2fcce72aa736c27ef4ec62db04ceb039a2425"} Oct 05 09:23:47 crc kubenswrapper[4846]: I1005 09:23:47.733149 4846 scope.go:117] "RemoveContainer" containerID="f46f3f7fe9829bdb7a499a84a9b2fcce72aa736c27ef4ec62db04ceb039a2425" Oct 05 09:23:47 crc kubenswrapper[4846]: I1005 09:23:47.733329 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-r8mf2" event={"ID":"dca9ba8b-113d-4f55-9a66-7350376e0918","Type":"ContainerDied","Data":"a0aae590fd4c01ad7905cc7dddb60eda34e34133bef9f5f62a82c5969f646f08"} Oct 05 09:23:47 crc kubenswrapper[4846]: I1005 09:23:47.758527 4846 scope.go:117] "RemoveContainer" containerID="4044341c9766860086af42ff6587fcb6439924cbb39773f4fa4b02724f19e59a" Oct 05 09:23:47 crc kubenswrapper[4846]: I1005 09:23:47.773588 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-r8mf2"] Oct 05 09:23:47 crc kubenswrapper[4846]: I1005 09:23:47.782603 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-r8mf2"] Oct 05 09:23:47 crc kubenswrapper[4846]: I1005 09:23:47.803439 4846 scope.go:117] "RemoveContainer" containerID="b2c0ccfaea9a4d5ac3a3ae9be998766b0e377f43405432e4f31557bc2e9ffdb8" Oct 05 09:23:47 crc kubenswrapper[4846]: I1005 09:23:47.851505 4846 scope.go:117] "RemoveContainer" containerID="f46f3f7fe9829bdb7a499a84a9b2fcce72aa736c27ef4ec62db04ceb039a2425" Oct 05 09:23:47 crc kubenswrapper[4846]: E1005 09:23:47.851963 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f46f3f7fe9829bdb7a499a84a9b2fcce72aa736c27ef4ec62db04ceb039a2425\": container with ID starting with f46f3f7fe9829bdb7a499a84a9b2fcce72aa736c27ef4ec62db04ceb039a2425 not found: ID does not exist" containerID="f46f3f7fe9829bdb7a499a84a9b2fcce72aa736c27ef4ec62db04ceb039a2425" Oct 05 09:23:47 crc kubenswrapper[4846]: I1005 09:23:47.851996 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f46f3f7fe9829bdb7a499a84a9b2fcce72aa736c27ef4ec62db04ceb039a2425"} err="failed to get container status \"f46f3f7fe9829bdb7a499a84a9b2fcce72aa736c27ef4ec62db04ceb039a2425\": rpc error: code = NotFound desc = could not find container \"f46f3f7fe9829bdb7a499a84a9b2fcce72aa736c27ef4ec62db04ceb039a2425\": container with ID starting with f46f3f7fe9829bdb7a499a84a9b2fcce72aa736c27ef4ec62db04ceb039a2425 not found: ID does not exist" Oct 05 09:23:47 crc kubenswrapper[4846]: I1005 09:23:47.852019 4846 scope.go:117] "RemoveContainer" containerID="4044341c9766860086af42ff6587fcb6439924cbb39773f4fa4b02724f19e59a" Oct 05 09:23:47 crc kubenswrapper[4846]: E1005 09:23:47.852328 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4044341c9766860086af42ff6587fcb6439924cbb39773f4fa4b02724f19e59a\": container with ID starting with 4044341c9766860086af42ff6587fcb6439924cbb39773f4fa4b02724f19e59a not found: ID does not exist" containerID="4044341c9766860086af42ff6587fcb6439924cbb39773f4fa4b02724f19e59a" Oct 05 09:23:47 crc kubenswrapper[4846]: I1005 09:23:47.852363 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4044341c9766860086af42ff6587fcb6439924cbb39773f4fa4b02724f19e59a"} err="failed to get container status \"4044341c9766860086af42ff6587fcb6439924cbb39773f4fa4b02724f19e59a\": rpc error: code = NotFound desc = could not find container \"4044341c9766860086af42ff6587fcb6439924cbb39773f4fa4b02724f19e59a\": container with ID starting with 4044341c9766860086af42ff6587fcb6439924cbb39773f4fa4b02724f19e59a not found: ID does not exist" Oct 05 09:23:47 crc kubenswrapper[4846]: I1005 09:23:47.852382 4846 scope.go:117] "RemoveContainer" containerID="b2c0ccfaea9a4d5ac3a3ae9be998766b0e377f43405432e4f31557bc2e9ffdb8" Oct 05 09:23:47 crc kubenswrapper[4846]: E1005 09:23:47.852793 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b2c0ccfaea9a4d5ac3a3ae9be998766b0e377f43405432e4f31557bc2e9ffdb8\": container with ID starting with b2c0ccfaea9a4d5ac3a3ae9be998766b0e377f43405432e4f31557bc2e9ffdb8 not found: ID does not exist" containerID="b2c0ccfaea9a4d5ac3a3ae9be998766b0e377f43405432e4f31557bc2e9ffdb8" Oct 05 09:23:47 crc kubenswrapper[4846]: I1005 09:23:47.852828 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b2c0ccfaea9a4d5ac3a3ae9be998766b0e377f43405432e4f31557bc2e9ffdb8"} err="failed to get container status \"b2c0ccfaea9a4d5ac3a3ae9be998766b0e377f43405432e4f31557bc2e9ffdb8\": rpc error: code = NotFound desc = could not find container \"b2c0ccfaea9a4d5ac3a3ae9be998766b0e377f43405432e4f31557bc2e9ffdb8\": container with ID starting with b2c0ccfaea9a4d5ac3a3ae9be998766b0e377f43405432e4f31557bc2e9ffdb8 not found: ID does not exist" Oct 05 09:23:48 crc kubenswrapper[4846]: I1005 09:23:48.513454 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dca9ba8b-113d-4f55-9a66-7350376e0918" path="/var/lib/kubelet/pods/dca9ba8b-113d-4f55-9a66-7350376e0918/volumes" Oct 05 09:23:52 crc kubenswrapper[4846]: I1005 09:23:52.498367 4846 scope.go:117] "RemoveContainer" containerID="2ddf4e92e4eafd2c8e76240cc0766ba587ed44b22d7136c95e5159c1c5e8c32e" Oct 05 09:23:52 crc kubenswrapper[4846]: E1005 09:23:52.499293 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 09:23:56 crc kubenswrapper[4846]: I1005 09:23:56.403479 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-pprrt" Oct 05 09:23:56 crc kubenswrapper[4846]: I1005 09:23:56.492295 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-pprrt" Oct 05 09:23:56 crc kubenswrapper[4846]: I1005 09:23:56.656387 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-pprrt"] Oct 05 09:23:57 crc kubenswrapper[4846]: I1005 09:23:57.858866 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-pprrt" podUID="f724ea50-bee0-4a90-a0f3-def9dbd56b0a" containerName="registry-server" containerID="cri-o://5726ba891cba54ccc0ba32ee24fbd7218f9131ef25b7ffddf0a8d8cf53df5e00" gracePeriod=2 Oct 05 09:23:58 crc kubenswrapper[4846]: I1005 09:23:58.358233 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-pprrt" Oct 05 09:23:58 crc kubenswrapper[4846]: I1005 09:23:58.524710 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f724ea50-bee0-4a90-a0f3-def9dbd56b0a-catalog-content\") pod \"f724ea50-bee0-4a90-a0f3-def9dbd56b0a\" (UID: \"f724ea50-bee0-4a90-a0f3-def9dbd56b0a\") " Oct 05 09:23:58 crc kubenswrapper[4846]: I1005 09:23:58.524907 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f724ea50-bee0-4a90-a0f3-def9dbd56b0a-utilities\") pod \"f724ea50-bee0-4a90-a0f3-def9dbd56b0a\" (UID: \"f724ea50-bee0-4a90-a0f3-def9dbd56b0a\") " Oct 05 09:23:58 crc kubenswrapper[4846]: I1005 09:23:58.525093 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xmpjb\" (UniqueName: \"kubernetes.io/projected/f724ea50-bee0-4a90-a0f3-def9dbd56b0a-kube-api-access-xmpjb\") pod \"f724ea50-bee0-4a90-a0f3-def9dbd56b0a\" (UID: \"f724ea50-bee0-4a90-a0f3-def9dbd56b0a\") " Oct 05 09:23:58 crc kubenswrapper[4846]: I1005 09:23:58.525661 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f724ea50-bee0-4a90-a0f3-def9dbd56b0a-utilities" (OuterVolumeSpecName: "utilities") pod "f724ea50-bee0-4a90-a0f3-def9dbd56b0a" (UID: "f724ea50-bee0-4a90-a0f3-def9dbd56b0a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 09:23:58 crc kubenswrapper[4846]: I1005 09:23:58.532244 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f724ea50-bee0-4a90-a0f3-def9dbd56b0a-kube-api-access-xmpjb" (OuterVolumeSpecName: "kube-api-access-xmpjb") pod "f724ea50-bee0-4a90-a0f3-def9dbd56b0a" (UID: "f724ea50-bee0-4a90-a0f3-def9dbd56b0a"). InnerVolumeSpecName "kube-api-access-xmpjb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 09:23:58 crc kubenswrapper[4846]: I1005 09:23:58.602928 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f724ea50-bee0-4a90-a0f3-def9dbd56b0a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f724ea50-bee0-4a90-a0f3-def9dbd56b0a" (UID: "f724ea50-bee0-4a90-a0f3-def9dbd56b0a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 09:23:58 crc kubenswrapper[4846]: I1005 09:23:58.627983 4846 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f724ea50-bee0-4a90-a0f3-def9dbd56b0a-utilities\") on node \"crc\" DevicePath \"\"" Oct 05 09:23:58 crc kubenswrapper[4846]: I1005 09:23:58.628086 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xmpjb\" (UniqueName: \"kubernetes.io/projected/f724ea50-bee0-4a90-a0f3-def9dbd56b0a-kube-api-access-xmpjb\") on node \"crc\" DevicePath \"\"" Oct 05 09:23:58 crc kubenswrapper[4846]: I1005 09:23:58.628118 4846 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f724ea50-bee0-4a90-a0f3-def9dbd56b0a-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 05 09:23:58 crc kubenswrapper[4846]: I1005 09:23:58.891189 4846 generic.go:334] "Generic (PLEG): container finished" podID="f724ea50-bee0-4a90-a0f3-def9dbd56b0a" containerID="5726ba891cba54ccc0ba32ee24fbd7218f9131ef25b7ffddf0a8d8cf53df5e00" exitCode=0 Oct 05 09:23:58 crc kubenswrapper[4846]: I1005 09:23:58.891233 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pprrt" event={"ID":"f724ea50-bee0-4a90-a0f3-def9dbd56b0a","Type":"ContainerDied","Data":"5726ba891cba54ccc0ba32ee24fbd7218f9131ef25b7ffddf0a8d8cf53df5e00"} Oct 05 09:23:58 crc kubenswrapper[4846]: I1005 09:23:58.891262 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pprrt" event={"ID":"f724ea50-bee0-4a90-a0f3-def9dbd56b0a","Type":"ContainerDied","Data":"cd0ea672880c96895116ffe009480f1a64ae07266d46d982e446a965f7330db6"} Oct 05 09:23:58 crc kubenswrapper[4846]: I1005 09:23:58.891280 4846 scope.go:117] "RemoveContainer" containerID="5726ba891cba54ccc0ba32ee24fbd7218f9131ef25b7ffddf0a8d8cf53df5e00" Oct 05 09:23:58 crc kubenswrapper[4846]: I1005 09:23:58.891448 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-pprrt" Oct 05 09:23:58 crc kubenswrapper[4846]: I1005 09:23:58.941364 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-pprrt"] Oct 05 09:23:58 crc kubenswrapper[4846]: I1005 09:23:58.950563 4846 scope.go:117] "RemoveContainer" containerID="9962bddd6e832f0e1055e1e0d7236e065de89bf0216c30e62d7eb52f9849a87a" Oct 05 09:23:58 crc kubenswrapper[4846]: I1005 09:23:58.956078 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-pprrt"] Oct 05 09:23:58 crc kubenswrapper[4846]: I1005 09:23:58.974895 4846 scope.go:117] "RemoveContainer" containerID="8ca55153530905c32876af2cab5c26abe18fbe153136ba0470926236f7c3c6fa" Oct 05 09:23:59 crc kubenswrapper[4846]: I1005 09:23:59.026370 4846 scope.go:117] "RemoveContainer" containerID="5726ba891cba54ccc0ba32ee24fbd7218f9131ef25b7ffddf0a8d8cf53df5e00" Oct 05 09:23:59 crc kubenswrapper[4846]: E1005 09:23:59.027316 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5726ba891cba54ccc0ba32ee24fbd7218f9131ef25b7ffddf0a8d8cf53df5e00\": container with ID starting with 5726ba891cba54ccc0ba32ee24fbd7218f9131ef25b7ffddf0a8d8cf53df5e00 not found: ID does not exist" containerID="5726ba891cba54ccc0ba32ee24fbd7218f9131ef25b7ffddf0a8d8cf53df5e00" Oct 05 09:23:59 crc kubenswrapper[4846]: I1005 09:23:59.027369 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5726ba891cba54ccc0ba32ee24fbd7218f9131ef25b7ffddf0a8d8cf53df5e00"} err="failed to get container status \"5726ba891cba54ccc0ba32ee24fbd7218f9131ef25b7ffddf0a8d8cf53df5e00\": rpc error: code = NotFound desc = could not find container \"5726ba891cba54ccc0ba32ee24fbd7218f9131ef25b7ffddf0a8d8cf53df5e00\": container with ID starting with 5726ba891cba54ccc0ba32ee24fbd7218f9131ef25b7ffddf0a8d8cf53df5e00 not found: ID does not exist" Oct 05 09:23:59 crc kubenswrapper[4846]: I1005 09:23:59.027401 4846 scope.go:117] "RemoveContainer" containerID="9962bddd6e832f0e1055e1e0d7236e065de89bf0216c30e62d7eb52f9849a87a" Oct 05 09:23:59 crc kubenswrapper[4846]: E1005 09:23:59.027893 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9962bddd6e832f0e1055e1e0d7236e065de89bf0216c30e62d7eb52f9849a87a\": container with ID starting with 9962bddd6e832f0e1055e1e0d7236e065de89bf0216c30e62d7eb52f9849a87a not found: ID does not exist" containerID="9962bddd6e832f0e1055e1e0d7236e065de89bf0216c30e62d7eb52f9849a87a" Oct 05 09:23:59 crc kubenswrapper[4846]: I1005 09:23:59.027929 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9962bddd6e832f0e1055e1e0d7236e065de89bf0216c30e62d7eb52f9849a87a"} err="failed to get container status \"9962bddd6e832f0e1055e1e0d7236e065de89bf0216c30e62d7eb52f9849a87a\": rpc error: code = NotFound desc = could not find container \"9962bddd6e832f0e1055e1e0d7236e065de89bf0216c30e62d7eb52f9849a87a\": container with ID starting with 9962bddd6e832f0e1055e1e0d7236e065de89bf0216c30e62d7eb52f9849a87a not found: ID does not exist" Oct 05 09:23:59 crc kubenswrapper[4846]: I1005 09:23:59.027956 4846 scope.go:117] "RemoveContainer" containerID="8ca55153530905c32876af2cab5c26abe18fbe153136ba0470926236f7c3c6fa" Oct 05 09:23:59 crc kubenswrapper[4846]: E1005 09:23:59.028422 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8ca55153530905c32876af2cab5c26abe18fbe153136ba0470926236f7c3c6fa\": container with ID starting with 8ca55153530905c32876af2cab5c26abe18fbe153136ba0470926236f7c3c6fa not found: ID does not exist" containerID="8ca55153530905c32876af2cab5c26abe18fbe153136ba0470926236f7c3c6fa" Oct 05 09:23:59 crc kubenswrapper[4846]: I1005 09:23:59.028457 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8ca55153530905c32876af2cab5c26abe18fbe153136ba0470926236f7c3c6fa"} err="failed to get container status \"8ca55153530905c32876af2cab5c26abe18fbe153136ba0470926236f7c3c6fa\": rpc error: code = NotFound desc = could not find container \"8ca55153530905c32876af2cab5c26abe18fbe153136ba0470926236f7c3c6fa\": container with ID starting with 8ca55153530905c32876af2cab5c26abe18fbe153136ba0470926236f7c3c6fa not found: ID does not exist" Oct 05 09:24:00 crc kubenswrapper[4846]: I1005 09:24:00.517537 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f724ea50-bee0-4a90-a0f3-def9dbd56b0a" path="/var/lib/kubelet/pods/f724ea50-bee0-4a90-a0f3-def9dbd56b0a/volumes" Oct 05 09:24:05 crc kubenswrapper[4846]: I1005 09:24:05.497930 4846 scope.go:117] "RemoveContainer" containerID="2ddf4e92e4eafd2c8e76240cc0766ba587ed44b22d7136c95e5159c1c5e8c32e" Oct 05 09:24:05 crc kubenswrapper[4846]: E1005 09:24:05.498937 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 09:24:18 crc kubenswrapper[4846]: I1005 09:24:18.507652 4846 scope.go:117] "RemoveContainer" containerID="2ddf4e92e4eafd2c8e76240cc0766ba587ed44b22d7136c95e5159c1c5e8c32e" Oct 05 09:24:18 crc kubenswrapper[4846]: E1005 09:24:18.508524 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 09:24:31 crc kubenswrapper[4846]: I1005 09:24:31.497924 4846 scope.go:117] "RemoveContainer" containerID="2ddf4e92e4eafd2c8e76240cc0766ba587ed44b22d7136c95e5159c1c5e8c32e" Oct 05 09:24:33 crc kubenswrapper[4846]: I1005 09:24:33.292121 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" event={"ID":"ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7","Type":"ContainerStarted","Data":"25000c03b37ea2bd1717c4fee8262aa1fc79b5e0033f2d47b41a9ddab188e4f3"} Oct 05 09:25:40 crc kubenswrapper[4846]: I1005 09:25:40.086531 4846 scope.go:117] "RemoveContainer" containerID="a22aa83760413f05b5cc8ff11761535de907053350fdf93d15fe9be94233cc63" Oct 05 09:25:40 crc kubenswrapper[4846]: I1005 09:25:40.140244 4846 scope.go:117] "RemoveContainer" containerID="63be0a54a6a1758e1e5bd71334414105c325f82b3cc45f854e0c3aff65ab31cb" Oct 05 09:26:40 crc kubenswrapper[4846]: I1005 09:26:40.292762 4846 scope.go:117] "RemoveContainer" containerID="54792b0b9e2f53a6566189ea3b286a204f4e9fc605021ed11336541a15ac8460" Oct 05 09:26:53 crc kubenswrapper[4846]: I1005 09:26:53.325318 4846 patch_prober.go:28] interesting pod/machine-config-daemon-fscvf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 09:26:53 crc kubenswrapper[4846]: I1005 09:26:53.326026 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 09:27:23 crc kubenswrapper[4846]: I1005 09:27:23.325423 4846 patch_prober.go:28] interesting pod/machine-config-daemon-fscvf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 09:27:23 crc kubenswrapper[4846]: I1005 09:27:23.326219 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 09:27:53 crc kubenswrapper[4846]: I1005 09:27:53.325682 4846 patch_prober.go:28] interesting pod/machine-config-daemon-fscvf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 09:27:53 crc kubenswrapper[4846]: I1005 09:27:53.326598 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 09:27:53 crc kubenswrapper[4846]: I1005 09:27:53.326686 4846 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" Oct 05 09:27:53 crc kubenswrapper[4846]: I1005 09:27:53.327837 4846 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"25000c03b37ea2bd1717c4fee8262aa1fc79b5e0033f2d47b41a9ddab188e4f3"} pod="openshift-machine-config-operator/machine-config-daemon-fscvf" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 05 09:27:53 crc kubenswrapper[4846]: I1005 09:27:53.327936 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" containerName="machine-config-daemon" containerID="cri-o://25000c03b37ea2bd1717c4fee8262aa1fc79b5e0033f2d47b41a9ddab188e4f3" gracePeriod=600 Oct 05 09:27:54 crc kubenswrapper[4846]: I1005 09:27:54.755866 4846 generic.go:334] "Generic (PLEG): container finished" podID="4e445bd2-1270-4c05-9e0b-bcdbec35fa74" containerID="dbbcf7bfea17739c7d2f5e98e099b63bbc5451902575a503bbfe34e745ba064b" exitCode=254 Oct 05 09:27:54 crc kubenswrapper[4846]: I1005 09:27:54.755994 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celldnhqr" event={"ID":"4e445bd2-1270-4c05-9e0b-bcdbec35fa74","Type":"ContainerDied","Data":"dbbcf7bfea17739c7d2f5e98e099b63bbc5451902575a503bbfe34e745ba064b"} Oct 05 09:27:54 crc kubenswrapper[4846]: I1005 09:27:54.761262 4846 generic.go:334] "Generic (PLEG): container finished" podID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" containerID="25000c03b37ea2bd1717c4fee8262aa1fc79b5e0033f2d47b41a9ddab188e4f3" exitCode=0 Oct 05 09:27:54 crc kubenswrapper[4846]: I1005 09:27:54.761316 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" event={"ID":"ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7","Type":"ContainerDied","Data":"25000c03b37ea2bd1717c4fee8262aa1fc79b5e0033f2d47b41a9ddab188e4f3"} Oct 05 09:27:54 crc kubenswrapper[4846]: I1005 09:27:54.761348 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" event={"ID":"ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7","Type":"ContainerStarted","Data":"8a13caff3cf7131d82ade0de5fe733ded9547a5758c374eab87ab478fe594e4d"} Oct 05 09:27:54 crc kubenswrapper[4846]: I1005 09:27:54.761370 4846 scope.go:117] "RemoveContainer" containerID="2ddf4e92e4eafd2c8e76240cc0766ba587ed44b22d7136c95e5159c1c5e8c32e" Oct 05 09:27:56 crc kubenswrapper[4846]: I1005 09:27:56.249240 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celldnhqr" Oct 05 09:27:56 crc kubenswrapper[4846]: I1005 09:27:56.368226 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/4e445bd2-1270-4c05-9e0b-bcdbec35fa74-nova-cell1-compute-config-1\") pod \"4e445bd2-1270-4c05-9e0b-bcdbec35fa74\" (UID: \"4e445bd2-1270-4c05-9e0b-bcdbec35fa74\") " Oct 05 09:27:56 crc kubenswrapper[4846]: I1005 09:27:56.368455 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4e445bd2-1270-4c05-9e0b-bcdbec35fa74-nova-cell1-combined-ca-bundle\") pod \"4e445bd2-1270-4c05-9e0b-bcdbec35fa74\" (UID: \"4e445bd2-1270-4c05-9e0b-bcdbec35fa74\") " Oct 05 09:27:56 crc kubenswrapper[4846]: I1005 09:27:56.368607 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/4e445bd2-1270-4c05-9e0b-bcdbec35fa74-nova-cell1-compute-config-0\") pod \"4e445bd2-1270-4c05-9e0b-bcdbec35fa74\" (UID: \"4e445bd2-1270-4c05-9e0b-bcdbec35fa74\") " Oct 05 09:27:56 crc kubenswrapper[4846]: I1005 09:27:56.368711 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/4e445bd2-1270-4c05-9e0b-bcdbec35fa74-nova-cells-global-config-0\") pod \"4e445bd2-1270-4c05-9e0b-bcdbec35fa74\" (UID: \"4e445bd2-1270-4c05-9e0b-bcdbec35fa74\") " Oct 05 09:27:56 crc kubenswrapper[4846]: I1005 09:27:56.368801 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k9xc7\" (UniqueName: \"kubernetes.io/projected/4e445bd2-1270-4c05-9e0b-bcdbec35fa74-kube-api-access-k9xc7\") pod \"4e445bd2-1270-4c05-9e0b-bcdbec35fa74\" (UID: \"4e445bd2-1270-4c05-9e0b-bcdbec35fa74\") " Oct 05 09:27:56 crc kubenswrapper[4846]: I1005 09:27:56.369166 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/4e445bd2-1270-4c05-9e0b-bcdbec35fa74-nova-migration-ssh-key-1\") pod \"4e445bd2-1270-4c05-9e0b-bcdbec35fa74\" (UID: \"4e445bd2-1270-4c05-9e0b-bcdbec35fa74\") " Oct 05 09:27:56 crc kubenswrapper[4846]: I1005 09:27:56.369303 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4e445bd2-1270-4c05-9e0b-bcdbec35fa74-ssh-key\") pod \"4e445bd2-1270-4c05-9e0b-bcdbec35fa74\" (UID: \"4e445bd2-1270-4c05-9e0b-bcdbec35fa74\") " Oct 05 09:27:56 crc kubenswrapper[4846]: I1005 09:27:56.369346 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4e445bd2-1270-4c05-9e0b-bcdbec35fa74-inventory\") pod \"4e445bd2-1270-4c05-9e0b-bcdbec35fa74\" (UID: \"4e445bd2-1270-4c05-9e0b-bcdbec35fa74\") " Oct 05 09:27:56 crc kubenswrapper[4846]: I1005 09:27:56.369400 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/4e445bd2-1270-4c05-9e0b-bcdbec35fa74-nova-migration-ssh-key-0\") pod \"4e445bd2-1270-4c05-9e0b-bcdbec35fa74\" (UID: \"4e445bd2-1270-4c05-9e0b-bcdbec35fa74\") " Oct 05 09:27:56 crc kubenswrapper[4846]: I1005 09:27:56.383595 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4e445bd2-1270-4c05-9e0b-bcdbec35fa74-kube-api-access-k9xc7" (OuterVolumeSpecName: "kube-api-access-k9xc7") pod "4e445bd2-1270-4c05-9e0b-bcdbec35fa74" (UID: "4e445bd2-1270-4c05-9e0b-bcdbec35fa74"). InnerVolumeSpecName "kube-api-access-k9xc7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 09:27:56 crc kubenswrapper[4846]: I1005 09:27:56.385518 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4e445bd2-1270-4c05-9e0b-bcdbec35fa74-nova-cell1-combined-ca-bundle" (OuterVolumeSpecName: "nova-cell1-combined-ca-bundle") pod "4e445bd2-1270-4c05-9e0b-bcdbec35fa74" (UID: "4e445bd2-1270-4c05-9e0b-bcdbec35fa74"). InnerVolumeSpecName "nova-cell1-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 09:27:56 crc kubenswrapper[4846]: I1005 09:27:56.406068 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4e445bd2-1270-4c05-9e0b-bcdbec35fa74-inventory" (OuterVolumeSpecName: "inventory") pod "4e445bd2-1270-4c05-9e0b-bcdbec35fa74" (UID: "4e445bd2-1270-4c05-9e0b-bcdbec35fa74"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 09:27:56 crc kubenswrapper[4846]: I1005 09:27:56.407987 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4e445bd2-1270-4c05-9e0b-bcdbec35fa74-nova-migration-ssh-key-1" (OuterVolumeSpecName: "nova-migration-ssh-key-1") pod "4e445bd2-1270-4c05-9e0b-bcdbec35fa74" (UID: "4e445bd2-1270-4c05-9e0b-bcdbec35fa74"). InnerVolumeSpecName "nova-migration-ssh-key-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 09:27:56 crc kubenswrapper[4846]: I1005 09:27:56.409507 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4e445bd2-1270-4c05-9e0b-bcdbec35fa74-nova-cell1-compute-config-0" (OuterVolumeSpecName: "nova-cell1-compute-config-0") pod "4e445bd2-1270-4c05-9e0b-bcdbec35fa74" (UID: "4e445bd2-1270-4c05-9e0b-bcdbec35fa74"). InnerVolumeSpecName "nova-cell1-compute-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 09:27:56 crc kubenswrapper[4846]: I1005 09:27:56.418447 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4e445bd2-1270-4c05-9e0b-bcdbec35fa74-nova-cells-global-config-0" (OuterVolumeSpecName: "nova-cells-global-config-0") pod "4e445bd2-1270-4c05-9e0b-bcdbec35fa74" (UID: "4e445bd2-1270-4c05-9e0b-bcdbec35fa74"). InnerVolumeSpecName "nova-cells-global-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 09:27:56 crc kubenswrapper[4846]: I1005 09:27:56.420166 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4e445bd2-1270-4c05-9e0b-bcdbec35fa74-nova-cell1-compute-config-1" (OuterVolumeSpecName: "nova-cell1-compute-config-1") pod "4e445bd2-1270-4c05-9e0b-bcdbec35fa74" (UID: "4e445bd2-1270-4c05-9e0b-bcdbec35fa74"). InnerVolumeSpecName "nova-cell1-compute-config-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 09:27:56 crc kubenswrapper[4846]: I1005 09:27:56.422595 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4e445bd2-1270-4c05-9e0b-bcdbec35fa74-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "4e445bd2-1270-4c05-9e0b-bcdbec35fa74" (UID: "4e445bd2-1270-4c05-9e0b-bcdbec35fa74"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 09:27:56 crc kubenswrapper[4846]: I1005 09:27:56.426362 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4e445bd2-1270-4c05-9e0b-bcdbec35fa74-nova-migration-ssh-key-0" (OuterVolumeSpecName: "nova-migration-ssh-key-0") pod "4e445bd2-1270-4c05-9e0b-bcdbec35fa74" (UID: "4e445bd2-1270-4c05-9e0b-bcdbec35fa74"). InnerVolumeSpecName "nova-migration-ssh-key-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 09:27:56 crc kubenswrapper[4846]: I1005 09:27:56.472472 4846 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/4e445bd2-1270-4c05-9e0b-bcdbec35fa74-nova-migration-ssh-key-1\") on node \"crc\" DevicePath \"\"" Oct 05 09:27:56 crc kubenswrapper[4846]: I1005 09:27:56.472518 4846 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4e445bd2-1270-4c05-9e0b-bcdbec35fa74-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 05 09:27:56 crc kubenswrapper[4846]: I1005 09:27:56.472533 4846 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4e445bd2-1270-4c05-9e0b-bcdbec35fa74-inventory\") on node \"crc\" DevicePath \"\"" Oct 05 09:27:56 crc kubenswrapper[4846]: I1005 09:27:56.472551 4846 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/4e445bd2-1270-4c05-9e0b-bcdbec35fa74-nova-migration-ssh-key-0\") on node \"crc\" DevicePath \"\"" Oct 05 09:27:56 crc kubenswrapper[4846]: I1005 09:27:56.472569 4846 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/4e445bd2-1270-4c05-9e0b-bcdbec35fa74-nova-cell1-compute-config-1\") on node \"crc\" DevicePath \"\"" Oct 05 09:27:56 crc kubenswrapper[4846]: I1005 09:27:56.472582 4846 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4e445bd2-1270-4c05-9e0b-bcdbec35fa74-nova-cell1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 09:27:56 crc kubenswrapper[4846]: I1005 09:27:56.472598 4846 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/4e445bd2-1270-4c05-9e0b-bcdbec35fa74-nova-cell1-compute-config-0\") on node \"crc\" DevicePath \"\"" Oct 05 09:27:56 crc kubenswrapper[4846]: I1005 09:27:56.472610 4846 reconciler_common.go:293] "Volume detached for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/4e445bd2-1270-4c05-9e0b-bcdbec35fa74-nova-cells-global-config-0\") on node \"crc\" DevicePath \"\"" Oct 05 09:27:56 crc kubenswrapper[4846]: I1005 09:27:56.472622 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k9xc7\" (UniqueName: \"kubernetes.io/projected/4e445bd2-1270-4c05-9e0b-bcdbec35fa74-kube-api-access-k9xc7\") on node \"crc\" DevicePath \"\"" Oct 05 09:27:56 crc kubenswrapper[4846]: I1005 09:27:56.810061 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celldnhqr" event={"ID":"4e445bd2-1270-4c05-9e0b-bcdbec35fa74","Type":"ContainerDied","Data":"4e094dc1e72e556a087a8e51a20b8654c7e32d2bb4dc6a2b32e16f933efda61e"} Oct 05 09:27:56 crc kubenswrapper[4846]: I1005 09:27:56.810852 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4e094dc1e72e556a087a8e51a20b8654c7e32d2bb4dc6a2b32e16f933efda61e" Oct 05 09:27:56 crc kubenswrapper[4846]: I1005 09:27:56.810126 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celldnhqr" Oct 05 09:28:04 crc kubenswrapper[4846]: I1005 09:28:04.032849 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellkhtf8"] Oct 05 09:28:04 crc kubenswrapper[4846]: E1005 09:28:04.033955 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dca9ba8b-113d-4f55-9a66-7350376e0918" containerName="extract-content" Oct 05 09:28:04 crc kubenswrapper[4846]: I1005 09:28:04.033972 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="dca9ba8b-113d-4f55-9a66-7350376e0918" containerName="extract-content" Oct 05 09:28:04 crc kubenswrapper[4846]: E1005 09:28:04.033992 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f724ea50-bee0-4a90-a0f3-def9dbd56b0a" containerName="registry-server" Oct 05 09:28:04 crc kubenswrapper[4846]: I1005 09:28:04.034002 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="f724ea50-bee0-4a90-a0f3-def9dbd56b0a" containerName="registry-server" Oct 05 09:28:04 crc kubenswrapper[4846]: E1005 09:28:04.034021 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f724ea50-bee0-4a90-a0f3-def9dbd56b0a" containerName="extract-content" Oct 05 09:28:04 crc kubenswrapper[4846]: I1005 09:28:04.034029 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="f724ea50-bee0-4a90-a0f3-def9dbd56b0a" containerName="extract-content" Oct 05 09:28:04 crc kubenswrapper[4846]: E1005 09:28:04.034040 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dca9ba8b-113d-4f55-9a66-7350376e0918" containerName="extract-utilities" Oct 05 09:28:04 crc kubenswrapper[4846]: I1005 09:28:04.034047 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="dca9ba8b-113d-4f55-9a66-7350376e0918" containerName="extract-utilities" Oct 05 09:28:04 crc kubenswrapper[4846]: E1005 09:28:04.034062 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4e445bd2-1270-4c05-9e0b-bcdbec35fa74" containerName="nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell1" Oct 05 09:28:04 crc kubenswrapper[4846]: I1005 09:28:04.034074 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="4e445bd2-1270-4c05-9e0b-bcdbec35fa74" containerName="nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell1" Oct 05 09:28:04 crc kubenswrapper[4846]: E1005 09:28:04.034103 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dca9ba8b-113d-4f55-9a66-7350376e0918" containerName="registry-server" Oct 05 09:28:04 crc kubenswrapper[4846]: I1005 09:28:04.034112 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="dca9ba8b-113d-4f55-9a66-7350376e0918" containerName="registry-server" Oct 05 09:28:04 crc kubenswrapper[4846]: E1005 09:28:04.034129 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f724ea50-bee0-4a90-a0f3-def9dbd56b0a" containerName="extract-utilities" Oct 05 09:28:04 crc kubenswrapper[4846]: I1005 09:28:04.034136 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="f724ea50-bee0-4a90-a0f3-def9dbd56b0a" containerName="extract-utilities" Oct 05 09:28:04 crc kubenswrapper[4846]: I1005 09:28:04.034434 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="f724ea50-bee0-4a90-a0f3-def9dbd56b0a" containerName="registry-server" Oct 05 09:28:04 crc kubenswrapper[4846]: I1005 09:28:04.034454 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="dca9ba8b-113d-4f55-9a66-7350376e0918" containerName="registry-server" Oct 05 09:28:04 crc kubenswrapper[4846]: I1005 09:28:04.034470 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="4e445bd2-1270-4c05-9e0b-bcdbec35fa74" containerName="nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell1" Oct 05 09:28:04 crc kubenswrapper[4846]: I1005 09:28:04.035563 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellkhtf8" Oct 05 09:28:04 crc kubenswrapper[4846]: I1005 09:28:04.037859 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 05 09:28:04 crc kubenswrapper[4846]: I1005 09:28:04.038317 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Oct 05 09:28:04 crc kubenswrapper[4846]: I1005 09:28:04.038861 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-migration-ssh-key" Oct 05 09:28:04 crc kubenswrapper[4846]: I1005 09:28:04.039240 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Oct 05 09:28:04 crc kubenswrapper[4846]: I1005 09:28:04.039452 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"nova-cells-global-config" Oct 05 09:28:04 crc kubenswrapper[4846]: I1005 09:28:04.039650 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-7687s" Oct 05 09:28:04 crc kubenswrapper[4846]: I1005 09:28:04.040326 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-compute-config" Oct 05 09:28:04 crc kubenswrapper[4846]: I1005 09:28:04.045393 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellkhtf8"] Oct 05 09:28:04 crc kubenswrapper[4846]: I1005 09:28:04.148921 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dd84ed43-c477-4b83-a735-142efa0d85ea-nova-cell1-combined-ca-bundle\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellkhtf8\" (UID: \"dd84ed43-c477-4b83-a735-142efa0d85ea\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellkhtf8" Oct 05 09:28:04 crc kubenswrapper[4846]: I1005 09:28:04.148997 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/dd84ed43-c477-4b83-a735-142efa0d85ea-nova-migration-ssh-key-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellkhtf8\" (UID: \"dd84ed43-c477-4b83-a735-142efa0d85ea\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellkhtf8" Oct 05 09:28:04 crc kubenswrapper[4846]: I1005 09:28:04.149038 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/dd84ed43-c477-4b83-a735-142efa0d85ea-nova-cells-global-config-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellkhtf8\" (UID: \"dd84ed43-c477-4b83-a735-142efa0d85ea\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellkhtf8" Oct 05 09:28:04 crc kubenswrapper[4846]: I1005 09:28:04.149122 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rvk6m\" (UniqueName: \"kubernetes.io/projected/dd84ed43-c477-4b83-a735-142efa0d85ea-kube-api-access-rvk6m\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellkhtf8\" (UID: \"dd84ed43-c477-4b83-a735-142efa0d85ea\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellkhtf8" Oct 05 09:28:04 crc kubenswrapper[4846]: I1005 09:28:04.149250 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/dd84ed43-c477-4b83-a735-142efa0d85ea-nova-cell1-compute-config-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellkhtf8\" (UID: \"dd84ed43-c477-4b83-a735-142efa0d85ea\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellkhtf8" Oct 05 09:28:04 crc kubenswrapper[4846]: I1005 09:28:04.149331 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/dd84ed43-c477-4b83-a735-142efa0d85ea-nova-migration-ssh-key-1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellkhtf8\" (UID: \"dd84ed43-c477-4b83-a735-142efa0d85ea\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellkhtf8" Oct 05 09:28:04 crc kubenswrapper[4846]: I1005 09:28:04.149383 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/dd84ed43-c477-4b83-a735-142efa0d85ea-ssh-key\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellkhtf8\" (UID: \"dd84ed43-c477-4b83-a735-142efa0d85ea\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellkhtf8" Oct 05 09:28:04 crc kubenswrapper[4846]: I1005 09:28:04.149417 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/dd84ed43-c477-4b83-a735-142efa0d85ea-nova-cell1-compute-config-1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellkhtf8\" (UID: \"dd84ed43-c477-4b83-a735-142efa0d85ea\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellkhtf8" Oct 05 09:28:04 crc kubenswrapper[4846]: I1005 09:28:04.149439 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/dd84ed43-c477-4b83-a735-142efa0d85ea-inventory\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellkhtf8\" (UID: \"dd84ed43-c477-4b83-a735-142efa0d85ea\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellkhtf8" Oct 05 09:28:04 crc kubenswrapper[4846]: I1005 09:28:04.251921 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/dd84ed43-c477-4b83-a735-142efa0d85ea-nova-migration-ssh-key-1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellkhtf8\" (UID: \"dd84ed43-c477-4b83-a735-142efa0d85ea\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellkhtf8" Oct 05 09:28:04 crc kubenswrapper[4846]: I1005 09:28:04.252485 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/dd84ed43-c477-4b83-a735-142efa0d85ea-ssh-key\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellkhtf8\" (UID: \"dd84ed43-c477-4b83-a735-142efa0d85ea\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellkhtf8" Oct 05 09:28:04 crc kubenswrapper[4846]: I1005 09:28:04.252569 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/dd84ed43-c477-4b83-a735-142efa0d85ea-nova-cell1-compute-config-1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellkhtf8\" (UID: \"dd84ed43-c477-4b83-a735-142efa0d85ea\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellkhtf8" Oct 05 09:28:04 crc kubenswrapper[4846]: I1005 09:28:04.252609 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/dd84ed43-c477-4b83-a735-142efa0d85ea-inventory\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellkhtf8\" (UID: \"dd84ed43-c477-4b83-a735-142efa0d85ea\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellkhtf8" Oct 05 09:28:04 crc kubenswrapper[4846]: I1005 09:28:04.252827 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dd84ed43-c477-4b83-a735-142efa0d85ea-nova-cell1-combined-ca-bundle\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellkhtf8\" (UID: \"dd84ed43-c477-4b83-a735-142efa0d85ea\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellkhtf8" Oct 05 09:28:04 crc kubenswrapper[4846]: I1005 09:28:04.252970 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/dd84ed43-c477-4b83-a735-142efa0d85ea-nova-migration-ssh-key-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellkhtf8\" (UID: \"dd84ed43-c477-4b83-a735-142efa0d85ea\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellkhtf8" Oct 05 09:28:04 crc kubenswrapper[4846]: I1005 09:28:04.253101 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/dd84ed43-c477-4b83-a735-142efa0d85ea-nova-cells-global-config-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellkhtf8\" (UID: \"dd84ed43-c477-4b83-a735-142efa0d85ea\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellkhtf8" Oct 05 09:28:04 crc kubenswrapper[4846]: I1005 09:28:04.253407 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rvk6m\" (UniqueName: \"kubernetes.io/projected/dd84ed43-c477-4b83-a735-142efa0d85ea-kube-api-access-rvk6m\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellkhtf8\" (UID: \"dd84ed43-c477-4b83-a735-142efa0d85ea\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellkhtf8" Oct 05 09:28:04 crc kubenswrapper[4846]: I1005 09:28:04.253601 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/dd84ed43-c477-4b83-a735-142efa0d85ea-nova-cell1-compute-config-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellkhtf8\" (UID: \"dd84ed43-c477-4b83-a735-142efa0d85ea\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellkhtf8" Oct 05 09:28:04 crc kubenswrapper[4846]: I1005 09:28:04.254645 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/dd84ed43-c477-4b83-a735-142efa0d85ea-nova-cells-global-config-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellkhtf8\" (UID: \"dd84ed43-c477-4b83-a735-142efa0d85ea\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellkhtf8" Oct 05 09:28:04 crc kubenswrapper[4846]: I1005 09:28:04.258001 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/dd84ed43-c477-4b83-a735-142efa0d85ea-inventory\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellkhtf8\" (UID: \"dd84ed43-c477-4b83-a735-142efa0d85ea\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellkhtf8" Oct 05 09:28:04 crc kubenswrapper[4846]: I1005 09:28:04.258589 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/dd84ed43-c477-4b83-a735-142efa0d85ea-nova-cell1-compute-config-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellkhtf8\" (UID: \"dd84ed43-c477-4b83-a735-142efa0d85ea\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellkhtf8" Oct 05 09:28:04 crc kubenswrapper[4846]: I1005 09:28:04.259113 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dd84ed43-c477-4b83-a735-142efa0d85ea-nova-cell1-combined-ca-bundle\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellkhtf8\" (UID: \"dd84ed43-c477-4b83-a735-142efa0d85ea\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellkhtf8" Oct 05 09:28:04 crc kubenswrapper[4846]: I1005 09:28:04.259115 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/dd84ed43-c477-4b83-a735-142efa0d85ea-ssh-key\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellkhtf8\" (UID: \"dd84ed43-c477-4b83-a735-142efa0d85ea\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellkhtf8" Oct 05 09:28:04 crc kubenswrapper[4846]: I1005 09:28:04.259866 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/dd84ed43-c477-4b83-a735-142efa0d85ea-nova-cell1-compute-config-1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellkhtf8\" (UID: \"dd84ed43-c477-4b83-a735-142efa0d85ea\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellkhtf8" Oct 05 09:28:04 crc kubenswrapper[4846]: I1005 09:28:04.263096 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/dd84ed43-c477-4b83-a735-142efa0d85ea-nova-migration-ssh-key-1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellkhtf8\" (UID: \"dd84ed43-c477-4b83-a735-142efa0d85ea\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellkhtf8" Oct 05 09:28:04 crc kubenswrapper[4846]: I1005 09:28:04.271809 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/dd84ed43-c477-4b83-a735-142efa0d85ea-nova-migration-ssh-key-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellkhtf8\" (UID: \"dd84ed43-c477-4b83-a735-142efa0d85ea\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellkhtf8" Oct 05 09:28:04 crc kubenswrapper[4846]: I1005 09:28:04.274223 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rvk6m\" (UniqueName: \"kubernetes.io/projected/dd84ed43-c477-4b83-a735-142efa0d85ea-kube-api-access-rvk6m\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellkhtf8\" (UID: \"dd84ed43-c477-4b83-a735-142efa0d85ea\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellkhtf8" Oct 05 09:28:04 crc kubenswrapper[4846]: I1005 09:28:04.356928 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellkhtf8" Oct 05 09:28:04 crc kubenswrapper[4846]: I1005 09:28:04.892551 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellkhtf8"] Oct 05 09:28:05 crc kubenswrapper[4846]: W1005 09:28:05.601609 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddd84ed43_c477_4b83_a735_142efa0d85ea.slice/crio-691967b9eb1be739fd4facde8e9d71385fb3afac95d4c22399d605f08ab8917b WatchSource:0}: Error finding container 691967b9eb1be739fd4facde8e9d71385fb3afac95d4c22399d605f08ab8917b: Status 404 returned error can't find the container with id 691967b9eb1be739fd4facde8e9d71385fb3afac95d4c22399d605f08ab8917b Oct 05 09:28:05 crc kubenswrapper[4846]: I1005 09:28:05.918783 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellkhtf8" event={"ID":"dd84ed43-c477-4b83-a735-142efa0d85ea","Type":"ContainerStarted","Data":"691967b9eb1be739fd4facde8e9d71385fb3afac95d4c22399d605f08ab8917b"} Oct 05 09:28:06 crc kubenswrapper[4846]: I1005 09:28:06.928711 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellkhtf8" event={"ID":"dd84ed43-c477-4b83-a735-142efa0d85ea","Type":"ContainerStarted","Data":"d048a1370172c453d06d63c92c5bb1ab5050a1f599aa0807d80874a1c341b808"} Oct 05 09:28:06 crc kubenswrapper[4846]: I1005 09:28:06.954968 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellkhtf8" podStartSLOduration=2.546825922 podStartE2EDuration="2.954903058s" podCreationTimestamp="2025-10-05 09:28:04 +0000 UTC" firstStartedPulling="2025-10-05 09:28:05.607390959 +0000 UTC m=+9607.848243734" lastFinishedPulling="2025-10-05 09:28:06.015468095 +0000 UTC m=+9608.256320870" observedRunningTime="2025-10-05 09:28:06.943893981 +0000 UTC m=+9609.184746786" watchObservedRunningTime="2025-10-05 09:28:06.954903058 +0000 UTC m=+9609.195755843" Oct 05 09:30:00 crc kubenswrapper[4846]: I1005 09:30:00.176467 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29327610-ppx5h"] Oct 05 09:30:00 crc kubenswrapper[4846]: I1005 09:30:00.179662 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29327610-ppx5h" Oct 05 09:30:00 crc kubenswrapper[4846]: I1005 09:30:00.182692 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 05 09:30:00 crc kubenswrapper[4846]: I1005 09:30:00.182901 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 05 09:30:00 crc kubenswrapper[4846]: I1005 09:30:00.191075 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29327610-ppx5h"] Oct 05 09:30:00 crc kubenswrapper[4846]: I1005 09:30:00.315012 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/885c33bf-3fb1-4e88-b5af-c4e9daaa9225-secret-volume\") pod \"collect-profiles-29327610-ppx5h\" (UID: \"885c33bf-3fb1-4e88-b5af-c4e9daaa9225\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327610-ppx5h" Oct 05 09:30:00 crc kubenswrapper[4846]: I1005 09:30:00.315262 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bpqx4\" (UniqueName: \"kubernetes.io/projected/885c33bf-3fb1-4e88-b5af-c4e9daaa9225-kube-api-access-bpqx4\") pod \"collect-profiles-29327610-ppx5h\" (UID: \"885c33bf-3fb1-4e88-b5af-c4e9daaa9225\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327610-ppx5h" Oct 05 09:30:00 crc kubenswrapper[4846]: I1005 09:30:00.315373 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/885c33bf-3fb1-4e88-b5af-c4e9daaa9225-config-volume\") pod \"collect-profiles-29327610-ppx5h\" (UID: \"885c33bf-3fb1-4e88-b5af-c4e9daaa9225\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327610-ppx5h" Oct 05 09:30:00 crc kubenswrapper[4846]: I1005 09:30:00.422598 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/885c33bf-3fb1-4e88-b5af-c4e9daaa9225-config-volume\") pod \"collect-profiles-29327610-ppx5h\" (UID: \"885c33bf-3fb1-4e88-b5af-c4e9daaa9225\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327610-ppx5h" Oct 05 09:30:00 crc kubenswrapper[4846]: I1005 09:30:00.422844 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/885c33bf-3fb1-4e88-b5af-c4e9daaa9225-secret-volume\") pod \"collect-profiles-29327610-ppx5h\" (UID: \"885c33bf-3fb1-4e88-b5af-c4e9daaa9225\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327610-ppx5h" Oct 05 09:30:00 crc kubenswrapper[4846]: I1005 09:30:00.423205 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bpqx4\" (UniqueName: \"kubernetes.io/projected/885c33bf-3fb1-4e88-b5af-c4e9daaa9225-kube-api-access-bpqx4\") pod \"collect-profiles-29327610-ppx5h\" (UID: \"885c33bf-3fb1-4e88-b5af-c4e9daaa9225\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327610-ppx5h" Oct 05 09:30:00 crc kubenswrapper[4846]: I1005 09:30:00.425605 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/885c33bf-3fb1-4e88-b5af-c4e9daaa9225-config-volume\") pod \"collect-profiles-29327610-ppx5h\" (UID: \"885c33bf-3fb1-4e88-b5af-c4e9daaa9225\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327610-ppx5h" Oct 05 09:30:00 crc kubenswrapper[4846]: I1005 09:30:00.430522 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/885c33bf-3fb1-4e88-b5af-c4e9daaa9225-secret-volume\") pod \"collect-profiles-29327610-ppx5h\" (UID: \"885c33bf-3fb1-4e88-b5af-c4e9daaa9225\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327610-ppx5h" Oct 05 09:30:00 crc kubenswrapper[4846]: I1005 09:30:00.449683 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bpqx4\" (UniqueName: \"kubernetes.io/projected/885c33bf-3fb1-4e88-b5af-c4e9daaa9225-kube-api-access-bpqx4\") pod \"collect-profiles-29327610-ppx5h\" (UID: \"885c33bf-3fb1-4e88-b5af-c4e9daaa9225\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327610-ppx5h" Oct 05 09:30:00 crc kubenswrapper[4846]: I1005 09:30:00.515593 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29327610-ppx5h" Oct 05 09:30:01 crc kubenswrapper[4846]: I1005 09:30:01.044868 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29327610-ppx5h"] Oct 05 09:30:01 crc kubenswrapper[4846]: I1005 09:30:01.395823 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29327610-ppx5h" event={"ID":"885c33bf-3fb1-4e88-b5af-c4e9daaa9225","Type":"ContainerStarted","Data":"3931819b64f771c962f4e227b9a6413253fd645107a48dd74373e2e0bbea1f03"} Oct 05 09:30:01 crc kubenswrapper[4846]: I1005 09:30:01.396157 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29327610-ppx5h" event={"ID":"885c33bf-3fb1-4e88-b5af-c4e9daaa9225","Type":"ContainerStarted","Data":"65b10fa8e095e5f90dd39168f67597adab8b8656cbd24a8346e386553cd37a44"} Oct 05 09:30:01 crc kubenswrapper[4846]: I1005 09:30:01.424828 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29327610-ppx5h" podStartSLOduration=1.42480665 podStartE2EDuration="1.42480665s" podCreationTimestamp="2025-10-05 09:30:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-05 09:30:01.413491915 +0000 UTC m=+9723.654344700" watchObservedRunningTime="2025-10-05 09:30:01.42480665 +0000 UTC m=+9723.665659435" Oct 05 09:30:02 crc kubenswrapper[4846]: I1005 09:30:02.411878 4846 generic.go:334] "Generic (PLEG): container finished" podID="885c33bf-3fb1-4e88-b5af-c4e9daaa9225" containerID="3931819b64f771c962f4e227b9a6413253fd645107a48dd74373e2e0bbea1f03" exitCode=0 Oct 05 09:30:02 crc kubenswrapper[4846]: I1005 09:30:02.412072 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29327610-ppx5h" event={"ID":"885c33bf-3fb1-4e88-b5af-c4e9daaa9225","Type":"ContainerDied","Data":"3931819b64f771c962f4e227b9a6413253fd645107a48dd74373e2e0bbea1f03"} Oct 05 09:30:03 crc kubenswrapper[4846]: I1005 09:30:03.799000 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29327610-ppx5h" Oct 05 09:30:03 crc kubenswrapper[4846]: I1005 09:30:03.908337 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/885c33bf-3fb1-4e88-b5af-c4e9daaa9225-config-volume\") pod \"885c33bf-3fb1-4e88-b5af-c4e9daaa9225\" (UID: \"885c33bf-3fb1-4e88-b5af-c4e9daaa9225\") " Oct 05 09:30:03 crc kubenswrapper[4846]: I1005 09:30:03.908471 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bpqx4\" (UniqueName: \"kubernetes.io/projected/885c33bf-3fb1-4e88-b5af-c4e9daaa9225-kube-api-access-bpqx4\") pod \"885c33bf-3fb1-4e88-b5af-c4e9daaa9225\" (UID: \"885c33bf-3fb1-4e88-b5af-c4e9daaa9225\") " Oct 05 09:30:03 crc kubenswrapper[4846]: I1005 09:30:03.908534 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/885c33bf-3fb1-4e88-b5af-c4e9daaa9225-secret-volume\") pod \"885c33bf-3fb1-4e88-b5af-c4e9daaa9225\" (UID: \"885c33bf-3fb1-4e88-b5af-c4e9daaa9225\") " Oct 05 09:30:03 crc kubenswrapper[4846]: I1005 09:30:03.909474 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/885c33bf-3fb1-4e88-b5af-c4e9daaa9225-config-volume" (OuterVolumeSpecName: "config-volume") pod "885c33bf-3fb1-4e88-b5af-c4e9daaa9225" (UID: "885c33bf-3fb1-4e88-b5af-c4e9daaa9225"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 09:30:03 crc kubenswrapper[4846]: I1005 09:30:03.933414 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/885c33bf-3fb1-4e88-b5af-c4e9daaa9225-kube-api-access-bpqx4" (OuterVolumeSpecName: "kube-api-access-bpqx4") pod "885c33bf-3fb1-4e88-b5af-c4e9daaa9225" (UID: "885c33bf-3fb1-4e88-b5af-c4e9daaa9225"). InnerVolumeSpecName "kube-api-access-bpqx4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 09:30:03 crc kubenswrapper[4846]: I1005 09:30:03.933575 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/885c33bf-3fb1-4e88-b5af-c4e9daaa9225-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "885c33bf-3fb1-4e88-b5af-c4e9daaa9225" (UID: "885c33bf-3fb1-4e88-b5af-c4e9daaa9225"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 09:30:04 crc kubenswrapper[4846]: I1005 09:30:04.014380 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bpqx4\" (UniqueName: \"kubernetes.io/projected/885c33bf-3fb1-4e88-b5af-c4e9daaa9225-kube-api-access-bpqx4\") on node \"crc\" DevicePath \"\"" Oct 05 09:30:04 crc kubenswrapper[4846]: I1005 09:30:04.014668 4846 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/885c33bf-3fb1-4e88-b5af-c4e9daaa9225-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 05 09:30:04 crc kubenswrapper[4846]: I1005 09:30:04.014679 4846 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/885c33bf-3fb1-4e88-b5af-c4e9daaa9225-config-volume\") on node \"crc\" DevicePath \"\"" Oct 05 09:30:04 crc kubenswrapper[4846]: I1005 09:30:04.440834 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29327610-ppx5h" event={"ID":"885c33bf-3fb1-4e88-b5af-c4e9daaa9225","Type":"ContainerDied","Data":"65b10fa8e095e5f90dd39168f67597adab8b8656cbd24a8346e386553cd37a44"} Oct 05 09:30:04 crc kubenswrapper[4846]: I1005 09:30:04.440882 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="65b10fa8e095e5f90dd39168f67597adab8b8656cbd24a8346e386553cd37a44" Oct 05 09:30:04 crc kubenswrapper[4846]: I1005 09:30:04.440948 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29327610-ppx5h" Oct 05 09:30:04 crc kubenswrapper[4846]: I1005 09:30:04.520630 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29327565-p7hbh"] Oct 05 09:30:04 crc kubenswrapper[4846]: I1005 09:30:04.520675 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29327565-p7hbh"] Oct 05 09:30:06 crc kubenswrapper[4846]: I1005 09:30:06.518960 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e9a16960-8bb2-43a3-bbc4-70a135eb2449" path="/var/lib/kubelet/pods/e9a16960-8bb2-43a3-bbc4-70a135eb2449/volumes" Oct 05 09:30:16 crc kubenswrapper[4846]: I1005 09:30:16.720608 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-222g8"] Oct 05 09:30:16 crc kubenswrapper[4846]: E1005 09:30:16.721900 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="885c33bf-3fb1-4e88-b5af-c4e9daaa9225" containerName="collect-profiles" Oct 05 09:30:16 crc kubenswrapper[4846]: I1005 09:30:16.721920 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="885c33bf-3fb1-4e88-b5af-c4e9daaa9225" containerName="collect-profiles" Oct 05 09:30:16 crc kubenswrapper[4846]: I1005 09:30:16.722269 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="885c33bf-3fb1-4e88-b5af-c4e9daaa9225" containerName="collect-profiles" Oct 05 09:30:16 crc kubenswrapper[4846]: I1005 09:30:16.724968 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-222g8" Oct 05 09:30:16 crc kubenswrapper[4846]: I1005 09:30:16.736805 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-222g8"] Oct 05 09:30:16 crc kubenswrapper[4846]: I1005 09:30:16.827302 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8a7dd7cc-03aa-4e36-b6f8-8db7c32ac209-utilities\") pod \"redhat-marketplace-222g8\" (UID: \"8a7dd7cc-03aa-4e36-b6f8-8db7c32ac209\") " pod="openshift-marketplace/redhat-marketplace-222g8" Oct 05 09:30:16 crc kubenswrapper[4846]: I1005 09:30:16.827409 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fd92g\" (UniqueName: \"kubernetes.io/projected/8a7dd7cc-03aa-4e36-b6f8-8db7c32ac209-kube-api-access-fd92g\") pod \"redhat-marketplace-222g8\" (UID: \"8a7dd7cc-03aa-4e36-b6f8-8db7c32ac209\") " pod="openshift-marketplace/redhat-marketplace-222g8" Oct 05 09:30:16 crc kubenswrapper[4846]: I1005 09:30:16.827442 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8a7dd7cc-03aa-4e36-b6f8-8db7c32ac209-catalog-content\") pod \"redhat-marketplace-222g8\" (UID: \"8a7dd7cc-03aa-4e36-b6f8-8db7c32ac209\") " pod="openshift-marketplace/redhat-marketplace-222g8" Oct 05 09:30:16 crc kubenswrapper[4846]: I1005 09:30:16.929211 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8a7dd7cc-03aa-4e36-b6f8-8db7c32ac209-utilities\") pod \"redhat-marketplace-222g8\" (UID: \"8a7dd7cc-03aa-4e36-b6f8-8db7c32ac209\") " pod="openshift-marketplace/redhat-marketplace-222g8" Oct 05 09:30:16 crc kubenswrapper[4846]: I1005 09:30:16.929355 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fd92g\" (UniqueName: \"kubernetes.io/projected/8a7dd7cc-03aa-4e36-b6f8-8db7c32ac209-kube-api-access-fd92g\") pod \"redhat-marketplace-222g8\" (UID: \"8a7dd7cc-03aa-4e36-b6f8-8db7c32ac209\") " pod="openshift-marketplace/redhat-marketplace-222g8" Oct 05 09:30:16 crc kubenswrapper[4846]: I1005 09:30:16.929394 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8a7dd7cc-03aa-4e36-b6f8-8db7c32ac209-catalog-content\") pod \"redhat-marketplace-222g8\" (UID: \"8a7dd7cc-03aa-4e36-b6f8-8db7c32ac209\") " pod="openshift-marketplace/redhat-marketplace-222g8" Oct 05 09:30:16 crc kubenswrapper[4846]: I1005 09:30:16.929738 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8a7dd7cc-03aa-4e36-b6f8-8db7c32ac209-utilities\") pod \"redhat-marketplace-222g8\" (UID: \"8a7dd7cc-03aa-4e36-b6f8-8db7c32ac209\") " pod="openshift-marketplace/redhat-marketplace-222g8" Oct 05 09:30:16 crc kubenswrapper[4846]: I1005 09:30:16.931232 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8a7dd7cc-03aa-4e36-b6f8-8db7c32ac209-catalog-content\") pod \"redhat-marketplace-222g8\" (UID: \"8a7dd7cc-03aa-4e36-b6f8-8db7c32ac209\") " pod="openshift-marketplace/redhat-marketplace-222g8" Oct 05 09:30:16 crc kubenswrapper[4846]: I1005 09:30:16.953224 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fd92g\" (UniqueName: \"kubernetes.io/projected/8a7dd7cc-03aa-4e36-b6f8-8db7c32ac209-kube-api-access-fd92g\") pod \"redhat-marketplace-222g8\" (UID: \"8a7dd7cc-03aa-4e36-b6f8-8db7c32ac209\") " pod="openshift-marketplace/redhat-marketplace-222g8" Oct 05 09:30:17 crc kubenswrapper[4846]: I1005 09:30:17.061390 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-222g8" Oct 05 09:30:17 crc kubenswrapper[4846]: I1005 09:30:17.549678 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-222g8"] Oct 05 09:30:17 crc kubenswrapper[4846]: I1005 09:30:17.624848 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-222g8" event={"ID":"8a7dd7cc-03aa-4e36-b6f8-8db7c32ac209","Type":"ContainerStarted","Data":"d12affc3cac753d4a2668342cf35844fc6a41e0c4c1070b10aa9aac6ced3ca17"} Oct 05 09:30:18 crc kubenswrapper[4846]: I1005 09:30:18.640584 4846 generic.go:334] "Generic (PLEG): container finished" podID="8a7dd7cc-03aa-4e36-b6f8-8db7c32ac209" containerID="88baf59fffe87b2e9e48a455622f53169ed742c1103bfe95f259cd27d7ad3bab" exitCode=0 Oct 05 09:30:18 crc kubenswrapper[4846]: I1005 09:30:18.640697 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-222g8" event={"ID":"8a7dd7cc-03aa-4e36-b6f8-8db7c32ac209","Type":"ContainerDied","Data":"88baf59fffe87b2e9e48a455622f53169ed742c1103bfe95f259cd27d7ad3bab"} Oct 05 09:30:18 crc kubenswrapper[4846]: I1005 09:30:18.644748 4846 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 05 09:30:19 crc kubenswrapper[4846]: I1005 09:30:19.653934 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-222g8" event={"ID":"8a7dd7cc-03aa-4e36-b6f8-8db7c32ac209","Type":"ContainerStarted","Data":"3f731088f119fe162d057f0748043f0ce19e98cfe41bb7f6da99e8ad520f77b5"} Oct 05 09:30:20 crc kubenswrapper[4846]: I1005 09:30:20.668354 4846 generic.go:334] "Generic (PLEG): container finished" podID="8a7dd7cc-03aa-4e36-b6f8-8db7c32ac209" containerID="3f731088f119fe162d057f0748043f0ce19e98cfe41bb7f6da99e8ad520f77b5" exitCode=0 Oct 05 09:30:20 crc kubenswrapper[4846]: I1005 09:30:20.668718 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-222g8" event={"ID":"8a7dd7cc-03aa-4e36-b6f8-8db7c32ac209","Type":"ContainerDied","Data":"3f731088f119fe162d057f0748043f0ce19e98cfe41bb7f6da99e8ad520f77b5"} Oct 05 09:30:21 crc kubenswrapper[4846]: I1005 09:30:21.714812 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-222g8" event={"ID":"8a7dd7cc-03aa-4e36-b6f8-8db7c32ac209","Type":"ContainerStarted","Data":"299cd03412c600a3afe59456638b54587c9fb5ef50f12950faf9f4927e3fdc60"} Oct 05 09:30:21 crc kubenswrapper[4846]: I1005 09:30:21.733792 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-222g8" podStartSLOduration=3.262421567 podStartE2EDuration="5.73377418s" podCreationTimestamp="2025-10-05 09:30:16 +0000 UTC" firstStartedPulling="2025-10-05 09:30:18.644329453 +0000 UTC m=+9740.885182238" lastFinishedPulling="2025-10-05 09:30:21.115682066 +0000 UTC m=+9743.356534851" observedRunningTime="2025-10-05 09:30:21.732124377 +0000 UTC m=+9743.972977162" watchObservedRunningTime="2025-10-05 09:30:21.73377418 +0000 UTC m=+9743.974626965" Oct 05 09:30:23 crc kubenswrapper[4846]: I1005 09:30:23.325354 4846 patch_prober.go:28] interesting pod/machine-config-daemon-fscvf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 09:30:23 crc kubenswrapper[4846]: I1005 09:30:23.325898 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 09:30:26 crc kubenswrapper[4846]: I1005 09:30:26.274964 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-z67q9"] Oct 05 09:30:26 crc kubenswrapper[4846]: I1005 09:30:26.279339 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-z67q9" Oct 05 09:30:26 crc kubenswrapper[4846]: I1005 09:30:26.294842 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-z67q9"] Oct 05 09:30:26 crc kubenswrapper[4846]: I1005 09:30:26.438695 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8jxbv\" (UniqueName: \"kubernetes.io/projected/b376bc05-9c1a-40ce-b237-f94b4774bcc8-kube-api-access-8jxbv\") pod \"community-operators-z67q9\" (UID: \"b376bc05-9c1a-40ce-b237-f94b4774bcc8\") " pod="openshift-marketplace/community-operators-z67q9" Oct 05 09:30:26 crc kubenswrapper[4846]: I1005 09:30:26.438834 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b376bc05-9c1a-40ce-b237-f94b4774bcc8-utilities\") pod \"community-operators-z67q9\" (UID: \"b376bc05-9c1a-40ce-b237-f94b4774bcc8\") " pod="openshift-marketplace/community-operators-z67q9" Oct 05 09:30:26 crc kubenswrapper[4846]: I1005 09:30:26.438931 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b376bc05-9c1a-40ce-b237-f94b4774bcc8-catalog-content\") pod \"community-operators-z67q9\" (UID: \"b376bc05-9c1a-40ce-b237-f94b4774bcc8\") " pod="openshift-marketplace/community-operators-z67q9" Oct 05 09:30:26 crc kubenswrapper[4846]: I1005 09:30:26.541412 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8jxbv\" (UniqueName: \"kubernetes.io/projected/b376bc05-9c1a-40ce-b237-f94b4774bcc8-kube-api-access-8jxbv\") pod \"community-operators-z67q9\" (UID: \"b376bc05-9c1a-40ce-b237-f94b4774bcc8\") " pod="openshift-marketplace/community-operators-z67q9" Oct 05 09:30:26 crc kubenswrapper[4846]: I1005 09:30:26.541483 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b376bc05-9c1a-40ce-b237-f94b4774bcc8-utilities\") pod \"community-operators-z67q9\" (UID: \"b376bc05-9c1a-40ce-b237-f94b4774bcc8\") " pod="openshift-marketplace/community-operators-z67q9" Oct 05 09:30:26 crc kubenswrapper[4846]: I1005 09:30:26.541546 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b376bc05-9c1a-40ce-b237-f94b4774bcc8-catalog-content\") pod \"community-operators-z67q9\" (UID: \"b376bc05-9c1a-40ce-b237-f94b4774bcc8\") " pod="openshift-marketplace/community-operators-z67q9" Oct 05 09:30:26 crc kubenswrapper[4846]: I1005 09:30:26.542121 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b376bc05-9c1a-40ce-b237-f94b4774bcc8-catalog-content\") pod \"community-operators-z67q9\" (UID: \"b376bc05-9c1a-40ce-b237-f94b4774bcc8\") " pod="openshift-marketplace/community-operators-z67q9" Oct 05 09:30:26 crc kubenswrapper[4846]: I1005 09:30:26.542129 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b376bc05-9c1a-40ce-b237-f94b4774bcc8-utilities\") pod \"community-operators-z67q9\" (UID: \"b376bc05-9c1a-40ce-b237-f94b4774bcc8\") " pod="openshift-marketplace/community-operators-z67q9" Oct 05 09:30:26 crc kubenswrapper[4846]: I1005 09:30:26.570722 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8jxbv\" (UniqueName: \"kubernetes.io/projected/b376bc05-9c1a-40ce-b237-f94b4774bcc8-kube-api-access-8jxbv\") pod \"community-operators-z67q9\" (UID: \"b376bc05-9c1a-40ce-b237-f94b4774bcc8\") " pod="openshift-marketplace/community-operators-z67q9" Oct 05 09:30:26 crc kubenswrapper[4846]: I1005 09:30:26.644593 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-z67q9" Oct 05 09:30:27 crc kubenswrapper[4846]: I1005 09:30:27.062356 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-222g8" Oct 05 09:30:27 crc kubenswrapper[4846]: I1005 09:30:27.062613 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-222g8" Oct 05 09:30:27 crc kubenswrapper[4846]: I1005 09:30:27.118041 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-222g8" Oct 05 09:30:27 crc kubenswrapper[4846]: I1005 09:30:27.261690 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-z67q9"] Oct 05 09:30:27 crc kubenswrapper[4846]: I1005 09:30:27.795504 4846 generic.go:334] "Generic (PLEG): container finished" podID="b376bc05-9c1a-40ce-b237-f94b4774bcc8" containerID="164a6d437682a03c44dbb9aff8b9128917e2b84f0d259817a307856bd0e39ba1" exitCode=0 Oct 05 09:30:27 crc kubenswrapper[4846]: I1005 09:30:27.795554 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-z67q9" event={"ID":"b376bc05-9c1a-40ce-b237-f94b4774bcc8","Type":"ContainerDied","Data":"164a6d437682a03c44dbb9aff8b9128917e2b84f0d259817a307856bd0e39ba1"} Oct 05 09:30:27 crc kubenswrapper[4846]: I1005 09:30:27.795772 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-z67q9" event={"ID":"b376bc05-9c1a-40ce-b237-f94b4774bcc8","Type":"ContainerStarted","Data":"91ede71fd5be2bc8b6fa98b5dc86122a3823d1e74fa3f275bb533e3f9402a275"} Oct 05 09:30:27 crc kubenswrapper[4846]: I1005 09:30:27.854634 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-222g8" Oct 05 09:30:28 crc kubenswrapper[4846]: I1005 09:30:28.807005 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-z67q9" event={"ID":"b376bc05-9c1a-40ce-b237-f94b4774bcc8","Type":"ContainerStarted","Data":"d26cfe4546858ac1eaefc2455e567f5484146ead8e2eea03d1db0edb25631650"} Oct 05 09:30:29 crc kubenswrapper[4846]: I1005 09:30:29.461011 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-222g8"] Oct 05 09:30:29 crc kubenswrapper[4846]: I1005 09:30:29.825691 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-222g8" podUID="8a7dd7cc-03aa-4e36-b6f8-8db7c32ac209" containerName="registry-server" containerID="cri-o://299cd03412c600a3afe59456638b54587c9fb5ef50f12950faf9f4927e3fdc60" gracePeriod=2 Oct 05 09:30:30 crc kubenswrapper[4846]: I1005 09:30:30.472695 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-222g8" Oct 05 09:30:30 crc kubenswrapper[4846]: I1005 09:30:30.665347 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8a7dd7cc-03aa-4e36-b6f8-8db7c32ac209-utilities\") pod \"8a7dd7cc-03aa-4e36-b6f8-8db7c32ac209\" (UID: \"8a7dd7cc-03aa-4e36-b6f8-8db7c32ac209\") " Oct 05 09:30:30 crc kubenswrapper[4846]: I1005 09:30:30.665443 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8a7dd7cc-03aa-4e36-b6f8-8db7c32ac209-catalog-content\") pod \"8a7dd7cc-03aa-4e36-b6f8-8db7c32ac209\" (UID: \"8a7dd7cc-03aa-4e36-b6f8-8db7c32ac209\") " Oct 05 09:30:30 crc kubenswrapper[4846]: I1005 09:30:30.665513 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fd92g\" (UniqueName: \"kubernetes.io/projected/8a7dd7cc-03aa-4e36-b6f8-8db7c32ac209-kube-api-access-fd92g\") pod \"8a7dd7cc-03aa-4e36-b6f8-8db7c32ac209\" (UID: \"8a7dd7cc-03aa-4e36-b6f8-8db7c32ac209\") " Oct 05 09:30:30 crc kubenswrapper[4846]: I1005 09:30:30.666709 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8a7dd7cc-03aa-4e36-b6f8-8db7c32ac209-utilities" (OuterVolumeSpecName: "utilities") pod "8a7dd7cc-03aa-4e36-b6f8-8db7c32ac209" (UID: "8a7dd7cc-03aa-4e36-b6f8-8db7c32ac209"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 09:30:30 crc kubenswrapper[4846]: I1005 09:30:30.677651 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8a7dd7cc-03aa-4e36-b6f8-8db7c32ac209-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8a7dd7cc-03aa-4e36-b6f8-8db7c32ac209" (UID: "8a7dd7cc-03aa-4e36-b6f8-8db7c32ac209"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 09:30:30 crc kubenswrapper[4846]: I1005 09:30:30.678405 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8a7dd7cc-03aa-4e36-b6f8-8db7c32ac209-kube-api-access-fd92g" (OuterVolumeSpecName: "kube-api-access-fd92g") pod "8a7dd7cc-03aa-4e36-b6f8-8db7c32ac209" (UID: "8a7dd7cc-03aa-4e36-b6f8-8db7c32ac209"). InnerVolumeSpecName "kube-api-access-fd92g". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 09:30:30 crc kubenswrapper[4846]: I1005 09:30:30.768304 4846 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8a7dd7cc-03aa-4e36-b6f8-8db7c32ac209-utilities\") on node \"crc\" DevicePath \"\"" Oct 05 09:30:30 crc kubenswrapper[4846]: I1005 09:30:30.768339 4846 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8a7dd7cc-03aa-4e36-b6f8-8db7c32ac209-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 05 09:30:30 crc kubenswrapper[4846]: I1005 09:30:30.768353 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fd92g\" (UniqueName: \"kubernetes.io/projected/8a7dd7cc-03aa-4e36-b6f8-8db7c32ac209-kube-api-access-fd92g\") on node \"crc\" DevicePath \"\"" Oct 05 09:30:30 crc kubenswrapper[4846]: I1005 09:30:30.836942 4846 generic.go:334] "Generic (PLEG): container finished" podID="b376bc05-9c1a-40ce-b237-f94b4774bcc8" containerID="d26cfe4546858ac1eaefc2455e567f5484146ead8e2eea03d1db0edb25631650" exitCode=0 Oct 05 09:30:30 crc kubenswrapper[4846]: I1005 09:30:30.837008 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-z67q9" event={"ID":"b376bc05-9c1a-40ce-b237-f94b4774bcc8","Type":"ContainerDied","Data":"d26cfe4546858ac1eaefc2455e567f5484146ead8e2eea03d1db0edb25631650"} Oct 05 09:30:30 crc kubenswrapper[4846]: I1005 09:30:30.841438 4846 generic.go:334] "Generic (PLEG): container finished" podID="8a7dd7cc-03aa-4e36-b6f8-8db7c32ac209" containerID="299cd03412c600a3afe59456638b54587c9fb5ef50f12950faf9f4927e3fdc60" exitCode=0 Oct 05 09:30:30 crc kubenswrapper[4846]: I1005 09:30:30.841478 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-222g8" event={"ID":"8a7dd7cc-03aa-4e36-b6f8-8db7c32ac209","Type":"ContainerDied","Data":"299cd03412c600a3afe59456638b54587c9fb5ef50f12950faf9f4927e3fdc60"} Oct 05 09:30:30 crc kubenswrapper[4846]: I1005 09:30:30.841505 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-222g8" event={"ID":"8a7dd7cc-03aa-4e36-b6f8-8db7c32ac209","Type":"ContainerDied","Data":"d12affc3cac753d4a2668342cf35844fc6a41e0c4c1070b10aa9aac6ced3ca17"} Oct 05 09:30:30 crc kubenswrapper[4846]: I1005 09:30:30.841526 4846 scope.go:117] "RemoveContainer" containerID="299cd03412c600a3afe59456638b54587c9fb5ef50f12950faf9f4927e3fdc60" Oct 05 09:30:30 crc kubenswrapper[4846]: I1005 09:30:30.841658 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-222g8" Oct 05 09:30:30 crc kubenswrapper[4846]: I1005 09:30:30.876472 4846 scope.go:117] "RemoveContainer" containerID="3f731088f119fe162d057f0748043f0ce19e98cfe41bb7f6da99e8ad520f77b5" Oct 05 09:30:30 crc kubenswrapper[4846]: I1005 09:30:30.888150 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-222g8"] Oct 05 09:30:30 crc kubenswrapper[4846]: I1005 09:30:30.896510 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-222g8"] Oct 05 09:30:30 crc kubenswrapper[4846]: I1005 09:30:30.912276 4846 scope.go:117] "RemoveContainer" containerID="88baf59fffe87b2e9e48a455622f53169ed742c1103bfe95f259cd27d7ad3bab" Oct 05 09:30:30 crc kubenswrapper[4846]: I1005 09:30:30.955424 4846 scope.go:117] "RemoveContainer" containerID="299cd03412c600a3afe59456638b54587c9fb5ef50f12950faf9f4927e3fdc60" Oct 05 09:30:30 crc kubenswrapper[4846]: E1005 09:30:30.955850 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"299cd03412c600a3afe59456638b54587c9fb5ef50f12950faf9f4927e3fdc60\": container with ID starting with 299cd03412c600a3afe59456638b54587c9fb5ef50f12950faf9f4927e3fdc60 not found: ID does not exist" containerID="299cd03412c600a3afe59456638b54587c9fb5ef50f12950faf9f4927e3fdc60" Oct 05 09:30:30 crc kubenswrapper[4846]: I1005 09:30:30.955879 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"299cd03412c600a3afe59456638b54587c9fb5ef50f12950faf9f4927e3fdc60"} err="failed to get container status \"299cd03412c600a3afe59456638b54587c9fb5ef50f12950faf9f4927e3fdc60\": rpc error: code = NotFound desc = could not find container \"299cd03412c600a3afe59456638b54587c9fb5ef50f12950faf9f4927e3fdc60\": container with ID starting with 299cd03412c600a3afe59456638b54587c9fb5ef50f12950faf9f4927e3fdc60 not found: ID does not exist" Oct 05 09:30:30 crc kubenswrapper[4846]: I1005 09:30:30.955904 4846 scope.go:117] "RemoveContainer" containerID="3f731088f119fe162d057f0748043f0ce19e98cfe41bb7f6da99e8ad520f77b5" Oct 05 09:30:30 crc kubenswrapper[4846]: E1005 09:30:30.956135 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3f731088f119fe162d057f0748043f0ce19e98cfe41bb7f6da99e8ad520f77b5\": container with ID starting with 3f731088f119fe162d057f0748043f0ce19e98cfe41bb7f6da99e8ad520f77b5 not found: ID does not exist" containerID="3f731088f119fe162d057f0748043f0ce19e98cfe41bb7f6da99e8ad520f77b5" Oct 05 09:30:30 crc kubenswrapper[4846]: I1005 09:30:30.956154 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3f731088f119fe162d057f0748043f0ce19e98cfe41bb7f6da99e8ad520f77b5"} err="failed to get container status \"3f731088f119fe162d057f0748043f0ce19e98cfe41bb7f6da99e8ad520f77b5\": rpc error: code = NotFound desc = could not find container \"3f731088f119fe162d057f0748043f0ce19e98cfe41bb7f6da99e8ad520f77b5\": container with ID starting with 3f731088f119fe162d057f0748043f0ce19e98cfe41bb7f6da99e8ad520f77b5 not found: ID does not exist" Oct 05 09:30:30 crc kubenswrapper[4846]: I1005 09:30:30.956165 4846 scope.go:117] "RemoveContainer" containerID="88baf59fffe87b2e9e48a455622f53169ed742c1103bfe95f259cd27d7ad3bab" Oct 05 09:30:30 crc kubenswrapper[4846]: E1005 09:30:30.956943 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"88baf59fffe87b2e9e48a455622f53169ed742c1103bfe95f259cd27d7ad3bab\": container with ID starting with 88baf59fffe87b2e9e48a455622f53169ed742c1103bfe95f259cd27d7ad3bab not found: ID does not exist" containerID="88baf59fffe87b2e9e48a455622f53169ed742c1103bfe95f259cd27d7ad3bab" Oct 05 09:30:30 crc kubenswrapper[4846]: I1005 09:30:30.956968 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"88baf59fffe87b2e9e48a455622f53169ed742c1103bfe95f259cd27d7ad3bab"} err="failed to get container status \"88baf59fffe87b2e9e48a455622f53169ed742c1103bfe95f259cd27d7ad3bab\": rpc error: code = NotFound desc = could not find container \"88baf59fffe87b2e9e48a455622f53169ed742c1103bfe95f259cd27d7ad3bab\": container with ID starting with 88baf59fffe87b2e9e48a455622f53169ed742c1103bfe95f259cd27d7ad3bab not found: ID does not exist" Oct 05 09:30:31 crc kubenswrapper[4846]: I1005 09:30:31.856074 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-z67q9" event={"ID":"b376bc05-9c1a-40ce-b237-f94b4774bcc8","Type":"ContainerStarted","Data":"5c5ceefdfe6c2c6c225fac0e472ce6bc816a8066dbf6dfa4aac780641fdb5387"} Oct 05 09:30:31 crc kubenswrapper[4846]: I1005 09:30:31.882992 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-z67q9" podStartSLOduration=2.454779961 podStartE2EDuration="5.882969058s" podCreationTimestamp="2025-10-05 09:30:26 +0000 UTC" firstStartedPulling="2025-10-05 09:30:27.797312421 +0000 UTC m=+9750.038165196" lastFinishedPulling="2025-10-05 09:30:31.225501518 +0000 UTC m=+9753.466354293" observedRunningTime="2025-10-05 09:30:31.876248993 +0000 UTC m=+9754.117101788" watchObservedRunningTime="2025-10-05 09:30:31.882969058 +0000 UTC m=+9754.123821843" Oct 05 09:30:32 crc kubenswrapper[4846]: I1005 09:30:32.512379 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8a7dd7cc-03aa-4e36-b6f8-8db7c32ac209" path="/var/lib/kubelet/pods/8a7dd7cc-03aa-4e36-b6f8-8db7c32ac209/volumes" Oct 05 09:30:36 crc kubenswrapper[4846]: I1005 09:30:36.644705 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-z67q9" Oct 05 09:30:36 crc kubenswrapper[4846]: I1005 09:30:36.645309 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-z67q9" Oct 05 09:30:36 crc kubenswrapper[4846]: I1005 09:30:36.706497 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-z67q9" Oct 05 09:30:36 crc kubenswrapper[4846]: I1005 09:30:36.988272 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-z67q9" Oct 05 09:30:37 crc kubenswrapper[4846]: I1005 09:30:37.060004 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-z67q9"] Oct 05 09:30:38 crc kubenswrapper[4846]: I1005 09:30:38.953322 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-z67q9" podUID="b376bc05-9c1a-40ce-b237-f94b4774bcc8" containerName="registry-server" containerID="cri-o://5c5ceefdfe6c2c6c225fac0e472ce6bc816a8066dbf6dfa4aac780641fdb5387" gracePeriod=2 Oct 05 09:30:39 crc kubenswrapper[4846]: I1005 09:30:39.506981 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-z67q9" Oct 05 09:30:39 crc kubenswrapper[4846]: I1005 09:30:39.672327 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b376bc05-9c1a-40ce-b237-f94b4774bcc8-catalog-content\") pod \"b376bc05-9c1a-40ce-b237-f94b4774bcc8\" (UID: \"b376bc05-9c1a-40ce-b237-f94b4774bcc8\") " Oct 05 09:30:39 crc kubenswrapper[4846]: I1005 09:30:39.672453 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8jxbv\" (UniqueName: \"kubernetes.io/projected/b376bc05-9c1a-40ce-b237-f94b4774bcc8-kube-api-access-8jxbv\") pod \"b376bc05-9c1a-40ce-b237-f94b4774bcc8\" (UID: \"b376bc05-9c1a-40ce-b237-f94b4774bcc8\") " Oct 05 09:30:39 crc kubenswrapper[4846]: I1005 09:30:39.672572 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b376bc05-9c1a-40ce-b237-f94b4774bcc8-utilities\") pod \"b376bc05-9c1a-40ce-b237-f94b4774bcc8\" (UID: \"b376bc05-9c1a-40ce-b237-f94b4774bcc8\") " Oct 05 09:30:39 crc kubenswrapper[4846]: I1005 09:30:39.673667 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b376bc05-9c1a-40ce-b237-f94b4774bcc8-utilities" (OuterVolumeSpecName: "utilities") pod "b376bc05-9c1a-40ce-b237-f94b4774bcc8" (UID: "b376bc05-9c1a-40ce-b237-f94b4774bcc8"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 09:30:39 crc kubenswrapper[4846]: I1005 09:30:39.678152 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b376bc05-9c1a-40ce-b237-f94b4774bcc8-kube-api-access-8jxbv" (OuterVolumeSpecName: "kube-api-access-8jxbv") pod "b376bc05-9c1a-40ce-b237-f94b4774bcc8" (UID: "b376bc05-9c1a-40ce-b237-f94b4774bcc8"). InnerVolumeSpecName "kube-api-access-8jxbv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 09:30:39 crc kubenswrapper[4846]: I1005 09:30:39.754523 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b376bc05-9c1a-40ce-b237-f94b4774bcc8-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b376bc05-9c1a-40ce-b237-f94b4774bcc8" (UID: "b376bc05-9c1a-40ce-b237-f94b4774bcc8"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 09:30:39 crc kubenswrapper[4846]: I1005 09:30:39.775160 4846 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b376bc05-9c1a-40ce-b237-f94b4774bcc8-utilities\") on node \"crc\" DevicePath \"\"" Oct 05 09:30:39 crc kubenswrapper[4846]: I1005 09:30:39.775205 4846 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b376bc05-9c1a-40ce-b237-f94b4774bcc8-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 05 09:30:39 crc kubenswrapper[4846]: I1005 09:30:39.775218 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8jxbv\" (UniqueName: \"kubernetes.io/projected/b376bc05-9c1a-40ce-b237-f94b4774bcc8-kube-api-access-8jxbv\") on node \"crc\" DevicePath \"\"" Oct 05 09:30:39 crc kubenswrapper[4846]: I1005 09:30:39.963067 4846 generic.go:334] "Generic (PLEG): container finished" podID="b376bc05-9c1a-40ce-b237-f94b4774bcc8" containerID="5c5ceefdfe6c2c6c225fac0e472ce6bc816a8066dbf6dfa4aac780641fdb5387" exitCode=0 Oct 05 09:30:39 crc kubenswrapper[4846]: I1005 09:30:39.963114 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-z67q9" event={"ID":"b376bc05-9c1a-40ce-b237-f94b4774bcc8","Type":"ContainerDied","Data":"5c5ceefdfe6c2c6c225fac0e472ce6bc816a8066dbf6dfa4aac780641fdb5387"} Oct 05 09:30:39 crc kubenswrapper[4846]: I1005 09:30:39.963144 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-z67q9" event={"ID":"b376bc05-9c1a-40ce-b237-f94b4774bcc8","Type":"ContainerDied","Data":"91ede71fd5be2bc8b6fa98b5dc86122a3823d1e74fa3f275bb533e3f9402a275"} Oct 05 09:30:39 crc kubenswrapper[4846]: I1005 09:30:39.963149 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-z67q9" Oct 05 09:30:39 crc kubenswrapper[4846]: I1005 09:30:39.963162 4846 scope.go:117] "RemoveContainer" containerID="5c5ceefdfe6c2c6c225fac0e472ce6bc816a8066dbf6dfa4aac780641fdb5387" Oct 05 09:30:39 crc kubenswrapper[4846]: I1005 09:30:39.992387 4846 scope.go:117] "RemoveContainer" containerID="d26cfe4546858ac1eaefc2455e567f5484146ead8e2eea03d1db0edb25631650" Oct 05 09:30:39 crc kubenswrapper[4846]: I1005 09:30:39.994555 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-z67q9"] Oct 05 09:30:40 crc kubenswrapper[4846]: I1005 09:30:40.015037 4846 scope.go:117] "RemoveContainer" containerID="164a6d437682a03c44dbb9aff8b9128917e2b84f0d259817a307856bd0e39ba1" Oct 05 09:30:40 crc kubenswrapper[4846]: I1005 09:30:40.016661 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-z67q9"] Oct 05 09:30:40 crc kubenswrapper[4846]: I1005 09:30:40.059856 4846 scope.go:117] "RemoveContainer" containerID="5c5ceefdfe6c2c6c225fac0e472ce6bc816a8066dbf6dfa4aac780641fdb5387" Oct 05 09:30:40 crc kubenswrapper[4846]: E1005 09:30:40.060280 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5c5ceefdfe6c2c6c225fac0e472ce6bc816a8066dbf6dfa4aac780641fdb5387\": container with ID starting with 5c5ceefdfe6c2c6c225fac0e472ce6bc816a8066dbf6dfa4aac780641fdb5387 not found: ID does not exist" containerID="5c5ceefdfe6c2c6c225fac0e472ce6bc816a8066dbf6dfa4aac780641fdb5387" Oct 05 09:30:40 crc kubenswrapper[4846]: I1005 09:30:40.060341 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5c5ceefdfe6c2c6c225fac0e472ce6bc816a8066dbf6dfa4aac780641fdb5387"} err="failed to get container status \"5c5ceefdfe6c2c6c225fac0e472ce6bc816a8066dbf6dfa4aac780641fdb5387\": rpc error: code = NotFound desc = could not find container \"5c5ceefdfe6c2c6c225fac0e472ce6bc816a8066dbf6dfa4aac780641fdb5387\": container with ID starting with 5c5ceefdfe6c2c6c225fac0e472ce6bc816a8066dbf6dfa4aac780641fdb5387 not found: ID does not exist" Oct 05 09:30:40 crc kubenswrapper[4846]: I1005 09:30:40.060383 4846 scope.go:117] "RemoveContainer" containerID="d26cfe4546858ac1eaefc2455e567f5484146ead8e2eea03d1db0edb25631650" Oct 05 09:30:40 crc kubenswrapper[4846]: E1005 09:30:40.060715 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d26cfe4546858ac1eaefc2455e567f5484146ead8e2eea03d1db0edb25631650\": container with ID starting with d26cfe4546858ac1eaefc2455e567f5484146ead8e2eea03d1db0edb25631650 not found: ID does not exist" containerID="d26cfe4546858ac1eaefc2455e567f5484146ead8e2eea03d1db0edb25631650" Oct 05 09:30:40 crc kubenswrapper[4846]: I1005 09:30:40.060749 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d26cfe4546858ac1eaefc2455e567f5484146ead8e2eea03d1db0edb25631650"} err="failed to get container status \"d26cfe4546858ac1eaefc2455e567f5484146ead8e2eea03d1db0edb25631650\": rpc error: code = NotFound desc = could not find container \"d26cfe4546858ac1eaefc2455e567f5484146ead8e2eea03d1db0edb25631650\": container with ID starting with d26cfe4546858ac1eaefc2455e567f5484146ead8e2eea03d1db0edb25631650 not found: ID does not exist" Oct 05 09:30:40 crc kubenswrapper[4846]: I1005 09:30:40.060772 4846 scope.go:117] "RemoveContainer" containerID="164a6d437682a03c44dbb9aff8b9128917e2b84f0d259817a307856bd0e39ba1" Oct 05 09:30:40 crc kubenswrapper[4846]: E1005 09:30:40.061085 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"164a6d437682a03c44dbb9aff8b9128917e2b84f0d259817a307856bd0e39ba1\": container with ID starting with 164a6d437682a03c44dbb9aff8b9128917e2b84f0d259817a307856bd0e39ba1 not found: ID does not exist" containerID="164a6d437682a03c44dbb9aff8b9128917e2b84f0d259817a307856bd0e39ba1" Oct 05 09:30:40 crc kubenswrapper[4846]: I1005 09:30:40.061155 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"164a6d437682a03c44dbb9aff8b9128917e2b84f0d259817a307856bd0e39ba1"} err="failed to get container status \"164a6d437682a03c44dbb9aff8b9128917e2b84f0d259817a307856bd0e39ba1\": rpc error: code = NotFound desc = could not find container \"164a6d437682a03c44dbb9aff8b9128917e2b84f0d259817a307856bd0e39ba1\": container with ID starting with 164a6d437682a03c44dbb9aff8b9128917e2b84f0d259817a307856bd0e39ba1 not found: ID does not exist" Oct 05 09:30:40 crc kubenswrapper[4846]: I1005 09:30:40.476622 4846 scope.go:117] "RemoveContainer" containerID="04048452dbf08a2f8c4be66289e063bae91b3f32184ebb728d5932fbd877ebaf" Oct 05 09:30:40 crc kubenswrapper[4846]: I1005 09:30:40.518901 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b376bc05-9c1a-40ce-b237-f94b4774bcc8" path="/var/lib/kubelet/pods/b376bc05-9c1a-40ce-b237-f94b4774bcc8/volumes" Oct 05 09:30:53 crc kubenswrapper[4846]: I1005 09:30:53.325554 4846 patch_prober.go:28] interesting pod/machine-config-daemon-fscvf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 09:30:53 crc kubenswrapper[4846]: I1005 09:30:53.326296 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 09:31:12 crc kubenswrapper[4846]: I1005 09:31:12.323807 4846 trace.go:236] Trace[1172700105]: "Calculate volume metrics of ovndbcluster-nb-etc-ovn for pod openstack/ovsdbserver-nb-2" (05-Oct-2025 09:31:11.172) (total time: 1151ms): Oct 05 09:31:12 crc kubenswrapper[4846]: Trace[1172700105]: [1.151630627s] [1.151630627s] END Oct 05 09:31:23 crc kubenswrapper[4846]: I1005 09:31:23.325055 4846 patch_prober.go:28] interesting pod/machine-config-daemon-fscvf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 09:31:23 crc kubenswrapper[4846]: I1005 09:31:23.325899 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 09:31:23 crc kubenswrapper[4846]: I1005 09:31:23.325946 4846 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" Oct 05 09:31:23 crc kubenswrapper[4846]: I1005 09:31:23.326796 4846 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"8a13caff3cf7131d82ade0de5fe733ded9547a5758c374eab87ab478fe594e4d"} pod="openshift-machine-config-operator/machine-config-daemon-fscvf" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 05 09:31:23 crc kubenswrapper[4846]: I1005 09:31:23.326862 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" containerName="machine-config-daemon" containerID="cri-o://8a13caff3cf7131d82ade0de5fe733ded9547a5758c374eab87ab478fe594e4d" gracePeriod=600 Oct 05 09:31:23 crc kubenswrapper[4846]: E1005 09:31:23.951258 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 09:31:24 crc kubenswrapper[4846]: I1005 09:31:24.488913 4846 generic.go:334] "Generic (PLEG): container finished" podID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" containerID="8a13caff3cf7131d82ade0de5fe733ded9547a5758c374eab87ab478fe594e4d" exitCode=0 Oct 05 09:31:24 crc kubenswrapper[4846]: I1005 09:31:24.489006 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" event={"ID":"ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7","Type":"ContainerDied","Data":"8a13caff3cf7131d82ade0de5fe733ded9547a5758c374eab87ab478fe594e4d"} Oct 05 09:31:24 crc kubenswrapper[4846]: I1005 09:31:24.489439 4846 scope.go:117] "RemoveContainer" containerID="25000c03b37ea2bd1717c4fee8262aa1fc79b5e0033f2d47b41a9ddab188e4f3" Oct 05 09:31:24 crc kubenswrapper[4846]: I1005 09:31:24.490375 4846 scope.go:117] "RemoveContainer" containerID="8a13caff3cf7131d82ade0de5fe733ded9547a5758c374eab87ab478fe594e4d" Oct 05 09:31:24 crc kubenswrapper[4846]: E1005 09:31:24.490795 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 09:31:34 crc kubenswrapper[4846]: I1005 09:31:34.609685 4846 generic.go:334] "Generic (PLEG): container finished" podID="dd84ed43-c477-4b83-a735-142efa0d85ea" containerID="d048a1370172c453d06d63c92c5bb1ab5050a1f599aa0807d80874a1c341b808" exitCode=0 Oct 05 09:31:34 crc kubenswrapper[4846]: I1005 09:31:34.610310 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellkhtf8" event={"ID":"dd84ed43-c477-4b83-a735-142efa0d85ea","Type":"ContainerDied","Data":"d048a1370172c453d06d63c92c5bb1ab5050a1f599aa0807d80874a1c341b808"} Oct 05 09:31:36 crc kubenswrapper[4846]: I1005 09:31:36.845334 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellkhtf8" Oct 05 09:31:36 crc kubenswrapper[4846]: I1005 09:31:36.907265 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/dd84ed43-c477-4b83-a735-142efa0d85ea-inventory\") pod \"dd84ed43-c477-4b83-a735-142efa0d85ea\" (UID: \"dd84ed43-c477-4b83-a735-142efa0d85ea\") " Oct 05 09:31:36 crc kubenswrapper[4846]: I1005 09:31:36.907413 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/dd84ed43-c477-4b83-a735-142efa0d85ea-nova-cell1-compute-config-0\") pod \"dd84ed43-c477-4b83-a735-142efa0d85ea\" (UID: \"dd84ed43-c477-4b83-a735-142efa0d85ea\") " Oct 05 09:31:36 crc kubenswrapper[4846]: I1005 09:31:36.907465 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/dd84ed43-c477-4b83-a735-142efa0d85ea-nova-migration-ssh-key-0\") pod \"dd84ed43-c477-4b83-a735-142efa0d85ea\" (UID: \"dd84ed43-c477-4b83-a735-142efa0d85ea\") " Oct 05 09:31:36 crc kubenswrapper[4846]: I1005 09:31:36.907521 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/dd84ed43-c477-4b83-a735-142efa0d85ea-ssh-key\") pod \"dd84ed43-c477-4b83-a735-142efa0d85ea\" (UID: \"dd84ed43-c477-4b83-a735-142efa0d85ea\") " Oct 05 09:31:36 crc kubenswrapper[4846]: I1005 09:31:36.907568 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/dd84ed43-c477-4b83-a735-142efa0d85ea-nova-cell1-compute-config-1\") pod \"dd84ed43-c477-4b83-a735-142efa0d85ea\" (UID: \"dd84ed43-c477-4b83-a735-142efa0d85ea\") " Oct 05 09:31:36 crc kubenswrapper[4846]: I1005 09:31:36.907637 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/dd84ed43-c477-4b83-a735-142efa0d85ea-nova-migration-ssh-key-1\") pod \"dd84ed43-c477-4b83-a735-142efa0d85ea\" (UID: \"dd84ed43-c477-4b83-a735-142efa0d85ea\") " Oct 05 09:31:36 crc kubenswrapper[4846]: I1005 09:31:36.907743 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/dd84ed43-c477-4b83-a735-142efa0d85ea-nova-cells-global-config-0\") pod \"dd84ed43-c477-4b83-a735-142efa0d85ea\" (UID: \"dd84ed43-c477-4b83-a735-142efa0d85ea\") " Oct 05 09:31:36 crc kubenswrapper[4846]: I1005 09:31:36.907812 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rvk6m\" (UniqueName: \"kubernetes.io/projected/dd84ed43-c477-4b83-a735-142efa0d85ea-kube-api-access-rvk6m\") pod \"dd84ed43-c477-4b83-a735-142efa0d85ea\" (UID: \"dd84ed43-c477-4b83-a735-142efa0d85ea\") " Oct 05 09:31:36 crc kubenswrapper[4846]: I1005 09:31:36.907870 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dd84ed43-c477-4b83-a735-142efa0d85ea-nova-cell1-combined-ca-bundle\") pod \"dd84ed43-c477-4b83-a735-142efa0d85ea\" (UID: \"dd84ed43-c477-4b83-a735-142efa0d85ea\") " Oct 05 09:31:36 crc kubenswrapper[4846]: I1005 09:31:36.931891 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dd84ed43-c477-4b83-a735-142efa0d85ea-kube-api-access-rvk6m" (OuterVolumeSpecName: "kube-api-access-rvk6m") pod "dd84ed43-c477-4b83-a735-142efa0d85ea" (UID: "dd84ed43-c477-4b83-a735-142efa0d85ea"). InnerVolumeSpecName "kube-api-access-rvk6m". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 09:31:36 crc kubenswrapper[4846]: I1005 09:31:36.954637 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dd84ed43-c477-4b83-a735-142efa0d85ea-nova-cell1-combined-ca-bundle" (OuterVolumeSpecName: "nova-cell1-combined-ca-bundle") pod "dd84ed43-c477-4b83-a735-142efa0d85ea" (UID: "dd84ed43-c477-4b83-a735-142efa0d85ea"). InnerVolumeSpecName "nova-cell1-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 09:31:36 crc kubenswrapper[4846]: I1005 09:31:36.961062 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dd84ed43-c477-4b83-a735-142efa0d85ea-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "dd84ed43-c477-4b83-a735-142efa0d85ea" (UID: "dd84ed43-c477-4b83-a735-142efa0d85ea"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 09:31:36 crc kubenswrapper[4846]: I1005 09:31:36.969221 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dd84ed43-c477-4b83-a735-142efa0d85ea-nova-cell1-compute-config-1" (OuterVolumeSpecName: "nova-cell1-compute-config-1") pod "dd84ed43-c477-4b83-a735-142efa0d85ea" (UID: "dd84ed43-c477-4b83-a735-142efa0d85ea"). InnerVolumeSpecName "nova-cell1-compute-config-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 09:31:36 crc kubenswrapper[4846]: I1005 09:31:36.974775 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dd84ed43-c477-4b83-a735-142efa0d85ea-nova-cells-global-config-0" (OuterVolumeSpecName: "nova-cells-global-config-0") pod "dd84ed43-c477-4b83-a735-142efa0d85ea" (UID: "dd84ed43-c477-4b83-a735-142efa0d85ea"). InnerVolumeSpecName "nova-cells-global-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 09:31:36 crc kubenswrapper[4846]: I1005 09:31:36.974873 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dd84ed43-c477-4b83-a735-142efa0d85ea-nova-migration-ssh-key-0" (OuterVolumeSpecName: "nova-migration-ssh-key-0") pod "dd84ed43-c477-4b83-a735-142efa0d85ea" (UID: "dd84ed43-c477-4b83-a735-142efa0d85ea"). InnerVolumeSpecName "nova-migration-ssh-key-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 09:31:36 crc kubenswrapper[4846]: I1005 09:31:36.977535 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dd84ed43-c477-4b83-a735-142efa0d85ea-inventory" (OuterVolumeSpecName: "inventory") pod "dd84ed43-c477-4b83-a735-142efa0d85ea" (UID: "dd84ed43-c477-4b83-a735-142efa0d85ea"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 09:31:36 crc kubenswrapper[4846]: I1005 09:31:36.983216 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dd84ed43-c477-4b83-a735-142efa0d85ea-nova-migration-ssh-key-1" (OuterVolumeSpecName: "nova-migration-ssh-key-1") pod "dd84ed43-c477-4b83-a735-142efa0d85ea" (UID: "dd84ed43-c477-4b83-a735-142efa0d85ea"). InnerVolumeSpecName "nova-migration-ssh-key-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 09:31:37 crc kubenswrapper[4846]: I1005 09:31:36.996803 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dd84ed43-c477-4b83-a735-142efa0d85ea-nova-cell1-compute-config-0" (OuterVolumeSpecName: "nova-cell1-compute-config-0") pod "dd84ed43-c477-4b83-a735-142efa0d85ea" (UID: "dd84ed43-c477-4b83-a735-142efa0d85ea"). InnerVolumeSpecName "nova-cell1-compute-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 09:31:37 crc kubenswrapper[4846]: I1005 09:31:37.010850 4846 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/dd84ed43-c477-4b83-a735-142efa0d85ea-nova-cell1-compute-config-0\") on node \"crc\" DevicePath \"\"" Oct 05 09:31:37 crc kubenswrapper[4846]: I1005 09:31:37.010890 4846 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/dd84ed43-c477-4b83-a735-142efa0d85ea-nova-migration-ssh-key-0\") on node \"crc\" DevicePath \"\"" Oct 05 09:31:37 crc kubenswrapper[4846]: I1005 09:31:37.010904 4846 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/dd84ed43-c477-4b83-a735-142efa0d85ea-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 05 09:31:37 crc kubenswrapper[4846]: I1005 09:31:37.010916 4846 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/dd84ed43-c477-4b83-a735-142efa0d85ea-nova-cell1-compute-config-1\") on node \"crc\" DevicePath \"\"" Oct 05 09:31:37 crc kubenswrapper[4846]: I1005 09:31:37.010928 4846 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/dd84ed43-c477-4b83-a735-142efa0d85ea-nova-migration-ssh-key-1\") on node \"crc\" DevicePath \"\"" Oct 05 09:31:37 crc kubenswrapper[4846]: I1005 09:31:37.010940 4846 reconciler_common.go:293] "Volume detached for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/dd84ed43-c477-4b83-a735-142efa0d85ea-nova-cells-global-config-0\") on node \"crc\" DevicePath \"\"" Oct 05 09:31:37 crc kubenswrapper[4846]: I1005 09:31:37.010951 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rvk6m\" (UniqueName: \"kubernetes.io/projected/dd84ed43-c477-4b83-a735-142efa0d85ea-kube-api-access-rvk6m\") on node \"crc\" DevicePath \"\"" Oct 05 09:31:37 crc kubenswrapper[4846]: I1005 09:31:37.010963 4846 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dd84ed43-c477-4b83-a735-142efa0d85ea-nova-cell1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 05 09:31:37 crc kubenswrapper[4846]: I1005 09:31:37.010975 4846 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/dd84ed43-c477-4b83-a735-142efa0d85ea-inventory\") on node \"crc\" DevicePath \"\"" Oct 05 09:31:37 crc kubenswrapper[4846]: I1005 09:31:37.646863 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellkhtf8" event={"ID":"dd84ed43-c477-4b83-a735-142efa0d85ea","Type":"ContainerDied","Data":"691967b9eb1be739fd4facde8e9d71385fb3afac95d4c22399d605f08ab8917b"} Oct 05 09:31:37 crc kubenswrapper[4846]: I1005 09:31:37.646919 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="691967b9eb1be739fd4facde8e9d71385fb3afac95d4c22399d605f08ab8917b" Oct 05 09:31:37 crc kubenswrapper[4846]: I1005 09:31:37.646953 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellkhtf8" Oct 05 09:31:38 crc kubenswrapper[4846]: I1005 09:31:38.504247 4846 scope.go:117] "RemoveContainer" containerID="8a13caff3cf7131d82ade0de5fe733ded9547a5758c374eab87ab478fe594e4d" Oct 05 09:31:38 crc kubenswrapper[4846]: E1005 09:31:38.505472 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 09:31:52 crc kubenswrapper[4846]: I1005 09:31:52.504420 4846 scope.go:117] "RemoveContainer" containerID="8a13caff3cf7131d82ade0de5fe733ded9547a5758c374eab87ab478fe594e4d" Oct 05 09:31:52 crc kubenswrapper[4846]: E1005 09:31:52.505360 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 09:32:06 crc kubenswrapper[4846]: I1005 09:32:06.498565 4846 scope.go:117] "RemoveContainer" containerID="8a13caff3cf7131d82ade0de5fe733ded9547a5758c374eab87ab478fe594e4d" Oct 05 09:32:06 crc kubenswrapper[4846]: E1005 09:32:06.499546 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 09:32:17 crc kubenswrapper[4846]: I1005 09:32:17.497701 4846 scope.go:117] "RemoveContainer" containerID="8a13caff3cf7131d82ade0de5fe733ded9547a5758c374eab87ab478fe594e4d" Oct 05 09:32:17 crc kubenswrapper[4846]: E1005 09:32:17.498585 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 09:32:29 crc kubenswrapper[4846]: I1005 09:32:29.498283 4846 scope.go:117] "RemoveContainer" containerID="8a13caff3cf7131d82ade0de5fe733ded9547a5758c374eab87ab478fe594e4d" Oct 05 09:32:29 crc kubenswrapper[4846]: E1005 09:32:29.499107 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 09:32:41 crc kubenswrapper[4846]: I1005 09:32:41.498483 4846 scope.go:117] "RemoveContainer" containerID="8a13caff3cf7131d82ade0de5fe733ded9547a5758c374eab87ab478fe594e4d" Oct 05 09:32:41 crc kubenswrapper[4846]: E1005 09:32:41.499654 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 09:32:53 crc kubenswrapper[4846]: I1005 09:32:53.497474 4846 scope.go:117] "RemoveContainer" containerID="8a13caff3cf7131d82ade0de5fe733ded9547a5758c374eab87ab478fe594e4d" Oct 05 09:32:53 crc kubenswrapper[4846]: E1005 09:32:53.498438 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 09:33:05 crc kubenswrapper[4846]: I1005 09:33:05.497926 4846 scope.go:117] "RemoveContainer" containerID="8a13caff3cf7131d82ade0de5fe733ded9547a5758c374eab87ab478fe594e4d" Oct 05 09:33:05 crc kubenswrapper[4846]: E1005 09:33:05.498943 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 09:33:06 crc kubenswrapper[4846]: I1005 09:33:06.847972 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-copy-data"] Oct 05 09:33:06 crc kubenswrapper[4846]: I1005 09:33:06.848666 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/mariadb-copy-data" podUID="1edceaaf-ed3d-4595-985f-91d045d101c7" containerName="adoption" containerID="cri-o://1647da17d92398b042c0915aded45cad06c01a4fd9210b9cf619d434c72384e0" gracePeriod=30 Oct 05 09:33:17 crc kubenswrapper[4846]: I1005 09:33:17.498437 4846 scope.go:117] "RemoveContainer" containerID="8a13caff3cf7131d82ade0de5fe733ded9547a5758c374eab87ab478fe594e4d" Oct 05 09:33:17 crc kubenswrapper[4846]: E1005 09:33:17.499592 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 09:33:30 crc kubenswrapper[4846]: I1005 09:33:30.498362 4846 scope.go:117] "RemoveContainer" containerID="8a13caff3cf7131d82ade0de5fe733ded9547a5758c374eab87ab478fe594e4d" Oct 05 09:33:30 crc kubenswrapper[4846]: E1005 09:33:30.499772 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 09:33:37 crc kubenswrapper[4846]: I1005 09:33:37.145927 4846 generic.go:334] "Generic (PLEG): container finished" podID="1edceaaf-ed3d-4595-985f-91d045d101c7" containerID="1647da17d92398b042c0915aded45cad06c01a4fd9210b9cf619d434c72384e0" exitCode=137 Oct 05 09:33:37 crc kubenswrapper[4846]: I1005 09:33:37.146049 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-copy-data" event={"ID":"1edceaaf-ed3d-4595-985f-91d045d101c7","Type":"ContainerDied","Data":"1647da17d92398b042c0915aded45cad06c01a4fd9210b9cf619d434c72384e0"} Oct 05 09:33:37 crc kubenswrapper[4846]: I1005 09:33:37.450915 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-copy-data" Oct 05 09:33:37 crc kubenswrapper[4846]: I1005 09:33:37.572697 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t7gqx\" (UniqueName: \"kubernetes.io/projected/1edceaaf-ed3d-4595-985f-91d045d101c7-kube-api-access-t7gqx\") pod \"1edceaaf-ed3d-4595-985f-91d045d101c7\" (UID: \"1edceaaf-ed3d-4595-985f-91d045d101c7\") " Oct 05 09:33:37 crc kubenswrapper[4846]: I1005 09:33:37.573605 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mariadb-data\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-8c01e9e7-1c79-4d8c-a6a6-9aa1977a5bc1\") pod \"1edceaaf-ed3d-4595-985f-91d045d101c7\" (UID: \"1edceaaf-ed3d-4595-985f-91d045d101c7\") " Oct 05 09:33:37 crc kubenswrapper[4846]: I1005 09:33:37.579154 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1edceaaf-ed3d-4595-985f-91d045d101c7-kube-api-access-t7gqx" (OuterVolumeSpecName: "kube-api-access-t7gqx") pod "1edceaaf-ed3d-4595-985f-91d045d101c7" (UID: "1edceaaf-ed3d-4595-985f-91d045d101c7"). InnerVolumeSpecName "kube-api-access-t7gqx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 09:33:37 crc kubenswrapper[4846]: I1005 09:33:37.601333 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-8c01e9e7-1c79-4d8c-a6a6-9aa1977a5bc1" (OuterVolumeSpecName: "mariadb-data") pod "1edceaaf-ed3d-4595-985f-91d045d101c7" (UID: "1edceaaf-ed3d-4595-985f-91d045d101c7"). InnerVolumeSpecName "pvc-8c01e9e7-1c79-4d8c-a6a6-9aa1977a5bc1". PluginName "kubernetes.io/csi", VolumeGidValue "" Oct 05 09:33:37 crc kubenswrapper[4846]: I1005 09:33:37.677587 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t7gqx\" (UniqueName: \"kubernetes.io/projected/1edceaaf-ed3d-4595-985f-91d045d101c7-kube-api-access-t7gqx\") on node \"crc\" DevicePath \"\"" Oct 05 09:33:37 crc kubenswrapper[4846]: I1005 09:33:37.677668 4846 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-8c01e9e7-1c79-4d8c-a6a6-9aa1977a5bc1\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-8c01e9e7-1c79-4d8c-a6a6-9aa1977a5bc1\") on node \"crc\" " Oct 05 09:33:37 crc kubenswrapper[4846]: I1005 09:33:37.724448 4846 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Oct 05 09:33:37 crc kubenswrapper[4846]: I1005 09:33:37.725446 4846 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-8c01e9e7-1c79-4d8c-a6a6-9aa1977a5bc1" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-8c01e9e7-1c79-4d8c-a6a6-9aa1977a5bc1") on node "crc" Oct 05 09:33:37 crc kubenswrapper[4846]: I1005 09:33:37.780310 4846 reconciler_common.go:293] "Volume detached for volume \"pvc-8c01e9e7-1c79-4d8c-a6a6-9aa1977a5bc1\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-8c01e9e7-1c79-4d8c-a6a6-9aa1977a5bc1\") on node \"crc\" DevicePath \"\"" Oct 05 09:33:38 crc kubenswrapper[4846]: I1005 09:33:38.163541 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-copy-data" event={"ID":"1edceaaf-ed3d-4595-985f-91d045d101c7","Type":"ContainerDied","Data":"3aa0c20bc7a0191ec1ca1e2908aac1495dcb12691c1b3c0032ff93e51adb1046"} Oct 05 09:33:38 crc kubenswrapper[4846]: I1005 09:33:38.163617 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-copy-data" Oct 05 09:33:38 crc kubenswrapper[4846]: I1005 09:33:38.163645 4846 scope.go:117] "RemoveContainer" containerID="1647da17d92398b042c0915aded45cad06c01a4fd9210b9cf619d434c72384e0" Oct 05 09:33:38 crc kubenswrapper[4846]: I1005 09:33:38.228681 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-copy-data"] Oct 05 09:33:38 crc kubenswrapper[4846]: I1005 09:33:38.238521 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-copy-data"] Oct 05 09:33:38 crc kubenswrapper[4846]: I1005 09:33:38.523837 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1edceaaf-ed3d-4595-985f-91d045d101c7" path="/var/lib/kubelet/pods/1edceaaf-ed3d-4595-985f-91d045d101c7/volumes" Oct 05 09:33:38 crc kubenswrapper[4846]: I1005 09:33:38.910475 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-copy-data"] Oct 05 09:33:38 crc kubenswrapper[4846]: I1005 09:33:38.910921 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-copy-data" podUID="f0768736-73ad-4ce5-be3c-1941d56cd760" containerName="adoption" containerID="cri-o://b2793fb5da2d48dbd09af9440210742861a05910349aabc2730a5269a46eecc4" gracePeriod=30 Oct 05 09:33:41 crc kubenswrapper[4846]: I1005 09:33:41.498818 4846 scope.go:117] "RemoveContainer" containerID="8a13caff3cf7131d82ade0de5fe733ded9547a5758c374eab87ab478fe594e4d" Oct 05 09:33:41 crc kubenswrapper[4846]: E1005 09:33:41.499650 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 09:33:44 crc kubenswrapper[4846]: I1005 09:33:44.406473 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-2dmlk"] Oct 05 09:33:44 crc kubenswrapper[4846]: E1005 09:33:44.407354 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8a7dd7cc-03aa-4e36-b6f8-8db7c32ac209" containerName="extract-utilities" Oct 05 09:33:44 crc kubenswrapper[4846]: I1005 09:33:44.407372 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="8a7dd7cc-03aa-4e36-b6f8-8db7c32ac209" containerName="extract-utilities" Oct 05 09:33:44 crc kubenswrapper[4846]: E1005 09:33:44.407396 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dd84ed43-c477-4b83-a735-142efa0d85ea" containerName="nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell1" Oct 05 09:33:44 crc kubenswrapper[4846]: I1005 09:33:44.407406 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="dd84ed43-c477-4b83-a735-142efa0d85ea" containerName="nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell1" Oct 05 09:33:44 crc kubenswrapper[4846]: E1005 09:33:44.407418 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b376bc05-9c1a-40ce-b237-f94b4774bcc8" containerName="registry-server" Oct 05 09:33:44 crc kubenswrapper[4846]: I1005 09:33:44.407426 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="b376bc05-9c1a-40ce-b237-f94b4774bcc8" containerName="registry-server" Oct 05 09:33:44 crc kubenswrapper[4846]: E1005 09:33:44.407439 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8a7dd7cc-03aa-4e36-b6f8-8db7c32ac209" containerName="registry-server" Oct 05 09:33:44 crc kubenswrapper[4846]: I1005 09:33:44.407447 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="8a7dd7cc-03aa-4e36-b6f8-8db7c32ac209" containerName="registry-server" Oct 05 09:33:44 crc kubenswrapper[4846]: E1005 09:33:44.407478 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b376bc05-9c1a-40ce-b237-f94b4774bcc8" containerName="extract-utilities" Oct 05 09:33:44 crc kubenswrapper[4846]: I1005 09:33:44.407487 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="b376bc05-9c1a-40ce-b237-f94b4774bcc8" containerName="extract-utilities" Oct 05 09:33:44 crc kubenswrapper[4846]: E1005 09:33:44.407500 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b376bc05-9c1a-40ce-b237-f94b4774bcc8" containerName="extract-content" Oct 05 09:33:44 crc kubenswrapper[4846]: I1005 09:33:44.407507 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="b376bc05-9c1a-40ce-b237-f94b4774bcc8" containerName="extract-content" Oct 05 09:33:44 crc kubenswrapper[4846]: E1005 09:33:44.407520 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1edceaaf-ed3d-4595-985f-91d045d101c7" containerName="adoption" Oct 05 09:33:44 crc kubenswrapper[4846]: I1005 09:33:44.407528 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="1edceaaf-ed3d-4595-985f-91d045d101c7" containerName="adoption" Oct 05 09:33:44 crc kubenswrapper[4846]: E1005 09:33:44.407537 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8a7dd7cc-03aa-4e36-b6f8-8db7c32ac209" containerName="extract-content" Oct 05 09:33:44 crc kubenswrapper[4846]: I1005 09:33:44.407544 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="8a7dd7cc-03aa-4e36-b6f8-8db7c32ac209" containerName="extract-content" Oct 05 09:33:44 crc kubenswrapper[4846]: I1005 09:33:44.407814 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="8a7dd7cc-03aa-4e36-b6f8-8db7c32ac209" containerName="registry-server" Oct 05 09:33:44 crc kubenswrapper[4846]: I1005 09:33:44.407830 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="1edceaaf-ed3d-4595-985f-91d045d101c7" containerName="adoption" Oct 05 09:33:44 crc kubenswrapper[4846]: I1005 09:33:44.407849 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="dd84ed43-c477-4b83-a735-142efa0d85ea" containerName="nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell1" Oct 05 09:33:44 crc kubenswrapper[4846]: I1005 09:33:44.407866 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="b376bc05-9c1a-40ce-b237-f94b4774bcc8" containerName="registry-server" Oct 05 09:33:44 crc kubenswrapper[4846]: I1005 09:33:44.409843 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2dmlk" Oct 05 09:33:44 crc kubenswrapper[4846]: I1005 09:33:44.419074 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-2dmlk"] Oct 05 09:33:44 crc kubenswrapper[4846]: I1005 09:33:44.534147 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2ba1b06e-663e-4d24-9d0d-f0ff29adde9b-utilities\") pod \"certified-operators-2dmlk\" (UID: \"2ba1b06e-663e-4d24-9d0d-f0ff29adde9b\") " pod="openshift-marketplace/certified-operators-2dmlk" Oct 05 09:33:44 crc kubenswrapper[4846]: I1005 09:33:44.534711 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2ba1b06e-663e-4d24-9d0d-f0ff29adde9b-catalog-content\") pod \"certified-operators-2dmlk\" (UID: \"2ba1b06e-663e-4d24-9d0d-f0ff29adde9b\") " pod="openshift-marketplace/certified-operators-2dmlk" Oct 05 09:33:44 crc kubenswrapper[4846]: I1005 09:33:44.535061 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lrfxt\" (UniqueName: \"kubernetes.io/projected/2ba1b06e-663e-4d24-9d0d-f0ff29adde9b-kube-api-access-lrfxt\") pod \"certified-operators-2dmlk\" (UID: \"2ba1b06e-663e-4d24-9d0d-f0ff29adde9b\") " pod="openshift-marketplace/certified-operators-2dmlk" Oct 05 09:33:44 crc kubenswrapper[4846]: I1005 09:33:44.637536 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2ba1b06e-663e-4d24-9d0d-f0ff29adde9b-utilities\") pod \"certified-operators-2dmlk\" (UID: \"2ba1b06e-663e-4d24-9d0d-f0ff29adde9b\") " pod="openshift-marketplace/certified-operators-2dmlk" Oct 05 09:33:44 crc kubenswrapper[4846]: I1005 09:33:44.637763 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2ba1b06e-663e-4d24-9d0d-f0ff29adde9b-catalog-content\") pod \"certified-operators-2dmlk\" (UID: \"2ba1b06e-663e-4d24-9d0d-f0ff29adde9b\") " pod="openshift-marketplace/certified-operators-2dmlk" Oct 05 09:33:44 crc kubenswrapper[4846]: I1005 09:33:44.637837 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lrfxt\" (UniqueName: \"kubernetes.io/projected/2ba1b06e-663e-4d24-9d0d-f0ff29adde9b-kube-api-access-lrfxt\") pod \"certified-operators-2dmlk\" (UID: \"2ba1b06e-663e-4d24-9d0d-f0ff29adde9b\") " pod="openshift-marketplace/certified-operators-2dmlk" Oct 05 09:33:44 crc kubenswrapper[4846]: I1005 09:33:44.638231 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2ba1b06e-663e-4d24-9d0d-f0ff29adde9b-utilities\") pod \"certified-operators-2dmlk\" (UID: \"2ba1b06e-663e-4d24-9d0d-f0ff29adde9b\") " pod="openshift-marketplace/certified-operators-2dmlk" Oct 05 09:33:44 crc kubenswrapper[4846]: I1005 09:33:44.638297 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2ba1b06e-663e-4d24-9d0d-f0ff29adde9b-catalog-content\") pod \"certified-operators-2dmlk\" (UID: \"2ba1b06e-663e-4d24-9d0d-f0ff29adde9b\") " pod="openshift-marketplace/certified-operators-2dmlk" Oct 05 09:33:44 crc kubenswrapper[4846]: I1005 09:33:44.668252 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lrfxt\" (UniqueName: \"kubernetes.io/projected/2ba1b06e-663e-4d24-9d0d-f0ff29adde9b-kube-api-access-lrfxt\") pod \"certified-operators-2dmlk\" (UID: \"2ba1b06e-663e-4d24-9d0d-f0ff29adde9b\") " pod="openshift-marketplace/certified-operators-2dmlk" Oct 05 09:33:44 crc kubenswrapper[4846]: I1005 09:33:44.737930 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2dmlk" Oct 05 09:33:45 crc kubenswrapper[4846]: I1005 09:33:45.355682 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-2dmlk"] Oct 05 09:33:46 crc kubenswrapper[4846]: I1005 09:33:46.248622 4846 generic.go:334] "Generic (PLEG): container finished" podID="2ba1b06e-663e-4d24-9d0d-f0ff29adde9b" containerID="449e5ed250cddb7ff09106a1d2655b161c6375e4e368d8832333f36d7996cd8f" exitCode=0 Oct 05 09:33:46 crc kubenswrapper[4846]: I1005 09:33:46.248736 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2dmlk" event={"ID":"2ba1b06e-663e-4d24-9d0d-f0ff29adde9b","Type":"ContainerDied","Data":"449e5ed250cddb7ff09106a1d2655b161c6375e4e368d8832333f36d7996cd8f"} Oct 05 09:33:46 crc kubenswrapper[4846]: I1005 09:33:46.250934 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2dmlk" event={"ID":"2ba1b06e-663e-4d24-9d0d-f0ff29adde9b","Type":"ContainerStarted","Data":"6d6a5c4d61b51dbb87c7a5ded6ffbd4e6f23de5722a63637603dfd5c6f6af9fc"} Oct 05 09:33:47 crc kubenswrapper[4846]: I1005 09:33:47.265441 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2dmlk" event={"ID":"2ba1b06e-663e-4d24-9d0d-f0ff29adde9b","Type":"ContainerStarted","Data":"8561884803d5e7b193f12a7f90f1a582ab155c26f991b1a20c8a25a1b4f802d3"} Oct 05 09:33:49 crc kubenswrapper[4846]: I1005 09:33:49.290093 4846 generic.go:334] "Generic (PLEG): container finished" podID="2ba1b06e-663e-4d24-9d0d-f0ff29adde9b" containerID="8561884803d5e7b193f12a7f90f1a582ab155c26f991b1a20c8a25a1b4f802d3" exitCode=0 Oct 05 09:33:49 crc kubenswrapper[4846]: I1005 09:33:49.290224 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2dmlk" event={"ID":"2ba1b06e-663e-4d24-9d0d-f0ff29adde9b","Type":"ContainerDied","Data":"8561884803d5e7b193f12a7f90f1a582ab155c26f991b1a20c8a25a1b4f802d3"} Oct 05 09:33:51 crc kubenswrapper[4846]: I1005 09:33:51.321003 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2dmlk" event={"ID":"2ba1b06e-663e-4d24-9d0d-f0ff29adde9b","Type":"ContainerStarted","Data":"6b8742ac90599080bcc978c0108a6ef8d60ba7af22f472082648c02fa065b481"} Oct 05 09:33:51 crc kubenswrapper[4846]: I1005 09:33:51.361106 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-2dmlk" podStartSLOduration=3.891062313 podStartE2EDuration="7.361074915s" podCreationTimestamp="2025-10-05 09:33:44 +0000 UTC" firstStartedPulling="2025-10-05 09:33:46.250332234 +0000 UTC m=+9948.491185009" lastFinishedPulling="2025-10-05 09:33:49.720344796 +0000 UTC m=+9951.961197611" observedRunningTime="2025-10-05 09:33:51.342545898 +0000 UTC m=+9953.583398713" watchObservedRunningTime="2025-10-05 09:33:51.361074915 +0000 UTC m=+9953.601927730" Oct 05 09:33:54 crc kubenswrapper[4846]: I1005 09:33:54.497395 4846 scope.go:117] "RemoveContainer" containerID="8a13caff3cf7131d82ade0de5fe733ded9547a5758c374eab87ab478fe594e4d" Oct 05 09:33:54 crc kubenswrapper[4846]: E1005 09:33:54.498039 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 09:33:54 crc kubenswrapper[4846]: I1005 09:33:54.738918 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-2dmlk" Oct 05 09:33:54 crc kubenswrapper[4846]: I1005 09:33:54.739011 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-2dmlk" Oct 05 09:33:54 crc kubenswrapper[4846]: I1005 09:33:54.837668 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-2dmlk" Oct 05 09:33:55 crc kubenswrapper[4846]: I1005 09:33:55.445284 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-2dmlk" Oct 05 09:33:55 crc kubenswrapper[4846]: I1005 09:33:55.503443 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-2dmlk"] Oct 05 09:33:57 crc kubenswrapper[4846]: I1005 09:33:57.389809 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-2dmlk" podUID="2ba1b06e-663e-4d24-9d0d-f0ff29adde9b" containerName="registry-server" containerID="cri-o://6b8742ac90599080bcc978c0108a6ef8d60ba7af22f472082648c02fa065b481" gracePeriod=2 Oct 05 09:33:58 crc kubenswrapper[4846]: I1005 09:33:58.406645 4846 generic.go:334] "Generic (PLEG): container finished" podID="2ba1b06e-663e-4d24-9d0d-f0ff29adde9b" containerID="6b8742ac90599080bcc978c0108a6ef8d60ba7af22f472082648c02fa065b481" exitCode=0 Oct 05 09:33:58 crc kubenswrapper[4846]: I1005 09:33:58.406755 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2dmlk" event={"ID":"2ba1b06e-663e-4d24-9d0d-f0ff29adde9b","Type":"ContainerDied","Data":"6b8742ac90599080bcc978c0108a6ef8d60ba7af22f472082648c02fa065b481"} Oct 05 09:33:58 crc kubenswrapper[4846]: I1005 09:33:58.407144 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2dmlk" event={"ID":"2ba1b06e-663e-4d24-9d0d-f0ff29adde9b","Type":"ContainerDied","Data":"6d6a5c4d61b51dbb87c7a5ded6ffbd4e6f23de5722a63637603dfd5c6f6af9fc"} Oct 05 09:33:58 crc kubenswrapper[4846]: I1005 09:33:58.407217 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6d6a5c4d61b51dbb87c7a5ded6ffbd4e6f23de5722a63637603dfd5c6f6af9fc" Oct 05 09:33:58 crc kubenswrapper[4846]: I1005 09:33:58.466221 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2dmlk" Oct 05 09:33:58 crc kubenswrapper[4846]: I1005 09:33:58.566309 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lrfxt\" (UniqueName: \"kubernetes.io/projected/2ba1b06e-663e-4d24-9d0d-f0ff29adde9b-kube-api-access-lrfxt\") pod \"2ba1b06e-663e-4d24-9d0d-f0ff29adde9b\" (UID: \"2ba1b06e-663e-4d24-9d0d-f0ff29adde9b\") " Oct 05 09:33:58 crc kubenswrapper[4846]: I1005 09:33:58.566502 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2ba1b06e-663e-4d24-9d0d-f0ff29adde9b-utilities\") pod \"2ba1b06e-663e-4d24-9d0d-f0ff29adde9b\" (UID: \"2ba1b06e-663e-4d24-9d0d-f0ff29adde9b\") " Oct 05 09:33:58 crc kubenswrapper[4846]: I1005 09:33:58.566729 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2ba1b06e-663e-4d24-9d0d-f0ff29adde9b-catalog-content\") pod \"2ba1b06e-663e-4d24-9d0d-f0ff29adde9b\" (UID: \"2ba1b06e-663e-4d24-9d0d-f0ff29adde9b\") " Oct 05 09:33:58 crc kubenswrapper[4846]: I1005 09:33:58.568456 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2ba1b06e-663e-4d24-9d0d-f0ff29adde9b-utilities" (OuterVolumeSpecName: "utilities") pod "2ba1b06e-663e-4d24-9d0d-f0ff29adde9b" (UID: "2ba1b06e-663e-4d24-9d0d-f0ff29adde9b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 09:33:58 crc kubenswrapper[4846]: I1005 09:33:58.574129 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2ba1b06e-663e-4d24-9d0d-f0ff29adde9b-kube-api-access-lrfxt" (OuterVolumeSpecName: "kube-api-access-lrfxt") pod "2ba1b06e-663e-4d24-9d0d-f0ff29adde9b" (UID: "2ba1b06e-663e-4d24-9d0d-f0ff29adde9b"). InnerVolumeSpecName "kube-api-access-lrfxt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 09:33:58 crc kubenswrapper[4846]: I1005 09:33:58.614001 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2ba1b06e-663e-4d24-9d0d-f0ff29adde9b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2ba1b06e-663e-4d24-9d0d-f0ff29adde9b" (UID: "2ba1b06e-663e-4d24-9d0d-f0ff29adde9b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 09:33:58 crc kubenswrapper[4846]: I1005 09:33:58.670110 4846 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2ba1b06e-663e-4d24-9d0d-f0ff29adde9b-utilities\") on node \"crc\" DevicePath \"\"" Oct 05 09:33:58 crc kubenswrapper[4846]: I1005 09:33:58.670160 4846 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2ba1b06e-663e-4d24-9d0d-f0ff29adde9b-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 05 09:33:58 crc kubenswrapper[4846]: I1005 09:33:58.670183 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lrfxt\" (UniqueName: \"kubernetes.io/projected/2ba1b06e-663e-4d24-9d0d-f0ff29adde9b-kube-api-access-lrfxt\") on node \"crc\" DevicePath \"\"" Oct 05 09:33:59 crc kubenswrapper[4846]: I1005 09:33:59.418858 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2dmlk" Oct 05 09:33:59 crc kubenswrapper[4846]: I1005 09:33:59.475963 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-2dmlk"] Oct 05 09:33:59 crc kubenswrapper[4846]: I1005 09:33:59.493376 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-2dmlk"] Oct 05 09:34:00 crc kubenswrapper[4846]: I1005 09:34:00.511022 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2ba1b06e-663e-4d24-9d0d-f0ff29adde9b" path="/var/lib/kubelet/pods/2ba1b06e-663e-4d24-9d0d-f0ff29adde9b/volumes" Oct 05 09:34:06 crc kubenswrapper[4846]: I1005 09:34:06.573452 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-xccjp"] Oct 05 09:34:06 crc kubenswrapper[4846]: E1005 09:34:06.575635 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2ba1b06e-663e-4d24-9d0d-f0ff29adde9b" containerName="extract-content" Oct 05 09:34:06 crc kubenswrapper[4846]: I1005 09:34:06.575730 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="2ba1b06e-663e-4d24-9d0d-f0ff29adde9b" containerName="extract-content" Oct 05 09:34:06 crc kubenswrapper[4846]: E1005 09:34:06.575792 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2ba1b06e-663e-4d24-9d0d-f0ff29adde9b" containerName="extract-utilities" Oct 05 09:34:06 crc kubenswrapper[4846]: I1005 09:34:06.575872 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="2ba1b06e-663e-4d24-9d0d-f0ff29adde9b" containerName="extract-utilities" Oct 05 09:34:06 crc kubenswrapper[4846]: E1005 09:34:06.575948 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2ba1b06e-663e-4d24-9d0d-f0ff29adde9b" containerName="registry-server" Oct 05 09:34:06 crc kubenswrapper[4846]: I1005 09:34:06.576024 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="2ba1b06e-663e-4d24-9d0d-f0ff29adde9b" containerName="registry-server" Oct 05 09:34:06 crc kubenswrapper[4846]: I1005 09:34:06.576392 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="2ba1b06e-663e-4d24-9d0d-f0ff29adde9b" containerName="registry-server" Oct 05 09:34:06 crc kubenswrapper[4846]: I1005 09:34:06.578558 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xccjp" Oct 05 09:34:06 crc kubenswrapper[4846]: I1005 09:34:06.601920 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-xccjp"] Oct 05 09:34:06 crc kubenswrapper[4846]: I1005 09:34:06.643952 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x2t9v\" (UniqueName: \"kubernetes.io/projected/38c22856-82b2-4ddc-9e43-0ef1399068ea-kube-api-access-x2t9v\") pod \"redhat-operators-xccjp\" (UID: \"38c22856-82b2-4ddc-9e43-0ef1399068ea\") " pod="openshift-marketplace/redhat-operators-xccjp" Oct 05 09:34:06 crc kubenswrapper[4846]: I1005 09:34:06.644053 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/38c22856-82b2-4ddc-9e43-0ef1399068ea-catalog-content\") pod \"redhat-operators-xccjp\" (UID: \"38c22856-82b2-4ddc-9e43-0ef1399068ea\") " pod="openshift-marketplace/redhat-operators-xccjp" Oct 05 09:34:06 crc kubenswrapper[4846]: I1005 09:34:06.644490 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/38c22856-82b2-4ddc-9e43-0ef1399068ea-utilities\") pod \"redhat-operators-xccjp\" (UID: \"38c22856-82b2-4ddc-9e43-0ef1399068ea\") " pod="openshift-marketplace/redhat-operators-xccjp" Oct 05 09:34:06 crc kubenswrapper[4846]: I1005 09:34:06.746926 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x2t9v\" (UniqueName: \"kubernetes.io/projected/38c22856-82b2-4ddc-9e43-0ef1399068ea-kube-api-access-x2t9v\") pod \"redhat-operators-xccjp\" (UID: \"38c22856-82b2-4ddc-9e43-0ef1399068ea\") " pod="openshift-marketplace/redhat-operators-xccjp" Oct 05 09:34:06 crc kubenswrapper[4846]: I1005 09:34:06.747014 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/38c22856-82b2-4ddc-9e43-0ef1399068ea-catalog-content\") pod \"redhat-operators-xccjp\" (UID: \"38c22856-82b2-4ddc-9e43-0ef1399068ea\") " pod="openshift-marketplace/redhat-operators-xccjp" Oct 05 09:34:06 crc kubenswrapper[4846]: I1005 09:34:06.747092 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/38c22856-82b2-4ddc-9e43-0ef1399068ea-utilities\") pod \"redhat-operators-xccjp\" (UID: \"38c22856-82b2-4ddc-9e43-0ef1399068ea\") " pod="openshift-marketplace/redhat-operators-xccjp" Oct 05 09:34:06 crc kubenswrapper[4846]: I1005 09:34:06.747572 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/38c22856-82b2-4ddc-9e43-0ef1399068ea-utilities\") pod \"redhat-operators-xccjp\" (UID: \"38c22856-82b2-4ddc-9e43-0ef1399068ea\") " pod="openshift-marketplace/redhat-operators-xccjp" Oct 05 09:34:06 crc kubenswrapper[4846]: I1005 09:34:06.748042 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/38c22856-82b2-4ddc-9e43-0ef1399068ea-catalog-content\") pod \"redhat-operators-xccjp\" (UID: \"38c22856-82b2-4ddc-9e43-0ef1399068ea\") " pod="openshift-marketplace/redhat-operators-xccjp" Oct 05 09:34:06 crc kubenswrapper[4846]: I1005 09:34:06.765524 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x2t9v\" (UniqueName: \"kubernetes.io/projected/38c22856-82b2-4ddc-9e43-0ef1399068ea-kube-api-access-x2t9v\") pod \"redhat-operators-xccjp\" (UID: \"38c22856-82b2-4ddc-9e43-0ef1399068ea\") " pod="openshift-marketplace/redhat-operators-xccjp" Oct 05 09:34:06 crc kubenswrapper[4846]: I1005 09:34:06.896823 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xccjp" Oct 05 09:34:07 crc kubenswrapper[4846]: I1005 09:34:07.378122 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-xccjp"] Oct 05 09:34:07 crc kubenswrapper[4846]: I1005 09:34:07.513464 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xccjp" event={"ID":"38c22856-82b2-4ddc-9e43-0ef1399068ea","Type":"ContainerStarted","Data":"2e5b4f82f749f0f89077c1f2c213fef68d8225e7615e719b1dc3e026db7534f7"} Oct 05 09:34:08 crc kubenswrapper[4846]: I1005 09:34:08.540812 4846 generic.go:334] "Generic (PLEG): container finished" podID="38c22856-82b2-4ddc-9e43-0ef1399068ea" containerID="d69c48b9cd7a9948e0959a4559a2e7eeb5949d132c46afcb0c71a9aa950e750e" exitCode=0 Oct 05 09:34:08 crc kubenswrapper[4846]: I1005 09:34:08.541144 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xccjp" event={"ID":"38c22856-82b2-4ddc-9e43-0ef1399068ea","Type":"ContainerDied","Data":"d69c48b9cd7a9948e0959a4559a2e7eeb5949d132c46afcb0c71a9aa950e750e"} Oct 05 09:34:09 crc kubenswrapper[4846]: I1005 09:34:09.498501 4846 scope.go:117] "RemoveContainer" containerID="8a13caff3cf7131d82ade0de5fe733ded9547a5758c374eab87ab478fe594e4d" Oct 05 09:34:09 crc kubenswrapper[4846]: E1005 09:34:09.499477 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 09:34:09 crc kubenswrapper[4846]: I1005 09:34:09.541410 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-copy-data" Oct 05 09:34:09 crc kubenswrapper[4846]: I1005 09:34:09.554677 4846 generic.go:334] "Generic (PLEG): container finished" podID="f0768736-73ad-4ce5-be3c-1941d56cd760" containerID="b2793fb5da2d48dbd09af9440210742861a05910349aabc2730a5269a46eecc4" exitCode=137 Oct 05 09:34:09 crc kubenswrapper[4846]: I1005 09:34:09.554756 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-copy-data" event={"ID":"f0768736-73ad-4ce5-be3c-1941d56cd760","Type":"ContainerDied","Data":"b2793fb5da2d48dbd09af9440210742861a05910349aabc2730a5269a46eecc4"} Oct 05 09:34:09 crc kubenswrapper[4846]: I1005 09:34:09.554766 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-copy-data" Oct 05 09:34:09 crc kubenswrapper[4846]: I1005 09:34:09.554929 4846 scope.go:117] "RemoveContainer" containerID="b2793fb5da2d48dbd09af9440210742861a05910349aabc2730a5269a46eecc4" Oct 05 09:34:09 crc kubenswrapper[4846]: I1005 09:34:09.555705 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-copy-data" event={"ID":"f0768736-73ad-4ce5-be3c-1941d56cd760","Type":"ContainerDied","Data":"199d01d1d7efc979b2c00c0d62bafdfbab10d8cf0ac1b73fd55f71c4407448ab"} Oct 05 09:34:09 crc kubenswrapper[4846]: I1005 09:34:09.588470 4846 scope.go:117] "RemoveContainer" containerID="b2793fb5da2d48dbd09af9440210742861a05910349aabc2730a5269a46eecc4" Oct 05 09:34:09 crc kubenswrapper[4846]: E1005 09:34:09.589542 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b2793fb5da2d48dbd09af9440210742861a05910349aabc2730a5269a46eecc4\": container with ID starting with b2793fb5da2d48dbd09af9440210742861a05910349aabc2730a5269a46eecc4 not found: ID does not exist" containerID="b2793fb5da2d48dbd09af9440210742861a05910349aabc2730a5269a46eecc4" Oct 05 09:34:09 crc kubenswrapper[4846]: I1005 09:34:09.589587 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b2793fb5da2d48dbd09af9440210742861a05910349aabc2730a5269a46eecc4"} err="failed to get container status \"b2793fb5da2d48dbd09af9440210742861a05910349aabc2730a5269a46eecc4\": rpc error: code = NotFound desc = could not find container \"b2793fb5da2d48dbd09af9440210742861a05910349aabc2730a5269a46eecc4\": container with ID starting with b2793fb5da2d48dbd09af9440210742861a05910349aabc2730a5269a46eecc4 not found: ID does not exist" Oct 05 09:34:09 crc kubenswrapper[4846]: I1005 09:34:09.610198 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-data\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-7d55f83c-6f76-43e2-9773-bc401946e9b8\") pod \"f0768736-73ad-4ce5-be3c-1941d56cd760\" (UID: \"f0768736-73ad-4ce5-be3c-1941d56cd760\") " Oct 05 09:34:09 crc kubenswrapper[4846]: I1005 09:34:09.610698 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-data-cert\" (UniqueName: \"kubernetes.io/secret/f0768736-73ad-4ce5-be3c-1941d56cd760-ovn-data-cert\") pod \"f0768736-73ad-4ce5-be3c-1941d56cd760\" (UID: \"f0768736-73ad-4ce5-be3c-1941d56cd760\") " Oct 05 09:34:09 crc kubenswrapper[4846]: I1005 09:34:09.610832 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fdfrw\" (UniqueName: \"kubernetes.io/projected/f0768736-73ad-4ce5-be3c-1941d56cd760-kube-api-access-fdfrw\") pod \"f0768736-73ad-4ce5-be3c-1941d56cd760\" (UID: \"f0768736-73ad-4ce5-be3c-1941d56cd760\") " Oct 05 09:34:09 crc kubenswrapper[4846]: I1005 09:34:09.619137 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f0768736-73ad-4ce5-be3c-1941d56cd760-ovn-data-cert" (OuterVolumeSpecName: "ovn-data-cert") pod "f0768736-73ad-4ce5-be3c-1941d56cd760" (UID: "f0768736-73ad-4ce5-be3c-1941d56cd760"). InnerVolumeSpecName "ovn-data-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 09:34:09 crc kubenswrapper[4846]: I1005 09:34:09.631900 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f0768736-73ad-4ce5-be3c-1941d56cd760-kube-api-access-fdfrw" (OuterVolumeSpecName: "kube-api-access-fdfrw") pod "f0768736-73ad-4ce5-be3c-1941d56cd760" (UID: "f0768736-73ad-4ce5-be3c-1941d56cd760"). InnerVolumeSpecName "kube-api-access-fdfrw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 09:34:09 crc kubenswrapper[4846]: I1005 09:34:09.632139 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-7d55f83c-6f76-43e2-9773-bc401946e9b8" (OuterVolumeSpecName: "ovn-data") pod "f0768736-73ad-4ce5-be3c-1941d56cd760" (UID: "f0768736-73ad-4ce5-be3c-1941d56cd760"). InnerVolumeSpecName "pvc-7d55f83c-6f76-43e2-9773-bc401946e9b8". PluginName "kubernetes.io/csi", VolumeGidValue "" Oct 05 09:34:09 crc kubenswrapper[4846]: I1005 09:34:09.714007 4846 reconciler_common.go:293] "Volume detached for volume \"ovn-data-cert\" (UniqueName: \"kubernetes.io/secret/f0768736-73ad-4ce5-be3c-1941d56cd760-ovn-data-cert\") on node \"crc\" DevicePath \"\"" Oct 05 09:34:09 crc kubenswrapper[4846]: I1005 09:34:09.714495 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fdfrw\" (UniqueName: \"kubernetes.io/projected/f0768736-73ad-4ce5-be3c-1941d56cd760-kube-api-access-fdfrw\") on node \"crc\" DevicePath \"\"" Oct 05 09:34:09 crc kubenswrapper[4846]: I1005 09:34:09.714535 4846 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-7d55f83c-6f76-43e2-9773-bc401946e9b8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-7d55f83c-6f76-43e2-9773-bc401946e9b8\") on node \"crc\" " Oct 05 09:34:09 crc kubenswrapper[4846]: I1005 09:34:09.740398 4846 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Oct 05 09:34:09 crc kubenswrapper[4846]: I1005 09:34:09.741381 4846 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-7d55f83c-6f76-43e2-9773-bc401946e9b8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-7d55f83c-6f76-43e2-9773-bc401946e9b8") on node "crc" Oct 05 09:34:09 crc kubenswrapper[4846]: I1005 09:34:09.817173 4846 reconciler_common.go:293] "Volume detached for volume \"pvc-7d55f83c-6f76-43e2-9773-bc401946e9b8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-7d55f83c-6f76-43e2-9773-bc401946e9b8\") on node \"crc\" DevicePath \"\"" Oct 05 09:34:09 crc kubenswrapper[4846]: I1005 09:34:09.904309 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-copy-data"] Oct 05 09:34:09 crc kubenswrapper[4846]: I1005 09:34:09.914891 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-copy-data"] Oct 05 09:34:10 crc kubenswrapper[4846]: I1005 09:34:10.510412 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f0768736-73ad-4ce5-be3c-1941d56cd760" path="/var/lib/kubelet/pods/f0768736-73ad-4ce5-be3c-1941d56cd760/volumes" Oct 05 09:34:10 crc kubenswrapper[4846]: I1005 09:34:10.567917 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xccjp" event={"ID":"38c22856-82b2-4ddc-9e43-0ef1399068ea","Type":"ContainerStarted","Data":"4d539beeafe32a167235623a05fb294e56175f47531244f485edb138faf10a58"} Oct 05 09:34:13 crc kubenswrapper[4846]: I1005 09:34:13.602812 4846 generic.go:334] "Generic (PLEG): container finished" podID="38c22856-82b2-4ddc-9e43-0ef1399068ea" containerID="4d539beeafe32a167235623a05fb294e56175f47531244f485edb138faf10a58" exitCode=0 Oct 05 09:34:13 crc kubenswrapper[4846]: I1005 09:34:13.602888 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xccjp" event={"ID":"38c22856-82b2-4ddc-9e43-0ef1399068ea","Type":"ContainerDied","Data":"4d539beeafe32a167235623a05fb294e56175f47531244f485edb138faf10a58"} Oct 05 09:34:15 crc kubenswrapper[4846]: I1005 09:34:15.626058 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xccjp" event={"ID":"38c22856-82b2-4ddc-9e43-0ef1399068ea","Type":"ContainerStarted","Data":"05ae53e0d64d62eef798aadabb12ff7e5cbfa84d6b35a4e5eb52503573e903b1"} Oct 05 09:34:15 crc kubenswrapper[4846]: I1005 09:34:15.652126 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-xccjp" podStartSLOduration=4.178980793 podStartE2EDuration="9.652109734s" podCreationTimestamp="2025-10-05 09:34:06 +0000 UTC" firstStartedPulling="2025-10-05 09:34:08.545373638 +0000 UTC m=+9970.786226453" lastFinishedPulling="2025-10-05 09:34:14.018502589 +0000 UTC m=+9976.259355394" observedRunningTime="2025-10-05 09:34:15.647246018 +0000 UTC m=+9977.888098803" watchObservedRunningTime="2025-10-05 09:34:15.652109734 +0000 UTC m=+9977.892962519" Oct 05 09:34:16 crc kubenswrapper[4846]: I1005 09:34:16.897258 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-xccjp" Oct 05 09:34:16 crc kubenswrapper[4846]: I1005 09:34:16.897467 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-xccjp" Oct 05 09:34:17 crc kubenswrapper[4846]: I1005 09:34:17.953143 4846 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-xccjp" podUID="38c22856-82b2-4ddc-9e43-0ef1399068ea" containerName="registry-server" probeResult="failure" output=< Oct 05 09:34:17 crc kubenswrapper[4846]: timeout: failed to connect service ":50051" within 1s Oct 05 09:34:17 crc kubenswrapper[4846]: > Oct 05 09:34:23 crc kubenswrapper[4846]: I1005 09:34:23.499658 4846 scope.go:117] "RemoveContainer" containerID="8a13caff3cf7131d82ade0de5fe733ded9547a5758c374eab87ab478fe594e4d" Oct 05 09:34:23 crc kubenswrapper[4846]: E1005 09:34:23.501085 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 09:34:27 crc kubenswrapper[4846]: I1005 09:34:27.775290 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-xccjp" Oct 05 09:34:27 crc kubenswrapper[4846]: I1005 09:34:27.831341 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-xccjp" Oct 05 09:34:28 crc kubenswrapper[4846]: I1005 09:34:28.028033 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-xccjp"] Oct 05 09:34:29 crc kubenswrapper[4846]: I1005 09:34:29.800959 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-xccjp" podUID="38c22856-82b2-4ddc-9e43-0ef1399068ea" containerName="registry-server" containerID="cri-o://05ae53e0d64d62eef798aadabb12ff7e5cbfa84d6b35a4e5eb52503573e903b1" gracePeriod=2 Oct 05 09:34:30 crc kubenswrapper[4846]: E1005 09:34:30.138722 4846 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod38c22856_82b2_4ddc_9e43_0ef1399068ea.slice/crio-conmon-05ae53e0d64d62eef798aadabb12ff7e5cbfa84d6b35a4e5eb52503573e903b1.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod38c22856_82b2_4ddc_9e43_0ef1399068ea.slice/crio-05ae53e0d64d62eef798aadabb12ff7e5cbfa84d6b35a4e5eb52503573e903b1.scope\": RecentStats: unable to find data in memory cache]" Oct 05 09:34:30 crc kubenswrapper[4846]: I1005 09:34:30.373144 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xccjp" Oct 05 09:34:30 crc kubenswrapper[4846]: I1005 09:34:30.541088 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/38c22856-82b2-4ddc-9e43-0ef1399068ea-catalog-content\") pod \"38c22856-82b2-4ddc-9e43-0ef1399068ea\" (UID: \"38c22856-82b2-4ddc-9e43-0ef1399068ea\") " Oct 05 09:34:30 crc kubenswrapper[4846]: I1005 09:34:30.541161 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2t9v\" (UniqueName: \"kubernetes.io/projected/38c22856-82b2-4ddc-9e43-0ef1399068ea-kube-api-access-x2t9v\") pod \"38c22856-82b2-4ddc-9e43-0ef1399068ea\" (UID: \"38c22856-82b2-4ddc-9e43-0ef1399068ea\") " Oct 05 09:34:30 crc kubenswrapper[4846]: I1005 09:34:30.541339 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/38c22856-82b2-4ddc-9e43-0ef1399068ea-utilities\") pod \"38c22856-82b2-4ddc-9e43-0ef1399068ea\" (UID: \"38c22856-82b2-4ddc-9e43-0ef1399068ea\") " Oct 05 09:34:30 crc kubenswrapper[4846]: I1005 09:34:30.543155 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/38c22856-82b2-4ddc-9e43-0ef1399068ea-utilities" (OuterVolumeSpecName: "utilities") pod "38c22856-82b2-4ddc-9e43-0ef1399068ea" (UID: "38c22856-82b2-4ddc-9e43-0ef1399068ea"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 09:34:30 crc kubenswrapper[4846]: I1005 09:34:30.550594 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/38c22856-82b2-4ddc-9e43-0ef1399068ea-kube-api-access-x2t9v" (OuterVolumeSpecName: "kube-api-access-x2t9v") pod "38c22856-82b2-4ddc-9e43-0ef1399068ea" (UID: "38c22856-82b2-4ddc-9e43-0ef1399068ea"). InnerVolumeSpecName "kube-api-access-x2t9v". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 09:34:30 crc kubenswrapper[4846]: I1005 09:34:30.644053 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2t9v\" (UniqueName: \"kubernetes.io/projected/38c22856-82b2-4ddc-9e43-0ef1399068ea-kube-api-access-x2t9v\") on node \"crc\" DevicePath \"\"" Oct 05 09:34:30 crc kubenswrapper[4846]: I1005 09:34:30.644101 4846 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/38c22856-82b2-4ddc-9e43-0ef1399068ea-utilities\") on node \"crc\" DevicePath \"\"" Oct 05 09:34:30 crc kubenswrapper[4846]: I1005 09:34:30.676730 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/38c22856-82b2-4ddc-9e43-0ef1399068ea-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "38c22856-82b2-4ddc-9e43-0ef1399068ea" (UID: "38c22856-82b2-4ddc-9e43-0ef1399068ea"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 09:34:30 crc kubenswrapper[4846]: I1005 09:34:30.747937 4846 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/38c22856-82b2-4ddc-9e43-0ef1399068ea-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 05 09:34:30 crc kubenswrapper[4846]: I1005 09:34:30.816879 4846 generic.go:334] "Generic (PLEG): container finished" podID="38c22856-82b2-4ddc-9e43-0ef1399068ea" containerID="05ae53e0d64d62eef798aadabb12ff7e5cbfa84d6b35a4e5eb52503573e903b1" exitCode=0 Oct 05 09:34:30 crc kubenswrapper[4846]: I1005 09:34:30.816998 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xccjp" Oct 05 09:34:30 crc kubenswrapper[4846]: I1005 09:34:30.817013 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xccjp" event={"ID":"38c22856-82b2-4ddc-9e43-0ef1399068ea","Type":"ContainerDied","Data":"05ae53e0d64d62eef798aadabb12ff7e5cbfa84d6b35a4e5eb52503573e903b1"} Oct 05 09:34:30 crc kubenswrapper[4846]: I1005 09:34:30.817499 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xccjp" event={"ID":"38c22856-82b2-4ddc-9e43-0ef1399068ea","Type":"ContainerDied","Data":"2e5b4f82f749f0f89077c1f2c213fef68d8225e7615e719b1dc3e026db7534f7"} Oct 05 09:34:30 crc kubenswrapper[4846]: I1005 09:34:30.817541 4846 scope.go:117] "RemoveContainer" containerID="05ae53e0d64d62eef798aadabb12ff7e5cbfa84d6b35a4e5eb52503573e903b1" Oct 05 09:34:30 crc kubenswrapper[4846]: I1005 09:34:30.868408 4846 scope.go:117] "RemoveContainer" containerID="4d539beeafe32a167235623a05fb294e56175f47531244f485edb138faf10a58" Oct 05 09:34:30 crc kubenswrapper[4846]: I1005 09:34:30.871976 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-xccjp"] Oct 05 09:34:30 crc kubenswrapper[4846]: I1005 09:34:30.893988 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-xccjp"] Oct 05 09:34:30 crc kubenswrapper[4846]: I1005 09:34:30.895339 4846 scope.go:117] "RemoveContainer" containerID="d69c48b9cd7a9948e0959a4559a2e7eeb5949d132c46afcb0c71a9aa950e750e" Oct 05 09:34:30 crc kubenswrapper[4846]: I1005 09:34:30.950845 4846 scope.go:117] "RemoveContainer" containerID="05ae53e0d64d62eef798aadabb12ff7e5cbfa84d6b35a4e5eb52503573e903b1" Oct 05 09:34:30 crc kubenswrapper[4846]: E1005 09:34:30.955501 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"05ae53e0d64d62eef798aadabb12ff7e5cbfa84d6b35a4e5eb52503573e903b1\": container with ID starting with 05ae53e0d64d62eef798aadabb12ff7e5cbfa84d6b35a4e5eb52503573e903b1 not found: ID does not exist" containerID="05ae53e0d64d62eef798aadabb12ff7e5cbfa84d6b35a4e5eb52503573e903b1" Oct 05 09:34:30 crc kubenswrapper[4846]: I1005 09:34:30.955575 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"05ae53e0d64d62eef798aadabb12ff7e5cbfa84d6b35a4e5eb52503573e903b1"} err="failed to get container status \"05ae53e0d64d62eef798aadabb12ff7e5cbfa84d6b35a4e5eb52503573e903b1\": rpc error: code = NotFound desc = could not find container \"05ae53e0d64d62eef798aadabb12ff7e5cbfa84d6b35a4e5eb52503573e903b1\": container with ID starting with 05ae53e0d64d62eef798aadabb12ff7e5cbfa84d6b35a4e5eb52503573e903b1 not found: ID does not exist" Oct 05 09:34:30 crc kubenswrapper[4846]: I1005 09:34:30.955623 4846 scope.go:117] "RemoveContainer" containerID="4d539beeafe32a167235623a05fb294e56175f47531244f485edb138faf10a58" Oct 05 09:34:30 crc kubenswrapper[4846]: E1005 09:34:30.957534 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4d539beeafe32a167235623a05fb294e56175f47531244f485edb138faf10a58\": container with ID starting with 4d539beeafe32a167235623a05fb294e56175f47531244f485edb138faf10a58 not found: ID does not exist" containerID="4d539beeafe32a167235623a05fb294e56175f47531244f485edb138faf10a58" Oct 05 09:34:30 crc kubenswrapper[4846]: I1005 09:34:30.957595 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4d539beeafe32a167235623a05fb294e56175f47531244f485edb138faf10a58"} err="failed to get container status \"4d539beeafe32a167235623a05fb294e56175f47531244f485edb138faf10a58\": rpc error: code = NotFound desc = could not find container \"4d539beeafe32a167235623a05fb294e56175f47531244f485edb138faf10a58\": container with ID starting with 4d539beeafe32a167235623a05fb294e56175f47531244f485edb138faf10a58 not found: ID does not exist" Oct 05 09:34:30 crc kubenswrapper[4846]: I1005 09:34:30.957628 4846 scope.go:117] "RemoveContainer" containerID="d69c48b9cd7a9948e0959a4559a2e7eeb5949d132c46afcb0c71a9aa950e750e" Oct 05 09:34:30 crc kubenswrapper[4846]: E1005 09:34:30.958122 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d69c48b9cd7a9948e0959a4559a2e7eeb5949d132c46afcb0c71a9aa950e750e\": container with ID starting with d69c48b9cd7a9948e0959a4559a2e7eeb5949d132c46afcb0c71a9aa950e750e not found: ID does not exist" containerID="d69c48b9cd7a9948e0959a4559a2e7eeb5949d132c46afcb0c71a9aa950e750e" Oct 05 09:34:30 crc kubenswrapper[4846]: I1005 09:34:30.958194 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d69c48b9cd7a9948e0959a4559a2e7eeb5949d132c46afcb0c71a9aa950e750e"} err="failed to get container status \"d69c48b9cd7a9948e0959a4559a2e7eeb5949d132c46afcb0c71a9aa950e750e\": rpc error: code = NotFound desc = could not find container \"d69c48b9cd7a9948e0959a4559a2e7eeb5949d132c46afcb0c71a9aa950e750e\": container with ID starting with d69c48b9cd7a9948e0959a4559a2e7eeb5949d132c46afcb0c71a9aa950e750e not found: ID does not exist" Oct 05 09:34:31 crc kubenswrapper[4846]: I1005 09:34:31.032905 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/tempest-tests-tempest"] Oct 05 09:34:31 crc kubenswrapper[4846]: E1005 09:34:31.033581 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="38c22856-82b2-4ddc-9e43-0ef1399068ea" containerName="extract-content" Oct 05 09:34:31 crc kubenswrapper[4846]: I1005 09:34:31.033605 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="38c22856-82b2-4ddc-9e43-0ef1399068ea" containerName="extract-content" Oct 05 09:34:31 crc kubenswrapper[4846]: E1005 09:34:31.033621 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="38c22856-82b2-4ddc-9e43-0ef1399068ea" containerName="registry-server" Oct 05 09:34:31 crc kubenswrapper[4846]: I1005 09:34:31.033631 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="38c22856-82b2-4ddc-9e43-0ef1399068ea" containerName="registry-server" Oct 05 09:34:31 crc kubenswrapper[4846]: E1005 09:34:31.033665 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="38c22856-82b2-4ddc-9e43-0ef1399068ea" containerName="extract-utilities" Oct 05 09:34:31 crc kubenswrapper[4846]: I1005 09:34:31.033675 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="38c22856-82b2-4ddc-9e43-0ef1399068ea" containerName="extract-utilities" Oct 05 09:34:31 crc kubenswrapper[4846]: E1005 09:34:31.033696 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f0768736-73ad-4ce5-be3c-1941d56cd760" containerName="adoption" Oct 05 09:34:31 crc kubenswrapper[4846]: I1005 09:34:31.033704 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="f0768736-73ad-4ce5-be3c-1941d56cd760" containerName="adoption" Oct 05 09:34:31 crc kubenswrapper[4846]: I1005 09:34:31.033956 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="38c22856-82b2-4ddc-9e43-0ef1399068ea" containerName="registry-server" Oct 05 09:34:31 crc kubenswrapper[4846]: I1005 09:34:31.033993 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="f0768736-73ad-4ce5-be3c-1941d56cd760" containerName="adoption" Oct 05 09:34:31 crc kubenswrapper[4846]: I1005 09:34:31.034975 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Oct 05 09:34:31 crc kubenswrapper[4846]: I1005 09:34:31.039436 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s0" Oct 05 09:34:31 crc kubenswrapper[4846]: I1005 09:34:31.039705 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-4c82f" Oct 05 09:34:31 crc kubenswrapper[4846]: I1005 09:34:31.039872 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-custom-data-s0" Oct 05 09:34:31 crc kubenswrapper[4846]: I1005 09:34:31.040074 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"test-operator-controller-priv-key" Oct 05 09:34:31 crc kubenswrapper[4846]: I1005 09:34:31.044393 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Oct 05 09:34:31 crc kubenswrapper[4846]: I1005 09:34:31.160319 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/fac403c1-c8cd-4675-ba46-59900a7cc18f-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"fac403c1-c8cd-4675-ba46-59900a7cc18f\") " pod="openstack/tempest-tests-tempest" Oct 05 09:34:31 crc kubenswrapper[4846]: I1005 09:34:31.160398 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/fac403c1-c8cd-4675-ba46-59900a7cc18f-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"fac403c1-c8cd-4675-ba46-59900a7cc18f\") " pod="openstack/tempest-tests-tempest" Oct 05 09:34:31 crc kubenswrapper[4846]: I1005 09:34:31.160494 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fac403c1-c8cd-4675-ba46-59900a7cc18f-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"fac403c1-c8cd-4675-ba46-59900a7cc18f\") " pod="openstack/tempest-tests-tempest" Oct 05 09:34:31 crc kubenswrapper[4846]: I1005 09:34:31.160549 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"tempest-tests-tempest\" (UID: \"fac403c1-c8cd-4675-ba46-59900a7cc18f\") " pod="openstack/tempest-tests-tempest" Oct 05 09:34:31 crc kubenswrapper[4846]: I1005 09:34:31.160584 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/fac403c1-c8cd-4675-ba46-59900a7cc18f-config-data\") pod \"tempest-tests-tempest\" (UID: \"fac403c1-c8cd-4675-ba46-59900a7cc18f\") " pod="openstack/tempest-tests-tempest" Oct 05 09:34:31 crc kubenswrapper[4846]: I1005 09:34:31.160607 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/fac403c1-c8cd-4675-ba46-59900a7cc18f-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"fac403c1-c8cd-4675-ba46-59900a7cc18f\") " pod="openstack/tempest-tests-tempest" Oct 05 09:34:31 crc kubenswrapper[4846]: I1005 09:34:31.160627 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h85lt\" (UniqueName: \"kubernetes.io/projected/fac403c1-c8cd-4675-ba46-59900a7cc18f-kube-api-access-h85lt\") pod \"tempest-tests-tempest\" (UID: \"fac403c1-c8cd-4675-ba46-59900a7cc18f\") " pod="openstack/tempest-tests-tempest" Oct 05 09:34:31 crc kubenswrapper[4846]: I1005 09:34:31.160931 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/fac403c1-c8cd-4675-ba46-59900a7cc18f-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"fac403c1-c8cd-4675-ba46-59900a7cc18f\") " pod="openstack/tempest-tests-tempest" Oct 05 09:34:31 crc kubenswrapper[4846]: I1005 09:34:31.161030 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/fac403c1-c8cd-4675-ba46-59900a7cc18f-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"fac403c1-c8cd-4675-ba46-59900a7cc18f\") " pod="openstack/tempest-tests-tempest" Oct 05 09:34:31 crc kubenswrapper[4846]: I1005 09:34:31.263532 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fac403c1-c8cd-4675-ba46-59900a7cc18f-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"fac403c1-c8cd-4675-ba46-59900a7cc18f\") " pod="openstack/tempest-tests-tempest" Oct 05 09:34:31 crc kubenswrapper[4846]: I1005 09:34:31.263668 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"tempest-tests-tempest\" (UID: \"fac403c1-c8cd-4675-ba46-59900a7cc18f\") " pod="openstack/tempest-tests-tempest" Oct 05 09:34:31 crc kubenswrapper[4846]: I1005 09:34:31.263702 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/fac403c1-c8cd-4675-ba46-59900a7cc18f-config-data\") pod \"tempest-tests-tempest\" (UID: \"fac403c1-c8cd-4675-ba46-59900a7cc18f\") " pod="openstack/tempest-tests-tempest" Oct 05 09:34:31 crc kubenswrapper[4846]: I1005 09:34:31.263731 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/fac403c1-c8cd-4675-ba46-59900a7cc18f-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"fac403c1-c8cd-4675-ba46-59900a7cc18f\") " pod="openstack/tempest-tests-tempest" Oct 05 09:34:31 crc kubenswrapper[4846]: I1005 09:34:31.263759 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h85lt\" (UniqueName: \"kubernetes.io/projected/fac403c1-c8cd-4675-ba46-59900a7cc18f-kube-api-access-h85lt\") pod \"tempest-tests-tempest\" (UID: \"fac403c1-c8cd-4675-ba46-59900a7cc18f\") " pod="openstack/tempest-tests-tempest" Oct 05 09:34:31 crc kubenswrapper[4846]: I1005 09:34:31.263919 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/fac403c1-c8cd-4675-ba46-59900a7cc18f-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"fac403c1-c8cd-4675-ba46-59900a7cc18f\") " pod="openstack/tempest-tests-tempest" Oct 05 09:34:31 crc kubenswrapper[4846]: I1005 09:34:31.263968 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/fac403c1-c8cd-4675-ba46-59900a7cc18f-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"fac403c1-c8cd-4675-ba46-59900a7cc18f\") " pod="openstack/tempest-tests-tempest" Oct 05 09:34:31 crc kubenswrapper[4846]: I1005 09:34:31.264032 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/fac403c1-c8cd-4675-ba46-59900a7cc18f-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"fac403c1-c8cd-4675-ba46-59900a7cc18f\") " pod="openstack/tempest-tests-tempest" Oct 05 09:34:31 crc kubenswrapper[4846]: I1005 09:34:31.264061 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/fac403c1-c8cd-4675-ba46-59900a7cc18f-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"fac403c1-c8cd-4675-ba46-59900a7cc18f\") " pod="openstack/tempest-tests-tempest" Oct 05 09:34:31 crc kubenswrapper[4846]: I1005 09:34:31.264341 4846 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"tempest-tests-tempest\" (UID: \"fac403c1-c8cd-4675-ba46-59900a7cc18f\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/tempest-tests-tempest" Oct 05 09:34:31 crc kubenswrapper[4846]: I1005 09:34:31.264961 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/fac403c1-c8cd-4675-ba46-59900a7cc18f-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"fac403c1-c8cd-4675-ba46-59900a7cc18f\") " pod="openstack/tempest-tests-tempest" Oct 05 09:34:31 crc kubenswrapper[4846]: I1005 09:34:31.265857 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/fac403c1-c8cd-4675-ba46-59900a7cc18f-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"fac403c1-c8cd-4675-ba46-59900a7cc18f\") " pod="openstack/tempest-tests-tempest" Oct 05 09:34:31 crc kubenswrapper[4846]: I1005 09:34:31.266216 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/fac403c1-c8cd-4675-ba46-59900a7cc18f-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"fac403c1-c8cd-4675-ba46-59900a7cc18f\") " pod="openstack/tempest-tests-tempest" Oct 05 09:34:31 crc kubenswrapper[4846]: I1005 09:34:31.266536 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/fac403c1-c8cd-4675-ba46-59900a7cc18f-config-data\") pod \"tempest-tests-tempest\" (UID: \"fac403c1-c8cd-4675-ba46-59900a7cc18f\") " pod="openstack/tempest-tests-tempest" Oct 05 09:34:31 crc kubenswrapper[4846]: I1005 09:34:31.270010 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/fac403c1-c8cd-4675-ba46-59900a7cc18f-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"fac403c1-c8cd-4675-ba46-59900a7cc18f\") " pod="openstack/tempest-tests-tempest" Oct 05 09:34:31 crc kubenswrapper[4846]: I1005 09:34:31.271282 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/fac403c1-c8cd-4675-ba46-59900a7cc18f-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"fac403c1-c8cd-4675-ba46-59900a7cc18f\") " pod="openstack/tempest-tests-tempest" Oct 05 09:34:31 crc kubenswrapper[4846]: I1005 09:34:31.272467 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fac403c1-c8cd-4675-ba46-59900a7cc18f-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"fac403c1-c8cd-4675-ba46-59900a7cc18f\") " pod="openstack/tempest-tests-tempest" Oct 05 09:34:31 crc kubenswrapper[4846]: I1005 09:34:31.297014 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h85lt\" (UniqueName: \"kubernetes.io/projected/fac403c1-c8cd-4675-ba46-59900a7cc18f-kube-api-access-h85lt\") pod \"tempest-tests-tempest\" (UID: \"fac403c1-c8cd-4675-ba46-59900a7cc18f\") " pod="openstack/tempest-tests-tempest" Oct 05 09:34:31 crc kubenswrapper[4846]: I1005 09:34:31.305286 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"tempest-tests-tempest\" (UID: \"fac403c1-c8cd-4675-ba46-59900a7cc18f\") " pod="openstack/tempest-tests-tempest" Oct 05 09:34:31 crc kubenswrapper[4846]: I1005 09:34:31.354857 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Oct 05 09:34:31 crc kubenswrapper[4846]: I1005 09:34:31.928116 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Oct 05 09:34:31 crc kubenswrapper[4846]: W1005 09:34:31.929327 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfac403c1_c8cd_4675_ba46_59900a7cc18f.slice/crio-b7d0f7c84d87c9a5ae9e1ca11f6598b1e30d5c4845659c44520b41e9d1c54dd5 WatchSource:0}: Error finding container b7d0f7c84d87c9a5ae9e1ca11f6598b1e30d5c4845659c44520b41e9d1c54dd5: Status 404 returned error can't find the container with id b7d0f7c84d87c9a5ae9e1ca11f6598b1e30d5c4845659c44520b41e9d1c54dd5 Oct 05 09:34:32 crc kubenswrapper[4846]: I1005 09:34:32.515845 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="38c22856-82b2-4ddc-9e43-0ef1399068ea" path="/var/lib/kubelet/pods/38c22856-82b2-4ddc-9e43-0ef1399068ea/volumes" Oct 05 09:34:32 crc kubenswrapper[4846]: I1005 09:34:32.883357 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"fac403c1-c8cd-4675-ba46-59900a7cc18f","Type":"ContainerStarted","Data":"b7d0f7c84d87c9a5ae9e1ca11f6598b1e30d5c4845659c44520b41e9d1c54dd5"} Oct 05 09:34:34 crc kubenswrapper[4846]: I1005 09:34:34.498307 4846 scope.go:117] "RemoveContainer" containerID="8a13caff3cf7131d82ade0de5fe733ded9547a5758c374eab87ab478fe594e4d" Oct 05 09:34:34 crc kubenswrapper[4846]: E1005 09:34:34.499286 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 09:34:45 crc kubenswrapper[4846]: I1005 09:34:45.497862 4846 scope.go:117] "RemoveContainer" containerID="8a13caff3cf7131d82ade0de5fe733ded9547a5758c374eab87ab478fe594e4d" Oct 05 09:34:45 crc kubenswrapper[4846]: E1005 09:34:45.498734 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 09:34:57 crc kubenswrapper[4846]: I1005 09:34:57.497719 4846 scope.go:117] "RemoveContainer" containerID="8a13caff3cf7131d82ade0de5fe733ded9547a5758c374eab87ab478fe594e4d" Oct 05 09:34:57 crc kubenswrapper[4846]: E1005 09:34:57.498713 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 09:35:09 crc kubenswrapper[4846]: I1005 09:35:09.498493 4846 scope.go:117] "RemoveContainer" containerID="8a13caff3cf7131d82ade0de5fe733ded9547a5758c374eab87ab478fe594e4d" Oct 05 09:35:09 crc kubenswrapper[4846]: E1005 09:35:09.499329 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 09:35:15 crc kubenswrapper[4846]: E1005 09:35:15.686643 4846 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.rdoproject.org/podified-antelope-centos9/openstack-tempest-all:88dc57612f447daadb492dcf3ad854ac" Oct 05 09:35:15 crc kubenswrapper[4846]: E1005 09:35:15.687299 4846 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.rdoproject.org/podified-antelope-centos9/openstack-tempest-all:88dc57612f447daadb492dcf3ad854ac" Oct 05 09:35:15 crc kubenswrapper[4846]: E1005 09:35:15.687587 4846 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:tempest-tests-tempest-tests-runner,Image:quay.rdoproject.org/podified-antelope-centos9/openstack-tempest-all:88dc57612f447daadb492dcf3ad854ac,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:test-operator-ephemeral-workdir,ReadOnly:false,MountPath:/var/lib/tempest,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-ephemeral-temporary,ReadOnly:false,MountPath:/tmp,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:false,MountPath:/etc/test_operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-logs,ReadOnly:false,MountPath:/var/lib/tempest/external_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/etc/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/var/lib/tempest/.config/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config-secret,ReadOnly:false,MountPath:/etc/openstack/secure.yaml,SubPath:secure.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ca-certs,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ssh-key,ReadOnly:false,MountPath:/var/lib/tempest/id_ecdsa,SubPath:ssh_key,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-h85lt,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42480,RunAsNonRoot:*false,ReadOnlyRootFilesystem:*false,AllowPrivilegeEscalation:*true,RunAsGroup:*42480,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-custom-data-s0,},Optional:nil,},SecretRef:nil,},EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-env-vars-s0,},Optional:nil,},SecretRef:nil,},},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod tempest-tests-tempest_openstack(fac403c1-c8cd-4675-ba46-59900a7cc18f): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 05 09:35:15 crc kubenswrapper[4846]: E1005 09:35:15.688813 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/tempest-tests-tempest" podUID="fac403c1-c8cd-4675-ba46-59900a7cc18f" Oct 05 09:35:16 crc kubenswrapper[4846]: E1005 09:35:16.403113 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-antelope-centos9/openstack-tempest-all:88dc57612f447daadb492dcf3ad854ac\\\"\"" pod="openstack/tempest-tests-tempest" podUID="fac403c1-c8cd-4675-ba46-59900a7cc18f" Oct 05 09:35:21 crc kubenswrapper[4846]: I1005 09:35:21.497617 4846 scope.go:117] "RemoveContainer" containerID="8a13caff3cf7131d82ade0de5fe733ded9547a5758c374eab87ab478fe594e4d" Oct 05 09:35:21 crc kubenswrapper[4846]: E1005 09:35:21.498617 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 09:35:28 crc kubenswrapper[4846]: I1005 09:35:28.522329 4846 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 05 09:35:28 crc kubenswrapper[4846]: I1005 09:35:28.764374 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s0" Oct 05 09:35:30 crc kubenswrapper[4846]: I1005 09:35:30.584262 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"fac403c1-c8cd-4675-ba46-59900a7cc18f","Type":"ContainerStarted","Data":"f38b40fca0e1a85108d450cbf8c3e04820b1957149b6e8013f5ca35446b8c5c2"} Oct 05 09:35:30 crc kubenswrapper[4846]: I1005 09:35:30.613936 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/tempest-tests-tempest" podStartSLOduration=3.784519075 podStartE2EDuration="1m0.613919581s" podCreationTimestamp="2025-10-05 09:34:30 +0000 UTC" firstStartedPulling="2025-10-05 09:34:31.932550958 +0000 UTC m=+9994.173403733" lastFinishedPulling="2025-10-05 09:35:28.761951464 +0000 UTC m=+10051.002804239" observedRunningTime="2025-10-05 09:35:30.605961406 +0000 UTC m=+10052.846814181" watchObservedRunningTime="2025-10-05 09:35:30.613919581 +0000 UTC m=+10052.854772356" Oct 05 09:35:32 crc kubenswrapper[4846]: I1005 09:35:32.498378 4846 scope.go:117] "RemoveContainer" containerID="8a13caff3cf7131d82ade0de5fe733ded9547a5758c374eab87ab478fe594e4d" Oct 05 09:35:32 crc kubenswrapper[4846]: E1005 09:35:32.499274 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 09:35:44 crc kubenswrapper[4846]: I1005 09:35:44.498319 4846 scope.go:117] "RemoveContainer" containerID="8a13caff3cf7131d82ade0de5fe733ded9547a5758c374eab87ab478fe594e4d" Oct 05 09:35:44 crc kubenswrapper[4846]: E1005 09:35:44.500032 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 09:35:57 crc kubenswrapper[4846]: I1005 09:35:57.497738 4846 scope.go:117] "RemoveContainer" containerID="8a13caff3cf7131d82ade0de5fe733ded9547a5758c374eab87ab478fe594e4d" Oct 05 09:35:57 crc kubenswrapper[4846]: E1005 09:35:57.498424 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 09:36:10 crc kubenswrapper[4846]: I1005 09:36:10.497832 4846 scope.go:117] "RemoveContainer" containerID="8a13caff3cf7131d82ade0de5fe733ded9547a5758c374eab87ab478fe594e4d" Oct 05 09:36:10 crc kubenswrapper[4846]: E1005 09:36:10.498556 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 09:36:24 crc kubenswrapper[4846]: I1005 09:36:24.497918 4846 scope.go:117] "RemoveContainer" containerID="8a13caff3cf7131d82ade0de5fe733ded9547a5758c374eab87ab478fe594e4d" Oct 05 09:36:25 crc kubenswrapper[4846]: I1005 09:36:25.160066 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" event={"ID":"ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7","Type":"ContainerStarted","Data":"181e8721a32b2b46b8326bcf6aaa4c81d02a0273b05206fc94cdaa1d4b495256"} Oct 05 09:38:53 crc kubenswrapper[4846]: I1005 09:38:53.325173 4846 patch_prober.go:28] interesting pod/machine-config-daemon-fscvf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 09:38:53 crc kubenswrapper[4846]: I1005 09:38:53.325746 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 09:39:23 crc kubenswrapper[4846]: I1005 09:39:23.325450 4846 patch_prober.go:28] interesting pod/machine-config-daemon-fscvf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 09:39:23 crc kubenswrapper[4846]: I1005 09:39:23.326012 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 09:39:53 crc kubenswrapper[4846]: I1005 09:39:53.324732 4846 patch_prober.go:28] interesting pod/machine-config-daemon-fscvf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 09:39:53 crc kubenswrapper[4846]: I1005 09:39:53.325339 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 09:39:53 crc kubenswrapper[4846]: I1005 09:39:53.325391 4846 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" Oct 05 09:39:53 crc kubenswrapper[4846]: I1005 09:39:53.326222 4846 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"181e8721a32b2b46b8326bcf6aaa4c81d02a0273b05206fc94cdaa1d4b495256"} pod="openshift-machine-config-operator/machine-config-daemon-fscvf" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 05 09:39:53 crc kubenswrapper[4846]: I1005 09:39:53.326289 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" containerName="machine-config-daemon" containerID="cri-o://181e8721a32b2b46b8326bcf6aaa4c81d02a0273b05206fc94cdaa1d4b495256" gracePeriod=600 Oct 05 09:39:54 crc kubenswrapper[4846]: I1005 09:39:54.201701 4846 generic.go:334] "Generic (PLEG): container finished" podID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" containerID="181e8721a32b2b46b8326bcf6aaa4c81d02a0273b05206fc94cdaa1d4b495256" exitCode=0 Oct 05 09:39:54 crc kubenswrapper[4846]: I1005 09:39:54.202357 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" event={"ID":"ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7","Type":"ContainerDied","Data":"181e8721a32b2b46b8326bcf6aaa4c81d02a0273b05206fc94cdaa1d4b495256"} Oct 05 09:39:54 crc kubenswrapper[4846]: I1005 09:39:54.202392 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" event={"ID":"ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7","Type":"ContainerStarted","Data":"43e1f18d9b5aea38dd1527fbb4fc557f1b37c59dfd2bdc2019c291047f4244b8"} Oct 05 09:39:54 crc kubenswrapper[4846]: I1005 09:39:54.202416 4846 scope.go:117] "RemoveContainer" containerID="8a13caff3cf7131d82ade0de5fe733ded9547a5758c374eab87ab478fe594e4d" Oct 05 09:40:21 crc kubenswrapper[4846]: I1005 09:40:21.242507 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-nc7fm"] Oct 05 09:40:21 crc kubenswrapper[4846]: I1005 09:40:21.246805 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-nc7fm" Oct 05 09:40:21 crc kubenswrapper[4846]: I1005 09:40:21.275081 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-nc7fm"] Oct 05 09:40:21 crc kubenswrapper[4846]: I1005 09:40:21.290784 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fkfxd\" (UniqueName: \"kubernetes.io/projected/9a47fe94-4500-48ab-8c50-2924eea46d65-kube-api-access-fkfxd\") pod \"redhat-marketplace-nc7fm\" (UID: \"9a47fe94-4500-48ab-8c50-2924eea46d65\") " pod="openshift-marketplace/redhat-marketplace-nc7fm" Oct 05 09:40:21 crc kubenswrapper[4846]: I1005 09:40:21.290972 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9a47fe94-4500-48ab-8c50-2924eea46d65-catalog-content\") pod \"redhat-marketplace-nc7fm\" (UID: \"9a47fe94-4500-48ab-8c50-2924eea46d65\") " pod="openshift-marketplace/redhat-marketplace-nc7fm" Oct 05 09:40:21 crc kubenswrapper[4846]: I1005 09:40:21.291129 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9a47fe94-4500-48ab-8c50-2924eea46d65-utilities\") pod \"redhat-marketplace-nc7fm\" (UID: \"9a47fe94-4500-48ab-8c50-2924eea46d65\") " pod="openshift-marketplace/redhat-marketplace-nc7fm" Oct 05 09:40:21 crc kubenswrapper[4846]: I1005 09:40:21.394034 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fkfxd\" (UniqueName: \"kubernetes.io/projected/9a47fe94-4500-48ab-8c50-2924eea46d65-kube-api-access-fkfxd\") pod \"redhat-marketplace-nc7fm\" (UID: \"9a47fe94-4500-48ab-8c50-2924eea46d65\") " pod="openshift-marketplace/redhat-marketplace-nc7fm" Oct 05 09:40:21 crc kubenswrapper[4846]: I1005 09:40:21.394150 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9a47fe94-4500-48ab-8c50-2924eea46d65-catalog-content\") pod \"redhat-marketplace-nc7fm\" (UID: \"9a47fe94-4500-48ab-8c50-2924eea46d65\") " pod="openshift-marketplace/redhat-marketplace-nc7fm" Oct 05 09:40:21 crc kubenswrapper[4846]: I1005 09:40:21.394247 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9a47fe94-4500-48ab-8c50-2924eea46d65-utilities\") pod \"redhat-marketplace-nc7fm\" (UID: \"9a47fe94-4500-48ab-8c50-2924eea46d65\") " pod="openshift-marketplace/redhat-marketplace-nc7fm" Oct 05 09:40:21 crc kubenswrapper[4846]: I1005 09:40:21.394784 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9a47fe94-4500-48ab-8c50-2924eea46d65-utilities\") pod \"redhat-marketplace-nc7fm\" (UID: \"9a47fe94-4500-48ab-8c50-2924eea46d65\") " pod="openshift-marketplace/redhat-marketplace-nc7fm" Oct 05 09:40:21 crc kubenswrapper[4846]: I1005 09:40:21.395047 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9a47fe94-4500-48ab-8c50-2924eea46d65-catalog-content\") pod \"redhat-marketplace-nc7fm\" (UID: \"9a47fe94-4500-48ab-8c50-2924eea46d65\") " pod="openshift-marketplace/redhat-marketplace-nc7fm" Oct 05 09:40:21 crc kubenswrapper[4846]: I1005 09:40:21.419127 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fkfxd\" (UniqueName: \"kubernetes.io/projected/9a47fe94-4500-48ab-8c50-2924eea46d65-kube-api-access-fkfxd\") pod \"redhat-marketplace-nc7fm\" (UID: \"9a47fe94-4500-48ab-8c50-2924eea46d65\") " pod="openshift-marketplace/redhat-marketplace-nc7fm" Oct 05 09:40:21 crc kubenswrapper[4846]: I1005 09:40:21.611349 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-nc7fm" Oct 05 09:40:22 crc kubenswrapper[4846]: I1005 09:40:22.512388 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-nc7fm"] Oct 05 09:40:22 crc kubenswrapper[4846]: I1005 09:40:22.565739 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nc7fm" event={"ID":"9a47fe94-4500-48ab-8c50-2924eea46d65","Type":"ContainerStarted","Data":"55e48d48caa6648ad2a8b85fe334d13b97c19b9f40190ee2f37aff11f3eb6ea5"} Oct 05 09:40:23 crc kubenswrapper[4846]: I1005 09:40:23.576604 4846 generic.go:334] "Generic (PLEG): container finished" podID="9a47fe94-4500-48ab-8c50-2924eea46d65" containerID="85cebeed4e8c7ed07c23b2eeedcd60a7b678ea0c55da62c9f586c9ab4a03cca8" exitCode=0 Oct 05 09:40:23 crc kubenswrapper[4846]: I1005 09:40:23.576709 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nc7fm" event={"ID":"9a47fe94-4500-48ab-8c50-2924eea46d65","Type":"ContainerDied","Data":"85cebeed4e8c7ed07c23b2eeedcd60a7b678ea0c55da62c9f586c9ab4a03cca8"} Oct 05 09:40:24 crc kubenswrapper[4846]: I1005 09:40:24.588768 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nc7fm" event={"ID":"9a47fe94-4500-48ab-8c50-2924eea46d65","Type":"ContainerStarted","Data":"dfdd50cf82cbf3b969b8157a740c593060da08fb94b37b19f3f5b8472fdcda08"} Oct 05 09:40:25 crc kubenswrapper[4846]: I1005 09:40:25.605222 4846 generic.go:334] "Generic (PLEG): container finished" podID="9a47fe94-4500-48ab-8c50-2924eea46d65" containerID="dfdd50cf82cbf3b969b8157a740c593060da08fb94b37b19f3f5b8472fdcda08" exitCode=0 Oct 05 09:40:25 crc kubenswrapper[4846]: I1005 09:40:25.605688 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nc7fm" event={"ID":"9a47fe94-4500-48ab-8c50-2924eea46d65","Type":"ContainerDied","Data":"dfdd50cf82cbf3b969b8157a740c593060da08fb94b37b19f3f5b8472fdcda08"} Oct 05 09:40:26 crc kubenswrapper[4846]: I1005 09:40:26.617827 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nc7fm" event={"ID":"9a47fe94-4500-48ab-8c50-2924eea46d65","Type":"ContainerStarted","Data":"976deb6f29fa64d7bb6c475baa46525ddd75892876d85d4a9dd46caa95c4221f"} Oct 05 09:40:26 crc kubenswrapper[4846]: I1005 09:40:26.636618 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-nc7fm" podStartSLOduration=3.139486231 podStartE2EDuration="5.636601078s" podCreationTimestamp="2025-10-05 09:40:21 +0000 UTC" firstStartedPulling="2025-10-05 09:40:23.57890061 +0000 UTC m=+10345.819753385" lastFinishedPulling="2025-10-05 09:40:26.076015457 +0000 UTC m=+10348.316868232" observedRunningTime="2025-10-05 09:40:26.632570604 +0000 UTC m=+10348.873423379" watchObservedRunningTime="2025-10-05 09:40:26.636601078 +0000 UTC m=+10348.877453853" Oct 05 09:40:31 crc kubenswrapper[4846]: I1005 09:40:31.611680 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-nc7fm" Oct 05 09:40:31 crc kubenswrapper[4846]: I1005 09:40:31.613379 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-nc7fm" Oct 05 09:40:31 crc kubenswrapper[4846]: I1005 09:40:31.687902 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-nc7fm" Oct 05 09:40:31 crc kubenswrapper[4846]: I1005 09:40:31.763837 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-nc7fm" Oct 05 09:40:31 crc kubenswrapper[4846]: I1005 09:40:31.942367 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-nc7fm"] Oct 05 09:40:33 crc kubenswrapper[4846]: I1005 09:40:33.687466 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-nc7fm" podUID="9a47fe94-4500-48ab-8c50-2924eea46d65" containerName="registry-server" containerID="cri-o://976deb6f29fa64d7bb6c475baa46525ddd75892876d85d4a9dd46caa95c4221f" gracePeriod=2 Oct 05 09:40:34 crc kubenswrapper[4846]: I1005 09:40:34.703005 4846 generic.go:334] "Generic (PLEG): container finished" podID="9a47fe94-4500-48ab-8c50-2924eea46d65" containerID="976deb6f29fa64d7bb6c475baa46525ddd75892876d85d4a9dd46caa95c4221f" exitCode=0 Oct 05 09:40:34 crc kubenswrapper[4846]: I1005 09:40:34.703325 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nc7fm" event={"ID":"9a47fe94-4500-48ab-8c50-2924eea46d65","Type":"ContainerDied","Data":"976deb6f29fa64d7bb6c475baa46525ddd75892876d85d4a9dd46caa95c4221f"} Oct 05 09:40:34 crc kubenswrapper[4846]: I1005 09:40:34.703399 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nc7fm" event={"ID":"9a47fe94-4500-48ab-8c50-2924eea46d65","Type":"ContainerDied","Data":"55e48d48caa6648ad2a8b85fe334d13b97c19b9f40190ee2f37aff11f3eb6ea5"} Oct 05 09:40:34 crc kubenswrapper[4846]: I1005 09:40:34.703419 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="55e48d48caa6648ad2a8b85fe334d13b97c19b9f40190ee2f37aff11f3eb6ea5" Oct 05 09:40:34 crc kubenswrapper[4846]: I1005 09:40:34.713146 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-nc7fm" Oct 05 09:40:34 crc kubenswrapper[4846]: I1005 09:40:34.868352 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9a47fe94-4500-48ab-8c50-2924eea46d65-catalog-content\") pod \"9a47fe94-4500-48ab-8c50-2924eea46d65\" (UID: \"9a47fe94-4500-48ab-8c50-2924eea46d65\") " Oct 05 09:40:34 crc kubenswrapper[4846]: I1005 09:40:34.883892 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fkfxd\" (UniqueName: \"kubernetes.io/projected/9a47fe94-4500-48ab-8c50-2924eea46d65-kube-api-access-fkfxd\") pod \"9a47fe94-4500-48ab-8c50-2924eea46d65\" (UID: \"9a47fe94-4500-48ab-8c50-2924eea46d65\") " Oct 05 09:40:34 crc kubenswrapper[4846]: I1005 09:40:34.883973 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9a47fe94-4500-48ab-8c50-2924eea46d65-utilities\") pod \"9a47fe94-4500-48ab-8c50-2924eea46d65\" (UID: \"9a47fe94-4500-48ab-8c50-2924eea46d65\") " Oct 05 09:40:34 crc kubenswrapper[4846]: I1005 09:40:34.884344 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9a47fe94-4500-48ab-8c50-2924eea46d65-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9a47fe94-4500-48ab-8c50-2924eea46d65" (UID: "9a47fe94-4500-48ab-8c50-2924eea46d65"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 09:40:34 crc kubenswrapper[4846]: I1005 09:40:34.885788 4846 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9a47fe94-4500-48ab-8c50-2924eea46d65-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 05 09:40:34 crc kubenswrapper[4846]: I1005 09:40:34.886495 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9a47fe94-4500-48ab-8c50-2924eea46d65-utilities" (OuterVolumeSpecName: "utilities") pod "9a47fe94-4500-48ab-8c50-2924eea46d65" (UID: "9a47fe94-4500-48ab-8c50-2924eea46d65"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 09:40:34 crc kubenswrapper[4846]: I1005 09:40:34.897952 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9a47fe94-4500-48ab-8c50-2924eea46d65-kube-api-access-fkfxd" (OuterVolumeSpecName: "kube-api-access-fkfxd") pod "9a47fe94-4500-48ab-8c50-2924eea46d65" (UID: "9a47fe94-4500-48ab-8c50-2924eea46d65"). InnerVolumeSpecName "kube-api-access-fkfxd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 09:40:34 crc kubenswrapper[4846]: I1005 09:40:34.990145 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fkfxd\" (UniqueName: \"kubernetes.io/projected/9a47fe94-4500-48ab-8c50-2924eea46d65-kube-api-access-fkfxd\") on node \"crc\" DevicePath \"\"" Oct 05 09:40:34 crc kubenswrapper[4846]: I1005 09:40:34.990203 4846 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9a47fe94-4500-48ab-8c50-2924eea46d65-utilities\") on node \"crc\" DevicePath \"\"" Oct 05 09:40:35 crc kubenswrapper[4846]: I1005 09:40:35.710946 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-nc7fm" Oct 05 09:40:35 crc kubenswrapper[4846]: I1005 09:40:35.759624 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-nc7fm"] Oct 05 09:40:35 crc kubenswrapper[4846]: I1005 09:40:35.774804 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-nc7fm"] Oct 05 09:40:36 crc kubenswrapper[4846]: I1005 09:40:36.508932 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9a47fe94-4500-48ab-8c50-2924eea46d65" path="/var/lib/kubelet/pods/9a47fe94-4500-48ab-8c50-2924eea46d65/volumes" Oct 05 09:40:41 crc kubenswrapper[4846]: I1005 09:40:41.235292 4846 scope.go:117] "RemoveContainer" containerID="8561884803d5e7b193f12a7f90f1a582ab155c26f991b1a20c8a25a1b4f802d3" Oct 05 09:40:41 crc kubenswrapper[4846]: I1005 09:40:41.261712 4846 scope.go:117] "RemoveContainer" containerID="6b8742ac90599080bcc978c0108a6ef8d60ba7af22f472082648c02fa065b481" Oct 05 09:40:41 crc kubenswrapper[4846]: I1005 09:40:41.304624 4846 scope.go:117] "RemoveContainer" containerID="449e5ed250cddb7ff09106a1d2655b161c6375e4e368d8832333f36d7996cd8f" Oct 05 09:41:15 crc kubenswrapper[4846]: I1005 09:41:15.074467 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-6tdvq"] Oct 05 09:41:15 crc kubenswrapper[4846]: E1005 09:41:15.078898 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a47fe94-4500-48ab-8c50-2924eea46d65" containerName="extract-content" Oct 05 09:41:15 crc kubenswrapper[4846]: I1005 09:41:15.078928 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a47fe94-4500-48ab-8c50-2924eea46d65" containerName="extract-content" Oct 05 09:41:15 crc kubenswrapper[4846]: E1005 09:41:15.078961 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a47fe94-4500-48ab-8c50-2924eea46d65" containerName="extract-utilities" Oct 05 09:41:15 crc kubenswrapper[4846]: I1005 09:41:15.078969 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a47fe94-4500-48ab-8c50-2924eea46d65" containerName="extract-utilities" Oct 05 09:41:15 crc kubenswrapper[4846]: E1005 09:41:15.078979 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a47fe94-4500-48ab-8c50-2924eea46d65" containerName="registry-server" Oct 05 09:41:15 crc kubenswrapper[4846]: I1005 09:41:15.078987 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a47fe94-4500-48ab-8c50-2924eea46d65" containerName="registry-server" Oct 05 09:41:15 crc kubenswrapper[4846]: I1005 09:41:15.079199 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="9a47fe94-4500-48ab-8c50-2924eea46d65" containerName="registry-server" Oct 05 09:41:15 crc kubenswrapper[4846]: I1005 09:41:15.080617 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6tdvq" Oct 05 09:41:15 crc kubenswrapper[4846]: I1005 09:41:15.121141 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-6tdvq"] Oct 05 09:41:15 crc kubenswrapper[4846]: I1005 09:41:15.137845 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/89320521-1f3e-4435-86fc-ef6c18706cbf-utilities\") pod \"community-operators-6tdvq\" (UID: \"89320521-1f3e-4435-86fc-ef6c18706cbf\") " pod="openshift-marketplace/community-operators-6tdvq" Oct 05 09:41:15 crc kubenswrapper[4846]: I1005 09:41:15.138080 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/89320521-1f3e-4435-86fc-ef6c18706cbf-catalog-content\") pod \"community-operators-6tdvq\" (UID: \"89320521-1f3e-4435-86fc-ef6c18706cbf\") " pod="openshift-marketplace/community-operators-6tdvq" Oct 05 09:41:15 crc kubenswrapper[4846]: I1005 09:41:15.138454 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dj7pg\" (UniqueName: \"kubernetes.io/projected/89320521-1f3e-4435-86fc-ef6c18706cbf-kube-api-access-dj7pg\") pod \"community-operators-6tdvq\" (UID: \"89320521-1f3e-4435-86fc-ef6c18706cbf\") " pod="openshift-marketplace/community-operators-6tdvq" Oct 05 09:41:15 crc kubenswrapper[4846]: I1005 09:41:15.240298 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/89320521-1f3e-4435-86fc-ef6c18706cbf-catalog-content\") pod \"community-operators-6tdvq\" (UID: \"89320521-1f3e-4435-86fc-ef6c18706cbf\") " pod="openshift-marketplace/community-operators-6tdvq" Oct 05 09:41:15 crc kubenswrapper[4846]: I1005 09:41:15.240478 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dj7pg\" (UniqueName: \"kubernetes.io/projected/89320521-1f3e-4435-86fc-ef6c18706cbf-kube-api-access-dj7pg\") pod \"community-operators-6tdvq\" (UID: \"89320521-1f3e-4435-86fc-ef6c18706cbf\") " pod="openshift-marketplace/community-operators-6tdvq" Oct 05 09:41:15 crc kubenswrapper[4846]: I1005 09:41:15.240520 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/89320521-1f3e-4435-86fc-ef6c18706cbf-utilities\") pod \"community-operators-6tdvq\" (UID: \"89320521-1f3e-4435-86fc-ef6c18706cbf\") " pod="openshift-marketplace/community-operators-6tdvq" Oct 05 09:41:15 crc kubenswrapper[4846]: I1005 09:41:15.240940 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/89320521-1f3e-4435-86fc-ef6c18706cbf-catalog-content\") pod \"community-operators-6tdvq\" (UID: \"89320521-1f3e-4435-86fc-ef6c18706cbf\") " pod="openshift-marketplace/community-operators-6tdvq" Oct 05 09:41:15 crc kubenswrapper[4846]: I1005 09:41:15.241103 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/89320521-1f3e-4435-86fc-ef6c18706cbf-utilities\") pod \"community-operators-6tdvq\" (UID: \"89320521-1f3e-4435-86fc-ef6c18706cbf\") " pod="openshift-marketplace/community-operators-6tdvq" Oct 05 09:41:15 crc kubenswrapper[4846]: I1005 09:41:15.897339 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dj7pg\" (UniqueName: \"kubernetes.io/projected/89320521-1f3e-4435-86fc-ef6c18706cbf-kube-api-access-dj7pg\") pod \"community-operators-6tdvq\" (UID: \"89320521-1f3e-4435-86fc-ef6c18706cbf\") " pod="openshift-marketplace/community-operators-6tdvq" Oct 05 09:41:16 crc kubenswrapper[4846]: I1005 09:41:16.027961 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6tdvq" Oct 05 09:41:16 crc kubenswrapper[4846]: I1005 09:41:16.537149 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-6tdvq"] Oct 05 09:41:17 crc kubenswrapper[4846]: I1005 09:41:17.096293 4846 generic.go:334] "Generic (PLEG): container finished" podID="89320521-1f3e-4435-86fc-ef6c18706cbf" containerID="b798cf588db95f4dd95e888e163cc83413be2b31c2eb1219f1d9de4eb31264b7" exitCode=0 Oct 05 09:41:17 crc kubenswrapper[4846]: I1005 09:41:17.096400 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6tdvq" event={"ID":"89320521-1f3e-4435-86fc-ef6c18706cbf","Type":"ContainerDied","Data":"b798cf588db95f4dd95e888e163cc83413be2b31c2eb1219f1d9de4eb31264b7"} Oct 05 09:41:17 crc kubenswrapper[4846]: I1005 09:41:17.097214 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6tdvq" event={"ID":"89320521-1f3e-4435-86fc-ef6c18706cbf","Type":"ContainerStarted","Data":"cd164b5682ca4605fc5e95cfbfe639e35ac4399c7ffaadc9ba2915f4961fd8a4"} Oct 05 09:41:17 crc kubenswrapper[4846]: I1005 09:41:17.099010 4846 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 05 09:41:19 crc kubenswrapper[4846]: I1005 09:41:19.135428 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6tdvq" event={"ID":"89320521-1f3e-4435-86fc-ef6c18706cbf","Type":"ContainerStarted","Data":"17da51ad0c2eb3eb26581614d08843b3fea49aa8b5d50fb645340a3a67cb400d"} Oct 05 09:41:20 crc kubenswrapper[4846]: I1005 09:41:20.149459 4846 generic.go:334] "Generic (PLEG): container finished" podID="89320521-1f3e-4435-86fc-ef6c18706cbf" containerID="17da51ad0c2eb3eb26581614d08843b3fea49aa8b5d50fb645340a3a67cb400d" exitCode=0 Oct 05 09:41:20 crc kubenswrapper[4846]: I1005 09:41:20.149564 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6tdvq" event={"ID":"89320521-1f3e-4435-86fc-ef6c18706cbf","Type":"ContainerDied","Data":"17da51ad0c2eb3eb26581614d08843b3fea49aa8b5d50fb645340a3a67cb400d"} Oct 05 09:41:21 crc kubenswrapper[4846]: I1005 09:41:21.160858 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6tdvq" event={"ID":"89320521-1f3e-4435-86fc-ef6c18706cbf","Type":"ContainerStarted","Data":"c24f8af726b345493631aff98222ac904bf578a3042c31a58782116e06ec1118"} Oct 05 09:41:21 crc kubenswrapper[4846]: I1005 09:41:21.183719 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-6tdvq" podStartSLOduration=2.736587611 podStartE2EDuration="6.183702913s" podCreationTimestamp="2025-10-05 09:41:15 +0000 UTC" firstStartedPulling="2025-10-05 09:41:17.098674675 +0000 UTC m=+10399.339527450" lastFinishedPulling="2025-10-05 09:41:20.545789967 +0000 UTC m=+10402.786642752" observedRunningTime="2025-10-05 09:41:21.178876455 +0000 UTC m=+10403.419729230" watchObservedRunningTime="2025-10-05 09:41:21.183702913 +0000 UTC m=+10403.424555688" Oct 05 09:41:26 crc kubenswrapper[4846]: I1005 09:41:26.028289 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-6tdvq" Oct 05 09:41:26 crc kubenswrapper[4846]: I1005 09:41:26.029705 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-6tdvq" Oct 05 09:41:26 crc kubenswrapper[4846]: I1005 09:41:26.077708 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-6tdvq" Oct 05 09:41:26 crc kubenswrapper[4846]: I1005 09:41:26.281307 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-6tdvq" Oct 05 09:41:26 crc kubenswrapper[4846]: I1005 09:41:26.331886 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-6tdvq"] Oct 05 09:41:28 crc kubenswrapper[4846]: I1005 09:41:28.246837 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-6tdvq" podUID="89320521-1f3e-4435-86fc-ef6c18706cbf" containerName="registry-server" containerID="cri-o://c24f8af726b345493631aff98222ac904bf578a3042c31a58782116e06ec1118" gracePeriod=2 Oct 05 09:41:29 crc kubenswrapper[4846]: I1005 09:41:29.234119 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6tdvq" Oct 05 09:41:29 crc kubenswrapper[4846]: I1005 09:41:29.264171 4846 generic.go:334] "Generic (PLEG): container finished" podID="89320521-1f3e-4435-86fc-ef6c18706cbf" containerID="c24f8af726b345493631aff98222ac904bf578a3042c31a58782116e06ec1118" exitCode=0 Oct 05 09:41:29 crc kubenswrapper[4846]: I1005 09:41:29.264224 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6tdvq" event={"ID":"89320521-1f3e-4435-86fc-ef6c18706cbf","Type":"ContainerDied","Data":"c24f8af726b345493631aff98222ac904bf578a3042c31a58782116e06ec1118"} Oct 05 09:41:29 crc kubenswrapper[4846]: I1005 09:41:29.264294 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6tdvq" event={"ID":"89320521-1f3e-4435-86fc-ef6c18706cbf","Type":"ContainerDied","Data":"cd164b5682ca4605fc5e95cfbfe639e35ac4399c7ffaadc9ba2915f4961fd8a4"} Oct 05 09:41:29 crc kubenswrapper[4846]: I1005 09:41:29.264313 4846 scope.go:117] "RemoveContainer" containerID="c24f8af726b345493631aff98222ac904bf578a3042c31a58782116e06ec1118" Oct 05 09:41:29 crc kubenswrapper[4846]: I1005 09:41:29.264246 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6tdvq" Oct 05 09:41:29 crc kubenswrapper[4846]: I1005 09:41:29.318382 4846 scope.go:117] "RemoveContainer" containerID="17da51ad0c2eb3eb26581614d08843b3fea49aa8b5d50fb645340a3a67cb400d" Oct 05 09:41:29 crc kubenswrapper[4846]: I1005 09:41:29.352469 4846 scope.go:117] "RemoveContainer" containerID="b798cf588db95f4dd95e888e163cc83413be2b31c2eb1219f1d9de4eb31264b7" Oct 05 09:41:29 crc kubenswrapper[4846]: I1005 09:41:29.402444 4846 scope.go:117] "RemoveContainer" containerID="c24f8af726b345493631aff98222ac904bf578a3042c31a58782116e06ec1118" Oct 05 09:41:29 crc kubenswrapper[4846]: E1005 09:41:29.402918 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c24f8af726b345493631aff98222ac904bf578a3042c31a58782116e06ec1118\": container with ID starting with c24f8af726b345493631aff98222ac904bf578a3042c31a58782116e06ec1118 not found: ID does not exist" containerID="c24f8af726b345493631aff98222ac904bf578a3042c31a58782116e06ec1118" Oct 05 09:41:29 crc kubenswrapper[4846]: I1005 09:41:29.402953 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c24f8af726b345493631aff98222ac904bf578a3042c31a58782116e06ec1118"} err="failed to get container status \"c24f8af726b345493631aff98222ac904bf578a3042c31a58782116e06ec1118\": rpc error: code = NotFound desc = could not find container \"c24f8af726b345493631aff98222ac904bf578a3042c31a58782116e06ec1118\": container with ID starting with c24f8af726b345493631aff98222ac904bf578a3042c31a58782116e06ec1118 not found: ID does not exist" Oct 05 09:41:29 crc kubenswrapper[4846]: I1005 09:41:29.402972 4846 scope.go:117] "RemoveContainer" containerID="17da51ad0c2eb3eb26581614d08843b3fea49aa8b5d50fb645340a3a67cb400d" Oct 05 09:41:29 crc kubenswrapper[4846]: E1005 09:41:29.403262 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"17da51ad0c2eb3eb26581614d08843b3fea49aa8b5d50fb645340a3a67cb400d\": container with ID starting with 17da51ad0c2eb3eb26581614d08843b3fea49aa8b5d50fb645340a3a67cb400d not found: ID does not exist" containerID="17da51ad0c2eb3eb26581614d08843b3fea49aa8b5d50fb645340a3a67cb400d" Oct 05 09:41:29 crc kubenswrapper[4846]: I1005 09:41:29.403291 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"17da51ad0c2eb3eb26581614d08843b3fea49aa8b5d50fb645340a3a67cb400d"} err="failed to get container status \"17da51ad0c2eb3eb26581614d08843b3fea49aa8b5d50fb645340a3a67cb400d\": rpc error: code = NotFound desc = could not find container \"17da51ad0c2eb3eb26581614d08843b3fea49aa8b5d50fb645340a3a67cb400d\": container with ID starting with 17da51ad0c2eb3eb26581614d08843b3fea49aa8b5d50fb645340a3a67cb400d not found: ID does not exist" Oct 05 09:41:29 crc kubenswrapper[4846]: I1005 09:41:29.403308 4846 scope.go:117] "RemoveContainer" containerID="b798cf588db95f4dd95e888e163cc83413be2b31c2eb1219f1d9de4eb31264b7" Oct 05 09:41:29 crc kubenswrapper[4846]: E1005 09:41:29.404292 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b798cf588db95f4dd95e888e163cc83413be2b31c2eb1219f1d9de4eb31264b7\": container with ID starting with b798cf588db95f4dd95e888e163cc83413be2b31c2eb1219f1d9de4eb31264b7 not found: ID does not exist" containerID="b798cf588db95f4dd95e888e163cc83413be2b31c2eb1219f1d9de4eb31264b7" Oct 05 09:41:29 crc kubenswrapper[4846]: I1005 09:41:29.404336 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b798cf588db95f4dd95e888e163cc83413be2b31c2eb1219f1d9de4eb31264b7"} err="failed to get container status \"b798cf588db95f4dd95e888e163cc83413be2b31c2eb1219f1d9de4eb31264b7\": rpc error: code = NotFound desc = could not find container \"b798cf588db95f4dd95e888e163cc83413be2b31c2eb1219f1d9de4eb31264b7\": container with ID starting with b798cf588db95f4dd95e888e163cc83413be2b31c2eb1219f1d9de4eb31264b7 not found: ID does not exist" Oct 05 09:41:29 crc kubenswrapper[4846]: I1005 09:41:29.424386 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/89320521-1f3e-4435-86fc-ef6c18706cbf-catalog-content\") pod \"89320521-1f3e-4435-86fc-ef6c18706cbf\" (UID: \"89320521-1f3e-4435-86fc-ef6c18706cbf\") " Oct 05 09:41:29 crc kubenswrapper[4846]: I1005 09:41:29.424644 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dj7pg\" (UniqueName: \"kubernetes.io/projected/89320521-1f3e-4435-86fc-ef6c18706cbf-kube-api-access-dj7pg\") pod \"89320521-1f3e-4435-86fc-ef6c18706cbf\" (UID: \"89320521-1f3e-4435-86fc-ef6c18706cbf\") " Oct 05 09:41:29 crc kubenswrapper[4846]: I1005 09:41:29.424739 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/89320521-1f3e-4435-86fc-ef6c18706cbf-utilities\") pod \"89320521-1f3e-4435-86fc-ef6c18706cbf\" (UID: \"89320521-1f3e-4435-86fc-ef6c18706cbf\") " Oct 05 09:41:29 crc kubenswrapper[4846]: I1005 09:41:29.427805 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/89320521-1f3e-4435-86fc-ef6c18706cbf-utilities" (OuterVolumeSpecName: "utilities") pod "89320521-1f3e-4435-86fc-ef6c18706cbf" (UID: "89320521-1f3e-4435-86fc-ef6c18706cbf"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 09:41:29 crc kubenswrapper[4846]: I1005 09:41:29.432466 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/89320521-1f3e-4435-86fc-ef6c18706cbf-kube-api-access-dj7pg" (OuterVolumeSpecName: "kube-api-access-dj7pg") pod "89320521-1f3e-4435-86fc-ef6c18706cbf" (UID: "89320521-1f3e-4435-86fc-ef6c18706cbf"). InnerVolumeSpecName "kube-api-access-dj7pg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 09:41:29 crc kubenswrapper[4846]: I1005 09:41:29.480245 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/89320521-1f3e-4435-86fc-ef6c18706cbf-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "89320521-1f3e-4435-86fc-ef6c18706cbf" (UID: "89320521-1f3e-4435-86fc-ef6c18706cbf"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 09:41:29 crc kubenswrapper[4846]: I1005 09:41:29.527720 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dj7pg\" (UniqueName: \"kubernetes.io/projected/89320521-1f3e-4435-86fc-ef6c18706cbf-kube-api-access-dj7pg\") on node \"crc\" DevicePath \"\"" Oct 05 09:41:29 crc kubenswrapper[4846]: I1005 09:41:29.527771 4846 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/89320521-1f3e-4435-86fc-ef6c18706cbf-utilities\") on node \"crc\" DevicePath \"\"" Oct 05 09:41:29 crc kubenswrapper[4846]: I1005 09:41:29.527816 4846 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/89320521-1f3e-4435-86fc-ef6c18706cbf-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 05 09:41:29 crc kubenswrapper[4846]: I1005 09:41:29.610345 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-6tdvq"] Oct 05 09:41:29 crc kubenswrapper[4846]: I1005 09:41:29.621781 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-6tdvq"] Oct 05 09:41:30 crc kubenswrapper[4846]: I1005 09:41:30.511752 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="89320521-1f3e-4435-86fc-ef6c18706cbf" path="/var/lib/kubelet/pods/89320521-1f3e-4435-86fc-ef6c18706cbf/volumes" Oct 05 09:41:53 crc kubenswrapper[4846]: I1005 09:41:53.324440 4846 patch_prober.go:28] interesting pod/machine-config-daemon-fscvf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 09:41:53 crc kubenswrapper[4846]: I1005 09:41:53.324864 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 09:42:23 crc kubenswrapper[4846]: I1005 09:42:23.324520 4846 patch_prober.go:28] interesting pod/machine-config-daemon-fscvf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 09:42:23 crc kubenswrapper[4846]: I1005 09:42:23.326243 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 09:42:53 crc kubenswrapper[4846]: I1005 09:42:53.325130 4846 patch_prober.go:28] interesting pod/machine-config-daemon-fscvf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 09:42:53 crc kubenswrapper[4846]: I1005 09:42:53.325758 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 09:42:53 crc kubenswrapper[4846]: I1005 09:42:53.325812 4846 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" Oct 05 09:42:53 crc kubenswrapper[4846]: I1005 09:42:53.326801 4846 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"43e1f18d9b5aea38dd1527fbb4fc557f1b37c59dfd2bdc2019c291047f4244b8"} pod="openshift-machine-config-operator/machine-config-daemon-fscvf" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 05 09:42:53 crc kubenswrapper[4846]: I1005 09:42:53.326948 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" containerName="machine-config-daemon" containerID="cri-o://43e1f18d9b5aea38dd1527fbb4fc557f1b37c59dfd2bdc2019c291047f4244b8" gracePeriod=600 Oct 05 09:42:53 crc kubenswrapper[4846]: E1005 09:42:53.454425 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 09:42:54 crc kubenswrapper[4846]: I1005 09:42:54.146237 4846 generic.go:334] "Generic (PLEG): container finished" podID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" containerID="43e1f18d9b5aea38dd1527fbb4fc557f1b37c59dfd2bdc2019c291047f4244b8" exitCode=0 Oct 05 09:42:54 crc kubenswrapper[4846]: I1005 09:42:54.146313 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" event={"ID":"ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7","Type":"ContainerDied","Data":"43e1f18d9b5aea38dd1527fbb4fc557f1b37c59dfd2bdc2019c291047f4244b8"} Oct 05 09:42:54 crc kubenswrapper[4846]: I1005 09:42:54.146621 4846 scope.go:117] "RemoveContainer" containerID="181e8721a32b2b46b8326bcf6aaa4c81d02a0273b05206fc94cdaa1d4b495256" Oct 05 09:42:54 crc kubenswrapper[4846]: I1005 09:42:54.147357 4846 scope.go:117] "RemoveContainer" containerID="43e1f18d9b5aea38dd1527fbb4fc557f1b37c59dfd2bdc2019c291047f4244b8" Oct 05 09:42:54 crc kubenswrapper[4846]: E1005 09:42:54.147697 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 09:43:05 crc kubenswrapper[4846]: I1005 09:43:05.497699 4846 scope.go:117] "RemoveContainer" containerID="43e1f18d9b5aea38dd1527fbb4fc557f1b37c59dfd2bdc2019c291047f4244b8" Oct 05 09:43:05 crc kubenswrapper[4846]: E1005 09:43:05.499222 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 09:43:17 crc kubenswrapper[4846]: I1005 09:43:17.498685 4846 scope.go:117] "RemoveContainer" containerID="43e1f18d9b5aea38dd1527fbb4fc557f1b37c59dfd2bdc2019c291047f4244b8" Oct 05 09:43:17 crc kubenswrapper[4846]: E1005 09:43:17.499414 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 09:43:32 crc kubenswrapper[4846]: I1005 09:43:32.498340 4846 scope.go:117] "RemoveContainer" containerID="43e1f18d9b5aea38dd1527fbb4fc557f1b37c59dfd2bdc2019c291047f4244b8" Oct 05 09:43:32 crc kubenswrapper[4846]: E1005 09:43:32.499157 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 09:43:46 crc kubenswrapper[4846]: I1005 09:43:46.498313 4846 scope.go:117] "RemoveContainer" containerID="43e1f18d9b5aea38dd1527fbb4fc557f1b37c59dfd2bdc2019c291047f4244b8" Oct 05 09:43:46 crc kubenswrapper[4846]: E1005 09:43:46.499374 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 09:44:00 crc kubenswrapper[4846]: I1005 09:44:00.499376 4846 scope.go:117] "RemoveContainer" containerID="43e1f18d9b5aea38dd1527fbb4fc557f1b37c59dfd2bdc2019c291047f4244b8" Oct 05 09:44:00 crc kubenswrapper[4846]: E1005 09:44:00.500507 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 09:44:08 crc kubenswrapper[4846]: I1005 09:44:08.803114 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-z4vbf"] Oct 05 09:44:08 crc kubenswrapper[4846]: E1005 09:44:08.804175 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="89320521-1f3e-4435-86fc-ef6c18706cbf" containerName="registry-server" Oct 05 09:44:08 crc kubenswrapper[4846]: I1005 09:44:08.804211 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="89320521-1f3e-4435-86fc-ef6c18706cbf" containerName="registry-server" Oct 05 09:44:08 crc kubenswrapper[4846]: E1005 09:44:08.804241 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="89320521-1f3e-4435-86fc-ef6c18706cbf" containerName="extract-utilities" Oct 05 09:44:08 crc kubenswrapper[4846]: I1005 09:44:08.804253 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="89320521-1f3e-4435-86fc-ef6c18706cbf" containerName="extract-utilities" Oct 05 09:44:08 crc kubenswrapper[4846]: E1005 09:44:08.804297 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="89320521-1f3e-4435-86fc-ef6c18706cbf" containerName="extract-content" Oct 05 09:44:08 crc kubenswrapper[4846]: I1005 09:44:08.804308 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="89320521-1f3e-4435-86fc-ef6c18706cbf" containerName="extract-content" Oct 05 09:44:08 crc kubenswrapper[4846]: I1005 09:44:08.804640 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="89320521-1f3e-4435-86fc-ef6c18706cbf" containerName="registry-server" Oct 05 09:44:08 crc kubenswrapper[4846]: I1005 09:44:08.807051 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-z4vbf" Oct 05 09:44:08 crc kubenswrapper[4846]: I1005 09:44:08.820536 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-z4vbf"] Oct 05 09:44:08 crc kubenswrapper[4846]: I1005 09:44:08.921905 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/618e874d-1fbf-4ea0-a065-774350222fd0-utilities\") pod \"certified-operators-z4vbf\" (UID: \"618e874d-1fbf-4ea0-a065-774350222fd0\") " pod="openshift-marketplace/certified-operators-z4vbf" Oct 05 09:44:08 crc kubenswrapper[4846]: I1005 09:44:08.921986 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/618e874d-1fbf-4ea0-a065-774350222fd0-catalog-content\") pod \"certified-operators-z4vbf\" (UID: \"618e874d-1fbf-4ea0-a065-774350222fd0\") " pod="openshift-marketplace/certified-operators-z4vbf" Oct 05 09:44:08 crc kubenswrapper[4846]: I1005 09:44:08.922032 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lxftf\" (UniqueName: \"kubernetes.io/projected/618e874d-1fbf-4ea0-a065-774350222fd0-kube-api-access-lxftf\") pod \"certified-operators-z4vbf\" (UID: \"618e874d-1fbf-4ea0-a065-774350222fd0\") " pod="openshift-marketplace/certified-operators-z4vbf" Oct 05 09:44:09 crc kubenswrapper[4846]: I1005 09:44:09.024239 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/618e874d-1fbf-4ea0-a065-774350222fd0-utilities\") pod \"certified-operators-z4vbf\" (UID: \"618e874d-1fbf-4ea0-a065-774350222fd0\") " pod="openshift-marketplace/certified-operators-z4vbf" Oct 05 09:44:09 crc kubenswrapper[4846]: I1005 09:44:09.024716 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/618e874d-1fbf-4ea0-a065-774350222fd0-catalog-content\") pod \"certified-operators-z4vbf\" (UID: \"618e874d-1fbf-4ea0-a065-774350222fd0\") " pod="openshift-marketplace/certified-operators-z4vbf" Oct 05 09:44:09 crc kubenswrapper[4846]: I1005 09:44:09.024793 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lxftf\" (UniqueName: \"kubernetes.io/projected/618e874d-1fbf-4ea0-a065-774350222fd0-kube-api-access-lxftf\") pod \"certified-operators-z4vbf\" (UID: \"618e874d-1fbf-4ea0-a065-774350222fd0\") " pod="openshift-marketplace/certified-operators-z4vbf" Oct 05 09:44:09 crc kubenswrapper[4846]: I1005 09:44:09.024712 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/618e874d-1fbf-4ea0-a065-774350222fd0-utilities\") pod \"certified-operators-z4vbf\" (UID: \"618e874d-1fbf-4ea0-a065-774350222fd0\") " pod="openshift-marketplace/certified-operators-z4vbf" Oct 05 09:44:09 crc kubenswrapper[4846]: I1005 09:44:09.024919 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/618e874d-1fbf-4ea0-a065-774350222fd0-catalog-content\") pod \"certified-operators-z4vbf\" (UID: \"618e874d-1fbf-4ea0-a065-774350222fd0\") " pod="openshift-marketplace/certified-operators-z4vbf" Oct 05 09:44:09 crc kubenswrapper[4846]: I1005 09:44:09.043650 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lxftf\" (UniqueName: \"kubernetes.io/projected/618e874d-1fbf-4ea0-a065-774350222fd0-kube-api-access-lxftf\") pod \"certified-operators-z4vbf\" (UID: \"618e874d-1fbf-4ea0-a065-774350222fd0\") " pod="openshift-marketplace/certified-operators-z4vbf" Oct 05 09:44:09 crc kubenswrapper[4846]: I1005 09:44:09.131574 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-z4vbf" Oct 05 09:44:09 crc kubenswrapper[4846]: W1005 09:44:09.770979 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod618e874d_1fbf_4ea0_a065_774350222fd0.slice/crio-3d78e7a90c012222c0ae6b45b58e6a93adb0584806cadaa06158fb9addd84ba5 WatchSource:0}: Error finding container 3d78e7a90c012222c0ae6b45b58e6a93adb0584806cadaa06158fb9addd84ba5: Status 404 returned error can't find the container with id 3d78e7a90c012222c0ae6b45b58e6a93adb0584806cadaa06158fb9addd84ba5 Oct 05 09:44:09 crc kubenswrapper[4846]: I1005 09:44:09.772609 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-z4vbf"] Oct 05 09:44:10 crc kubenswrapper[4846]: I1005 09:44:10.017898 4846 generic.go:334] "Generic (PLEG): container finished" podID="618e874d-1fbf-4ea0-a065-774350222fd0" containerID="e9ec438ed3f0ba233a71049cd162dd34ae952b2b096d3a3762f0cd9da0b1db32" exitCode=0 Oct 05 09:44:10 crc kubenswrapper[4846]: I1005 09:44:10.017992 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-z4vbf" event={"ID":"618e874d-1fbf-4ea0-a065-774350222fd0","Type":"ContainerDied","Data":"e9ec438ed3f0ba233a71049cd162dd34ae952b2b096d3a3762f0cd9da0b1db32"} Oct 05 09:44:10 crc kubenswrapper[4846]: I1005 09:44:10.018379 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-z4vbf" event={"ID":"618e874d-1fbf-4ea0-a065-774350222fd0","Type":"ContainerStarted","Data":"3d78e7a90c012222c0ae6b45b58e6a93adb0584806cadaa06158fb9addd84ba5"} Oct 05 09:44:11 crc kubenswrapper[4846]: I1005 09:44:11.029785 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-z4vbf" event={"ID":"618e874d-1fbf-4ea0-a065-774350222fd0","Type":"ContainerStarted","Data":"a68defebaddf47f281d75d7601daea8b8f658431106b24bbf2ee25b1cf4741dd"} Oct 05 09:44:12 crc kubenswrapper[4846]: I1005 09:44:12.039755 4846 generic.go:334] "Generic (PLEG): container finished" podID="618e874d-1fbf-4ea0-a065-774350222fd0" containerID="a68defebaddf47f281d75d7601daea8b8f658431106b24bbf2ee25b1cf4741dd" exitCode=0 Oct 05 09:44:12 crc kubenswrapper[4846]: I1005 09:44:12.039825 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-z4vbf" event={"ID":"618e874d-1fbf-4ea0-a065-774350222fd0","Type":"ContainerDied","Data":"a68defebaddf47f281d75d7601daea8b8f658431106b24bbf2ee25b1cf4741dd"} Oct 05 09:44:12 crc kubenswrapper[4846]: I1005 09:44:12.609292 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-nszwd"] Oct 05 09:44:12 crc kubenswrapper[4846]: I1005 09:44:12.611587 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-nszwd" Oct 05 09:44:12 crc kubenswrapper[4846]: I1005 09:44:12.623762 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-nszwd"] Oct 05 09:44:12 crc kubenswrapper[4846]: I1005 09:44:12.707923 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/483dc857-1f8b-4d00-bcfb-11e07773600e-utilities\") pod \"redhat-operators-nszwd\" (UID: \"483dc857-1f8b-4d00-bcfb-11e07773600e\") " pod="openshift-marketplace/redhat-operators-nszwd" Oct 05 09:44:12 crc kubenswrapper[4846]: I1005 09:44:12.707985 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/483dc857-1f8b-4d00-bcfb-11e07773600e-catalog-content\") pod \"redhat-operators-nszwd\" (UID: \"483dc857-1f8b-4d00-bcfb-11e07773600e\") " pod="openshift-marketplace/redhat-operators-nszwd" Oct 05 09:44:12 crc kubenswrapper[4846]: I1005 09:44:12.708029 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d9zsd\" (UniqueName: \"kubernetes.io/projected/483dc857-1f8b-4d00-bcfb-11e07773600e-kube-api-access-d9zsd\") pod \"redhat-operators-nszwd\" (UID: \"483dc857-1f8b-4d00-bcfb-11e07773600e\") " pod="openshift-marketplace/redhat-operators-nszwd" Oct 05 09:44:12 crc kubenswrapper[4846]: I1005 09:44:12.810261 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/483dc857-1f8b-4d00-bcfb-11e07773600e-utilities\") pod \"redhat-operators-nszwd\" (UID: \"483dc857-1f8b-4d00-bcfb-11e07773600e\") " pod="openshift-marketplace/redhat-operators-nszwd" Oct 05 09:44:12 crc kubenswrapper[4846]: I1005 09:44:12.810482 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/483dc857-1f8b-4d00-bcfb-11e07773600e-catalog-content\") pod \"redhat-operators-nszwd\" (UID: \"483dc857-1f8b-4d00-bcfb-11e07773600e\") " pod="openshift-marketplace/redhat-operators-nszwd" Oct 05 09:44:12 crc kubenswrapper[4846]: I1005 09:44:12.810629 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d9zsd\" (UniqueName: \"kubernetes.io/projected/483dc857-1f8b-4d00-bcfb-11e07773600e-kube-api-access-d9zsd\") pod \"redhat-operators-nszwd\" (UID: \"483dc857-1f8b-4d00-bcfb-11e07773600e\") " pod="openshift-marketplace/redhat-operators-nszwd" Oct 05 09:44:12 crc kubenswrapper[4846]: I1005 09:44:12.811574 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/483dc857-1f8b-4d00-bcfb-11e07773600e-utilities\") pod \"redhat-operators-nszwd\" (UID: \"483dc857-1f8b-4d00-bcfb-11e07773600e\") " pod="openshift-marketplace/redhat-operators-nszwd" Oct 05 09:44:12 crc kubenswrapper[4846]: I1005 09:44:12.811935 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/483dc857-1f8b-4d00-bcfb-11e07773600e-catalog-content\") pod \"redhat-operators-nszwd\" (UID: \"483dc857-1f8b-4d00-bcfb-11e07773600e\") " pod="openshift-marketplace/redhat-operators-nszwd" Oct 05 09:44:12 crc kubenswrapper[4846]: I1005 09:44:12.846120 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d9zsd\" (UniqueName: \"kubernetes.io/projected/483dc857-1f8b-4d00-bcfb-11e07773600e-kube-api-access-d9zsd\") pod \"redhat-operators-nszwd\" (UID: \"483dc857-1f8b-4d00-bcfb-11e07773600e\") " pod="openshift-marketplace/redhat-operators-nszwd" Oct 05 09:44:12 crc kubenswrapper[4846]: I1005 09:44:12.942530 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-nszwd" Oct 05 09:44:13 crc kubenswrapper[4846]: I1005 09:44:13.057907 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-z4vbf" event={"ID":"618e874d-1fbf-4ea0-a065-774350222fd0","Type":"ContainerStarted","Data":"099eaf2990fe6c332cba37cfbcd50a706a6b47bffbb78d0b3d6ec6d7ccf3be07"} Oct 05 09:44:13 crc kubenswrapper[4846]: I1005 09:44:13.419445 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-z4vbf" podStartSLOduration=2.847648455 podStartE2EDuration="5.419422506s" podCreationTimestamp="2025-10-05 09:44:08 +0000 UTC" firstStartedPulling="2025-10-05 09:44:10.020572458 +0000 UTC m=+10572.261425233" lastFinishedPulling="2025-10-05 09:44:12.592346509 +0000 UTC m=+10574.833199284" observedRunningTime="2025-10-05 09:44:13.078221942 +0000 UTC m=+10575.319074737" watchObservedRunningTime="2025-10-05 09:44:13.419422506 +0000 UTC m=+10575.660275281" Oct 05 09:44:13 crc kubenswrapper[4846]: I1005 09:44:13.427097 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-nszwd"] Oct 05 09:44:13 crc kubenswrapper[4846]: W1005 09:44:13.799496 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod483dc857_1f8b_4d00_bcfb_11e07773600e.slice/crio-454f8f4a2efc03dc45f99c233e4999f84dc379c6da442794efd8fef0c85a6a68 WatchSource:0}: Error finding container 454f8f4a2efc03dc45f99c233e4999f84dc379c6da442794efd8fef0c85a6a68: Status 404 returned error can't find the container with id 454f8f4a2efc03dc45f99c233e4999f84dc379c6da442794efd8fef0c85a6a68 Oct 05 09:44:14 crc kubenswrapper[4846]: I1005 09:44:14.069608 4846 generic.go:334] "Generic (PLEG): container finished" podID="483dc857-1f8b-4d00-bcfb-11e07773600e" containerID="efae05574094b1c1b5b5d1f33f32ab4bcaff45aab3c84d403e0c6cef6c4c7574" exitCode=0 Oct 05 09:44:14 crc kubenswrapper[4846]: I1005 09:44:14.069728 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nszwd" event={"ID":"483dc857-1f8b-4d00-bcfb-11e07773600e","Type":"ContainerDied","Data":"efae05574094b1c1b5b5d1f33f32ab4bcaff45aab3c84d403e0c6cef6c4c7574"} Oct 05 09:44:14 crc kubenswrapper[4846]: I1005 09:44:14.069995 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nszwd" event={"ID":"483dc857-1f8b-4d00-bcfb-11e07773600e","Type":"ContainerStarted","Data":"454f8f4a2efc03dc45f99c233e4999f84dc379c6da442794efd8fef0c85a6a68"} Oct 05 09:44:15 crc kubenswrapper[4846]: I1005 09:44:15.498198 4846 scope.go:117] "RemoveContainer" containerID="43e1f18d9b5aea38dd1527fbb4fc557f1b37c59dfd2bdc2019c291047f4244b8" Oct 05 09:44:15 crc kubenswrapper[4846]: E1005 09:44:15.498644 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 09:44:16 crc kubenswrapper[4846]: I1005 09:44:16.103754 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nszwd" event={"ID":"483dc857-1f8b-4d00-bcfb-11e07773600e","Type":"ContainerStarted","Data":"8135938d9894ba83720bce36365dc9a09185a214bdd81a700e6b48c6eb02c8ea"} Oct 05 09:44:18 crc kubenswrapper[4846]: I1005 09:44:18.131201 4846 generic.go:334] "Generic (PLEG): container finished" podID="483dc857-1f8b-4d00-bcfb-11e07773600e" containerID="8135938d9894ba83720bce36365dc9a09185a214bdd81a700e6b48c6eb02c8ea" exitCode=0 Oct 05 09:44:18 crc kubenswrapper[4846]: I1005 09:44:18.131386 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nszwd" event={"ID":"483dc857-1f8b-4d00-bcfb-11e07773600e","Type":"ContainerDied","Data":"8135938d9894ba83720bce36365dc9a09185a214bdd81a700e6b48c6eb02c8ea"} Oct 05 09:44:19 crc kubenswrapper[4846]: I1005 09:44:19.131547 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-z4vbf" Oct 05 09:44:19 crc kubenswrapper[4846]: I1005 09:44:19.131966 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-z4vbf" Oct 05 09:44:19 crc kubenswrapper[4846]: I1005 09:44:19.147729 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nszwd" event={"ID":"483dc857-1f8b-4d00-bcfb-11e07773600e","Type":"ContainerStarted","Data":"27bc09278e10f6008696f2b5c9038661a5e45bc6c6db5b090e4bfd81917c163c"} Oct 05 09:44:19 crc kubenswrapper[4846]: I1005 09:44:19.172788 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-nszwd" podStartSLOduration=2.6642962900000002 podStartE2EDuration="7.172770793s" podCreationTimestamp="2025-10-05 09:44:12 +0000 UTC" firstStartedPulling="2025-10-05 09:44:14.07241916 +0000 UTC m=+10576.313271935" lastFinishedPulling="2025-10-05 09:44:18.580893673 +0000 UTC m=+10580.821746438" observedRunningTime="2025-10-05 09:44:19.169676821 +0000 UTC m=+10581.410529606" watchObservedRunningTime="2025-10-05 09:44:19.172770793 +0000 UTC m=+10581.413623578" Oct 05 09:44:19 crc kubenswrapper[4846]: I1005 09:44:19.193163 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-z4vbf" Oct 05 09:44:20 crc kubenswrapper[4846]: I1005 09:44:20.216769 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-z4vbf" Oct 05 09:44:20 crc kubenswrapper[4846]: I1005 09:44:20.970820 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-z4vbf"] Oct 05 09:44:22 crc kubenswrapper[4846]: I1005 09:44:22.174748 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-z4vbf" podUID="618e874d-1fbf-4ea0-a065-774350222fd0" containerName="registry-server" containerID="cri-o://099eaf2990fe6c332cba37cfbcd50a706a6b47bffbb78d0b3d6ec6d7ccf3be07" gracePeriod=2 Oct 05 09:44:22 crc kubenswrapper[4846]: I1005 09:44:22.719152 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-z4vbf" Oct 05 09:44:22 crc kubenswrapper[4846]: I1005 09:44:22.757639 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/618e874d-1fbf-4ea0-a065-774350222fd0-catalog-content\") pod \"618e874d-1fbf-4ea0-a065-774350222fd0\" (UID: \"618e874d-1fbf-4ea0-a065-774350222fd0\") " Oct 05 09:44:22 crc kubenswrapper[4846]: I1005 09:44:22.757754 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lxftf\" (UniqueName: \"kubernetes.io/projected/618e874d-1fbf-4ea0-a065-774350222fd0-kube-api-access-lxftf\") pod \"618e874d-1fbf-4ea0-a065-774350222fd0\" (UID: \"618e874d-1fbf-4ea0-a065-774350222fd0\") " Oct 05 09:44:22 crc kubenswrapper[4846]: I1005 09:44:22.757845 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/618e874d-1fbf-4ea0-a065-774350222fd0-utilities\") pod \"618e874d-1fbf-4ea0-a065-774350222fd0\" (UID: \"618e874d-1fbf-4ea0-a065-774350222fd0\") " Oct 05 09:44:22 crc kubenswrapper[4846]: I1005 09:44:22.758925 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/618e874d-1fbf-4ea0-a065-774350222fd0-utilities" (OuterVolumeSpecName: "utilities") pod "618e874d-1fbf-4ea0-a065-774350222fd0" (UID: "618e874d-1fbf-4ea0-a065-774350222fd0"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 09:44:22 crc kubenswrapper[4846]: I1005 09:44:22.759390 4846 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/618e874d-1fbf-4ea0-a065-774350222fd0-utilities\") on node \"crc\" DevicePath \"\"" Oct 05 09:44:22 crc kubenswrapper[4846]: I1005 09:44:22.764102 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/618e874d-1fbf-4ea0-a065-774350222fd0-kube-api-access-lxftf" (OuterVolumeSpecName: "kube-api-access-lxftf") pod "618e874d-1fbf-4ea0-a065-774350222fd0" (UID: "618e874d-1fbf-4ea0-a065-774350222fd0"). InnerVolumeSpecName "kube-api-access-lxftf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 09:44:22 crc kubenswrapper[4846]: I1005 09:44:22.803736 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/618e874d-1fbf-4ea0-a065-774350222fd0-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "618e874d-1fbf-4ea0-a065-774350222fd0" (UID: "618e874d-1fbf-4ea0-a065-774350222fd0"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 09:44:22 crc kubenswrapper[4846]: I1005 09:44:22.861765 4846 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/618e874d-1fbf-4ea0-a065-774350222fd0-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 05 09:44:22 crc kubenswrapper[4846]: I1005 09:44:22.862012 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lxftf\" (UniqueName: \"kubernetes.io/projected/618e874d-1fbf-4ea0-a065-774350222fd0-kube-api-access-lxftf\") on node \"crc\" DevicePath \"\"" Oct 05 09:44:22 crc kubenswrapper[4846]: I1005 09:44:22.943411 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-nszwd" Oct 05 09:44:22 crc kubenswrapper[4846]: I1005 09:44:22.944398 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-nszwd" Oct 05 09:44:23 crc kubenswrapper[4846]: I1005 09:44:23.214789 4846 generic.go:334] "Generic (PLEG): container finished" podID="618e874d-1fbf-4ea0-a065-774350222fd0" containerID="099eaf2990fe6c332cba37cfbcd50a706a6b47bffbb78d0b3d6ec6d7ccf3be07" exitCode=0 Oct 05 09:44:23 crc kubenswrapper[4846]: I1005 09:44:23.214905 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-z4vbf" event={"ID":"618e874d-1fbf-4ea0-a065-774350222fd0","Type":"ContainerDied","Data":"099eaf2990fe6c332cba37cfbcd50a706a6b47bffbb78d0b3d6ec6d7ccf3be07"} Oct 05 09:44:23 crc kubenswrapper[4846]: I1005 09:44:23.214971 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-z4vbf" event={"ID":"618e874d-1fbf-4ea0-a065-774350222fd0","Type":"ContainerDied","Data":"3d78e7a90c012222c0ae6b45b58e6a93adb0584806cadaa06158fb9addd84ba5"} Oct 05 09:44:23 crc kubenswrapper[4846]: I1005 09:44:23.214996 4846 scope.go:117] "RemoveContainer" containerID="099eaf2990fe6c332cba37cfbcd50a706a6b47bffbb78d0b3d6ec6d7ccf3be07" Oct 05 09:44:23 crc kubenswrapper[4846]: I1005 09:44:23.215424 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-z4vbf" Oct 05 09:44:23 crc kubenswrapper[4846]: I1005 09:44:23.240894 4846 scope.go:117] "RemoveContainer" containerID="a68defebaddf47f281d75d7601daea8b8f658431106b24bbf2ee25b1cf4741dd" Oct 05 09:44:23 crc kubenswrapper[4846]: I1005 09:44:23.263427 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-z4vbf"] Oct 05 09:44:23 crc kubenswrapper[4846]: I1005 09:44:23.274944 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-z4vbf"] Oct 05 09:44:23 crc kubenswrapper[4846]: I1005 09:44:23.280723 4846 scope.go:117] "RemoveContainer" containerID="e9ec438ed3f0ba233a71049cd162dd34ae952b2b096d3a3762f0cd9da0b1db32" Oct 05 09:44:23 crc kubenswrapper[4846]: I1005 09:44:23.312881 4846 scope.go:117] "RemoveContainer" containerID="099eaf2990fe6c332cba37cfbcd50a706a6b47bffbb78d0b3d6ec6d7ccf3be07" Oct 05 09:44:23 crc kubenswrapper[4846]: E1005 09:44:23.313262 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"099eaf2990fe6c332cba37cfbcd50a706a6b47bffbb78d0b3d6ec6d7ccf3be07\": container with ID starting with 099eaf2990fe6c332cba37cfbcd50a706a6b47bffbb78d0b3d6ec6d7ccf3be07 not found: ID does not exist" containerID="099eaf2990fe6c332cba37cfbcd50a706a6b47bffbb78d0b3d6ec6d7ccf3be07" Oct 05 09:44:23 crc kubenswrapper[4846]: I1005 09:44:23.313293 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"099eaf2990fe6c332cba37cfbcd50a706a6b47bffbb78d0b3d6ec6d7ccf3be07"} err="failed to get container status \"099eaf2990fe6c332cba37cfbcd50a706a6b47bffbb78d0b3d6ec6d7ccf3be07\": rpc error: code = NotFound desc = could not find container \"099eaf2990fe6c332cba37cfbcd50a706a6b47bffbb78d0b3d6ec6d7ccf3be07\": container with ID starting with 099eaf2990fe6c332cba37cfbcd50a706a6b47bffbb78d0b3d6ec6d7ccf3be07 not found: ID does not exist" Oct 05 09:44:23 crc kubenswrapper[4846]: I1005 09:44:23.313321 4846 scope.go:117] "RemoveContainer" containerID="a68defebaddf47f281d75d7601daea8b8f658431106b24bbf2ee25b1cf4741dd" Oct 05 09:44:23 crc kubenswrapper[4846]: E1005 09:44:23.313549 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a68defebaddf47f281d75d7601daea8b8f658431106b24bbf2ee25b1cf4741dd\": container with ID starting with a68defebaddf47f281d75d7601daea8b8f658431106b24bbf2ee25b1cf4741dd not found: ID does not exist" containerID="a68defebaddf47f281d75d7601daea8b8f658431106b24bbf2ee25b1cf4741dd" Oct 05 09:44:23 crc kubenswrapper[4846]: I1005 09:44:23.313601 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a68defebaddf47f281d75d7601daea8b8f658431106b24bbf2ee25b1cf4741dd"} err="failed to get container status \"a68defebaddf47f281d75d7601daea8b8f658431106b24bbf2ee25b1cf4741dd\": rpc error: code = NotFound desc = could not find container \"a68defebaddf47f281d75d7601daea8b8f658431106b24bbf2ee25b1cf4741dd\": container with ID starting with a68defebaddf47f281d75d7601daea8b8f658431106b24bbf2ee25b1cf4741dd not found: ID does not exist" Oct 05 09:44:23 crc kubenswrapper[4846]: I1005 09:44:23.313634 4846 scope.go:117] "RemoveContainer" containerID="e9ec438ed3f0ba233a71049cd162dd34ae952b2b096d3a3762f0cd9da0b1db32" Oct 05 09:44:23 crc kubenswrapper[4846]: E1005 09:44:23.313923 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e9ec438ed3f0ba233a71049cd162dd34ae952b2b096d3a3762f0cd9da0b1db32\": container with ID starting with e9ec438ed3f0ba233a71049cd162dd34ae952b2b096d3a3762f0cd9da0b1db32 not found: ID does not exist" containerID="e9ec438ed3f0ba233a71049cd162dd34ae952b2b096d3a3762f0cd9da0b1db32" Oct 05 09:44:23 crc kubenswrapper[4846]: I1005 09:44:23.313973 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e9ec438ed3f0ba233a71049cd162dd34ae952b2b096d3a3762f0cd9da0b1db32"} err="failed to get container status \"e9ec438ed3f0ba233a71049cd162dd34ae952b2b096d3a3762f0cd9da0b1db32\": rpc error: code = NotFound desc = could not find container \"e9ec438ed3f0ba233a71049cd162dd34ae952b2b096d3a3762f0cd9da0b1db32\": container with ID starting with e9ec438ed3f0ba233a71049cd162dd34ae952b2b096d3a3762f0cd9da0b1db32 not found: ID does not exist" Oct 05 09:44:23 crc kubenswrapper[4846]: I1005 09:44:23.999665 4846 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-nszwd" podUID="483dc857-1f8b-4d00-bcfb-11e07773600e" containerName="registry-server" probeResult="failure" output=< Oct 05 09:44:23 crc kubenswrapper[4846]: timeout: failed to connect service ":50051" within 1s Oct 05 09:44:23 crc kubenswrapper[4846]: > Oct 05 09:44:24 crc kubenswrapper[4846]: I1005 09:44:24.512055 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="618e874d-1fbf-4ea0-a065-774350222fd0" path="/var/lib/kubelet/pods/618e874d-1fbf-4ea0-a065-774350222fd0/volumes" Oct 05 09:44:30 crc kubenswrapper[4846]: I1005 09:44:30.498705 4846 scope.go:117] "RemoveContainer" containerID="43e1f18d9b5aea38dd1527fbb4fc557f1b37c59dfd2bdc2019c291047f4244b8" Oct 05 09:44:30 crc kubenswrapper[4846]: E1005 09:44:30.500863 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 09:44:33 crc kubenswrapper[4846]: I1005 09:44:33.024771 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-nszwd" Oct 05 09:44:33 crc kubenswrapper[4846]: I1005 09:44:33.079728 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-nszwd" Oct 05 09:44:34 crc kubenswrapper[4846]: I1005 09:44:34.574936 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-nszwd"] Oct 05 09:44:34 crc kubenswrapper[4846]: I1005 09:44:34.575649 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-nszwd" podUID="483dc857-1f8b-4d00-bcfb-11e07773600e" containerName="registry-server" containerID="cri-o://27bc09278e10f6008696f2b5c9038661a5e45bc6c6db5b090e4bfd81917c163c" gracePeriod=2 Oct 05 09:44:35 crc kubenswrapper[4846]: I1005 09:44:35.125472 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-nszwd" Oct 05 09:44:35 crc kubenswrapper[4846]: I1005 09:44:35.247944 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/483dc857-1f8b-4d00-bcfb-11e07773600e-catalog-content\") pod \"483dc857-1f8b-4d00-bcfb-11e07773600e\" (UID: \"483dc857-1f8b-4d00-bcfb-11e07773600e\") " Oct 05 09:44:35 crc kubenswrapper[4846]: I1005 09:44:35.248529 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/483dc857-1f8b-4d00-bcfb-11e07773600e-utilities\") pod \"483dc857-1f8b-4d00-bcfb-11e07773600e\" (UID: \"483dc857-1f8b-4d00-bcfb-11e07773600e\") " Oct 05 09:44:35 crc kubenswrapper[4846]: I1005 09:44:35.249836 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/483dc857-1f8b-4d00-bcfb-11e07773600e-utilities" (OuterVolumeSpecName: "utilities") pod "483dc857-1f8b-4d00-bcfb-11e07773600e" (UID: "483dc857-1f8b-4d00-bcfb-11e07773600e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 09:44:35 crc kubenswrapper[4846]: I1005 09:44:35.250273 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d9zsd\" (UniqueName: \"kubernetes.io/projected/483dc857-1f8b-4d00-bcfb-11e07773600e-kube-api-access-d9zsd\") pod \"483dc857-1f8b-4d00-bcfb-11e07773600e\" (UID: \"483dc857-1f8b-4d00-bcfb-11e07773600e\") " Oct 05 09:44:35 crc kubenswrapper[4846]: I1005 09:44:35.252593 4846 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/483dc857-1f8b-4d00-bcfb-11e07773600e-utilities\") on node \"crc\" DevicePath \"\"" Oct 05 09:44:35 crc kubenswrapper[4846]: I1005 09:44:35.257501 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/483dc857-1f8b-4d00-bcfb-11e07773600e-kube-api-access-d9zsd" (OuterVolumeSpecName: "kube-api-access-d9zsd") pod "483dc857-1f8b-4d00-bcfb-11e07773600e" (UID: "483dc857-1f8b-4d00-bcfb-11e07773600e"). InnerVolumeSpecName "kube-api-access-d9zsd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 09:44:35 crc kubenswrapper[4846]: I1005 09:44:35.334579 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/483dc857-1f8b-4d00-bcfb-11e07773600e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "483dc857-1f8b-4d00-bcfb-11e07773600e" (UID: "483dc857-1f8b-4d00-bcfb-11e07773600e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 09:44:35 crc kubenswrapper[4846]: I1005 09:44:35.353592 4846 generic.go:334] "Generic (PLEG): container finished" podID="483dc857-1f8b-4d00-bcfb-11e07773600e" containerID="27bc09278e10f6008696f2b5c9038661a5e45bc6c6db5b090e4bfd81917c163c" exitCode=0 Oct 05 09:44:35 crc kubenswrapper[4846]: I1005 09:44:35.353668 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nszwd" event={"ID":"483dc857-1f8b-4d00-bcfb-11e07773600e","Type":"ContainerDied","Data":"27bc09278e10f6008696f2b5c9038661a5e45bc6c6db5b090e4bfd81917c163c"} Oct 05 09:44:35 crc kubenswrapper[4846]: I1005 09:44:35.353742 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nszwd" event={"ID":"483dc857-1f8b-4d00-bcfb-11e07773600e","Type":"ContainerDied","Data":"454f8f4a2efc03dc45f99c233e4999f84dc379c6da442794efd8fef0c85a6a68"} Oct 05 09:44:35 crc kubenswrapper[4846]: I1005 09:44:35.353770 4846 scope.go:117] "RemoveContainer" containerID="27bc09278e10f6008696f2b5c9038661a5e45bc6c6db5b090e4bfd81917c163c" Oct 05 09:44:35 crc kubenswrapper[4846]: I1005 09:44:35.354023 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-nszwd" Oct 05 09:44:35 crc kubenswrapper[4846]: I1005 09:44:35.358547 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d9zsd\" (UniqueName: \"kubernetes.io/projected/483dc857-1f8b-4d00-bcfb-11e07773600e-kube-api-access-d9zsd\") on node \"crc\" DevicePath \"\"" Oct 05 09:44:35 crc kubenswrapper[4846]: I1005 09:44:35.358591 4846 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/483dc857-1f8b-4d00-bcfb-11e07773600e-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 05 09:44:35 crc kubenswrapper[4846]: I1005 09:44:35.387557 4846 scope.go:117] "RemoveContainer" containerID="8135938d9894ba83720bce36365dc9a09185a214bdd81a700e6b48c6eb02c8ea" Oct 05 09:44:35 crc kubenswrapper[4846]: I1005 09:44:35.402078 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-nszwd"] Oct 05 09:44:35 crc kubenswrapper[4846]: I1005 09:44:35.410808 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-nszwd"] Oct 05 09:44:35 crc kubenswrapper[4846]: I1005 09:44:35.431534 4846 scope.go:117] "RemoveContainer" containerID="efae05574094b1c1b5b5d1f33f32ab4bcaff45aab3c84d403e0c6cef6c4c7574" Oct 05 09:44:35 crc kubenswrapper[4846]: I1005 09:44:35.465553 4846 scope.go:117] "RemoveContainer" containerID="27bc09278e10f6008696f2b5c9038661a5e45bc6c6db5b090e4bfd81917c163c" Oct 05 09:44:35 crc kubenswrapper[4846]: E1005 09:44:35.465981 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"27bc09278e10f6008696f2b5c9038661a5e45bc6c6db5b090e4bfd81917c163c\": container with ID starting with 27bc09278e10f6008696f2b5c9038661a5e45bc6c6db5b090e4bfd81917c163c not found: ID does not exist" containerID="27bc09278e10f6008696f2b5c9038661a5e45bc6c6db5b090e4bfd81917c163c" Oct 05 09:44:35 crc kubenswrapper[4846]: I1005 09:44:35.466023 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"27bc09278e10f6008696f2b5c9038661a5e45bc6c6db5b090e4bfd81917c163c"} err="failed to get container status \"27bc09278e10f6008696f2b5c9038661a5e45bc6c6db5b090e4bfd81917c163c\": rpc error: code = NotFound desc = could not find container \"27bc09278e10f6008696f2b5c9038661a5e45bc6c6db5b090e4bfd81917c163c\": container with ID starting with 27bc09278e10f6008696f2b5c9038661a5e45bc6c6db5b090e4bfd81917c163c not found: ID does not exist" Oct 05 09:44:35 crc kubenswrapper[4846]: I1005 09:44:35.466050 4846 scope.go:117] "RemoveContainer" containerID="8135938d9894ba83720bce36365dc9a09185a214bdd81a700e6b48c6eb02c8ea" Oct 05 09:44:35 crc kubenswrapper[4846]: E1005 09:44:35.466469 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8135938d9894ba83720bce36365dc9a09185a214bdd81a700e6b48c6eb02c8ea\": container with ID starting with 8135938d9894ba83720bce36365dc9a09185a214bdd81a700e6b48c6eb02c8ea not found: ID does not exist" containerID="8135938d9894ba83720bce36365dc9a09185a214bdd81a700e6b48c6eb02c8ea" Oct 05 09:44:35 crc kubenswrapper[4846]: I1005 09:44:35.466504 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8135938d9894ba83720bce36365dc9a09185a214bdd81a700e6b48c6eb02c8ea"} err="failed to get container status \"8135938d9894ba83720bce36365dc9a09185a214bdd81a700e6b48c6eb02c8ea\": rpc error: code = NotFound desc = could not find container \"8135938d9894ba83720bce36365dc9a09185a214bdd81a700e6b48c6eb02c8ea\": container with ID starting with 8135938d9894ba83720bce36365dc9a09185a214bdd81a700e6b48c6eb02c8ea not found: ID does not exist" Oct 05 09:44:35 crc kubenswrapper[4846]: I1005 09:44:35.466525 4846 scope.go:117] "RemoveContainer" containerID="efae05574094b1c1b5b5d1f33f32ab4bcaff45aab3c84d403e0c6cef6c4c7574" Oct 05 09:44:35 crc kubenswrapper[4846]: E1005 09:44:35.466904 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"efae05574094b1c1b5b5d1f33f32ab4bcaff45aab3c84d403e0c6cef6c4c7574\": container with ID starting with efae05574094b1c1b5b5d1f33f32ab4bcaff45aab3c84d403e0c6cef6c4c7574 not found: ID does not exist" containerID="efae05574094b1c1b5b5d1f33f32ab4bcaff45aab3c84d403e0c6cef6c4c7574" Oct 05 09:44:35 crc kubenswrapper[4846]: I1005 09:44:35.466930 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"efae05574094b1c1b5b5d1f33f32ab4bcaff45aab3c84d403e0c6cef6c4c7574"} err="failed to get container status \"efae05574094b1c1b5b5d1f33f32ab4bcaff45aab3c84d403e0c6cef6c4c7574\": rpc error: code = NotFound desc = could not find container \"efae05574094b1c1b5b5d1f33f32ab4bcaff45aab3c84d403e0c6cef6c4c7574\": container with ID starting with efae05574094b1c1b5b5d1f33f32ab4bcaff45aab3c84d403e0c6cef6c4c7574 not found: ID does not exist" Oct 05 09:44:36 crc kubenswrapper[4846]: I1005 09:44:36.509748 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="483dc857-1f8b-4d00-bcfb-11e07773600e" path="/var/lib/kubelet/pods/483dc857-1f8b-4d00-bcfb-11e07773600e/volumes" Oct 05 09:44:45 crc kubenswrapper[4846]: I1005 09:44:45.497389 4846 scope.go:117] "RemoveContainer" containerID="43e1f18d9b5aea38dd1527fbb4fc557f1b37c59dfd2bdc2019c291047f4244b8" Oct 05 09:44:45 crc kubenswrapper[4846]: E1005 09:44:45.498295 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 09:44:59 crc kubenswrapper[4846]: I1005 09:44:59.497400 4846 scope.go:117] "RemoveContainer" containerID="43e1f18d9b5aea38dd1527fbb4fc557f1b37c59dfd2bdc2019c291047f4244b8" Oct 05 09:44:59 crc kubenswrapper[4846]: E1005 09:44:59.498157 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 09:45:00 crc kubenswrapper[4846]: I1005 09:45:00.187104 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29327625-gfctl"] Oct 05 09:45:00 crc kubenswrapper[4846]: E1005 09:45:00.187706 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="618e874d-1fbf-4ea0-a065-774350222fd0" containerName="extract-utilities" Oct 05 09:45:00 crc kubenswrapper[4846]: I1005 09:45:00.187732 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="618e874d-1fbf-4ea0-a065-774350222fd0" containerName="extract-utilities" Oct 05 09:45:00 crc kubenswrapper[4846]: E1005 09:45:00.187745 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="618e874d-1fbf-4ea0-a065-774350222fd0" containerName="extract-content" Oct 05 09:45:00 crc kubenswrapper[4846]: I1005 09:45:00.187753 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="618e874d-1fbf-4ea0-a065-774350222fd0" containerName="extract-content" Oct 05 09:45:00 crc kubenswrapper[4846]: E1005 09:45:00.187775 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="483dc857-1f8b-4d00-bcfb-11e07773600e" containerName="registry-server" Oct 05 09:45:00 crc kubenswrapper[4846]: I1005 09:45:00.187785 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="483dc857-1f8b-4d00-bcfb-11e07773600e" containerName="registry-server" Oct 05 09:45:00 crc kubenswrapper[4846]: E1005 09:45:00.187814 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="483dc857-1f8b-4d00-bcfb-11e07773600e" containerName="extract-utilities" Oct 05 09:45:00 crc kubenswrapper[4846]: I1005 09:45:00.187822 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="483dc857-1f8b-4d00-bcfb-11e07773600e" containerName="extract-utilities" Oct 05 09:45:00 crc kubenswrapper[4846]: E1005 09:45:00.187848 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="618e874d-1fbf-4ea0-a065-774350222fd0" containerName="registry-server" Oct 05 09:45:00 crc kubenswrapper[4846]: I1005 09:45:00.187855 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="618e874d-1fbf-4ea0-a065-774350222fd0" containerName="registry-server" Oct 05 09:45:00 crc kubenswrapper[4846]: E1005 09:45:00.187871 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="483dc857-1f8b-4d00-bcfb-11e07773600e" containerName="extract-content" Oct 05 09:45:00 crc kubenswrapper[4846]: I1005 09:45:00.187876 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="483dc857-1f8b-4d00-bcfb-11e07773600e" containerName="extract-content" Oct 05 09:45:00 crc kubenswrapper[4846]: I1005 09:45:00.188075 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="483dc857-1f8b-4d00-bcfb-11e07773600e" containerName="registry-server" Oct 05 09:45:00 crc kubenswrapper[4846]: I1005 09:45:00.188112 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="618e874d-1fbf-4ea0-a065-774350222fd0" containerName="registry-server" Oct 05 09:45:00 crc kubenswrapper[4846]: I1005 09:45:00.188956 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29327625-gfctl" Oct 05 09:45:00 crc kubenswrapper[4846]: I1005 09:45:00.190950 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 05 09:45:00 crc kubenswrapper[4846]: I1005 09:45:00.191029 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 05 09:45:00 crc kubenswrapper[4846]: I1005 09:45:00.200284 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29327625-gfctl"] Oct 05 09:45:00 crc kubenswrapper[4846]: I1005 09:45:00.234369 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b3aaea93-f35b-4da5-a94d-e9d49e7266c6-secret-volume\") pod \"collect-profiles-29327625-gfctl\" (UID: \"b3aaea93-f35b-4da5-a94d-e9d49e7266c6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327625-gfctl" Oct 05 09:45:00 crc kubenswrapper[4846]: I1005 09:45:00.234519 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7fjfl\" (UniqueName: \"kubernetes.io/projected/b3aaea93-f35b-4da5-a94d-e9d49e7266c6-kube-api-access-7fjfl\") pod \"collect-profiles-29327625-gfctl\" (UID: \"b3aaea93-f35b-4da5-a94d-e9d49e7266c6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327625-gfctl" Oct 05 09:45:00 crc kubenswrapper[4846]: I1005 09:45:00.234639 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b3aaea93-f35b-4da5-a94d-e9d49e7266c6-config-volume\") pod \"collect-profiles-29327625-gfctl\" (UID: \"b3aaea93-f35b-4da5-a94d-e9d49e7266c6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327625-gfctl" Oct 05 09:45:00 crc kubenswrapper[4846]: I1005 09:45:00.336161 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b3aaea93-f35b-4da5-a94d-e9d49e7266c6-config-volume\") pod \"collect-profiles-29327625-gfctl\" (UID: \"b3aaea93-f35b-4da5-a94d-e9d49e7266c6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327625-gfctl" Oct 05 09:45:00 crc kubenswrapper[4846]: I1005 09:45:00.336302 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b3aaea93-f35b-4da5-a94d-e9d49e7266c6-secret-volume\") pod \"collect-profiles-29327625-gfctl\" (UID: \"b3aaea93-f35b-4da5-a94d-e9d49e7266c6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327625-gfctl" Oct 05 09:45:00 crc kubenswrapper[4846]: I1005 09:45:00.336378 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7fjfl\" (UniqueName: \"kubernetes.io/projected/b3aaea93-f35b-4da5-a94d-e9d49e7266c6-kube-api-access-7fjfl\") pod \"collect-profiles-29327625-gfctl\" (UID: \"b3aaea93-f35b-4da5-a94d-e9d49e7266c6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327625-gfctl" Oct 05 09:45:00 crc kubenswrapper[4846]: I1005 09:45:00.338049 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b3aaea93-f35b-4da5-a94d-e9d49e7266c6-config-volume\") pod \"collect-profiles-29327625-gfctl\" (UID: \"b3aaea93-f35b-4da5-a94d-e9d49e7266c6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327625-gfctl" Oct 05 09:45:00 crc kubenswrapper[4846]: I1005 09:45:00.347049 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b3aaea93-f35b-4da5-a94d-e9d49e7266c6-secret-volume\") pod \"collect-profiles-29327625-gfctl\" (UID: \"b3aaea93-f35b-4da5-a94d-e9d49e7266c6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327625-gfctl" Oct 05 09:45:00 crc kubenswrapper[4846]: I1005 09:45:00.363667 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7fjfl\" (UniqueName: \"kubernetes.io/projected/b3aaea93-f35b-4da5-a94d-e9d49e7266c6-kube-api-access-7fjfl\") pod \"collect-profiles-29327625-gfctl\" (UID: \"b3aaea93-f35b-4da5-a94d-e9d49e7266c6\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29327625-gfctl" Oct 05 09:45:00 crc kubenswrapper[4846]: I1005 09:45:00.505999 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29327625-gfctl" Oct 05 09:45:00 crc kubenswrapper[4846]: I1005 09:45:00.978539 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29327625-gfctl"] Oct 05 09:45:01 crc kubenswrapper[4846]: I1005 09:45:01.705808 4846 generic.go:334] "Generic (PLEG): container finished" podID="b3aaea93-f35b-4da5-a94d-e9d49e7266c6" containerID="e864bc4baf7a7496e1fb3d3751d288594a4919b1a345be4b93bc1325e4b78b76" exitCode=0 Oct 05 09:45:01 crc kubenswrapper[4846]: I1005 09:45:01.705865 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29327625-gfctl" event={"ID":"b3aaea93-f35b-4da5-a94d-e9d49e7266c6","Type":"ContainerDied","Data":"e864bc4baf7a7496e1fb3d3751d288594a4919b1a345be4b93bc1325e4b78b76"} Oct 05 09:45:01 crc kubenswrapper[4846]: I1005 09:45:01.706215 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29327625-gfctl" event={"ID":"b3aaea93-f35b-4da5-a94d-e9d49e7266c6","Type":"ContainerStarted","Data":"80f57ccc1f2c850ed2b863f4e1ca6514d888b79bf66fc43f15a580e381683abb"} Oct 05 09:45:03 crc kubenswrapper[4846]: I1005 09:45:03.155736 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29327625-gfctl" Oct 05 09:45:03 crc kubenswrapper[4846]: I1005 09:45:03.221905 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b3aaea93-f35b-4da5-a94d-e9d49e7266c6-secret-volume\") pod \"b3aaea93-f35b-4da5-a94d-e9d49e7266c6\" (UID: \"b3aaea93-f35b-4da5-a94d-e9d49e7266c6\") " Oct 05 09:45:03 crc kubenswrapper[4846]: I1005 09:45:03.222983 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b3aaea93-f35b-4da5-a94d-e9d49e7266c6-config-volume\") pod \"b3aaea93-f35b-4da5-a94d-e9d49e7266c6\" (UID: \"b3aaea93-f35b-4da5-a94d-e9d49e7266c6\") " Oct 05 09:45:03 crc kubenswrapper[4846]: I1005 09:45:03.223078 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7fjfl\" (UniqueName: \"kubernetes.io/projected/b3aaea93-f35b-4da5-a94d-e9d49e7266c6-kube-api-access-7fjfl\") pod \"b3aaea93-f35b-4da5-a94d-e9d49e7266c6\" (UID: \"b3aaea93-f35b-4da5-a94d-e9d49e7266c6\") " Oct 05 09:45:03 crc kubenswrapper[4846]: I1005 09:45:03.224051 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b3aaea93-f35b-4da5-a94d-e9d49e7266c6-config-volume" (OuterVolumeSpecName: "config-volume") pod "b3aaea93-f35b-4da5-a94d-e9d49e7266c6" (UID: "b3aaea93-f35b-4da5-a94d-e9d49e7266c6"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 09:45:03 crc kubenswrapper[4846]: I1005 09:45:03.227282 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b3aaea93-f35b-4da5-a94d-e9d49e7266c6-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "b3aaea93-f35b-4da5-a94d-e9d49e7266c6" (UID: "b3aaea93-f35b-4da5-a94d-e9d49e7266c6"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 09:45:03 crc kubenswrapper[4846]: I1005 09:45:03.227597 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b3aaea93-f35b-4da5-a94d-e9d49e7266c6-kube-api-access-7fjfl" (OuterVolumeSpecName: "kube-api-access-7fjfl") pod "b3aaea93-f35b-4da5-a94d-e9d49e7266c6" (UID: "b3aaea93-f35b-4da5-a94d-e9d49e7266c6"). InnerVolumeSpecName "kube-api-access-7fjfl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 09:45:03 crc kubenswrapper[4846]: I1005 09:45:03.326578 4846 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b3aaea93-f35b-4da5-a94d-e9d49e7266c6-config-volume\") on node \"crc\" DevicePath \"\"" Oct 05 09:45:03 crc kubenswrapper[4846]: I1005 09:45:03.326636 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7fjfl\" (UniqueName: \"kubernetes.io/projected/b3aaea93-f35b-4da5-a94d-e9d49e7266c6-kube-api-access-7fjfl\") on node \"crc\" DevicePath \"\"" Oct 05 09:45:03 crc kubenswrapper[4846]: I1005 09:45:03.326660 4846 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b3aaea93-f35b-4da5-a94d-e9d49e7266c6-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 05 09:45:03 crc kubenswrapper[4846]: I1005 09:45:03.731025 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29327625-gfctl" event={"ID":"b3aaea93-f35b-4da5-a94d-e9d49e7266c6","Type":"ContainerDied","Data":"80f57ccc1f2c850ed2b863f4e1ca6514d888b79bf66fc43f15a580e381683abb"} Oct 05 09:45:03 crc kubenswrapper[4846]: I1005 09:45:03.731069 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="80f57ccc1f2c850ed2b863f4e1ca6514d888b79bf66fc43f15a580e381683abb" Oct 05 09:45:03 crc kubenswrapper[4846]: I1005 09:45:03.731595 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29327625-gfctl" Oct 05 09:45:04 crc kubenswrapper[4846]: I1005 09:45:04.264891 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29327580-shk64"] Oct 05 09:45:04 crc kubenswrapper[4846]: I1005 09:45:04.272911 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29327580-shk64"] Oct 05 09:45:04 crc kubenswrapper[4846]: I1005 09:45:04.513926 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b885295e-73f2-4fc4-8c83-149b88556acd" path="/var/lib/kubelet/pods/b885295e-73f2-4fc4-8c83-149b88556acd/volumes" Oct 05 09:45:10 crc kubenswrapper[4846]: I1005 09:45:10.811991 4846 generic.go:334] "Generic (PLEG): container finished" podID="fac403c1-c8cd-4675-ba46-59900a7cc18f" containerID="f38b40fca0e1a85108d450cbf8c3e04820b1957149b6e8013f5ca35446b8c5c2" exitCode=0 Oct 05 09:45:10 crc kubenswrapper[4846]: I1005 09:45:10.812104 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"fac403c1-c8cd-4675-ba46-59900a7cc18f","Type":"ContainerDied","Data":"f38b40fca0e1a85108d450cbf8c3e04820b1957149b6e8013f5ca35446b8c5c2"} Oct 05 09:45:12 crc kubenswrapper[4846]: I1005 09:45:12.163018 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Oct 05 09:45:12 crc kubenswrapper[4846]: I1005 09:45:12.347832 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/fac403c1-c8cd-4675-ba46-59900a7cc18f-openstack-config\") pod \"fac403c1-c8cd-4675-ba46-59900a7cc18f\" (UID: \"fac403c1-c8cd-4675-ba46-59900a7cc18f\") " Oct 05 09:45:12 crc kubenswrapper[4846]: I1005 09:45:12.347946 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fac403c1-c8cd-4675-ba46-59900a7cc18f-ssh-key\") pod \"fac403c1-c8cd-4675-ba46-59900a7cc18f\" (UID: \"fac403c1-c8cd-4675-ba46-59900a7cc18f\") " Oct 05 09:45:12 crc kubenswrapper[4846]: I1005 09:45:12.348039 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/fac403c1-c8cd-4675-ba46-59900a7cc18f-test-operator-ephemeral-workdir\") pod \"fac403c1-c8cd-4675-ba46-59900a7cc18f\" (UID: \"fac403c1-c8cd-4675-ba46-59900a7cc18f\") " Oct 05 09:45:12 crc kubenswrapper[4846]: I1005 09:45:12.348074 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-logs\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"fac403c1-c8cd-4675-ba46-59900a7cc18f\" (UID: \"fac403c1-c8cd-4675-ba46-59900a7cc18f\") " Oct 05 09:45:12 crc kubenswrapper[4846]: I1005 09:45:12.348126 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/fac403c1-c8cd-4675-ba46-59900a7cc18f-config-data\") pod \"fac403c1-c8cd-4675-ba46-59900a7cc18f\" (UID: \"fac403c1-c8cd-4675-ba46-59900a7cc18f\") " Oct 05 09:45:12 crc kubenswrapper[4846]: I1005 09:45:12.348219 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h85lt\" (UniqueName: \"kubernetes.io/projected/fac403c1-c8cd-4675-ba46-59900a7cc18f-kube-api-access-h85lt\") pod \"fac403c1-c8cd-4675-ba46-59900a7cc18f\" (UID: \"fac403c1-c8cd-4675-ba46-59900a7cc18f\") " Oct 05 09:45:12 crc kubenswrapper[4846]: I1005 09:45:12.348267 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/fac403c1-c8cd-4675-ba46-59900a7cc18f-ca-certs\") pod \"fac403c1-c8cd-4675-ba46-59900a7cc18f\" (UID: \"fac403c1-c8cd-4675-ba46-59900a7cc18f\") " Oct 05 09:45:12 crc kubenswrapper[4846]: I1005 09:45:12.348292 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/fac403c1-c8cd-4675-ba46-59900a7cc18f-test-operator-ephemeral-temporary\") pod \"fac403c1-c8cd-4675-ba46-59900a7cc18f\" (UID: \"fac403c1-c8cd-4675-ba46-59900a7cc18f\") " Oct 05 09:45:12 crc kubenswrapper[4846]: I1005 09:45:12.348340 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/fac403c1-c8cd-4675-ba46-59900a7cc18f-openstack-config-secret\") pod \"fac403c1-c8cd-4675-ba46-59900a7cc18f\" (UID: \"fac403c1-c8cd-4675-ba46-59900a7cc18f\") " Oct 05 09:45:12 crc kubenswrapper[4846]: I1005 09:45:12.350037 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fac403c1-c8cd-4675-ba46-59900a7cc18f-test-operator-ephemeral-temporary" (OuterVolumeSpecName: "test-operator-ephemeral-temporary") pod "fac403c1-c8cd-4675-ba46-59900a7cc18f" (UID: "fac403c1-c8cd-4675-ba46-59900a7cc18f"). InnerVolumeSpecName "test-operator-ephemeral-temporary". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 09:45:12 crc kubenswrapper[4846]: I1005 09:45:12.350339 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fac403c1-c8cd-4675-ba46-59900a7cc18f-config-data" (OuterVolumeSpecName: "config-data") pod "fac403c1-c8cd-4675-ba46-59900a7cc18f" (UID: "fac403c1-c8cd-4675-ba46-59900a7cc18f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 09:45:12 crc kubenswrapper[4846]: I1005 09:45:12.356161 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage07-crc" (OuterVolumeSpecName: "test-operator-logs") pod "fac403c1-c8cd-4675-ba46-59900a7cc18f" (UID: "fac403c1-c8cd-4675-ba46-59900a7cc18f"). InnerVolumeSpecName "local-storage07-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 05 09:45:12 crc kubenswrapper[4846]: I1005 09:45:12.357405 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fac403c1-c8cd-4675-ba46-59900a7cc18f-test-operator-ephemeral-workdir" (OuterVolumeSpecName: "test-operator-ephemeral-workdir") pod "fac403c1-c8cd-4675-ba46-59900a7cc18f" (UID: "fac403c1-c8cd-4675-ba46-59900a7cc18f"). InnerVolumeSpecName "test-operator-ephemeral-workdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 09:45:12 crc kubenswrapper[4846]: I1005 09:45:12.360403 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fac403c1-c8cd-4675-ba46-59900a7cc18f-kube-api-access-h85lt" (OuterVolumeSpecName: "kube-api-access-h85lt") pod "fac403c1-c8cd-4675-ba46-59900a7cc18f" (UID: "fac403c1-c8cd-4675-ba46-59900a7cc18f"). InnerVolumeSpecName "kube-api-access-h85lt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 09:45:12 crc kubenswrapper[4846]: I1005 09:45:12.388131 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fac403c1-c8cd-4675-ba46-59900a7cc18f-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "fac403c1-c8cd-4675-ba46-59900a7cc18f" (UID: "fac403c1-c8cd-4675-ba46-59900a7cc18f"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 09:45:12 crc kubenswrapper[4846]: I1005 09:45:12.391776 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fac403c1-c8cd-4675-ba46-59900a7cc18f-ca-certs" (OuterVolumeSpecName: "ca-certs") pod "fac403c1-c8cd-4675-ba46-59900a7cc18f" (UID: "fac403c1-c8cd-4675-ba46-59900a7cc18f"). InnerVolumeSpecName "ca-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 09:45:12 crc kubenswrapper[4846]: I1005 09:45:12.414359 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fac403c1-c8cd-4675-ba46-59900a7cc18f-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "fac403c1-c8cd-4675-ba46-59900a7cc18f" (UID: "fac403c1-c8cd-4675-ba46-59900a7cc18f"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 05 09:45:12 crc kubenswrapper[4846]: I1005 09:45:12.423112 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fac403c1-c8cd-4675-ba46-59900a7cc18f-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "fac403c1-c8cd-4675-ba46-59900a7cc18f" (UID: "fac403c1-c8cd-4675-ba46-59900a7cc18f"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 05 09:45:12 crc kubenswrapper[4846]: I1005 09:45:12.451068 4846 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/fac403c1-c8cd-4675-ba46-59900a7cc18f-openstack-config\") on node \"crc\" DevicePath \"\"" Oct 05 09:45:12 crc kubenswrapper[4846]: I1005 09:45:12.451112 4846 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fac403c1-c8cd-4675-ba46-59900a7cc18f-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 05 09:45:12 crc kubenswrapper[4846]: I1005 09:45:12.451129 4846 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/fac403c1-c8cd-4675-ba46-59900a7cc18f-test-operator-ephemeral-workdir\") on node \"crc\" DevicePath \"\"" Oct 05 09:45:12 crc kubenswrapper[4846]: I1005 09:45:12.451177 4846 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" " Oct 05 09:45:12 crc kubenswrapper[4846]: I1005 09:45:12.451205 4846 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/fac403c1-c8cd-4675-ba46-59900a7cc18f-config-data\") on node \"crc\" DevicePath \"\"" Oct 05 09:45:12 crc kubenswrapper[4846]: I1005 09:45:12.451218 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h85lt\" (UniqueName: \"kubernetes.io/projected/fac403c1-c8cd-4675-ba46-59900a7cc18f-kube-api-access-h85lt\") on node \"crc\" DevicePath \"\"" Oct 05 09:45:12 crc kubenswrapper[4846]: I1005 09:45:12.451232 4846 reconciler_common.go:293] "Volume detached for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/fac403c1-c8cd-4675-ba46-59900a7cc18f-ca-certs\") on node \"crc\" DevicePath \"\"" Oct 05 09:45:12 crc kubenswrapper[4846]: I1005 09:45:12.451245 4846 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/fac403c1-c8cd-4675-ba46-59900a7cc18f-test-operator-ephemeral-temporary\") on node \"crc\" DevicePath \"\"" Oct 05 09:45:12 crc kubenswrapper[4846]: I1005 09:45:12.451258 4846 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/fac403c1-c8cd-4675-ba46-59900a7cc18f-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Oct 05 09:45:12 crc kubenswrapper[4846]: I1005 09:45:12.485224 4846 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage07-crc" (UniqueName: "kubernetes.io/local-volume/local-storage07-crc") on node "crc" Oct 05 09:45:12 crc kubenswrapper[4846]: I1005 09:45:12.498262 4846 scope.go:117] "RemoveContainer" containerID="43e1f18d9b5aea38dd1527fbb4fc557f1b37c59dfd2bdc2019c291047f4244b8" Oct 05 09:45:12 crc kubenswrapper[4846]: E1005 09:45:12.498557 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 09:45:12 crc kubenswrapper[4846]: I1005 09:45:12.552524 4846 reconciler_common.go:293] "Volume detached for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" DevicePath \"\"" Oct 05 09:45:12 crc kubenswrapper[4846]: I1005 09:45:12.839795 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"fac403c1-c8cd-4675-ba46-59900a7cc18f","Type":"ContainerDied","Data":"b7d0f7c84d87c9a5ae9e1ca11f6598b1e30d5c4845659c44520b41e9d1c54dd5"} Oct 05 09:45:12 crc kubenswrapper[4846]: I1005 09:45:12.839828 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b7d0f7c84d87c9a5ae9e1ca11f6598b1e30d5c4845659c44520b41e9d1c54dd5" Oct 05 09:45:12 crc kubenswrapper[4846]: I1005 09:45:12.840304 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Oct 05 09:45:17 crc kubenswrapper[4846]: I1005 09:45:17.315026 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Oct 05 09:45:17 crc kubenswrapper[4846]: E1005 09:45:17.316402 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fac403c1-c8cd-4675-ba46-59900a7cc18f" containerName="tempest-tests-tempest-tests-runner" Oct 05 09:45:17 crc kubenswrapper[4846]: I1005 09:45:17.316427 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="fac403c1-c8cd-4675-ba46-59900a7cc18f" containerName="tempest-tests-tempest-tests-runner" Oct 05 09:45:17 crc kubenswrapper[4846]: E1005 09:45:17.316486 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b3aaea93-f35b-4da5-a94d-e9d49e7266c6" containerName="collect-profiles" Oct 05 09:45:17 crc kubenswrapper[4846]: I1005 09:45:17.316500 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="b3aaea93-f35b-4da5-a94d-e9d49e7266c6" containerName="collect-profiles" Oct 05 09:45:17 crc kubenswrapper[4846]: I1005 09:45:17.316877 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="b3aaea93-f35b-4da5-a94d-e9d49e7266c6" containerName="collect-profiles" Oct 05 09:45:17 crc kubenswrapper[4846]: I1005 09:45:17.316920 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="fac403c1-c8cd-4675-ba46-59900a7cc18f" containerName="tempest-tests-tempest-tests-runner" Oct 05 09:45:17 crc kubenswrapper[4846]: I1005 09:45:17.318033 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 05 09:45:17 crc kubenswrapper[4846]: I1005 09:45:17.323502 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-4c82f" Oct 05 09:45:17 crc kubenswrapper[4846]: I1005 09:45:17.323877 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Oct 05 09:45:17 crc kubenswrapper[4846]: I1005 09:45:17.463257 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gzz7z\" (UniqueName: \"kubernetes.io/projected/a936c21f-704d-4e2d-bbb5-da92c2e8666b-kube-api-access-gzz7z\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"a936c21f-704d-4e2d-bbb5-da92c2e8666b\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 05 09:45:17 crc kubenswrapper[4846]: I1005 09:45:17.464167 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"a936c21f-704d-4e2d-bbb5-da92c2e8666b\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 05 09:45:17 crc kubenswrapper[4846]: I1005 09:45:17.566880 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gzz7z\" (UniqueName: \"kubernetes.io/projected/a936c21f-704d-4e2d-bbb5-da92c2e8666b-kube-api-access-gzz7z\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"a936c21f-704d-4e2d-bbb5-da92c2e8666b\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 05 09:45:17 crc kubenswrapper[4846]: I1005 09:45:17.567032 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"a936c21f-704d-4e2d-bbb5-da92c2e8666b\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 05 09:45:17 crc kubenswrapper[4846]: I1005 09:45:17.567753 4846 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"a936c21f-704d-4e2d-bbb5-da92c2e8666b\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 05 09:45:17 crc kubenswrapper[4846]: I1005 09:45:17.603149 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gzz7z\" (UniqueName: \"kubernetes.io/projected/a936c21f-704d-4e2d-bbb5-da92c2e8666b-kube-api-access-gzz7z\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"a936c21f-704d-4e2d-bbb5-da92c2e8666b\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 05 09:45:17 crc kubenswrapper[4846]: I1005 09:45:17.620824 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"a936c21f-704d-4e2d-bbb5-da92c2e8666b\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 05 09:45:17 crc kubenswrapper[4846]: I1005 09:45:17.659967 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 05 09:45:18 crc kubenswrapper[4846]: I1005 09:45:18.132294 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Oct 05 09:45:18 crc kubenswrapper[4846]: I1005 09:45:18.922830 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"a936c21f-704d-4e2d-bbb5-da92c2e8666b","Type":"ContainerStarted","Data":"3598298fbc8c8c623509325f6ba2f80bb4b568cc408e176670760042e0279671"} Oct 05 09:45:20 crc kubenswrapper[4846]: I1005 09:45:20.946017 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"a936c21f-704d-4e2d-bbb5-da92c2e8666b","Type":"ContainerStarted","Data":"2db336c2d90c32e641c6bafec507826cb88c10207ad4181256ad95d0b3f317cb"} Oct 05 09:45:20 crc kubenswrapper[4846]: I1005 09:45:20.970050 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" podStartSLOduration=2.94367146 podStartE2EDuration="3.970030467s" podCreationTimestamp="2025-10-05 09:45:17 +0000 UTC" firstStartedPulling="2025-10-05 09:45:18.709025298 +0000 UTC m=+10640.949878113" lastFinishedPulling="2025-10-05 09:45:19.735384355 +0000 UTC m=+10641.976237120" observedRunningTime="2025-10-05 09:45:20.965836297 +0000 UTC m=+10643.206689082" watchObservedRunningTime="2025-10-05 09:45:20.970030467 +0000 UTC m=+10643.210883252" Oct 05 09:45:25 crc kubenswrapper[4846]: I1005 09:45:25.497938 4846 scope.go:117] "RemoveContainer" containerID="43e1f18d9b5aea38dd1527fbb4fc557f1b37c59dfd2bdc2019c291047f4244b8" Oct 05 09:45:25 crc kubenswrapper[4846]: E1005 09:45:25.499070 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 09:45:40 crc kubenswrapper[4846]: I1005 09:45:40.498979 4846 scope.go:117] "RemoveContainer" containerID="43e1f18d9b5aea38dd1527fbb4fc557f1b37c59dfd2bdc2019c291047f4244b8" Oct 05 09:45:40 crc kubenswrapper[4846]: E1005 09:45:40.502486 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 09:45:41 crc kubenswrapper[4846]: I1005 09:45:41.508357 4846 scope.go:117] "RemoveContainer" containerID="24d95c8a26fdaa4b41a0694536bd8091b81e5b92cd86f9eaa23e21b644f20d45" Oct 05 09:45:52 crc kubenswrapper[4846]: I1005 09:45:52.498224 4846 scope.go:117] "RemoveContainer" containerID="43e1f18d9b5aea38dd1527fbb4fc557f1b37c59dfd2bdc2019c291047f4244b8" Oct 05 09:45:52 crc kubenswrapper[4846]: E1005 09:45:52.499705 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 09:46:03 crc kubenswrapper[4846]: I1005 09:46:03.499041 4846 scope.go:117] "RemoveContainer" containerID="43e1f18d9b5aea38dd1527fbb4fc557f1b37c59dfd2bdc2019c291047f4244b8" Oct 05 09:46:03 crc kubenswrapper[4846]: E1005 09:46:03.500240 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 09:46:10 crc kubenswrapper[4846]: I1005 09:46:10.060763 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-755h7/must-gather-kl77b"] Oct 05 09:46:10 crc kubenswrapper[4846]: I1005 09:46:10.066630 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-755h7/must-gather-kl77b" Oct 05 09:46:10 crc kubenswrapper[4846]: I1005 09:46:10.068694 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-755h7"/"kube-root-ca.crt" Oct 05 09:46:10 crc kubenswrapper[4846]: I1005 09:46:10.069124 4846 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-755h7"/"default-dockercfg-m5wtz" Oct 05 09:46:10 crc kubenswrapper[4846]: I1005 09:46:10.069513 4846 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-755h7"/"openshift-service-ca.crt" Oct 05 09:46:10 crc kubenswrapper[4846]: I1005 09:46:10.075484 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-755h7/must-gather-kl77b"] Oct 05 09:46:10 crc kubenswrapper[4846]: I1005 09:46:10.131751 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/a2867fd9-a2a3-44f0-849f-8d7ce90e0051-must-gather-output\") pod \"must-gather-kl77b\" (UID: \"a2867fd9-a2a3-44f0-849f-8d7ce90e0051\") " pod="openshift-must-gather-755h7/must-gather-kl77b" Oct 05 09:46:10 crc kubenswrapper[4846]: I1005 09:46:10.131809 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dpzlm\" (UniqueName: \"kubernetes.io/projected/a2867fd9-a2a3-44f0-849f-8d7ce90e0051-kube-api-access-dpzlm\") pod \"must-gather-kl77b\" (UID: \"a2867fd9-a2a3-44f0-849f-8d7ce90e0051\") " pod="openshift-must-gather-755h7/must-gather-kl77b" Oct 05 09:46:10 crc kubenswrapper[4846]: I1005 09:46:10.234605 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/a2867fd9-a2a3-44f0-849f-8d7ce90e0051-must-gather-output\") pod \"must-gather-kl77b\" (UID: \"a2867fd9-a2a3-44f0-849f-8d7ce90e0051\") " pod="openshift-must-gather-755h7/must-gather-kl77b" Oct 05 09:46:10 crc kubenswrapper[4846]: I1005 09:46:10.234668 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dpzlm\" (UniqueName: \"kubernetes.io/projected/a2867fd9-a2a3-44f0-849f-8d7ce90e0051-kube-api-access-dpzlm\") pod \"must-gather-kl77b\" (UID: \"a2867fd9-a2a3-44f0-849f-8d7ce90e0051\") " pod="openshift-must-gather-755h7/must-gather-kl77b" Oct 05 09:46:10 crc kubenswrapper[4846]: I1005 09:46:10.235016 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/a2867fd9-a2a3-44f0-849f-8d7ce90e0051-must-gather-output\") pod \"must-gather-kl77b\" (UID: \"a2867fd9-a2a3-44f0-849f-8d7ce90e0051\") " pod="openshift-must-gather-755h7/must-gather-kl77b" Oct 05 09:46:10 crc kubenswrapper[4846]: I1005 09:46:10.252304 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dpzlm\" (UniqueName: \"kubernetes.io/projected/a2867fd9-a2a3-44f0-849f-8d7ce90e0051-kube-api-access-dpzlm\") pod \"must-gather-kl77b\" (UID: \"a2867fd9-a2a3-44f0-849f-8d7ce90e0051\") " pod="openshift-must-gather-755h7/must-gather-kl77b" Oct 05 09:46:10 crc kubenswrapper[4846]: I1005 09:46:10.391157 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-755h7/must-gather-kl77b" Oct 05 09:46:10 crc kubenswrapper[4846]: I1005 09:46:10.881154 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-755h7/must-gather-kl77b"] Oct 05 09:46:11 crc kubenswrapper[4846]: I1005 09:46:11.570594 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-755h7/must-gather-kl77b" event={"ID":"a2867fd9-a2a3-44f0-849f-8d7ce90e0051","Type":"ContainerStarted","Data":"3afdda9d434b2b489bdfb20d838dbec66aa78657673397f4568bd5f2328bb45d"} Oct 05 09:46:14 crc kubenswrapper[4846]: I1005 09:46:14.497616 4846 scope.go:117] "RemoveContainer" containerID="43e1f18d9b5aea38dd1527fbb4fc557f1b37c59dfd2bdc2019c291047f4244b8" Oct 05 09:46:14 crc kubenswrapper[4846]: E1005 09:46:14.498482 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 09:46:15 crc kubenswrapper[4846]: I1005 09:46:15.613220 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-755h7/must-gather-kl77b" event={"ID":"a2867fd9-a2a3-44f0-849f-8d7ce90e0051","Type":"ContainerStarted","Data":"8070a727bfd7ab91e6005ea90b4578d7eeaaf47f8c870839f0c0dfe7d7e6986f"} Oct 05 09:46:15 crc kubenswrapper[4846]: I1005 09:46:15.613595 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-755h7/must-gather-kl77b" event={"ID":"a2867fd9-a2a3-44f0-849f-8d7ce90e0051","Type":"ContainerStarted","Data":"72f62f82142890e94f0525a8a258b8963373444c09159f16c4037d0be1e62fae"} Oct 05 09:46:15 crc kubenswrapper[4846]: I1005 09:46:15.636090 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-755h7/must-gather-kl77b" podStartSLOduration=1.96589926 podStartE2EDuration="5.636069649s" podCreationTimestamp="2025-10-05 09:46:10 +0000 UTC" firstStartedPulling="2025-10-05 09:46:10.88716083 +0000 UTC m=+10693.128013615" lastFinishedPulling="2025-10-05 09:46:14.557331219 +0000 UTC m=+10696.798184004" observedRunningTime="2025-10-05 09:46:15.633816559 +0000 UTC m=+10697.874669344" watchObservedRunningTime="2025-10-05 09:46:15.636069649 +0000 UTC m=+10697.876922434" Oct 05 09:46:20 crc kubenswrapper[4846]: I1005 09:46:20.561527 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-755h7/crc-debug-msgt2"] Oct 05 09:46:20 crc kubenswrapper[4846]: I1005 09:46:20.565324 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-755h7/crc-debug-msgt2" Oct 05 09:46:20 crc kubenswrapper[4846]: I1005 09:46:20.659812 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tmz5t\" (UniqueName: \"kubernetes.io/projected/eb9f0bd2-8856-4ffe-998a-2de236c90504-kube-api-access-tmz5t\") pod \"crc-debug-msgt2\" (UID: \"eb9f0bd2-8856-4ffe-998a-2de236c90504\") " pod="openshift-must-gather-755h7/crc-debug-msgt2" Oct 05 09:46:20 crc kubenswrapper[4846]: I1005 09:46:20.659987 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/eb9f0bd2-8856-4ffe-998a-2de236c90504-host\") pod \"crc-debug-msgt2\" (UID: \"eb9f0bd2-8856-4ffe-998a-2de236c90504\") " pod="openshift-must-gather-755h7/crc-debug-msgt2" Oct 05 09:46:20 crc kubenswrapper[4846]: I1005 09:46:20.762893 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/eb9f0bd2-8856-4ffe-998a-2de236c90504-host\") pod \"crc-debug-msgt2\" (UID: \"eb9f0bd2-8856-4ffe-998a-2de236c90504\") " pod="openshift-must-gather-755h7/crc-debug-msgt2" Oct 05 09:46:20 crc kubenswrapper[4846]: I1005 09:46:20.763063 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/eb9f0bd2-8856-4ffe-998a-2de236c90504-host\") pod \"crc-debug-msgt2\" (UID: \"eb9f0bd2-8856-4ffe-998a-2de236c90504\") " pod="openshift-must-gather-755h7/crc-debug-msgt2" Oct 05 09:46:20 crc kubenswrapper[4846]: I1005 09:46:20.763346 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tmz5t\" (UniqueName: \"kubernetes.io/projected/eb9f0bd2-8856-4ffe-998a-2de236c90504-kube-api-access-tmz5t\") pod \"crc-debug-msgt2\" (UID: \"eb9f0bd2-8856-4ffe-998a-2de236c90504\") " pod="openshift-must-gather-755h7/crc-debug-msgt2" Oct 05 09:46:20 crc kubenswrapper[4846]: I1005 09:46:20.783621 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tmz5t\" (UniqueName: \"kubernetes.io/projected/eb9f0bd2-8856-4ffe-998a-2de236c90504-kube-api-access-tmz5t\") pod \"crc-debug-msgt2\" (UID: \"eb9f0bd2-8856-4ffe-998a-2de236c90504\") " pod="openshift-must-gather-755h7/crc-debug-msgt2" Oct 05 09:46:20 crc kubenswrapper[4846]: I1005 09:46:20.883008 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-755h7/crc-debug-msgt2" Oct 05 09:46:20 crc kubenswrapper[4846]: I1005 09:46:20.927793 4846 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 05 09:46:21 crc kubenswrapper[4846]: I1005 09:46:21.674306 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-755h7/crc-debug-msgt2" event={"ID":"eb9f0bd2-8856-4ffe-998a-2de236c90504","Type":"ContainerStarted","Data":"2b6951c40e34d3912adbbd34773634318cdfb830af2f93b1a2aa170ae361e9cf"} Oct 05 09:46:29 crc kubenswrapper[4846]: I1005 09:46:29.497655 4846 scope.go:117] "RemoveContainer" containerID="43e1f18d9b5aea38dd1527fbb4fc557f1b37c59dfd2bdc2019c291047f4244b8" Oct 05 09:46:29 crc kubenswrapper[4846]: E1005 09:46:29.498425 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 09:46:32 crc kubenswrapper[4846]: I1005 09:46:32.778689 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-755h7/crc-debug-msgt2" event={"ID":"eb9f0bd2-8856-4ffe-998a-2de236c90504","Type":"ContainerStarted","Data":"4424237257eb8339c7cee2720444adcb8d71efbc263d3fb8d250188251b93605"} Oct 05 09:46:32 crc kubenswrapper[4846]: I1005 09:46:32.792223 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-755h7/crc-debug-msgt2" podStartSLOduration=1.57716534 podStartE2EDuration="12.792201176s" podCreationTimestamp="2025-10-05 09:46:20 +0000 UTC" firstStartedPulling="2025-10-05 09:46:20.927560016 +0000 UTC m=+10703.168412791" lastFinishedPulling="2025-10-05 09:46:32.142595852 +0000 UTC m=+10714.383448627" observedRunningTime="2025-10-05 09:46:32.789819093 +0000 UTC m=+10715.030671868" watchObservedRunningTime="2025-10-05 09:46:32.792201176 +0000 UTC m=+10715.033053951" Oct 05 09:46:41 crc kubenswrapper[4846]: I1005 09:46:41.498281 4846 scope.go:117] "RemoveContainer" containerID="43e1f18d9b5aea38dd1527fbb4fc557f1b37c59dfd2bdc2019c291047f4244b8" Oct 05 09:46:41 crc kubenswrapper[4846]: E1005 09:46:41.499719 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 09:46:41 crc kubenswrapper[4846]: I1005 09:46:41.607860 4846 scope.go:117] "RemoveContainer" containerID="976deb6f29fa64d7bb6c475baa46525ddd75892876d85d4a9dd46caa95c4221f" Oct 05 09:46:41 crc kubenswrapper[4846]: I1005 09:46:41.668048 4846 scope.go:117] "RemoveContainer" containerID="85cebeed4e8c7ed07c23b2eeedcd60a7b678ea0c55da62c9f586c9ab4a03cca8" Oct 05 09:46:42 crc kubenswrapper[4846]: I1005 09:46:42.141965 4846 scope.go:117] "RemoveContainer" containerID="dfdd50cf82cbf3b969b8157a740c593060da08fb94b37b19f3f5b8472fdcda08" Oct 05 09:46:53 crc kubenswrapper[4846]: I1005 09:46:53.498309 4846 scope.go:117] "RemoveContainer" containerID="43e1f18d9b5aea38dd1527fbb4fc557f1b37c59dfd2bdc2019c291047f4244b8" Oct 05 09:46:53 crc kubenswrapper[4846]: E1005 09:46:53.499332 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 09:47:08 crc kubenswrapper[4846]: I1005 09:47:08.511362 4846 scope.go:117] "RemoveContainer" containerID="43e1f18d9b5aea38dd1527fbb4fc557f1b37c59dfd2bdc2019c291047f4244b8" Oct 05 09:47:08 crc kubenswrapper[4846]: E1005 09:47:08.512430 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 09:47:21 crc kubenswrapper[4846]: I1005 09:47:21.497522 4846 scope.go:117] "RemoveContainer" containerID="43e1f18d9b5aea38dd1527fbb4fc557f1b37c59dfd2bdc2019c291047f4244b8" Oct 05 09:47:21 crc kubenswrapper[4846]: E1005 09:47:21.498172 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 09:47:36 crc kubenswrapper[4846]: I1005 09:47:36.500968 4846 scope.go:117] "RemoveContainer" containerID="43e1f18d9b5aea38dd1527fbb4fc557f1b37c59dfd2bdc2019c291047f4244b8" Oct 05 09:47:36 crc kubenswrapper[4846]: E1005 09:47:36.503835 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 09:47:48 crc kubenswrapper[4846]: I1005 09:47:48.506170 4846 scope.go:117] "RemoveContainer" containerID="43e1f18d9b5aea38dd1527fbb4fc557f1b37c59dfd2bdc2019c291047f4244b8" Oct 05 09:47:48 crc kubenswrapper[4846]: E1005 09:47:48.506873 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 09:47:49 crc kubenswrapper[4846]: I1005 09:47:49.994107 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_alertmanager-metric-storage-0_e459c893-d9e7-44de-a985-49b29f59f459/init-config-reloader/0.log" Oct 05 09:47:50 crc kubenswrapper[4846]: I1005 09:47:50.175619 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_alertmanager-metric-storage-0_e459c893-d9e7-44de-a985-49b29f59f459/init-config-reloader/0.log" Oct 05 09:47:50 crc kubenswrapper[4846]: I1005 09:47:50.303801 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_alertmanager-metric-storage-0_e459c893-d9e7-44de-a985-49b29f59f459/alertmanager/0.log" Oct 05 09:47:50 crc kubenswrapper[4846]: I1005 09:47:50.359618 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_alertmanager-metric-storage-0_e459c893-d9e7-44de-a985-49b29f59f459/config-reloader/0.log" Oct 05 09:47:50 crc kubenswrapper[4846]: I1005 09:47:50.570245 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_aodh-0_c2926ee6-0f75-41e1-8140-cd2a2bfbad2b/aodh-api/0.log" Oct 05 09:47:50 crc kubenswrapper[4846]: I1005 09:47:50.720282 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_aodh-0_c2926ee6-0f75-41e1-8140-cd2a2bfbad2b/aodh-evaluator/0.log" Oct 05 09:47:50 crc kubenswrapper[4846]: I1005 09:47:50.815758 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_aodh-0_c2926ee6-0f75-41e1-8140-cd2a2bfbad2b/aodh-listener/0.log" Oct 05 09:47:50 crc kubenswrapper[4846]: I1005 09:47:50.920722 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_aodh-0_c2926ee6-0f75-41e1-8140-cd2a2bfbad2b/aodh-notifier/0.log" Oct 05 09:47:51 crc kubenswrapper[4846]: I1005 09:47:51.097141 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-5d595db478-zkhxm_8dd27928-1e7a-42b1-ae7f-7957e42cf4c1/barbican-api/0.log" Oct 05 09:47:51 crc kubenswrapper[4846]: I1005 09:47:51.279043 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-5d595db478-zkhxm_8dd27928-1e7a-42b1-ae7f-7957e42cf4c1/barbican-api-log/0.log" Oct 05 09:47:51 crc kubenswrapper[4846]: I1005 09:47:51.483735 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-595f76b44d-4pljq_f93477d4-a2ba-4b9c-bfa9-2db3393596e5/barbican-keystone-listener/0.log" Oct 05 09:47:51 crc kubenswrapper[4846]: I1005 09:47:51.863784 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-595f76b44d-4pljq_f93477d4-a2ba-4b9c-bfa9-2db3393596e5/barbican-keystone-listener-log/0.log" Oct 05 09:47:51 crc kubenswrapper[4846]: I1005 09:47:51.976156 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-67855dd78c-jwwd4_43fb6020-9a27-4839-9a56-a23a03d21b84/barbican-worker/0.log" Oct 05 09:47:52 crc kubenswrapper[4846]: I1005 09:47:52.030004 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-67855dd78c-jwwd4_43fb6020-9a27-4839-9a56-a23a03d21b84/barbican-worker-log/0.log" Oct 05 09:47:52 crc kubenswrapper[4846]: I1005 09:47:52.212623 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-openstack-openstack-cell1-rxf6k_f55e5fdc-aff8-4ebe-afa7-a570e18bf335/bootstrap-openstack-openstack-cell1/0.log" Oct 05 09:47:52 crc kubenswrapper[4846]: I1005 09:47:52.404711 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_31f6d8fa-326e-46cb-83ca-9294b1e20903/ceilometer-notification-agent/0.log" Oct 05 09:47:52 crc kubenswrapper[4846]: I1005 09:47:52.432764 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_31f6d8fa-326e-46cb-83ca-9294b1e20903/ceilometer-central-agent/0.log" Oct 05 09:47:52 crc kubenswrapper[4846]: I1005 09:47:52.535350 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_31f6d8fa-326e-46cb-83ca-9294b1e20903/proxy-httpd/0.log" Oct 05 09:47:52 crc kubenswrapper[4846]: I1005 09:47:52.575147 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_31f6d8fa-326e-46cb-83ca-9294b1e20903/sg-core/0.log" Oct 05 09:47:52 crc kubenswrapper[4846]: I1005 09:47:52.785228 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_8d38c463-603e-4e77-9af8-d9329875a9a3/cinder-api-log/0.log" Oct 05 09:47:52 crc kubenswrapper[4846]: I1005 09:47:52.805518 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_8d38c463-603e-4e77-9af8-d9329875a9a3/cinder-api/0.log" Oct 05 09:47:53 crc kubenswrapper[4846]: I1005 09:47:53.030046 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_70e68e86-4002-41af-ac0c-7426ae1b96d3/cinder-scheduler/0.log" Oct 05 09:47:53 crc kubenswrapper[4846]: I1005 09:47:53.055504 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_70e68e86-4002-41af-ac0c-7426ae1b96d3/probe/0.log" Oct 05 09:47:53 crc kubenswrapper[4846]: I1005 09:47:53.239595 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-openstack-openstack-cell1-l9jt5_58a90b64-2441-4ccf-8655-a5a582dee3cc/configure-network-openstack-openstack-cell1/0.log" Oct 05 09:47:53 crc kubenswrapper[4846]: I1005 09:47:53.351768 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-openstack-openstack-cell1-wzg2n_c9d6416a-ce5b-41a2-b49b-a67573544f06/configure-os-openstack-openstack-cell1/0.log" Oct 05 09:47:53 crc kubenswrapper[4846]: I1005 09:47:53.544413 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-8869d4b55-9z82r_a17b7828-72b0-4715-9aa6-484f29bc3125/init/0.log" Oct 05 09:47:53 crc kubenswrapper[4846]: I1005 09:47:53.760455 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-8869d4b55-9z82r_a17b7828-72b0-4715-9aa6-484f29bc3125/dnsmasq-dns/0.log" Oct 05 09:47:53 crc kubenswrapper[4846]: I1005 09:47:53.763351 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-8869d4b55-9z82r_a17b7828-72b0-4715-9aa6-484f29bc3125/init/0.log" Oct 05 09:47:53 crc kubenswrapper[4846]: I1005 09:47:53.951934 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_download-cache-openstack-openstack-cell1-w242f_7ecab93f-583b-4a2f-b000-10f96324d002/download-cache-openstack-openstack-cell1/0.log" Oct 05 09:47:54 crc kubenswrapper[4846]: I1005 09:47:54.059442 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_a913b914-fe8a-4d4a-9969-7f4df6313f02/glance-httpd/0.log" Oct 05 09:47:54 crc kubenswrapper[4846]: I1005 09:47:54.140967 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_a913b914-fe8a-4d4a-9969-7f4df6313f02/glance-log/0.log" Oct 05 09:47:54 crc kubenswrapper[4846]: I1005 09:47:54.241791 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_25d587cf-9f44-4c05-9ff9-c427f8337de5/glance-httpd/0.log" Oct 05 09:47:54 crc kubenswrapper[4846]: I1005 09:47:54.324602 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_25d587cf-9f44-4c05-9ff9-c427f8337de5/glance-log/0.log" Oct 05 09:47:55 crc kubenswrapper[4846]: I1005 09:47:55.081383 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_heat-engine-54b8fb86bb-q8qps_28512aa9-3230-44b3-a1f2-55247d128483/heat-engine/0.log" Oct 05 09:47:55 crc kubenswrapper[4846]: I1005 09:47:55.169122 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_heat-api-69bf66cb7b-c4q2z_2db59028-72ef-4d73-84b1-d6bc6af749a5/heat-api/0.log" Oct 05 09:47:55 crc kubenswrapper[4846]: I1005 09:47:55.356192 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_heat-cfnapi-77cc459545-h5rt6_ed7a6ddf-cae5-48a3-931f-85c7287d77d8/heat-cfnapi/0.log" Oct 05 09:47:55 crc kubenswrapper[4846]: I1005 09:47:55.465808 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-85d5597856-frbmx_82fe0928-8008-48ce-8993-41ff90d89bcd/horizon/0.log" Oct 05 09:47:55 crc kubenswrapper[4846]: I1005 09:47:55.699929 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-openstack-openstack-cell1-45cc8_dda9bfde-287c-49de-aacc-57e6b1ef2950/install-certs-openstack-openstack-cell1/0.log" Oct 05 09:47:55 crc kubenswrapper[4846]: I1005 09:47:55.915448 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-openstack-openstack-cell1-l4pz6_b4fc0719-2125-4505-bd67-3052dcfd3f1a/install-os-openstack-openstack-cell1/0.log" Oct 05 09:47:56 crc kubenswrapper[4846]: I1005 09:47:55.991604 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-85d5597856-frbmx_82fe0928-8008-48ce-8993-41ff90d89bcd/horizon-log/0.log" Oct 05 09:47:56 crc kubenswrapper[4846]: I1005 09:47:56.309735 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29327581-f9wqf_9cb75497-a34f-4ff3-8140-4be43217b17f/keystone-cron/0.log" Oct 05 09:47:56 crc kubenswrapper[4846]: I1005 09:47:56.524758 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_f89bcf16-c362-44a0-9869-551cea0beacc/kube-state-metrics/0.log" Oct 05 09:47:56 crc kubenswrapper[4846]: I1005 09:47:56.736976 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-openstack-openstack-cell1-2f7vt_2f3f1db8-f42b-4f2b-aeb3-05b9523aa4ea/libvirt-openstack-openstack-cell1/0.log" Oct 05 09:47:56 crc kubenswrapper[4846]: I1005 09:47:56.824928 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-575454b8f7-64pz9_617bdbc8-22e4-43b8-b344-7a76842c40e8/keystone-api/0.log" Oct 05 09:47:57 crc kubenswrapper[4846]: I1005 09:47:57.243077 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-dhcp-openstack-openstack-cell1-hl2xj_3d69d7c2-0c3a-4850-9ad0-e9654b3e8166/neutron-dhcp-openstack-openstack-cell1/0.log" Oct 05 09:47:57 crc kubenswrapper[4846]: I1005 09:47:57.828292 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-fdc58b465-qpttk_e8c40de6-3f63-4f56-a480-0fbf007507a5/neutron-httpd/0.log" Oct 05 09:47:58 crc kubenswrapper[4846]: I1005 09:47:58.038535 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-fdc58b465-qpttk_e8c40de6-3f63-4f56-a480-0fbf007507a5/neutron-api/0.log" Oct 05 09:47:58 crc kubenswrapper[4846]: I1005 09:47:58.236258 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-openstack-openstack-cell1-xnhm4_22023b6c-2796-4c84-96ad-105022a558fb/neutron-metadata-openstack-openstack-cell1/0.log" Oct 05 09:47:58 crc kubenswrapper[4846]: I1005 09:47:58.496753 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-sriov-openstack-openstack-cell1-ds9h4_e844e15c-7dd3-4056-8727-a24c9b65dbd8/neutron-sriov-openstack-openstack-cell1/0.log" Oct 05 09:47:59 crc kubenswrapper[4846]: I1005 09:47:59.053114 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_563f91c3-63c3-4dc0-886a-e6a1a8a5672e/nova-api-log/0.log" Oct 05 09:47:59 crc kubenswrapper[4846]: I1005 09:47:59.309131 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_563f91c3-63c3-4dc0-886a-e6a1a8a5672e/nova-api-api/0.log" Oct 05 09:47:59 crc kubenswrapper[4846]: I1005 09:47:59.553503 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_74d21914-81a3-47ce-8a85-3d52430d9422/nova-cell0-conductor-conductor/0.log" Oct 05 09:47:59 crc kubenswrapper[4846]: I1005 09:47:59.825645 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_0fccb0ce-ab1d-495e-b080-8c1310a2297a/nova-cell1-conductor-conductor/0.log" Oct 05 09:48:00 crc kubenswrapper[4846]: I1005 09:48:00.103923 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_78d644fc-c8ab-4b0f-bdf4-772bdca9adf3/nova-cell1-novncproxy-novncproxy/0.log" Oct 05 09:48:00 crc kubenswrapper[4846]: I1005 09:48:00.518680 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celldnhqr_4e445bd2-1270-4c05-9e0b-bcdbec35fa74/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell1/0.log" Oct 05 09:48:00 crc kubenswrapper[4846]: I1005 09:48:00.910052 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellkhtf8_dd84ed43-c477-4b83-a735-142efa0d85ea/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell1/0.log" Oct 05 09:48:01 crc kubenswrapper[4846]: I1005 09:48:01.269890 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-openstack-openstack-cell1-kjczt_6b81e192-94b8-4a4b-918b-feab5d5a7860/nova-cell1-openstack-openstack-cell1/0.log" Oct 05 09:48:01 crc kubenswrapper[4846]: I1005 09:48:01.497561 4846 scope.go:117] "RemoveContainer" containerID="43e1f18d9b5aea38dd1527fbb4fc557f1b37c59dfd2bdc2019c291047f4244b8" Oct 05 09:48:01 crc kubenswrapper[4846]: I1005 09:48:01.538121 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_95c1c757-bd73-4e9a-9d5e-59c5970491cf/nova-metadata-log/0.log" Oct 05 09:48:02 crc kubenswrapper[4846]: I1005 09:48:02.162835 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_81e95869-36b9-41c5-a1cc-e3cea1471c26/nova-scheduler-scheduler/0.log" Oct 05 09:48:02 crc kubenswrapper[4846]: I1005 09:48:02.557785 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_95c1c757-bd73-4e9a-9d5e-59c5970491cf/nova-metadata-metadata/0.log" Oct 05 09:48:02 crc kubenswrapper[4846]: I1005 09:48:02.707399 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" event={"ID":"ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7","Type":"ContainerStarted","Data":"71dd5b6fb2a9c6f83610e28480b9546168c48746147b1078114492dbe0d834b1"} Oct 05 09:48:02 crc kubenswrapper[4846]: I1005 09:48:02.727803 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_643edc27-5b17-45e9-9286-f74591843690/mysql-bootstrap/0.log" Oct 05 09:48:02 crc kubenswrapper[4846]: I1005 09:48:02.844535 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_643edc27-5b17-45e9-9286-f74591843690/mysql-bootstrap/0.log" Oct 05 09:48:02 crc kubenswrapper[4846]: I1005 09:48:02.927058 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_643edc27-5b17-45e9-9286-f74591843690/galera/0.log" Oct 05 09:48:03 crc kubenswrapper[4846]: I1005 09:48:03.160982 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_771f44c5-2cc5-431e-9569-eb7da776756d/mysql-bootstrap/0.log" Oct 05 09:48:03 crc kubenswrapper[4846]: I1005 09:48:03.292613 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_771f44c5-2cc5-431e-9569-eb7da776756d/mysql-bootstrap/0.log" Oct 05 09:48:03 crc kubenswrapper[4846]: I1005 09:48:03.411963 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_771f44c5-2cc5-431e-9569-eb7da776756d/galera/0.log" Oct 05 09:48:03 crc kubenswrapper[4846]: I1005 09:48:03.611214 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_e45f7b43-2271-4c62-a0b5-a66462317e84/openstackclient/0.log" Oct 05 09:48:03 crc kubenswrapper[4846]: I1005 09:48:03.840193 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_4bbb24b3-544b-47db-9058-47aa640afa64/openstack-network-exporter/0.log" Oct 05 09:48:03 crc kubenswrapper[4846]: I1005 09:48:03.985570 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_4bbb24b3-544b-47db-9058-47aa640afa64/ovn-northd/0.log" Oct 05 09:48:04 crc kubenswrapper[4846]: I1005 09:48:04.275489 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-openstack-openstack-cell1-5bb6g_b6c4e5a0-3410-406d-bb32-6046cf340cf1/ovn-openstack-openstack-cell1/0.log" Oct 05 09:48:04 crc kubenswrapper[4846]: I1005 09:48:04.482862 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_52e7bf3e-f881-4ff2-aae6-40aed0788d54/openstack-network-exporter/0.log" Oct 05 09:48:04 crc kubenswrapper[4846]: I1005 09:48:04.621198 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_52e7bf3e-f881-4ff2-aae6-40aed0788d54/ovsdbserver-nb/0.log" Oct 05 09:48:04 crc kubenswrapper[4846]: I1005 09:48:04.842351 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-1_f46c4128-97fe-45d4-ac45-dc6e018294ba/openstack-network-exporter/0.log" Oct 05 09:48:04 crc kubenswrapper[4846]: I1005 09:48:04.907042 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-1_f46c4128-97fe-45d4-ac45-dc6e018294ba/ovsdbserver-nb/0.log" Oct 05 09:48:05 crc kubenswrapper[4846]: I1005 09:48:05.100958 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-2_fcda94d9-8d6f-4aa3-a60b-2ed20e1b3552/openstack-network-exporter/0.log" Oct 05 09:48:05 crc kubenswrapper[4846]: I1005 09:48:05.255873 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-2_fcda94d9-8d6f-4aa3-a60b-2ed20e1b3552/ovsdbserver-nb/0.log" Oct 05 09:48:05 crc kubenswrapper[4846]: I1005 09:48:05.431977 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_cfc772ff-5d34-4838-8077-aa4da607881d/openstack-network-exporter/0.log" Oct 05 09:48:05 crc kubenswrapper[4846]: I1005 09:48:05.441622 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_8b5a8f28-85f4-41e5-8da2-f7fe7a0d1220/memcached/0.log" Oct 05 09:48:05 crc kubenswrapper[4846]: I1005 09:48:05.551070 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_cfc772ff-5d34-4838-8077-aa4da607881d/ovsdbserver-sb/0.log" Oct 05 09:48:05 crc kubenswrapper[4846]: I1005 09:48:05.627047 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-1_cb861427-61ff-4885-bfc3-af718ddd3b7e/openstack-network-exporter/0.log" Oct 05 09:48:05 crc kubenswrapper[4846]: I1005 09:48:05.731077 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-1_cb861427-61ff-4885-bfc3-af718ddd3b7e/ovsdbserver-sb/0.log" Oct 05 09:48:05 crc kubenswrapper[4846]: I1005 09:48:05.823601 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-2_6d99cbc7-c521-437e-9225-eaf5d0abdc51/openstack-network-exporter/0.log" Oct 05 09:48:05 crc kubenswrapper[4846]: I1005 09:48:05.935652 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-2_6d99cbc7-c521-437e-9225-eaf5d0abdc51/ovsdbserver-sb/0.log" Oct 05 09:48:06 crc kubenswrapper[4846]: I1005 09:48:06.223773 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-6fb6f8b89d-dvr8k_fbf245a7-f5e8-4ddf-89eb-737454886513/placement-api/0.log" Oct 05 09:48:06 crc kubenswrapper[4846]: I1005 09:48:06.322689 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-6fb6f8b89d-dvr8k_fbf245a7-f5e8-4ddf-89eb-737454886513/placement-log/0.log" Oct 05 09:48:06 crc kubenswrapper[4846]: I1005 09:48:06.410480 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_pre-adoption-validation-openstack-pre-adoption-openstack-c672lz_4ca8fa61-f7ee-4f10-b047-589a5558c05b/pre-adoption-validation-openstack-pre-adoption-openstack-cell1/0.log" Oct 05 09:48:06 crc kubenswrapper[4846]: I1005 09:48:06.569449 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_8e051aea-4a36-4dd1-9d60-60842891f011/init-config-reloader/0.log" Oct 05 09:48:06 crc kubenswrapper[4846]: I1005 09:48:06.714460 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_8e051aea-4a36-4dd1-9d60-60842891f011/init-config-reloader/0.log" Oct 05 09:48:06 crc kubenswrapper[4846]: I1005 09:48:06.738774 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_8e051aea-4a36-4dd1-9d60-60842891f011/config-reloader/0.log" Oct 05 09:48:06 crc kubenswrapper[4846]: I1005 09:48:06.763952 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_8e051aea-4a36-4dd1-9d60-60842891f011/prometheus/0.log" Oct 05 09:48:06 crc kubenswrapper[4846]: I1005 09:48:06.879502 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_8e051aea-4a36-4dd1-9d60-60842891f011/thanos-sidecar/0.log" Oct 05 09:48:06 crc kubenswrapper[4846]: I1005 09:48:06.933672 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_df56f237-d65a-4dbf-a3cf-ee1c496d7ce4/setup-container/0.log" Oct 05 09:48:07 crc kubenswrapper[4846]: I1005 09:48:07.125367 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_df56f237-d65a-4dbf-a3cf-ee1c496d7ce4/setup-container/0.log" Oct 05 09:48:07 crc kubenswrapper[4846]: I1005 09:48:07.150538 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_df56f237-d65a-4dbf-a3cf-ee1c496d7ce4/rabbitmq/0.log" Oct 05 09:48:07 crc kubenswrapper[4846]: I1005 09:48:07.343494 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_a8b2a717-7ead-4166-b460-3d9de6f3acec/setup-container/0.log" Oct 05 09:48:07 crc kubenswrapper[4846]: I1005 09:48:07.515674 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_a8b2a717-7ead-4166-b460-3d9de6f3acec/setup-container/0.log" Oct 05 09:48:07 crc kubenswrapper[4846]: I1005 09:48:07.550668 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_a8b2a717-7ead-4166-b460-3d9de6f3acec/rabbitmq/0.log" Oct 05 09:48:07 crc kubenswrapper[4846]: I1005 09:48:07.713234 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-openstack-openstack-cell1-fdct8_1a2802cf-4055-4f27-af97-e9dbd5266a3d/reboot-os-openstack-openstack-cell1/0.log" Oct 05 09:48:07 crc kubenswrapper[4846]: I1005 09:48:07.835688 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-openstack-openstack-cell1-m2snw_fe44fec2-66d0-4b82-909b-3a27f42340db/run-os-openstack-openstack-cell1/0.log" Oct 05 09:48:08 crc kubenswrapper[4846]: I1005 09:48:08.004552 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-openstack-x82zm_a87e92c9-28ef-42d1-a662-dea1a080b896/ssh-known-hosts-openstack/0.log" Oct 05 09:48:08 crc kubenswrapper[4846]: I1005 09:48:08.245392 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-5b5cc989cf-btgr5_d3b7b761-5203-4ed5-8fc1-4f4a3d572592/proxy-server/0.log" Oct 05 09:48:08 crc kubenswrapper[4846]: I1005 09:48:08.420692 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-ring-rebalance-vr9g5_77e9ec78-ecc3-4953-9c33-f04b3f50b890/swift-ring-rebalance/0.log" Oct 05 09:48:08 crc kubenswrapper[4846]: I1005 09:48:08.443399 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-5b5cc989cf-btgr5_d3b7b761-5203-4ed5-8fc1-4f4a3d572592/proxy-httpd/0.log" Oct 05 09:48:08 crc kubenswrapper[4846]: I1005 09:48:08.674355 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-openstack-openstack-cell1-fl275_8b360825-bbe2-4522-b0b0-1f0bf6c4bc9f/telemetry-openstack-openstack-cell1/0.log" Oct 05 09:48:08 crc kubenswrapper[4846]: I1005 09:48:08.852133 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest_fac403c1-c8cd-4675-ba46-59900a7cc18f/tempest-tests-tempest-tests-runner/0.log" Oct 05 09:48:08 crc kubenswrapper[4846]: I1005 09:48:08.926017 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-tempest-tempest-tests-tempest_a936c21f-704d-4e2d-bbb5-da92c2e8666b/test-operator-logs-container/0.log" Oct 05 09:48:09 crc kubenswrapper[4846]: I1005 09:48:09.119879 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tripleo-cleanup-tripleo-cleanup-openstack-cell1-ptzpk_19baf6c0-a301-49fe-be7b-6682904106b9/tripleo-cleanup-tripleo-cleanup-openstack-cell1/0.log" Oct 05 09:48:09 crc kubenswrapper[4846]: I1005 09:48:09.292613 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-openstack-openstack-cell1-fkcjx_c7fe64f3-1811-49fb-bc47-c9a5a1060954/validate-network-openstack-openstack-cell1/0.log" Oct 05 09:49:02 crc kubenswrapper[4846]: I1005 09:49:02.408306 4846 generic.go:334] "Generic (PLEG): container finished" podID="eb9f0bd2-8856-4ffe-998a-2de236c90504" containerID="4424237257eb8339c7cee2720444adcb8d71efbc263d3fb8d250188251b93605" exitCode=0 Oct 05 09:49:02 crc kubenswrapper[4846]: I1005 09:49:02.408501 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-755h7/crc-debug-msgt2" event={"ID":"eb9f0bd2-8856-4ffe-998a-2de236c90504","Type":"ContainerDied","Data":"4424237257eb8339c7cee2720444adcb8d71efbc263d3fb8d250188251b93605"} Oct 05 09:49:03 crc kubenswrapper[4846]: I1005 09:49:03.556914 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-755h7/crc-debug-msgt2" Oct 05 09:49:03 crc kubenswrapper[4846]: I1005 09:49:03.606608 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-755h7/crc-debug-msgt2"] Oct 05 09:49:03 crc kubenswrapper[4846]: I1005 09:49:03.618698 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-755h7/crc-debug-msgt2"] Oct 05 09:49:03 crc kubenswrapper[4846]: I1005 09:49:03.670955 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tmz5t\" (UniqueName: \"kubernetes.io/projected/eb9f0bd2-8856-4ffe-998a-2de236c90504-kube-api-access-tmz5t\") pod \"eb9f0bd2-8856-4ffe-998a-2de236c90504\" (UID: \"eb9f0bd2-8856-4ffe-998a-2de236c90504\") " Oct 05 09:49:03 crc kubenswrapper[4846]: I1005 09:49:03.671131 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/eb9f0bd2-8856-4ffe-998a-2de236c90504-host\") pod \"eb9f0bd2-8856-4ffe-998a-2de236c90504\" (UID: \"eb9f0bd2-8856-4ffe-998a-2de236c90504\") " Oct 05 09:49:03 crc kubenswrapper[4846]: I1005 09:49:03.671306 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/eb9f0bd2-8856-4ffe-998a-2de236c90504-host" (OuterVolumeSpecName: "host") pod "eb9f0bd2-8856-4ffe-998a-2de236c90504" (UID: "eb9f0bd2-8856-4ffe-998a-2de236c90504"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 05 09:49:03 crc kubenswrapper[4846]: I1005 09:49:03.671944 4846 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/eb9f0bd2-8856-4ffe-998a-2de236c90504-host\") on node \"crc\" DevicePath \"\"" Oct 05 09:49:03 crc kubenswrapper[4846]: I1005 09:49:03.683226 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eb9f0bd2-8856-4ffe-998a-2de236c90504-kube-api-access-tmz5t" (OuterVolumeSpecName: "kube-api-access-tmz5t") pod "eb9f0bd2-8856-4ffe-998a-2de236c90504" (UID: "eb9f0bd2-8856-4ffe-998a-2de236c90504"). InnerVolumeSpecName "kube-api-access-tmz5t". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 09:49:03 crc kubenswrapper[4846]: I1005 09:49:03.774209 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tmz5t\" (UniqueName: \"kubernetes.io/projected/eb9f0bd2-8856-4ffe-998a-2de236c90504-kube-api-access-tmz5t\") on node \"crc\" DevicePath \"\"" Oct 05 09:49:04 crc kubenswrapper[4846]: I1005 09:49:04.434069 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2b6951c40e34d3912adbbd34773634318cdfb830af2f93b1a2aa170ae361e9cf" Oct 05 09:49:04 crc kubenswrapper[4846]: I1005 09:49:04.434114 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-755h7/crc-debug-msgt2" Oct 05 09:49:04 crc kubenswrapper[4846]: I1005 09:49:04.508298 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="eb9f0bd2-8856-4ffe-998a-2de236c90504" path="/var/lib/kubelet/pods/eb9f0bd2-8856-4ffe-998a-2de236c90504/volumes" Oct 05 09:49:04 crc kubenswrapper[4846]: I1005 09:49:04.854771 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-755h7/crc-debug-dxxdp"] Oct 05 09:49:04 crc kubenswrapper[4846]: E1005 09:49:04.855379 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eb9f0bd2-8856-4ffe-998a-2de236c90504" containerName="container-00" Oct 05 09:49:04 crc kubenswrapper[4846]: I1005 09:49:04.855400 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="eb9f0bd2-8856-4ffe-998a-2de236c90504" containerName="container-00" Oct 05 09:49:04 crc kubenswrapper[4846]: I1005 09:49:04.855761 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="eb9f0bd2-8856-4ffe-998a-2de236c90504" containerName="container-00" Oct 05 09:49:04 crc kubenswrapper[4846]: I1005 09:49:04.856828 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-755h7/crc-debug-dxxdp" Oct 05 09:49:05 crc kubenswrapper[4846]: I1005 09:49:05.006153 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hcjnz\" (UniqueName: \"kubernetes.io/projected/0d325705-466c-4f52-bb47-0566c8bdbdbf-kube-api-access-hcjnz\") pod \"crc-debug-dxxdp\" (UID: \"0d325705-466c-4f52-bb47-0566c8bdbdbf\") " pod="openshift-must-gather-755h7/crc-debug-dxxdp" Oct 05 09:49:05 crc kubenswrapper[4846]: I1005 09:49:05.006330 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/0d325705-466c-4f52-bb47-0566c8bdbdbf-host\") pod \"crc-debug-dxxdp\" (UID: \"0d325705-466c-4f52-bb47-0566c8bdbdbf\") " pod="openshift-must-gather-755h7/crc-debug-dxxdp" Oct 05 09:49:05 crc kubenswrapper[4846]: I1005 09:49:05.109074 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hcjnz\" (UniqueName: \"kubernetes.io/projected/0d325705-466c-4f52-bb47-0566c8bdbdbf-kube-api-access-hcjnz\") pod \"crc-debug-dxxdp\" (UID: \"0d325705-466c-4f52-bb47-0566c8bdbdbf\") " pod="openshift-must-gather-755h7/crc-debug-dxxdp" Oct 05 09:49:05 crc kubenswrapper[4846]: I1005 09:49:05.109622 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/0d325705-466c-4f52-bb47-0566c8bdbdbf-host\") pod \"crc-debug-dxxdp\" (UID: \"0d325705-466c-4f52-bb47-0566c8bdbdbf\") " pod="openshift-must-gather-755h7/crc-debug-dxxdp" Oct 05 09:49:05 crc kubenswrapper[4846]: I1005 09:49:05.109760 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/0d325705-466c-4f52-bb47-0566c8bdbdbf-host\") pod \"crc-debug-dxxdp\" (UID: \"0d325705-466c-4f52-bb47-0566c8bdbdbf\") " pod="openshift-must-gather-755h7/crc-debug-dxxdp" Oct 05 09:49:05 crc kubenswrapper[4846]: I1005 09:49:05.141166 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hcjnz\" (UniqueName: \"kubernetes.io/projected/0d325705-466c-4f52-bb47-0566c8bdbdbf-kube-api-access-hcjnz\") pod \"crc-debug-dxxdp\" (UID: \"0d325705-466c-4f52-bb47-0566c8bdbdbf\") " pod="openshift-must-gather-755h7/crc-debug-dxxdp" Oct 05 09:49:05 crc kubenswrapper[4846]: I1005 09:49:05.181934 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-755h7/crc-debug-dxxdp" Oct 05 09:49:05 crc kubenswrapper[4846]: I1005 09:49:05.446131 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-755h7/crc-debug-dxxdp" event={"ID":"0d325705-466c-4f52-bb47-0566c8bdbdbf","Type":"ContainerStarted","Data":"fd5e22157da2784468290fbd8360c83b1b45e680865969e1926ed00959535cee"} Oct 05 09:49:06 crc kubenswrapper[4846]: I1005 09:49:06.461739 4846 generic.go:334] "Generic (PLEG): container finished" podID="0d325705-466c-4f52-bb47-0566c8bdbdbf" containerID="4d9eb16a8d8865ce8122ac96654f74dc80d62f8000a2603516b2ff52c19fd384" exitCode=0 Oct 05 09:49:06 crc kubenswrapper[4846]: I1005 09:49:06.461802 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-755h7/crc-debug-dxxdp" event={"ID":"0d325705-466c-4f52-bb47-0566c8bdbdbf","Type":"ContainerDied","Data":"4d9eb16a8d8865ce8122ac96654f74dc80d62f8000a2603516b2ff52c19fd384"} Oct 05 09:49:07 crc kubenswrapper[4846]: I1005 09:49:07.631855 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-755h7/crc-debug-dxxdp" Oct 05 09:49:07 crc kubenswrapper[4846]: I1005 09:49:07.663298 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hcjnz\" (UniqueName: \"kubernetes.io/projected/0d325705-466c-4f52-bb47-0566c8bdbdbf-kube-api-access-hcjnz\") pod \"0d325705-466c-4f52-bb47-0566c8bdbdbf\" (UID: \"0d325705-466c-4f52-bb47-0566c8bdbdbf\") " Oct 05 09:49:07 crc kubenswrapper[4846]: I1005 09:49:07.663465 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/0d325705-466c-4f52-bb47-0566c8bdbdbf-host\") pod \"0d325705-466c-4f52-bb47-0566c8bdbdbf\" (UID: \"0d325705-466c-4f52-bb47-0566c8bdbdbf\") " Oct 05 09:49:07 crc kubenswrapper[4846]: I1005 09:49:07.663641 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0d325705-466c-4f52-bb47-0566c8bdbdbf-host" (OuterVolumeSpecName: "host") pod "0d325705-466c-4f52-bb47-0566c8bdbdbf" (UID: "0d325705-466c-4f52-bb47-0566c8bdbdbf"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 05 09:49:07 crc kubenswrapper[4846]: I1005 09:49:07.664457 4846 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/0d325705-466c-4f52-bb47-0566c8bdbdbf-host\") on node \"crc\" DevicePath \"\"" Oct 05 09:49:07 crc kubenswrapper[4846]: I1005 09:49:07.668975 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0d325705-466c-4f52-bb47-0566c8bdbdbf-kube-api-access-hcjnz" (OuterVolumeSpecName: "kube-api-access-hcjnz") pod "0d325705-466c-4f52-bb47-0566c8bdbdbf" (UID: "0d325705-466c-4f52-bb47-0566c8bdbdbf"). InnerVolumeSpecName "kube-api-access-hcjnz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 09:49:07 crc kubenswrapper[4846]: I1005 09:49:07.766021 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hcjnz\" (UniqueName: \"kubernetes.io/projected/0d325705-466c-4f52-bb47-0566c8bdbdbf-kube-api-access-hcjnz\") on node \"crc\" DevicePath \"\"" Oct 05 09:49:08 crc kubenswrapper[4846]: I1005 09:49:08.481638 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-755h7/crc-debug-dxxdp" event={"ID":"0d325705-466c-4f52-bb47-0566c8bdbdbf","Type":"ContainerDied","Data":"fd5e22157da2784468290fbd8360c83b1b45e680865969e1926ed00959535cee"} Oct 05 09:49:08 crc kubenswrapper[4846]: I1005 09:49:08.482073 4846 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fd5e22157da2784468290fbd8360c83b1b45e680865969e1926ed00959535cee" Oct 05 09:49:08 crc kubenswrapper[4846]: I1005 09:49:08.481707 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-755h7/crc-debug-dxxdp" Oct 05 09:49:19 crc kubenswrapper[4846]: I1005 09:49:19.529193 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-755h7/crc-debug-dxxdp"] Oct 05 09:49:19 crc kubenswrapper[4846]: I1005 09:49:19.536458 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-755h7/crc-debug-dxxdp"] Oct 05 09:49:20 crc kubenswrapper[4846]: I1005 09:49:20.521877 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0d325705-466c-4f52-bb47-0566c8bdbdbf" path="/var/lib/kubelet/pods/0d325705-466c-4f52-bb47-0566c8bdbdbf/volumes" Oct 05 09:49:20 crc kubenswrapper[4846]: I1005 09:49:20.780020 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-755h7/crc-debug-tbp9k"] Oct 05 09:49:20 crc kubenswrapper[4846]: E1005 09:49:20.780726 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0d325705-466c-4f52-bb47-0566c8bdbdbf" containerName="container-00" Oct 05 09:49:20 crc kubenswrapper[4846]: I1005 09:49:20.780751 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="0d325705-466c-4f52-bb47-0566c8bdbdbf" containerName="container-00" Oct 05 09:49:20 crc kubenswrapper[4846]: I1005 09:49:20.781246 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="0d325705-466c-4f52-bb47-0566c8bdbdbf" containerName="container-00" Oct 05 09:49:20 crc kubenswrapper[4846]: I1005 09:49:20.783561 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-755h7/crc-debug-tbp9k" Oct 05 09:49:20 crc kubenswrapper[4846]: I1005 09:49:20.890200 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/5d4c72b3-2d52-4b41-a7ed-acedfcdb7633-host\") pod \"crc-debug-tbp9k\" (UID: \"5d4c72b3-2d52-4b41-a7ed-acedfcdb7633\") " pod="openshift-must-gather-755h7/crc-debug-tbp9k" Oct 05 09:49:20 crc kubenswrapper[4846]: I1005 09:49:20.891580 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zplkn\" (UniqueName: \"kubernetes.io/projected/5d4c72b3-2d52-4b41-a7ed-acedfcdb7633-kube-api-access-zplkn\") pod \"crc-debug-tbp9k\" (UID: \"5d4c72b3-2d52-4b41-a7ed-acedfcdb7633\") " pod="openshift-must-gather-755h7/crc-debug-tbp9k" Oct 05 09:49:20 crc kubenswrapper[4846]: I1005 09:49:20.994127 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/5d4c72b3-2d52-4b41-a7ed-acedfcdb7633-host\") pod \"crc-debug-tbp9k\" (UID: \"5d4c72b3-2d52-4b41-a7ed-acedfcdb7633\") " pod="openshift-must-gather-755h7/crc-debug-tbp9k" Oct 05 09:49:20 crc kubenswrapper[4846]: I1005 09:49:20.994290 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/5d4c72b3-2d52-4b41-a7ed-acedfcdb7633-host\") pod \"crc-debug-tbp9k\" (UID: \"5d4c72b3-2d52-4b41-a7ed-acedfcdb7633\") " pod="openshift-must-gather-755h7/crc-debug-tbp9k" Oct 05 09:49:20 crc kubenswrapper[4846]: I1005 09:49:20.994422 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zplkn\" (UniqueName: \"kubernetes.io/projected/5d4c72b3-2d52-4b41-a7ed-acedfcdb7633-kube-api-access-zplkn\") pod \"crc-debug-tbp9k\" (UID: \"5d4c72b3-2d52-4b41-a7ed-acedfcdb7633\") " pod="openshift-must-gather-755h7/crc-debug-tbp9k" Oct 05 09:49:21 crc kubenswrapper[4846]: I1005 09:49:21.020587 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zplkn\" (UniqueName: \"kubernetes.io/projected/5d4c72b3-2d52-4b41-a7ed-acedfcdb7633-kube-api-access-zplkn\") pod \"crc-debug-tbp9k\" (UID: \"5d4c72b3-2d52-4b41-a7ed-acedfcdb7633\") " pod="openshift-must-gather-755h7/crc-debug-tbp9k" Oct 05 09:49:21 crc kubenswrapper[4846]: I1005 09:49:21.110647 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-755h7/crc-debug-tbp9k" Oct 05 09:49:21 crc kubenswrapper[4846]: W1005 09:49:21.152400 4846 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5d4c72b3_2d52_4b41_a7ed_acedfcdb7633.slice/crio-058a50a5921f54327df1a7340df79676f4dbb06d4b36ebb5500710f20b9a6cf3 WatchSource:0}: Error finding container 058a50a5921f54327df1a7340df79676f4dbb06d4b36ebb5500710f20b9a6cf3: Status 404 returned error can't find the container with id 058a50a5921f54327df1a7340df79676f4dbb06d4b36ebb5500710f20b9a6cf3 Oct 05 09:49:21 crc kubenswrapper[4846]: I1005 09:49:21.617820 4846 generic.go:334] "Generic (PLEG): container finished" podID="5d4c72b3-2d52-4b41-a7ed-acedfcdb7633" containerID="fc40fb770020887235a4d1af0793055b92ff4c9ca8d626d2e8f88c272f763793" exitCode=0 Oct 05 09:49:21 crc kubenswrapper[4846]: I1005 09:49:21.617885 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-755h7/crc-debug-tbp9k" event={"ID":"5d4c72b3-2d52-4b41-a7ed-acedfcdb7633","Type":"ContainerDied","Data":"fc40fb770020887235a4d1af0793055b92ff4c9ca8d626d2e8f88c272f763793"} Oct 05 09:49:21 crc kubenswrapper[4846]: I1005 09:49:21.618242 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-755h7/crc-debug-tbp9k" event={"ID":"5d4c72b3-2d52-4b41-a7ed-acedfcdb7633","Type":"ContainerStarted","Data":"058a50a5921f54327df1a7340df79676f4dbb06d4b36ebb5500710f20b9a6cf3"} Oct 05 09:49:21 crc kubenswrapper[4846]: I1005 09:49:21.674870 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-755h7/crc-debug-tbp9k"] Oct 05 09:49:21 crc kubenswrapper[4846]: I1005 09:49:21.690676 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-755h7/crc-debug-tbp9k"] Oct 05 09:49:22 crc kubenswrapper[4846]: I1005 09:49:22.728839 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-755h7/crc-debug-tbp9k" Oct 05 09:49:22 crc kubenswrapper[4846]: I1005 09:49:22.839051 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/5d4c72b3-2d52-4b41-a7ed-acedfcdb7633-host\") pod \"5d4c72b3-2d52-4b41-a7ed-acedfcdb7633\" (UID: \"5d4c72b3-2d52-4b41-a7ed-acedfcdb7633\") " Oct 05 09:49:22 crc kubenswrapper[4846]: I1005 09:49:22.839514 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zplkn\" (UniqueName: \"kubernetes.io/projected/5d4c72b3-2d52-4b41-a7ed-acedfcdb7633-kube-api-access-zplkn\") pod \"5d4c72b3-2d52-4b41-a7ed-acedfcdb7633\" (UID: \"5d4c72b3-2d52-4b41-a7ed-acedfcdb7633\") " Oct 05 09:49:22 crc kubenswrapper[4846]: I1005 09:49:22.839264 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5d4c72b3-2d52-4b41-a7ed-acedfcdb7633-host" (OuterVolumeSpecName: "host") pod "5d4c72b3-2d52-4b41-a7ed-acedfcdb7633" (UID: "5d4c72b3-2d52-4b41-a7ed-acedfcdb7633"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 05 09:49:22 crc kubenswrapper[4846]: I1005 09:49:22.840799 4846 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/5d4c72b3-2d52-4b41-a7ed-acedfcdb7633-host\") on node \"crc\" DevicePath \"\"" Oct 05 09:49:22 crc kubenswrapper[4846]: I1005 09:49:22.848526 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5d4c72b3-2d52-4b41-a7ed-acedfcdb7633-kube-api-access-zplkn" (OuterVolumeSpecName: "kube-api-access-zplkn") pod "5d4c72b3-2d52-4b41-a7ed-acedfcdb7633" (UID: "5d4c72b3-2d52-4b41-a7ed-acedfcdb7633"). InnerVolumeSpecName "kube-api-access-zplkn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 09:49:22 crc kubenswrapper[4846]: I1005 09:49:22.943775 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zplkn\" (UniqueName: \"kubernetes.io/projected/5d4c72b3-2d52-4b41-a7ed-acedfcdb7633-kube-api-access-zplkn\") on node \"crc\" DevicePath \"\"" Oct 05 09:49:23 crc kubenswrapper[4846]: I1005 09:49:23.428955 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_0da5b2e9368362304c733e302da704023e73b0b2df8ed109170f4705a8lxrbm_1a853c3b-02bb-4ad2-8f91-2629d9eade31/util/0.log" Oct 05 09:49:23 crc kubenswrapper[4846]: I1005 09:49:23.600194 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_0da5b2e9368362304c733e302da704023e73b0b2df8ed109170f4705a8lxrbm_1a853c3b-02bb-4ad2-8f91-2629d9eade31/pull/0.log" Oct 05 09:49:23 crc kubenswrapper[4846]: I1005 09:49:23.611289 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_0da5b2e9368362304c733e302da704023e73b0b2df8ed109170f4705a8lxrbm_1a853c3b-02bb-4ad2-8f91-2629d9eade31/util/0.log" Oct 05 09:49:23 crc kubenswrapper[4846]: I1005 09:49:23.620372 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_0da5b2e9368362304c733e302da704023e73b0b2df8ed109170f4705a8lxrbm_1a853c3b-02bb-4ad2-8f91-2629d9eade31/pull/0.log" Oct 05 09:49:23 crc kubenswrapper[4846]: I1005 09:49:23.637227 4846 scope.go:117] "RemoveContainer" containerID="fc40fb770020887235a4d1af0793055b92ff4c9ca8d626d2e8f88c272f763793" Oct 05 09:49:23 crc kubenswrapper[4846]: I1005 09:49:23.637297 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-755h7/crc-debug-tbp9k" Oct 05 09:49:23 crc kubenswrapper[4846]: I1005 09:49:23.781483 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_0da5b2e9368362304c733e302da704023e73b0b2df8ed109170f4705a8lxrbm_1a853c3b-02bb-4ad2-8f91-2629d9eade31/util/0.log" Oct 05 09:49:23 crc kubenswrapper[4846]: I1005 09:49:23.800249 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_0da5b2e9368362304c733e302da704023e73b0b2df8ed109170f4705a8lxrbm_1a853c3b-02bb-4ad2-8f91-2629d9eade31/extract/0.log" Oct 05 09:49:23 crc kubenswrapper[4846]: I1005 09:49:23.804818 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_0da5b2e9368362304c733e302da704023e73b0b2df8ed109170f4705a8lxrbm_1a853c3b-02bb-4ad2-8f91-2629d9eade31/pull/0.log" Oct 05 09:49:23 crc kubenswrapper[4846]: I1005 09:49:23.949076 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-5b974f6766-khk5k_f626116b-4f3a-48cb-ad75-c6550333ee6e/kube-rbac-proxy/0.log" Oct 05 09:49:24 crc kubenswrapper[4846]: I1005 09:49:24.013233 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-84bd8f6848-b4rhd_58088fc8-5aaa-422a-84cd-d4ef7d10e989/kube-rbac-proxy/0.log" Oct 05 09:49:24 crc kubenswrapper[4846]: I1005 09:49:24.071266 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-5b974f6766-khk5k_f626116b-4f3a-48cb-ad75-c6550333ee6e/manager/0.log" Oct 05 09:49:24 crc kubenswrapper[4846]: I1005 09:49:24.229854 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-84bd8f6848-b4rhd_58088fc8-5aaa-422a-84cd-d4ef7d10e989/manager/0.log" Oct 05 09:49:24 crc kubenswrapper[4846]: I1005 09:49:24.244287 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-58d86cd59d-gnb7x_ac8cdf70-77c0-44c9-b61b-11325d91b698/kube-rbac-proxy/0.log" Oct 05 09:49:24 crc kubenswrapper[4846]: I1005 09:49:24.277826 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-58d86cd59d-gnb7x_ac8cdf70-77c0-44c9-b61b-11325d91b698/manager/0.log" Oct 05 09:49:24 crc kubenswrapper[4846]: I1005 09:49:24.414769 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-698456cdc6-w4fst_c1b6f754-a4fc-4981-a722-415569d3e7a3/kube-rbac-proxy/0.log" Oct 05 09:49:24 crc kubenswrapper[4846]: I1005 09:49:24.506809 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5d4c72b3-2d52-4b41-a7ed-acedfcdb7633" path="/var/lib/kubelet/pods/5d4c72b3-2d52-4b41-a7ed-acedfcdb7633/volumes" Oct 05 09:49:24 crc kubenswrapper[4846]: I1005 09:49:24.522614 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-698456cdc6-w4fst_c1b6f754-a4fc-4981-a722-415569d3e7a3/manager/0.log" Oct 05 09:49:24 crc kubenswrapper[4846]: I1005 09:49:24.612651 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5c497dbdb-g4bkx_931bcbda-d249-4625-a93c-bfa49dcb38ae/kube-rbac-proxy/0.log" Oct 05 09:49:24 crc kubenswrapper[4846]: I1005 09:49:24.694476 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-6675647785-d2n5v_a1899a06-0cb1-44aa-ba1e-d6876051bab1/kube-rbac-proxy/0.log" Oct 05 09:49:24 crc kubenswrapper[4846]: I1005 09:49:24.705396 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5c497dbdb-g4bkx_931bcbda-d249-4625-a93c-bfa49dcb38ae/manager/0.log" Oct 05 09:49:24 crc kubenswrapper[4846]: I1005 09:49:24.821329 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-6675647785-d2n5v_a1899a06-0cb1-44aa-ba1e-d6876051bab1/manager/0.log" Oct 05 09:49:24 crc kubenswrapper[4846]: I1005 09:49:24.928682 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-84788b6bc5-pg8c6_f0825244-e759-46a4-a16b-4d993dac339d/kube-rbac-proxy/0.log" Oct 05 09:49:25 crc kubenswrapper[4846]: I1005 09:49:25.113833 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-6f5894c49f-fkx4r_92c93055-428a-4d41-b034-a805b68f3e19/kube-rbac-proxy/0.log" Oct 05 09:49:25 crc kubenswrapper[4846]: I1005 09:49:25.178666 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-84788b6bc5-pg8c6_f0825244-e759-46a4-a16b-4d993dac339d/manager/0.log" Oct 05 09:49:25 crc kubenswrapper[4846]: I1005 09:49:25.186902 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-6f5894c49f-fkx4r_92c93055-428a-4d41-b034-a805b68f3e19/manager/0.log" Oct 05 09:49:25 crc kubenswrapper[4846]: I1005 09:49:25.280810 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-57c9cdcf57-j9742_6ff6c74d-2f8b-42d0-af60-88baa4399b32/kube-rbac-proxy/0.log" Oct 05 09:49:25 crc kubenswrapper[4846]: I1005 09:49:25.387444 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-7cb48dbc-cvtkp_e46909e0-1642-436c-b597-167a42e0a72d/kube-rbac-proxy/0.log" Oct 05 09:49:25 crc kubenswrapper[4846]: I1005 09:49:25.429645 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-57c9cdcf57-j9742_6ff6c74d-2f8b-42d0-af60-88baa4399b32/manager/0.log" Oct 05 09:49:25 crc kubenswrapper[4846]: I1005 09:49:25.518413 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-7cb48dbc-cvtkp_e46909e0-1642-436c-b597-167a42e0a72d/manager/0.log" Oct 05 09:49:25 crc kubenswrapper[4846]: I1005 09:49:25.721987 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-d6c9dc5bc-2fb7l_71de4f09-b26f-443c-9142-7edbc0d18e38/kube-rbac-proxy/0.log" Oct 05 09:49:25 crc kubenswrapper[4846]: I1005 09:49:25.790475 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-d6c9dc5bc-2fb7l_71de4f09-b26f-443c-9142-7edbc0d18e38/manager/0.log" Oct 05 09:49:25 crc kubenswrapper[4846]: I1005 09:49:25.918273 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-69b956fbf6-xsltv_644a3a04-e2d4-4352-9bac-e8d661fd61b0/kube-rbac-proxy/0.log" Oct 05 09:49:25 crc kubenswrapper[4846]: I1005 09:49:25.972088 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-69b956fbf6-xsltv_644a3a04-e2d4-4352-9bac-e8d661fd61b0/manager/0.log" Oct 05 09:49:26 crc kubenswrapper[4846]: I1005 09:49:26.110745 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-6c9b57c67-ggtrm_49e72a92-e750-4046-9894-f3a0359531b2/kube-rbac-proxy/0.log" Oct 05 09:49:26 crc kubenswrapper[4846]: I1005 09:49:26.201304 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-69f59f9d8-s4qf7_15b8ca14-e689-430c-84b2-40da6f1e83a5/kube-rbac-proxy/0.log" Oct 05 09:49:26 crc kubenswrapper[4846]: I1005 09:49:26.291640 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-6c9b57c67-ggtrm_49e72a92-e750-4046-9894-f3a0359531b2/manager/0.log" Oct 05 09:49:26 crc kubenswrapper[4846]: I1005 09:49:26.298583 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-69f59f9d8-s4qf7_15b8ca14-e689-430c-84b2-40da6f1e83a5/manager/0.log" Oct 05 09:49:26 crc kubenswrapper[4846]: I1005 09:49:26.385750 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-66cc85b5d5lsczg_eb017e3a-1627-4dc7-bda3-1faec8b72739/kube-rbac-proxy/0.log" Oct 05 09:49:26 crc kubenswrapper[4846]: I1005 09:49:26.487557 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-66cc85b5d5lsczg_eb017e3a-1627-4dc7-bda3-1faec8b72739/manager/0.log" Oct 05 09:49:26 crc kubenswrapper[4846]: I1005 09:49:26.562783 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-7cfc658b9-22b7k_d7419a91-cc67-4cdb-b8a4-d41480e5854f/kube-rbac-proxy/0.log" Oct 05 09:49:26 crc kubenswrapper[4846]: I1005 09:49:26.753863 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-677d5bb784-mzr42_af8c317e-4f15-4f06-9a50-198efd07cc4b/kube-rbac-proxy/0.log" Oct 05 09:49:26 crc kubenswrapper[4846]: I1005 09:49:26.955347 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-677d5bb784-mzr42_af8c317e-4f15-4f06-9a50-198efd07cc4b/operator/0.log" Oct 05 09:49:27 crc kubenswrapper[4846]: I1005 09:49:27.074650 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-c968bb45-rmwgr_945ab775-ffae-41cb-a315-5e05cf65484d/kube-rbac-proxy/0.log" Oct 05 09:49:27 crc kubenswrapper[4846]: I1005 09:49:27.120066 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-4d9cv_21f31e13-73d1-4eb0-bd66-5878496bfdb2/registry-server/0.log" Oct 05 09:49:27 crc kubenswrapper[4846]: I1005 09:49:27.266099 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-c968bb45-rmwgr_945ab775-ffae-41cb-a315-5e05cf65484d/manager/0.log" Oct 05 09:49:27 crc kubenswrapper[4846]: I1005 09:49:27.353376 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-66f6d6849b-62zd4_95847a57-fbe8-43f2-b4b4-e630665f17e3/kube-rbac-proxy/0.log" Oct 05 09:49:27 crc kubenswrapper[4846]: I1005 09:49:27.416756 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-66f6d6849b-62zd4_95847a57-fbe8-43f2-b4b4-e630665f17e3/manager/0.log" Oct 05 09:49:27 crc kubenswrapper[4846]: I1005 09:49:27.514834 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-5f97d8c699-pdqdx_aae22754-9bda-494a-afb8-42debef021dd/operator/0.log" Oct 05 09:49:27 crc kubenswrapper[4846]: I1005 09:49:27.629766 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-76d5577b-pkjcv_ab1677d6-3cb0-4dce-8e8d-4d2b67025e6d/kube-rbac-proxy/0.log" Oct 05 09:49:27 crc kubenswrapper[4846]: I1005 09:49:27.785826 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-76d5577b-pkjcv_ab1677d6-3cb0-4dce-8e8d-4d2b67025e6d/manager/0.log" Oct 05 09:49:27 crc kubenswrapper[4846]: I1005 09:49:27.836196 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-f589c7597-jzsnd_3053ab31-f463-49b7-9b32-d37b0506462b/kube-rbac-proxy/0.log" Oct 05 09:49:27 crc kubenswrapper[4846]: I1005 09:49:27.939997 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-6bb6dcddc-4krwr_b7a3163e-66e4-406c-9285-e2d125024bc8/kube-rbac-proxy/0.log" Oct 05 09:49:28 crc kubenswrapper[4846]: I1005 09:49:28.094705 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-6bb6dcddc-4krwr_b7a3163e-66e4-406c-9285-e2d125024bc8/manager/0.log" Oct 05 09:49:28 crc kubenswrapper[4846]: I1005 09:49:28.204492 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-f589c7597-jzsnd_3053ab31-f463-49b7-9b32-d37b0506462b/manager/0.log" Oct 05 09:49:28 crc kubenswrapper[4846]: I1005 09:49:28.209170 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-5d98cc5575-blkpr_d4de72d9-7d9c-4d90-ba8a-4179d9028033/kube-rbac-proxy/0.log" Oct 05 09:49:28 crc kubenswrapper[4846]: I1005 09:49:28.306759 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-5d98cc5575-blkpr_d4de72d9-7d9c-4d90-ba8a-4179d9028033/manager/0.log" Oct 05 09:49:28 crc kubenswrapper[4846]: I1005 09:49:28.735424 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-7cfc658b9-22b7k_d7419a91-cc67-4cdb-b8a4-d41480e5854f/manager/0.log" Oct 05 09:49:45 crc kubenswrapper[4846]: I1005 09:49:45.140840 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-2x9nl_57e551f8-d9da-4692-b8ae-f7ff6829fe25/control-plane-machine-set-operator/0.log" Oct 05 09:49:45 crc kubenswrapper[4846]: I1005 09:49:45.320500 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-vwx77_f743f03b-3f08-4da0-b53e-1cb33ab516b9/kube-rbac-proxy/0.log" Oct 05 09:49:45 crc kubenswrapper[4846]: I1005 09:49:45.373539 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-vwx77_f743f03b-3f08-4da0-b53e-1cb33ab516b9/machine-api-operator/0.log" Oct 05 09:49:58 crc kubenswrapper[4846]: I1005 09:49:58.366273 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-7d4cc89fcb-84jrb_a1c9433e-11a8-47fb-91ee-4da4b43facf3/cert-manager-controller/0.log" Oct 05 09:49:58 crc kubenswrapper[4846]: I1005 09:49:58.501160 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-7d9f95dbf-ht8qn_7520b957-0745-432e-a8d5-37e2f9e7405b/cert-manager-cainjector/0.log" Oct 05 09:49:58 crc kubenswrapper[4846]: I1005 09:49:58.568357 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-d969966f-fb5qg_4823450d-b1c6-4d08-9796-109059819ec0/cert-manager-webhook/0.log" Oct 05 09:50:11 crc kubenswrapper[4846]: I1005 09:50:11.033143 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-6b874cbd85-ltpns_624c96f8-0f60-48c9-abd3-5c9594a525a7/nmstate-console-plugin/0.log" Oct 05 09:50:11 crc kubenswrapper[4846]: I1005 09:50:11.254112 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-5dcf6_8383ac5a-9f8c-4bdc-aded-25b6b9cb837a/nmstate-handler/0.log" Oct 05 09:50:11 crc kubenswrapper[4846]: I1005 09:50:11.255358 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-fdff9cb8d-dlm67_62d96d39-477a-468c-8b8b-02521146f2ba/kube-rbac-proxy/0.log" Oct 05 09:50:11 crc kubenswrapper[4846]: I1005 09:50:11.272548 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-fdff9cb8d-dlm67_62d96d39-477a-468c-8b8b-02521146f2ba/nmstate-metrics/0.log" Oct 05 09:50:11 crc kubenswrapper[4846]: I1005 09:50:11.416032 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-858ddd8f98-lwsfd_46fb70d7-c48a-4d57-990f-5e583f96fa16/nmstate-operator/0.log" Oct 05 09:50:11 crc kubenswrapper[4846]: I1005 09:50:11.470961 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-6cdbc54649-h5krq_ceef70f7-1471-4dd1-aabc-85d885a302da/nmstate-webhook/0.log" Oct 05 09:50:23 crc kubenswrapper[4846]: I1005 09:50:23.325305 4846 patch_prober.go:28] interesting pod/machine-config-daemon-fscvf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 09:50:23 crc kubenswrapper[4846]: I1005 09:50:23.325880 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 09:50:26 crc kubenswrapper[4846]: I1005 09:50:26.844862 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-68d546b9d8-vhpgd_3541826d-961c-4259-be16-51f4dc05aec6/kube-rbac-proxy/0.log" Oct 05 09:50:27 crc kubenswrapper[4846]: I1005 09:50:27.076667 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-kz2bc_ab2bc1f2-31f3-4793-926a-e649b556d8de/cp-frr-files/0.log" Oct 05 09:50:27 crc kubenswrapper[4846]: I1005 09:50:27.218363 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-68d546b9d8-vhpgd_3541826d-961c-4259-be16-51f4dc05aec6/controller/0.log" Oct 05 09:50:27 crc kubenswrapper[4846]: I1005 09:50:27.256417 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-kz2bc_ab2bc1f2-31f3-4793-926a-e649b556d8de/cp-reloader/0.log" Oct 05 09:50:27 crc kubenswrapper[4846]: I1005 09:50:27.280482 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-kz2bc_ab2bc1f2-31f3-4793-926a-e649b556d8de/cp-metrics/0.log" Oct 05 09:50:27 crc kubenswrapper[4846]: I1005 09:50:27.281352 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-kz2bc_ab2bc1f2-31f3-4793-926a-e649b556d8de/cp-frr-files/0.log" Oct 05 09:50:27 crc kubenswrapper[4846]: I1005 09:50:27.400663 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-kz2bc_ab2bc1f2-31f3-4793-926a-e649b556d8de/cp-reloader/0.log" Oct 05 09:50:27 crc kubenswrapper[4846]: I1005 09:50:27.567734 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-kz2bc_ab2bc1f2-31f3-4793-926a-e649b556d8de/cp-frr-files/0.log" Oct 05 09:50:27 crc kubenswrapper[4846]: I1005 09:50:27.581544 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-kz2bc_ab2bc1f2-31f3-4793-926a-e649b556d8de/cp-reloader/0.log" Oct 05 09:50:27 crc kubenswrapper[4846]: I1005 09:50:27.618143 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-kz2bc_ab2bc1f2-31f3-4793-926a-e649b556d8de/cp-metrics/0.log" Oct 05 09:50:27 crc kubenswrapper[4846]: I1005 09:50:27.634003 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-kz2bc_ab2bc1f2-31f3-4793-926a-e649b556d8de/cp-metrics/0.log" Oct 05 09:50:27 crc kubenswrapper[4846]: I1005 09:50:27.861608 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-kz2bc_ab2bc1f2-31f3-4793-926a-e649b556d8de/cp-metrics/0.log" Oct 05 09:50:27 crc kubenswrapper[4846]: I1005 09:50:27.912132 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-kz2bc_ab2bc1f2-31f3-4793-926a-e649b556d8de/cp-frr-files/0.log" Oct 05 09:50:27 crc kubenswrapper[4846]: I1005 09:50:27.927838 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-kz2bc_ab2bc1f2-31f3-4793-926a-e649b556d8de/controller/0.log" Oct 05 09:50:27 crc kubenswrapper[4846]: I1005 09:50:27.928699 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-kz2bc_ab2bc1f2-31f3-4793-926a-e649b556d8de/cp-reloader/0.log" Oct 05 09:50:28 crc kubenswrapper[4846]: I1005 09:50:28.103234 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-kz2bc_ab2bc1f2-31f3-4793-926a-e649b556d8de/frr-metrics/0.log" Oct 05 09:50:28 crc kubenswrapper[4846]: I1005 09:50:28.136696 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-kz2bc_ab2bc1f2-31f3-4793-926a-e649b556d8de/kube-rbac-proxy/0.log" Oct 05 09:50:28 crc kubenswrapper[4846]: I1005 09:50:28.164723 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-kz2bc_ab2bc1f2-31f3-4793-926a-e649b556d8de/kube-rbac-proxy-frr/0.log" Oct 05 09:50:28 crc kubenswrapper[4846]: I1005 09:50:28.313688 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-kz2bc_ab2bc1f2-31f3-4793-926a-e649b556d8de/reloader/0.log" Oct 05 09:50:28 crc kubenswrapper[4846]: I1005 09:50:28.365540 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-64bf5d555-sf9k8_0a4508c6-6dff-467f-9cac-ef5c7b921984/frr-k8s-webhook-server/0.log" Oct 05 09:50:28 crc kubenswrapper[4846]: I1005 09:50:28.658252 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-8fccc95cf-j9fdr_cbc67894-51b7-4d7d-92b2-eb8b70a77af5/manager/0.log" Oct 05 09:50:28 crc kubenswrapper[4846]: I1005 09:50:28.799847 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-7bd844db7f-6hpbl_15c86c60-ec26-4e57-bb15-cd3bc1843e80/webhook-server/0.log" Oct 05 09:50:28 crc kubenswrapper[4846]: I1005 09:50:28.995259 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-czf4m_dad102f2-8887-4a01-98b1-be88b2836eb5/kube-rbac-proxy/0.log" Oct 05 09:50:29 crc kubenswrapper[4846]: I1005 09:50:29.817916 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-czf4m_dad102f2-8887-4a01-98b1-be88b2836eb5/speaker/0.log" Oct 05 09:50:30 crc kubenswrapper[4846]: I1005 09:50:30.976789 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-kz2bc_ab2bc1f2-31f3-4793-926a-e649b556d8de/frr/0.log" Oct 05 09:50:37 crc kubenswrapper[4846]: I1005 09:50:37.259882 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-gv5pf"] Oct 05 09:50:37 crc kubenswrapper[4846]: E1005 09:50:37.261012 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5d4c72b3-2d52-4b41-a7ed-acedfcdb7633" containerName="container-00" Oct 05 09:50:37 crc kubenswrapper[4846]: I1005 09:50:37.261027 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="5d4c72b3-2d52-4b41-a7ed-acedfcdb7633" containerName="container-00" Oct 05 09:50:37 crc kubenswrapper[4846]: I1005 09:50:37.261285 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="5d4c72b3-2d52-4b41-a7ed-acedfcdb7633" containerName="container-00" Oct 05 09:50:37 crc kubenswrapper[4846]: I1005 09:50:37.262949 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-gv5pf" Oct 05 09:50:37 crc kubenswrapper[4846]: I1005 09:50:37.283754 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-gv5pf"] Oct 05 09:50:37 crc kubenswrapper[4846]: I1005 09:50:37.424799 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v9hrs\" (UniqueName: \"kubernetes.io/projected/9617e8e6-7fd5-47c6-9afb-e6160221ddea-kube-api-access-v9hrs\") pod \"redhat-marketplace-gv5pf\" (UID: \"9617e8e6-7fd5-47c6-9afb-e6160221ddea\") " pod="openshift-marketplace/redhat-marketplace-gv5pf" Oct 05 09:50:37 crc kubenswrapper[4846]: I1005 09:50:37.424974 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9617e8e6-7fd5-47c6-9afb-e6160221ddea-catalog-content\") pod \"redhat-marketplace-gv5pf\" (UID: \"9617e8e6-7fd5-47c6-9afb-e6160221ddea\") " pod="openshift-marketplace/redhat-marketplace-gv5pf" Oct 05 09:50:37 crc kubenswrapper[4846]: I1005 09:50:37.425007 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9617e8e6-7fd5-47c6-9afb-e6160221ddea-utilities\") pod \"redhat-marketplace-gv5pf\" (UID: \"9617e8e6-7fd5-47c6-9afb-e6160221ddea\") " pod="openshift-marketplace/redhat-marketplace-gv5pf" Oct 05 09:50:37 crc kubenswrapper[4846]: I1005 09:50:37.527940 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9617e8e6-7fd5-47c6-9afb-e6160221ddea-catalog-content\") pod \"redhat-marketplace-gv5pf\" (UID: \"9617e8e6-7fd5-47c6-9afb-e6160221ddea\") " pod="openshift-marketplace/redhat-marketplace-gv5pf" Oct 05 09:50:37 crc kubenswrapper[4846]: I1005 09:50:37.528029 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9617e8e6-7fd5-47c6-9afb-e6160221ddea-utilities\") pod \"redhat-marketplace-gv5pf\" (UID: \"9617e8e6-7fd5-47c6-9afb-e6160221ddea\") " pod="openshift-marketplace/redhat-marketplace-gv5pf" Oct 05 09:50:37 crc kubenswrapper[4846]: I1005 09:50:37.528257 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v9hrs\" (UniqueName: \"kubernetes.io/projected/9617e8e6-7fd5-47c6-9afb-e6160221ddea-kube-api-access-v9hrs\") pod \"redhat-marketplace-gv5pf\" (UID: \"9617e8e6-7fd5-47c6-9afb-e6160221ddea\") " pod="openshift-marketplace/redhat-marketplace-gv5pf" Oct 05 09:50:37 crc kubenswrapper[4846]: I1005 09:50:37.528576 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9617e8e6-7fd5-47c6-9afb-e6160221ddea-catalog-content\") pod \"redhat-marketplace-gv5pf\" (UID: \"9617e8e6-7fd5-47c6-9afb-e6160221ddea\") " pod="openshift-marketplace/redhat-marketplace-gv5pf" Oct 05 09:50:37 crc kubenswrapper[4846]: I1005 09:50:37.528664 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9617e8e6-7fd5-47c6-9afb-e6160221ddea-utilities\") pod \"redhat-marketplace-gv5pf\" (UID: \"9617e8e6-7fd5-47c6-9afb-e6160221ddea\") " pod="openshift-marketplace/redhat-marketplace-gv5pf" Oct 05 09:50:37 crc kubenswrapper[4846]: I1005 09:50:37.558342 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v9hrs\" (UniqueName: \"kubernetes.io/projected/9617e8e6-7fd5-47c6-9afb-e6160221ddea-kube-api-access-v9hrs\") pod \"redhat-marketplace-gv5pf\" (UID: \"9617e8e6-7fd5-47c6-9afb-e6160221ddea\") " pod="openshift-marketplace/redhat-marketplace-gv5pf" Oct 05 09:50:37 crc kubenswrapper[4846]: I1005 09:50:37.581981 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-gv5pf" Oct 05 09:50:38 crc kubenswrapper[4846]: I1005 09:50:38.093428 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-gv5pf"] Oct 05 09:50:38 crc kubenswrapper[4846]: I1005 09:50:38.456492 4846 generic.go:334] "Generic (PLEG): container finished" podID="9617e8e6-7fd5-47c6-9afb-e6160221ddea" containerID="72f64ea7a880c126e7c51eb6aaa61a8bb5bbd42707f5fbd95a39f3cefa9f48f5" exitCode=0 Oct 05 09:50:38 crc kubenswrapper[4846]: I1005 09:50:38.456546 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gv5pf" event={"ID":"9617e8e6-7fd5-47c6-9afb-e6160221ddea","Type":"ContainerDied","Data":"72f64ea7a880c126e7c51eb6aaa61a8bb5bbd42707f5fbd95a39f3cefa9f48f5"} Oct 05 09:50:38 crc kubenswrapper[4846]: I1005 09:50:38.456577 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gv5pf" event={"ID":"9617e8e6-7fd5-47c6-9afb-e6160221ddea","Type":"ContainerStarted","Data":"bd65ee98cf7691ba9ea9a4cc384a3152aef3efb704c3daeb6a28d0ded0e97d8e"} Oct 05 09:50:39 crc kubenswrapper[4846]: I1005 09:50:39.471500 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gv5pf" event={"ID":"9617e8e6-7fd5-47c6-9afb-e6160221ddea","Type":"ContainerStarted","Data":"9544ee25d84f14dc160b79653d2b46930d4697897ced6a941352a55cf1877f84"} Oct 05 09:50:40 crc kubenswrapper[4846]: I1005 09:50:40.489851 4846 generic.go:334] "Generic (PLEG): container finished" podID="9617e8e6-7fd5-47c6-9afb-e6160221ddea" containerID="9544ee25d84f14dc160b79653d2b46930d4697897ced6a941352a55cf1877f84" exitCode=0 Oct 05 09:50:40 crc kubenswrapper[4846]: I1005 09:50:40.489916 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gv5pf" event={"ID":"9617e8e6-7fd5-47c6-9afb-e6160221ddea","Type":"ContainerDied","Data":"9544ee25d84f14dc160b79653d2b46930d4697897ced6a941352a55cf1877f84"} Oct 05 09:50:41 crc kubenswrapper[4846]: I1005 09:50:41.503025 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gv5pf" event={"ID":"9617e8e6-7fd5-47c6-9afb-e6160221ddea","Type":"ContainerStarted","Data":"50d1f1fb3f134cad8831d188a34bd3dafb9eabb2a382d73d7688087ba240fefd"} Oct 05 09:50:41 crc kubenswrapper[4846]: I1005 09:50:41.531708 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-gv5pf" podStartSLOduration=2.069198419 podStartE2EDuration="4.5316927s" podCreationTimestamp="2025-10-05 09:50:37 +0000 UTC" firstStartedPulling="2025-10-05 09:50:38.458664393 +0000 UTC m=+10960.699517168" lastFinishedPulling="2025-10-05 09:50:40.921158684 +0000 UTC m=+10963.162011449" observedRunningTime="2025-10-05 09:50:41.524289674 +0000 UTC m=+10963.765142449" watchObservedRunningTime="2025-10-05 09:50:41.5316927 +0000 UTC m=+10963.772545465" Oct 05 09:50:45 crc kubenswrapper[4846]: I1005 09:50:45.028109 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb697lj68_57b21e86-380b-4403-a4f4-5196748e08e3/util/0.log" Oct 05 09:50:45 crc kubenswrapper[4846]: I1005 09:50:45.211192 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb697lj68_57b21e86-380b-4403-a4f4-5196748e08e3/util/0.log" Oct 05 09:50:45 crc kubenswrapper[4846]: I1005 09:50:45.221871 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb697lj68_57b21e86-380b-4403-a4f4-5196748e08e3/pull/0.log" Oct 05 09:50:45 crc kubenswrapper[4846]: I1005 09:50:45.268975 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb697lj68_57b21e86-380b-4403-a4f4-5196748e08e3/pull/0.log" Oct 05 09:50:45 crc kubenswrapper[4846]: I1005 09:50:45.427756 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb697lj68_57b21e86-380b-4403-a4f4-5196748e08e3/util/0.log" Oct 05 09:50:45 crc kubenswrapper[4846]: I1005 09:50:45.456756 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb697lj68_57b21e86-380b-4403-a4f4-5196748e08e3/pull/0.log" Oct 05 09:50:45 crc kubenswrapper[4846]: I1005 09:50:45.559634 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb697lj68_57b21e86-380b-4403-a4f4-5196748e08e3/extract/0.log" Oct 05 09:50:45 crc kubenswrapper[4846]: I1005 09:50:45.699652 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2xndcw_18f66388-66f4-4457-877f-d47dff48bb98/util/0.log" Oct 05 09:50:45 crc kubenswrapper[4846]: I1005 09:50:45.844690 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2xndcw_18f66388-66f4-4457-877f-d47dff48bb98/pull/0.log" Oct 05 09:50:45 crc kubenswrapper[4846]: I1005 09:50:45.844869 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2xndcw_18f66388-66f4-4457-877f-d47dff48bb98/pull/0.log" Oct 05 09:50:45 crc kubenswrapper[4846]: I1005 09:50:45.882257 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2xndcw_18f66388-66f4-4457-877f-d47dff48bb98/util/0.log" Oct 05 09:50:46 crc kubenswrapper[4846]: I1005 09:50:46.051923 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2xndcw_18f66388-66f4-4457-877f-d47dff48bb98/util/0.log" Oct 05 09:50:46 crc kubenswrapper[4846]: I1005 09:50:46.080729 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2xndcw_18f66388-66f4-4457-877f-d47dff48bb98/extract/0.log" Oct 05 09:50:46 crc kubenswrapper[4846]: I1005 09:50:46.110050 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2xndcw_18f66388-66f4-4457-877f-d47dff48bb98/pull/0.log" Oct 05 09:50:46 crc kubenswrapper[4846]: I1005 09:50:46.214407 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dpmkhb_a5383419-49af-4680-9793-d361430cd80a/util/0.log" Oct 05 09:50:46 crc kubenswrapper[4846]: I1005 09:50:46.450243 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dpmkhb_a5383419-49af-4680-9793-d361430cd80a/pull/0.log" Oct 05 09:50:46 crc kubenswrapper[4846]: I1005 09:50:46.485457 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dpmkhb_a5383419-49af-4680-9793-d361430cd80a/util/0.log" Oct 05 09:50:46 crc kubenswrapper[4846]: I1005 09:50:46.490616 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dpmkhb_a5383419-49af-4680-9793-d361430cd80a/pull/0.log" Oct 05 09:50:46 crc kubenswrapper[4846]: I1005 09:50:46.608751 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dpmkhb_a5383419-49af-4680-9793-d361430cd80a/util/0.log" Oct 05 09:50:46 crc kubenswrapper[4846]: I1005 09:50:46.610285 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dpmkhb_a5383419-49af-4680-9793-d361430cd80a/pull/0.log" Oct 05 09:50:46 crc kubenswrapper[4846]: I1005 09:50:46.699839 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2dpmkhb_a5383419-49af-4680-9793-d361430cd80a/extract/0.log" Oct 05 09:50:46 crc kubenswrapper[4846]: I1005 09:50:46.803459 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-7fskx_04f12f2c-eb1a-4b2a-a5d4-bb930a6e278f/extract-utilities/0.log" Oct 05 09:50:46 crc kubenswrapper[4846]: I1005 09:50:46.968537 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-7fskx_04f12f2c-eb1a-4b2a-a5d4-bb930a6e278f/extract-utilities/0.log" Oct 05 09:50:46 crc kubenswrapper[4846]: I1005 09:50:46.974716 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-7fskx_04f12f2c-eb1a-4b2a-a5d4-bb930a6e278f/extract-content/0.log" Oct 05 09:50:47 crc kubenswrapper[4846]: I1005 09:50:47.009391 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-7fskx_04f12f2c-eb1a-4b2a-a5d4-bb930a6e278f/extract-content/0.log" Oct 05 09:50:47 crc kubenswrapper[4846]: I1005 09:50:47.107474 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-7fskx_04f12f2c-eb1a-4b2a-a5d4-bb930a6e278f/extract-utilities/0.log" Oct 05 09:50:47 crc kubenswrapper[4846]: I1005 09:50:47.168653 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-7fskx_04f12f2c-eb1a-4b2a-a5d4-bb930a6e278f/extract-content/0.log" Oct 05 09:50:47 crc kubenswrapper[4846]: I1005 09:50:47.343760 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-2cz9z_78cd470d-a948-4332-a64f-6a0da13f295c/extract-utilities/0.log" Oct 05 09:50:47 crc kubenswrapper[4846]: I1005 09:50:47.521460 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-2cz9z_78cd470d-a948-4332-a64f-6a0da13f295c/extract-content/0.log" Oct 05 09:50:47 crc kubenswrapper[4846]: I1005 09:50:47.582240 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-gv5pf" Oct 05 09:50:47 crc kubenswrapper[4846]: I1005 09:50:47.582286 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-gv5pf" Oct 05 09:50:47 crc kubenswrapper[4846]: I1005 09:50:47.631622 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-2cz9z_78cd470d-a948-4332-a64f-6a0da13f295c/extract-content/0.log" Oct 05 09:50:47 crc kubenswrapper[4846]: I1005 09:50:47.631648 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-gv5pf" Oct 05 09:50:47 crc kubenswrapper[4846]: I1005 09:50:47.631785 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-2cz9z_78cd470d-a948-4332-a64f-6a0da13f295c/extract-utilities/0.log" Oct 05 09:50:47 crc kubenswrapper[4846]: I1005 09:50:47.859656 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-2cz9z_78cd470d-a948-4332-a64f-6a0da13f295c/extract-content/0.log" Oct 05 09:50:47 crc kubenswrapper[4846]: I1005 09:50:47.883797 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-2cz9z_78cd470d-a948-4332-a64f-6a0da13f295c/extract-utilities/0.log" Oct 05 09:50:48 crc kubenswrapper[4846]: I1005 09:50:48.222655 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c65gkl_efcc5448-8dd3-4e98-8729-042643de817d/util/0.log" Oct 05 09:50:48 crc kubenswrapper[4846]: I1005 09:50:48.488923 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c65gkl_efcc5448-8dd3-4e98-8729-042643de817d/pull/0.log" Oct 05 09:50:48 crc kubenswrapper[4846]: I1005 09:50:48.492496 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c65gkl_efcc5448-8dd3-4e98-8729-042643de817d/util/0.log" Oct 05 09:50:48 crc kubenswrapper[4846]: I1005 09:50:48.503317 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c65gkl_efcc5448-8dd3-4e98-8729-042643de817d/pull/0.log" Oct 05 09:50:48 crc kubenswrapper[4846]: I1005 09:50:48.610995 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-gv5pf" Oct 05 09:50:48 crc kubenswrapper[4846]: I1005 09:50:48.652719 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-gv5pf"] Oct 05 09:50:48 crc kubenswrapper[4846]: I1005 09:50:48.727744 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c65gkl_efcc5448-8dd3-4e98-8729-042643de817d/extract/0.log" Oct 05 09:50:48 crc kubenswrapper[4846]: I1005 09:50:48.728599 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c65gkl_efcc5448-8dd3-4e98-8729-042643de817d/pull/0.log" Oct 05 09:50:48 crc kubenswrapper[4846]: I1005 09:50:48.748024 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c65gkl_efcc5448-8dd3-4e98-8729-042643de817d/util/0.log" Oct 05 09:50:48 crc kubenswrapper[4846]: I1005 09:50:48.811670 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-7fskx_04f12f2c-eb1a-4b2a-a5d4-bb930a6e278f/registry-server/0.log" Oct 05 09:50:48 crc kubenswrapper[4846]: I1005 09:50:48.916557 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-4kl8r_b5257dc4-0faa-4e06-aeea-b25504581b7e/marketplace-operator/0.log" Oct 05 09:50:49 crc kubenswrapper[4846]: I1005 09:50:49.038697 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-gv5pf_9617e8e6-7fd5-47c6-9afb-e6160221ddea/extract-utilities/0.log" Oct 05 09:50:49 crc kubenswrapper[4846]: I1005 09:50:49.272899 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-gv5pf_9617e8e6-7fd5-47c6-9afb-e6160221ddea/extract-content/0.log" Oct 05 09:50:49 crc kubenswrapper[4846]: I1005 09:50:49.290017 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-gv5pf_9617e8e6-7fd5-47c6-9afb-e6160221ddea/extract-content/0.log" Oct 05 09:50:49 crc kubenswrapper[4846]: I1005 09:50:49.330407 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-gv5pf_9617e8e6-7fd5-47c6-9afb-e6160221ddea/extract-utilities/0.log" Oct 05 09:50:49 crc kubenswrapper[4846]: I1005 09:50:49.391989 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-2cz9z_78cd470d-a948-4332-a64f-6a0da13f295c/registry-server/0.log" Oct 05 09:50:49 crc kubenswrapper[4846]: I1005 09:50:49.482137 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-gv5pf_9617e8e6-7fd5-47c6-9afb-e6160221ddea/registry-server/0.log" Oct 05 09:50:49 crc kubenswrapper[4846]: I1005 09:50:49.485037 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-gv5pf_9617e8e6-7fd5-47c6-9afb-e6160221ddea/extract-content/0.log" Oct 05 09:50:49 crc kubenswrapper[4846]: I1005 09:50:49.485490 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-gv5pf_9617e8e6-7fd5-47c6-9afb-e6160221ddea/extract-utilities/0.log" Oct 05 09:50:49 crc kubenswrapper[4846]: I1005 09:50:49.570457 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-mkttk_01709201-a99b-4e92-adc8-868c6d272ac8/extract-utilities/0.log" Oct 05 09:50:49 crc kubenswrapper[4846]: I1005 09:50:49.749948 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-mkttk_01709201-a99b-4e92-adc8-868c6d272ac8/extract-content/0.log" Oct 05 09:50:49 crc kubenswrapper[4846]: I1005 09:50:49.751731 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-mkttk_01709201-a99b-4e92-adc8-868c6d272ac8/extract-content/0.log" Oct 05 09:50:49 crc kubenswrapper[4846]: I1005 09:50:49.758500 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-mkttk_01709201-a99b-4e92-adc8-868c6d272ac8/extract-utilities/0.log" Oct 05 09:50:49 crc kubenswrapper[4846]: I1005 09:50:49.957888 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-mkttk_01709201-a99b-4e92-adc8-868c6d272ac8/extract-utilities/0.log" Oct 05 09:50:50 crc kubenswrapper[4846]: I1005 09:50:50.051573 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-2mvqk_e6edda1b-9b24-48db-829d-9fd38c6faadb/extract-utilities/0.log" Oct 05 09:50:50 crc kubenswrapper[4846]: I1005 09:50:50.051573 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-mkttk_01709201-a99b-4e92-adc8-868c6d272ac8/extract-content/0.log" Oct 05 09:50:50 crc kubenswrapper[4846]: I1005 09:50:50.291299 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-2mvqk_e6edda1b-9b24-48db-829d-9fd38c6faadb/extract-content/0.log" Oct 05 09:50:50 crc kubenswrapper[4846]: I1005 09:50:50.327288 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-2mvqk_e6edda1b-9b24-48db-829d-9fd38c6faadb/extract-content/0.log" Oct 05 09:50:50 crc kubenswrapper[4846]: I1005 09:50:50.341027 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-2mvqk_e6edda1b-9b24-48db-829d-9fd38c6faadb/extract-utilities/0.log" Oct 05 09:50:50 crc kubenswrapper[4846]: I1005 09:50:50.373278 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-mkttk_01709201-a99b-4e92-adc8-868c6d272ac8/registry-server/0.log" Oct 05 09:50:50 crc kubenswrapper[4846]: I1005 09:50:50.503226 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-2mvqk_e6edda1b-9b24-48db-829d-9fd38c6faadb/extract-content/0.log" Oct 05 09:50:50 crc kubenswrapper[4846]: I1005 09:50:50.536506 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-2mvqk_e6edda1b-9b24-48db-829d-9fd38c6faadb/extract-utilities/0.log" Oct 05 09:50:50 crc kubenswrapper[4846]: I1005 09:50:50.581878 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-gv5pf" podUID="9617e8e6-7fd5-47c6-9afb-e6160221ddea" containerName="registry-server" containerID="cri-o://50d1f1fb3f134cad8831d188a34bd3dafb9eabb2a382d73d7688087ba240fefd" gracePeriod=2 Oct 05 09:50:51 crc kubenswrapper[4846]: I1005 09:50:51.100326 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-gv5pf" Oct 05 09:50:51 crc kubenswrapper[4846]: I1005 09:50:51.213124 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v9hrs\" (UniqueName: \"kubernetes.io/projected/9617e8e6-7fd5-47c6-9afb-e6160221ddea-kube-api-access-v9hrs\") pod \"9617e8e6-7fd5-47c6-9afb-e6160221ddea\" (UID: \"9617e8e6-7fd5-47c6-9afb-e6160221ddea\") " Oct 05 09:50:51 crc kubenswrapper[4846]: I1005 09:50:51.213268 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9617e8e6-7fd5-47c6-9afb-e6160221ddea-catalog-content\") pod \"9617e8e6-7fd5-47c6-9afb-e6160221ddea\" (UID: \"9617e8e6-7fd5-47c6-9afb-e6160221ddea\") " Oct 05 09:50:51 crc kubenswrapper[4846]: I1005 09:50:51.213369 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9617e8e6-7fd5-47c6-9afb-e6160221ddea-utilities\") pod \"9617e8e6-7fd5-47c6-9afb-e6160221ddea\" (UID: \"9617e8e6-7fd5-47c6-9afb-e6160221ddea\") " Oct 05 09:50:51 crc kubenswrapper[4846]: I1005 09:50:51.214650 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9617e8e6-7fd5-47c6-9afb-e6160221ddea-utilities" (OuterVolumeSpecName: "utilities") pod "9617e8e6-7fd5-47c6-9afb-e6160221ddea" (UID: "9617e8e6-7fd5-47c6-9afb-e6160221ddea"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 09:50:51 crc kubenswrapper[4846]: I1005 09:50:51.219667 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9617e8e6-7fd5-47c6-9afb-e6160221ddea-kube-api-access-v9hrs" (OuterVolumeSpecName: "kube-api-access-v9hrs") pod "9617e8e6-7fd5-47c6-9afb-e6160221ddea" (UID: "9617e8e6-7fd5-47c6-9afb-e6160221ddea"). InnerVolumeSpecName "kube-api-access-v9hrs". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 09:50:51 crc kubenswrapper[4846]: I1005 09:50:51.237665 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9617e8e6-7fd5-47c6-9afb-e6160221ddea-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9617e8e6-7fd5-47c6-9afb-e6160221ddea" (UID: "9617e8e6-7fd5-47c6-9afb-e6160221ddea"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 09:50:51 crc kubenswrapper[4846]: I1005 09:50:51.315367 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v9hrs\" (UniqueName: \"kubernetes.io/projected/9617e8e6-7fd5-47c6-9afb-e6160221ddea-kube-api-access-v9hrs\") on node \"crc\" DevicePath \"\"" Oct 05 09:50:51 crc kubenswrapper[4846]: I1005 09:50:51.315402 4846 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9617e8e6-7fd5-47c6-9afb-e6160221ddea-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 05 09:50:51 crc kubenswrapper[4846]: I1005 09:50:51.315415 4846 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9617e8e6-7fd5-47c6-9afb-e6160221ddea-utilities\") on node \"crc\" DevicePath \"\"" Oct 05 09:50:51 crc kubenswrapper[4846]: I1005 09:50:51.591731 4846 generic.go:334] "Generic (PLEG): container finished" podID="9617e8e6-7fd5-47c6-9afb-e6160221ddea" containerID="50d1f1fb3f134cad8831d188a34bd3dafb9eabb2a382d73d7688087ba240fefd" exitCode=0 Oct 05 09:50:51 crc kubenswrapper[4846]: I1005 09:50:51.591774 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gv5pf" event={"ID":"9617e8e6-7fd5-47c6-9afb-e6160221ddea","Type":"ContainerDied","Data":"50d1f1fb3f134cad8831d188a34bd3dafb9eabb2a382d73d7688087ba240fefd"} Oct 05 09:50:51 crc kubenswrapper[4846]: I1005 09:50:51.591813 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-gv5pf" Oct 05 09:50:51 crc kubenswrapper[4846]: I1005 09:50:51.592140 4846 scope.go:117] "RemoveContainer" containerID="50d1f1fb3f134cad8831d188a34bd3dafb9eabb2a382d73d7688087ba240fefd" Oct 05 09:50:51 crc kubenswrapper[4846]: I1005 09:50:51.592058 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gv5pf" event={"ID":"9617e8e6-7fd5-47c6-9afb-e6160221ddea","Type":"ContainerDied","Data":"bd65ee98cf7691ba9ea9a4cc384a3152aef3efb704c3daeb6a28d0ded0e97d8e"} Oct 05 09:50:51 crc kubenswrapper[4846]: I1005 09:50:51.634372 4846 scope.go:117] "RemoveContainer" containerID="9544ee25d84f14dc160b79653d2b46930d4697897ced6a941352a55cf1877f84" Oct 05 09:50:51 crc kubenswrapper[4846]: I1005 09:50:51.634422 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-gv5pf"] Oct 05 09:50:51 crc kubenswrapper[4846]: I1005 09:50:51.651011 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-gv5pf"] Oct 05 09:50:51 crc kubenswrapper[4846]: I1005 09:50:51.655840 4846 scope.go:117] "RemoveContainer" containerID="72f64ea7a880c126e7c51eb6aaa61a8bb5bbd42707f5fbd95a39f3cefa9f48f5" Oct 05 09:50:51 crc kubenswrapper[4846]: I1005 09:50:51.702429 4846 scope.go:117] "RemoveContainer" containerID="50d1f1fb3f134cad8831d188a34bd3dafb9eabb2a382d73d7688087ba240fefd" Oct 05 09:50:51 crc kubenswrapper[4846]: E1005 09:50:51.703349 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"50d1f1fb3f134cad8831d188a34bd3dafb9eabb2a382d73d7688087ba240fefd\": container with ID starting with 50d1f1fb3f134cad8831d188a34bd3dafb9eabb2a382d73d7688087ba240fefd not found: ID does not exist" containerID="50d1f1fb3f134cad8831d188a34bd3dafb9eabb2a382d73d7688087ba240fefd" Oct 05 09:50:51 crc kubenswrapper[4846]: I1005 09:50:51.703408 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"50d1f1fb3f134cad8831d188a34bd3dafb9eabb2a382d73d7688087ba240fefd"} err="failed to get container status \"50d1f1fb3f134cad8831d188a34bd3dafb9eabb2a382d73d7688087ba240fefd\": rpc error: code = NotFound desc = could not find container \"50d1f1fb3f134cad8831d188a34bd3dafb9eabb2a382d73d7688087ba240fefd\": container with ID starting with 50d1f1fb3f134cad8831d188a34bd3dafb9eabb2a382d73d7688087ba240fefd not found: ID does not exist" Oct 05 09:50:51 crc kubenswrapper[4846]: I1005 09:50:51.703482 4846 scope.go:117] "RemoveContainer" containerID="9544ee25d84f14dc160b79653d2b46930d4697897ced6a941352a55cf1877f84" Oct 05 09:50:51 crc kubenswrapper[4846]: E1005 09:50:51.703956 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9544ee25d84f14dc160b79653d2b46930d4697897ced6a941352a55cf1877f84\": container with ID starting with 9544ee25d84f14dc160b79653d2b46930d4697897ced6a941352a55cf1877f84 not found: ID does not exist" containerID="9544ee25d84f14dc160b79653d2b46930d4697897ced6a941352a55cf1877f84" Oct 05 09:50:51 crc kubenswrapper[4846]: I1005 09:50:51.704001 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9544ee25d84f14dc160b79653d2b46930d4697897ced6a941352a55cf1877f84"} err="failed to get container status \"9544ee25d84f14dc160b79653d2b46930d4697897ced6a941352a55cf1877f84\": rpc error: code = NotFound desc = could not find container \"9544ee25d84f14dc160b79653d2b46930d4697897ced6a941352a55cf1877f84\": container with ID starting with 9544ee25d84f14dc160b79653d2b46930d4697897ced6a941352a55cf1877f84 not found: ID does not exist" Oct 05 09:50:51 crc kubenswrapper[4846]: I1005 09:50:51.704029 4846 scope.go:117] "RemoveContainer" containerID="72f64ea7a880c126e7c51eb6aaa61a8bb5bbd42707f5fbd95a39f3cefa9f48f5" Oct 05 09:50:51 crc kubenswrapper[4846]: E1005 09:50:51.704594 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"72f64ea7a880c126e7c51eb6aaa61a8bb5bbd42707f5fbd95a39f3cefa9f48f5\": container with ID starting with 72f64ea7a880c126e7c51eb6aaa61a8bb5bbd42707f5fbd95a39f3cefa9f48f5 not found: ID does not exist" containerID="72f64ea7a880c126e7c51eb6aaa61a8bb5bbd42707f5fbd95a39f3cefa9f48f5" Oct 05 09:50:51 crc kubenswrapper[4846]: I1005 09:50:51.704640 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"72f64ea7a880c126e7c51eb6aaa61a8bb5bbd42707f5fbd95a39f3cefa9f48f5"} err="failed to get container status \"72f64ea7a880c126e7c51eb6aaa61a8bb5bbd42707f5fbd95a39f3cefa9f48f5\": rpc error: code = NotFound desc = could not find container \"72f64ea7a880c126e7c51eb6aaa61a8bb5bbd42707f5fbd95a39f3cefa9f48f5\": container with ID starting with 72f64ea7a880c126e7c51eb6aaa61a8bb5bbd42707f5fbd95a39f3cefa9f48f5 not found: ID does not exist" Oct 05 09:50:51 crc kubenswrapper[4846]: I1005 09:50:51.794578 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-2mvqk_e6edda1b-9b24-48db-829d-9fd38c6faadb/registry-server/0.log" Oct 05 09:50:52 crc kubenswrapper[4846]: I1005 09:50:52.508526 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9617e8e6-7fd5-47c6-9afb-e6160221ddea" path="/var/lib/kubelet/pods/9617e8e6-7fd5-47c6-9afb-e6160221ddea/volumes" Oct 05 09:50:53 crc kubenswrapper[4846]: I1005 09:50:53.324685 4846 patch_prober.go:28] interesting pod/machine-config-daemon-fscvf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 09:50:53 crc kubenswrapper[4846]: I1005 09:50:53.325361 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 09:51:04 crc kubenswrapper[4846]: I1005 09:51:04.728860 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-7c8cf85677-gjc7q_bb1e8f48-88f8-4623-9503-3589e855967a/prometheus-operator/0.log" Oct 05 09:51:04 crc kubenswrapper[4846]: I1005 09:51:04.901508 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-59bdf9d4c4-mrtq4_d7e575d9-c148-487e-ad29-10c42e593221/prometheus-operator-admission-webhook/0.log" Oct 05 09:51:04 crc kubenswrapper[4846]: I1005 09:51:04.941494 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-59bdf9d4c4-44zmx_3f591385-e1df-4dbd-a42d-a96f6a866cad/prometheus-operator-admission-webhook/0.log" Oct 05 09:51:05 crc kubenswrapper[4846]: I1005 09:51:05.086786 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_observability-operator-cc5f78dfc-95rb8_c96f8bbb-d1ef-4345-8ce7-8e72d543e47e/operator/0.log" Oct 05 09:51:05 crc kubenswrapper[4846]: I1005 09:51:05.197221 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_perses-operator-54bc95c9fb-llcsj_a17876c6-28e5-4ac8-9ebd-ed1be29bb964/perses-operator/0.log" Oct 05 09:51:13 crc kubenswrapper[4846]: E1005 09:51:13.725217 4846 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.111:43756->38.102.83.111:43661: write tcp 38.102.83.111:43756->38.102.83.111:43661: write: broken pipe Oct 05 09:51:23 crc kubenswrapper[4846]: I1005 09:51:23.324739 4846 patch_prober.go:28] interesting pod/machine-config-daemon-fscvf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 09:51:23 crc kubenswrapper[4846]: I1005 09:51:23.325261 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 09:51:23 crc kubenswrapper[4846]: I1005 09:51:23.325304 4846 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" Oct 05 09:51:23 crc kubenswrapper[4846]: I1005 09:51:23.326212 4846 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"71dd5b6fb2a9c6f83610e28480b9546168c48746147b1078114492dbe0d834b1"} pod="openshift-machine-config-operator/machine-config-daemon-fscvf" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 05 09:51:23 crc kubenswrapper[4846]: I1005 09:51:23.326284 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" containerName="machine-config-daemon" containerID="cri-o://71dd5b6fb2a9c6f83610e28480b9546168c48746147b1078114492dbe0d834b1" gracePeriod=600 Oct 05 09:51:23 crc kubenswrapper[4846]: I1005 09:51:23.948636 4846 generic.go:334] "Generic (PLEG): container finished" podID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" containerID="71dd5b6fb2a9c6f83610e28480b9546168c48746147b1078114492dbe0d834b1" exitCode=0 Oct 05 09:51:23 crc kubenswrapper[4846]: I1005 09:51:23.948724 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" event={"ID":"ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7","Type":"ContainerDied","Data":"71dd5b6fb2a9c6f83610e28480b9546168c48746147b1078114492dbe0d834b1"} Oct 05 09:51:23 crc kubenswrapper[4846]: I1005 09:51:23.948948 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" event={"ID":"ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7","Type":"ContainerStarted","Data":"44058d22fd2292f0f5655e0c5f43cdadbd639ab93a4c59ff686a1cf9a143858a"} Oct 05 09:51:23 crc kubenswrapper[4846]: I1005 09:51:23.948970 4846 scope.go:117] "RemoveContainer" containerID="43e1f18d9b5aea38dd1527fbb4fc557f1b37c59dfd2bdc2019c291047f4244b8" Oct 05 09:51:25 crc kubenswrapper[4846]: E1005 09:51:25.453417 4846 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.111:33146->38.102.83.111:43661: write tcp 38.102.83.111:33146->38.102.83.111:43661: write: broken pipe Oct 05 09:52:42 crc kubenswrapper[4846]: I1005 09:52:42.456616 4846 scope.go:117] "RemoveContainer" containerID="4424237257eb8339c7cee2720444adcb8d71efbc263d3fb8d250188251b93605" Oct 05 09:53:23 crc kubenswrapper[4846]: I1005 09:53:23.324620 4846 patch_prober.go:28] interesting pod/machine-config-daemon-fscvf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 09:53:23 crc kubenswrapper[4846]: I1005 09:53:23.325260 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 09:53:39 crc kubenswrapper[4846]: I1005 09:53:39.805957 4846 generic.go:334] "Generic (PLEG): container finished" podID="a2867fd9-a2a3-44f0-849f-8d7ce90e0051" containerID="72f62f82142890e94f0525a8a258b8963373444c09159f16c4037d0be1e62fae" exitCode=0 Oct 05 09:53:39 crc kubenswrapper[4846]: I1005 09:53:39.806019 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-755h7/must-gather-kl77b" event={"ID":"a2867fd9-a2a3-44f0-849f-8d7ce90e0051","Type":"ContainerDied","Data":"72f62f82142890e94f0525a8a258b8963373444c09159f16c4037d0be1e62fae"} Oct 05 09:53:39 crc kubenswrapper[4846]: I1005 09:53:39.808156 4846 scope.go:117] "RemoveContainer" containerID="72f62f82142890e94f0525a8a258b8963373444c09159f16c4037d0be1e62fae" Oct 05 09:53:40 crc kubenswrapper[4846]: I1005 09:53:40.764246 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-755h7_must-gather-kl77b_a2867fd9-a2a3-44f0-849f-8d7ce90e0051/gather/0.log" Oct 05 09:53:50 crc kubenswrapper[4846]: I1005 09:53:50.328666 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-755h7/must-gather-kl77b"] Oct 05 09:53:50 crc kubenswrapper[4846]: I1005 09:53:50.329592 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-755h7/must-gather-kl77b" podUID="a2867fd9-a2a3-44f0-849f-8d7ce90e0051" containerName="copy" containerID="cri-o://8070a727bfd7ab91e6005ea90b4578d7eeaaf47f8c870839f0c0dfe7d7e6986f" gracePeriod=2 Oct 05 09:53:50 crc kubenswrapper[4846]: I1005 09:53:50.339411 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-755h7/must-gather-kl77b"] Oct 05 09:53:50 crc kubenswrapper[4846]: I1005 09:53:50.800503 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-755h7_must-gather-kl77b_a2867fd9-a2a3-44f0-849f-8d7ce90e0051/copy/0.log" Oct 05 09:53:50 crc kubenswrapper[4846]: I1005 09:53:50.801664 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-755h7/must-gather-kl77b" Oct 05 09:53:50 crc kubenswrapper[4846]: I1005 09:53:50.949589 4846 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-755h7_must-gather-kl77b_a2867fd9-a2a3-44f0-849f-8d7ce90e0051/copy/0.log" Oct 05 09:53:50 crc kubenswrapper[4846]: I1005 09:53:50.950137 4846 generic.go:334] "Generic (PLEG): container finished" podID="a2867fd9-a2a3-44f0-849f-8d7ce90e0051" containerID="8070a727bfd7ab91e6005ea90b4578d7eeaaf47f8c870839f0c0dfe7d7e6986f" exitCode=143 Oct 05 09:53:50 crc kubenswrapper[4846]: I1005 09:53:50.950213 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-755h7/must-gather-kl77b" Oct 05 09:53:50 crc kubenswrapper[4846]: I1005 09:53:50.950234 4846 scope.go:117] "RemoveContainer" containerID="8070a727bfd7ab91e6005ea90b4578d7eeaaf47f8c870839f0c0dfe7d7e6986f" Oct 05 09:53:50 crc kubenswrapper[4846]: I1005 09:53:50.951413 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/a2867fd9-a2a3-44f0-849f-8d7ce90e0051-must-gather-output\") pod \"a2867fd9-a2a3-44f0-849f-8d7ce90e0051\" (UID: \"a2867fd9-a2a3-44f0-849f-8d7ce90e0051\") " Oct 05 09:53:50 crc kubenswrapper[4846]: I1005 09:53:50.951556 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dpzlm\" (UniqueName: \"kubernetes.io/projected/a2867fd9-a2a3-44f0-849f-8d7ce90e0051-kube-api-access-dpzlm\") pod \"a2867fd9-a2a3-44f0-849f-8d7ce90e0051\" (UID: \"a2867fd9-a2a3-44f0-849f-8d7ce90e0051\") " Oct 05 09:53:50 crc kubenswrapper[4846]: I1005 09:53:50.962880 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a2867fd9-a2a3-44f0-849f-8d7ce90e0051-kube-api-access-dpzlm" (OuterVolumeSpecName: "kube-api-access-dpzlm") pod "a2867fd9-a2a3-44f0-849f-8d7ce90e0051" (UID: "a2867fd9-a2a3-44f0-849f-8d7ce90e0051"). InnerVolumeSpecName "kube-api-access-dpzlm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 09:53:50 crc kubenswrapper[4846]: I1005 09:53:50.973222 4846 scope.go:117] "RemoveContainer" containerID="72f62f82142890e94f0525a8a258b8963373444c09159f16c4037d0be1e62fae" Oct 05 09:53:51 crc kubenswrapper[4846]: I1005 09:53:51.053851 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dpzlm\" (UniqueName: \"kubernetes.io/projected/a2867fd9-a2a3-44f0-849f-8d7ce90e0051-kube-api-access-dpzlm\") on node \"crc\" DevicePath \"\"" Oct 05 09:53:51 crc kubenswrapper[4846]: I1005 09:53:51.106118 4846 scope.go:117] "RemoveContainer" containerID="8070a727bfd7ab91e6005ea90b4578d7eeaaf47f8c870839f0c0dfe7d7e6986f" Oct 05 09:53:51 crc kubenswrapper[4846]: E1005 09:53:51.106592 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8070a727bfd7ab91e6005ea90b4578d7eeaaf47f8c870839f0c0dfe7d7e6986f\": container with ID starting with 8070a727bfd7ab91e6005ea90b4578d7eeaaf47f8c870839f0c0dfe7d7e6986f not found: ID does not exist" containerID="8070a727bfd7ab91e6005ea90b4578d7eeaaf47f8c870839f0c0dfe7d7e6986f" Oct 05 09:53:51 crc kubenswrapper[4846]: I1005 09:53:51.106632 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8070a727bfd7ab91e6005ea90b4578d7eeaaf47f8c870839f0c0dfe7d7e6986f"} err="failed to get container status \"8070a727bfd7ab91e6005ea90b4578d7eeaaf47f8c870839f0c0dfe7d7e6986f\": rpc error: code = NotFound desc = could not find container \"8070a727bfd7ab91e6005ea90b4578d7eeaaf47f8c870839f0c0dfe7d7e6986f\": container with ID starting with 8070a727bfd7ab91e6005ea90b4578d7eeaaf47f8c870839f0c0dfe7d7e6986f not found: ID does not exist" Oct 05 09:53:51 crc kubenswrapper[4846]: I1005 09:53:51.106657 4846 scope.go:117] "RemoveContainer" containerID="72f62f82142890e94f0525a8a258b8963373444c09159f16c4037d0be1e62fae" Oct 05 09:53:51 crc kubenswrapper[4846]: E1005 09:53:51.107028 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"72f62f82142890e94f0525a8a258b8963373444c09159f16c4037d0be1e62fae\": container with ID starting with 72f62f82142890e94f0525a8a258b8963373444c09159f16c4037d0be1e62fae not found: ID does not exist" containerID="72f62f82142890e94f0525a8a258b8963373444c09159f16c4037d0be1e62fae" Oct 05 09:53:51 crc kubenswrapper[4846]: I1005 09:53:51.107065 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"72f62f82142890e94f0525a8a258b8963373444c09159f16c4037d0be1e62fae"} err="failed to get container status \"72f62f82142890e94f0525a8a258b8963373444c09159f16c4037d0be1e62fae\": rpc error: code = NotFound desc = could not find container \"72f62f82142890e94f0525a8a258b8963373444c09159f16c4037d0be1e62fae\": container with ID starting with 72f62f82142890e94f0525a8a258b8963373444c09159f16c4037d0be1e62fae not found: ID does not exist" Oct 05 09:53:51 crc kubenswrapper[4846]: I1005 09:53:51.208083 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a2867fd9-a2a3-44f0-849f-8d7ce90e0051-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "a2867fd9-a2a3-44f0-849f-8d7ce90e0051" (UID: "a2867fd9-a2a3-44f0-849f-8d7ce90e0051"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 09:53:51 crc kubenswrapper[4846]: I1005 09:53:51.257442 4846 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/a2867fd9-a2a3-44f0-849f-8d7ce90e0051-must-gather-output\") on node \"crc\" DevicePath \"\"" Oct 05 09:53:52 crc kubenswrapper[4846]: I1005 09:53:52.515545 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a2867fd9-a2a3-44f0-849f-8d7ce90e0051" path="/var/lib/kubelet/pods/a2867fd9-a2a3-44f0-849f-8d7ce90e0051/volumes" Oct 05 09:53:53 crc kubenswrapper[4846]: I1005 09:53:53.324922 4846 patch_prober.go:28] interesting pod/machine-config-daemon-fscvf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 09:53:53 crc kubenswrapper[4846]: I1005 09:53:53.325207 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 09:54:23 crc kubenswrapper[4846]: I1005 09:54:23.324632 4846 patch_prober.go:28] interesting pod/machine-config-daemon-fscvf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 05 09:54:23 crc kubenswrapper[4846]: I1005 09:54:23.325306 4846 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 05 09:54:23 crc kubenswrapper[4846]: I1005 09:54:23.325372 4846 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" Oct 05 09:54:23 crc kubenswrapper[4846]: I1005 09:54:23.326437 4846 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"44058d22fd2292f0f5655e0c5f43cdadbd639ab93a4c59ff686a1cf9a143858a"} pod="openshift-machine-config-operator/machine-config-daemon-fscvf" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 05 09:54:23 crc kubenswrapper[4846]: I1005 09:54:23.326541 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" containerName="machine-config-daemon" containerID="cri-o://44058d22fd2292f0f5655e0c5f43cdadbd639ab93a4c59ff686a1cf9a143858a" gracePeriod=600 Oct 05 09:54:23 crc kubenswrapper[4846]: I1005 09:54:23.408237 4846 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-96s59"] Oct 05 09:54:23 crc kubenswrapper[4846]: E1005 09:54:23.408666 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9617e8e6-7fd5-47c6-9afb-e6160221ddea" containerName="registry-server" Oct 05 09:54:23 crc kubenswrapper[4846]: I1005 09:54:23.408683 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="9617e8e6-7fd5-47c6-9afb-e6160221ddea" containerName="registry-server" Oct 05 09:54:23 crc kubenswrapper[4846]: E1005 09:54:23.408706 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9617e8e6-7fd5-47c6-9afb-e6160221ddea" containerName="extract-content" Oct 05 09:54:23 crc kubenswrapper[4846]: I1005 09:54:23.408713 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="9617e8e6-7fd5-47c6-9afb-e6160221ddea" containerName="extract-content" Oct 05 09:54:23 crc kubenswrapper[4846]: E1005 09:54:23.408730 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a2867fd9-a2a3-44f0-849f-8d7ce90e0051" containerName="copy" Oct 05 09:54:23 crc kubenswrapper[4846]: I1005 09:54:23.408736 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="a2867fd9-a2a3-44f0-849f-8d7ce90e0051" containerName="copy" Oct 05 09:54:23 crc kubenswrapper[4846]: E1005 09:54:23.408748 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9617e8e6-7fd5-47c6-9afb-e6160221ddea" containerName="extract-utilities" Oct 05 09:54:23 crc kubenswrapper[4846]: I1005 09:54:23.408755 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="9617e8e6-7fd5-47c6-9afb-e6160221ddea" containerName="extract-utilities" Oct 05 09:54:23 crc kubenswrapper[4846]: E1005 09:54:23.408775 4846 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a2867fd9-a2a3-44f0-849f-8d7ce90e0051" containerName="gather" Oct 05 09:54:23 crc kubenswrapper[4846]: I1005 09:54:23.408781 4846 state_mem.go:107] "Deleted CPUSet assignment" podUID="a2867fd9-a2a3-44f0-849f-8d7ce90e0051" containerName="gather" Oct 05 09:54:23 crc kubenswrapper[4846]: I1005 09:54:23.408967 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="a2867fd9-a2a3-44f0-849f-8d7ce90e0051" containerName="gather" Oct 05 09:54:23 crc kubenswrapper[4846]: I1005 09:54:23.409011 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="9617e8e6-7fd5-47c6-9afb-e6160221ddea" containerName="registry-server" Oct 05 09:54:23 crc kubenswrapper[4846]: I1005 09:54:23.409026 4846 memory_manager.go:354] "RemoveStaleState removing state" podUID="a2867fd9-a2a3-44f0-849f-8d7ce90e0051" containerName="copy" Oct 05 09:54:23 crc kubenswrapper[4846]: I1005 09:54:23.410495 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-96s59" Oct 05 09:54:23 crc kubenswrapper[4846]: I1005 09:54:23.443073 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-96s59"] Oct 05 09:54:23 crc kubenswrapper[4846]: E1005 09:54:23.458783 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 09:54:23 crc kubenswrapper[4846]: I1005 09:54:23.555392 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8a6f84ee-81b8-4c94-97da-bfa381b7dded-utilities\") pod \"redhat-operators-96s59\" (UID: \"8a6f84ee-81b8-4c94-97da-bfa381b7dded\") " pod="openshift-marketplace/redhat-operators-96s59" Oct 05 09:54:23 crc kubenswrapper[4846]: I1005 09:54:23.555717 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8a6f84ee-81b8-4c94-97da-bfa381b7dded-catalog-content\") pod \"redhat-operators-96s59\" (UID: \"8a6f84ee-81b8-4c94-97da-bfa381b7dded\") " pod="openshift-marketplace/redhat-operators-96s59" Oct 05 09:54:23 crc kubenswrapper[4846]: I1005 09:54:23.555859 4846 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rz42q\" (UniqueName: \"kubernetes.io/projected/8a6f84ee-81b8-4c94-97da-bfa381b7dded-kube-api-access-rz42q\") pod \"redhat-operators-96s59\" (UID: \"8a6f84ee-81b8-4c94-97da-bfa381b7dded\") " pod="openshift-marketplace/redhat-operators-96s59" Oct 05 09:54:23 crc kubenswrapper[4846]: I1005 09:54:23.658949 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8a6f84ee-81b8-4c94-97da-bfa381b7dded-utilities\") pod \"redhat-operators-96s59\" (UID: \"8a6f84ee-81b8-4c94-97da-bfa381b7dded\") " pod="openshift-marketplace/redhat-operators-96s59" Oct 05 09:54:23 crc kubenswrapper[4846]: I1005 09:54:23.659017 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8a6f84ee-81b8-4c94-97da-bfa381b7dded-catalog-content\") pod \"redhat-operators-96s59\" (UID: \"8a6f84ee-81b8-4c94-97da-bfa381b7dded\") " pod="openshift-marketplace/redhat-operators-96s59" Oct 05 09:54:23 crc kubenswrapper[4846]: I1005 09:54:23.659445 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8a6f84ee-81b8-4c94-97da-bfa381b7dded-catalog-content\") pod \"redhat-operators-96s59\" (UID: \"8a6f84ee-81b8-4c94-97da-bfa381b7dded\") " pod="openshift-marketplace/redhat-operators-96s59" Oct 05 09:54:23 crc kubenswrapper[4846]: I1005 09:54:23.659639 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8a6f84ee-81b8-4c94-97da-bfa381b7dded-utilities\") pod \"redhat-operators-96s59\" (UID: \"8a6f84ee-81b8-4c94-97da-bfa381b7dded\") " pod="openshift-marketplace/redhat-operators-96s59" Oct 05 09:54:23 crc kubenswrapper[4846]: I1005 09:54:23.660307 4846 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rz42q\" (UniqueName: \"kubernetes.io/projected/8a6f84ee-81b8-4c94-97da-bfa381b7dded-kube-api-access-rz42q\") pod \"redhat-operators-96s59\" (UID: \"8a6f84ee-81b8-4c94-97da-bfa381b7dded\") " pod="openshift-marketplace/redhat-operators-96s59" Oct 05 09:54:23 crc kubenswrapper[4846]: I1005 09:54:23.682694 4846 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rz42q\" (UniqueName: \"kubernetes.io/projected/8a6f84ee-81b8-4c94-97da-bfa381b7dded-kube-api-access-rz42q\") pod \"redhat-operators-96s59\" (UID: \"8a6f84ee-81b8-4c94-97da-bfa381b7dded\") " pod="openshift-marketplace/redhat-operators-96s59" Oct 05 09:54:23 crc kubenswrapper[4846]: I1005 09:54:23.729283 4846 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-96s59" Oct 05 09:54:24 crc kubenswrapper[4846]: I1005 09:54:24.214800 4846 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-96s59"] Oct 05 09:54:24 crc kubenswrapper[4846]: I1005 09:54:24.337150 4846 generic.go:334] "Generic (PLEG): container finished" podID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" containerID="44058d22fd2292f0f5655e0c5f43cdadbd639ab93a4c59ff686a1cf9a143858a" exitCode=0 Oct 05 09:54:24 crc kubenswrapper[4846]: I1005 09:54:24.337396 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" event={"ID":"ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7","Type":"ContainerDied","Data":"44058d22fd2292f0f5655e0c5f43cdadbd639ab93a4c59ff686a1cf9a143858a"} Oct 05 09:54:24 crc kubenswrapper[4846]: I1005 09:54:24.338510 4846 scope.go:117] "RemoveContainer" containerID="71dd5b6fb2a9c6f83610e28480b9546168c48746147b1078114492dbe0d834b1" Oct 05 09:54:24 crc kubenswrapper[4846]: I1005 09:54:24.339214 4846 scope.go:117] "RemoveContainer" containerID="44058d22fd2292f0f5655e0c5f43cdadbd639ab93a4c59ff686a1cf9a143858a" Oct 05 09:54:24 crc kubenswrapper[4846]: E1005 09:54:24.339524 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 09:54:24 crc kubenswrapper[4846]: I1005 09:54:24.341015 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-96s59" event={"ID":"8a6f84ee-81b8-4c94-97da-bfa381b7dded","Type":"ContainerStarted","Data":"944c8ee850ebc5fcdbbb4d90faf899690976b6c2a5498adec559b0ad660b0e53"} Oct 05 09:54:25 crc kubenswrapper[4846]: I1005 09:54:25.380276 4846 generic.go:334] "Generic (PLEG): container finished" podID="8a6f84ee-81b8-4c94-97da-bfa381b7dded" containerID="76f893d8efdf7607dbde2bd2aa8349b42afc0ae8cc91462030ab7f82c791cc6f" exitCode=0 Oct 05 09:54:25 crc kubenswrapper[4846]: I1005 09:54:25.380380 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-96s59" event={"ID":"8a6f84ee-81b8-4c94-97da-bfa381b7dded","Type":"ContainerDied","Data":"76f893d8efdf7607dbde2bd2aa8349b42afc0ae8cc91462030ab7f82c791cc6f"} Oct 05 09:54:25 crc kubenswrapper[4846]: I1005 09:54:25.387222 4846 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 05 09:54:26 crc kubenswrapper[4846]: I1005 09:54:26.407436 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-96s59" event={"ID":"8a6f84ee-81b8-4c94-97da-bfa381b7dded","Type":"ContainerStarted","Data":"580d4e82838b37756000564838a7f5004450a50ac786d75e31a9c480372ce086"} Oct 05 09:54:29 crc kubenswrapper[4846]: I1005 09:54:29.451436 4846 generic.go:334] "Generic (PLEG): container finished" podID="8a6f84ee-81b8-4c94-97da-bfa381b7dded" containerID="580d4e82838b37756000564838a7f5004450a50ac786d75e31a9c480372ce086" exitCode=0 Oct 05 09:54:29 crc kubenswrapper[4846]: I1005 09:54:29.451520 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-96s59" event={"ID":"8a6f84ee-81b8-4c94-97da-bfa381b7dded","Type":"ContainerDied","Data":"580d4e82838b37756000564838a7f5004450a50ac786d75e31a9c480372ce086"} Oct 05 09:54:30 crc kubenswrapper[4846]: I1005 09:54:30.463606 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-96s59" event={"ID":"8a6f84ee-81b8-4c94-97da-bfa381b7dded","Type":"ContainerStarted","Data":"772b999c43ba755f873908ca69ae4d1377d7d5e01859187c7a53bd541beb0eb1"} Oct 05 09:54:30 crc kubenswrapper[4846]: I1005 09:54:30.495588 4846 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-96s59" podStartSLOduration=2.9552211330000002 podStartE2EDuration="7.495564456s" podCreationTimestamp="2025-10-05 09:54:23 +0000 UTC" firstStartedPulling="2025-10-05 09:54:25.386992337 +0000 UTC m=+11187.627845112" lastFinishedPulling="2025-10-05 09:54:29.92733566 +0000 UTC m=+11192.168188435" observedRunningTime="2025-10-05 09:54:30.487733409 +0000 UTC m=+11192.728586184" watchObservedRunningTime="2025-10-05 09:54:30.495564456 +0000 UTC m=+11192.736417231" Oct 05 09:54:33 crc kubenswrapper[4846]: I1005 09:54:33.730219 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-96s59" Oct 05 09:54:33 crc kubenswrapper[4846]: I1005 09:54:33.732023 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-96s59" Oct 05 09:54:34 crc kubenswrapper[4846]: I1005 09:54:34.778084 4846 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-96s59" podUID="8a6f84ee-81b8-4c94-97da-bfa381b7dded" containerName="registry-server" probeResult="failure" output=< Oct 05 09:54:34 crc kubenswrapper[4846]: timeout: failed to connect service ":50051" within 1s Oct 05 09:54:34 crc kubenswrapper[4846]: > Oct 05 09:54:35 crc kubenswrapper[4846]: I1005 09:54:35.498701 4846 scope.go:117] "RemoveContainer" containerID="44058d22fd2292f0f5655e0c5f43cdadbd639ab93a4c59ff686a1cf9a143858a" Oct 05 09:54:35 crc kubenswrapper[4846]: E1005 09:54:35.499313 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 09:54:43 crc kubenswrapper[4846]: I1005 09:54:43.817205 4846 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-96s59" Oct 05 09:54:43 crc kubenswrapper[4846]: I1005 09:54:43.908151 4846 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-96s59" Oct 05 09:54:44 crc kubenswrapper[4846]: I1005 09:54:44.069522 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-96s59"] Oct 05 09:54:45 crc kubenswrapper[4846]: I1005 09:54:45.672939 4846 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-96s59" podUID="8a6f84ee-81b8-4c94-97da-bfa381b7dded" containerName="registry-server" containerID="cri-o://772b999c43ba755f873908ca69ae4d1377d7d5e01859187c7a53bd541beb0eb1" gracePeriod=2 Oct 05 09:54:46 crc kubenswrapper[4846]: I1005 09:54:46.240777 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-96s59" Oct 05 09:54:46 crc kubenswrapper[4846]: I1005 09:54:46.297411 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8a6f84ee-81b8-4c94-97da-bfa381b7dded-utilities\") pod \"8a6f84ee-81b8-4c94-97da-bfa381b7dded\" (UID: \"8a6f84ee-81b8-4c94-97da-bfa381b7dded\") " Oct 05 09:54:46 crc kubenswrapper[4846]: I1005 09:54:46.297521 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8a6f84ee-81b8-4c94-97da-bfa381b7dded-catalog-content\") pod \"8a6f84ee-81b8-4c94-97da-bfa381b7dded\" (UID: \"8a6f84ee-81b8-4c94-97da-bfa381b7dded\") " Oct 05 09:54:46 crc kubenswrapper[4846]: I1005 09:54:46.297622 4846 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rz42q\" (UniqueName: \"kubernetes.io/projected/8a6f84ee-81b8-4c94-97da-bfa381b7dded-kube-api-access-rz42q\") pod \"8a6f84ee-81b8-4c94-97da-bfa381b7dded\" (UID: \"8a6f84ee-81b8-4c94-97da-bfa381b7dded\") " Oct 05 09:54:46 crc kubenswrapper[4846]: I1005 09:54:46.298639 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8a6f84ee-81b8-4c94-97da-bfa381b7dded-utilities" (OuterVolumeSpecName: "utilities") pod "8a6f84ee-81b8-4c94-97da-bfa381b7dded" (UID: "8a6f84ee-81b8-4c94-97da-bfa381b7dded"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 09:54:46 crc kubenswrapper[4846]: I1005 09:54:46.311690 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8a6f84ee-81b8-4c94-97da-bfa381b7dded-kube-api-access-rz42q" (OuterVolumeSpecName: "kube-api-access-rz42q") pod "8a6f84ee-81b8-4c94-97da-bfa381b7dded" (UID: "8a6f84ee-81b8-4c94-97da-bfa381b7dded"). InnerVolumeSpecName "kube-api-access-rz42q". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 05 09:54:46 crc kubenswrapper[4846]: I1005 09:54:46.401335 4846 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8a6f84ee-81b8-4c94-97da-bfa381b7dded-utilities\") on node \"crc\" DevicePath \"\"" Oct 05 09:54:46 crc kubenswrapper[4846]: I1005 09:54:46.401377 4846 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rz42q\" (UniqueName: \"kubernetes.io/projected/8a6f84ee-81b8-4c94-97da-bfa381b7dded-kube-api-access-rz42q\") on node \"crc\" DevicePath \"\"" Oct 05 09:54:46 crc kubenswrapper[4846]: I1005 09:54:46.408536 4846 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8a6f84ee-81b8-4c94-97da-bfa381b7dded-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8a6f84ee-81b8-4c94-97da-bfa381b7dded" (UID: "8a6f84ee-81b8-4c94-97da-bfa381b7dded"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 05 09:54:46 crc kubenswrapper[4846]: I1005 09:54:46.502733 4846 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8a6f84ee-81b8-4c94-97da-bfa381b7dded-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 05 09:54:46 crc kubenswrapper[4846]: I1005 09:54:46.694613 4846 generic.go:334] "Generic (PLEG): container finished" podID="8a6f84ee-81b8-4c94-97da-bfa381b7dded" containerID="772b999c43ba755f873908ca69ae4d1377d7d5e01859187c7a53bd541beb0eb1" exitCode=0 Oct 05 09:54:46 crc kubenswrapper[4846]: I1005 09:54:46.694690 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-96s59" event={"ID":"8a6f84ee-81b8-4c94-97da-bfa381b7dded","Type":"ContainerDied","Data":"772b999c43ba755f873908ca69ae4d1377d7d5e01859187c7a53bd541beb0eb1"} Oct 05 09:54:46 crc kubenswrapper[4846]: I1005 09:54:46.694706 4846 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-96s59" Oct 05 09:54:46 crc kubenswrapper[4846]: I1005 09:54:46.694743 4846 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-96s59" event={"ID":"8a6f84ee-81b8-4c94-97da-bfa381b7dded","Type":"ContainerDied","Data":"944c8ee850ebc5fcdbbb4d90faf899690976b6c2a5498adec559b0ad660b0e53"} Oct 05 09:54:46 crc kubenswrapper[4846]: I1005 09:54:46.694784 4846 scope.go:117] "RemoveContainer" containerID="772b999c43ba755f873908ca69ae4d1377d7d5e01859187c7a53bd541beb0eb1" Oct 05 09:54:46 crc kubenswrapper[4846]: I1005 09:54:46.729056 4846 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-96s59"] Oct 05 09:54:46 crc kubenswrapper[4846]: I1005 09:54:46.738709 4846 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-96s59"] Oct 05 09:54:46 crc kubenswrapper[4846]: I1005 09:54:46.739851 4846 scope.go:117] "RemoveContainer" containerID="580d4e82838b37756000564838a7f5004450a50ac786d75e31a9c480372ce086" Oct 05 09:54:46 crc kubenswrapper[4846]: I1005 09:54:46.765101 4846 scope.go:117] "RemoveContainer" containerID="76f893d8efdf7607dbde2bd2aa8349b42afc0ae8cc91462030ab7f82c791cc6f" Oct 05 09:54:46 crc kubenswrapper[4846]: I1005 09:54:46.822870 4846 scope.go:117] "RemoveContainer" containerID="772b999c43ba755f873908ca69ae4d1377d7d5e01859187c7a53bd541beb0eb1" Oct 05 09:54:46 crc kubenswrapper[4846]: E1005 09:54:46.823515 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"772b999c43ba755f873908ca69ae4d1377d7d5e01859187c7a53bd541beb0eb1\": container with ID starting with 772b999c43ba755f873908ca69ae4d1377d7d5e01859187c7a53bd541beb0eb1 not found: ID does not exist" containerID="772b999c43ba755f873908ca69ae4d1377d7d5e01859187c7a53bd541beb0eb1" Oct 05 09:54:46 crc kubenswrapper[4846]: I1005 09:54:46.823570 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"772b999c43ba755f873908ca69ae4d1377d7d5e01859187c7a53bd541beb0eb1"} err="failed to get container status \"772b999c43ba755f873908ca69ae4d1377d7d5e01859187c7a53bd541beb0eb1\": rpc error: code = NotFound desc = could not find container \"772b999c43ba755f873908ca69ae4d1377d7d5e01859187c7a53bd541beb0eb1\": container with ID starting with 772b999c43ba755f873908ca69ae4d1377d7d5e01859187c7a53bd541beb0eb1 not found: ID does not exist" Oct 05 09:54:46 crc kubenswrapper[4846]: I1005 09:54:46.823600 4846 scope.go:117] "RemoveContainer" containerID="580d4e82838b37756000564838a7f5004450a50ac786d75e31a9c480372ce086" Oct 05 09:54:46 crc kubenswrapper[4846]: E1005 09:54:46.824101 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"580d4e82838b37756000564838a7f5004450a50ac786d75e31a9c480372ce086\": container with ID starting with 580d4e82838b37756000564838a7f5004450a50ac786d75e31a9c480372ce086 not found: ID does not exist" containerID="580d4e82838b37756000564838a7f5004450a50ac786d75e31a9c480372ce086" Oct 05 09:54:46 crc kubenswrapper[4846]: I1005 09:54:46.824137 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"580d4e82838b37756000564838a7f5004450a50ac786d75e31a9c480372ce086"} err="failed to get container status \"580d4e82838b37756000564838a7f5004450a50ac786d75e31a9c480372ce086\": rpc error: code = NotFound desc = could not find container \"580d4e82838b37756000564838a7f5004450a50ac786d75e31a9c480372ce086\": container with ID starting with 580d4e82838b37756000564838a7f5004450a50ac786d75e31a9c480372ce086 not found: ID does not exist" Oct 05 09:54:46 crc kubenswrapper[4846]: I1005 09:54:46.824162 4846 scope.go:117] "RemoveContainer" containerID="76f893d8efdf7607dbde2bd2aa8349b42afc0ae8cc91462030ab7f82c791cc6f" Oct 05 09:54:46 crc kubenswrapper[4846]: E1005 09:54:46.824556 4846 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"76f893d8efdf7607dbde2bd2aa8349b42afc0ae8cc91462030ab7f82c791cc6f\": container with ID starting with 76f893d8efdf7607dbde2bd2aa8349b42afc0ae8cc91462030ab7f82c791cc6f not found: ID does not exist" containerID="76f893d8efdf7607dbde2bd2aa8349b42afc0ae8cc91462030ab7f82c791cc6f" Oct 05 09:54:46 crc kubenswrapper[4846]: I1005 09:54:46.824612 4846 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"76f893d8efdf7607dbde2bd2aa8349b42afc0ae8cc91462030ab7f82c791cc6f"} err="failed to get container status \"76f893d8efdf7607dbde2bd2aa8349b42afc0ae8cc91462030ab7f82c791cc6f\": rpc error: code = NotFound desc = could not find container \"76f893d8efdf7607dbde2bd2aa8349b42afc0ae8cc91462030ab7f82c791cc6f\": container with ID starting with 76f893d8efdf7607dbde2bd2aa8349b42afc0ae8cc91462030ab7f82c791cc6f not found: ID does not exist" Oct 05 09:54:47 crc kubenswrapper[4846]: I1005 09:54:47.499347 4846 scope.go:117] "RemoveContainer" containerID="44058d22fd2292f0f5655e0c5f43cdadbd639ab93a4c59ff686a1cf9a143858a" Oct 05 09:54:47 crc kubenswrapper[4846]: E1005 09:54:47.500561 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 09:54:48 crc kubenswrapper[4846]: I1005 09:54:48.513828 4846 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8a6f84ee-81b8-4c94-97da-bfa381b7dded" path="/var/lib/kubelet/pods/8a6f84ee-81b8-4c94-97da-bfa381b7dded/volumes" Oct 05 09:55:02 crc kubenswrapper[4846]: I1005 09:55:02.497938 4846 scope.go:117] "RemoveContainer" containerID="44058d22fd2292f0f5655e0c5f43cdadbd639ab93a4c59ff686a1cf9a143858a" Oct 05 09:55:02 crc kubenswrapper[4846]: E1005 09:55:02.498973 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" Oct 05 09:55:13 crc kubenswrapper[4846]: I1005 09:55:13.498283 4846 scope.go:117] "RemoveContainer" containerID="44058d22fd2292f0f5655e0c5f43cdadbd639ab93a4c59ff686a1cf9a143858a" Oct 05 09:55:13 crc kubenswrapper[4846]: E1005 09:55:13.499170 4846 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-fscvf_openshift-machine-config-operator(ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7)\"" pod="openshift-machine-config-operator/machine-config-daemon-fscvf" podUID="ae2aab07-74cf-4d7f-8d67-bc9aaa0fcab7" var/home/core/zuul-output/logs/crc-cloud-workdir-crc-all-logs.tar.gz0000644000175000000000000000005515070440216024444 0ustar coreroot‹íÁ  ÷Om7 €7šÞ'(var/home/core/zuul-output/logs/crc-cloud/0000755000175000000000000000000015070440216017361 5ustar corerootvar/home/core/zuul-output/artifacts/0000755000175000017500000000000015070411706016506 5ustar corecorevar/home/core/zuul-output/docs/0000755000175000017500000000000015070411706015456 5ustar corecore